2 * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved.
3 * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved.
5 * This copyrighted material is made available to anyone wishing to use,
6 * modify, copy, or redistribute it subject to the terms and conditions
7 * of the GNU General Public License v.2.
10 #include <linux/sched.h>
11 #include <linux/slab.h>
12 #include <linux/spinlock.h>
13 #include <linux/completion.h>
14 #include <linux/buffer_head.h>
15 #include <linux/pagemap.h>
16 #include <linux/uio.h>
17 #include <linux/blkdev.h>
19 #include <linux/smp_lock.h>
21 #include <linux/gfs2_ondisk.h>
22 #include <linux/ext2_fs.h>
23 #include <linux/crc32.h>
24 #include <linux/iflags.h>
25 #include <asm/uaccess.h>
28 #include "lm_interface.h"
46 /* "bad" is for NFS support */
47 struct filldir_bad_entry
{
49 unsigned int fbe_length
;
51 struct gfs2_inum fbe_inum
;
52 unsigned int fbe_type
;
56 struct gfs2_sbd
*fdb_sbd
;
58 struct filldir_bad_entry
*fdb_entry
;
59 unsigned int fdb_entry_num
;
60 unsigned int fdb_entry_off
;
63 unsigned int fdb_name_size
;
64 unsigned int fdb_name_off
;
67 /* For regular, non-NFS */
69 struct gfs2_sbd
*fdr_sbd
;
72 filldir_t fdr_filldir
;
77 * Most fields left uninitialised to catch anybody who tries to
78 * use them. f_flags set to prevent file_accessed() from touching
79 * any other part of this. Its use is purely as a flag so that we
80 * know (in readpage()) whether or not do to locking.
82 struct file gfs2_internal_file_sentinal
= {
83 .f_flags
= O_NOATIME
|O_RDONLY
,
86 static int gfs2_read_actor(read_descriptor_t
*desc
, struct page
*page
,
87 unsigned long offset
, unsigned long size
)
90 unsigned long count
= desc
->count
;
96 memcpy(desc
->arg
.buf
, kaddr
+ offset
, size
);
99 desc
->count
= count
- size
;
100 desc
->written
+= size
;
101 desc
->arg
.buf
+= size
;
105 int gfs2_internal_read(struct gfs2_inode
*ip
, struct file_ra_state
*ra_state
,
106 char *buf
, loff_t
*pos
, unsigned size
)
108 struct inode
*inode
= &ip
->i_inode
;
109 read_descriptor_t desc
;
114 do_generic_mapping_read(inode
->i_mapping
, ra_state
,
115 &gfs2_internal_file_sentinal
, pos
, &desc
,
117 return desc
.written
? desc
.written
: desc
.error
;
121 * gfs2_llseek - seek to a location in a file
123 * @offset: the offset
124 * @origin: Where to seek from (SEEK_SET, SEEK_CUR, or SEEK_END)
126 * SEEK_END requires the glock for the file because it references the
129 * Returns: The new offset, or errno
132 static loff_t
gfs2_llseek(struct file
*file
, loff_t offset
, int origin
)
134 struct gfs2_inode
*ip
= GFS2_I(file
->f_mapping
->host
);
135 struct gfs2_holder i_gh
;
139 error
= gfs2_glock_nq_init(ip
->i_gl
, LM_ST_SHARED
, LM_FLAG_ANY
,
142 error
= remote_llseek(file
, offset
, origin
);
143 gfs2_glock_dq_uninit(&i_gh
);
146 error
= remote_llseek(file
, offset
, origin
);
152 static ssize_t
gfs2_direct_IO_read(struct kiocb
*iocb
, const struct iovec
*iov
,
153 loff_t offset
, unsigned long nr_segs
)
155 struct file
*file
= iocb
->ki_filp
;
156 struct address_space
*mapping
= file
->f_mapping
;
159 retval
= filemap_write_and_wait(mapping
);
161 retval
= mapping
->a_ops
->direct_IO(READ
, iocb
, iov
, offset
,
168 * __gfs2_file_aio_read - The main GFS2 read function
170 * N.B. This is almost, but not quite the same as __generic_file_aio_read()
171 * the important subtle different being that inode->i_size isn't valid
172 * unless we are holding a lock, and we do this _only_ on the O_DIRECT
173 * path since otherwise locking is done entirely at the page cache
176 static ssize_t
__gfs2_file_aio_read(struct kiocb
*iocb
,
177 const struct iovec
*iov
,
178 unsigned long nr_segs
, loff_t
*ppos
)
180 struct file
*filp
= iocb
->ki_filp
;
181 struct gfs2_inode
*ip
= GFS2_I(filp
->f_mapping
->host
);
182 struct gfs2_holder gh
;
188 for (seg
= 0; seg
< nr_segs
; seg
++) {
189 const struct iovec
*iv
= &iov
[seg
];
192 * If any segment has a negative length, or the cumulative
193 * length ever wraps negative then return -EINVAL.
195 count
+= iv
->iov_len
;
196 if (unlikely((ssize_t
)(count
|iv
->iov_len
) < 0))
198 if (access_ok(VERIFY_WRITE
, iv
->iov_base
, iv
->iov_len
))
203 count
-= iv
->iov_len
; /* This segment is no good */
207 /* coalesce the iovecs and go direct-to-BIO for O_DIRECT */
208 if (filp
->f_flags
& O_DIRECT
) {
209 loff_t pos
= *ppos
, size
;
210 struct address_space
*mapping
;
213 mapping
= filp
->f_mapping
;
214 inode
= mapping
->host
;
217 goto out
; /* skip atime */
219 gfs2_holder_init(ip
->i_gl
, LM_ST_SHARED
, GL_ATIME
, &gh
);
220 retval
= gfs2_glock_nq_m_atime(1, &gh
);
223 if (gfs2_is_stuffed(ip
)) {
224 gfs2_glock_dq_m(1, &gh
);
225 gfs2_holder_uninit(&gh
);
226 goto fallback_to_normal
;
228 size
= i_size_read(inode
);
230 retval
= gfs2_direct_IO_read(iocb
, iov
, pos
, nr_segs
);
231 if (retval
> 0 && !is_sync_kiocb(iocb
))
232 retval
= -EIOCBQUEUED
;
234 *ppos
= pos
+ retval
;
237 gfs2_glock_dq_m(1, &gh
);
238 gfs2_holder_uninit(&gh
);
245 for (seg
= 0; seg
< nr_segs
; seg
++) {
246 read_descriptor_t desc
;
249 desc
.arg
.buf
= iov
[seg
].iov_base
;
250 desc
.count
= iov
[seg
].iov_len
;
254 do_generic_file_read(filp
,ppos
,&desc
,file_read_actor
);
255 retval
+= desc
.written
;
257 retval
= retval
?: desc
.error
;
267 * gfs2_read - Read bytes from a file
268 * @file: The file to read from
269 * @buf: The buffer to copy into
270 * @size: The amount of data requested
271 * @offset: The current file offset
273 * Outputs: Offset - updated according to number of bytes read
275 * Returns: The number of bytes read, errno on failure
278 static ssize_t
gfs2_read(struct file
*filp
, char __user
*buf
, size_t size
,
281 struct iovec local_iov
= { .iov_base
= buf
, .iov_len
= size
};
285 init_sync_kiocb(&kiocb
, filp
);
286 ret
= __gfs2_file_aio_read(&kiocb
, &local_iov
, 1, offset
);
287 if (-EIOCBQUEUED
== ret
)
288 ret
= wait_on_sync_kiocb(&kiocb
);
292 static ssize_t
gfs2_file_readv(struct file
*filp
, const struct iovec
*iov
,
293 unsigned long nr_segs
, loff_t
*ppos
)
298 init_sync_kiocb(&kiocb
, filp
);
299 ret
= __gfs2_file_aio_read(&kiocb
, iov
, nr_segs
, ppos
);
300 if (-EIOCBQUEUED
== ret
)
301 ret
= wait_on_sync_kiocb(&kiocb
);
305 static ssize_t
gfs2_file_aio_read(struct kiocb
*iocb
, char __user
*buf
,
306 size_t count
, loff_t pos
)
308 struct iovec local_iov
= { .iov_base
= buf
, .iov_len
= count
};
310 BUG_ON(iocb
->ki_pos
!= pos
);
311 return __gfs2_file_aio_read(iocb
, &local_iov
, 1, &iocb
->ki_pos
);
316 * filldir_reg_func - Report a directory entry to the caller of gfs2_dir_read()
317 * @opaque: opaque data used by the function
318 * @name: the name of the directory entry
319 * @length: the length of the name
320 * @offset: the entry's offset in the directory
321 * @inum: the inode number the entry points to
322 * @type: the type of inode the entry points to
324 * Returns: 0 on success, 1 if buffer full
327 static int filldir_reg_func(void *opaque
, const char *name
, unsigned int length
,
328 uint64_t offset
, struct gfs2_inum
*inum
,
331 struct filldir_reg
*fdr
= (struct filldir_reg
*)opaque
;
332 struct gfs2_sbd
*sdp
= fdr
->fdr_sbd
;
335 error
= fdr
->fdr_filldir(fdr
->fdr_opaque
, name
, length
, offset
,
336 inum
->no_formal_ino
, type
);
340 if (fdr
->fdr_prefetch
&& !(length
== 1 && *name
== '.')) {
341 gfs2_glock_prefetch_num(sdp
,
342 inum
->no_addr
, &gfs2_inode_glops
,
343 LM_ST_SHARED
, LM_FLAG_TRY
| LM_FLAG_ANY
);
344 gfs2_glock_prefetch_num(sdp
,
345 inum
->no_addr
, &gfs2_iopen_glops
,
346 LM_ST_SHARED
, LM_FLAG_TRY
);
353 * readdir_reg - Read directory entries from a directory
354 * @file: The directory to read from
355 * @dirent: Buffer for dirents
356 * @filldir: Function used to do the copying
361 static int readdir_reg(struct file
*file
, void *dirent
, filldir_t filldir
)
363 struct inode
*dir
= file
->f_mapping
->host
;
364 struct gfs2_inode
*dip
= GFS2_I(dir
);
365 struct filldir_reg fdr
;
366 struct gfs2_holder d_gh
;
367 uint64_t offset
= file
->f_pos
;
370 fdr
.fdr_sbd
= GFS2_SB(dir
);
371 fdr
.fdr_prefetch
= 1;
372 fdr
.fdr_filldir
= filldir
;
373 fdr
.fdr_opaque
= dirent
;
375 gfs2_holder_init(dip
->i_gl
, LM_ST_SHARED
, GL_ATIME
, &d_gh
);
376 error
= gfs2_glock_nq_atime(&d_gh
);
378 gfs2_holder_uninit(&d_gh
);
382 error
= gfs2_dir_read(dir
, &offset
, &fdr
, filldir_reg_func
);
384 gfs2_glock_dq_uninit(&d_gh
);
386 file
->f_pos
= offset
;
392 * filldir_bad_func - Report a directory entry to the caller of gfs2_dir_read()
393 * @opaque: opaque data used by the function
394 * @name: the name of the directory entry
395 * @length: the length of the name
396 * @offset: the entry's offset in the directory
397 * @inum: the inode number the entry points to
398 * @type: the type of inode the entry points to
400 * For supporting NFS.
402 * Returns: 0 on success, 1 if buffer full
405 static int filldir_bad_func(void *opaque
, const char *name
, unsigned int length
,
406 uint64_t offset
, struct gfs2_inum
*inum
,
409 struct filldir_bad
*fdb
= (struct filldir_bad
*)opaque
;
410 struct gfs2_sbd
*sdp
= fdb
->fdb_sbd
;
411 struct filldir_bad_entry
*fbe
;
413 if (fdb
->fdb_entry_off
== fdb
->fdb_entry_num
||
414 fdb
->fdb_name_off
+ length
> fdb
->fdb_name_size
)
417 fbe
= &fdb
->fdb_entry
[fdb
->fdb_entry_off
];
418 fbe
->fbe_name
= fdb
->fdb_name
+ fdb
->fdb_name_off
;
419 memcpy(fbe
->fbe_name
, name
, length
);
420 fbe
->fbe_length
= length
;
421 fbe
->fbe_offset
= offset
;
422 fbe
->fbe_inum
= *inum
;
423 fbe
->fbe_type
= type
;
425 fdb
->fdb_entry_off
++;
426 fdb
->fdb_name_off
+= length
;
428 if (!(length
== 1 && *name
== '.')) {
429 gfs2_glock_prefetch_num(sdp
,
430 inum
->no_addr
, &gfs2_inode_glops
,
431 LM_ST_SHARED
, LM_FLAG_TRY
| LM_FLAG_ANY
);
432 gfs2_glock_prefetch_num(sdp
,
433 inum
->no_addr
, &gfs2_iopen_glops
,
434 LM_ST_SHARED
, LM_FLAG_TRY
);
441 * readdir_bad - Read directory entries from a directory
442 * @file: The directory to read from
443 * @dirent: Buffer for dirents
444 * @filldir: Function used to do the copying
446 * For supporting NFS.
451 static int readdir_bad(struct file
*file
, void *dirent
, filldir_t filldir
)
453 struct inode
*dir
= file
->f_mapping
->host
;
454 struct gfs2_inode
*dip
= GFS2_I(dir
);
455 struct gfs2_sbd
*sdp
= GFS2_SB(dir
);
456 struct filldir_reg fdr
;
457 unsigned int entries
, size
;
458 struct filldir_bad
*fdb
;
459 struct gfs2_holder d_gh
;
460 uint64_t offset
= file
->f_pos
;
462 struct filldir_bad_entry
*fbe
;
465 entries
= gfs2_tune_get(sdp
, gt_entries_per_readdir
);
466 size
= sizeof(struct filldir_bad
) +
467 entries
* (sizeof(struct filldir_bad_entry
) + GFS2_FAST_NAME_SIZE
);
469 fdb
= kzalloc(size
, GFP_KERNEL
);
474 fdb
->fdb_entry
= (struct filldir_bad_entry
*)(fdb
+ 1);
475 fdb
->fdb_entry_num
= entries
;
476 fdb
->fdb_name
= ((char *)fdb
) + sizeof(struct filldir_bad
) +
477 entries
* sizeof(struct filldir_bad_entry
);
478 fdb
->fdb_name_size
= entries
* GFS2_FAST_NAME_SIZE
;
480 gfs2_holder_init(dip
->i_gl
, LM_ST_SHARED
, GL_ATIME
, &d_gh
);
481 error
= gfs2_glock_nq_atime(&d_gh
);
483 gfs2_holder_uninit(&d_gh
);
487 error
= gfs2_dir_read(dir
, &offset
, fdb
, filldir_bad_func
);
489 gfs2_glock_dq_uninit(&d_gh
);
492 fdr
.fdr_prefetch
= 0;
493 fdr
.fdr_filldir
= filldir
;
494 fdr
.fdr_opaque
= dirent
;
496 for (x
= 0; x
< fdb
->fdb_entry_off
; x
++) {
497 fbe
= &fdb
->fdb_entry
[x
];
499 error
= filldir_reg_func(&fdr
,
500 fbe
->fbe_name
, fbe
->fbe_length
,
502 &fbe
->fbe_inum
, fbe
->fbe_type
);
504 file
->f_pos
= fbe
->fbe_offset
;
510 file
->f_pos
= offset
;
519 * gfs2_readdir - Read directory entries from a directory
520 * @file: The directory to read from
521 * @dirent: Buffer for dirents
522 * @filldir: Function used to do the copying
527 static int gfs2_readdir(struct file
*file
, void *dirent
, filldir_t filldir
)
531 if (strcmp(current
->comm
, "nfsd") != 0)
532 error
= readdir_reg(file
, dirent
, filldir
);
534 error
= readdir_bad(file
, dirent
, filldir
);
539 static const u32 iflags_to_gfs2
[32] = {
540 [iflag_Sync
] = GFS2_DIF_SYNC
,
541 [iflag_Immutable
] = GFS2_DIF_IMMUTABLE
,
542 [iflag_Append
] = GFS2_DIF_APPENDONLY
,
543 [iflag_NoAtime
] = GFS2_DIF_NOATIME
,
544 [iflag_Index
] = GFS2_DIF_EXHASH
,
545 [iflag_JournalData
] = GFS2_DIF_JDATA
,
546 [iflag_DirectIO
] = GFS2_DIF_DIRECTIO
,
549 static const u32 gfs2_to_iflags
[32] = {
550 [gfs2fl_Sync
] = IFLAG_SYNC
,
551 [gfs2fl_Immutable
] = IFLAG_IMMUTABLE
,
552 [gfs2fl_AppendOnly
] = IFLAG_APPEND
,
553 [gfs2fl_NoAtime
] = IFLAG_NOATIME
,
554 [gfs2fl_ExHash
] = IFLAG_INDEX
,
555 [gfs2fl_Jdata
] = IFLAG_JOURNAL_DATA
,
556 [gfs2fl_Directio
] = IFLAG_DIRECTIO
,
557 [gfs2fl_InheritDirectio
] = IFLAG_DIRECTIO
,
558 [gfs2fl_InheritJdata
] = IFLAG_JOURNAL_DATA
,
561 static int gfs2_get_flags(struct file
*filp
, u32 __user
*ptr
)
563 struct inode
*inode
= filp
->f_dentry
->d_inode
;
564 struct gfs2_inode
*ip
= GFS2_I(inode
);
565 struct gfs2_holder gh
;
569 gfs2_holder_init(ip
->i_gl
, LM_ST_SHARED
, GL_ATIME
, &gh
);
570 error
= gfs2_glock_nq_m_atime(1, &gh
);
574 iflags
= iflags_cvt(gfs2_to_iflags
, ip
->i_di
.di_flags
);
575 if (put_user(iflags
, ptr
))
578 gfs2_glock_dq_m(1, &gh
);
579 gfs2_holder_uninit(&gh
);
583 /* Flags that can be set by user space */
584 #define GFS2_FLAGS_USER_SET (GFS2_DIF_JDATA| \
586 GFS2_DIF_IMMUTABLE| \
587 GFS2_DIF_APPENDONLY| \
591 GFS2_DIF_INHERIT_DIRECTIO| \
592 GFS2_DIF_INHERIT_JDATA)
595 * gfs2_set_flags - set flags on an inode
597 * @flags: The flags to set
598 * @mask: Indicates which flags are valid
601 static int do_gfs2_set_flags(struct file
*filp
, u32 reqflags
, u32 mask
)
603 struct inode
*inode
= filp
->f_dentry
->d_inode
;
604 struct gfs2_inode
*ip
= GFS2_I(inode
);
605 struct gfs2_sbd
*sdp
= GFS2_SB(inode
);
606 struct buffer_head
*bh
;
607 struct gfs2_holder gh
;
609 u32 new_flags
, flags
;
611 gfs2_holder_init(ip
->i_gl
, LM_ST_EXCLUSIVE
, 0, &gh
);
612 error
= gfs2_glock_nq_init(ip
->i_gl
, LM_ST_EXCLUSIVE
, 0, &gh
);
614 gfs2_holder_uninit(&gh
);
618 flags
= ip
->i_di
.di_flags
;
619 new_flags
= (flags
& ~mask
) | (reqflags
& mask
);
620 if ((new_flags
^ flags
) == 0)
623 if (S_ISDIR(inode
->i_mode
)) {
624 if ((new_flags
^ flags
) & GFS2_DIF_JDATA
)
625 new_flags
^= (GFS2_DIF_JDATA
|GFS2_DIF_INHERIT_JDATA
);
626 if ((new_flags
^ flags
) & GFS2_DIF_DIRECTIO
)
627 new_flags
^= (GFS2_DIF_DIRECTIO
|GFS2_DIF_INHERIT_DIRECTIO
);
631 if ((new_flags
^ flags
) & ~GFS2_FLAGS_USER_SET
)
635 if (IS_IMMUTABLE(inode
) && (new_flags
& GFS2_DIF_IMMUTABLE
))
637 if (IS_APPEND(inode
) && (new_flags
& GFS2_DIF_APPENDONLY
))
639 if (((new_flags
^ flags
) & GFS2_DIF_IMMUTABLE
) &&
640 !capable(CAP_LINUX_IMMUTABLE
))
642 if (!IS_IMMUTABLE(inode
)) {
643 error
= gfs2_repermission(inode
, MAY_WRITE
, NULL
);
648 error
= gfs2_trans_begin(sdp
, RES_DINODE
, 0);
651 error
= gfs2_meta_inode_buffer(ip
, &bh
);
654 gfs2_trans_add_bh(ip
->i_gl
, bh
, 1);
655 ip
->i_di
.di_flags
= new_flags
;
656 gfs2_dinode_out(&ip
->i_di
, bh
->b_data
);
661 gfs2_glock_dq_uninit(&gh
);
665 static int gfs2_set_flags(struct file
*filp
, u32 __user
*ptr
)
667 u32 iflags
, gfsflags
;
668 if (get_user(iflags
, ptr
))
670 gfsflags
= iflags_cvt(iflags_to_gfs2
, iflags
);
671 return do_gfs2_set_flags(filp
, gfsflags
, ~0);
674 static long gfs2_ioctl(struct file
*filp
, unsigned int cmd
, unsigned long arg
)
678 return gfs2_get_flags(filp
, (u32 __user
*)arg
);
680 return gfs2_set_flags(filp
, (u32 __user
*)arg
);
688 * @file: The file to map
689 * @vma: The VMA which described the mapping
691 * Returns: 0 or error code
694 static int gfs2_mmap(struct file
*file
, struct vm_area_struct
*vma
)
696 struct gfs2_inode
*ip
= GFS2_I(file
->f_mapping
->host
);
697 struct gfs2_holder i_gh
;
700 gfs2_holder_init(ip
->i_gl
, LM_ST_SHARED
, GL_ATIME
, &i_gh
);
701 error
= gfs2_glock_nq_atime(&i_gh
);
703 gfs2_holder_uninit(&i_gh
);
707 /* This is VM_MAYWRITE instead of VM_WRITE because a call
708 to mprotect() can turn on VM_WRITE later. */
710 if ((vma
->vm_flags
& (VM_MAYSHARE
| VM_MAYWRITE
)) ==
711 (VM_MAYSHARE
| VM_MAYWRITE
))
712 vma
->vm_ops
= &gfs2_vm_ops_sharewrite
;
714 vma
->vm_ops
= &gfs2_vm_ops_private
;
716 gfs2_glock_dq_uninit(&i_gh
);
722 * gfs2_open - open a file
723 * @inode: the inode to open
724 * @file: the struct file for this opening
729 static int gfs2_open(struct inode
*inode
, struct file
*file
)
731 struct gfs2_inode
*ip
= GFS2_I(inode
);
732 struct gfs2_holder i_gh
;
733 struct gfs2_file
*fp
;
736 fp
= kzalloc(sizeof(struct gfs2_file
), GFP_KERNEL
);
740 mutex_init(&fp
->f_fl_mutex
);
742 gfs2_assert_warn(GFS2_SB(inode
), !file
->private_data
);
743 file
->private_data
= fp
;
745 if (S_ISREG(ip
->i_di
.di_mode
)) {
746 error
= gfs2_glock_nq_init(ip
->i_gl
, LM_ST_SHARED
, LM_FLAG_ANY
,
751 if (!(file
->f_flags
& O_LARGEFILE
) &&
752 ip
->i_di
.di_size
> MAX_NON_LFS
) {
757 /* Listen to the Direct I/O flag */
759 if (ip
->i_di
.di_flags
& GFS2_DIF_DIRECTIO
)
760 file
->f_flags
|= O_DIRECT
;
762 gfs2_glock_dq_uninit(&i_gh
);
768 gfs2_glock_dq_uninit(&i_gh
);
771 file
->private_data
= NULL
;
778 * gfs2_close - called to close a struct file
779 * @inode: the inode the struct file belongs to
780 * @file: the struct file being closed
785 static int gfs2_close(struct inode
*inode
, struct file
*file
)
787 struct gfs2_sbd
*sdp
= inode
->i_sb
->s_fs_info
;
788 struct gfs2_file
*fp
;
790 fp
= file
->private_data
;
791 file
->private_data
= NULL
;
793 if (gfs2_assert_warn(sdp
, fp
))
802 * gfs2_fsync - sync the dirty data for a file (across the cluster)
803 * @file: the file that points to the dentry (we ignore this)
804 * @dentry: the dentry that points to the inode to sync
809 static int gfs2_fsync(struct file
*file
, struct dentry
*dentry
, int datasync
)
811 struct gfs2_inode
*ip
= GFS2_I(dentry
->d_inode
);
813 gfs2_log_flush(ip
->i_gl
->gl_sbd
, ip
->i_gl
);
819 * gfs2_lock - acquire/release a posix lock on a file
820 * @file: the file pointer
821 * @cmd: either modify or retrieve lock state, possibly wait
822 * @fl: type and range of lock
827 static int gfs2_lock(struct file
*file
, int cmd
, struct file_lock
*fl
)
829 struct gfs2_inode
*ip
= GFS2_I(file
->f_mapping
->host
);
830 struct gfs2_sbd
*sdp
= GFS2_SB(file
->f_mapping
->host
);
831 struct lm_lockname name
=
832 { .ln_number
= ip
->i_num
.no_addr
,
833 .ln_type
= LM_TYPE_PLOCK
};
835 if (!(fl
->fl_flags
& FL_POSIX
))
837 if ((ip
->i_di
.di_mode
& (S_ISGID
| S_IXGRP
)) == S_ISGID
)
840 if (sdp
->sd_args
.ar_localflocks
) {
842 struct file_lock tmp
;
844 ret
= posix_test_lock(file
, fl
, &tmp
);
845 fl
->fl_type
= F_UNLCK
;
847 memcpy(fl
, &tmp
, sizeof(struct file_lock
));
850 return posix_lock_file_wait(file
, fl
);
855 return gfs2_lm_plock_get(sdp
, &name
, file
, fl
);
856 else if (fl
->fl_type
== F_UNLCK
)
857 return gfs2_lm_punlock(sdp
, &name
, file
, fl
);
859 return gfs2_lm_plock(sdp
, &name
, file
, cmd
, fl
);
863 * gfs2_sendfile - Send bytes to a file or socket
864 * @in_file: The file to read from
865 * @out_file: The file to write to
866 * @count: The amount of data
867 * @offset: The beginning file offset
869 * Outputs: offset - updated according to number of bytes read
871 * Returns: The number of bytes sent, errno on failure
874 static ssize_t
gfs2_sendfile(struct file
*in_file
, loff_t
*offset
, size_t count
,
875 read_actor_t actor
, void *target
)
877 return generic_file_sendfile(in_file
, offset
, count
, actor
, target
);
880 static int do_flock(struct file
*file
, int cmd
, struct file_lock
*fl
)
882 struct gfs2_file
*fp
= file
->private_data
;
883 struct gfs2_holder
*fl_gh
= &fp
->f_fl_gh
;
884 struct gfs2_inode
*ip
= GFS2_I(file
->f_dentry
->d_inode
);
885 struct gfs2_glock
*gl
;
890 state
= (fl
->fl_type
== F_WRLCK
) ? LM_ST_EXCLUSIVE
: LM_ST_SHARED
;
891 flags
= ((IS_SETLKW(cmd
)) ? 0 : LM_FLAG_TRY
) | GL_EXACT
| GL_NOCACHE
;
893 mutex_lock(&fp
->f_fl_mutex
);
897 if (fl_gh
->gh_state
== state
)
900 flock_lock_file_wait(file
,
901 &(struct file_lock
){.fl_type
= F_UNLCK
});
902 gfs2_glock_dq_uninit(fl_gh
);
904 error
= gfs2_glock_get(GFS2_SB(&ip
->i_inode
),
905 ip
->i_num
.no_addr
, &gfs2_flock_glops
,
911 gfs2_holder_init(gl
, state
, flags
, fl_gh
);
914 error
= gfs2_glock_nq(fl_gh
);
916 gfs2_holder_uninit(fl_gh
);
917 if (error
== GLR_TRYFAILED
)
920 error
= flock_lock_file_wait(file
, fl
);
921 gfs2_assert_warn(GFS2_SB(&ip
->i_inode
), !error
);
925 mutex_unlock(&fp
->f_fl_mutex
);
930 static void do_unflock(struct file
*file
, struct file_lock
*fl
)
932 struct gfs2_file
*fp
= file
->private_data
;
933 struct gfs2_holder
*fl_gh
= &fp
->f_fl_gh
;
935 mutex_lock(&fp
->f_fl_mutex
);
936 flock_lock_file_wait(file
, fl
);
938 gfs2_glock_dq_uninit(fl_gh
);
939 mutex_unlock(&fp
->f_fl_mutex
);
943 * gfs2_flock - acquire/release a flock lock on a file
944 * @file: the file pointer
945 * @cmd: either modify or retrieve lock state, possibly wait
946 * @fl: type and range of lock
951 static int gfs2_flock(struct file
*file
, int cmd
, struct file_lock
*fl
)
953 struct gfs2_inode
*ip
= GFS2_I(file
->f_mapping
->host
);
954 struct gfs2_sbd
*sdp
= GFS2_SB(file
->f_mapping
->host
);
956 if (!(fl
->fl_flags
& FL_FLOCK
))
958 if ((ip
->i_di
.di_mode
& (S_ISGID
| S_IXGRP
)) == S_ISGID
)
961 if (sdp
->sd_args
.ar_localflocks
)
962 return flock_lock_file_wait(file
, fl
);
964 if (fl
->fl_type
== F_UNLCK
) {
965 do_unflock(file
, fl
);
968 return do_flock(file
, cmd
, fl
);
971 struct file_operations gfs2_file_fops
= {
972 .llseek
= gfs2_llseek
,
974 .readv
= gfs2_file_readv
,
975 .aio_read
= gfs2_file_aio_read
,
976 .write
= generic_file_write
,
977 .writev
= generic_file_writev
,
978 .aio_write
= generic_file_aio_write
,
979 .unlocked_ioctl
= gfs2_ioctl
,
982 .release
= gfs2_close
,
985 .sendfile
= gfs2_sendfile
,
987 .splice_read
= generic_file_splice_read
,
988 .splice_write
= generic_file_splice_write
,
991 struct file_operations gfs2_dir_fops
= {
992 .readdir
= gfs2_readdir
,
993 .unlocked_ioctl
= gfs2_ioctl
,
995 .release
= gfs2_close
,