4 * Copyright (C) 1992 Rick Sladkey
7 #include <linux/file.h>
8 #include <linux/falloc.h>
9 #include <linux/nfs_fs.h>
10 #include <uapi/linux/btrfs.h> /* BTRFS_IOC_CLONE/BTRFS_IOC_CLONE_RANGE */
11 #include "delegation.h"
19 #ifdef CONFIG_NFS_V4_2
23 #define NFSDBG_FACILITY NFSDBG_FILE
26 nfs4_file_open(struct inode
*inode
, struct file
*filp
)
28 struct nfs_open_context
*ctx
;
29 struct dentry
*dentry
= filp
->f_path
.dentry
;
30 struct dentry
*parent
= NULL
;
32 unsigned openflags
= filp
->f_flags
;
37 * If no cached dentry exists or if it's negative, NFSv4 handled the
38 * opens in ->lookup() or ->create().
40 * We only get this far for a cached positive dentry. We skipped
41 * revalidation, so handle it here by dropping the dentry and returning
42 * -EOPENSTALE. The VFS will retry the lookup/create/open.
45 dprintk("NFS: open file(%pd2)\n", dentry
);
47 err
= nfs_check_flags(openflags
);
51 if ((openflags
& O_ACCMODE
) == 3)
54 /* We can't create new files here */
55 openflags
&= ~(O_CREAT
|O_EXCL
);
57 parent
= dget_parent(dentry
);
58 dir
= d_inode(parent
);
60 ctx
= alloc_nfs_open_context(filp
->f_path
.dentry
, filp
->f_mode
);
65 attr
.ia_valid
= ATTR_OPEN
;
66 if (openflags
& O_TRUNC
) {
67 attr
.ia_valid
|= ATTR_SIZE
;
69 nfs_sync_inode(inode
);
72 inode
= NFS_PROTO(dir
)->open_context(dir
, ctx
, openflags
, &attr
, NULL
);
86 if (inode
!= d_inode(dentry
))
89 nfs_set_verifier(dentry
, nfs_save_change_attribute(dir
));
90 nfs_file_set_open_context(filp
, ctx
);
91 nfs_fscache_open_file(inode
, filp
);
95 put_nfs_open_context(ctx
);
107 * Flush all dirty pages, and check for write errors.
110 nfs4_file_flush(struct file
*file
, fl_owner_t id
)
112 struct inode
*inode
= file_inode(file
);
114 dprintk("NFS: flush(%pD2)\n", file
);
116 nfs_inc_stats(inode
, NFSIOS_VFSFLUSH
);
117 if ((file
->f_mode
& FMODE_WRITE
) == 0)
121 * If we're holding a write delegation, then check if we're required
122 * to flush the i/o on close. If not, then just start the i/o now.
124 if (!nfs4_delegation_flush_on_close(inode
))
125 return filemap_fdatawrite(file
->f_mapping
);
127 /* Flush writes to the server and return any errors */
128 return vfs_fsync(file
, 0);
132 nfs4_file_fsync(struct file
*file
, loff_t start
, loff_t end
, int datasync
)
135 struct inode
*inode
= file_inode(file
);
137 trace_nfs_fsync_enter(inode
);
139 nfs_inode_dio_wait(inode
);
141 ret
= filemap_write_and_wait_range(inode
->i_mapping
, start
, end
);
145 ret
= nfs_file_fsync_commit(file
, start
, end
, datasync
);
147 ret
= pnfs_sync_inode(inode
, !!datasync
);
150 * If nfs_file_fsync_commit detected a server reboot, then
151 * resend all dirty pages that might have been covered by
152 * the NFS_CONTEXT_RESEND_WRITES flag
156 } while (ret
== -EAGAIN
);
158 trace_nfs_fsync_exit(inode
, ret
);
162 #ifdef CONFIG_NFS_V4_2
163 static loff_t
nfs4_file_llseek(struct file
*filep
, loff_t offset
, int whence
)
170 ret
= nfs42_proc_llseek(filep
, offset
, whence
);
171 if (ret
!= -ENOTSUPP
)
174 return nfs_file_llseek(filep
, offset
, whence
);
178 static long nfs42_fallocate(struct file
*filep
, int mode
, loff_t offset
, loff_t len
)
180 struct inode
*inode
= file_inode(filep
);
183 if (!S_ISREG(inode
->i_mode
))
186 if ((mode
!= 0) && (mode
!= (FALLOC_FL_PUNCH_HOLE
| FALLOC_FL_KEEP_SIZE
)))
189 ret
= inode_newsize_ok(inode
, offset
+ len
);
193 if (mode
& FALLOC_FL_PUNCH_HOLE
)
194 return nfs42_proc_deallocate(filep
, offset
, len
);
195 return nfs42_proc_allocate(filep
, offset
, len
);
198 static int nfs42_clone_file_range(struct file
*src_file
, loff_t src_off
,
199 struct file
*dst_file
, loff_t dst_off
, u64 count
)
201 struct inode
*dst_inode
= file_inode(dst_file
);
202 struct nfs_server
*server
= NFS_SERVER(dst_inode
);
203 struct inode
*src_inode
= file_inode(src_file
);
204 unsigned int bs
= server
->clone_blksize
;
205 bool same_inode
= false;
208 /* check alignment w.r.t. clone_blksize */
211 if (!IS_ALIGNED(src_off
, bs
) || !IS_ALIGNED(dst_off
, bs
))
213 if (!IS_ALIGNED(count
, bs
) && i_size_read(src_inode
) != (src_off
+ count
))
217 if (src_inode
== dst_inode
)
220 /* XXX: do we lock at all? what if server needs CB_RECALL_LAYOUT? */
222 inode_lock(src_inode
);
223 } else if (dst_inode
< src_inode
) {
224 inode_lock_nested(dst_inode
, I_MUTEX_PARENT
);
225 inode_lock_nested(src_inode
, I_MUTEX_CHILD
);
227 inode_lock_nested(src_inode
, I_MUTEX_PARENT
);
228 inode_lock_nested(dst_inode
, I_MUTEX_CHILD
);
231 /* flush all pending writes on both src and dst so that server
232 * has the latest data */
233 ret
= nfs_sync_inode(src_inode
);
236 ret
= nfs_sync_inode(dst_inode
);
240 ret
= nfs42_proc_clone(src_file
, dst_file
, src_off
, dst_off
, count
);
242 /* truncate inode page cache of the dst range so that future reads can fetch
243 * new data from server */
245 truncate_inode_pages_range(&dst_inode
->i_data
, dst_off
, dst_off
+ count
- 1);
249 inode_unlock(src_inode
);
250 } else if (dst_inode
< src_inode
) {
251 inode_unlock(src_inode
);
252 inode_unlock(dst_inode
);
254 inode_unlock(dst_inode
);
255 inode_unlock(src_inode
);
260 #endif /* CONFIG_NFS_V4_2 */
262 const struct file_operations nfs4_file_operations
= {
263 .read_iter
= nfs_file_read
,
264 .write_iter
= nfs_file_write
,
265 .mmap
= nfs_file_mmap
,
266 .open
= nfs4_file_open
,
267 .flush
= nfs4_file_flush
,
268 .release
= nfs_file_release
,
269 .fsync
= nfs4_file_fsync
,
272 .splice_read
= nfs_file_splice_read
,
273 .splice_write
= iter_file_splice_write
,
274 .check_flags
= nfs_check_flags
,
275 .setlease
= simple_nosetlease
,
276 #ifdef CONFIG_NFS_V4_2
277 .llseek
= nfs4_file_llseek
,
278 .fallocate
= nfs42_fallocate
,
279 .clone_file_range
= nfs42_clone_file_range
,
281 .llseek
= nfs_file_llseek
,