Merge branch 'for-linus-4.7' of git://git.kernel.org/pub/scm/linux/kernel/git/mason...
[deliverable/linux.git] / fs / hfsplus / super.c
1 /*
2 * linux/fs/hfsplus/super.c
3 *
4 * Copyright (C) 2001
5 * Brad Boyer (flar@allandria.com)
6 * (C) 2003 Ardis Technologies <roman@ardistech.com>
7 *
8 */
9
10 #include <linux/module.h>
11 #include <linux/init.h>
12 #include <linux/pagemap.h>
13 #include <linux/blkdev.h>
14 #include <linux/backing-dev.h>
15 #include <linux/fs.h>
16 #include <linux/slab.h>
17 #include <linux/vfs.h>
18 #include <linux/nls.h>
19
20 static struct inode *hfsplus_alloc_inode(struct super_block *sb);
21 static void hfsplus_destroy_inode(struct inode *inode);
22
23 #include "hfsplus_fs.h"
24 #include "xattr.h"
25
26 static int hfsplus_system_read_inode(struct inode *inode)
27 {
28 struct hfsplus_vh *vhdr = HFSPLUS_SB(inode->i_sb)->s_vhdr;
29
30 switch (inode->i_ino) {
31 case HFSPLUS_EXT_CNID:
32 hfsplus_inode_read_fork(inode, &vhdr->ext_file);
33 inode->i_mapping->a_ops = &hfsplus_btree_aops;
34 break;
35 case HFSPLUS_CAT_CNID:
36 hfsplus_inode_read_fork(inode, &vhdr->cat_file);
37 inode->i_mapping->a_ops = &hfsplus_btree_aops;
38 break;
39 case HFSPLUS_ALLOC_CNID:
40 hfsplus_inode_read_fork(inode, &vhdr->alloc_file);
41 inode->i_mapping->a_ops = &hfsplus_aops;
42 break;
43 case HFSPLUS_START_CNID:
44 hfsplus_inode_read_fork(inode, &vhdr->start_file);
45 break;
46 case HFSPLUS_ATTR_CNID:
47 hfsplus_inode_read_fork(inode, &vhdr->attr_file);
48 inode->i_mapping->a_ops = &hfsplus_btree_aops;
49 break;
50 default:
51 return -EIO;
52 }
53
54 return 0;
55 }
56
57 struct inode *hfsplus_iget(struct super_block *sb, unsigned long ino)
58 {
59 struct hfs_find_data fd;
60 struct inode *inode;
61 int err;
62
63 inode = iget_locked(sb, ino);
64 if (!inode)
65 return ERR_PTR(-ENOMEM);
66 if (!(inode->i_state & I_NEW))
67 return inode;
68
69 INIT_LIST_HEAD(&HFSPLUS_I(inode)->open_dir_list);
70 spin_lock_init(&HFSPLUS_I(inode)->open_dir_lock);
71 mutex_init(&HFSPLUS_I(inode)->extents_lock);
72 HFSPLUS_I(inode)->flags = 0;
73 HFSPLUS_I(inode)->extent_state = 0;
74 HFSPLUS_I(inode)->rsrc_inode = NULL;
75 atomic_set(&HFSPLUS_I(inode)->opencnt, 0);
76
77 if (inode->i_ino >= HFSPLUS_FIRSTUSER_CNID ||
78 inode->i_ino == HFSPLUS_ROOT_CNID) {
79 err = hfs_find_init(HFSPLUS_SB(inode->i_sb)->cat_tree, &fd);
80 if (!err) {
81 err = hfsplus_find_cat(inode->i_sb, inode->i_ino, &fd);
82 if (!err)
83 err = hfsplus_cat_read_inode(inode, &fd);
84 hfs_find_exit(&fd);
85 }
86 } else {
87 err = hfsplus_system_read_inode(inode);
88 }
89
90 if (err) {
91 iget_failed(inode);
92 return ERR_PTR(err);
93 }
94
95 unlock_new_inode(inode);
96 return inode;
97 }
98
99 static int hfsplus_system_write_inode(struct inode *inode)
100 {
101 struct hfsplus_sb_info *sbi = HFSPLUS_SB(inode->i_sb);
102 struct hfsplus_vh *vhdr = sbi->s_vhdr;
103 struct hfsplus_fork_raw *fork;
104 struct hfs_btree *tree = NULL;
105
106 switch (inode->i_ino) {
107 case HFSPLUS_EXT_CNID:
108 fork = &vhdr->ext_file;
109 tree = sbi->ext_tree;
110 break;
111 case HFSPLUS_CAT_CNID:
112 fork = &vhdr->cat_file;
113 tree = sbi->cat_tree;
114 break;
115 case HFSPLUS_ALLOC_CNID:
116 fork = &vhdr->alloc_file;
117 break;
118 case HFSPLUS_START_CNID:
119 fork = &vhdr->start_file;
120 break;
121 case HFSPLUS_ATTR_CNID:
122 fork = &vhdr->attr_file;
123 tree = sbi->attr_tree;
124 break;
125 default:
126 return -EIO;
127 }
128
129 if (fork->total_size != cpu_to_be64(inode->i_size)) {
130 set_bit(HFSPLUS_SB_WRITEBACKUP, &sbi->flags);
131 hfsplus_mark_mdb_dirty(inode->i_sb);
132 }
133 hfsplus_inode_write_fork(inode, fork);
134 if (tree) {
135 int err = hfs_btree_write(tree);
136
137 if (err) {
138 pr_err("b-tree write err: %d, ino %lu\n",
139 err, inode->i_ino);
140 return err;
141 }
142 }
143 return 0;
144 }
145
146 static int hfsplus_write_inode(struct inode *inode,
147 struct writeback_control *wbc)
148 {
149 int err;
150
151 hfs_dbg(INODE, "hfsplus_write_inode: %lu\n", inode->i_ino);
152
153 err = hfsplus_ext_write_extent(inode);
154 if (err)
155 return err;
156
157 if (inode->i_ino >= HFSPLUS_FIRSTUSER_CNID ||
158 inode->i_ino == HFSPLUS_ROOT_CNID)
159 return hfsplus_cat_write_inode(inode);
160 else
161 return hfsplus_system_write_inode(inode);
162 }
163
164 static void hfsplus_evict_inode(struct inode *inode)
165 {
166 hfs_dbg(INODE, "hfsplus_evict_inode: %lu\n", inode->i_ino);
167 truncate_inode_pages_final(&inode->i_data);
168 clear_inode(inode);
169 if (HFSPLUS_IS_RSRC(inode)) {
170 HFSPLUS_I(HFSPLUS_I(inode)->rsrc_inode)->rsrc_inode = NULL;
171 iput(HFSPLUS_I(inode)->rsrc_inode);
172 }
173 }
174
175 static int hfsplus_sync_fs(struct super_block *sb, int wait)
176 {
177 struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
178 struct hfsplus_vh *vhdr = sbi->s_vhdr;
179 int write_backup = 0;
180 int error, error2;
181
182 if (!wait)
183 return 0;
184
185 hfs_dbg(SUPER, "hfsplus_sync_fs\n");
186
187 /*
188 * Explicitly write out the special metadata inodes.
189 *
190 * While these special inodes are marked as hashed and written
191 * out peridocically by the flusher threads we redirty them
192 * during writeout of normal inodes, and thus the life lock
193 * prevents us from getting the latest state to disk.
194 */
195 error = filemap_write_and_wait(sbi->cat_tree->inode->i_mapping);
196 error2 = filemap_write_and_wait(sbi->ext_tree->inode->i_mapping);
197 if (!error)
198 error = error2;
199 if (sbi->attr_tree) {
200 error2 =
201 filemap_write_and_wait(sbi->attr_tree->inode->i_mapping);
202 if (!error)
203 error = error2;
204 }
205 error2 = filemap_write_and_wait(sbi->alloc_file->i_mapping);
206 if (!error)
207 error = error2;
208
209 mutex_lock(&sbi->vh_mutex);
210 mutex_lock(&sbi->alloc_mutex);
211 vhdr->free_blocks = cpu_to_be32(sbi->free_blocks);
212 vhdr->next_cnid = cpu_to_be32(sbi->next_cnid);
213 vhdr->folder_count = cpu_to_be32(sbi->folder_count);
214 vhdr->file_count = cpu_to_be32(sbi->file_count);
215
216 if (test_and_clear_bit(HFSPLUS_SB_WRITEBACKUP, &sbi->flags)) {
217 memcpy(sbi->s_backup_vhdr, sbi->s_vhdr, sizeof(*sbi->s_vhdr));
218 write_backup = 1;
219 }
220
221 error2 = hfsplus_submit_bio(sb,
222 sbi->part_start + HFSPLUS_VOLHEAD_SECTOR,
223 sbi->s_vhdr_buf, NULL, WRITE_SYNC);
224 if (!error)
225 error = error2;
226 if (!write_backup)
227 goto out;
228
229 error2 = hfsplus_submit_bio(sb,
230 sbi->part_start + sbi->sect_count - 2,
231 sbi->s_backup_vhdr_buf, NULL, WRITE_SYNC);
232 if (!error)
233 error2 = error;
234 out:
235 mutex_unlock(&sbi->alloc_mutex);
236 mutex_unlock(&sbi->vh_mutex);
237
238 if (!test_bit(HFSPLUS_SB_NOBARRIER, &sbi->flags))
239 blkdev_issue_flush(sb->s_bdev, GFP_KERNEL, NULL);
240
241 return error;
242 }
243
244 static void delayed_sync_fs(struct work_struct *work)
245 {
246 int err;
247 struct hfsplus_sb_info *sbi;
248
249 sbi = container_of(work, struct hfsplus_sb_info, sync_work.work);
250
251 spin_lock(&sbi->work_lock);
252 sbi->work_queued = 0;
253 spin_unlock(&sbi->work_lock);
254
255 err = hfsplus_sync_fs(sbi->alloc_file->i_sb, 1);
256 if (err)
257 pr_err("delayed sync fs err %d\n", err);
258 }
259
260 void hfsplus_mark_mdb_dirty(struct super_block *sb)
261 {
262 struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
263 unsigned long delay;
264
265 if (sb->s_flags & MS_RDONLY)
266 return;
267
268 spin_lock(&sbi->work_lock);
269 if (!sbi->work_queued) {
270 delay = msecs_to_jiffies(dirty_writeback_interval * 10);
271 queue_delayed_work(system_long_wq, &sbi->sync_work, delay);
272 sbi->work_queued = 1;
273 }
274 spin_unlock(&sbi->work_lock);
275 }
276
277 static void hfsplus_put_super(struct super_block *sb)
278 {
279 struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
280
281 hfs_dbg(SUPER, "hfsplus_put_super\n");
282
283 cancel_delayed_work_sync(&sbi->sync_work);
284
285 if (!(sb->s_flags & MS_RDONLY) && sbi->s_vhdr) {
286 struct hfsplus_vh *vhdr = sbi->s_vhdr;
287
288 vhdr->modify_date = hfsp_now2mt();
289 vhdr->attributes |= cpu_to_be32(HFSPLUS_VOL_UNMNT);
290 vhdr->attributes &= cpu_to_be32(~HFSPLUS_VOL_INCNSTNT);
291
292 hfsplus_sync_fs(sb, 1);
293 }
294
295 hfs_btree_close(sbi->attr_tree);
296 hfs_btree_close(sbi->cat_tree);
297 hfs_btree_close(sbi->ext_tree);
298 iput(sbi->alloc_file);
299 iput(sbi->hidden_dir);
300 kfree(sbi->s_vhdr_buf);
301 kfree(sbi->s_backup_vhdr_buf);
302 unload_nls(sbi->nls);
303 kfree(sb->s_fs_info);
304 sb->s_fs_info = NULL;
305 }
306
307 static int hfsplus_statfs(struct dentry *dentry, struct kstatfs *buf)
308 {
309 struct super_block *sb = dentry->d_sb;
310 struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
311 u64 id = huge_encode_dev(sb->s_bdev->bd_dev);
312
313 buf->f_type = HFSPLUS_SUPER_MAGIC;
314 buf->f_bsize = sb->s_blocksize;
315 buf->f_blocks = sbi->total_blocks << sbi->fs_shift;
316 buf->f_bfree = sbi->free_blocks << sbi->fs_shift;
317 buf->f_bavail = buf->f_bfree;
318 buf->f_files = 0xFFFFFFFF;
319 buf->f_ffree = 0xFFFFFFFF - sbi->next_cnid;
320 buf->f_fsid.val[0] = (u32)id;
321 buf->f_fsid.val[1] = (u32)(id >> 32);
322 buf->f_namelen = HFSPLUS_MAX_STRLEN;
323
324 return 0;
325 }
326
327 static int hfsplus_remount(struct super_block *sb, int *flags, char *data)
328 {
329 sync_filesystem(sb);
330 if ((*flags & MS_RDONLY) == (sb->s_flags & MS_RDONLY))
331 return 0;
332 if (!(*flags & MS_RDONLY)) {
333 struct hfsplus_vh *vhdr = HFSPLUS_SB(sb)->s_vhdr;
334 int force = 0;
335
336 if (!hfsplus_parse_options_remount(data, &force))
337 return -EINVAL;
338
339 if (!(vhdr->attributes & cpu_to_be32(HFSPLUS_VOL_UNMNT))) {
340 pr_warn("filesystem was not cleanly unmounted, running fsck.hfsplus is recommended. leaving read-only.\n");
341 sb->s_flags |= MS_RDONLY;
342 *flags |= MS_RDONLY;
343 } else if (force) {
344 /* nothing */
345 } else if (vhdr->attributes &
346 cpu_to_be32(HFSPLUS_VOL_SOFTLOCK)) {
347 pr_warn("filesystem is marked locked, leaving read-only.\n");
348 sb->s_flags |= MS_RDONLY;
349 *flags |= MS_RDONLY;
350 } else if (vhdr->attributes &
351 cpu_to_be32(HFSPLUS_VOL_JOURNALED)) {
352 pr_warn("filesystem is marked journaled, leaving read-only.\n");
353 sb->s_flags |= MS_RDONLY;
354 *flags |= MS_RDONLY;
355 }
356 }
357 return 0;
358 }
359
360 static const struct super_operations hfsplus_sops = {
361 .alloc_inode = hfsplus_alloc_inode,
362 .destroy_inode = hfsplus_destroy_inode,
363 .write_inode = hfsplus_write_inode,
364 .evict_inode = hfsplus_evict_inode,
365 .put_super = hfsplus_put_super,
366 .sync_fs = hfsplus_sync_fs,
367 .statfs = hfsplus_statfs,
368 .remount_fs = hfsplus_remount,
369 .show_options = hfsplus_show_options,
370 };
371
372 static int hfsplus_fill_super(struct super_block *sb, void *data, int silent)
373 {
374 struct hfsplus_vh *vhdr;
375 struct hfsplus_sb_info *sbi;
376 hfsplus_cat_entry entry;
377 struct hfs_find_data fd;
378 struct inode *root, *inode;
379 struct qstr str;
380 struct nls_table *nls = NULL;
381 u64 last_fs_block, last_fs_page;
382 int err;
383
384 err = -ENOMEM;
385 sbi = kzalloc(sizeof(*sbi), GFP_KERNEL);
386 if (!sbi)
387 goto out;
388
389 sb->s_fs_info = sbi;
390 mutex_init(&sbi->alloc_mutex);
391 mutex_init(&sbi->vh_mutex);
392 spin_lock_init(&sbi->work_lock);
393 INIT_DELAYED_WORK(&sbi->sync_work, delayed_sync_fs);
394 hfsplus_fill_defaults(sbi);
395
396 err = -EINVAL;
397 if (!hfsplus_parse_options(data, sbi)) {
398 pr_err("unable to parse mount options\n");
399 goto out_unload_nls;
400 }
401
402 /* temporarily use utf8 to correctly find the hidden dir below */
403 nls = sbi->nls;
404 sbi->nls = load_nls("utf8");
405 if (!sbi->nls) {
406 pr_err("unable to load nls for utf8\n");
407 goto out_unload_nls;
408 }
409
410 /* Grab the volume header */
411 if (hfsplus_read_wrapper(sb)) {
412 if (!silent)
413 pr_warn("unable to find HFS+ superblock\n");
414 goto out_unload_nls;
415 }
416 vhdr = sbi->s_vhdr;
417
418 /* Copy parts of the volume header into the superblock */
419 sb->s_magic = HFSPLUS_VOLHEAD_SIG;
420 if (be16_to_cpu(vhdr->version) < HFSPLUS_MIN_VERSION ||
421 be16_to_cpu(vhdr->version) > HFSPLUS_CURRENT_VERSION) {
422 pr_err("wrong filesystem version\n");
423 goto out_free_vhdr;
424 }
425 sbi->total_blocks = be32_to_cpu(vhdr->total_blocks);
426 sbi->free_blocks = be32_to_cpu(vhdr->free_blocks);
427 sbi->next_cnid = be32_to_cpu(vhdr->next_cnid);
428 sbi->file_count = be32_to_cpu(vhdr->file_count);
429 sbi->folder_count = be32_to_cpu(vhdr->folder_count);
430 sbi->data_clump_blocks =
431 be32_to_cpu(vhdr->data_clump_sz) >> sbi->alloc_blksz_shift;
432 if (!sbi->data_clump_blocks)
433 sbi->data_clump_blocks = 1;
434 sbi->rsrc_clump_blocks =
435 be32_to_cpu(vhdr->rsrc_clump_sz) >> sbi->alloc_blksz_shift;
436 if (!sbi->rsrc_clump_blocks)
437 sbi->rsrc_clump_blocks = 1;
438
439 err = -EFBIG;
440 last_fs_block = sbi->total_blocks - 1;
441 last_fs_page = (last_fs_block << sbi->alloc_blksz_shift) >>
442 PAGE_SHIFT;
443
444 if ((last_fs_block > (sector_t)(~0ULL) >> (sbi->alloc_blksz_shift - 9)) ||
445 (last_fs_page > (pgoff_t)(~0ULL))) {
446 pr_err("filesystem size too large\n");
447 goto out_free_vhdr;
448 }
449
450 /* Set up operations so we can load metadata */
451 sb->s_op = &hfsplus_sops;
452 sb->s_maxbytes = MAX_LFS_FILESIZE;
453
454 if (!(vhdr->attributes & cpu_to_be32(HFSPLUS_VOL_UNMNT))) {
455 pr_warn("Filesystem was not cleanly unmounted, running fsck.hfsplus is recommended. mounting read-only.\n");
456 sb->s_flags |= MS_RDONLY;
457 } else if (test_and_clear_bit(HFSPLUS_SB_FORCE, &sbi->flags)) {
458 /* nothing */
459 } else if (vhdr->attributes & cpu_to_be32(HFSPLUS_VOL_SOFTLOCK)) {
460 pr_warn("Filesystem is marked locked, mounting read-only.\n");
461 sb->s_flags |= MS_RDONLY;
462 } else if ((vhdr->attributes & cpu_to_be32(HFSPLUS_VOL_JOURNALED)) &&
463 !(sb->s_flags & MS_RDONLY)) {
464 pr_warn("write access to a journaled filesystem is not supported, use the force option at your own risk, mounting read-only.\n");
465 sb->s_flags |= MS_RDONLY;
466 }
467
468 err = -EINVAL;
469
470 /* Load metadata objects (B*Trees) */
471 sbi->ext_tree = hfs_btree_open(sb, HFSPLUS_EXT_CNID);
472 if (!sbi->ext_tree) {
473 pr_err("failed to load extents file\n");
474 goto out_free_vhdr;
475 }
476 sbi->cat_tree = hfs_btree_open(sb, HFSPLUS_CAT_CNID);
477 if (!sbi->cat_tree) {
478 pr_err("failed to load catalog file\n");
479 goto out_close_ext_tree;
480 }
481 atomic_set(&sbi->attr_tree_state, HFSPLUS_EMPTY_ATTR_TREE);
482 if (vhdr->attr_file.total_blocks != 0) {
483 sbi->attr_tree = hfs_btree_open(sb, HFSPLUS_ATTR_CNID);
484 if (!sbi->attr_tree) {
485 pr_err("failed to load attributes file\n");
486 goto out_close_cat_tree;
487 }
488 atomic_set(&sbi->attr_tree_state, HFSPLUS_VALID_ATTR_TREE);
489 }
490 sb->s_xattr = hfsplus_xattr_handlers;
491
492 inode = hfsplus_iget(sb, HFSPLUS_ALLOC_CNID);
493 if (IS_ERR(inode)) {
494 pr_err("failed to load allocation file\n");
495 err = PTR_ERR(inode);
496 goto out_close_attr_tree;
497 }
498 sbi->alloc_file = inode;
499
500 /* Load the root directory */
501 root = hfsplus_iget(sb, HFSPLUS_ROOT_CNID);
502 if (IS_ERR(root)) {
503 pr_err("failed to load root directory\n");
504 err = PTR_ERR(root);
505 goto out_put_alloc_file;
506 }
507
508 sb->s_d_op = &hfsplus_dentry_operations;
509 sb->s_root = d_make_root(root);
510 if (!sb->s_root) {
511 err = -ENOMEM;
512 goto out_put_alloc_file;
513 }
514
515 str.len = sizeof(HFSP_HIDDENDIR_NAME) - 1;
516 str.name = HFSP_HIDDENDIR_NAME;
517 err = hfs_find_init(sbi->cat_tree, &fd);
518 if (err)
519 goto out_put_root;
520 err = hfsplus_cat_build_key(sb, fd.search_key, HFSPLUS_ROOT_CNID, &str);
521 if (unlikely(err < 0))
522 goto out_put_root;
523 if (!hfs_brec_read(&fd, &entry, sizeof(entry))) {
524 hfs_find_exit(&fd);
525 if (entry.type != cpu_to_be16(HFSPLUS_FOLDER))
526 goto out_put_root;
527 inode = hfsplus_iget(sb, be32_to_cpu(entry.folder.id));
528 if (IS_ERR(inode)) {
529 err = PTR_ERR(inode);
530 goto out_put_root;
531 }
532 sbi->hidden_dir = inode;
533 } else
534 hfs_find_exit(&fd);
535
536 if (!(sb->s_flags & MS_RDONLY)) {
537 /*
538 * H+LX == hfsplusutils, H+Lx == this driver, H+lx is unused
539 * all three are registered with Apple for our use
540 */
541 vhdr->last_mount_vers = cpu_to_be32(HFSP_MOUNT_VERSION);
542 vhdr->modify_date = hfsp_now2mt();
543 be32_add_cpu(&vhdr->write_count, 1);
544 vhdr->attributes &= cpu_to_be32(~HFSPLUS_VOL_UNMNT);
545 vhdr->attributes |= cpu_to_be32(HFSPLUS_VOL_INCNSTNT);
546 hfsplus_sync_fs(sb, 1);
547
548 if (!sbi->hidden_dir) {
549 mutex_lock(&sbi->vh_mutex);
550 sbi->hidden_dir = hfsplus_new_inode(sb, S_IFDIR);
551 if (!sbi->hidden_dir) {
552 mutex_unlock(&sbi->vh_mutex);
553 err = -ENOMEM;
554 goto out_put_root;
555 }
556 err = hfsplus_create_cat(sbi->hidden_dir->i_ino, root,
557 &str, sbi->hidden_dir);
558 if (err) {
559 mutex_unlock(&sbi->vh_mutex);
560 goto out_put_hidden_dir;
561 }
562
563 err = hfsplus_init_inode_security(sbi->hidden_dir,
564 root, &str);
565 if (err == -EOPNOTSUPP)
566 err = 0; /* Operation is not supported. */
567 else if (err) {
568 /*
569 * Try to delete anyway without
570 * error analysis.
571 */
572 hfsplus_delete_cat(sbi->hidden_dir->i_ino,
573 root, &str);
574 mutex_unlock(&sbi->vh_mutex);
575 goto out_put_hidden_dir;
576 }
577
578 mutex_unlock(&sbi->vh_mutex);
579 hfsplus_mark_inode_dirty(sbi->hidden_dir,
580 HFSPLUS_I_CAT_DIRTY);
581 }
582 }
583
584 unload_nls(sbi->nls);
585 sbi->nls = nls;
586 return 0;
587
588 out_put_hidden_dir:
589 iput(sbi->hidden_dir);
590 out_put_root:
591 dput(sb->s_root);
592 sb->s_root = NULL;
593 out_put_alloc_file:
594 iput(sbi->alloc_file);
595 out_close_attr_tree:
596 hfs_btree_close(sbi->attr_tree);
597 out_close_cat_tree:
598 hfs_btree_close(sbi->cat_tree);
599 out_close_ext_tree:
600 hfs_btree_close(sbi->ext_tree);
601 out_free_vhdr:
602 kfree(sbi->s_vhdr_buf);
603 kfree(sbi->s_backup_vhdr_buf);
604 out_unload_nls:
605 unload_nls(sbi->nls);
606 unload_nls(nls);
607 kfree(sbi);
608 out:
609 return err;
610 }
611
612 MODULE_AUTHOR("Brad Boyer");
613 MODULE_DESCRIPTION("Extended Macintosh Filesystem");
614 MODULE_LICENSE("GPL");
615
616 static struct kmem_cache *hfsplus_inode_cachep;
617
618 static struct inode *hfsplus_alloc_inode(struct super_block *sb)
619 {
620 struct hfsplus_inode_info *i;
621
622 i = kmem_cache_alloc(hfsplus_inode_cachep, GFP_KERNEL);
623 return i ? &i->vfs_inode : NULL;
624 }
625
626 static void hfsplus_i_callback(struct rcu_head *head)
627 {
628 struct inode *inode = container_of(head, struct inode, i_rcu);
629
630 kmem_cache_free(hfsplus_inode_cachep, HFSPLUS_I(inode));
631 }
632
633 static void hfsplus_destroy_inode(struct inode *inode)
634 {
635 call_rcu(&inode->i_rcu, hfsplus_i_callback);
636 }
637
638 #define HFSPLUS_INODE_SIZE sizeof(struct hfsplus_inode_info)
639
640 static struct dentry *hfsplus_mount(struct file_system_type *fs_type,
641 int flags, const char *dev_name, void *data)
642 {
643 return mount_bdev(fs_type, flags, dev_name, data, hfsplus_fill_super);
644 }
645
646 static struct file_system_type hfsplus_fs_type = {
647 .owner = THIS_MODULE,
648 .name = "hfsplus",
649 .mount = hfsplus_mount,
650 .kill_sb = kill_block_super,
651 .fs_flags = FS_REQUIRES_DEV,
652 };
653 MODULE_ALIAS_FS("hfsplus");
654
655 static void hfsplus_init_once(void *p)
656 {
657 struct hfsplus_inode_info *i = p;
658
659 inode_init_once(&i->vfs_inode);
660 }
661
662 static int __init init_hfsplus_fs(void)
663 {
664 int err;
665
666 hfsplus_inode_cachep = kmem_cache_create("hfsplus_icache",
667 HFSPLUS_INODE_SIZE, 0, SLAB_HWCACHE_ALIGN|SLAB_ACCOUNT,
668 hfsplus_init_once);
669 if (!hfsplus_inode_cachep)
670 return -ENOMEM;
671 err = hfsplus_create_attr_tree_cache();
672 if (err)
673 goto destroy_inode_cache;
674 err = register_filesystem(&hfsplus_fs_type);
675 if (err)
676 goto destroy_attr_tree_cache;
677 return 0;
678
679 destroy_attr_tree_cache:
680 hfsplus_destroy_attr_tree_cache();
681
682 destroy_inode_cache:
683 kmem_cache_destroy(hfsplus_inode_cachep);
684
685 return err;
686 }
687
688 static void __exit exit_hfsplus_fs(void)
689 {
690 unregister_filesystem(&hfsplus_fs_type);
691
692 /*
693 * Make sure all delayed rcu free inodes are flushed before we
694 * destroy cache.
695 */
696 rcu_barrier();
697 hfsplus_destroy_attr_tree_cache();
698 kmem_cache_destroy(hfsplus_inode_cachep);
699 }
700
701 module_init(init_hfsplus_fs)
702 module_exit(exit_hfsplus_fs)
This page took 0.082016 seconds and 5 git commands to generate.