mirror of https://github.com/Qortal/Brooklyn
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
689 lines
19 KiB
689 lines
19 KiB
// SPDX-License-Identifier: GPL-2.0 |
|
/* |
|
* linux/fs/hfsplus/inode.c |
|
* |
|
* Copyright (C) 2001 |
|
* Brad Boyer ([email protected]) |
|
* (C) 2003 Ardis Technologies <[email protected]> |
|
* |
|
* Inode handling routines |
|
*/ |
|
|
|
#include <linux/blkdev.h> |
|
#include <linux/mm.h> |
|
#include <linux/fs.h> |
|
#include <linux/pagemap.h> |
|
#include <linux/mpage.h> |
|
#include <linux/sched.h> |
|
#include <linux/cred.h> |
|
#include <linux/uio.h> |
|
#include <linux/fileattr.h> |
|
|
|
#include "hfsplus_fs.h" |
|
#include "hfsplus_raw.h" |
|
#include "xattr.h" |
|
|
|
static int hfsplus_read_folio(struct file *file, struct folio *folio) |
|
{ |
|
return block_read_full_folio(folio, hfsplus_get_block); |
|
} |
|
|
|
static int hfsplus_writepage(struct page *page, struct writeback_control *wbc) |
|
{ |
|
return block_write_full_page(page, hfsplus_get_block, wbc); |
|
} |
|
|
|
static void hfsplus_write_failed(struct address_space *mapping, loff_t to) |
|
{ |
|
struct inode *inode = mapping->host; |
|
|
|
if (to > inode->i_size) { |
|
truncate_pagecache(inode, inode->i_size); |
|
hfsplus_file_truncate(inode); |
|
} |
|
} |
|
|
|
int hfsplus_write_begin(struct file *file, struct address_space *mapping, |
|
loff_t pos, unsigned len, struct page **pagep, void **fsdata) |
|
{ |
|
int ret; |
|
|
|
*pagep = NULL; |
|
ret = cont_write_begin(file, mapping, pos, len, pagep, fsdata, |
|
hfsplus_get_block, |
|
&HFSPLUS_I(mapping->host)->phys_size); |
|
if (unlikely(ret)) |
|
hfsplus_write_failed(mapping, pos + len); |
|
|
|
return ret; |
|
} |
|
|
|
static sector_t hfsplus_bmap(struct address_space *mapping, sector_t block) |
|
{ |
|
return generic_block_bmap(mapping, block, hfsplus_get_block); |
|
} |
|
|
|
static bool hfsplus_release_folio(struct folio *folio, gfp_t mask) |
|
{ |
|
struct inode *inode = folio->mapping->host; |
|
struct super_block *sb = inode->i_sb; |
|
struct hfs_btree *tree; |
|
struct hfs_bnode *node; |
|
u32 nidx; |
|
int i; |
|
bool res = true; |
|
|
|
switch (inode->i_ino) { |
|
case HFSPLUS_EXT_CNID: |
|
tree = HFSPLUS_SB(sb)->ext_tree; |
|
break; |
|
case HFSPLUS_CAT_CNID: |
|
tree = HFSPLUS_SB(sb)->cat_tree; |
|
break; |
|
case HFSPLUS_ATTR_CNID: |
|
tree = HFSPLUS_SB(sb)->attr_tree; |
|
break; |
|
default: |
|
BUG(); |
|
return false; |
|
} |
|
if (!tree) |
|
return false; |
|
if (tree->node_size >= PAGE_SIZE) { |
|
nidx = folio->index >> |
|
(tree->node_size_shift - PAGE_SHIFT); |
|
spin_lock(&tree->hash_lock); |
|
node = hfs_bnode_findhash(tree, nidx); |
|
if (!node) |
|
; |
|
else if (atomic_read(&node->refcnt)) |
|
res = false; |
|
if (res && node) { |
|
hfs_bnode_unhash(node); |
|
hfs_bnode_free(node); |
|
} |
|
spin_unlock(&tree->hash_lock); |
|
} else { |
|
nidx = folio->index << |
|
(PAGE_SHIFT - tree->node_size_shift); |
|
i = 1 << (PAGE_SHIFT - tree->node_size_shift); |
|
spin_lock(&tree->hash_lock); |
|
do { |
|
node = hfs_bnode_findhash(tree, nidx++); |
|
if (!node) |
|
continue; |
|
if (atomic_read(&node->refcnt)) { |
|
res = false; |
|
break; |
|
} |
|
hfs_bnode_unhash(node); |
|
hfs_bnode_free(node); |
|
} while (--i && nidx < tree->node_count); |
|
spin_unlock(&tree->hash_lock); |
|
} |
|
return res ? try_to_free_buffers(folio) : false; |
|
} |
|
|
|
static ssize_t hfsplus_direct_IO(struct kiocb *iocb, struct iov_iter *iter) |
|
{ |
|
struct file *file = iocb->ki_filp; |
|
struct address_space *mapping = file->f_mapping; |
|
struct inode *inode = mapping->host; |
|
size_t count = iov_iter_count(iter); |
|
ssize_t ret; |
|
|
|
ret = blockdev_direct_IO(iocb, inode, iter, hfsplus_get_block); |
|
|
|
/* |
|
* In case of error extending write may have instantiated a few |
|
* blocks outside i_size. Trim these off again. |
|
*/ |
|
if (unlikely(iov_iter_rw(iter) == WRITE && ret < 0)) { |
|
loff_t isize = i_size_read(inode); |
|
loff_t end = iocb->ki_pos + count; |
|
|
|
if (end > isize) |
|
hfsplus_write_failed(mapping, end); |
|
} |
|
|
|
return ret; |
|
} |
|
|
|
static int hfsplus_writepages(struct address_space *mapping, |
|
struct writeback_control *wbc) |
|
{ |
|
return mpage_writepages(mapping, wbc, hfsplus_get_block); |
|
} |
|
|
|
const struct address_space_operations hfsplus_btree_aops = { |
|
.dirty_folio = block_dirty_folio, |
|
.invalidate_folio = block_invalidate_folio, |
|
.read_folio = hfsplus_read_folio, |
|
.writepage = hfsplus_writepage, |
|
.write_begin = hfsplus_write_begin, |
|
.write_end = generic_write_end, |
|
.bmap = hfsplus_bmap, |
|
.release_folio = hfsplus_release_folio, |
|
}; |
|
|
|
const struct address_space_operations hfsplus_aops = { |
|
.dirty_folio = block_dirty_folio, |
|
.invalidate_folio = block_invalidate_folio, |
|
.read_folio = hfsplus_read_folio, |
|
.writepage = hfsplus_writepage, |
|
.write_begin = hfsplus_write_begin, |
|
.write_end = generic_write_end, |
|
.bmap = hfsplus_bmap, |
|
.direct_IO = hfsplus_direct_IO, |
|
.writepages = hfsplus_writepages, |
|
}; |
|
|
|
const struct dentry_operations hfsplus_dentry_operations = { |
|
.d_hash = hfsplus_hash_dentry, |
|
.d_compare = hfsplus_compare_dentry, |
|
}; |
|
|
|
static void hfsplus_get_perms(struct inode *inode, |
|
struct hfsplus_perm *perms, int dir) |
|
{ |
|
struct hfsplus_sb_info *sbi = HFSPLUS_SB(inode->i_sb); |
|
u16 mode; |
|
|
|
mode = be16_to_cpu(perms->mode); |
|
|
|
i_uid_write(inode, be32_to_cpu(perms->owner)); |
|
if (!i_uid_read(inode) && !mode) |
|
inode->i_uid = sbi->uid; |
|
|
|
i_gid_write(inode, be32_to_cpu(perms->group)); |
|
if (!i_gid_read(inode) && !mode) |
|
inode->i_gid = sbi->gid; |
|
|
|
if (dir) { |
|
mode = mode ? (mode & S_IALLUGO) : (S_IRWXUGO & ~(sbi->umask)); |
|
mode |= S_IFDIR; |
|
} else if (!mode) |
|
mode = S_IFREG | ((S_IRUGO|S_IWUGO) & ~(sbi->umask)); |
|
inode->i_mode = mode; |
|
|
|
HFSPLUS_I(inode)->userflags = perms->userflags; |
|
if (perms->rootflags & HFSPLUS_FLG_IMMUTABLE) |
|
inode->i_flags |= S_IMMUTABLE; |
|
else |
|
inode->i_flags &= ~S_IMMUTABLE; |
|
if (perms->rootflags & HFSPLUS_FLG_APPEND) |
|
inode->i_flags |= S_APPEND; |
|
else |
|
inode->i_flags &= ~S_APPEND; |
|
} |
|
|
|
static int hfsplus_file_open(struct inode *inode, struct file *file) |
|
{ |
|
if (HFSPLUS_IS_RSRC(inode)) |
|
inode = HFSPLUS_I(inode)->rsrc_inode; |
|
if (!(file->f_flags & O_LARGEFILE) && i_size_read(inode) > MAX_NON_LFS) |
|
return -EOVERFLOW; |
|
atomic_inc(&HFSPLUS_I(inode)->opencnt); |
|
return 0; |
|
} |
|
|
|
static int hfsplus_file_release(struct inode *inode, struct file *file) |
|
{ |
|
struct super_block *sb = inode->i_sb; |
|
|
|
if (HFSPLUS_IS_RSRC(inode)) |
|
inode = HFSPLUS_I(inode)->rsrc_inode; |
|
if (atomic_dec_and_test(&HFSPLUS_I(inode)->opencnt)) { |
|
inode_lock(inode); |
|
hfsplus_file_truncate(inode); |
|
if (inode->i_flags & S_DEAD) { |
|
hfsplus_delete_cat(inode->i_ino, |
|
HFSPLUS_SB(sb)->hidden_dir, NULL); |
|
hfsplus_delete_inode(inode); |
|
} |
|
inode_unlock(inode); |
|
} |
|
return 0; |
|
} |
|
|
|
static int hfsplus_setattr(struct user_namespace *mnt_userns, |
|
struct dentry *dentry, struct iattr *attr) |
|
{ |
|
struct inode *inode = d_inode(dentry); |
|
int error; |
|
|
|
error = setattr_prepare(&init_user_ns, dentry, attr); |
|
if (error) |
|
return error; |
|
|
|
if ((attr->ia_valid & ATTR_SIZE) && |
|
attr->ia_size != i_size_read(inode)) { |
|
inode_dio_wait(inode); |
|
if (attr->ia_size > inode->i_size) { |
|
error = generic_cont_expand_simple(inode, |
|
attr->ia_size); |
|
if (error) |
|
return error; |
|
} |
|
truncate_setsize(inode, attr->ia_size); |
|
hfsplus_file_truncate(inode); |
|
inode->i_mtime = inode->i_ctime = current_time(inode); |
|
} |
|
|
|
setattr_copy(&init_user_ns, inode, attr); |
|
mark_inode_dirty(inode); |
|
|
|
return 0; |
|
} |
|
|
|
int hfsplus_getattr(struct user_namespace *mnt_userns, const struct path *path, |
|
struct kstat *stat, u32 request_mask, |
|
unsigned int query_flags) |
|
{ |
|
struct inode *inode = d_inode(path->dentry); |
|
struct hfsplus_inode_info *hip = HFSPLUS_I(inode); |
|
|
|
if (request_mask & STATX_BTIME) { |
|
stat->result_mask |= STATX_BTIME; |
|
stat->btime = hfsp_mt2ut(hip->create_date); |
|
} |
|
|
|
if (inode->i_flags & S_APPEND) |
|
stat->attributes |= STATX_ATTR_APPEND; |
|
if (inode->i_flags & S_IMMUTABLE) |
|
stat->attributes |= STATX_ATTR_IMMUTABLE; |
|
if (hip->userflags & HFSPLUS_FLG_NODUMP) |
|
stat->attributes |= STATX_ATTR_NODUMP; |
|
|
|
stat->attributes_mask |= STATX_ATTR_APPEND | STATX_ATTR_IMMUTABLE | |
|
STATX_ATTR_NODUMP; |
|
|
|
generic_fillattr(&init_user_ns, inode, stat); |
|
return 0; |
|
} |
|
|
|
int hfsplus_file_fsync(struct file *file, loff_t start, loff_t end, |
|
int datasync) |
|
{ |
|
struct inode *inode = file->f_mapping->host; |
|
struct hfsplus_inode_info *hip = HFSPLUS_I(inode); |
|
struct hfsplus_sb_info *sbi = HFSPLUS_SB(inode->i_sb); |
|
int error = 0, error2; |
|
|
|
error = file_write_and_wait_range(file, start, end); |
|
if (error) |
|
return error; |
|
inode_lock(inode); |
|
|
|
/* |
|
* Sync inode metadata into the catalog and extent trees. |
|
*/ |
|
sync_inode_metadata(inode, 1); |
|
|
|
/* |
|
* And explicitly write out the btrees. |
|
*/ |
|
if (test_and_clear_bit(HFSPLUS_I_CAT_DIRTY, &hip->flags)) |
|
error = filemap_write_and_wait(sbi->cat_tree->inode->i_mapping); |
|
|
|
if (test_and_clear_bit(HFSPLUS_I_EXT_DIRTY, &hip->flags)) { |
|
error2 = |
|
filemap_write_and_wait(sbi->ext_tree->inode->i_mapping); |
|
if (!error) |
|
error = error2; |
|
} |
|
|
|
if (test_and_clear_bit(HFSPLUS_I_ATTR_DIRTY, &hip->flags)) { |
|
if (sbi->attr_tree) { |
|
error2 = |
|
filemap_write_and_wait( |
|
sbi->attr_tree->inode->i_mapping); |
|
if (!error) |
|
error = error2; |
|
} else { |
|
pr_err("sync non-existent attributes tree\n"); |
|
} |
|
} |
|
|
|
if (test_and_clear_bit(HFSPLUS_I_ALLOC_DIRTY, &hip->flags)) { |
|
error2 = filemap_write_and_wait(sbi->alloc_file->i_mapping); |
|
if (!error) |
|
error = error2; |
|
} |
|
|
|
if (!test_bit(HFSPLUS_SB_NOBARRIER, &sbi->flags)) |
|
blkdev_issue_flush(inode->i_sb->s_bdev); |
|
|
|
inode_unlock(inode); |
|
|
|
return error; |
|
} |
|
|
|
static const struct inode_operations hfsplus_file_inode_operations = { |
|
.setattr = hfsplus_setattr, |
|
.getattr = hfsplus_getattr, |
|
.listxattr = hfsplus_listxattr, |
|
.fileattr_get = hfsplus_fileattr_get, |
|
.fileattr_set = hfsplus_fileattr_set, |
|
}; |
|
|
|
static const struct file_operations hfsplus_file_operations = { |
|
.llseek = generic_file_llseek, |
|
.read_iter = generic_file_read_iter, |
|
.write_iter = generic_file_write_iter, |
|
.mmap = generic_file_mmap, |
|
.splice_read = generic_file_splice_read, |
|
.fsync = hfsplus_file_fsync, |
|
.open = hfsplus_file_open, |
|
.release = hfsplus_file_release, |
|
.unlocked_ioctl = hfsplus_ioctl, |
|
}; |
|
|
|
struct inode *hfsplus_new_inode(struct super_block *sb, struct inode *dir, |
|
umode_t mode) |
|
{ |
|
struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb); |
|
struct inode *inode = new_inode(sb); |
|
struct hfsplus_inode_info *hip; |
|
|
|
if (!inode) |
|
return NULL; |
|
|
|
inode->i_ino = sbi->next_cnid++; |
|
inode_init_owner(&init_user_ns, inode, dir, mode); |
|
set_nlink(inode, 1); |
|
inode->i_mtime = inode->i_atime = inode->i_ctime = current_time(inode); |
|
|
|
hip = HFSPLUS_I(inode); |
|
INIT_LIST_HEAD(&hip->open_dir_list); |
|
spin_lock_init(&hip->open_dir_lock); |
|
mutex_init(&hip->extents_lock); |
|
atomic_set(&hip->opencnt, 0); |
|
hip->extent_state = 0; |
|
hip->flags = 0; |
|
hip->userflags = 0; |
|
hip->subfolders = 0; |
|
memset(hip->first_extents, 0, sizeof(hfsplus_extent_rec)); |
|
memset(hip->cached_extents, 0, sizeof(hfsplus_extent_rec)); |
|
hip->alloc_blocks = 0; |
|
hip->first_blocks = 0; |
|
hip->cached_start = 0; |
|
hip->cached_blocks = 0; |
|
hip->phys_size = 0; |
|
hip->fs_blocks = 0; |
|
hip->rsrc_inode = NULL; |
|
if (S_ISDIR(inode->i_mode)) { |
|
inode->i_size = 2; |
|
sbi->folder_count++; |
|
inode->i_op = &hfsplus_dir_inode_operations; |
|
inode->i_fop = &hfsplus_dir_operations; |
|
} else if (S_ISREG(inode->i_mode)) { |
|
sbi->file_count++; |
|
inode->i_op = &hfsplus_file_inode_operations; |
|
inode->i_fop = &hfsplus_file_operations; |
|
inode->i_mapping->a_ops = &hfsplus_aops; |
|
hip->clump_blocks = sbi->data_clump_blocks; |
|
} else if (S_ISLNK(inode->i_mode)) { |
|
sbi->file_count++; |
|
inode->i_op = &page_symlink_inode_operations; |
|
inode_nohighmem(inode); |
|
inode->i_mapping->a_ops = &hfsplus_aops; |
|
hip->clump_blocks = 1; |
|
} else |
|
sbi->file_count++; |
|
insert_inode_hash(inode); |
|
mark_inode_dirty(inode); |
|
hfsplus_mark_mdb_dirty(sb); |
|
|
|
return inode; |
|
} |
|
|
|
void hfsplus_delete_inode(struct inode *inode) |
|
{ |
|
struct super_block *sb = inode->i_sb; |
|
|
|
if (S_ISDIR(inode->i_mode)) { |
|
HFSPLUS_SB(sb)->folder_count--; |
|
hfsplus_mark_mdb_dirty(sb); |
|
return; |
|
} |
|
HFSPLUS_SB(sb)->file_count--; |
|
if (S_ISREG(inode->i_mode)) { |
|
if (!inode->i_nlink) { |
|
inode->i_size = 0; |
|
hfsplus_file_truncate(inode); |
|
} |
|
} else if (S_ISLNK(inode->i_mode)) { |
|
inode->i_size = 0; |
|
hfsplus_file_truncate(inode); |
|
} |
|
hfsplus_mark_mdb_dirty(sb); |
|
} |
|
|
|
void hfsplus_inode_read_fork(struct inode *inode, struct hfsplus_fork_raw *fork) |
|
{ |
|
struct super_block *sb = inode->i_sb; |
|
struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb); |
|
struct hfsplus_inode_info *hip = HFSPLUS_I(inode); |
|
u32 count; |
|
int i; |
|
|
|
memcpy(&hip->first_extents, &fork->extents, sizeof(hfsplus_extent_rec)); |
|
for (count = 0, i = 0; i < 8; i++) |
|
count += be32_to_cpu(fork->extents[i].block_count); |
|
hip->first_blocks = count; |
|
memset(hip->cached_extents, 0, sizeof(hfsplus_extent_rec)); |
|
hip->cached_start = 0; |
|
hip->cached_blocks = 0; |
|
|
|
hip->alloc_blocks = be32_to_cpu(fork->total_blocks); |
|
hip->phys_size = inode->i_size = be64_to_cpu(fork->total_size); |
|
hip->fs_blocks = |
|
(inode->i_size + sb->s_blocksize - 1) >> sb->s_blocksize_bits; |
|
inode_set_bytes(inode, hip->fs_blocks << sb->s_blocksize_bits); |
|
hip->clump_blocks = |
|
be32_to_cpu(fork->clump_size) >> sbi->alloc_blksz_shift; |
|
if (!hip->clump_blocks) { |
|
hip->clump_blocks = HFSPLUS_IS_RSRC(inode) ? |
|
sbi->rsrc_clump_blocks : |
|
sbi->data_clump_blocks; |
|
} |
|
} |
|
|
|
void hfsplus_inode_write_fork(struct inode *inode, |
|
struct hfsplus_fork_raw *fork) |
|
{ |
|
memcpy(&fork->extents, &HFSPLUS_I(inode)->first_extents, |
|
sizeof(hfsplus_extent_rec)); |
|
fork->total_size = cpu_to_be64(inode->i_size); |
|
fork->total_blocks = cpu_to_be32(HFSPLUS_I(inode)->alloc_blocks); |
|
} |
|
|
|
int hfsplus_cat_read_inode(struct inode *inode, struct hfs_find_data *fd) |
|
{ |
|
hfsplus_cat_entry entry; |
|
int res = 0; |
|
u16 type; |
|
|
|
type = hfs_bnode_read_u16(fd->bnode, fd->entryoffset); |
|
|
|
HFSPLUS_I(inode)->linkid = 0; |
|
if (type == HFSPLUS_FOLDER) { |
|
struct hfsplus_cat_folder *folder = &entry.folder; |
|
|
|
WARN_ON(fd->entrylength < sizeof(struct hfsplus_cat_folder)); |
|
hfs_bnode_read(fd->bnode, &entry, fd->entryoffset, |
|
sizeof(struct hfsplus_cat_folder)); |
|
hfsplus_get_perms(inode, &folder->permissions, 1); |
|
set_nlink(inode, 1); |
|
inode->i_size = 2 + be32_to_cpu(folder->valence); |
|
inode->i_atime = hfsp_mt2ut(folder->access_date); |
|
inode->i_mtime = hfsp_mt2ut(folder->content_mod_date); |
|
inode->i_ctime = hfsp_mt2ut(folder->attribute_mod_date); |
|
HFSPLUS_I(inode)->create_date = folder->create_date; |
|
HFSPLUS_I(inode)->fs_blocks = 0; |
|
if (folder->flags & cpu_to_be16(HFSPLUS_HAS_FOLDER_COUNT)) { |
|
HFSPLUS_I(inode)->subfolders = |
|
be32_to_cpu(folder->subfolders); |
|
} |
|
inode->i_op = &hfsplus_dir_inode_operations; |
|
inode->i_fop = &hfsplus_dir_operations; |
|
} else if (type == HFSPLUS_FILE) { |
|
struct hfsplus_cat_file *file = &entry.file; |
|
|
|
WARN_ON(fd->entrylength < sizeof(struct hfsplus_cat_file)); |
|
hfs_bnode_read(fd->bnode, &entry, fd->entryoffset, |
|
sizeof(struct hfsplus_cat_file)); |
|
|
|
hfsplus_inode_read_fork(inode, HFSPLUS_IS_RSRC(inode) ? |
|
&file->rsrc_fork : &file->data_fork); |
|
hfsplus_get_perms(inode, &file->permissions, 0); |
|
set_nlink(inode, 1); |
|
if (S_ISREG(inode->i_mode)) { |
|
if (file->permissions.dev) |
|
set_nlink(inode, |
|
be32_to_cpu(file->permissions.dev)); |
|
inode->i_op = &hfsplus_file_inode_operations; |
|
inode->i_fop = &hfsplus_file_operations; |
|
inode->i_mapping->a_ops = &hfsplus_aops; |
|
} else if (S_ISLNK(inode->i_mode)) { |
|
inode->i_op = &page_symlink_inode_operations; |
|
inode_nohighmem(inode); |
|
inode->i_mapping->a_ops = &hfsplus_aops; |
|
} else { |
|
init_special_inode(inode, inode->i_mode, |
|
be32_to_cpu(file->permissions.dev)); |
|
} |
|
inode->i_atime = hfsp_mt2ut(file->access_date); |
|
inode->i_mtime = hfsp_mt2ut(file->content_mod_date); |
|
inode->i_ctime = hfsp_mt2ut(file->attribute_mod_date); |
|
HFSPLUS_I(inode)->create_date = file->create_date; |
|
} else { |
|
pr_err("bad catalog entry used to create inode\n"); |
|
res = -EIO; |
|
} |
|
return res; |
|
} |
|
|
|
int hfsplus_cat_write_inode(struct inode *inode) |
|
{ |
|
struct inode *main_inode = inode; |
|
struct hfs_find_data fd; |
|
hfsplus_cat_entry entry; |
|
|
|
if (HFSPLUS_IS_RSRC(inode)) |
|
main_inode = HFSPLUS_I(inode)->rsrc_inode; |
|
|
|
if (!main_inode->i_nlink) |
|
return 0; |
|
|
|
if (hfs_find_init(HFSPLUS_SB(main_inode->i_sb)->cat_tree, &fd)) |
|
/* panic? */ |
|
return -EIO; |
|
|
|
if (hfsplus_find_cat(main_inode->i_sb, main_inode->i_ino, &fd)) |
|
/* panic? */ |
|
goto out; |
|
|
|
if (S_ISDIR(main_inode->i_mode)) { |
|
struct hfsplus_cat_folder *folder = &entry.folder; |
|
|
|
WARN_ON(fd.entrylength < sizeof(struct hfsplus_cat_folder)); |
|
hfs_bnode_read(fd.bnode, &entry, fd.entryoffset, |
|
sizeof(struct hfsplus_cat_folder)); |
|
/* simple node checks? */ |
|
hfsplus_cat_set_perms(inode, &folder->permissions); |
|
folder->access_date = hfsp_ut2mt(inode->i_atime); |
|
folder->content_mod_date = hfsp_ut2mt(inode->i_mtime); |
|
folder->attribute_mod_date = hfsp_ut2mt(inode->i_ctime); |
|
folder->valence = cpu_to_be32(inode->i_size - 2); |
|
if (folder->flags & cpu_to_be16(HFSPLUS_HAS_FOLDER_COUNT)) { |
|
folder->subfolders = |
|
cpu_to_be32(HFSPLUS_I(inode)->subfolders); |
|
} |
|
hfs_bnode_write(fd.bnode, &entry, fd.entryoffset, |
|
sizeof(struct hfsplus_cat_folder)); |
|
} else if (HFSPLUS_IS_RSRC(inode)) { |
|
struct hfsplus_cat_file *file = &entry.file; |
|
hfs_bnode_read(fd.bnode, &entry, fd.entryoffset, |
|
sizeof(struct hfsplus_cat_file)); |
|
hfsplus_inode_write_fork(inode, &file->rsrc_fork); |
|
hfs_bnode_write(fd.bnode, &entry, fd.entryoffset, |
|
sizeof(struct hfsplus_cat_file)); |
|
} else { |
|
struct hfsplus_cat_file *file = &entry.file; |
|
|
|
WARN_ON(fd.entrylength < sizeof(struct hfsplus_cat_file)); |
|
hfs_bnode_read(fd.bnode, &entry, fd.entryoffset, |
|
sizeof(struct hfsplus_cat_file)); |
|
hfsplus_inode_write_fork(inode, &file->data_fork); |
|
hfsplus_cat_set_perms(inode, &file->permissions); |
|
if (HFSPLUS_FLG_IMMUTABLE & |
|
(file->permissions.rootflags | |
|
file->permissions.userflags)) |
|
file->flags |= cpu_to_be16(HFSPLUS_FILE_LOCKED); |
|
else |
|
file->flags &= cpu_to_be16(~HFSPLUS_FILE_LOCKED); |
|
file->access_date = hfsp_ut2mt(inode->i_atime); |
|
file->content_mod_date = hfsp_ut2mt(inode->i_mtime); |
|
file->attribute_mod_date = hfsp_ut2mt(inode->i_ctime); |
|
hfs_bnode_write(fd.bnode, &entry, fd.entryoffset, |
|
sizeof(struct hfsplus_cat_file)); |
|
} |
|
|
|
set_bit(HFSPLUS_I_CAT_DIRTY, &HFSPLUS_I(inode)->flags); |
|
out: |
|
hfs_find_exit(&fd); |
|
return 0; |
|
} |
|
|
|
int hfsplus_fileattr_get(struct dentry *dentry, struct fileattr *fa) |
|
{ |
|
struct inode *inode = d_inode(dentry); |
|
struct hfsplus_inode_info *hip = HFSPLUS_I(inode); |
|
unsigned int flags = 0; |
|
|
|
if (inode->i_flags & S_IMMUTABLE) |
|
flags |= FS_IMMUTABLE_FL; |
|
if (inode->i_flags & S_APPEND) |
|
flags |= FS_APPEND_FL; |
|
if (hip->userflags & HFSPLUS_FLG_NODUMP) |
|
flags |= FS_NODUMP_FL; |
|
|
|
fileattr_fill_flags(fa, flags); |
|
|
|
return 0; |
|
} |
|
|
|
int hfsplus_fileattr_set(struct user_namespace *mnt_userns, |
|
struct dentry *dentry, struct fileattr *fa) |
|
{ |
|
struct inode *inode = d_inode(dentry); |
|
struct hfsplus_inode_info *hip = HFSPLUS_I(inode); |
|
unsigned int new_fl = 0; |
|
|
|
if (fileattr_has_fsx(fa)) |
|
return -EOPNOTSUPP; |
|
|
|
/* don't silently ignore unsupported ext2 flags */ |
|
if (fa->flags & ~(FS_IMMUTABLE_FL|FS_APPEND_FL|FS_NODUMP_FL)) |
|
return -EOPNOTSUPP; |
|
|
|
if (fa->flags & FS_IMMUTABLE_FL) |
|
new_fl |= S_IMMUTABLE; |
|
|
|
if (fa->flags & FS_APPEND_FL) |
|
new_fl |= S_APPEND; |
|
|
|
inode_set_flags(inode, new_fl, S_IMMUTABLE | S_APPEND); |
|
|
|
if (fa->flags & FS_NODUMP_FL) |
|
hip->userflags |= HFSPLUS_FLG_NODUMP; |
|
else |
|
hip->userflags &= ~HFSPLUS_FLG_NODUMP; |
|
|
|
inode->i_ctime = current_time(inode); |
|
mark_inode_dirty(inode); |
|
|
|
return 0; |
|
}
|
|
|