mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-12-27 08:45:26 -05:00
Pull vfs inode updates from Christian Brauner:
"Features:
- Hide inode->i_state behind accessors. Open-coded accesses prevent
asserting they are done correctly. One obvious aspect is locking,
but significantly more can be checked. For example it can be
detected when the code is clearing flags which are already missing,
or is setting flags when it is illegal (e.g., I_FREEING when
->i_count > 0)
- Provide accessors for ->i_state, converts all filesystems using
coccinelle and manual conversions (btrfs, ceph, smb, f2fs, gfs2,
overlayfs, nilfs2, xfs), and makes plain ->i_state access fail to
compile
- Rework I_NEW handling to operate without fences, simplifying the
code after the accessor infrastructure is in place
Cleanups:
- Move wait_on_inode() from writeback.h to fs.h
- Spell out fenced ->i_state accesses with explicit smp_wmb/smp_rmb
for clarity
- Cosmetic fixes to LRU handling
- Push list presence check into inode_io_list_del()
- Touch up predicts in __d_lookup_rcu()
- ocfs2: retire ocfs2_drop_inode() and I_WILL_FREE usage
- Assert on ->i_count in iput_final()
- Assert ->i_lock held in __iget()
Fixes:
- Add missing fences to I_NEW handling"
* tag 'vfs-6.19-rc1.inode' of git://git.kernel.org/pub/scm/linux/kernel/git/vfs/vfs: (22 commits)
dcache: touch up predicts in __d_lookup_rcu()
fs: push list presence check into inode_io_list_del()
fs: cosmetic fixes to lru handling
fs: rework I_NEW handling to operate without fences
fs: make plain ->i_state access fail to compile
xfs: use the new ->i_state accessors
nilfs2: use the new ->i_state accessors
overlayfs: use the new ->i_state accessors
gfs2: use the new ->i_state accessors
f2fs: use the new ->i_state accessors
smb: use the new ->i_state accessors
ceph: use the new ->i_state accessors
btrfs: use the new ->i_state accessors
Manual conversion to use ->i_state accessors of all places not covered by coccinelle
Coccinelle-based conversion to use ->i_state accessors
fs: provide accessors for ->i_state
fs: spell out fenced ->i_state accesses with explicit smp_wmb/smp_rmb
fs: move wait_on_inode() from writeback.h to fs.h
fs: add missing fences to I_NEW handling
ocfs2: retire ocfs2_drop_inode() and I_WILL_FREE usage
...
1145 lines
29 KiB
C
1145 lines
29 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* (C) 2001 Clemson University and The University of Chicago
|
|
* Copyright 2018 Omnibond Systems, L.L.C.
|
|
*
|
|
* See COPYING in top-level directory.
|
|
*/
|
|
|
|
/*
|
|
* Linux VFS inode operations.
|
|
*/
|
|
|
|
#include <linux/blkdev.h>
|
|
#include <linux/fileattr.h>
|
|
#include "protocol.h"
|
|
#include "orangefs-kernel.h"
|
|
#include "orangefs-bufmap.h"
|
|
|
|
static int orangefs_writepage_locked(struct folio *folio,
|
|
struct writeback_control *wbc)
|
|
{
|
|
struct inode *inode = folio->mapping->host;
|
|
struct orangefs_write_range *wr = NULL;
|
|
struct iov_iter iter;
|
|
struct bio_vec bv;
|
|
size_t wlen;
|
|
ssize_t ret;
|
|
loff_t len, off;
|
|
|
|
folio_start_writeback(folio);
|
|
|
|
len = i_size_read(inode);
|
|
if (folio->private) {
|
|
wr = folio->private;
|
|
off = wr->pos;
|
|
if ((off + wr->len > len) && (off <= len))
|
|
wlen = len - off;
|
|
else
|
|
wlen = wr->len;
|
|
if (wlen == 0)
|
|
wlen = wr->len;
|
|
} else {
|
|
WARN_ON(1);
|
|
off = folio_pos(folio);
|
|
wlen = folio_size(folio);
|
|
|
|
if (wlen > len - off)
|
|
wlen = len - off;
|
|
}
|
|
|
|
WARN_ON(wlen == 0);
|
|
bvec_set_folio(&bv, folio, wlen, offset_in_folio(folio, off));
|
|
iov_iter_bvec(&iter, ITER_SOURCE, &bv, 1, wlen);
|
|
|
|
ret = wait_for_direct_io(ORANGEFS_IO_WRITE, inode, &off, &iter, wlen,
|
|
len, wr, NULL, NULL);
|
|
if (ret < 0) {
|
|
mapping_set_error(folio->mapping, ret);
|
|
} else {
|
|
ret = 0;
|
|
}
|
|
kfree(folio_detach_private(folio));
|
|
return ret;
|
|
}
|
|
|
|
struct orangefs_writepages {
|
|
loff_t off;
|
|
size_t len;
|
|
kuid_t uid;
|
|
kgid_t gid;
|
|
int maxpages;
|
|
int nfolios;
|
|
struct address_space *mapping;
|
|
struct folio **folios;
|
|
struct bio_vec *bv;
|
|
};
|
|
|
|
static int orangefs_writepages_work(struct orangefs_writepages *ow,
|
|
struct writeback_control *wbc)
|
|
{
|
|
struct inode *inode = ow->mapping->host;
|
|
struct orangefs_write_range *wrp, wr;
|
|
struct iov_iter iter;
|
|
ssize_t ret;
|
|
size_t start;
|
|
loff_t len, off;
|
|
int i;
|
|
|
|
len = i_size_read(inode);
|
|
|
|
start = offset_in_folio(ow->folios[0], ow->off);
|
|
for (i = 0; i < ow->nfolios; i++) {
|
|
folio_start_writeback(ow->folios[i]);
|
|
bvec_set_folio(&ow->bv[i], ow->folios[i],
|
|
folio_size(ow->folios[i]) - start, start);
|
|
start = 0;
|
|
}
|
|
iov_iter_bvec(&iter, ITER_SOURCE, ow->bv, ow->nfolios, ow->len);
|
|
|
|
WARN_ON(ow->off >= len);
|
|
if (ow->off + ow->len > len)
|
|
ow->len = len - ow->off;
|
|
|
|
off = ow->off;
|
|
wr.uid = ow->uid;
|
|
wr.gid = ow->gid;
|
|
ret = wait_for_direct_io(ORANGEFS_IO_WRITE, inode, &off, &iter, ow->len,
|
|
0, &wr, NULL, NULL);
|
|
if (ret < 0)
|
|
mapping_set_error(ow->mapping, ret);
|
|
else
|
|
ret = 0;
|
|
|
|
for (i = 0; i < ow->nfolios; i++) {
|
|
wrp = folio_detach_private(ow->folios[i]);
|
|
kfree(wrp);
|
|
folio_end_writeback(ow->folios[i]);
|
|
folio_unlock(ow->folios[i]);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int orangefs_writepages_callback(struct folio *folio,
|
|
struct writeback_control *wbc, struct orangefs_writepages *ow)
|
|
{
|
|
struct orangefs_write_range *wr = folio->private;
|
|
int ret;
|
|
|
|
if (!wr) {
|
|
folio_unlock(folio);
|
|
/* It's not private so there's nothing to write, right? */
|
|
printk("writepages_callback not private!\n");
|
|
BUG();
|
|
return 0;
|
|
}
|
|
|
|
ret = -1;
|
|
if (ow->nfolios == 0) {
|
|
ow->off = wr->pos;
|
|
ow->len = wr->len;
|
|
ow->uid = wr->uid;
|
|
ow->gid = wr->gid;
|
|
ow->folios[ow->nfolios++] = folio;
|
|
ret = 0;
|
|
goto done;
|
|
}
|
|
if (!uid_eq(ow->uid, wr->uid) || !gid_eq(ow->gid, wr->gid)) {
|
|
orangefs_writepages_work(ow, wbc);
|
|
ow->nfolios = 0;
|
|
ret = -1;
|
|
goto done;
|
|
}
|
|
if (ow->off + ow->len == wr->pos) {
|
|
ow->len += wr->len;
|
|
ow->folios[ow->nfolios++] = folio;
|
|
ret = 0;
|
|
goto done;
|
|
}
|
|
done:
|
|
if (ret == -1) {
|
|
if (ow->nfolios) {
|
|
orangefs_writepages_work(ow, wbc);
|
|
ow->nfolios = 0;
|
|
}
|
|
ret = orangefs_writepage_locked(folio, wbc);
|
|
mapping_set_error(folio->mapping, ret);
|
|
folio_unlock(folio);
|
|
folio_end_writeback(folio);
|
|
} else {
|
|
if (ow->nfolios == ow->maxpages) {
|
|
orangefs_writepages_work(ow, wbc);
|
|
ow->nfolios = 0;
|
|
}
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static int orangefs_writepages(struct address_space *mapping,
|
|
struct writeback_control *wbc)
|
|
{
|
|
struct orangefs_writepages *ow;
|
|
struct blk_plug plug;
|
|
int error;
|
|
struct folio *folio = NULL;
|
|
|
|
ow = kzalloc(sizeof(struct orangefs_writepages), GFP_KERNEL);
|
|
if (!ow)
|
|
return -ENOMEM;
|
|
ow->maxpages = orangefs_bufmap_size_query()/PAGE_SIZE;
|
|
ow->folios = kcalloc(ow->maxpages, sizeof(struct folio *), GFP_KERNEL);
|
|
if (!ow->folios) {
|
|
kfree(ow);
|
|
return -ENOMEM;
|
|
}
|
|
ow->bv = kcalloc(ow->maxpages, sizeof(struct bio_vec), GFP_KERNEL);
|
|
if (!ow->bv) {
|
|
kfree(ow->folios);
|
|
kfree(ow);
|
|
return -ENOMEM;
|
|
}
|
|
ow->mapping = mapping;
|
|
blk_start_plug(&plug);
|
|
while ((folio = writeback_iter(mapping, wbc, folio, &error)))
|
|
error = orangefs_writepages_callback(folio, wbc, ow);
|
|
if (ow->nfolios)
|
|
error = orangefs_writepages_work(ow, wbc);
|
|
blk_finish_plug(&plug);
|
|
kfree(ow->folios);
|
|
kfree(ow->bv);
|
|
kfree(ow);
|
|
return error;
|
|
}
|
|
|
|
static int orangefs_launder_folio(struct folio *);
|
|
|
|
static void orangefs_readahead(struct readahead_control *rac)
|
|
{
|
|
loff_t offset;
|
|
struct iov_iter iter;
|
|
struct inode *inode = rac->mapping->host;
|
|
struct xarray *i_pages;
|
|
struct folio *folio;
|
|
loff_t new_start = readahead_pos(rac);
|
|
int ret;
|
|
size_t new_len = 0;
|
|
|
|
loff_t bytes_remaining = inode->i_size - readahead_pos(rac);
|
|
loff_t pages_remaining = bytes_remaining / PAGE_SIZE;
|
|
|
|
if (pages_remaining >= 1024)
|
|
new_len = 4194304;
|
|
else if (pages_remaining > readahead_count(rac))
|
|
new_len = bytes_remaining;
|
|
|
|
if (new_len)
|
|
readahead_expand(rac, new_start, new_len);
|
|
|
|
offset = readahead_pos(rac);
|
|
i_pages = &rac->mapping->i_pages;
|
|
|
|
iov_iter_xarray(&iter, ITER_DEST, i_pages, offset, readahead_length(rac));
|
|
|
|
/* read in the pages. */
|
|
if ((ret = wait_for_direct_io(ORANGEFS_IO_READ, inode,
|
|
&offset, &iter, readahead_length(rac),
|
|
inode->i_size, NULL, NULL, rac->file)) < 0)
|
|
gossip_debug(GOSSIP_FILE_DEBUG,
|
|
"%s: wait_for_direct_io failed. \n", __func__);
|
|
else
|
|
ret = 0;
|
|
|
|
/* clean up. */
|
|
while ((folio = readahead_folio(rac))) {
|
|
if (!ret)
|
|
folio_mark_uptodate(folio);
|
|
folio_unlock(folio);
|
|
}
|
|
}
|
|
|
|
static int orangefs_read_folio(struct file *file, struct folio *folio)
|
|
{
|
|
struct inode *inode = folio->mapping->host;
|
|
struct iov_iter iter;
|
|
struct bio_vec bv;
|
|
ssize_t ret;
|
|
loff_t off; /* offset of this folio in the file */
|
|
|
|
if (folio_test_dirty(folio))
|
|
orangefs_launder_folio(folio);
|
|
|
|
off = folio_pos(folio);
|
|
bvec_set_folio(&bv, folio, folio_size(folio), 0);
|
|
iov_iter_bvec(&iter, ITER_DEST, &bv, 1, folio_size(folio));
|
|
|
|
ret = wait_for_direct_io(ORANGEFS_IO_READ, inode, &off, &iter,
|
|
folio_size(folio), inode->i_size, NULL, NULL, file);
|
|
/* this will only zero remaining unread portions of the folio data */
|
|
iov_iter_zero(~0U, &iter);
|
|
/* takes care of potential aliasing */
|
|
flush_dcache_folio(folio);
|
|
if (ret > 0)
|
|
ret = 0;
|
|
folio_end_read(folio, ret == 0);
|
|
return ret;
|
|
}
|
|
|
|
static int orangefs_write_begin(const struct kiocb *iocb,
|
|
struct address_space *mapping, loff_t pos,
|
|
unsigned len, struct folio **foliop,
|
|
void **fsdata)
|
|
{
|
|
struct orangefs_write_range *wr;
|
|
struct folio *folio;
|
|
int ret;
|
|
|
|
folio = __filemap_get_folio(mapping, pos / PAGE_SIZE, FGP_WRITEBEGIN,
|
|
mapping_gfp_mask(mapping));
|
|
if (IS_ERR(folio))
|
|
return PTR_ERR(folio);
|
|
|
|
*foliop = folio;
|
|
|
|
if (folio_test_dirty(folio) && !folio_test_private(folio)) {
|
|
/*
|
|
* Should be impossible. If it happens, launder the page
|
|
* since we don't know what's dirty. This will WARN in
|
|
* orangefs_writepage_locked.
|
|
*/
|
|
ret = orangefs_launder_folio(folio);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
if (folio_test_private(folio)) {
|
|
struct orangefs_write_range *wr;
|
|
wr = folio_get_private(folio);
|
|
if (wr->pos + wr->len == pos &&
|
|
uid_eq(wr->uid, current_fsuid()) &&
|
|
gid_eq(wr->gid, current_fsgid())) {
|
|
wr->len += len;
|
|
goto okay;
|
|
} else {
|
|
wr->pos = pos;
|
|
wr->len = len;
|
|
ret = orangefs_launder_folio(folio);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
wr = kmalloc(sizeof *wr, GFP_KERNEL);
|
|
if (!wr)
|
|
return -ENOMEM;
|
|
|
|
wr->pos = pos;
|
|
wr->len = len;
|
|
wr->uid = current_fsuid();
|
|
wr->gid = current_fsgid();
|
|
folio_attach_private(folio, wr);
|
|
okay:
|
|
return 0;
|
|
}
|
|
|
|
static int orangefs_write_end(const struct kiocb *iocb,
|
|
struct address_space *mapping,
|
|
loff_t pos, unsigned len, unsigned copied,
|
|
struct folio *folio, void *fsdata)
|
|
{
|
|
struct inode *inode = folio->mapping->host;
|
|
loff_t last_pos = pos + copied;
|
|
|
|
/*
|
|
* No need to use i_size_read() here, the i_size
|
|
* cannot change under us because we hold the i_mutex.
|
|
*/
|
|
if (last_pos > inode->i_size)
|
|
i_size_write(inode, last_pos);
|
|
|
|
/* zero the stale part of the folio if we did a short copy */
|
|
if (!folio_test_uptodate(folio)) {
|
|
unsigned from = pos & (PAGE_SIZE - 1);
|
|
if (copied < len) {
|
|
folio_zero_range(folio, from + copied, len - copied);
|
|
}
|
|
/* Set fully written pages uptodate. */
|
|
if (pos == folio_pos(folio) &&
|
|
(len == PAGE_SIZE || pos + len == inode->i_size)) {
|
|
folio_zero_segment(folio, from + copied, PAGE_SIZE);
|
|
folio_mark_uptodate(folio);
|
|
}
|
|
}
|
|
|
|
folio_mark_dirty(folio);
|
|
folio_unlock(folio);
|
|
folio_put(folio);
|
|
|
|
mark_inode_dirty_sync(file_inode(iocb->ki_filp));
|
|
return copied;
|
|
}
|
|
|
|
static void orangefs_invalidate_folio(struct folio *folio,
|
|
size_t offset, size_t length)
|
|
{
|
|
struct orangefs_write_range *wr = folio_get_private(folio);
|
|
|
|
if (offset == 0 && length == PAGE_SIZE) {
|
|
kfree(folio_detach_private(folio));
|
|
return;
|
|
/* write range entirely within invalidate range (or equal) */
|
|
} else if (folio_pos(folio) + offset <= wr->pos &&
|
|
wr->pos + wr->len <= folio_pos(folio) + offset + length) {
|
|
kfree(folio_detach_private(folio));
|
|
/* XXX is this right? only caller in fs */
|
|
folio_cancel_dirty(folio);
|
|
return;
|
|
/* invalidate range chops off end of write range */
|
|
} else if (wr->pos < folio_pos(folio) + offset &&
|
|
wr->pos + wr->len <= folio_pos(folio) + offset + length &&
|
|
folio_pos(folio) + offset < wr->pos + wr->len) {
|
|
size_t x;
|
|
x = wr->pos + wr->len - (folio_pos(folio) + offset);
|
|
WARN_ON(x > wr->len);
|
|
wr->len -= x;
|
|
wr->uid = current_fsuid();
|
|
wr->gid = current_fsgid();
|
|
/* invalidate range chops off beginning of write range */
|
|
} else if (folio_pos(folio) + offset <= wr->pos &&
|
|
folio_pos(folio) + offset + length < wr->pos + wr->len &&
|
|
wr->pos < folio_pos(folio) + offset + length) {
|
|
size_t x;
|
|
x = folio_pos(folio) + offset + length - wr->pos;
|
|
WARN_ON(x > wr->len);
|
|
wr->pos += x;
|
|
wr->len -= x;
|
|
wr->uid = current_fsuid();
|
|
wr->gid = current_fsgid();
|
|
/* invalidate range entirely within write range (punch hole) */
|
|
} else if (wr->pos < folio_pos(folio) + offset &&
|
|
folio_pos(folio) + offset + length < wr->pos + wr->len) {
|
|
/* XXX what do we do here... should not WARN_ON */
|
|
WARN_ON(1);
|
|
/* punch hole */
|
|
/*
|
|
* should we just ignore this and write it out anyway?
|
|
* it hardly makes sense
|
|
*/
|
|
return;
|
|
/* non-overlapping ranges */
|
|
} else {
|
|
/* WARN if they do overlap */
|
|
if (!((folio_pos(folio) + offset + length <= wr->pos) ^
|
|
(wr->pos + wr->len <= folio_pos(folio) + offset))) {
|
|
WARN_ON(1);
|
|
printk("invalidate range offset %llu length %zu\n",
|
|
folio_pos(folio) + offset, length);
|
|
printk("write range offset %llu length %zu\n",
|
|
wr->pos, wr->len);
|
|
}
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Above there are returns where wr is freed or where we WARN.
|
|
* Thus the following runs if wr was modified above.
|
|
*/
|
|
|
|
orangefs_launder_folio(folio);
|
|
}
|
|
|
|
static bool orangefs_release_folio(struct folio *folio, gfp_t foo)
|
|
{
|
|
return !folio_test_private(folio);
|
|
}
|
|
|
|
static void orangefs_free_folio(struct folio *folio)
|
|
{
|
|
kfree(folio_detach_private(folio));
|
|
}
|
|
|
|
static int orangefs_launder_folio(struct folio *folio)
|
|
{
|
|
int r = 0;
|
|
struct writeback_control wbc = {
|
|
.sync_mode = WB_SYNC_ALL,
|
|
.nr_to_write = 0,
|
|
};
|
|
folio_wait_writeback(folio);
|
|
if (folio_clear_dirty_for_io(folio)) {
|
|
r = orangefs_writepage_locked(folio, &wbc);
|
|
folio_end_writeback(folio);
|
|
}
|
|
return r;
|
|
}
|
|
|
|
static ssize_t orangefs_direct_IO(struct kiocb *iocb,
|
|
struct iov_iter *iter)
|
|
{
|
|
/*
|
|
* Comment from original do_readv_writev:
|
|
* Common entry point for read/write/readv/writev
|
|
* This function will dispatch it to either the direct I/O
|
|
* or buffered I/O path depending on the mount options and/or
|
|
* augmented/extended metadata attached to the file.
|
|
* Note: File extended attributes override any mount options.
|
|
*/
|
|
struct file *file = iocb->ki_filp;
|
|
loff_t pos = iocb->ki_pos;
|
|
enum ORANGEFS_io_type type = iov_iter_rw(iter) == WRITE ?
|
|
ORANGEFS_IO_WRITE : ORANGEFS_IO_READ;
|
|
loff_t *offset = &pos;
|
|
struct inode *inode = file->f_mapping->host;
|
|
struct orangefs_inode_s *orangefs_inode = ORANGEFS_I(inode);
|
|
struct orangefs_khandle *handle = &orangefs_inode->refn.khandle;
|
|
size_t count = iov_iter_count(iter);
|
|
ssize_t total_count = 0;
|
|
ssize_t ret = -EINVAL;
|
|
|
|
gossip_debug(GOSSIP_FILE_DEBUG,
|
|
"%s-BEGIN(%pU): count(%d) after estimate_max_iovecs.\n",
|
|
__func__,
|
|
handle,
|
|
(int)count);
|
|
|
|
if (type == ORANGEFS_IO_WRITE) {
|
|
gossip_debug(GOSSIP_FILE_DEBUG,
|
|
"%s(%pU): proceeding with offset : %llu, "
|
|
"size %d\n",
|
|
__func__,
|
|
handle,
|
|
llu(*offset),
|
|
(int)count);
|
|
}
|
|
|
|
if (count == 0) {
|
|
ret = 0;
|
|
goto out;
|
|
}
|
|
|
|
while (iov_iter_count(iter)) {
|
|
size_t each_count = iov_iter_count(iter);
|
|
size_t amt_complete;
|
|
|
|
/* how much to transfer in this loop iteration */
|
|
if (each_count > orangefs_bufmap_size_query())
|
|
each_count = orangefs_bufmap_size_query();
|
|
|
|
gossip_debug(GOSSIP_FILE_DEBUG,
|
|
"%s(%pU): size of each_count(%d)\n",
|
|
__func__,
|
|
handle,
|
|
(int)each_count);
|
|
gossip_debug(GOSSIP_FILE_DEBUG,
|
|
"%s(%pU): BEFORE wait_for_io: offset is %d\n",
|
|
__func__,
|
|
handle,
|
|
(int)*offset);
|
|
|
|
ret = wait_for_direct_io(type, inode, offset, iter,
|
|
each_count, 0, NULL, NULL, file);
|
|
gossip_debug(GOSSIP_FILE_DEBUG,
|
|
"%s(%pU): return from wait_for_io:%d\n",
|
|
__func__,
|
|
handle,
|
|
(int)ret);
|
|
|
|
if (ret < 0)
|
|
goto out;
|
|
|
|
*offset += ret;
|
|
total_count += ret;
|
|
amt_complete = ret;
|
|
|
|
gossip_debug(GOSSIP_FILE_DEBUG,
|
|
"%s(%pU): AFTER wait_for_io: offset is %d\n",
|
|
__func__,
|
|
handle,
|
|
(int)*offset);
|
|
|
|
/*
|
|
* if we got a short I/O operations,
|
|
* fall out and return what we got so far
|
|
*/
|
|
if (amt_complete < each_count)
|
|
break;
|
|
} /*end while */
|
|
|
|
out:
|
|
if (total_count > 0)
|
|
ret = total_count;
|
|
if (ret > 0) {
|
|
if (type == ORANGEFS_IO_READ) {
|
|
file_accessed(file);
|
|
} else {
|
|
file_update_time(file);
|
|
if (*offset > i_size_read(inode))
|
|
i_size_write(inode, *offset);
|
|
}
|
|
}
|
|
|
|
gossip_debug(GOSSIP_FILE_DEBUG,
|
|
"%s(%pU): Value(%d) returned.\n",
|
|
__func__,
|
|
handle,
|
|
(int)ret);
|
|
|
|
return ret;
|
|
}
|
|
|
|
/** ORANGEFS2 implementation of address space operations */
|
|
static const struct address_space_operations orangefs_address_operations = {
|
|
.readahead = orangefs_readahead,
|
|
.read_folio = orangefs_read_folio,
|
|
.writepages = orangefs_writepages,
|
|
.dirty_folio = filemap_dirty_folio,
|
|
.write_begin = orangefs_write_begin,
|
|
.write_end = orangefs_write_end,
|
|
.invalidate_folio = orangefs_invalidate_folio,
|
|
.release_folio = orangefs_release_folio,
|
|
.free_folio = orangefs_free_folio,
|
|
.migrate_folio = filemap_migrate_folio,
|
|
.launder_folio = orangefs_launder_folio,
|
|
.direct_IO = orangefs_direct_IO,
|
|
};
|
|
|
|
vm_fault_t orangefs_page_mkwrite(struct vm_fault *vmf)
|
|
{
|
|
struct folio *folio = page_folio(vmf->page);
|
|
struct inode *inode = file_inode(vmf->vma->vm_file);
|
|
struct orangefs_inode_s *orangefs_inode = ORANGEFS_I(inode);
|
|
unsigned long *bitlock = &orangefs_inode->bitlock;
|
|
vm_fault_t ret;
|
|
struct orangefs_write_range *wr;
|
|
|
|
sb_start_pagefault(inode->i_sb);
|
|
|
|
if (wait_on_bit(bitlock, 1, TASK_KILLABLE)) {
|
|
ret = VM_FAULT_RETRY;
|
|
goto out;
|
|
}
|
|
|
|
folio_lock(folio);
|
|
if (folio_test_dirty(folio) && !folio_test_private(folio)) {
|
|
/*
|
|
* Should be impossible. If it happens, launder the folio
|
|
* since we don't know what's dirty. This will WARN in
|
|
* orangefs_writepage_locked.
|
|
*/
|
|
if (orangefs_launder_folio(folio)) {
|
|
ret = VM_FAULT_LOCKED|VM_FAULT_RETRY;
|
|
goto out;
|
|
}
|
|
}
|
|
if (folio_test_private(folio)) {
|
|
wr = folio_get_private(folio);
|
|
if (uid_eq(wr->uid, current_fsuid()) &&
|
|
gid_eq(wr->gid, current_fsgid())) {
|
|
wr->pos = page_offset(vmf->page);
|
|
wr->len = PAGE_SIZE;
|
|
goto okay;
|
|
} else {
|
|
if (orangefs_launder_folio(folio)) {
|
|
ret = VM_FAULT_LOCKED|VM_FAULT_RETRY;
|
|
goto out;
|
|
}
|
|
}
|
|
}
|
|
wr = kmalloc(sizeof *wr, GFP_KERNEL);
|
|
if (!wr) {
|
|
ret = VM_FAULT_LOCKED|VM_FAULT_RETRY;
|
|
goto out;
|
|
}
|
|
wr->pos = page_offset(vmf->page);
|
|
wr->len = PAGE_SIZE;
|
|
wr->uid = current_fsuid();
|
|
wr->gid = current_fsgid();
|
|
folio_attach_private(folio, wr);
|
|
okay:
|
|
|
|
file_update_time(vmf->vma->vm_file);
|
|
if (folio->mapping != inode->i_mapping) {
|
|
folio_unlock(folio);
|
|
ret = VM_FAULT_LOCKED|VM_FAULT_NOPAGE;
|
|
goto out;
|
|
}
|
|
|
|
/*
|
|
* We mark the folio dirty already here so that when freeze is in
|
|
* progress, we are guaranteed that writeback during freezing will
|
|
* see the dirty folio and writeprotect it again.
|
|
*/
|
|
folio_mark_dirty(folio);
|
|
folio_wait_stable(folio);
|
|
ret = VM_FAULT_LOCKED;
|
|
out:
|
|
sb_end_pagefault(inode->i_sb);
|
|
return ret;
|
|
}
|
|
|
|
static int orangefs_setattr_size(struct inode *inode, struct iattr *iattr)
|
|
{
|
|
struct orangefs_inode_s *orangefs_inode = ORANGEFS_I(inode);
|
|
struct orangefs_kernel_op_s *new_op;
|
|
loff_t orig_size;
|
|
int ret = -EINVAL;
|
|
|
|
gossip_debug(GOSSIP_INODE_DEBUG,
|
|
"%s: %pU: Handle is %pU | fs_id %d | size is %llu\n",
|
|
__func__,
|
|
get_khandle_from_ino(inode),
|
|
&orangefs_inode->refn.khandle,
|
|
orangefs_inode->refn.fs_id,
|
|
iattr->ia_size);
|
|
|
|
/* Ensure that we have a up to date size, so we know if it changed. */
|
|
ret = orangefs_inode_getattr(inode, ORANGEFS_GETATTR_SIZE);
|
|
if (ret == -ESTALE)
|
|
ret = -EIO;
|
|
if (ret) {
|
|
gossip_err("%s: orangefs_inode_getattr failed, ret:%d:.\n",
|
|
__func__, ret);
|
|
return ret;
|
|
}
|
|
orig_size = i_size_read(inode);
|
|
|
|
/* This is truncate_setsize in a different order. */
|
|
truncate_pagecache(inode, iattr->ia_size);
|
|
i_size_write(inode, iattr->ia_size);
|
|
if (iattr->ia_size > orig_size)
|
|
pagecache_isize_extended(inode, orig_size, iattr->ia_size);
|
|
|
|
new_op = op_alloc(ORANGEFS_VFS_OP_TRUNCATE);
|
|
if (!new_op)
|
|
return -ENOMEM;
|
|
|
|
new_op->upcall.req.truncate.refn = orangefs_inode->refn;
|
|
new_op->upcall.req.truncate.size = (__s64) iattr->ia_size;
|
|
|
|
ret = service_operation(new_op,
|
|
__func__,
|
|
get_interruptible_flag(inode));
|
|
|
|
/*
|
|
* the truncate has no downcall members to retrieve, but
|
|
* the status value tells us if it went through ok or not
|
|
*/
|
|
gossip_debug(GOSSIP_INODE_DEBUG, "%s: ret:%d:\n", __func__, ret);
|
|
|
|
op_release(new_op);
|
|
|
|
if (ret != 0)
|
|
return ret;
|
|
|
|
if (orig_size != i_size_read(inode))
|
|
iattr->ia_valid |= ATTR_CTIME | ATTR_MTIME;
|
|
|
|
return ret;
|
|
}
|
|
|
|
int __orangefs_setattr(struct inode *inode, struct iattr *iattr)
|
|
{
|
|
int ret;
|
|
|
|
if (iattr->ia_valid & ATTR_MODE) {
|
|
if (iattr->ia_mode & (S_ISVTX)) {
|
|
if (is_root_handle(inode)) {
|
|
/*
|
|
* allow sticky bit to be set on root (since
|
|
* it shows up that way by default anyhow),
|
|
* but don't show it to the server
|
|
*/
|
|
iattr->ia_mode -= S_ISVTX;
|
|
} else {
|
|
gossip_debug(GOSSIP_UTILS_DEBUG,
|
|
"User attempted to set sticky bit on non-root directory; returning EINVAL.\n");
|
|
ret = -EINVAL;
|
|
goto out;
|
|
}
|
|
}
|
|
if (iattr->ia_mode & (S_ISUID)) {
|
|
gossip_debug(GOSSIP_UTILS_DEBUG,
|
|
"Attempting to set setuid bit (not supported); returning EINVAL.\n");
|
|
ret = -EINVAL;
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
if (iattr->ia_valid & ATTR_SIZE) {
|
|
ret = orangefs_setattr_size(inode, iattr);
|
|
if (ret)
|
|
goto out;
|
|
}
|
|
|
|
again:
|
|
spin_lock(&inode->i_lock);
|
|
if (ORANGEFS_I(inode)->attr_valid) {
|
|
if (uid_eq(ORANGEFS_I(inode)->attr_uid, current_fsuid()) &&
|
|
gid_eq(ORANGEFS_I(inode)->attr_gid, current_fsgid())) {
|
|
ORANGEFS_I(inode)->attr_valid = iattr->ia_valid;
|
|
} else {
|
|
spin_unlock(&inode->i_lock);
|
|
write_inode_now(inode, 1);
|
|
goto again;
|
|
}
|
|
} else {
|
|
ORANGEFS_I(inode)->attr_valid = iattr->ia_valid;
|
|
ORANGEFS_I(inode)->attr_uid = current_fsuid();
|
|
ORANGEFS_I(inode)->attr_gid = current_fsgid();
|
|
}
|
|
setattr_copy(&nop_mnt_idmap, inode, iattr);
|
|
spin_unlock(&inode->i_lock);
|
|
mark_inode_dirty(inode);
|
|
|
|
ret = 0;
|
|
out:
|
|
return ret;
|
|
}
|
|
|
|
int __orangefs_setattr_mode(struct dentry *dentry, struct iattr *iattr)
|
|
{
|
|
int ret;
|
|
struct inode *inode = d_inode(dentry);
|
|
|
|
ret = __orangefs_setattr(inode, iattr);
|
|
/* change mode on a file that has ACLs */
|
|
if (!ret && (iattr->ia_valid & ATTR_MODE))
|
|
ret = posix_acl_chmod(&nop_mnt_idmap, dentry, inode->i_mode);
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Change attributes of an object referenced by dentry.
|
|
*/
|
|
int orangefs_setattr(struct mnt_idmap *idmap, struct dentry *dentry,
|
|
struct iattr *iattr)
|
|
{
|
|
int ret;
|
|
gossip_debug(GOSSIP_INODE_DEBUG, "__orangefs_setattr: called on %pd\n",
|
|
dentry);
|
|
ret = setattr_prepare(&nop_mnt_idmap, dentry, iattr);
|
|
if (ret)
|
|
goto out;
|
|
ret = __orangefs_setattr_mode(dentry, iattr);
|
|
sync_inode_metadata(d_inode(dentry), 1);
|
|
out:
|
|
gossip_debug(GOSSIP_INODE_DEBUG, "orangefs_setattr: returning %d\n",
|
|
ret);
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Obtain attributes of an object given a dentry
|
|
*/
|
|
int orangefs_getattr(struct mnt_idmap *idmap, const struct path *path,
|
|
struct kstat *stat, u32 request_mask, unsigned int flags)
|
|
{
|
|
int ret;
|
|
struct inode *inode = path->dentry->d_inode;
|
|
|
|
gossip_debug(GOSSIP_INODE_DEBUG,
|
|
"orangefs_getattr: called on %pd mask %u\n",
|
|
path->dentry, request_mask);
|
|
|
|
ret = orangefs_inode_getattr(inode,
|
|
request_mask & STATX_SIZE ? ORANGEFS_GETATTR_SIZE : 0);
|
|
if (ret == 0) {
|
|
generic_fillattr(&nop_mnt_idmap, request_mask, inode, stat);
|
|
|
|
/* override block size reported to stat */
|
|
if (!(request_mask & STATX_SIZE))
|
|
stat->result_mask &= ~STATX_SIZE;
|
|
|
|
generic_fill_statx_attr(inode, stat);
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
int orangefs_permission(struct mnt_idmap *idmap,
|
|
struct inode *inode, int mask)
|
|
{
|
|
int ret;
|
|
|
|
if (mask & MAY_NOT_BLOCK)
|
|
return -ECHILD;
|
|
|
|
gossip_debug(GOSSIP_INODE_DEBUG, "%s: refreshing\n", __func__);
|
|
|
|
/* Make sure the permission (and other common attrs) are up to date. */
|
|
ret = orangefs_inode_getattr(inode, 0);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
return generic_permission(&nop_mnt_idmap, inode, mask);
|
|
}
|
|
|
|
int orangefs_update_time(struct inode *inode, int flags)
|
|
{
|
|
struct iattr iattr;
|
|
|
|
gossip_debug(GOSSIP_INODE_DEBUG, "orangefs_update_time: %pU\n",
|
|
get_khandle_from_ino(inode));
|
|
|
|
flags = inode_update_timestamps(inode, flags);
|
|
|
|
memset(&iattr, 0, sizeof iattr);
|
|
if (flags & S_ATIME)
|
|
iattr.ia_valid |= ATTR_ATIME;
|
|
if (flags & S_CTIME)
|
|
iattr.ia_valid |= ATTR_CTIME;
|
|
if (flags & S_MTIME)
|
|
iattr.ia_valid |= ATTR_MTIME;
|
|
return __orangefs_setattr(inode, &iattr);
|
|
}
|
|
|
|
static int orangefs_fileattr_get(struct dentry *dentry, struct file_kattr *fa)
|
|
{
|
|
u64 val = 0;
|
|
int ret;
|
|
|
|
gossip_debug(GOSSIP_FILE_DEBUG, "%s: called on %pd\n", __func__,
|
|
dentry);
|
|
|
|
ret = orangefs_inode_getxattr(d_inode(dentry),
|
|
"user.pvfs2.meta_hint",
|
|
&val, sizeof(val));
|
|
if (ret < 0 && ret != -ENODATA)
|
|
return ret;
|
|
|
|
gossip_debug(GOSSIP_FILE_DEBUG, "%s: flags=%u\n", __func__, (u32) val);
|
|
|
|
fileattr_fill_flags(fa, val);
|
|
return 0;
|
|
}
|
|
|
|
static int orangefs_fileattr_set(struct mnt_idmap *idmap,
|
|
struct dentry *dentry, struct file_kattr *fa)
|
|
{
|
|
u64 val = 0;
|
|
|
|
gossip_debug(GOSSIP_FILE_DEBUG, "%s: called on %pd\n", __func__,
|
|
dentry);
|
|
/*
|
|
* ORANGEFS_MIRROR_FL is set internally when the mirroring mode is
|
|
* turned on for a file. The user is not allowed to turn on this bit,
|
|
* but the bit is present if the user first gets the flags and then
|
|
* updates the flags with some new settings. So, we ignore it in the
|
|
* following edit. bligon.
|
|
*/
|
|
if (fileattr_has_fsx(fa) ||
|
|
(fa->flags & ~(FS_IMMUTABLE_FL | FS_APPEND_FL | FS_NOATIME_FL | ORANGEFS_MIRROR_FL))) {
|
|
gossip_err("%s: only supports setting one of FS_IMMUTABLE_FL|FS_APPEND_FL|FS_NOATIME_FL\n",
|
|
__func__);
|
|
return -EOPNOTSUPP;
|
|
}
|
|
val = fa->flags;
|
|
gossip_debug(GOSSIP_FILE_DEBUG, "%s: flags=%u\n", __func__, (u32) val);
|
|
return orangefs_inode_setxattr(d_inode(dentry),
|
|
"user.pvfs2.meta_hint",
|
|
&val, sizeof(val), 0);
|
|
}
|
|
|
|
/* ORANGEFS2 implementation of VFS inode operations for files */
|
|
static const struct inode_operations orangefs_file_inode_operations = {
|
|
.get_inode_acl = orangefs_get_acl,
|
|
.set_acl = orangefs_set_acl,
|
|
.setattr = orangefs_setattr,
|
|
.getattr = orangefs_getattr,
|
|
.listxattr = orangefs_listxattr,
|
|
.permission = orangefs_permission,
|
|
.update_time = orangefs_update_time,
|
|
.fileattr_get = orangefs_fileattr_get,
|
|
.fileattr_set = orangefs_fileattr_set,
|
|
};
|
|
|
|
static int orangefs_init_iops(struct inode *inode)
|
|
{
|
|
inode->i_mapping->a_ops = &orangefs_address_operations;
|
|
|
|
switch (inode->i_mode & S_IFMT) {
|
|
case S_IFREG:
|
|
inode->i_op = &orangefs_file_inode_operations;
|
|
inode->i_fop = &orangefs_file_operations;
|
|
break;
|
|
case S_IFLNK:
|
|
inode->i_op = &orangefs_symlink_inode_operations;
|
|
break;
|
|
case S_IFDIR:
|
|
inode->i_op = &orangefs_dir_inode_operations;
|
|
inode->i_fop = &orangefs_dir_operations;
|
|
break;
|
|
default:
|
|
gossip_debug(GOSSIP_INODE_DEBUG,
|
|
"%s: unsupported mode\n",
|
|
__func__);
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Given an ORANGEFS object identifier (fsid, handle), convert it into
|
|
* a ino_t type that will be used as a hash-index from where the handle will
|
|
* be searched for in the VFS hash table of inodes.
|
|
*/
|
|
static inline ino_t orangefs_handle_hash(struct orangefs_object_kref *ref)
|
|
{
|
|
if (!ref)
|
|
return 0;
|
|
return orangefs_khandle_to_ino(&(ref->khandle));
|
|
}
|
|
|
|
/*
|
|
* Called to set up an inode from iget5_locked.
|
|
*/
|
|
static int orangefs_set_inode(struct inode *inode, void *data)
|
|
{
|
|
struct orangefs_object_kref *ref = (struct orangefs_object_kref *) data;
|
|
ORANGEFS_I(inode)->refn.fs_id = ref->fs_id;
|
|
ORANGEFS_I(inode)->refn.khandle = ref->khandle;
|
|
ORANGEFS_I(inode)->attr_valid = 0;
|
|
hash_init(ORANGEFS_I(inode)->xattr_cache);
|
|
ORANGEFS_I(inode)->mapping_time = jiffies - 1;
|
|
ORANGEFS_I(inode)->bitlock = 0;
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Called to determine if handles match.
|
|
*/
|
|
static int orangefs_test_inode(struct inode *inode, void *data)
|
|
{
|
|
struct orangefs_object_kref *ref = (struct orangefs_object_kref *) data;
|
|
struct orangefs_inode_s *orangefs_inode = NULL;
|
|
|
|
orangefs_inode = ORANGEFS_I(inode);
|
|
/* test handles and fs_ids... */
|
|
return (!ORANGEFS_khandle_cmp(&(orangefs_inode->refn.khandle),
|
|
&(ref->khandle)) &&
|
|
orangefs_inode->refn.fs_id == ref->fs_id);
|
|
}
|
|
|
|
/*
|
|
* Front-end to lookup the inode-cache maintained by the VFS using the ORANGEFS
|
|
* file handle.
|
|
*
|
|
* @sb: the file system super block instance.
|
|
* @ref: The ORANGEFS object for which we are trying to locate an inode.
|
|
*/
|
|
struct inode *orangefs_iget(struct super_block *sb,
|
|
struct orangefs_object_kref *ref)
|
|
{
|
|
struct inode *inode = NULL;
|
|
unsigned long hash;
|
|
int error;
|
|
|
|
hash = orangefs_handle_hash(ref);
|
|
inode = iget5_locked(sb,
|
|
hash,
|
|
orangefs_test_inode,
|
|
orangefs_set_inode,
|
|
ref);
|
|
|
|
if (!inode)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
if (!(inode_state_read_once(inode) & I_NEW))
|
|
return inode;
|
|
|
|
error = orangefs_inode_getattr(inode, ORANGEFS_GETATTR_NEW);
|
|
if (error) {
|
|
iget_failed(inode);
|
|
return ERR_PTR(error);
|
|
}
|
|
|
|
inode->i_ino = hash; /* needed for stat etc */
|
|
orangefs_init_iops(inode);
|
|
unlock_new_inode(inode);
|
|
|
|
gossip_debug(GOSSIP_INODE_DEBUG,
|
|
"iget handle %pU, fsid %d hash %ld i_ino %lu\n",
|
|
&ref->khandle,
|
|
ref->fs_id,
|
|
hash,
|
|
inode->i_ino);
|
|
|
|
return inode;
|
|
}
|
|
|
|
/*
|
|
* Allocate an inode for a newly created file and insert it into the inode hash.
|
|
*/
|
|
struct inode *orangefs_new_inode(struct super_block *sb, struct inode *dir,
|
|
umode_t mode, dev_t dev, struct orangefs_object_kref *ref)
|
|
{
|
|
struct posix_acl *acl = NULL, *default_acl = NULL;
|
|
unsigned long hash = orangefs_handle_hash(ref);
|
|
struct inode *inode;
|
|
int error;
|
|
|
|
gossip_debug(GOSSIP_INODE_DEBUG,
|
|
"%s:(sb is %p | MAJOR(dev)=%u | MINOR(dev)=%u mode=%o)\n",
|
|
__func__,
|
|
sb,
|
|
MAJOR(dev),
|
|
MINOR(dev),
|
|
mode);
|
|
|
|
inode = new_inode(sb);
|
|
if (!inode)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
error = posix_acl_create(dir, &mode, &default_acl, &acl);
|
|
if (error)
|
|
goto out_iput;
|
|
|
|
orangefs_set_inode(inode, ref);
|
|
inode->i_ino = hash; /* needed for stat etc */
|
|
|
|
error = orangefs_inode_getattr(inode, ORANGEFS_GETATTR_NEW);
|
|
if (error)
|
|
goto out_iput;
|
|
|
|
orangefs_init_iops(inode);
|
|
inode->i_rdev = dev;
|
|
|
|
if (default_acl) {
|
|
error = __orangefs_set_acl(inode, default_acl,
|
|
ACL_TYPE_DEFAULT);
|
|
if (error)
|
|
goto out_iput;
|
|
}
|
|
|
|
if (acl) {
|
|
error = __orangefs_set_acl(inode, acl, ACL_TYPE_ACCESS);
|
|
if (error)
|
|
goto out_iput;
|
|
}
|
|
|
|
error = insert_inode_locked4(inode, hash, orangefs_test_inode, ref);
|
|
if (error < 0)
|
|
goto out_iput;
|
|
|
|
gossip_debug(GOSSIP_INODE_DEBUG,
|
|
"Initializing ACL's for inode %pU\n",
|
|
get_khandle_from_ino(inode));
|
|
if (mode != inode->i_mode) {
|
|
struct iattr iattr = {
|
|
.ia_mode = mode,
|
|
.ia_valid = ATTR_MODE,
|
|
};
|
|
inode->i_mode = mode;
|
|
__orangefs_setattr(inode, &iattr);
|
|
__posix_acl_chmod(&acl, GFP_KERNEL, inode->i_mode);
|
|
}
|
|
posix_acl_release(acl);
|
|
posix_acl_release(default_acl);
|
|
return inode;
|
|
|
|
out_iput:
|
|
iput(inode);
|
|
posix_acl_release(acl);
|
|
posix_acl_release(default_acl);
|
|
return ERR_PTR(error);
|
|
}
|