updated: [566bad4] Fix compiling error on kernels > 2.6.23: error: unknown field âsendfileâ specified in initializer
Patrick Winnertz
winnie at debian.org
Fri Jun 5 13:57:56 UTC 2009
The following commit has been merged in the master branch:
commit 566bad4c5ad84f69bb3f4f5656806a0955598c08
Author: Patrick Winnertz <winnie at debian.org>
Date: Fri Jun 5 10:20:13 2009 +0200
Fix compiling error on kernels > 2.6.23: error: unknown field ‘sendfile’ specified in initializer
Signed-off-by: Patrick Winnertz <winnie at debian.org>
diff --git a/debian/patches/bio_errors.dpatch b/debian/patches/bio_errors.dpatch
index 9e6c171..4bba9cd 100755
--- a/debian/patches/bio_errors.dpatch
+++ b/debian/patches/bio_errors.dpatch
@@ -6,9 +6,9 @@
@DPATCH@
diff -urNad lustre~/lustre/autoconf/lustre-core.m4 lustre/lustre/autoconf/lustre-core.m4
---- lustre~/lustre/autoconf/lustre-core.m4 2009-06-02 11:38:56.000000000 +0200
-+++ lustre/lustre/autoconf/lustre-core.m4 2009-06-04 15:16:45.000000000 +0200
-@@ -1390,6 +1390,22 @@
+--- lustre~/lustre/autoconf/lustre-core.m4 2009-06-05 10:12:07.000000000 +0200
++++ lustre/lustre/autoconf/lustre-core.m4 2009-06-05 10:18:54.000000000 +0200
+@@ -1390,6 +1390,41 @@
])
])
@@ -28,21 +28,41 @@ diff -urNad lustre~/lustre/autoconf/lustre-core.m4 lustre/lustre/autoconf/lustre
+])
+])
+
++# 2.6.23 change .sendfile to .splice_read
++# RHEL4 (-92 kernel) have both sendfile and .splice_read API
++AC_DEFUN([LC_KERNEL_SENDFILE],
++[AC_MSG_CHECKING([if kernel has .sendfile])
++LB_LINUX_TRY_COMPILE([
++ #include <linux/fs.h>
++],[
++ struct file_operations file;
++
++ file.sendfile = NULL;
++], [
++ AC_MSG_RESULT([yes])
++ AC_DEFINE(HAVE_KERNEL_SENDFILE, 1,
++ [kernel has .sendfile])
++],[
++ AC_MSG_RESULT([no])
++])
++])
++
#
# LC_PROG_LINUX
#
-@@ -1503,6 +1519,8 @@
+@@ -1503,6 +1538,9 @@
LC_UNREGISTER_BLKDEV_RETURN_INT
LC_KERNEL_SPLICE_READ
LC_HAVE_EXPORTFS_H
+ LC_BIO_ENDIO_2ARG
++ LC_KERNEL_SENDFILE
+
])
#
diff -urNad lustre~/lustre/include/linux/lustre_compat25.h lustre/lustre/include/linux/lustre_compat25.h
---- lustre~/lustre/include/linux/lustre_compat25.h 2009-06-02 11:38:56.000000000 +0200
-+++ lustre/lustre/include/linux/lustre_compat25.h 2009-06-04 15:12:23.000000000 +0200
+--- lustre~/lustre/include/linux/lustre_compat25.h 2009-06-05 10:12:07.000000000 +0200
++++ lustre/lustre/include/linux/lustre_compat25.h 2009-06-05 10:12:08.000000000 +0200
@@ -472,5 +472,13 @@
#define sysctl_vfs_cache_pressure 100
#endif
@@ -57,9 +77,3484 @@ diff -urNad lustre~/lustre/include/linux/lustre_compat25.h lustre/lustre/include
+
#endif /* __KERNEL__ */
#endif /* _COMPAT25_H */
+diff -urNad lustre~/lustre/llite/file.c lustre/lustre/llite/file.c
+--- lustre~/lustre/llite/file.c 2009-06-05 10:12:07.000000000 +0200
++++ lustre/lustre/llite/file.c 2009-06-05 10:15:12.000000000 +0200
+@@ -3262,7 +3262,9 @@
+ .release = ll_file_release,
+ .mmap = ll_file_mmap,
+ .llseek = ll_file_seek,
++#ifdef HAVE_KERNEL_SENDFILE
+ .sendfile = ll_file_sendfile,
++#endif
+ .fsync = ll_fsync,
+ };
+
+@@ -3284,7 +3286,9 @@
+ .release = ll_file_release,
+ .mmap = ll_file_mmap,
+ .llseek = ll_file_seek,
++#ifdef HAVE_KERNEL_SENDFILE
+ .sendfile = ll_file_sendfile,
++#endif
+ .fsync = ll_fsync,
+ #ifdef HAVE_F_OP_FLOCK
+ .flock = ll_file_flock,
+@@ -3311,7 +3315,9 @@
+ .release = ll_file_release,
+ .mmap = ll_file_mmap,
+ .llseek = ll_file_seek,
++#ifdef HAVE_KERNEL_SENDFILE
+ .sendfile = ll_file_sendfile,
++#endif
+ .fsync = ll_fsync,
+ #ifdef HAVE_F_OP_FLOCK
+ .flock = ll_file_noflock,
+diff -urNad lustre~/lustre/llite/file.c~ lustre/lustre/llite/file.c~
+--- lustre~/lustre/llite/file.c~ 1970-01-01 01:00:00.000000000 +0100
++++ lustre/lustre/llite/file.c~ 2009-06-05 10:13:34.000000000 +0200
+@@ -0,0 +1,3438 @@
++/* -*- mode: c; c-basic-offset: 8; indent-tabs-mode: nil; -*-
++ * vim:expandtab:shiftwidth=8:tabstop=8:
++ *
++ * GPL HEADER START
++ *
++ * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
++ *
++ * This program is free software; you can redistribute it and/or modify
++ * it under the terms of the GNU General Public License version 2 only,
++ * as published by the Free Software Foundation.
++ *
++ * This program is distributed in the hope that it will be useful, but
++ * WITHOUT ANY WARRANTY; without even the implied warranty of
++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
++ * General Public License version 2 for more details (a copy is included
++ * in the LICENSE file that accompanied this code).
++ *
++ * You should have received a copy of the GNU General Public License
++ * version 2 along with this program; If not, see
++ * http://www.sun.com/software/products/lustre/docs/GPLv2.pdf
++ *
++ * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara,
++ * CA 95054 USA or visit www.sun.com if you need additional information or
++ * have any questions.
++ *
++ * GPL HEADER END
++ */
++/*
++ * Copyright 2008 Sun Microsystems, Inc. All rights reserved
++ * Use is subject to license terms.
++ */
++/*
++ * This file is part of Lustre, http://www.lustre.org/
++ * Lustre is a trademark of Sun Microsystems, Inc.
++ *
++ * lustre/llite/file.c
++ *
++ * Author: Peter Braam <braam at clusterfs.com>
++ * Author: Phil Schwan <phil at clusterfs.com>
++ * Author: Andreas Dilger <adilger at clusterfs.com>
++ */
++
++#define DEBUG_SUBSYSTEM S_LLITE
++#include <lustre_dlm.h>
++#include <lustre_lite.h>
++#include <linux/pagemap.h>
++#include <linux/file.h>
++#include <linux/posix_acl.h>
++#include "llite_internal.h"
++#include <lustre/ll_fiemap.h>
++
++/* also used by llite/special.c:ll_special_open() */
++struct ll_file_data *ll_file_data_get(void)
++{
++ struct ll_file_data *fd;
++
++ OBD_SLAB_ALLOC_PTR(fd, ll_file_data_slab);
++ return fd;
++}
++
++static void ll_file_data_put(struct ll_file_data *fd)
++{
++ if (fd != NULL)
++ OBD_SLAB_FREE_PTR(fd, ll_file_data_slab);
++}
++
++static int ll_close_inode_openhandle(struct inode *inode,
++ struct obd_client_handle *och)
++{
++ struct ptlrpc_request *req = NULL;
++ struct obd_device *obd;
++ struct obdo *oa;
++ struct mdc_op_data data = { { 0 } };
++ int rc;
++ ENTRY;
++
++ obd = class_exp2obd(ll_i2mdcexp(inode));
++ if (obd == NULL) {
++ CERROR("Invalid MDC connection handle "LPX64"\n",
++ ll_i2mdcexp(inode)->exp_handle.h_cookie);
++ GOTO(out, rc = 0);
++ }
++
++ /*
++ * here we check if this is forced umount. If so this is called on
++ * canceling "open lock" and we do not call mdc_close() in this case, as
++ * it will not be successful, as import is already deactivated.
++ */
++ if (obd->obd_force)
++ GOTO(out, rc = 0);
++
++ OBDO_ALLOC(oa);
++ if (!oa)
++ RETURN(-ENOMEM); // XXX We leak openhandle and request here.
++
++ oa->o_id = inode->i_ino;
++ oa->o_valid = OBD_MD_FLID;
++ obdo_from_inode(oa, inode, OBD_MD_FLTYPE | OBD_MD_FLMODE |
++ OBD_MD_FLSIZE | OBD_MD_FLBLOCKS |
++ OBD_MD_FLATIME | OBD_MD_FLMTIME |
++ OBD_MD_FLCTIME);
++ if (ll_is_inode_dirty(inode)) {
++ oa->o_flags = MDS_BFLAG_UNCOMMITTED_WRITES;
++ oa->o_valid |= OBD_MD_FLFLAGS;
++ }
++ ll_inode2fid(&data.fid1, inode);
++ rc = mdc_close(ll_i2mdcexp(inode), &data, oa, och, &req);
++ if (rc == EAGAIN) {
++ /* We are the last writer, so the MDS has instructed us to get
++ * the file size and any write cookies, then close again. */
++ ll_queue_done_writing(inode);
++ rc = 0;
++ } else if (rc) {
++ CERROR("inode %lu mdc close failed: rc = %d\n",
++ inode->i_ino, rc);
++ }
++
++ OBDO_FREE(oa);
++
++ if (rc == 0) {
++ rc = ll_objects_destroy(req, inode);
++ if (rc)
++ CERROR("inode %lu ll_objects destroy: rc = %d\n",
++ inode->i_ino, rc);
++ }
++
++ ptlrpc_req_finished(req); /* This is close request */
++ EXIT;
++out:
++ mdc_clear_open_replay_data(och);
++
++ return rc;
++}
++
++int ll_mdc_real_close(struct inode *inode, int flags)
++{
++ struct ll_inode_info *lli = ll_i2info(inode);
++ int rc = 0;
++ struct obd_client_handle **och_p;
++ struct obd_client_handle *och;
++ __u64 *och_usecount;
++
++ ENTRY;
++
++ if (flags & FMODE_WRITE) {
++ och_p = &lli->lli_mds_write_och;
++ och_usecount = &lli->lli_open_fd_write_count;
++ } else if (flags & FMODE_EXEC) {
++ och_p = &lli->lli_mds_exec_och;
++ och_usecount = &lli->lli_open_fd_exec_count;
++ } else {
++ LASSERT(flags & FMODE_READ);
++ och_p = &lli->lli_mds_read_och;
++ och_usecount = &lli->lli_open_fd_read_count;
++ }
++
++ down(&lli->lli_och_sem);
++ if (*och_usecount) { /* There are still users of this handle, so
++ skip freeing it. */
++ up(&lli->lli_och_sem);
++ RETURN(0);
++ }
++ och=*och_p;
++ *och_p = NULL;
++ up(&lli->lli_och_sem);
++
++ if (och) { /* There might be a race and somebody have freed this och
++ already */
++ rc = ll_close_inode_openhandle(inode, och);
++ och->och_fh.cookie = DEAD_HANDLE_MAGIC;
++ OBD_FREE(och, sizeof *och);
++ }
++
++ RETURN(rc);
++}
++
++int ll_mdc_close(struct obd_export *mdc_exp, struct inode *inode,
++ struct file *file)
++{
++ struct ll_file_data *fd = LUSTRE_FPRIVATE(file);
++ struct ll_inode_info *lli = ll_i2info(inode);
++ int rc = 0;
++ ENTRY;
++
++ /* clear group lock, if present */
++ if (unlikely(fd->fd_flags & LL_FILE_GROUP_LOCKED)) {
++ struct lov_stripe_md *lsm = ll_i2info(inode)->lli_smd;
++ fd->fd_flags &= ~(LL_FILE_GROUP_LOCKED|LL_FILE_IGNORE_LOCK);
++ rc = ll_extent_unlock(fd, inode, lsm, LCK_GROUP,
++ &fd->fd_cwlockh);
++ }
++
++ /* Let's see if we have good enough OPEN lock on the file and if
++ we can skip talking to MDS */
++ if (file->f_dentry->d_inode) { /* Can this ever be false? */
++ int lockmode;
++ int flags = LDLM_FL_BLOCK_GRANTED | LDLM_FL_TEST_LOCK;
++ struct lustre_handle lockh;
++ struct inode *inode = file->f_dentry->d_inode;
++ struct ldlm_res_id file_res_id;
++
++ ldlm_policy_data_t policy = {.l_inodebits={MDS_INODELOCK_OPEN}};
++ fid_build_reg_res_name(ll_inode_lu_fid(inode), &file_res_id);
++
++ down(&lli->lli_och_sem);
++ if (fd->fd_omode & FMODE_WRITE) {
++ lockmode = LCK_CW;
++ LASSERT(lli->lli_open_fd_write_count);
++ lli->lli_open_fd_write_count--;
++ } else if (fd->fd_omode & FMODE_EXEC) {
++ lockmode = LCK_PR;
++ LASSERT(lli->lli_open_fd_exec_count);
++ lli->lli_open_fd_exec_count--;
++ } else {
++ lockmode = LCK_CR;
++ LASSERT(lli->lli_open_fd_read_count);
++ lli->lli_open_fd_read_count--;
++ }
++ up(&lli->lli_och_sem);
++
++ if (!ldlm_lock_match(mdc_exp->exp_obd->obd_namespace, flags,
++ &file_res_id, LDLM_IBITS, &policy,lockmode,
++ &lockh)) {
++ rc = ll_mdc_real_close(file->f_dentry->d_inode,
++ fd->fd_omode);
++ }
++ } else {
++ CERROR("Releasing a file %p with negative dentry %p. Name %s",
++ file, file->f_dentry, file->f_dentry->d_name.name);
++ }
++
++ LUSTRE_FPRIVATE(file) = NULL;
++ ll_file_data_put(fd);
++
++ RETURN(rc);
++}
++
++int lov_test_and_clear_async_rc(struct lov_stripe_md *lsm);
++
++/* While this returns an error code, fput() the caller does not, so we need
++ * to make every effort to clean up all of our state here. Also, applications
++ * rarely check close errors and even if an error is returned they will not
++ * re-try the close call.
++ */
++int ll_file_release(struct inode *inode, struct file *file)
++{
++ struct ll_file_data *fd;
++ struct ll_sb_info *sbi = ll_i2sbi(inode);
++ struct ll_inode_info *lli = ll_i2info(inode);
++ struct lov_stripe_md *lsm = lli->lli_smd;
++ int rc;
++ ENTRY;
++
++ CDEBUG(D_VFSTRACE, "VFS Op:inode=%lu/%u(%p)\n", inode->i_ino,
++ inode->i_generation, inode);
++
++
++ if (inode->i_sb->s_root != file->f_dentry)
++ ll_stats_ops_tally(sbi, LPROC_LL_RELEASE, 1);
++ fd = LUSTRE_FPRIVATE(file);
++ LASSERT(fd != NULL);
++
++ /* The last ref on @file, maybe not the the owner pid of statahead.
++ * Different processes can open the same dir, "ll_opendir_key" means:
++ * it is me that should stop the statahead thread. */
++ if (lli->lli_opendir_key == fd && lli->lli_opendir_pid != 0)
++ ll_stop_statahead(inode, lli->lli_opendir_key);
++
++ if (inode->i_sb->s_root == file->f_dentry) {
++ LUSTRE_FPRIVATE(file) = NULL;
++ ll_file_data_put(fd);
++ RETURN(0);
++ }
++
++ if (lsm)
++ lov_test_and_clear_async_rc(lsm);
++ lli->lli_async_rc = 0;
++
++ /* Ensure that dirty pages are flushed out with the right creds */
++ if (file->f_mode & FMODE_WRITE)
++ filemap_fdatawrite(file->f_mapping);
++
++ rc = ll_mdc_close(sbi->ll_mdc_exp, inode, file);
++ RETURN(rc);
++}
++
++static int ll_intent_file_open(struct file *file, void *lmm,
++ int lmmsize, struct lookup_intent *itp)
++{
++ struct ll_sb_info *sbi = ll_i2sbi(file->f_dentry->d_inode);
++ struct mdc_op_data data = { { 0 } };
++ struct dentry *parent = file->f_dentry->d_parent;
++ const char *name = file->f_dentry->d_name.name;
++ const int len = file->f_dentry->d_name.len;
++ struct inode *inode = file->f_dentry->d_inode;
++ struct ptlrpc_request *req;
++ int rc;
++ ENTRY;
++
++ if (!parent)
++ RETURN(-ENOENT);
++
++ ll_prepare_mdc_op_data(&data, parent->d_inode, inode,
++ name, len, O_RDWR, NULL);
++
++ /* Usually we come here only for NFSD, and we want open lock.
++ But we can also get here with pre 2.6.15 patchless kernels, and in
++ that case that lock is also ok */
++ /* We can also get here if there was cached open handle in revalidate_it
++ * but it disappeared while we were getting from there to ll_file_open.
++ * But this means this file was closed and immediatelly opened which
++ * makes a good candidate for using OPEN lock */
++ /* If lmmsize & lmm are not 0, we are just setting stripe info
++ * parameters. No need for the open lock */
++ if (!lmm && !lmmsize)
++ itp->it_flags |= MDS_OPEN_LOCK;
++
++ rc = mdc_intent_lock(sbi->ll_mdc_exp, &data, lmm, lmmsize, itp,
++ 0 /*unused */, &req, ll_mdc_blocking_ast, 0);
++ if (rc == -ESTALE) {
++ /* reason for keep own exit path - don`t flood log
++ * with messages with -ESTALE errors.
++ */
++ if (!it_disposition(itp, DISP_OPEN_OPEN) ||
++ it_open_error(DISP_OPEN_OPEN, itp))
++ GOTO(out, rc);
++ ll_release_openhandle(file->f_dentry, itp);
++ GOTO(out, rc);
++ }
++
++ if (rc != 0 || it_open_error(DISP_OPEN_OPEN, itp)) {
++ rc = rc ? rc : it_open_error(DISP_OPEN_OPEN, itp);
++ CDEBUG(D_VFSTRACE, "lock enqueue: err: %d\n", rc);
++ GOTO(out, rc);
++ }
++
++ if (itp->d.lustre.it_lock_mode)
++ mdc_set_lock_data(&itp->d.lustre.it_lock_handle,
++ inode);
++
++ rc = ll_prep_inode(sbi->ll_osc_exp, &file->f_dentry->d_inode,
++ req, DLM_REPLY_REC_OFF, NULL);
++out:
++ ptlrpc_req_finished(itp->d.lustre.it_data);
++ it_clear_disposition(itp, DISP_ENQ_COMPLETE);
++ ll_intent_drop_lock(itp);
++
++ RETURN(rc);
++}
++
++
++static void ll_och_fill(struct ll_inode_info *lli, struct lookup_intent *it,
++ struct obd_client_handle *och)
++{
++ struct ptlrpc_request *req = it->d.lustre.it_data;
++ struct mds_body *body;
++
++ LASSERT(och);
++
++ body = lustre_msg_buf(req->rq_repmsg, DLM_REPLY_REC_OFF, sizeof(*body));
++ LASSERT(body != NULL); /* reply already checked out */
++ /* and swabbed in mdc_enqueue */
++ LASSERT(lustre_rep_swabbed(req, DLM_REPLY_REC_OFF));
++
++ memcpy(&och->och_fh, &body->handle, sizeof(body->handle));
++ och->och_magic = OBD_CLIENT_HANDLE_MAGIC;
++ lli->lli_io_epoch = body->io_epoch;
++
++ mdc_set_open_replay_data(och, it->d.lustre.it_data);
++}
++
++int ll_local_open(struct file *file, struct lookup_intent *it,
++ struct ll_file_data *fd, struct obd_client_handle *och)
++{
++ ENTRY;
++
++ LASSERT(!LUSTRE_FPRIVATE(file));
++
++ LASSERT(fd != NULL);
++
++ if (och)
++ ll_och_fill(ll_i2info(file->f_dentry->d_inode), it, och);
++ LUSTRE_FPRIVATE(file) = fd;
++ ll_readahead_init(file->f_dentry->d_inode, &fd->fd_ras);
++ fd->fd_omode = it->it_flags;
++
++ RETURN(0);
++}
++
++/* Open a file, and (for the very first open) create objects on the OSTs at
++ * this time. If opened with O_LOV_DELAY_CREATE, then we don't do the object
++ * creation or open until ll_lov_setstripe() ioctl is called. We grab
++ * lli_open_sem to ensure no other process will create objects, send the
++ * stripe MD to the MDS, or try to destroy the objects if that fails.
++ *
++ * If we already have the stripe MD locally then we don't request it in
++ * mdc_open(), by passing a lmm_size = 0.
++ *
++ * It is up to the application to ensure no other processes open this file
++ * in the O_LOV_DELAY_CREATE case, or the default striping pattern will be
++ * used. We might be able to avoid races of that sort by getting lli_open_sem
++ * before returning in the O_LOV_DELAY_CREATE case and dropping it here
++ * or in ll_file_release(), but I'm not sure that is desirable/necessary.
++ */
++int ll_file_open(struct inode *inode, struct file *file)
++{
++ struct ll_inode_info *lli = ll_i2info(inode);
++ struct lookup_intent *it, oit = { .it_op = IT_OPEN,
++ .it_flags = file->f_flags };
++ struct lov_stripe_md *lsm;
++ struct ptlrpc_request *req = NULL;
++ struct obd_client_handle **och_p;
++ __u64 *och_usecount;
++ struct ll_file_data *fd;
++ int rc = 0, opendir_set = 0;
++ ENTRY;
++
++ CDEBUG(D_VFSTRACE, "VFS Op:inode=%lu/%u(%p), flags %o\n", inode->i_ino,
++ inode->i_generation, inode, file->f_flags);
++
++#ifdef HAVE_VFS_INTENT_PATCHES
++ it = file->f_it;
++#else
++ it = file->private_data; /* XXX: compat macro */
++ file->private_data = NULL; /* prevent ll_local_open assertion */
++#endif
++
++ fd = ll_file_data_get();
++ if (fd == NULL)
++ RETURN(-ENOMEM);
++
++ if (S_ISDIR(inode->i_mode)) {
++again:
++ spin_lock(&lli->lli_lock);
++ if (lli->lli_opendir_key == NULL && lli->lli_opendir_pid == 0) {
++ LASSERT(lli->lli_sai == NULL);
++ lli->lli_opendir_key = fd;
++ lli->lli_opendir_pid = cfs_curproc_pid();
++ opendir_set = 1;
++ } else if (unlikely(lli->lli_opendir_pid == cfs_curproc_pid() &&
++ lli->lli_opendir_key != NULL)) {
++ /* Two cases for this:
++ * (1) The same process open such directory many times.
++ * (2) The old process opened the directory, and exited
++ * before its children processes. Then new process
++ * with the same pid opens such directory before the
++ * old process's children processes exit.
++ * reset stat ahead for such cases. */
++ spin_unlock(&lli->lli_lock);
++ CDEBUG(D_INFO, "Conflict statahead for %.*s %lu/%u"
++ " reset it.\n", file->f_dentry->d_name.len,
++ file->f_dentry->d_name.name,
++ inode->i_ino, inode->i_generation);
++ ll_stop_statahead(inode, lli->lli_opendir_key);
++ goto again;
++ }
++ spin_unlock(&lli->lli_lock);
++ }
++
++ if (inode->i_sb->s_root == file->f_dentry) {
++ LUSTRE_FPRIVATE(file) = fd;
++ RETURN(0);
++ }
++
++ if (!it || !it->d.lustre.it_disposition) {
++ /* Convert f_flags into access mode. We cannot use file->f_mode,
++ * because everything but O_ACCMODE mask was stripped from it */
++ if ((oit.it_flags + 1) & O_ACCMODE)
++ oit.it_flags++;
++ if (file->f_flags & O_TRUNC)
++ oit.it_flags |= FMODE_WRITE;
++
++ /* kernel only call f_op->open in dentry_open. filp_open calls
++ * dentry_open after call to open_namei that checks permissions.
++ * Only nfsd_open call dentry_open directly without checking
++ * permissions and because of that this code below is safe. */
++ if (oit.it_flags & FMODE_WRITE)
++ oit.it_flags |= MDS_OPEN_OWNEROVERRIDE;
++
++ /* We do not want O_EXCL here, presumably we opened the file
++ * already? XXX - NFS implications? */
++ oit.it_flags &= ~O_EXCL;
++
++ it = &oit;
++ }
++
++restart:
++ /* Let's see if we have file open on MDS already. */
++ if (it->it_flags & FMODE_WRITE) {
++ och_p = &lli->lli_mds_write_och;
++ och_usecount = &lli->lli_open_fd_write_count;
++ } else if (it->it_flags & FMODE_EXEC) {
++ och_p = &lli->lli_mds_exec_och;
++ och_usecount = &lli->lli_open_fd_exec_count;
++ } else {
++ och_p = &lli->lli_mds_read_och;
++ och_usecount = &lli->lli_open_fd_read_count;
++ }
++
++ LASSERTF(it->it_flags != 0, "it %p dist %d \n", it,
++ it->d.lustre.it_disposition);
++
++ down(&lli->lli_och_sem);
++ if (*och_p) { /* Open handle is present */
++ if (it_disposition(it, DISP_OPEN_OPEN)) {
++ /* Well, there's extra open request that we do not need,
++ let's close it somehow. This will decref request. */
++ rc = it_open_error(DISP_OPEN_OPEN, it);
++ if (rc) {
++ up(&lli->lli_och_sem);
++ ll_file_data_put(fd);
++ GOTO(out_openerr, rc);
++ }
++ ll_release_openhandle(file->f_dentry, it);
++ lprocfs_counter_incr(ll_i2sbi(inode)->ll_stats,
++ LPROC_LL_OPEN);
++ }
++ (*och_usecount)++;
++
++ rc = ll_local_open(file, it, fd, NULL);
++
++ LASSERTF(rc == 0, "rc = %d\n", rc);
++ } else {
++ LASSERT(*och_usecount == 0);
++ if (!it->d.lustre.it_disposition) {
++ /* We cannot just request lock handle now, new ELC code
++ means that one of other OPEN locks for this file
++ could be cancelled, and since blocking ast handler
++ would attempt to grab och_sem as well, that would
++ result in a deadlock */
++ up(&lli->lli_och_sem);
++ it->it_flags |= O_CHECK_STALE;
++ rc = ll_intent_file_open(file, NULL, 0, it);
++ it->it_flags &= ~O_CHECK_STALE;
++ if (rc) {
++ ll_file_data_put(fd);
++ GOTO(out_openerr, rc);
++ }
++
++ mdc_set_lock_data(&it->d.lustre.it_lock_handle,
++ file->f_dentry->d_inode);
++ goto restart;
++ }
++
++ OBD_ALLOC(*och_p, sizeof (struct obd_client_handle));
++ if (!*och_p) {
++ ll_file_data_put(fd);
++ GOTO(out_och_free, rc = -ENOMEM);
++ }
++ (*och_usecount)++;
++ req = it->d.lustre.it_data;
++
++ /* mdc_intent_lock() didn't get a request ref if there was an
++ * open error, so don't do cleanup on the request here
++ * (bug 3430) */
++ /* XXX (green): Should not we bail out on any error here, not
++ * just open error? */
++ rc = it_open_error(DISP_OPEN_OPEN, it);
++ if (rc) {
++ ll_file_data_put(fd);
++ GOTO(out_och_free, rc);
++ }
++
++ ll_stats_ops_tally(ll_i2sbi(inode), LPROC_LL_OPEN, 1);
++ rc = ll_local_open(file, it, fd, *och_p);
++ LASSERTF(rc == 0, "rc = %d\n", rc);
++ }
++ up(&lli->lli_och_sem);
++
++ /* Must do this outside lli_och_sem lock to prevent deadlock where
++ different kind of OPEN lock for this same inode gets cancelled
++ by ldlm_cancel_lru */
++ if (!S_ISREG(inode->i_mode))
++ GOTO(out, rc);
++
++ lsm = lli->lli_smd;
++ if (lsm == NULL) {
++ if (file->f_flags & O_LOV_DELAY_CREATE ||
++ !(file->f_mode & FMODE_WRITE)) {
++ CDEBUG(D_INODE, "object creation was delayed\n");
++ GOTO(out, rc);
++ }
++ }
++ file->f_flags &= ~O_LOV_DELAY_CREATE;
++ GOTO(out, rc);
++ out:
++ ptlrpc_req_finished(req);
++ if (req)
++ it_clear_disposition(it, DISP_ENQ_OPEN_REF);
++ if (rc == 0) {
++ ll_open_complete(inode);
++ } else {
++out_och_free:
++ if (*och_p) {
++ OBD_FREE(*och_p, sizeof (struct obd_client_handle));
++ *och_p = NULL; /* OBD_FREE writes some magic there */
++ (*och_usecount)--;
++ }
++ up(&lli->lli_och_sem);
++out_openerr:
++ if (opendir_set != 0)
++ ll_stop_statahead(inode, lli->lli_opendir_key);
++ }
++
++ return rc;
++}
++
++/* Fills the obdo with the attributes for the inode defined by lsm */
++int ll_lsm_getattr(struct obd_export *exp, struct lov_stripe_md *lsm,
++ struct obdo *oa)
++{
++ struct ptlrpc_request_set *set;
++ struct obd_info oinfo = { { { 0 } } };
++ int rc;
++ ENTRY;
++
++ LASSERT(lsm != NULL);
++
++ memset(oa, 0, sizeof *oa);
++ oinfo.oi_md = lsm;
++ oinfo.oi_oa = oa;
++ oa->o_id = lsm->lsm_object_id;
++ oa->o_gr = lsm->lsm_object_gr;
++ oa->o_mode = S_IFREG;
++ oa->o_valid = OBD_MD_FLID | OBD_MD_FLTYPE | OBD_MD_FLSIZE |
++ OBD_MD_FLBLOCKS | OBD_MD_FLBLKSZ | OBD_MD_FLMTIME |
++ OBD_MD_FLCTIME | OBD_MD_FLGROUP;
++
++ set = ptlrpc_prep_set();
++ if (set == NULL) {
++ rc = -ENOMEM;
++ } else {
++ rc = obd_getattr_async(exp, &oinfo, set);
++ if (rc == 0)
++ rc = ptlrpc_set_wait(set);
++ ptlrpc_set_destroy(set);
++ }
++ if (rc)
++ RETURN(rc);
++
++ oa->o_valid &= (OBD_MD_FLBLOCKS | OBD_MD_FLBLKSZ | OBD_MD_FLMTIME |
++ OBD_MD_FLCTIME | OBD_MD_FLSIZE);
++ RETURN(0);
++}
++
++static int ll_lock_to_stripe_offset(struct inode *inode, struct ldlm_lock *lock)
++{
++ struct ll_inode_info *lli = ll_i2info(inode);
++ struct lov_stripe_md *lsm = lli->lli_smd;
++ struct obd_export *exp = ll_i2obdexp(inode);
++ struct {
++ char name[16];
++ struct ldlm_lock *lock;
++ } key = { .name = KEY_LOCK_TO_STRIPE, .lock = lock };
++ __u32 stripe, vallen = sizeof(stripe);
++ struct lov_oinfo *loinfo;
++ int rc;
++ ENTRY;
++
++ if (lsm->lsm_stripe_count == 1)
++ GOTO(check, stripe = 0);
++
++ /* get our offset in the lov */
++ rc = obd_get_info(exp, sizeof(key), &key, &vallen, &stripe, lsm);
++ if (rc != 0) {
++ CERROR("obd_get_info: rc = %d\n", rc);
++ RETURN(rc);
++ }
++ LASSERT(stripe < lsm->lsm_stripe_count);
++
++check:
++ loinfo = lsm->lsm_oinfo[stripe];
++ if (!osc_res_name_eq(loinfo->loi_id, loinfo->loi_gr,
++ &lock->l_resource->lr_name)) {
++ LDLM_ERROR(lock, "resource doesn't match object "LPU64"/"LPU64,
++ loinfo->loi_id, loinfo->loi_gr);
++ RETURN(-ELDLM_NO_LOCK_DATA);
++ }
++
++ RETURN(stripe);
++}
++
++/* Get extra page reference to ensure it is not going away */
++void ll_pin_extent_cb(void *data)
++{
++ struct page *page = data;
++
++ page_cache_get(page);
++
++ return;
++}
++/* Flush the page from page cache for an extent as its canceled.
++ * Page to remove is delivered as @data.
++ *
++ * No one can dirty the extent until we've finished our work and they cannot
++ * enqueue another lock. The DLM protects us from ll_file_read/write here,
++ * but other kernel actors could have pages locked.
++ *
++ * If @discard is set, there is no need to write the page if it is dirty.
++ *
++ * Called with the DLM lock held. */
++int ll_page_removal_cb(void *data, int discard)
++{
++ int rc;
++ struct page *page = data;
++ struct address_space *mapping;
++
++ ENTRY;
++
++ /* We have page reference already from ll_pin_page */
++ lock_page(page);
++
++ /* Already truncated by somebody */
++ if (!page->mapping)
++ GOTO(out, rc = 0);
++
++ mapping = page->mapping;
++
++ ll_teardown_mmaps(mapping,
++ (__u64)page->index << PAGE_CACHE_SHIFT,
++ ((__u64)page->index<<PAGE_CACHE_SHIFT)|
++ ~PAGE_CACHE_MASK);
++ LL_CDEBUG_PAGE(D_PAGE, page, "removing page\n");
++ if (!discard && PageWriteback(page))
++ wait_on_page_writeback(page);
++
++ if (!discard && clear_page_dirty_for_io(page)) {
++ rc = ll_call_writepage(page->mapping->host, page);
++ /* either waiting for io to complete or reacquiring
++ * the lock that the failed writepage released */
++ lock_page(page);
++ wait_on_page_writeback(page);
++ if (rc < 0) {
++ CERROR("writepage inode %lu(%p) of page %p "
++ "failed: %d\n", mapping->host->i_ino,
++ mapping->host, page, rc);
++ if (rc == -ENOSPC)
++ set_bit(AS_ENOSPC, &mapping->flags);
++ else
++ set_bit(AS_EIO, &mapping->flags);
++ }
++ }
++ if (page->mapping != NULL) {
++ struct ll_async_page *llap = llap_cast_private(page);
++ // checking again to account for writeback's lock_page()
++ LL_CDEBUG_PAGE(D_PAGE, page, "truncating\n");
++ if (llap)
++ ll_ra_accounting(llap, page->mapping);
++ ll_truncate_complete_page(page);
++ }
++ EXIT;
++out:
++ LASSERT(!PageWriteback(page));
++ unlock_page(page);
++ page_cache_release(page);
++
++ return 0;
++}
++
++int ll_extent_lock_cancel_cb(struct ldlm_lock *lock, struct ldlm_lock_desc *new,
++ void *data, int flag)
++{
++ struct inode *inode;
++ struct ll_inode_info *lli;
++ struct lov_stripe_md *lsm;
++ int stripe;
++ __u64 kms;
++
++ ENTRY;
++
++ if ((unsigned long)data > 0 && (unsigned long)data < 0x1000) {
++ LDLM_ERROR(lock, "cancelling lock with bad data %p", data);
++ LBUG();
++ }
++
++ inode = ll_inode_from_lock(lock);
++ if (inode == NULL)
++ RETURN(0);
++ lli = ll_i2info(inode);
++ if (lli == NULL)
++ GOTO(iput, 0);
++ if (lli->lli_smd == NULL)
++ GOTO(iput, 0);
++ lsm = lli->lli_smd;
++
++ stripe = ll_lock_to_stripe_offset(inode, lock);
++ if (stripe < 0)
++ GOTO(iput, 0);
++
++ lov_stripe_lock(lsm);
++ lock_res_and_lock(lock);
++ kms = ldlm_extent_shift_kms(lock,
++ lsm->lsm_oinfo[stripe]->loi_kms);
++
++ if (lsm->lsm_oinfo[stripe]->loi_kms != kms)
++ LDLM_DEBUG(lock, "updating kms from "LPU64" to "LPU64,
++ lsm->lsm_oinfo[stripe]->loi_kms, kms);
++ lsm->lsm_oinfo[stripe]->loi_kms = kms;
++ unlock_res_and_lock(lock);
++ lov_stripe_unlock(lsm);
++ ll_try_done_writing(inode);
++ EXIT;
++iput:
++ iput(inode);
++
++ return 0;
++}
++
++#if 0
++int ll_async_completion_ast(struct ldlm_lock *lock, int flags, void *data)
++{
++ /* XXX ALLOCATE - 160 bytes */
++ struct inode *inode = ll_inode_from_lock(lock);
++ struct ll_inode_info *lli = ll_i2info(inode);
++ struct lustre_handle lockh = { 0 };
++ struct ost_lvb *lvb;
++ int stripe;
++ ENTRY;
++
++ if (flags & (LDLM_FL_BLOCK_WAIT | LDLM_FL_BLOCK_GRANTED |
++ LDLM_FL_BLOCK_CONV)) {
++ LBUG(); /* not expecting any blocked async locks yet */
++ LDLM_DEBUG(lock, "client-side async enqueue returned a blocked "
++ "lock, returning");
++ ldlm_lock_dump(D_OTHER, lock, 0);
++ ldlm_reprocess_all(lock->l_resource);
++ RETURN(0);
++ }
++
++ LDLM_DEBUG(lock, "client-side async enqueue: granted/glimpsed");
++
++ stripe = ll_lock_to_stripe_offset(inode, lock);
++ if (stripe < 0)
++ goto iput;
++
++ if (lock->l_lvb_len) {
++ struct lov_stripe_md *lsm = lli->lli_smd;
++ __u64 kms;
++ lvb = lock->l_lvb_data;
++ lsm->lsm_oinfo[stripe].loi_rss = lvb->lvb_size;
++
++ lock_res_and_lock(lock);
++ ll_inode_size_lock(inode, 1);
++ kms = MAX(lsm->lsm_oinfo[stripe].loi_kms, lvb->lvb_size);
++ kms = ldlm_extent_shift_kms(NULL, kms);
++ if (lsm->lsm_oinfo[stripe].loi_kms != kms)
++ LDLM_DEBUG(lock, "updating kms from "LPU64" to "LPU64,
++ lsm->lsm_oinfo[stripe].loi_kms, kms);
++ lsm->lsm_oinfo[stripe].loi_kms = kms;
++ ll_inode_size_unlock(inode, 1);
++ unlock_res_and_lock(lock);
++ }
++
++iput:
++ iput(inode);
++ wake_up(&lock->l_waitq);
++
++ ldlm_lock2handle(lock, &lockh);
++ ldlm_lock_decref(&lockh, LCK_PR);
++ RETURN(0);
++}
++#endif
++
++static int ll_glimpse_callback(struct ldlm_lock *lock, void *reqp)
++{
++ struct ptlrpc_request *req = reqp;
++ struct inode *inode = ll_inode_from_lock(lock);
++ struct ll_inode_info *lli;
++ struct lov_stripe_md *lsm;
++ struct ost_lvb *lvb;
++ int rc, stripe;
++ int size[2] = { sizeof(struct ptlrpc_body), sizeof(*lvb) };
++ ENTRY;
++
++ if (inode == NULL)
++ GOTO(out, rc = -ELDLM_NO_LOCK_DATA);
++ lli = ll_i2info(inode);
++ if (lli == NULL)
++ GOTO(iput, rc = -ELDLM_NO_LOCK_DATA);
++ lsm = lli->lli_smd;
++ if (lsm == NULL)
++ GOTO(iput, rc = -ELDLM_NO_LOCK_DATA);
++
++ /* First, find out which stripe index this lock corresponds to. */
++ stripe = ll_lock_to_stripe_offset(inode, lock);
++ if (stripe < 0)
++ GOTO(iput, rc = -ELDLM_NO_LOCK_DATA);
++
++ rc = lustre_pack_reply(req, 2, size, NULL);
++ if (rc)
++ GOTO(iput, rc);
++
++ lvb = lustre_msg_buf(req->rq_repmsg, REPLY_REC_OFF, sizeof(*lvb));
++ lvb->lvb_size = lli->lli_smd->lsm_oinfo[stripe]->loi_kms;
++ lvb->lvb_mtime = LTIME_S(inode->i_mtime);
++ lvb->lvb_atime = LTIME_S(inode->i_atime);
++ lvb->lvb_ctime = LTIME_S(inode->i_ctime);
++
++ LDLM_DEBUG(lock, "i_size: %llu -> stripe number %u -> kms "LPU64
++ " atime "LPU64", mtime "LPU64", ctime "LPU64,
++ i_size_read(inode), stripe, lvb->lvb_size, lvb->lvb_atime,
++ lvb->lvb_mtime, lvb->lvb_ctime);
++ iput:
++ iput(inode);
++
++ out:
++ /* These errors are normal races, so we don't want to fill the console
++ * with messages by calling ptlrpc_error() */
++ if (rc == -ELDLM_NO_LOCK_DATA)
++ lustre_pack_reply(req, 1, NULL, NULL);
++
++ req->rq_status = rc;
++ return rc;
++}
++
++int ll_glimpse_ioctl(struct ll_sb_info *sbi, struct lov_stripe_md *lsm,
++ lstat_t *st)
++{
++ struct lustre_handle lockh = { 0 };
++ struct ldlm_enqueue_info einfo = { 0 };
++ struct obd_info oinfo = { { { 0 } } };
++ struct ost_lvb lvb;
++ int rc;
++
++ ENTRY;
++
++ einfo.ei_type = LDLM_EXTENT;
++ einfo.ei_mode = LCK_PR;
++ einfo.ei_cb_bl = osc_extent_blocking_cb;
++ einfo.ei_cb_cp = ldlm_completion_ast;
++ einfo.ei_cb_gl = ll_glimpse_callback;
++ einfo.ei_cbdata = NULL;
++
++ oinfo.oi_policy.l_extent.end = OBD_OBJECT_EOF;
++ oinfo.oi_lockh = &lockh;
++ oinfo.oi_md = lsm;
++ oinfo.oi_flags = LDLM_FL_HAS_INTENT;
++
++ rc = obd_enqueue_rqset(sbi->ll_osc_exp, &oinfo, &einfo);
++ if (rc == -ENOENT)
++ RETURN(rc);
++ if (rc != 0) {
++ CERROR("obd_enqueue returned rc %d, "
++ "returning -EIO\n", rc);
++ RETURN(rc > 0 ? -EIO : rc);
++ }
++
++ lov_stripe_lock(lsm);
++ memset(&lvb, 0, sizeof(lvb));
++ obd_merge_lvb(sbi->ll_osc_exp, lsm, &lvb, 0);
++ st->st_size = lvb.lvb_size;
++ st->st_blocks = lvb.lvb_blocks;
++ st->st_mtime = lvb.lvb_mtime;
++ st->st_atime = lvb.lvb_atime;
++ st->st_ctime = lvb.lvb_ctime;
++ lov_stripe_unlock(lsm);
++
++ RETURN(rc);
++}
++
++/* NB: obd_merge_lvb will prefer locally cached writes if they extend the
++ * file (because it prefers KMS over RSS when larger) */
++int ll_glimpse_size(struct inode *inode, int ast_flags)
++{
++ struct ll_inode_info *lli = ll_i2info(inode);
++ struct ll_sb_info *sbi = ll_i2sbi(inode);
++ struct lustre_handle lockh = { 0 };
++ struct ldlm_enqueue_info einfo = { 0 };
++ struct obd_info oinfo = { { { 0 } } };
++ struct ost_lvb lvb;
++ int rc;
++ ENTRY;
++
++ CDEBUG(D_DLMTRACE, "Glimpsing inode %lu\n", inode->i_ino);
++
++ if (!lli->lli_smd) {
++ CDEBUG(D_DLMTRACE, "No objects for inode %lu\n", inode->i_ino);
++ RETURN(0);
++ }
++
++ /* NOTE: this looks like DLM lock request, but it may not be one. Due
++ * to LDLM_FL_HAS_INTENT flag, this is glimpse request, that
++ * won't revoke any conflicting DLM locks held. Instead,
++ * ll_glimpse_callback() will be called on each client
++ * holding a DLM lock against this file, and resulting size
++ * will be returned for each stripe. DLM lock on [0, EOF] is
++ * acquired only if there were no conflicting locks. */
++ einfo.ei_type = LDLM_EXTENT;
++ einfo.ei_mode = LCK_PR;
++ einfo.ei_cb_bl = osc_extent_blocking_cb;
++ einfo.ei_cb_cp = ldlm_completion_ast;
++ einfo.ei_cb_gl = ll_glimpse_callback;
++ einfo.ei_cbdata = inode;
++
++ oinfo.oi_policy.l_extent.end = OBD_OBJECT_EOF;
++ oinfo.oi_lockh = &lockh;
++ oinfo.oi_md = lli->lli_smd;
++ oinfo.oi_flags = ast_flags | LDLM_FL_HAS_INTENT;
++
++ rc = obd_enqueue_rqset(sbi->ll_osc_exp, &oinfo, &einfo);
++ if (rc == -ENOENT)
++ RETURN(rc);
++ if (rc != 0) {
++ CERROR("obd_enqueue returned rc %d, returning -EIO\n", rc);
++ RETURN(rc > 0 ? -EIO : rc);
++ }
++
++ ll_inode_size_lock(inode, 1);
++ inode_init_lvb(inode, &lvb);
++ /* merge timestamps the most resently obtained from mds with
++ timestamps obtained from osts */
++ lvb.lvb_atime = lli->lli_lvb.lvb_atime;
++ lvb.lvb_mtime = lli->lli_lvb.lvb_mtime;
++ lvb.lvb_ctime = lli->lli_lvb.lvb_ctime;
++ rc = obd_merge_lvb(sbi->ll_osc_exp, lli->lli_smd, &lvb, 0);
++ i_size_write(inode, lvb.lvb_size);
++ inode->i_blocks = lvb.lvb_blocks;
++ LTIME_S(inode->i_mtime) = lvb.lvb_mtime;
++ LTIME_S(inode->i_atime) = lvb.lvb_atime;
++ LTIME_S(inode->i_ctime) = lvb.lvb_ctime;
++ ll_inode_size_unlock(inode, 1);
++
++ CDEBUG(D_DLMTRACE, "glimpse: size: %llu, blocks: %llu\n",
++ i_size_read(inode), (long long)inode->i_blocks);
++
++ RETURN(rc);
++}
++
++int ll_extent_lock(struct ll_file_data *fd, struct inode *inode,
++ struct lov_stripe_md *lsm, int mode,
++ ldlm_policy_data_t *policy, struct lustre_handle *lockh,
++ int ast_flags)
++{
++ struct ll_sb_info *sbi = ll_i2sbi(inode);
++ struct ost_lvb lvb;
++ struct ldlm_enqueue_info einfo = { 0 };
++ struct obd_info oinfo = { { { 0 } } };
++ int rc;
++ ENTRY;
++
++ LASSERT(!lustre_handle_is_used(lockh));
++ LASSERT(lsm != NULL);
++
++ /* don't drop the mmapped file to LRU */
++ if (mapping_mapped(inode->i_mapping))
++ ast_flags |= LDLM_FL_NO_LRU;
++
++ /* XXX phil: can we do this? won't it screw the file size up? */
++ if ((fd && (fd->fd_flags & LL_FILE_IGNORE_LOCK)) ||
++ (sbi->ll_flags & LL_SBI_NOLCK))
++ RETURN(0);
++
++ CDEBUG(D_DLMTRACE, "Locking inode %lu, start "LPU64" end "LPU64"\n",
++ inode->i_ino, policy->l_extent.start, policy->l_extent.end);
++
++ einfo.ei_type = LDLM_EXTENT;
++ einfo.ei_mode = mode;
++ einfo.ei_cb_bl = osc_extent_blocking_cb;
++ einfo.ei_cb_cp = ldlm_completion_ast;
++ einfo.ei_cb_gl = ll_glimpse_callback;
++ einfo.ei_cbdata = inode;
++
++ oinfo.oi_policy = *policy;
++ oinfo.oi_lockh = lockh;
++ oinfo.oi_md = lsm;
++ oinfo.oi_flags = ast_flags;
++
++ rc = obd_enqueue(sbi->ll_osc_exp, &oinfo, &einfo, NULL);
++ *policy = oinfo.oi_policy;
++ if (rc > 0)
++ rc = -EIO;
++
++ ll_inode_size_lock(inode, 1);
++ inode_init_lvb(inode, &lvb);
++ obd_merge_lvb(sbi->ll_osc_exp, lsm, &lvb, 1);
++
++ if (policy->l_extent.start == 0 &&
++ policy->l_extent.end == OBD_OBJECT_EOF) {
++ /* vmtruncate()->ll_truncate() first sets the i_size and then
++ * the kms under both a DLM lock and the
++ * ll_inode_size_lock(). If we don't get the
++ * ll_inode_size_lock() here we can match the DLM lock and
++ * reset i_size from the kms before the truncating path has
++ * updated the kms. generic_file_write can then trust the
++ * stale i_size when doing appending writes and effectively
++ * cancel the result of the truncate. Getting the
++ * ll_inode_size_lock() after the enqueue maintains the DLM
++ * -> ll_inode_size_lock() acquiring order. */
++ i_size_write(inode, lvb.lvb_size);
++ CDEBUG(D_INODE, "inode=%lu, updating i_size %llu\n",
++ inode->i_ino, i_size_read(inode));
++ }
++
++ if (rc == 0) {
++ LTIME_S(inode->i_mtime) = lvb.lvb_mtime;
++ LTIME_S(inode->i_atime) = lvb.lvb_atime;
++ LTIME_S(inode->i_ctime) = lvb.lvb_ctime;
++ }
++ ll_inode_size_unlock(inode, 1);
++
++ RETURN(rc);
++}
++
++int ll_extent_unlock(struct ll_file_data *fd, struct inode *inode,
++ struct lov_stripe_md *lsm, int mode,
++ struct lustre_handle *lockh)
++{
++ struct ll_sb_info *sbi = ll_i2sbi(inode);
++ int rc;
++ ENTRY;
++
++ /* XXX phil: can we do this? won't it screw the file size up? */
++ if ((fd && (fd->fd_flags & LL_FILE_IGNORE_LOCK)) ||
++ (sbi->ll_flags & LL_SBI_NOLCK))
++ RETURN(0);
++
++ rc = obd_cancel(sbi->ll_osc_exp, lsm, mode, lockh);
++
++ RETURN(rc);
++}
++
++static void ll_set_file_contended(struct inode *inode)
++{
++ struct ll_inode_info *lli = ll_i2info(inode);
++
++ lli->lli_contention_time = cfs_time_current();
++ set_bit(LLI_F_CONTENDED, &lli->lli_flags);
++}
++
++void ll_clear_file_contended(struct inode *inode)
++{
++ struct ll_inode_info *lli = ll_i2info(inode);
++
++ clear_bit(LLI_F_CONTENDED, &lli->lli_flags);
++}
++
++static int ll_is_file_contended(struct file *file)
++{
++ struct inode *inode = file->f_dentry->d_inode;
++ struct ll_inode_info *lli = ll_i2info(inode);
++ struct ll_sb_info *sbi = ll_i2sbi(inode);
++ struct ll_file_data *fd = LUSTRE_FPRIVATE(file);
++ ENTRY;
++
++ if (!(sbi->ll_lco.lco_flags & OBD_CONNECT_SRVLOCK)) {
++ CDEBUG(D_INFO, "the server does not support SRVLOCK feature,"
++ " osc connect flags = 0x"LPX64"\n",
++ sbi->ll_lco.lco_flags);
++ RETURN(0);
++ }
++ if (fd && (fd->fd_flags & LL_FILE_IGNORE_LOCK))
++ RETURN(0);
++ if (test_bit(LLI_F_CONTENDED, &lli->lli_flags)) {
++ cfs_time_t cur_time = cfs_time_current();
++ cfs_time_t retry_time;
++
++ retry_time = cfs_time_add(
++ lli->lli_contention_time,
++ cfs_time_seconds(sbi->ll_contention_time));
++ if (cfs_time_after(cur_time, retry_time)) {
++ ll_clear_file_contended(inode);
++ RETURN(0);
++ }
++ RETURN(1);
++ }
++ RETURN(0);
++}
++
++static int ll_file_get_tree_lock_iov(struct ll_lock_tree *tree,
++ struct file *file, const struct iovec *iov,
++ unsigned long nr_segs,
++ obd_off start, obd_off end, int rw)
++{
++ int append;
++ int tree_locked = 0;
++ int rc;
++ struct inode * inode = file->f_dentry->d_inode;
++ ENTRY;
++
++ append = (rw == OBD_BRW_WRITE) && (file->f_flags & O_APPEND);
++
++ if (append || !ll_is_file_contended(file)) {
++ struct ll_lock_tree_node *node;
++ int ast_flags;
++
++ ast_flags = append ? 0 : LDLM_FL_DENY_ON_CONTENTION;
++ if (file->f_flags & O_NONBLOCK)
++ ast_flags |= LDLM_FL_BLOCK_NOWAIT;
++ node = ll_node_from_inode(inode, start, end,
++ (rw == OBD_BRW_WRITE) ? LCK_PW : LCK_PR);
++ if (IS_ERR(node)) {
++ rc = PTR_ERR(node);
++ GOTO(out, rc);
++ }
++ tree->lt_fd = LUSTRE_FPRIVATE(file);
++ rc = ll_tree_lock_iov(tree, node, iov, nr_segs, ast_flags);
++ if (rc == 0)
++ tree_locked = 1;
++ else if (rc == -EUSERS)
++ ll_set_file_contended(inode);
++ else
++ GOTO(out, rc);
++ }
++ RETURN(tree_locked);
++out:
++ return rc;
++}
++
++/* XXX: exact copy from kernel code (__generic_file_aio_write_nolock from rhel4)
++ */
++static size_t ll_file_get_iov_count(const struct iovec *iov,
++ unsigned long *nr_segs)
++{
++ size_t count = 0;
++ unsigned long seg;
++
++ for (seg = 0; seg < *nr_segs; seg++) {
++ const struct iovec *iv = &iov[seg];
++
++ /*
++ * If any segment has a negative length, or the cumulative
++ * length ever wraps negative then return -EINVAL.
++ */
++ count += iv->iov_len;
++ if (unlikely((ssize_t)(count|iv->iov_len) < 0))
++ return -EINVAL;
++ if (access_ok(VERIFY_WRITE, iv->iov_base, iv->iov_len))
++ continue;
++ if (seg == 0)
++ return -EFAULT;
++ *nr_segs = seg;
++ count -= iv->iov_len; /* This segment is no good */
++ break;
++ }
++ return count;
++}
++
++static int iov_copy_update(unsigned long *nr_segs, const struct iovec **iov_out,
++ unsigned long *nrsegs_copy,
++ struct iovec *iov_copy, size_t *offset,
++ size_t size)
++{
++ int i;
++ const struct iovec *iov = *iov_out;
++ for (i = 0; i < *nr_segs;
++ i++) {
++ const struct iovec *iv = &iov[i];
++ struct iovec *ivc = &iov_copy[i];
++ *ivc = *iv;
++ if (i == 0) {
++ ivc->iov_len -= *offset;
++ ivc->iov_base += *offset;
++ }
++ if (ivc->iov_len >= size) {
++ ivc->iov_len = size;
++ if (i == 0)
++ *offset += size;
++ else
++ *offset = size;
++ break;
++ }
++ size -= ivc->iov_len;
++ }
++ *iov_out += i;
++ *nr_segs -= i;
++ *nrsegs_copy = i + 1;
++
++ return 0;
++}
++
++static int ll_reget_short_lock(struct page *page, int rw,
++ obd_off start, obd_off end,
++ void **cookie)
++{
++ struct ll_async_page *llap;
++ struct obd_export *exp;
++ struct inode *inode = page->mapping->host;
++
++ ENTRY;
++
++ exp = ll_i2obdexp(inode);
++ if (exp == NULL)
++ RETURN(0);
++
++ llap = llap_cast_private(page);
++ if (llap == NULL)
++ RETURN(0);
++
++ RETURN(obd_reget_short_lock(exp, ll_i2info(inode)->lli_smd,
++ &llap->llap_cookie, rw, start, end,
++ cookie));
++}
++
++static void ll_release_short_lock(struct inode *inode, obd_off end,
++ void *cookie, int rw)
++{
++ struct obd_export *exp;
++ int rc;
++
++ exp = ll_i2obdexp(inode);
++ if (exp == NULL)
++ return;
++
++ rc = obd_release_short_lock(exp, ll_i2info(inode)->lli_smd, end,
++ cookie, rw);
++ if (rc < 0)
++ CERROR("unlock failed (%d)\n", rc);
++}
++
++static inline int ll_file_get_fast_lock(struct file *file,
++ obd_off ppos, obd_off end,
++ const struct iovec *iov,
++ unsigned long nr_segs,
++ void **cookie, int rw)
++{
++ int rc = 0, seg;
++ struct page *page;
++
++ ENTRY;
++
++ /* we would like this read request to be lockfree */
++ for (seg = 0; seg < nr_segs; seg++) {
++ const struct iovec *iv = &iov[seg];
++ if (ll_region_mapped((unsigned long)iv->iov_base, iv->iov_len))
++ GOTO(out, rc);
++ }
++
++ page = find_lock_page(file->f_dentry->d_inode->i_mapping,
++ ppos >> CFS_PAGE_SHIFT);
++ if (page) {
++ if (ll_reget_short_lock(page, rw, ppos, end, cookie))
++ rc = 1;
++
++ unlock_page(page);
++ page_cache_release(page);
++ }
++
++out:
++ RETURN(rc);
++}
++
++static inline void ll_file_put_fast_lock(struct inode *inode, obd_off end,
++ void *cookie, int rw)
++{
++ ll_release_short_lock(inode, end, cookie, rw);
++}
++
++enum ll_lock_style {
++ LL_LOCK_STYLE_NOLOCK = 0,
++ LL_LOCK_STYLE_FASTLOCK = 1,
++ LL_LOCK_STYLE_TREELOCK = 2
++};
++
++static inline int ll_file_get_lock(struct file *file, obd_off ppos,
++ obd_off end, const struct iovec *iov,
++ unsigned long nr_segs, void **cookie,
++ struct ll_lock_tree *tree, int rw)
++{
++ int rc;
++
++ ENTRY;
++
++ if (ll_file_get_fast_lock(file, ppos, end, iov, nr_segs, cookie, rw))
++ RETURN(LL_LOCK_STYLE_FASTLOCK);
++
++ rc = ll_file_get_tree_lock_iov(tree, file, iov, nr_segs,
++ ppos, end, rw);
++ /* rc: 1 for tree lock, 0 for no lock, <0 for error */
++ switch (rc) {
++ case 1:
++ RETURN(LL_LOCK_STYLE_TREELOCK);
++ case 0:
++ RETURN(LL_LOCK_STYLE_NOLOCK);
++ }
++
++ /* an error happened if we reached this point, rc = -errno here */
++ RETURN(rc);
++}
++
++static inline void ll_file_put_lock(struct inode *inode, obd_off end,
++ enum ll_lock_style lock_style,
++ void *cookie, struct ll_lock_tree *tree,
++ int rw)
++
++{
++ switch (lock_style) {
++ case LL_LOCK_STYLE_TREELOCK:
++ ll_tree_unlock(tree);
++ break;
++ case LL_LOCK_STYLE_FASTLOCK:
++ ll_file_put_fast_lock(inode, end, cookie, rw);
++ break;
++ default:
++ CERROR("invalid locking style (%d)\n", lock_style);
++ }
++}
++
++#ifdef HAVE_FILE_READV
++static ssize_t ll_file_readv(struct file *file, const struct iovec *iov,
++ unsigned long nr_segs, loff_t *ppos)
++{
++#else
++static ssize_t ll_file_aio_read(struct kiocb *iocb, const struct iovec *iov,
++ unsigned long nr_segs, loff_t pos)
++{
++ struct file *file = iocb->ki_filp;
++ loff_t *ppos = &iocb->ki_pos;
++#endif
++ struct inode *inode = file->f_dentry->d_inode;
++ struct ll_inode_info *lli = ll_i2info(inode);
++ struct lov_stripe_md *lsm = lli->lli_smd;
++ struct ll_sb_info *sbi = ll_i2sbi(inode);
++ struct ll_lock_tree tree;
++ struct ost_lvb lvb;
++ struct ll_ra_read bead;
++ int ra = 0;
++ obd_off end;
++ ssize_t retval, chunk, sum = 0;
++ int lock_style;
++ struct iovec *iov_copy = NULL;
++ unsigned long nrsegs_copy, nrsegs_orig = 0;
++ size_t count, iov_offset = 0;
++ __u64 kms;
++ void *cookie;
++ ENTRY;
++
++ count = ll_file_get_iov_count(iov, &nr_segs);
++ CDEBUG(D_VFSTRACE, "VFS Op:inode=%lu/%u(%p),size="LPSZ",offset=%Ld\n",
++ inode->i_ino, inode->i_generation, inode, count, *ppos);
++ /* "If nbyte is 0, read() will return 0 and have no other results."
++ * -- Single Unix Spec */
++ if (count == 0)
++ RETURN(0);
++
++ ll_stats_ops_tally(sbi, LPROC_LL_READ_BYTES, count);
++
++ if (!lsm) {
++ /* Read on file with no objects should return zero-filled
++ * buffers up to file size (we can get non-zero sizes with
++ * mknod + truncate, then opening file for read. This is a
++ * common pattern in NFS case, it seems). Bug 6243 */
++ int notzeroed;
++ /* Since there are no objects on OSTs, we have nothing to get
++ * lock on and so we are forced to access inode->i_size
++ * unguarded */
++
++ /* Read beyond end of file */
++ if (*ppos >= i_size_read(inode))
++ RETURN(0);
++
++ if (count > i_size_read(inode) - *ppos)
++ count = i_size_read(inode) - *ppos;
++ /* Make sure to correctly adjust the file pos pointer for
++ * EFAULT case */
++ for (nrsegs_copy = 0; nrsegs_copy < nr_segs; nrsegs_copy++) {
++ const struct iovec *iv = &iov[nrsegs_copy];
++
++ if (count < iv->iov_len)
++ chunk = count;
++ else
++ chunk = iv->iov_len;
++ notzeroed = clear_user(iv->iov_base, chunk);
++ sum += (chunk - notzeroed);
++ count -= (chunk - notzeroed);
++ if (notzeroed || !count)
++ break;
++ }
++ *ppos += sum;
++ if (!sum)
++ RETURN(-EFAULT);
++ RETURN(sum);
++ }
++
++repeat:
++ if (sbi->ll_max_rw_chunk != 0) {
++ /* first, let's know the end of the current stripe */
++ end = *ppos;
++ obd_extent_calc(sbi->ll_osc_exp, lsm, OBD_CALC_STRIPE_END,&end);
++
++ /* correct, the end is beyond the request */
++ if (end > *ppos + count - 1)
++ end = *ppos + count - 1;
++
++ /* and chunk shouldn't be too large even if striping is wide */
++ if (end - *ppos > sbi->ll_max_rw_chunk)
++ end = *ppos + sbi->ll_max_rw_chunk - 1;
++
++ chunk = end - *ppos + 1;
++ if ((count == chunk) && (iov_offset == 0)) {
++ if (iov_copy)
++ OBD_FREE(iov_copy, sizeof(*iov) * nrsegs_orig);
++
++ iov_copy = (struct iovec *)iov;
++ nrsegs_copy = nr_segs;
++ } else {
++ if (!iov_copy) {
++ nrsegs_orig = nr_segs;
++ OBD_ALLOC(iov_copy, sizeof(*iov) * nr_segs);
++ if (!iov_copy)
++ GOTO(out, retval = -ENOMEM);
++ }
++
++ iov_copy_update(&nr_segs, &iov, &nrsegs_copy, iov_copy,
++ &iov_offset, chunk);
++ }
++ } else {
++ end = *ppos + count - 1;
++ iov_copy = (struct iovec *)iov;
++ nrsegs_copy = nr_segs;
++ }
++
++ lock_style = ll_file_get_lock(file, (obd_off)(*ppos), end,
++ iov_copy, nrsegs_copy, &cookie, &tree,
++ OBD_BRW_READ);
++ if (lock_style < 0)
++ GOTO(out, retval = lock_style);
++
++ ll_inode_size_lock(inode, 1);
++ /*
++ * Consistency guarantees: following possibilities exist for the
++ * relation between region being read and real file size at this
++ * moment:
++ *
++ * (A): the region is completely inside of the file;
++ *
++ * (B-x): x bytes of region are inside of the file, the rest is
++ * outside;
++ *
++ * (C): the region is completely outside of the file.
++ *
++ * This classification is stable under DLM lock acquired by
++ * ll_tree_lock() above, because to change class, other client has to
++ * take DLM lock conflicting with our lock. Also, any updates to
++ * ->i_size by other threads on this client are serialized by
++ * ll_inode_size_lock(). This guarantees that short reads are handled
++ * correctly in the face of concurrent writes and truncates.
++ */
++ inode_init_lvb(inode, &lvb);
++ obd_merge_lvb(ll_i2sbi(inode)->ll_osc_exp, lsm, &lvb, 1);
++ kms = lvb.lvb_size;
++ if (*ppos + count - 1 > kms) {
++ /* A glimpse is necessary to determine whether we return a
++ * short read (B) or some zeroes at the end of the buffer (C) */
++ ll_inode_size_unlock(inode, 1);
++ retval = ll_glimpse_size(inode, LDLM_FL_BLOCK_GRANTED);
++ if (retval) {
++ if (lock_style != LL_LOCK_STYLE_NOLOCK)
++ ll_file_put_lock(inode, end, lock_style,
++ cookie, &tree, OBD_BRW_READ);
++ goto out;
++ }
++ } else {
++ /* region is within kms and, hence, within real file size (A).
++ * We need to increase i_size to cover the read region so that
++ * generic_file_read() will do its job, but that doesn't mean
++ * the kms size is _correct_, it is only the _minimum_ size.
++ * If someone does a stat they will get the correct size which
++ * will always be >= the kms value here. b=11081 */
++ if (i_size_read(inode) < kms)
++ i_size_write(inode, kms);
++ ll_inode_size_unlock(inode, 1);
++ }
++
++ chunk = end - *ppos + 1;
++ CDEBUG(D_INODE,"Read ino %lu, "LPSZ" bytes, offset %lld, i_size %llu\n",
++ inode->i_ino, chunk, *ppos, i_size_read(inode));
++
++ /* turn off the kernel's read-ahead */
++ if (lock_style != LL_LOCK_STYLE_NOLOCK) {
++ /* read under locks
++ *
++ * 1. update inode's atime as long as concurrent stat
++ * (via ll_glimpse_size) might bring out-of-date ones
++ *
++ * 2. update lsm so that next stat (via
++ * ll_glimpse_size) could get correct values in lsm */
++ struct ost_lvb xtimes;
++
++ lov_stripe_lock(lsm);
++ LTIME_S(inode->i_atime) = LTIME_S(CURRENT_TIME);
++ xtimes.lvb_atime = LTIME_S(inode->i_atime);
++ obd_update_lvb(sbi->ll_osc_exp, lsm, &xtimes,
++ OBD_MD_FLATIME);
++ lov_stripe_unlock(lsm);
++
++ file->f_ra.ra_pages = 0;
++ /* initialize read-ahead window once per syscall */
++ if (ra == 0) {
++ ra = 1;
++ bead.lrr_start = *ppos >> CFS_PAGE_SHIFT;
++ bead.lrr_count = (count + CFS_PAGE_SIZE - 1) >> CFS_PAGE_SHIFT;
++ ll_ra_read_in(file, &bead);
++ }
++
++ /* BUG: 5972 */
++ file_accessed(file);
++#ifdef HAVE_FILE_READV
++ retval = generic_file_readv(file, iov_copy, nrsegs_copy, ppos);
++#else
++ retval = generic_file_aio_read(iocb, iov_copy, nrsegs_copy,
++ *ppos);
++#endif
++ ll_file_put_lock(inode, end, lock_style, cookie,
++ &tree, OBD_BRW_READ);
++ } else {
++ /* lockless read
++ *
++ * current time will get into request as atime
++ * (lustre/osc/osc_request.c:osc_build_request())
++ */
++ retval = ll_file_lockless_io(file, iov_copy, nrsegs_copy, ppos,
++ READ, chunk);
++ }
++ ll_rw_stats_tally(sbi, current->pid, file, count, 0);
++ if (retval > 0) {
++ count -= retval;
++ sum += retval;
++ if (retval == chunk && count > 0)
++ goto repeat;
++ }
++
++ out:
++ if (ra != 0)
++ ll_ra_read_ex(file, &bead);
++ retval = (sum > 0) ? sum : retval;
++
++ if (iov_copy && iov_copy != iov)
++ OBD_FREE(iov_copy, sizeof(*iov) * nrsegs_orig);
++
++ RETURN(retval);
++}
++
++static ssize_t ll_file_read(struct file *file, char *buf, size_t count,
++ loff_t *ppos)
++{
++ struct iovec local_iov = { .iov_base = (void __user *)buf,
++ .iov_len = count };
++#ifdef HAVE_FILE_READV
++ return ll_file_readv(file, &local_iov, 1, ppos);
++#else
++ struct kiocb kiocb;
++ ssize_t ret;
++
++ init_sync_kiocb(&kiocb, file);
++ kiocb.ki_pos = *ppos;
++ kiocb.ki_left = count;
++
++ ret = ll_file_aio_read(&kiocb, &local_iov, 1, kiocb.ki_pos);
++ *ppos = kiocb.ki_pos;
++ return ret;
++#endif
++}
++
++/*
++ * Write to a file (through the page cache).
++ */
++#ifdef HAVE_FILE_WRITEV
++static ssize_t ll_file_writev(struct file *file, const struct iovec *iov,
++ unsigned long nr_segs, loff_t *ppos)
++{
++#else /* AIO stuff */
++static ssize_t ll_file_aio_write(struct kiocb *iocb, const struct iovec *iov,
++ unsigned long nr_segs, loff_t pos)
++{
++ struct file *file = iocb->ki_filp;
++ loff_t *ppos = &iocb->ki_pos;
++#endif
++ struct inode *inode = file->f_dentry->d_inode;
++ struct ll_sb_info *sbi = ll_i2sbi(inode);
++ struct lov_stripe_md *lsm = ll_i2info(inode)->lli_smd;
++ struct ll_lock_tree tree;
++ loff_t maxbytes = ll_file_maxbytes(inode);
++ loff_t lock_start, lock_end, end;
++ ssize_t retval, chunk, sum = 0;
++ int tree_locked;
++ struct iovec *iov_copy = NULL;
++ unsigned long nrsegs_copy, nrsegs_orig = 0;
++ size_t count, iov_offset = 0;
++ ENTRY;
++
++ count = ll_file_get_iov_count(iov, &nr_segs);
++
++ CDEBUG(D_VFSTRACE, "VFS Op:inode=%lu/%u(%p),size="LPSZ",offset=%Ld\n",
++ inode->i_ino, inode->i_generation, inode, count, *ppos);
++
++ SIGNAL_MASK_ASSERT(); /* XXX BUG 1511 */
++
++ /* POSIX, but surprised the VFS doesn't check this already */
++ if (count == 0)
++ RETURN(0);
++
++ /* If file was opened for LL_IOC_LOV_SETSTRIPE but the ioctl wasn't
++ * called on the file, don't fail the below assertion (bug 2388). */
++ if (file->f_flags & O_LOV_DELAY_CREATE &&
++ ll_i2info(inode)->lli_smd == NULL)
++ RETURN(-EBADF);
++
++ LASSERT(ll_i2info(inode)->lli_smd != NULL);
++
++ down(&ll_i2info(inode)->lli_write_sem);
++
++repeat:
++ chunk = 0; /* just to fix gcc's warning */
++ end = *ppos + count - 1;
++
++ if (file->f_flags & O_APPEND) {
++ lock_start = 0;
++ lock_end = OBD_OBJECT_EOF;
++ iov_copy = (struct iovec *)iov;
++ nrsegs_copy = nr_segs;
++ } else if (sbi->ll_max_rw_chunk != 0) {
++ /* first, let's know the end of the current stripe */
++ end = *ppos;
++ obd_extent_calc(sbi->ll_osc_exp, lsm, OBD_CALC_STRIPE_END,
++ (obd_off *)&end);
++
++ /* correct, the end is beyond the request */
++ if (end > *ppos + count - 1)
++ end = *ppos + count - 1;
++
++ /* and chunk shouldn't be too large even if striping is wide */
++ if (end - *ppos > sbi->ll_max_rw_chunk)
++ end = *ppos + sbi->ll_max_rw_chunk - 1;
++ lock_start = *ppos;
++ lock_end = end;
++ chunk = end - *ppos + 1;
++ if ((count == chunk) && (iov_offset == 0)) {
++ if (iov_copy)
++ OBD_FREE(iov_copy, sizeof(*iov) * nrsegs_orig);
++
++ iov_copy = (struct iovec *)iov;
++ nrsegs_copy = nr_segs;
++ } else {
++ if (!iov_copy) {
++ nrsegs_orig = nr_segs;
++ OBD_ALLOC(iov_copy, sizeof(*iov) * nr_segs);
++ if (!iov_copy)
++ GOTO(out, retval = -ENOMEM);
++ }
++ iov_copy_update(&nr_segs, &iov, &nrsegs_copy, iov_copy,
++ &iov_offset, chunk);
++ }
++ } else {
++ lock_start = *ppos;
++ lock_end = end;
++ iov_copy = (struct iovec *)iov;
++ nrsegs_copy = nr_segs;
++ }
++
++ tree_locked = ll_file_get_tree_lock_iov(&tree, file, iov_copy,
++ nrsegs_copy,
++ (obd_off)lock_start,
++ (obd_off)lock_end,
++ OBD_BRW_WRITE);
++ if (tree_locked < 0)
++ GOTO(out, retval = tree_locked);
++
++ /* This is ok, g_f_w will overwrite this under i_sem if it races
++ * with a local truncate, it just makes our maxbyte checking easier.
++ * The i_size value gets updated in ll_extent_lock() as a consequence
++ * of the [0,EOF] extent lock we requested above. */
++ if (file->f_flags & O_APPEND) {
++ *ppos = i_size_read(inode);
++ end = *ppos + count - 1;
++ }
++
++ if (*ppos >= maxbytes) {
++ send_sig(SIGXFSZ, current, 0);
++ GOTO(out_unlock, retval = -EFBIG);
++ }
++ if (end > maxbytes - 1)
++ end = maxbytes - 1;
++
++ /* generic_file_write handles O_APPEND after getting i_mutex */
++ chunk = end - *ppos + 1;
++ CDEBUG(D_INFO, "Writing inode %lu, "LPSZ" bytes, offset %Lu\n",
++ inode->i_ino, chunk, *ppos);
++ if (tree_locked) {
++ /* write under locks
++ *
++ * 1. update inode's mtime and ctime as long as
++ * concurrent stat (via ll_glimpse_size) might bring
++ * out-of-date ones
++ *
++ * 2. update lsm so that next stat (via
++ * ll_glimpse_size) could get correct values in lsm */
++ struct ost_lvb xtimes;
++
++ lov_stripe_lock(lsm);
++ LTIME_S(inode->i_mtime) = LTIME_S(CURRENT_TIME);
++ LTIME_S(inode->i_ctime) = LTIME_S(CURRENT_TIME);
++ xtimes.lvb_mtime = LTIME_S(inode->i_mtime);
++ xtimes.lvb_ctime = LTIME_S(inode->i_ctime);
++ obd_update_lvb(sbi->ll_osc_exp, lsm, &xtimes,
++ OBD_MD_FLMTIME | OBD_MD_FLCTIME);
++ lov_stripe_unlock(lsm);
++
++#ifdef HAVE_FILE_WRITEV
++ retval = generic_file_writev(file, iov_copy, nrsegs_copy, ppos);
++#else
++ retval = generic_file_aio_write(iocb, iov_copy, nrsegs_copy,
++ *ppos);
++#endif
++ } else {
++ /* lockless write
++ *
++ * current time will get into request as mtime and
++ * ctime (lustre/osc/osc_request.c:osc_build_request())
++ */
++ retval = ll_file_lockless_io(file, iov_copy, nrsegs_copy,
++ ppos, WRITE, chunk);
++ }
++ ll_rw_stats_tally(ll_i2sbi(inode), current->pid, file, chunk, 1);
++
++out_unlock:
++ if (tree_locked)
++ ll_tree_unlock(&tree);
++
++out:
++ if (retval > 0) {
++ count -= retval;
++ sum += retval;
++ if (retval == chunk && count > 0)
++ goto repeat;
++ }
++
++ up(&ll_i2info(inode)->lli_write_sem);
++
++ if (iov_copy && iov_copy != iov)
++ OBD_FREE(iov_copy, sizeof(*iov) * nrsegs_orig);
++
++ retval = (sum > 0) ? sum : retval;
++ ll_stats_ops_tally(ll_i2sbi(inode), LPROC_LL_WRITE_BYTES,
++ retval > 0 ? retval : 0);
++ RETURN(retval);
++}
++
++static ssize_t ll_file_write(struct file *file, const char *buf, size_t count,
++ loff_t *ppos)
++{
++ struct iovec local_iov = { .iov_base = (void __user *)buf,
++ .iov_len = count };
++
++#ifdef HAVE_FILE_WRITEV
++ return ll_file_writev(file, &local_iov, 1, ppos);
++#else
++ struct kiocb kiocb;
++ ssize_t ret;
++
++ init_sync_kiocb(&kiocb, file);
++ kiocb.ki_pos = *ppos;
++ kiocb.ki_left = count;
++
++ ret = ll_file_aio_write(&kiocb, &local_iov, 1, kiocb.ki_pos);
++ *ppos = kiocb.ki_pos;
++
++ return ret;
++#endif
++}
++
++/*
++ * Send file content (through pagecache) somewhere with helper
++ */
++static ssize_t ll_file_sendfile(struct file *in_file, loff_t *ppos,size_t count,
++ read_actor_t actor, void *target)
++{
++ struct inode *inode = in_file->f_dentry->d_inode;
++ struct ll_inode_info *lli = ll_i2info(inode);
++ struct lov_stripe_md *lsm = lli->lli_smd;
++ struct ll_lock_tree tree;
++ struct ll_lock_tree_node *node;
++ struct ost_lvb lvb;
++ struct ll_ra_read bead;
++ int rc;
++ ssize_t retval;
++ __u64 kms;
++ ENTRY;
++ CDEBUG(D_VFSTRACE, "VFS Op:inode=%lu/%u(%p),size="LPSZ",offset=%Ld\n",
++ inode->i_ino, inode->i_generation, inode, count, *ppos);
++
++ /* "If nbyte is 0, read() will return 0 and have no other results."
++ * -- Single Unix Spec */
++ if (count == 0)
++ RETURN(0);
++
++ ll_stats_ops_tally(ll_i2sbi(inode), LPROC_LL_READ_BYTES, count);
++ /* turn off the kernel's read-ahead */
++ in_file->f_ra.ra_pages = 0;
++
++ /* File with no objects, nothing to lock */
++ if (!lsm)
++ RETURN(generic_file_sendfile(in_file, ppos, count, actor, target));
++
++ node = ll_node_from_inode(inode, *ppos, *ppos + count - 1, LCK_PR);
++ if (IS_ERR(node))
++ RETURN(PTR_ERR(node));
++
++ tree.lt_fd = LUSTRE_FPRIVATE(in_file);
++ rc = ll_tree_lock(&tree, node, NULL, count,
++ in_file->f_flags & O_NONBLOCK?LDLM_FL_BLOCK_NOWAIT:0);
++ if (rc != 0)
++ RETURN(rc);
++
++ ll_clear_file_contended(inode);
++ ll_inode_size_lock(inode, 1);
++ /*
++ * Consistency guarantees: following possibilities exist for the
++ * relation between region being read and real file size at this
++ * moment:
++ *
++ * (A): the region is completely inside of the file;
++ *
++ * (B-x): x bytes of region are inside of the file, the rest is
++ * outside;
++ *
++ * (C): the region is completely outside of the file.
++ *
++ * This classification is stable under DLM lock acquired by
++ * ll_tree_lock() above, because to change class, other client has to
++ * take DLM lock conflicting with our lock. Also, any updates to
++ * ->i_size by other threads on this client are serialized by
++ * ll_inode_size_lock(). This guarantees that short reads are handled
++ * correctly in the face of concurrent writes and truncates.
++ */
++ inode_init_lvb(inode, &lvb);
++ obd_merge_lvb(ll_i2sbi(inode)->ll_osc_exp, lsm, &lvb, 1);
++ kms = lvb.lvb_size;
++ if (*ppos + count - 1 > kms) {
++ /* A glimpse is necessary to determine whether we return a
++ * short read (B) or some zeroes at the end of the buffer (C) */
++ ll_inode_size_unlock(inode, 1);
++ retval = ll_glimpse_size(inode, LDLM_FL_BLOCK_GRANTED);
++ if (retval)
++ goto out;
++ } else {
++ /* region is within kms and, hence, within real file size (A) */
++ i_size_write(inode, kms);
++ ll_inode_size_unlock(inode, 1);
++ }
++
++ CDEBUG(D_INFO, "Send ino %lu, "LPSZ" bytes, offset %lld, i_size %llu\n",
++ inode->i_ino, count, *ppos, i_size_read(inode));
++
++ bead.lrr_start = *ppos >> CFS_PAGE_SHIFT;
++ bead.lrr_count = (count + CFS_PAGE_SIZE - 1) >> CFS_PAGE_SHIFT;
++ ll_ra_read_in(in_file, &bead);
++ /* BUG: 5972 */
++ file_accessed(in_file);
++ retval = generic_file_sendfile(in_file, ppos, count, actor, target);
++ ll_ra_read_ex(in_file, &bead);
++
++ out:
++ ll_tree_unlock(&tree);
++ RETURN(retval);
++}
++
++static int ll_lov_recreate_obj(struct inode *inode, struct file *file,
++ unsigned long arg)
++{
++ struct ll_inode_info *lli = ll_i2info(inode);
++ struct obd_export *exp = ll_i2obdexp(inode);
++ struct ll_recreate_obj ucreatp;
++ struct obd_trans_info oti = { 0 };
++ struct obdo *oa = NULL;
++ int lsm_size;
++ int rc = 0;
++ struct lov_stripe_md *lsm, *lsm2;
++ ENTRY;
++
++ if (!cfs_capable(CFS_CAP_SYS_ADMIN))
++ RETURN(-EPERM);
++
++ rc = copy_from_user(&ucreatp, (struct ll_recreate_obj *)arg,
++ sizeof(struct ll_recreate_obj));
++ if (rc) {
++ RETURN(-EFAULT);
++ }
++ OBDO_ALLOC(oa);
++ if (oa == NULL)
++ RETURN(-ENOMEM);
++
++ down(&lli->lli_size_sem);
++ lsm = lli->lli_smd;
++ if (lsm == NULL)
++ GOTO(out, rc = -ENOENT);
++ lsm_size = sizeof(*lsm) + (sizeof(struct lov_oinfo) *
++ (lsm->lsm_stripe_count));
++
++ OBD_ALLOC(lsm2, lsm_size);
++ if (lsm2 == NULL)
++ GOTO(out, rc = -ENOMEM);
++
++ oa->o_id = ucreatp.lrc_id;
++ oa->o_nlink = ucreatp.lrc_ost_idx;
++ oa->o_flags |= OBD_FL_RECREATE_OBJS;
++ oa->o_valid = OBD_MD_FLID | OBD_MD_FLFLAGS;
++ obdo_from_inode(oa, inode, OBD_MD_FLTYPE | OBD_MD_FLATIME |
++ OBD_MD_FLMTIME | OBD_MD_FLCTIME);
++
++ memcpy(lsm2, lsm, lsm_size);
++ rc = obd_create(exp, oa, &lsm2, &oti);
++
++ OBD_FREE(lsm2, lsm_size);
++ GOTO(out, rc);
++out:
++ up(&lli->lli_size_sem);
++ OBDO_FREE(oa);
++ return rc;
++}
++
++int ll_lov_setstripe_ea_info(struct inode *inode, struct file *file,
++ int flags, struct lov_user_md *lum,
++ int lum_size)
++{
++ struct ll_inode_info *lli = ll_i2info(inode);
++ struct lov_stripe_md *lsm;
++ struct lookup_intent oit = {.it_op = IT_OPEN, .it_flags = flags};
++ int rc = 0;
++ ENTRY;
++
++ down(&lli->lli_size_sem);
++ lsm = lli->lli_smd;
++ if (lsm) {
++ up(&lli->lli_size_sem);
++ CDEBUG(D_IOCTL, "stripe already exists for ino %lu\n",
++ inode->i_ino);
++ RETURN(-EEXIST);
++ }
++
++ rc = ll_intent_file_open(file, lum, lum_size, &oit);
++ if (rc)
++ GOTO(out, rc);
++ if (it_disposition(&oit, DISP_LOOKUP_NEG))
++ GOTO(out_req_free, rc = -ENOENT);
++ rc = oit.d.lustre.it_status;
++ if (rc < 0)
++ GOTO(out_req_free, rc);
++
++ ll_release_openhandle(file->f_dentry, &oit);
++
++ out:
++ up(&lli->lli_size_sem);
++ ll_intent_release(&oit);
++ RETURN(rc);
++out_req_free:
++ ptlrpc_req_finished((struct ptlrpc_request *) oit.d.lustre.it_data);
++ goto out;
++}
++
++int ll_lov_getstripe_ea_info(struct inode *inode, const char *filename,
++ struct lov_mds_md **lmmp, int *lmm_size,
++ struct ptlrpc_request **request)
++{
++ struct ll_sb_info *sbi = ll_i2sbi(inode);
++ struct ll_fid fid;
++ struct mds_body *body;
++ struct lov_mds_md *lmm = NULL;
++ struct ptlrpc_request *req = NULL;
++ int rc, lmmsize;
++
++ ll_inode2fid(&fid, inode);
++
++ rc = ll_get_max_mdsize(sbi, &lmmsize);
++ if (rc)
++ RETURN(rc);
++
++ rc = mdc_getattr_name(sbi->ll_mdc_exp, &fid,
++ filename, strlen(filename) + 1,
++ OBD_MD_FLEASIZE | OBD_MD_FLDIREA,
++ lmmsize, &req);
++ if (rc < 0) {
++ CDEBUG(D_INFO, "mdc_getattr_name failed "
++ "on %s: rc %d\n", filename, rc);
++ GOTO(out, rc);
++ }
++
++ body = lustre_msg_buf(req->rq_repmsg, REPLY_REC_OFF,
++ sizeof(*body));
++ LASSERT(body != NULL); /* checked by mdc_getattr_name */
++ /* swabbed by mdc_getattr_name */
++ LASSERT(lustre_rep_swabbed(req, REPLY_REC_OFF));
++
++ lmmsize = body->eadatasize;
++
++ if (!(body->valid & (OBD_MD_FLEASIZE | OBD_MD_FLDIREA)) ||
++ lmmsize == 0) {
++ GOTO(out, rc = -ENODATA);
++ }
++
++ lmm = lustre_msg_buf(req->rq_repmsg, REPLY_REC_OFF + 1,
++ lmmsize);
++ LASSERT(lmm != NULL);
++ LASSERT(lustre_rep_swabbed(req, REPLY_REC_OFF + 1));
++
++ if ((lmm->lmm_magic != cpu_to_le32(LOV_MAGIC_V1)) &&
++ (lmm->lmm_magic != cpu_to_le32(LOV_MAGIC_V3)) &&
++ (lmm->lmm_magic != cpu_to_le32(LOV_MAGIC_JOIN))) {
++ GOTO(out, rc = -EPROTO);
++ }
++ /*
++ * This is coming from the MDS, so is probably in
++ * little endian. We convert it to host endian before
++ * passing it to userspace.
++ */
++ if (LOV_MAGIC != cpu_to_le32(LOV_MAGIC)) {
++ /* if function called for directory - we should
++ * avoid swab not existent lsm objects */
++ if ((lmm->lmm_magic == cpu_to_le32(LOV_MAGIC_V1)) ||
++ (lmm->lmm_magic == cpu_to_le32(LOV_MAGIC_V3))) {
++ lustre_swab_lov_user_md((struct lov_user_md*)lmm);
++ if (S_ISREG(body->mode))
++ lustre_swab_lov_user_md_objects(
++ (struct lov_user_md*)lmm);
++ } else if (lmm->lmm_magic == cpu_to_le32(LOV_MAGIC_JOIN)) {
++ lustre_swab_lov_user_md_join((struct lov_user_md_join *)lmm);
++ }
++ }
++
++ if (lmm->lmm_magic == LOV_MAGIC_JOIN) {
++ struct lov_stripe_md *lsm;
++ struct lov_user_md_join *lmj;
++ int lmj_size, i, aindex = 0;
++
++ rc = obd_unpackmd(sbi->ll_osc_exp, &lsm, lmm, lmmsize);
++ if (rc < 0)
++ GOTO(out, rc = -ENOMEM);
++ rc = obd_checkmd(sbi->ll_osc_exp, sbi->ll_mdc_exp, lsm);
++ if (rc)
++ GOTO(out_free_memmd, rc);
++
++ lmj_size = sizeof(struct lov_user_md_join) +
++ lsm->lsm_stripe_count *
++ sizeof(struct lov_user_ost_data_join);
++ OBD_ALLOC(lmj, lmj_size);
++ if (!lmj)
++ GOTO(out_free_memmd, rc = -ENOMEM);
++
++ memcpy(lmj, lmm, sizeof(struct lov_user_md_join));
++ for (i = 0; i < lsm->lsm_stripe_count; i++) {
++ struct lov_extent *lex =
++ &lsm->lsm_array->lai_ext_array[aindex];
++
++ if (lex->le_loi_idx + lex->le_stripe_count <= i)
++ aindex ++;
++ CDEBUG(D_INFO, "aindex %d i %d l_extent_start "
++ LPU64" len %d\n", aindex, i,
++ lex->le_start, (int)lex->le_len);
++ lmj->lmm_objects[i].l_extent_start =
++ lex->le_start;
++
++ if ((int)lex->le_len == -1)
++ lmj->lmm_objects[i].l_extent_end = -1;
++ else
++ lmj->lmm_objects[i].l_extent_end =
++ lex->le_start + lex->le_len;
++ lmj->lmm_objects[i].l_object_id =
++ lsm->lsm_oinfo[i]->loi_id;
++ lmj->lmm_objects[i].l_object_gr =
++ lsm->lsm_oinfo[i]->loi_gr;
++ lmj->lmm_objects[i].l_ost_gen =
++ lsm->lsm_oinfo[i]->loi_ost_gen;
++ lmj->lmm_objects[i].l_ost_idx =
++ lsm->lsm_oinfo[i]->loi_ost_idx;
++ }
++ lmm = (struct lov_mds_md *)lmj;
++ lmmsize = lmj_size;
++out_free_memmd:
++ obd_free_memmd(sbi->ll_osc_exp, &lsm);
++ }
++out:
++ *lmmp = lmm;
++ *lmm_size = lmmsize;
++ *request = req;
++ return rc;
++}
++static int ll_lov_setea(struct inode *inode, struct file *file,
++ unsigned long arg)
++{
++ int flags = MDS_OPEN_HAS_OBJS | FMODE_WRITE;
++ struct lov_user_md *lump;
++ int lum_size = sizeof(struct lov_user_md) +
++ sizeof(struct lov_user_ost_data);
++ int rc;
++ ENTRY;
++
++ if (!cfs_capable(CFS_CAP_SYS_ADMIN))
++ RETURN(-EPERM);
++
++ OBD_ALLOC(lump, lum_size);
++ if (lump == NULL) {
++ RETURN(-ENOMEM);
++ }
++ rc = copy_from_user(lump, (struct lov_user_md *)arg, lum_size);
++ if (rc) {
++ OBD_FREE(lump, lum_size);
++ RETURN(-EFAULT);
++ }
++
++ rc = ll_lov_setstripe_ea_info(inode, file, flags, lump, lum_size);
++
++ OBD_FREE(lump, lum_size);
++ RETURN(rc);
++}
++
++static int ll_lov_setstripe(struct inode *inode, struct file *file,
++ unsigned long arg)
++{
++ struct lov_user_md_v3 lumv3;
++ struct lov_user_md_v1 *lumv1 = (struct lov_user_md_v1 *)&lumv3;
++ struct lov_user_md_v1 *lumv1p = (struct lov_user_md_v1 *)arg;
++ struct lov_user_md_v3 *lumv3p = (struct lov_user_md_v3 *)arg;
++ int lum_size;
++ int rc;
++ int flags = FMODE_WRITE;
++ ENTRY;
++
++ /* first try with v1 which is smaller than v3 */
++ lum_size = sizeof(struct lov_user_md_v1);
++ rc = copy_from_user(lumv1, lumv1p, lum_size);
++ if (rc)
++ RETURN(-EFAULT);
++
++ if (lumv1->lmm_magic == LOV_USER_MAGIC_V3) {
++ lum_size = sizeof(struct lov_user_md_v3);
++ rc = copy_from_user(&lumv3, lumv3p, lum_size);
++ if (rc)
++ RETURN(-EFAULT);
++ }
++
++ rc = ll_lov_setstripe_ea_info(inode, file, flags, lumv1, lum_size);
++ if (rc == 0) {
++ put_user(0, &lumv1p->lmm_stripe_count);
++ rc = obd_iocontrol(LL_IOC_LOV_GETSTRIPE, ll_i2obdexp(inode),
++ 0, ll_i2info(inode)->lli_smd,
++ (void *)arg);
++ }
++ RETURN(rc);
++}
++
++static int ll_lov_getstripe(struct inode *inode, unsigned long arg)
++{
++ struct lov_stripe_md *lsm = ll_i2info(inode)->lli_smd;
++
++ if (!lsm)
++ RETURN(-ENODATA);
++
++ return obd_iocontrol(LL_IOC_LOV_GETSTRIPE, ll_i2obdexp(inode), 0, lsm,
++ (void *)arg);
++}
++
++static int ll_get_grouplock(struct inode *inode, struct file *file,
++ unsigned long arg)
++{
++ struct ll_file_data *fd = LUSTRE_FPRIVATE(file);
++ ldlm_policy_data_t policy = { .l_extent = { .start = 0,
++ .end = OBD_OBJECT_EOF}};
++ struct lustre_handle lockh = { 0 };
++ struct ll_inode_info *lli = ll_i2info(inode);
++ struct lov_stripe_md *lsm = lli->lli_smd;
++ int flags = 0, rc;
++ ENTRY;
++
++ if (fd->fd_flags & LL_FILE_GROUP_LOCKED) {
++ RETURN(-EINVAL);
++ }
++
++ policy.l_extent.gid = arg;
++ if (file->f_flags & O_NONBLOCK)
++ flags = LDLM_FL_BLOCK_NOWAIT;
++
++ rc = ll_extent_lock(fd, inode, lsm, LCK_GROUP, &policy, &lockh, flags);
++ if (rc)
++ RETURN(rc);
++
++ fd->fd_flags |= LL_FILE_GROUP_LOCKED|LL_FILE_IGNORE_LOCK;
++ fd->fd_gid = arg;
++ memcpy(&fd->fd_cwlockh, &lockh, sizeof(lockh));
++
++ RETURN(0);
++}
++
++static int ll_put_grouplock(struct inode *inode, struct file *file,
++ unsigned long arg)
++{
++ struct ll_file_data *fd = LUSTRE_FPRIVATE(file);
++ struct ll_inode_info *lli = ll_i2info(inode);
++ struct lov_stripe_md *lsm = lli->lli_smd;
++ int rc;
++ ENTRY;
++
++ if (!(fd->fd_flags & LL_FILE_GROUP_LOCKED)) {
++ /* Ugh, it's already unlocked. */
++ RETURN(-EINVAL);
++ }
++
++ if (fd->fd_gid != arg) /* Ugh? Unlocking with different gid? */
++ RETURN(-EINVAL);
++
++ fd->fd_flags &= ~(LL_FILE_GROUP_LOCKED|LL_FILE_IGNORE_LOCK);
++
++ rc = ll_extent_unlock(fd, inode, lsm, LCK_GROUP, &fd->fd_cwlockh);
++ if (rc)
++ RETURN(rc);
++
++ fd->fd_gid = 0;
++ memset(&fd->fd_cwlockh, 0, sizeof(fd->fd_cwlockh));
++
++ RETURN(0);
++}
++
++#if LUSTRE_FIX >= 50
++static int join_sanity_check(struct inode *head, struct inode *tail)
++{
++ ENTRY;
++ if ((ll_i2sbi(head)->ll_flags & LL_SBI_JOIN) == 0) {
++ CERROR("server do not support join \n");
++ RETURN(-EINVAL);
++ }
++ if (!S_ISREG(tail->i_mode) || !S_ISREG(head->i_mode)) {
++ CERROR("tail ino %lu and ino head %lu must be regular\n",
++ head->i_ino, tail->i_ino);
++ RETURN(-EINVAL);
++ }
++ if (head->i_ino == tail->i_ino) {
++ CERROR("file %lu can not be joined to itself \n", head->i_ino);
++ RETURN(-EINVAL);
++ }
++ if (i_size_read(head) % JOIN_FILE_ALIGN) {
++ CERROR("hsize %llu must be times of 64K\n", i_size_read(head));
++ RETURN(-EINVAL);
++ }
++ RETURN(0);
++}
++
++static int join_file(struct inode *head_inode, struct file *head_filp,
++ struct file *tail_filp)
++{
++ struct dentry *tail_dentry = tail_filp->f_dentry;
++ struct lookup_intent oit = {.it_op = IT_OPEN,
++ .it_flags = head_filp->f_flags|O_JOIN_FILE};
++ struct ldlm_enqueue_info einfo = { LDLM_IBITS, LCK_PW,
++ ll_mdc_blocking_ast, ldlm_completion_ast, NULL, NULL };
++
++ struct lustre_handle lockh;
++ struct mdc_op_data *op_data;
++ int rc;
++ loff_t data;
++ ENTRY;
++
++ tail_dentry = tail_filp->f_dentry;
++
++ OBD_ALLOC_PTR(op_data);
++ if (op_data == NULL) {
++ RETURN(-ENOMEM);
++ }
++
++ data = i_size_read(head_inode);
++ ll_prepare_mdc_op_data(op_data, head_inode,
++ tail_dentry->d_parent->d_inode,
++ tail_dentry->d_name.name,
++ tail_dentry->d_name.len, 0, &data);
++ rc = mdc_enqueue(ll_i2mdcexp(head_inode), &einfo, &oit,
++ op_data, &lockh, NULL, 0, 0);
++
++ if (rc < 0)
++ GOTO(out, rc);
++
++ rc = oit.d.lustre.it_status;
++
++ if (rc < 0 || it_open_error(DISP_OPEN_OPEN, &oit)) {
++ rc = rc ? rc : it_open_error(DISP_OPEN_OPEN, &oit);
++ ptlrpc_req_finished((struct ptlrpc_request *)
++ oit.d.lustre.it_data);
++ GOTO(out, rc);
++ }
++
++ if (oit.d.lustre.it_lock_mode) { /* If we got lock - release it right
++ * away */
++ ldlm_lock_decref(&lockh, oit.d.lustre.it_lock_mode);
++ oit.d.lustre.it_lock_mode = 0;
++ }
++ ptlrpc_req_finished((struct ptlrpc_request *) oit.d.lustre.it_data);
++ it_clear_disposition(&oit, DISP_ENQ_COMPLETE);
++ ll_release_openhandle(head_filp->f_dentry, &oit);
++out:
++ if (op_data)
++ OBD_FREE_PTR(op_data);
++ ll_intent_release(&oit);
++ RETURN(rc);
++}
++
++static int ll_file_join(struct inode *head, struct file *filp,
++ char *filename_tail)
++{
++ struct inode *tail = NULL, *first = NULL, *second = NULL;
++ struct dentry *tail_dentry;
++ struct file *tail_filp, *first_filp, *second_filp;
++ struct ll_lock_tree first_tree, second_tree;
++ struct ll_lock_tree_node *first_node, *second_node;
++ struct ll_inode_info *hlli = ll_i2info(head), *tlli;
++ int rc = 0, cleanup_phase = 0;
++ ENTRY;
++
++ CDEBUG(D_VFSTRACE, "VFS Op:head=%lu/%u(%p) tail %s\n",
++ head->i_ino, head->i_generation, head, filename_tail);
++
++ tail_filp = filp_open(filename_tail, O_WRONLY, 0644);
++ if (IS_ERR(tail_filp)) {
++ CERROR("Can not open tail file %s", filename_tail);
++ rc = PTR_ERR(tail_filp);
++ GOTO(cleanup, rc);
++ }
++ tail = igrab(tail_filp->f_dentry->d_inode);
++
++ tlli = ll_i2info(tail);
++ tail_dentry = tail_filp->f_dentry;
++ LASSERT(tail_dentry);
++ cleanup_phase = 1;
++
++ /*reorder the inode for lock sequence*/
++ first = head->i_ino > tail->i_ino ? head : tail;
++ second = head->i_ino > tail->i_ino ? tail : head;
++ first_filp = head->i_ino > tail->i_ino ? filp : tail_filp;
++ second_filp = head->i_ino > tail->i_ino ? tail_filp : filp;
++
++ CDEBUG(D_INFO, "reorder object from %lu:%lu to %lu:%lu \n",
++ head->i_ino, tail->i_ino, first->i_ino, second->i_ino);
++ first_node = ll_node_from_inode(first, 0, OBD_OBJECT_EOF, LCK_EX);
++ if (IS_ERR(first_node)){
++ rc = PTR_ERR(first_node);
++ GOTO(cleanup, rc);
++ }
++ first_tree.lt_fd = first_filp->private_data;
++ rc = ll_tree_lock(&first_tree, first_node, NULL, 0, 0);
++ if (rc != 0)
++ GOTO(cleanup, rc);
++ cleanup_phase = 2;
++
++ second_node = ll_node_from_inode(second, 0, OBD_OBJECT_EOF, LCK_EX);
++ if (IS_ERR(second_node)){
++ rc = PTR_ERR(second_node);
++ GOTO(cleanup, rc);
++ }
++ second_tree.lt_fd = second_filp->private_data;
++ rc = ll_tree_lock(&second_tree, second_node, NULL, 0, 0);
++ if (rc != 0)
++ GOTO(cleanup, rc);
++ cleanup_phase = 3;
++
++ rc = join_sanity_check(head, tail);
++ if (rc)
++ GOTO(cleanup, rc);
++
++ rc = join_file(head, filp, tail_filp);
++ if (rc)
++ GOTO(cleanup, rc);
++cleanup:
++ switch (cleanup_phase) {
++ case 3:
++ ll_tree_unlock(&second_tree);
++ obd_cancel_unused(ll_i2obdexp(second),
++ ll_i2info(second)->lli_smd, 0, NULL);
++ case 2:
++ ll_tree_unlock(&first_tree);
++ obd_cancel_unused(ll_i2obdexp(first),
++ ll_i2info(first)->lli_smd, 0, NULL);
++ case 1:
++ filp_close(tail_filp, 0);
++ if (tail)
++ iput(tail);
++ if (head && rc == 0) {
++ obd_free_memmd(ll_i2sbi(head)->ll_osc_exp,
++ &hlli->lli_smd);
++ hlli->lli_smd = NULL;
++ }
++ case 0:
++ break;
++ default:
++ CERROR("invalid cleanup_phase %d\n", cleanup_phase);
++ LBUG();
++ }
++ RETURN(rc);
++}
++#endif /* LUSTRE_FIX >= 50 */
++
++/**
++ * Close inode open handle
++ *
++ * \param dentry [in] dentry which contains the inode
++ * \param it [in,out] intent which contains open info and result
++ *
++ * \retval 0 success
++ * \retval <0 failure
++ */
++int ll_release_openhandle(struct dentry *dentry, struct lookup_intent *it)
++{
++ struct inode *inode = dentry->d_inode;
++ struct obd_client_handle *och;
++ int rc;
++ ENTRY;
++
++ LASSERT(inode);
++
++ /* Root ? Do nothing. */
++ if (dentry->d_inode->i_sb->s_root == dentry)
++ RETURN(0);
++
++ /* No open handle to close? Move away */
++ if (!it_disposition(it, DISP_OPEN_OPEN))
++ RETURN(0);
++
++ LASSERT(it_open_error(DISP_OPEN_OPEN, it) == 0);
++
++ OBD_ALLOC(och, sizeof(*och));
++ if (!och)
++ GOTO(out, rc = -ENOMEM);
++
++ ll_och_fill(ll_i2info(inode), it, och);
++
++ rc = ll_close_inode_openhandle(inode, och);
++
++ OBD_FREE(och, sizeof(*och));
++ out:
++ /* this one is in place of ll_file_open */
++ if (it_disposition(it, DISP_ENQ_OPEN_REF))
++ ptlrpc_req_finished(it->d.lustre.it_data);
++ it_clear_disposition(it, DISP_ENQ_OPEN_REF);
++ RETURN(rc);
++}
++
++int ll_fiemap(struct inode *inode, struct ll_user_fiemap *fiemap,
++ int num_bytes)
++{
++ struct obd_export *exp = ll_i2obdexp(inode);
++ struct lov_stripe_md *lsm = ll_i2info(inode)->lli_smd;
++ struct ll_fiemap_info_key fm_key = { .name = KEY_FIEMAP, };
++ int vallen = num_bytes;
++ int rc;
++ ENTRY;
++
++ /* If the stripe_count > 1 and the application does not understand
++ * DEVICE_ORDER flag, then it cannot interpret the extents correctly.
++ */
++ if (lsm->lsm_stripe_count > 1 &&
++ !(fiemap->fm_flags & FIEMAP_FLAG_DEVICE_ORDER))
++ return -EOPNOTSUPP;
++
++ fm_key.oa.o_id = lsm->lsm_object_id;
++ fm_key.oa.o_valid = OBD_MD_FLID;
++
++ obdo_from_inode(&fm_key.oa, inode, OBD_MD_FLFID | OBD_MD_FLSIZE);
++
++ /* If filesize is 0, then there would be no objects for mapping */
++ if (fm_key.oa.o_size == 0) {
++ fiemap->fm_mapped_extents = 0;
++ RETURN(0);
++ }
++
++ memcpy(&fm_key.fiemap, fiemap, sizeof(*fiemap));
++
++ rc = obd_get_info(exp, sizeof(fm_key), &fm_key, &vallen, fiemap, lsm);
++ if (rc)
++ CERROR("obd_get_info failed: rc = %d\n", rc);
++
++ RETURN(rc);
++}
++
++int ll_file_ioctl(struct inode *inode, struct file *file, unsigned int cmd,
++ unsigned long arg)
++{
++ struct ll_file_data *fd = LUSTRE_FPRIVATE(file);
++ int flags;
++ ENTRY;
++
++ CDEBUG(D_VFSTRACE, "VFS Op:inode=%lu/%u(%p),cmd=%x\n", inode->i_ino,
++ inode->i_generation, inode, cmd);
++ ll_stats_ops_tally(ll_i2sbi(inode), LPROC_LL_IOCTL, 1);
++
++ /* asm-ppc{,64} declares TCGETS, et. al. as type 't' not 'T' */
++ if (_IOC_TYPE(cmd) == 'T' || _IOC_TYPE(cmd) == 't') /* tty ioctls */
++ RETURN(-ENOTTY);
++
++ switch(cmd) {
++ case LL_IOC_GETFLAGS:
++ /* Get the current value of the file flags */
++ return put_user(fd->fd_flags, (int *)arg);
++ case LL_IOC_SETFLAGS:
++ case LL_IOC_CLRFLAGS:
++ /* Set or clear specific file flags */
++ /* XXX This probably needs checks to ensure the flags are
++ * not abused, and to handle any flag side effects.
++ */
++ if (get_user(flags, (int *) arg))
++ RETURN(-EFAULT);
++
++ if (cmd == LL_IOC_SETFLAGS) {
++ if ((flags & LL_FILE_IGNORE_LOCK) &&
++ !(file->f_flags & O_DIRECT)) {
++ CERROR("%s: unable to disable locking on "
++ "non-O_DIRECT file\n", current->comm);
++ RETURN(-EINVAL);
++ }
++
++ fd->fd_flags |= flags;
++ } else {
++ fd->fd_flags &= ~flags;
++ }
++ RETURN(0);
++ case LL_IOC_LOV_SETSTRIPE:
++ RETURN(ll_lov_setstripe(inode, file, arg));
++ case LL_IOC_LOV_SETEA:
++ RETURN(ll_lov_setea(inode, file, arg));
++ case LL_IOC_LOV_GETSTRIPE:
++ RETURN(ll_lov_getstripe(inode, arg));
++ case LL_IOC_RECREATE_OBJ:
++ RETURN(ll_lov_recreate_obj(inode, file, arg));
++ case EXT3_IOC_FIEMAP: {
++ struct ll_user_fiemap *fiemap_s;
++ size_t num_bytes, ret_bytes;
++ unsigned int extent_count;
++ int rc = 0;
++
++ /* Get the extent count so we can calculate the size of
++ * required fiemap buffer */
++ if (get_user(extent_count,
++ &((struct ll_user_fiemap __user *)arg)->fm_extent_count))
++ RETURN(-EFAULT);
++ num_bytes = sizeof(*fiemap_s) + (extent_count *
++ sizeof(struct ll_fiemap_extent));
++ OBD_VMALLOC(fiemap_s, num_bytes);
++ if (fiemap_s == NULL)
++ RETURN(-ENOMEM);
++
++ if (copy_from_user(fiemap_s,(struct ll_user_fiemap __user *)arg,
++ sizeof(*fiemap_s)))
++ GOTO(error, rc = -EFAULT);
++
++ if (fiemap_s->fm_flags & ~LUSTRE_FIEMAP_FLAGS_COMPAT) {
++ fiemap_s->fm_flags = fiemap_s->fm_flags &
++ ~LUSTRE_FIEMAP_FLAGS_COMPAT;
++ if (copy_to_user((char *)arg, fiemap_s,
++ sizeof(*fiemap_s)))
++ GOTO(error, rc = -EFAULT);
++
++ GOTO(error, rc = -EBADR);
++ }
++
++ /* If fm_extent_count is non-zero, read the first extent since
++ * it is used to calculate end_offset and device from previous
++ * fiemap call. */
++ if (extent_count) {
++ if (copy_from_user(&fiemap_s->fm_extents[0],
++ (char __user *)arg + sizeof(*fiemap_s),
++ sizeof(struct ll_fiemap_extent)))
++ GOTO(error, rc = -EFAULT);
++ }
++
++ if (fiemap_s->fm_flags & FIEMAP_FLAG_SYNC) {
++ int rc;
++
++ rc = filemap_fdatawrite(inode->i_mapping);
++ if (rc)
++ GOTO(error, rc);
++ }
++
++ rc = ll_fiemap(inode, fiemap_s, num_bytes);
++ if (rc)
++ GOTO(error, rc);
++
++ ret_bytes = sizeof(struct ll_user_fiemap);
++
++ if (extent_count != 0)
++ ret_bytes += (fiemap_s->fm_mapped_extents *
++ sizeof(struct ll_fiemap_extent));
++
++ if (copy_to_user((void *)arg, fiemap_s, ret_bytes))
++ rc = -EFAULT;
++
++error:
++ OBD_VFREE(fiemap_s, num_bytes);
++ RETURN(rc);
++ }
++ case EXT3_IOC_GETFLAGS:
++ case EXT3_IOC_SETFLAGS:
++ RETURN(ll_iocontrol(inode, file, cmd, arg));
++ case EXT3_IOC_GETVERSION_OLD:
++ case EXT3_IOC_GETVERSION:
++ RETURN(put_user(inode->i_generation, (int *)arg));
++ case LL_IOC_JOIN: {
++#if LUSTRE_FIX >= 50
++ /* Allow file join in beta builds to allow debuggging */
++ char *ftail;
++ int rc;
++
++ ftail = getname((const char *)arg);
++ if (IS_ERR(ftail))
++ RETURN(PTR_ERR(ftail));
++ rc = ll_file_join(inode, file, ftail);
++ putname(ftail);
++ RETURN(rc);
++#else
++ CWARN("file join is not supported in this version of Lustre\n");
++ RETURN(-ENOTTY);
++#endif
++ }
++ case LL_IOC_GROUP_LOCK:
++ RETURN(ll_get_grouplock(inode, file, arg));
++ case LL_IOC_GROUP_UNLOCK:
++ RETURN(ll_put_grouplock(inode, file, arg));
++ case IOC_OBD_STATFS:
++ RETURN(ll_obd_statfs(inode, (void *)arg));
++ case OBD_IOC_GETNAME_OLD:
++ case OBD_IOC_GETNAME: {
++ struct obd_device *obd =
++ class_exp2obd(ll_i2sbi(inode)->ll_osc_exp);
++ if (!obd)
++ RETURN(-EFAULT);
++ if (copy_to_user((void *)arg, obd->obd_name,
++ strlen(obd->obd_name) + 1))
++ RETURN (-EFAULT);
++ RETURN(0);
++ }
++
++ /* We need to special case any other ioctls we want to handle,
++ * to send them to the MDS/OST as appropriate and to properly
++ * network encode the arg field.
++ case EXT3_IOC_SETVERSION_OLD:
++ case EXT3_IOC_SETVERSION:
++ */
++ default: {
++ int err;
++
++ if (LLIOC_STOP ==
++ ll_iocontrol_call(inode, file, cmd, arg, &err))
++ RETURN(err);
++
++ RETURN(obd_iocontrol(cmd, ll_i2obdexp(inode), 0, NULL,
++ (void *)arg));
++ }
++ }
++}
++
++loff_t ll_file_seek(struct file *file, loff_t offset, int origin)
++{
++ struct inode *inode = file->f_dentry->d_inode;
++ struct ll_inode_info *lli = ll_i2info(inode);
++ struct lov_stripe_md *lsm = lli->lli_smd;
++ loff_t retval;
++ ENTRY;
++ retval = offset + ((origin == 2) ? i_size_read(inode) :
++ (origin == 1) ? file->f_pos : 0);
++ CDEBUG(D_VFSTRACE, "VFS Op:inode=%lu/%u(%p), to=%Lu=%#Lx(%s)\n",
++ inode->i_ino, inode->i_generation, inode, retval, retval,
++ origin == 2 ? "SEEK_END": origin == 1 ? "SEEK_CUR" : "SEEK_SET");
++ ll_stats_ops_tally(ll_i2sbi(inode), LPROC_LL_LLSEEK, 1);
++
++ if (origin == 2) { /* SEEK_END */
++ int nonblock = 0, rc;
++
++ if (file->f_flags & O_NONBLOCK)
++ nonblock = LDLM_FL_BLOCK_NOWAIT;
++
++ if (lsm != NULL) {
++ rc = ll_glimpse_size(inode, nonblock);
++ if (rc != 0)
++ RETURN(rc);
++ }
++
++ ll_inode_size_lock(inode, 0);
++ offset += i_size_read(inode);
++ ll_inode_size_unlock(inode, 0);
++ } else if (origin == 1) { /* SEEK_CUR */
++ offset += file->f_pos;
++ }
++
++ retval = -EINVAL;
++ if (offset >= 0 && offset <= ll_file_maxbytes(inode)) {
++ if (offset != file->f_pos) {
++ file->f_pos = offset;
++ file->f_version = 0;
++ }
++ retval = offset;
++ }
++
++ RETURN(retval);
++}
++
++int ll_fsync(struct file *file, struct dentry *dentry, int data)
++{
++ struct inode *inode = dentry->d_inode;
++ struct ll_inode_info *lli = ll_i2info(inode);
++ struct lov_stripe_md *lsm = lli->lli_smd;
++ struct ll_fid fid;
++ struct ptlrpc_request *req;
++ int rc, err;
++ ENTRY;
++ CDEBUG(D_VFSTRACE, "VFS Op:inode=%lu/%u(%p)\n", inode->i_ino,
++ inode->i_generation, inode);
++ ll_stats_ops_tally(ll_i2sbi(inode), LPROC_LL_FSYNC, 1);
++
++ /* fsync's caller has already called _fdata{sync,write}, we want
++ * that IO to finish before calling the osc and mdc sync methods */
++ rc = filemap_fdatawait(inode->i_mapping);
++
++ /* catch async errors that were recorded back when async writeback
++ * failed for pages in this mapping. */
++ err = lli->lli_async_rc;
++ lli->lli_async_rc = 0;
++ if (rc == 0)
++ rc = err;
++ if (lsm) {
++ err = lov_test_and_clear_async_rc(lsm);
++ if (rc == 0)
++ rc = err;
++ }
++
++ ll_inode2fid(&fid, inode);
++ err = mdc_sync(ll_i2sbi(inode)->ll_mdc_exp, &fid, &req);
++ if (!rc)
++ rc = err;
++ if (!err)
++ ptlrpc_req_finished(req);
++
++ if (data && lsm) {
++ struct obd_info *oinfo;
++
++ OBD_ALLOC_PTR(oinfo);
++ if (!oinfo)
++ RETURN(rc ? rc : -ENOMEM);
++ OBDO_ALLOC(oinfo->oi_oa);
++ if (!oinfo->oi_oa) {
++ OBD_FREE_PTR(oinfo);
++ RETURN(rc ? rc : -ENOMEM);
++ }
++ oinfo->oi_oa->o_id = lsm->lsm_object_id;
++ oinfo->oi_oa->o_gr = lsm->lsm_object_gr;
++ oinfo->oi_oa->o_valid = OBD_MD_FLID | OBD_MD_FLGROUP;
++ obdo_from_inode(oinfo->oi_oa, inode,
++ OBD_MD_FLTYPE | OBD_MD_FLATIME |
++ OBD_MD_FLMTIME | OBD_MD_FLCTIME);
++ oinfo->oi_md = lsm;
++ err = obd_sync_rqset(ll_i2sbi(inode)->ll_osc_exp, oinfo,
++ 0, OBD_OBJECT_EOF);
++ if (!rc)
++ rc = err;
++ OBDO_FREE(oinfo->oi_oa);
++ OBD_FREE_PTR(oinfo);
++ }
++
++ RETURN(rc);
++}
++
++int ll_file_flock(struct file *file, int cmd, struct file_lock *file_lock)
++{
++ struct inode *inode = file->f_dentry->d_inode;
++ struct ll_sb_info *sbi = ll_i2sbi(inode);
++ struct lu_fid *fid = ll_inode_lu_fid(inode);
++ struct ldlm_res_id res_id =
++ { .name = { fid_seq(fid),
++ fid_oid(fid),
++ fid_ver(fid),
++ LDLM_FLOCK} };
++ struct ldlm_enqueue_info einfo = { LDLM_FLOCK, 0, NULL,
++ ldlm_flock_completion_ast, NULL, file_lock };
++ struct lustre_handle lockh = {0};
++ ldlm_policy_data_t flock;
++ int flags = 0;
++ int rc;
++ ENTRY;
++
++ CDEBUG(D_VFSTRACE, "VFS Op:inode=%lu file_lock=%p\n",
++ inode->i_ino, file_lock);
++ ll_stats_ops_tally(ll_i2sbi(inode), LPROC_LL_FLOCK, 1);
++
++ if (fid_is_igif(fid)) {
++ /* If this is an IGIF inode, we need to keep the 1.6-style
++ * flock mapping for compatibility. If it is a proper FID
++ * then we know any other client accessing it must also be
++ * accessing it as a FID and can use the CMD-style flock. */
++ res_id.name[2] = LDLM_FLOCK;
++ res_id.name[3] = 0;
++ }
++
++ if (file_lock->fl_flags & FL_FLOCK) {
++ LASSERT((cmd == F_SETLKW) || (cmd == F_SETLK));
++ /* set missing params for flock() calls */
++ file_lock->fl_end = OFFSET_MAX;
++ file_lock->fl_pid = current->tgid;
++ }
++ flock.l_flock.pid = file_lock->fl_pid;
++ flock.l_flock.start = file_lock->fl_start;
++ flock.l_flock.end = file_lock->fl_end;
++
++ switch (file_lock->fl_type) {
++ case F_RDLCK:
++ einfo.ei_mode = LCK_PR;
++ break;
++ case F_UNLCK:
++ /* An unlock request may or may not have any relation to
++ * existing locks so we may not be able to pass a lock handle
++ * via a normal ldlm_lock_cancel() request. The request may even
++ * unlock a byte range in the middle of an existing lock. In
++ * order to process an unlock request we need all of the same
++ * information that is given with a normal read or write record
++ * lock request. To avoid creating another ldlm unlock (cancel)
++ * message we'll treat a LCK_NL flock request as an unlock. */
++ einfo.ei_mode = LCK_NL;
++ break;
++ case F_WRLCK:
++ einfo.ei_mode = LCK_PW;
++ break;
++ default:
++ CERROR("unknown fcntl lock type: %d\n", file_lock->fl_type);
++ RETURN (-EINVAL);
++ }
++
++ switch (cmd) {
++ case F_SETLKW:
++#ifdef F_SETLKW64
++ case F_SETLKW64:
++#endif
++ flags = 0;
++ break;
++ case F_SETLK:
++#ifdef F_SETLK64
++ case F_SETLK64:
++#endif
++ flags = LDLM_FL_BLOCK_NOWAIT;
++ break;
++ case F_GETLK:
++#ifdef F_GETLK64
++ case F_GETLK64:
++#endif
++ flags = LDLM_FL_TEST_LOCK;
++ /* Save the old mode so that if the mode in the lock changes we
++ * can decrement the appropriate reader or writer refcount. */
++ file_lock->fl_type = einfo.ei_mode;
++ break;
++ default:
++ CERROR("unknown fcntl lock command: %d\n", cmd);
++ RETURN (-EINVAL);
++ }
++
++ CDEBUG(D_DLMTRACE, "inode=%lu, pid=%u, flags=%#x, mode=%u, "
++ "start="LPU64", end="LPU64"\n", inode->i_ino, flock.l_flock.pid,
++ flags, einfo.ei_mode, flock.l_flock.start, flock.l_flock.end);
++
++ rc = ldlm_cli_enqueue(sbi->ll_mdc_exp, NULL, &einfo, res_id,
++ &flock, &flags, NULL, 0, NULL, &lockh, 0);
++ if ((file_lock->fl_flags & FL_FLOCK) &&
++ (rc == 0 || file_lock->fl_type == F_UNLCK))
++ ll_flock_lock_file_wait(file, file_lock, (cmd == F_SETLKW));
++#ifdef HAVE_F_OP_FLOCK
++ if ((file_lock->fl_flags & FL_POSIX) &&
++ (rc == 0 || file_lock->fl_type == F_UNLCK) &&
++ !(flags & LDLM_FL_TEST_LOCK))
++ posix_lock_file_wait(file, file_lock);
++#endif
++
++ RETURN(rc);
++}
++
++int ll_file_noflock(struct file *file, int cmd, struct file_lock *file_lock)
++{
++ ENTRY;
++
++ RETURN(-ENOSYS);
++}
++
++int ll_have_md_lock(struct inode *inode, __u64 bits)
++{
++ struct lustre_handle lockh;
++ struct ldlm_res_id res_id;
++ struct obd_device *obddev;
++ ldlm_policy_data_t policy = { .l_inodebits = {bits}};
++ int flags;
++ ENTRY;
++
++ if (!inode)
++ RETURN(0);
++
++ obddev = ll_i2mdcexp(inode)->exp_obd;
++ fid_build_reg_res_name(ll_inode_lu_fid(inode), &res_id);
++
++ CDEBUG(D_INFO, "trying to match res "LPU64":"LPU64":"LPU64"\n",
++ res_id.name[0],
++ res_id.name[1],
++ res_id.name[2]);
++
++ flags = LDLM_FL_BLOCK_GRANTED | LDLM_FL_CBPENDING | LDLM_FL_TEST_LOCK;
++ if (ldlm_lock_match(obddev->obd_namespace, flags, &res_id, LDLM_IBITS,
++ &policy, LCK_CR|LCK_CW|LCK_PR|LCK_PW, &lockh)) {
++ RETURN(1);
++ }
++
++ RETURN(0);
++}
++
++static int ll_inode_revalidate_fini(struct inode *inode, int rc) {
++ if (rc == -ENOENT) { /* Already unlinked. Just update nlink
++ * and return success */
++ inode->i_nlink = 0;
++ /* This path cannot be hit for regular files unless in
++ * case of obscure races, so no need to to validate
++ * size. */
++ if (!S_ISREG(inode->i_mode) &&
++ !S_ISDIR(inode->i_mode))
++ return 0;
++ }
++
++ if (rc) {
++ CERROR("failure %d inode %lu\n", rc, inode->i_ino);
++ return -abs(rc);
++
++ }
++
++ return 0;
++}
++
++int ll_inode_revalidate_it(struct dentry *dentry, struct lookup_intent *it)
++{
++ struct inode *inode = dentry->d_inode;
++ struct ptlrpc_request *req = NULL;
++ struct obd_export *exp;
++ int rc;
++ ENTRY;
++
++ if (!inode) {
++ CERROR("REPORT THIS LINE TO PETER\n");
++ RETURN(0);
++ }
++ CDEBUG(D_VFSTRACE, "VFS Op:inode=%lu/%u(%p),name=%s\n",
++ inode->i_ino, inode->i_generation, inode, dentry->d_name.name);
++
++ exp = ll_i2mdcexp(inode);
++
++ if (exp->exp_connect_flags & OBD_CONNECT_ATTRFID) {
++ struct lookup_intent oit = { .it_op = IT_GETATTR };
++ struct mdc_op_data op_data = { { 0 } };
++
++ /* Call getattr by fid, so do not provide name at all. */
++ ll_prepare_mdc_op_data(&op_data, dentry->d_parent->d_inode,
++ dentry->d_inode, NULL, 0, 0, NULL);
++ oit.it_flags |= O_CHECK_STALE;
++ rc = mdc_intent_lock(exp, &op_data, NULL, 0,
++ /* we are not interested in name
++ based lookup */
++ &oit, 0, &req,
++ ll_mdc_blocking_ast, 0);
++ oit.it_flags &= ~O_CHECK_STALE;
++ if (rc < 0) {
++ rc = ll_inode_revalidate_fini(inode, rc);
++ GOTO (out, rc);
++ }
++
++ rc = revalidate_it_finish(req, DLM_REPLY_REC_OFF, &oit, dentry);
++ if (rc != 0) {
++ ll_intent_release(&oit);
++ GOTO(out, rc);
++ }
++
++ /* Unlinked? Unhash dentry, so it is not picked up later by
++ do_lookup() -> ll_revalidate_it(). We cannot use d_drop
++ here to preserve get_cwd functionality on 2.6.
++ Bug 10503 */
++ if (!dentry->d_inode->i_nlink) {
++ spin_lock(&ll_lookup_lock);
++ spin_lock(&dcache_lock);
++ ll_drop_dentry(dentry);
++ spin_unlock(&dcache_lock);
++ spin_unlock(&ll_lookup_lock);
++ }
++
++ ll_lookup_finish_locks(&oit, dentry);
++ } else if (!ll_have_md_lock(dentry->d_inode,
++ MDS_INODELOCK_UPDATE|MDS_INODELOCK_LOOKUP)) {
++ struct ll_sb_info *sbi = ll_i2sbi(dentry->d_inode);
++ struct ll_fid fid;
++ obd_valid valid = OBD_MD_FLGETATTR;
++ int ealen = 0;
++
++ if (S_ISREG(inode->i_mode)) {
++ rc = ll_get_max_mdsize(sbi, &ealen);
++ if (rc)
++ RETURN(rc);
++ valid |= OBD_MD_FLEASIZE | OBD_MD_FLMODEASIZE;
++ }
++ ll_inode2fid(&fid, inode);
++ rc = mdc_getattr(sbi->ll_mdc_exp, &fid, valid, ealen, &req);
++ if (rc) {
++ rc = ll_inode_revalidate_fini(inode, rc);
++ RETURN(rc);
++ }
++
++ rc = ll_prep_inode(sbi->ll_osc_exp, &inode, req, REPLY_REC_OFF,
++ NULL);
++ if (rc)
++ GOTO(out, rc);
++ }
++
++ /* if object not yet allocated, don't validate size */
++ if (ll_i2info(inode)->lli_smd == NULL) {
++ LTIME_S(inode->i_atime) = ll_i2info(inode)->lli_lvb.lvb_atime;
++ LTIME_S(inode->i_mtime) = ll_i2info(inode)->lli_lvb.lvb_mtime;
++ LTIME_S(inode->i_ctime) = ll_i2info(inode)->lli_lvb.lvb_ctime;
++ GOTO(out, rc = 0);
++ }
++
++ /* ll_glimpse_size will prefer locally cached writes if they extend
++ * the file */
++ rc = ll_glimpse_size(inode, 0);
++
++out:
++ ptlrpc_req_finished(req);
++ RETURN(rc);
++}
++
++int ll_getattr_it(struct vfsmount *mnt, struct dentry *de,
++ struct lookup_intent *it, struct kstat *stat)
++{
++ struct inode *inode = de->d_inode;
++ int res = 0;
++
++ res = ll_inode_revalidate_it(de, it);
++ ll_stats_ops_tally(ll_i2sbi(inode), LPROC_LL_GETATTR, 1);
++
++ if (res)
++ return res;
++
++ stat->dev = inode->i_sb->s_dev;
++ stat->ino = inode->i_ino;
++ stat->mode = inode->i_mode;
++ stat->nlink = inode->i_nlink;
++ stat->uid = inode->i_uid;
++ stat->gid = inode->i_gid;
++ stat->rdev = kdev_t_to_nr(inode->i_rdev);
++ stat->atime = inode->i_atime;
++ stat->mtime = inode->i_mtime;
++ stat->ctime = inode->i_ctime;
++#ifdef HAVE_INODE_BLKSIZE
++ stat->blksize = inode->i_blksize;
++#else
++ stat->blksize = 1<<inode->i_blkbits;
++#endif
++
++ ll_inode_size_lock(inode, 0);
++ stat->size = i_size_read(inode);
++ stat->blocks = inode->i_blocks;
++ ll_inode_size_unlock(inode, 0);
++
++ return 0;
++}
++int ll_getattr(struct vfsmount *mnt, struct dentry *de, struct kstat *stat)
++{
++ struct lookup_intent it = { .it_op = IT_GETATTR };
++
++ return ll_getattr_it(mnt, de, &it, stat);
++}
++
++static
++int lustre_check_acl(struct inode *inode, int mask)
++{
++#ifdef CONFIG_FS_POSIX_ACL
++ struct ll_inode_info *lli = ll_i2info(inode);
++ struct posix_acl *acl;
++ int rc;
++ ENTRY;
++
++ spin_lock(&lli->lli_lock);
++ acl = posix_acl_dup(lli->lli_posix_acl);
++ spin_unlock(&lli->lli_lock);
++
++ if (!acl)
++ RETURN(-EAGAIN);
++
++ rc = posix_acl_permission(inode, acl, mask);
++ posix_acl_release(acl);
++
++ RETURN(rc);
++#else
++ return -EAGAIN;
++#endif
++}
++
++#if (LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,10))
++int ll_inode_permission(struct inode *inode, int mask, struct nameidata *nd)
++{
++ CDEBUG(D_VFSTRACE, "VFS Op:inode=%lu/%u(%p), mask %o\n",
++ inode->i_ino, inode->i_generation, inode, mask);
++
++ ll_stats_ops_tally(ll_i2sbi(inode), LPROC_LL_INODE_PERM, 1);
++ return generic_permission(inode, mask, lustre_check_acl);
++}
++#else
++#if (LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,0))
++int ll_inode_permission(struct inode *inode, int mask, struct nameidata *nd)
++#else
++int ll_inode_permission(struct inode *inode, int mask)
++#endif
++{
++ int mode = inode->i_mode;
++ int rc;
++
++ CDEBUG(D_VFSTRACE, "VFS Op:inode=%lu/%u(%p), mask %o\n",
++ inode->i_ino, inode->i_generation, inode, mask);
++ ll_stats_ops_tally(ll_i2sbi(inode), LPROC_LL_INODE_PERM, 1);
++
++ if ((mask & MAY_WRITE) && IS_RDONLY(inode) &&
++ (S_ISREG(mode) || S_ISDIR(mode) || S_ISLNK(mode)))
++ return -EROFS;
++ if ((mask & MAY_WRITE) && IS_IMMUTABLE(inode))
++ return -EACCES;
++ if (current->fsuid == inode->i_uid) {
++ mode >>= 6;
++ } else if (1) {
++ if (((mode >> 3) & mask & S_IRWXO) != mask)
++ goto check_groups;
++ rc = lustre_check_acl(inode, mask);
++ if (rc == -EAGAIN)
++ goto check_groups;
++ if (rc == -EACCES)
++ goto check_capabilities;
++ return rc;
++ } else {
++check_groups:
++ if (in_group_p(inode->i_gid))
++ mode >>= 3;
++ }
++ if ((mode & mask & S_IRWXO) == mask)
++ return 0;
++
++check_capabilities:
++ if (!(mask & MAY_EXEC) ||
++ (inode->i_mode & S_IXUGO) || S_ISDIR(inode->i_mode))
++ if (cfs_capable(CFS_CAP_DAC_OVERRIDE))
++ return 0;
++
++ if (cfs_capable(CFS_CAP_DAC_READ_SEARCH) && ((mask == MAY_READ) ||
++ (S_ISDIR(inode->i_mode) && !(mask & MAY_WRITE))))
++ return 0;
++
++ return -EACCES;
++}
++#endif
++
++/* -o localflock - only provides locally consistent flock locks */
++struct file_operations ll_file_operations = {
++ .read = ll_file_read,
++#ifdef HAVE_FILE_READV
++ .readv = ll_file_readv,
++#else
++ .aio_read = ll_file_aio_read,
++#endif
++ .write = ll_file_write,
++#ifdef HAVE_FILE_WRITEV
++ .writev = ll_file_writev,
++#else
++ .aio_write = ll_file_aio_write,
++#endif
++ .ioctl = ll_file_ioctl,
++ .open = ll_file_open,
++ .release = ll_file_release,
++ .mmap = ll_file_mmap,
++ .llseek = ll_file_seek,
++#ifdef HAVE_KERNEL_SENDFILE
++ .sendfile = ll_file_sendfile,
++#endif
++ .fsync = ll_fsync,
++};
++
++struct file_operations ll_file_operations_flock = {
++ .read = ll_file_read,
++#ifdef HAVE_FILE_READV
++ .readv = ll_file_readv,
++#else
++ .aio_read = ll_file_aio_read,
++#endif
++ .write = ll_file_write,
++#ifdef HAVE_FILE_WRITEV
++ .writev = ll_file_writev,
++#else
++ .aio_write = ll_file_aio_write,
++#endif
++ .ioctl = ll_file_ioctl,
++ .open = ll_file_open,
++ .release = ll_file_release,
++ .mmap = ll_file_mmap,
++ .llseek = ll_file_seek,
++#ifdef HAVE_KERNEL_SENDFILE
++ .sendfile = ll_file_sendfile,
++#endif
++ .fsync = ll_fsync,
++#ifdef HAVE_F_OP_FLOCK
++ .flock = ll_file_flock,
++#endif
++ .lock = ll_file_flock
++};
++
++/* These are for -o noflock - to return ENOSYS on flock calls */
++struct file_operations ll_file_operations_noflock = {
++ .read = ll_file_read,
++#ifdef HAVE_FILE_READV
++ .readv = ll_file_readv,
++#else
++ .aio_read = ll_file_aio_read,
++#endif
++ .write = ll_file_write,
++#ifdef HAVE_FILE_WRITEV
++ .writev = ll_file_writev,
++#else
++ .aio_write = ll_file_aio_write,
++#endif
++ .ioctl = ll_file_ioctl,
++ .open = ll_file_open,
++ .release = ll_file_release,
++ .mmap = ll_file_mmap,
++ .llseek = ll_file_seek,
++ .sendfile = ll_file_sendfile,
++ .fsync = ll_fsync,
++#ifdef HAVE_F_OP_FLOCK
++ .flock = ll_file_noflock,
++#endif
++ .lock = ll_file_noflock
++};
++
++struct inode_operations ll_file_inode_operations = {
++#ifdef HAVE_VFS_INTENT_PATCHES
++ .setattr_raw = ll_setattr_raw,
++#endif
++ .setattr = ll_setattr,
++ .truncate = ll_truncate,
++ .getattr = ll_getattr,
++ .permission = ll_inode_permission,
++ .setxattr = ll_setxattr,
++ .getxattr = ll_getxattr,
++ .listxattr = ll_listxattr,
++ .removexattr = ll_removexattr,
++};
++
++/* dynamic ioctl number support routins */
++static struct llioc_ctl_data {
++ struct rw_semaphore ioc_sem;
++ struct list_head ioc_head;
++} llioc = {
++ __RWSEM_INITIALIZER(llioc.ioc_sem),
++ CFS_LIST_HEAD_INIT(llioc.ioc_head)
++};
++
++
++struct llioc_data {
++ struct list_head iocd_list;
++ unsigned int iocd_size;
++ llioc_callback_t iocd_cb;
++ unsigned int iocd_count;
++ unsigned int iocd_cmd[0];
++};
++
++void *ll_iocontrol_register(llioc_callback_t cb, int count, unsigned int *cmd)
++{
++ unsigned int size;
++ struct llioc_data *in_data = NULL;
++ ENTRY;
++
++ if (cb == NULL || cmd == NULL ||
++ count > LLIOC_MAX_CMD || count < 0)
++ RETURN(NULL);
++
++ size = sizeof(*in_data) + count * sizeof(unsigned int);
++ OBD_ALLOC(in_data, size);
++ if (in_data == NULL)
++ RETURN(NULL);
++
++ memset(in_data, 0, sizeof(*in_data));
++ in_data->iocd_size = size;
++ in_data->iocd_cb = cb;
++ in_data->iocd_count = count;
++ memcpy(in_data->iocd_cmd, cmd, sizeof(unsigned int) * count);
++
++ down_write(&llioc.ioc_sem);
++ list_add_tail(&in_data->iocd_list, &llioc.ioc_head);
++ up_write(&llioc.ioc_sem);
++
++ RETURN(in_data);
++}
++
++void ll_iocontrol_unregister(void *magic)
++{
++ struct llioc_data *tmp;
++
++ if (magic == NULL)
++ return;
++
++ down_write(&llioc.ioc_sem);
++ list_for_each_entry(tmp, &llioc.ioc_head, iocd_list) {
++ if (tmp == magic) {
++ unsigned int size = tmp->iocd_size;
++
++ list_del(&tmp->iocd_list);
++ up_write(&llioc.ioc_sem);
++
++ OBD_FREE(tmp, size);
++ return;
++ }
++ }
++ up_write(&llioc.ioc_sem);
++
++ CWARN("didn't find iocontrol register block with magic: %p\n", magic);
++}
++
++EXPORT_SYMBOL(ll_iocontrol_register);
++EXPORT_SYMBOL(ll_iocontrol_unregister);
++
++enum llioc_iter ll_iocontrol_call(struct inode *inode, struct file *file,
++ unsigned int cmd, unsigned long arg, int *rcp)
++{
++ enum llioc_iter ret = LLIOC_CONT;
++ struct llioc_data *data;
++ int rc = -EINVAL, i;
++
++ down_read(&llioc.ioc_sem);
++ list_for_each_entry(data, &llioc.ioc_head, iocd_list) {
++ for (i = 0; i < data->iocd_count; i++) {
++ if (cmd != data->iocd_cmd[i])
++ continue;
++
++ ret = data->iocd_cb(inode, file, cmd, arg, data, &rc);
++ break;
++ }
++
++ if (ret == LLIOC_STOP)
++ break;
++ }
++ up_read(&llioc.ioc_sem);
++
++ if (rcp)
++ *rcp = rc;
++ return ret;
++}
diff -urNad lustre~/lustre/llite/lloop.c lustre/lustre/llite/lloop.c
--- lustre~/lustre/llite/lloop.c 2009-06-02 11:38:20.000000000 +0200
-+++ lustre/lustre/llite/lloop.c 2009-06-04 15:12:41.000000000 +0200
++++ lustre/lustre/llite/lloop.c 2009-06-05 10:12:08.000000000 +0200
@@ -347,7 +347,7 @@
loop_add_bio(lo, old_bio);
return 0;
--
Lustre Debian Packaging
More information about the Pkg-lustre-svn-commit
mailing list