]> git.proxmox.com Git - mirror_zfs.git/blobdiff - module/zfs/zfs_vnops.c
Fix a potential use-after-free in zfs_setsecattr()
[mirror_zfs.git] / module / zfs / zfs_vnops.c
index e21fe1acae2a208be5c8a1aae39fc6aacd2e752f..5377da401126cab7829778af3b101a8c0113d3ef 100644 (file)
@@ -6,7 +6,7 @@
  * You may not use this file except in compliance with the License.
  *
  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
- * or http://www.opensolaris.org/os/licensing.
+ * or https://opensource.org/licenses/CDDL-1.0.
  * See the License for the specific language governing permissions
  * and limitations under the License.
  *
 
 /*
  * Copyright (c) 2005, 2010, Oracle and/or its affiliates. All rights reserved.
- * Copyright (c) 2012, 2015 by Delphix. All rights reserved.
+ * Copyright (c) 2012, 2018 by Delphix. All rights reserved.
  * Copyright (c) 2015 by Chunwei Chen. All rights reserved.
  * Copyright 2017 Nexenta Systems, Inc.
+ * Copyright (c) 2021, 2022 by Pawel Jakub Dawidek
  */
 
 /* Portions Copyright 2007 Jeremy Teo */
 /* Portions Copyright 2010 Robert Milkowski */
 
-
 #include <sys/types.h>
 #include <sys/param.h>
 #include <sys/time.h>
-#include <sys/systm.h>
 #include <sys/sysmacros.h>
-#include <sys/resource.h>
 #include <sys/vfs.h>
-#include <sys/vfs_opreg.h>
+#include <sys/uio_impl.h>
 #include <sys/file.h>
 #include <sys/stat.h>
 #include <sys/kmem.h>
-#include <sys/taskq.h>
-#include <sys/uio.h>
-#include <sys/vmsystm.h>
-#include <sys/atomic.h>
-#include <vm/pvn.h>
-#include <sys/pathname.h>
 #include <sys/cmn_err.h>
 #include <sys/errno.h>
-#include <sys/unistd.h>
 #include <sys/zfs_dir.h>
 #include <sys/zfs_acl.h>
 #include <sys/zfs_ioctl.h>
 #include <sys/fs/zfs.h>
 #include <sys/dmu.h>
 #include <sys/dmu_objset.h>
+#include <sys/dsl_crypt.h>
 #include <sys/spa.h>
 #include <sys/txg.h>
 #include <sys/dbuf.h>
-#include <sys/zap.h>
-#include <sys/sa.h>
-#include <sys/dirent.h>
 #include <sys/policy.h>
-#include <sys/sunddi.h>
-#include <sys/sid.h>
-#include <sys/mode.h>
-#include "fs/fs_subr.h"
-#include <sys/zfs_ctldir.h>
-#include <sys/zfs_fuid.h>
-#include <sys/zfs_sa.h>
+#include <sys/zfeature.h>
 #include <sys/zfs_vnops.h>
-#include <sys/dnlc.h>
-#include <sys/zfs_rlock.h>
-#include <sys/extdirent.h>
-#include <sys/kidmap.h>
-#include <sys/cred.h>
-#include <sys/attr.h>
-#include <sys/zpl.h>
-#include <sys/zil.h>
-
-/*
- * Programming rules.
- *
- * Each vnode op performs some logical unit of work.  To do this, the ZPL must
- * properly lock its in-core state, create a DMU transaction, do the work,
- * record this work in the intent log (ZIL), commit the DMU transaction,
- * and wait for the intent log to commit if it is a synchronous operation.
- * Moreover, the vnode ops must work in both normal and log replay context.
- * The ordering of events is important to avoid deadlocks and references
- * to freed memory.  The example below illustrates the following Big Rules:
- *
- *  (1) A check must be made in each zfs thread for a mounted file system.
- *     This is done avoiding races using ZFS_ENTER(zfsvfs).
- *      A ZFS_EXIT(zfsvfs) is needed before all returns.  Any znodes
- *      must be checked with ZFS_VERIFY_ZP(zp).  Both of these macros
- *      can return EIO from the calling function.
- *
- *  (2)        iput() should always be the last thing except for zil_commit()
- *     (if necessary) and ZFS_EXIT(). This is for 3 reasons:
- *     First, if it's the last reference, the vnode/znode
- *     can be freed, so the zp may point to freed memory.  Second, the last
- *     reference will call zfs_zinactive(), which may induce a lot of work --
- *     pushing cached pages (which acquires range locks) and syncing out
- *     cached atime changes.  Third, zfs_zinactive() may require a new tx,
- *     which could deadlock the system if you were already holding one.
- *     If you must call iput() within a tx then use zfs_iput_async().
- *
- *  (3)        All range locks must be grabbed before calling dmu_tx_assign(),
- *     as they can span dmu_tx_assign() calls.
- *
- *  (4) If ZPL locks are held, pass TXG_NOWAIT as the second argument to
- *      dmu_tx_assign().  This is critical because we don't want to block
- *      while holding locks.
- *
- *     If no ZPL locks are held (aside from ZFS_ENTER()), use TXG_WAIT.  This
- *     reduces lock contention and CPU usage when we must wait (note that if
- *     throughput is constrained by the storage, nearly every transaction
- *     must wait).
- *
- *      Note, in particular, that if a lock is sometimes acquired before
- *      the tx assigns, and sometimes after (e.g. z_lock), then failing
- *      to use a non-blocking assign can deadlock the system.  The scenario:
- *
- *     Thread A has grabbed a lock before calling dmu_tx_assign().
- *     Thread B is in an already-assigned tx, and blocks for this lock.
- *     Thread A calls dmu_tx_assign(TXG_WAIT) and blocks in txg_wait_open()
- *     forever, because the previous txg can't quiesce until B's tx commits.
- *
- *     If dmu_tx_assign() returns ERESTART and zfsvfs->z_assign is TXG_NOWAIT,
- *     then drop all locks, call dmu_tx_wait(), and try again.  On subsequent
- *     calls to dmu_tx_assign(), pass TXG_NOTHROTTLE in addition to TXG_NOWAIT,
- *     to indicate that this operation has already called dmu_tx_wait().
- *     This will ensure that we don't retry forever, waiting a short bit
- *     each time.
- *
- *  (5)        If the operation succeeded, generate the intent log entry for it
- *     before dropping locks.  This ensures that the ordering of events
- *     in the intent log matches the order in which they actually occurred.
- *     During ZIL replay the zfs_log_* functions will update the sequence
- *     number to indicate the zil transaction has replayed.
- *
- *  (6)        At the end of each vnode op, the DMU tx must always commit,
- *     regardless of whether there were any errors.
- *
- *  (7)        After dropping all locks, invoke zil_commit(zilog, foid)
- *     to ensure that synchronous semantics are provided when necessary.
- *
- * In general, this is how things should be ordered in each vnode op:
- *
- *     ZFS_ENTER(zfsvfs);              // exit if unmounted
- * top:
- *     zfs_dirent_lock(&dl, ...)       // lock directory entry (may igrab())
- *     rw_enter(...);                  // grab any other locks you need
- *     tx = dmu_tx_create(...);        // get DMU tx
- *     dmu_tx_hold_*();                // hold each object you might modify
- *     error = dmu_tx_assign(tx, (waited ? TXG_NOTHROTTLE : 0) | TXG_NOWAIT);
- *     if (error) {
- *             rw_exit(...);           // drop locks
- *             zfs_dirent_unlock(dl);  // unlock directory entry
- *             iput(...);              // release held vnodes
- *             if (error == ERESTART) {
- *                     waited = B_TRUE;
- *                     dmu_tx_wait(tx);
- *                     dmu_tx_abort(tx);
- *                     goto top;
- *             }
- *             dmu_tx_abort(tx);       // abort DMU tx
- *             ZFS_EXIT(zfsvfs);       // finished in zfs
- *             return (error);         // really out of space
- *     }
- *     error = do_real_work();         // do whatever this VOP does
- *     if (error == 0)
- *             zfs_log_*(...);         // on success, make ZIL entry
- *     dmu_tx_commit(tx);              // commit DMU tx -- error or not
- *     rw_exit(...);                   // drop locks
- *     zfs_dirent_unlock(dl);          // unlock directory entry
- *     iput(...);                      // release held vnodes
- *     zil_commit(zilog, foid);        // synchronous when necessary
- *     ZFS_EXIT(zfsvfs);               // finished in zfs
- *     return (error);                 // done, report error
- */
+#include <sys/zfs_quota.h>
+#include <sys/zfs_vfsops.h>
+#include <sys/zfs_znode.h>
 
-/*
- * Virus scanning is unsupported.  It would be possible to add a hook
- * here to performance the required virus scan.  This could be done
- * entirely in the kernel or potentially as an update to invoke a
- * scanning utility.
- */
-static int
-zfs_vscan(struct inode *ip, cred_t *cr, int async)
-{
-       return (0);
-}
 
-/* ARGSUSED */
 int
-zfs_open(struct inode *ip, int mode, int flag, cred_t *cr)
+zfs_fsync(znode_t *zp, int syncflag, cred_t *cr)
 {
-       znode_t *zp = ITOZ(ip);
-       zfsvfs_t *zfsvfs = ITOZSB(ip);
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(zp);
-
-       /* Honor ZFS_APPENDONLY file attribute */
-       if ((mode & FMODE_WRITE) && (zp->z_pflags & ZFS_APPENDONLY) &&
-           ((flag & O_APPEND) == 0)) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EPERM));
-       }
+       int error = 0;
+       zfsvfs_t *zfsvfs = ZTOZSB(zp);
 
-       /* Virus scan eligible files on open */
-       if (!zfs_has_ctldir(zp) && zfsvfs->z_vscan && S_ISREG(ip->i_mode) &&
-           !(zp->z_pflags & ZFS_AV_QUARANTINED) && zp->z_size > 0) {
-               if (zfs_vscan(ip, cr, 0) != 0) {
-                       ZFS_EXIT(zfsvfs);
-                       return (SET_ERROR(EACCES));
-               }
+       if (zfsvfs->z_os->os_sync != ZFS_SYNC_DISABLED) {
+               if ((error = zfs_enter_verify_zp(zfsvfs, zp, FTAG)) != 0)
+                       return (error);
+               atomic_inc_32(&zp->z_sync_writes_cnt);
+               zil_commit(zfsvfs->z_log, zp->z_id);
+               atomic_dec_32(&zp->z_sync_writes_cnt);
+               zfs_exit(zfsvfs, FTAG);
        }
-
-       /* Keep a count of the synchronous opens in the znode */
-       if (flag & O_SYNC)
-               atomic_inc_32(&zp->z_sync_cnt);
-
-       ZFS_EXIT(zfsvfs);
-       return (0);
+       return (error);
 }
 
-/* ARGSUSED */
-int
-zfs_close(struct inode *ip, int flag, cred_t *cr)
-{
-       znode_t *zp = ITOZ(ip);
-       zfsvfs_t *zfsvfs = ITOZSB(ip);
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(zp);
-
-       /* Decrement the synchronous opens in the znode */
-       if (flag & O_SYNC)
-               atomic_dec_32(&zp->z_sync_cnt);
-
-       if (!zfs_has_ctldir(zp) && zfsvfs->z_vscan && S_ISREG(ip->i_mode) &&
-           !(zp->z_pflags & ZFS_AV_QUARANTINED) && zp->z_size > 0)
-               VERIFY(zfs_vscan(ip, cr, 1) == 0);
-
-       ZFS_EXIT(zfsvfs);
-       return (0);
-}
 
 #if defined(SEEK_HOLE) && defined(SEEK_DATA)
 /*
@@ -256,9 +83,9 @@ zfs_close(struct inode *ip, int flag, cred_t *cr)
  * data (cmd == SEEK_DATA). "off" is an in/out parameter.
  */
 static int
-zfs_holey_common(struct inode *ip, int cmd, loff_t *off)
+zfs_holey_common(znode_t *zp, ulong_t cmd, loff_t *off)
 {
-       znode_t *zp = ITOZ(ip);
+       zfs_locked_range_t *lr;
        uint64_t noff = (uint64_t)*off; /* new offset */
        uint64_t file_sz;
        int error;
@@ -269,17 +96,23 @@ zfs_holey_common(struct inode *ip, int cmd, loff_t *off)
                return (SET_ERROR(ENXIO));
        }
 
-       if (cmd == SEEK_HOLE)
+       if (cmd == F_SEEK_HOLE)
                hole = B_TRUE;
        else
                hole = B_FALSE;
 
+       /* Flush any mmap()'d data to disk */
+       if (zn_has_cached_data(zp, 0, file_sz - 1))
+               zn_flush_cached_data(zp, B_FALSE);
+
+       lr = zfs_rangelock_enter(&zp->z_rangelock, 0, UINT64_MAX, RL_READER);
        error = dmu_offset_next(ZTOZSB(zp)->z_os, zp->z_id, hole, &noff);
+       zfs_rangelock_exit(lr);
 
        if (error == ESRCH)
                return (SET_ERROR(ENXIO));
 
-       /* file was dirty, so fall back to using generic logic */
+       /* File was dirty, so fall back to using generic logic */
        if (error == EBUSY) {
                if (hole)
                        *off = file_sz;
@@ -306,134 +139,58 @@ zfs_holey_common(struct inode *ip, int cmd, loff_t *off)
 }
 
 int
-zfs_holey(struct inode *ip, int cmd, loff_t *off)
+zfs_holey(znode_t *zp, ulong_t cmd, loff_t *off)
 {
-       znode_t *zp = ITOZ(ip);
-       zfsvfs_t *zfsvfs = ITOZSB(ip);
+       zfsvfs_t *zfsvfs = ZTOZSB(zp);
        int error;
 
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(zp);
+       if ((error = zfs_enter_verify_zp(zfsvfs, zp, FTAG)) != 0)
+               return (error);
 
-       error = zfs_holey_common(ip, cmd, off);
+       error = zfs_holey_common(zp, cmd, off);
 
-       ZFS_EXIT(zfsvfs);
+       zfs_exit(zfsvfs, FTAG);
        return (error);
 }
 #endif /* SEEK_HOLE && SEEK_DATA */
 
-#if defined(_KERNEL)
-/*
- * When a file is memory mapped, we must keep the IO data synchronized
- * between the DMU cache and the memory mapped pages.  What this means:
- *
- * On Write:   If we find a memory mapped page, we write to *both*
- *             the page and the dmu buffer.
- */
-static void
-update_pages(struct inode *ip, int64_t start, int len,
-    objset_t *os, uint64_t oid)
-{
-       struct address_space *mp = ip->i_mapping;
-       struct page *pp;
-       uint64_t nbytes;
-       int64_t off;
-       void *pb;
-
-       off = start & (PAGE_SIZE-1);
-       for (start &= PAGE_MASK; len > 0; start += PAGE_SIZE) {
-               nbytes = MIN(PAGE_SIZE - off, len);
-
-               pp = find_lock_page(mp, start >> PAGE_SHIFT);
-               if (pp) {
-                       if (mapping_writably_mapped(mp))
-                               flush_dcache_page(pp);
-
-                       pb = kmap(pp);
-                       (void) dmu_read(os, oid, start+off, nbytes, pb+off,
-                           DMU_READ_PREFETCH);
-                       kunmap(pp);
-
-                       if (mapping_writably_mapped(mp))
-                               flush_dcache_page(pp);
-
-                       mark_page_accessed(pp);
-                       SetPageUptodate(pp);
-                       ClearPageError(pp);
-                       unlock_page(pp);
-                       put_page(pp);
-               }
-
-               len -= nbytes;
-               off = 0;
-       }
-}
-
-/*
- * When a file is memory mapped, we must keep the IO data synchronized
- * between the DMU cache and the memory mapped pages.  What this means:
- *
- * On Read:    We "read" preferentially from memory mapped pages,
- *             else we default from the dmu buffer.
- *
- * NOTE: We will always "break up" the IO into PAGESIZE uiomoves when
- *      the file is memory mapped.
- */
-static int
-mappedread(struct inode *ip, int nbytes, uio_t *uio)
+int
+zfs_access(znode_t *zp, int mode, int flag, cred_t *cr)
 {
-       struct address_space *mp = ip->i_mapping;
-       struct page *pp;
-       znode_t *zp = ITOZ(ip);
-       int64_t start, off;
-       uint64_t bytes;
-       int len = nbytes;
-       int error = 0;
-       void *pb;
-
-       start = uio->uio_loffset;
-       off = start & (PAGE_SIZE-1);
-       for (start &= PAGE_MASK; len > 0; start += PAGE_SIZE) {
-               bytes = MIN(PAGE_SIZE - off, len);
-
-               pp = find_lock_page(mp, start >> PAGE_SHIFT);
-               if (pp) {
-                       ASSERT(PageUptodate(pp));
-
-                       pb = kmap(pp);
-                       error = uiomove(pb + off, bytes, UIO_READ, uio);
-                       kunmap(pp);
+       zfsvfs_t *zfsvfs = ZTOZSB(zp);
+       int error;
 
-                       if (mapping_writably_mapped(mp))
-                               flush_dcache_page(pp);
+       if ((error = zfs_enter_verify_zp(zfsvfs, zp, FTAG)) != 0)
+               return (error);
 
-                       mark_page_accessed(pp);
-                       unlock_page(pp);
-                       put_page(pp);
-               } else {
-                       error = dmu_read_uio_dbuf(sa_get_db(zp->z_sa_hdl),
-                           uio, bytes);
-               }
+       if (flag & V_ACE_MASK)
+#if defined(__linux__)
+               error = zfs_zaccess(zp, mode, flag, B_FALSE, cr,
+                   zfs_init_idmap);
+#else
+               error = zfs_zaccess(zp, mode, flag, B_FALSE, cr,
+                   NULL);
+#endif
+       else
+#if defined(__linux__)
+               error = zfs_zaccess_rwx(zp, mode, flag, cr, zfs_init_idmap);
+#else
+               error = zfs_zaccess_rwx(zp, mode, flag, cr, NULL);
+#endif
 
-               len -= bytes;
-               off = 0;
-               if (error)
-                       break;
-       }
+       zfs_exit(zfsvfs, FTAG);
        return (error);
 }
-#endif /* _KERNEL */
 
-unsigned long zfs_read_chunk_size = 1024 * 1024; /* Tunable */
-unsigned long zfs_delete_blocks = DMU_MAX_DELETEBLKCNT;
+static uint64_t zfs_vnops_read_chunk_size = 1024 * 1024; /* Tunable */
 
 /*
  * Read bytes from specified file into supplied buffer.
  *
- *     IN:     ip      - inode of file to be read from.
+ *     IN:     zp      - inode of file to be read from.
  *             uio     - structure supplying read location, range info,
  *                       and return buffer.
- *             ioflag  - FSYNC flags; used to provide FRSYNC semantics.
+ *             ioflag  - O_SYNC flags; used to provide FRSYNC semantics.
  *                       O_DIRECT flag; used to bypass page cache.
  *             cr      - credentials of caller.
  *
@@ -444,106 +201,92 @@ unsigned long zfs_delete_blocks = DMU_MAX_DELETEBLKCNT;
  * Side Effects:
  *     inode - atime updated if byte count > 0
  */
-/* ARGSUSED */
 int
-zfs_read(struct inode *ip, uio_t *uio, int ioflag, cred_t *cr)
+zfs_read(struct znode *zp, zfs_uio_t *uio, int ioflag, cred_t *cr)
 {
-       znode_t         *zp = ITOZ(ip);
-       zfsvfs_t        *zfsvfs = ITOZSB(ip);
-       ssize_t         n, nbytes;
-       int             error = 0;
-       rl_t            *rl;
-#ifdef HAVE_UIO_ZEROCOPY
-       xuio_t          *xuio = NULL;
-#endif /* HAVE_UIO_ZEROCOPY */
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(zp);
+       (void) cr;
+       int error = 0;
+       boolean_t frsync = B_FALSE;
+
+       zfsvfs_t *zfsvfs = ZTOZSB(zp);
+       if ((error = zfs_enter_verify_zp(zfsvfs, zp, FTAG)) != 0)
+               return (error);
 
        if (zp->z_pflags & ZFS_AV_QUARANTINED) {
-               ZFS_EXIT(zfsvfs);
+               zfs_exit(zfsvfs, FTAG);
                return (SET_ERROR(EACCES));
        }
 
+       /* We don't copy out anything useful for directories. */
+       if (Z_ISDIR(ZTOTYPE(zp))) {
+               zfs_exit(zfsvfs, FTAG);
+               return (SET_ERROR(EISDIR));
+       }
+
        /*
         * Validate file offset
         */
-       if (uio->uio_loffset < (offset_t)0) {
-               ZFS_EXIT(zfsvfs);
+       if (zfs_uio_offset(uio) < (offset_t)0) {
+               zfs_exit(zfsvfs, FTAG);
                return (SET_ERROR(EINVAL));
        }
 
        /*
         * Fasttrack empty reads
         */
-       if (uio->uio_resid == 0) {
-               ZFS_EXIT(zfsvfs);
+       if (zfs_uio_resid(uio) == 0) {
+               zfs_exit(zfsvfs, FTAG);
                return (0);
        }
 
+#ifdef FRSYNC
        /*
         * If we're in FRSYNC mode, sync out this znode before reading it.
         * Only do this for non-snapshots.
+        *
+        * Some platforms do not support FRSYNC and instead map it
+        * to O_SYNC, which results in unnecessary calls to zil_commit. We
+        * only honor FRSYNC requests on platforms which support it.
         */
+       frsync = !!(ioflag & FRSYNC);
+#endif
        if (zfsvfs->z_log &&
-           (ioflag & FRSYNC || zfsvfs->z_os->os_sync == ZFS_SYNC_ALWAYS))
+           (frsync || zfsvfs->z_os->os_sync == ZFS_SYNC_ALWAYS))
                zil_commit(zfsvfs->z_log, zp->z_id);
 
        /*
         * Lock the range against changes.
         */
-       rl = zfs_range_lock(&zp->z_range_lock, uio->uio_loffset, uio->uio_resid,
-           RL_READER);
+       zfs_locked_range_t *lr = zfs_rangelock_enter(&zp->z_rangelock,
+           zfs_uio_offset(uio), zfs_uio_resid(uio), RL_READER);
 
        /*
         * If we are reading past end-of-file we can skip
         * to the end; but we might still need to set atime.
         */
-       if (uio->uio_loffset >= zp->z_size) {
+       if (zfs_uio_offset(uio) >= zp->z_size) {
                error = 0;
                goto out;
        }
 
-       ASSERT(uio->uio_loffset < zp->z_size);
-       n = MIN(uio->uio_resid, zp->z_size - uio->uio_loffset);
-
-#ifdef HAVE_UIO_ZEROCOPY
-       if ((uio->uio_extflg == UIO_XUIO) &&
-           (((xuio_t *)uio)->xu_type == UIOTYPE_ZEROCOPY)) {
-               int nblk;
-               int blksz = zp->z_blksz;
-               uint64_t offset = uio->uio_loffset;
-
-               xuio = (xuio_t *)uio;
-               if ((ISP2(blksz))) {
-                       nblk = (P2ROUNDUP(offset + n, blksz) - P2ALIGN(offset,
-                           blksz)) / blksz;
-               } else {
-                       ASSERT(offset + n <= blksz);
-                       nblk = 1;
-               }
-               (void) dmu_xuio_init(xuio, nblk);
-
-               if (vn_has_cached_data(ip)) {
-                       /*
-                        * For simplicity, we always allocate a full buffer
-                        * even if we only expect to read a portion of a block.
-                        */
-                       while (--nblk >= 0) {
-                               (void) dmu_xuio_add(xuio,
-                                   dmu_request_arcbuf(sa_get_db(zp->z_sa_hdl),
-                                   blksz), 0, blksz);
-                       }
-               }
-       }
-#endif /* HAVE_UIO_ZEROCOPY */
+       ASSERT(zfs_uio_offset(uio) < zp->z_size);
+#if defined(__linux__)
+       ssize_t start_offset = zfs_uio_offset(uio);
+#endif
+       ssize_t n = MIN(zfs_uio_resid(uio), zp->z_size - zfs_uio_offset(uio));
+       ssize_t start_resid = n;
 
        while (n > 0) {
-               nbytes = MIN(n, zfs_read_chunk_size -
-                   P2PHASE(uio->uio_loffset, zfs_read_chunk_size));
-
-               if (zp->z_is_mapped && !(ioflag & O_DIRECT)) {
-                       error = mappedread(ip, nbytes, uio);
+               ssize_t nbytes = MIN(n, zfs_vnops_read_chunk_size -
+                   P2PHASE(zfs_uio_offset(uio), zfs_vnops_read_chunk_size));
+#ifdef UIO_NOCOPY
+               if (zfs_uio_segflg(uio) == UIO_NOCOPY)
+                       error = mappedread_sf(zp, nbytes, uio);
+               else
+#endif
+               if (zn_has_cached_data(zp, zfs_uio_offset(uio),
+                   zfs_uio_offset(uio) + nbytes - 1) && !(ioflag & O_DIRECT)) {
+                       error = mappedread(zp, nbytes, uio);
                } else {
                        error = dmu_read_uio_dbuf(sa_get_db(zp->z_sa_hdl),
                            uio, nbytes);
@@ -553,25 +296,97 @@ zfs_read(struct inode *ip, uio_t *uio, int ioflag, cred_t *cr)
                        /* convert checksum errors into IO errors */
                        if (error == ECKSUM)
                                error = SET_ERROR(EIO);
+
+#if defined(__linux__)
+                       /*
+                        * if we actually read some bytes, bubbling EFAULT
+                        * up to become EAGAIN isn't what we want here...
+                        *
+                        * ...on Linux, at least. On FBSD, doing this breaks.
+                        */
+                       if (error == EFAULT &&
+                           (zfs_uio_offset(uio) - start_offset) != 0)
+                               error = 0;
+#endif
                        break;
                }
 
                n -= nbytes;
        }
+
+       int64_t nread = start_resid - n;
+       dataset_kstats_update_read_kstats(&zfsvfs->z_kstat, nread);
+       task_io_account_read(nread);
 out:
-       zfs_range_unlock(rl);
+       zfs_rangelock_exit(lr);
 
-       ZFS_EXIT(zfsvfs);
+       ZFS_ACCESSTIME_STAMP(zfsvfs, zp);
+       zfs_exit(zfsvfs, FTAG);
        return (error);
 }
 
+static void
+zfs_clear_setid_bits_if_necessary(zfsvfs_t *zfsvfs, znode_t *zp, cred_t *cr,
+    uint64_t *clear_setid_bits_txgp, dmu_tx_t *tx)
+{
+       zilog_t *zilog = zfsvfs->z_log;
+       const uint64_t uid = KUID_TO_SUID(ZTOUID(zp));
+
+       ASSERT(clear_setid_bits_txgp != NULL);
+       ASSERT(tx != NULL);
+
+       /*
+        * Clear Set-UID/Set-GID bits on successful write if not
+        * privileged and at least one of the execute bits is set.
+        *
+        * It would be nice to do this after all writes have
+        * been done, but that would still expose the ISUID/ISGID
+        * to another app after the partial write is committed.
+        *
+        * Note: we don't call zfs_fuid_map_id() here because
+        * user 0 is not an ephemeral uid.
+        */
+       mutex_enter(&zp->z_acl_lock);
+       if ((zp->z_mode & (S_IXUSR | (S_IXUSR >> 3) | (S_IXUSR >> 6))) != 0 &&
+           (zp->z_mode & (S_ISUID | S_ISGID)) != 0 &&
+           secpolicy_vnode_setid_retain(zp, cr,
+           ((zp->z_mode & S_ISUID) != 0 && uid == 0)) != 0) {
+               uint64_t newmode;
+
+               zp->z_mode &= ~(S_ISUID | S_ISGID);
+               newmode = zp->z_mode;
+               (void) sa_update(zp->z_sa_hdl, SA_ZPL_MODE(zfsvfs),
+                   (void *)&newmode, sizeof (uint64_t), tx);
+
+               mutex_exit(&zp->z_acl_lock);
+
+               /*
+                * Make sure SUID/SGID bits will be removed when we replay the
+                * log. If the setid bits are keep coming back, don't log more
+                * than one TX_SETATTR per transaction group.
+                */
+               if (*clear_setid_bits_txgp != dmu_tx_get_txg(tx)) {
+                       vattr_t va = {0};
+
+                       va.va_mask = ATTR_MODE;
+                       va.va_nodeid = zp->z_id;
+                       va.va_mode = newmode;
+                       zfs_log_setattr(zilog, tx, TX_SETATTR, zp, &va,
+                           ATTR_MODE, NULL);
+                       *clear_setid_bits_txgp = dmu_tx_get_txg(tx);
+               }
+       } else {
+               mutex_exit(&zp->z_acl_lock);
+       }
+}
+
 /*
  * Write the bytes to a file.
  *
- *     IN:     ip      - inode of file to be written to.
+ *     IN:     zp      - znode of file to be written to.
  *             uio     - structure supplying write location, range info,
  *                       and data buffer.
- *             ioflag  - FAPPEND flag set if in append mode.
+ *             ioflag  - O_APPEND flag set if in append mode.
  *                       O_DIRECT flag; used to bypass page cache.
  *             cr      - credentials of caller.
  *
@@ -583,51 +398,27 @@ out:
  * Timestamps:
  *     ip - ctime|mtime updated if byte count > 0
  */
-
-/* ARGSUSED */
 int
-zfs_write(struct inode *ip, uio_t *uio, int ioflag, cred_t *cr)
+zfs_write(znode_t *zp, zfs_uio_t *uio, int ioflag, cred_t *cr)
 {
-       znode_t         *zp = ITOZ(ip);
-       rlim64_t        limit = uio->uio_limit;
-       ssize_t         start_resid = uio->uio_resid;
-       ssize_t         tx_bytes;
-       uint64_t        end_size;
-       dmu_tx_t        *tx;
-       zfsvfs_t        *zfsvfs = ZTOZSB(zp);
-       zilog_t         *zilog;
-       offset_t        woff;
-       ssize_t         n, nbytes;
-       rl_t            *rl;
-       int             max_blksz = zfsvfs->z_max_blksz;
-       int             error = 0;
-       arc_buf_t       *abuf;
-       const iovec_t   *aiov = NULL;
-       xuio_t          *xuio = NULL;
-       int             write_eof;
-       int             count = 0;
-       sa_bulk_attr_t  bulk[4];
-       uint64_t        mtime[2], ctime[2];
-       uint32_t        uid;
-#ifdef HAVE_UIO_ZEROCOPY
-       int             i_iov = 0;
-       const iovec_t   *iovp = uio->uio_iov;
-       ASSERTV(int     iovcnt = uio->uio_iovcnt);
-#endif
+       int error = 0, error1;
+       ssize_t start_resid = zfs_uio_resid(uio);
+       uint64_t clear_setid_bits_txg = 0;
 
        /*
         * Fasttrack empty write
         */
-       n = start_resid;
+       ssize_t n = start_resid;
        if (n == 0)
                return (0);
 
-       if (limit == RLIM64_INFINITY || limit > MAXOFFSET_T)
-               limit = MAXOFFSET_T;
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(zp);
+       zfsvfs_t *zfsvfs = ZTOZSB(zp);
+       if ((error = zfs_enter_verify_zp(zfsvfs, zp, FTAG)) != 0)
+               return (error);
 
+       sa_bulk_attr_t bulk[4];
+       int count = 0;
+       uint64_t mtime[2], ctime[2];
        SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_MTIME(zfsvfs), NULL, &mtime, 16);
        SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_CTIME(zfsvfs), NULL, &ctime, 16);
        SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_SIZE(zfsvfs), NULL,
@@ -640,55 +431,53 @@ zfs_write(struct inode *ip, uio_t *uio, int ioflag, cred_t *cr)
         * so check it explicitly here.
         */
        if (zfs_is_readonly(zfsvfs)) {
-               ZFS_EXIT(zfsvfs);
+               zfs_exit(zfsvfs, FTAG);
                return (SET_ERROR(EROFS));
        }
 
        /*
-        * If immutable or not appending then return EPERM
+        * If immutable or not appending then return EPERM.
+        * Intentionally allow ZFS_READONLY through here.
+        * See zfs_zaccess_common()
         */
-       if ((zp->z_pflags & (ZFS_IMMUTABLE | ZFS_READONLY)) ||
-           ((zp->z_pflags & ZFS_APPENDONLY) && !(ioflag & FAPPEND) &&
-           (uio->uio_loffset < zp->z_size))) {
-               ZFS_EXIT(zfsvfs);
+       if ((zp->z_pflags & ZFS_IMMUTABLE) ||
+           ((zp->z_pflags & ZFS_APPENDONLY) && !(ioflag & O_APPEND) &&
+           (zfs_uio_offset(uio) < zp->z_size))) {
+               zfs_exit(zfsvfs, FTAG);
                return (SET_ERROR(EPERM));
        }
 
-       zilog = zfsvfs->z_log;
-
        /*
         * Validate file offset
         */
-       woff = ioflag & FAPPEND ? zp->z_size : uio->uio_loffset;
+       offset_t woff = ioflag & O_APPEND ? zp->z_size : zfs_uio_offset(uio);
        if (woff < 0) {
-               ZFS_EXIT(zfsvfs);
+               zfs_exit(zfsvfs, FTAG);
                return (SET_ERROR(EINVAL));
        }
 
        /*
         * Pre-fault the pages to ensure slow (eg NFS) pages
         * don't hold up txg.
-        * Skip this if uio contains loaned arc_buf.
         */
-#ifdef HAVE_UIO_ZEROCOPY
-       if ((uio->uio_extflg == UIO_XUIO) &&
-           (((xuio_t *)uio)->xu_type == UIOTYPE_ZEROCOPY))
-               xuio = (xuio_t *)uio;
-       else
-#endif
-               uio_prefaultpages(MIN(n, max_blksz), uio);
+       ssize_t pfbytes = MIN(n, DMU_MAX_ACCESS >> 1);
+       if (zfs_uio_prefaultpages(pfbytes, uio)) {
+               zfs_exit(zfsvfs, FTAG);
+               return (SET_ERROR(EFAULT));
+       }
 
        /*
         * If in append mode, set the io offset pointer to eof.
         */
-       if (ioflag & FAPPEND) {
+       zfs_locked_range_t *lr;
+       if (ioflag & O_APPEND) {
                /*
                 * Obtain an appending range lock to guarantee file append
                 * semantics.  We reset the write offset once we have the lock.
                 */
-               rl = zfs_range_lock(&zp->z_range_lock, 0, n, RL_APPEND);
-               woff = rl->r_off;
-               if (rl->r_len == UINT64_MAX) {
+               lr = zfs_rangelock_enter(&zp->z_rangelock, 0, n, RL_APPEND);
+               woff = lr->lr_offset;
+               if (lr->lr_length == UINT64_MAX) {
                        /*
                         * We overlocked the file because this write will cause
                         * the file block size to increase.
@@ -696,29 +485,41 @@ zfs_write(struct inode *ip, uio_t *uio, int ioflag, cred_t *cr)
                         */
                        woff = zp->z_size;
                }
-               uio->uio_loffset = woff;
+               zfs_uio_setoffset(uio, woff);
        } else {
                /*
                 * Note that if the file block size will change as a result of
                 * this write, then this range lock will lock the entire file
                 * so that we can re-write the block safely.
                 */
-               rl = zfs_range_lock(&zp->z_range_lock, woff, n, RL_WRITER);
+               lr = zfs_rangelock_enter(&zp->z_rangelock, woff, n, RL_WRITER);
+       }
+
+       if (zn_rlimit_fsize_uio(zp, uio)) {
+               zfs_rangelock_exit(lr);
+               zfs_exit(zfsvfs, FTAG);
+               return (SET_ERROR(EFBIG));
        }
 
+       const rlim64_t limit = MAXOFFSET_T;
+
        if (woff >= limit) {
-               zfs_range_unlock(rl);
-               ZFS_EXIT(zfsvfs);
+               zfs_rangelock_exit(lr);
+               zfs_exit(zfsvfs, FTAG);
                return (SET_ERROR(EFBIG));
        }
 
-       if ((woff + n) > limit || woff > (limit - n))
+       if (n > limit - woff)
                n = limit - woff;
 
-       /* Will this write extend the file length? */
-       write_eof = (woff + n > zp->z_size);
+       uint64_t end_size = MAX(zp->z_size, woff + n);
+       zilog_t *zilog = zfsvfs->z_log;
+       boolean_t commit = (ioflag & (O_SYNC | O_DSYNC)) ||
+           (zfsvfs->z_os->os_sync == ZFS_SYNC_ALWAYS);
 
-       end_size = MAX(zp->z_size, woff + n);
+       const uint64_t uid = KUID_TO_SUID(ZTOUID(zp));
+       const uint64_t gid = KGID_TO_SGID(ZTOGID(zp));
+       const uint64_t projid = zp->z_projid;
 
        /*
         * Write the file in reasonable size chunks.  Each chunk is written
@@ -726,32 +527,42 @@ zfs_write(struct inode *ip, uio_t *uio, int ioflag, cred_t *cr)
         * and allows us to do more fine-grained space accounting.
         */
        while (n > 0) {
-               abuf = NULL;
-               woff = uio->uio_loffset;
-               if (zfs_owner_overquota(zfsvfs, zp, B_FALSE) ||
-                   zfs_owner_overquota(zfsvfs, zp, B_TRUE)) {
-                       if (abuf != NULL)
-                               dmu_return_arcbuf(abuf);
+               woff = zfs_uio_offset(uio);
+
+               if (zfs_id_overblockquota(zfsvfs, DMU_USERUSED_OBJECT, uid) ||
+                   zfs_id_overblockquota(zfsvfs, DMU_GROUPUSED_OBJECT, gid) ||
+                   (projid != ZFS_DEFAULT_PROJID &&
+                   zfs_id_overblockquota(zfsvfs, DMU_PROJECTUSED_OBJECT,
+                   projid))) {
                        error = SET_ERROR(EDQUOT);
                        break;
                }
 
-               if (xuio && abuf == NULL) {
-#ifdef HAVE_UIO_ZEROCOPY
-                       ASSERT(i_iov < iovcnt);
-                       ASSERT3U(uio->uio_segflg, !=, UIO_BVEC);
-                       aiov = &iovp[i_iov];
-                       abuf = dmu_xuio_arcbuf(xuio, i_iov);
-                       dmu_xuio_clear(xuio, i_iov);
-                       ASSERT((aiov->iov_base == abuf->b_data) ||
-                           ((char *)aiov->iov_base - (char *)abuf->b_data +
-                           aiov->iov_len == arc_buf_size(abuf)));
-                       i_iov++;
-#endif
-               } else if (abuf == NULL && n >= max_blksz &&
-                   woff >= zp->z_size &&
-                   P2PHASE(woff, max_blksz) == 0 &&
-                   zp->z_blksz == max_blksz) {
+               uint64_t blksz;
+               if (lr->lr_length == UINT64_MAX && zp->z_size <= zp->z_blksz) {
+                       if (zp->z_blksz > zfsvfs->z_max_blksz &&
+                           !ISP2(zp->z_blksz)) {
+                               /*
+                                * File's blocksize is already larger than the
+                                * "recordsize" property.  Only let it grow to
+                                * the next power of 2.
+                                */
+                               blksz = 1 << highbit64(zp->z_blksz);
+                       } else {
+                               blksz = zfsvfs->z_max_blksz;
+                       }
+                       blksz = MIN(blksz, P2ROUNDUP(end_size,
+                           SPA_MINBLOCKSIZE));
+                       blksz = MAX(blksz, zp->z_blksz);
+               } else {
+                       blksz = zp->z_blksz;
+               }
+
+               arc_buf_t *abuf = NULL;
+               ssize_t nbytes = n;
+               if (n >= blksz && woff >= zp->z_size &&
+                   P2PHASE(woff, blksz) == 0 &&
+                   (blksz >= SPA_OLD_MAXBLOCKSIZE || n < 4 * blksz)) {
                        /*
                         * This write covers a full block.  "Borrow" a buffer
                         * from the dmu so that we can fill it before we enter
@@ -759,26 +570,37 @@ zfs_write(struct inode *ip, uio_t *uio, int ioflag, cred_t *cr)
                         * holding up the transaction if the data copy hangs
                         * up on a pagefault (e.g., from an NFS server mapping).
                         */
-                       size_t cbytes;
-
                        abuf = dmu_request_arcbuf(sa_get_db(zp->z_sa_hdl),
-                           max_blksz);
+                           blksz);
                        ASSERT(abuf != NULL);
-                       ASSERT(arc_buf_size(abuf) == max_blksz);
-                       if ((error = uiocopy(abuf->b_data, max_blksz,
-                           UIO_WRITE, uio, &cbytes))) {
+                       ASSERT(arc_buf_size(abuf) == blksz);
+                       if ((error = zfs_uiocopy(abuf->b_data, blksz,
+                           UIO_WRITE, uio, &nbytes))) {
                                dmu_return_arcbuf(abuf);
                                break;
                        }
-                       ASSERT(cbytes == max_blksz);
+                       ASSERT3S(nbytes, ==, blksz);
+               } else {
+                       nbytes = MIN(n, (DMU_MAX_ACCESS >> 1) -
+                           P2PHASE(woff, blksz));
+                       if (pfbytes < nbytes) {
+                               if (zfs_uio_prefaultpages(nbytes, uio)) {
+                                       error = SET_ERROR(EFAULT);
+                                       break;
+                               }
+                               pfbytes = nbytes;
+                       }
                }
 
                /*
                 * Start a transaction.
                 */
-               tx = dmu_tx_create(zfsvfs->z_os);
+               dmu_tx_t *tx = dmu_tx_create(zfsvfs->z_os);
                dmu_tx_hold_sa(tx, zp->z_sa_hdl, B_FALSE);
-               dmu_tx_hold_write(tx, zp->z_id, woff, MIN(n, max_blksz));
+               dmu_buf_impl_t *db = (dmu_buf_impl_t *)sa_get_db(zp->z_sa_hdl);
+               DB_DNODE_ENTER(db);
+               dmu_tx_hold_write_by_dnode(tx, DB_DNODE(db), woff, nbytes);
+               DB_DNODE_EXIT(db);
                zfs_sa_upgrade_txholds(tx, zp);
                error = dmu_tx_assign(tx, TXG_WAIT);
                if (error) {
@@ -789,69 +611,86 @@ zfs_write(struct inode *ip, uio_t *uio, int ioflag, cred_t *cr)
                }
 
                /*
-                * If zfs_range_lock() over-locked we grow the blocksize
-                * and then reduce the lock range.  This will only happen
-                * on the first iteration since zfs_range_reduce() will
-                * shrink down r_len to the appropriate size.
+                * NB: We must call zfs_clear_setid_bits_if_necessary before
+                * committing the transaction!
                 */
-               if (rl->r_len == UINT64_MAX) {
-                       uint64_t new_blksz;
-
-                       if (zp->z_blksz > max_blksz) {
-                               /*
-                                * File's blocksize is already larger than the
-                                * "recordsize" property.  Only let it grow to
-                                * the next power of 2.
-                                */
-                               ASSERT(!ISP2(zp->z_blksz));
-                               new_blksz = MIN(end_size,
-                                   1 << highbit64(zp->z_blksz));
-                       } else {
-                               new_blksz = MIN(end_size, max_blksz);
-                       }
-                       zfs_grow_blocksize(zp, new_blksz, tx);
-                       zfs_range_reduce(rl, woff, n);
-               }
 
                /*
-                * XXX - should we really limit each write to z_max_blksz?
-                * Perhaps we should use SPA_MAXBLOCKSIZE chunks?
+                * If rangelock_enter() over-locked we grow the blocksize
+                * and then reduce the lock range.  This will only happen
+                * on the first iteration since rangelock_reduce() will
+                * shrink down lr_length to the appropriate size.
                 */
-               nbytes = MIN(n, max_blksz - P2PHASE(woff, max_blksz));
+               if (lr->lr_length == UINT64_MAX) {
+                       zfs_grow_blocksize(zp, blksz, tx);
+                       zfs_rangelock_reduce(lr, woff, n);
+               }
 
+               ssize_t tx_bytes;
                if (abuf == NULL) {
-                       tx_bytes = uio->uio_resid;
+                       tx_bytes = zfs_uio_resid(uio);
+                       zfs_uio_fault_disable(uio, B_TRUE);
                        error = dmu_write_uio_dbuf(sa_get_db(zp->z_sa_hdl),
                            uio, nbytes, tx);
-                       tx_bytes -= uio->uio_resid;
+                       zfs_uio_fault_disable(uio, B_FALSE);
+#ifdef __linux__
+                       if (error == EFAULT) {
+                               zfs_clear_setid_bits_if_necessary(zfsvfs, zp,
+                                   cr, &clear_setid_bits_txg, tx);
+                               dmu_tx_commit(tx);
+                               /*
+                                * Account for partial writes before
+                                * continuing the loop.
+                                * Update needs to occur before the next
+                                * zfs_uio_prefaultpages, or prefaultpages may
+                                * error, and we may break the loop early.
+                                */
+                               n -= tx_bytes - zfs_uio_resid(uio);
+                               pfbytes -= tx_bytes - zfs_uio_resid(uio);
+                               continue;
+                       }
+#endif
+                       /*
+                        * On FreeBSD, EFAULT should be propagated back to the
+                        * VFS, which will handle faulting and will retry.
+                        */
+                       if (error != 0 && error != EFAULT) {
+                               zfs_clear_setid_bits_if_necessary(zfsvfs, zp,
+                                   cr, &clear_setid_bits_txg, tx);
+                               dmu_tx_commit(tx);
+                               break;
+                       }
+                       tx_bytes -= zfs_uio_resid(uio);
                } else {
-                       tx_bytes = nbytes;
-                       ASSERT(xuio == NULL || tx_bytes == aiov->iov_len);
                        /*
-                        * If this is not a full block write, but we are
-                        * extending the file past EOF and this data starts
-                        * block-aligned, use assign_arcbuf().  Otherwise,
-                        * write via dmu_write().
+                        * Thus, we're writing a full block at a block-aligned
+                        * offset and extending the file past EOF.
+                        *
+                        * dmu_assign_arcbuf_by_dbuf() will directly assign the
+                        * arc buffer to a dbuf.
                         */
-                       if (tx_bytes < max_blksz && (!write_eof ||
-                           aiov->iov_base != abuf->b_data)) {
-                               ASSERT(xuio);
-                               dmu_write(zfsvfs->z_os, zp->z_id, woff,
-                                   /* cppcheck-suppress nullPointer */
-                                   aiov->iov_len, aiov->iov_base, tx);
+                       error = dmu_assign_arcbuf_by_dbuf(
+                           sa_get_db(zp->z_sa_hdl), woff, abuf, tx);
+                       if (error != 0) {
+                               /*
+                                * XXX This might not be necessary if
+                                * dmu_assign_arcbuf_by_dbuf is guaranteed
+                                * to be atomic.
+                                */
+                               zfs_clear_setid_bits_if_necessary(zfsvfs, zp,
+                                   cr, &clear_setid_bits_txg, tx);
                                dmu_return_arcbuf(abuf);
-                               xuio_stat_wbuf_copied();
-                       } else {
-                               ASSERT(xuio || tx_bytes == max_blksz);
-                               dmu_assign_arcbuf_by_dbuf(
-                                   sa_get_db(zp->z_sa_hdl), woff, abuf, tx);
+                               dmu_tx_commit(tx);
+                               break;
                        }
-                       ASSERT(tx_bytes <= uio->uio_resid);
-                       uioskip(uio, tx_bytes);
+                       ASSERT3S(nbytes, <=, zfs_uio_resid(uio));
+                       zfs_uioskip(uio, nbytes);
+                       tx_bytes = nbytes;
                }
-               if (tx_bytes && zp->z_is_mapped && !(ioflag & O_DIRECT)) {
-                       update_pages(ip, woff,
-                           tx_bytes, zfsvfs->z_os, zp->z_id);
+               if (tx_bytes &&
+                   zn_has_cached_data(zp, woff, woff + tx_bytes - 1) &&
+                   !(ioflag & O_DIRECT)) {
+                       update_pages(zp, woff, tx_bytes, zfsvfs->z_os);
                }
 
                /*
@@ -866,31 +705,8 @@ zfs_write(struct inode *ip, uio_t *uio, int ioflag, cred_t *cr)
                        break;
                }
 
-               /*
-                * Clear Set-UID/Set-GID bits on successful write if not
-                * privileged and at least one of the execute bits is set.
-                *
-                * It would be nice to to this after all writes have
-                * been done, but that would still expose the ISUID/ISGID
-                * to another app after the partial write is committed.
-                *
-                * Note: we don't call zfs_fuid_map_id() here because
-                * user 0 is not an ephemeral uid.
-                */
-               mutex_enter(&zp->z_acl_lock);
-               uid = KUID_TO_SUID(ip->i_uid);
-               if ((zp->z_mode & (S_IXUSR | (S_IXUSR >> 3) |
-                   (S_IXUSR >> 6))) != 0 &&
-                   (zp->z_mode & (S_ISUID | S_ISGID)) != 0 &&
-                   secpolicy_vnode_setid_retain(cr,
-                   ((zp->z_mode & S_ISUID) != 0 && uid == 0)) != 0) {
-                       uint64_t newmode;
-                       zp->z_mode &= ~(S_ISUID | S_ISGID);
-                       ip->i_mode = newmode = zp->z_mode;
-                       (void) sa_update(zp->z_sa_hdl, SA_ZPL_MODE(zfsvfs),
-                           (void *)&newmode, sizeof (uint64_t), tx);
-               }
-               mutex_exit(&zp->z_acl_lock);
+               zfs_clear_setid_bits_if_necessary(zfsvfs, zp, cr,
+                   &clear_setid_bits_txg, tx);
 
                zfs_tstamp_update_setup(zp, CONTENT_MODIFIED, mtime, ctime);
 
@@ -898,10 +714,10 @@ zfs_write(struct inode *ip, uio_t *uio, int ioflag, cred_t *cr)
                 * Update the file size (zp_size) if it has changed;
                 * account for possible concurrent updates.
                 */
-               while ((end_size = zp->z_size) < uio->uio_loffset) {
+               while ((end_size = zp->z_size) < zfs_uio_offset(uio)) {
                        (void) atomic_cas_64(&zp->z_size, end_size,
-                           uio->uio_loffset);
-                       ASSERT(error == 0);
+                           zfs_uio_offset(uio));
+                       ASSERT(error == 0 || error == EFAULT);
                }
                /*
                 * If we are replaying and eof is non zero then force
@@ -911,93 +727,100 @@ zfs_write(struct inode *ip, uio_t *uio, int ioflag, cred_t *cr)
                if (zfsvfs->z_replay && zfsvfs->z_replay_eof != 0)
                        zp->z_size = zfsvfs->z_replay_eof;
 
-               error = sa_bulk_update(zp->z_sa_hdl, bulk, count, tx);
+               error1 = sa_bulk_update(zp->z_sa_hdl, bulk, count, tx);
+               if (error1 != 0)
+                       /* Avoid clobbering EFAULT. */
+                       error = error1;
 
-               zfs_log_write(zilog, tx, TX_WRITE, zp, woff, tx_bytes, ioflag,
+               /*
+                * NB: During replay, the TX_SETATTR record logged by
+                * zfs_clear_setid_bits_if_necessary must precede any of
+                * the TX_WRITE records logged here.
+                */
+               zfs_log_write(zilog, tx, TX_WRITE, zp, woff, tx_bytes, commit,
                    NULL, NULL);
+
                dmu_tx_commit(tx);
 
                if (error != 0)
                        break;
-               ASSERT(tx_bytes == nbytes);
+               ASSERT3S(tx_bytes, ==, nbytes);
                n -= nbytes;
-
-               if (!xuio && n > 0)
-                       uio_prefaultpages(MIN(n, max_blksz), uio);
+               pfbytes -= nbytes;
        }
 
-       zfs_inode_update(zp);
-       zfs_range_unlock(rl);
+       zfs_znode_update_vfs(zp);
+       zfs_rangelock_exit(lr);
 
        /*
-        * If we're in replay mode, or we made no progress, return error.
-        * Otherwise, it's at least a partial write, so it's successful.
+        * If we're in replay mode, or we made no progress, or the
+        * uio data is inaccessible return an error.  Otherwise, it's
+        * at least a partial write, so it's successful.
         */
-       if (zfsvfs->z_replay || uio->uio_resid == start_resid) {
-               ZFS_EXIT(zfsvfs);
+       if (zfsvfs->z_replay || zfs_uio_resid(uio) == start_resid ||
+           error == EFAULT) {
+               zfs_exit(zfsvfs, FTAG);
                return (error);
        }
 
-       if (ioflag & (FSYNC | FDSYNC) ||
-           zfsvfs->z_os->os_sync == ZFS_SYNC_ALWAYS)
+       if (commit)
                zil_commit(zilog, zp->z_id);
 
-       ZFS_EXIT(zfsvfs);
+       const int64_t nwritten = start_resid - zfs_uio_resid(uio);
+       dataset_kstats_update_write_kstats(&zfsvfs->z_kstat, nwritten);
+       task_io_account_write(nwritten);
+
+       zfs_exit(zfsvfs, FTAG);
        return (0);
 }
 
-/*
- * Drop a reference on the passed inode asynchronously. This ensures
- * that the caller will never drop the last reference on an inode in
- * the current context. Doing so while holding open a tx could result
- * in a deadlock if iput_final() re-enters the filesystem code.
- */
-void
-zfs_iput_async(struct inode *ip)
+int
+zfs_getsecattr(znode_t *zp, vsecattr_t *vsecp, int flag, cred_t *cr)
 {
-       objset_t *os = ITOZSB(ip)->z_os;
+       zfsvfs_t *zfsvfs = ZTOZSB(zp);
+       int error;
+       boolean_t skipaclchk = (flag & ATTR_NOACLCHECK) ? B_TRUE : B_FALSE;
 
-       ASSERT(atomic_read(&ip->i_count) > 0);
-       ASSERT(os != NULL);
+       if ((error = zfs_enter_verify_zp(zfsvfs, zp, FTAG)) != 0)
+               return (error);
+       error = zfs_getacl(zp, vsecp, skipaclchk, cr);
+       zfs_exit(zfsvfs, FTAG);
 
-       if (atomic_read(&ip->i_count) == 1)
-               VERIFY(taskq_dispatch(dsl_pool_iput_taskq(dmu_objset_pool(os)),
-                   (task_func_t *)iput, ip, TQ_SLEEP) != TASKQID_INVALID);
-       else
-               iput(ip);
+       return (error);
 }
 
-void
-zfs_get_done(zgd_t *zgd, int error)
+int
+zfs_setsecattr(znode_t *zp, vsecattr_t *vsecp, int flag, cred_t *cr)
 {
-       znode_t *zp = zgd->zgd_private;
-
-       if (zgd->zgd_db)
-               dmu_buf_rele(zgd->zgd_db, zgd);
-
-       zfs_range_unlock(zgd->zgd_rl);
+       zfsvfs_t *zfsvfs = ZTOZSB(zp);
+       int error;
+       boolean_t skipaclchk = (flag & ATTR_NOACLCHECK) ? B_TRUE : B_FALSE;
+       zilog_t *zilog;
 
-       /*
-        * Release the vnode asynchronously as we currently have the
-        * txg stopped from syncing.
-        */
-       zfs_iput_async(ZTOI(zp));
+       if ((error = zfs_enter_verify_zp(zfsvfs, zp, FTAG)) != 0)
+               return (error);
+       zilog = zfsvfs->z_log;
+       error = zfs_setacl(zp, vsecp, skipaclchk, cr);
 
-       if (error == 0 && zgd->zgd_bp)
-               zil_lwb_add_block(zgd->zgd_lwb, zgd->zgd_bp);
+       if (zfsvfs->z_os->os_sync == ZFS_SYNC_ALWAYS)
+               zil_commit(zilog, 0);
 
-       kmem_free(zgd, sizeof (zgd_t));
+       zfs_exit(zfsvfs, FTAG);
+       return (error);
 }
 
-#ifdef DEBUG
+#ifdef ZFS_DEBUG
 static int zil_fault_io = 0;
 #endif
 
+static void zfs_get_done(zgd_t *zgd, int error);
+
 /*
  * Get data to generate a TX_WRITE intent log record.
  */
 int
-zfs_get_data(void *arg, lr_write_t *lr, char *buf, struct lwb *lwb, zio_t *zio)
+zfs_get_data(void *arg, uint64_t gen, lr_write_t *lr, char *buf,
+    struct lwb *lwb, zio_t *zio)
 {
        zfsvfs_t *zfsvfs = arg;
        objset_t *os = zfsvfs->z_os;
@@ -1008,9 +831,9 @@ zfs_get_data(void *arg, lr_write_t *lr, char *buf, struct lwb *lwb, zio_t *zio)
        dmu_buf_t *db;
        zgd_t *zgd;
        int error = 0;
+       uint64_t zp_gen;
 
        ASSERT3P(lwb, !=, NULL);
-       ASSERT3P(zio, !=, NULL);
        ASSERT3U(size, !=, 0);
 
        /*
@@ -1023,11 +846,21 @@ zfs_get_data(void *arg, lr_write_t *lr, char *buf, struct lwb *lwb, zio_t *zio)
                 * Release the vnode asynchronously as we currently have the
                 * txg stopped from syncing.
                 */
-               zfs_iput_async(ZTOI(zp));
+               zfs_zrele_async(zp);
+               return (SET_ERROR(ENOENT));
+       }
+       /* check if generation number matches */
+       if (sa_lookup(zp->z_sa_hdl, SA_ZPL_GEN(zfsvfs), &zp_gen,
+           sizeof (zp_gen)) != 0) {
+               zfs_zrele_async(zp);
+               return (SET_ERROR(EIO));
+       }
+       if (zp_gen != gen) {
+               zfs_zrele_async(zp);
                return (SET_ERROR(ENOENT));
        }
 
-       zgd = (zgd_t *)kmem_zalloc(sizeof (zgd_t), KM_SLEEP);
+       zgd = kmem_zalloc(sizeof (zgd_t), KM_SLEEP);
        zgd->zgd_lwb = lwb;
        zgd->zgd_private = zp;
 
@@ -1039,8 +872,8 @@ zfs_get_data(void *arg, lr_write_t *lr, char *buf, struct lwb *lwb, zio_t *zio)
         * we don't have to write the data twice.
         */
        if (buf != NULL) { /* immediate write */
-               zgd->zgd_rl = zfs_range_lock(&zp->z_range_lock, offset, size,
-                   RL_READER);
+               zgd->zgd_lr = zfs_rangelock_enter(&zp->z_rangelock,
+                   offset, size, RL_READER);
                /* test for truncation needs to be done while range locked */
                if (offset >= zp->z_size) {
                        error = SET_ERROR(ENOENT);
@@ -1050,6 +883,7 @@ zfs_get_data(void *arg, lr_write_t *lr, char *buf, struct lwb *lwb, zio_t *zio)
                }
                ASSERT(error == 0 || error == ENOENT);
        } else { /* indirect write */
+               ASSERT3P(zio, !=, NULL);
                /*
                 * Have to lock the whole block to ensure when it's
                 * written out and its checksum is being calculated
@@ -1061,25 +895,25 @@ zfs_get_data(void *arg, lr_write_t *lr, char *buf, struct lwb *lwb, zio_t *zio)
                        size = zp->z_blksz;
                        blkoff = ISP2(size) ? P2PHASE(offset, size) : offset;
                        offset -= blkoff;
-                       zgd->zgd_rl = zfs_range_lock(&zp->z_range_lock, offset,
-                           size, RL_READER);
+                       zgd->zgd_lr = zfs_rangelock_enter(&zp->z_rangelock,
+                           offset, size, RL_READER);
                        if (zp->z_blksz == size)
                                break;
                        offset += blkoff;
-                       zfs_range_unlock(zgd->zgd_rl);
+                       zfs_rangelock_exit(zgd->zgd_lr);
                }
                /* test for truncation needs to be done while range locked */
                if (lr->lr_offset >= zp->z_size)
                        error = SET_ERROR(ENOENT);
-#ifdef DEBUG
+#ifdef ZFS_DEBUG
                if (zil_fault_io) {
                        error = SET_ERROR(EIO);
                        zil_fault_io = 0;
                }
 #endif
                if (error == 0)
-                       error = dmu_buf_hold(os, object, offset, zgd, &db,
-                           DMU_READ_NO_PREFETCH);
+                       error = dmu_buf_hold_noread(os, object, offset, zgd,
+                           &db);
 
                if (error == 0) {
                        blkptr_t *bp = &lr->lr_blkptr;
@@ -1105,6 +939,14 @@ zfs_get_data(void *arg, lr_write_t *lr, char *buf, struct lwb *lwb, zio_t *zio)
 
                        if (error == EALREADY) {
                                lr->lr_common.lrc_txtype = TX_WRITE2;
+                               /*
+                                * TX_WRITE2 relies on the data previously
+                                * written by the TX_WRITE that caused
+                                * EALREADY.  We zero out the BP because
+                                * it is the old, currently-on-disk BP.
+                                */
+                               zgd->zgd_bp = NULL;
+                               BP_ZERO(bp);
                                error = 0;
                        }
                }
@@ -1115,3878 +957,559 @@ zfs_get_data(void *arg, lr_write_t *lr, char *buf, struct lwb *lwb, zio_t *zio)
        return (error);
 }
 
-/*ARGSUSED*/
-int
-zfs_access(struct inode *ip, int mode, int flag, cred_t *cr)
-{
-       znode_t *zp = ITOZ(ip);
-       zfsvfs_t *zfsvfs = ITOZSB(ip);
-       int error;
 
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(zp);
+static void
+zfs_get_done(zgd_t *zgd, int error)
+{
+       (void) error;
+       znode_t *zp = zgd->zgd_private;
 
-       if (flag & V_ACE_MASK)
-               error = zfs_zaccess(zp, mode, flag, B_FALSE, cr);
-       else
-               error = zfs_zaccess_rwx(zp, mode, flag, cr);
+       if (zgd->zgd_db)
+               dmu_buf_rele(zgd->zgd_db, zgd);
 
-       ZFS_EXIT(zfsvfs);
-       return (error);
-}
-
-/*
- * Lookup an entry in a directory, or an extended attribute directory.
- * If it exists, return a held inode reference for it.
- *
- *     IN:     dip     - inode of directory to search.
- *             nm      - name of entry to lookup.
- *             flags   - LOOKUP_XATTR set if looking for an attribute.
- *             cr      - credentials of caller.
- *             direntflags - directory lookup flags
- *             realpnp - returned pathname.
- *
- *     OUT:    ipp     - inode of located entry, NULL if not found.
- *
- *     RETURN: 0 on success, error code on failure.
- *
- * Timestamps:
- *     NA
- */
-/* ARGSUSED */
-int
-zfs_lookup(struct inode *dip, char *nm, struct inode **ipp, int flags,
-    cred_t *cr, int *direntflags, pathname_t *realpnp)
-{
-       znode_t *zdp = ITOZ(dip);
-       zfsvfs_t *zfsvfs = ITOZSB(dip);
-       int error = 0;
+       zfs_rangelock_exit(zgd->zgd_lr);
 
        /*
-        * Fast path lookup, however we must skip DNLC lookup
-        * for case folding or normalizing lookups because the
-        * DNLC code only stores the passed in name.  This means
-        * creating 'a' and removing 'A' on a case insensitive
-        * file system would work, but DNLC still thinks 'a'
-        * exists and won't let you create it again on the next
-        * pass through fast path.
+        * Release the vnode asynchronously as we currently have the
+        * txg stopped from syncing.
         */
-       if (!(flags & (LOOKUP_XATTR | FIGNORECASE))) {
+       zfs_zrele_async(zp);
 
-               if (!S_ISDIR(dip->i_mode)) {
-                       return (SET_ERROR(ENOTDIR));
-               } else if (zdp->z_sa_hdl == NULL) {
-                       return (SET_ERROR(EIO));
-               }
+       kmem_free(zgd, sizeof (zgd_t));
+}
 
-               if (nm[0] == 0 || (nm[0] == '.' && nm[1] == '\0')) {
-                       error = zfs_fastaccesschk_execute(zdp, cr);
-                       if (!error) {
-                               *ipp = dip;
-                               igrab(*ipp);
-                               return (0);
-                       }
-                       return (error);
-#ifdef HAVE_DNLC
-               } else if (!zdp->z_zfsvfs->z_norm &&
-                   (zdp->z_zfsvfs->z_case == ZFS_CASE_SENSITIVE)) {
+static int
+zfs_enter_two(zfsvfs_t *zfsvfs1, zfsvfs_t *zfsvfs2, const char *tag)
+{
+       int error;
 
-                       vnode_t *tvp = dnlc_lookup(dvp, nm);
+       /* Swap. Not sure if the order of zfs_enter()s is important. */
+       if (zfsvfs1 > zfsvfs2) {
+               zfsvfs_t *tmpzfsvfs;
 
-                       if (tvp) {
-                               error = zfs_fastaccesschk_execute(zdp, cr);
-                               if (error) {
-                                       iput(tvp);
-                                       return (error);
-                               }
-                               if (tvp == DNLC_NO_VNODE) {
-                                       iput(tvp);
-                                       return (SET_ERROR(ENOENT));
-                               } else {
-                                       *vpp = tvp;
-                                       return (specvp_check(vpp, cr));
-                               }
-                       }
-#endif /* HAVE_DNLC */
+               tmpzfsvfs = zfsvfs2;
+               zfsvfs2 = zfsvfs1;
+               zfsvfs1 = tmpzfsvfs;
+       }
+
+       error = zfs_enter(zfsvfs1, tag);
+       if (error != 0)
+               return (error);
+       if (zfsvfs1 != zfsvfs2) {
+               error = zfs_enter(zfsvfs2, tag);
+               if (error != 0) {
+                       zfs_exit(zfsvfs1, tag);
+                       return (error);
                }
        }
 
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(zdp);
+       return (0);
+}
 
-       *ipp = NULL;
+static void
+zfs_exit_two(zfsvfs_t *zfsvfs1, zfsvfs_t *zfsvfs2, const char *tag)
+{
 
-       if (flags & LOOKUP_XATTR) {
-               /*
-                * We don't allow recursive attributes..
-                * Maybe someday we will.
-                */
-               if (zdp->z_pflags & ZFS_XATTR) {
-                       ZFS_EXIT(zfsvfs);
-                       return (SET_ERROR(EINVAL));
-               }
+       zfs_exit(zfsvfs1, tag);
+       if (zfsvfs1 != zfsvfs2)
+               zfs_exit(zfsvfs2, tag);
+}
 
-               if ((error = zfs_get_xattrdir(zdp, ipp, cr, flags))) {
-                       ZFS_EXIT(zfsvfs);
-                       return (error);
-               }
+/*
+ * We split each clone request in chunks that can fit into a single ZIL
+ * log entry. Each ZIL log entry can fit 130816 bytes for a block cloning
+ * operation (see zil_max_log_data() and zfs_log_clone_range()). This gives
+ * us room for storing 1022 block pointers.
+ *
+ * On success, the function return the number of bytes copied in *lenp.
+ * Note, it doesn't return how much bytes are left to be copied.
+ * On errors which are caused by any file system limitations or
+ * brt limitations `EINVAL` is returned. In the most cases a user
+ * requested bad parameters, it could be possible to clone the file but
+ * some parameters don't match the requirements.
+ */
+int
+zfs_clone_range(znode_t *inzp, uint64_t *inoffp, znode_t *outzp,
+    uint64_t *outoffp, uint64_t *lenp, cred_t *cr)
+{
+       zfsvfs_t        *inzfsvfs, *outzfsvfs;
+       objset_t        *inos, *outos;
+       zfs_locked_range_t *inlr, *outlr;
+       dmu_buf_impl_t  *db;
+       dmu_tx_t        *tx;
+       zilog_t         *zilog;
+       uint64_t        inoff, outoff, len, done;
+       uint64_t        outsize, size;
+       int             error;
+       int             count = 0;
+       sa_bulk_attr_t  bulk[3];
+       uint64_t        mtime[2], ctime[2];
+       uint64_t        uid, gid, projid;
+       blkptr_t        *bps;
+       size_t          maxblocks, nbps;
+       uint_t          inblksz;
+       uint64_t        clear_setid_bits_txg = 0;
 
-               /*
-                * Do we have permission to get into attribute directory?
-                */
+       inoff = *inoffp;
+       outoff = *outoffp;
+       len = *lenp;
+       done = 0;
 
-               if ((error = zfs_zaccess(ITOZ(*ipp), ACE_EXECUTE, 0,
-                   B_FALSE, cr))) {
-                       iput(*ipp);
-                       *ipp = NULL;
-               }
+       inzfsvfs = ZTOZSB(inzp);
+       outzfsvfs = ZTOZSB(outzp);
 
-               ZFS_EXIT(zfsvfs);
+       /*
+        * We need to call zfs_enter() potentially on two different datasets,
+        * so we need a dedicated function for that.
+        */
+       error = zfs_enter_two(inzfsvfs, outzfsvfs, FTAG);
+       if (error != 0)
                return (error);
-       }
 
-       if (!S_ISDIR(dip->i_mode)) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(ENOTDIR));
+       inos = inzfsvfs->z_os;
+       outos = outzfsvfs->z_os;
+
+       /*
+        * Both source and destination have to belong to the same storage pool.
+        */
+       if (dmu_objset_spa(inos) != dmu_objset_spa(outos)) {
+               zfs_exit_two(inzfsvfs, outzfsvfs, FTAG);
+               return (SET_ERROR(EXDEV));
        }
 
        /*
-        * Check accessibility of directory.
+        * outos and inos belongs to the same storage pool.
+        * see a few lines above, only one check.
         */
+       if (!spa_feature_is_enabled(dmu_objset_spa(outos),
+           SPA_FEATURE_BLOCK_CLONING)) {
+               zfs_exit_two(inzfsvfs, outzfsvfs, FTAG);
+               return (SET_ERROR(EOPNOTSUPP));
+       }
 
-       if ((error = zfs_zaccess(zdp, ACE_EXECUTE, 0, B_FALSE, cr))) {
-               ZFS_EXIT(zfsvfs);
-               return (error);
+       ASSERT(!outzfsvfs->z_replay);
+
+       /*
+        * Block cloning from an unencrypted dataset into an encrypted
+        * dataset and vice versa is not supported.
+        */
+       if (inos->os_encrypted != outos->os_encrypted) {
+               zfs_exit_two(inzfsvfs, outzfsvfs, FTAG);
+               return (SET_ERROR(EXDEV));
        }
 
-       if (zfsvfs->z_utf8 && u8_validate(nm, strlen(nm),
-           NULL, U8_VALIDATE_ENTIRE, &error) < 0) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EILSEQ));
+       /*
+        * Cloning across encrypted datasets is possible only if they
+        * share the same master key.
+        */
+       if (inos != outos && inos->os_encrypted &&
+           !dmu_objset_crypto_key_equal(inos, outos)) {
+               zfs_exit_two(inzfsvfs, outzfsvfs, FTAG);
+               return (SET_ERROR(EXDEV));
        }
 
-       error = zfs_dirlook(zdp, nm, ipp, flags, direntflags, realpnp);
-       if ((error == 0) && (*ipp))
-               zfs_inode_update(ITOZ(*ipp));
+       error = zfs_verify_zp(inzp);
+       if (error == 0)
+               error = zfs_verify_zp(outzp);
+       if (error != 0) {
+               zfs_exit_two(inzfsvfs, outzfsvfs, FTAG);
+               return (error);
+       }
 
-       ZFS_EXIT(zfsvfs);
-       return (error);
-}
+       /*
+        * We don't copy source file's flags that's why we don't allow to clone
+        * files that are in quarantine.
+        */
+       if (inzp->z_pflags & ZFS_AV_QUARANTINED) {
+               zfs_exit_two(inzfsvfs, outzfsvfs, FTAG);
+               return (SET_ERROR(EACCES));
+       }
 
-/*
- * Attempt to create a new entry in a directory.  If the entry
- * already exists, truncate the file if permissible, else return
- * an error.  Return the ip of the created or trunc'd file.
- *
- *     IN:     dip     - inode of directory to put new file entry in.
- *             name    - name of new file entry.
- *             vap     - attributes of new file.
- *             excl    - flag indicating exclusive or non-exclusive mode.
- *             mode    - mode to open file with.
- *             cr      - credentials of caller.
- *             flag    - large file flag [UNUSED].
- *             vsecp   - ACL to be set
- *
- *     OUT:    ipp     - inode of created or trunc'd entry.
- *
- *     RETURN: 0 on success, error code on failure.
- *
- * Timestamps:
- *     dip - ctime|mtime updated if new entry created
- *      ip - ctime|mtime always, atime if new
- */
+       if (inoff >= inzp->z_size) {
+               *lenp = 0;
+               zfs_exit_two(inzfsvfs, outzfsvfs, FTAG);
+               return (0);
+       }
+       if (len > inzp->z_size - inoff) {
+               len = inzp->z_size - inoff;
+       }
+       if (len == 0) {
+               *lenp = 0;
+               zfs_exit_two(inzfsvfs, outzfsvfs, FTAG);
+               return (0);
+       }
 
-/* ARGSUSED */
-int
-zfs_create(struct inode *dip, char *name, vattr_t *vap, int excl,
-    int mode, struct inode **ipp, cred_t *cr, int flag, vsecattr_t *vsecp)
-{
-       znode_t         *zp, *dzp = ITOZ(dip);
-       zfsvfs_t        *zfsvfs = ITOZSB(dip);
-       zilog_t         *zilog;
-       objset_t        *os;
-       zfs_dirlock_t   *dl;
-       dmu_tx_t        *tx;
-       int             error;
-       uid_t           uid;
-       gid_t           gid;
-       zfs_acl_ids_t   acl_ids;
-       boolean_t       fuid_dirtied;
-       boolean_t       have_acl = B_FALSE;
-       boolean_t       waited = B_FALSE;
+       /*
+        * Callers might not be able to detect properly that we are read-only,
+        * so check it explicitly here.
+        */
+       if (zfs_is_readonly(outzfsvfs)) {
+               zfs_exit_two(inzfsvfs, outzfsvfs, FTAG);
+               return (SET_ERROR(EROFS));
+       }
 
        /*
-        * If we have an ephemeral id, ACL, or XVATTR then
-        * make sure file system is at proper version
+        * If immutable or not appending then return EPERM.
+        * Intentionally allow ZFS_READONLY through here.
+        * See zfs_zaccess_common()
         */
+       if ((outzp->z_pflags & ZFS_IMMUTABLE) != 0) {
+               zfs_exit_two(inzfsvfs, outzfsvfs, FTAG);
+               return (SET_ERROR(EPERM));
+       }
 
-       gid = crgetgid(cr);
-       uid = crgetuid(cr);
+       /*
+        * No overlapping if we are cloning within the same file.
+        */
+       if (inzp == outzp) {
+               if (inoff < outoff + len && outoff < inoff + len) {
+                       zfs_exit_two(inzfsvfs, outzfsvfs, FTAG);
+                       return (SET_ERROR(EINVAL));
+               }
+       }
 
-       if (zfsvfs->z_use_fuids == B_FALSE &&
-           (vsecp || IS_EPHEMERAL(uid) || IS_EPHEMERAL(gid)))
-               return (SET_ERROR(EINVAL));
+       /*
+        * Maintain predictable lock order.
+        */
+       if (inzp < outzp || (inzp == outzp && inoff < outoff)) {
+               inlr = zfs_rangelock_enter(&inzp->z_rangelock, inoff, len,
+                   RL_READER);
+               outlr = zfs_rangelock_enter(&outzp->z_rangelock, outoff, len,
+                   RL_WRITER);
+       } else {
+               outlr = zfs_rangelock_enter(&outzp->z_rangelock, outoff, len,
+                   RL_WRITER);
+               inlr = zfs_rangelock_enter(&inzp->z_rangelock, inoff, len,
+                   RL_READER);
+       }
 
-       if (name == NULL)
-               return (SET_ERROR(EINVAL));
+       inblksz = inzp->z_blksz;
 
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(dzp);
-       os = zfsvfs->z_os;
-       zilog = zfsvfs->z_log;
+       /*
+        * We cannot clone into a file with different block size if we can't
+        * grow it (block size is already bigger, has more than one block, or
+        * not locked for growth).  There are other possible reasons for the
+        * grow to fail, but we cover what we can before opening transaction
+        * and the rest detect after we try to do it.
+        */
+       if (inblksz < outzp->z_blksz) {
+               error = SET_ERROR(EINVAL);
+               goto unlock;
+       }
+       if (inblksz != outzp->z_blksz && (outzp->z_size > outzp->z_blksz ||
+           outlr->lr_length != UINT64_MAX)) {
+               error = SET_ERROR(EINVAL);
+               goto unlock;
+       }
 
-       if (zfsvfs->z_utf8 && u8_validate(name, strlen(name),
-           NULL, U8_VALIDATE_ENTIRE, &error) < 0) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EILSEQ));
+       /*
+        * Block size must be power-of-2 if destination offset != 0.
+        * There can be no multiple blocks of non-power-of-2 size.
+        */
+       if (outoff != 0 && !ISP2(inblksz)) {
+               error = SET_ERROR(EINVAL);
+               goto unlock;
        }
 
-       if (vap->va_mask & ATTR_XVATTR) {
-               if ((error = secpolicy_xvattr((xvattr_t *)vap,
-                   crgetuid(cr), cr, vap->va_mode)) != 0) {
-                       ZFS_EXIT(zfsvfs);
-                       return (error);
-               }
+       /*
+        * Offsets and len must be at block boundries.
+        */
+       if ((inoff % inblksz) != 0 || (outoff % inblksz) != 0) {
+               error = SET_ERROR(EINVAL);
+               goto unlock;
+       }
+       /*
+        * Length must be multipe of blksz, except for the end of the file.
+        */
+       if ((len % inblksz) != 0 &&
+           (len < inzp->z_size - inoff || len < outzp->z_size - outoff)) {
+               error = SET_ERROR(EINVAL);
+               goto unlock;
        }
 
-top:
-       *ipp = NULL;
-       if (*name == '\0') {
-               /*
-                * Null component name refers to the directory itself.
-                */
-               igrab(dip);
-               zp = dzp;
-               dl = NULL;
-               error = 0;
-       } else {
-               /* possible igrab(zp) */
-               int zflg = 0;
+       /*
+        * If we are copying only one block and it is smaller than recordsize
+        * property, do not allow destination to grow beyond one block if it
+        * is not there yet.  Otherwise the destination will get stuck with
+        * that block size forever, that can be as small as 512 bytes, no
+        * matter how big the destination grow later.
+        */
+       if (len <= inblksz && inblksz < outzfsvfs->z_max_blksz &&
+           outzp->z_size <= inblksz && outoff + len > inblksz) {
+               error = SET_ERROR(EINVAL);
+               goto unlock;
+       }
 
-               if (flag & FIGNORECASE)
-                       zflg |= ZCILOOK;
+       error = zn_rlimit_fsize(outoff + len);
+       if (error != 0) {
+               goto unlock;
+       }
 
-               error = zfs_dirent_lock(&dl, dzp, name, &zp, zflg,
-                   NULL, NULL);
-               if (error) {
-                       if (have_acl)
-                               zfs_acl_ids_free(&acl_ids);
-                       if (strcmp(name, "..") == 0)
-                               error = SET_ERROR(EISDIR);
-                       ZFS_EXIT(zfsvfs);
-                       return (error);
-               }
+       if (inoff >= MAXOFFSET_T || outoff >= MAXOFFSET_T) {
+               error = SET_ERROR(EFBIG);
+               goto unlock;
        }
 
-       if (zp == NULL) {
-               uint64_t txtype;
+       SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_MTIME(outzfsvfs), NULL,
+           &mtime, 16);
+       SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_CTIME(outzfsvfs), NULL,
+           &ctime, 16);
+       SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_SIZE(outzfsvfs), NULL,
+           &outzp->z_size, 8);
 
-               /*
-                * Create a new file object and update the directory
-                * to reference it.
-                */
-               if ((error = zfs_zaccess(dzp, ACE_ADD_FILE, 0, B_FALSE, cr))) {
-                       if (have_acl)
-                               zfs_acl_ids_free(&acl_ids);
-                       goto out;
-               }
+       zilog = outzfsvfs->z_log;
+       maxblocks = zil_max_log_data(zilog, sizeof (lr_clone_range_t)) /
+           sizeof (bps[0]);
 
-               /*
-                * We only support the creation of regular files in
-                * extended attribute directories.
-                */
+       uid = KUID_TO_SUID(ZTOUID(outzp));
+       gid = KGID_TO_SGID(ZTOGID(outzp));
+       projid = outzp->z_projid;
 
-               if ((dzp->z_pflags & ZFS_XATTR) && !S_ISREG(vap->va_mode)) {
-                       if (have_acl)
-                               zfs_acl_ids_free(&acl_ids);
-                       error = SET_ERROR(EINVAL);
-                       goto out;
-               }
+       bps = vmem_alloc(sizeof (bps[0]) * maxblocks, KM_SLEEP);
 
-               if (!have_acl && (error = zfs_acl_ids_create(dzp, 0, vap,
-                   cr, vsecp, &acl_ids)) != 0)
-                       goto out;
-               have_acl = B_TRUE;
+       /*
+        * Clone the file in reasonable size chunks.  Each chunk is cloned
+        * in a separate transaction; this keeps the intent log records small
+        * and allows us to do more fine-grained space accounting.
+        */
+       while (len > 0) {
+               size = MIN(inblksz * maxblocks, len);
 
-               if (zfs_acl_ids_overquota(zfsvfs, &acl_ids)) {
-                       zfs_acl_ids_free(&acl_ids);
+               if (zfs_id_overblockquota(outzfsvfs, DMU_USERUSED_OBJECT,
+                   uid) ||
+                   zfs_id_overblockquota(outzfsvfs, DMU_GROUPUSED_OBJECT,
+                   gid) ||
+                   (projid != ZFS_DEFAULT_PROJID &&
+                   zfs_id_overblockquota(outzfsvfs, DMU_PROJECTUSED_OBJECT,
+                   projid))) {
                        error = SET_ERROR(EDQUOT);
-                       goto out;
+                       break;
                }
 
-               tx = dmu_tx_create(os);
-
-               dmu_tx_hold_sa_create(tx, acl_ids.z_aclp->z_acl_bytes +
-                   ZFS_SA_BASE_ATTR_SIZE);
+               nbps = maxblocks;
+               error = dmu_read_l0_bps(inos, inzp->z_id, inoff, size, bps,
+                   &nbps);
+               if (error != 0) {
+                       /*
+                        * If we are trying to clone a block that was created
+                        * in the current transaction group, error will be
+                        * EAGAIN here, which we can just return to the caller
+                        * so it can fallback if it likes.
+                        */
+                       break;
+               }
 
-               fuid_dirtied = zfsvfs->z_fuid_dirty;
-               if (fuid_dirtied)
-                       zfs_fuid_txhold(zfsvfs, tx);
-               dmu_tx_hold_zap(tx, dzp->z_id, TRUE, name);
-               dmu_tx_hold_sa(tx, dzp->z_sa_hdl, B_FALSE);
-               if (!zfsvfs->z_use_sa &&
-                   acl_ids.z_aclp->z_acl_bytes > ZFS_ACE_SPACE) {
-                       dmu_tx_hold_write(tx, DMU_NEW_OBJECT,
-                           0, acl_ids.z_aclp->z_acl_bytes);
+               /*
+                * Start a transaction.
+                */
+               tx = dmu_tx_create(outos);
+               dmu_tx_hold_sa(tx, outzp->z_sa_hdl, B_FALSE);
+               db = (dmu_buf_impl_t *)sa_get_db(outzp->z_sa_hdl);
+               DB_DNODE_ENTER(db);
+               dmu_tx_hold_clone_by_dnode(tx, DB_DNODE(db), outoff, size);
+               DB_DNODE_EXIT(db);
+               zfs_sa_upgrade_txholds(tx, outzp);
+               error = dmu_tx_assign(tx, TXG_WAIT);
+               if (error != 0) {
+                       dmu_tx_abort(tx);
+                       break;
                }
 
-               error = dmu_tx_assign(tx,
-                   (waited ? TXG_NOTHROTTLE : 0) | TXG_NOWAIT);
-               if (error) {
-                       zfs_dirent_unlock(dl);
-                       if (error == ERESTART) {
-                               waited = B_TRUE;
-                               dmu_tx_wait(tx);
+               /*
+                * Copy source znode's block size. This is done only if the
+                * whole znode is locked (see zfs_rangelock_cb()) and only
+                * on the first iteration since zfs_rangelock_reduce() will
+                * shrink down lr_length to the appropriate size.
+                */
+               if (outlr->lr_length == UINT64_MAX) {
+                       zfs_grow_blocksize(outzp, inblksz, tx);
+
+                       /*
+                        * Block growth may fail for many reasons we can not
+                        * predict here.  If it happen the cloning is doomed.
+                        */
+                       if (inblksz != outzp->z_blksz) {
+                               error = SET_ERROR(EINVAL);
                                dmu_tx_abort(tx);
-                               goto top;
+                               break;
                        }
-                       zfs_acl_ids_free(&acl_ids);
-                       dmu_tx_abort(tx);
-                       ZFS_EXIT(zfsvfs);
-                       return (error);
-               }
-               zfs_mknode(dzp, vap, tx, cr, 0, &zp, &acl_ids);
 
-               error = zfs_link_create(dl, zp, tx, ZNEW);
-               if (error != 0) {
                        /*
-                        * Since, we failed to add the directory entry for it,
-                        * delete the newly created dnode.
+                        * Round range lock up to the block boundary, so we
+                        * prevent appends until we are done.
                         */
-                       zfs_znode_delete(zp, tx);
-                       remove_inode_hash(ZTOI(zp));
-                       zfs_acl_ids_free(&acl_ids);
-                       dmu_tx_commit(tx);
-                       goto out;
+                       zfs_rangelock_reduce(outlr, outoff,
+                           ((len - 1) / inblksz + 1) * inblksz);
                }
 
-               if (fuid_dirtied)
-                       zfs_fuid_sync(zfsvfs, tx);
+               error = dmu_brt_clone(outos, outzp->z_id, outoff, size, tx,
+                   bps, nbps);
+               if (error != 0) {
+                       dmu_tx_commit(tx);
+                       break;
+               }
 
-               txtype = zfs_log_create_txtype(Z_FILE, vsecp, vap);
-               if (flag & FIGNORECASE)
-                       txtype |= TX_CI;
-               zfs_log_create(zilog, tx, txtype, dzp, zp, name,
-                   vsecp, acl_ids.z_fuidp, vap);
-               zfs_acl_ids_free(&acl_ids);
-               dmu_tx_commit(tx);
-       } else {
-               int aflags = (flag & FAPPEND) ? V_APPEND : 0;
+               zfs_clear_setid_bits_if_necessary(outzfsvfs, outzp, cr,
+                   &clear_setid_bits_txg, tx);
 
-               if (have_acl)
-                       zfs_acl_ids_free(&acl_ids);
-               have_acl = B_FALSE;
+               zfs_tstamp_update_setup(outzp, CONTENT_MODIFIED, mtime, ctime);
 
                /*
-                * A directory entry already exists for this name.
-                */
-               /*
-                * Can't truncate an existing file if in exclusive mode.
-                */
-               if (excl) {
-                       error = SET_ERROR(EEXIST);
-                       goto out;
-               }
-               /*
-                * Can't open a directory for writing.
-                */
-               if (S_ISDIR(ZTOI(zp)->i_mode)) {
-                       error = SET_ERROR(EISDIR);
-                       goto out;
-               }
-               /*
-                * Verify requested access to file.
+                * Update the file size (zp_size) if it has changed;
+                * account for possible concurrent updates.
                 */
-               if (mode && (error = zfs_zaccess_rwx(zp, mode, aflags, cr))) {
-                       goto out;
+               while ((outsize = outzp->z_size) < outoff + size) {
+                       (void) atomic_cas_64(&outzp->z_size, outsize,
+                           outoff + size);
                }
 
-               mutex_enter(&dzp->z_lock);
-               dzp->z_seq++;
-               mutex_exit(&dzp->z_lock);
+               error = sa_bulk_update(outzp->z_sa_hdl, bulk, count, tx);
+
+               zfs_log_clone_range(zilog, tx, TX_CLONE_RANGE, outzp, outoff,
+                   size, inblksz, bps, nbps);
+
+               dmu_tx_commit(tx);
+
+               if (error != 0)
+                       break;
+
+               inoff += size;
+               outoff += size;
+               len -= size;
+               done += size;
+       }
 
+       vmem_free(bps, sizeof (bps[0]) * maxblocks);
+       zfs_znode_update_vfs(outzp);
+
+unlock:
+       zfs_rangelock_exit(outlr);
+       zfs_rangelock_exit(inlr);
+
+       if (done > 0) {
                /*
-                * Truncate regular files if requested.
+                * If we have made at least partial progress, reset the error.
                 */
-               if (S_ISREG(ZTOI(zp)->i_mode) &&
-                   (vap->va_mask & ATTR_SIZE) && (vap->va_size == 0)) {
-                       /* we can't hold any locks when calling zfs_freesp() */
-                       if (dl) {
-                               zfs_dirent_unlock(dl);
-                               dl = NULL;
-                       }
-                       error = zfs_freesp(zp, 0, 0, mode, TRUE);
-               }
-       }
-out:
+               error = 0;
+
+               ZFS_ACCESSTIME_STAMP(inzfsvfs, inzp);
 
-       if (dl)
-               zfs_dirent_unlock(dl);
+               if (outos->os_sync == ZFS_SYNC_ALWAYS) {
+                       zil_commit(zilog, outzp->z_id);
+               }
 
-       if (error) {
-               if (zp)
-                       iput(ZTOI(zp));
+               *inoffp += done;
+               *outoffp += done;
+               *lenp = done;
        } else {
-               zfs_inode_update(dzp);
-               zfs_inode_update(zp);
-               *ipp = ZTOI(zp);
+               /*
+                * If we made no progress, there must be a good reason.
+                * EOF is handled explicitly above, before the loop.
+                */
+               ASSERT3S(error, !=, 0);
        }
 
-       if (zfsvfs->z_os->os_sync == ZFS_SYNC_ALWAYS)
-               zil_commit(zilog, 0);
+       zfs_exit_two(inzfsvfs, outzfsvfs, FTAG);
 
-       ZFS_EXIT(zfsvfs);
        return (error);
 }
 
-/* ARGSUSED */
+/*
+ * Usual pattern would be to call zfs_clone_range() from zfs_replay_clone(),
+ * but we cannot do that, because when replaying we don't have source znode
+ * available. This is why we need a dedicated replay function.
+ */
 int
-zfs_tmpfile(struct inode *dip, vattr_t *vap, int excl,
-    int mode, struct inode **ipp, cred_t *cr, int flag, vsecattr_t *vsecp)
+zfs_clone_range_replay(znode_t *zp, uint64_t off, uint64_t len, uint64_t blksz,
+    const blkptr_t *bps, size_t nbps)
 {
-       znode_t         *zp = NULL, *dzp = ITOZ(dip);
-       zfsvfs_t        *zfsvfs = ITOZSB(dip);
-       objset_t        *os;
+       zfsvfs_t        *zfsvfs;
+       dmu_buf_impl_t  *db;
        dmu_tx_t        *tx;
        int             error;
-       uid_t           uid;
-       gid_t           gid;
-       zfs_acl_ids_t   acl_ids;
-       boolean_t       fuid_dirtied;
-       boolean_t       have_acl = B_FALSE;
-       boolean_t       waited = B_FALSE;
+       int             count = 0;
+       sa_bulk_attr_t  bulk[3];
+       uint64_t        mtime[2], ctime[2];
 
-       /*
-        * If we have an ephemeral id, ACL, or XVATTR then
-        * make sure file system is at proper version
-        */
+       ASSERT3U(off, <, MAXOFFSET_T);
+       ASSERT3U(len, >, 0);
+       ASSERT3U(nbps, >, 0);
 
-       gid = crgetgid(cr);
-       uid = crgetuid(cr);
+       zfsvfs = ZTOZSB(zp);
 
-       if (zfsvfs->z_use_fuids == B_FALSE &&
-           (vsecp || IS_EPHEMERAL(uid) || IS_EPHEMERAL(gid)))
-               return (SET_ERROR(EINVAL));
+       ASSERT(spa_feature_is_enabled(dmu_objset_spa(zfsvfs->z_os),
+           SPA_FEATURE_BLOCK_CLONING));
 
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(dzp);
-       os = zfsvfs->z_os;
+       if ((error = zfs_enter_verify_zp(zfsvfs, zp, FTAG)) != 0)
+               return (error);
 
-       if (vap->va_mask & ATTR_XVATTR) {
-               if ((error = secpolicy_xvattr((xvattr_t *)vap,
-                   crgetuid(cr), cr, vap->va_mode)) != 0) {
-                       ZFS_EXIT(zfsvfs);
-                       return (error);
-               }
+       ASSERT(zfsvfs->z_replay);
+       ASSERT(!zfs_is_readonly(zfsvfs));
+
+       if ((off % blksz) != 0) {
+               zfs_exit(zfsvfs, FTAG);
+               return (SET_ERROR(EINVAL));
        }
 
-top:
-       *ipp = NULL;
+       SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_MTIME(zfsvfs), NULL, &mtime, 16);
+       SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_CTIME(zfsvfs), NULL, &ctime, 16);
+       SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_SIZE(zfsvfs), NULL,
+           &zp->z_size, 8);
 
        /*
-        * Create a new file object and update the directory
-        * to reference it.
+        * Start a transaction.
         */
-       if ((error = zfs_zaccess(dzp, ACE_ADD_FILE, 0, B_FALSE, cr))) {
-               if (have_acl)
-                       zfs_acl_ids_free(&acl_ids);
-               goto out;
+       tx = dmu_tx_create(zfsvfs->z_os);
+
+       dmu_tx_hold_sa(tx, zp->z_sa_hdl, B_FALSE);
+       db = (dmu_buf_impl_t *)sa_get_db(zp->z_sa_hdl);
+       DB_DNODE_ENTER(db);
+       dmu_tx_hold_clone_by_dnode(tx, DB_DNODE(db), off, len);
+       DB_DNODE_EXIT(db);
+       zfs_sa_upgrade_txholds(tx, zp);
+       error = dmu_tx_assign(tx, TXG_WAIT);
+       if (error != 0) {
+               dmu_tx_abort(tx);
+               zfs_exit(zfsvfs, FTAG);
+               return (error);
        }
 
-       if (!have_acl && (error = zfs_acl_ids_create(dzp, 0, vap,
-           cr, vsecp, &acl_ids)) != 0)
-               goto out;
-       have_acl = B_TRUE;
+       if (zp->z_blksz < blksz)
+               zfs_grow_blocksize(zp, blksz, tx);
 
-       if (zfs_acl_ids_overquota(zfsvfs, &acl_ids)) {
-               zfs_acl_ids_free(&acl_ids);
-               error = SET_ERROR(EDQUOT);
-               goto out;
-       }
+       dmu_brt_clone(zfsvfs->z_os, zp->z_id, off, len, tx, bps, nbps);
 
-       tx = dmu_tx_create(os);
+       zfs_tstamp_update_setup(zp, CONTENT_MODIFIED, mtime, ctime);
 
-       dmu_tx_hold_sa_create(tx, acl_ids.z_aclp->z_acl_bytes +
-           ZFS_SA_BASE_ATTR_SIZE);
-       dmu_tx_hold_zap(tx, zfsvfs->z_unlinkedobj, FALSE, NULL);
+       if (zp->z_size < off + len)
+               zp->z_size = off + len;
 
-       fuid_dirtied = zfsvfs->z_fuid_dirty;
-       if (fuid_dirtied)
-               zfs_fuid_txhold(zfsvfs, tx);
-       if (!zfsvfs->z_use_sa &&
-           acl_ids.z_aclp->z_acl_bytes > ZFS_ACE_SPACE) {
-               dmu_tx_hold_write(tx, DMU_NEW_OBJECT,
-                   0, acl_ids.z_aclp->z_acl_bytes);
-       }
-       error = dmu_tx_assign(tx, (waited ? TXG_NOTHROTTLE : 0) | TXG_NOWAIT);
-       if (error) {
-               if (error == ERESTART) {
-                       waited = B_TRUE;
-                       dmu_tx_wait(tx);
-                       dmu_tx_abort(tx);
-                       goto top;
-               }
-               zfs_acl_ids_free(&acl_ids);
-               dmu_tx_abort(tx);
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-       zfs_mknode(dzp, vap, tx, cr, IS_TMPFILE, &zp, &acl_ids);
-
-       if (fuid_dirtied)
-               zfs_fuid_sync(zfsvfs, tx);
-
-       /* Add to unlinked set */
-       zp->z_unlinked = 1;
-       zfs_unlinked_add(zp, tx);
-       zfs_acl_ids_free(&acl_ids);
-       dmu_tx_commit(tx);
-out:
-
-       if (error) {
-               if (zp)
-                       iput(ZTOI(zp));
-       } else {
-               zfs_inode_update(dzp);
-               zfs_inode_update(zp);
-               *ipp = ZTOI(zp);
-       }
-
-       ZFS_EXIT(zfsvfs);
-       return (error);
-}
-
-/*
- * Remove an entry from a directory.
- *
- *     IN:     dip     - inode of directory to remove entry from.
- *             name    - name of entry to remove.
- *             cr      - credentials of caller.
- *
- *     RETURN: 0 if success
- *             error code if failure
- *
- * Timestamps:
- *     dip - ctime|mtime
- *      ip - ctime (if nlink > 0)
- */
-
-uint64_t null_xattr = 0;
-
-/*ARGSUSED*/
-int
-zfs_remove(struct inode *dip, char *name, cred_t *cr, int flags)
-{
-       znode_t         *zp, *dzp = ITOZ(dip);
-       znode_t         *xzp;
-       struct inode    *ip;
-       zfsvfs_t        *zfsvfs = ITOZSB(dip);
-       zilog_t         *zilog;
-       uint64_t        acl_obj, xattr_obj;
-       uint64_t        xattr_obj_unlinked = 0;
-       uint64_t        obj = 0;
-       uint64_t        links;
-       zfs_dirlock_t   *dl;
-       dmu_tx_t        *tx;
-       boolean_t       may_delete_now, delete_now = FALSE;
-       boolean_t       unlinked, toobig = FALSE;
-       uint64_t        txtype;
-       pathname_t      *realnmp = NULL;
-       pathname_t      realnm;
-       int             error;
-       int             zflg = ZEXISTS;
-       boolean_t       waited = B_FALSE;
-
-       if (name == NULL)
-               return (SET_ERROR(EINVAL));
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(dzp);
-       zilog = zfsvfs->z_log;
-
-       if (flags & FIGNORECASE) {
-               zflg |= ZCILOOK;
-               pn_alloc(&realnm);
-               realnmp = &realnm;
-       }
-
-top:
-       xattr_obj = 0;
-       xzp = NULL;
-       /*
-        * Attempt to lock directory; fail if entry doesn't exist.
-        */
-       if ((error = zfs_dirent_lock(&dl, dzp, name, &zp, zflg,
-           NULL, realnmp))) {
-               if (realnmp)
-                       pn_free(realnmp);
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-
-       ip = ZTOI(zp);
-
-       if ((error = zfs_zaccess_delete(dzp, zp, cr))) {
-               goto out;
-       }
-
-       /*
-        * Need to use rmdir for removing directories.
-        */
-       if (S_ISDIR(ip->i_mode)) {
-               error = SET_ERROR(EPERM);
-               goto out;
-       }
-
-#ifdef HAVE_DNLC
-       if (realnmp)
-               dnlc_remove(dvp, realnmp->pn_buf);
-       else
-               dnlc_remove(dvp, name);
-#endif /* HAVE_DNLC */
-
-       mutex_enter(&zp->z_lock);
-       may_delete_now = atomic_read(&ip->i_count) == 1 && !(zp->z_is_mapped);
-       mutex_exit(&zp->z_lock);
-
-       /*
-        * We may delete the znode now, or we may put it in the unlinked set;
-        * it depends on whether we're the last link, and on whether there are
-        * other holds on the inode.  So we dmu_tx_hold() the right things to
-        * allow for either case.
-        */
-       obj = zp->z_id;
-       tx = dmu_tx_create(zfsvfs->z_os);
-       dmu_tx_hold_zap(tx, dzp->z_id, FALSE, name);
-       dmu_tx_hold_sa(tx, zp->z_sa_hdl, B_FALSE);
-       zfs_sa_upgrade_txholds(tx, zp);
-       zfs_sa_upgrade_txholds(tx, dzp);
-       if (may_delete_now) {
-               toobig = zp->z_size > zp->z_blksz * zfs_delete_blocks;
-               /* if the file is too big, only hold_free a token amount */
-               dmu_tx_hold_free(tx, zp->z_id, 0,
-                   (toobig ? DMU_MAX_ACCESS : DMU_OBJECT_END));
-       }
-
-       /* are there any extended attributes? */
-       error = sa_lookup(zp->z_sa_hdl, SA_ZPL_XATTR(zfsvfs),
-           &xattr_obj, sizeof (xattr_obj));
-       if (error == 0 && xattr_obj) {
-               error = zfs_zget(zfsvfs, xattr_obj, &xzp);
-               ASSERT0(error);
-               dmu_tx_hold_sa(tx, zp->z_sa_hdl, B_TRUE);
-               dmu_tx_hold_sa(tx, xzp->z_sa_hdl, B_FALSE);
-       }
-
-       mutex_enter(&zp->z_lock);
-       if ((acl_obj = zfs_external_acl(zp)) != 0 && may_delete_now)
-               dmu_tx_hold_free(tx, acl_obj, 0, DMU_OBJECT_END);
-       mutex_exit(&zp->z_lock);
-
-       /* charge as an update -- would be nice not to charge at all */
-       dmu_tx_hold_zap(tx, zfsvfs->z_unlinkedobj, FALSE, NULL);
-
-       /*
-        * Mark this transaction as typically resulting in a net free of space
-        */
-       dmu_tx_mark_netfree(tx);
-
-       error = dmu_tx_assign(tx, (waited ? TXG_NOTHROTTLE : 0) | TXG_NOWAIT);
-       if (error) {
-               zfs_dirent_unlock(dl);
-               if (error == ERESTART) {
-                       waited = B_TRUE;
-                       dmu_tx_wait(tx);
-                       dmu_tx_abort(tx);
-                       iput(ip);
-                       if (xzp)
-                               iput(ZTOI(xzp));
-                       goto top;
-               }
-               if (realnmp)
-                       pn_free(realnmp);
-               dmu_tx_abort(tx);
-               iput(ip);
-               if (xzp)
-                       iput(ZTOI(xzp));
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-
-       /*
-        * Remove the directory entry.
-        */
-       error = zfs_link_destroy(dl, zp, tx, zflg, &unlinked);
-
-       if (error) {
-               dmu_tx_commit(tx);
-               goto out;
-       }
-
-       if (unlinked) {
-               /*
-                * Hold z_lock so that we can make sure that the ACL obj
-                * hasn't changed.  Could have been deleted due to
-                * zfs_sa_upgrade().
-                */
-               mutex_enter(&zp->z_lock);
-               (void) sa_lookup(zp->z_sa_hdl, SA_ZPL_XATTR(zfsvfs),
-                   &xattr_obj_unlinked, sizeof (xattr_obj_unlinked));
-               delete_now = may_delete_now && !toobig &&
-                   atomic_read(&ip->i_count) == 1 && !(zp->z_is_mapped) &&
-                   xattr_obj == xattr_obj_unlinked && zfs_external_acl(zp) ==
-                   acl_obj;
-       }
-
-       if (delete_now) {
-               if (xattr_obj_unlinked) {
-                       ASSERT3U(ZTOI(xzp)->i_nlink, ==, 2);
-                       mutex_enter(&xzp->z_lock);
-                       xzp->z_unlinked = 1;
-                       clear_nlink(ZTOI(xzp));
-                       links = 0;
-                       error = sa_update(xzp->z_sa_hdl, SA_ZPL_LINKS(zfsvfs),
-                           &links, sizeof (links), tx);
-                       ASSERT3U(error,  ==,  0);
-                       mutex_exit(&xzp->z_lock);
-                       zfs_unlinked_add(xzp, tx);
-
-                       if (zp->z_is_sa)
-                               error = sa_remove(zp->z_sa_hdl,
-                                   SA_ZPL_XATTR(zfsvfs), tx);
-                       else
-                               error = sa_update(zp->z_sa_hdl,
-                                   SA_ZPL_XATTR(zfsvfs), &null_xattr,
-                                   sizeof (uint64_t), tx);
-                       ASSERT0(error);
-               }
-               /*
-                * Add to the unlinked set because a new reference could be
-                * taken concurrently resulting in a deferred destruction.
-                */
-               zfs_unlinked_add(zp, tx);
-               mutex_exit(&zp->z_lock);
-       } else if (unlinked) {
-               mutex_exit(&zp->z_lock);
-               zfs_unlinked_add(zp, tx);
-       }
-
-       txtype = TX_REMOVE;
-       if (flags & FIGNORECASE)
-               txtype |= TX_CI;
-       zfs_log_remove(zilog, tx, txtype, dzp, name, obj);
-
-       dmu_tx_commit(tx);
-out:
-       if (realnmp)
-               pn_free(realnmp);
-
-       zfs_dirent_unlock(dl);
-       zfs_inode_update(dzp);
-       zfs_inode_update(zp);
-
-       if (delete_now)
-               iput(ip);
-       else
-               zfs_iput_async(ip);
-
-       if (xzp) {
-               zfs_inode_update(xzp);
-               zfs_iput_async(ZTOI(xzp));
-       }
-
-       if (zfsvfs->z_os->os_sync == ZFS_SYNC_ALWAYS)
-               zil_commit(zilog, 0);
-
-       ZFS_EXIT(zfsvfs);
-       return (error);
-}
-
-/*
- * Create a new directory and insert it into dip using the name
- * provided.  Return a pointer to the inserted directory.
- *
- *     IN:     dip     - inode of directory to add subdir to.
- *             dirname - name of new directory.
- *             vap     - attributes of new directory.
- *             cr      - credentials of caller.
- *             vsecp   - ACL to be set
- *
- *     OUT:    ipp     - inode of created directory.
- *
- *     RETURN: 0 if success
- *             error code if failure
- *
- * Timestamps:
- *     dip - ctime|mtime updated
- *     ipp - ctime|mtime|atime updated
- */
-/*ARGSUSED*/
-int
-zfs_mkdir(struct inode *dip, char *dirname, vattr_t *vap, struct inode **ipp,
-    cred_t *cr, int flags, vsecattr_t *vsecp)
-{
-       znode_t         *zp, *dzp = ITOZ(dip);
-       zfsvfs_t        *zfsvfs = ITOZSB(dip);
-       zilog_t         *zilog;
-       zfs_dirlock_t   *dl;
-       uint64_t        txtype;
-       dmu_tx_t        *tx;
-       int             error;
-       int             zf = ZNEW;
-       uid_t           uid;
-       gid_t           gid = crgetgid(cr);
-       zfs_acl_ids_t   acl_ids;
-       boolean_t       fuid_dirtied;
-       boolean_t       waited = B_FALSE;
-
-       ASSERT(S_ISDIR(vap->va_mode));
-
-       /*
-        * If we have an ephemeral id, ACL, or XVATTR then
-        * make sure file system is at proper version
-        */
-
-       uid = crgetuid(cr);
-       if (zfsvfs->z_use_fuids == B_FALSE &&
-           (vsecp || IS_EPHEMERAL(uid) || IS_EPHEMERAL(gid)))
-               return (SET_ERROR(EINVAL));
-
-       if (dirname == NULL)
-               return (SET_ERROR(EINVAL));
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(dzp);
-       zilog = zfsvfs->z_log;
-
-       if (dzp->z_pflags & ZFS_XATTR) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EINVAL));
-       }
-
-       if (zfsvfs->z_utf8 && u8_validate(dirname,
-           strlen(dirname), NULL, U8_VALIDATE_ENTIRE, &error) < 0) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EILSEQ));
-       }
-       if (flags & FIGNORECASE)
-               zf |= ZCILOOK;
-
-       if (vap->va_mask & ATTR_XVATTR) {
-               if ((error = secpolicy_xvattr((xvattr_t *)vap,
-                   crgetuid(cr), cr, vap->va_mode)) != 0) {
-                       ZFS_EXIT(zfsvfs);
-                       return (error);
-               }
-       }
-
-       if ((error = zfs_acl_ids_create(dzp, 0, vap, cr,
-           vsecp, &acl_ids)) != 0) {
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-       /*
-        * First make sure the new directory doesn't exist.
-        *
-        * Existence is checked first to make sure we don't return
-        * EACCES instead of EEXIST which can cause some applications
-        * to fail.
-        */
-top:
-       *ipp = NULL;
-
-       if ((error = zfs_dirent_lock(&dl, dzp, dirname, &zp, zf,
-           NULL, NULL))) {
-               zfs_acl_ids_free(&acl_ids);
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-
-       if ((error = zfs_zaccess(dzp, ACE_ADD_SUBDIRECTORY, 0, B_FALSE, cr))) {
-               zfs_acl_ids_free(&acl_ids);
-               zfs_dirent_unlock(dl);
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-
-       if (zfs_acl_ids_overquota(zfsvfs, &acl_ids)) {
-               zfs_acl_ids_free(&acl_ids);
-               zfs_dirent_unlock(dl);
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EDQUOT));
-       }
-
-       /*
-        * Add a new entry to the directory.
-        */
-       tx = dmu_tx_create(zfsvfs->z_os);
-       dmu_tx_hold_zap(tx, dzp->z_id, TRUE, dirname);
-       dmu_tx_hold_zap(tx, DMU_NEW_OBJECT, FALSE, NULL);
-       fuid_dirtied = zfsvfs->z_fuid_dirty;
-       if (fuid_dirtied)
-               zfs_fuid_txhold(zfsvfs, tx);
-       if (!zfsvfs->z_use_sa && acl_ids.z_aclp->z_acl_bytes > ZFS_ACE_SPACE) {
-               dmu_tx_hold_write(tx, DMU_NEW_OBJECT, 0,
-                   acl_ids.z_aclp->z_acl_bytes);
-       }
-
-       dmu_tx_hold_sa_create(tx, acl_ids.z_aclp->z_acl_bytes +
-           ZFS_SA_BASE_ATTR_SIZE);
-
-       error = dmu_tx_assign(tx, (waited ? TXG_NOTHROTTLE : 0) | TXG_NOWAIT);
-       if (error) {
-               zfs_dirent_unlock(dl);
-               if (error == ERESTART) {
-                       waited = B_TRUE;
-                       dmu_tx_wait(tx);
-                       dmu_tx_abort(tx);
-                       goto top;
-               }
-               zfs_acl_ids_free(&acl_ids);
-               dmu_tx_abort(tx);
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-
-       /*
-        * Create new node.
-        */
-       zfs_mknode(dzp, vap, tx, cr, 0, &zp, &acl_ids);
-
-       /*
-        * Now put new name in parent dir.
-        */
-       error = zfs_link_create(dl, zp, tx, ZNEW);
-       if (error != 0) {
-               zfs_znode_delete(zp, tx);
-               remove_inode_hash(ZTOI(zp));
-               goto out;
-       }
-
-       if (fuid_dirtied)
-               zfs_fuid_sync(zfsvfs, tx);
-
-       *ipp = ZTOI(zp);
-
-       txtype = zfs_log_create_txtype(Z_DIR, vsecp, vap);
-       if (flags & FIGNORECASE)
-               txtype |= TX_CI;
-       zfs_log_create(zilog, tx, txtype, dzp, zp, dirname, vsecp,
-           acl_ids.z_fuidp, vap);
-
-out:
-       zfs_acl_ids_free(&acl_ids);
-
-       dmu_tx_commit(tx);
-
-       zfs_dirent_unlock(dl);
-
-       if (zfsvfs->z_os->os_sync == ZFS_SYNC_ALWAYS)
-               zil_commit(zilog, 0);
-
-       if (error != 0) {
-               iput(ZTOI(zp));
-       } else {
-               zfs_inode_update(dzp);
-               zfs_inode_update(zp);
-       }
-       ZFS_EXIT(zfsvfs);
-       return (error);
-}
-
-/*
- * Remove a directory subdir entry.  If the current working
- * directory is the same as the subdir to be removed, the
- * remove will fail.
- *
- *     IN:     dip     - inode of directory to remove from.
- *             name    - name of directory to be removed.
- *             cwd     - inode of current working directory.
- *             cr      - credentials of caller.
- *             flags   - case flags
- *
- *     RETURN: 0 on success, error code on failure.
- *
- * Timestamps:
- *     dip - ctime|mtime updated
- */
-/*ARGSUSED*/
-int
-zfs_rmdir(struct inode *dip, char *name, struct inode *cwd, cred_t *cr,
-    int flags)
-{
-       znode_t         *dzp = ITOZ(dip);
-       znode_t         *zp;
-       struct inode    *ip;
-       zfsvfs_t        *zfsvfs = ITOZSB(dip);
-       zilog_t         *zilog;
-       zfs_dirlock_t   *dl;
-       dmu_tx_t        *tx;
-       int             error;
-       int             zflg = ZEXISTS;
-       boolean_t       waited = B_FALSE;
-
-       if (name == NULL)
-               return (SET_ERROR(EINVAL));
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(dzp);
-       zilog = zfsvfs->z_log;
-
-       if (flags & FIGNORECASE)
-               zflg |= ZCILOOK;
-top:
-       zp = NULL;
-
-       /*
-        * Attempt to lock directory; fail if entry doesn't exist.
-        */
-       if ((error = zfs_dirent_lock(&dl, dzp, name, &zp, zflg,
-           NULL, NULL))) {
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-
-       ip = ZTOI(zp);
-
-       if ((error = zfs_zaccess_delete(dzp, zp, cr))) {
-               goto out;
-       }
-
-       if (!S_ISDIR(ip->i_mode)) {
-               error = SET_ERROR(ENOTDIR);
-               goto out;
-       }
-
-       if (ip == cwd) {
-               error = SET_ERROR(EINVAL);
-               goto out;
-       }
-
-       /*
-        * Grab a lock on the directory to make sure that no one is
-        * trying to add (or lookup) entries while we are removing it.
-        */
-       rw_enter(&zp->z_name_lock, RW_WRITER);
-
-       /*
-        * Grab a lock on the parent pointer to make sure we play well
-        * with the treewalk and directory rename code.
-        */
-       rw_enter(&zp->z_parent_lock, RW_WRITER);
-
-       tx = dmu_tx_create(zfsvfs->z_os);
-       dmu_tx_hold_zap(tx, dzp->z_id, FALSE, name);
-       dmu_tx_hold_sa(tx, zp->z_sa_hdl, B_FALSE);
-       dmu_tx_hold_zap(tx, zfsvfs->z_unlinkedobj, FALSE, NULL);
-       zfs_sa_upgrade_txholds(tx, zp);
-       zfs_sa_upgrade_txholds(tx, dzp);
-       dmu_tx_mark_netfree(tx);
-       error = dmu_tx_assign(tx, (waited ? TXG_NOTHROTTLE : 0) | TXG_NOWAIT);
-       if (error) {
-               rw_exit(&zp->z_parent_lock);
-               rw_exit(&zp->z_name_lock);
-               zfs_dirent_unlock(dl);
-               if (error == ERESTART) {
-                       waited = B_TRUE;
-                       dmu_tx_wait(tx);
-                       dmu_tx_abort(tx);
-                       iput(ip);
-                       goto top;
-               }
-               dmu_tx_abort(tx);
-               iput(ip);
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-
-       error = zfs_link_destroy(dl, zp, tx, zflg, NULL);
-
-       if (error == 0) {
-               uint64_t txtype = TX_RMDIR;
-               if (flags & FIGNORECASE)
-                       txtype |= TX_CI;
-               zfs_log_remove(zilog, tx, txtype, dzp, name, ZFS_NO_OBJECT);
-       }
-
-       dmu_tx_commit(tx);
-
-       rw_exit(&zp->z_parent_lock);
-       rw_exit(&zp->z_name_lock);
-out:
-       zfs_dirent_unlock(dl);
-
-       zfs_inode_update(dzp);
-       zfs_inode_update(zp);
-       iput(ip);
-
-       if (zfsvfs->z_os->os_sync == ZFS_SYNC_ALWAYS)
-               zil_commit(zilog, 0);
-
-       ZFS_EXIT(zfsvfs);
-       return (error);
-}
-
-/*
- * Read as many directory entries as will fit into the provided
- * dirent buffer from the given directory cursor position.
- *
- *     IN:     ip      - inode of directory to read.
- *             dirent  - buffer for directory entries.
- *
- *     OUT:    dirent  - filler buffer of directory entries.
- *
- *     RETURN: 0 if success
- *             error code if failure
- *
- * Timestamps:
- *     ip - atime updated
- *
- * Note that the low 4 bits of the cookie returned by zap is always zero.
- * This allows us to use the low range for "special" directory entries:
- * We use 0 for '.', and 1 for '..'.  If this is the root of the filesystem,
- * we use the offset 2 for the '.zfs' directory.
- */
-/* ARGSUSED */
-int
-zfs_readdir(struct inode *ip, struct dir_context *ctx, cred_t *cr)
-{
-       znode_t         *zp = ITOZ(ip);
-       zfsvfs_t        *zfsvfs = ITOZSB(ip);
-       objset_t        *os;
-       zap_cursor_t    zc;
-       zap_attribute_t zap;
-       int             error;
-       uint8_t         prefetch;
-       uint8_t         type;
-       int             done = 0;
-       uint64_t        parent;
-       uint64_t        offset; /* must be unsigned; checks for < 1 */
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(zp);
-
-       if ((error = sa_lookup(zp->z_sa_hdl, SA_ZPL_PARENT(zfsvfs),
-           &parent, sizeof (parent))) != 0)
-               goto out;
-
-       /*
-        * Quit if directory has been removed (posix)
-        */
-       if (zp->z_unlinked)
-               goto out;
-
-       error = 0;
-       os = zfsvfs->z_os;
-       offset = ctx->pos;
-       prefetch = zp->z_zn_prefetch;
-
-       /*
-        * Initialize the iterator cursor.
-        */
-       if (offset <= 3) {
-               /*
-                * Start iteration from the beginning of the directory.
-                */
-               zap_cursor_init(&zc, os, zp->z_id);
-       } else {
-               /*
-                * The offset is a serialized cursor.
-                */
-               zap_cursor_init_serialized(&zc, os, zp->z_id, offset);
-       }
-
-       /*
-        * Transform to file-system independent format
-        */
-       while (!done) {
-               uint64_t objnum;
-               /*
-                * Special case `.', `..', and `.zfs'.
-                */
-               if (offset == 0) {
-                       (void) strcpy(zap.za_name, ".");
-                       zap.za_normalization_conflict = 0;
-                       objnum = zp->z_id;
-                       type = DT_DIR;
-               } else if (offset == 1) {
-                       (void) strcpy(zap.za_name, "..");
-                       zap.za_normalization_conflict = 0;
-                       objnum = parent;
-                       type = DT_DIR;
-               } else if (offset == 2 && zfs_show_ctldir(zp)) {
-                       (void) strcpy(zap.za_name, ZFS_CTLDIR_NAME);
-                       zap.za_normalization_conflict = 0;
-                       objnum = ZFSCTL_INO_ROOT;
-                       type = DT_DIR;
-               } else {
-                       /*
-                        * Grab next entry.
-                        */
-                       if ((error = zap_cursor_retrieve(&zc, &zap))) {
-                               if (error == ENOENT)
-                                       break;
-                               else
-                                       goto update;
-                       }
-
-                       /*
-                        * Allow multiple entries provided the first entry is
-                        * the object id.  Non-zpl consumers may safely make
-                        * use of the additional space.
-                        *
-                        * XXX: This should be a feature flag for compatibility
-                        */
-                       if (zap.za_integer_length != 8 ||
-                           zap.za_num_integers == 0) {
-                               cmn_err(CE_WARN, "zap_readdir: bad directory "
-                                   "entry, obj = %lld, offset = %lld, "
-                                   "length = %d, num = %lld\n",
-                                   (u_longlong_t)zp->z_id,
-                                   (u_longlong_t)offset,
-                                   zap.za_integer_length,
-                                   (u_longlong_t)zap.za_num_integers);
-                               error = SET_ERROR(ENXIO);
-                               goto update;
-                       }
-
-                       objnum = ZFS_DIRENT_OBJ(zap.za_first_integer);
-                       type = ZFS_DIRENT_TYPE(zap.za_first_integer);
-               }
-
-               done = !dir_emit(ctx, zap.za_name, strlen(zap.za_name),
-                   objnum, type);
-               if (done)
-                       break;
-
-               /* Prefetch znode */
-               if (prefetch) {
-                       dmu_prefetch(os, objnum, 0, 0, 0,
-                           ZIO_PRIORITY_SYNC_READ);
-               }
-
-               /*
-                * Move to the next entry, fill in the previous offset.
-                */
-               if (offset > 2 || (offset == 2 && !zfs_show_ctldir(zp))) {
-                       zap_cursor_advance(&zc);
-                       offset = zap_cursor_serialize(&zc);
-               } else {
-                       offset += 1;
-               }
-               ctx->pos = offset;
-       }
-       zp->z_zn_prefetch = B_FALSE; /* a lookup will re-enable pre-fetching */
-
-update:
-       zap_cursor_fini(&zc);
-       if (error == ENOENT)
-               error = 0;
-out:
-       ZFS_EXIT(zfsvfs);
-
-       return (error);
-}
-
-ulong_t zfs_fsync_sync_cnt = 4;
-
-int
-zfs_fsync(struct inode *ip, int syncflag, cred_t *cr)
-{
-       znode_t *zp = ITOZ(ip);
-       zfsvfs_t *zfsvfs = ITOZSB(ip);
-
-       (void) tsd_set(zfs_fsyncer_key, (void *)zfs_fsync_sync_cnt);
-
-       if (zfsvfs->z_os->os_sync != ZFS_SYNC_DISABLED) {
-               ZFS_ENTER(zfsvfs);
-               ZFS_VERIFY_ZP(zp);
-               zil_commit(zfsvfs->z_log, zp->z_id);
-               ZFS_EXIT(zfsvfs);
-       }
-       tsd_set(zfs_fsyncer_key, NULL);
-
-       return (0);
-}
-
-
-/*
- * Get the requested file attributes and place them in the provided
- * vattr structure.
- *
- *     IN:     ip      - inode of file.
- *             vap     - va_mask identifies requested attributes.
- *                       If ATTR_XVATTR set, then optional attrs are requested
- *             flags   - ATTR_NOACLCHECK (CIFS server context)
- *             cr      - credentials of caller.
- *
- *     OUT:    vap     - attribute values.
- *
- *     RETURN: 0 (always succeeds)
- */
-/* ARGSUSED */
-int
-zfs_getattr(struct inode *ip, vattr_t *vap, int flags, cred_t *cr)
-{
-       znode_t *zp = ITOZ(ip);
-       zfsvfs_t *zfsvfs = ITOZSB(ip);
-       int     error = 0;
-       uint64_t links;
-       uint64_t atime[2], mtime[2], ctime[2];
-       xvattr_t *xvap = (xvattr_t *)vap;       /* vap may be an xvattr_t * */
-       xoptattr_t *xoap = NULL;
-       boolean_t skipaclchk = (flags & ATTR_NOACLCHECK) ? B_TRUE : B_FALSE;
-       sa_bulk_attr_t bulk[3];
-       int count = 0;
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(zp);
-
-       zfs_fuid_map_ids(zp, cr, &vap->va_uid, &vap->va_gid);
-
-       SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_ATIME(zfsvfs), NULL, &atime, 16);
-       SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_MTIME(zfsvfs), NULL, &mtime, 16);
-       SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_CTIME(zfsvfs), NULL, &ctime, 16);
-
-       if ((error = sa_bulk_lookup(zp->z_sa_hdl, bulk, count)) != 0) {
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-
-       /*
-        * If ACL is trivial don't bother looking for ACE_READ_ATTRIBUTES.
-        * Also, if we are the owner don't bother, since owner should
-        * always be allowed to read basic attributes of file.
-        */
-       if (!(zp->z_pflags & ZFS_ACL_TRIVIAL) &&
-           (vap->va_uid != crgetuid(cr))) {
-               if ((error = zfs_zaccess(zp, ACE_READ_ATTRIBUTES, 0,
-                   skipaclchk, cr))) {
-                       ZFS_EXIT(zfsvfs);
-                       return (error);
-               }
-       }
-
-       /*
-        * Return all attributes.  It's cheaper to provide the answer
-        * than to determine whether we were asked the question.
-        */
-
-       mutex_enter(&zp->z_lock);
-       vap->va_type = vn_mode_to_vtype(zp->z_mode);
-       vap->va_mode = zp->z_mode;
-       vap->va_fsid = ZTOI(zp)->i_sb->s_dev;
-       vap->va_nodeid = zp->z_id;
-       if ((zp->z_id == zfsvfs->z_root) && zfs_show_ctldir(zp))
-               links = ZTOI(zp)->i_nlink + 1;
-       else
-               links = ZTOI(zp)->i_nlink;
-       vap->va_nlink = MIN(links, ZFS_LINK_MAX);
-       vap->va_size = i_size_read(ip);
-       vap->va_rdev = ip->i_rdev;
-       vap->va_seq = ip->i_generation;
-
-       /*
-        * Add in any requested optional attributes and the create time.
-        * Also set the corresponding bits in the returned attribute bitmap.
-        */
-       if ((xoap = xva_getxoptattr(xvap)) != NULL && zfsvfs->z_use_fuids) {
-               if (XVA_ISSET_REQ(xvap, XAT_ARCHIVE)) {
-                       xoap->xoa_archive =
-                           ((zp->z_pflags & ZFS_ARCHIVE) != 0);
-                       XVA_SET_RTN(xvap, XAT_ARCHIVE);
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_READONLY)) {
-                       xoap->xoa_readonly =
-                           ((zp->z_pflags & ZFS_READONLY) != 0);
-                       XVA_SET_RTN(xvap, XAT_READONLY);
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_SYSTEM)) {
-                       xoap->xoa_system =
-                           ((zp->z_pflags & ZFS_SYSTEM) != 0);
-                       XVA_SET_RTN(xvap, XAT_SYSTEM);
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_HIDDEN)) {
-                       xoap->xoa_hidden =
-                           ((zp->z_pflags & ZFS_HIDDEN) != 0);
-                       XVA_SET_RTN(xvap, XAT_HIDDEN);
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_NOUNLINK)) {
-                       xoap->xoa_nounlink =
-                           ((zp->z_pflags & ZFS_NOUNLINK) != 0);
-                       XVA_SET_RTN(xvap, XAT_NOUNLINK);
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_IMMUTABLE)) {
-                       xoap->xoa_immutable =
-                           ((zp->z_pflags & ZFS_IMMUTABLE) != 0);
-                       XVA_SET_RTN(xvap, XAT_IMMUTABLE);
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_APPENDONLY)) {
-                       xoap->xoa_appendonly =
-                           ((zp->z_pflags & ZFS_APPENDONLY) != 0);
-                       XVA_SET_RTN(xvap, XAT_APPENDONLY);
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_NODUMP)) {
-                       xoap->xoa_nodump =
-                           ((zp->z_pflags & ZFS_NODUMP) != 0);
-                       XVA_SET_RTN(xvap, XAT_NODUMP);
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_OPAQUE)) {
-                       xoap->xoa_opaque =
-                           ((zp->z_pflags & ZFS_OPAQUE) != 0);
-                       XVA_SET_RTN(xvap, XAT_OPAQUE);
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_AV_QUARANTINED)) {
-                       xoap->xoa_av_quarantined =
-                           ((zp->z_pflags & ZFS_AV_QUARANTINED) != 0);
-                       XVA_SET_RTN(xvap, XAT_AV_QUARANTINED);
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_AV_MODIFIED)) {
-                       xoap->xoa_av_modified =
-                           ((zp->z_pflags & ZFS_AV_MODIFIED) != 0);
-                       XVA_SET_RTN(xvap, XAT_AV_MODIFIED);
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_AV_SCANSTAMP) &&
-                   S_ISREG(ip->i_mode)) {
-                       zfs_sa_get_scanstamp(zp, xvap);
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_CREATETIME)) {
-                       uint64_t times[2];
-
-                       (void) sa_lookup(zp->z_sa_hdl, SA_ZPL_CRTIME(zfsvfs),
-                           times, sizeof (times));
-                       ZFS_TIME_DECODE(&xoap->xoa_createtime, times);
-                       XVA_SET_RTN(xvap, XAT_CREATETIME);
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_REPARSE)) {
-                       xoap->xoa_reparse = ((zp->z_pflags & ZFS_REPARSE) != 0);
-                       XVA_SET_RTN(xvap, XAT_REPARSE);
-               }
-               if (XVA_ISSET_REQ(xvap, XAT_GEN)) {
-                       xoap->xoa_generation = ip->i_generation;
-                       XVA_SET_RTN(xvap, XAT_GEN);
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_OFFLINE)) {
-                       xoap->xoa_offline =
-                           ((zp->z_pflags & ZFS_OFFLINE) != 0);
-                       XVA_SET_RTN(xvap, XAT_OFFLINE);
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_SPARSE)) {
-                       xoap->xoa_sparse =
-                           ((zp->z_pflags & ZFS_SPARSE) != 0);
-                       XVA_SET_RTN(xvap, XAT_SPARSE);
-               }
-       }
-
-       ZFS_TIME_DECODE(&vap->va_atime, atime);
-       ZFS_TIME_DECODE(&vap->va_mtime, mtime);
-       ZFS_TIME_DECODE(&vap->va_ctime, ctime);
-
-       mutex_exit(&zp->z_lock);
-
-       sa_object_size(zp->z_sa_hdl, &vap->va_blksize, &vap->va_nblocks);
-
-       if (zp->z_blksz == 0) {
-               /*
-                * Block size hasn't been set; suggest maximal I/O transfers.
-                */
-               vap->va_blksize = zfsvfs->z_max_blksz;
-       }
-
-       ZFS_EXIT(zfsvfs);
-       return (0);
-}
-
-/*
- * Get the basic file attributes and place them in the provided kstat
- * structure.  The inode is assumed to be the authoritative source
- * for most of the attributes.  However, the znode currently has the
- * authoritative atime, blksize, and block count.
- *
- *     IN:     ip      - inode of file.
- *
- *     OUT:    sp      - kstat values.
- *
- *     RETURN: 0 (always succeeds)
- */
-/* ARGSUSED */
-int
-zfs_getattr_fast(struct inode *ip, struct kstat *sp)
-{
-       znode_t *zp = ITOZ(ip);
-       zfsvfs_t *zfsvfs = ITOZSB(ip);
-       uint32_t blksize;
-       u_longlong_t nblocks;
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(zp);
-
-       mutex_enter(&zp->z_lock);
-
-       generic_fillattr(ip, sp);
-
-       sa_object_size(zp->z_sa_hdl, &blksize, &nblocks);
-       sp->blksize = blksize;
-       sp->blocks = nblocks;
-
-       if (unlikely(zp->z_blksz == 0)) {
-               /*
-                * Block size hasn't been set; suggest maximal I/O transfers.
-                */
-               sp->blksize = zfsvfs->z_max_blksz;
-       }
-
-       mutex_exit(&zp->z_lock);
-
-       /*
-        * Required to prevent NFS client from detecting different inode
-        * numbers of snapshot root dentry before and after snapshot mount.
-        */
-       if (zfsvfs->z_issnap) {
-               if (ip->i_sb->s_root->d_inode == ip)
-                       sp->ino = ZFSCTL_INO_SNAPDIRS -
-                           dmu_objset_id(zfsvfs->z_os);
-       }
-
-       ZFS_EXIT(zfsvfs);
-
-       return (0);
-}
-
-/*
- * Set the file attributes to the values contained in the
- * vattr structure.
- *
- *     IN:     ip      - inode of file to be modified.
- *             vap     - new attribute values.
- *                       If ATTR_XVATTR set, then optional attrs are being set
- *             flags   - ATTR_UTIME set if non-default time values provided.
- *                     - ATTR_NOACLCHECK (CIFS context only).
- *             cr      - credentials of caller.
- *
- *     RETURN: 0 if success
- *             error code if failure
- *
- * Timestamps:
- *     ip - ctime updated, mtime updated if size changed.
- */
-/* ARGSUSED */
-int
-zfs_setattr(struct inode *ip, vattr_t *vap, int flags, cred_t *cr)
-{
-       znode_t         *zp = ITOZ(ip);
-       zfsvfs_t        *zfsvfs = ITOZSB(ip);
-       zilog_t         *zilog;
-       dmu_tx_t        *tx;
-       vattr_t         oldva;
-       xvattr_t        *tmpxvattr;
-       uint_t          mask = vap->va_mask;
-       uint_t          saved_mask = 0;
-       int             trim_mask = 0;
-       uint64_t        new_mode;
-       uint64_t        new_kuid = 0, new_kgid = 0, new_uid, new_gid;
-       uint64_t        xattr_obj;
-       uint64_t        mtime[2], ctime[2], atime[2];
-       znode_t         *attrzp;
-       int             need_policy = FALSE;
-       int             err, err2;
-       zfs_fuid_info_t *fuidp = NULL;
-       xvattr_t *xvap = (xvattr_t *)vap;       /* vap may be an xvattr_t * */
-       xoptattr_t      *xoap;
-       zfs_acl_t       *aclp;
-       boolean_t skipaclchk = (flags & ATTR_NOACLCHECK) ? B_TRUE : B_FALSE;
-       boolean_t       fuid_dirtied = B_FALSE;
-       sa_bulk_attr_t  *bulk, *xattr_bulk;
-       int             count = 0, xattr_count = 0;
-
-       if (mask == 0)
-               return (0);
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(zp);
-
-       zilog = zfsvfs->z_log;
-
-       /*
-        * Make sure that if we have ephemeral uid/gid or xvattr specified
-        * that file system is at proper version level
-        */
-
-       if (zfsvfs->z_use_fuids == B_FALSE &&
-           (((mask & ATTR_UID) && IS_EPHEMERAL(vap->va_uid)) ||
-           ((mask & ATTR_GID) && IS_EPHEMERAL(vap->va_gid)) ||
-           (mask & ATTR_XVATTR))) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EINVAL));
-       }
-
-       if (mask & ATTR_SIZE && S_ISDIR(ip->i_mode)) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EISDIR));
-       }
-
-       if (mask & ATTR_SIZE && !S_ISREG(ip->i_mode) && !S_ISFIFO(ip->i_mode)) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EINVAL));
-       }
-
-       /*
-        * If this is an xvattr_t, then get a pointer to the structure of
-        * optional attributes.  If this is NULL, then we have a vattr_t.
-        */
-       xoap = xva_getxoptattr(xvap);
-
-       tmpxvattr = kmem_alloc(sizeof (xvattr_t), KM_SLEEP);
-       xva_init(tmpxvattr);
-
-       bulk = kmem_alloc(sizeof (sa_bulk_attr_t) * 7, KM_SLEEP);
-       xattr_bulk = kmem_alloc(sizeof (sa_bulk_attr_t) * 7, KM_SLEEP);
-
-       /*
-        * Immutable files can only alter immutable bit and atime
-        */
-       if ((zp->z_pflags & ZFS_IMMUTABLE) &&
-           ((mask & (ATTR_SIZE|ATTR_UID|ATTR_GID|ATTR_MTIME|ATTR_MODE)) ||
-           ((mask & ATTR_XVATTR) && XVA_ISSET_REQ(xvap, XAT_CREATETIME)))) {
-               err = SET_ERROR(EPERM);
-               goto out3;
-       }
-
-       if ((mask & ATTR_SIZE) && (zp->z_pflags & ZFS_READONLY)) {
-               err = SET_ERROR(EPERM);
-               goto out3;
-       }
-
-       /*
-        * Verify timestamps doesn't overflow 32 bits.
-        * ZFS can handle large timestamps, but 32bit syscalls can't
-        * handle times greater than 2039.  This check should be removed
-        * once large timestamps are fully supported.
-        */
-       if (mask & (ATTR_ATIME | ATTR_MTIME)) {
-               if (((mask & ATTR_ATIME) &&
-                   TIMESPEC_OVERFLOW(&vap->va_atime)) ||
-                   ((mask & ATTR_MTIME) &&
-                   TIMESPEC_OVERFLOW(&vap->va_mtime))) {
-                       err = SET_ERROR(EOVERFLOW);
-                       goto out3;
-               }
-       }
-
-top:
-       attrzp = NULL;
-       aclp = NULL;
-
-       /* Can this be moved to before the top label? */
-       if (zfs_is_readonly(zfsvfs)) {
-               err = SET_ERROR(EROFS);
-               goto out3;
-       }
-
-       /*
-        * First validate permissions
-        */
-
-       if (mask & ATTR_SIZE) {
-               err = zfs_zaccess(zp, ACE_WRITE_DATA, 0, skipaclchk, cr);
-               if (err)
-                       goto out3;
-
-               /*
-                * XXX - Note, we are not providing any open
-                * mode flags here (like FNDELAY), so we may
-                * block if there are locks present... this
-                * should be addressed in openat().
-                */
-               /* XXX - would it be OK to generate a log record here? */
-               err = zfs_freesp(zp, vap->va_size, 0, 0, FALSE);
-               if (err)
-                       goto out3;
-       }
-
-       if (mask & (ATTR_ATIME|ATTR_MTIME) ||
-           ((mask & ATTR_XVATTR) && (XVA_ISSET_REQ(xvap, XAT_HIDDEN) ||
-           XVA_ISSET_REQ(xvap, XAT_READONLY) ||
-           XVA_ISSET_REQ(xvap, XAT_ARCHIVE) ||
-           XVA_ISSET_REQ(xvap, XAT_OFFLINE) ||
-           XVA_ISSET_REQ(xvap, XAT_SPARSE) ||
-           XVA_ISSET_REQ(xvap, XAT_CREATETIME) ||
-           XVA_ISSET_REQ(xvap, XAT_SYSTEM)))) {
-               need_policy = zfs_zaccess(zp, ACE_WRITE_ATTRIBUTES, 0,
-                   skipaclchk, cr);
-       }
-
-       if (mask & (ATTR_UID|ATTR_GID)) {
-               int     idmask = (mask & (ATTR_UID|ATTR_GID));
-               int     take_owner;
-               int     take_group;
-
-               /*
-                * NOTE: even if a new mode is being set,
-                * we may clear S_ISUID/S_ISGID bits.
-                */
-
-               if (!(mask & ATTR_MODE))
-                       vap->va_mode = zp->z_mode;
-
-               /*
-                * Take ownership or chgrp to group we are a member of
-                */
-
-               take_owner = (mask & ATTR_UID) && (vap->va_uid == crgetuid(cr));
-               take_group = (mask & ATTR_GID) &&
-                   zfs_groupmember(zfsvfs, vap->va_gid, cr);
-
-               /*
-                * If both ATTR_UID and ATTR_GID are set then take_owner and
-                * take_group must both be set in order to allow taking
-                * ownership.
-                *
-                * Otherwise, send the check through secpolicy_vnode_setattr()
-                *
-                */
-
-               if (((idmask == (ATTR_UID|ATTR_GID)) &&
-                   take_owner && take_group) ||
-                   ((idmask == ATTR_UID) && take_owner) ||
-                   ((idmask == ATTR_GID) && take_group)) {
-                       if (zfs_zaccess(zp, ACE_WRITE_OWNER, 0,
-                           skipaclchk, cr) == 0) {
-                               /*
-                                * Remove setuid/setgid for non-privileged users
-                                */
-                               (void) secpolicy_setid_clear(vap, cr);
-                               trim_mask = (mask & (ATTR_UID|ATTR_GID));
-                       } else {
-                               need_policy =  TRUE;
-                       }
-               } else {
-                       need_policy =  TRUE;
-               }
-       }
-
-       mutex_enter(&zp->z_lock);
-       oldva.va_mode = zp->z_mode;
-       zfs_fuid_map_ids(zp, cr, &oldva.va_uid, &oldva.va_gid);
-       if (mask & ATTR_XVATTR) {
-               /*
-                * Update xvattr mask to include only those attributes
-                * that are actually changing.
-                *
-                * the bits will be restored prior to actually setting
-                * the attributes so the caller thinks they were set.
-                */
-               if (XVA_ISSET_REQ(xvap, XAT_APPENDONLY)) {
-                       if (xoap->xoa_appendonly !=
-                           ((zp->z_pflags & ZFS_APPENDONLY) != 0)) {
-                               need_policy = TRUE;
-                       } else {
-                               XVA_CLR_REQ(xvap, XAT_APPENDONLY);
-                               XVA_SET_REQ(tmpxvattr, XAT_APPENDONLY);
-                       }
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_NOUNLINK)) {
-                       if (xoap->xoa_nounlink !=
-                           ((zp->z_pflags & ZFS_NOUNLINK) != 0)) {
-                               need_policy = TRUE;
-                       } else {
-                               XVA_CLR_REQ(xvap, XAT_NOUNLINK);
-                               XVA_SET_REQ(tmpxvattr, XAT_NOUNLINK);
-                       }
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_IMMUTABLE)) {
-                       if (xoap->xoa_immutable !=
-                           ((zp->z_pflags & ZFS_IMMUTABLE) != 0)) {
-                               need_policy = TRUE;
-                       } else {
-                               XVA_CLR_REQ(xvap, XAT_IMMUTABLE);
-                               XVA_SET_REQ(tmpxvattr, XAT_IMMUTABLE);
-                       }
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_NODUMP)) {
-                       if (xoap->xoa_nodump !=
-                           ((zp->z_pflags & ZFS_NODUMP) != 0)) {
-                               need_policy = TRUE;
-                       } else {
-                               XVA_CLR_REQ(xvap, XAT_NODUMP);
-                               XVA_SET_REQ(tmpxvattr, XAT_NODUMP);
-                       }
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_AV_MODIFIED)) {
-                       if (xoap->xoa_av_modified !=
-                           ((zp->z_pflags & ZFS_AV_MODIFIED) != 0)) {
-                               need_policy = TRUE;
-                       } else {
-                               XVA_CLR_REQ(xvap, XAT_AV_MODIFIED);
-                               XVA_SET_REQ(tmpxvattr, XAT_AV_MODIFIED);
-                       }
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_AV_QUARANTINED)) {
-                       if ((!S_ISREG(ip->i_mode) &&
-                           xoap->xoa_av_quarantined) ||
-                           xoap->xoa_av_quarantined !=
-                           ((zp->z_pflags & ZFS_AV_QUARANTINED) != 0)) {
-                               need_policy = TRUE;
-                       } else {
-                               XVA_CLR_REQ(xvap, XAT_AV_QUARANTINED);
-                               XVA_SET_REQ(tmpxvattr, XAT_AV_QUARANTINED);
-                       }
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_REPARSE)) {
-                       mutex_exit(&zp->z_lock);
-                       err = SET_ERROR(EPERM);
-                       goto out3;
-               }
-
-               if (need_policy == FALSE &&
-                   (XVA_ISSET_REQ(xvap, XAT_AV_SCANSTAMP) ||
-                   XVA_ISSET_REQ(xvap, XAT_OPAQUE))) {
-                       need_policy = TRUE;
-               }
-       }
-
-       mutex_exit(&zp->z_lock);
-
-       if (mask & ATTR_MODE) {
-               if (zfs_zaccess(zp, ACE_WRITE_ACL, 0, skipaclchk, cr) == 0) {
-                       err = secpolicy_setid_setsticky_clear(ip, vap,
-                           &oldva, cr);
-                       if (err)
-                               goto out3;
-
-                       trim_mask |= ATTR_MODE;
-               } else {
-                       need_policy = TRUE;
-               }
-       }
-
-       if (need_policy) {
-               /*
-                * If trim_mask is set then take ownership
-                * has been granted or write_acl is present and user
-                * has the ability to modify mode.  In that case remove
-                * UID|GID and or MODE from mask so that
-                * secpolicy_vnode_setattr() doesn't revoke it.
-                */
-
-               if (trim_mask) {
-                       saved_mask = vap->va_mask;
-                       vap->va_mask &= ~trim_mask;
-               }
-               err = secpolicy_vnode_setattr(cr, ip, vap, &oldva, flags,
-                   (int (*)(void *, int, cred_t *))zfs_zaccess_unix, zp);
-               if (err)
-                       goto out3;
-
-               if (trim_mask)
-                       vap->va_mask |= saved_mask;
-       }
-
-       /*
-        * secpolicy_vnode_setattr, or take ownership may have
-        * changed va_mask
-        */
-       mask = vap->va_mask;
-
-       if ((mask & (ATTR_UID | ATTR_GID))) {
-               err = sa_lookup(zp->z_sa_hdl, SA_ZPL_XATTR(zfsvfs),
-                   &xattr_obj, sizeof (xattr_obj));
-
-               if (err == 0 && xattr_obj) {
-                       err = zfs_zget(ZTOZSB(zp), xattr_obj, &attrzp);
-                       if (err)
-                               goto out2;
-               }
-               if (mask & ATTR_UID) {
-                       new_kuid = zfs_fuid_create(zfsvfs,
-                           (uint64_t)vap->va_uid, cr, ZFS_OWNER, &fuidp);
-                       if (new_kuid != KUID_TO_SUID(ZTOI(zp)->i_uid) &&
-                           zfs_fuid_overquota(zfsvfs, B_FALSE, new_kuid)) {
-                               if (attrzp)
-                                       iput(ZTOI(attrzp));
-                               err = SET_ERROR(EDQUOT);
-                               goto out2;
-                       }
-               }
-
-               if (mask & ATTR_GID) {
-                       new_kgid = zfs_fuid_create(zfsvfs,
-                           (uint64_t)vap->va_gid, cr, ZFS_GROUP, &fuidp);
-                       if (new_kgid != KGID_TO_SGID(ZTOI(zp)->i_gid) &&
-                           zfs_fuid_overquota(zfsvfs, B_TRUE, new_kgid)) {
-                               if (attrzp)
-                                       iput(ZTOI(attrzp));
-                               err = SET_ERROR(EDQUOT);
-                               goto out2;
-                       }
-               }
-       }
-       tx = dmu_tx_create(zfsvfs->z_os);
-
-       if (mask & ATTR_MODE) {
-               uint64_t pmode = zp->z_mode;
-               uint64_t acl_obj;
-               new_mode = (pmode & S_IFMT) | (vap->va_mode & ~S_IFMT);
-
-               zfs_acl_chmod_setattr(zp, &aclp, new_mode);
-
-               mutex_enter(&zp->z_lock);
-               if (!zp->z_is_sa && ((acl_obj = zfs_external_acl(zp)) != 0)) {
-                       /*
-                        * Are we upgrading ACL from old V0 format
-                        * to V1 format?
-                        */
-                       if (zfsvfs->z_version >= ZPL_VERSION_FUID &&
-                           zfs_znode_acl_version(zp) ==
-                           ZFS_ACL_VERSION_INITIAL) {
-                               dmu_tx_hold_free(tx, acl_obj, 0,
-                                   DMU_OBJECT_END);
-                               dmu_tx_hold_write(tx, DMU_NEW_OBJECT,
-                                   0, aclp->z_acl_bytes);
-                       } else {
-                               dmu_tx_hold_write(tx, acl_obj, 0,
-                                   aclp->z_acl_bytes);
-                       }
-               } else if (!zp->z_is_sa && aclp->z_acl_bytes > ZFS_ACE_SPACE) {
-                       dmu_tx_hold_write(tx, DMU_NEW_OBJECT,
-                           0, aclp->z_acl_bytes);
-               }
-               mutex_exit(&zp->z_lock);
-               dmu_tx_hold_sa(tx, zp->z_sa_hdl, B_TRUE);
-       } else {
-               if ((mask & ATTR_XVATTR) &&
-                   XVA_ISSET_REQ(xvap, XAT_AV_SCANSTAMP))
-                       dmu_tx_hold_sa(tx, zp->z_sa_hdl, B_TRUE);
-               else
-                       dmu_tx_hold_sa(tx, zp->z_sa_hdl, B_FALSE);
-       }
-
-       if (attrzp) {
-               dmu_tx_hold_sa(tx, attrzp->z_sa_hdl, B_FALSE);
-       }
-
-       fuid_dirtied = zfsvfs->z_fuid_dirty;
-       if (fuid_dirtied)
-               zfs_fuid_txhold(zfsvfs, tx);
-
-       zfs_sa_upgrade_txholds(tx, zp);
-
-       err = dmu_tx_assign(tx, TXG_WAIT);
-       if (err)
-               goto out;
-
-       count = 0;
-       /*
-        * Set each attribute requested.
-        * We group settings according to the locks they need to acquire.
-        *
-        * Note: you cannot set ctime directly, although it will be
-        * updated as a side-effect of calling this function.
-        */
-
-
-       if (mask & (ATTR_UID|ATTR_GID|ATTR_MODE))
-               mutex_enter(&zp->z_acl_lock);
-       mutex_enter(&zp->z_lock);
-
-       SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_FLAGS(zfsvfs), NULL,
-           &zp->z_pflags, sizeof (zp->z_pflags));
-
-       if (attrzp) {
-               if (mask & (ATTR_UID|ATTR_GID|ATTR_MODE))
-                       mutex_enter(&attrzp->z_acl_lock);
-               mutex_enter(&attrzp->z_lock);
-               SA_ADD_BULK_ATTR(xattr_bulk, xattr_count,
-                   SA_ZPL_FLAGS(zfsvfs), NULL, &attrzp->z_pflags,
-                   sizeof (attrzp->z_pflags));
-       }
-
-       if (mask & (ATTR_UID|ATTR_GID)) {
-
-               if (mask & ATTR_UID) {
-                       ZTOI(zp)->i_uid = SUID_TO_KUID(new_kuid);
-                       new_uid = zfs_uid_read(ZTOI(zp));
-                       SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_UID(zfsvfs), NULL,
-                           &new_uid, sizeof (new_uid));
-                       if (attrzp) {
-                               SA_ADD_BULK_ATTR(xattr_bulk, xattr_count,
-                                   SA_ZPL_UID(zfsvfs), NULL, &new_uid,
-                                   sizeof (new_uid));
-                               ZTOI(attrzp)->i_uid = SUID_TO_KUID(new_uid);
-                       }
-               }
-
-               if (mask & ATTR_GID) {
-                       ZTOI(zp)->i_gid = SGID_TO_KGID(new_kgid);
-                       new_gid = zfs_gid_read(ZTOI(zp));
-                       SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_GID(zfsvfs),
-                           NULL, &new_gid, sizeof (new_gid));
-                       if (attrzp) {
-                               SA_ADD_BULK_ATTR(xattr_bulk, xattr_count,
-                                   SA_ZPL_GID(zfsvfs), NULL, &new_gid,
-                                   sizeof (new_gid));
-                               ZTOI(attrzp)->i_gid = SGID_TO_KGID(new_kgid);
-                       }
-               }
-               if (!(mask & ATTR_MODE)) {
-                       SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_MODE(zfsvfs),
-                           NULL, &new_mode, sizeof (new_mode));
-                       new_mode = zp->z_mode;
-               }
-               err = zfs_acl_chown_setattr(zp);
-               ASSERT(err == 0);
-               if (attrzp) {
-                       err = zfs_acl_chown_setattr(attrzp);
-                       ASSERT(err == 0);
-               }
-       }
-
-       if (mask & ATTR_MODE) {
-               SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_MODE(zfsvfs), NULL,
-                   &new_mode, sizeof (new_mode));
-               zp->z_mode = ZTOI(zp)->i_mode = new_mode;
-               ASSERT3P(aclp, !=, NULL);
-               err = zfs_aclset_common(zp, aclp, cr, tx);
-               ASSERT0(err);
-               if (zp->z_acl_cached)
-                       zfs_acl_free(zp->z_acl_cached);
-               zp->z_acl_cached = aclp;
-               aclp = NULL;
-       }
-
-       if ((mask & ATTR_ATIME) || zp->z_atime_dirty) {
-               zp->z_atime_dirty = 0;
-               ZFS_TIME_ENCODE(&ip->i_atime, atime);
-               SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_ATIME(zfsvfs), NULL,
-                   &atime, sizeof (atime));
-       }
-
-       if (mask & (ATTR_MTIME | ATTR_SIZE)) {
-               ZFS_TIME_ENCODE(&vap->va_mtime, mtime);
-               ZTOI(zp)->i_mtime = timespec_trunc(vap->va_mtime,
-                   ZTOI(zp)->i_sb->s_time_gran);
-
-               SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_MTIME(zfsvfs), NULL,
-                   mtime, sizeof (mtime));
-       }
-
-       if (mask & (ATTR_CTIME | ATTR_SIZE)) {
-               ZFS_TIME_ENCODE(&vap->va_ctime, ctime);
-               ZTOI(zp)->i_ctime = timespec_trunc(vap->va_ctime,
-                   ZTOI(zp)->i_sb->s_time_gran);
-               SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_CTIME(zfsvfs), NULL,
-                   ctime, sizeof (ctime));
-       }
-
-       if (attrzp && mask) {
-               SA_ADD_BULK_ATTR(xattr_bulk, xattr_count,
-                   SA_ZPL_CTIME(zfsvfs), NULL, &ctime,
-                   sizeof (ctime));
-       }
-
-       /*
-        * Do this after setting timestamps to prevent timestamp
-        * update from toggling bit
-        */
-
-       if (xoap && (mask & ATTR_XVATTR)) {
-
-               /*
-                * restore trimmed off masks
-                * so that return masks can be set for caller.
-                */
-
-               if (XVA_ISSET_REQ(tmpxvattr, XAT_APPENDONLY)) {
-                       XVA_SET_REQ(xvap, XAT_APPENDONLY);
-               }
-               if (XVA_ISSET_REQ(tmpxvattr, XAT_NOUNLINK)) {
-                       XVA_SET_REQ(xvap, XAT_NOUNLINK);
-               }
-               if (XVA_ISSET_REQ(tmpxvattr, XAT_IMMUTABLE)) {
-                       XVA_SET_REQ(xvap, XAT_IMMUTABLE);
-               }
-               if (XVA_ISSET_REQ(tmpxvattr, XAT_NODUMP)) {
-                       XVA_SET_REQ(xvap, XAT_NODUMP);
-               }
-               if (XVA_ISSET_REQ(tmpxvattr, XAT_AV_MODIFIED)) {
-                       XVA_SET_REQ(xvap, XAT_AV_MODIFIED);
-               }
-               if (XVA_ISSET_REQ(tmpxvattr, XAT_AV_QUARANTINED)) {
-                       XVA_SET_REQ(xvap, XAT_AV_QUARANTINED);
-               }
-
-               if (XVA_ISSET_REQ(xvap, XAT_AV_SCANSTAMP))
-                       ASSERT(S_ISREG(ip->i_mode));
-
-               zfs_xvattr_set(zp, xvap, tx);
-       }
-
-       if (fuid_dirtied)
-               zfs_fuid_sync(zfsvfs, tx);
-
-       if (mask != 0)
-               zfs_log_setattr(zilog, tx, TX_SETATTR, zp, vap, mask, fuidp);
-
-       mutex_exit(&zp->z_lock);
-       if (mask & (ATTR_UID|ATTR_GID|ATTR_MODE))
-               mutex_exit(&zp->z_acl_lock);
-
-       if (attrzp) {
-               if (mask & (ATTR_UID|ATTR_GID|ATTR_MODE))
-                       mutex_exit(&attrzp->z_acl_lock);
-               mutex_exit(&attrzp->z_lock);
-       }
-out:
-       if (err == 0 && attrzp) {
-               err2 = sa_bulk_update(attrzp->z_sa_hdl, xattr_bulk,
-                   xattr_count, tx);
-               ASSERT(err2 == 0);
-       }
-
-       if (aclp)
-               zfs_acl_free(aclp);
-
-       if (fuidp) {
-               zfs_fuid_info_free(fuidp);
-               fuidp = NULL;
-       }
-
-       if (err) {
-               dmu_tx_abort(tx);
-               if (attrzp)
-                       iput(ZTOI(attrzp));
-               if (err == ERESTART)
-                       goto top;
-       } else {
-               err2 = sa_bulk_update(zp->z_sa_hdl, bulk, count, tx);
-               dmu_tx_commit(tx);
-               if (attrzp)
-                       iput(ZTOI(attrzp));
-               zfs_inode_update(zp);
-       }
-
-out2:
-       if (zfsvfs->z_os->os_sync == ZFS_SYNC_ALWAYS)
-               zil_commit(zilog, 0);
-
-out3:
-       kmem_free(xattr_bulk, sizeof (sa_bulk_attr_t) * 7);
-       kmem_free(bulk, sizeof (sa_bulk_attr_t) * 7);
-       kmem_free(tmpxvattr, sizeof (xvattr_t));
-       ZFS_EXIT(zfsvfs);
-       return (err);
-}
-
-typedef struct zfs_zlock {
-       krwlock_t       *zl_rwlock;     /* lock we acquired */
-       znode_t         *zl_znode;      /* znode we held */
-       struct zfs_zlock *zl_next;      /* next in list */
-} zfs_zlock_t;
-
-/*
- * Drop locks and release vnodes that were held by zfs_rename_lock().
- */
-static void
-zfs_rename_unlock(zfs_zlock_t **zlpp)
-{
-       zfs_zlock_t *zl;
-
-       while ((zl = *zlpp) != NULL) {
-               if (zl->zl_znode != NULL)
-                       zfs_iput_async(ZTOI(zl->zl_znode));
-               rw_exit(zl->zl_rwlock);
-               *zlpp = zl->zl_next;
-               kmem_free(zl, sizeof (*zl));
-       }
-}
-
-/*
- * Search back through the directory tree, using the ".." entries.
- * Lock each directory in the chain to prevent concurrent renames.
- * Fail any attempt to move a directory into one of its own descendants.
- * XXX - z_parent_lock can overlap with map or grow locks
- */
-static int
-zfs_rename_lock(znode_t *szp, znode_t *tdzp, znode_t *sdzp, zfs_zlock_t **zlpp)
-{
-       zfs_zlock_t     *zl;
-       znode_t         *zp = tdzp;
-       uint64_t        rootid = ZTOZSB(zp)->z_root;
-       uint64_t        oidp = zp->z_id;
-       krwlock_t       *rwlp = &szp->z_parent_lock;
-       krw_t           rw = RW_WRITER;
-
-       /*
-        * First pass write-locks szp and compares to zp->z_id.
-        * Later passes read-lock zp and compare to zp->z_parent.
-        */
-       do {
-               if (!rw_tryenter(rwlp, rw)) {
-                       /*
-                        * Another thread is renaming in this path.
-                        * Note that if we are a WRITER, we don't have any
-                        * parent_locks held yet.
-                        */
-                       if (rw == RW_READER && zp->z_id > szp->z_id) {
-                               /*
-                                * Drop our locks and restart
-                                */
-                               zfs_rename_unlock(&zl);
-                               *zlpp = NULL;
-                               zp = tdzp;
-                               oidp = zp->z_id;
-                               rwlp = &szp->z_parent_lock;
-                               rw = RW_WRITER;
-                               continue;
-                       } else {
-                               /*
-                                * Wait for other thread to drop its locks
-                                */
-                               rw_enter(rwlp, rw);
-                       }
-               }
-
-               zl = kmem_alloc(sizeof (*zl), KM_SLEEP);
-               zl->zl_rwlock = rwlp;
-               zl->zl_znode = NULL;
-               zl->zl_next = *zlpp;
-               *zlpp = zl;
-
-               if (oidp == szp->z_id)          /* We're a descendant of szp */
-                       return (SET_ERROR(EINVAL));
-
-               if (oidp == rootid)             /* We've hit the top */
-                       return (0);
-
-               if (rw == RW_READER) {          /* i.e. not the first pass */
-                       int error = zfs_zget(ZTOZSB(zp), oidp, &zp);
-                       if (error)
-                               return (error);
-                       zl->zl_znode = zp;
-               }
-               (void) sa_lookup(zp->z_sa_hdl, SA_ZPL_PARENT(ZTOZSB(zp)),
-                   &oidp, sizeof (oidp));
-               rwlp = &zp->z_parent_lock;
-               rw = RW_READER;
-
-       } while (zp->z_id != sdzp->z_id);
-
-       return (0);
-}
-
-/*
- * Move an entry from the provided source directory to the target
- * directory.  Change the entry name as indicated.
- *
- *     IN:     sdip    - Source directory containing the "old entry".
- *             snm     - Old entry name.
- *             tdip    - Target directory to contain the "new entry".
- *             tnm     - New entry name.
- *             cr      - credentials of caller.
- *             flags   - case flags
- *
- *     RETURN: 0 on success, error code on failure.
- *
- * Timestamps:
- *     sdip,tdip - ctime|mtime updated
- */
-/*ARGSUSED*/
-int
-zfs_rename(struct inode *sdip, char *snm, struct inode *tdip, char *tnm,
-    cred_t *cr, int flags)
-{
-       znode_t         *tdzp, *szp, *tzp;
-       znode_t         *sdzp = ITOZ(sdip);
-       zfsvfs_t        *zfsvfs = ITOZSB(sdip);
-       zilog_t         *zilog;
-       zfs_dirlock_t   *sdl, *tdl;
-       dmu_tx_t        *tx;
-       zfs_zlock_t     *zl;
-       int             cmp, serr, terr;
-       int             error = 0;
-       int             zflg = 0;
-       boolean_t       waited = B_FALSE;
-
-       if (snm == NULL || tnm == NULL)
-               return (SET_ERROR(EINVAL));
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(sdzp);
-       zilog = zfsvfs->z_log;
-
-       tdzp = ITOZ(tdip);
-       ZFS_VERIFY_ZP(tdzp);
-
-       /*
-        * We check i_sb because snapshots and the ctldir must have different
-        * super blocks.
-        */
-       if (tdip->i_sb != sdip->i_sb || zfsctl_is_node(tdip)) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EXDEV));
-       }
-
-       if (zfsvfs->z_utf8 && u8_validate(tnm,
-           strlen(tnm), NULL, U8_VALIDATE_ENTIRE, &error) < 0) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EILSEQ));
-       }
-
-       if (flags & FIGNORECASE)
-               zflg |= ZCILOOK;
-
-top:
-       szp = NULL;
-       tzp = NULL;
-       zl = NULL;
-
-       /*
-        * This is to prevent the creation of links into attribute space
-        * by renaming a linked file into/outof an attribute directory.
-        * See the comment in zfs_link() for why this is considered bad.
-        */
-       if ((tdzp->z_pflags & ZFS_XATTR) != (sdzp->z_pflags & ZFS_XATTR)) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EINVAL));
-       }
-
-       /*
-        * Lock source and target directory entries.  To prevent deadlock,
-        * a lock ordering must be defined.  We lock the directory with
-        * the smallest object id first, or if it's a tie, the one with
-        * the lexically first name.
-        */
-       if (sdzp->z_id < tdzp->z_id) {
-               cmp = -1;
-       } else if (sdzp->z_id > tdzp->z_id) {
-               cmp = 1;
-       } else {
-               /*
-                * First compare the two name arguments without
-                * considering any case folding.
-                */
-               int nofold = (zfsvfs->z_norm & ~U8_TEXTPREP_TOUPPER);
-
-               cmp = u8_strcmp(snm, tnm, 0, nofold, U8_UNICODE_LATEST, &error);
-               ASSERT(error == 0 || !zfsvfs->z_utf8);
-               if (cmp == 0) {
-                       /*
-                        * POSIX: "If the old argument and the new argument
-                        * both refer to links to the same existing file,
-                        * the rename() function shall return successfully
-                        * and perform no other action."
-                        */
-                       ZFS_EXIT(zfsvfs);
-                       return (0);
-               }
-               /*
-                * If the file system is case-folding, then we may
-                * have some more checking to do.  A case-folding file
-                * system is either supporting mixed case sensitivity
-                * access or is completely case-insensitive.  Note
-                * that the file system is always case preserving.
-                *
-                * In mixed sensitivity mode case sensitive behavior
-                * is the default.  FIGNORECASE must be used to
-                * explicitly request case insensitive behavior.
-                *
-                * If the source and target names provided differ only
-                * by case (e.g., a request to rename 'tim' to 'Tim'),
-                * we will treat this as a special case in the
-                * case-insensitive mode: as long as the source name
-                * is an exact match, we will allow this to proceed as
-                * a name-change request.
-                */
-               if ((zfsvfs->z_case == ZFS_CASE_INSENSITIVE ||
-                   (zfsvfs->z_case == ZFS_CASE_MIXED &&
-                   flags & FIGNORECASE)) &&
-                   u8_strcmp(snm, tnm, 0, zfsvfs->z_norm, U8_UNICODE_LATEST,
-                   &error) == 0) {
-                       /*
-                        * case preserving rename request, require exact
-                        * name matches
-                        */
-                       zflg |= ZCIEXACT;
-                       zflg &= ~ZCILOOK;
-               }
-       }
-
-       /*
-        * If the source and destination directories are the same, we should
-        * grab the z_name_lock of that directory only once.
-        */
-       if (sdzp == tdzp) {
-               zflg |= ZHAVELOCK;
-               rw_enter(&sdzp->z_name_lock, RW_READER);
-       }
-
-       if (cmp < 0) {
-               serr = zfs_dirent_lock(&sdl, sdzp, snm, &szp,
-                   ZEXISTS | zflg, NULL, NULL);
-               terr = zfs_dirent_lock(&tdl,
-                   tdzp, tnm, &tzp, ZRENAMING | zflg, NULL, NULL);
-       } else {
-               terr = zfs_dirent_lock(&tdl,
-                   tdzp, tnm, &tzp, zflg, NULL, NULL);
-               serr = zfs_dirent_lock(&sdl,
-                   sdzp, snm, &szp, ZEXISTS | ZRENAMING | zflg,
-                   NULL, NULL);
-       }
-
-       if (serr) {
-               /*
-                * Source entry invalid or not there.
-                */
-               if (!terr) {
-                       zfs_dirent_unlock(tdl);
-                       if (tzp)
-                               iput(ZTOI(tzp));
-               }
-
-               if (sdzp == tdzp)
-                       rw_exit(&sdzp->z_name_lock);
-
-               if (strcmp(snm, "..") == 0)
-                       serr = EINVAL;
-               ZFS_EXIT(zfsvfs);
-               return (serr);
-       }
-       if (terr) {
-               zfs_dirent_unlock(sdl);
-               iput(ZTOI(szp));
-
-               if (sdzp == tdzp)
-                       rw_exit(&sdzp->z_name_lock);
-
-               if (strcmp(tnm, "..") == 0)
-                       terr = EINVAL;
-               ZFS_EXIT(zfsvfs);
-               return (terr);
-       }
-
-       /*
-        * Must have write access at the source to remove the old entry
-        * and write access at the target to create the new entry.
-        * Note that if target and source are the same, this can be
-        * done in a single check.
-        */
-
-       if ((error = zfs_zaccess_rename(sdzp, szp, tdzp, tzp, cr)))
-               goto out;
-
-       if (S_ISDIR(ZTOI(szp)->i_mode)) {
-               /*
-                * Check to make sure rename is valid.
-                * Can't do a move like this: /usr/a/b to /usr/a/b/c/d
-                */
-               if ((error = zfs_rename_lock(szp, tdzp, sdzp, &zl)))
-                       goto out;
-       }
-
-       /*
-        * Does target exist?
-        */
-       if (tzp) {
-               /*
-                * Source and target must be the same type.
-                */
-               if (S_ISDIR(ZTOI(szp)->i_mode)) {
-                       if (!S_ISDIR(ZTOI(tzp)->i_mode)) {
-                               error = SET_ERROR(ENOTDIR);
-                               goto out;
-                       }
-               } else {
-                       if (S_ISDIR(ZTOI(tzp)->i_mode)) {
-                               error = SET_ERROR(EISDIR);
-                               goto out;
-                       }
-               }
-               /*
-                * POSIX dictates that when the source and target
-                * entries refer to the same file object, rename
-                * must do nothing and exit without error.
-                */
-               if (szp->z_id == tzp->z_id) {
-                       error = 0;
-                       goto out;
-               }
-       }
-
-       tx = dmu_tx_create(zfsvfs->z_os);
-       dmu_tx_hold_sa(tx, szp->z_sa_hdl, B_FALSE);
-       dmu_tx_hold_sa(tx, sdzp->z_sa_hdl, B_FALSE);
-       dmu_tx_hold_zap(tx, sdzp->z_id, FALSE, snm);
-       dmu_tx_hold_zap(tx, tdzp->z_id, TRUE, tnm);
-       if (sdzp != tdzp) {
-               dmu_tx_hold_sa(tx, tdzp->z_sa_hdl, B_FALSE);
-               zfs_sa_upgrade_txholds(tx, tdzp);
-       }
-       if (tzp) {
-               dmu_tx_hold_sa(tx, tzp->z_sa_hdl, B_FALSE);
-               zfs_sa_upgrade_txholds(tx, tzp);
-       }
-
-       zfs_sa_upgrade_txholds(tx, szp);
-       dmu_tx_hold_zap(tx, zfsvfs->z_unlinkedobj, FALSE, NULL);
-       error = dmu_tx_assign(tx, (waited ? TXG_NOTHROTTLE : 0) | TXG_NOWAIT);
-       if (error) {
-               if (zl != NULL)
-                       zfs_rename_unlock(&zl);
-               zfs_dirent_unlock(sdl);
-               zfs_dirent_unlock(tdl);
-
-               if (sdzp == tdzp)
-                       rw_exit(&sdzp->z_name_lock);
-
-               if (error == ERESTART) {
-                       waited = B_TRUE;
-                       dmu_tx_wait(tx);
-                       dmu_tx_abort(tx);
-                       iput(ZTOI(szp));
-                       if (tzp)
-                               iput(ZTOI(tzp));
-                       goto top;
-               }
-               dmu_tx_abort(tx);
-               iput(ZTOI(szp));
-               if (tzp)
-                       iput(ZTOI(tzp));
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-
-       if (tzp)        /* Attempt to remove the existing target */
-               error = zfs_link_destroy(tdl, tzp, tx, zflg, NULL);
-
-       if (error == 0) {
-               error = zfs_link_create(tdl, szp, tx, ZRENAMING);
-               if (error == 0) {
-                       szp->z_pflags |= ZFS_AV_MODIFIED;
-
-                       error = sa_update(szp->z_sa_hdl, SA_ZPL_FLAGS(zfsvfs),
-                           (void *)&szp->z_pflags, sizeof (uint64_t), tx);
-                       ASSERT0(error);
-
-                       error = zfs_link_destroy(sdl, szp, tx, ZRENAMING, NULL);
-                       if (error == 0) {
-                               zfs_log_rename(zilog, tx, TX_RENAME |
-                                   (flags & FIGNORECASE ? TX_CI : 0), sdzp,
-                                   sdl->dl_name, tdzp, tdl->dl_name, szp);
-                       } else {
-                               /*
-                                * At this point, we have successfully created
-                                * the target name, but have failed to remove
-                                * the source name.  Since the create was done
-                                * with the ZRENAMING flag, there are
-                                * complications; for one, the link count is
-                                * wrong.  The easiest way to deal with this
-                                * is to remove the newly created target, and
-                                * return the original error.  This must
-                                * succeed; fortunately, it is very unlikely to
-                                * fail, since we just created it.
-                                */
-                               VERIFY3U(zfs_link_destroy(tdl, szp, tx,
-                                   ZRENAMING, NULL), ==, 0);
-                       }
-               } else {
-                       /*
-                        * If we had removed the existing target, subsequent
-                        * call to zfs_link_create() to add back the same entry
-                        * but, the new dnode (szp) should not fail.
-                        */
-                       ASSERT(tzp == NULL);
-               }
-       }
-
-       dmu_tx_commit(tx);
-out:
-       if (zl != NULL)
-               zfs_rename_unlock(&zl);
-
-       zfs_dirent_unlock(sdl);
-       zfs_dirent_unlock(tdl);
-
-       zfs_inode_update(sdzp);
-       if (sdzp == tdzp)
-               rw_exit(&sdzp->z_name_lock);
-
-       if (sdzp != tdzp)
-               zfs_inode_update(tdzp);
-
-       zfs_inode_update(szp);
-       iput(ZTOI(szp));
-       if (tzp) {
-               zfs_inode_update(tzp);
-               iput(ZTOI(tzp));
-       }
-
-       if (zfsvfs->z_os->os_sync == ZFS_SYNC_ALWAYS)
-               zil_commit(zilog, 0);
-
-       ZFS_EXIT(zfsvfs);
-       return (error);
-}
-
-/*
- * Insert the indicated symbolic reference entry into the directory.
- *
- *     IN:     dip     - Directory to contain new symbolic link.
- *             link    - Name for new symlink entry.
- *             vap     - Attributes of new entry.
- *             target  - Target path of new symlink.
- *
- *             cr      - credentials of caller.
- *             flags   - case flags
- *
- *     RETURN: 0 on success, error code on failure.
- *
- * Timestamps:
- *     dip - ctime|mtime updated
- */
-/*ARGSUSED*/
-int
-zfs_symlink(struct inode *dip, char *name, vattr_t *vap, char *link,
-    struct inode **ipp, cred_t *cr, int flags)
-{
-       znode_t         *zp, *dzp = ITOZ(dip);
-       zfs_dirlock_t   *dl;
-       dmu_tx_t        *tx;
-       zfsvfs_t        *zfsvfs = ITOZSB(dip);
-       zilog_t         *zilog;
-       uint64_t        len = strlen(link);
-       int             error;
-       int             zflg = ZNEW;
-       zfs_acl_ids_t   acl_ids;
-       boolean_t       fuid_dirtied;
-       uint64_t        txtype = TX_SYMLINK;
-       boolean_t       waited = B_FALSE;
-
-       ASSERT(S_ISLNK(vap->va_mode));
-
-       if (name == NULL)
-               return (SET_ERROR(EINVAL));
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(dzp);
-       zilog = zfsvfs->z_log;
-
-       if (zfsvfs->z_utf8 && u8_validate(name, strlen(name),
-           NULL, U8_VALIDATE_ENTIRE, &error) < 0) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EILSEQ));
-       }
-       if (flags & FIGNORECASE)
-               zflg |= ZCILOOK;
-
-       if (len > MAXPATHLEN) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(ENAMETOOLONG));
-       }
-
-       if ((error = zfs_acl_ids_create(dzp, 0,
-           vap, cr, NULL, &acl_ids)) != 0) {
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-top:
-       *ipp = NULL;
-
-       /*
-        * Attempt to lock directory; fail if entry already exists.
-        */
-       error = zfs_dirent_lock(&dl, dzp, name, &zp, zflg, NULL, NULL);
-       if (error) {
-               zfs_acl_ids_free(&acl_ids);
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-
-       if ((error = zfs_zaccess(dzp, ACE_ADD_FILE, 0, B_FALSE, cr))) {
-               zfs_acl_ids_free(&acl_ids);
-               zfs_dirent_unlock(dl);
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-
-       if (zfs_acl_ids_overquota(zfsvfs, &acl_ids)) {
-               zfs_acl_ids_free(&acl_ids);
-               zfs_dirent_unlock(dl);
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EDQUOT));
-       }
-       tx = dmu_tx_create(zfsvfs->z_os);
-       fuid_dirtied = zfsvfs->z_fuid_dirty;
-       dmu_tx_hold_write(tx, DMU_NEW_OBJECT, 0, MAX(1, len));
-       dmu_tx_hold_zap(tx, dzp->z_id, TRUE, name);
-       dmu_tx_hold_sa_create(tx, acl_ids.z_aclp->z_acl_bytes +
-           ZFS_SA_BASE_ATTR_SIZE + len);
-       dmu_tx_hold_sa(tx, dzp->z_sa_hdl, B_FALSE);
-       if (!zfsvfs->z_use_sa && acl_ids.z_aclp->z_acl_bytes > ZFS_ACE_SPACE) {
-               dmu_tx_hold_write(tx, DMU_NEW_OBJECT, 0,
-                   acl_ids.z_aclp->z_acl_bytes);
-       }
-       if (fuid_dirtied)
-               zfs_fuid_txhold(zfsvfs, tx);
-       error = dmu_tx_assign(tx, (waited ? TXG_NOTHROTTLE : 0) | TXG_NOWAIT);
-       if (error) {
-               zfs_dirent_unlock(dl);
-               if (error == ERESTART) {
-                       waited = B_TRUE;
-                       dmu_tx_wait(tx);
-                       dmu_tx_abort(tx);
-                       goto top;
-               }
-               zfs_acl_ids_free(&acl_ids);
-               dmu_tx_abort(tx);
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-
-       /*
-        * Create a new object for the symlink.
-        * for version 4 ZPL datsets the symlink will be an SA attribute
-        */
-       zfs_mknode(dzp, vap, tx, cr, 0, &zp, &acl_ids);
-
-       if (fuid_dirtied)
-               zfs_fuid_sync(zfsvfs, tx);
-
-       mutex_enter(&zp->z_lock);
-       if (zp->z_is_sa)
-               error = sa_update(zp->z_sa_hdl, SA_ZPL_SYMLINK(zfsvfs),
-                   link, len, tx);
-       else
-               zfs_sa_symlink(zp, link, len, tx);
-       mutex_exit(&zp->z_lock);
-
-       zp->z_size = len;
-       (void) sa_update(zp->z_sa_hdl, SA_ZPL_SIZE(zfsvfs),
-           &zp->z_size, sizeof (zp->z_size), tx);
-       /*
-        * Insert the new object into the directory.
-        */
-       error = zfs_link_create(dl, zp, tx, ZNEW);
-       if (error != 0) {
-               zfs_znode_delete(zp, tx);
-               remove_inode_hash(ZTOI(zp));
-       } else {
-               if (flags & FIGNORECASE)
-                       txtype |= TX_CI;
-               zfs_log_symlink(zilog, tx, txtype, dzp, zp, name, link);
-
-               zfs_inode_update(dzp);
-               zfs_inode_update(zp);
-       }
-
-       zfs_acl_ids_free(&acl_ids);
-
-       dmu_tx_commit(tx);
-
-       zfs_dirent_unlock(dl);
-
-       if (error == 0) {
-               *ipp = ZTOI(zp);
-
-               if (zfsvfs->z_os->os_sync == ZFS_SYNC_ALWAYS)
-                       zil_commit(zilog, 0);
-       } else {
-               iput(ZTOI(zp));
-       }
-
-       ZFS_EXIT(zfsvfs);
-       return (error);
-}
-
-/*
- * Return, in the buffer contained in the provided uio structure,
- * the symbolic path referred to by ip.
- *
- *     IN:     ip      - inode of symbolic link
- *             uio     - structure to contain the link path.
- *             cr      - credentials of caller.
- *
- *     RETURN: 0 if success
- *             error code if failure
- *
- * Timestamps:
- *     ip - atime updated
- */
-/* ARGSUSED */
-int
-zfs_readlink(struct inode *ip, uio_t *uio, cred_t *cr)
-{
-       znode_t         *zp = ITOZ(ip);
-       zfsvfs_t        *zfsvfs = ITOZSB(ip);
-       int             error;
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(zp);
-
-       mutex_enter(&zp->z_lock);
-       if (zp->z_is_sa)
-               error = sa_lookup_uio(zp->z_sa_hdl,
-                   SA_ZPL_SYMLINK(zfsvfs), uio);
-       else
-               error = zfs_sa_readlink(zp, uio);
-       mutex_exit(&zp->z_lock);
-
-       ZFS_EXIT(zfsvfs);
-       return (error);
-}
-
-/*
- * Insert a new entry into directory tdip referencing sip.
- *
- *     IN:     tdip    - Directory to contain new entry.
- *             sip     - inode of new entry.
- *             name    - name of new entry.
- *             cr      - credentials of caller.
- *
- *     RETURN: 0 if success
- *             error code if failure
- *
- * Timestamps:
- *     tdip - ctime|mtime updated
- *      sip - ctime updated
- */
-/* ARGSUSED */
-int
-zfs_link(struct inode *tdip, struct inode *sip, char *name, cred_t *cr,
-    int flags)
-{
-       znode_t         *dzp = ITOZ(tdip);
-       znode_t         *tzp, *szp;
-       zfsvfs_t        *zfsvfs = ITOZSB(tdip);
-       zilog_t         *zilog;
-       zfs_dirlock_t   *dl;
-       dmu_tx_t        *tx;
-       int             error;
-       int             zf = ZNEW;
-       uint64_t        parent;
-       uid_t           owner;
-       boolean_t       waited = B_FALSE;
-       boolean_t       is_tmpfile = 0;
-       uint64_t        txg;
-#ifdef HAVE_TMPFILE
-       is_tmpfile = (sip->i_nlink == 0 && (sip->i_state & I_LINKABLE));
-#endif
-       ASSERT(S_ISDIR(tdip->i_mode));
-
-       if (name == NULL)
-               return (SET_ERROR(EINVAL));
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(dzp);
-       zilog = zfsvfs->z_log;
-
-       /*
-        * POSIX dictates that we return EPERM here.
-        * Better choices include ENOTSUP or EISDIR.
-        */
-       if (S_ISDIR(sip->i_mode)) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EPERM));
-       }
-
-       szp = ITOZ(sip);
-       ZFS_VERIFY_ZP(szp);
-
-       /*
-        * We check i_sb because snapshots and the ctldir must have different
-        * super blocks.
-        */
-       if (sip->i_sb != tdip->i_sb || zfsctl_is_node(sip)) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EXDEV));
-       }
-
-       /* Prevent links to .zfs/shares files */
-
-       if ((error = sa_lookup(szp->z_sa_hdl, SA_ZPL_PARENT(zfsvfs),
-           &parent, sizeof (uint64_t))) != 0) {
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-       if (parent == zfsvfs->z_shares_dir) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EPERM));
-       }
-
-       if (zfsvfs->z_utf8 && u8_validate(name,
-           strlen(name), NULL, U8_VALIDATE_ENTIRE, &error) < 0) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EILSEQ));
-       }
-       if (flags & FIGNORECASE)
-               zf |= ZCILOOK;
-
-       /*
-        * We do not support links between attributes and non-attributes
-        * because of the potential security risk of creating links
-        * into "normal" file space in order to circumvent restrictions
-        * imposed in attribute space.
-        */
-       if ((szp->z_pflags & ZFS_XATTR) != (dzp->z_pflags & ZFS_XATTR)) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EINVAL));
-       }
-
-       owner = zfs_fuid_map_id(zfsvfs, KUID_TO_SUID(sip->i_uid),
-           cr, ZFS_OWNER);
-       if (owner != crgetuid(cr) && secpolicy_basic_link(cr) != 0) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EPERM));
-       }
-
-       if ((error = zfs_zaccess(dzp, ACE_ADD_FILE, 0, B_FALSE, cr))) {
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-
-top:
-       /*
-        * Attempt to lock directory; fail if entry already exists.
-        */
-       error = zfs_dirent_lock(&dl, dzp, name, &tzp, zf, NULL, NULL);
-       if (error) {
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-
-       tx = dmu_tx_create(zfsvfs->z_os);
-       dmu_tx_hold_sa(tx, szp->z_sa_hdl, B_FALSE);
-       dmu_tx_hold_zap(tx, dzp->z_id, TRUE, name);
-       if (is_tmpfile)
-               dmu_tx_hold_zap(tx, zfsvfs->z_unlinkedobj, FALSE, NULL);
-
-       zfs_sa_upgrade_txholds(tx, szp);
-       zfs_sa_upgrade_txholds(tx, dzp);
-       error = dmu_tx_assign(tx, (waited ? TXG_NOTHROTTLE : 0) | TXG_NOWAIT);
-       if (error) {
-               zfs_dirent_unlock(dl);
-               if (error == ERESTART) {
-                       waited = B_TRUE;
-                       dmu_tx_wait(tx);
-                       dmu_tx_abort(tx);
-                       goto top;
-               }
-               dmu_tx_abort(tx);
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-       /* unmark z_unlinked so zfs_link_create will not reject */
-       if (is_tmpfile)
-               szp->z_unlinked = 0;
-       error = zfs_link_create(dl, szp, tx, 0);
-
-       if (error == 0) {
-               uint64_t txtype = TX_LINK;
-               /*
-                * tmpfile is created to be in z_unlinkedobj, so remove it.
-                * Also, we don't log in ZIL, be cause all previous file
-                * operation on the tmpfile are ignored by ZIL. Instead we
-                * always wait for txg to sync to make sure all previous
-                * operation are sync safe.
-                */
-               if (is_tmpfile) {
-                       VERIFY(zap_remove_int(zfsvfs->z_os,
-                           zfsvfs->z_unlinkedobj, szp->z_id, tx) == 0);
-               } else {
-                       if (flags & FIGNORECASE)
-                               txtype |= TX_CI;
-                       zfs_log_link(zilog, tx, txtype, dzp, szp, name);
-               }
-       } else if (is_tmpfile) {
-               /* restore z_unlinked since when linking failed */
-               szp->z_unlinked = 1;
-       }
-       txg = dmu_tx_get_txg(tx);
-       dmu_tx_commit(tx);
-
-       zfs_dirent_unlock(dl);
-
-       if (!is_tmpfile && zfsvfs->z_os->os_sync == ZFS_SYNC_ALWAYS)
-               zil_commit(zilog, 0);
-
-       if (is_tmpfile)
-               txg_wait_synced(dmu_objset_pool(zfsvfs->z_os), txg);
-
-       zfs_inode_update(dzp);
-       zfs_inode_update(szp);
-       ZFS_EXIT(zfsvfs);
-       return (error);
-}
-
-static void
-zfs_putpage_commit_cb(void *arg)
-{
-       struct page *pp = arg;
-
-       ClearPageError(pp);
-       end_page_writeback(pp);
-}
-
-/*
- * Push a page out to disk, once the page is on stable storage the
- * registered commit callback will be run as notification of completion.
- *
- *     IN:     ip      - page mapped for inode.
- *             pp      - page to push (page is locked)
- *             wbc     - writeback control data
- *
- *     RETURN: 0 if success
- *             error code if failure
- *
- * Timestamps:
- *     ip - ctime|mtime updated
- */
-/* ARGSUSED */
-int
-zfs_putpage(struct inode *ip, struct page *pp, struct writeback_control *wbc)
-{
-       znode_t         *zp = ITOZ(ip);
-       zfsvfs_t        *zfsvfs = ITOZSB(ip);
-       loff_t          offset;
-       loff_t          pgoff;
-       unsigned int    pglen;
-       rl_t            *rl;
-       dmu_tx_t        *tx;
-       caddr_t         va;
-       int             err = 0;
-       uint64_t        mtime[2], ctime[2];
-       sa_bulk_attr_t  bulk[3];
-       int             cnt = 0;
-       struct address_space *mapping;
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(zp);
-
-       ASSERT(PageLocked(pp));
-
-       pgoff = page_offset(pp);        /* Page byte-offset in file */
-       offset = i_size_read(ip);       /* File length in bytes */
-       pglen = MIN(PAGE_SIZE,          /* Page length in bytes */
-           P2ROUNDUP(offset, PAGE_SIZE)-pgoff);
-
-       /* Page is beyond end of file */
-       if (pgoff >= offset) {
-               unlock_page(pp);
-               ZFS_EXIT(zfsvfs);
-               return (0);
-       }
-
-       /* Truncate page length to end of file */
-       if (pgoff + pglen > offset)
-               pglen = offset - pgoff;
-
-#if 0
-       /*
-        * FIXME: Allow mmap writes past its quota.  The correct fix
-        * is to register a page_mkwrite() handler to count the page
-        * against its quota when it is about to be dirtied.
-        */
-       if (zfs_owner_overquota(zfsvfs, zp, B_FALSE) ||
-           zfs_owner_overquota(zfsvfs, zp, B_TRUE)) {
-               err = EDQUOT;
-       }
-#endif
-
-       /*
-        * The ordering here is critical and must adhere to the following
-        * rules in order to avoid deadlocking in either zfs_read() or
-        * zfs_free_range() due to a lock inversion.
-        *
-        * 1) The page must be unlocked prior to acquiring the range lock.
-        *    This is critical because zfs_read() calls find_lock_page()
-        *    which may block on the page lock while holding the range lock.
-        *
-        * 2) Before setting or clearing write back on a page the range lock
-        *    must be held in order to prevent a lock inversion with the
-        *    zfs_free_range() function.
-        *
-        * This presents a problem because upon entering this function the
-        * page lock is already held.  To safely acquire the range lock the
-        * page lock must be dropped.  This creates a window where another
-        * process could truncate, invalidate, dirty, or write out the page.
-        *
-        * Therefore, after successfully reacquiring the range and page locks
-        * the current page state is checked.  In the common case everything
-        * will be as is expected and it can be written out.  However, if
-        * the page state has changed it must be handled accordingly.
-        */
-       mapping = pp->mapping;
-       redirty_page_for_writepage(wbc, pp);
-       unlock_page(pp);
-
-       rl = zfs_range_lock(&zp->z_range_lock, pgoff, pglen, RL_WRITER);
-       lock_page(pp);
-
-       /* Page mapping changed or it was no longer dirty, we're done */
-       if (unlikely((mapping != pp->mapping) || !PageDirty(pp))) {
-               unlock_page(pp);
-               zfs_range_unlock(rl);
-               ZFS_EXIT(zfsvfs);
-               return (0);
-       }
-
-       /* Another process started write block if required */
-       if (PageWriteback(pp)) {
-               unlock_page(pp);
-               zfs_range_unlock(rl);
-
-               if (wbc->sync_mode != WB_SYNC_NONE)
-                       wait_on_page_writeback(pp);
-
-               ZFS_EXIT(zfsvfs);
-               return (0);
-       }
-
-       /* Clear the dirty flag the required locks are held */
-       if (!clear_page_dirty_for_io(pp)) {
-               unlock_page(pp);
-               zfs_range_unlock(rl);
-               ZFS_EXIT(zfsvfs);
-               return (0);
-       }
-
-       /*
-        * Counterpart for redirty_page_for_writepage() above.  This page
-        * was in fact not skipped and should not be counted as if it were.
-        */
-       wbc->pages_skipped--;
-       set_page_writeback(pp);
-       unlock_page(pp);
-
-       tx = dmu_tx_create(zfsvfs->z_os);
-       dmu_tx_hold_write(tx, zp->z_id, pgoff, pglen);
-       dmu_tx_hold_sa(tx, zp->z_sa_hdl, B_FALSE);
-       zfs_sa_upgrade_txholds(tx, zp);
-
-       err = dmu_tx_assign(tx, TXG_NOWAIT);
-       if (err != 0) {
-               if (err == ERESTART)
-                       dmu_tx_wait(tx);
-
-               dmu_tx_abort(tx);
-               __set_page_dirty_nobuffers(pp);
-               ClearPageError(pp);
-               end_page_writeback(pp);
-               zfs_range_unlock(rl);
-               ZFS_EXIT(zfsvfs);
-               return (err);
-       }
-
-       va = kmap(pp);
-       ASSERT3U(pglen, <=, PAGE_SIZE);
-       dmu_write(zfsvfs->z_os, zp->z_id, pgoff, pglen, va, tx);
-       kunmap(pp);
-
-       SA_ADD_BULK_ATTR(bulk, cnt, SA_ZPL_MTIME(zfsvfs), NULL, &mtime, 16);
-       SA_ADD_BULK_ATTR(bulk, cnt, SA_ZPL_CTIME(zfsvfs), NULL, &ctime, 16);
-       SA_ADD_BULK_ATTR(bulk, cnt, SA_ZPL_FLAGS(zfsvfs), NULL,
-           &zp->z_pflags, 8);
-
-       /* Preserve the mtime and ctime provided by the inode */
-       ZFS_TIME_ENCODE(&ip->i_mtime, mtime);
-       ZFS_TIME_ENCODE(&ip->i_ctime, ctime);
-       zp->z_atime_dirty = 0;
-       zp->z_seq++;
-
-       err = sa_bulk_update(zp->z_sa_hdl, bulk, cnt, tx);
-
-       zfs_log_write(zfsvfs->z_log, tx, TX_WRITE, zp, pgoff, pglen, 0,
-           zfs_putpage_commit_cb, pp);
-       dmu_tx_commit(tx);
-
-       zfs_range_unlock(rl);
-
-       if (wbc->sync_mode != WB_SYNC_NONE) {
-               /*
-                * Note that this is rarely called under writepages(), because
-                * writepages() normally handles the entire commit for
-                * performance reasons.
-                */
-               zil_commit(zfsvfs->z_log, zp->z_id);
-       }
-
-       ZFS_EXIT(zfsvfs);
-       return (err);
-}
-
-/*
- * Update the system attributes when the inode has been dirtied.  For the
- * moment we only update the mode, atime, mtime, and ctime.
- */
-int
-zfs_dirty_inode(struct inode *ip, int flags)
-{
-       znode_t         *zp = ITOZ(ip);
-       zfsvfs_t        *zfsvfs = ITOZSB(ip);
-       dmu_tx_t        *tx;
-       uint64_t        mode, atime[2], mtime[2], ctime[2];
-       sa_bulk_attr_t  bulk[4];
-       int             error = 0;
-       int             cnt = 0;
-
-       if (zfs_is_readonly(zfsvfs) || dmu_objset_is_snapshot(zfsvfs->z_os))
-               return (0);
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(zp);
-
-#ifdef I_DIRTY_TIME
-       /*
-        * This is the lazytime semantic indroduced in Linux 4.0
-        * This flag will only be called from update_time when lazytime is set.
-        * (Note, I_DIRTY_SYNC will also set if not lazytime)
-        * Fortunately mtime and ctime are managed within ZFS itself, so we
-        * only need to dirty atime.
-        */
-       if (flags == I_DIRTY_TIME) {
-               zp->z_atime_dirty = 1;
-               goto out;
-       }
-#endif
-
-       tx = dmu_tx_create(zfsvfs->z_os);
-
-       dmu_tx_hold_sa(tx, zp->z_sa_hdl, B_FALSE);
-       zfs_sa_upgrade_txholds(tx, zp);
-
-       error = dmu_tx_assign(tx, TXG_WAIT);
-       if (error) {
-               dmu_tx_abort(tx);
-               goto out;
-       }
-
-       mutex_enter(&zp->z_lock);
-       zp->z_atime_dirty = 0;
-
-       SA_ADD_BULK_ATTR(bulk, cnt, SA_ZPL_MODE(zfsvfs), NULL, &mode, 8);
-       SA_ADD_BULK_ATTR(bulk, cnt, SA_ZPL_ATIME(zfsvfs), NULL, &atime, 16);
-       SA_ADD_BULK_ATTR(bulk, cnt, SA_ZPL_MTIME(zfsvfs), NULL, &mtime, 16);
-       SA_ADD_BULK_ATTR(bulk, cnt, SA_ZPL_CTIME(zfsvfs), NULL, &ctime, 16);
-
-       /* Preserve the mode, mtime and ctime provided by the inode */
-       ZFS_TIME_ENCODE(&ip->i_atime, atime);
-       ZFS_TIME_ENCODE(&ip->i_mtime, mtime);
-       ZFS_TIME_ENCODE(&ip->i_ctime, ctime);
-       mode = ip->i_mode;
-
-       zp->z_mode = mode;
-
-       error = sa_bulk_update(zp->z_sa_hdl, bulk, cnt, tx);
-       mutex_exit(&zp->z_lock);
-
-       dmu_tx_commit(tx);
-out:
-       ZFS_EXIT(zfsvfs);
-       return (error);
-}
-
-/*ARGSUSED*/
-void
-zfs_inactive(struct inode *ip)
-{
-       znode_t *zp = ITOZ(ip);
-       zfsvfs_t *zfsvfs = ITOZSB(ip);
-       uint64_t atime[2];
-       int error;
-       int need_unlock = 0;
-
-       /* Only read lock if we haven't already write locked, e.g. rollback */
-       if (!RW_WRITE_HELD(&zfsvfs->z_teardown_inactive_lock)) {
-               need_unlock = 1;
-               rw_enter(&zfsvfs->z_teardown_inactive_lock, RW_READER);
-       }
-       if (zp->z_sa_hdl == NULL) {
-               if (need_unlock)
-                       rw_exit(&zfsvfs->z_teardown_inactive_lock);
-               return;
-       }
-
-       if (zp->z_atime_dirty && zp->z_unlinked == 0) {
-               dmu_tx_t *tx = dmu_tx_create(zfsvfs->z_os);
-
-               dmu_tx_hold_sa(tx, zp->z_sa_hdl, B_FALSE);
-               zfs_sa_upgrade_txholds(tx, zp);
-               error = dmu_tx_assign(tx, TXG_WAIT);
-               if (error) {
-                       dmu_tx_abort(tx);
-               } else {
-                       ZFS_TIME_ENCODE(&ip->i_atime, atime);
-                       mutex_enter(&zp->z_lock);
-                       (void) sa_update(zp->z_sa_hdl, SA_ZPL_ATIME(zfsvfs),
-                           (void *)&atime, sizeof (atime), tx);
-                       zp->z_atime_dirty = 0;
-                       mutex_exit(&zp->z_lock);
-                       dmu_tx_commit(tx);
-               }
-       }
-
-       zfs_zinactive(zp);
-       if (need_unlock)
-               rw_exit(&zfsvfs->z_teardown_inactive_lock);
-}
-
-/*
- * Bounds-check the seek operation.
- *
- *     IN:     ip      - inode seeking within
- *             ooff    - old file offset
- *             noffp   - pointer to new file offset
- *             ct      - caller context
- *
- *     RETURN: 0 if success
- *             EINVAL if new offset invalid
- */
-/* ARGSUSED */
-int
-zfs_seek(struct inode *ip, offset_t ooff, offset_t *noffp)
-{
-       if (S_ISDIR(ip->i_mode))
-               return (0);
-       return ((*noffp < 0 || *noffp > MAXOFFSET_T) ? EINVAL : 0);
-}
-
-/*
- * Fill pages with data from the disk.
- */
-static int
-zfs_fillpage(struct inode *ip, struct page *pl[], int nr_pages)
-{
-       znode_t *zp = ITOZ(ip);
-       zfsvfs_t *zfsvfs = ITOZSB(ip);
-       objset_t *os;
-       struct page *cur_pp;
-       u_offset_t io_off, total;
-       size_t io_len;
-       loff_t i_size;
-       unsigned page_idx;
-       int err;
-
-       os = zfsvfs->z_os;
-       io_len = nr_pages << PAGE_SHIFT;
-       i_size = i_size_read(ip);
-       io_off = page_offset(pl[0]);
-
-       if (io_off + io_len > i_size)
-               io_len = i_size - io_off;
-
-       /*
-        * Iterate over list of pages and read each page individually.
-        */
-       page_idx = 0;
-       for (total = io_off + io_len; io_off < total; io_off += PAGESIZE) {
-               caddr_t va;
-
-               cur_pp = pl[page_idx++];
-               va = kmap(cur_pp);
-               err = dmu_read(os, zp->z_id, io_off, PAGESIZE, va,
-                   DMU_READ_PREFETCH);
-               kunmap(cur_pp);
-               if (err) {
-                       /* convert checksum errors into IO errors */
-                       if (err == ECKSUM)
-                               err = SET_ERROR(EIO);
-                       return (err);
-               }
-       }
-
-       return (0);
-}
-
-/*
- * Uses zfs_fillpage to read data from the file and fill the pages.
- *
- *     IN:     ip       - inode of file to get data from.
- *             pl       - list of pages to read
- *             nr_pages - number of pages to read
- *
- *     RETURN: 0 on success, error code on failure.
- *
- * Timestamps:
- *     vp - atime updated
- */
-/* ARGSUSED */
-int
-zfs_getpage(struct inode *ip, struct page *pl[], int nr_pages)
-{
-       znode_t  *zp  = ITOZ(ip);
-       zfsvfs_t *zfsvfs = ITOZSB(ip);
-       int      err;
-
-       if (pl == NULL)
-               return (0);
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(zp);
-
-       err = zfs_fillpage(ip, pl, nr_pages);
-
-       ZFS_EXIT(zfsvfs);
-       return (err);
-}
-
-/*
- * Check ZFS specific permissions to memory map a section of a file.
- *
- *     IN:     ip      - inode of the file to mmap
- *             off     - file offset
- *             addrp   - start address in memory region
- *             len     - length of memory region
- *             vm_flags- address flags
- *
- *     RETURN: 0 if success
- *             error code if failure
- */
-/*ARGSUSED*/
-int
-zfs_map(struct inode *ip, offset_t off, caddr_t *addrp, size_t len,
-    unsigned long vm_flags)
-{
-       znode_t  *zp = ITOZ(ip);
-       zfsvfs_t *zfsvfs = ITOZSB(ip);
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(zp);
-
-       if ((vm_flags & VM_WRITE) && (zp->z_pflags &
-           (ZFS_IMMUTABLE | ZFS_READONLY | ZFS_APPENDONLY))) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EPERM));
-       }
-
-       if ((vm_flags & (VM_READ | VM_EXEC)) &&
-           (zp->z_pflags & ZFS_AV_QUARANTINED)) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EACCES));
-       }
-
-       if (off < 0 || len > MAXOFFSET_T - off) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(ENXIO));
-       }
-
-       ZFS_EXIT(zfsvfs);
-       return (0);
-}
-
-/*
- * convoff - converts the given data (start, whence) to the
- * given whence.
- */
-int
-convoff(struct inode *ip, flock64_t *lckdat, int  whence, offset_t offset)
-{
-       vattr_t vap;
-       int error;
-
-       if ((lckdat->l_whence == 2) || (whence == 2)) {
-               if ((error = zfs_getattr(ip, &vap, 0, CRED())))
-                       return (error);
-       }
-
-       switch (lckdat->l_whence) {
-       case 1:
-               lckdat->l_start += offset;
-               break;
-       case 2:
-               lckdat->l_start += vap.va_size;
-               /* FALLTHRU */
-       case 0:
-               break;
-       default:
-               return (SET_ERROR(EINVAL));
-       }
-
-       if (lckdat->l_start < 0)
-               return (SET_ERROR(EINVAL));
-
-       switch (whence) {
-       case 1:
-               lckdat->l_start -= offset;
-               break;
-       case 2:
-               lckdat->l_start -= vap.va_size;
-               /* FALLTHRU */
-       case 0:
-               break;
-       default:
-               return (SET_ERROR(EINVAL));
-       }
-
-       lckdat->l_whence = (short)whence;
-       return (0);
-}
-
-/*
- * Free or allocate space in a file.  Currently, this function only
- * supports the `F_FREESP' command.  However, this command is somewhat
- * misnamed, as its functionality includes the ability to allocate as
- * well as free space.
- *
- *     IN:     ip      - inode of file to free data in.
- *             cmd     - action to take (only F_FREESP supported).
- *             bfp     - section of file to free/alloc.
- *             flag    - current file open mode flags.
- *             offset  - current file offset.
- *             cr      - credentials of caller [UNUSED].
- *
- *     RETURN: 0 on success, error code on failure.
- *
- * Timestamps:
- *     ip - ctime|mtime updated
- */
-/* ARGSUSED */
-int
-zfs_space(struct inode *ip, int cmd, flock64_t *bfp, int flag,
-    offset_t offset, cred_t *cr)
-{
-       znode_t         *zp = ITOZ(ip);
-       zfsvfs_t        *zfsvfs = ITOZSB(ip);
-       uint64_t        off, len;
-       int             error;
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(zp);
-
-       if (cmd != F_FREESP) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EINVAL));
-       }
-
-       /*
-        * Callers might not be able to detect properly that we are read-only,
-        * so check it explicitly here.
-        */
-       if (zfs_is_readonly(zfsvfs)) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EROFS));
-       }
-
-       if ((error = convoff(ip, bfp, 0, offset))) {
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-
-       if (bfp->l_len < 0) {
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EINVAL));
-       }
+       error = sa_bulk_update(zp->z_sa_hdl, bulk, count, tx);
 
        /*
-        * Permissions aren't checked on Solaris because on this OS
-        * zfs_space() can only be called with an opened file handle.
-        * On Linux we can get here through truncate_range() which
-        * operates directly on inodes, so we need to check access rights.
+        * zil_replaying() not only check if we are replaying ZIL, but also
+        * updates the ZIL header to record replay progress.
         */
-       if ((error = zfs_zaccess(zp, ACE_WRITE_DATA, 0, B_FALSE, cr))) {
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-
-       off = bfp->l_start;
-       len = bfp->l_len; /* 0 means from off to end of file */
-
-       error = zfs_freesp(zp, off, len, flag, TRUE);
-
-       ZFS_EXIT(zfsvfs);
-       return (error);
-}
-
-/*ARGSUSED*/
-int
-zfs_fid(struct inode *ip, fid_t *fidp)
-{
-       znode_t         *zp = ITOZ(ip);
-       zfsvfs_t        *zfsvfs = ITOZSB(ip);
-       uint32_t        gen;
-       uint64_t        gen64;
-       uint64_t        object = zp->z_id;
-       zfid_short_t    *zfid;
-       int             size, i, error;
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(zp);
-
-       if ((error = sa_lookup(zp->z_sa_hdl, SA_ZPL_GEN(zfsvfs),
-           &gen64, sizeof (uint64_t))) != 0) {
-               ZFS_EXIT(zfsvfs);
-               return (error);
-       }
-
-       gen = (uint32_t)gen64;
-
-       size = SHORT_FID_LEN;
-
-       zfid = (zfid_short_t *)fidp;
-
-       zfid->zf_len = size;
-
-       for (i = 0; i < sizeof (zfid->zf_object); i++)
-               zfid->zf_object[i] = (uint8_t)(object >> (8 * i));
-
-       /* Must have a non-zero generation number to distinguish from .zfs */
-       if (gen == 0)
-               gen = 1;
-       for (i = 0; i < sizeof (zfid->zf_gen); i++)
-               zfid->zf_gen[i] = (uint8_t)(gen >> (8 * i));
-
-       ZFS_EXIT(zfsvfs);
-       return (0);
-}
-
-/*ARGSUSED*/
-int
-zfs_getsecattr(struct inode *ip, vsecattr_t *vsecp, int flag, cred_t *cr)
-{
-       znode_t *zp = ITOZ(ip);
-       zfsvfs_t *zfsvfs = ITOZSB(ip);
-       int error;
-       boolean_t skipaclchk = (flag & ATTR_NOACLCHECK) ? B_TRUE : B_FALSE;
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(zp);
-       error = zfs_getacl(zp, vsecp, skipaclchk, cr);
-       ZFS_EXIT(zfsvfs);
-
-       return (error);
-}
-
-/*ARGSUSED*/
-int
-zfs_setsecattr(struct inode *ip, vsecattr_t *vsecp, int flag, cred_t *cr)
-{
-       znode_t *zp = ITOZ(ip);
-       zfsvfs_t *zfsvfs = ITOZSB(ip);
-       int error;
-       boolean_t skipaclchk = (flag & ATTR_NOACLCHECK) ? B_TRUE : B_FALSE;
-       zilog_t *zilog = zfsvfs->z_log;
+       VERIFY(zil_replaying(zfsvfs->z_log, tx));
 
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(zp);
+       dmu_tx_commit(tx);
 
-       error = zfs_setacl(zp, vsecp, skipaclchk, cr);
+       zfs_znode_update_vfs(zp);
 
-       if (zfsvfs->z_os->os_sync == ZFS_SYNC_ALWAYS)
-               zil_commit(zilog, 0);
+       zfs_exit(zfsvfs, FTAG);
 
-       ZFS_EXIT(zfsvfs);
        return (error);
 }
 
-#ifdef HAVE_UIO_ZEROCOPY
-/*
- * Tunable, both must be a power of 2.
- *
- * zcr_blksz_min: the smallest read we may consider to loan out an arcbuf
- * zcr_blksz_max: if set to less than the file block size, allow loaning out of
- *             an arcbuf for a partial block read
- */
-int zcr_blksz_min = (1 << 10); /* 1K */
-int zcr_blksz_max = (1 << 17); /* 128K */
-
-/*ARGSUSED*/
-static int
-zfs_reqzcbuf(struct inode *ip, enum uio_rw ioflag, xuio_t *xuio, cred_t *cr)
-{
-       znode_t *zp = ITOZ(ip);
-       zfsvfs_t *zfsvfs = ITOZSB(ip);
-       int max_blksz = zfsvfs->z_max_blksz;
-       uio_t *uio = &xuio->xu_uio;
-       ssize_t size = uio->uio_resid;
-       offset_t offset = uio->uio_loffset;
-       int blksz;
-       int fullblk, i;
-       arc_buf_t *abuf;
-       ssize_t maxsize;
-       int preamble, postamble;
-
-       if (xuio->xu_type != UIOTYPE_ZEROCOPY)
-               return (SET_ERROR(EINVAL));
-
-       ZFS_ENTER(zfsvfs);
-       ZFS_VERIFY_ZP(zp);
-       switch (ioflag) {
-       case UIO_WRITE:
-               /*
-                * Loan out an arc_buf for write if write size is bigger than
-                * max_blksz, and the file's block size is also max_blksz.
-                */
-               blksz = max_blksz;
-               if (size < blksz || zp->z_blksz != blksz) {
-                       ZFS_EXIT(zfsvfs);
-                       return (SET_ERROR(EINVAL));
-               }
-               /*
-                * Caller requests buffers for write before knowing where the
-                * write offset might be (e.g. NFS TCP write).
-                */
-               if (offset == -1) {
-                       preamble = 0;
-               } else {
-                       preamble = P2PHASE(offset, blksz);
-                       if (preamble) {
-                               preamble = blksz - preamble;
-                               size -= preamble;
-                       }
-               }
-
-               postamble = P2PHASE(size, blksz);
-               size -= postamble;
-
-               fullblk = size / blksz;
-               (void) dmu_xuio_init(xuio,
-                   (preamble != 0) + fullblk + (postamble != 0));
-
-               /*
-                * Have to fix iov base/len for partial buffers.  They
-                * currently represent full arc_buf's.
-                */
-               if (preamble) {
-                       /* data begins in the middle of the arc_buf */
-                       abuf = dmu_request_arcbuf(sa_get_db(zp->z_sa_hdl),
-                           blksz);
-                       ASSERT(abuf);
-                       (void) dmu_xuio_add(xuio, abuf,
-                           blksz - preamble, preamble);
-               }
-
-               for (i = 0; i < fullblk; i++) {
-                       abuf = dmu_request_arcbuf(sa_get_db(zp->z_sa_hdl),
-                           blksz);
-                       ASSERT(abuf);
-                       (void) dmu_xuio_add(xuio, abuf, 0, blksz);
-               }
-
-               if (postamble) {
-                       /* data ends in the middle of the arc_buf */
-                       abuf = dmu_request_arcbuf(sa_get_db(zp->z_sa_hdl),
-                           blksz);
-                       ASSERT(abuf);
-                       (void) dmu_xuio_add(xuio, abuf, 0, postamble);
-               }
-               break;
-       case UIO_READ:
-               /*
-                * Loan out an arc_buf for read if the read size is larger than
-                * the current file block size.  Block alignment is not
-                * considered.  Partial arc_buf will be loaned out for read.
-                */
-               blksz = zp->z_blksz;
-               if (blksz < zcr_blksz_min)
-                       blksz = zcr_blksz_min;
-               if (blksz > zcr_blksz_max)
-                       blksz = zcr_blksz_max;
-               /* avoid potential complexity of dealing with it */
-               if (blksz > max_blksz) {
-                       ZFS_EXIT(zfsvfs);
-                       return (SET_ERROR(EINVAL));
-               }
-
-               maxsize = zp->z_size - uio->uio_loffset;
-               if (size > maxsize)
-                       size = maxsize;
-
-               if (size < blksz) {
-                       ZFS_EXIT(zfsvfs);
-                       return (SET_ERROR(EINVAL));
-               }
-               break;
-       default:
-               ZFS_EXIT(zfsvfs);
-               return (SET_ERROR(EINVAL));
-       }
-
-       uio->uio_extflg = UIO_XUIO;
-       XUIO_XUZC_RW(xuio) = ioflag;
-       ZFS_EXIT(zfsvfs);
-       return (0);
-}
-
-/*ARGSUSED*/
-static int
-zfs_retzcbuf(struct inode *ip, xuio_t *xuio, cred_t *cr)
-{
-       int i;
-       arc_buf_t *abuf;
-       int ioflag = XUIO_XUZC_RW(xuio);
-
-       ASSERT(xuio->xu_type == UIOTYPE_ZEROCOPY);
-
-       i = dmu_xuio_cnt(xuio);
-       while (i-- > 0) {
-               abuf = dmu_xuio_arcbuf(xuio, i);
-               /*
-                * if abuf == NULL, it must be a write buffer
-                * that has been returned in zfs_write().
-                */
-               if (abuf)
-                       dmu_return_arcbuf(abuf);
-               ASSERT(abuf || ioflag == UIO_WRITE);
-       }
-
-       dmu_xuio_fini(xuio);
-       return (0);
-}
-#endif /* HAVE_UIO_ZEROCOPY */
-
-#if defined(_KERNEL) && defined(HAVE_SPL)
-EXPORT_SYMBOL(zfs_open);
-EXPORT_SYMBOL(zfs_close);
-EXPORT_SYMBOL(zfs_read);
-EXPORT_SYMBOL(zfs_write);
 EXPORT_SYMBOL(zfs_access);
-EXPORT_SYMBOL(zfs_lookup);
-EXPORT_SYMBOL(zfs_create);
-EXPORT_SYMBOL(zfs_tmpfile);
-EXPORT_SYMBOL(zfs_remove);
-EXPORT_SYMBOL(zfs_mkdir);
-EXPORT_SYMBOL(zfs_rmdir);
-EXPORT_SYMBOL(zfs_readdir);
 EXPORT_SYMBOL(zfs_fsync);
-EXPORT_SYMBOL(zfs_getattr);
-EXPORT_SYMBOL(zfs_getattr_fast);
-EXPORT_SYMBOL(zfs_setattr);
-EXPORT_SYMBOL(zfs_rename);
-EXPORT_SYMBOL(zfs_symlink);
-EXPORT_SYMBOL(zfs_readlink);
-EXPORT_SYMBOL(zfs_link);
-EXPORT_SYMBOL(zfs_inactive);
-EXPORT_SYMBOL(zfs_space);
-EXPORT_SYMBOL(zfs_fid);
+EXPORT_SYMBOL(zfs_holey);
+EXPORT_SYMBOL(zfs_read);
+EXPORT_SYMBOL(zfs_write);
 EXPORT_SYMBOL(zfs_getsecattr);
 EXPORT_SYMBOL(zfs_setsecattr);
-EXPORT_SYMBOL(zfs_getpage);
-EXPORT_SYMBOL(zfs_putpage);
-EXPORT_SYMBOL(zfs_dirty_inode);
-EXPORT_SYMBOL(zfs_map);
-
-/* CSTYLED */
-module_param(zfs_delete_blocks, ulong, 0644);
-MODULE_PARM_DESC(zfs_delete_blocks, "Delete files larger than N blocks async");
-module_param(zfs_read_chunk_size, long, 0644);
-MODULE_PARM_DESC(zfs_read_chunk_size, "Bytes to read per chunk");
-#endif
+EXPORT_SYMBOL(zfs_clone_range);
+EXPORT_SYMBOL(zfs_clone_range_replay);
+
+ZFS_MODULE_PARAM(zfs_vnops, zfs_vnops_, read_chunk_size, U64, ZMOD_RW,
+       "Bytes to read per chunk");