]> git.proxmox.com Git - mirror_zfs.git/blame - zfs/lib/libzpool/zfs_znode.c
Remove stray stub kernel files which should be brought in my linux-kernel-module...
[mirror_zfs.git] / zfs / lib / libzpool / zfs_znode.c
CommitLineData
34dc7c2f
BB
1/*
2 * CDDL HEADER START
3 *
4 * The contents of this file are subject to the terms of the
5 * Common Development and Distribution License (the "License").
6 * You may not use this file except in compliance with the License.
7 *
8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9 * or http://www.opensolaris.org/os/licensing.
10 * See the License for the specific language governing permissions
11 * and limitations under the License.
12 *
13 * When distributing Covered Code, include this CDDL HEADER in each
14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15 * If applicable, add the following below this CDDL HEADER, with the
16 * fields enclosed by brackets "[]" replaced with your own identifying
17 * information: Portions Copyright [yyyy] [name of copyright owner]
18 *
19 * CDDL HEADER END
20 */
21/*
22 * Copyright 2008 Sun Microsystems, Inc. All rights reserved.
23 * Use is subject to license terms.
24 */
25
26/* Portions Copyright 2007 Jeremy Teo */
27
28#pragma ident "@(#)zfs_znode.c 1.34 08/04/27 SMI"
29
30#ifdef _KERNEL
31#include <sys/types.h>
32#include <sys/param.h>
33#include <sys/time.h>
34#include <sys/systm.h>
35#include <sys/sysmacros.h>
36#include <sys/resource.h>
37#include <sys/mntent.h>
38#include <sys/mkdev.h>
39#include <sys/u8_textprep.h>
40#include <sys/dsl_dataset.h>
41#include <sys/vfs.h>
42#include <sys/vfs_opreg.h>
43#include <sys/vnode.h>
44#include <sys/file.h>
45#include <sys/kmem.h>
46#include <sys/errno.h>
47#include <sys/unistd.h>
48#include <sys/mode.h>
49#include <sys/atomic.h>
50#include <vm/pvn.h>
51#include "fs/fs_subr.h"
52#include <sys/zfs_dir.h>
53#include <sys/zfs_acl.h>
54#include <sys/zfs_ioctl.h>
55#include <sys/zfs_rlock.h>
56#include <sys/zfs_fuid.h>
57#include <sys/fs/zfs.h>
58#include <sys/kidmap.h>
59#endif /* _KERNEL */
60
61#include <sys/dmu.h>
62#include <sys/refcount.h>
63#include <sys/stat.h>
64#include <sys/zap.h>
65#include <sys/zfs_znode.h>
66
67#include "zfs_prop.h"
68
69/*
70 * Functions needed for userland (ie: libzpool) are not put under
71 * #ifdef_KERNEL; the rest of the functions have dependencies
72 * (such as VFS logic) that will not compile easily in userland.
73 */
74#ifdef _KERNEL
75struct kmem_cache *znode_cache = NULL;
76
77/*ARGSUSED*/
78static void
79znode_evict_error(dmu_buf_t *dbuf, void *user_ptr)
80{
81 /*
82 * We should never drop all dbuf refs without first clearing
83 * the eviction callback.
84 */
85 panic("evicting znode %p\n", user_ptr);
86}
87
88/*ARGSUSED*/
89static int
90zfs_znode_cache_constructor(void *buf, void *cdrarg, int kmflags)
91{
92 znode_t *zp = buf;
93
94 zp->z_vnode = vn_alloc(KM_SLEEP);
95 zp->z_vnode->v_data = (caddr_t)zp;
96 mutex_init(&zp->z_lock, NULL, MUTEX_DEFAULT, NULL);
97 rw_init(&zp->z_map_lock, NULL, RW_DEFAULT, NULL);
98 rw_init(&zp->z_parent_lock, NULL, RW_DEFAULT, NULL);
99 rw_init(&zp->z_name_lock, NULL, RW_DEFAULT, NULL);
100 mutex_init(&zp->z_acl_lock, NULL, MUTEX_DEFAULT, NULL);
101
102 mutex_init(&zp->z_range_lock, NULL, MUTEX_DEFAULT, NULL);
103 avl_create(&zp->z_range_avl, zfs_range_compare,
104 sizeof (rl_t), offsetof(rl_t, r_node));
105
106 zp->z_dbuf = NULL;
107 zp->z_dirlocks = 0;
108 return (0);
109}
110
111/*ARGSUSED*/
112static void
113zfs_znode_cache_destructor(void *buf, void *cdarg)
114{
115 znode_t *zp = buf;
116
117 ASSERT(zp->z_dirlocks == 0);
118 mutex_destroy(&zp->z_lock);
119 rw_destroy(&zp->z_map_lock);
120 rw_destroy(&zp->z_parent_lock);
121 rw_destroy(&zp->z_name_lock);
122 mutex_destroy(&zp->z_acl_lock);
123 avl_destroy(&zp->z_range_avl);
124 mutex_destroy(&zp->z_range_lock);
125
126 ASSERT(zp->z_dbuf == NULL);
127 ASSERT(ZTOV(zp)->v_count == 0);
128 vn_free(ZTOV(zp));
129}
130
131void
132zfs_znode_init(void)
133{
134 /*
135 * Initialize zcache
136 */
137 ASSERT(znode_cache == NULL);
138 znode_cache = kmem_cache_create("zfs_znode_cache",
139 sizeof (znode_t), 0, zfs_znode_cache_constructor,
140 zfs_znode_cache_destructor, NULL, NULL, NULL, 0);
141}
142
143void
144zfs_znode_fini(void)
145{
146 /*
147 * Cleanup vfs & vnode ops
148 */
149 zfs_remove_op_tables();
150
151 /*
152 * Cleanup zcache
153 */
154 if (znode_cache)
155 kmem_cache_destroy(znode_cache);
156 znode_cache = NULL;
157}
158
159struct vnodeops *zfs_dvnodeops;
160struct vnodeops *zfs_fvnodeops;
161struct vnodeops *zfs_symvnodeops;
162struct vnodeops *zfs_xdvnodeops;
163struct vnodeops *zfs_evnodeops;
164
165void
166zfs_remove_op_tables()
167{
168 /*
169 * Remove vfs ops
170 */
171 ASSERT(zfsfstype);
172 (void) vfs_freevfsops_by_type(zfsfstype);
173 zfsfstype = 0;
174
175 /*
176 * Remove vnode ops
177 */
178 if (zfs_dvnodeops)
179 vn_freevnodeops(zfs_dvnodeops);
180 if (zfs_fvnodeops)
181 vn_freevnodeops(zfs_fvnodeops);
182 if (zfs_symvnodeops)
183 vn_freevnodeops(zfs_symvnodeops);
184 if (zfs_xdvnodeops)
185 vn_freevnodeops(zfs_xdvnodeops);
186 if (zfs_evnodeops)
187 vn_freevnodeops(zfs_evnodeops);
188
189 zfs_dvnodeops = NULL;
190 zfs_fvnodeops = NULL;
191 zfs_symvnodeops = NULL;
192 zfs_xdvnodeops = NULL;
193 zfs_evnodeops = NULL;
194}
195
196extern const fs_operation_def_t zfs_dvnodeops_template[];
197extern const fs_operation_def_t zfs_fvnodeops_template[];
198extern const fs_operation_def_t zfs_xdvnodeops_template[];
199extern const fs_operation_def_t zfs_symvnodeops_template[];
200extern const fs_operation_def_t zfs_evnodeops_template[];
201
202int
203zfs_create_op_tables()
204{
205 int error;
206
207 /*
208 * zfs_dvnodeops can be set if mod_remove() calls mod_installfs()
209 * due to a failure to remove the the 2nd modlinkage (zfs_modldrv).
210 * In this case we just return as the ops vectors are already set up.
211 */
212 if (zfs_dvnodeops)
213 return (0);
214
215 error = vn_make_ops(MNTTYPE_ZFS, zfs_dvnodeops_template,
216 &zfs_dvnodeops);
217 if (error)
218 return (error);
219
220 error = vn_make_ops(MNTTYPE_ZFS, zfs_fvnodeops_template,
221 &zfs_fvnodeops);
222 if (error)
223 return (error);
224
225 error = vn_make_ops(MNTTYPE_ZFS, zfs_symvnodeops_template,
226 &zfs_symvnodeops);
227 if (error)
228 return (error);
229
230 error = vn_make_ops(MNTTYPE_ZFS, zfs_xdvnodeops_template,
231 &zfs_xdvnodeops);
232 if (error)
233 return (error);
234
235 error = vn_make_ops(MNTTYPE_ZFS, zfs_evnodeops_template,
236 &zfs_evnodeops);
237
238 return (error);
239}
240
241/*
242 * zfs_init_fs - Initialize the zfsvfs struct and the file system
243 * incore "master" object. Verify version compatibility.
244 */
245int
246zfs_init_fs(zfsvfs_t *zfsvfs, znode_t **zpp, cred_t *cr)
247{
248 extern int zfsfstype;
249
250 objset_t *os = zfsvfs->z_os;
251 int i, error;
252 dmu_object_info_t doi;
253 uint64_t fsid_guid;
254 uint64_t zval;
255
256 *zpp = NULL;
257
258 /*
259 * XXX - hack to auto-create the pool root filesystem at
260 * the first attempted mount.
261 */
262 if (dmu_object_info(os, MASTER_NODE_OBJ, &doi) == ENOENT) {
263 dmu_tx_t *tx = dmu_tx_create(os);
264 uint64_t zpl_version;
265 nvlist_t *zprops;
266
267 dmu_tx_hold_zap(tx, DMU_NEW_OBJECT, TRUE, NULL); /* master */
268 dmu_tx_hold_zap(tx, DMU_NEW_OBJECT, TRUE, NULL); /* del queue */
269 dmu_tx_hold_bonus(tx, DMU_NEW_OBJECT); /* root node */
270 error = dmu_tx_assign(tx, TXG_WAIT);
271 ASSERT3U(error, ==, 0);
272 if (spa_version(dmu_objset_spa(os)) >= SPA_VERSION_FUID)
273 zpl_version = ZPL_VERSION;
274 else
275 zpl_version = ZPL_VERSION_FUID - 1;
276
277 VERIFY(nvlist_alloc(&zprops, NV_UNIQUE_NAME, KM_SLEEP) == 0);
278 VERIFY(nvlist_add_uint64(zprops,
279 zfs_prop_to_name(ZFS_PROP_VERSION), zpl_version) == 0);
280 zfs_create_fs(os, cr, zprops, tx);
281 nvlist_free(zprops);
282 dmu_tx_commit(tx);
283 }
284
285 error = zfs_get_zplprop(os, ZFS_PROP_VERSION, &zfsvfs->z_version);
286 if (error) {
287 return (error);
288 } else if (zfsvfs->z_version > ZPL_VERSION) {
289 (void) printf("Mismatched versions: File system "
290 "is version %llu on-disk format, which is "
291 "incompatible with this software version %lld!",
292 (u_longlong_t)zfsvfs->z_version, ZPL_VERSION);
293 return (ENOTSUP);
294 }
295
296 if ((error = zfs_get_zplprop(os, ZFS_PROP_NORMALIZE, &zval)) != 0)
297 return (error);
298 zfsvfs->z_norm = (int)zval;
299 if ((error = zfs_get_zplprop(os, ZFS_PROP_UTF8ONLY, &zval)) != 0)
300 return (error);
301 zfsvfs->z_utf8 = (zval != 0);
302 if ((error = zfs_get_zplprop(os, ZFS_PROP_CASE, &zval)) != 0)
303 return (error);
304 zfsvfs->z_case = (uint_t)zval;
305 /*
306 * Fold case on file systems that are always or sometimes case
307 * insensitive.
308 */
309 if (zfsvfs->z_case == ZFS_CASE_INSENSITIVE ||
310 zfsvfs->z_case == ZFS_CASE_MIXED)
311 zfsvfs->z_norm |= U8_TEXTPREP_TOUPPER;
312
313 /*
314 * The fsid is 64 bits, composed of an 8-bit fs type, which
315 * separates our fsid from any other filesystem types, and a
316 * 56-bit objset unique ID. The objset unique ID is unique to
317 * all objsets open on this system, provided by unique_create().
318 * The 8-bit fs type must be put in the low bits of fsid[1]
319 * because that's where other Solaris filesystems put it.
320 */
321 fsid_guid = dmu_objset_fsid_guid(os);
322 ASSERT((fsid_guid & ~((1ULL<<56)-1)) == 0);
323 zfsvfs->z_vfs->vfs_fsid.val[0] = fsid_guid;
324 zfsvfs->z_vfs->vfs_fsid.val[1] = ((fsid_guid>>32) << 8) |
325 zfsfstype & 0xFF;
326
327 error = zap_lookup(os, MASTER_NODE_OBJ, ZFS_ROOT_OBJ, 8, 1,
328 &zfsvfs->z_root);
329 if (error)
330 return (error);
331 ASSERT(zfsvfs->z_root != 0);
332
333 error = zap_lookup(os, MASTER_NODE_OBJ, ZFS_UNLINKED_SET, 8, 1,
334 &zfsvfs->z_unlinkedobj);
335 if (error)
336 return (error);
337
338 /*
339 * Initialize zget mutex's
340 */
341 for (i = 0; i != ZFS_OBJ_MTX_SZ; i++)
342 mutex_init(&zfsvfs->z_hold_mtx[i], NULL, MUTEX_DEFAULT, NULL);
343
344 error = zfs_zget(zfsvfs, zfsvfs->z_root, zpp);
345 if (error) {
346 /*
347 * On error, we destroy the mutexes here since it's not
348 * possible for the caller to determine if the mutexes were
349 * initialized properly.
350 */
351 for (i = 0; i != ZFS_OBJ_MTX_SZ; i++)
352 mutex_destroy(&zfsvfs->z_hold_mtx[i]);
353 return (error);
354 }
355 ASSERT3U((*zpp)->z_id, ==, zfsvfs->z_root);
356 error = zap_lookup(os, MASTER_NODE_OBJ, ZFS_FUID_TABLES, 8, 1,
357 &zfsvfs->z_fuid_obj);
358 if (error == ENOENT)
359 error = 0;
360
361 return (0);
362}
363
364/*
365 * define a couple of values we need available
366 * for both 64 and 32 bit environments.
367 */
368#ifndef NBITSMINOR64
369#define NBITSMINOR64 32
370#endif
371#ifndef MAXMAJ64
372#define MAXMAJ64 0xffffffffUL
373#endif
374#ifndef MAXMIN64
375#define MAXMIN64 0xffffffffUL
376#endif
377
378/*
379 * Create special expldev for ZFS private use.
380 * Can't use standard expldev since it doesn't do
381 * what we want. The standard expldev() takes a
382 * dev32_t in LP64 and expands it to a long dev_t.
383 * We need an interface that takes a dev32_t in ILP32
384 * and expands it to a long dev_t.
385 */
386static uint64_t
387zfs_expldev(dev_t dev)
388{
389#ifndef _LP64
390 major_t major = (major_t)dev >> NBITSMINOR32 & MAXMAJ32;
391 return (((uint64_t)major << NBITSMINOR64) |
392 ((minor_t)dev & MAXMIN32));
393#else
394 return (dev);
395#endif
396}
397
398/*
399 * Special cmpldev for ZFS private use.
400 * Can't use standard cmpldev since it takes
401 * a long dev_t and compresses it to dev32_t in
402 * LP64. We need to do a compaction of a long dev_t
403 * to a dev32_t in ILP32.
404 */
405dev_t
406zfs_cmpldev(uint64_t dev)
407{
408#ifndef _LP64
409 minor_t minor = (minor_t)dev & MAXMIN64;
410 major_t major = (major_t)(dev >> NBITSMINOR64) & MAXMAJ64;
411
412 if (major > MAXMAJ32 || minor > MAXMIN32)
413 return (NODEV32);
414
415 return (((dev32_t)major << NBITSMINOR32) | minor);
416#else
417 return (dev);
418#endif
419}
420
421static void
422zfs_znode_dmu_init(znode_t *zp, dmu_buf_t *db)
423{
424 znode_t *nzp;
425 zfsvfs_t *zfsvfs = zp->z_zfsvfs;
426
427 ASSERT(MUTEX_HELD(ZFS_OBJ_MUTEX(zp)));
428
429 mutex_enter(&zp->z_lock);
430
431 ASSERT(zp->z_dbuf == NULL);
432 zp->z_dbuf = db;
433 nzp = dmu_buf_set_user_ie(db, zp, &zp->z_phys, znode_evict_error);
434
435 /*
436 * there should be no
437 * concurrent zgets on this object.
438 */
439 if (nzp != NULL)
440 panic("existing znode %p for dbuf %p", nzp, db);
441
442 /*
443 * Slap on VROOT if we are the root znode
444 */
445 if (zp->z_id == zfsvfs->z_root)
446 ZTOV(zp)->v_flag |= VROOT;
447
448 mutex_exit(&zp->z_lock);
449 vn_exists(ZTOV(zp));
450}
451
452void
453zfs_znode_dmu_fini(znode_t *zp)
454{
455 dmu_buf_t *db = zp->z_dbuf;
456 ASSERT(MUTEX_HELD(ZFS_OBJ_MUTEX(zp)) || zp->z_unlinked ||
457 RW_WRITE_HELD(&zp->z_zfsvfs->z_teardown_inactive_lock));
458 ASSERT(zp->z_dbuf != NULL);
459 zp->z_dbuf = NULL;
460 VERIFY(zp == dmu_buf_update_user(db, zp, NULL, NULL, NULL));
461 dmu_buf_rele(db, NULL);
462}
463
464/*
465 * Construct a new znode/vnode and intialize.
466 *
467 * This does not do a call to dmu_set_user() that is
468 * up to the caller to do, in case you don't want to
469 * return the znode
470 */
471static znode_t *
472zfs_znode_alloc(zfsvfs_t *zfsvfs, dmu_buf_t *db, int blksz)
473{
474 znode_t *zp;
475 vnode_t *vp;
476
477 zp = kmem_cache_alloc(znode_cache, KM_SLEEP);
478
479 ASSERT(zp->z_dirlocks == NULL);
480 ASSERT(zp->z_dbuf == NULL);
481
482 zp->z_phys = NULL;
483 zp->z_zfsvfs = zfsvfs;
484 zp->z_unlinked = 0;
485 zp->z_atime_dirty = 0;
486 zp->z_mapcnt = 0;
487 zp->z_last_itx = 0;
488 zp->z_id = db->db_object;
489 zp->z_blksz = blksz;
490 zp->z_seq = 0x7A4653;
491 zp->z_sync_cnt = 0;
492
493 vp = ZTOV(zp);
494 vn_reinit(vp);
495
496 zfs_znode_dmu_init(zp, db);
497
498 zp->z_gen = zp->z_phys->zp_gen;
499
500 mutex_enter(&zfsvfs->z_znodes_lock);
501 list_insert_tail(&zfsvfs->z_all_znodes, zp);
502 mutex_exit(&zfsvfs->z_znodes_lock);
503
504 vp->v_vfsp = zfsvfs->z_parent->z_vfs;
505 vp->v_type = IFTOVT((mode_t)zp->z_phys->zp_mode);
506
507 switch (vp->v_type) {
508 case VDIR:
509 if (zp->z_phys->zp_flags & ZFS_XATTR) {
510 vn_setops(vp, zfs_xdvnodeops);
511 vp->v_flag |= V_XATTRDIR;
512 } else {
513 vn_setops(vp, zfs_dvnodeops);
514 }
515 zp->z_zn_prefetch = B_TRUE; /* z_prefetch default is enabled */
516 break;
517 case VBLK:
518 case VCHR:
519 vp->v_rdev = zfs_cmpldev(zp->z_phys->zp_rdev);
520 /*FALLTHROUGH*/
521 case VFIFO:
522 case VSOCK:
523 case VDOOR:
524 vn_setops(vp, zfs_fvnodeops);
525 break;
526 case VREG:
527 vp->v_flag |= VMODSORT;
528 vn_setops(vp, zfs_fvnodeops);
529 break;
530 case VLNK:
531 vn_setops(vp, zfs_symvnodeops);
532 break;
533 default:
534 vn_setops(vp, zfs_evnodeops);
535 break;
536 }
537
538 VFS_HOLD(zfsvfs->z_vfs);
539 return (zp);
540}
541
542/*
543 * Create a new DMU object to hold a zfs znode.
544 *
545 * IN: dzp - parent directory for new znode
546 * vap - file attributes for new znode
547 * tx - dmu transaction id for zap operations
548 * cr - credentials of caller
549 * flag - flags:
550 * IS_ROOT_NODE - new object will be root
551 * IS_XATTR - new object is an attribute
552 * IS_REPLAY - intent log replay
553 * bonuslen - length of bonus buffer
554 * setaclp - File/Dir initial ACL
555 * fuidp - Tracks fuid allocation.
556 *
557 * OUT: zpp - allocated znode
558 *
559 */
560void
561zfs_mknode(znode_t *dzp, vattr_t *vap, dmu_tx_t *tx, cred_t *cr,
562 uint_t flag, znode_t **zpp, int bonuslen, zfs_acl_t *setaclp,
563 zfs_fuid_info_t **fuidp)
564{
565 dmu_buf_t *db;
566 znode_phys_t *pzp;
567 zfsvfs_t *zfsvfs = dzp->z_zfsvfs;
568 timestruc_t now;
569 uint64_t gen, obj;
570 int err;
571
572 ASSERT(vap && (vap->va_mask & (AT_TYPE|AT_MODE)) == (AT_TYPE|AT_MODE));
573
574 if (zfsvfs->z_assign >= TXG_INITIAL) { /* ZIL replay */
575 obj = vap->va_nodeid;
576 flag |= IS_REPLAY;
577 now = vap->va_ctime; /* see zfs_replay_create() */
578 gen = vap->va_nblocks; /* ditto */
579 } else {
580 obj = 0;
581 gethrestime(&now);
582 gen = dmu_tx_get_txg(tx);
583 }
584
585 /*
586 * Create a new DMU object.
587 */
588 /*
589 * There's currently no mechanism for pre-reading the blocks that will
590 * be to needed allocate a new object, so we accept the small chance
591 * that there will be an i/o error and we will fail one of the
592 * assertions below.
593 */
594 if (vap->va_type == VDIR) {
595 if (flag & IS_REPLAY) {
596 err = zap_create_claim_norm(zfsvfs->z_os, obj,
597 zfsvfs->z_norm, DMU_OT_DIRECTORY_CONTENTS,
598 DMU_OT_ZNODE, sizeof (znode_phys_t) + bonuslen, tx);
599 ASSERT3U(err, ==, 0);
600 } else {
601 obj = zap_create_norm(zfsvfs->z_os,
602 zfsvfs->z_norm, DMU_OT_DIRECTORY_CONTENTS,
603 DMU_OT_ZNODE, sizeof (znode_phys_t) + bonuslen, tx);
604 }
605 } else {
606 if (flag & IS_REPLAY) {
607 err = dmu_object_claim(zfsvfs->z_os, obj,
608 DMU_OT_PLAIN_FILE_CONTENTS, 0,
609 DMU_OT_ZNODE, sizeof (znode_phys_t) + bonuslen, tx);
610 ASSERT3U(err, ==, 0);
611 } else {
612 obj = dmu_object_alloc(zfsvfs->z_os,
613 DMU_OT_PLAIN_FILE_CONTENTS, 0,
614 DMU_OT_ZNODE, sizeof (znode_phys_t) + bonuslen, tx);
615 }
616 }
617 VERIFY(0 == dmu_bonus_hold(zfsvfs->z_os, obj, NULL, &db));
618 dmu_buf_will_dirty(db, tx);
619
620 /*
621 * Initialize the znode physical data to zero.
622 */
623 ASSERT(db->db_size >= sizeof (znode_phys_t));
624 bzero(db->db_data, db->db_size);
625 pzp = db->db_data;
626
627 /*
628 * If this is the root, fix up the half-initialized parent pointer
629 * to reference the just-allocated physical data area.
630 */
631 if (flag & IS_ROOT_NODE) {
632 dzp->z_dbuf = db;
633 dzp->z_phys = pzp;
634 dzp->z_id = obj;
635 }
636
637 /*
638 * If parent is an xattr, so am I.
639 */
640 if (dzp->z_phys->zp_flags & ZFS_XATTR)
641 flag |= IS_XATTR;
642
643 if (vap->va_type == VBLK || vap->va_type == VCHR) {
644 pzp->zp_rdev = zfs_expldev(vap->va_rdev);
645 }
646
647 if (zfsvfs->z_use_fuids)
648 pzp->zp_flags = ZFS_ARCHIVE | ZFS_AV_MODIFIED;
649
650 if (vap->va_type == VDIR) {
651 pzp->zp_size = 2; /* contents ("." and "..") */
652 pzp->zp_links = (flag & (IS_ROOT_NODE | IS_XATTR)) ? 2 : 1;
653 }
654
655 pzp->zp_parent = dzp->z_id;
656 if (flag & IS_XATTR)
657 pzp->zp_flags |= ZFS_XATTR;
658
659 pzp->zp_gen = gen;
660
661 ZFS_TIME_ENCODE(&now, pzp->zp_crtime);
662 ZFS_TIME_ENCODE(&now, pzp->zp_ctime);
663
664 if (vap->va_mask & AT_ATIME) {
665 ZFS_TIME_ENCODE(&vap->va_atime, pzp->zp_atime);
666 } else {
667 ZFS_TIME_ENCODE(&now, pzp->zp_atime);
668 }
669
670 if (vap->va_mask & AT_MTIME) {
671 ZFS_TIME_ENCODE(&vap->va_mtime, pzp->zp_mtime);
672 } else {
673 ZFS_TIME_ENCODE(&now, pzp->zp_mtime);
674 }
675
676 pzp->zp_mode = MAKEIMODE(vap->va_type, vap->va_mode);
677 if (!(flag & IS_ROOT_NODE)) {
678 ZFS_OBJ_HOLD_ENTER(zfsvfs, obj)
679 *zpp = zfs_znode_alloc(zfsvfs, db, 0);
680 ZFS_OBJ_HOLD_EXIT(zfsvfs, obj);
681 } else {
682 /*
683 * If we are creating the root node, the "parent" we
684 * passed in is the znode for the root.
685 */
686 *zpp = dzp;
687 }
688 zfs_perm_init(*zpp, dzp, flag, vap, tx, cr, setaclp, fuidp);
689}
690
691void
692zfs_xvattr_set(znode_t *zp, xvattr_t *xvap)
693{
694 xoptattr_t *xoap;
695
696 xoap = xva_getxoptattr(xvap);
697 ASSERT(xoap);
698
699 if (XVA_ISSET_REQ(xvap, XAT_CREATETIME)) {
700 ZFS_TIME_ENCODE(&xoap->xoa_createtime, zp->z_phys->zp_crtime);
701 XVA_SET_RTN(xvap, XAT_CREATETIME);
702 }
703 if (XVA_ISSET_REQ(xvap, XAT_READONLY)) {
704 ZFS_ATTR_SET(zp, ZFS_READONLY, xoap->xoa_readonly);
705 XVA_SET_RTN(xvap, XAT_READONLY);
706 }
707 if (XVA_ISSET_REQ(xvap, XAT_HIDDEN)) {
708 ZFS_ATTR_SET(zp, ZFS_HIDDEN, xoap->xoa_hidden);
709 XVA_SET_RTN(xvap, XAT_HIDDEN);
710 }
711 if (XVA_ISSET_REQ(xvap, XAT_SYSTEM)) {
712 ZFS_ATTR_SET(zp, ZFS_SYSTEM, xoap->xoa_system);
713 XVA_SET_RTN(xvap, XAT_SYSTEM);
714 }
715 if (XVA_ISSET_REQ(xvap, XAT_ARCHIVE)) {
716 ZFS_ATTR_SET(zp, ZFS_ARCHIVE, xoap->xoa_archive);
717 XVA_SET_RTN(xvap, XAT_ARCHIVE);
718 }
719 if (XVA_ISSET_REQ(xvap, XAT_IMMUTABLE)) {
720 ZFS_ATTR_SET(zp, ZFS_IMMUTABLE, xoap->xoa_immutable);
721 XVA_SET_RTN(xvap, XAT_IMMUTABLE);
722 }
723 if (XVA_ISSET_REQ(xvap, XAT_NOUNLINK)) {
724 ZFS_ATTR_SET(zp, ZFS_NOUNLINK, xoap->xoa_nounlink);
725 XVA_SET_RTN(xvap, XAT_NOUNLINK);
726 }
727 if (XVA_ISSET_REQ(xvap, XAT_APPENDONLY)) {
728 ZFS_ATTR_SET(zp, ZFS_APPENDONLY, xoap->xoa_appendonly);
729 XVA_SET_RTN(xvap, XAT_APPENDONLY);
730 }
731 if (XVA_ISSET_REQ(xvap, XAT_NODUMP)) {
732 ZFS_ATTR_SET(zp, ZFS_NODUMP, xoap->xoa_nodump);
733 XVA_SET_RTN(xvap, XAT_NODUMP);
734 }
735 if (XVA_ISSET_REQ(xvap, XAT_OPAQUE)) {
736 ZFS_ATTR_SET(zp, ZFS_OPAQUE, xoap->xoa_opaque);
737 XVA_SET_RTN(xvap, XAT_OPAQUE);
738 }
739 if (XVA_ISSET_REQ(xvap, XAT_AV_QUARANTINED)) {
740 ZFS_ATTR_SET(zp, ZFS_AV_QUARANTINED,
741 xoap->xoa_av_quarantined);
742 XVA_SET_RTN(xvap, XAT_AV_QUARANTINED);
743 }
744 if (XVA_ISSET_REQ(xvap, XAT_AV_MODIFIED)) {
745 ZFS_ATTR_SET(zp, ZFS_AV_MODIFIED, xoap->xoa_av_modified);
746 XVA_SET_RTN(xvap, XAT_AV_MODIFIED);
747 }
748 if (XVA_ISSET_REQ(xvap, XAT_AV_SCANSTAMP)) {
749 (void) memcpy(zp->z_phys + 1, xoap->xoa_av_scanstamp,
750 sizeof (xoap->xoa_av_scanstamp));
751 zp->z_phys->zp_flags |= ZFS_BONUS_SCANSTAMP;
752 XVA_SET_RTN(xvap, XAT_AV_SCANSTAMP);
753 }
754}
755
756int
757zfs_zget(zfsvfs_t *zfsvfs, uint64_t obj_num, znode_t **zpp)
758{
759 dmu_object_info_t doi;
760 dmu_buf_t *db;
761 znode_t *zp;
762 int err;
763
764 *zpp = NULL;
765
766 ZFS_OBJ_HOLD_ENTER(zfsvfs, obj_num);
767
768 err = dmu_bonus_hold(zfsvfs->z_os, obj_num, NULL, &db);
769 if (err) {
770 ZFS_OBJ_HOLD_EXIT(zfsvfs, obj_num);
771 return (err);
772 }
773
774 dmu_object_info_from_db(db, &doi);
775 if (doi.doi_bonus_type != DMU_OT_ZNODE ||
776 doi.doi_bonus_size < sizeof (znode_phys_t)) {
777 dmu_buf_rele(db, NULL);
778 ZFS_OBJ_HOLD_EXIT(zfsvfs, obj_num);
779 return (EINVAL);
780 }
781
782 zp = dmu_buf_get_user(db);
783 if (zp != NULL) {
784 mutex_enter(&zp->z_lock);
785
786 /*
787 * Since we do immediate eviction of the z_dbuf, we
788 * should never find a dbuf with a znode that doesn't
789 * know about the dbuf.
790 */
791 ASSERT3P(zp->z_dbuf, ==, db);
792 ASSERT3U(zp->z_id, ==, obj_num);
793 if (zp->z_unlinked) {
794 err = ENOENT;
795 } else {
796 VN_HOLD(ZTOV(zp));
797 *zpp = zp;
798 err = 0;
799 }
800 dmu_buf_rele(db, NULL);
801 mutex_exit(&zp->z_lock);
802 ZFS_OBJ_HOLD_EXIT(zfsvfs, obj_num);
803 return (err);
804 }
805
806 /*
807 * Not found create new znode/vnode
808 */
809 zp = zfs_znode_alloc(zfsvfs, db, doi.doi_data_block_size);
810 ZFS_OBJ_HOLD_EXIT(zfsvfs, obj_num);
811 *zpp = zp;
812 return (0);
813}
814
815int
816zfs_rezget(znode_t *zp)
817{
818 zfsvfs_t *zfsvfs = zp->z_zfsvfs;
819 dmu_object_info_t doi;
820 dmu_buf_t *db;
821 uint64_t obj_num = zp->z_id;
822 int err;
823
824 ZFS_OBJ_HOLD_ENTER(zfsvfs, obj_num);
825
826 err = dmu_bonus_hold(zfsvfs->z_os, obj_num, NULL, &db);
827 if (err) {
828 ZFS_OBJ_HOLD_EXIT(zfsvfs, obj_num);
829 return (err);
830 }
831
832 dmu_object_info_from_db(db, &doi);
833 if (doi.doi_bonus_type != DMU_OT_ZNODE ||
834 doi.doi_bonus_size < sizeof (znode_phys_t)) {
835 dmu_buf_rele(db, NULL);
836 ZFS_OBJ_HOLD_EXIT(zfsvfs, obj_num);
837 return (EINVAL);
838 }
839
840 if (((znode_phys_t *)db->db_data)->zp_gen != zp->z_gen) {
841 dmu_buf_rele(db, NULL);
842 ZFS_OBJ_HOLD_EXIT(zfsvfs, obj_num);
843 return (EIO);
844 }
845
846 zfs_znode_dmu_init(zp, db);
847 zp->z_unlinked = (zp->z_phys->zp_links == 0);
848 zp->z_blksz = doi.doi_data_block_size;
849
850 ZFS_OBJ_HOLD_EXIT(zfsvfs, obj_num);
851
852 return (0);
853}
854
855void
856zfs_znode_delete(znode_t *zp, dmu_tx_t *tx)
857{
858 zfsvfs_t *zfsvfs = zp->z_zfsvfs;
859 uint64_t obj = zp->z_id;
860
861 ZFS_OBJ_HOLD_ENTER(zfsvfs, obj);
862 if (zp->z_phys->zp_acl.z_acl_extern_obj) {
863 VERIFY(0 == dmu_object_free(zfsvfs->z_os,
864 zp->z_phys->zp_acl.z_acl_extern_obj, tx));
865 }
866 VERIFY(0 == dmu_object_free(zfsvfs->z_os, obj, tx));
867 zfs_znode_dmu_fini(zp);
868 ZFS_OBJ_HOLD_EXIT(zfsvfs, obj);
869 zfs_znode_free(zp);
870}
871
872void
873zfs_zinactive(znode_t *zp)
874{
875 vnode_t *vp = ZTOV(zp);
876 zfsvfs_t *zfsvfs = zp->z_zfsvfs;
877 uint64_t z_id = zp->z_id;
878
879 ASSERT(zp->z_dbuf && zp->z_phys);
880
881 /*
882 * Don't allow a zfs_zget() while were trying to release this znode
883 */
884 ZFS_OBJ_HOLD_ENTER(zfsvfs, z_id);
885
886 mutex_enter(&zp->z_lock);
887 mutex_enter(&vp->v_lock);
888 vp->v_count--;
889 if (vp->v_count > 0 || vn_has_cached_data(vp)) {
890 /*
891 * If the hold count is greater than zero, somebody has
892 * obtained a new reference on this znode while we were
893 * processing it here, so we are done. If we still have
894 * mapped pages then we are also done, since we don't
895 * want to inactivate the znode until the pages get pushed.
896 *
897 * XXX - if vn_has_cached_data(vp) is true, but count == 0,
898 * this seems like it would leave the znode hanging with
899 * no chance to go inactive...
900 */
901 mutex_exit(&vp->v_lock);
902 mutex_exit(&zp->z_lock);
903 ZFS_OBJ_HOLD_EXIT(zfsvfs, z_id);
904 return;
905 }
906 mutex_exit(&vp->v_lock);
907
908 /*
909 * If this was the last reference to a file with no links,
910 * remove the file from the file system.
911 */
912 if (zp->z_unlinked) {
913 mutex_exit(&zp->z_lock);
914 ZFS_OBJ_HOLD_EXIT(zfsvfs, z_id);
915 zfs_rmnode(zp);
916 return;
917 }
918 mutex_exit(&zp->z_lock);
919 zfs_znode_dmu_fini(zp);
920 ZFS_OBJ_HOLD_EXIT(zfsvfs, z_id);
921 zfs_znode_free(zp);
922}
923
924void
925zfs_znode_free(znode_t *zp)
926{
927 zfsvfs_t *zfsvfs = zp->z_zfsvfs;
928
929 vn_invalid(ZTOV(zp));
930
931 mutex_enter(&zfsvfs->z_znodes_lock);
932 list_remove(&zfsvfs->z_all_znodes, zp);
933 mutex_exit(&zfsvfs->z_znodes_lock);
934
935 kmem_cache_free(znode_cache, zp);
936
937 VFS_RELE(zfsvfs->z_vfs);
938}
939
940void
941zfs_time_stamper_locked(znode_t *zp, uint_t flag, dmu_tx_t *tx)
942{
943 timestruc_t now;
944
945 ASSERT(MUTEX_HELD(&zp->z_lock));
946
947 gethrestime(&now);
948
949 if (tx) {
950 dmu_buf_will_dirty(zp->z_dbuf, tx);
951 zp->z_atime_dirty = 0;
952 zp->z_seq++;
953 } else {
954 zp->z_atime_dirty = 1;
955 }
956
957 if (flag & AT_ATIME)
958 ZFS_TIME_ENCODE(&now, zp->z_phys->zp_atime);
959
960 if (flag & AT_MTIME) {
961 ZFS_TIME_ENCODE(&now, zp->z_phys->zp_mtime);
962 if (zp->z_zfsvfs->z_use_fuids)
963 zp->z_phys->zp_flags |= (ZFS_ARCHIVE | ZFS_AV_MODIFIED);
964 }
965
966 if (flag & AT_CTIME) {
967 ZFS_TIME_ENCODE(&now, zp->z_phys->zp_ctime);
968 if (zp->z_zfsvfs->z_use_fuids)
969 zp->z_phys->zp_flags |= ZFS_ARCHIVE;
970 }
971}
972
973/*
974 * Update the requested znode timestamps with the current time.
975 * If we are in a transaction, then go ahead and mark the znode
976 * dirty in the transaction so the timestamps will go to disk.
977 * Otherwise, we will get pushed next time the znode is updated
978 * in a transaction, or when this znode eventually goes inactive.
979 *
980 * Why is this OK?
981 * 1 - Only the ACCESS time is ever updated outside of a transaction.
982 * 2 - Multiple consecutive updates will be collapsed into a single
983 * znode update by the transaction grouping semantics of the DMU.
984 */
985void
986zfs_time_stamper(znode_t *zp, uint_t flag, dmu_tx_t *tx)
987{
988 mutex_enter(&zp->z_lock);
989 zfs_time_stamper_locked(zp, flag, tx);
990 mutex_exit(&zp->z_lock);
991}
992
993/*
994 * Grow the block size for a file.
995 *
996 * IN: zp - znode of file to free data in.
997 * size - requested block size
998 * tx - open transaction.
999 *
1000 * NOTE: this function assumes that the znode is write locked.
1001 */
1002void
1003zfs_grow_blocksize(znode_t *zp, uint64_t size, dmu_tx_t *tx)
1004{
1005 int error;
1006 u_longlong_t dummy;
1007
1008 if (size <= zp->z_blksz)
1009 return;
1010 /*
1011 * If the file size is already greater than the current blocksize,
1012 * we will not grow. If there is more than one block in a file,
1013 * the blocksize cannot change.
1014 */
1015 if (zp->z_blksz && zp->z_phys->zp_size > zp->z_blksz)
1016 return;
1017
1018 error = dmu_object_set_blocksize(zp->z_zfsvfs->z_os, zp->z_id,
1019 size, 0, tx);
1020 if (error == ENOTSUP)
1021 return;
1022 ASSERT3U(error, ==, 0);
1023
1024 /* What blocksize did we actually get? */
1025 dmu_object_size_from_db(zp->z_dbuf, &zp->z_blksz, &dummy);
1026}
1027
1028/*
1029 * This is a dummy interface used when pvn_vplist_dirty() should *not*
1030 * be calling back into the fs for a putpage(). E.g.: when truncating
1031 * a file, the pages being "thrown away* don't need to be written out.
1032 */
1033/* ARGSUSED */
1034static int
1035zfs_no_putpage(vnode_t *vp, page_t *pp, u_offset_t *offp, size_t *lenp,
1036 int flags, cred_t *cr)
1037{
1038 ASSERT(0);
1039 return (0);
1040}
1041
1042/*
1043 * Free space in a file.
1044 *
1045 * IN: zp - znode of file to free data in.
1046 * off - start of section to free.
1047 * len - length of section to free (0 => to EOF).
1048 * flag - current file open mode flags.
1049 *
1050 * RETURN: 0 if success
1051 * error code if failure
1052 */
1053int
1054zfs_freesp(znode_t *zp, uint64_t off, uint64_t len, int flag, boolean_t log)
1055{
1056 vnode_t *vp = ZTOV(zp);
1057 dmu_tx_t *tx;
1058 zfsvfs_t *zfsvfs = zp->z_zfsvfs;
1059 zilog_t *zilog = zfsvfs->z_log;
1060 rl_t *rl;
1061 uint64_t end = off + len;
1062 uint64_t size, new_blksz;
1063 uint64_t pflags = zp->z_phys->zp_flags;
1064 int error;
1065
1066 if ((pflags & (ZFS_IMMUTABLE|ZFS_READONLY)) ||
1067 off < zp->z_phys->zp_size && (pflags & ZFS_APPENDONLY))
1068 return (EPERM);
1069
1070 if (ZTOV(zp)->v_type == VFIFO)
1071 return (0);
1072
1073 /*
1074 * If we will change zp_size then lock the whole file,
1075 * otherwise just lock the range being freed.
1076 */
1077 if (len == 0 || off + len > zp->z_phys->zp_size) {
1078 rl = zfs_range_lock(zp, 0, UINT64_MAX, RL_WRITER);
1079 } else {
1080 rl = zfs_range_lock(zp, off, len, RL_WRITER);
1081 /* recheck, in case zp_size changed */
1082 if (off + len > zp->z_phys->zp_size) {
1083 /* lost race: file size changed, lock whole file */
1084 zfs_range_unlock(rl);
1085 rl = zfs_range_lock(zp, 0, UINT64_MAX, RL_WRITER);
1086 }
1087 }
1088
1089 /*
1090 * Nothing to do if file already at desired length.
1091 */
1092 size = zp->z_phys->zp_size;
1093 if (len == 0 && size == off && off != 0) {
1094 zfs_range_unlock(rl);
1095 return (0);
1096 }
1097
1098 /*
1099 * Check for any locks in the region to be freed.
1100 */
1101 if (MANDLOCK(vp, (mode_t)zp->z_phys->zp_mode)) {
1102 uint64_t start = off;
1103 uint64_t extent = len;
1104
1105 if (off > size) {
1106 start = size;
1107 extent += off - size;
1108 } else if (len == 0) {
1109 extent = size - off;
1110 }
1111 if (error = chklock(vp, FWRITE, start, extent, flag, NULL)) {
1112 zfs_range_unlock(rl);
1113 return (error);
1114 }
1115 }
1116
1117 tx = dmu_tx_create(zfsvfs->z_os);
1118 dmu_tx_hold_bonus(tx, zp->z_id);
1119 new_blksz = 0;
1120 if (end > size &&
1121 (!ISP2(zp->z_blksz) || zp->z_blksz < zfsvfs->z_max_blksz)) {
1122 /*
1123 * We are growing the file past the current block size.
1124 */
1125 if (zp->z_blksz > zp->z_zfsvfs->z_max_blksz) {
1126 ASSERT(!ISP2(zp->z_blksz));
1127 new_blksz = MIN(end, SPA_MAXBLOCKSIZE);
1128 } else {
1129 new_blksz = MIN(end, zp->z_zfsvfs->z_max_blksz);
1130 }
1131 dmu_tx_hold_write(tx, zp->z_id, 0, MIN(end, new_blksz));
1132 } else if (off < size) {
1133 /*
1134 * If len == 0, we are truncating the file.
1135 */
1136 dmu_tx_hold_free(tx, zp->z_id, off, len ? len : DMU_OBJECT_END);
1137 }
1138
1139 error = dmu_tx_assign(tx, zfsvfs->z_assign);
1140 if (error) {
1141 if (error == ERESTART && zfsvfs->z_assign == TXG_NOWAIT)
1142 dmu_tx_wait(tx);
1143 dmu_tx_abort(tx);
1144 zfs_range_unlock(rl);
1145 return (error);
1146 }
1147
1148 if (new_blksz)
1149 zfs_grow_blocksize(zp, new_blksz, tx);
1150
1151 if (end > size || len == 0)
1152 zp->z_phys->zp_size = end;
1153
1154 if (off < size) {
1155 objset_t *os = zfsvfs->z_os;
1156 uint64_t rlen = len;
1157
1158 if (len == 0)
1159 rlen = -1;
1160 else if (end > size)
1161 rlen = size - off;
1162 VERIFY(0 == dmu_free_range(os, zp->z_id, off, rlen, tx));
1163 }
1164
1165 if (log) {
1166 zfs_time_stamper(zp, CONTENT_MODIFIED, tx);
1167 zfs_log_truncate(zilog, tx, TX_TRUNCATE, zp, off, len);
1168 }
1169
1170 zfs_range_unlock(rl);
1171
1172 dmu_tx_commit(tx);
1173
1174 /*
1175 * Clear any mapped pages in the truncated region. This has to
1176 * happen outside of the transaction to avoid the possibility of
1177 * a deadlock with someone trying to push a page that we are
1178 * about to invalidate.
1179 */
1180 rw_enter(&zp->z_map_lock, RW_WRITER);
1181 if (off < size && vn_has_cached_data(vp)) {
1182 page_t *pp;
1183 uint64_t start = off & PAGEMASK;
1184 int poff = off & PAGEOFFSET;
1185
1186 if (poff != 0 && (pp = page_lookup(vp, start, SE_SHARED))) {
1187 /*
1188 * We need to zero a partial page.
1189 */
1190 pagezero(pp, poff, PAGESIZE - poff);
1191 start += PAGESIZE;
1192 page_unlock(pp);
1193 }
1194 error = pvn_vplist_dirty(vp, start, zfs_no_putpage,
1195 B_INVAL | B_TRUNC, NULL);
1196 ASSERT(error == 0);
1197 }
1198 rw_exit(&zp->z_map_lock);
1199
1200 return (0);
1201}
1202
1203void
1204zfs_create_fs(objset_t *os, cred_t *cr, nvlist_t *zplprops, dmu_tx_t *tx)
1205{
1206 zfsvfs_t zfsvfs;
1207 uint64_t moid, doid;
1208 uint64_t version = 0;
1209 uint64_t sense = ZFS_CASE_SENSITIVE;
1210 uint64_t norm = 0;
1211 nvpair_t *elem;
1212 int error;
1213 znode_t *rootzp = NULL;
1214 vnode_t *vp;
1215 vattr_t vattr;
1216 znode_t *zp;
1217
1218 /*
1219 * First attempt to create master node.
1220 */
1221 /*
1222 * In an empty objset, there are no blocks to read and thus
1223 * there can be no i/o errors (which we assert below).
1224 */
1225 moid = MASTER_NODE_OBJ;
1226 error = zap_create_claim(os, moid, DMU_OT_MASTER_NODE,
1227 DMU_OT_NONE, 0, tx);
1228 ASSERT(error == 0);
1229
1230 /*
1231 * Set starting attributes.
1232 */
1233 elem = NULL;
1234 while ((elem = nvlist_next_nvpair(zplprops, elem)) != NULL) {
1235 /* For the moment we expect all zpl props to be uint64_ts */
1236 uint64_t val;
1237 char *name;
1238
1239 ASSERT(nvpair_type(elem) == DATA_TYPE_UINT64);
1240 VERIFY(nvpair_value_uint64(elem, &val) == 0);
1241 name = nvpair_name(elem);
1242 if (strcmp(name, zfs_prop_to_name(ZFS_PROP_VERSION)) == 0) {
1243 version = val;
1244 error = zap_update(os, moid, ZPL_VERSION_STR,
1245 8, 1, &version, tx);
1246 } else {
1247 error = zap_update(os, moid, name, 8, 1, &val, tx);
1248 }
1249 ASSERT(error == 0);
1250 if (strcmp(name, zfs_prop_to_name(ZFS_PROP_NORMALIZE)) == 0)
1251 norm = val;
1252 else if (strcmp(name, zfs_prop_to_name(ZFS_PROP_CASE)) == 0)
1253 sense = val;
1254 }
1255 ASSERT(version != 0);
1256
1257 /*
1258 * Create a delete queue.
1259 */
1260 doid = zap_create(os, DMU_OT_UNLINKED_SET, DMU_OT_NONE, 0, tx);
1261
1262 error = zap_add(os, moid, ZFS_UNLINKED_SET, 8, 1, &doid, tx);
1263 ASSERT(error == 0);
1264
1265 /*
1266 * Create root znode. Create minimal znode/vnode/zfsvfs
1267 * to allow zfs_mknode to work.
1268 */
1269 vattr.va_mask = AT_MODE|AT_UID|AT_GID|AT_TYPE;
1270 vattr.va_type = VDIR;
1271 vattr.va_mode = S_IFDIR|0755;
1272 vattr.va_uid = crgetuid(cr);
1273 vattr.va_gid = crgetgid(cr);
1274
1275 rootzp = kmem_cache_alloc(znode_cache, KM_SLEEP);
1276 rootzp->z_zfsvfs = &zfsvfs;
1277 rootzp->z_unlinked = 0;
1278 rootzp->z_atime_dirty = 0;
1279
1280 vp = ZTOV(rootzp);
1281 vn_reinit(vp);
1282 vp->v_type = VDIR;
1283
1284 bzero(&zfsvfs, sizeof (zfsvfs_t));
1285
1286 zfsvfs.z_os = os;
1287 zfsvfs.z_assign = TXG_NOWAIT;
1288 zfsvfs.z_parent = &zfsvfs;
1289 zfsvfs.z_version = version;
1290 zfsvfs.z_use_fuids = USE_FUIDS(version, os);
1291 zfsvfs.z_norm = norm;
1292 /*
1293 * Fold case on file systems that are always or sometimes case
1294 * insensitive.
1295 */
1296 if (sense == ZFS_CASE_INSENSITIVE || sense == ZFS_CASE_MIXED)
1297 zfsvfs.z_norm |= U8_TEXTPREP_TOUPPER;
1298
1299 mutex_init(&zfsvfs.z_znodes_lock, NULL, MUTEX_DEFAULT, NULL);
1300 list_create(&zfsvfs.z_all_znodes, sizeof (znode_t),
1301 offsetof(znode_t, z_link_node));
1302
1303 zfs_mknode(rootzp, &vattr, tx, cr, IS_ROOT_NODE, &zp, 0, NULL, NULL);
1304 ASSERT3P(zp, ==, rootzp);
1305 error = zap_add(os, moid, ZFS_ROOT_OBJ, 8, 1, &rootzp->z_id, tx);
1306 ASSERT(error == 0);
1307
1308 ZTOV(rootzp)->v_count = 0;
1309 dmu_buf_rele(rootzp->z_dbuf, NULL);
1310 rootzp->z_dbuf = NULL;
1311 kmem_cache_free(znode_cache, rootzp);
1312}
1313
1314#endif /* _KERNEL */
1315/*
1316 * Given an object number, return its parent object number and whether
1317 * or not the object is an extended attribute directory.
1318 */
1319static int
1320zfs_obj_to_pobj(objset_t *osp, uint64_t obj, uint64_t *pobjp, int *is_xattrdir)
1321{
1322 dmu_buf_t *db;
1323 dmu_object_info_t doi;
1324 znode_phys_t *zp;
1325 int error;
1326
1327 if ((error = dmu_bonus_hold(osp, obj, FTAG, &db)) != 0)
1328 return (error);
1329
1330 dmu_object_info_from_db(db, &doi);
1331 if (doi.doi_bonus_type != DMU_OT_ZNODE ||
1332 doi.doi_bonus_size < sizeof (znode_phys_t)) {
1333 dmu_buf_rele(db, FTAG);
1334 return (EINVAL);
1335 }
1336
1337 zp = db->db_data;
1338 *pobjp = zp->zp_parent;
1339 *is_xattrdir = ((zp->zp_flags & ZFS_XATTR) != 0) &&
1340 S_ISDIR(zp->zp_mode);
1341 dmu_buf_rele(db, FTAG);
1342
1343 return (0);
1344}
1345
1346int
1347zfs_obj_to_path(objset_t *osp, uint64_t obj, char *buf, int len)
1348{
1349 char *path = buf + len - 1;
1350 int error;
1351
1352 *path = '\0';
1353
1354 for (;;) {
1355 uint64_t pobj;
1356 char component[MAXNAMELEN + 2];
1357 size_t complen;
1358 int is_xattrdir;
1359
1360 if ((error = zfs_obj_to_pobj(osp, obj, &pobj,
1361 &is_xattrdir)) != 0)
1362 break;
1363
1364 if (pobj == obj) {
1365 if (path[0] != '/')
1366 *--path = '/';
1367 break;
1368 }
1369
1370 component[0] = '/';
1371 if (is_xattrdir) {
1372 (void) sprintf(component + 1, "<xattrdir>");
1373 } else {
1374 error = zap_value_search(osp, pobj, obj,
1375 ZFS_DIRENT_OBJ(-1ULL), component + 1);
1376 if (error != 0)
1377 break;
1378 }
1379
1380 complen = strlen(component);
1381 path -= complen;
1382 ASSERT(path >= buf);
1383 bcopy(component, path, complen);
1384 obj = pobj;
1385 }
1386
1387 if (error == 0)
1388 (void) memmove(buf, path, buf + len - path);
1389 return (error);
1390}