Deleted Added
sdiff udiff text old ( 178485 ) new ( 178491 )
full compact
1/*-
2 * Copyright (c) 1994 Jan-Simon Pendry
3 * Copyright (c) 1994
4 * The Regents of the University of California. All rights reserved.
5 * Copyright (c) 2005, 2006 Masanori Ozawa <ozawa@ongs.co.jp>, ONGS Inc.
6 * Copyright (c) 2006 Daichi Goto <daichi@freebsd.org>
7 *
8 * This code is derived from software contributed to Berkeley by
9 * Jan-Simon Pendry.
10 *
11 * Redistribution and use in source and binary forms, with or without
12 * modification, are permitted provided that the following conditions
13 * are met:
14 * 1. Redistributions of source code must retain the above copyright
15 * notice, this list of conditions and the following disclaimer.
16 * 2. Redistributions in binary form must reproduce the above copyright
17 * notice, this list of conditions and the following disclaimer in the
18 * documentation and/or other materials provided with the distribution.
19 * 4. Neither the name of the University nor the names of its contributors
20 * may be used to endorse or promote products derived from this software
21 * without specific prior written permission.
22 *
23 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
24 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
25 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
26 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
27 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
28 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
29 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
30 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
31 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
32 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
33 * SUCH DAMAGE.
34 *
35 * @(#)union_subr.c 8.20 (Berkeley) 5/20/95
36 * $FreeBSD: head/sys/fs/unionfs/union_subr.c 178491 2008-04-25 11:37:20Z daichi $
37 */
38
39#include <sys/param.h>
40#include <sys/systm.h>
41#include <sys/kernel.h>
42#include <sys/lock.h>
43#include <sys/mutex.h>
44#include <sys/malloc.h>
45#include <sys/mount.h>
46#include <sys/namei.h>
47#include <sys/proc.h>
48#include <sys/vnode.h>
49#include <sys/dirent.h>
50#include <sys/fcntl.h>
51#include <sys/filedesc.h>
52#include <sys/stat.h>
53#include <sys/resourcevar.h>
54
55#ifdef MAC
56#include <sys/mac.h>
57#endif
58
59#include <vm/uma.h>
60
61#include <fs/unionfs/union.h>
62
63#define NUNIONFSNODECACHE 16
64
65static MALLOC_DEFINE(M_UNIONFSHASH, "UNIONFS hash", "UNIONFS hash table");
66MALLOC_DEFINE(M_UNIONFSNODE, "UNIONFS node", "UNIONFS vnode private part");
67MALLOC_DEFINE(M_UNIONFSPATH, "UNIONFS path", "UNIONFS path private part");
68
69/*
70 * Initialize
71 */
72int
73unionfs_init(struct vfsconf *vfsp)
74{
75 UNIONFSDEBUG("unionfs_init\n"); /* printed during system boot */
76 return (0);
77}
78
79/*
80 * Uninitialize
81 */
82int
83unionfs_uninit(struct vfsconf *vfsp)
84{
85 return (0);
86}
87
88static struct unionfs_node_hashhead *
89unionfs_get_hashhead(struct vnode *dvp, char *path)
90{
91 int count;
92 char hash;
93 struct unionfs_node *unp;
94
95 hash = 0;
96 unp = VTOUNIONFS(dvp);
97 if (path != NULL) {
98 for (count = 0; path[count]; count++)
99 hash += path[count];
100 }
101
102 return (&(unp->un_hashtbl[hash & (unp->un_hashmask)]));
103}
104
105/*
106 * Get the cached vnode. (only VDIR)
107 */
108static struct vnode *
109unionfs_get_cached_vdir(struct vnode *uvp, struct vnode *lvp,
110 struct vnode *dvp, char *path)
111{
112 struct unionfs_node_hashhead *hd;
113 struct unionfs_node *unp;
114 struct vnode *vp;
115
116 KASSERT((uvp == NULLVP || uvp->v_type == VDIR || uvp->v_type == VSOCK),
117 ("unionfs_get_cached_vdir: v_type != VDIR/VSOCK"));
118 KASSERT((lvp == NULLVP || lvp->v_type == VDIR || lvp->v_type == VSOCK),
119 ("unionfs_get_cached_vdir: v_type != VDIR/VSOCK"));
120
121 VI_LOCK(dvp);
122 hd = unionfs_get_hashhead(dvp, path);
123 LIST_FOREACH(unp, hd, un_hash) {
124 if (!strcmp(unp->un_path, path)) {
125 vp = UNIONFSTOV(unp);
126 VI_LOCK_FLAGS(vp, MTX_DUPOK);
127 VI_UNLOCK(dvp);
128 vp->v_iflag &= ~VI_OWEINACT;
129 if ((vp->v_iflag & (VI_DOOMED | VI_DOINGINACT)) != 0) {
130 VI_UNLOCK(vp);
131 vp = NULLVP;
132 } else
133 VI_UNLOCK(vp);
134 return (vp);
135 }
136 }
137 VI_UNLOCK(dvp);
138
139 return (NULLVP);
140}
141
142/*
143 * Add the new vnode into cache. (only VDIR)
144 */
145static struct vnode *
146unionfs_ins_cached_vdir(struct unionfs_node *uncp,
147 struct vnode *dvp, char *path)
148{
149 struct unionfs_node_hashhead *hd;
150 struct unionfs_node *unp;
151 struct vnode *vp;
152
153 KASSERT((uncp->un_uppervp==NULLVP || uncp->un_uppervp->v_type==VDIR ||
154 uncp->un_uppervp->v_type==VSOCK),
155 ("unionfs_ins_cached_vdir: v_type != VDIR/VSOCK"));
156 KASSERT((uncp->un_lowervp==NULLVP || uncp->un_lowervp->v_type==VDIR ||
157 uncp->un_lowervp->v_type==VSOCK),
158 ("unionfs_ins_cached_vdir: v_type != VDIR/VSOCK"));
159
160 VI_LOCK(dvp);
161 hd = unionfs_get_hashhead(dvp, path);
162 LIST_FOREACH(unp, hd, un_hash) {
163 if (!strcmp(unp->un_path, path)) {
164 vp = UNIONFSTOV(unp);
165 VI_LOCK_FLAGS(vp, MTX_DUPOK);
166 vp->v_iflag &= ~VI_OWEINACT;
167 if ((vp->v_iflag & (VI_DOOMED | VI_DOINGINACT)) != 0) {
168 LIST_INSERT_HEAD(hd, uncp, un_hash);
169 VI_UNLOCK(vp);
170 vp = NULLVP;
171 } else
172 VI_UNLOCK(vp);
173 VI_UNLOCK(dvp);
174 return (vp);
175 }
176 }
177
178 LIST_INSERT_HEAD(hd, uncp, un_hash);
179 VI_UNLOCK(dvp);
180
181 return (NULLVP);
182}
183
184/*
185 * Remove the vnode. (only VDIR)
186 */
187static void
188unionfs_rem_cached_vdir(struct unionfs_node *unp, struct vnode *dvp)
189{
190 KASSERT((unp != NULL), ("unionfs_rem_cached_vdir: null node"));
191 KASSERT((dvp != NULLVP),
192 ("unionfs_rem_cached_vdir: null parent vnode"));
193 KASSERT((unp->un_hash.le_prev != NULL),
194 ("unionfs_rem_cached_vdir: null hash"));
195
196 VI_LOCK(dvp);
197 LIST_REMOVE(unp, un_hash);
198 VI_UNLOCK(dvp);
199}
200
201/*
202 * Make a new or get existing unionfs node.
203 *
204 * uppervp and lowervp should be unlocked. Because if new unionfs vnode is
205 * locked, uppervp or lowervp is locked too. In order to prevent dead lock,
206 * you should not lock plurality simultaneously.
207 */
208int
209unionfs_nodeget(struct mount *mp, struct vnode *uppervp,
210 struct vnode *lowervp, struct vnode *dvp,
211 struct vnode **vpp, struct componentname *cnp,
212 struct thread *td)
213{
214 struct unionfs_mount *ump;
215 struct unionfs_node *unp;
216 struct vnode *vp;
217 int error;
218 int lkflags;
219 enum vtype vt;
220 char *path;
221
222 ump = MOUNTTOUNIONFSMOUNT(mp);
223 lkflags = (cnp ? cnp->cn_lkflags : 0);
224 path = (cnp ? cnp->cn_nameptr : NULL);
225 *vpp = NULLVP;
226
227 if (uppervp == NULLVP && lowervp == NULLVP)
228 panic("unionfs_nodeget: upper and lower is null");
229
230 vt = (uppervp != NULLVP ? uppervp->v_type : lowervp->v_type);
231
232 /* If it has no ISLASTCN flag, path check is skipped. */
233 if (cnp && !(cnp->cn_flags & ISLASTCN))
234 path = NULL;
235
236 /* check the vdir cache */
237 if (path != NULL && dvp != NULLVP && (vt == VDIR || vt == VSOCK)) {
238 vp = unionfs_get_cached_vdir(uppervp, lowervp, dvp, path);
239 if (vp != NULLVP) {
240 vref(vp);
241 *vpp = vp;
242 goto unionfs_nodeget_out;
243 }
244 }
245
246 if ((uppervp == NULLVP || ump->um_uppervp != uppervp) ||
247 (lowervp == NULLVP || ump->um_lowervp != lowervp)) {
248 /* dvp will be NULLVP only in case of root vnode. */
249 if (dvp == NULLVP)
250 return (EINVAL);
251 }
252
253 /*
254 * Do the MALLOC before the getnewvnode since doing so afterward
255 * might cause a bogus v_data pointer to get dereferenced elsewhere
256 * if MALLOC should block.
257 */
258 MALLOC(unp, struct unionfs_node *, sizeof(struct unionfs_node),
259 M_UNIONFSNODE, M_WAITOK | M_ZERO);
260
261 error = getnewvnode("unionfs", mp, &unionfs_vnodeops, &vp);
262 if (error != 0) {
263 FREE(unp, M_UNIONFSNODE);
264 return (error);
265 }
266 error = insmntque(vp, mp); /* XXX: Too early for mpsafe fs */
267 if (error != 0) {
268 FREE(unp, M_UNIONFSNODE);
269 return (error);
270 }
271 if (dvp != NULLVP)
272 vref(dvp);
273 if (uppervp != NULLVP)
274 vref(uppervp);
275 if (lowervp != NULLVP)
276 vref(lowervp);
277
278 switch (vt) {
279 case VDIR:
280 unp->un_hashtbl = hashinit(NUNIONFSNODECACHE, M_UNIONFSHASH,
281 &(unp->un_hashmask));
282 break;
283 case VSOCK:
284 if (uppervp != NULLVP)
285 vp->v_socket = uppervp->v_socket;
286 else
287 vp->v_socket = lowervp->v_socket;
288 break;
289 default:
290 break;
291 }
292
293 unp->un_vnode = vp;
294 unp->un_uppervp = uppervp;
295 unp->un_lowervp = lowervp;
296 unp->un_dvp = dvp;
297 if (uppervp != NULLVP)
298 vp->v_vnlock = uppervp->v_vnlock;
299 else
300 vp->v_vnlock = lowervp->v_vnlock;
301
302 if (path != NULL) {
303 unp->un_path = (char *)
304 malloc(cnp->cn_namelen +1, M_UNIONFSPATH, M_WAITOK|M_ZERO);
305 bcopy(cnp->cn_nameptr, unp->un_path, cnp->cn_namelen);
306 unp->un_path[cnp->cn_namelen] = '\0';
307 }
308 vp->v_type = vt;
309 vp->v_data = unp;
310
311 if ((uppervp != NULLVP && ump->um_uppervp == uppervp) &&
312 (lowervp != NULLVP && ump->um_lowervp == lowervp))
313 vp->v_vflag |= VV_ROOT;
314
315 if (path != NULL && dvp != NULLVP && (vt == VDIR || vt == VSOCK))
316 *vpp = unionfs_ins_cached_vdir(unp, dvp, path);
317 if ((*vpp) != NULLVP) {
318 if (dvp != NULLVP)
319 vrele(dvp);
320 if (uppervp != NULLVP)
321 vrele(uppervp);
322 if (lowervp != NULLVP)
323 vrele(lowervp);
324
325 unp->un_uppervp = NULLVP;
326 unp->un_lowervp = NULLVP;
327 unp->un_dvp = NULLVP;
328 vrele(vp);
329 vp = *vpp;
330 vref(vp);
331 } else
332 *vpp = vp;
333
334unionfs_nodeget_out:
335 if (lkflags & LK_TYPE_MASK)
336 vn_lock(vp, lkflags | LK_RETRY);
337
338 return (0);
339}
340
341/*
342 * Clean up the unionfs node.
343 */
344void
345unionfs_noderem(struct vnode *vp, struct thread *td)
346{
347 int vfslocked;
348 struct unionfs_node *unp;
349 struct unionfs_node_status *unsp, *unsp_tmp;
350 struct vnode *lvp;
351 struct vnode *uvp;
352 struct vnode *dvp;
353
354 /*
355 * Use the interlock to protect the clearing of v_data to
356 * prevent faults in unionfs_lock().
357 */
358 VI_LOCK(vp);
359 unp = VTOUNIONFS(vp);
360 lvp = unp->un_lowervp;
361 uvp = unp->un_uppervp;
362 dvp = unp->un_dvp;
363 unp->un_lowervp = unp->un_uppervp = NULLVP;
364
365 vp->v_vnlock = &(vp->v_lock);
366 vp->v_data = NULL;
367 lockmgr(vp->v_vnlock, LK_EXCLUSIVE | LK_INTERLOCK, VI_MTX(vp));
368 if (lvp != NULLVP)
369 VOP_UNLOCK(lvp, 0);
370 if (uvp != NULLVP)
371 VOP_UNLOCK(uvp, 0);
372 vp->v_object = NULL;
373
374 if (unp->un_path != NULL && dvp != NULLVP &&
375 (vp->v_type == VDIR || vp->v_type == VSOCK))
376 unionfs_rem_cached_vdir(unp, dvp);
377
378 if (lvp != NULLVP) {
379 vfslocked = VFS_LOCK_GIANT(lvp->v_mount);
380 vrele(lvp);
381 VFS_UNLOCK_GIANT(vfslocked);
382 }
383 if (uvp != NULLVP) {
384 vfslocked = VFS_LOCK_GIANT(uvp->v_mount);
385 vrele(uvp);
386 VFS_UNLOCK_GIANT(vfslocked);
387 }
388 if (dvp != NULLVP) {
389 vfslocked = VFS_LOCK_GIANT(dvp->v_mount);
390 vrele(dvp);
391 VFS_UNLOCK_GIANT(vfslocked);
392 unp->un_dvp = NULLVP;
393 }
394 if (unp->un_path != NULL) {
395 free(unp->un_path, M_UNIONFSPATH);
396 unp->un_path = NULL;
397 }
398
399 if (unp->un_hashtbl != NULL)
400 hashdestroy(unp->un_hashtbl, M_UNIONFSHASH, unp->un_hashmask);
401
402 LIST_FOREACH_SAFE(unsp, &(unp->un_unshead), uns_list, unsp_tmp) {
403 LIST_REMOVE(unsp, uns_list);
404 free(unsp, M_TEMP);
405 }
406 FREE(unp, M_UNIONFSNODE);
407}
408
409/*
410 * Get the unionfs node status.
411 * You need exclusive lock this vnode.
412 */
413void
414unionfs_get_node_status(struct unionfs_node *unp, struct thread *td,
415 struct unionfs_node_status **unspp)
416{
417 struct unionfs_node_status *unsp;
418 pid_t pid = td->td_proc->p_pid;
419
420 KASSERT(NULL != unspp, ("null pointer"));
421 ASSERT_VOP_ELOCKED(UNIONFSTOV(unp), "unionfs_get_node_status");
422
423 LIST_FOREACH(unsp, &(unp->un_unshead), uns_list) {
424 if (unsp->uns_pid == pid) {
425 *unspp = unsp;
426 return;
427 }
428 }
429
430 /* create a new unionfs node status */
431 MALLOC(unsp, struct unionfs_node_status *,
432 sizeof(struct unionfs_node_status), M_TEMP, M_WAITOK | M_ZERO);
433
434 unsp->uns_pid = pid;
435 LIST_INSERT_HEAD(&(unp->un_unshead), unsp, uns_list);
436
437 *unspp = unsp;
438}
439
440/*
441 * Remove the unionfs node status, if you can.
442 * You need exclusive lock this vnode.
443 */
444void
445unionfs_tryrem_node_status(struct unionfs_node *unp,
446 struct unionfs_node_status *unsp)
447{
448 KASSERT(NULL != unsp, ("null pointer"));
449 ASSERT_VOP_ELOCKED(UNIONFSTOV(unp), "unionfs_get_node_status");
450
451 if (0 < unsp->uns_lower_opencnt || 0 < unsp->uns_upper_opencnt)
452 return;
453
454 LIST_REMOVE(unsp, uns_list);
455 free(unsp, M_TEMP);
456}
457
458/*
459 * Create upper node attr.
460 */
461void
462unionfs_create_uppervattr_core(struct unionfs_mount *ump,
463 struct vattr *lva,
464 struct vattr *uva,
465 struct thread *td)
466{
467 VATTR_NULL(uva);
468 uva->va_type = lva->va_type;
469 uva->va_atime = lva->va_atime;
470 uva->va_mtime = lva->va_mtime;
471 uva->va_ctime = lva->va_ctime;
472
473 switch (ump->um_copymode) {
474 case UNIONFS_TRANSPARENT:
475 uva->va_mode = lva->va_mode;
476 uva->va_uid = lva->va_uid;
477 uva->va_gid = lva->va_gid;
478 break;
479 case UNIONFS_MASQUERADE:
480 if (ump->um_uid == lva->va_uid) {
481 uva->va_mode = lva->va_mode & 077077;
482 uva->va_mode |= (lva->va_type == VDIR ? ump->um_udir : ump->um_ufile) & 0700;
483 uva->va_uid = lva->va_uid;
484 uva->va_gid = lva->va_gid;
485 } else {
486 uva->va_mode = (lva->va_type == VDIR ? ump->um_udir : ump->um_ufile);
487 uva->va_uid = ump->um_uid;
488 uva->va_gid = ump->um_gid;
489 }
490 break;
491 default: /* UNIONFS_TRADITIONAL */
492 FILEDESC_SLOCK(td->td_proc->p_fd);
493 uva->va_mode = 0777 & ~td->td_proc->p_fd->fd_cmask;
494 FILEDESC_SUNLOCK(td->td_proc->p_fd);
495 uva->va_uid = ump->um_uid;
496 uva->va_gid = ump->um_gid;
497 break;
498 }
499}
500
501/*
502 * Create upper node attr.
503 */
504int
505unionfs_create_uppervattr(struct unionfs_mount *ump,
506 struct vnode *lvp,
507 struct vattr *uva,
508 struct ucred *cred,
509 struct thread *td)
510{
511 int error;
512 struct vattr lva;
513
514 if ((error = VOP_GETATTR(lvp, &lva, cred, td)))
515 return (error);
516
517 unionfs_create_uppervattr_core(ump, &lva, uva, td);
518
519 return (error);
520}
521
522/*
523 * relookup
524 *
525 * dvp should be locked on entry and will be locked on return.
526 *
527 * If an error is returned, *vpp will be invalid, otherwise it will hold a
528 * locked, referenced vnode. If *vpp == dvp then remember that only one
529 * LK_EXCLUSIVE lock is held.
530 */
531static int
532unionfs_relookup(struct vnode *dvp, struct vnode **vpp,
533 struct componentname *cnp, struct componentname *cn,
534 struct thread *td, char *path, int pathlen, u_long nameiop)
535{
536 int error;
537
538 cn->cn_namelen = pathlen;
539 cn->cn_pnbuf = uma_zalloc(namei_zone, M_WAITOK);
540 bcopy(path, cn->cn_pnbuf, pathlen);
541 cn->cn_pnbuf[pathlen] = '\0';
542
543 cn->cn_nameiop = nameiop;
544 cn->cn_flags = (LOCKPARENT | LOCKLEAF | HASBUF | SAVENAME | ISLASTCN);
545 cn->cn_lkflags = LK_EXCLUSIVE;
546 cn->cn_thread = td;
547 cn->cn_cred = cnp->cn_cred;
548
549 cn->cn_nameptr = cn->cn_pnbuf;
550 cn->cn_consume = cnp->cn_consume;
551
552 if (nameiop == DELETE)
553 cn->cn_flags |= (cnp->cn_flags & (DOWHITEOUT | SAVESTART));
554 else if (RENAME == nameiop)
555 cn->cn_flags |= (cnp->cn_flags & SAVESTART);
556
557 vref(dvp);
558 VOP_UNLOCK(dvp, 0);
559
560 if ((error = relookup(dvp, vpp, cn))) {
561 uma_zfree(namei_zone, cn->cn_pnbuf);
562 cn->cn_flags &= ~HASBUF;
563 vn_lock(dvp, LK_EXCLUSIVE | LK_RETRY);
564 } else
565 vrele(dvp);
566
567 return (error);
568}
569
570/*
571 * relookup for CREATE namei operation.
572 *
573 * dvp is unionfs vnode. dvp should be locked.
574 *
575 * If it called 'unionfs_copyfile' function by unionfs_link etc,
576 * VOP_LOOKUP information is broken.
577 * So it need relookup in order to create link etc.
578 */
579int
580unionfs_relookup_for_create(struct vnode *dvp, struct componentname *cnp,
581 struct thread *td)
582{
583 int error;
584 struct vnode *udvp;
585 struct vnode *vp;
586 struct componentname cn;
587
588 udvp = UNIONFSVPTOUPPERVP(dvp);
589 vp = NULLVP;
590
591 error = unionfs_relookup(udvp, &vp, cnp, &cn, td, cnp->cn_nameptr,
592 strlen(cnp->cn_nameptr), CREATE);
593 if (error)
594 return (error);
595
596 if (vp != NULLVP) {
597 if (udvp == vp)
598 vrele(vp);
599 else
600 vput(vp);
601
602 error = EEXIST;
603 }
604
605 if (cn.cn_flags & HASBUF) {
606 uma_zfree(namei_zone, cn.cn_pnbuf);
607 cn.cn_flags &= ~HASBUF;
608 }
609
610 if (!error) {
611 cn.cn_flags |= (cnp->cn_flags & HASBUF);
612 cnp->cn_flags = cn.cn_flags;
613 }
614
615 return (error);
616}
617
618/*
619 * relookup for DELETE namei operation.
620 *
621 * dvp is unionfs vnode. dvp should be locked.
622 */
623int
624unionfs_relookup_for_delete(struct vnode *dvp, struct componentname *cnp,
625 struct thread *td)
626{
627 int error;
628 struct vnode *udvp;
629 struct vnode *vp;
630 struct componentname cn;
631
632 udvp = UNIONFSVPTOUPPERVP(dvp);
633 vp = NULLVP;
634
635 error = unionfs_relookup(udvp, &vp, cnp, &cn, td, cnp->cn_nameptr,
636 strlen(cnp->cn_nameptr), DELETE);
637 if (error)
638 return (error);
639
640 if (vp == NULLVP)
641 error = ENOENT;
642 else {
643 if (udvp == vp)
644 vrele(vp);
645 else
646 vput(vp);
647 }
648
649 if (cn.cn_flags & HASBUF) {
650 uma_zfree(namei_zone, cn.cn_pnbuf);
651 cn.cn_flags &= ~HASBUF;
652 }
653
654 if (!error) {
655 cn.cn_flags |= (cnp->cn_flags & HASBUF);
656 cnp->cn_flags = cn.cn_flags;
657 }
658
659 return (error);
660}
661
662/*
663 * relookup for RENAME namei operation.
664 *
665 * dvp is unionfs vnode. dvp should be locked.
666 */
667int
668unionfs_relookup_for_rename(struct vnode *dvp, struct componentname *cnp,
669 struct thread *td)
670{
671 int error;
672 struct vnode *udvp;
673 struct vnode *vp;
674 struct componentname cn;
675
676 udvp = UNIONFSVPTOUPPERVP(dvp);
677 vp = NULLVP;
678
679 error = unionfs_relookup(udvp, &vp, cnp, &cn, td, cnp->cn_nameptr,
680 strlen(cnp->cn_nameptr), RENAME);
681 if (error)
682 return (error);
683
684 if (vp != NULLVP) {
685 if (udvp == vp)
686 vrele(vp);
687 else
688 vput(vp);
689 }
690
691 if (cn.cn_flags & HASBUF) {
692 uma_zfree(namei_zone, cn.cn_pnbuf);
693 cn.cn_flags &= ~HASBUF;
694 }
695
696 if (!error) {
697 cn.cn_flags |= (cnp->cn_flags & HASBUF);
698 cnp->cn_flags = cn.cn_flags;
699 }
700
701 return (error);
702
703}
704
705/*
706 * Update the unionfs_node.
707 *
708 * uvp is new locked upper vnode. unionfs vnode's lock will be exchanged to the
709 * uvp's lock and lower's lock will be unlocked.
710 */
711static void
712unionfs_node_update(struct unionfs_node *unp, struct vnode *uvp,
713 struct thread *td)
714{
715 unsigned count, lockrec;
716 struct vnode *vp;
717 struct vnode *lvp;
718 struct vnode *dvp;
719
720 vp = UNIONFSTOV(unp);
721 lvp = unp->un_lowervp;
722 ASSERT_VOP_ELOCKED(lvp, "unionfs_node_update");
723 dvp = unp->un_dvp;
724
725 /*
726 * lock update
727 */
728 VI_LOCK(vp);
729 unp->un_uppervp = uvp;
730 vp->v_vnlock = uvp->v_vnlock;
731 VI_UNLOCK(vp);
732 lockrec = lvp->v_vnlock->lk_recurse;
733 for (count = 0; count < lockrec; count++)
734 vn_lock(uvp, LK_EXCLUSIVE | LK_CANRECURSE | LK_RETRY);
735
736 /*
737 * cache update
738 */
739 if (unp->un_path != NULL && dvp != NULLVP &&
740 (vp->v_type == VDIR || vp->v_type == VSOCK)) {
741 static struct unionfs_node_hashhead *hd;
742
743 VI_LOCK(dvp);
744 hd = unionfs_get_hashhead(dvp, unp->un_path);
745 LIST_REMOVE(unp, un_hash);
746 LIST_INSERT_HEAD(hd, unp, un_hash);
747 VI_UNLOCK(dvp);
748 }
749}
750
751/*
752 * Create a new shadow dir.
753 *
754 * udvp should be locked on entry and will be locked on return.
755 *
756 * If no error returned, unp will be updated.
757 */
758int
759unionfs_mkshadowdir(struct unionfs_mount *ump, struct vnode *udvp,
760 struct unionfs_node *unp, struct componentname *cnp,
761 struct thread *td)
762{
763 int error;
764 struct vnode *lvp;
765 struct vnode *uvp;
766 struct vattr va;
767 struct vattr lva;
768 struct componentname cn;
769 struct mount *mp;
770 struct ucred *cred;
771 struct ucred *credbk;
772 struct uidinfo *rootinfo;
773
774 if (unp->un_uppervp != NULLVP)
775 return (EEXIST);
776
777 lvp = unp->un_lowervp;
778 uvp = NULLVP;
779 credbk = cnp->cn_cred;
780
781 /* Authority change to root */
782 rootinfo = uifind((uid_t)0);
783 cred = crdup(cnp->cn_cred);
784 chgproccnt(cred->cr_ruidinfo, 1, 0);
785 change_euid(cred, rootinfo);
786 change_ruid(cred, rootinfo);
787 change_svuid(cred, (uid_t)0);
788 uifree(rootinfo);
789 cnp->cn_cred = cred;
790
791 memset(&cn, 0, sizeof(cn));
792
793 if ((error = VOP_GETATTR(lvp, &lva, cnp->cn_cred, td)))
794 goto unionfs_mkshadowdir_abort;
795
796 if ((error = unionfs_relookup(udvp, &uvp, cnp, &cn, td, cnp->cn_nameptr, cnp->cn_namelen, CREATE)))
797 goto unionfs_mkshadowdir_abort;
798 if (uvp != NULLVP) {
799 if (udvp == uvp)
800 vrele(uvp);
801 else
802 vput(uvp);
803
804 error = EEXIST;
805 goto unionfs_mkshadowdir_free_out;
806 }
807
808 if ((error = vn_start_write(udvp, &mp, V_WAIT | PCATCH)))
809 goto unionfs_mkshadowdir_free_out;
810 if ((error = VOP_LEASE(udvp, td, cn.cn_cred, LEASE_WRITE))) {
811 vn_finished_write(mp);
812 goto unionfs_mkshadowdir_free_out;
813 }
814 unionfs_create_uppervattr_core(ump, &lva, &va, td);
815
816 error = VOP_MKDIR(udvp, &uvp, &cn, &va);
817
818 if (!error) {
819 unionfs_node_update(unp, uvp, td);
820
821 /*
822 * XXX The bug which cannot set uid/gid was corrected.
823 * Ignore errors.
824 */
825 va.va_type = VNON;
826 VOP_SETATTR(uvp, &va, cn.cn_cred, td);
827 }
828 vn_finished_write(mp);
829
830unionfs_mkshadowdir_free_out:
831 if (cn.cn_flags & HASBUF) {
832 uma_zfree(namei_zone, cn.cn_pnbuf);
833 cn.cn_flags &= ~HASBUF;
834 }
835
836unionfs_mkshadowdir_abort:
837 cnp->cn_cred = credbk;
838 chgproccnt(cred->cr_ruidinfo, -1, 0);
839 crfree(cred);
840
841 return (error);
842}
843
844/*
845 * Create a new whiteout.
846 *
847 * dvp should be locked on entry and will be locked on return.
848 */
849int
850unionfs_mkwhiteout(struct vnode *dvp, struct componentname *cnp,
851 struct thread *td, char *path)
852{
853 int error;
854 struct vnode *wvp;
855 struct componentname cn;
856 struct mount *mp;
857
858 if (path == NULL)
859 path = cnp->cn_nameptr;
860
861 wvp = NULLVP;
862 if ((error = unionfs_relookup(dvp, &wvp, cnp, &cn, td, path, strlen(path), CREATE)))
863 return (error);
864 if (wvp != NULLVP) {
865 if (cn.cn_flags & HASBUF) {
866 uma_zfree(namei_zone, cn.cn_pnbuf);
867 cn.cn_flags &= ~HASBUF;
868 }
869 if (dvp == wvp)
870 vrele(wvp);
871 else
872 vput(wvp);
873
874 return (EEXIST);
875 }
876
877 if ((error = vn_start_write(dvp, &mp, V_WAIT | PCATCH)))
878 goto unionfs_mkwhiteout_free_out;
879 if (!(error = VOP_LEASE(dvp, td, td->td_ucred, LEASE_WRITE)))
880 error = VOP_WHITEOUT(dvp, &cn, CREATE);
881
882 vn_finished_write(mp);
883
884unionfs_mkwhiteout_free_out:
885 if (cn.cn_flags & HASBUF) {
886 uma_zfree(namei_zone, cn.cn_pnbuf);
887 cn.cn_flags &= ~HASBUF;
888 }
889
890 return (error);
891}
892
893/*
894 * Create a new vnode for create a new shadow file.
895 *
896 * If an error is returned, *vpp will be invalid, otherwise it will hold a
897 * locked, referenced and opened vnode.
898 *
899 * unp is never updated.
900 */
901static int
902unionfs_vn_create_on_upper(struct vnode **vpp, struct vnode *udvp,
903 struct unionfs_node *unp, struct vattr *uvap,
904 struct thread *td)
905{
906 struct unionfs_mount *ump;
907 struct vnode *vp;
908 struct vnode *lvp;
909 struct ucred *cred;
910 struct vattr lva;
911 int fmode;
912 int error;
913 struct componentname cn;
914
915 ump = MOUNTTOUNIONFSMOUNT(UNIONFSTOV(unp)->v_mount);
916 vp = NULLVP;
917 lvp = unp->un_lowervp;
918 cred = td->td_ucred;
919 fmode = FFLAGS(O_WRONLY | O_CREAT | O_TRUNC | O_EXCL);
920 error = 0;
921
922 if ((error = VOP_GETATTR(lvp, &lva, cred, td)) != 0)
923 return (error);
924 unionfs_create_uppervattr_core(ump, &lva, uvap, td);
925
926 if (unp->un_path == NULL)
927 panic("unionfs: un_path is null");
928
929 cn.cn_namelen = strlen(unp->un_path);
930 cn.cn_pnbuf = uma_zalloc(namei_zone, M_WAITOK);
931 bcopy(unp->un_path, cn.cn_pnbuf, cn.cn_namelen + 1);
932 cn.cn_nameiop = CREATE;
933 cn.cn_flags = (LOCKPARENT | LOCKLEAF | HASBUF | SAVENAME | ISLASTCN);
934 cn.cn_lkflags = LK_EXCLUSIVE;
935 cn.cn_thread = td;
936 cn.cn_cred = cred;
937 cn.cn_nameptr = cn.cn_pnbuf;
938 cn.cn_consume = 0;
939
940 vref(udvp);
941 if ((error = relookup(udvp, &vp, &cn)) != 0)
942 goto unionfs_vn_create_on_upper_free_out2;
943 vrele(udvp);
944
945 if (vp != NULLVP) {
946 if (vp == udvp)
947 vrele(vp);
948 else
949 vput(vp);
950 error = EEXIST;
951 goto unionfs_vn_create_on_upper_free_out1;
952 }
953
954 if ((error = VOP_LEASE(udvp, td, cred, LEASE_WRITE)) != 0)
955 goto unionfs_vn_create_on_upper_free_out1;
956
957 if ((error = VOP_CREATE(udvp, &vp, &cn, uvap)) != 0)
958 goto unionfs_vn_create_on_upper_free_out1;
959
960 if ((error = VOP_OPEN(vp, fmode, cred, td, NULL)) != 0) {
961 vput(vp);
962 goto unionfs_vn_create_on_upper_free_out1;
963 }
964 vp->v_writecount++;
965 *vpp = vp;
966
967unionfs_vn_create_on_upper_free_out1:
968 VOP_UNLOCK(udvp, 0);
969
970unionfs_vn_create_on_upper_free_out2:
971 if (cn.cn_flags & HASBUF) {
972 uma_zfree(namei_zone, cn.cn_pnbuf);
973 cn.cn_flags &= ~HASBUF;
974 }
975
976 return (error);
977}
978
979/*
980 * Copy from lvp to uvp.
981 *
982 * lvp and uvp should be locked and opened on entry and will be locked and
983 * opened on return.
984 */
985static int
986unionfs_copyfile_core(struct vnode *lvp, struct vnode *uvp,
987 struct ucred *cred, struct thread *td)
988{
989 int error;
990 off_t offset;
991 int count;
992 int bufoffset;
993 char *buf;
994 struct uio uio;
995 struct iovec iov;
996
997 error = 0;
998 memset(&uio, 0, sizeof(uio));
999
1000 uio.uio_td = td;
1001 uio.uio_segflg = UIO_SYSSPACE;
1002 uio.uio_offset = 0;
1003
1004 if ((error = VOP_LEASE(lvp, td, cred, LEASE_READ)) != 0)
1005 return (error);
1006 if ((error = VOP_LEASE(uvp, td, cred, LEASE_WRITE)) != 0)
1007 return (error);
1008 buf = malloc(MAXBSIZE, M_TEMP, M_WAITOK);
1009
1010 while (error == 0) {
1011 offset = uio.uio_offset;
1012
1013 uio.uio_iov = &iov;
1014 uio.uio_iovcnt = 1;
1015 iov.iov_base = buf;
1016 iov.iov_len = MAXBSIZE;
1017 uio.uio_resid = iov.iov_len;
1018 uio.uio_rw = UIO_READ;
1019
1020 if ((error = VOP_READ(lvp, &uio, 0, cred)) != 0)
1021 break;
1022 if ((count = MAXBSIZE - uio.uio_resid) == 0)
1023 break;
1024
1025 bufoffset = 0;
1026 while (bufoffset < count) {
1027 uio.uio_iov = &iov;
1028 uio.uio_iovcnt = 1;
1029 iov.iov_base = buf + bufoffset;
1030 iov.iov_len = count - bufoffset;
1031 uio.uio_offset = offset + bufoffset;
1032 uio.uio_resid = iov.iov_len;
1033 uio.uio_rw = UIO_WRITE;
1034
1035 if ((error = VOP_WRITE(uvp, &uio, 0, cred)) != 0)
1036 break;
1037
1038 bufoffset += (count - bufoffset) - uio.uio_resid;
1039 }
1040
1041 uio.uio_offset = offset + bufoffset;
1042 }
1043
1044 free(buf, M_TEMP);
1045
1046 return (error);
1047}
1048
1049/*
1050 * Copy file from lower to upper.
1051 *
1052 * If you need copy of the contents, set 1 to docopy. Otherwise, set 0 to
1053 * docopy.
1054 *
1055 * If no error returned, unp will be updated.
1056 */
1057int
1058unionfs_copyfile(struct unionfs_node *unp, int docopy, struct ucred *cred,
1059 struct thread *td)
1060{
1061 int error;
1062 struct mount *mp;
1063 struct vnode *udvp;
1064 struct vnode *lvp;
1065 struct vnode *uvp;
1066 struct vattr uva;
1067
1068 lvp = unp->un_lowervp;
1069 uvp = NULLVP;
1070
1071 if ((UNIONFSTOV(unp)->v_mount->mnt_flag & MNT_RDONLY))
1072 return (EROFS);
1073 if (unp->un_dvp == NULLVP)
1074 return (EINVAL);
1075 if (unp->un_uppervp != NULLVP)
1076 return (EEXIST);
1077 udvp = VTOUNIONFS(unp->un_dvp)->un_uppervp;
1078 if (udvp == NULLVP)
1079 return (EROFS);
1080 if ((udvp->v_mount->mnt_flag & MNT_RDONLY))
1081 return (EROFS);
1082
1083 error = VOP_ACCESS(lvp, VREAD, cred, td);
1084 if (error != 0)
1085 return (error);
1086
1087 if ((error = vn_start_write(udvp, &mp, V_WAIT | PCATCH)) != 0)
1088 return (error);
1089 error = unionfs_vn_create_on_upper(&uvp, udvp, unp, &uva, td);
1090 if (error != 0) {
1091 vn_finished_write(mp);
1092 return (error);
1093 }
1094
1095 if (docopy != 0) {
1096 error = VOP_OPEN(lvp, FREAD, cred, td, NULL);
1097 if (error == 0) {
1098 error = unionfs_copyfile_core(lvp, uvp, cred, td);
1099 VOP_CLOSE(lvp, FREAD, cred, td);
1100 }
1101 }
1102 VOP_CLOSE(uvp, FWRITE, cred, td);
1103 uvp->v_writecount--;
1104
1105 vn_finished_write(mp);
1106
1107 if (error == 0) {
1108 /* Reset the attributes. Ignore errors. */
1109 uva.va_type = VNON;
1110 VOP_SETATTR(uvp, &uva, cred, td);
1111 }
1112
1113 unionfs_node_update(unp, uvp, td);
1114
1115 return (error);
1116}
1117
1118/*
1119 * It checks whether vp can rmdir. (check empty)
1120 *
1121 * vp is unionfs vnode.
1122 * vp should be locked.
1123 */
1124int
1125unionfs_check_rmdir(struct vnode *vp, struct ucred *cred, struct thread *td)
1126{
1127 int error;
1128 int eofflag;
1129 int lookuperr;
1130 struct vnode *uvp;
1131 struct vnode *lvp;
1132 struct vnode *tvp;
1133 struct vattr va;
1134 struct componentname cn;
1135 /*
1136 * The size of buf needs to be larger than DIRBLKSIZ.
1137 */
1138 char buf[256 * 6];
1139 struct dirent *dp;
1140 struct dirent *edp;
1141 struct uio uio;
1142 struct iovec iov;
1143
1144 ASSERT_VOP_ELOCKED(vp, "unionfs_check_rmdir");
1145
1146 eofflag = 0;
1147 uvp = UNIONFSVPTOUPPERVP(vp);
1148 lvp = UNIONFSVPTOLOWERVP(vp);
1149
1150 /* check opaque */
1151 if ((error = VOP_GETATTR(uvp, &va, cred, td)) != 0)
1152 return (error);
1153 if (va.va_flags & OPAQUE)
1154 return (0);
1155
1156 /* open vnode */
1157#ifdef MAC
1158 if ((error = mac_vnode_check_open(cred, vp, VEXEC|VREAD)) != 0)
1159 return (error);
1160#endif
1161 if ((error = VOP_ACCESS(vp, VEXEC|VREAD, cred, td)) != 0)
1162 return (error);
1163 if ((error = VOP_OPEN(vp, FREAD, cred, td, NULL)) != 0)
1164 return (error);
1165
1166 uio.uio_rw = UIO_READ;
1167 uio.uio_segflg = UIO_SYSSPACE;
1168 uio.uio_td = td;
1169 uio.uio_offset = 0;
1170
1171#ifdef MAC
1172 error = mac_vnode_check_readdir(td->td_ucred, lvp);
1173#endif
1174 while (!error && !eofflag) {
1175 iov.iov_base = buf;
1176 iov.iov_len = sizeof(buf);
1177 uio.uio_iov = &iov;
1178 uio.uio_iovcnt = 1;
1179 uio.uio_resid = iov.iov_len;
1180
1181 error = VOP_READDIR(lvp, &uio, cred, &eofflag, NULL, NULL);
1182 if (error != 0)
1183 break;
1184 if (eofflag == 0 && uio.uio_resid == sizeof(buf)) {
1185#ifdef DIAGNOSTIC
1186 panic("bad readdir response from lower FS.");
1187#endif
1188 break;
1189 }
1190
1191 edp = (struct dirent*)&buf[sizeof(buf) - uio.uio_resid];
1192 for (dp = (struct dirent*)buf; !error && dp < edp;
1193 dp = (struct dirent*)((caddr_t)dp + dp->d_reclen)) {
1194 if (dp->d_type == DT_WHT ||
1195 (dp->d_namlen == 1 && dp->d_name[0] == '.') ||
1196 (dp->d_namlen == 2 && !bcmp(dp->d_name, "..", 2)))
1197 continue;
1198
1199 cn.cn_namelen = dp->d_namlen;
1200 cn.cn_pnbuf = NULL;
1201 cn.cn_nameptr = dp->d_name;
1202 cn.cn_nameiop = LOOKUP;
1203 cn.cn_flags = (LOCKPARENT | LOCKLEAF | SAVENAME | RDONLY | ISLASTCN);
1204 cn.cn_lkflags = LK_EXCLUSIVE;
1205 cn.cn_thread = td;
1206 cn.cn_cred = cred;
1207 cn.cn_consume = 0;
1208
1209 /*
1210 * check entry in lower.
1211 * Sometimes, readdir function returns
1212 * wrong entry.
1213 */
1214 lookuperr = VOP_LOOKUP(lvp, &tvp, &cn);
1215
1216 if (!lookuperr)
1217 vput(tvp);
1218 else
1219 continue; /* skip entry */
1220
1221 /*
1222 * check entry
1223 * If it has no exist/whiteout entry in upper,
1224 * directory is not empty.
1225 */
1226 cn.cn_flags = (LOCKPARENT | LOCKLEAF | SAVENAME | RDONLY | ISLASTCN);
1227 lookuperr = VOP_LOOKUP(uvp, &tvp, &cn);
1228
1229 if (!lookuperr)
1230 vput(tvp);
1231
1232 /* ignore exist or whiteout entry */
1233 if (!lookuperr ||
1234 (lookuperr == ENOENT && (cn.cn_flags & ISWHITEOUT)))
1235 continue;
1236
1237 error = ENOTEMPTY;
1238 }
1239 }
1240
1241 /* close vnode */
1242 VOP_CLOSE(vp, FREAD, cred, td);
1243
1244 return (error);
1245}
1246
1247#ifdef DIAGNOSTIC
1248
1249struct vnode *
1250unionfs_checkuppervp(struct vnode *vp, char *fil, int lno)
1251{
1252 struct unionfs_node *unp;
1253
1254 unp = VTOUNIONFS(vp);
1255
1256#ifdef notyet
1257 if (vp->v_op != unionfs_vnodeop_p) {
1258 printf("unionfs_checkuppervp: on non-unionfs-node.\n");
1259#ifdef KDB
1260 kdb_enter(KDB_WHY_UNIONFS,
1261 "unionfs_checkuppervp: on non-unionfs-node.\n");
1262#endif
1263 panic("unionfs_checkuppervp");
1264 };
1265#endif
1266 return (unp->un_uppervp);
1267}
1268
1269struct vnode *
1270unionfs_checklowervp(struct vnode *vp, char *fil, int lno)
1271{
1272 struct unionfs_node *unp;
1273
1274 unp = VTOUNIONFS(vp);
1275
1276#ifdef notyet
1277 if (vp->v_op != unionfs_vnodeop_p) {
1278 printf("unionfs_checklowervp: on non-unionfs-node.\n");
1279#ifdef KDB
1280 kdb_enter(KDB_WHY_UNIONFS,
1281 "unionfs_checklowervp: on non-unionfs-node.\n");
1282#endif
1283 panic("unionfs_checklowervp");
1284 };
1285#endif
1286 return (unp->un_lowervp);
1287}
1288#endif