1/*
2 * CDDL HEADER START
3 *
4 * The contents of this file are subject to the terms of the
5 * Common Development and Distribution License (the "License").
6 * You may not use this file except in compliance with the License.
7 *
8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9 * or https://opensource.org/licenses/CDDL-1.0.
10 * See the License for the specific language governing permissions
11 * and limitations under the License.
12 *
13 * When distributing Covered Code, include this CDDL HEADER in each
14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15 * If applicable, add the following below this CDDL HEADER, with the
16 * fields enclosed by brackets "[]" replaced with your own identifying
17 * information: Portions Copyright [yyyy] [name of copyright owner]
18 *
19 * CDDL HEADER END
20 */
21/*
22 * Copyright (c) 2005, 2010, Oracle and/or its affiliates. All rights reserved.
23 * Copyright (c) 2015, 2018 by Delphix. All rights reserved.
24 * Copyright (c) 2022 by Pawel Jakub Dawidek
25 */
26
27
28#include <sys/types.h>
29#include <sys/param.h>
30#include <sys/sysmacros.h>
31#include <sys/cmn_err.h>
32#include <sys/kmem.h>
33#include <sys/thread.h>
34#include <sys/file.h>
35#include <sys/vfs.h>
36#include <sys/zfs_znode.h>
37#include <sys/zfs_dir.h>
38#include <sys/zil.h>
39#include <sys/zil_impl.h>
40#include <sys/byteorder.h>
41#include <sys/policy.h>
42#include <sys/stat.h>
43#include <sys/acl.h>
44#include <sys/dmu.h>
45#include <sys/dbuf.h>
46#include <sys/spa.h>
47#include <sys/zfs_fuid.h>
48#include <sys/dsl_dataset.h>
49
50/*
51 * These zfs_log_* functions must be called within a dmu tx, in one
52 * of 2 contexts depending on zilog->z_replay:
53 *
54 * Non replay mode
55 * ---------------
56 * We need to record the transaction so that if it is committed to
57 * the Intent Log then it can be replayed.  An intent log transaction
58 * structure (itx_t) is allocated and all the information necessary to
59 * possibly replay the transaction is saved in it. The itx is then assigned
60 * a sequence number and inserted in the in-memory list anchored in the zilog.
61 *
62 * Replay mode
63 * -----------
64 * We need to mark the intent log record as replayed in the log header.
65 * This is done in the same transaction as the replay so that they
66 * commit atomically.
67 */
68
69int
70zfs_log_create_txtype(zil_create_t type, vsecattr_t *vsecp, vattr_t *vap)
71{
72	int isxvattr = (vap->va_mask & ATTR_XVATTR);
73	switch (type) {
74	case Z_FILE:
75		if (vsecp == NULL && !isxvattr)
76			return (TX_CREATE);
77		if (vsecp && isxvattr)
78			return (TX_CREATE_ACL_ATTR);
79		if (vsecp)
80			return (TX_CREATE_ACL);
81		else
82			return (TX_CREATE_ATTR);
83	case Z_DIR:
84		if (vsecp == NULL && !isxvattr)
85			return (TX_MKDIR);
86		if (vsecp && isxvattr)
87			return (TX_MKDIR_ACL_ATTR);
88		if (vsecp)
89			return (TX_MKDIR_ACL);
90		else
91			return (TX_MKDIR_ATTR);
92	case Z_XATTRDIR:
93		return (TX_MKXATTR);
94	}
95	ASSERT(0);
96	return (TX_MAX_TYPE);
97}
98
99/*
100 * build up the log data necessary for logging xvattr_t
101 * First lr_attr_t is initialized.  following the lr_attr_t
102 * is the mapsize and attribute bitmap copied from the xvattr_t.
103 * Following the bitmap and bitmapsize two 64 bit words are reserved
104 * for the create time which may be set.  Following the create time
105 * records a single 64 bit integer which has the bits to set on
106 * replay for the xvattr.
107 */
108static void
109zfs_log_xvattr(lr_attr_t *lrattr, xvattr_t *xvap)
110{
111	xoptattr_t *xoap;
112
113	xoap = xva_getxoptattr(xvap);
114	ASSERT(xoap);
115
116	lrattr->lr_attr_masksize = xvap->xva_mapsize;
117	uint32_t *bitmap = &lrattr->lr_attr_bitmap;
118	for (int i = 0; i != xvap->xva_mapsize; i++, bitmap++)
119		*bitmap = xvap->xva_reqattrmap[i];
120
121	lr_attr_end_t *end = (lr_attr_end_t *)bitmap;
122	end->lr_attr_attrs = 0;
123	end->lr_attr_crtime[0] = 0;
124	end->lr_attr_crtime[1] = 0;
125	memset(end->lr_attr_scanstamp, 0, AV_SCANSTAMP_SZ);
126
127	if (XVA_ISSET_REQ(xvap, XAT_READONLY))
128		end->lr_attr_attrs |= (xoap->xoa_readonly == 0) ? 0 :
129		    XAT0_READONLY;
130	if (XVA_ISSET_REQ(xvap, XAT_HIDDEN))
131		end->lr_attr_attrs |= (xoap->xoa_hidden == 0) ? 0 :
132		    XAT0_HIDDEN;
133	if (XVA_ISSET_REQ(xvap, XAT_SYSTEM))
134		end->lr_attr_attrs |= (xoap->xoa_system == 0) ? 0 :
135		    XAT0_SYSTEM;
136	if (XVA_ISSET_REQ(xvap, XAT_ARCHIVE))
137		end->lr_attr_attrs |= (xoap->xoa_archive == 0) ? 0 :
138		    XAT0_ARCHIVE;
139	if (XVA_ISSET_REQ(xvap, XAT_IMMUTABLE))
140		end->lr_attr_attrs |= (xoap->xoa_immutable == 0) ? 0 :
141		    XAT0_IMMUTABLE;
142	if (XVA_ISSET_REQ(xvap, XAT_NOUNLINK))
143		end->lr_attr_attrs |= (xoap->xoa_nounlink == 0) ? 0 :
144		    XAT0_NOUNLINK;
145	if (XVA_ISSET_REQ(xvap, XAT_APPENDONLY))
146		end->lr_attr_attrs |= (xoap->xoa_appendonly == 0) ? 0 :
147		    XAT0_APPENDONLY;
148	if (XVA_ISSET_REQ(xvap, XAT_OPAQUE))
149		end->lr_attr_attrs |= (xoap->xoa_opaque == 0) ? 0 :
150		    XAT0_APPENDONLY;
151	if (XVA_ISSET_REQ(xvap, XAT_NODUMP))
152		end->lr_attr_attrs |= (xoap->xoa_nodump == 0) ? 0 :
153		    XAT0_NODUMP;
154	if (XVA_ISSET_REQ(xvap, XAT_AV_QUARANTINED))
155		end->lr_attr_attrs |= (xoap->xoa_av_quarantined == 0) ? 0 :
156		    XAT0_AV_QUARANTINED;
157	if (XVA_ISSET_REQ(xvap, XAT_AV_MODIFIED))
158		end->lr_attr_attrs |= (xoap->xoa_av_modified == 0) ? 0 :
159		    XAT0_AV_MODIFIED;
160	if (XVA_ISSET_REQ(xvap, XAT_CREATETIME))
161		ZFS_TIME_ENCODE(&xoap->xoa_createtime, end->lr_attr_crtime);
162	if (XVA_ISSET_REQ(xvap, XAT_AV_SCANSTAMP)) {
163		ASSERT(!XVA_ISSET_REQ(xvap, XAT_PROJID));
164
165		memcpy(end->lr_attr_scanstamp, xoap->xoa_av_scanstamp,
166		    AV_SCANSTAMP_SZ);
167	} else if (XVA_ISSET_REQ(xvap, XAT_PROJID)) {
168		/*
169		 * XAT_PROJID and XAT_AV_SCANSTAMP will never be valid
170		 * at the same time, so we can share the same space.
171		 */
172		memcpy(end->lr_attr_scanstamp, &xoap->xoa_projid,
173		    sizeof (uint64_t));
174	}
175	if (XVA_ISSET_REQ(xvap, XAT_REPARSE))
176		end->lr_attr_attrs |= (xoap->xoa_reparse == 0) ? 0 :
177		    XAT0_REPARSE;
178	if (XVA_ISSET_REQ(xvap, XAT_OFFLINE))
179		end->lr_attr_attrs |= (xoap->xoa_offline == 0) ? 0 :
180		    XAT0_OFFLINE;
181	if (XVA_ISSET_REQ(xvap, XAT_SPARSE))
182		end->lr_attr_attrs |= (xoap->xoa_sparse == 0) ? 0 :
183		    XAT0_SPARSE;
184	if (XVA_ISSET_REQ(xvap, XAT_PROJINHERIT))
185		end->lr_attr_attrs |= (xoap->xoa_projinherit == 0) ? 0 :
186		    XAT0_PROJINHERIT;
187}
188
189static void *
190zfs_log_fuid_ids(zfs_fuid_info_t *fuidp, void *start)
191{
192	zfs_fuid_t *zfuid;
193	uint64_t *fuidloc = start;
194
195	/* First copy in the ACE FUIDs */
196	for (zfuid = list_head(&fuidp->z_fuids); zfuid;
197	    zfuid = list_next(&fuidp->z_fuids, zfuid)) {
198		*fuidloc++ = zfuid->z_logfuid;
199	}
200	return (fuidloc);
201}
202
203
204static void *
205zfs_log_fuid_domains(zfs_fuid_info_t *fuidp, void *start)
206{
207	zfs_fuid_domain_t *zdomain;
208
209	/* now copy in the domain info, if any */
210	if (fuidp->z_domain_str_sz != 0) {
211		for (zdomain = list_head(&fuidp->z_domains); zdomain;
212		    zdomain = list_next(&fuidp->z_domains, zdomain)) {
213			memcpy(start, zdomain->z_domain,
214			    strlen(zdomain->z_domain) + 1);
215			start = (caddr_t)start +
216			    strlen(zdomain->z_domain) + 1;
217		}
218	}
219	return (start);
220}
221
222/*
223 * If zp is an xattr node, check whether the xattr owner is unlinked.
224 * We don't want to log anything if the owner is unlinked.
225 */
226static int
227zfs_xattr_owner_unlinked(znode_t *zp)
228{
229	int unlinked = 0;
230	znode_t *dzp;
231#ifdef __FreeBSD__
232	znode_t *tzp = zp;
233
234	/*
235	 * zrele drops the vnode lock which violates the VOP locking contract
236	 * on FreeBSD. See comment at the top of zfs_replay.c for more detail.
237	 */
238	/*
239	 * if zp is XATTR node, keep walking up via z_xattr_parent until we
240	 * get the owner
241	 */
242	while (tzp->z_pflags & ZFS_XATTR) {
243		ASSERT3U(zp->z_xattr_parent, !=, 0);
244		if (zfs_zget(ZTOZSB(tzp), tzp->z_xattr_parent, &dzp) != 0) {
245			unlinked = 1;
246			break;
247		}
248
249		if (tzp != zp)
250			zrele(tzp);
251		tzp = dzp;
252		unlinked = tzp->z_unlinked;
253	}
254	if (tzp != zp)
255		zrele(tzp);
256#else
257	zhold(zp);
258	/*
259	 * if zp is XATTR node, keep walking up via z_xattr_parent until we
260	 * get the owner
261	 */
262	while (zp->z_pflags & ZFS_XATTR) {
263		ASSERT3U(zp->z_xattr_parent, !=, 0);
264		if (zfs_zget(ZTOZSB(zp), zp->z_xattr_parent, &dzp) != 0) {
265			unlinked = 1;
266			break;
267		}
268
269		zrele(zp);
270		zp = dzp;
271		unlinked = zp->z_unlinked;
272	}
273	zrele(zp);
274#endif
275	return (unlinked);
276}
277
278/*
279 * Handles TX_CREATE, TX_CREATE_ATTR, TX_MKDIR, TX_MKDIR_ATTR and
280 * TK_MKXATTR transactions.
281 *
282 * TX_CREATE and TX_MKDIR are standard creates, but they may have FUID
283 * domain information appended prior to the name.  In this case the
284 * uid/gid in the log record will be a log centric FUID.
285 *
286 * TX_CREATE_ACL_ATTR and TX_MKDIR_ACL_ATTR handle special creates that
287 * may contain attributes, ACL and optional fuid information.
288 *
289 * TX_CREATE_ACL and TX_MKDIR_ACL handle special creates that specify
290 * and ACL and normal users/groups in the ACEs.
291 *
292 * There may be an optional xvattr attribute information similar
293 * to zfs_log_setattr.
294 *
295 * Also, after the file name "domain" strings may be appended.
296 */
297void
298zfs_log_create(zilog_t *zilog, dmu_tx_t *tx, uint64_t txtype,
299    znode_t *dzp, znode_t *zp, const char *name, vsecattr_t *vsecp,
300    zfs_fuid_info_t *fuidp, vattr_t *vap)
301{
302	itx_t *itx;
303	lr_create_t *lr;
304	lr_acl_create_t *lracl;
305	size_t aclsize = 0;
306	size_t xvatsize = 0;
307	size_t txsize;
308	xvattr_t *xvap = (xvattr_t *)vap;
309	void *end;
310	size_t lrsize;
311	size_t namesize = strlen(name) + 1;
312	size_t fuidsz = 0;
313
314	if (zil_replaying(zilog, tx) || zfs_xattr_owner_unlinked(dzp))
315		return;
316
317	/*
318	 * If we have FUIDs present then add in space for
319	 * domains and ACE fuid's if any.
320	 */
321	if (fuidp) {
322		fuidsz += fuidp->z_domain_str_sz;
323		fuidsz += fuidp->z_fuid_cnt * sizeof (uint64_t);
324	}
325
326	if (vap->va_mask & ATTR_XVATTR)
327		xvatsize = ZIL_XVAT_SIZE(xvap->xva_mapsize);
328
329	if ((int)txtype == TX_CREATE_ATTR || (int)txtype == TX_MKDIR_ATTR ||
330	    (int)txtype == TX_CREATE || (int)txtype == TX_MKDIR ||
331	    (int)txtype == TX_MKXATTR) {
332		txsize = sizeof (*lr) + namesize + fuidsz + xvatsize;
333		lrsize = sizeof (*lr);
334	} else {
335		txsize =
336		    sizeof (lr_acl_create_t) + namesize + fuidsz +
337		    ZIL_ACE_LENGTH(aclsize) + xvatsize;
338		lrsize = sizeof (lr_acl_create_t);
339	}
340
341	itx = zil_itx_create(txtype, txsize);
342
343	lr = (lr_create_t *)&itx->itx_lr;
344	lr->lr_doid = dzp->z_id;
345	lr->lr_foid = zp->z_id;
346	/* Store dnode slot count in 8 bits above object id. */
347	LR_FOID_SET_SLOTS(lr->lr_foid, zp->z_dnodesize >> DNODE_SHIFT);
348	lr->lr_mode = zp->z_mode;
349	if (!IS_EPHEMERAL(KUID_TO_SUID(ZTOUID(zp)))) {
350		lr->lr_uid = (uint64_t)KUID_TO_SUID(ZTOUID(zp));
351	} else {
352		lr->lr_uid = fuidp->z_fuid_owner;
353	}
354	if (!IS_EPHEMERAL(KGID_TO_SGID(ZTOGID(zp)))) {
355		lr->lr_gid = (uint64_t)KGID_TO_SGID(ZTOGID(zp));
356	} else {
357		lr->lr_gid = fuidp->z_fuid_group;
358	}
359	(void) sa_lookup(zp->z_sa_hdl, SA_ZPL_GEN(ZTOZSB(zp)), &lr->lr_gen,
360	    sizeof (uint64_t));
361	(void) sa_lookup(zp->z_sa_hdl, SA_ZPL_CRTIME(ZTOZSB(zp)),
362	    lr->lr_crtime, sizeof (uint64_t) * 2);
363
364	if (sa_lookup(zp->z_sa_hdl, SA_ZPL_RDEV(ZTOZSB(zp)), &lr->lr_rdev,
365	    sizeof (lr->lr_rdev)) != 0)
366		lr->lr_rdev = 0;
367
368	/*
369	 * Fill in xvattr info if any
370	 */
371	if (vap->va_mask & ATTR_XVATTR) {
372		zfs_log_xvattr((lr_attr_t *)((caddr_t)lr + lrsize), xvap);
373		end = (caddr_t)lr + lrsize + xvatsize;
374	} else {
375		end = (caddr_t)lr + lrsize;
376	}
377
378	/* Now fill in any ACL info */
379
380	if (vsecp) {
381		lracl = (lr_acl_create_t *)&itx->itx_lr;
382		lracl->lr_aclcnt = vsecp->vsa_aclcnt;
383		lracl->lr_acl_bytes = aclsize;
384		lracl->lr_domcnt = fuidp ? fuidp->z_domain_cnt : 0;
385		lracl->lr_fuidcnt  = fuidp ? fuidp->z_fuid_cnt : 0;
386		if (vsecp->vsa_aclflags & VSA_ACE_ACLFLAGS)
387			lracl->lr_acl_flags = (uint64_t)vsecp->vsa_aclflags;
388		else
389			lracl->lr_acl_flags = 0;
390
391		memcpy(end, vsecp->vsa_aclentp, aclsize);
392		end = (caddr_t)end + ZIL_ACE_LENGTH(aclsize);
393	}
394
395	/* drop in FUID info */
396	if (fuidp) {
397		end = zfs_log_fuid_ids(fuidp, end);
398		end = zfs_log_fuid_domains(fuidp, end);
399	}
400	/*
401	 * Now place file name in log record
402	 */
403	memcpy(end, name, namesize);
404
405	zil_itx_assign(zilog, itx, tx);
406}
407
408/*
409 * Handles both TX_REMOVE and TX_RMDIR transactions.
410 */
411void
412zfs_log_remove(zilog_t *zilog, dmu_tx_t *tx, uint64_t txtype,
413    znode_t *dzp, const char *name, uint64_t foid, boolean_t unlinked)
414{
415	itx_t *itx;
416	lr_remove_t *lr;
417	size_t namesize = strlen(name) + 1;
418
419	if (zil_replaying(zilog, tx) || zfs_xattr_owner_unlinked(dzp))
420		return;
421
422	itx = zil_itx_create(txtype, sizeof (*lr) + namesize);
423	lr = (lr_remove_t *)&itx->itx_lr;
424	lr->lr_doid = dzp->z_id;
425	memcpy(lr + 1, name, namesize);
426
427	itx->itx_oid = foid;
428
429	/*
430	 * Object ids can be re-instantiated in the next txg so
431	 * remove any async transactions to avoid future leaks.
432	 * This can happen if a fsync occurs on the re-instantiated
433	 * object for a WR_INDIRECT or WR_NEED_COPY write, which gets
434	 * the new file data and flushes a write record for the old object.
435	 */
436	if (unlinked) {
437		ASSERT((txtype & ~TX_CI) == TX_REMOVE);
438		zil_remove_async(zilog, foid);
439	}
440	zil_itx_assign(zilog, itx, tx);
441}
442
443/*
444 * Handles TX_LINK transactions.
445 */
446void
447zfs_log_link(zilog_t *zilog, dmu_tx_t *tx, uint64_t txtype,
448    znode_t *dzp, znode_t *zp, const char *name)
449{
450	itx_t *itx;
451	lr_link_t *lr;
452	size_t namesize = strlen(name) + 1;
453
454	if (zil_replaying(zilog, tx))
455		return;
456
457	itx = zil_itx_create(txtype, sizeof (*lr) + namesize);
458	lr = (lr_link_t *)&itx->itx_lr;
459	lr->lr_doid = dzp->z_id;
460	lr->lr_link_obj = zp->z_id;
461	memcpy(lr + 1, name, namesize);
462
463	zil_itx_assign(zilog, itx, tx);
464}
465
466/*
467 * Handles TX_SYMLINK transactions.
468 */
469void
470zfs_log_symlink(zilog_t *zilog, dmu_tx_t *tx, uint64_t txtype,
471    znode_t *dzp, znode_t *zp, const char *name, const char *link)
472{
473	itx_t *itx;
474	lr_create_t *lr;
475	size_t namesize = strlen(name) + 1;
476	size_t linksize = strlen(link) + 1;
477
478	if (zil_replaying(zilog, tx))
479		return;
480
481	itx = zil_itx_create(txtype, sizeof (*lr) + namesize + linksize);
482	lr = (lr_create_t *)&itx->itx_lr;
483	lr->lr_doid = dzp->z_id;
484	lr->lr_foid = zp->z_id;
485	lr->lr_uid = KUID_TO_SUID(ZTOUID(zp));
486	lr->lr_gid = KGID_TO_SGID(ZTOGID(zp));
487	lr->lr_mode = zp->z_mode;
488	(void) sa_lookup(zp->z_sa_hdl, SA_ZPL_GEN(ZTOZSB(zp)), &lr->lr_gen,
489	    sizeof (uint64_t));
490	(void) sa_lookup(zp->z_sa_hdl, SA_ZPL_CRTIME(ZTOZSB(zp)),
491	    lr->lr_crtime, sizeof (uint64_t) * 2);
492	memcpy((char *)(lr + 1), name, namesize);
493	memcpy((char *)(lr + 1) + namesize, link, linksize);
494
495	zil_itx_assign(zilog, itx, tx);
496}
497
498static void
499do_zfs_log_rename(zilog_t *zilog, dmu_tx_t *tx, uint64_t txtype, znode_t *sdzp,
500    const char *sname, znode_t *tdzp, const char *dname, znode_t *szp)
501{
502	itx_t *itx;
503	lr_rename_t *lr;
504	size_t snamesize = strlen(sname) + 1;
505	size_t dnamesize = strlen(dname) + 1;
506
507	if (zil_replaying(zilog, tx))
508		return;
509
510	itx = zil_itx_create(txtype, sizeof (*lr) + snamesize + dnamesize);
511	lr = (lr_rename_t *)&itx->itx_lr;
512	lr->lr_sdoid = sdzp->z_id;
513	lr->lr_tdoid = tdzp->z_id;
514	memcpy((char *)(lr + 1), sname, snamesize);
515	memcpy((char *)(lr + 1) + snamesize, dname, dnamesize);
516	itx->itx_oid = szp->z_id;
517
518	zil_itx_assign(zilog, itx, tx);
519}
520
521/*
522 * Handles TX_RENAME transactions.
523 */
524void
525zfs_log_rename(zilog_t *zilog, dmu_tx_t *tx, uint64_t txtype, znode_t *sdzp,
526    const char *sname, znode_t *tdzp, const char *dname, znode_t *szp)
527{
528	txtype |= TX_RENAME;
529	do_zfs_log_rename(zilog, tx, txtype, sdzp, sname, tdzp, dname, szp);
530}
531
532/*
533 * Handles TX_RENAME_EXCHANGE transactions.
534 */
535void
536zfs_log_rename_exchange(zilog_t *zilog, dmu_tx_t *tx, uint64_t txtype,
537    znode_t *sdzp, const char *sname, znode_t *tdzp, const char *dname,
538    znode_t *szp)
539{
540	txtype |= TX_RENAME_EXCHANGE;
541	do_zfs_log_rename(zilog, tx, txtype, sdzp, sname, tdzp, dname, szp);
542}
543
544/*
545 * Handles TX_RENAME_WHITEOUT transactions.
546 *
547 * Unfortunately we cannot reuse do_zfs_log_rename because we we need to call
548 * zfs_mknode() on replay which requires stashing bits as with TX_CREATE.
549 */
550void
551zfs_log_rename_whiteout(zilog_t *zilog, dmu_tx_t *tx, uint64_t txtype,
552    znode_t *sdzp, const char *sname, znode_t *tdzp, const char *dname,
553    znode_t *szp, znode_t *wzp)
554{
555	itx_t *itx;
556	lr_rename_whiteout_t *lr;
557	size_t snamesize = strlen(sname) + 1;
558	size_t dnamesize = strlen(dname) + 1;
559
560	if (zil_replaying(zilog, tx))
561		return;
562
563	txtype |= TX_RENAME_WHITEOUT;
564	itx = zil_itx_create(txtype, sizeof (*lr) + snamesize + dnamesize);
565	lr = (lr_rename_whiteout_t *)&itx->itx_lr;
566	lr->lr_rename.lr_sdoid = sdzp->z_id;
567	lr->lr_rename.lr_tdoid = tdzp->z_id;
568
569	/*
570	 * RENAME_WHITEOUT will create an entry at the source znode, so we need
571	 * to store the same data that the equivalent call to zfs_log_create()
572	 * would.
573	 */
574	lr->lr_wfoid = wzp->z_id;
575	LR_FOID_SET_SLOTS(lr->lr_wfoid, wzp->z_dnodesize >> DNODE_SHIFT);
576	(void) sa_lookup(wzp->z_sa_hdl, SA_ZPL_GEN(ZTOZSB(wzp)), &lr->lr_wgen,
577	    sizeof (uint64_t));
578	(void) sa_lookup(wzp->z_sa_hdl, SA_ZPL_CRTIME(ZTOZSB(wzp)),
579	    lr->lr_wcrtime, sizeof (uint64_t) * 2);
580	lr->lr_wmode = wzp->z_mode;
581	lr->lr_wuid = (uint64_t)KUID_TO_SUID(ZTOUID(wzp));
582	lr->lr_wgid = (uint64_t)KGID_TO_SGID(ZTOGID(wzp));
583
584	/*
585	 * This rdev will always be makdevice(0, 0) but because the ZIL log and
586	 * replay code needs to be platform independent (and there is no
587	 * platform independent makdev()) we need to copy the one created
588	 * during the rename operation.
589	 */
590	(void) sa_lookup(wzp->z_sa_hdl, SA_ZPL_RDEV(ZTOZSB(wzp)), &lr->lr_wrdev,
591	    sizeof (lr->lr_wrdev));
592
593	memcpy((char *)(lr + 1), sname, snamesize);
594	memcpy((char *)(lr + 1) + snamesize, dname, dnamesize);
595	itx->itx_oid = szp->z_id;
596
597	zil_itx_assign(zilog, itx, tx);
598}
599
600/*
601 * zfs_log_write() handles TX_WRITE transactions. The specified callback is
602 * called as soon as the write is on stable storage (be it via a DMU sync or a
603 * ZIL commit).
604 */
605static int64_t zfs_immediate_write_sz = 32768;
606
607void
608zfs_log_write(zilog_t *zilog, dmu_tx_t *tx, int txtype,
609    znode_t *zp, offset_t off, ssize_t resid, boolean_t commit,
610    zil_callback_t callback, void *callback_data)
611{
612	dmu_buf_impl_t *db = (dmu_buf_impl_t *)sa_get_db(zp->z_sa_hdl);
613	uint32_t blocksize = zp->z_blksz;
614	itx_wr_state_t write_state;
615	uint64_t gen = 0;
616	ssize_t size = resid;
617
618	if (zil_replaying(zilog, tx) || zp->z_unlinked ||
619	    zfs_xattr_owner_unlinked(zp)) {
620		if (callback != NULL)
621			callback(callback_data);
622		return;
623	}
624
625	if (zilog->zl_logbias == ZFS_LOGBIAS_THROUGHPUT)
626		write_state = WR_INDIRECT;
627	else if (!spa_has_slogs(zilog->zl_spa) &&
628	    resid >= zfs_immediate_write_sz)
629		write_state = WR_INDIRECT;
630	else if (commit)
631		write_state = WR_COPIED;
632	else
633		write_state = WR_NEED_COPY;
634
635	(void) sa_lookup(zp->z_sa_hdl, SA_ZPL_GEN(ZTOZSB(zp)), &gen,
636	    sizeof (gen));
637
638	while (resid) {
639		itx_t *itx;
640		lr_write_t *lr;
641		itx_wr_state_t wr_state = write_state;
642		ssize_t len = resid;
643
644		/*
645		 * A WR_COPIED record must fit entirely in one log block.
646		 * Large writes can use WR_NEED_COPY, which the ZIL will
647		 * split into multiple records across several log blocks
648		 * if necessary.
649		 */
650		if (wr_state == WR_COPIED &&
651		    resid > zil_max_copied_data(zilog))
652			wr_state = WR_NEED_COPY;
653		else if (wr_state == WR_INDIRECT)
654			len = MIN(blocksize - P2PHASE(off, blocksize), resid);
655
656		itx = zil_itx_create(txtype, sizeof (*lr) +
657		    (wr_state == WR_COPIED ? len : 0));
658		lr = (lr_write_t *)&itx->itx_lr;
659
660		/*
661		 * For WR_COPIED records, copy the data into the lr_write_t.
662		 */
663		if (wr_state == WR_COPIED) {
664			int err;
665			DB_DNODE_ENTER(db);
666			err = dmu_read_by_dnode(DB_DNODE(db), off, len, lr + 1,
667			    DMU_READ_NO_PREFETCH);
668			if (err != 0) {
669				zil_itx_destroy(itx);
670				itx = zil_itx_create(txtype, sizeof (*lr));
671				lr = (lr_write_t *)&itx->itx_lr;
672				wr_state = WR_NEED_COPY;
673			}
674			DB_DNODE_EXIT(db);
675		}
676
677		itx->itx_wr_state = wr_state;
678		lr->lr_foid = zp->z_id;
679		lr->lr_offset = off;
680		lr->lr_length = len;
681		lr->lr_blkoff = 0;
682		BP_ZERO(&lr->lr_blkptr);
683
684		itx->itx_private = ZTOZSB(zp);
685		itx->itx_sync = (zp->z_sync_cnt != 0);
686		itx->itx_gen = gen;
687
688		itx->itx_callback = callback;
689		itx->itx_callback_data = callback_data;
690		zil_itx_assign(zilog, itx, tx);
691
692		off += len;
693		resid -= len;
694	}
695
696	if (write_state == WR_COPIED || write_state == WR_NEED_COPY) {
697		dsl_pool_wrlog_count(zilog->zl_dmu_pool, size, tx->tx_txg);
698	}
699}
700
701/*
702 * Handles TX_TRUNCATE transactions.
703 */
704void
705zfs_log_truncate(zilog_t *zilog, dmu_tx_t *tx, int txtype,
706    znode_t *zp, uint64_t off, uint64_t len)
707{
708	itx_t *itx;
709	lr_truncate_t *lr;
710
711	if (zil_replaying(zilog, tx) || zp->z_unlinked ||
712	    zfs_xattr_owner_unlinked(zp))
713		return;
714
715	itx = zil_itx_create(txtype, sizeof (*lr));
716	lr = (lr_truncate_t *)&itx->itx_lr;
717	lr->lr_foid = zp->z_id;
718	lr->lr_offset = off;
719	lr->lr_length = len;
720
721	itx->itx_sync = (zp->z_sync_cnt != 0);
722	zil_itx_assign(zilog, itx, tx);
723}
724
725/*
726 * Handles TX_SETATTR transactions.
727 */
728void
729zfs_log_setattr(zilog_t *zilog, dmu_tx_t *tx, int txtype,
730    znode_t *zp, vattr_t *vap, uint_t mask_applied, zfs_fuid_info_t *fuidp)
731{
732	itx_t		*itx;
733	lr_setattr_t	*lr;
734	xvattr_t	*xvap = (xvattr_t *)vap;
735	size_t		recsize = sizeof (lr_setattr_t);
736	void		*start;
737
738	if (zil_replaying(zilog, tx) || zp->z_unlinked)
739		return;
740
741	/*
742	 * If XVATTR set, then log record size needs to allow
743	 * for lr_attr_t + xvattr mask, mapsize and create time
744	 * plus actual attribute values
745	 */
746	if (vap->va_mask & ATTR_XVATTR)
747		recsize = sizeof (*lr) + ZIL_XVAT_SIZE(xvap->xva_mapsize);
748
749	if (fuidp)
750		recsize += fuidp->z_domain_str_sz;
751
752	itx = zil_itx_create(txtype, recsize);
753	lr = (lr_setattr_t *)&itx->itx_lr;
754	lr->lr_foid = zp->z_id;
755	lr->lr_mask = (uint64_t)mask_applied;
756	lr->lr_mode = (uint64_t)vap->va_mode;
757	if ((mask_applied & ATTR_UID) && IS_EPHEMERAL(vap->va_uid))
758		lr->lr_uid = fuidp->z_fuid_owner;
759	else
760		lr->lr_uid = (uint64_t)vap->va_uid;
761
762	if ((mask_applied & ATTR_GID) && IS_EPHEMERAL(vap->va_gid))
763		lr->lr_gid = fuidp->z_fuid_group;
764	else
765		lr->lr_gid = (uint64_t)vap->va_gid;
766
767	lr->lr_size = (uint64_t)vap->va_size;
768	ZFS_TIME_ENCODE(&vap->va_atime, lr->lr_atime);
769	ZFS_TIME_ENCODE(&vap->va_mtime, lr->lr_mtime);
770	start = (lr_setattr_t *)(lr + 1);
771	if (vap->va_mask & ATTR_XVATTR) {
772		zfs_log_xvattr((lr_attr_t *)start, xvap);
773		start = (caddr_t)start + ZIL_XVAT_SIZE(xvap->xva_mapsize);
774	}
775
776	/*
777	 * Now stick on domain information if any on end
778	 */
779
780	if (fuidp)
781		(void) zfs_log_fuid_domains(fuidp, start);
782
783	itx->itx_sync = (zp->z_sync_cnt != 0);
784	zil_itx_assign(zilog, itx, tx);
785}
786
787/*
788 * Handles TX_SETSAXATTR transactions.
789 */
790void
791zfs_log_setsaxattr(zilog_t *zilog, dmu_tx_t *tx, int txtype,
792    znode_t *zp, const char *name, const void *value, size_t size)
793{
794	itx_t		*itx;
795	lr_setsaxattr_t	*lr;
796	size_t		recsize = sizeof (lr_setsaxattr_t);
797	void		*xattrstart;
798	int		namelen;
799
800	if (zil_replaying(zilog, tx) || zp->z_unlinked)
801		return;
802
803	namelen = strlen(name) + 1;
804	recsize += (namelen + size);
805	itx = zil_itx_create(txtype, recsize);
806	lr = (lr_setsaxattr_t *)&itx->itx_lr;
807	lr->lr_foid = zp->z_id;
808	xattrstart = (char *)(lr + 1);
809	memcpy(xattrstart, name, namelen);
810	if (value != NULL) {
811		memcpy((char *)xattrstart + namelen, value, size);
812		lr->lr_size = size;
813	} else {
814		lr->lr_size = 0;
815	}
816
817	itx->itx_sync = (zp->z_sync_cnt != 0);
818	zil_itx_assign(zilog, itx, tx);
819}
820
821/*
822 * Handles TX_ACL transactions.
823 */
824void
825zfs_log_acl(zilog_t *zilog, dmu_tx_t *tx, znode_t *zp,
826    vsecattr_t *vsecp, zfs_fuid_info_t *fuidp)
827{
828	itx_t *itx;
829	lr_acl_v0_t *lrv0;
830	lr_acl_t *lr;
831	int txtype;
832	int lrsize;
833	size_t txsize;
834	size_t aclbytes = vsecp->vsa_aclentsz;
835
836	if (zil_replaying(zilog, tx) || zp->z_unlinked)
837		return;
838
839	txtype = (ZTOZSB(zp)->z_version < ZPL_VERSION_FUID) ?
840	    TX_ACL_V0 : TX_ACL;
841
842	if (txtype == TX_ACL)
843		lrsize = sizeof (*lr);
844	else
845		lrsize = sizeof (*lrv0);
846
847	txsize = lrsize +
848	    ((txtype == TX_ACL) ? ZIL_ACE_LENGTH(aclbytes) : aclbytes) +
849	    (fuidp ? fuidp->z_domain_str_sz : 0) +
850	    sizeof (uint64_t) * (fuidp ? fuidp->z_fuid_cnt : 0);
851
852	itx = zil_itx_create(txtype, txsize);
853
854	lr = (lr_acl_t *)&itx->itx_lr;
855	lr->lr_foid = zp->z_id;
856	if (txtype == TX_ACL) {
857		lr->lr_acl_bytes = aclbytes;
858		lr->lr_domcnt = fuidp ? fuidp->z_domain_cnt : 0;
859		lr->lr_fuidcnt = fuidp ? fuidp->z_fuid_cnt : 0;
860		if (vsecp->vsa_mask & VSA_ACE_ACLFLAGS)
861			lr->lr_acl_flags = (uint64_t)vsecp->vsa_aclflags;
862		else
863			lr->lr_acl_flags = 0;
864	}
865	lr->lr_aclcnt = (uint64_t)vsecp->vsa_aclcnt;
866
867	if (txtype == TX_ACL_V0) {
868		lrv0 = (lr_acl_v0_t *)lr;
869		memcpy(lrv0 + 1, vsecp->vsa_aclentp, aclbytes);
870	} else {
871		void *start = (ace_t *)(lr + 1);
872
873		memcpy(start, vsecp->vsa_aclentp, aclbytes);
874
875		start = (caddr_t)start + ZIL_ACE_LENGTH(aclbytes);
876
877		if (fuidp) {
878			start = zfs_log_fuid_ids(fuidp, start);
879			(void) zfs_log_fuid_domains(fuidp, start);
880		}
881	}
882
883	itx->itx_sync = (zp->z_sync_cnt != 0);
884	zil_itx_assign(zilog, itx, tx);
885}
886
887/*
888 * Handles TX_CLONE_RANGE transactions.
889 */
890void
891zfs_log_clone_range(zilog_t *zilog, dmu_tx_t *tx, int txtype, znode_t *zp,
892    uint64_t off, uint64_t len, uint64_t blksz, const blkptr_t *bps,
893    size_t nbps)
894{
895	itx_t *itx;
896	lr_clone_range_t *lr;
897	uint64_t partlen, max_log_data;
898	size_t partnbps;
899
900	if (zil_replaying(zilog, tx) || zp->z_unlinked)
901		return;
902
903	max_log_data = zil_max_log_data(zilog, sizeof (lr_clone_range_t));
904
905	while (nbps > 0) {
906		partnbps = MIN(nbps, max_log_data / sizeof (bps[0]));
907		partlen = partnbps * blksz;
908		ASSERT3U(partlen, <, len + blksz);
909		partlen = MIN(partlen, len);
910
911		itx = zil_itx_create(txtype,
912		    sizeof (*lr) + sizeof (bps[0]) * partnbps);
913		lr = (lr_clone_range_t *)&itx->itx_lr;
914		lr->lr_foid = zp->z_id;
915		lr->lr_offset = off;
916		lr->lr_length = partlen;
917		lr->lr_blksz = blksz;
918		lr->lr_nbps = partnbps;
919		memcpy(lr->lr_bps, bps, sizeof (bps[0]) * partnbps);
920
921		itx->itx_sync = (zp->z_sync_cnt != 0);
922
923		zil_itx_assign(zilog, itx, tx);
924
925		bps += partnbps;
926		ASSERT3U(nbps, >=, partnbps);
927		nbps -= partnbps;
928		off += partlen;
929		ASSERT3U(len, >=, partlen);
930		len -= partlen;
931	}
932}
933
934ZFS_MODULE_PARAM(zfs, zfs_, immediate_write_sz, S64, ZMOD_RW,
935	"Largest data block to write to zil");
936