uu_list.c revision 168404
1/*
2 * CDDL HEADER START
3 *
4 * The contents of this file are subject to the terms of the
5 * Common Development and Distribution License, Version 1.0 only
6 * (the "License").  You may not use this file except in compliance
7 * with the License.
8 *
9 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
10 * or http://www.opensolaris.org/os/licensing.
11 * See the License for the specific language governing permissions
12 * and limitations under the License.
13 *
14 * When distributing Covered Code, include this CDDL HEADER in each
15 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
16 * If applicable, add the following below this CDDL HEADER, with the
17 * fields enclosed by brackets "[]" replaced with your own identifying
18 * information: Portions Copyright [yyyy] [name of copyright owner]
19 *
20 * CDDL HEADER END
21 */
22/*
23 * Copyright 2005 Sun Microsystems, Inc.  All rights reserved.
24 * Use is subject to license terms.
25 */
26
27#pragma ident	"%Z%%M%	%I%	%E% SMI"
28
29#include "libuutil_common.h"
30
31#include <stdlib.h>
32#include <string.h>
33#include <unistd.h>
34#include <sys/time.h>
35
36#define	ELEM_TO_NODE(lp, e) \
37	((uu_list_node_impl_t *)((uintptr_t)(e) + (lp)->ul_offset))
38
39#define	NODE_TO_ELEM(lp, n) \
40	((void *)((uintptr_t)(n) - (lp)->ul_offset))
41
42/*
43 * uu_list_index_ts define a location for insertion.  They are simply a
44 * pointer to the object after the insertion point.  We store a mark
45 * in the low-bits of the index, to help prevent mistakes.
46 *
47 * When debugging, the index mark changes on every insert and delete, to
48 * catch stale references.
49 */
50#define	INDEX_MAX		(sizeof (uintptr_t) - 1)
51#define	INDEX_NEXT(m)		(((m) == INDEX_MAX)? 1 : ((m) + 1) & INDEX_MAX)
52
53#define	INDEX_TO_NODE(i)	((uu_list_node_impl_t *)((i) & ~INDEX_MAX))
54#define	NODE_TO_INDEX(p, n)	(((uintptr_t)(n) & ~INDEX_MAX) | (p)->ul_index)
55#define	INDEX_VALID(p, i)	(((i) & INDEX_MAX) == (p)->ul_index)
56#define	INDEX_CHECK(i)		(((i) & INDEX_MAX) != 0)
57
58#define	POOL_TO_MARKER(pp) ((void *)((uintptr_t)(pp) | 1))
59
60static uu_list_pool_t	uu_null_lpool = { &uu_null_lpool, &uu_null_lpool };
61static pthread_mutex_t	uu_lpool_list_lock = PTHREAD_MUTEX_INITIALIZER;
62
63uu_list_pool_t *
64uu_list_pool_create(const char *name, size_t objsize,
65    size_t nodeoffset, uu_compare_fn_t *compare_func, uint32_t flags)
66{
67	uu_list_pool_t *pp, *next, *prev;
68
69	if (name == NULL ||
70	    uu_check_name(name, UU_NAME_DOMAIN) == -1 ||
71	    nodeoffset + sizeof (uu_list_node_t) > objsize) {
72		uu_set_error(UU_ERROR_INVALID_ARGUMENT);
73		return (NULL);
74	}
75
76	if (flags & ~UU_LIST_POOL_DEBUG) {
77		uu_set_error(UU_ERROR_UNKNOWN_FLAG);
78		return (NULL);
79	}
80
81	pp = uu_zalloc(sizeof (uu_list_pool_t));
82	if (pp == NULL) {
83		uu_set_error(UU_ERROR_NO_MEMORY);
84		return (NULL);
85	}
86
87	(void) strlcpy(pp->ulp_name, name, sizeof (pp->ulp_name));
88	pp->ulp_nodeoffset = nodeoffset;
89	pp->ulp_objsize = objsize;
90	pp->ulp_cmp = compare_func;
91	if (flags & UU_LIST_POOL_DEBUG)
92		pp->ulp_debug = 1;
93	pp->ulp_last_index = 0;
94
95	(void) pthread_mutex_init(&pp->ulp_lock, NULL);
96
97	pp->ulp_null_list.ul_next_enc = UU_PTR_ENCODE(&pp->ulp_null_list);
98	pp->ulp_null_list.ul_prev_enc = UU_PTR_ENCODE(&pp->ulp_null_list);
99
100	(void) pthread_mutex_lock(&uu_lpool_list_lock);
101	pp->ulp_next = next = &uu_null_lpool;
102	pp->ulp_prev = prev = next->ulp_prev;
103	next->ulp_prev = pp;
104	prev->ulp_next = pp;
105	(void) pthread_mutex_unlock(&uu_lpool_list_lock);
106
107	return (pp);
108}
109
110void
111uu_list_pool_destroy(uu_list_pool_t *pp)
112{
113	if (pp->ulp_debug) {
114		if (pp->ulp_null_list.ul_next_enc !=
115		    UU_PTR_ENCODE(&pp->ulp_null_list) ||
116		    pp->ulp_null_list.ul_prev_enc !=
117		    UU_PTR_ENCODE(&pp->ulp_null_list)) {
118			uu_panic("uu_list_pool_destroy: Pool \"%.*s\" (%p) has "
119			    "outstanding lists, or is corrupt.\n",
120			    sizeof (pp->ulp_name), pp->ulp_name, pp);
121		}
122	}
123	(void) pthread_mutex_lock(&uu_lpool_list_lock);
124	pp->ulp_next->ulp_prev = pp->ulp_prev;
125	pp->ulp_prev->ulp_next = pp->ulp_next;
126	(void) pthread_mutex_unlock(&uu_lpool_list_lock);
127	pp->ulp_prev = NULL;
128	pp->ulp_next = NULL;
129	uu_free(pp);
130}
131
132void
133uu_list_node_init(void *base, uu_list_node_t *np_arg, uu_list_pool_t *pp)
134{
135	uu_list_node_impl_t *np = (uu_list_node_impl_t *)np_arg;
136
137	if (pp->ulp_debug) {
138		uintptr_t offset = (uintptr_t)np - (uintptr_t)base;
139		if (offset + sizeof (*np) > pp->ulp_objsize) {
140			uu_panic("uu_list_node_init(%p, %p, %p (\"%s\")): "
141			    "offset %ld doesn't fit in object (size %ld)\n",
142			    base, np, pp, pp->ulp_name, offset,
143			    pp->ulp_objsize);
144		}
145		if (offset != pp->ulp_nodeoffset) {
146			uu_panic("uu_list_node_init(%p, %p, %p (\"%s\")): "
147			    "offset %ld doesn't match pool's offset (%ld)\n",
148			    base, np, pp, pp->ulp_name, offset,
149			    pp->ulp_objsize);
150		}
151	}
152	np->uln_next = POOL_TO_MARKER(pp);
153	np->uln_prev = NULL;
154}
155
156void
157uu_list_node_fini(void *base, uu_list_node_t *np_arg, uu_list_pool_t *pp)
158{
159	uu_list_node_impl_t *np = (uu_list_node_impl_t *)np_arg;
160
161	if (pp->ulp_debug) {
162		if (np->uln_next == NULL &&
163		    np->uln_prev == NULL) {
164			uu_panic("uu_list_node_fini(%p, %p, %p (\"%s\")): "
165			    "node already finied\n",
166			    base, np_arg, pp, pp->ulp_name);
167		}
168		if (np->uln_next != POOL_TO_MARKER(pp) ||
169		    np->uln_prev != NULL) {
170			uu_panic("uu_list_node_fini(%p, %p, %p (\"%s\")): "
171			    "node corrupt or on list\n",
172			    base, np_arg, pp, pp->ulp_name);
173		}
174	}
175	np->uln_next = NULL;
176	np->uln_prev = NULL;
177}
178
179uu_list_t *
180uu_list_create(uu_list_pool_t *pp, void *parent, uint32_t flags)
181{
182	uu_list_t *lp, *next, *prev;
183
184	if (flags & ~(UU_LIST_DEBUG | UU_LIST_SORTED)) {
185		uu_set_error(UU_ERROR_UNKNOWN_FLAG);
186		return (NULL);
187	}
188
189	if ((flags & UU_LIST_SORTED) && pp->ulp_cmp == NULL) {
190		if (pp->ulp_debug)
191			uu_panic("uu_list_create(%p, ...): requested "
192			    "UU_LIST_SORTED, but pool has no comparison func\n",
193			    pp);
194		uu_set_error(UU_ERROR_NOT_SUPPORTED);
195		return (NULL);
196	}
197
198	lp = uu_zalloc(sizeof (*lp));
199	if (lp == NULL) {
200		uu_set_error(UU_ERROR_NO_MEMORY);
201		return (NULL);
202	}
203
204	lp->ul_pool = pp;
205	lp->ul_parent_enc = UU_PTR_ENCODE(parent);
206	lp->ul_offset = pp->ulp_nodeoffset;
207	lp->ul_debug = pp->ulp_debug || (flags & UU_LIST_DEBUG);
208	lp->ul_sorted = (flags & UU_LIST_SORTED);
209	lp->ul_numnodes = 0;
210	lp->ul_index = (pp->ulp_last_index = INDEX_NEXT(pp->ulp_last_index));
211
212	lp->ul_null_node.uln_next = &lp->ul_null_node;
213	lp->ul_null_node.uln_prev = &lp->ul_null_node;
214
215	lp->ul_null_walk.ulw_next = &lp->ul_null_walk;
216	lp->ul_null_walk.ulw_prev = &lp->ul_null_walk;
217
218	(void) pthread_mutex_lock(&pp->ulp_lock);
219	next = &pp->ulp_null_list;
220	prev = UU_PTR_DECODE(next->ul_prev_enc);
221	lp->ul_next_enc = UU_PTR_ENCODE(next);
222	lp->ul_prev_enc = UU_PTR_ENCODE(prev);
223	next->ul_prev_enc = UU_PTR_ENCODE(lp);
224	prev->ul_next_enc = UU_PTR_ENCODE(lp);
225	(void) pthread_mutex_unlock(&pp->ulp_lock);
226
227	return (lp);
228}
229
230void
231uu_list_destroy(uu_list_t *lp)
232{
233	uu_list_pool_t *pp = lp->ul_pool;
234
235	if (lp->ul_debug) {
236		if (lp->ul_null_node.uln_next != &lp->ul_null_node ||
237		    lp->ul_null_node.uln_prev != &lp->ul_null_node) {
238			uu_panic("uu_list_destroy(%p):  list not empty\n",
239			    lp);
240		}
241		if (lp->ul_numnodes != 0) {
242			uu_panic("uu_list_destroy(%p):  numnodes is nonzero, "
243			    "but list is empty\n", lp);
244		}
245		if (lp->ul_null_walk.ulw_next != &lp->ul_null_walk ||
246		    lp->ul_null_walk.ulw_prev != &lp->ul_null_walk) {
247			uu_panic("uu_list_destroy(%p):  outstanding walkers\n",
248			    lp);
249		}
250	}
251
252	(void) pthread_mutex_lock(&pp->ulp_lock);
253	UU_LIST_PTR(lp->ul_next_enc)->ul_prev_enc = lp->ul_prev_enc;
254	UU_LIST_PTR(lp->ul_prev_enc)->ul_next_enc = lp->ul_next_enc;
255	(void) pthread_mutex_unlock(&pp->ulp_lock);
256	lp->ul_prev_enc = UU_PTR_ENCODE(NULL);
257	lp->ul_next_enc = UU_PTR_ENCODE(NULL);
258	lp->ul_pool = NULL;
259	uu_free(lp);
260}
261
262static void
263list_insert(uu_list_t *lp, uu_list_node_impl_t *np, uu_list_node_impl_t *prev,
264    uu_list_node_impl_t *next)
265{
266	if (lp->ul_debug) {
267		if (next->uln_prev != prev || prev->uln_next != next)
268			uu_panic("insert(%p): internal error: %p and %p not "
269			    "neighbors\n", lp, next, prev);
270
271		if (np->uln_next != POOL_TO_MARKER(lp->ul_pool) ||
272		    np->uln_prev != NULL) {
273			uu_panic("insert(%p): elem %p node %p corrupt, "
274			    "not initialized, or already in a list.\n",
275			    lp, NODE_TO_ELEM(lp, np), np);
276		}
277		/*
278		 * invalidate outstanding uu_list_index_ts.
279		 */
280		lp->ul_index = INDEX_NEXT(lp->ul_index);
281	}
282	np->uln_next = next;
283	np->uln_prev = prev;
284	next->uln_prev = np;
285	prev->uln_next = np;
286
287	lp->ul_numnodes++;
288}
289
290void
291uu_list_insert(uu_list_t *lp, void *elem, uu_list_index_t idx)
292{
293	uu_list_node_impl_t *np;
294
295	np = INDEX_TO_NODE(idx);
296	if (np == NULL)
297		np = &lp->ul_null_node;
298
299	if (lp->ul_debug) {
300		if (!INDEX_VALID(lp, idx))
301			uu_panic("uu_list_insert(%p, %p, %p): %s\n",
302			    lp, elem, idx,
303			    INDEX_CHECK(idx)? "outdated index" :
304			    "invalid index");
305		if (np->uln_prev == NULL)
306			uu_panic("uu_list_insert(%p, %p, %p): out-of-date "
307			    "index\n", lp, elem, idx);
308	}
309
310	list_insert(lp, ELEM_TO_NODE(lp, elem), np->uln_prev, np);
311}
312
313void *
314uu_list_find(uu_list_t *lp, void *elem, void *private, uu_list_index_t *out)
315{
316	int sorted = lp->ul_sorted;
317	uu_compare_fn_t *func = lp->ul_pool->ulp_cmp;
318	uu_list_node_impl_t *np;
319
320	if (func == NULL) {
321		if (out != NULL)
322			*out = 0;
323		uu_set_error(UU_ERROR_NOT_SUPPORTED);
324		return (NULL);
325	}
326	for (np = lp->ul_null_node.uln_next; np != &lp->ul_null_node;
327	    np = np->uln_next) {
328		void *ep = NODE_TO_ELEM(lp, np);
329		int cmp = func(ep, elem, private);
330		if (cmp == 0) {
331			if (out != NULL)
332				*out = NODE_TO_INDEX(lp, np);
333			return (ep);
334		}
335		if (sorted && cmp > 0) {
336			if (out != NULL)
337				*out = NODE_TO_INDEX(lp, np);
338			return (NULL);
339		}
340	}
341	if (out != NULL)
342		*out = NODE_TO_INDEX(lp, 0);
343	return (NULL);
344}
345
346void *
347uu_list_nearest_next(uu_list_t *lp, uu_list_index_t idx)
348{
349	uu_list_node_impl_t *np = INDEX_TO_NODE(idx);
350
351	if (np == NULL)
352		np = &lp->ul_null_node;
353
354	if (lp->ul_debug) {
355		if (!INDEX_VALID(lp, idx))
356			uu_panic("uu_list_nearest_next(%p, %p): %s\n",
357			    lp, idx, INDEX_CHECK(idx)? "outdated index" :
358			    "invalid index");
359		if (np->uln_prev == NULL)
360			uu_panic("uu_list_nearest_next(%p, %p): out-of-date "
361			    "index\n", lp, idx);
362	}
363
364	if (np == &lp->ul_null_node)
365		return (NULL);
366	else
367		return (NODE_TO_ELEM(lp, np));
368}
369
370void *
371uu_list_nearest_prev(uu_list_t *lp, uu_list_index_t idx)
372{
373	uu_list_node_impl_t *np = INDEX_TO_NODE(idx);
374
375	if (np == NULL)
376		np = &lp->ul_null_node;
377
378	if (lp->ul_debug) {
379		if (!INDEX_VALID(lp, idx))
380			uu_panic("uu_list_nearest_prev(%p, %p): %s\n",
381			    lp, idx, INDEX_CHECK(idx)? "outdated index" :
382			    "invalid index");
383		if (np->uln_prev == NULL)
384			uu_panic("uu_list_nearest_prev(%p, %p): out-of-date "
385			    "index\n", lp, idx);
386	}
387
388	if ((np = np->uln_prev) == &lp->ul_null_node)
389		return (NULL);
390	else
391		return (NODE_TO_ELEM(lp, np));
392}
393
394static void
395list_walk_init(uu_list_walk_t *wp, uu_list_t *lp, uint32_t flags)
396{
397	uu_list_walk_t *next, *prev;
398
399	int robust = (flags & UU_WALK_ROBUST);
400	int direction = (flags & UU_WALK_REVERSE)? -1 : 1;
401
402	(void) memset(wp, 0, sizeof (*wp));
403	wp->ulw_list = lp;
404	wp->ulw_robust = robust;
405	wp->ulw_dir = direction;
406	if (direction > 0)
407		wp->ulw_next_result = lp->ul_null_node.uln_next;
408	else
409		wp->ulw_next_result = lp->ul_null_node.uln_prev;
410
411	if (lp->ul_debug || robust) {
412		wp->ulw_next = next = &lp->ul_null_walk;
413		wp->ulw_prev = prev = next->ulw_prev;
414		next->ulw_prev = wp;
415		prev->ulw_next = wp;
416	}
417}
418
419static uu_list_node_impl_t *
420list_walk_advance(uu_list_walk_t *wp, uu_list_t *lp)
421{
422	uu_list_node_impl_t *np = wp->ulw_next_result;
423	uu_list_node_impl_t *next;
424
425	if (np == &lp->ul_null_node)
426		return (NULL);
427
428	next = (wp->ulw_dir > 0)? np->uln_next : np->uln_prev;
429
430	wp->ulw_next_result = next;
431	return (np);
432}
433
434static void
435list_walk_fini(uu_list_walk_t *wp)
436{
437	/* GLXXX debugging? */
438	if (wp->ulw_next != NULL) {
439		wp->ulw_next->ulw_prev = wp->ulw_prev;
440		wp->ulw_prev->ulw_next = wp->ulw_next;
441		wp->ulw_next = NULL;
442		wp->ulw_prev = NULL;
443	}
444	wp->ulw_list = NULL;
445	wp->ulw_next_result = NULL;
446}
447
448uu_list_walk_t *
449uu_list_walk_start(uu_list_t *lp, uint32_t flags)
450{
451	uu_list_walk_t *wp;
452
453	if (flags & ~(UU_WALK_ROBUST | UU_WALK_REVERSE)) {
454		uu_set_error(UU_ERROR_UNKNOWN_FLAG);
455		return (NULL);
456	}
457
458	wp = uu_zalloc(sizeof (*wp));
459	if (wp == NULL) {
460		uu_set_error(UU_ERROR_NO_MEMORY);
461		return (NULL);
462	}
463
464	list_walk_init(wp, lp, flags);
465	return (wp);
466}
467
468void *
469uu_list_walk_next(uu_list_walk_t *wp)
470{
471	uu_list_t *lp = wp->ulw_list;
472	uu_list_node_impl_t *np = list_walk_advance(wp, lp);
473
474	if (np == NULL)
475		return (NULL);
476
477	return (NODE_TO_ELEM(lp, np));
478}
479
480void
481uu_list_walk_end(uu_list_walk_t *wp)
482{
483	list_walk_fini(wp);
484	uu_free(wp);
485}
486
487int
488uu_list_walk(uu_list_t *lp, uu_walk_fn_t *func, void *private, uint32_t flags)
489{
490	uu_list_node_impl_t *np;
491
492	int status = UU_WALK_NEXT;
493
494	int robust = (flags & UU_WALK_ROBUST);
495	int reverse = (flags & UU_WALK_REVERSE);
496
497	if (flags & ~(UU_WALK_ROBUST | UU_WALK_REVERSE)) {
498		uu_set_error(UU_ERROR_UNKNOWN_FLAG);
499		return (-1);
500	}
501
502	if (lp->ul_debug || robust) {
503		uu_list_walk_t my_walk;
504		void *e;
505
506		list_walk_init(&my_walk, lp, flags);
507		while (status == UU_WALK_NEXT &&
508		    (e = uu_list_walk_next(&my_walk)) != NULL)
509			status = (*func)(e, private);
510		list_walk_fini(&my_walk);
511	} else {
512		if (!reverse) {
513			for (np = lp->ul_null_node.uln_next;
514			    status == UU_WALK_NEXT && np != &lp->ul_null_node;
515			    np = np->uln_next) {
516				status = (*func)(NODE_TO_ELEM(lp, np), private);
517			}
518		} else {
519			for (np = lp->ul_null_node.uln_prev;
520			    status == UU_WALK_NEXT && np != &lp->ul_null_node;
521			    np = np->uln_prev) {
522				status = (*func)(NODE_TO_ELEM(lp, np), private);
523			}
524		}
525	}
526	if (status >= 0)
527		return (0);
528	uu_set_error(UU_ERROR_CALLBACK_FAILED);
529	return (-1);
530}
531
532void
533uu_list_remove(uu_list_t *lp, void *elem)
534{
535	uu_list_node_impl_t *np = ELEM_TO_NODE(lp, elem);
536	uu_list_walk_t *wp;
537
538	if (lp->ul_debug) {
539		if (np->uln_prev == NULL)
540			uu_panic("uu_list_remove(%p, %p): elem not on list\n",
541			    lp, elem);
542		/*
543		 * invalidate outstanding uu_list_index_ts.
544		 */
545		lp->ul_index = INDEX_NEXT(lp->ul_index);
546	}
547
548	/*
549	 * robust walkers must be advanced.  In debug mode, non-robust
550	 * walkers are also on the list.  If there are any, it's an error.
551	 */
552	for (wp = lp->ul_null_walk.ulw_next; wp != &lp->ul_null_walk;
553	    wp = wp->ulw_next) {
554		if (wp->ulw_robust) {
555			if (np == wp->ulw_next_result)
556				(void) list_walk_advance(wp, lp);
557		} else if (wp->ulw_next_result != NULL) {
558			uu_panic("uu_list_remove(%p, %p): active non-robust "
559			    "walker\n", lp, elem);
560		}
561	}
562
563	np->uln_next->uln_prev = np->uln_prev;
564	np->uln_prev->uln_next = np->uln_next;
565
566	lp->ul_numnodes--;
567
568	np->uln_next = POOL_TO_MARKER(lp->ul_pool);
569	np->uln_prev = NULL;
570}
571
572void *
573uu_list_teardown(uu_list_t *lp, void **cookie)
574{
575	void *ep;
576
577	/*
578	 * XXX: disable list modification until list is empty
579	 */
580	if (lp->ul_debug && *cookie != NULL)
581		uu_panic("uu_list_teardown(%p, %p): unexpected cookie\n", lp,
582		    cookie);
583
584	ep = uu_list_first(lp);
585	if (ep)
586		uu_list_remove(lp, ep);
587	return (ep);
588}
589
590int
591uu_list_insert_before(uu_list_t *lp, void *target, void *elem)
592{
593	uu_list_node_impl_t *np = ELEM_TO_NODE(lp, target);
594
595	if (target == NULL)
596		np = &lp->ul_null_node;
597
598	if (lp->ul_debug) {
599		if (np->uln_prev == NULL)
600			uu_panic("uu_list_insert_before(%p, %p, %p): %p is "
601			    "not currently on a list\n",
602			    lp, target, elem, target);
603	}
604	if (lp->ul_sorted) {
605		if (lp->ul_debug)
606			uu_panic("uu_list_insert_before(%p, ...): list is "
607			    "UU_LIST_SORTED\n", lp);
608		uu_set_error(UU_ERROR_NOT_SUPPORTED);
609		return (-1);
610	}
611
612	list_insert(lp, ELEM_TO_NODE(lp, elem), np->uln_prev, np);
613	return (0);
614}
615
616int
617uu_list_insert_after(uu_list_t *lp, void *target, void *elem)
618{
619	uu_list_node_impl_t *np = ELEM_TO_NODE(lp, target);
620
621	if (target == NULL)
622		np = &lp->ul_null_node;
623
624	if (lp->ul_debug) {
625		if (np->uln_prev == NULL)
626			uu_panic("uu_list_insert_after(%p, %p, %p): %p is "
627			    "not currently on a list\n",
628			    lp, target, elem, target);
629	}
630	if (lp->ul_sorted) {
631		if (lp->ul_debug)
632			uu_panic("uu_list_insert_after(%p, ...): list is "
633			    "UU_LIST_SORTED\n", lp);
634		uu_set_error(UU_ERROR_NOT_SUPPORTED);
635		return (-1);
636	}
637
638	list_insert(lp, ELEM_TO_NODE(lp, elem), np, np->uln_next);
639	return (0);
640}
641
642size_t
643uu_list_numnodes(uu_list_t *lp)
644{
645	return (lp->ul_numnodes);
646}
647
648void *
649uu_list_first(uu_list_t *lp)
650{
651	uu_list_node_impl_t *n = lp->ul_null_node.uln_next;
652	if (n == &lp->ul_null_node)
653		return (NULL);
654	return (NODE_TO_ELEM(lp, n));
655}
656
657void *
658uu_list_last(uu_list_t *lp)
659{
660	uu_list_node_impl_t *n = lp->ul_null_node.uln_prev;
661	if (n == &lp->ul_null_node)
662		return (NULL);
663	return (NODE_TO_ELEM(lp, n));
664}
665
666void *
667uu_list_next(uu_list_t *lp, void *elem)
668{
669	uu_list_node_impl_t *n = ELEM_TO_NODE(lp, elem);
670
671	n = n->uln_next;
672	if (n == &lp->ul_null_node)
673		return (NULL);
674	return (NODE_TO_ELEM(lp, n));
675}
676
677void *
678uu_list_prev(uu_list_t *lp, void *elem)
679{
680	uu_list_node_impl_t *n = ELEM_TO_NODE(lp, elem);
681
682	n = n->uln_prev;
683	if (n == &lp->ul_null_node)
684		return (NULL);
685	return (NODE_TO_ELEM(lp, n));
686}
687
688/*
689 * called from uu_lockup() and uu_release(), as part of our fork1()-safety.
690 */
691void
692uu_list_lockup(void)
693{
694	uu_list_pool_t *pp;
695
696	(void) pthread_mutex_lock(&uu_lpool_list_lock);
697	for (pp = uu_null_lpool.ulp_next; pp != &uu_null_lpool;
698	    pp = pp->ulp_next)
699		(void) pthread_mutex_lock(&pp->ulp_lock);
700}
701
702void
703uu_list_release(void)
704{
705	uu_list_pool_t *pp;
706
707	for (pp = uu_null_lpool.ulp_next; pp != &uu_null_lpool;
708	    pp = pp->ulp_next)
709		(void) pthread_mutex_unlock(&pp->ulp_lock);
710	(void) pthread_mutex_unlock(&uu_lpool_list_lock);
711}
712