1/* $Header: /p/tcsh/cvsroot/tcsh/tc.alloc.c,v 3.50 2011/12/30 20:55:24 christos Exp $ */
2/*
3 * tc.alloc.c (Caltech) 2/21/82
4 * Chris Kingsley, kingsley@cit-20.
5 *
6 * This is a very fast storage allocator.  It allocates blocks of a small
7 * number of different sizes, and keeps free lists of each size.  Blocks that
8 * don't exactly fit are passed up to the next larger size.  In this
9 * implementation, the available sizes are 2^n-4 (or 2^n-12) bytes long.
10 * This is designed for use in a program that uses vast quantities of memory,
11 * but bombs when it runs out.
12 */
13/*-
14 * Copyright (c) 1980, 1991 The Regents of the University of California.
15 * All rights reserved.
16 *
17 * Redistribution and use in source and binary forms, with or without
18 * modification, are permitted provided that the following conditions
19 * are met:
20 * 1. Redistributions of source code must retain the above copyright
21 *    notice, this list of conditions and the following disclaimer.
22 * 2. Redistributions in binary form must reproduce the above copyright
23 *    notice, this list of conditions and the following disclaimer in the
24 *    documentation and/or other materials provided with the distribution.
25 * 3. Neither the name of the University nor the names of its contributors
26 *    may be used to endorse or promote products derived from this software
27 *    without specific prior written permission.
28 *
29 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
30 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
31 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
32 * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
33 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
34 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
35 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
36 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
37 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
38 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
39 * SUCH DAMAGE.
40 */
41#include "sh.h"
42#ifdef HAVE_MALLINFO
43#include <malloc.h>
44#endif
45
46RCSID("$tcsh: tc.alloc.c,v 3.50 2011/12/30 20:55:24 christos Exp $")
47
48#define RCHECK
49#define DEBUG
50
51static char   *memtop = NULL;		/* PWP: top of current memory */
52static char   *membot = NULL;		/* PWP: bottom of allocatable memory */
53
54int dont_free = 0;
55
56#ifdef WINNT_NATIVE
57# define malloc		fmalloc
58# define free		ffree
59# define calloc		fcalloc
60# define realloc	frealloc
61#endif /* WINNT_NATIVE */
62
63#if !defined(DEBUG) || defined(SYSMALLOC)
64static void
65out_of_memory (void)
66{
67    static const char msg[] = "Out of memory\n";
68
69    write(didfds ? 2 : SHDIAG, msg, strlen(msg));
70    _exit(1);
71}
72#endif
73
74#ifndef SYSMALLOC
75
76#ifdef SX
77extern void* sbrk();
78#endif
79/*
80 * Lots of os routines are busted and try to free invalid pointers.
81 * Although our free routine is smart enough and it will pick bad
82 * pointers most of the time, in cases where we know we are going to get
83 * a bad pointer, we'd rather leak.
84 */
85
86#ifndef NULL
87#define	NULL 0
88#endif
89
90typedef unsigned char U_char;	/* we don't really have signed chars */
91typedef unsigned int U_int;
92typedef unsigned short U_short;
93typedef unsigned long U_long;
94
95
96/*
97 * The overhead on a block is at least 4 bytes.  When free, this space
98 * contains a pointer to the next free block, and the bottom two bits must
99 * be zero.  When in use, the first byte is set to MAGIC, and the second
100 * byte is the size index.  The remaining bytes are for alignment.
101 * If range checking is enabled and the size of the block fits
102 * in two bytes, then the top two bytes hold the size of the requested block
103 * plus the range checking words, and the header word MINUS ONE.
104 */
105
106
107#define MEMALIGN(a) (((a) + ROUNDUP) & ~ROUNDUP)
108
109union overhead {
110    union overhead *ov_next;	/* when free */
111    struct {
112	U_char  ovu_magic;	/* magic number */
113	U_char  ovu_index;	/* bucket # */
114#ifdef RCHECK
115	U_short ovu_size;	/* actual block size */
116	U_int   ovu_rmagic;	/* range magic number */
117#endif
118    }       ovu;
119#define	ov_magic	ovu.ovu_magic
120#define	ov_index	ovu.ovu_index
121#define	ov_size		ovu.ovu_size
122#define	ov_rmagic	ovu.ovu_rmagic
123};
124
125#define	MAGIC		0xfd	/* magic # on accounting info */
126#define RMAGIC		0x55555555	/* magic # on range info */
127#ifdef RCHECK
128#define	RSLOP		sizeof (U_int)
129#else
130#define	RSLOP		0
131#endif
132
133
134#define ROUNDUP	7
135
136/*
137 * nextf[i] is the pointer to the next free block of size 2^(i+3).  The
138 * smallest allocatable block is 8 bytes.  The overhead information
139 * precedes the data area returned to the user.
140 */
141#define	NBUCKETS ((sizeof(long) << 3) - 3)
142static union overhead *nextf[NBUCKETS] IZERO_STRUCT;
143
144/*
145 * nmalloc[i] is the difference between the number of mallocs and frees
146 * for a given block size.
147 */
148static U_int nmalloc[NBUCKETS] IZERO_STRUCT;
149
150#ifndef lint
151static	int	findbucket	(union overhead *, int);
152static	void	morecore	(int);
153#endif
154
155
156#ifdef DEBUG
157# define CHECK(a, str, p) \
158    if (a) { \
159	xprintf(str, p);	\
160	xprintf(" (memtop = %p membot = %p)\n", memtop, membot);	\
161	abort(); \
162    }
163#else
164# define CHECK(a, str, p) \
165    if (a) { \
166	xprintf(str, p);	\
167	xprintf(" (memtop = %p membot = %p)\n", memtop, membot);	\
168	return; \
169    }
170#endif
171
172memalign_t
173malloc(size_t nbytes)
174{
175#ifndef lint
176    union overhead *p;
177    int bucket = 0;
178    unsigned shiftr;
179
180    /*
181     * Convert amount of memory requested into closest block size stored in
182     * hash buckets which satisfies request.  Account for space used per block
183     * for accounting.
184     */
185#ifdef SUNOS4
186    /*
187     * SunOS localtime() overwrites the 9th byte on an 8 byte malloc()....
188     * so we get one more...
189     * From Michael Schroeder: This is not true. It depends on the
190     * timezone string. In Europe it can overwrite the 13th byte on a
191     * 12 byte malloc.
192     * So we punt and we always allocate an extra byte.
193     */
194    nbytes++;
195#endif
196
197    nbytes = MEMALIGN(MEMALIGN(sizeof(union overhead)) + nbytes + RSLOP);
198    shiftr = (nbytes - 1) >> 2;
199
200    /* apart from this loop, this is O(1) */
201    while ((shiftr >>= 1) != 0)
202	bucket++;
203    /*
204     * If nothing in hash bucket right now, request more memory from the
205     * system.
206     */
207    if (nextf[bucket] == NULL)
208	morecore(bucket);
209    if ((p = nextf[bucket]) == NULL) {
210	child++;
211#ifndef DEBUG
212	out_of_memory();
213#else
214	showall(NULL, NULL);
215	xprintf(CGETS(19, 1, "nbytes=%zu: Out of memory\n"), nbytes);
216	abort();
217#endif
218	/* fool lint */
219	return ((memalign_t) 0);
220    }
221    /* remove from linked list */
222    nextf[bucket] = nextf[bucket]->ov_next;
223    p->ov_magic = MAGIC;
224    p->ov_index = bucket;
225    nmalloc[bucket]++;
226#ifdef RCHECK
227    /*
228     * Record allocated size of block and bound space with magic numbers.
229     */
230    p->ov_size = (p->ov_index <= 13) ? nbytes - 1 : 0;
231    p->ov_rmagic = RMAGIC;
232    *((U_int *) (((caddr_t) p) + nbytes - RSLOP)) = RMAGIC;
233#endif
234    return ((memalign_t) (((caddr_t) p) + MEMALIGN(sizeof(union overhead))));
235#else
236    if (nbytes)
237	return ((memalign_t) 0);
238    else
239	return ((memalign_t) 0);
240#endif /* !lint */
241}
242
243#ifndef lint
244/*
245 * Allocate more memory to the indicated bucket.
246 */
247static void
248morecore(int bucket)
249{
250    union overhead *op;
251    int rnu;		/* 2^rnu bytes will be requested */
252    int nblks;		/* become nblks blocks of the desired size */
253    int siz;
254
255    if (nextf[bucket])
256	return;
257    /*
258     * Insure memory is allocated on a page boundary.  Should make getpageize
259     * call?
260     */
261    op = (union overhead *) sbrk(0);
262    memtop = (char *) op;
263    if (membot == NULL)
264	membot = memtop;
265    if ((long) op & 0x3ff) {
266	memtop = sbrk((int) (1024 - ((long) op & 0x3ff)));
267	memtop += (long) (1024 - ((long) op & 0x3ff));
268    }
269
270    /* take 2k unless the block is bigger than that */
271    rnu = (bucket <= 8) ? 11 : bucket + 3;
272    nblks = 1 << (rnu - (bucket + 3));	/* how many blocks to get */
273    memtop = sbrk(1 << rnu);	/* PWP */
274    op = (union overhead *) memtop;
275    /* no more room! */
276    if ((long) op == -1)
277	return;
278    memtop += (long) (1 << rnu);
279    /*
280     * Round up to minimum allocation size boundary and deduct from block count
281     * to reflect.
282     */
283    if (((U_long) op) & ROUNDUP) {
284	op = (union overhead *) (((U_long) op + (ROUNDUP + 1)) & ~ROUNDUP);
285	nblks--;
286    }
287    /*
288     * Add new memory allocated to that on free list for this hash bucket.
289     */
290    nextf[bucket] = op;
291    siz = 1 << (bucket + 3);
292    while (--nblks > 0) {
293	op->ov_next = (union overhead *) (((caddr_t) op) + siz);
294	op = (union overhead *) (((caddr_t) op) + siz);
295    }
296    op->ov_next = NULL;
297}
298
299#endif
300
301void
302free(ptr_t cp)
303{
304#ifndef lint
305    int size;
306    union overhead *op;
307
308    /*
309     * the don't free flag is there so that we avoid os bugs in routines
310     * that free invalid pointers!
311     */
312    if (cp == NULL || dont_free)
313	return;
314    CHECK(!memtop || !membot,
315	  CGETS(19, 2, "free(%p) called before any allocations."), cp);
316    CHECK(cp > (ptr_t) memtop,
317	  CGETS(19, 3, "free(%p) above top of memory."), cp);
318    CHECK(cp < (ptr_t) membot,
319	  CGETS(19, 4, "free(%p) below bottom of memory."), cp);
320    op = (union overhead *) (((caddr_t) cp) - MEMALIGN(sizeof(union overhead)));
321    CHECK(op->ov_magic != MAGIC,
322	  CGETS(19, 5, "free(%p) bad block."), cp);
323
324#ifdef RCHECK
325    if (op->ov_index <= 13)
326	CHECK(*(U_int *) ((caddr_t) op + op->ov_size + 1 - RSLOP) != RMAGIC,
327	      CGETS(19, 6, "free(%p) bad range check."), cp);
328#endif
329    CHECK(op->ov_index >= NBUCKETS,
330	  CGETS(19, 7, "free(%p) bad block index."), cp);
331    size = op->ov_index;
332    op->ov_next = nextf[size];
333    nextf[size] = op;
334
335    nmalloc[size]--;
336
337#else
338    if (cp == NULL)
339	return;
340#endif
341}
342
343memalign_t
344calloc(size_t i, size_t j)
345{
346#ifndef lint
347    char *cp;
348
349    i *= j;
350    cp = xmalloc(i);
351    memset(cp, 0, i);
352
353    return ((memalign_t) cp);
354#else
355    if (i && j)
356	return ((memalign_t) 0);
357    else
358	return ((memalign_t) 0);
359#endif
360}
361
362/*
363 * When a program attempts "storage compaction" as mentioned in the
364 * old malloc man page, it realloc's an already freed block.  Usually
365 * this is the last block it freed; occasionally it might be farther
366 * back.  We have to search all the free lists for the block in order
367 * to determine its bucket: 1st we make one pass thru the lists
368 * checking only the first block in each; if that fails we search
369 * ``realloc_srchlen'' blocks in each list for a match (the variable
370 * is extern so the caller can modify it).  If that fails we just copy
371 * however many bytes was given to realloc() and hope it's not huge.
372 */
373#ifndef lint
374/* 4 should be plenty, -1 =>'s whole list */
375static int     realloc_srchlen = 4;
376#endif /* lint */
377
378memalign_t
379realloc(ptr_t cp, size_t nbytes)
380{
381#ifndef lint
382    U_int onb;
383    union overhead *op;
384    ptr_t res;
385    int i;
386    int     was_alloced = 0;
387
388    if (cp == NULL)
389	return (malloc(nbytes));
390    op = (union overhead *) (((caddr_t) cp) - MEMALIGN(sizeof(union overhead)));
391    if (op->ov_magic == MAGIC) {
392	was_alloced++;
393	i = op->ov_index;
394    }
395    else
396	/*
397	 * Already free, doing "compaction".
398	 *
399	 * Search for the old block of memory on the free list.  First, check the
400	 * most common case (last element free'd), then (this failing) the last
401	 * ``realloc_srchlen'' items free'd. If all lookups fail, then assume
402	 * the size of the memory block being realloc'd is the smallest
403	 * possible.
404	 */
405	if ((i = findbucket(op, 1)) < 0 &&
406	    (i = findbucket(op, realloc_srchlen)) < 0)
407	    i = 0;
408
409    onb = MEMALIGN(nbytes + MEMALIGN(sizeof(union overhead)) + RSLOP);
410
411    /* avoid the copy if same size block */
412    if (was_alloced && (onb <= (U_int) (1 << (i + 3))) &&
413	(onb > (U_int) (1 << (i + 2)))) {
414#ifdef RCHECK
415	/* JMR: formerly this wasn't updated ! */
416	nbytes = MEMALIGN(MEMALIGN(sizeof(union overhead))+nbytes+RSLOP);
417	*((U_int *) (((caddr_t) op) + nbytes - RSLOP)) = RMAGIC;
418	op->ov_rmagic = RMAGIC;
419	op->ov_size = (op->ov_index <= 13) ? nbytes - 1 : 0;
420#endif
421	return ((memalign_t) cp);
422    }
423    if ((res = malloc(nbytes)) == NULL)
424	return ((memalign_t) NULL);
425    if (cp != res) {		/* common optimization */
426	/*
427	 * christos: this used to copy nbytes! It should copy the
428	 * smaller of the old and new size
429	 */
430	onb = (1 << (i + 3)) - MEMALIGN(sizeof(union overhead)) - RSLOP;
431	(void) memmove(res, cp, onb < nbytes ? onb : nbytes);
432    }
433    if (was_alloced)
434	free(cp);
435    return ((memalign_t) res);
436#else
437    if (cp && nbytes)
438	return ((memalign_t) 0);
439    else
440	return ((memalign_t) 0);
441#endif /* !lint */
442}
443
444/*
445 * On linux, _nss_nis_setnetgrent() calls this function to determine
446 * the usable size of the pointer passed, but this is not a portable
447 * API, so we cannot use our malloc replacement without providing one.
448 * Thanks a lot glibc!
449 */
450#ifdef __linux__
451#define M_U_S_CONST
452#else
453#define M_U_S_CONST
454#endif
455size_t malloc_usable_size(M_U_S_CONST void *);
456size_t
457malloc_usable_size(M_U_S_CONST void *ptr)
458{
459    const union overhead *op = (const union overhead *)
460	(((const char *) ptr) - MEMALIGN(sizeof(*op)));
461    if (op->ov_magic == MAGIC)
462	    return 1 << (op->ov_index + 2);
463    else
464	    return 0;
465}
466
467
468#ifndef lint
469/*
470 * Search ``srchlen'' elements of each free list for a block whose
471 * header starts at ``freep''.  If srchlen is -1 search the whole list.
472 * Return bucket number, or -1 if not found.
473 */
474static int
475findbucket(union overhead *freep, int srchlen)
476{
477    union overhead *p;
478    size_t i;
479    int j;
480
481    for (i = 0; i < NBUCKETS; i++) {
482	j = 0;
483	for (p = nextf[i]; p && j != srchlen; p = p->ov_next) {
484	    if (p == freep)
485		return (i);
486	    j++;
487	}
488    }
489    return (-1);
490}
491
492#endif
493
494
495#else				/* SYSMALLOC */
496
497/**
498 ** ``Protected versions'' of malloc, realloc, calloc, and free
499 **
500 ** On many systems:
501 **
502 ** 1. malloc(0) is bad
503 ** 2. free(0) is bad
504 ** 3. realloc(0, n) is bad
505 ** 4. realloc(n, 0) is bad
506 **
507 ** Also we call our error routine if we run out of memory.
508 **/
509memalign_t
510smalloc(size_t n)
511{
512    ptr_t   ptr;
513
514    n = n ? n : 1;
515
516#ifdef HAVE_SBRK
517    if (membot == NULL)
518	membot = sbrk(0);
519#endif /* HAVE_SBRK */
520
521    if ((ptr = malloc(n)) == NULL)
522	out_of_memory();
523#ifndef HAVE_SBRK
524    if (memtop < ((char *) ptr) + n)
525	memtop = ((char *) ptr) + n;
526    if (membot == NULL)
527	membot = ptr;
528#endif /* !HAVE_SBRK */
529    return ((memalign_t) ptr);
530}
531
532memalign_t
533srealloc(ptr_t p, size_t n)
534{
535    ptr_t   ptr;
536
537    n = n ? n : 1;
538
539#ifdef HAVE_SBRK
540    if (membot == NULL)
541	membot = sbrk(0);
542#endif /* HAVE_SBRK */
543
544    if ((ptr = (p ? realloc(p, n) : malloc(n))) == NULL)
545	out_of_memory();
546#ifndef HAVE_SBRK
547    if (memtop < ((char *) ptr) + n)
548	memtop = ((char *) ptr) + n;
549    if (membot == NULL)
550	membot = ptr;
551#endif /* !HAVE_SBRK */
552    return ((memalign_t) ptr);
553}
554
555memalign_t
556scalloc(size_t s, size_t n)
557{
558    ptr_t   ptr;
559
560    n *= s;
561    n = n ? n : 1;
562
563#ifdef HAVE_SBRK
564    if (membot == NULL)
565	membot = sbrk(0);
566#endif /* HAVE_SBRK */
567
568    if ((ptr = malloc(n)) == NULL)
569	out_of_memory();
570
571    memset (ptr, 0, n);
572
573#ifndef HAVE_SBRK
574    if (memtop < ((char *) ptr) + n)
575	memtop = ((char *) ptr) + n;
576    if (membot == NULL)
577	membot = ptr;
578#endif /* !HAVE_SBRK */
579
580    return ((memalign_t) ptr);
581}
582
583void
584sfree(ptr_t p)
585{
586    if (p && !dont_free)
587	free(p);
588}
589
590#endif /* SYSMALLOC */
591
592/*
593 * mstats - print out statistics about malloc
594 *
595 * Prints two lines of numbers, one showing the length of the free list
596 * for each size category, the second showing the number of mallocs -
597 * frees for each size category.
598 */
599/*ARGSUSED*/
600void
601showall(Char **v, struct command *c)
602{
603#ifndef SYSMALLOC
604    size_t i, j;
605    union overhead *p;
606    int     totfree = 0, totused = 0;
607
608    xprintf(CGETS(19, 8, "%s current memory allocation:\nfree:\t"), progname);
609    for (i = 0; i < NBUCKETS; i++) {
610	for (j = 0, p = nextf[i]; p; p = p->ov_next, j++)
611	    continue;
612	xprintf(" %4zd", j);
613	totfree += j * (1 << (i + 3));
614    }
615    xprintf("\n%s:\t", CGETS(19, 9, "used"));
616    for (i = 0; i < NBUCKETS; i++) {
617	xprintf(" %4d", nmalloc[i]);
618	totused += nmalloc[i] * (1 << (i + 3));
619    }
620    xprintf(CGETS(19, 10, "\n\tTotal in use: %d, total free: %d\n"),
621	    totused, totfree);
622    xprintf(CGETS(19, 11,
623	    "\tAllocated memory from 0x%lx to 0x%lx.  Real top at 0x%lx\n"),
624	    (unsigned long) membot, (unsigned long) memtop,
625	    (unsigned long) sbrk(0));
626#else /* SYSMALLOC */
627#ifndef HAVE_MALLINFO
628#ifdef HAVE_SBRK
629    memtop = sbrk(0);
630#endif /* HAVE_SBRK */
631    xprintf(CGETS(19, 12, "Allocated memory from 0x%lx to 0x%lx (%ld).\n"),
632	    (unsigned long) membot, (unsigned long) memtop,
633	    (unsigned long) (memtop - membot));
634#else /* HAVE_MALLINFO */
635    struct mallinfo mi;
636
637    mi = mallinfo();
638    xprintf(CGETS(19, 13, "%s current memory allocation:\n"), progname);
639    xprintf(CGETS(19, 14, "Total space allocated from system: %d\n"), mi.arena);
640    xprintf(CGETS(19, 15, "Number of non-inuse chunks: %d\n"), mi.ordblks);
641    xprintf(CGETS(19, 16, "Number of mmapped regions: %d\n"), mi.hblks);
642    xprintf(CGETS(19, 17, "Total space in mmapped regions: %d\n"), mi.hblkhd);
643    xprintf(CGETS(19, 18, "Total allocated space: %d\n"), mi.uordblks);
644    xprintf(CGETS(19, 19, "Total non-inuse space: %d\n"), mi.fordblks);
645    xprintf(CGETS(19, 20, "Top-most, releasable space: %d\n"), mi.keepcost);
646#endif /* HAVE_MALLINFO */
647#endif /* SYSMALLOC */
648    USE(c);
649    USE(v);
650}
651