Fix mdoc(7)/man(7) mix up.
[netbsd-mini2440.git] / lib / libbsdmalloc / malloc.c
blob200a8ae64df9dfbc4f23d30d9946d21b0923a6af
1 /* $NetBSD: malloc.c,v 1.1 2003/04/21 22:21:06 elric Exp $ */
3 /*
4 * Copyright (c) 1983, 1993
5 * The Regents of the University of California. All rights reserved.
7 * Redistribution and use in source and binary forms, with or without
8 * modification, are permitted provided that the following conditions
9 * are met:
10 * 1. Redistributions of source code must retain the above copyright
11 * notice, this list of conditions and the following disclaimer.
12 * 2. Redistributions in binary form must reproduce the above copyright
13 * notice, this list of conditions and the following disclaimer in the
14 * documentation and/or other materials provided with the distribution.
15 * 3. Neither the name of the University nor the names of its contributors
16 * may be used to endorse or promote products derived from this software
17 * without specific prior written permission.
19 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
20 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
21 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
22 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
23 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
24 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
25 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
26 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
27 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
28 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
29 * SUCH DAMAGE.
32 #include <sys/cdefs.h>
33 #if defined(LIBC_SCCS) && !defined(lint)
34 #if 0
35 static char sccsid[] = "@(#)malloc.c 8.1 (Berkeley) 6/4/93";
36 #else
37 __RCSID("$NetBSD: malloc.c,v 1.1 2003/04/21 22:21:06 elric Exp $");
38 #endif
39 #endif /* LIBC_SCCS and not lint */
42 * malloc.c (Caltech) 2/21/82
43 * Chris Kingsley, kingsley@cit-20.
45 * This is a very fast storage allocator. It allocates blocks of a small
46 * number of different sizes, and keeps free lists of each size. Blocks that
47 * don't exactly fit are passed up to the next larger size. In this
48 * implementation, the available sizes are 2^n-4 (or 2^n-10) bytes long.
49 * This is designed for use in a virtual memory environment.
52 #include <sys/types.h>
53 #if defined(DEBUG) || defined(RCHECK)
54 #include <sys/uio.h>
55 #endif
56 #if defined(RCHECK) || defined(MSTATS)
57 #include <stdio.h>
58 #endif
59 #include <stdlib.h>
60 #include <string.h>
61 #include <unistd.h>
62 #include "reentrant.h"
66 * The overhead on a block is at least 4 bytes. When free, this space
67 * contains a pointer to the next free block, and the bottom two bits must
68 * be zero. When in use, the first byte is set to MAGIC, and the second
69 * byte is the size index. The remaining bytes are for alignment.
70 * If range checking is enabled then a second word holds the size of the
71 * requested block, less 1, rounded up to a multiple of sizeof(RMAGIC).
72 * The order of elements is critical: ov_magic must overlay the low order
73 * bits of ov_next, and ov_magic can not be a valid ov_next bit pattern.
75 union overhead {
76 union overhead *ov_next; /* when free */
77 struct {
78 u_char ovu_magic; /* magic number */
79 u_char ovu_index; /* bucket # */
80 #ifdef RCHECK
81 u_short ovu_rmagic; /* range magic number */
82 u_long ovu_size; /* actual block size */
83 #endif
84 } ovu;
85 #define ov_magic ovu.ovu_magic
86 #define ov_index ovu.ovu_index
87 #define ov_rmagic ovu.ovu_rmagic
88 #define ov_size ovu.ovu_size
91 #define MAGIC 0xef /* magic # on accounting info */
92 #ifdef RCHECK
93 #define RMAGIC 0x5555 /* magic # on range info */
94 #endif
96 #ifdef RCHECK
97 #define RSLOP sizeof (u_short)
98 #else
99 #define RSLOP 0
100 #endif
103 * nextf[i] is the pointer to the next free block of size 2^(i+3). The
104 * smallest allocatable block is 8 bytes. The overhead information
105 * precedes the data area returned to the user.
107 #define NBUCKETS 30
108 static union overhead *nextf[NBUCKETS];
110 static long pagesz; /* page size */
111 static int pagebucket; /* page size bucket */
113 #ifdef MSTATS
115 * nmalloc[i] is the difference between the number of mallocs and frees
116 * for a given block size.
118 static u_int nmalloc[NBUCKETS];
119 #endif
121 #ifdef _REENT
122 static mutex_t malloc_mutex = MUTEX_INITIALIZER;
123 #endif
125 static void morecore __P((int));
126 static int findbucket __P((union overhead *, int));
127 #ifdef MSTATS
128 void mstats __P((const char *));
129 #endif
131 #if defined(DEBUG) || defined(RCHECK)
132 #define ASSERT(p) if (!(p)) botch(__STRING(p))
134 static void botch __P((const char *));
137 * NOTE: since this may be called while malloc_mutex is locked, stdio must not
138 * be used in this function.
140 static void
141 botch(s)
142 const char *s;
144 struct iovec iov[3];
146 iov[0].iov_base = "\nassertion botched: ";
147 iov[0].iov_len = 20;
148 iov[1].iov_base = (void *)s;
149 iov[1].iov_len = strlen(s);
150 iov[2].iov_base = "\n";
151 iov[2].iov_len = 1;
154 * This place deserves a word of warning: a cancellation point will
155 * occur when executing writev(), and we might be still owning
156 * malloc_mutex. At this point we need to disable cancellation
157 * until `after' abort() because i) establishing a cancellation handler
158 * might, depending on the implementation, result in another malloc()
159 * to be executed, and ii) it is really not desirable to let execution
160 * continue. `Fix me.'
162 * Note that holding mutex_lock during abort() is safe.
165 (void)writev(STDERR_FILENO, iov, 3);
166 abort();
168 #else
169 #define ASSERT(p)
170 #endif
172 void *
173 malloc(nbytes)
174 size_t nbytes;
176 union overhead *op;
177 int bucket;
178 long n;
179 unsigned amt;
181 mutex_lock(&malloc_mutex);
184 * First time malloc is called, setup page size and
185 * align break pointer so all data will be page aligned.
187 if (pagesz == 0) {
188 pagesz = n = getpagesize();
189 ASSERT(pagesz > 0);
190 op = (union overhead *)(void *)sbrk(0);
191 n = n - sizeof (*op) - ((long)op & (n - 1));
192 if (n < 0)
193 n += pagesz;
194 if (n) {
195 if (sbrk((int)n) == (void *)-1) {
196 mutex_unlock(&malloc_mutex);
197 return (NULL);
200 bucket = 0;
201 amt = 8;
202 while (pagesz > amt) {
203 amt <<= 1;
204 bucket++;
206 pagebucket = bucket;
209 * Convert amount of memory requested into closest block size
210 * stored in hash buckets which satisfies request.
211 * Account for space used per block for accounting.
213 if (nbytes <= (n = pagesz - sizeof (*op) - RSLOP)) {
214 #ifndef RCHECK
215 amt = 8; /* size of first bucket */
216 bucket = 0;
217 #else
218 amt = 16; /* size of first bucket */
219 bucket = 1;
220 #endif
221 n = -((long)sizeof (*op) + RSLOP);
222 } else {
223 amt = (unsigned)pagesz;
224 bucket = pagebucket;
226 while (nbytes > amt + n) {
227 amt <<= 1;
228 if (amt == 0)
229 return (NULL);
230 bucket++;
233 * If nothing in hash bucket right now,
234 * request more memory from the system.
236 if ((op = nextf[bucket]) == NULL) {
237 morecore(bucket);
238 if ((op = nextf[bucket]) == NULL) {
239 mutex_unlock(&malloc_mutex);
240 return (NULL);
243 /* remove from linked list */
244 nextf[bucket] = op->ov_next;
245 op->ov_magic = MAGIC;
246 op->ov_index = bucket;
247 #ifdef MSTATS
248 nmalloc[bucket]++;
249 #endif
250 mutex_unlock(&malloc_mutex);
251 #ifdef RCHECK
253 * Record allocated size of block and
254 * bound space with magic numbers.
256 op->ov_size = (nbytes + RSLOP - 1) & ~(RSLOP - 1);
257 op->ov_rmagic = RMAGIC;
258 *(u_short *)((caddr_t)(op + 1) + op->ov_size) = RMAGIC;
259 #endif
260 return ((void *)(op + 1));
264 * Allocate more memory to the indicated bucket.
266 static void
267 morecore(bucket)
268 int bucket;
270 union overhead *op;
271 long sz; /* size of desired block */
272 long amt; /* amount to allocate */
273 long nblks; /* how many blocks we get */
276 * sbrk_size <= 0 only for big, FLUFFY, requests (about
277 * 2^30 bytes on a VAX, I think) or for a negative arg.
279 sz = 1 << (bucket + 3);
280 #ifdef DEBUG
281 ASSERT(sz > 0);
282 #else
283 if (sz <= 0)
284 return;
285 #endif
286 if (sz < pagesz) {
287 amt = pagesz;
288 nblks = amt / sz;
289 } else {
290 amt = sz + pagesz;
291 nblks = 1;
293 op = (union overhead *)(void *)sbrk((int)amt);
294 /* no more room! */
295 if ((long)op == -1)
296 return;
298 * Add new memory allocated to that on
299 * free list for this hash bucket.
301 nextf[bucket] = op;
302 while (--nblks > 0) {
303 op->ov_next =
304 (union overhead *)(void *)((caddr_t)(void *)op+(size_t)sz);
305 op = op->ov_next;
309 void
310 free(cp)
311 void *cp;
313 long size;
314 union overhead *op;
316 if (cp == NULL)
317 return;
318 op = (union overhead *)(void *)((caddr_t)cp - sizeof (union overhead));
319 #ifdef DEBUG
320 ASSERT(op->ov_magic == MAGIC); /* make sure it was in use */
321 #else
322 if (op->ov_magic != MAGIC)
323 return; /* sanity */
324 #endif
325 #ifdef RCHECK
326 ASSERT(op->ov_rmagic == RMAGIC);
327 ASSERT(*(u_short *)((caddr_t)(op + 1) + op->ov_size) == RMAGIC);
328 #endif
329 size = op->ov_index;
330 ASSERT(size < NBUCKETS);
331 mutex_lock(&malloc_mutex);
332 op->ov_next = nextf[(unsigned int)size];/* also clobbers ov_magic */
333 nextf[(unsigned int)size] = op;
334 #ifdef MSTATS
335 nmalloc[(size_t)size]--;
336 #endif
337 mutex_unlock(&malloc_mutex);
341 * When a program attempts "storage compaction" as mentioned in the
342 * old malloc man page, it realloc's an already freed block. Usually
343 * this is the last block it freed; occasionally it might be farther
344 * back. We have to search all the free lists for the block in order
345 * to determine its bucket: 1st we make one pass thru the lists
346 * checking only the first block in each; if that fails we search
347 * ``__realloc_srchlen'' blocks in each list for a match (the variable
348 * is extern so the caller can modify it). If that fails we just copy
349 * however many bytes was given to realloc() and hope it's not huge.
351 int __realloc_srchlen = 4; /* 4 should be plenty, -1 =>'s whole list */
353 void *
354 realloc(cp, nbytes)
355 void *cp;
356 size_t nbytes;
358 u_long onb;
359 long i;
360 union overhead *op;
361 char *res;
362 int was_alloced = 0;
364 if (cp == NULL)
365 return (malloc(nbytes));
366 if (nbytes == 0) {
367 free (cp);
368 return (NULL);
370 op = (union overhead *)(void *)((caddr_t)cp - sizeof (union overhead));
371 mutex_lock(&malloc_mutex);
372 if (op->ov_magic == MAGIC) {
373 was_alloced++;
374 i = op->ov_index;
375 } else {
377 * Already free, doing "compaction".
379 * Search for the old block of memory on the
380 * free list. First, check the most common
381 * case (last element free'd), then (this failing)
382 * the last ``__realloc_srchlen'' items free'd.
383 * If all lookups fail, then assume the size of
384 * the memory block being realloc'd is the
385 * largest possible (so that all "nbytes" of new
386 * memory are copied into). Note that this could cause
387 * a memory fault if the old area was tiny, and the moon
388 * is gibbous. However, that is very unlikely.
390 if ((i = findbucket(op, 1)) < 0 &&
391 (i = findbucket(op, __realloc_srchlen)) < 0)
392 i = NBUCKETS;
394 onb = (u_long)1 << (u_long)(i + 3);
395 if (onb < pagesz)
396 onb -= sizeof (*op) + RSLOP;
397 else
398 onb += pagesz - sizeof (*op) - RSLOP;
399 /* avoid the copy if same size block */
400 if (was_alloced) {
401 if (i) {
402 i = (long)1 << (long)(i + 2);
403 if (i < pagesz)
404 i -= sizeof (*op) + RSLOP;
405 else
406 i += pagesz - sizeof (*op) - RSLOP;
408 if (nbytes <= onb && nbytes > i) {
409 #ifdef RCHECK
410 op->ov_size = (nbytes + RSLOP - 1) & ~(RSLOP - 1);
411 *(u_short *)((caddr_t)(op + 1) + op->ov_size) = RMAGIC;
412 #endif
413 mutex_unlock(&malloc_mutex);
414 return (cp);
417 #ifndef _REENT
418 else
419 free(cp);
420 #endif
422 mutex_unlock(&malloc_mutex);
423 if ((res = malloc(nbytes)) == NULL) {
424 #ifdef _REENT
425 free(cp);
426 #endif
427 return (NULL);
429 #ifndef _REENT
430 if (cp != res) /* common optimization if "compacting" */
431 (void)memmove(res, cp, (size_t)((nbytes < onb) ? nbytes : onb));
432 #else
433 (void)memmove(res, cp, (size_t)((nbytes < onb) ? nbytes : onb));
434 free(cp);
435 #endif
436 return (res);
440 * Search ``srchlen'' elements of each free list for a block whose
441 * header starts at ``freep''. If srchlen is -1 search the whole list.
442 * Return bucket number, or -1 if not found.
444 static int
445 findbucket(freep, srchlen)
446 union overhead *freep;
447 int srchlen;
449 union overhead *p;
450 int i, j;
452 for (i = 0; i < NBUCKETS; i++) {
453 j = 0;
454 for (p = nextf[i]; p && j != srchlen; p = p->ov_next) {
455 if (p == freep)
456 return (i);
457 j++;
460 return (-1);
463 #ifdef MSTATS
465 * mstats - print out statistics about malloc
467 * Prints two lines of numbers, one showing the length of the free list
468 * for each size category, the second showing the number of mallocs -
469 * frees for each size category.
471 void
472 mstats(s)
473 char *s;
475 int i, j;
476 union overhead *p;
477 int totfree = 0,
478 totused = 0;
480 fprintf(stderr, "Memory allocation statistics %s\nfree:\t", s);
481 for (i = 0; i < NBUCKETS; i++) {
482 for (j = 0, p = nextf[i]; p; p = p->ov_next, j++)
484 fprintf(stderr, " %d", j);
485 totfree += j * (1 << (i + 3));
487 fprintf(stderr, "\nused:\t");
488 for (i = 0; i < NBUCKETS; i++) {
489 fprintf(stderr, " %d", nmalloc[i]);
490 totused += nmalloc[i] * (1 << (i + 3));
492 fprintf(stderr, "\n\tTotal in use: %d, total free: %d\n",
493 totused, totfree);
495 #endif