diff options
author | Otto Moerbeek <otto@cvs.openbsd.org> | 2023-11-04 11:02:36 +0000 |
---|---|---|
committer | Otto Moerbeek <otto@cvs.openbsd.org> | 2023-11-04 11:02:36 +0000 |
commit | 3680995327d35e1778d72c9c719bd996c69f8d65 (patch) | |
tree | d461c57a72566abb02055394521ea143fc496400 /lib/libc/stdlib | |
parent | 5a47507cb2a63d53857d345e7d561f1e5bdb5ee6 (diff) |
KNF plus fixed a few signed vs unsigned compares (that we actually
not real problems)
Diffstat (limited to 'lib/libc/stdlib')
-rw-r--r-- | lib/libc/stdlib/malloc.c | 55 |
1 files changed, 33 insertions, 22 deletions
diff --git a/lib/libc/stdlib/malloc.c b/lib/libc/stdlib/malloc.c index 81ad79dfd30..9da180d8148 100644 --- a/lib/libc/stdlib/malloc.c +++ b/lib/libc/stdlib/malloc.c @@ -1,4 +1,4 @@ -/* $OpenBSD: malloc.c,v 1.292 2023/10/26 17:59:16 otto Exp $ */ +/* $OpenBSD: malloc.c,v 1.293 2023/11/04 11:02:35 otto Exp $ */ /* * Copyright (c) 2008, 2010, 2011, 2016, 2023 Otto Moerbeek <otto@drijf.net> * Copyright (c) 2012 Matthew Dempsky <matthew@openbsd.org> @@ -215,7 +215,8 @@ struct chunk_info { u_short bits[CHUNK_INFO_TAIL]; /* which chunks are free */ }; -#define CHUNK_FREE(i, n) ((i)->bits[(n) / MALLOC_BITS] & (1U << ((n) % MALLOC_BITS))) +#define CHUNK_FREE(i, n) ((i)->bits[(n) / MALLOC_BITS] & \ + (1U << ((n) % MALLOC_BITS))) struct malloc_readonly { /* Main bookkeeping information */ @@ -232,7 +233,7 @@ struct malloc_readonly { u_int junk_loc; /* variation in location of junk */ size_t malloc_guard; /* use guard pages after allocations? */ #ifdef MALLOC_STATS - int malloc_stats; /* save callers, dump leak report at end */ + int malloc_stats; /* save callers, dump leak report */ int malloc_verbose; /* dump verbose statistics at end */ #define DO_STATS mopts.malloc_stats #else @@ -542,7 +543,7 @@ omalloc_init(void) static void omalloc_poolinit(struct dir_info *d, int mmap_flag) { - int i, j; + u_int i, j; d->r = NULL; d->rbytesused = sizeof(d->rbytes); @@ -597,7 +598,8 @@ omalloc_grow(struct dir_info *d) } if (d->regions_total > 0) { - oldpsz = PAGEROUND(d->regions_total * sizeof(struct region_info)); + oldpsz = PAGEROUND(d->regions_total * + sizeof(struct region_info)); /* clear to avoid meta info ending up in the cache */ unmap(d, d->r, oldpsz, oldpsz); } @@ -995,7 +997,8 @@ alloc_chunk_info(struct dir_info *d, u_int bucket) for (i = 0; i < count; i++, q += size) { p = (struct chunk_info *)q; - LIST_INSERT_HEAD(&d->chunk_info_list[bucket], p, entries); + LIST_INSERT_HEAD(&d->chunk_info_list[bucket], p, + entries); } } p = LIST_FIRST(&d->chunk_info_list[bucket]); @@ -1023,7 +1026,8 @@ omalloc_make_chunks(struct dir_info *d, u_int bucket, u_int listnum) ff = map(d, MALLOC_PAGESIZE, 0); if (ff == MAP_FAILED) goto err; - memset(ff, 0, sizeof(void *) * MALLOC_PAGESIZE / B2ALLOC(bucket)); + memset(ff, 0, sizeof(void *) * MALLOC_PAGESIZE / + B2ALLOC(bucket)); } /* memory protect the page allocated in the malloc(0) case */ @@ -1405,14 +1409,14 @@ _malloc_init(int from_rthreads) sz = mopts.malloc_mutexes * sizeof(*d) + 2 * MALLOC_PAGESIZE; if ((p = MMAPNONE(sz, 0)) == MAP_FAILED) wrterror(NULL, "malloc_init mmap1 failed"); - if (mprotect(p + MALLOC_PAGESIZE, mopts.malloc_mutexes * sizeof(*d), - PROT_READ | PROT_WRITE)) + if (mprotect(p + MALLOC_PAGESIZE, mopts.malloc_mutexes * + sizeof(*d), PROT_READ | PROT_WRITE)) wrterror(NULL, "malloc_init mprotect1 failed"); if (mimmutable(p, sz)) wrterror(NULL, "malloc_init mimmutable1 failed"); - d_avail = (((mopts.malloc_mutexes * sizeof(*d) + MALLOC_PAGEMASK) & - ~MALLOC_PAGEMASK) - (mopts.malloc_mutexes * sizeof(*d))) >> - MALLOC_MINSHIFT; + d_avail = (((mopts.malloc_mutexes * sizeof(*d) + + MALLOC_PAGEMASK) & ~MALLOC_PAGEMASK) - + (mopts.malloc_mutexes * sizeof(*d))) >> MALLOC_MINSHIFT; d = (struct dir_info *)(p + MALLOC_PAGESIZE + (arc4random_uniform(d_avail) << MALLOC_MINSHIFT)); STATS_ADD(d[1].malloc_used, sz); @@ -1422,9 +1426,12 @@ _malloc_init(int from_rthreads) if (((uintptr_t)&malloc_readonly & MALLOC_PAGEMASK) == 0) { if (mprotect(&malloc_readonly, sizeof(malloc_readonly), PROT_READ)) - wrterror(NULL, "malloc_init mprotect r/o failed"); - if (mimmutable(&malloc_readonly, sizeof(malloc_readonly))) - wrterror(NULL, "malloc_init mimmutable r/o failed"); + wrterror(NULL, + "malloc_init mprotect r/o failed"); + if (mimmutable(&malloc_readonly, + sizeof(malloc_readonly))) + wrterror(NULL, + "malloc_init mimmutable r/o failed"); } } @@ -1458,7 +1465,8 @@ _malloc_init(int from_rthreads) wrterror(NULL, "malloc_init mmap2 failed"); if (mimmutable(p, sz)) - wrterror(NULL, "malloc_init mimmutable2 failed"); + wrterror(NULL, + "malloc_init mimmutable2 failed"); for (j = 0; j < MAX_SMALLCACHEABLE_SIZE; j++) { d->smallcache[j].pages = p; p = (char *)p + d->smallcache[j].max * @@ -1535,7 +1543,8 @@ findpool(void *p, struct dir_info *argpool, struct dir_info **foundpool, if (r == NULL) { u_int i, nmutexes; - nmutexes = mopts.malloc_pool[1]->malloc_mt ? mopts.malloc_mutexes : 2; + nmutexes = mopts.malloc_pool[1]->malloc_mt ? + mopts.malloc_mutexes : 2; for (i = 1; i < nmutexes; i++) { u_int j = (argpool->mutex + i) & (nmutexes - 1); @@ -1813,7 +1822,8 @@ orealloc(struct dir_info **argpool, void *p, size_t newsz) size_t needed = rnewsz - roldsz; STATS_INC(pool->cheap_realloc_tries); - q = MMAPA(hint, needed, MAP_FIXED | __MAP_NOREPLACE | pool->mmap_flag); + q = MMAPA(hint, needed, MAP_FIXED | + __MAP_NOREPLACE | pool->mmap_flag); if (q == hint) { STATS_ADD(pool->malloc_used, needed); if (pool->malloc_junk == 2) @@ -2030,7 +2040,8 @@ orecallocarray(struct dir_info **argpool, void *p, size_t oldsize, wrterror(pool, "recorded size %zu < %zu", sz - mopts.malloc_guard, oldsize); if (oldsize < (sz - mopts.malloc_guard) / 2) - wrterror(pool, "recorded size %zu inconsistent with %zu", + wrterror(pool, + "recorded size %zu inconsistent with %zu", sz - mopts.malloc_guard, oldsize); } @@ -2383,7 +2394,7 @@ ulog(const char *format, ...) va_end(ap); if (len < 0) return; - if (len > KTR_USER_MAXLEN - filled) + if ((size_t)len > KTR_USER_MAXLEN - filled) len = KTR_USER_MAXLEN - filled; filled += len; if (filled > 0) { @@ -2516,7 +2527,7 @@ dump_chunk(struct leaktree* leaks, struct chunk_info *p, void **f, static void dump_free_chunk_info(struct dir_info *d, struct leaktree *leaks) { - int i, j, count; + u_int i, j, count; struct chunk_info *p; ulog("Free chunk structs:\n"); @@ -2639,7 +2650,7 @@ malloc_dump0(int poolno, struct dir_info *pool, struct leaktree *leaks) void malloc_dump(void) { - int i; + u_int i; int saved_errno = errno; /* XXX leak when run multiple times */ |