summaryrefslogtreecommitdiff
path: root/sys
diff options
context:
space:
mode:
authorStefan Kempf <stefan@cvs.openbsd.org>2016-05-22 16:18:27 +0000
committerStefan Kempf <stefan@cvs.openbsd.org>2016-05-22 16:18:27 +0000
commit8319ea69173e999cfecb98aff5509713d6dec3ce (patch)
tree66ed248c6b44de7eba442afabfb3ee3cd08bd03b /sys
parentc53cb24a8434d426b2e17022a9b263787b29cb47 (diff)
Make amaps use less kernel memory
This is achieved by grouping amap slots into chunks that are allocated on-demand by pool(9). Endless "fltamapcopy" loops because of kmem shortage should be solved now. The kmem savings are also important to later enable vmm(4) to use larged shared memory mappings for guest VM RAM. This adapts libkvm also because the amap structure layout has changed. Testing and fix of libkvm glitch in initial diff by tb@ Feedback and "time to get this in" kettenis@
Diffstat (limited to 'sys')
-rw-r--r--sys/uvm/uvm_amap.c553
-rw-r--r--sys/uvm/uvm_amap.h89
2 files changed, 460 insertions, 182 deletions
diff --git a/sys/uvm/uvm_amap.c b/sys/uvm/uvm_amap.c
index 977a95a2154..a333f4bd8a3 100644
--- a/sys/uvm/uvm_amap.c
+++ b/sys/uvm/uvm_amap.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm_amap.c,v 1.68 2016/05/08 16:29:57 stefan Exp $ */
+/* $OpenBSD: uvm_amap.c,v 1.69 2016/05/22 16:18:26 stefan Exp $ */
/* $NetBSD: uvm_amap.c,v 1.27 2000/11/25 06:27:59 chs Exp $ */
/*
@@ -44,20 +44,19 @@
#include <uvm/uvm_swap.h>
/*
- * pool for allocation of vm_map structures. note that in order to
+ * pools for allocation of vm_amap structures. note that in order to
* avoid an endless loop, the amap pool's allocator cannot allocate
* memory from an amap (it currently goes through the kernel uobj, so
* we are ok).
*/
struct pool uvm_amap_pool;
-
-/* Pools for amap slots for the most common amap slot sizes */
-struct pool uvm_amap_slot_pools[UVM_AMAP_CHUNK];
+struct pool uvm_small_amap_pool[UVM_AMAP_CHUNK];
+struct pool uvm_amap_chunk_pool;
LIST_HEAD(, vm_amap) amap_list;
-static char amap_slot_pool_names[UVM_AMAP_CHUNK][13];
+static char amap_small_pool_names[UVM_AMAP_CHUNK][9];
#define MALLOC_SLOT_UNIT (2 * sizeof(int) + sizeof(struct vm_anon *))
@@ -69,6 +68,8 @@ static struct vm_amap *amap_alloc1(int, int, int);
static __inline void amap_list_insert(struct vm_amap *);
static __inline void amap_list_remove(struct vm_amap *);
+struct vm_amap_chunk *amap_chunk_get(struct vm_amap *, int, int, int);
+
static __inline void
amap_list_insert(struct vm_amap *amap)
{
@@ -81,6 +82,72 @@ amap_list_remove(struct vm_amap *amap)
LIST_REMOVE(amap, am_list);
}
+/*
+ * amap_chunk_get: lookup a chunk for slot. if create is non-zero,
+ * the chunk is created if it does not yet exist.
+ *
+ * => returns the chunk on success or NULL on error
+ */
+struct vm_amap_chunk *
+amap_chunk_get(struct vm_amap *amap, int slot, int create, int waitf)
+{
+ int bucket = UVM_AMAP_BUCKET(amap, slot);
+ int baseslot = AMAP_BASE_SLOT(slot);
+ int n;
+ struct vm_amap_chunk *chunk, *newchunk, *pchunk = NULL;
+
+ if (UVM_AMAP_SMALL(amap))
+ return &amap->am_small;
+
+ for (chunk = amap->am_buckets[bucket]; chunk != NULL;
+ chunk = TAILQ_NEXT(chunk, ac_list)) {
+ if (UVM_AMAP_BUCKET(amap, chunk->ac_baseslot) != bucket)
+ break;
+ if (chunk->ac_baseslot == baseslot)
+ return chunk;
+ pchunk = chunk;
+ }
+ if (!create)
+ return NULL;
+
+ if (amap->am_nslot - baseslot >= UVM_AMAP_CHUNK)
+ n = UVM_AMAP_CHUNK;
+ else
+ n = amap->am_nslot - baseslot;
+
+ newchunk = pool_get(&uvm_amap_chunk_pool, waitf | PR_ZERO);
+ if (newchunk == NULL)
+ return NULL;
+
+ if (pchunk == NULL) {
+ TAILQ_INSERT_TAIL(&amap->am_chunks, newchunk, ac_list);
+ KASSERT(amap->am_buckets[bucket] == NULL);
+ amap->am_buckets[bucket] = newchunk;
+ } else
+ TAILQ_INSERT_AFTER(&amap->am_chunks, pchunk, newchunk,
+ ac_list);
+
+ amap->am_ncused++;
+ newchunk->ac_baseslot = baseslot;
+ newchunk->ac_nslot = n;
+ return newchunk;
+}
+
+void
+amap_chunk_free(struct vm_amap *amap, struct vm_amap_chunk *chunk)
+{
+ int bucket = UVM_AMAP_BUCKET(amap, chunk->ac_baseslot);
+
+ if (UVM_AMAP_SMALL(amap))
+ return;
+
+ TAILQ_REMOVE(&amap->am_chunks, chunk, ac_list);
+ if (amap->am_buckets[bucket] == chunk)
+ amap->am_buckets[bucket] = NULL;
+ pool_put(&uvm_amap_chunk_pool, chunk);
+ amap->am_ncused--;
+}
+
#ifdef UVM_AMAP_PPREF
/*
* what is ppref? ppref is an _optional_ amap feature which is used
@@ -157,19 +224,28 @@ void
amap_init(void)
{
int i;
+ size_t size;
/* Initialize the vm_amap pool. */
pool_init(&uvm_amap_pool, sizeof(struct vm_amap), 0, 0, PR_WAITOK,
"amappl", NULL);
pool_sethiwat(&uvm_amap_pool, 4096);
- for (i = 0; i < nitems(uvm_amap_slot_pools); i++) {
- snprintf(amap_slot_pool_names[i],
- sizeof(amap_slot_pool_names[0]), "amapslotpl%d", i + 1);
- pool_init(&uvm_amap_slot_pools[i], (i + 1) * MALLOC_SLOT_UNIT,
- 0, 0, PR_WAITOK, amap_slot_pool_names[i], NULL);
- pool_sethiwat(&uvm_amap_slot_pools[i], 4096);
+ /* initialize small amap pools */
+ for (i = 0; i < nitems(uvm_small_amap_pool); i++) {
+ snprintf(amap_small_pool_names[i],
+ sizeof(amap_small_pool_names[0]), "amappl%d", i + 1);
+ size = offsetof(struct vm_amap, am_small.ac_anon) +
+ (i + 1) * sizeof(struct vm_anon *);
+ pool_init(&uvm_small_amap_pool[i], size, 0, 0, 0,
+ amap_small_pool_names[i], NULL);
}
+
+ pool_init(&uvm_amap_chunk_pool,
+ sizeof(struct vm_amap_chunk) +
+ UVM_AMAP_CHUNK * sizeof(struct vm_anon *), 0, 0, 0,
+ "amapchunkpl", NULL);
+ pool_sethiwat(&uvm_amap_chunk_pool, 4096);
}
/*
@@ -180,9 +256,50 @@ static inline struct vm_amap *
amap_alloc1(int slots, int waitf, int lazyalloc)
{
struct vm_amap *amap;
+ struct vm_amap_chunk *chunk, *tmp;
+ int chunks, chunkperbucket = 1, hashshift = 0;
+ int buckets, i, n;
+ int pwaitf = (waitf == M_WAITOK) ? PR_WAITOK : PR_NOWAIT;
+
+ chunks = roundup(slots, UVM_AMAP_CHUNK) / UVM_AMAP_CHUNK;
+
+ if (lazyalloc) {
+ /*
+ * Basically, the amap is a hash map where the number of
+ * buckets is fixed. We select the number of buckets using the
+ * following strategy:
+ *
+ * 1. The maximal number of entries to search in a bucket upon
+ * a collision should be less than or equal to
+ * log2(slots / UVM_AMAP_CHUNK). This is the worst-case number
+ * of lookups we would have if we could chunk the amap. The
+ * log2(n) comes from the fact that amaps are chunked by
+ * splitting up their vm_map_entries and organizing those
+ * in a binary search tree.
+ *
+ * 2. The maximal number of entries in a bucket must be a
+ * power of two.
+ *
+ * The maximal number of entries per bucket is used to hash
+ * a slot to a bucket.
+ *
+ * In the future, this strategy could be refined to make it
+ * even harder/impossible that the total amount of KVA needed
+ * for the hash buckets of all amaps to exceed the maximal
+ * amount of KVA memory reserved for amaps.
+ */
+ chunkperbucket = 1 << hashshift;
+ while ((1 << chunkperbucket) * 2 <= chunks) {
+ hashshift++;
+ chunkperbucket = 1 << hashshift;
+ }
+ }
- amap = pool_get(&uvm_amap_pool, (waitf == M_WAITOK) ? PR_WAITOK
- : PR_NOWAIT);
+ if (slots > UVM_AMAP_CHUNK)
+ amap = pool_get(&uvm_amap_pool, pwaitf);
+ else
+ amap = pool_get(&uvm_small_amap_pool[slots - 1],
+ pwaitf | PR_ZERO);
if (amap == NULL)
return(NULL);
@@ -196,25 +313,48 @@ amap_alloc1(int slots, int waitf, int lazyalloc)
amap->am_nslot = slots;
amap->am_nused = 0;
- if (slots > UVM_AMAP_CHUNK)
- amap->am_slots = malloc(slots * MALLOC_SLOT_UNIT,
- M_UVMAMAP, waitf);
- else
- amap->am_slots = pool_get(
- &uvm_amap_slot_pools[slots - 1],
- (waitf == M_WAITOK) ? PR_WAITOK : PR_NOWAIT);
+ if (UVM_AMAP_SMALL(amap))
+ return (amap);
- if (amap->am_slots == NULL)
+ amap->am_ncused = 0;
+ TAILQ_INIT(&amap->am_chunks);
+ amap->am_hashshift = hashshift;
+ amap->am_buckets = NULL;
+
+ buckets = howmany(chunks, chunkperbucket);
+ amap->am_buckets = mallocarray(buckets, sizeof(amap->am_buckets),
+ M_UVMAMAP, waitf | (lazyalloc ? M_ZERO : 0));
+ if (amap->am_buckets == NULL)
goto fail1;
- amap->am_bckptr = (int *)(((char *)amap->am_slots) + slots *
- sizeof(int));
- amap->am_anon = (struct vm_anon **)(((char *)amap->am_bckptr) +
- slots * sizeof(int));
+ if (!lazyalloc) {
+ for (i = 0; i < buckets; i++) {
+ if (i == buckets - 1) {
+ n = slots % UVM_AMAP_CHUNK;
+ if (n == 0)
+ n = UVM_AMAP_CHUNK;
+ } else
+ n = UVM_AMAP_CHUNK;
+
+ chunk = pool_get(&uvm_amap_chunk_pool,
+ PR_ZERO | pwaitf);
+ if (chunk == NULL)
+ goto fail1;
+
+ amap->am_buckets[i] = chunk;
+ amap->am_ncused++;
+ chunk->ac_baseslot = i * UVM_AMAP_CHUNK;
+ chunk->ac_nslot = n;
+ TAILQ_INSERT_TAIL(&amap->am_chunks, chunk, ac_list);
+ }
+ }
return(amap);
fail1:
+ free(amap->am_buckets, M_UVMAMAP, 0);
+ TAILQ_FOREACH_SAFE(chunk, &amap->am_chunks, ac_list, tmp)
+ pool_put(&uvm_amap_chunk_pool, chunk);
pool_put(&uvm_amap_pool, amap);
return (NULL);
}
@@ -234,11 +374,8 @@ amap_alloc(vaddr_t sz, int waitf, int lazyalloc)
AMAP_B2SLOT(slots, sz); /* load slots */
amap = amap_alloc1(slots, waitf, lazyalloc);
- if (amap) {
- memset(amap->am_anon, 0,
- amap->am_nslot * sizeof(struct vm_anon *));
+ if (amap)
amap_list_insert(amap);
- }
return(amap);
}
@@ -252,22 +389,24 @@ amap_alloc(vaddr_t sz, int waitf, int lazyalloc)
void
amap_free(struct vm_amap *amap)
{
+ struct vm_amap_chunk *chunk, *tmp;
KASSERT(amap->am_ref == 0 && amap->am_nused == 0);
KASSERT((amap->am_flags & AMAP_SWAPOFF) == 0);
- if (amap->am_nslot > UVM_AMAP_CHUNK)
- free(amap->am_slots, M_UVMAMAP, 0);
- else
- pool_put(&uvm_amap_slot_pools[amap->am_nslot - 1],
- amap->am_slots);
-
#ifdef UVM_AMAP_PPREF
if (amap->am_ppref && amap->am_ppref != PPREF_NONE)
free(amap->am_ppref, M_UVMAMAP, 0);
#endif
- pool_put(&uvm_amap_pool, amap);
+ if (UVM_AMAP_SMALL(amap))
+ pool_put(&uvm_small_amap_pool[amap->am_nslot - 1], amap);
+ else {
+ TAILQ_FOREACH_SAFE(chunk, &amap->am_chunks, ac_list, tmp)
+ pool_put(&uvm_amap_chunk_pool, chunk);
+ free(amap->am_buckets, M_UVMAMAP, 0);
+ pool_put(&uvm_amap_pool, amap);
+ }
}
/*
@@ -280,8 +419,9 @@ amap_free(struct vm_amap *amap)
void
amap_wipeout(struct vm_amap *amap)
{
- int lcv, slot;
+ int slot;
struct vm_anon *anon;
+ struct vm_amap_chunk *chunk;
KASSERT(amap->am_ref == 0);
@@ -291,19 +431,25 @@ amap_wipeout(struct vm_amap *amap)
}
amap_list_remove(amap);
- for (lcv = 0 ; lcv < amap->am_nused ; lcv++) {
- int refs;
+ AMAP_CHUNK_FOREACH(chunk, amap) {
+ int i, refs, map = chunk->ac_usedmap;
- slot = amap->am_slots[lcv];
- anon = amap->am_anon[slot];
+ for (i = ffs(map); i != 0; i = ffs(map)) {
+ slot = i - 1;
+ map ^= 1 << slot;
+ anon = chunk->ac_anon[slot];
- if (anon == NULL || anon->an_ref == 0)
- panic("amap_wipeout: corrupt amap");
+ if (anon == NULL || anon->an_ref == 0)
+ panic("amap_wipeout: corrupt amap");
- refs = --anon->an_ref;
- if (refs == 0) {
- /* we had the last reference to a vm_anon. free it. */
- uvm_anfree(anon);
+ refs = --anon->an_ref;
+ if (refs == 0) {
+ /*
+ * we had the last reference to a vm_anon.
+ * free it.
+ */
+ uvm_anfree(anon);
+ }
}
}
@@ -332,6 +478,8 @@ amap_copy(struct vm_map *map, struct vm_map_entry *entry, int waitf,
struct vm_amap *amap, *srcamap;
int slots, lcv, lazyalloc = 0;
vaddr_t chunksize;
+ int i, j, k, n, srcslot;
+ struct vm_amap_chunk *chunk = NULL, *srcchunk = NULL;
/* is there a map to copy? if not, create one from scratch. */
if (entry->aref.ar_amap == NULL) {
@@ -379,6 +527,9 @@ amap_copy(struct vm_map *map, struct vm_map_entry *entry, int waitf,
/* looks like we need to copy the map. */
AMAP_B2SLOT(slots, entry->end - entry->start);
+ if (!UVM_AMAP_SMALL(entry->aref.ar_amap) &&
+ entry->aref.ar_amap->am_hashshift != 0)
+ lazyalloc = 1;
amap = amap_alloc1(slots, waitf, lazyalloc);
if (amap == NULL)
return;
@@ -398,17 +549,39 @@ amap_copy(struct vm_map *map, struct vm_map_entry *entry, int waitf,
}
/* we must copy it now. */
- for (lcv = 0 ; lcv < slots; lcv++) {
- amap->am_anon[lcv] =
- srcamap->am_anon[entry->aref.ar_pageoff + lcv];
- if (amap->am_anon[lcv] == NULL)
+ for (lcv = 0; lcv < slots; lcv += n) {
+ srcslot = entry->aref.ar_pageoff + lcv;
+ i = UVM_AMAP_SLOTIDX(lcv);
+ j = UVM_AMAP_SLOTIDX(srcslot);
+ n = UVM_AMAP_CHUNK;
+ if (i > j)
+ n -= i;
+ else
+ n -= j;
+ if (lcv + n > slots)
+ n = slots - lcv;
+
+ srcchunk = amap_chunk_get(srcamap, srcslot, 0, PR_NOWAIT);
+ if (srcchunk == NULL)
continue;
- amap->am_anon[lcv]->an_ref++;
- amap->am_bckptr[lcv] = amap->am_nused;
- amap->am_slots[amap->am_nused] = lcv;
- amap->am_nused++;
+
+ chunk = amap_chunk_get(amap, lcv, 1, PR_NOWAIT);
+ if (chunk == NULL) {
+ amap->am_ref = 0;
+ amap_wipeout(amap);
+ return;
+ }
+
+ for (k = 0; k < n; i++, j++, k++) {
+ chunk->ac_anon[i] = srcchunk->ac_anon[j];
+ if (chunk->ac_anon[i] == NULL)
+ continue;
+
+ chunk->ac_usedmap |= (1 << i);
+ chunk->ac_anon[i]->an_ref++;
+ amap->am_nused++;
+ }
}
- KASSERT(lcv == amap->am_nslot);
/*
* drop our reference to the old amap (srcamap).
@@ -452,9 +625,10 @@ void
amap_cow_now(struct vm_map *map, struct vm_map_entry *entry)
{
struct vm_amap *amap = entry->aref.ar_amap;
- int lcv, slot;
+ int slot;
struct vm_anon *anon, *nanon;
struct vm_page *pg, *npg;
+ struct vm_amap_chunk *chunk;
/*
* note that if we wait, we must ReStart the "lcv" for loop because
@@ -462,22 +636,27 @@ amap_cow_now(struct vm_map *map, struct vm_map_entry *entry)
* am_anon[] array on us.
*/
ReStart:
- for (lcv = 0 ; lcv < amap->am_nused ; lcv++) {
- /* get the page */
- slot = amap->am_slots[lcv];
- anon = amap->am_anon[slot];
- pg = anon->an_page;
+ AMAP_CHUNK_FOREACH(chunk, amap) {
+ int i, map = chunk->ac_usedmap;
- /* page must be resident since parent is wired */
- if (pg == NULL)
- panic("amap_cow_now: non-resident wired page"
- " in anon %p", anon);
+ for (i = ffs(map); i != 0; i = ffs(map)) {
+ slot = i - 1;
+ map ^= 1 << slot;
+ anon = chunk->ac_anon[slot];
+ pg = anon->an_page;
+
+ /* page must be resident since parent is wired */
+ if (pg == NULL)
+ panic("amap_cow_now: non-resident wired page"
+ " in anon %p", anon);
+
+ /*
+ * if the anon ref count is one, we are safe (the child
+ * has exclusive access to the page).
+ */
+ if (anon->an_ref <= 1)
+ continue;
- /*
- * if the anon ref count is one, we are safe (the child has
- * exclusive access to the page).
- */
- if (anon->an_ref > 1) {
/*
* if the page is busy then we have to wait for
* it and then restart.
@@ -507,14 +686,14 @@ ReStart:
uvm_wait("cownowpage");
goto ReStart;
}
-
+
/*
* got it... now we can copy the data and replace anon
* with our new one...
*/
uvm_pagecopy(pg, npg); /* old -> new */
anon->an_ref--; /* can't drop to zero */
- amap->am_anon[slot] = nanon; /* replace */
+ chunk->ac_anon[slot] = nanon; /* replace */
/*
* drop PG_BUSY on new page ... since we have had its
@@ -658,61 +837,83 @@ amap_pp_adjref(struct vm_amap *amap, int curslot, vsize_t slotlen, int adjval)
void
amap_wiperange(struct vm_amap *amap, int slotoff, int slots)
{
- int byanon, lcv, stop, curslot, ptr, slotend;
+ int curslot, i, map, bybucket;
+ int bucket, startbucket, endbucket;
+ int startbase, endbase;
struct vm_anon *anon;
+ struct vm_amap_chunk *chunk, *nchunk;
/*
- * we can either traverse the amap by am_anon or by am_slots depending
- * on which is cheaper. decide now.
+ * we can either traverse the amap by am_chunks or by am_buckets
+ * depending on which is cheaper. decide now.
*/
- if (slots < amap->am_nused) {
- byanon = TRUE;
- lcv = slotoff;
- stop = slotoff + slots;
+ startbucket = UVM_AMAP_BUCKET(amap, slotoff);
+ endbucket = UVM_AMAP_BUCKET(amap, slotoff + slots - 1);
+
+ if (UVM_AMAP_SMALL(amap)) {
+ bybucket = FALSE;
+ chunk = &amap->am_small;
+ } else if (endbucket + 1 - startbucket >= amap->am_ncused) {
+ bybucket = FALSE;
+ chunk = TAILQ_FIRST(&amap->am_chunks);
} else {
- byanon = FALSE;
- lcv = 0;
- stop = amap->am_nused;
- slotend = slotoff + slots;
+ bybucket = TRUE;
+ bucket = startbucket;
+ chunk = amap->am_buckets[bucket];
}
- while (lcv < stop) {
- int refs;
+ startbase = AMAP_BASE_SLOT(slotoff);
+ endbase = AMAP_BASE_SLOT(slotoff + slots - 1);
+ for (;;) {
+ if (chunk == NULL || (bybucket &&
+ UVM_AMAP_BUCKET(amap, chunk->ac_baseslot) != bucket)) {
+ if (!bybucket || bucket >= endbucket)
+ break;
+ bucket++;
+ chunk = amap->am_buckets[bucket];
+ continue;
+ } else if (!bybucket) {
+ if (chunk->ac_baseslot + chunk->ac_nslot <= slotoff)
+ goto next;
+ if (chunk->ac_baseslot >= slotoff + slots)
+ goto next;
+ }
- if (byanon) {
- curslot = lcv++; /* lcv advances here */
- if (amap->am_anon[curslot] == NULL)
- continue;
- } else {
- curslot = amap->am_slots[lcv];
- if (curslot < slotoff || curslot >= slotend) {
- lcv++; /* lcv advances here */
- continue;
+ map = chunk->ac_usedmap;
+ if (startbase == chunk->ac_baseslot)
+ map &= ~((1 << (slotoff - startbase)) - 1);
+ if (endbase == chunk->ac_baseslot)
+ map &= (1 << (slotoff + slots - endbase)) - 1;
+
+ for (i = ffs(map); i != 0; i = ffs(map)) {
+ int refs;
+
+ curslot = i - 1;
+ map ^= 1 << curslot;
+ chunk->ac_usedmap ^= 1 << curslot;
+ anon = chunk->ac_anon[curslot];
+
+ /* remove it from the amap */
+ chunk->ac_anon[curslot] = NULL;
+
+ amap->am_nused--;
+
+ /* drop anon reference count */
+ refs = --anon->an_ref;
+ if (refs == 0) {
+ /*
+ * we just eliminated the last reference to an
+ * anon. free it.
+ */
+ uvm_anfree(anon);
}
- stop--; /* drop stop, since anon will be removed */
}
- anon = amap->am_anon[curslot];
-
- /* remove it from the amap */
- amap->am_anon[curslot] = NULL;
- ptr = amap->am_bckptr[curslot];
- if (ptr != (amap->am_nused - 1)) {
- amap->am_slots[ptr] =
- amap->am_slots[amap->am_nused - 1];
- amap->am_bckptr[amap->am_slots[ptr]] =
- ptr; /* back ptr. */
- }
- amap->am_nused--;
- /* drop anon reference count */
- refs = --anon->an_ref;
- if (refs == 0) {
- /*
- * we just eliminated the last reference to an anon.
- * free it.
- */
- uvm_anfree(anon);
- }
+next:
+ nchunk = TAILQ_NEXT(chunk, ac_list);
+ if (chunk->ac_usedmap == 0)
+ amap_chunk_free(amap, chunk);
+ chunk = nchunk;
}
}
@@ -734,31 +935,38 @@ amap_swap_off(int startslot, int endslot)
boolean_t rv = FALSE;
for (am = LIST_FIRST(&amap_list); am != NULL && !rv; am = am_next) {
- int i;
+ int i, map;
+ struct vm_amap_chunk *chunk;
- for (i = 0; i < am->am_nused; i++) {
- int slot;
- int swslot;
- struct vm_anon *anon;
+again:
+ AMAP_CHUNK_FOREACH(chunk, am) {
+ map = chunk->ac_usedmap;
- slot = am->am_slots[i];
- anon = am->am_anon[slot];
+ for (i = ffs(map); i != 0; i = ffs(map)) {
+ int swslot;
+ int slot = i - 1;
+ struct vm_anon *anon;
- swslot = anon->an_swslot;
- if (swslot < startslot || endslot <= swslot) {
- continue;
- }
+ map ^= 1 << slot;
+ anon = chunk->ac_anon[slot];
- am->am_flags |= AMAP_SWAPOFF;
+ swslot = anon->an_swslot;
+ if (swslot < startslot || endslot <= swslot) {
+ continue;
+ }
- rv = uvm_anon_pagein(anon);
+ am->am_flags |= AMAP_SWAPOFF;
- am->am_flags &= ~AMAP_SWAPOFF;
- if (rv || amap_refs(am) == 0)
- break;
- i = 0;
+ rv = uvm_anon_pagein(anon);
+
+ am->am_flags &= ~AMAP_SWAPOFF;
+ if (rv || amap_refs(am) == 0)
+ goto nextamap;
+ goto again;
+ }
}
+nextamap:
am_next = LIST_NEXT(am, am_list);
if (amap_refs(am) == 0)
amap_wipeout(am);
@@ -775,6 +983,7 @@ amap_lookup(struct vm_aref *aref, vaddr_t offset)
{
int slot;
struct vm_amap *amap = aref->ar_amap;
+ struct vm_amap_chunk *chunk;
AMAP_B2SLOT(slot, offset);
slot += aref->ar_pageoff;
@@ -782,7 +991,11 @@ amap_lookup(struct vm_aref *aref, vaddr_t offset)
if (slot >= amap->am_nslot)
panic("amap_lookup: offset out of range");
- return(amap->am_anon[slot]);
+ chunk = amap_chunk_get(amap, slot, 0, PR_NOWAIT);
+ if (chunk == NULL)
+ return NULL;
+
+ return chunk->ac_anon[UVM_AMAP_SLOTIDX(slot)];
}
/*
@@ -794,8 +1007,9 @@ void
amap_lookups(struct vm_aref *aref, vaddr_t offset,
struct vm_anon **anons, int npages)
{
- int slot;
+ int i, lcv, n, slot;
struct vm_amap *amap = aref->ar_amap;
+ struct vm_amap_chunk *chunk = NULL;
AMAP_B2SLOT(slot, offset);
slot += aref->ar_pageoff;
@@ -803,9 +1017,19 @@ amap_lookups(struct vm_aref *aref, vaddr_t offset,
if ((slot + (npages - 1)) >= amap->am_nslot)
panic("amap_lookups: offset out of range");
- memcpy(anons, &amap->am_anon[slot], npages * sizeof(struct vm_anon *));
+ for (i = 0, lcv = slot; lcv < slot + npages; i += n, lcv += n) {
+ n = UVM_AMAP_CHUNK - UVM_AMAP_SLOTIDX(lcv);
+ if (lcv + n > slot + npages)
+ n = slot + npages - lcv;
- return;
+ chunk = amap_chunk_get(amap, lcv, 0, PR_NOWAIT);
+ if (chunk == NULL)
+ memset(&anons[i], 0, n * sizeof(*anons[i]));
+ else
+ memcpy(&anons[i],
+ &chunk->ac_anon[UVM_AMAP_SLOTIDX(lcv)],
+ n * sizeof(*anons[i]));
+ }
}
/*
@@ -816,6 +1040,18 @@ amap_lookups(struct vm_aref *aref, vaddr_t offset,
void
amap_populate(struct vm_aref *aref, vaddr_t offset)
{
+ int slot;
+ struct vm_amap *amap = aref->ar_amap;
+ struct vm_amap_chunk *chunk;
+
+ AMAP_B2SLOT(slot, offset);
+ slot += aref->ar_pageoff;
+
+ if (slot >= amap->am_nslot)
+ panic("amap_populate: offset out of range");
+
+ chunk = amap_chunk_get(amap, slot, 1, PR_WAITOK);
+ KASSERT(chunk != NULL);
}
/*
@@ -829,33 +1065,37 @@ amap_add(struct vm_aref *aref, vaddr_t offset, struct vm_anon *anon,
{
int slot;
struct vm_amap *amap = aref->ar_amap;
+ struct vm_amap_chunk *chunk;
AMAP_B2SLOT(slot, offset);
slot += aref->ar_pageoff;
if (slot >= amap->am_nslot)
panic("amap_add: offset out of range");
+ chunk = amap_chunk_get(amap, slot, 1, PR_NOWAIT);
+ if (chunk == NULL)
+ return 1;
+ slot = UVM_AMAP_SLOTIDX(slot);
if (replace) {
- if (amap->am_anon[slot] == NULL)
+ if (chunk->ac_anon[slot] == NULL)
panic("amap_add: replacing null anon");
- if (amap->am_anon[slot]->an_page != NULL &&
+ if (chunk->ac_anon[slot]->an_page != NULL &&
(amap->am_flags & AMAP_SHARED) != 0) {
- pmap_page_protect(amap->am_anon[slot]->an_page,
+ pmap_page_protect(chunk->ac_anon[slot]->an_page,
PROT_NONE);
/*
* XXX: suppose page is supposed to be wired somewhere?
*/
}
} else { /* !replace */
- if (amap->am_anon[slot] != NULL)
+ if (chunk->ac_anon[slot] != NULL)
panic("amap_add: slot in use");
- amap->am_bckptr[slot] = amap->am_nused;
- amap->am_slots[amap->am_nused] = slot;
+ chunk->ac_usedmap |= 1 << slot;
amap->am_nused++;
}
- amap->am_anon[slot] = anon;
+ chunk->ac_anon[slot] = anon;
return 0;
}
@@ -866,26 +1106,29 @@ amap_add(struct vm_aref *aref, vaddr_t offset, struct vm_anon *anon,
void
amap_unadd(struct vm_aref *aref, vaddr_t offset)
{
- int ptr, slot;
+ int slot;
struct vm_amap *amap = aref->ar_amap;
+ struct vm_amap_chunk *chunk;
AMAP_B2SLOT(slot, offset);
slot += aref->ar_pageoff;
if (slot >= amap->am_nslot)
panic("amap_unadd: offset out of range");
+ chunk = amap_chunk_get(amap, slot, 0, PR_NOWAIT);
+ if (chunk == NULL)
+ panic("amap_unadd: chunk for slot %d not present", slot);
- if (amap->am_anon[slot] == NULL)
+ slot = UVM_AMAP_SLOTIDX(slot);
+ if (chunk->ac_anon[slot] == NULL)
panic("amap_unadd: nothing there");
- amap->am_anon[slot] = NULL;
- ptr = amap->am_bckptr[slot];
-
- if (ptr != (amap->am_nused - 1)) { /* swap to keep slots contig? */
- amap->am_slots[ptr] = amap->am_slots[amap->am_nused - 1];
- amap->am_bckptr[amap->am_slots[ptr]] = ptr; /* back link */
- }
+ chunk->ac_anon[slot] = NULL;
+ chunk->ac_usedmap &= ~(1 << slot);
amap->am_nused--;
+
+ if (chunk->ac_usedmap == 0)
+ amap_chunk_free(amap, chunk);
}
/*
diff --git a/sys/uvm/uvm_amap.h b/sys/uvm/uvm_amap.h
index 1c5abfb8ab1..542a1ece6c3 100644
--- a/sys/uvm/uvm_amap.h
+++ b/sys/uvm/uvm_amap.h
@@ -1,4 +1,4 @@
-/* $OpenBSD: uvm_amap.h,v 1.26 2016/05/08 16:29:57 stefan Exp $ */
+/* $OpenBSD: uvm_amap.h,v 1.27 2016/05/22 16:18:26 stefan Exp $ */
/* $NetBSD: uvm_amap.h,v 1.14 2001/02/18 21:19:08 chs Exp $ */
/*
@@ -120,49 +120,71 @@ boolean_t amap_swap_off(int, int);
#define UVM_AMAP_PPREF /* track partial references */
/*
- * here is the definition of the vm_amap structure for this implementation.
+ * here is the definition of the vm_amap structure and helper structures for
+ * this implementation.
*/
+struct vm_amap_chunk {
+ TAILQ_ENTRY(vm_amap_chunk) ac_list;
+ int ac_baseslot;
+ uint16_t ac_usedmap;
+ uint16_t ac_nslot;
+ struct vm_anon *ac_anon[];
+};
+
struct vm_amap {
int am_ref; /* reference count */
int am_flags; /* flags */
int am_nslot; /* # of slots currently in map */
int am_nused; /* # of slots currently in use */
- int *am_slots; /* contig array of active slots */
- int *am_bckptr; /* back pointer array to am_slots */
- struct vm_anon **am_anon; /* array of anonymous pages */
#ifdef UVM_AMAP_PPREF
int *am_ppref; /* per page reference count (if !NULL) */
#endif
LIST_ENTRY(vm_amap) am_list;
+
+ union {
+ struct {
+ struct vm_amap_chunk **amn_buckets;
+ TAILQ_HEAD(, vm_amap_chunk) amn_chunks;
+ int amn_ncused; /* # of chunkers currently in use */
+ int amn_hashshift; /* shift count to hash slot to bucket */
+ } ami_normal;
+
+ /*
+ * MUST be last element in vm_amap because it contains a
+ * variably sized array element.
+ */
+ struct vm_amap_chunk ami_small;
+ } am_impl;
+
+#define am_buckets am_impl.ami_normal.amn_buckets
+#define am_chunks am_impl.ami_normal.amn_chunks
+#define am_ncused am_impl.ami_normal.amn_ncused
+#define am_hashshift am_impl.ami_normal.amn_hashshift
+
+#define am_small am_impl.ami_small
};
/*
- * note that am_slots, am_bckptr, and am_anon are arrays. this allows
- * fast lookup of pages based on their virual address at the expense of
- * some extra memory. in the future we should be smarter about memory
- * usage and fall back to a non-array based implementation on systems
- * that are short of memory (XXXCDC).
+ * The entries in an amap are called slots. For example an amap that
+ * covers four pages is said to have four slots.
*
- * the entries in the array are called slots... for example an amap that
- * covers four pages of virtual memory is said to have four slots. here
- * is an example of the array usage for a four slot amap. note that only
- * slots one and three have anons assigned to them. "D/C" means that we
- * "don't care" about the value.
- *
- * 0 1 2 3
- * am_anon: NULL, anon0, NULL, anon1 (actual pointers to anons)
- * am_bckptr: D/C, 1, D/C, 0 (points to am_slots entry)
+ * The slots of an amap are clustered into chunks of UVM_AMAP_CHUNK
+ * slots each. The data structure of a chunk is vm_amap_chunk.
+ * Every chunk contains an array of pointers to vm_anon, and a bitmap
+ * is used to represent which of the slots are in use.
*
- * am_slots: 3, 1, D/C, D/C (says slots 3 and 1 are in use)
- *
- * note that am_bckptr is D/C if the slot in am_anon is set to NULL.
- * to find the entry in am_slots for an anon, look at am_bckptr[slot],
- * thus the entry for slot 3 in am_slots[] is at am_slots[am_bckptr[3]].
- * in general, if am_anon[X] is non-NULL, then the following must be
- * true: am_slots[am_bckptr[X]] == X
+ * Small amaps of up to UVM_AMAP_CHUNK slots have the chunk directly
+ * embedded in the amap structure.
*
- * note that am_slots is always contig-packed.
+ * amaps with more slots are normal amaps and organize chunks in a hash
+ * table. The hash table is organized as an array of buckets.
+ * All chunks of the amap are additionally stored in a linked list.
+ * Chunks that belong to the same hash bucket are stored in the list
+ * consecutively. When all slots in a chunk are unused, the chunk is freed.
+ *
+ * For large amaps, the bucket array can grow large. See the description
+ * below how large bucket arrays are avoided.
*/
/*
@@ -205,6 +227,11 @@ struct vm_amap {
#define UVM_AMAP_LARGE 256 /* # of slots in "large" amap */
#define UVM_AMAP_CHUNK 16 /* # of slots to chunk large amaps in */
+#define UVM_AMAP_SMALL(amap) ((amap)->am_nslot <= UVM_AMAP_CHUNK)
+#define UVM_AMAP_SLOTIDX(slot) ((slot) % UVM_AMAP_CHUNK)
+#define UVM_AMAP_BUCKET(amap, slot) \
+ (((slot) / UVM_AMAP_CHUNK) >> (amap)->am_hashshift)
+
#ifdef _KERNEL
/*
@@ -217,6 +244,14 @@ struct vm_amap {
(S) = (B) >> PAGE_SHIFT; \
}
+#define AMAP_CHUNK_FOREACH(chunk, amap) \
+ for (chunk = (UVM_AMAP_SMALL(amap) ? \
+ &(amap)->am_small : TAILQ_FIRST(&(amap)->am_chunks)); \
+ (chunk) != NULL; (chunk) = TAILQ_NEXT(chunk, ac_list))
+
+#define AMAP_BASE_SLOT(slot) \
+ (((slot) / UVM_AMAP_CHUNK) * UVM_AMAP_CHUNK)
+
/*
* flags macros
*/