diff options
author | Niklas Hallqvist <niklas@cvs.openbsd.org> | 1997-09-22 20:44:54 +0000 |
---|---|---|
committer | Niklas Hallqvist <niklas@cvs.openbsd.org> | 1997-09-22 20:44:54 +0000 |
commit | 5e4b014712c2454a243e3ce76e5cefe718a93d8a (patch) | |
tree | e3240a42a16a07f7b365abbab985f9d81c150044 /sys/vm | |
parent | fad8e16d2411526ff8d913825a3424dc13ca91d6 (diff) |
From NetBSD: vm_page_alloc_memory giving continuous memory
Diffstat (limited to 'sys/vm')
-rw-r--r-- | sys/vm/vm_page.c | 346 | ||||
-rw-r--r-- | sys/vm/vm_page.h | 29 |
2 files changed, 325 insertions, 50 deletions
diff --git a/sys/vm/vm_page.c b/sys/vm/vm_page.c index 55241f87fad..c2a2285d8db 100644 --- a/sys/vm/vm_page.c +++ b/sys/vm/vm_page.c @@ -1,5 +1,44 @@ -/* $OpenBSD: vm_page.c,v 1.6 1997/07/25 06:03:10 mickey Exp $ */ -/* $NetBSD: vm_page.c,v 1.28 1996/02/05 01:54:05 christos Exp $ */ +/* $OpenBSD: vm_page.c,v 1.7 1997/09/22 20:44:52 niklas Exp $ */ +/* $NetBSD: vm_page.c,v 1.31 1997/06/06 23:10:23 thorpej Exp $ */ + +#define VM_PAGE_ALLOC_MEMORY_STATS + +/*- + * Copyright (c) 1997 The NetBSD Foundation, Inc. + * All rights reserved. + * + * This code is derived from software contributed to The NetBSD Foundation + * by Jason R. Thorpe of the Numerical Aerospace Simulation Facility, + * NASA Ames Research Center. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * 3. All advertising materials mentioning features or use of this software + * must display the following acknowledgement: + * This product includes software developed by the NetBSD + * Foundation, Inc. and its contributors. + * 4. Neither the name of The NetBSD Foundation nor the names of its + * contributors may be used to endorse or promote products derived + * from this software without specific prior written permission. + * + * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS + * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED + * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR + * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS + * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS + * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN + * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) + * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE + * POSSIBILITY OF SUCH DAMAGE. + */ /* * Copyright (c) 1991, 1993 @@ -110,6 +149,7 @@ simple_lock_data_t vm_page_queue_free_lock; boolean_t vm_page_startup_initialized; vm_page_t vm_page_array; +int vm_page_count; #ifndef MACHINE_NONCONTIG long first_page; long last_page; @@ -117,7 +157,6 @@ vm_offset_t first_phys_addr; vm_offset_t last_phys_addr; #else u_long first_page; -int vm_page_count; #endif /* MACHINE_NONCONTIG */ vm_size_t page_mask; int page_shift; @@ -168,7 +207,6 @@ vm_page_bootstrap(startp, endp) extern vm_offset_t kentry_data; extern vm_size_t kentry_data_size; - /* * Initialize the locks */ @@ -195,13 +233,13 @@ vm_page_bootstrap(startp, endp) * map (they should use their own maps). */ kentry_data_size = round_page(MAX_KMAP*sizeof(struct vm_map) + - MAX_KMAPENT*sizeof(struct vm_map_entry)); - kentry_data = (vm_offset_t) pmap_steal_memory(kentry_data_size); + MAX_KMAPENT*sizeof(struct vm_map_entry)); + kentry_data = (vm_offset_t)pmap_steal_memory(kentry_data_size); /* * Validate these zone addresses. */ - bzero((caddr_t) kentry_data, kentry_data_size); + bzero((caddr_t)kentry_data, kentry_data_size); /* * Allocate (and initialize) the virtual-to-physical @@ -270,13 +308,12 @@ vm_page_startup(start, end) { register vm_page_t m; register struct pglist *bucket; - vm_size_t npages; + int npages; int i; vm_offset_t pa; extern vm_offset_t kentry_data; extern vm_size_t kentry_data_size; - /* * Initialize the locks */ @@ -340,7 +377,7 @@ vm_page_startup(start, end) * map (they should use their own maps). */ kentry_data_size = round_page(MAX_KMAP*sizeof(struct vm_map) + - MAX_KMAPENT*sizeof(struct vm_map_entry)); + MAX_KMAPENT*sizeof(struct vm_map_entry)); kentry_data = (vm_offset_t) pmap_bootstrap_alloc(kentry_data_size); /* @@ -348,35 +385,36 @@ vm_page_startup(start, end) * available for use (taking into account the overhead * of a page structure per page). */ - cnt.v_free_count = npages = (*end - *start + sizeof(struct vm_page)) - / (PAGE_SIZE + sizeof(struct vm_page)); + cnt.v_free_count = vm_page_count = (*end - *start + + sizeof(struct vm_page)) / (PAGE_SIZE + sizeof(struct vm_page)); /* * Record the extent of physical memory that the * virtual memory system manages. */ first_page = *start; - first_page += npages*sizeof(struct vm_page); + first_page += vm_page_count * sizeof(struct vm_page); first_page = atop(round_page(first_page)); - last_page = first_page + npages - 1; + last_page = first_page + vm_page_count - 1; first_phys_addr = ptoa(first_page); - last_phys_addr = ptoa(last_page) + PAGE_MASK; - + last_phys_addr = ptoa(last_page) + PAGE_MASK; /* * Allocate and clear the mem entry structures. */ m = vm_page_array = (vm_page_t) - pmap_bootstrap_alloc(npages * sizeof(struct vm_page)); + pmap_bootstrap_alloc(vm_page_count * sizeof(struct vm_page)); + bzero(vm_page_array, vm_page_count * sizeof(struct vm_page)); /* * Initialize the mem entry structures now, and * put them in the free queue. */ pa = first_phys_addr; + npages = vm_page_count; while (npages--) { - m->flags = 0; + m->flags = PG_FREE; m->object = NULL; m->phys_addr = pa; TAILQ_INSERT_TAIL(&vm_page_queue_free, m, pageq); @@ -406,30 +444,22 @@ pmap_steal_memory(size) { vm_offset_t addr, vaddr, paddr; -#ifdef i386 /* XXX i386 calls pmap_steal_memory before vm_mem_init() */ - if (cnt.v_page_size == 0) /* XXX */ - vm_set_page_size(); -#endif - /* * We round the size to an integer multiple. */ - size = (size + 3) &~ 3; /* XXX */ /* * If this is the first call to pmap_steal_memory, * we have to initialize ourself. */ - - if (virtual_space_start == virtual_space_end) { + if (virtual_space_start == virtual_space_end) { pmap_virtual_space(&virtual_space_start, &virtual_space_end); /* * The initial values must be aligned properly, and * we don't trust the pmap module to do it right. */ - virtual_space_start = round_page(virtual_space_start); virtual_space_end = trunc_page(virtual_space_end); } @@ -437,17 +467,14 @@ pmap_steal_memory(size) /* * Allocate virtual memory for this request. */ - addr = virtual_space_start; virtual_space_start += size; /* * Allocate and map physical pages to back new virtual pages. */ - - for (vaddr = round_page(addr); - vaddr < addr + size; - vaddr += PAGE_SIZE) { + for (vaddr = round_page(addr); vaddr < addr + size; + vaddr += PAGE_SIZE) { if (!pmap_next_page(&paddr)) panic("pmap_steal_memory"); @@ -455,9 +482,8 @@ pmap_steal_memory(size) * XXX Logically, these mappings should be wired, * but some pmap modules barf if they are. */ - pmap_enter(pmap_kernel(), vaddr, paddr, - VM_PROT_READ|VM_PROT_WRITE, FALSE); + VM_PROT_READ|VM_PROT_WRITE, FALSE); } return addr; @@ -494,16 +520,15 @@ pmap_startup(startp, endp) vm_page_array = (vm_page_t) pmap_steal_memory(vm_page_count * sizeof(*vm_page_array)); + bzero(vm_page_array, vm_page_count * sizeof(*vm_page_array)); #ifdef DIAGNOSTIC /* * Initialize everyting in case the holes are stepped in, * and set PA to something that will cause a panic... */ - for (i = 0; i < vm_page_count; i++) { - bzero(&vm_page_array[i], sizeof(*vm_page_array)); + for (i = 0; i < vm_page_count; i++) vm_page_array[i].phys_addr = 0xdeadbeef; - } #endif /* @@ -587,7 +612,7 @@ vm_page_insert(mem, object, offset) simple_lock(&bucket_lock); TAILQ_INSERT_TAIL(bucket, mem, hashq); simple_unlock(&bucket_lock); - (void) splx(spl); + (void)splx(spl); /* * Now link into the object's list of backed pages. @@ -604,7 +629,7 @@ vm_page_insert(mem, object, offset) /* * vm_page_remove: [ internal use only ] - * NOTE: used by device pager as well -wfj + * XXX: used by device pager as well * * Removes the given mem entry from the object/offset-page * table and the object page list. @@ -707,8 +732,9 @@ vm_page_rename(mem, new_object, new_offset) if (mem->object == new_object) return; - vm_page_lock_queues(); /* keep page from moving out from - under pageout daemon */ + /* Keep page from moving out from under pageout daemon */ + vm_page_lock_queues(); + vm_page_remove(mem); vm_page_insert(mem, new_object, new_offset); vm_page_unlock_queues(); @@ -759,7 +785,7 @@ vm_page_alloc(object, offset) */ if (cnt.v_free_count < cnt.v_free_min || (cnt.v_free_count < cnt.v_free_target && - cnt.v_inactive_count < cnt.v_inactive_target)) + cnt.v_inactive_count < cnt.v_inactive_target)) thread_wakeup(&vm_pages_needed); return (mem); } @@ -794,6 +820,7 @@ vm_page_free(mem) spl = splimp(); simple_lock(&vm_page_queue_free_lock); + mem->flags |= PG_FREE; TAILQ_INSERT_TAIL(&vm_page_queue_free, mem, pageq); cnt.v_free_count++; @@ -958,3 +985,238 @@ vm_page_copy(src_m, dest_m) dest_m->flags &= ~PG_CLEAN; pmap_copy_page(VM_PAGE_TO_PHYS(src_m), VM_PAGE_TO_PHYS(dest_m)); } + +#ifdef VM_PAGE_ALLOC_MEMORY_STATS +#define STAT_INCR(v) (v)++ +#define STAT_DECR(v) do { \ + if ((v) == 0) \ + printf("%s:%d -- Already 0!\n", __FILE__, __LINE__); \ + else \ + (v)--; \ + } while (0) +u_long vm_page_alloc_memory_npages; +#else +#define STAT_INCR(v) +#define STAT_DECR(v) +#endif + +/* + * vm_page_alloc_memory: + * + * Allocate physical pages conforming to the restrictions + * provided: + * + * size The size of the allocation, + * rounded to page size. + * + * low The low address of the allowed + * allocation range. + * + * high The high address of the allowed + * allocation range. + * + * alignment Allocation must be aligned to this + * power-of-two boundary. + * + * boundary No segment in the allocation may + * cross this power-of-two boundary + * (relative to zero). + * + * The allocated pages are placed at the tail of `rlist'; `rlist' + * is assumed to be properly initialized by the caller. The + * number of memory segments that the allocated memory may + * occupy is specified in the `nsegs' arguement. + * + * Returns 0 on success or an errno value to indicate mode + * of failure. + * + * XXX This implementation could be improved. It only + * XXX allocates a single segment. + */ +int +vm_page_alloc_memory(size, low, high, alignment, boundary, rlist, nsegs, + waitok) + vm_size_t size; + vm_offset_t low, high, alignment, boundary; + struct pglist *rlist; + int nsegs, waitok; +{ + vm_offset_t try, idxpa, lastidxpa; + int s, tryidx, idx, end, error; + vm_page_t m; + u_long pagemask; +#ifdef DEBUG + vm_page_t tp; +#endif + +#ifdef DIAGNOSTIC + if ((alignment & (alignment - 1)) != 0) + panic("vm_page_alloc_memory: alignment must be power of 2"); + + if ((boundary & (boundary - 1)) != 0) + panic("vm_page_alloc_memory: boundary must be power of 2"); +#endif + + /* + * Our allocations are always page granularity, so our alignment + * must be, too. + */ + if (alignment < PAGE_SIZE) + alignment = PAGE_SIZE; + + size = round_page(size); + try = roundup(low, alignment); + + if (boundary != 0 && boundary < size) + return (EINVAL); + + pagemask = ~(boundary - 1); + + /* Default to "lose". */ + error = ENOMEM; + + /* + * Block all memory allocation and lock the free list. + */ + s = splimp(); + simple_lock(&vm_page_queue_free_lock); + + /* Are there even any free pages? */ + if (vm_page_queue_free.tqh_first == NULL) + goto out; + + for (;; try += alignment) { + if (try + size > high) { + /* + * We've run past the allowable range. + */ + goto out; + } + + /* + * Make sure this is a managed physical page. + */ + if (IS_VM_PHYSADDR(try) == 0) + continue; + + tryidx = idx = VM_PAGE_INDEX(try); + end = idx + (size / PAGE_SIZE); + if (end > vm_page_count) { + /* + * No more physical memory. + */ + goto out; + } + + /* + * Found a suitable starting page. See of the range + * is free. + */ + for (; idx < end; idx++) { + if (VM_PAGE_IS_FREE(&vm_page_array[idx]) == 0) { + /* + * Page not available. + */ + break; + } + + idxpa = VM_PAGE_TO_PHYS(&vm_page_array[idx]); + + /* + * Make sure this is a managed physical page. + * XXX Necessary? I guess only if there + * XXX are holes in the vm_page_array[]. + */ + if (IS_VM_PHYSADDR(idxpa) == 0) + break; + + if (idx > tryidx) { + lastidxpa = + VM_PAGE_TO_PHYS(&vm_page_array[idx - 1]); + + if ((lastidxpa + PAGE_SIZE) != idxpa) { + /* + * Region not contiguous. + */ + break; + } + if (boundary != 0 && + ((lastidxpa ^ idxpa) & pagemask) != 0) { + /* + * Region crosses boundary. + */ + break; + } + } + } + + if (idx == end) { + /* + * Woo hoo! Found one. + */ + break; + } + } + + /* + * Okay, we have a chunk of memory that conforms to + * the requested constraints. + */ + idx = tryidx; + while (idx < end) { + m = &vm_page_array[idx]; +#ifdef DEBUG + for (tp = vm_page_queue_free.tqh_first; tp != NULL; + tp = tp->pageq.tqe_next) { + if (tp == m) + break; + } + if (tp == NULL) + panic("vm_page_alloc_memory: page not on freelist"); +#endif + TAILQ_REMOVE(&vm_page_queue_free, m, pageq); + cnt.v_free_count--; + m->flags = PG_CLEAN; + m->object = NULL; + m->wire_count = 0; + TAILQ_INSERT_TAIL(rlist, m, pageq); + idx++; + STAT_INCR(vm_page_alloc_memory_npages); + } + error = 0; + + out: + simple_unlock(&vm_page_queue_free_lock); + splx(s); + return (error); +} + +/* + * vm_page_free_memory: + * + * Free a list of pages previously allocated by vm_page_alloc_memory(). + * The pages are assumed to have no mappings. + */ +void +vm_page_free_memory(list) + struct pglist *list; +{ + vm_page_t m; + int s; + + /* + * Block all memory allocation and lock the free list. + */ + s = splimp(); + simple_lock(&vm_page_queue_free_lock); + + while ((m = list->tqh_first) != NULL) { + TAILQ_REMOVE(list, m, pageq); + m->flags = PG_FREE; + TAILQ_INSERT_TAIL(&vm_page_queue_free, m, pageq); + STAT_DECR(vm_page_alloc_memory_npages); + } + + simple_unlock(&vm_page_queue_free_lock); + splx(s); +} diff --git a/sys/vm/vm_page.h b/sys/vm/vm_page.h index 14a529cf9fe..e4780cdbb8b 100644 --- a/sys/vm/vm_page.h +++ b/sys/vm/vm_page.h @@ -1,5 +1,5 @@ -/* $OpenBSD: vm_page.h,v 1.3 1996/08/13 22:22:17 niklas Exp $ */ -/* $NetBSD: vm_page.h,v 1.18 1995/03/26 20:39:13 jtc Exp $ */ +/* $OpenBSD: vm_page.h,v 1.4 1997/09/22 20:44:53 niklas Exp $ */ +/* $NetBSD: vm_page.h,v 1.20 1997/06/06 23:10:25 thorpej Exp $ */ /* * Copyright (c) 1991, 1993 @@ -140,6 +140,7 @@ struct vm_page { #define PG_FAKE 0x0200 /* page is placeholder for pagein (O) */ #define PG_FILLED 0x0400 /* client flag to set when filled */ #define PG_DIRTY 0x0800 /* client flag to set when dirty */ +#define PG_FREE 0x1000 /* XXX page is on free list */ #define PG_FAULTING 0x2000 /* page is being faulted in */ #define PG_PAGEROWNED 0x4000 /* DEBUG: async paging op in progress */ #define PG_PTPAGE 0x8000 /* DEBUG: is a user page table page */ @@ -220,16 +221,24 @@ int vm_page_count; /* How many pages do we manage? */ #define IS_VM_PHYSADDR(pa) \ ((pa) >= first_phys_addr && (pa) <= last_phys_addr) -#define PHYS_TO_VM_PAGE(pa) \ - (&vm_page_array[atop(pa) - first_page ]) +#define VM_PAGE_INDEX(pa) \ + (atop((pa)) - first_page) #else -#define IS_VM_PHYSADDR(pa) \ - (pmap_page_index(pa) >= 0) +#define IS_VM_PHYSADDR(pa) \ +({ \ + int __pmapidx = pmap_page_index(pa); \ + (__pmapidx >= 0 && __pmapidx >= first_page); \ +}) -#define PHYS_TO_VM_PAGE(pa) \ - (&vm_page_array[pmap_page_index(pa) - first_page]) +#define VM_PAGE_INDEX(pa) \ + (pmap_page_index((pa)) - first_page) #endif /* MACHINE_NONCONTIG */ +#define PHYS_TO_VM_PAGE(pa) \ + (&vm_page_array[VM_PAGE_INDEX((pa))]) + +#define VM_PAGE_IS_FREE(entry) ((entry)->flags & PG_FREE) + extern simple_lock_data_t vm_page_queue_lock; /* lock on active and inactive page queues */ @@ -286,6 +295,10 @@ void pmap_startup __P((vm_offset_t *, vm_offset_t *)); void vm_page_activate __P((vm_page_t)); vm_page_t vm_page_alloc __P((vm_object_t, vm_offset_t)); +int vm_page_alloc_memory __P((vm_size_t, vm_offset_t, + vm_offset_t, vm_offset_t, vm_offset_t, + struct pglist *, int, int)); +void vm_page_free_memory __P((struct pglist *)); #ifdef MACHINE_NONCONTIG void vm_page_bootstrap __P((vm_offset_t *, vm_offset_t *)); #endif |