summaryrefslogtreecommitdiff
path: root/sys/arch
diff options
context:
space:
mode:
authorArtur Grabowski <art@cvs.openbsd.org>2007-05-15 16:38:34 +0000
committerArtur Grabowski <art@cvs.openbsd.org>2007-05-15 16:38:34 +0000
commit87ee5a23019011123c5ba3ade55d7bfcef71e452 (patch)
treed93963e7628b78cab376f04a924c886a37094b8d /sys/arch
parentf8dea286e8cb137402f9ebcc1528d9b8e7e92b45 (diff)
Switch amd64 to VM_PAGE_MD. Mostly just imitating i386. flags in pg_flags.
deraadt@ ok
Diffstat (limited to 'sys/arch')
-rw-r--r--sys/arch/amd64/amd64/pmap.c350
-rw-r--r--sys/arch/amd64/include/pmap.h23
-rw-r--r--sys/arch/amd64/include/vmparam.h17
3 files changed, 123 insertions, 267 deletions
diff --git a/sys/arch/amd64/amd64/pmap.c b/sys/arch/amd64/amd64/pmap.c
index dfc1dcdb089..df7488d00d0 100644
--- a/sys/arch/amd64/amd64/pmap.c
+++ b/sys/arch/amd64/amd64/pmap.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: pmap.c,v 1.20 2007/04/13 18:57:49 art Exp $ */
+/* $OpenBSD: pmap.c,v 1.21 2007/05/15 16:38:33 art Exp $ */
/* $NetBSD: pmap.c,v 1.3 2003/05/08 18:13:13 thorpej Exp $ */
/*
@@ -163,12 +163,6 @@
*
* - struct pmap: describes the address space of one thread
* - struct pv_entry: describes one <PMAP,VA> mapping of a PA
- * - struct pv_head: there is one pv_head per managed page of
- * physical memory. the pv_head points to a list of pv_entry
- * structures which describe all the <PMAP,VA> pairs that this
- * page is mapped in. this is critical for page based operations
- * such as pmap_page_protect() [change protection on _all_ mappings
- * of a page]
* - pmap_remove_record: a list of virtual addresses whose mappings
* have been changed. used for TLB flushing.
*/
@@ -223,12 +217,6 @@
* in the alternate PTE space (since that is determined by the
* entry in the PDP).
*
- * - pvh_lock (per pv_head)
- * this lock protects the pv_entry list which is chained off the
- * pv_head structure for a specific managed PA. it is locked
- * when traversing the list (e.g. adding/removing mappings,
- * syncing R/M bits, etc.)
- *
* - pmaps_lock
* this lock protects the list of active pmaps (headed by "pmaps").
* we lock it when adding or removing pmaps from this list.
@@ -389,7 +377,7 @@ extern int end;
* local prototypes
*/
-void pmap_enter_pv(struct pv_head *, struct pv_entry *, struct pmap *,
+void pmap_enter_pv(struct vm_page *, struct pv_entry *, struct pmap *,
vaddr_t, struct vm_page *);
struct vm_page *pmap_get_ptp(struct pmap *, vaddr_t, pd_entry_t **);
struct vm_page *pmap_find_ptp(struct pmap *, vaddr_t, paddr_t, int);
@@ -398,7 +386,7 @@ void pmap_free_ptp(struct pmap *, struct vm_page *,
void pmap_freepage(struct pmap *, struct vm_page *, int);
static boolean_t pmap_is_active(struct pmap *, int);
void pmap_map_ptes(struct pmap *, pt_entry_t **, pd_entry_t ***);
-struct pv_entry *pmap_remove_pv(struct pv_head *, struct pmap *, vaddr_t);
+struct pv_entry *pmap_remove_pv(struct vm_page *, struct pmap *, vaddr_t);
void pmap_do_remove(struct pmap *, vaddr_t, vaddr_t, int);
boolean_t pmap_remove_pte(struct pmap *, struct vm_page *, pt_entry_t *,
vaddr_t, int32_t *, int);
@@ -413,6 +401,8 @@ boolean_t pmap_pdes_valid(vaddr_t, pd_entry_t **, pd_entry_t *);
void pmap_alloc_level(pd_entry_t **, vaddr_t, int, long *);
void pmap_apte_flush(struct pmap *pmap);
+void pmap_sync_flags_pte(struct vm_page *, u_long);
+
/*
* p m a p i n l i n e h e l p e r f u n c t i o n s
*/
@@ -440,6 +430,21 @@ pmap_is_active(struct pmap *pmap, int cpu_id)
(pmap->pm_cpus & (1U << cpu_id)) != 0);
}
+static __inline u_int
+pmap_pte2flags(u_long pte)
+{
+ return (((pte & PG_U) ? PG_PMAP_REF : 0) |
+ ((pte & PG_M) ? PG_PMAP_MOD : 0));
+}
+
+void
+pmap_sync_flags_pte(struct vm_page *pg, u_long pte)
+{
+ if (pte & (PG_U|PG_M)) {
+ atomic_setbits_int(&pg->pg_flags, pmap_pte2flags(pte));
+ }
+}
+
void
pmap_apte_flush(struct pmap *pmap)
{
@@ -890,68 +895,8 @@ pmap_prealloc_lowmem_ptps(void)
void
pmap_init(void)
{
- int lcv;
- unsigned int npages, i;
struct vm_page *pg;
- vaddr_t addr;
- vsize_t s;
-
- /*
- * compute the number of pages we have and then allocate RAM
- * for each pages' pv_head and saved attributes.
- */
-
- npages = 0;
- for (lcv = 0 ; lcv < vm_nphysseg ; lcv++)
- npages += (vm_physmem[lcv].end - vm_physmem[lcv].start);
- s = (vsize_t) (sizeof(struct pv_head) * npages +
- sizeof(char) * npages);
- s = round_page(s);
- addr = (vaddr_t) uvm_km_zalloc(kernel_map, s);
- if (addr == 0)
- panic("pmap_init: unable to allocate pv_heads");
-
- /*
- * init all pv_head's and attrs in one memset
- */
-
- /* allocate pv_head stuff first */
- for (lcv = 0 ; lcv < vm_nphysseg ; lcv++) {
- vm_physmem[lcv].pmseg.pvhead = (struct pv_head *) addr;
- addr = (vaddr_t)(vm_physmem[lcv].pmseg.pvhead +
- (vm_physmem[lcv].end - vm_physmem[lcv].start));
- for (i = 0;
- i < (vm_physmem[lcv].end - vm_physmem[lcv].start); i++) {
- simple_lock_init(
- &vm_physmem[lcv].pmseg.pvhead[i].pvh_lock);
- }
- }
-
- /* now allocate attrs */
- for (lcv = 0 ; lcv < vm_nphysseg ; lcv++) {
- vm_physmem[lcv].pmseg.attrs = (unsigned char *)addr;
- addr = (vaddr_t)(vm_physmem[lcv].pmseg.attrs +
- (vm_physmem[lcv].end - vm_physmem[lcv].start));
- }
-
-#ifdef LOCKDEBUG
- /*
- * Now, initialize all the pv_head locks.
- * We only do this if LOCKDEBUG because we know that initialized locks
- * are always all-zero if !LOCKDEBUG.
- */
- for (lcv = 0; lcv < vm_nphysseg ; lcv++) {
- int off, npages;
- struct pmap_physseg *pmsegp;
-
- npages = vm_physmem[lcv].end - vm_physmem[lcv].start;
- pmsegp = &vm_physmem[lcv].pmseg;
-
- for (off = 0; off <npages; off++)
- simple_lock_init(&pmsegp->pvhead[off].pvh_lock);
-
- }
-#endif
+ int i;
pg = uvm_pagealloc(NULL, 0, NULL, UVM_PGA_USERESERVE);
if (pg == NULL)
@@ -978,18 +923,14 @@ pmap_init(void)
/*
* main pv_entry manipulation functions:
- * pmap_enter_pv: enter a mapping onto a pv_head list
- * pmap_remove_pv: remove a mapping from a pv_head list
- *
- * NOTE: pmap_enter_pv expects to lock the pvh itself
- * pmap_remove_pv expects the caller to lock the pvh before calling
+ * pmap_enter_pv: enter a mapping onto a pv list
+ * pmap_remove_pv: remove a mapping from a pv list
*/
/*
- * pmap_enter_pv: enter a mapping onto a pv_head lst
+ * pmap_enter_pv: enter a mapping onto a pv list
*
* => caller should have pmap locked
- * => we will gain the lock on the pv_head and allocate the new pv_entry
* => caller should adjust ptp's wire_count before calling
*
* pve: preallocated pve for us to use
@@ -997,41 +938,36 @@ pmap_init(void)
*/
void
-pmap_enter_pv(struct pv_head *pvh, struct pv_entry *pve, struct pmap *pmap,
+pmap_enter_pv(struct vm_page *pg, struct pv_entry *pve, struct pmap *pmap,
vaddr_t va, struct vm_page *ptp)
{
pve->pv_pmap = pmap;
pve->pv_va = va;
pve->pv_ptp = ptp; /* NULL for kernel pmap */
- simple_lock(&pvh->pvh_lock); /* lock pv_head */
- pve->pv_next = pvh->pvh_list; /* add to ... */
- pvh->pvh_list = pve; /* ... locked list */
- simple_unlock(&pvh->pvh_lock); /* unlock, done! */
+ pve->pv_next = pg->mdpage.pv_list; /* add to ... */
+ pg->mdpage.pv_list = pve; /* ... list */
}
/*
* pmap_remove_pv: try to remove a mapping from a pv_list
*
* => pmap should be locked
- * => caller should hold lock on pv_head [so that attrs can be adjusted]
* => caller should adjust ptp's wire_count and free PTP if needed
* => we return the removed pve
*/
struct pv_entry *
-pmap_remove_pv(struct pv_head *pvh, struct pmap *pmap, vaddr_t va)
+pmap_remove_pv(struct vm_page *pg, struct pmap *pmap, vaddr_t va)
{
struct pv_entry *pve, **prevptr;
- prevptr = &pvh->pvh_list; /* previous pv_entry pointer */
- pve = *prevptr;
- while (pve) {
+ prevptr = &pg->mdpage.pv_list;
+ while ((pve = *prevptr) != NULL) {
if (pve->pv_pmap == pmap && pve->pv_va == va) { /* match? */
*prevptr = pve->pv_next; /* remove it! */
break;
}
prevptr = &pve->pv_next; /* previous pointer */
- pve = pve->pv_next; /* advance */
}
return(pve); /* return removed pve */
}
@@ -1690,8 +1626,8 @@ pmap_remove_ptes(struct pmap *pmap, struct vm_page *ptp, vaddr_t ptpva,
{
struct pv_entry *pve;
pt_entry_t *pte = (pt_entry_t *) ptpva;
+ struct vm_page *pg;
pt_entry_t opte;
- int bank, off;
/*
* note that ptpva points to the PTE that maps startva. this may
@@ -1722,34 +1658,31 @@ pmap_remove_ptes(struct pmap *pmap, struct vm_page *ptp, vaddr_t ptpva,
if (ptp)
ptp->wire_count--; /* dropping a PTE */
+ pg = PHYS_TO_VM_PAGE(opte & PG_FRAME);
+
/*
- * if we are not on a pv_head list we are done.
+ * if we are not on a pv list we are done.
*/
if ((opte & PG_PVLIST) == 0) {
#ifdef DIAGNOSTIC
- if (vm_physseg_find(atop(opte & PG_FRAME), &off)
- != -1)
+ if (pg != NULL)
panic("pmap_remove_ptes: managed page without "
"PG_PVLIST for 0x%lx", startva);
#endif
continue;
}
- bank = vm_physseg_find(atop(opte & PG_FRAME), &off);
#ifdef DIAGNOSTIC
- if (bank == -1)
+ if (pg == NULL)
panic("pmap_remove_ptes: unmanaged page marked "
"PG_PVLIST, va = 0x%lx, pa = 0x%lx",
startva, (u_long)(opte & PG_FRAME));
#endif
/* sync R/M bits */
- simple_lock(&vm_physmem[bank].pmseg.pvhead[off].pvh_lock);
- vm_physmem[bank].pmseg.attrs[off] |= (opte & (PG_U|PG_M));
- pve = pmap_remove_pv(&vm_physmem[bank].pmseg.pvhead[off], pmap,
- startva);
- simple_unlock(&vm_physmem[bank].pmseg.pvhead[off].pvh_lock);
+ pmap_sync_flags_pte(pg, opte);
+ pve = pmap_remove_pv(pg, pmap, startva);
if (pve) {
pool_put(&pmap_pv_pool, pve);
@@ -1774,9 +1707,9 @@ boolean_t
pmap_remove_pte(struct pmap *pmap, struct vm_page *ptp, pt_entry_t *pte,
vaddr_t va, int32_t *cpumaskp, int flags)
{
- pt_entry_t opte;
- int bank, off;
struct pv_entry *pve;
+ struct vm_page *pg;
+ pt_entry_t opte;
if (!pmap_valid_entry(*pte))
return(FALSE); /* VA not mapped */
@@ -1796,34 +1729,30 @@ pmap_remove_pte(struct pmap *pmap, struct vm_page *ptp, pt_entry_t *pte,
pmap_tlb_shootdown(pmap, va, opte, cpumaskp);
+ pg = PHYS_TO_VM_PAGE(opte & PG_FRAME);
+
/*
- * if we are not on a pv_head list we are done.
+ * if we are not on a pv list we are done.
*/
-
if ((opte & PG_PVLIST) == 0) {
#ifdef DIAGNOSTIC
- if (vm_physseg_find(atop(opte & PG_FRAME), &off) != -1) {
+ if (pg != NULL)
panic("pmap_remove_pte: managed page without "
"PG_PVLIST for 0x%lx", va);
- }
#endif
return(TRUE);
}
- bank = vm_physseg_find(atop(opte & PG_FRAME), &off);
#ifdef DIAGNOSTIC
- if (bank == -1)
+ if (pg == NULL)
panic("pmap_remove_pte: unmanaged page marked "
"PG_PVLIST, va = 0x%lx, pa = 0x%lx", va,
(u_long)(opte & PG_FRAME));
#endif
/* sync R/M bits */
- simple_lock(&vm_physmem[bank].pmseg.pvhead[off].pvh_lock);
- vm_physmem[bank].pmseg.attrs[off] |= (opte & (PG_U|PG_M));
- pve = pmap_remove_pv(&vm_physmem[bank].pmseg.pvhead[off], pmap, va);
- simple_unlock(&vm_physmem[bank].pmseg.pvhead[off].pvh_lock);
-
+ pmap_sync_flags_pte(pg, opte);
+ pve = pmap_remove_pv(pg, pmap, va);
if (pve)
pool_put(&pmap_pv_pool, pve);
return(TRUE);
@@ -1858,10 +1787,6 @@ pmap_do_remove(struct pmap *pmap, vaddr_t sva, vaddr_t eva, int flags)
struct vm_page *ptp;
int32_t cpumask = 0;
- /*
- * we lock in the pmap => pv_head direction
- */
-
PMAP_MAP_TO_HEAD_LOCK();
pmap_map_ptes(pmap, &ptes, &pdes); /* locks pmap */
@@ -1972,16 +1897,13 @@ pmap_do_remove(struct pmap *pmap, vaddr_t sva, vaddr_t eva, int flags)
/*
* pmap_page_remove: remove a managed vm_page from all pmaps that map it
*
- * => we set pv_head => pmap locking
* => R/M bits are sync'd back to attrs
*/
void
pmap_page_remove(struct vm_page *pg)
{
- int bank, off;
- struct pv_head *pvh;
- struct pv_entry *pve, *npve, **prevptr;
+ struct pv_entry *pve;
pt_entry_t *ptes, opte;
pd_entry_t **pdes;
#ifdef DIAGNOSTIC
@@ -1989,27 +1911,11 @@ pmap_page_remove(struct vm_page *pg)
#endif
int32_t cpumask = 0;
- /* XXX: vm_page should either contain pv_head or have a pointer to it */
- bank = vm_physseg_find(atop(VM_PAGE_TO_PHYS(pg)), &off);
- if (bank == -1) {
- printf("pmap_page_remove: unmanaged page?\n");
- return;
- }
-
- pvh = &vm_physmem[bank].pmseg.pvhead[off];
- if (pvh->pvh_list == NULL) {
- return;
- }
-
- /* set pv_head => pmap locking */
PMAP_HEAD_TO_MAP_LOCK();
- /* XXX: needed if we hold head->map lock? */
- simple_lock(&pvh->pvh_lock);
+ while ((pve = pg->mdpage.pv_list) != NULL) {
+ pg->mdpage.pv_list = pve->pv_next;
- for (prevptr = &pvh->pvh_list, pve = pvh->pvh_list;
- pve != NULL; pve = npve) {
- npve = pve->pv_next;
pmap_map_ptes(pve->pv_pmap, &ptes, &pdes); /* locks pmap */
#ifdef DIAGNOSTIC
@@ -2026,7 +1932,7 @@ pmap_page_remove(struct vm_page *pg)
}
#endif
- /* atomically save the old PTE and zap! it */
+ /* atomically save the old PTE and zap it */
opte = pmap_pte_set(&ptes[pl1_i(pve->pv_va)], 0);
if (opte & PG_W)
@@ -2035,8 +1941,7 @@ pmap_page_remove(struct vm_page *pg)
pmap_tlb_shootdown(pve->pv_pmap, pve->pv_va, opte, &cpumask);
- /* sync R/M bits */
- vm_physmem[bank].pmseg.attrs[off] |= (opte & (PG_U|PG_M));
+ pmap_sync_flags_pte(pg, opte);
/* update the PTP reference count. free if last reference. */
if (pve->pv_ptp) {
@@ -2047,12 +1952,9 @@ pmap_page_remove(struct vm_page *pg)
}
}
pmap_unmap_ptes(pve->pv_pmap); /* unlocks pmap */
- *prevptr = npve; /* remove it */
pool_put(&pmap_pv_pool, pve);
}
- pvh->pvh_list = NULL;
- simple_unlock(&pvh->pvh_lock);
PMAP_HEAD_TO_MAP_UNLOCK();
pmap_tlb_shootnow(cpumask);
}
@@ -2066,101 +1968,65 @@ pmap_page_remove(struct vm_page *pg)
/*
* pmap_test_attrs: test a page's attributes
- *
- * => we set pv_head => pmap locking
*/
boolean_t
-pmap_test_attrs(struct vm_page *pg, unsigned testbits)
+pmap_test_attrs(struct vm_page *pg, unsigned int testbits)
{
- int bank, off;
- unsigned char *myattrs;
- struct pv_head *pvh;
struct pv_entry *pve;
pt_entry_t *ptes, pte;
pd_entry_t **pdes;
+ u_long mybits, testflags;
- /* XXX: vm_page should either contain pv_head or have a pointer to it */
- bank = vm_physseg_find(atop(VM_PAGE_TO_PHYS(pg)), &off);
- if (bank == -1) {
- printf("pmap_test_attrs: unmanaged page?\n");
- return(FALSE);
- }
-
- /*
- * before locking: see if attributes are already set and if so,
- * return!
- */
+ testflags = pmap_pte2flags(testbits);
- myattrs = &vm_physmem[bank].pmseg.attrs[off];
- if (*myattrs & testbits)
- return(TRUE);
-
- /* test to see if there is a list before bothering to lock */
- pvh = &vm_physmem[bank].pmseg.pvhead[off];
- if (pvh->pvh_list == NULL) {
- return(FALSE);
- }
+ if (pg->pg_flags & testflags)
+ return (TRUE);
- /* nope, gonna have to do it the hard way */
PMAP_HEAD_TO_MAP_LOCK();
- /* XXX: needed if we hold head->map lock? */
- simple_lock(&pvh->pvh_lock);
-
- for (pve = pvh->pvh_list; pve != NULL && (*myattrs & testbits) == 0;
- pve = pve->pv_next) {
+ mybits = 0;
+ for (pve = pg->mdpage.pv_list; pve != NULL && mybits == 0;
+ pve = pve->pv_next) {
pmap_map_ptes(pve->pv_pmap, &ptes, &pdes);
pte = ptes[pl1_i(pve->pv_va)];
pmap_unmap_ptes(pve->pv_pmap);
- *myattrs |= pte;
+ mybits |= (pte & testbits);
}
+ PMAP_HEAD_TO_MAP_UNLOCK();
- /*
- * note that we will exit the for loop with a non-null pve if
- * we have found the bits we are testing for.
- */
+ if (mybits == 0)
+ return (FALSE);
- simple_unlock(&pvh->pvh_lock);
- PMAP_HEAD_TO_MAP_UNLOCK();
- return((*myattrs & testbits) != 0);
+ atomic_setbits_int(&pg->pg_flags, pmap_pte2flags(mybits));
+
+ return (TRUE);
}
/*
* pmap_clear_attrs: change a page's attributes
*
- * => we set pv_head => pmap locking
* => we return TRUE if we cleared one of the bits we were asked to
*/
boolean_t
-pmap_clear_attrs(struct vm_page *pg, unsigned clearbits)
+pmap_clear_attrs(struct vm_page *pg, unsigned long clearbits)
{
- int bank, off;
- unsigned result;
- struct pv_head *pvh;
struct pv_entry *pve;
pt_entry_t *ptes, opte;
pd_entry_t **pdes;
- unsigned char *myattrs;
int32_t cpumask = 0;
+ u_long clearflags;
+ int result;
- /* XXX: vm_page should either contain pv_head or have a pointer to it */
- bank = vm_physseg_find(atop(VM_PAGE_TO_PHYS(pg)), &off);
- if (bank == -1) {
- printf("pmap_change_attrs: unmanaged page?\n");
- return(FALSE);
- }
+ clearflags = pmap_pte2flags(clearbits);
PMAP_HEAD_TO_MAP_LOCK();
- pvh = &vm_physmem[bank].pmseg.pvhead[off];
- /* XXX: needed if we hold head->map lock? */
- simple_lock(&pvh->pvh_lock);
- myattrs = &vm_physmem[bank].pmseg.attrs[off];
- result = *myattrs & clearbits;
- *myattrs &= ~clearbits;
+ result = pg->pg_flags & clearflags;
+ if (result)
+ atomic_clearbits_int(&pg->pg_flags, clearflags);
- for (pve = pvh->pvh_list; pve != NULL; pve = pve->pv_next) {
+ for (pve = pg->mdpage.pv_list; pve != NULL; pve = pve->pv_next) {
pmap_map_ptes(pve->pv_pmap, &ptes, &pdes); /* locks pmap */
#ifdef DIAGNOSTIC
if (!pmap_pdes_valid(pve->pv_va, pdes, NULL))
@@ -2170,7 +2036,7 @@ pmap_clear_attrs(struct vm_page *pg, unsigned clearbits)
opte = ptes[pl1_i(pve->pv_va)];
if (opte & clearbits) {
- result |= (opte & clearbits);
+ result = 1;
pmap_pte_clearbits(&ptes[pl1_i(pve->pv_va)],
(opte & clearbits));
pmap_tlb_shootdown(pve->pv_pmap, pve->pv_va, opte,
@@ -2179,12 +2045,11 @@ pmap_clear_attrs(struct vm_page *pg, unsigned clearbits)
pmap_unmap_ptes(pve->pv_pmap); /* unlocks pmap */
}
- simple_unlock(&pvh->pvh_lock);
PMAP_HEAD_TO_MAP_UNLOCK();
pmap_tlb_shootnow(cpumask);
- return(result != 0);
+ return (result != 0);
}
/*
@@ -2355,7 +2220,6 @@ pmap_collect(struct pmap *pmap)
* pmap_enter: enter a mapping into a pmap
*
* => must be done "now" ... no lazy-evaluation
- * => we set pmap => pv_head locking
*/
int
@@ -2363,12 +2227,11 @@ pmap_enter(struct pmap *pmap, vaddr_t va, paddr_t pa, vm_prot_t prot, int flags)
{
pt_entry_t *ptes, opte, npte;
pd_entry_t **pdes;
- struct vm_page *ptp;
- struct pv_head *pvh;
- struct pv_entry *pve;
- int bank, off, error;
+ struct vm_page *ptp, *pg = NULL;
+ struct pv_entry *pve = NULL;
int ptpdelta, wireddelta, resdelta;
boolean_t wired = (flags & PMAP_WIRED) != 0;
+ int error;
#ifdef DIAGNOSTIC
/* sanity check: totally out of range? */
@@ -2436,20 +2299,22 @@ pmap_enter(struct pmap *pmap, vaddr_t va, paddr_t pa, vm_prot_t prot, int flags)
/* if this is on the PVLIST, sync R/M bit */
if (opte & PG_PVLIST) {
- bank = vm_physseg_find(atop(pa), &off);
+ pg = PHYS_TO_VM_PAGE(pa);
#ifdef DIAGNOSTIC
- if (bank == -1)
+ if (pg == NULL)
panic("pmap_enter: same pa PG_PVLIST "
"mapping with unmanaged page "
"pa = 0x%lx (0x%lx)", pa,
atop(pa));
#endif
- pvh = &vm_physmem[bank].pmseg.pvhead[off];
- simple_lock(&pvh->pvh_lock);
- vm_physmem[bank].pmseg.attrs[off] |= opte;
- simple_unlock(&pvh->pvh_lock);
+ pmap_sync_flags_pte(pg, opte);
} else {
- pvh = NULL; /* ensure !PG_PVLIST */
+#ifdef DIAGNOSTIC
+ if (PHYS_TO_VM_PAGE(pa) != NULL)
+ panic("pmap_enter: same pa, managed "
+ "page, no PG_VLIST pa: 0x%lx\n",
+ pa);
+#endif
}
goto enter_now;
}
@@ -2464,20 +2329,16 @@ pmap_enter(struct pmap *pmap, vaddr_t va, paddr_t pa, vm_prot_t prot, int flags)
*/
if (opte & PG_PVLIST) {
- bank = vm_physseg_find(atop(opte & PG_FRAME), &off);
+ pg = PHYS_TO_VM_PAGE(opte & PG_FRAME);
#ifdef DIAGNOSTIC
- if (bank == -1)
+ if (pg == NULL)
panic("pmap_enter: PG_PVLIST mapping with "
"unmanaged page "
"pa = 0x%lx (0x%lx)", pa, atop(pa));
#endif
- pvh = &vm_physmem[bank].pmseg.pvhead[off];
- simple_lock(&pvh->pvh_lock);
- pve = pmap_remove_pv(pvh, pmap, va);
- vm_physmem[bank].pmseg.attrs[off] |= opte;
- simple_unlock(&pvh->pvh_lock);
- } else {
- pve = NULL;
+ pmap_sync_flags_pte(pg, opte);
+ pve = pmap_remove_pv(pg, pmap, va);
+ pg = NULL; /* This is not the page we are looking for */
}
} else { /* opte not valid */
pve = NULL;
@@ -2499,9 +2360,10 @@ pmap_enter(struct pmap *pmap, vaddr_t va, paddr_t pa, vm_prot_t prot, int flags)
* if this entry is to be on a pvlist, enter it now.
*/
- bank = vm_physseg_find(atop(pa), &off);
- if (pmap_initialized && bank != -1) {
- pvh = &vm_physmem[bank].pmseg.pvhead[off];
+ if (pmap_initialized)
+ pg = PHYS_TO_VM_PAGE(pa);
+
+ if (pg != NULL) {
if (pve == NULL) {
pve = pool_get(&pmap_pv_pool, PR_NOWAIT);
if (pve == NULL) {
@@ -2512,27 +2374,28 @@ pmap_enter(struct pmap *pmap, vaddr_t va, paddr_t pa, vm_prot_t prot, int flags)
panic("pmap_enter: no pv entries available");
}
}
- /* lock pvh when adding */
- pmap_enter_pv(pvh, pve, pmap, va, ptp);
+ pmap_enter_pv(pg, pve, pmap, va, ptp);
} else {
-
/* new mapping is not PG_PVLIST. free pve if we've got one */
- pvh = NULL; /* ensure !PG_PVLIST */
if (pve)
pool_put(&pmap_pv_pool, pve);
}
enter_now:
/*
- * at this point pvh is !NULL if we want the PG_PVLIST bit set
+ * at this point pg is !NULL if we want the PG_PVLIST bit set
*/
pmap->pm_stats.resident_count += resdelta;
pmap->pm_stats.wired_count += wireddelta;
if (ptp)
ptp->wire_count += ptpdelta;
+
+ if (pg != PHYS_TO_VM_PAGE(pa))
+ panic("wtf?");
+
npte = pa | protection_codes[prot] | PG_V;
- if (pvh)
+ if (pg != NULL)
npte |= PG_PVLIST;
if (wired)
npte |= PG_W;
@@ -2736,9 +2599,6 @@ pmap_dump(struct pmap *pmap, vaddr_t sva, vaddr_t eva)
if (eva > VM_MAXUSER_ADDRESS || eva <= sva)
eva = VM_MAXUSER_ADDRESS;
- /*
- * we lock in the pmap => pv_head direction
- */
PMAP_MAP_TO_HEAD_LOCK();
pmap_map_ptes(pmap, &ptes, &pdes); /* locks pmap */
diff --git a/sys/arch/amd64/include/pmap.h b/sys/arch/amd64/include/pmap.h
index 7df3bf23f89..6a53ba4384a 100644
--- a/sys/arch/amd64/include/pmap.h
+++ b/sys/arch/amd64/include/pmap.h
@@ -1,4 +1,4 @@
-/* $OpenBSD: pmap.h,v 1.10 2006/03/15 20:01:58 miod Exp $ */
+/* $OpenBSD: pmap.h,v 1.11 2007/05/15 16:38:33 art Exp $ */
/* $NetBSD: pmap.h,v 1.1 2003/04/26 18:39:46 fvdl Exp $ */
/*
@@ -328,20 +328,15 @@ struct pmap {
#define PMF_USER_LDT 0x01 /* pmap has user-set LDT */
/*
- * for each managed physical page we maintain a list of <PMAP,VA>'s
- * which it is mapped at. the list is headed by a pv_head structure.
- * there is one pv_head per managed phys page (allocated at boot time).
- * the pv_head structure points to a list of pv_entry structures (each
- * describes one mapping).
+ * We keep mod/ref flags in struct vm_page->pg_flags.
*/
+#define PG_PMAP_MOD PG_PMAP0
+#define PG_PMAP_REF PG_PMAP1
-struct pv_entry;
-
-struct pv_head {
- struct simplelock pvh_lock; /* locks every pv on this list */
- struct pv_entry *pvh_list; /* head of list (locked by pvh_lock) */
-};
-
+/*
+ * for each managed physical page we maintain a list of <PMAP,VA>'s
+ * which it is mapped at.
+ */
struct pv_entry { /* locked by its list's pvh_lock */
struct pv_entry *pv_next; /* next entry */
struct pmap *pv_pmap; /* the pmap */
@@ -403,7 +398,7 @@ extern pd_entry_t *pdes[];
*/
void pmap_bootstrap(vaddr_t, paddr_t);
-boolean_t pmap_clear_attrs(struct vm_page *, unsigned);
+boolean_t pmap_clear_attrs(struct vm_page *, unsigned long);
static void pmap_page_protect(struct vm_page *, vm_prot_t);
void pmap_page_remove (struct vm_page *);
static void pmap_protect(struct pmap *, vaddr_t,
diff --git a/sys/arch/amd64/include/vmparam.h b/sys/arch/amd64/include/vmparam.h
index 37a98f0a33d..f18e8697277 100644
--- a/sys/arch/amd64/include/vmparam.h
+++ b/sys/arch/amd64/include/vmparam.h
@@ -1,4 +1,4 @@
-/* $OpenBSD: vmparam.h,v 1.6 2007/03/29 03:17:31 weingart Exp $ */
+/* $OpenBSD: vmparam.h,v 1.7 2007/05/15 16:38:33 art Exp $ */
/* $NetBSD: vmparam.h,v 1.1 2003/04/26 18:39:49 fvdl Exp $ */
/*-
@@ -107,13 +107,14 @@
#define VM_FREELIST_DEFAULT 0
#define VM_FREELIST_FIRST16 1
-/*
- * pmap specific data stored in the vm_physmem[] array
- */
-#define __HAVE_PMAP_PHYSSEG
-struct pmap_physseg {
- struct pv_head *pvhead; /* pv_head array */
- unsigned char *attrs; /* attrs array */
+#define __HAVE_VM_PAGE_MD
+struct pv_entry;
+struct vm_page_md {
+ struct pv_entry *pv_list;
};
+#define VM_MDPAGE_INIT(pg) do { \
+ (pg)->mdpage.pv_list = NULL; \
+} while (0)
+
#endif /* _VMPARAM_H_ */