summaryrefslogtreecommitdiff
path: root/sys/arch
diff options
context:
space:
mode:
authorTom Cosgrove <tom@cvs.openbsd.org>2007-02-20 21:15:02 +0000
committerTom Cosgrove <tom@cvs.openbsd.org>2007-02-20 21:15:02 +0000
commitb5de0662c0ae590547e7ddf1190bad604e2d2d32 (patch)
tree24d8349b344e6d332fcbcfcb1db55cf1ca3417ee /sys/arch
parentd084ccaa91178ec96027c48218e22002215bedba (diff)
Revert PAE pmap for now, until the strange bug is found. This stops
the freezes many of us are seeing (especially on amd64 machines running OpenBSD/i386). Much testing by nick@ (as always - thanks!), hugh@, ian@, kettenis@ and Sam Smith (s (at) msmith (dot) net). Requested by, input from, and ok deraadt@ ok art@, kettenis@, miod@
Diffstat (limited to 'sys/arch')
-rw-r--r--sys/arch/i386/conf/files.i3863
-rw-r--r--sys/arch/i386/i386/autoconf.c5
-rw-r--r--sys/arch/i386/i386/cpu.c7
-rw-r--r--sys/arch/i386/i386/db_memrw.c31
-rw-r--r--sys/arch/i386/i386/genassym.cf3
-rw-r--r--sys/arch/i386/i386/kgdb_machdep.c6
-rw-r--r--sys/arch/i386/i386/kvm86.c5
-rw-r--r--sys/arch/i386/i386/lapic.c8
-rw-r--r--sys/arch/i386/i386/locore.s51
-rw-r--r--sys/arch/i386/i386/machdep.c238
-rw-r--r--sys/arch/i386/i386/mpbios.c4
-rw-r--r--sys/arch/i386/i386/mptramp.s20
-rw-r--r--sys/arch/i386/i386/pmap.c706
-rw-r--r--sys/arch/i386/i386/vm_machdep.c16
-rw-r--r--sys/arch/i386/include/_types.h6
-rw-r--r--sys/arch/i386/include/bus.h12
-rw-r--r--sys/arch/i386/include/cpu.h6
-rw-r--r--sys/arch/i386/include/loadfile_machdep.h4
-rw-r--r--sys/arch/i386/include/param.h10
-rw-r--r--sys/arch/i386/include/pmap.h363
-rw-r--r--sys/arch/i386/include/pte.h107
-rw-r--r--sys/arch/i386/include/vmparam.h15
-rw-r--r--sys/arch/i386/pci/pci_addr_fixup.c20
-rw-r--r--sys/arch/i386/pci/pci_machdep.c4
-rw-r--r--sys/arch/i386/pci/pcibios.c4
25 files changed, 670 insertions, 984 deletions
diff --git a/sys/arch/i386/conf/files.i386 b/sys/arch/i386/conf/files.i386
index e0c62190346..35aee08457f 100644
--- a/sys/arch/i386/conf/files.i386
+++ b/sys/arch/i386/conf/files.i386
@@ -1,4 +1,4 @@
-# $OpenBSD: files.i386,v 1.153 2007/02/17 17:38:37 tom Exp $
+# $OpenBSD: files.i386,v 1.154 2007/02/20 21:15:01 tom Exp $
#
# new style config file for i386 architecture
#
@@ -33,7 +33,6 @@ file arch/i386/i386/k6_mem.c mtrr
file arch/i386/i386/microtime.s
file arch/i386/i386/p4tcc.c !small_kernel & i686_cpu
file arch/i386/i386/pmap.c
-file arch/i386/i386/pmapae.c !small_kernel
file arch/i386/i386/powernow.c !small_kernel & i586_cpu
file arch/i386/i386/powernow-k7.c !small_kernel & i686_cpu
file arch/i386/i386/powernow-k8.c !small_kernel & i686_cpu
diff --git a/sys/arch/i386/i386/autoconf.c b/sys/arch/i386/i386/autoconf.c
index 59b314022b5..0721333c06e 100644
--- a/sys/arch/i386/i386/autoconf.c
+++ b/sys/arch/i386/i386/autoconf.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: autoconf.c,v 1.69 2006/11/29 22:40:13 miod Exp $ */
+/* $OpenBSD: autoconf.c,v 1.70 2007/02/20 21:15:01 tom Exp $ */
/* $NetBSD: autoconf.c,v 1.20 1996/05/03 19:41:56 christos Exp $ */
/*-
@@ -124,9 +124,6 @@ cpu_configure(void)
kvm86_init();
#endif
-#ifndef SMALL_KERNEL
- pmap_bootstrap_pae();
-#endif
if (config_rootfound("mainbus", NULL) == NULL)
panic("cpu_configure: mainbus not configured");
diff --git a/sys/arch/i386/i386/cpu.c b/sys/arch/i386/i386/cpu.c
index 9f601ebcd5f..aab0c99c860 100644
--- a/sys/arch/i386/i386/cpu.c
+++ b/sys/arch/i386/i386/cpu.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: cpu.c,v 1.20 2006/06/10 17:50:30 gwk Exp $ */
+/* $OpenBSD: cpu.c,v 1.21 2007/02/20 21:15:01 tom Exp $ */
/* $NetBSD: cpu.c,v 1.1.2.7 2000/06/26 02:04:05 sommerfeld Exp $ */
/*-
@@ -248,7 +248,8 @@ cpu_attach(struct device *parent, struct device *self, void *aux)
pcb->pcb_tss.tss_esp = kstack + USPACE - 16 -
sizeof (struct trapframe);
pcb->pcb_pmap = pmap_kernel();
- pcb->pcb_cr3 = pcb->pcb_pmap->pm_pdirpa;
+ pcb->pcb_cr3 = vtophys((vaddr_t)pcb->pcb_pmap->pm_pdir);
+ /* pcb->pcb_cr3 = pcb->pcb_pmap->pm_pdir - KERNBASE; XXX ??? */
cpu_default_ldt(ci); /* Use the `global' ldt until one alloc'd */
#endif
@@ -409,7 +410,7 @@ cpu_boot_secondary(struct cpu_info *ci)
printf("%s: starting", ci->ci_dev.dv_xname);
/* XXX move elsewhere, not per CPU. */
- mp_pdirpa = kpm->pm_pdirpa;
+ mp_pdirpa = vtophys((vaddr_t)kpm->pm_pdir);
pcb = ci->ci_idle_pcb;
diff --git a/sys/arch/i386/i386/db_memrw.c b/sys/arch/i386/i386/db_memrw.c
index f490f1e02e1..6ff8a7eb835 100644
--- a/sys/arch/i386/i386/db_memrw.c
+++ b/sys/arch/i386/i386/db_memrw.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: db_memrw.c,v 1.11 2006/05/11 13:21:11 mickey Exp $ */
+/* $OpenBSD: db_memrw.c,v 1.12 2007/02/20 21:15:01 tom Exp $ */
/* $NetBSD: db_memrw.c,v 1.6 1999/04/12 20:38:19 pk Exp $ */
/*
@@ -63,19 +63,28 @@ db_read_bytes(vaddr_t addr, size_t size, char *data)
void
db_write_bytes(vaddr_t addr, size_t size, char *data)
{
- extern char etext;
- u_int32_t bits, bits1;
- vaddr_t addr1 = 0;
char *dst;
+ pt_entry_t *ptep0 = 0;
+ pt_entry_t oldmap0 = { 0 };
+ vaddr_t addr1;
+ pt_entry_t *ptep1 = 0;
+ pt_entry_t oldmap1 = { 0 };
+ extern char etext;
+
if (addr >= VM_MIN_KERNEL_ADDRESS &&
addr < (vaddr_t)&etext) {
- bits = pmap_pte_setbits(addr, PG_RW, 0) & PG_RW;
+ ptep0 = kvtopte(addr);
+ oldmap0 = *ptep0;
+ *(int *)ptep0 |= /* INTEL_PTE_WRITE */ PG_RW;
addr1 = trunc_page(addr + size - 1);
- if (trunc_page(addr) != addr1)
+ if (trunc_page(addr) != addr1) {
/* data crosses a page boundary */
- bits1 = pmap_pte_setbits(addr1, PG_RW, 0) & PG_RW;
+ ptep1 = kvtopte(addr1);
+ oldmap1 = *ptep1;
+ *(int *)ptep1 |= /* INTEL_PTE_WRITE */ PG_RW;
+ }
tlbflush();
}
@@ -84,10 +93,10 @@ db_write_bytes(vaddr_t addr, size_t size, char *data)
while (size-- > 0)
*dst++ = *data++;
- if (addr1) {
- pmap_pte_setbits(addr, 0, bits ^ PG_RW);
- if (trunc_page(addr) != addr1)
- pmap_pte_setbits(addr1, 0, bits1 ^ PG_RW);
+ if (ptep0) {
+ *ptep0 = oldmap0;
+ if (ptep1)
+ *ptep1 = oldmap1;
tlbflush();
}
}
diff --git a/sys/arch/i386/i386/genassym.cf b/sys/arch/i386/i386/genassym.cf
index cb328569a85..b47fb186d00 100644
--- a/sys/arch/i386/i386/genassym.cf
+++ b/sys/arch/i386/i386/genassym.cf
@@ -1,4 +1,4 @@
-# $OpenBSD: genassym.cf,v 1.24 2006/04/27 15:37:50 mickey Exp $
+# $OpenBSD: genassym.cf,v 1.25 2007/02/20 21:15:01 tom Exp $
#
# Copyright (c) 1982, 1990 The Regents of the University of California.
# All rights reserved.
@@ -80,6 +80,7 @@ export PDSLOT_KERN
export PDSLOT_PTE
export PDSLOT_APTE
export NKPTP_MIN
+export NKPTP_MAX
# values for virtual memory
export VM_MAXUSER_ADDRESS
diff --git a/sys/arch/i386/i386/kgdb_machdep.c b/sys/arch/i386/i386/kgdb_machdep.c
index 85706ba7866..5d74f8123c7 100644
--- a/sys/arch/i386/i386/kgdb_machdep.c
+++ b/sys/arch/i386/i386/kgdb_machdep.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: kgdb_machdep.c,v 1.7 2006/09/19 11:06:33 jsg Exp $ */
+/* $OpenBSD: kgdb_machdep.c,v 1.8 2007/02/20 21:15:01 tom Exp $ */
/* $NetBSD: kgdb_machdep.c,v 1.6 1998/08/13 21:36:03 thorpej Exp $ */
/*-
@@ -95,13 +95,15 @@ int
kgdb_acc(vaddr_t va, size_t len)
{
vaddr_t last_va;
+ pt_entry_t *pte;
last_va = va + len;
va &= ~PGOFSET;
last_va &= ~PGOFSET;
do {
- if ((pmap_pte_bits(va) & PG_V) == 0)
+ pte = kvtopte(va);
+ if ((*pte & PG_V) == 0)
return (0);
va += NBPG;
} while (va < last_va);
diff --git a/sys/arch/i386/i386/kvm86.c b/sys/arch/i386/i386/kvm86.c
index 2ccea3bcd8b..4a42bc44b99 100644
--- a/sys/arch/i386/i386/kvm86.c
+++ b/sys/arch/i386/i386/kvm86.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: kvm86.c,v 1.2 2006/11/27 16:00:19 gwk Exp $ */
+/* $OpenBSD: kvm86.c,v 1.3 2007/02/20 21:15:01 tom Exp $ */
/* $NetBSD: kvm86.c,v 1.10 2005/12/26 19:23:59 perry Exp $ */
/*
* Copyright (c) 2002
@@ -48,9 +48,6 @@ extern void kvm86_ret(struct trapframe *, int);
#define PGTABLE_SIZE ((1024 + 64) * 1024 / PAGE_SIZE)
-/* XXX: not here */
-typedef u_int32_t pt_entry_t;
-
struct kvm86_data {
pt_entry_t pgtbl[PGTABLE_SIZE];
diff --git a/sys/arch/i386/i386/lapic.c b/sys/arch/i386/i386/lapic.c
index 7ca7fc9ef0a..bd74f8d5f10 100644
--- a/sys/arch/i386/i386/lapic.c
+++ b/sys/arch/i386/i386/lapic.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: lapic.c,v 1.12 2007/02/19 11:59:00 tom Exp $ */
+/* $OpenBSD: lapic.c,v 1.13 2007/02/20 21:15:01 tom Exp $ */
/* $NetBSD: lapic.c,v 1.1.2.8 2000/02/23 06:10:50 sommerfeld Exp $ */
/*-
@@ -78,8 +78,9 @@ void
lapic_map(lapic_base)
paddr_t lapic_base;
{
- vaddr_t va = (vaddr_t)&local_apic;
int s;
+ pt_entry_t *pte;
+ vaddr_t va = (vaddr_t)&local_apic;
disable_intr();
s = lapic_tpr;
@@ -93,7 +94,8 @@ lapic_map(lapic_base)
* might have changed the value of cpu_number()..
*/
- pmap_pte_set(va, lapic_base, PG_RW | PG_V | PG_N);
+ pte = kvtopte(va);
+ *pte = lapic_base | PG_RW | PG_V | PG_N;
invlpg(va);
#ifdef MULTIPROCESSOR
diff --git a/sys/arch/i386/i386/locore.s b/sys/arch/i386/i386/locore.s
index 020ae8e0b45..fd5d7bc1ddc 100644
--- a/sys/arch/i386/i386/locore.s
+++ b/sys/arch/i386/i386/locore.s
@@ -1,4 +1,4 @@
-/* $OpenBSD: locore.s,v 1.105 2006/12/08 21:35:05 dim Exp $ */
+/* $OpenBSD: locore.s,v 1.106 2007/02/20 21:15:01 tom Exp $ */
/* $NetBSD: locore.s,v 1.145 1996/05/03 19:41:19 christos Exp $ */
/*-
@@ -201,7 +201,6 @@
.globl _C_LABEL(cpu_cache_ecx), _C_LABEL(cpu_cache_edx)
.globl _C_LABEL(cold), _C_LABEL(cnvmem), _C_LABEL(extmem)
.globl _C_LABEL(esym)
- .globl _C_LABEL(nkptp_max)
.globl _C_LABEL(boothowto), _C_LABEL(bootdev), _C_LABEL(atdevbase)
.globl _C_LABEL(proc0paddr), _C_LABEL(PTDpaddr), _C_LABEL(PTDsize)
.globl _C_LABEL(gdt)
@@ -535,9 +534,9 @@ try586: /* Use the `cpuid' instruction. */
* 0 1 2 3
*/
#define PROC0PDIR ((0) * NBPG)
-#define PROC0STACK ((4) * NBPG)
-#define SYSMAP ((4+UPAGES) * NBPG)
-#define TABLESIZE ((4+UPAGES) * NBPG) /* + _C_LABEL(nkpde) * NBPG */
+#define PROC0STACK ((1) * NBPG)
+#define SYSMAP ((1+UPAGES) * NBPG)
+#define TABLESIZE ((1+UPAGES) * NBPG) /* + _C_LABEL(nkpde) * NBPG */
/* Find end of kernel image. */
movl $RELOC(_C_LABEL(end)),%edi
@@ -565,9 +564,9 @@ try586: /* Use the `cpuid' instruction. */
jge 1f
movl $NKPTP_MIN,%ecx # set at min
jmp 2f
-1: cmpl RELOC(_C_LABEL(nkptp_max)),%ecx # larger than max?
+1: cmpl $NKPTP_MAX,%ecx # larger than max?
jle 2f
- movl RELOC(_C_LABEL(nkptp_max)),%ecx
+ movl $NKPTP_MAX,%ecx
2: movl %ecx,RELOC(_C_LABEL(nkpde)) # and store it back
/* Clear memory for bootstrap tables. */
@@ -652,8 +651,6 @@ try586: /* Use the `cpuid' instruction. */
/* Install a PDE recursively mapping page directory as a page table! */
leal (PROC0PDIR+PG_V|PG_KW)(%esi),%eax # pte for ptd
movl %eax,(PROC0PDIR+PDSLOT_PTE*4)(%esi) # recursive PD slot
- addl $NBPG, %eax # pte for ptd[1]
- movl %eax,(PROC0PDIR+(PDSLOT_PTE+1)*4)(%esi) # recursive PD slot
/* Save phys. addr of PTD, for libkvm. */
movl %esi,RELOC(_C_LABEL(PTDpaddr))
@@ -2305,42 +2302,6 @@ ENTRY(i686_pagezero)
ret
#endif
-#ifndef SMALL_KERNEL
-/*
- * int cpu_paenable(void *);
- */
-ENTRY(cpu_paenable)
- movl $-1, %eax
- testl $CPUID_PAE, _C_LABEL(cpu_feature)
- jz 1f
-
- pushl %esi
- pushl %edi
- movl 12(%esp), %esi
- movl %cr3, %edi
- orl $0xfe0, %edi /* PDPT will be in the last four slots! */
- movl %edi, %cr3
- addl $KERNBASE, %edi /* and make it back virtual again */
- movl $8, %ecx
- cld
- rep
- movsl
- movl %cr4, %eax
- orl $CR4_PAE, %eax
- movl %eax, %cr4 /* BANG!!! */
- movl 12(%esp), %eax
- subl $KERNBASE, %eax
- movl %eax, %cr3 /* reload real PDPT */
- movl $4*NBPG, %eax
- movl %eax, _C_LABEL(PTDsize)
-
- xorl %eax, %eax
- popl %edi
- popl %esi
-1:
- ret
-#endif /* !SMALL_KERNEL */
-
#if NLAPIC > 0
#include <i386/i386/apicvec.s>
#endif
diff --git a/sys/arch/i386/i386/machdep.c b/sys/arch/i386/i386/machdep.c
index 60f984e5f38..74af9afb577 100644
--- a/sys/arch/i386/i386/machdep.c
+++ b/sys/arch/i386/i386/machdep.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: machdep.c,v 1.377 2007/02/17 23:59:03 marco Exp $ */
+/* $OpenBSD: machdep.c,v 1.378 2007/02/20 21:15:01 tom Exp $ */
/* $NetBSD: machdep.c,v 1.214 1996/11/10 03:16:17 thorpej Exp $ */
/*-
@@ -218,9 +218,6 @@ int bufcachepercent = BUFCACHEPERCENT;
extern int boothowto;
int physmem;
-#ifndef SMALL_KERNEL
-int pae_copy;
-#endif
struct dumpmem {
paddr_t start;
@@ -246,7 +243,7 @@ int i386_has_sse2;
int i386_has_xcrypt;
bootarg_t *bootargp;
-paddr_t avail_end, avail_end2;
+paddr_t avail_end;
struct vm_map *exec_map = NULL;
struct vm_map *phys_map = NULL;
@@ -328,12 +325,6 @@ int allowaperture = 0;
#endif
#endif
-#ifdef I686_PAE
-int cpu_pae = 1;
-#else
-int cpu_pae = 0;
-#endif
-
void winchip_cpu_setup(struct cpu_info *);
void amd_family5_setperf_setup(struct cpu_info *);
void amd_family5_setup(struct cpu_info *);
@@ -2849,6 +2840,7 @@ fix_f00f(void)
struct region_descriptor region;
vaddr_t va;
void *p;
+ pt_entry_t *pte;
/* Allocate two new pages */
va = uvm_km_zalloc(kernel_map, NBPG*2);
@@ -2863,7 +2855,8 @@ fix_f00f(void)
GCODE_SEL);
/* Map first page RO */
- pmap_pte_setbits(va, 0, PG_RW);
+ pte = PTE_BASE + atop(va);
+ *pte &= ~PG_RW;
/* Reload idtr */
setregion(&region, idt, sizeof(idt_region) - 1);
@@ -3018,11 +3011,11 @@ init386(paddr_t first_avail)
if (bootargc > NBPG)
panic("too many boot args");
- if (extent_alloc_region(iomem_ex, (u_long)bootargv, bootargc,
+ if (extent_alloc_region(iomem_ex, (paddr_t)bootargv, bootargc,
EX_NOWAIT))
panic("cannot reserve /boot args memory");
- pmap_enter(pmap_kernel(), (vaddr_t)bootargp, (u_long)bootargv,
+ pmap_enter(pmap_kernel(), (vaddr_t)bootargp, (paddr_t)bootargv,
VM_PROT_READ|VM_PROT_WRITE,
VM_PROT_READ|VM_PROT_WRITE|PMAP_WIRED);
@@ -3035,6 +3028,15 @@ init386(paddr_t first_avail)
if (bios_memmap == NULL)
panic("no BIOS memory map supplied");
#endif
+
+#if defined(MULTIPROCESSOR)
+ /* install the page after boot args as PT page for first 4M */
+ pmap_enter(pmap_kernel(), (u_long)vtopte(0),
+ round_page((vaddr_t)(bootargv + bootargc)),
+ VM_PROT_READ|VM_PROT_WRITE,
+ VM_PROT_READ|VM_PROT_WRITE|PMAP_WIRED);
+ memset(vtopte(0), 0, NBPG); /* make sure it is clean before using */
+#endif
/*
* account all the memory passed in the map from /boot
@@ -3048,12 +3050,27 @@ init386(paddr_t first_avail)
for(i = 0, im = bios_memmap; im->type != BIOS_MAP_END; im++)
if (im->type == BIOS_MAP_FREE) {
paddr_t a, e;
+#ifdef DEBUG
+ printf(" %llx-%llx", im->addr, im->addr + im->size);
+#endif
+
+ if (im->addr >= 0x100000000ULL) {
+#ifdef DEBUG
+ printf("-H");
+#endif
+ continue;
+ }
a = round_page(im->addr);
- e = trunc_page(im->addr + im->size);
+ if (im->addr + im->size <= 0xfffff000ULL)
+ e = trunc_page(im->addr + im->size);
+ else {
#ifdef DEBUG
- printf(" %llx-%llx", a, e);
+ printf("-T");
#endif
+ e = 0xfffff000;
+ }
+
/* skip first eight pages */
if (a < 8 * NBPG)
a = 8 * NBPG;
@@ -3073,16 +3090,7 @@ init386(paddr_t first_avail)
continue;
}
- if (a >= 0x100000000ULL) {
-#ifdef DEBUG
- printf("-H");
-#endif
- if (!cpu_pae)
- continue;
- }
-
- if (e <= 0x100000000ULL &&
- extent_alloc_region(iomem_ex, a, e - a, EX_NOWAIT))
+ if (extent_alloc_region(iomem_ex, a, e - a, EX_NOWAIT))
/* XXX What should we do? */
printf("\nWARNING: CAN'T ALLOCATE RAM (%x-%x)"
" FROM IOMEM EXTENT MAP!\n", a, e);
@@ -3091,15 +3099,11 @@ init386(paddr_t first_avail)
dumpmem[i].start = atop(a);
dumpmem[i].end = atop(e);
i++;
- avail_end2 = MAX(avail_end2, e);
- if (avail_end2 < 0x100000000ULL)
- avail_end = avail_end2;
+ avail_end = max(avail_end, e);
}
ndumpmem = i;
avail_end -= round_page(MSGBUFSIZE);
- if (avail_end2 < 0x100000000ULL)
- avail_end2 = avail_end;
#ifdef DEBUG
printf(": %lx\n", avail_end);
@@ -3130,34 +3134,30 @@ init386(paddr_t first_avail)
e = dumpmem[i].end;
if (a < atop(first_avail) && e > atop(first_avail))
a = atop(first_avail);
- if (a < atop(avail_end) && e > atop(avail_end))
+ if (e > atop(avail_end))
e = atop(avail_end);
if (a < e) {
if (a < atop(16 * 1024 * 1024)) {
lim = MIN(atop(16 * 1024 * 1024), e);
#ifdef DEBUG
- printf(" %llx-%llx (<16M)", a, lim);
+- printf(" %x-%x (<16M)", a, lim);
#endif
uvm_page_physload(a, lim, a, lim,
VM_FREELIST_FIRST16);
if (e > lim) {
#ifdef DEBUG
- printf(" %llx-%llx", lim, e);
+- printf(" %x-%x", lim, e);
#endif
uvm_page_physload(lim, e, lim, e,
VM_FREELIST_DEFAULT);
}
} else {
#ifdef DEBUG
- printf(" %llx-%llx", a, e);
+- printf(" %x-%x", a, e);
#endif
- if (a >= atop(0x100000000ULL))
- uvm_page_physload(a, e, a, a - 1,
- VM_FREELIST_ABOVE4G);
- else
- uvm_page_physload(a, e, a, e,
- VM_FREELIST_DEFAULT);
+ uvm_page_physload(a, e, a, e,
+ VM_FREELIST_DEFAULT);
}
}
}
@@ -3566,8 +3566,8 @@ bus_mem_add_mapping(bus_addr_t bpa, bus_size_t size, int cacheable,
bus_space_handle_t *bshp)
{
u_long pa, endpa;
- u_int32_t bits;
vaddr_t va;
+ pt_entry_t *pte;
bus_size_t map_size;
#ifdef MULTIPROCESSOR
u_int32_t cpumask = 0;
@@ -3599,12 +3599,13 @@ bus_mem_add_mapping(bus_addr_t bpa, bus_size_t size, int cacheable,
* on those machines.
*/
if (cpu_class != CPUCLASS_386) {
+ pte = kvtopte(va);
if (cacheable)
- bits = pmap_pte_setbits(va, 0, PG_N);
+ *pte &= ~PG_N;
else
- bits = pmap_pte_setbits(va, PG_N, 0);
+ *pte |= PG_N;
#ifdef MULTIPROCESSOR
- pmap_tlb_shootdown(pmap_kernel(), va, bits,
+ pmap_tlb_shootdown(pmap_kernel(), va, *pte,
&cpumask);
#else
pmap_update_pg(va);
@@ -3624,7 +3625,7 @@ bus_space_unmap(bus_space_tag_t t, bus_space_handle_t bsh, bus_size_t size)
{
struct extent *ex;
u_long va, endva;
- paddr_t bpa;
+ bus_addr_t bpa;
/*
* Find the correct extent and bus physical address.
@@ -3634,7 +3635,7 @@ bus_space_unmap(bus_space_tag_t t, bus_space_handle_t bsh, bus_size_t size)
bpa = bsh;
} else if (t == I386_BUS_SPACE_MEM) {
ex = iomem_ex;
- bpa = (u_long)ISA_PHYSADDR(bsh);
+ bpa = (bus_addr_t)ISA_PHYSADDR(bsh);
if (IOM_BEGIN <= bpa && bpa <= IOM_END)
goto ok;
@@ -3670,7 +3671,7 @@ _bus_space_unmap(bus_space_tag_t t, bus_space_handle_t bsh, bus_size_t size,
bus_addr_t *adrp)
{
u_long va, endva;
- paddr_t bpa;
+ bus_addr_t bpa;
/*
* Find the correct bus physical address.
@@ -3678,7 +3679,7 @@ _bus_space_unmap(bus_space_tag_t t, bus_space_handle_t bsh, bus_size_t size,
if (t == I386_BUS_SPACE_IO) {
bpa = bsh;
} else if (t == I386_BUS_SPACE_MEM) {
- bpa = (u_long)ISA_PHYSADDR(bsh);
+ bpa = (bus_addr_t)ISA_PHYSADDR(bsh);
if (IOM_BEGIN <= bpa && bpa <= IOM_END)
goto ok;
@@ -3732,7 +3733,6 @@ _bus_dmamap_create(bus_dma_tag_t t, bus_size_t size, int nsegments,
struct i386_bus_dmamap *map;
void *mapstore;
size_t mapsize;
- int npages;
/*
* Allocate and initialize the DMA map. The end of the map
@@ -3748,17 +3748,6 @@ _bus_dmamap_create(bus_dma_tag_t t, bus_size_t size, int nsegments,
*/
mapsize = sizeof(struct i386_bus_dmamap) +
(sizeof(bus_dma_segment_t) * (nsegments - 1));
- npages = 0;
-#ifndef SMALL_KERNEL
- if (avail_end2 > avail_end &&
- (flags & (BUS_DMA_64BIT|BUS_DMA_24BIT)) == 0) {
- /* this many pages plus one in case we get split */
- npages = round_page(size) / PAGE_SIZE + 1;
- if (npages < nsegments) /* looks stupid, but possible */
- npages = nsegments;
- mapsize += sizeof(struct vm_page *) * npages;
- }
-#endif /* !SMALL_KERNEL */
if ((mapstore = malloc(mapsize, M_DEVBUF,
(flags & BUS_DMA_NOWAIT) ? M_NOWAIT : M_WAITOK)) == NULL)
return (ENOMEM);
@@ -3769,55 +3758,10 @@ _bus_dmamap_create(bus_dma_tag_t t, bus_size_t size, int nsegments,
map->_dm_segcnt = nsegments;
map->_dm_maxsegsz = maxsegsz;
map->_dm_boundary = boundary;
- map->_dm_pages = npages? (void *)&map->dm_segs[nsegments] : NULL;
- map->_dm_npages = npages;
map->_dm_flags = flags & ~(BUS_DMA_WAITOK|BUS_DMA_NOWAIT);
map->dm_mapsize = 0; /* no valid mappings */
map->dm_nsegs = 0;
-#ifndef SMALL_KERNEL
- if (npages) {
- struct pglist mlist;
- vaddr_t va;
- int error;
-
- size = npages << PGSHIFT;
- va = uvm_km_valloc(kernel_map, size);
- if (va == 0) {
- map->_dm_npages = 0;
- free(map, M_DEVBUF);
- return (ENOMEM);
- }
-
- TAILQ_INIT(&mlist);
- /* if not a 64bit map -- allocate some bouncy-bouncy */
- error = uvm_pglistalloc(size,
- round_page(ISA_DMA_BOUNCE_THRESHOLD), 0xfffff000,
- PAGE_SIZE, boundary, &mlist, nsegments,
- (flags & BUS_DMA_NOWAIT) == 0);
- if (error) {
- map->_dm_npages = 0;
- uvm_km_free(kernel_map, (vaddr_t)va, size);
- free(map, M_DEVBUF);
- return (ENOMEM);
- } else {
- struct vm_page **pg = map->_dm_pages;
-
- npages--;
- *pg = TAILQ_FIRST(&mlist);
- pmap_kenter_pa(va, VM_PAGE_TO_PHYS(*pg),
- VM_PROT_READ | VM_PROT_WRITE | PMAP_WIRED);
- for (pg++, va += PAGE_SIZE; npages--;
- pg++, va += PAGE_SIZE) {
- *pg = TAILQ_NEXT(pg[-1], pageq);
- pmap_kenter_pa(va, VM_PAGE_TO_PHYS(*pg),
- VM_PROT_READ | VM_PROT_WRITE | PMAP_WIRED);
- }
- }
- map->_dm_pgva = va;
- }
-#endif /* !SMALL_KERNEL */
-
*dmamp = map;
return (0);
}
@@ -3841,7 +3785,7 @@ int
_bus_dmamap_load(bus_dma_tag_t t, bus_dmamap_t map, void *buf,
bus_size_t buflen, struct proc *p, int flags)
{
- paddr_t lastaddr;
+ bus_addr_t lastaddr;
int seg, error;
/*
@@ -4009,7 +3953,6 @@ _bus_dmamap_unload(bus_dma_tag_t t, bus_dmamap_t map)
*/
map->dm_mapsize = 0;
map->dm_nsegs = 0;
- map->_dm_nused = 0;
}
/*
@@ -4017,43 +3960,10 @@ _bus_dmamap_unload(bus_dma_tag_t t, bus_dmamap_t map)
* by bus-specific DMA map synchronization functions.
*/
void
-_bus_dmamap_sync(bus_dma_tag_t t, bus_dmamap_t map, bus_addr_t offset,
+_bus_dmamap_sync(bus_dma_tag_t t, bus_dmamap_t map, bus_addr_t addr,
bus_size_t size, int op)
{
-#ifndef SMALL_KERNEL
- bus_dma_segment_t *sg;
- int i, off = offset;
- bus_size_t l;
-
- /* scan the segment list performing necessary copies */
- if (!(map->_dm_flags & BUS_DMA_64BIT) && map->_dm_nused) {
- for (i = map->_dm_segcnt, sg = map->dm_segs;
- size && i--; sg++) {
- if (off >= sg->ds_len) {
- off -= sg->ds_len;
- continue;
- }
-
- l = sg->ds_len - off;
- if (l > size)
- l = size;
- size -= l;
- if (sg->ds_addr2) {
- if (op & BUS_DMASYNC_POSTREAD) {
- bcopy((void *)(sg->ds_va2 + off),
- (void *)(sg->ds_va + off), l);
- pae_copy++;
- }
- if (op & BUS_DMASYNC_PREWRITE) {
- bcopy((void *)(sg->ds_va + off),
- (void *)(sg->ds_va2 + off), l);
- pae_copy++;
- }
- }
- off = 0;
- }
- }
-#endif /* !SMALL_KERNEL */
+ /* Nothing to do here. */
}
/*
@@ -4197,8 +4107,8 @@ _bus_dmamap_load_buffer(bus_dma_tag_t t, bus_dmamap_t map, void *buf,
int first)
{
bus_size_t sgsize;
- paddr_t curaddr, lastaddr, oaddr, baddr, bmask;
- vaddr_t pgva, vaddr = (vaddr_t)buf;
+ bus_addr_t curaddr, lastaddr, baddr, bmask;
+ vaddr_t vaddr = (vaddr_t)buf;
int seg;
pmap_t pmap;
@@ -4214,24 +4124,7 @@ _bus_dmamap_load_buffer(bus_dma_tag_t t, bus_dmamap_t map, void *buf,
/*
* Get the physical address for this segment.
*/
- pmap_extract(pmap, vaddr, &curaddr);
- oaddr = 0;
- pgva = 0;
-#ifndef SMALL_KERNEL
- if (!(map->_dm_flags & BUS_DMA_64BIT) &&
- curaddr >= 0x100000000ULL) {
- struct vm_page *pg;
- int page, off;
-
- if (map->_dm_nused + 1 >= map->_dm_npages)
- return (ENOMEM);
- off = vaddr & PAGE_MASK;
- pg = map->_dm_pages[page = map->_dm_nused++];
- oaddr = curaddr;
- curaddr = VM_PAGE_TO_PHYS(pg) + off;
- pgva = map->_dm_pgva + (page << PGSHIFT) + off;
- }
-#endif /* !SMALL_KERNEL */
+ pmap_extract(pmap, vaddr, (paddr_t *)&curaddr);
/*
* Compute the segment size, and adjust counts.
@@ -4255,10 +4148,7 @@ _bus_dmamap_load_buffer(bus_dma_tag_t t, bus_dmamap_t map, void *buf,
*/
if (first) {
map->dm_segs[seg].ds_addr = curaddr;
- map->dm_segs[seg].ds_addr2 = oaddr;
map->dm_segs[seg].ds_len = sgsize;
- map->dm_segs[seg].ds_va = vaddr;
- map->dm_segs[seg].ds_va2 = pgva;
first = 0;
} else {
if (curaddr == lastaddr &&
@@ -4272,10 +4162,7 @@ _bus_dmamap_load_buffer(bus_dma_tag_t t, bus_dmamap_t map, void *buf,
if (++seg >= map->_dm_segcnt)
break;
map->dm_segs[seg].ds_addr = curaddr;
- map->dm_segs[seg].ds_addr2 = oaddr;
map->dm_segs[seg].ds_len = sgsize;
- map->dm_segs[seg].ds_va = vaddr;
- map->dm_segs[seg].ds_va2 = pgva;
}
}
@@ -4311,19 +4198,6 @@ _bus_dmamem_alloc_range(bus_dma_tag_t t, bus_size_t size, bus_size_t alignment,
/* Always round the size. */
size = round_page(size);
- if (flags & BUS_DMA_64BIT) {
- if (high > 0x100000000ULL && low < 0x100000000ULL)
- low = 0x100000000ULL;
- } else if (high > 0x100000000ULL) {
- if (low >= 0x100000000ULL) {
-#ifdef DIAGNOSTIC
- printf("_bus_dmamem_alloc_range: "
- "32bit request in above 4GB space\n");
-#endif
- return (EINVAL);
- } else
- high = 0x100000000ULL;
- }
TAILQ_INIT(&mlist);
/*
diff --git a/sys/arch/i386/i386/mpbios.c b/sys/arch/i386/i386/mpbios.c
index 415e5c0145c..5157a990767 100644
--- a/sys/arch/i386/i386/mpbios.c
+++ b/sys/arch/i386/i386/mpbios.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: mpbios.c,v 1.20 2007/01/23 21:17:18 kettenis Exp $ */
+/* $OpenBSD: mpbios.c,v 1.21 2007/02/20 21:15:01 tom Exp $ */
/* $NetBSD: mpbios.c,v 1.2 2002/10/01 12:56:57 fvdl Exp $ */
/*-
@@ -1024,7 +1024,7 @@ mpbios_ioapic(const u_int8_t *ent, struct device *self)
aaa.aaa_name = "ioapic";
aaa.apic_id = entry->apic_id;
aaa.apic_version = entry->apic_version;
- aaa.apic_address = (u_long)entry->apic_address;
+ aaa.apic_address = (paddr_t)entry->apic_address;
aaa.apic_vecbase = -1;
aaa.flags = (mp_fps->mpfb2 & 0x80) ? IOAPIC_PICMODE : IOAPIC_VWIRE;
diff --git a/sys/arch/i386/i386/mptramp.s b/sys/arch/i386/i386/mptramp.s
index 7ec527bd359..b68ec50ed12 100644
--- a/sys/arch/i386/i386/mptramp.s
+++ b/sys/arch/i386/i386/mptramp.s
@@ -1,4 +1,4 @@
-/* $OpenBSD: mptramp.s,v 1.7 2006/10/16 15:51:26 tom Exp $ */
+/* $OpenBSD: mptramp.s,v 1.8 2007/02/20 21:15:01 tom Exp $ */
/*-
* Copyright (c) 2000 The NetBSD Foundation, Inc.
@@ -165,20 +165,10 @@ _TRMP_LABEL(mp_startup)
/* Load base of page directory and enable mapping. */
movl %ecx,%cr3 # load ptd addr into mmu
-#ifndef SMALL_KERNEL
- movl $_C_LABEL(pmap_pte_set_pae),%eax
- cmpl RELOC(_C_LABEL(pmap_pte_set_p)),%eax
- jne nopae
-
- movl %cr4,%eax
- orl $CR4_PAE,%eax
- movl %eax, %cr4
-nopae:
-#endif
- movl %cr0,%eax # get control word
- # enable paging & NPX emulation
- orl $(CR0_PE|CR0_PG|CR0_NE|CR0_TS|CR0_EM|CR0_MP|CR0_WP),%eax
- movl %eax,%cr0 # and let's page NOW!
+ movl %cr0,%eax # get control word
+ # enable paging & NPX emulation
+ orl $(CR0_PE|CR0_PG|CR0_NE|CR0_TS|CR0_EM|CR0_MP|CR0_WP),%eax
+ movl %eax,%cr0 # and let's page NOW!
#ifdef MPDEBUG
leal _C_LABEL(cpu_trace),%edi
diff --git a/sys/arch/i386/i386/pmap.c b/sys/arch/i386/i386/pmap.c
index 0871e629546..6b30caaa9e3 100644
--- a/sys/arch/i386/i386/pmap.c
+++ b/sys/arch/i386/i386/pmap.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: pmap.c,v 1.96 2007/02/03 16:48:23 miod Exp $ */
+/* $OpenBSD: pmap.c,v 1.97 2007/02/20 21:15:01 tom Exp $ */
/* $NetBSD: pmap.c,v 1.91 2000/06/02 17:46:37 thorpej Exp $ */
/*
@@ -81,6 +81,15 @@
#include <stand/boot/bootarg.h>
/*
+ * general info:
+ *
+ * - for an explanation of how the i386 MMU hardware works see
+ * the comments in <machine/pte.h>.
+ *
+ * - for an explanation of the general memory structure used by
+ * this pmap (including the recursive mapping), see the comments
+ * in <machine/pmap.h>.
+ *
* this file contains the code for the "pmap module." the module's
* job is to manage the hardware's virtual to physical address mappings.
* note that there are two levels of mapping in the VM system:
@@ -118,181 +127,6 @@
* its pv_entrys.
*/
/*
- * i386 MMU hardware structure:
- *
- * the i386 MMU is a two-level MMU which maps 4GB of virtual memory.
- * the pagesize is 4K (4096 [0x1000] bytes), although newer pentium
- * processors can support a 4MB pagesize as well.
- *
- * the first level table (segment table?) is called a "page directory"
- * and it contains 1024 page directory entries (PDEs). each PDE is
- * 4 bytes (an int), so a PD fits in a single 4K page. this page is
- * the page directory page (PDP). each PDE in a PDP maps 4MB of space
- * (1024 * 4MB = 4GB). a PDE contains the physical address of the
- * second level table: the page table. or, if 4MB pages are being used,
- * then the PDE contains the PA of the 4MB page being mapped.
- *
- * a page table consists of 1024 page table entries (PTEs). each PTE is
- * 4 bytes (an int), so a page table also fits in a single 4K page. a
- * 4K page being used as a page table is called a page table page (PTP).
- * each PTE in a PTP maps one 4K page (1024 * 4K = 4MB). a PTE contains
- * the physical address of the page it maps and some flag bits (described
- * below).
- *
- * the processor has a special register, "cr3", which points to the
- * the PDP which is currently controlling the mappings of the virtual
- * address space.
- *
- * the following picture shows the translation process for a 4K page:
- *
- * %cr3 register [PA of PDP]
- * |
- * |
- * | bits <31-22> of VA bits <21-12> of VA bits <11-0>
- * | index the PDP (0 - 1023) index the PTP are the page offset
- * | | | |
- * | v | |
- * +--->+----------+ | |
- * | PD Page | PA of v |
- * | |---PTP-------->+------------+ |
- * | 1024 PDE | | page table |--PTE--+ |
- * | entries | | (aka PTP) | | |
- * +----------+ | 1024 PTE | | |
- * | entries | | |
- * +------------+ | |
- * | |
- * bits <31-12> bits <11-0>
- * p h y s i c a l a d d r
- *
- * the i386 caches PTEs in a TLB. it is important to flush out old
- * TLB mappings when making a change to a mappings. writing to the
- * %cr3 will flush the entire TLB. newer processors also have an
- * instruction that will invalidate the mapping of a single page (which
- * is useful if you are changing a single mappings because it preserves
- * all the cached TLB entries).
- *
- * as shows, bits 31-12 of the PTE contain PA of the page being mapped.
- * the rest of the PTE is defined as follows:
- * bit# name use
- * 11 n/a available for OS use, hardware ignores it
- * 10 n/a available for OS use, hardware ignores it
- * 9 n/a available for OS use, hardware ignores it
- * 8 G global bit (see discussion below)
- * 7 PS page size [for PDEs] (0=4k, 1=4M <if supported>)
- * 6 D dirty (modified) page
- * 5 A accessed (referenced) page
- * 4 PCD cache disable
- * 3 PWT prevent write through (cache)
- * 2 U/S user/supervisor bit (0=supervisor only, 1=both u&s)
- * 1 R/W read/write bit (0=read only, 1=read-write)
- * 0 P present (valid)
- *
- * notes:
- * - on the i386 the R/W bit is ignored if processor is in supervisor
- * state (bug!)
- * - PS is only supported on newer processors
- * - PTEs with the G bit are global in the sense that they are not
- * flushed from the TLB when %cr3 is written (to flush, use the
- * "flush single page" instruction). this is only supported on
- * newer processors. this bit can be used to keep the kernel's
- * TLB entries around while context switching. since the kernel
- * is mapped into all processes at the same place it does not make
- * sense to flush these entries when switching from one process'
- * pmap to another.
- */
-/*
- * A pmap describes a process' 4GB virtual address space. This
- * virtual address space can be broken up into 1024 4MB regions which
- * are described by PDEs in the PDP. The PDEs are defined as follows:
- *
- * Ranges are inclusive -> exclusive, just like vm_map_entry start/end.
- * The following assumes that KERNBASE is 0xd0000000.
- *
- * PDE#s VA range Usage
- * 0->831 0x0 -> 0xcfc00000 user address space, note that the
- * max user address is 0xcfbfe000
- * the final two pages in the last 4MB
- * used to be reserved for the UAREA
- * but now are no longer used.
- * 831 0xcfc00000-> recursive mapping of PDP (used for
- * 0xd0000000 linear mapping of PTPs).
- * 832->1023 0xd0000000-> kernel address space (constant
- * 0xffc00000 across all pmaps/processes).
- * 1023 0xffc00000-> "alternate" recursive PDP mapping
- * <end> (for other pmaps).
- *
- *
- * Note: A recursive PDP mapping provides a way to map all the PTEs for
- * a 4GB address space into a linear chunk of virtual memory. In other
- * words, the PTE for page 0 is the first int mapped into the 4MB recursive
- * area. The PTE for page 1 is the second int. The very last int in the
- * 4MB range is the PTE that maps VA 0xffffe000 (the last page in a 4GB
- * address).
- *
- * All pmaps' PDs must have the same values in slots 832->1023 so that
- * the kernel is always mapped in every process. These values are loaded
- * into the PD at pmap creation time.
- *
- * At any one time only one pmap can be active on a processor. This is
- * the pmap whose PDP is pointed to by processor register %cr3. This pmap
- * will have all its PTEs mapped into memory at the recursive mapping
- * point (slot #831 as show above). When the pmap code wants to find the
- * PTE for a virtual address, all it has to do is the following:
- *
- * Address of PTE = (831 * 4MB) + (VA / NBPG) * sizeof(pt_entry_t)
- * = 0xcfc00000 + (VA / 4096) * 4
- *
- * What happens if the pmap layer is asked to perform an operation
- * on a pmap that is not the one which is currently active? In that
- * case we take the PA of the PDP of non-active pmap and put it in
- * slot 1023 of the active pmap. This causes the non-active pmap's
- * PTEs to get mapped in the final 4MB of the 4GB address space
- * (e.g. starting at 0xffc00000).
- *
- * The following figure shows the effects of the recursive PDP mapping:
- *
- * PDP (%cr3)
- * +----+
- * | 0| -> PTP#0 that maps VA 0x0 -> 0x400000
- * | |
- * | |
- * | 831| -> points back to PDP (%cr3) mapping VA 0xcfc00000 -> 0xd0000000
- * | 832| -> first kernel PTP (maps 0xd0000000 -> 0xe0400000)
- * | |
- * |1023| -> points to alternate pmap's PDP (maps 0xffc00000 -> end)
- * +----+
- *
- * Note that the PDE#831 VA (0xcfc00000) is defined as "PTE_BASE".
- * Note that the PDE#1023 VA (0xffc00000) is defined as "APTE_BASE".
- *
- * Starting at VA 0xcfc00000 the current active PDP (%cr3) acts as a
- * PTP:
- *
- * PTP#831 == PDP(%cr3) => maps VA 0xcfc00000 -> 0xd0000000
- * +----+
- * | 0| -> maps the contents of PTP#0 at VA 0xcfc00000->0xcfc01000
- * | |
- * | |
- * | 831| -> maps the contents of PTP#831 (the PDP) at VA 0xcff3f000
- * | 832| -> maps the contents of first kernel PTP
- * | |
- * |1023|
- * +----+
- *
- * Note that mapping of the PDP at PTP#831's VA (0xcff3f000) is
- * defined as "PDP_BASE".... within that mapping there are two
- * defines:
- * "PDP_PDE" (0xcff3fcfc) is the VA of the PDE in the PDP
- * which points back to itself.
- * "APDP_PDE" (0xcff3fffc) is the VA of the PDE in the PDP which
- * establishes the recursive mapping of the alternate pmap.
- * To set the alternate PDP, one just has to put the correct
- * PA info in *APDP_PDE.
- *
- * Note that in the APTE_BASE space, the APDP appears at VA
- * "APDP_BASE" (0xfffff000).
- */
-/*
* memory allocation
*
* - there are three data structures that we must dynamically allocate:
@@ -384,79 +218,6 @@ struct simplelock pmaps_lock;
#define PMAP_HEAD_TO_MAP_LOCK() /* null */
#define PMAP_HEAD_TO_MAP_UNLOCK() /* null */
-#define PG_FRAME 0xfffff000 /* page frame mask */
-#define PG_LGFRAME 0xffc00000 /* large (4M) page frame mask */
-
-/*
- * The following defines give the virtual addresses of various MMU
- * data structures:
- * PTE_BASE and APTE_BASE: the base VA of the linear PTE mappings
- * PTD_BASE and APTD_BASE: the base VA of the recursive mapping of the PTD
- * PDP_PDE and APDP_PDE: the VA of the PDE that points back to the PDP/APDP
- */
-#define PTE_BASE ((pt_entry_t *) (PDSLOT_PTE * NBPD) )
-#define APTE_BASE ((pt_entry_t *) (PDSLOT_APTE * NBPD) )
-#define PDP_BASE ((pd_entry_t *)(((char *)PTE_BASE) + (PDSLOT_PTE * NBPG)))
-#define APDP_BASE ((pd_entry_t *)(((char *)APTE_BASE) + (PDSLOT_APTE * NBPG)))
-#define PDP_PDE (PDP_BASE + PDSLOT_PTE)
-#define APDP_PDE (PDP_BASE + PDSLOT_APTE)
-
-#define PDOFSET (NBPD-1) /* mask for non-PD part of VA */
-#define PTES_PER_PTP (NBPD / NBPG) /* # of PTEs in a PTP */
-
-/*
- * various address macros
- *
- * vtopte: return a pointer to the PTE mapping a VA
- *
- */
-#define vtopte(VA) (PTE_BASE + atop((vaddr_t)VA))
-
-/*
- * Mach derived conversion macros
- */
-#define i386_round_pdr(x) ((((unsigned)(x)) + PDOFSET) & ~PDOFSET)
-
-/*
- * pdei/ptei: generate index into PDP/PTP from a VA
- */
-#define PD_MASK 0xffc00000 /* page directory address bits */
-#define PT_MASK 0x003ff000 /* page table address bits */
-#define pdei(VA) (((VA) & PD_MASK) >> PDSHIFT)
-#define ptei(VA) (((VA) & PT_MASK) >> PGSHIFT)
-
-/*
- * PTP macros:
- * A PTP's index is the PD index of the PDE that points to it.
- * A PTP's offset is the byte-offset in the PTE space that this PTP is at.
- * A PTP's VA is the first VA mapped by that PTP.
- *
- * Note that NBPG == number of bytes in a PTP (4096 bytes == 1024 entries)
- * NBPD == number of bytes a PTP can map (4MB)
- */
-
-#define ptp_i2o(I) ((I) * NBPG) /* index => offset */
-#define ptp_o2i(O) ((O) / NBPG) /* offset => index */
-#define ptp_i2v(I) ((I) * NBPD) /* index => VA */
-#define ptp_v2i(V) ((V) / NBPD) /* VA => index (same as pdei) */
-
-/*
- * Access PD and PT
- */
-#define PDE(pm,i) (((pd_entry_t *)(pm)->pm_pdir)[(i)])
-
-/*
- * here we define the data types for PDEs and PTEs
- */
-typedef u_int32_t pd_entry_t; /* PDE */
-typedef u_int32_t pt_entry_t; /* PTE */
-
-/*
- * Number of PTE's per cache line. 4 byte pte, 32-byte cache line
- * Used to avoid false sharing of cache lines.
- */
-#define NPTECL 8
-
/*
* TLB Shootdown:
*
@@ -476,13 +237,13 @@ struct pmap_tlb_shootdown_job {
TAILQ_ENTRY(pmap_tlb_shootdown_job) pj_list;
vaddr_t pj_va; /* virtual address */
pmap_t pj_pmap; /* the pmap which maps the address */
- u_int32_t pj_pte; /* the PTE bits */
+ pt_entry_t pj_pte; /* the PTE bits */
struct pmap_tlb_shootdown_job *pj_nextfree;
};
struct pmap_tlb_shootdown_q {
TAILQ_HEAD(, pmap_tlb_shootdown_job) pq_head;
- u_int32_t pq_pte; /* aggregate PTE bits */
+ int pq_pte; /* aggregate PTE bits */
int pq_count; /* number of pending requests */
struct mutex pq_mutex; /* mutex on queue */
int pq_flushg; /* pending flush global */
@@ -504,8 +265,7 @@ struct pmap_tlb_shootdown_job *pj_page, *pj_free;
* global data structures
*/
-struct pmap kernel_pmap_store /* the kernel's pmap (proc0) */
- __attribute__((aligned(32)));
+struct pmap kernel_pmap_store; /* the kernel's pmap (proc0) */
/*
* nkpde is the number of kernel PTPs allocated for the kernel at
@@ -515,7 +275,6 @@ struct pmap kernel_pmap_store /* the kernel's pmap (proc0) */
*/
int nkpde = NKPTP;
-int nkptp_max = 1024 - (KERNBASE/NBPD) - 1; /* largest value (-1 for APTP space) */
#ifdef NKPDE
#error "obsolete NKPDE: use NKPTP"
#endif
@@ -541,8 +300,8 @@ paddr_t hole_end; /* PA of end of "hole" */
* other data structures
*/
-u_int32_t protection_codes[8]; /* maps MI prot to i386 prot code */
-boolean_t pmap_initialized = FALSE; /* pmap_init done yet? */
+static pt_entry_t protection_codes[8]; /* maps MI prot to i386 prot code */
+static boolean_t pmap_initialized = FALSE; /* pmap_init done yet? */
/*
* the following two vaddr_t's are used during system startup
@@ -551,10 +310,8 @@ boolean_t pmap_initialized = FALSE; /* pmap_init done yet? */
* VM space is turned over to the kernel_map vm_map.
*/
-vaddr_t virtual_avail; /* VA of first free KVA */
-vaddr_t virtual_end; /* VA of last free KVA */
-
-vaddr_t vm_max_address = (PDSLOT_PTE << PDSHIFT) + (PDSLOT_PTE << PGSHIFT);
+static vaddr_t virtual_avail; /* VA of first free KVA */
+static vaddr_t virtual_end; /* VA of last free KVA */
/*
* pv_page management structures: locked by pvalloc_lock
@@ -575,8 +332,8 @@ static vaddr_t pv_cachedva; /* cached VA for later use */
* linked list of all non-kernel pmaps
*/
-struct pmap_head pmaps;
-struct pmap *pmaps_hand = NULL; /* used by pmap_steal_ptp */
+static struct pmap_head pmaps;
+static struct pmap *pmaps_hand = NULL; /* used by pmap_steal_ptp */
/*
* pool that pmap structures are allocated from
@@ -603,7 +360,7 @@ struct pool pmap_pmap_pool;
*/
static pt_entry_t *csrc_pte, *cdst_pte, *zero_pte, *ptp_pte;
-caddr_t pmap_csrcp, pmap_cdstp, pmap_zerop, pmap_ptpp;
+static caddr_t csrcp, cdstp, zerop, ptpp;
caddr_t vmmap; /* XXX: used by mem.c... it should really uvm_map_reserve it */
#if defined(I586_CPU)
@@ -617,27 +374,46 @@ extern vaddr_t pentium_idt_vaddr;
*/
struct pv_entry *pmap_add_pvpage(struct pv_page *, boolean_t);
+struct vm_page *pmap_alloc_ptp(struct pmap *, int, boolean_t);
+struct pv_entry *pmap_alloc_pv(struct pmap *, int); /* see codes below */
#define ALLOCPV_NEED 0 /* need PV now */
#define ALLOCPV_TRY 1 /* just try to allocate, don't steal */
#define ALLOCPV_NONEED 2 /* don't need PV, just growing cache */
struct pv_entry *pmap_alloc_pvpage(struct pmap *, int);
-struct vm_page *pmap_alloc_ptp_86(struct pmap *, int, boolean_t);
-struct vm_page *pmap_get_ptp_86(struct pmap *, int, boolean_t);
-struct vm_page *pmap_steal_ptp_86(struct uvm_object *, vaddr_t);
-pt_entry_t *pmap_map_ptes_86(struct pmap *);
-void pmap_unmap_ptes_86(struct pmap *);
-boolean_t pmap_remove_pte_86(struct pmap *, struct vm_page *,
- pt_entry_t *, vaddr_t, int32_t *);
-void pmap_remove_ptes_86(struct pmap *, struct vm_page *, vaddr_t,
- vaddr_t, vaddr_t, int32_t *);
-vaddr_t pmap_tmpmap_pa_86(paddr_t);
-pt_entry_t *pmap_tmpmap_pvepte_86(struct pv_entry *);
-void pmap_tmpunmap_pa_86(void);
-void pmap_tmpunmap_pvepte_86(struct pv_entry *);
-boolean_t pmap_try_steal_pv_86(struct pv_head *,
- struct pv_entry *, struct pv_entry *);
-
-void pmap_release(pmap_t);
+void pmap_enter_pv(struct pv_head *,
+ struct pv_entry *, struct pmap *,
+ vaddr_t, struct vm_page *);
+void pmap_free_pv(struct pmap *, struct pv_entry *);
+void pmap_free_pvs(struct pmap *, struct pv_entry *);
+void pmap_free_pv_doit(struct pv_entry *);
+void pmap_free_pvpage(void);
+struct vm_page *pmap_get_ptp(struct pmap *, int, boolean_t);
+boolean_t pmap_is_curpmap(struct pmap *);
+boolean_t pmap_is_active(struct pmap *, int);
+pt_entry_t *pmap_map_ptes(struct pmap *);
+struct pv_entry *pmap_remove_pv(struct pv_head *, struct pmap *,
+ vaddr_t);
+boolean_t pmap_remove_pte(struct pmap *, struct vm_page *, pt_entry_t *,
+ vaddr_t, int32_t *);
+void pmap_remove_ptes(struct pmap *, struct vm_page *, vaddr_t,
+ vaddr_t, vaddr_t, int32_t *);
+struct vm_page *pmap_steal_ptp(struct uvm_object *, vaddr_t);
+vaddr_t pmap_tmpmap_pa(paddr_t);
+pt_entry_t *pmap_tmpmap_pvepte(struct pv_entry *);
+void pmap_tmpunmap_pa(void);
+void pmap_tmpunmap_pvepte(struct pv_entry *);
+void pmap_apte_flush(struct pmap *);
+boolean_t pmap_try_steal_pv(struct pv_head *,
+ struct pv_entry *,
+ struct pv_entry *);
+void pmap_unmap_ptes(struct pmap *);
+void pmap_exec_account(struct pmap *, vaddr_t, pt_entry_t,
+ pt_entry_t);
+
+void pmap_pinit(pmap_t);
+void pmap_release(pmap_t);
+
+void pmap_zero_phys(paddr_t);
void setcslimit(struct pmap *, struct trapframe *, struct pcb *, vaddr_t);
@@ -677,13 +453,13 @@ pmap_is_active(pmap, cpu_id)
*/
vaddr_t
-pmap_tmpmap_pa_86(paddr_t pa)
+pmap_tmpmap_pa(paddr_t pa)
{
#ifdef MULTIPROCESSOR
int id = cpu_number();
#endif
pt_entry_t *ptpte = PTESLEW(ptp_pte, id);
- caddr_t ptpva = VASLEW(pmap_ptpp, id);
+ caddr_t ptpva = VASLEW(ptpp, id);
#if defined(DIAGNOSTIC)
if (*ptpte)
panic("pmap_tmpmap_pa: ptp_pte in use?");
@@ -697,13 +473,13 @@ pmap_tmpmap_pa_86(paddr_t pa)
*/
void
-pmap_tmpunmap_pa_86()
+pmap_tmpunmap_pa()
{
#ifdef MULTIPROCESSOR
int id = cpu_number();
#endif
pt_entry_t *ptpte = PTESLEW(ptp_pte, id);
- caddr_t ptpva = VASLEW(pmap_ptpp, id);
+ caddr_t ptpva = VASLEW(ptpp, id);
#if defined(DIAGNOSTIC)
if (!pmap_valid_entry(*ptpte))
panic("pmap_tmpunmap_pa: our pte invalid?");
@@ -724,7 +500,7 @@ pmap_tmpunmap_pa_86()
*/
pt_entry_t *
-pmap_tmpmap_pvepte_86(struct pv_entry *pve)
+pmap_tmpmap_pvepte(struct pv_entry *pve)
{
#ifdef DIAGNOSTIC
if (pve->pv_pmap == pmap_kernel())
@@ -735,7 +511,7 @@ pmap_tmpmap_pvepte_86(struct pv_entry *pve)
if (pmap_is_curpmap(pve->pv_pmap))
return(vtopte(pve->pv_va));
- return(((pt_entry_t *)pmap_tmpmap_pa_86(VM_PAGE_TO_PHYS(pve->pv_ptp)))
+ return(((pt_entry_t *)pmap_tmpmap_pa(VM_PAGE_TO_PHYS(pve->pv_ptp)))
+ ptei((unsigned)pve->pv_va));
}
@@ -744,13 +520,13 @@ pmap_tmpmap_pvepte_86(struct pv_entry *pve)
*/
void
-pmap_tmpunmap_pvepte_86(struct pv_entry *pve)
+pmap_tmpunmap_pvepte(struct pv_entry *pve)
{
/* was it current pmap? if so, return */
if (pmap_is_curpmap(pve->pv_pmap))
return;
- pmap_tmpunmap_pa_86();
+ pmap_tmpunmap_pa();
}
void
@@ -793,7 +569,7 @@ pmap_apte_flush(struct pmap *pmap)
*/
pt_entry_t *
-pmap_map_ptes_86(struct pmap *pmap)
+pmap_map_ptes(struct pmap *pmap)
{
pd_entry_t opde;
@@ -832,7 +608,7 @@ pmap_map_ptes_86(struct pmap *pmap)
*/
void
-pmap_unmap_ptes_86(struct pmap *pmap)
+pmap_unmap_ptes(struct pmap *pmap)
{
if (pmap == pmap_kernel())
return;
@@ -851,7 +627,7 @@ pmap_unmap_ptes_86(struct pmap *pmap)
void
pmap_exec_account(struct pmap *pm, vaddr_t va,
- u_int32_t opte, u_int32_t npte)
+ pt_entry_t opte, pt_entry_t npte)
{
if (curproc == NULL || curproc->p_vmspace == NULL ||
pm != vm_map_pmap(&curproc->p_vmspace->vm_map))
@@ -925,48 +701,6 @@ pmap_exec_fixup(struct vm_map *map, struct trapframe *tf, struct pcb *pcb)
return (1);
}
-u_int32_t
-pmap_pte_set_86(vaddr_t va, paddr_t pa, u_int32_t bits)
-{
- pt_entry_t pte, *ptep = vtopte(va);
-
- pte = i386_atomic_testset_ul(ptep, pa | bits); /* zap! */
- return (pte & ~PG_FRAME);
-}
-
-u_int32_t
-pmap_pte_setbits_86(vaddr_t va, u_int32_t set, u_int32_t clr)
-{
- pt_entry_t *ptep = vtopte(va);
- pt_entry_t pte = *ptep;
-
- *ptep = (pte | set) & ~clr;
- return (pte & ~PG_FRAME);
-
-}
-
-u_int32_t
-pmap_pte_bits_86(vaddr_t va)
-{
- pt_entry_t *ptep = vtopte(va);
-
- return (*ptep & ~PG_FRAME);
-}
-
-paddr_t
-pmap_pte_paddr_86(vaddr_t va)
-{
- pt_entry_t *ptep = vtopte(va);
-
- return (*ptep & PG_FRAME);
-}
-
-paddr_t
-vtophys(vaddr_t va)
-{
- return ((*vtopte(va) & PG_FRAME) | (va & ~PG_FRAME));
-}
-
void
setcslimit(struct pmap *pm, struct trapframe *tf, struct pcb *pcb,
vaddr_t limit)
@@ -1021,15 +755,16 @@ setcslimit(struct pmap *pm, struct trapframe *tf, struct pcb *pcb,
void
pmap_kenter_pa(vaddr_t va, paddr_t pa, vm_prot_t prot)
{
- u_int32_t bits;
+ pt_entry_t *pte, opte, npte;
- bits = pmap_pte_set(va, pa, ((prot & VM_PROT_WRITE)? PG_RW : PG_RO) |
- PG_V | pmap_pg_g);
- if (pmap_valid_entry(bits)) {
+ pte = vtopte(va);
+ npte = pa | ((prot & VM_PROT_WRITE)? PG_RW : PG_RO) | PG_V | pmap_pg_g;
+ opte = i386_atomic_testset_ul(pte, npte); /* zap! */
+ if (pmap_valid_entry(opte)) {
#ifdef MULTIPROCESSOR
int32_t cpumask = 0;
- pmap_tlb_shootdown(pmap_kernel(), va, bits, &cpumask);
+ pmap_tlb_shootdown(pmap_kernel(), va, opte, &cpumask);
pmap_tlb_shootnow(cpumask);
#else
/* Don't bother deferring in the single CPU case. */
@@ -1050,21 +785,25 @@ pmap_kenter_pa(vaddr_t va, paddr_t pa, vm_prot_t prot)
void
pmap_kremove(vaddr_t va, vsize_t len)
{
- u_int32_t bits;
+ pt_entry_t *pte, opte;
#ifdef MULTIPROCESSOR
int32_t cpumask = 0;
#endif
len >>= PAGE_SHIFT;
for ( /* null */ ; len ; len--, va += PAGE_SIZE) {
- bits = pmap_pte_set(va, 0, 0);
+ if (va < VM_MIN_KERNEL_ADDRESS)
+ pte = vtopte(va);
+ else
+ pte = kvtopte(va);
+ opte = i386_atomic_testset_ul(pte, 0); /* zap! */
#ifdef DIAGNOSTIC
- if (bits & PG_PVLIST)
+ if (opte & PG_PVLIST)
panic("pmap_kremove: PG_PVLIST mapping for 0x%lx", va);
#endif
- if ((bits & (PG_V | PG_U)) == (PG_V | PG_U))
+ if ((opte & (PG_V | PG_U)) == (PG_V | PG_U))
#ifdef MULTIPROCESSOR
- pmap_tlb_shootdown(pmap_kernel(), va, bits, &cpumask);
+ pmap_tlb_shootdown(pmap_kernel(), va, opte, &cpumask);
#else
pmap_update_pg(va);
#endif
@@ -1165,8 +904,8 @@ pmap_bootstrap(vaddr_t kva_start)
kpm->pm_obj.uo_npages = 0;
kpm->pm_obj.uo_refs = 1;
bzero(&kpm->pm_list, sizeof(kpm->pm_list)); /* pm_list not used */
- kpm->pm_pdir = (vaddr_t)(proc0.p_addr->u_pcb.pcb_cr3 + KERNBASE);
- kpm->pm_pdirpa = proc0.p_addr->u_pcb.pcb_cr3;
+ kpm->pm_pdir = (pd_entry_t *)(proc0.p_addr->u_pcb.pcb_cr3 + KERNBASE);
+ kpm->pm_pdirpa = (u_int32_t) proc0.p_addr->u_pcb.pcb_cr3;
kpm->pm_stats.wired_count = kpm->pm_stats.resident_count =
atop(kva_start - VM_MIN_KERNEL_ADDRESS);
@@ -1209,27 +948,27 @@ pmap_bootstrap(vaddr_t kva_start)
* as well; we could waste less space if we knew the largest
* CPU ID beforehand.
*/
- pmap_csrcp = (caddr_t) virtual_avail; csrc_pte = pte;
+ csrcp = (caddr_t) virtual_avail; csrc_pte = pte;
- pmap_cdstp = (caddr_t) virtual_avail+PAGE_SIZE; cdst_pte = pte+1;
+ cdstp = (caddr_t) virtual_avail+PAGE_SIZE; cdst_pte = pte+1;
- pmap_zerop = (caddr_t) virtual_avail+PAGE_SIZE*2; zero_pte = pte+2;
+ zerop = (caddr_t) virtual_avail+PAGE_SIZE*2; zero_pte = pte+2;
- pmap_ptpp = (caddr_t) virtual_avail+PAGE_SIZE*3; ptp_pte = pte+3;
+ ptpp = (caddr_t) virtual_avail+PAGE_SIZE*3; ptp_pte = pte+3;
virtual_avail += PAGE_SIZE * I386_MAXPROCS * NPTECL;
pte += I386_MAXPROCS * NPTECL;
#else
- pmap_csrcp = (caddr_t) virtual_avail; csrc_pte = pte; /* allocate */
+ csrcp = (caddr_t) virtual_avail; csrc_pte = pte; /* allocate */
virtual_avail += PAGE_SIZE; pte++; /* advance */
- pmap_cdstp = (caddr_t) virtual_avail; cdst_pte = pte;
+ cdstp = (caddr_t) virtual_avail; cdst_pte = pte;
virtual_avail += PAGE_SIZE; pte++;
- pmap_zerop = (caddr_t) virtual_avail; zero_pte = pte;
+ zerop = (caddr_t) virtual_avail; zero_pte = pte;
virtual_avail += PAGE_SIZE; pte++;
- pmap_ptpp = (caddr_t) virtual_avail; ptp_pte = pte;
+ ptpp = (caddr_t) virtual_avail; ptp_pte = pte;
virtual_avail += PAGE_SIZE; pte++;
#endif
@@ -1263,7 +1002,7 @@ pmap_bootstrap(vaddr_t kva_start)
* initialize the pmap pool.
*/
- pool_init(&pmap_pmap_pool, sizeof(struct pmap), 32, 0, 0, "pmappl",
+ pool_init(&pmap_pmap_pool, sizeof(struct pmap), 0, 0, 0, "pmappl",
&pool_allocator_nointr);
/*
@@ -1277,14 +1016,6 @@ pmap_bootstrap(vaddr_t kva_start)
mtx_init(&pmap_tlb_shootdown_q[i].pq_mutex, IPL_IPI);
}
-#if defined(MULTIPROCESSOR)
- /* install the page after boot args as PT page for first 4M */
- pmap_enter(pmap_kernel(), (u_long)vtopte(0),
- round_page((vaddr_t)(bootargv + bootargc)),
- VM_PROT_READ|VM_PROT_WRITE, VM_PROT_READ|VM_PROT_WRITE|PMAP_WIRED);
- memset(vtopte(0), 0, NBPG); /* make sure it is clean before using */
-#endif
-
/*
* ensure the TLB is sync'd with reality by flushing it...
*/
@@ -1588,7 +1319,7 @@ steal_one:
*/
boolean_t
-pmap_try_steal_pv_86(struct pv_head *pvh, struct pv_entry *cpv,
+pmap_try_steal_pv(struct pv_head *pvh, struct pv_entry *cpv,
struct pv_entry *prevpv)
{
pt_entry_t *ptep, opte;
@@ -1609,7 +1340,7 @@ pmap_try_steal_pv_86(struct pv_head *pvh, struct pv_entry *cpv,
* mapping from the pmap.
*/
- ptep = pmap_tmpmap_pvepte_86(cpv);
+ ptep = pmap_tmpmap_pvepte(cpv);
if (*ptep & PG_W) {
ptep = NULL; /* wired page, avoid stealing this one */
} else {
@@ -1622,7 +1353,7 @@ pmap_try_steal_pv_86(struct pv_head *pvh, struct pv_entry *cpv,
if (pmap_is_curpmap(cpv->pv_pmap))
pmap_update_pg(cpv->pv_va);
#endif
- pmap_tmpunmap_pvepte_86(cpv);
+ pmap_tmpunmap_pvepte(cpv);
}
if (ptep == NULL) {
simple_unlock(&cpv->pv_pmap->pm_obj.vmobjlock);
@@ -1904,7 +1635,7 @@ pmap_remove_pv(struct pv_head *pvh, struct pmap *pmap, vaddr_t va)
*/
struct vm_page *
-pmap_alloc_ptp_86(struct pmap *pmap, int pde_index, boolean_t just_try)
+pmap_alloc_ptp(struct pmap *pmap, int pde_index, boolean_t just_try)
{
struct vm_page *ptp;
@@ -1913,7 +1644,7 @@ pmap_alloc_ptp_86(struct pmap *pmap, int pde_index, boolean_t just_try)
if (ptp == NULL) {
if (just_try)
return(NULL);
- ptp = pmap_steal_ptp_86(&pmap->pm_obj, ptp_i2o(pde_index));
+ ptp = pmap_steal_ptp(&pmap->pm_obj, ptp_i2o(pde_index));
if (ptp == NULL) {
return (NULL);
}
@@ -1924,8 +1655,8 @@ pmap_alloc_ptp_86(struct pmap *pmap, int pde_index, boolean_t just_try)
/* got one! */
ptp->flags &= ~PG_BUSY; /* never busy */
ptp->wire_count = 1; /* no mappings yet */
- PDE(pmap, pde_index) =
- (pd_entry_t) (VM_PAGE_TO_PHYS(ptp) | PG_u | PG_RW | PG_V);
+ pmap->pm_pdir[pde_index] =
+ (pd_entry_t) (VM_PAGE_TO_PHYS(ptp) | PG_u | PG_RW | PG_V);
pmap->pm_stats.resident_count++; /* count PTP as resident */
pmap->pm_ptphint = ptp;
return(ptp);
@@ -1943,7 +1674,7 @@ pmap_alloc_ptp_86(struct pmap *pmap, int pde_index, boolean_t just_try)
*/
struct vm_page *
-pmap_steal_ptp_86(struct uvm_object *obj, vaddr_t offset)
+pmap_steal_ptp(struct uvm_object *obj, vaddr_t offset)
{
struct vm_page *ptp = NULL;
struct pmap *firstpmap;
@@ -1977,21 +1708,21 @@ pmap_steal_ptp_86(struct uvm_object *obj, vaddr_t offset)
idx = ptp_o2i(ptp->offset);
#ifdef DIAGNOSTIC
if (VM_PAGE_TO_PHYS(ptp) !=
- (PDE(pmaps_hand, idx) & PG_FRAME))
+ (pmaps_hand->pm_pdir[idx] & PG_FRAME))
panic("pmap_steal_ptp: PTP mismatch!");
#endif
ptes = (pt_entry_t *)
- pmap_tmpmap_pa_86(VM_PAGE_TO_PHYS(ptp));
+ pmap_tmpmap_pa(VM_PAGE_TO_PHYS(ptp));
for (lcv = 0 ; lcv < PTES_PER_PTP ; lcv++)
if ((ptes[lcv] & (PG_V|PG_W)) ==
(PG_V|PG_W))
break;
if (lcv == PTES_PER_PTP)
- pmap_remove_ptes_86(pmaps_hand, ptp,
+ pmap_remove_ptes(pmaps_hand, ptp,
(vaddr_t)ptes, ptp_i2v(idx),
ptp_i2v(idx+1), &cpumask);
- pmap_tmpunmap_pa_86();
+ pmap_tmpunmap_pa();
if (lcv != PTES_PER_PTP)
/* wired, try next PTP */
@@ -2001,7 +1732,7 @@ pmap_steal_ptp_86(struct uvm_object *obj, vaddr_t offset)
* got it!!!
*/
- PDE(pmaps_hand, idx) = 0; /* zap! */
+ pmaps_hand->pm_pdir[idx] = 0; /* zap! */
pmaps_hand->pm_stats.resident_count--;
#ifdef MULTIPROCESSOR
pmap_apte_flush(pmaps_hand);
@@ -2045,15 +1776,15 @@ pmap_steal_ptp_86(struct uvm_object *obj, vaddr_t offset)
*/
struct vm_page *
-pmap_get_ptp_86(struct pmap *pmap, int pde_index, boolean_t just_try)
+pmap_get_ptp(struct pmap *pmap, int pde_index, boolean_t just_try)
{
struct vm_page *ptp;
- if (pmap_valid_entry(PDE(pmap, pde_index))) {
+ if (pmap_valid_entry(pmap->pm_pdir[pde_index])) {
/* valid... check hint (saves us a PA->PG lookup) */
if (pmap->pm_ptphint &&
- (PDE(pmap, pde_index) & PG_FRAME) ==
+ (pmap->pm_pdir[pde_index] & PG_FRAME) ==
VM_PAGE_TO_PHYS(pmap->pm_ptphint))
return(pmap->pm_ptphint);
@@ -2067,7 +1798,7 @@ pmap_get_ptp_86(struct pmap *pmap, int pde_index, boolean_t just_try)
}
/* allocate a new PTP (updates ptphint) */
- return(pmap_alloc_ptp_86(pmap, pde_index, just_try));
+ return(pmap_alloc_ptp(pmap, pde_index, just_try));
}
/*
@@ -2087,7 +1818,17 @@ pmap_create(void)
struct pmap *pmap;
pmap = pool_get(&pmap_pmap_pool, PR_WAITOK);
+ pmap_pinit(pmap);
+ return(pmap);
+}
+/*
+ * pmap_pinit: given a zero'd pmap structure, init it.
+ */
+
+void
+pmap_pinit(struct pmap *pmap)
+{
/* init uvm_object */
simple_lock_init(&pmap->pm_obj.vmobjlock);
pmap->pm_obj.pgops = NULL; /* currently not a mappable object */
@@ -2100,37 +1841,26 @@ pmap_create(void)
pmap->pm_hiexec = 0;
pmap->pm_flags = 0;
- /* init the LDT */
- pmap->pm_ldt = NULL;
- pmap->pm_ldt_len = 0;
- pmap->pm_ldt_sel = GSEL(GLDT_SEL, SEL_KPL);
-
setsegment(&pmap->pm_codeseg, 0, atop(I386_MAX_EXE_ADDR) - 1,
SDT_MEMERA, SEL_UPL, 1, 1);
- pmap_pinit_pd(pmap);
- return(pmap);
-}
-
-/*
- * pmap_pinit: given a zero'd pmap structure, init it.
- */
-
-void
-pmap_pinit_pd_86(struct pmap *pmap)
-{
/* allocate PDP */
- pmap->pm_pdir = uvm_km_alloc(kernel_map, NBPG);
+ pmap->pm_pdir = (pd_entry_t *) uvm_km_alloc(kernel_map, NBPG);
if (pmap->pm_pdir == NULL)
panic("pmap_pinit: kernel_map out of virtual space!");
- pmap_extract(pmap_kernel(), (vaddr_t)pmap->pm_pdir, &pmap->pm_pdirpa);
- pmap->pm_pdirsize = NBPG;
+ (void) pmap_extract(pmap_kernel(), (vaddr_t)pmap->pm_pdir,
+ (paddr_t *)&pmap->pm_pdirpa);
/* init PDP */
/* zero init area */
- bzero((void *)pmap->pm_pdir, PDSLOT_PTE * sizeof(pd_entry_t));
+ bzero(pmap->pm_pdir, PDSLOT_PTE * sizeof(pd_entry_t));
/* put in recursive PDE to map the PTEs */
- PDE(pmap, PDSLOT_PTE) = pmap->pm_pdirpa | PG_V | PG_KW;
+ pmap->pm_pdir[PDSLOT_PTE] = pmap->pm_pdirpa | PG_V | PG_KW;
+
+ /* init the LDT */
+ pmap->pm_ldt = NULL;
+ pmap->pm_ldt_len = 0;
+ pmap->pm_ldt_sel = GSEL(GLDT_SEL, SEL_KPL);
/*
* we need to lock pmaps_lock to prevent nkpde from changing on
@@ -2140,10 +1870,10 @@ pmap_pinit_pd_86(struct pmap *pmap)
*/
simple_lock(&pmaps_lock);
/* put in kernel VM PDEs */
- bcopy(&PDP_BASE[PDSLOT_KERN], &PDE(pmap, PDSLOT_KERN),
+ bcopy(&PDP_BASE[PDSLOT_KERN], &pmap->pm_pdir[PDSLOT_KERN],
nkpde * sizeof(pd_entry_t));
/* zero the rest */
- bzero(&PDE(pmap, PDSLOT_KERN + nkpde),
+ bzero(&pmap->pm_pdir[PDSLOT_KERN + nkpde],
NBPG - ((PDSLOT_KERN + nkpde) * sizeof(pd_entry_t)));
LIST_INSERT_HEAD(&pmaps, pmap, pm_list);
simple_unlock(&pmaps_lock);
@@ -2221,7 +1951,7 @@ pmap_release(struct pmap *pmap)
* MULTIPROCESSOR -- no need to flush out of other processors'
* APTE space because we do that in pmap_unmap_ptes().
*/
- uvm_km_free(kernel_map, pmap->pm_pdir, pmap->pm_pdirsize);
+ uvm_km_free(kernel_map, (vaddr_t)pmap->pm_pdir, NBPG);
#ifdef USER_LDT
if (pmap->pm_flags & PMF_USER_LDT) {
@@ -2402,15 +2132,15 @@ pmap_deactivate(struct proc *p)
*/
boolean_t
-pmap_extract_86(struct pmap *pmap, vaddr_t va, paddr_t *pap)
+pmap_extract(struct pmap *pmap, vaddr_t va, paddr_t *pap)
{
paddr_t retval;
pt_entry_t *ptes;
- if (PDE(pmap, pdei(va))) {
- ptes = pmap_map_ptes_86(pmap);
+ if (pmap->pm_pdir[pdei(va)]) {
+ ptes = pmap_map_ptes(pmap);
retval = (paddr_t)(ptes[atop(va)] & PG_FRAME);
- pmap_unmap_ptes_86(pmap);
+ pmap_unmap_ptes(pmap);
if (pap != NULL)
*pap = retval | (va & ~PG_FRAME);
return (TRUE);
@@ -2446,13 +2176,13 @@ pmap_zero_page(struct vm_page *pg)
* initialized.
*/
void
-pmap_zero_phys_86(paddr_t pa)
+pmap_zero_phys(paddr_t pa)
{
#ifdef MULTIPROCESSOR
int id = cpu_number();
#endif
pt_entry_t *zpte = PTESLEW(zero_pte, id);
- caddr_t zerova = VASLEW(pmap_zerop, id);
+ caddr_t zerova = VASLEW(zerop, id);
#ifdef DIAGNOSTIC
if (*zpte)
@@ -2470,13 +2200,13 @@ pmap_zero_phys_86(paddr_t pa)
*/
boolean_t
-pmap_zero_page_uncached_86(paddr_t pa)
+pmap_zero_page_uncached(paddr_t pa)
{
#ifdef MULTIPROCESSOR
int id = cpu_number();
#endif
pt_entry_t *zpte = PTESLEW(zero_pte, id);
- caddr_t zerova = VASLEW(pmap_zerop, id);
+ caddr_t zerova = VASLEW(zerop, id);
#ifdef DIAGNOSTIC
if (*zpte)
@@ -2497,7 +2227,7 @@ pmap_zero_page_uncached_86(paddr_t pa)
*/
void
-pmap_copy_page_86(struct vm_page *srcpg, struct vm_page *dstpg)
+pmap_copy_page(struct vm_page *srcpg, struct vm_page *dstpg)
{
paddr_t srcpa = VM_PAGE_TO_PHYS(srcpg);
paddr_t dstpa = VM_PAGE_TO_PHYS(dstpg);
@@ -2506,8 +2236,8 @@ pmap_copy_page_86(struct vm_page *srcpg, struct vm_page *dstpg)
#endif
pt_entry_t *spte = PTESLEW(csrc_pte,id);
pt_entry_t *dpte = PTESLEW(cdst_pte,id);
- caddr_t csrcva = VASLEW(pmap_csrcp, id);
- caddr_t cdstva = VASLEW(pmap_cdstp, id);
+ caddr_t csrcva = VASLEW(csrcp, id);
+ caddr_t cdstva = VASLEW(cdstp, id);
#ifdef DIAGNOSTIC
if (*spte || *dpte)
@@ -2541,7 +2271,7 @@ pmap_copy_page_86(struct vm_page *srcpg, struct vm_page *dstpg)
*/
void
-pmap_remove_ptes_86(struct pmap *pmap, struct vm_page *ptp, vaddr_t ptpva,
+pmap_remove_ptes(struct pmap *pmap, struct vm_page *ptp, vaddr_t ptpva,
vaddr_t startva, vaddr_t endva, int32_t *cpumaskp)
{
struct pv_entry *pv_tofree = NULL; /* list of pv_entrys to free */
@@ -2600,7 +2330,7 @@ pmap_remove_ptes_86(struct pmap *pmap, struct vm_page *ptp, vaddr_t ptpva,
#ifdef DIAGNOSTIC
if (bank == -1)
panic("pmap_remove_ptes: unmanaged page marked "
- "PG_PVLIST, va = 0x%lx, pa = 0x%llx",
+ "PG_PVLIST, va = 0x%lx, pa = 0x%lx",
startva, (u_long)(opte & PG_FRAME));
#endif
@@ -2634,7 +2364,7 @@ pmap_remove_ptes_86(struct pmap *pmap, struct vm_page *ptp, vaddr_t ptpva,
*/
boolean_t
-pmap_remove_pte_86(struct pmap *pmap, struct vm_page *ptp, pt_entry_t *pte,
+pmap_remove_pte(struct pmap *pmap, struct vm_page *ptp, pt_entry_t *pte,
vaddr_t va, int32_t *cpumaskp)
{
pt_entry_t opte;
@@ -2681,7 +2411,7 @@ pmap_remove_pte_86(struct pmap *pmap, struct vm_page *ptp, pt_entry_t *pte,
#ifdef DIAGNOSTIC
if (bank == -1)
panic("pmap_remove_pte: unmanaged page marked "
- "PG_PVLIST, va = 0x%lx, pa = 0x%llx", va,
+ "PG_PVLIST, va = 0x%lx, pa = 0x%lx", va,
(u_long)(opte & PG_FRAME));
#endif
@@ -2703,7 +2433,7 @@ pmap_remove_pte_86(struct pmap *pmap, struct vm_page *ptp, pt_entry_t *pte,
*/
void
-pmap_remove_86(struct pmap *pmap, vaddr_t sva, vaddr_t eva)
+pmap_remove(struct pmap *pmap, vaddr_t sva, vaddr_t eva)
{
pt_entry_t *ptes, opte;
boolean_t result;
@@ -2717,7 +2447,7 @@ pmap_remove_86(struct pmap *pmap, vaddr_t sva, vaddr_t eva)
*/
PMAP_MAP_TO_HEAD_LOCK();
- ptes = pmap_map_ptes_86(pmap); /* locks pmap */
+ ptes = pmap_map_ptes(pmap); /* locks pmap */
/*
* removing one page? take shortcut function.
@@ -2725,10 +2455,10 @@ pmap_remove_86(struct pmap *pmap, vaddr_t sva, vaddr_t eva)
if (sva + PAGE_SIZE == eva) {
- if (pmap_valid_entry(PDE(pmap, pdei(sva)))) {
+ if (pmap_valid_entry(pmap->pm_pdir[pdei(sva)])) {
/* PA of the PTP */
- ptppa = PDE(pmap, pdei(sva)) & PG_FRAME;
+ ptppa = pmap->pm_pdir[pdei(sva)] & PG_FRAME;
/* get PTP if non-kernel mapping */
@@ -2751,7 +2481,7 @@ pmap_remove_86(struct pmap *pmap, vaddr_t sva, vaddr_t eva)
}
/* do it! */
- result = pmap_remove_pte_86(pmap, ptp,
+ result = pmap_remove_pte(pmap, ptp,
&ptes[atop(sva)], sva, &cpumask);
/*
@@ -2762,7 +2492,7 @@ pmap_remove_86(struct pmap *pmap, vaddr_t sva, vaddr_t eva)
if (result && ptp && ptp->wire_count <= 1) {
/* zap! */
opte = i386_atomic_testset_ul(
- &PDE(pmap, pdei(sva)), 0);
+ &pmap->pm_pdir[pdei(sva)], 0);
#ifdef MULTIPROCESSOR
/*
* XXXthorpej Redundant shootdown can happen
@@ -2793,7 +2523,7 @@ pmap_remove_86(struct pmap *pmap, vaddr_t sva, vaddr_t eva)
}
}
pmap_tlb_shootnow(cpumask);
- pmap_unmap_ptes_86(pmap); /* unlock pmap */
+ pmap_unmap_ptes(pmap); /* unlock pmap */
PMAP_MAP_TO_HEAD_UNLOCK();
return;
}
@@ -2823,12 +2553,12 @@ pmap_remove_86(struct pmap *pmap, vaddr_t sva, vaddr_t eva)
/* XXXCDC: ugly hack to avoid freeing PDP here */
continue;
- if (!pmap_valid_entry(PDE(pmap, pdei(sva))))
+ if (!pmap_valid_entry(pmap->pm_pdir[pdei(sva)]))
/* valid block? */
continue;
/* PA of the PTP */
- ptppa = PDE(pmap, pdei(sva)) & PG_FRAME;
+ ptppa = (pmap->pm_pdir[pdei(sva)] & PG_FRAME);
/* get PTP if non-kernel mapping */
if (pmap == pmap_kernel()) {
@@ -2847,14 +2577,14 @@ pmap_remove_86(struct pmap *pmap, vaddr_t sva, vaddr_t eva)
#endif
}
}
- pmap_remove_ptes_86(pmap, ptp, (vaddr_t)&ptes[atop(sva)],
+ pmap_remove_ptes(pmap, ptp, (vaddr_t)&ptes[atop(sva)],
sva, blkendva, &cpumask);
/* if PTP is no longer being used, free it! */
if (ptp && ptp->wire_count <= 1) {
/* zap! */
opte = i386_atomic_testset_ul(
- &PDE(pmap, pdei(sva)), 0);
+ &pmap->pm_pdir[pdei(sva)], 0);
#if defined(MULTIPROCESSOR)
/*
* XXXthorpej Redundant shootdown can happen here
@@ -2883,7 +2613,7 @@ pmap_remove_86(struct pmap *pmap, vaddr_t sva, vaddr_t eva)
}
pmap_tlb_shootnow(cpumask);
- pmap_unmap_ptes_86(pmap);
+ pmap_unmap_ptes(pmap);
PMAP_MAP_TO_HEAD_UNLOCK();
}
@@ -2895,7 +2625,7 @@ pmap_remove_86(struct pmap *pmap, vaddr_t sva, vaddr_t eva)
*/
void
-pmap_page_remove_86(struct vm_page *pg)
+pmap_page_remove(struct vm_page *pg)
{
int bank, off;
struct pv_head *pvh;
@@ -2922,19 +2652,19 @@ pmap_page_remove_86(struct vm_page *pg)
simple_lock(&pvh->pvh_lock);
for (pve = pvh->pvh_list ; pve != NULL ; pve = pve->pv_next) {
- ptes = pmap_map_ptes_86(pve->pv_pmap); /* locks pmap */
+ ptes = pmap_map_ptes(pve->pv_pmap); /* locks pmap */
#ifdef DIAGNOSTIC
if (pve->pv_va >= uvm.pager_sva && pve->pv_va < uvm.pager_eva)
printf("pmap_page_remove: found pager VA on pv_list\n");
- if (pve->pv_ptp && (PDE(pve->pv_pmap,
- pdei(pve->pv_va)) & PG_FRAME) !=
- VM_PAGE_TO_PHYS(pve->pv_ptp)) {
+ if (pve->pv_ptp && (pve->pv_pmap->pm_pdir[pdei(pve->pv_va)] &
+ PG_FRAME)
+ != VM_PAGE_TO_PHYS(pve->pv_ptp)) {
printf("pmap_page_remove: pg=%p: va=%lx, pv_ptp=%p\n",
pg, pve->pv_va, pve->pv_ptp);
printf("pmap_page_remove: PTP's phys addr: "
"actual=%x, recorded=%lx\n",
- (PDE(pve->pv_pmap, pdei(pve->pv_va)) &
+ (pve->pv_pmap->pm_pdir[pdei(pve->pv_va)] &
PG_FRAME), VM_PAGE_TO_PHYS(pve->pv_ptp));
panic("pmap_page_remove: mapped managed page has "
"invalid pv_ptp field");
@@ -2970,7 +2700,8 @@ pmap_page_remove_86(struct vm_page *pg)
/* zap! */
opte = i386_atomic_testset_ul(
- &PDE(pve->pv_pmap, pdei(pve->pv_va)), 0);
+ &pve->pv_pmap->pm_pdir[pdei(pve->pv_va)],
+ 0);
pmap_tlb_shootdown(curpcb->pcb_pmap,
((vaddr_t)ptes) + pve->pv_ptp->offset,
opte, &cpumask);
@@ -2992,7 +2723,7 @@ pmap_page_remove_86(struct vm_page *pg)
uvm_pagefree(pve->pv_ptp);
}
}
- pmap_unmap_ptes_86(pve->pv_pmap); /* unlocks pmap */
+ pmap_unmap_ptes(pve->pv_pmap); /* unlocks pmap */
}
pmap_free_pvs(NULL, pvh->pvh_list);
pvh->pvh_list = NULL;
@@ -3015,7 +2746,7 @@ pmap_page_remove_86(struct vm_page *pg)
*/
boolean_t
-pmap_test_attrs_86(struct vm_page *pg, int testbits)
+pmap_test_attrs(struct vm_page *pg, int testbits)
{
int bank, off;
char *myattrs;
@@ -3052,9 +2783,9 @@ pmap_test_attrs_86(struct vm_page *pg, int testbits)
for (pve = pvh->pvh_list; pve != NULL && (*myattrs & testbits) == 0;
pve = pve->pv_next) {
- ptes = pmap_map_ptes_86(pve->pv_pmap);
+ ptes = pmap_map_ptes(pve->pv_pmap);
pte = ptes[atop(pve->pv_va)];
- pmap_unmap_ptes_86(pve->pv_pmap);
+ pmap_unmap_ptes(pve->pv_pmap);
*myattrs |= pte;
}
@@ -3076,7 +2807,7 @@ pmap_test_attrs_86(struct vm_page *pg, int testbits)
*/
boolean_t
-pmap_change_attrs_86(struct vm_page *pg, int setbits, int clearbits)
+pmap_change_attrs(struct vm_page *pg, int setbits, int clearbits)
{
u_int32_t result;
int bank, off;
@@ -3104,12 +2835,12 @@ pmap_change_attrs_86(struct vm_page *pg, int setbits, int clearbits)
for (pve = pvh->pvh_list; pve != NULL; pve = pve->pv_next) {
#ifdef DIAGNOSTIC
- if (!pmap_valid_entry(PDE(pve->pv_pmap, pdei(pve->pv_va))))
+ if (!pmap_valid_entry(pve->pv_pmap->pm_pdir[pdei(pve->pv_va)]))
panic("pmap_change_attrs: mapping without PTP "
"detected");
#endif
- ptes = pmap_map_ptes_86(pve->pv_pmap); /* locks pmap */
+ ptes = pmap_map_ptes(pve->pv_pmap); /* locks pmap */
npte = ptes[atop(pve->pv_va)];
result |= (npte & clearbits);
npte = (npte | setbits) & ~clearbits;
@@ -3119,7 +2850,7 @@ pmap_change_attrs_86(struct vm_page *pg, int setbits, int clearbits)
pmap_tlb_shootdown(pve->pv_pmap,
atop(pve->pv_va), opte, &cpumask);
}
- pmap_unmap_ptes_86(pve->pv_pmap); /* unlocks pmap */
+ pmap_unmap_ptes(pve->pv_pmap); /* unlocks pmap */
}
simple_unlock(&pvh->pvh_lock);
@@ -3155,7 +2886,7 @@ pmap_change_attrs_86(struct vm_page *pg, int setbits, int clearbits)
*/
void
-pmap_write_protect_86(struct pmap *pmap, vaddr_t sva, vaddr_t eva,
+pmap_write_protect(struct pmap *pmap, vaddr_t sva, vaddr_t eva,
vm_prot_t prot)
{
pt_entry_t *ptes, *spte, *epte, npte;
@@ -3163,7 +2894,7 @@ pmap_write_protect_86(struct pmap *pmap, vaddr_t sva, vaddr_t eva,
u_int32_t md_prot;
int32_t cpumask = 0;
- ptes = pmap_map_ptes_86(pmap); /* locks pmap */
+ ptes = pmap_map_ptes(pmap); /* locks pmap */
/* should be ok, but just in case ... */
sva &= PG_FRAME;
@@ -3189,7 +2920,7 @@ pmap_write_protect_86(struct pmap *pmap, vaddr_t sva, vaddr_t eva,
continue;
/* empty block? */
- if (!pmap_valid_entry(PDE(pmap, pdei(sva))))
+ if (!pmap_valid_entry(pmap->pm_pdir[pdei(sva)]))
continue;
md_prot = protection_codes[prot];
@@ -3218,7 +2949,7 @@ pmap_write_protect_86(struct pmap *pmap, vaddr_t sva, vaddr_t eva,
}
pmap_tlb_shootnow(cpumask);
- pmap_unmap_ptes_86(pmap); /* unlocks pmap */
+ pmap_unmap_ptes(pmap); /* unlocks pmap */
}
/*
@@ -3232,12 +2963,12 @@ pmap_write_protect_86(struct pmap *pmap, vaddr_t sva, vaddr_t eva,
*/
void
-pmap_unwire_86(struct pmap *pmap, vaddr_t va)
+pmap_unwire(struct pmap *pmap, vaddr_t va)
{
pt_entry_t *ptes;
- if (pmap_valid_entry(PDE(pmap, pdei(va)))) {
- ptes = pmap_map_ptes_86(pmap); /* locks pmap */
+ if (pmap_valid_entry(pmap->pm_pdir[pdei(va)])) {
+ ptes = pmap_map_ptes(pmap); /* locks pmap */
#ifdef DIAGNOSTIC
if (!pmap_valid_entry(ptes[atop(va)]))
@@ -3253,7 +2984,7 @@ pmap_unwire_86(struct pmap *pmap, vaddr_t va)
"didn't change!\n", pmap, va);
}
#endif
- pmap_unmap_ptes_86(pmap); /* unlocks map */
+ pmap_unmap_ptes(pmap); /* unlocks map */
}
#ifdef DIAGNOSTIC
else {
@@ -3299,7 +3030,7 @@ pmap_collect(struct pmap *pmap)
*/
int
-pmap_enter_86(struct pmap *pmap, vaddr_t va, paddr_t pa,
+pmap_enter(struct pmap *pmap, vaddr_t va, paddr_t pa,
vm_prot_t prot, int flags)
{
pt_entry_t *ptes, opte, npte;
@@ -3319,7 +3050,7 @@ pmap_enter_86(struct pmap *pmap, vaddr_t va, paddr_t pa,
/* sanity check: kernel PTPs should already have been pre-allocated */
if (va >= VM_MIN_KERNEL_ADDRESS &&
- !pmap_valid_entry(PDE(pmap, pdei(va))))
+ !pmap_valid_entry(pmap->pm_pdir[pdei(va)]))
panic("pmap_enter: missing kernel PTP!");
#endif
@@ -3330,11 +3061,11 @@ pmap_enter_86(struct pmap *pmap, vaddr_t va, paddr_t pa,
* map in ptes and get a pointer to our PTP (unless we are the kernel)
*/
- ptes = pmap_map_ptes_86(pmap); /* locks pmap */
+ ptes = pmap_map_ptes(pmap); /* locks pmap */
if (pmap == pmap_kernel()) {
ptp = NULL;
} else {
- ptp = pmap_get_ptp_86(pmap, pdei(va), FALSE);
+ ptp = pmap_get_ptp(pmap, pdei(va), FALSE);
if (ptp == NULL) {
if (flags & PMAP_CANFAIL) {
error = ENOMEM;
@@ -3376,7 +3107,7 @@ pmap_enter_86(struct pmap *pmap, vaddr_t va, paddr_t pa,
if (bank == -1)
panic("pmap_enter: same pa PG_PVLIST "
"mapping with unmanaged page "
- "pa = 0x%llx (0x%lx)", pa,
+ "pa = 0x%lx (0x%lx)", pa,
atop(pa));
#endif
pvh = &vm_physmem[bank].pmseg.pvhead[off];
@@ -3404,7 +3135,7 @@ pmap_enter_86(struct pmap *pmap, vaddr_t va, paddr_t pa,
if (bank == -1)
panic("pmap_enter: PG_PVLIST mapping with "
"unmanaged page "
- "pa = 0x%llx (0x%lx)", pa, atop(pa));
+ "pa = 0x%lx (0x%lx)", pa, atop(pa));
#endif
pvh = &vm_physmem[bank].pmseg.pvhead[off];
simple_lock(&pvh->pvh_lock);
@@ -3490,7 +3221,7 @@ enter_now:
error = 0;
out:
- pmap_unmap_ptes_86(pmap);
+ pmap_unmap_ptes(pmap);
PMAP_MAP_TO_HEAD_UNLOCK();
return error;
@@ -3504,7 +3235,7 @@ out:
*/
vaddr_t
-pmap_growkernel_86(vaddr_t maxkvaddr)
+pmap_growkernel(vaddr_t maxkvaddr)
{
struct pmap *kpm = pmap_kernel(), *pm;
int needed_kpde; /* needed number of kernel PTPs */
@@ -3535,9 +3266,10 @@ pmap_growkernel_86(vaddr_t maxkvaddr)
if (uvm_page_physget(&ptaddr) == FALSE)
panic("pmap_growkernel: out of memory");
- pmap_zero_phys_86(ptaddr);
+ pmap_zero_phys(ptaddr);
- PDE(kpm, PDSLOT_KERN + nkpde) = ptaddr | PG_RW | PG_V;
+ kpm->pm_pdir[PDSLOT_KERN + nkpde] =
+ ptaddr | PG_RW | PG_V;
/* count PTP as resident */
kpm->pm_stats.resident_count++;
@@ -3550,18 +3282,18 @@ pmap_growkernel_86(vaddr_t maxkvaddr)
* INVOKED WHILE pmap_init() IS RUNNING!
*/
- if (pmap_alloc_ptp_86(kpm, PDSLOT_KERN+nkpde, FALSE) == NULL) {
+ if (pmap_alloc_ptp(kpm, PDSLOT_KERN+nkpde, FALSE) == NULL) {
panic("pmap_growkernel: alloc ptp failed");
}
/* PG_u not for kernel */
- PDE(kpm, PDSLOT_KERN + nkpde) &= ~PG_u;
+ kpm->pm_pdir[PDSLOT_KERN + nkpde] &= ~PG_u;
/* distribute new kernel PTP to all active pmaps */
simple_lock(&pmaps_lock);
LIST_FOREACH(pm, &pmaps, pm_list) {
- PDE(pm, PDSLOT_KERN + nkpde) =
- PDE(kpm, PDSLOT_KERN + nkpde);
+ pm->pm_pdir[PDSLOT_KERN + nkpde] =
+ kpm->pm_pdir[PDSLOT_KERN + nkpde];
}
simple_unlock(&pmaps_lock);
}
@@ -3574,7 +3306,7 @@ out:
}
#ifdef DEBUG
-void pmap_dump_86(struct pmap *, vaddr_t, vaddr_t);
+void pmap_dump(struct pmap *, vaddr_t, vaddr_t);
/*
* pmap_dump: dump all the mappings from a pmap
@@ -3583,7 +3315,7 @@ void pmap_dump_86(struct pmap *, vaddr_t, vaddr_t);
*/
void
-pmap_dump_86(struct pmap *pmap, vaddr_t sva, vaddr_t eva)
+pmap_dump(struct pmap *pmap, vaddr_t sva, vaddr_t eva)
{
pt_entry_t *ptes, *pte;
vaddr_t blkendva;
@@ -3601,7 +3333,7 @@ pmap_dump_86(struct pmap *pmap, vaddr_t sva, vaddr_t eva)
*/
PMAP_MAP_TO_HEAD_LOCK();
- ptes = pmap_map_ptes_86(pmap); /* locks pmap */
+ ptes = pmap_map_ptes(pmap); /* locks pmap */
/*
* dumping a range of pages: we dump in PTP sized blocks (4MB)
@@ -3615,7 +3347,7 @@ pmap_dump_86(struct pmap *pmap, vaddr_t sva, vaddr_t eva)
blkendva = eva;
/* valid block? */
- if (!pmap_valid_entry(PDE(pmap, pdei(sva))))
+ if (!pmap_valid_entry(pmap->pm_pdir[pdei(sva)]))
continue;
pte = &ptes[atop(sva)];
@@ -3626,7 +3358,7 @@ pmap_dump_86(struct pmap *pmap, vaddr_t sva, vaddr_t eva)
sva, *pte, *pte & PG_FRAME);
}
}
- pmap_unmap_ptes_86(pmap);
+ pmap_unmap_ptes(pmap);
PMAP_MAP_TO_HEAD_UNLOCK();
}
#endif
@@ -3692,7 +3424,7 @@ pmap_tlb_shootnow(int32_t cpumask)
* Cause the TLB entry for pmap/va to be shot down.
*/
void
-pmap_tlb_shootdown(pmap_t pmap, vaddr_t va, u_int32_t pte, int32_t *cpumaskp)
+pmap_tlb_shootdown(pmap_t pmap, vaddr_t va, pt_entry_t pte, int32_t *cpumaskp)
{
struct cpu_info *ci, *self;
struct pmap_tlb_shootdown_q *pq;
@@ -3912,31 +3644,3 @@ pmap_tlb_shootdown_job_put(struct pmap_tlb_shootdown_q *pq,
pq->pq_count--;
}
-
-#ifndef SMALL_KERNEL
-u_int32_t (*pmap_pte_set_p)(vaddr_t, paddr_t, u_int32_t) =
- pmap_pte_set_86;
-u_int32_t (*pmap_pte_setbits_p)(vaddr_t, u_int32_t, u_int32_t) =
- pmap_pte_setbits_86;
-u_int32_t (*pmap_pte_bits_p)(vaddr_t) = pmap_pte_bits_86;
-paddr_t (*pmap_pte_paddr_p)(vaddr_t) = pmap_pte_paddr_86;
-boolean_t (*pmap_change_attrs_p)(struct vm_page *, int, int) =
- pmap_change_attrs_86;
-int (*pmap_enter_p)(pmap_t, vaddr_t, paddr_t, vm_prot_t, int) =
- pmap_enter_86;
-boolean_t (*pmap_extract_p)(pmap_t, vaddr_t, paddr_t *) = pmap_extract_86;
-vaddr_t (*pmap_growkernel_p)(vaddr_t) = pmap_growkernel_86;
-void (*pmap_page_remove_p)(struct vm_page *) = pmap_page_remove_86;
-void (*pmap_remove_p)(struct pmap *, vaddr_t, vaddr_t) = pmap_remove_86;
-boolean_t (*pmap_test_attrs_p)(struct vm_page *, int) = pmap_test_attrs_86;
-void (*pmap_unwire_p)(struct pmap *, vaddr_t) = pmap_unwire_86;
-void (*pmap_write_protect_p)(struct pmap *, vaddr_t, vaddr_t, vm_prot_t) =
- pmap_write_protect_86;
-void (*pmap_pinit_pd_p)(pmap_t) = pmap_pinit_pd_86;
-void (*pmap_zero_phys_p)(paddr_t) = pmap_zero_phys_86;
-boolean_t (*pmap_zero_page_uncached_p)(paddr_t) = pmap_zero_page_uncached_86;
-void (*pmap_copy_page_p)(struct vm_page *, struct vm_page *) =
- pmap_copy_page_86;
-boolean_t (*pmap_try_steal_pv_p)(struct pv_head *, struct pv_entry *,
- struct pv_entry *) = pmap_try_steal_pv_86;
-#endif /* !SMALL_KERNEL */
diff --git a/sys/arch/i386/i386/vm_machdep.c b/sys/arch/i386/i386/vm_machdep.c
index c62f3682c9d..7bc934d292f 100644
--- a/sys/arch/i386/i386/vm_machdep.c
+++ b/sys/arch/i386/i386/vm_machdep.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: vm_machdep.c,v 1.47 2006/11/29 12:26:13 miod Exp $ */
+/* $OpenBSD: vm_machdep.c,v 1.48 2007/02/20 21:15:01 tom Exp $ */
/* $NetBSD: vm_machdep.c,v 1.61 1996/05/03 19:42:35 christos Exp $ */
/*-
@@ -217,7 +217,8 @@ cpu_coredump(struct proc *p, struct vnode *vp, struct ucred *cred,
void
pagemove(caddr_t from, caddr_t to, size_t size)
{
- u_int32_t ofpte, otpte;
+ pt_entry_t *fpte, *tpte;
+ pt_entry_t ofpte, otpte;
#ifdef MULTIPROCESSOR
u_int32_t cpumask = 0;
#endif
@@ -226,12 +227,13 @@ pagemove(caddr_t from, caddr_t to, size_t size)
if ((size & PAGE_MASK) != 0)
panic("pagemove");
#endif
+ fpte = kvtopte((vaddr_t)from);
+ tpte = kvtopte((vaddr_t)to);
while (size > 0) {
- ofpte = pmap_pte_bits((vaddr_t)from);
- otpte = pmap_pte_bits((vaddr_t)to);
- pmap_pte_set((vaddr_t)to,
- pmap_pte_paddr((vaddr_t)from), ofpte);
- pmap_pte_set((vaddr_t)from, 0, 0);
+ ofpte = *fpte;
+ otpte = *tpte;
+ *tpte++ = *fpte;
+ *fpte++ = 0;
#if defined(I386_CPU) && !defined(MULTIPROCESSOR)
if (cpu_class != CPUCLASS_386)
#endif
diff --git a/sys/arch/i386/include/_types.h b/sys/arch/i386/include/_types.h
index f731aefd89f..cef2a372f29 100644
--- a/sys/arch/i386/include/_types.h
+++ b/sys/arch/i386/include/_types.h
@@ -1,4 +1,4 @@
-/* $OpenBSD: _types.h,v 1.3 2006/04/27 15:37:53 mickey Exp $ */
+/* $OpenBSD: _types.h,v 1.4 2007/02/20 21:15:01 tom Exp $ */
/*-
* Copyright (c) 1990, 1993
@@ -86,9 +86,9 @@ typedef __int32_t __register_t;
/* VM system types */
typedef unsigned long __vaddr_t;
+typedef unsigned long __paddr_t;
typedef unsigned long __vsize_t;
-typedef unsigned long long __paddr_t;
-typedef unsigned long long __psize_t;
+typedef unsigned long __psize_t;
/* Standard system types */
typedef int __clock_t;
diff --git a/sys/arch/i386/include/bus.h b/sys/arch/i386/include/bus.h
index 9900c76d4f0..5c5caf3e6c7 100644
--- a/sys/arch/i386/include/bus.h
+++ b/sys/arch/i386/include/bus.h
@@ -1,4 +1,4 @@
-/* $OpenBSD: bus.h,v 1.39 2006/04/27 15:37:53 mickey Exp $ */
+/* $OpenBSD: bus.h,v 1.40 2007/02/20 21:15:01 tom Exp $ */
/* $NetBSD: bus.h,v 1.6 1996/11/10 03:19:25 thorpej Exp $ */
/*-
@@ -741,7 +741,7 @@ void bus_space_free(bus_space_tag_t t, bus_space_handle_t bsh,
#define BUS_DMA_COHERENT 0x004 /* hint: map memory DMA coherent */
#define BUS_DMA_BUS1 0x010 /* placeholders for bus functions... */
#define BUS_DMA_BUS2 0x020
-#define BUS_DMA_64BIT 0x040 /* large memory high segment is ok */
+#define BUS_DMA_BUS3 0x040
#define BUS_DMA_24BIT 0x080 /* isadma map */
#define BUS_DMA_STREAMING 0x100 /* hint: sequential, unidirectional */
#define BUS_DMA_READ 0x200 /* mapping is device -> memory only */
@@ -771,10 +771,7 @@ typedef struct i386_bus_dmamap *bus_dmamap_t;
*/
struct i386_bus_dma_segment {
bus_addr_t ds_addr; /* DMA address */
- paddr_t ds_addr2; /* replacement store */
bus_size_t ds_len; /* length of transfer */
- vaddr_t ds_va; /* mapped loaded data */
- vaddr_t ds_va2; /* mapped replacement data */
};
typedef struct i386_bus_dma_segment bus_dma_segment_t;
@@ -866,11 +863,6 @@ struct i386_bus_dmamap {
void *_dm_cookie; /* cookie for bus-specific functions */
- struct vm_page **_dm_pages; /* replacement pages */
- vaddr_t _dm_pgva; /* those above -- mapped */
- int _dm_npages; /* number of pages allocated */
- int _dm_nused; /* number of pages replaced */
-
/*
* PUBLIC MEMBERS: these are used by machine-independent code.
*/
diff --git a/sys/arch/i386/include/cpu.h b/sys/arch/i386/include/cpu.h
index b9ac96eca8a..153e6f2c608 100644
--- a/sys/arch/i386/include/cpu.h
+++ b/sys/arch/i386/include/cpu.h
@@ -1,4 +1,4 @@
-/* $OpenBSD: cpu.h,v 1.86 2006/12/24 20:30:35 miod Exp $ */
+/* $OpenBSD: cpu.h,v 1.87 2007/02/20 21:15:01 tom Exp $ */
/* $NetBSD: cpu.h,v 1.35 1996/05/05 19:29:26 christos Exp $ */
/*-
@@ -420,10 +420,6 @@ int kvtop(caddr_t);
void vm86_gpfault(struct proc *, int);
#endif /* VM86 */
-#ifndef SMALL_KERNEL
-int cpu_paenable(void *);
-#endif /* !SMALL_KERNEL */
-
#ifdef GENERIC
/* swapgeneric.c */
void setconf(void);
diff --git a/sys/arch/i386/include/loadfile_machdep.h b/sys/arch/i386/include/loadfile_machdep.h
index 5903231fc58..4990ae48bbd 100644
--- a/sys/arch/i386/include/loadfile_machdep.h
+++ b/sys/arch/i386/include/loadfile_machdep.h
@@ -1,4 +1,4 @@
-/* $OpenBSD: loadfile_machdep.h,v 1.2 2006/04/27 15:37:53 mickey Exp $ */
+/* $OpenBSD: loadfile_machdep.h,v 1.3 2007/02/20 21:15:01 tom Exp $ */
/* $NetBSD: loadfile_machdep.h,v 1.1 1999/04/29 03:17:12 tsubai Exp $ */
/*-
@@ -43,7 +43,7 @@
#define LOAD_KERNEL (LOAD_ALL & ~LOAD_TEXTA)
#define COUNT_KERNEL (COUNT_ALL & ~COUNT_TEXTA)
-#define LOADADDR(a) (((u_long)(a) + (u_long)offset)&0xfffffff)
+#define LOADADDR(a) ((((u_long)(a)) + offset)&0xfffffff)
#define ALIGNENTRY(a) ((u_long)(a))
#define READ(f, b, c) read((f), (void *)LOADADDR(b), (c))
#define BCOPY(s, d, c) memcpy((void *)LOADADDR(d), (void *)(s), (c))
diff --git a/sys/arch/i386/include/param.h b/sys/arch/i386/include/param.h
index cc288c2239b..161973c6f36 100644
--- a/sys/arch/i386/include/param.h
+++ b/sys/arch/i386/include/param.h
@@ -1,4 +1,4 @@
-/* $OpenBSD: param.h,v 1.38 2006/11/30 21:20:41 dim Exp $ */
+/* $OpenBSD: param.h,v 1.39 2007/02/20 21:15:01 tom Exp $ */
/* $NetBSD: param.h,v 1.29 1996/03/04 05:04:26 cgd Exp $ */
/*-
@@ -75,6 +75,8 @@
#define PAGE_SIZE (1 << PAGE_SHIFT)
#define PAGE_MASK (PAGE_SIZE - 1)
+#define NPTEPG (NBPG/(sizeof (pt_entry_t)))
+
/*
* Start of kernel virtual space. Remember to alter the memory and
* page table layout description in pmap.h when changing this.
@@ -129,3 +131,9 @@
/* bytes to disk blocks */
#define dbtob(x) ((x) << DEV_BSHIFT)
#define btodb(x) ((x) >> DEV_BSHIFT)
+
+/*
+ * Mach derived conversion macros
+ */
+#define i386_round_pdr(x) ((((unsigned)(x)) + PDOFSET) & ~PDOFSET)
+#define i386_trunc_pdr(x) ((unsigned)(x) & ~PDOFSET)
diff --git a/sys/arch/i386/include/pmap.h b/sys/arch/i386/include/pmap.h
index b1e1c2ec4b8..405edfe6ada 100644
--- a/sys/arch/i386/include/pmap.h
+++ b/sys/arch/i386/include/pmap.h
@@ -1,4 +1,4 @@
-/* $OpenBSD: pmap.h,v 1.42 2006/04/27 15:37:53 mickey Exp $ */
+/* $OpenBSD: pmap.h,v 1.43 2007/02/20 21:15:01 tom Exp $ */
/* $NetBSD: pmap.h,v 1.44 2000/04/24 17:18:18 thorpej Exp $ */
/*
@@ -47,11 +47,123 @@
#include <uvm/uvm_object.h>
/*
- * The following defines identify the slots used as described in pmap.c .
+ * See pte.h for a description of i386 MMU terminology and hardware
+ * interface.
+ *
+ * A pmap describes a process' 4GB virtual address space. This
+ * virtual address space can be broken up into 1024 4MB regions which
+ * are described by PDEs in the PDP. The PDEs are defined as follows:
+ *
+ * Ranges are inclusive -> exclusive, just like vm_map_entry start/end.
+ * The following assumes that KERNBASE is 0xd0000000.
+ *
+ * PDE#s VA range Usage
+ * 0->831 0x0 -> 0xcfc00000 user address space, note that the
+ * max user address is 0xcfbfe000
+ * the final two pages in the last 4MB
+ * used to be reserved for the UAREA
+ * but now are no longer used.
+ * 831 0xcfc00000-> recursive mapping of PDP (used for
+ * 0xd0000000 linear mapping of PTPs).
+ * 832->1023 0xd0000000-> kernel address space (constant
+ * 0xffc00000 across all pmaps/processes).
+ * 1023 0xffc00000-> "alternate" recursive PDP mapping
+ * <end> (for other pmaps).
+ *
+ *
+ * Note: A recursive PDP mapping provides a way to map all the PTEs for
+ * a 4GB address space into a linear chunk of virtual memory. In other
+ * words, the PTE for page 0 is the first int mapped into the 4MB recursive
+ * area. The PTE for page 1 is the second int. The very last int in the
+ * 4MB range is the PTE that maps VA 0xffffe000 (the last page in a 4GB
+ * address).
+ *
+ * All pmaps' PDs must have the same values in slots 832->1023 so that
+ * the kernel is always mapped in every process. These values are loaded
+ * into the PD at pmap creation time.
+ *
+ * At any one time only one pmap can be active on a processor. This is
+ * the pmap whose PDP is pointed to by processor register %cr3. This pmap
+ * will have all its PTEs mapped into memory at the recursive mapping
+ * point (slot #831 as show above). When the pmap code wants to find the
+ * PTE for a virtual address, all it has to do is the following:
+ *
+ * Address of PTE = (831 * 4MB) + (VA / NBPG) * sizeof(pt_entry_t)
+ * = 0xcfc00000 + (VA / 4096) * 4
+ *
+ * What happens if the pmap layer is asked to perform an operation
+ * on a pmap that is not the one which is currently active? In that
+ * case we take the PA of the PDP of non-active pmap and put it in
+ * slot 1023 of the active pmap. This causes the non-active pmap's
+ * PTEs to get mapped in the final 4MB of the 4GB address space
+ * (e.g. starting at 0xffc00000).
+ *
+ * The following figure shows the effects of the recursive PDP mapping:
+ *
+ * PDP (%cr3)
+ * +----+
+ * | 0| -> PTP#0 that maps VA 0x0 -> 0x400000
+ * | |
+ * | |
+ * | 831| -> points back to PDP (%cr3) mapping VA 0xcfc00000 -> 0xd0000000
+ * | 832| -> first kernel PTP (maps 0xd0000000 -> 0xe0400000)
+ * | |
+ * |1023| -> points to alternate pmap's PDP (maps 0xffc00000 -> end)
+ * +----+
+ *
+ * Note that the PDE#831 VA (0xcfc00000) is defined as "PTE_BASE".
+ * Note that the PDE#1023 VA (0xffc00000) is defined as "APTE_BASE".
+ *
+ * Starting at VA 0xcfc00000 the current active PDP (%cr3) acts as a
+ * PTP:
+ *
+ * PTP#831 == PDP(%cr3) => maps VA 0xcfc00000 -> 0xd0000000
+ * +----+
+ * | 0| -> maps the contents of PTP#0 at VA 0xcfc00000->0xcfc01000
+ * | |
+ * | |
+ * | 831| -> maps the contents of PTP#831 (the PDP) at VA 0xcff3f000
+ * | 832| -> maps the contents of first kernel PTP
+ * | |
+ * |1023|
+ * +----+
+ *
+ * Note that mapping of the PDP at PTP#831's VA (0xcff3f000) is
+ * defined as "PDP_BASE".... within that mapping there are two
+ * defines:
+ * "PDP_PDE" (0xcff3fcfc) is the VA of the PDE in the PDP
+ * which points back to itself.
+ * "APDP_PDE" (0xcff3fffc) is the VA of the PDE in the PDP which
+ * establishes the recursive mapping of the alternate pmap.
+ * To set the alternate PDP, one just has to put the correct
+ * PA info in *APDP_PDE.
+ *
+ * Note that in the APTE_BASE space, the APDP appears at VA
+ * "APDP_BASE" (0xfffff000).
*/
-#define PDSLOT_PTE ((KERNBASE/NBPD)-2) /* 830: for recursive PDP map */
-#define PDSLOT_KERN (KERNBASE/NBPD) /* 832: start of kernel space */
-#define PDSLOT_APTE ((unsigned)1022) /* 1022: alternative recursive slot */
+
+/*
+ * The following defines identify the slots used as described above.
+ */
+
+#define PDSLOT_PTE ((KERNBASE/NBPD)-1) /* 831: for recursive PDP map */
+#define PDSLOT_KERN (KERNBASE/NBPD) /* 832: start of kernel space */
+#define PDSLOT_APTE ((unsigned)1023) /* 1023: alternative recursive slot */
+
+/*
+ * The following defines give the virtual addresses of various MMU
+ * data structures:
+ * PTE_BASE and APTE_BASE: the base VA of the linear PTE mappings
+ * PTD_BASE and APTD_BASE: the base VA of the recursive mapping of the PTD
+ * PDP_PDE and APDP_PDE: the VA of the PDE that points back to the PDP/APDP
+ */
+
+#define PTE_BASE ((pt_entry_t *) (PDSLOT_PTE * NBPD) )
+#define APTE_BASE ((pt_entry_t *) (PDSLOT_APTE * NBPD) )
+#define PDP_BASE ((pd_entry_t *)(((char *)PTE_BASE) + (PDSLOT_PTE * NBPG)))
+#define APDP_BASE ((pd_entry_t *)(((char *)APTE_BASE) + (PDSLOT_APTE * NBPG)))
+#define PDP_PDE (PDP_BASE + PDSLOT_PTE)
+#define APDP_PDE (PDP_BASE + PDSLOT_APTE)
/*
* The following define determines how many PTPs should be set up for the
@@ -59,10 +171,55 @@
* get the VM system running. Once the VM system is running, the
* pmap module can add more PTPs to the kernel area on demand.
*/
+
#ifndef NKPTP
-#define NKPTP 8 /* 16/32MB to start */
+#define NKPTP 4 /* 16MB to start */
#endif
#define NKPTP_MIN 4 /* smallest value we allow */
+#define NKPTP_MAX (1024 - (KERNBASE/NBPD) - 1)
+ /* largest value (-1 for APTP space) */
+
+/*
+ * various address macros
+ *
+ * vtopte: return a pointer to the PTE mapping a VA
+ * kvtopte: same as above (takes a KVA, but doesn't matter with this pmap)
+ * ptetov: given a pointer to a PTE, return the VA that it maps
+ * vtophys: translate a VA to the PA mapped to it
+ *
+ * plus alternative versions of the above
+ */
+
+#define vtopte(VA) (PTE_BASE + atop(VA))
+#define kvtopte(VA) vtopte(VA)
+#define ptetov(PT) (ptoa(PT - PTE_BASE))
+#define vtophys(VA) ((*vtopte(VA) & PG_FRAME) | \
+ ((unsigned)(VA) & ~PG_FRAME))
+#define avtopte(VA) (APTE_BASE + atop(VA))
+#define ptetoav(PT) (ptoa(PT - APTE_BASE))
+#define avtophys(VA) ((*avtopte(VA) & PG_FRAME) | \
+ ((unsigned)(VA) & ~PG_FRAME))
+
+/*
+ * pdei/ptei: generate index into PDP/PTP from a VA
+ */
+#define pdei(VA) (((VA) & PD_MASK) >> PDSHIFT)
+#define ptei(VA) (((VA) & PT_MASK) >> PGSHIFT)
+
+/*
+ * PTP macros:
+ * A PTP's index is the PD index of the PDE that points to it.
+ * A PTP's offset is the byte-offset in the PTE space that this PTP is at.
+ * A PTP's VA is the first VA mapped by that PTP.
+ *
+ * Note that NBPG == number of bytes in a PTP (4096 bytes == 1024 entries)
+ * NBPD == number of bytes a PTP can map (4MB)
+ */
+
+#define ptp_i2o(I) ((I) * NBPG) /* index => offset */
+#define ptp_o2i(O) ((O) / NBPG) /* offset => index */
+#define ptp_i2v(I) ((I) * NBPD) /* index => VA */
+#define ptp_v2i(V) ((V) / NBPD) /* VA => index (same as pdei) */
/*
* PG_AVAIL usage: we make use of the ignored bits of the PTE
@@ -72,6 +229,12 @@
#define PG_PVLIST PG_AVAIL2 /* mapping has entry on pvlist */
#define PG_X PG_AVAIL3 /* executable mapping */
+/*
+ * Number of PTE's per cache line. 4 byte pte, 32-byte cache line
+ * Used to avoid false sharing of cache lines.
+ */
+#define NPTECL 8
+
#ifdef _KERNEL
/*
* pmap data structures: see pmap.c for details of locking.
@@ -94,15 +257,13 @@ LIST_HEAD(pmap_head, pmap); /* struct pmap_head: head of a pmap list */
*/
struct pmap {
- paddr_t pm_pdidx[4]; /* PDIEs for PAE mode */
- paddr_t pm_pdirpa; /* PA of PD (read-only after create) */
- vaddr_t pm_pdir; /* VA of PD (lck by object lock) */
- int pm_pdirsize; /* PD size (4k vs 16k on pae */
struct uvm_object pm_obj; /* object (lck by object lock) */
#define pm_lock pm_obj.vmobjlock
LIST_ENTRY(pmap) pm_list; /* list (lck by pm_list lock) */
+ pd_entry_t *pm_pdir; /* VA of PD (lck by object lock) */
+ paddr_t pm_pdirpa; /* PA of PD (read-only after create) */
struct vm_page *pm_ptphint; /* pointer to a PTP in our pmap */
- struct pmap_statistics pm_stats;/* pmap stats (lck by object lock) */
+ struct pmap_statistics pm_stats; /* pmap stats (lck by object lock) */
vaddr_t pm_hiexec; /* highest executable mapping */
int pm_flags; /* see below */
@@ -172,185 +333,67 @@ struct pv_page {
/*
* global kernel variables
*/
-extern char PTD[];
+
+extern pd_entry_t PTD[];
+
+/* PTDpaddr: is the physical address of the kernel's PDP */
+extern u_int32_t PTDpaddr;
+
extern struct pmap kernel_pmap_store; /* kernel pmap */
-extern int nkptp_max;
+extern int nkpde; /* current # of PDEs for kernel */
+extern int pmap_pg_g; /* do we support PG_G? */
/*
- * Our dual-pmap design requires to play a pointer-and-seek.
- * Although being nice folks we are handle single-pmap kernels special.
+ * Macros
*/
-#define PMAP_EXCLUDE_DECLS /* tells uvm_pmap.h *not* to include decls */
-/*
- * Dumb macros
- */
#define pmap_kernel() (&kernel_pmap_store)
#define pmap_resident_count(pmap) ((pmap)->pm_stats.resident_count)
#define pmap_update(pm) /* nada */
-#define pmap_clear_modify(pg) pmap_change_attrs(pg, 0, PG_M)
-#define pmap_clear_reference(pg) pmap_change_attrs(pg, 0, PG_U)
-#define pmap_copy(DP,SP,D,L,S) /* nicht */
-#define pmap_is_modified(pg) pmap_test_attrs(pg, PG_M)
-#define pmap_is_referenced(pg) pmap_test_attrs(pg, PG_U)
-#define pmap_phys_address(ppn) ptoa(ppn)
-#define pmap_valid_entry(E) ((E) & PG_V) /* is PDE or PTE valid? */
+#define pmap_clear_modify(pg) pmap_change_attrs(pg, 0, PG_M)
+#define pmap_clear_reference(pg) pmap_change_attrs(pg, 0, PG_U)
+#define pmap_copy(DP,SP,D,L,S)
+#define pmap_is_modified(pg) pmap_test_attrs(pg, PG_M)
+#define pmap_is_referenced(pg) pmap_test_attrs(pg, PG_U)
+#define pmap_phys_address(ppn) ptoa(ppn)
+#define pmap_valid_entry(E) ((E) & PG_V) /* is PDE or PTE valid? */
+
+#define pmap_proc_iflush(p,va,len) /* nothing */
+#define pmap_unuse_final(p) /* nothing */
-#define pmap_proc_iflush(p,va,len) /* nothing */
-#define pmap_unuse_final(p) /* 4anaEB u nycToTa */
/*
* Prototypes
*/
+
void pmap_bootstrap(vaddr_t);
-void pmap_bootstrap_pae(void);
-void pmap_virtual_space(vaddr_t *, vaddr_t *);
-void pmap_init(void);
-struct pmap * pmap_create(void);
-void pmap_destroy(struct pmap *);
-void pmap_reference(struct pmap *);
-void pmap_fork(struct pmap *, struct pmap *);
-void pmap_collect(struct pmap *);
-void pmap_activate(struct proc *);
-void pmap_deactivate(struct proc *);
-void pmap_kenter_pa(vaddr_t, paddr_t, vm_prot_t);
-void pmap_kremove(vaddr_t, vsize_t);
-void pmap_zero_page(struct vm_page *);
-void pmap_copy_page(struct vm_page *, struct vm_page *);
-
-struct pv_entry*pmap_alloc_pv(struct pmap *, int);
-void pmap_enter_pv(struct pv_head *, struct pv_entry *,
- struct pmap *, vaddr_t, struct vm_page *);
-void pmap_free_pv(struct pmap *, struct pv_entry *);
-void pmap_free_pvs(struct pmap *, struct pv_entry *);
-void pmap_free_pv_doit(struct pv_entry *);
-void pmap_free_pvpage(void);
+boolean_t pmap_change_attrs(struct vm_page *, int, int);
static void pmap_page_protect(struct vm_page *, vm_prot_t);
-static void pmap_protect(struct pmap *, vaddr_t, vaddr_t, vm_prot_t);
+void pmap_page_remove(struct vm_page *);
+static void pmap_protect(struct pmap *, vaddr_t,
+ vaddr_t, vm_prot_t);
+void pmap_remove(struct pmap *, vaddr_t, vaddr_t);
+boolean_t pmap_test_attrs(struct vm_page *, int);
static void pmap_update_pg(vaddr_t);
-static void pmap_update_2pg(vaddr_t, vaddr_t);
+static void pmap_update_2pg(vaddr_t,vaddr_t);
+void pmap_write_protect(struct pmap *, vaddr_t,
+ vaddr_t, vm_prot_t);
int pmap_exec_fixup(struct vm_map *, struct trapframe *,
struct pcb *);
-void pmap_exec_account(struct pmap *, vaddr_t, u_int32_t,
- u_int32_t);
vaddr_t reserve_dumppages(vaddr_t); /* XXX: not a pmap fn */
-paddr_t vtophys(vaddr_t va);
-void pmap_tlb_shootdown(pmap_t, vaddr_t, u_int32_t, int32_t *);
+void pmap_tlb_shootdown(pmap_t, vaddr_t, pt_entry_t, int32_t *);
void pmap_tlb_shootnow(int32_t);
void pmap_do_tlb_shootdown(struct cpu_info *);
-boolean_t pmap_is_curpmap(struct pmap *);
-boolean_t pmap_is_active(struct pmap *, int);
-void pmap_apte_flush(struct pmap *);
-struct pv_entry *pmap_remove_pv(struct pv_head *, struct pmap *, vaddr_t);
-
-#ifdef SMALL_KERNEL
-#define pmap_pte_set_86 pmap_pte_set
-#define pmap_pte_setbits_86 pmap_pte_setbits
-#define pmap_pte_bits_86 pmap_pte_bits
-#define pmap_pte_paddr_86 pmap_pte_paddr
-#define pmap_change_attrs_86 pmap_change_attrs
-#define pmap_enter_86 pmap_enter
-#define pmap_extract_86 pmap_extract
-#define pmap_growkernel_86 pmap_growkernel
-#define pmap_page_remove_86 pmap_page_remove
-#define pmap_remove_86 pmap_remove
-#define pmap_test_attrs_86 pmap_test_attrs
-#define pmap_unwire_86 pmap_unwire
-#define pmap_write_protect_86 pmap_write_protect
-#define pmap_pinit_pd_86 pmap_pinit_pd
-#define pmap_zero_phys_86 pmap_zero_phys
-#define pmap_zero_page_uncached_86 pmap_zero_page_uncached
-#define pmap_copy_page_86 pmap_copy_page
-#define pmap_try_steal_pv_86 pmap_try_steal_pv
-#else
-extern u_int32_t (*pmap_pte_set_p)(vaddr_t, paddr_t, u_int32_t);
-extern u_int32_t (*pmap_pte_setbits_p)(vaddr_t, u_int32_t, u_int32_t);
-extern u_int32_t (*pmap_pte_bits_p)(vaddr_t);
-extern paddr_t (*pmap_pte_paddr_p)(vaddr_t);
-extern boolean_t (*pmap_change_attrs_p)(struct vm_page *, int, int);
-extern int (*pmap_enter_p)(pmap_t, vaddr_t, paddr_t, vm_prot_t, int);
-extern boolean_t (*pmap_extract_p)(pmap_t, vaddr_t, paddr_t *);
-extern vaddr_t (*pmap_growkernel_p)(vaddr_t);
-extern void (*pmap_page_remove_p)(struct vm_page *);
-extern void (*pmap_remove_p)(struct pmap *, vaddr_t, vaddr_t);
-extern boolean_t (*pmap_test_attrs_p)(struct vm_page *, int);
-extern void (*pmap_unwire_p)(struct pmap *, vaddr_t);
-extern void (*pmap_write_protect_p)(struct pmap*, vaddr_t, vaddr_t, vm_prot_t);
-extern void (*pmap_pinit_pd_p)(pmap_t);
-extern void (*pmap_zero_phys_p)(paddr_t);
-extern boolean_t (*pmap_zero_page_uncached_p)(paddr_t);
-extern void (*pmap_copy_page_p)(struct vm_page *, struct vm_page *);
-extern boolean_t (*pmap_try_steal_pv_p)(struct pv_head *pvh,
- struct pv_entry *cpv, struct pv_entry *prevpv);
-
-u_int32_t pmap_pte_set_pae(vaddr_t, paddr_t, u_int32_t);
-u_int32_t pmap_pte_setbits_pae(vaddr_t, u_int32_t, u_int32_t);
-u_int32_t pmap_pte_bits_pae(vaddr_t);
-paddr_t pmap_pte_paddr_pae(vaddr_t);
-boolean_t pmap_try_steal_pv_pae(struct pv_head *pvh, struct pv_entry *cpv,
- struct pv_entry *prevpv);
-boolean_t pmap_change_attrs_pae(struct vm_page *, int, int);
-int pmap_enter_pae(pmap_t, vaddr_t, paddr_t, vm_prot_t, int);
-boolean_t pmap_extract_pae(pmap_t, vaddr_t, paddr_t *);
-vaddr_t pmap_growkernel_pae(vaddr_t);
-void pmap_page_remove_pae(struct vm_page *);
-void pmap_remove_pae(struct pmap *, vaddr_t, vaddr_t);
-boolean_t pmap_test_attrs_pae(struct vm_page *, int);
-void pmap_unwire_pae(struct pmap *, vaddr_t);
-void pmap_write_protect_pae(struct pmap *, vaddr_t, vaddr_t, vm_prot_t);
-void pmap_pinit_pd_pae(pmap_t);
-void pmap_zero_phys_pae(paddr_t);
-boolean_t pmap_zero_page_uncached_pae(paddr_t);
-void pmap_copy_page_pae(struct vm_page *, struct vm_page *);
-
-#define pmap_pte_set (*pmap_pte_set_p)
-#define pmap_pte_setbits (*pmap_pte_setbits_p)
-#define pmap_pte_bits (*pmap_pte_bits_p)
-#define pmap_pte_paddr (*pmap_pte_paddr_p)
-#define pmap_change_attrs (*pmap_change_attrs_p)
-#define pmap_enter (*pmap_enter_p)
-#define pmap_extract (*pmap_extract_p)
-#define pmap_growkernel (*pmap_growkernel_p)
-#define pmap_page_remove (*pmap_page_remove_p)
-#define pmap_remove (*pmap_remove_p)
-#define pmap_test_attrs (*pmap_test_attrs_p)
-#define pmap_unwire (*pmap_unwire_p)
-#define pmap_write_protect (*pmap_write_protect_p)
-#define pmap_pinit_pd (*pmap_pinit_pd_p)
-#define pmap_zero_phys (*pmap_zero_phys_p)
-#define pmap_zero_page_uncached (*pmap_zero_page_uncached_p)
-#define pmap_copy_page (*pmap_copy_page_p)
-#define pmap_try_steal_pv (*pmap_try_steal_pv_p)
-#endif
-
-u_int32_t pmap_pte_set_86(vaddr_t, paddr_t, u_int32_t);
-u_int32_t pmap_pte_setbits_86(vaddr_t, u_int32_t, u_int32_t);
-u_int32_t pmap_pte_bits_86(vaddr_t);
-paddr_t pmap_pte_paddr_86(vaddr_t);
-boolean_t pmap_try_steal_pv_86(struct pv_head *pvh, struct pv_entry *cpv,
- struct pv_entry *prevpv);
-boolean_t pmap_change_attrs_86(struct vm_page *, int, int);
-int pmap_enter_86(pmap_t, vaddr_t, paddr_t, vm_prot_t, int);
-boolean_t pmap_extract_86(pmap_t, vaddr_t, paddr_t *);
-vaddr_t pmap_growkernel_86(vaddr_t);
-void pmap_page_remove_86(struct vm_page *);
-void pmap_remove_86(struct pmap *, vaddr_t, vaddr_t);
-boolean_t pmap_test_attrs_86(struct vm_page *, int);
-void pmap_unwire_86(struct pmap *, vaddr_t);
-void pmap_write_protect_86(struct pmap *, vaddr_t, vaddr_t, vm_prot_t);
-void pmap_pinit_pd_86(pmap_t);
-void pmap_zero_phys_86(paddr_t);
-boolean_t pmap_zero_page_uncached_86(paddr_t);
-void pmap_copy_page_86(struct vm_page *, struct vm_page *);
#define PMAP_GROWKERNEL /* turn on pmap_growkernel interface */
/*
* Do idle page zero'ing uncached to avoid polluting the cache.
*/
+boolean_t pmap_zero_page_uncached(paddr_t);
#define PMAP_PAGEIDLEZERO(pg) pmap_zero_page_uncached(VM_PAGE_TO_PHYS(pg))
/*
diff --git a/sys/arch/i386/include/pte.h b/sys/arch/i386/include/pte.h
index 73a3bc3e7b0..a3b3542824d 100644
--- a/sys/arch/i386/include/pte.h
+++ b/sys/arch/i386/include/pte.h
@@ -1,4 +1,4 @@
-/* $OpenBSD: pte.h,v 1.8 2006/04/27 15:37:53 mickey Exp $ */
+/* $OpenBSD: pte.h,v 1.9 2007/02/20 21:15:01 tom Exp $ */
/* $NetBSD: pte.h,v 1.11 1998/02/06 21:58:05 thorpej Exp $ */
/*
@@ -45,11 +45,114 @@
#define _I386_PTE_H_
/*
+ * i386 MMU hardware structure:
+ *
+ * the i386 MMU is a two-level MMU which maps 4GB of virtual memory.
+ * the pagesize is 4K (4096 [0x1000] bytes), although newer pentium
+ * processors can support a 4MB pagesize as well.
+ *
+ * the first level table (segment table?) is called a "page directory"
+ * and it contains 1024 page directory entries (PDEs). each PDE is
+ * 4 bytes (an int), so a PD fits in a single 4K page. this page is
+ * the page directory page (PDP). each PDE in a PDP maps 4MB of space
+ * (1024 * 4MB = 4GB). a PDE contains the physical address of the
+ * second level table: the page table. or, if 4MB pages are being used,
+ * then the PDE contains the PA of the 4MB page being mapped.
+ *
+ * a page table consists of 1024 page table entries (PTEs). each PTE is
+ * 4 bytes (an int), so a page table also fits in a single 4K page. a
+ * 4K page being used as a page table is called a page table page (PTP).
+ * each PTE in a PTP maps one 4K page (1024 * 4K = 4MB). a PTE contains
+ * the physical address of the page it maps and some flag bits (described
+ * below).
+ *
+ * the processor has a special register, "cr3", which points to the
+ * the PDP which is currently controlling the mappings of the virtual
+ * address space.
+ *
+ * the following picture shows the translation process for a 4K page:
+ *
+ * %cr3 register [PA of PDP]
+ * |
+ * |
+ * | bits <31-22> of VA bits <21-12> of VA bits <11-0>
+ * | index the PDP (0 - 1023) index the PTP are the page offset
+ * | | | |
+ * | v | |
+ * +--->+----------+ | |
+ * | PD Page | PA of v |
+ * | |---PTP-------->+------------+ |
+ * | 1024 PDE | | page table |--PTE--+ |
+ * | entries | | (aka PTP) | | |
+ * +----------+ | 1024 PTE | | |
+ * | entries | | |
+ * +------------+ | |
+ * | |
+ * bits <31-12> bits <11-0>
+ * p h y s i c a l a d d r
+ *
+ * the i386 caches PTEs in a TLB. it is important to flush out old
+ * TLB mappings when making a change to a mappings. writing to the
+ * %cr3 will flush the entire TLB. newer processors also have an
+ * instruction that will invalidate the mapping of a single page (which
+ * is useful if you are changing a single mappings because it preserves
+ * all the cached TLB entries).
+ *
+ * as shows, bits 31-12 of the PTE contain PA of the page being mapped.
+ * the rest of the PTE is defined as follows:
+ * bit# name use
+ * 11 n/a available for OS use, hardware ignores it
+ * 10 n/a available for OS use, hardware ignores it
+ * 9 n/a available for OS use, hardware ignores it
+ * 8 G global bit (see discussion below)
+ * 7 PS page size [for PDEs] (0=4k, 1=4M <if supported>)
+ * 6 D dirty (modified) page
+ * 5 A accessed (referenced) page
+ * 4 PCD cache disable
+ * 3 PWT prevent write through (cache)
+ * 2 U/S user/supervisor bit (0=supervisor only, 1=both u&s)
+ * 1 R/W read/write bit (0=read only, 1=read-write)
+ * 0 P present (valid)
+ *
+ * notes:
+ * - on the i386 the R/W bit is ignored if processor is in supervisor
+ * state (bug!)
+ * - PS is only supported on newer processors
+ * - PTEs with the G bit are global in the sense that they are not
+ * flushed from the TLB when %cr3 is written (to flush, use the
+ * "flush single page" instruction). this is only supported on
+ * newer processors. this bit can be used to keep the kernel's
+ * TLB entries around while context switching. since the kernel
+ * is mapped into all processes at the same place it does not make
+ * sense to flush these entries when switching from one process'
+ * pmap to another.
+ */
+
+#if !defined(_LOCORE)
+
+/*
+ * here we define the data types for PDEs and PTEs
+ */
+
+typedef u_int32_t pd_entry_t; /* PDE */
+typedef u_int32_t pt_entry_t; /* PTE */
+
+#endif
+
+/*
* now we define various for playing with virtual addresses
*/
#define PDSHIFT 22 /* offset of PD index in VA */
#define NBPD (1 << PDSHIFT) /* # bytes mapped by PD (4MB) */
+#define PDOFSET (NBPD-1) /* mask for non-PD part of VA */
+#if 0 /* not used? */
+#define NPTEPD (NBPD / NBPG) /* # of PTEs in a PD */
+#else
+#define PTES_PER_PTP (NBPD / NBPG) /* # of PTEs in a PTP */
+#endif
+#define PD_MASK 0xffc00000 /* page directory address bits */
+#define PT_MASK 0x003ff000 /* page table address bits */
/*
* here we define the bits of the PDE/PTE, as described above:
@@ -70,6 +173,8 @@
#define PG_AVAIL1 0x00000200 /* ignored by hardware */
#define PG_AVAIL2 0x00000400 /* ignored by hardware */
#define PG_AVAIL3 0x00000800 /* ignored by hardware */
+#define PG_FRAME 0xfffff000 /* page frame mask */
+#define PG_LGFRAME 0xffc00000 /* large (4M) page frame mask */
/*
* various short-hand protection codes
diff --git a/sys/arch/i386/include/vmparam.h b/sys/arch/i386/include/vmparam.h
index 6174c378725..8d7dd4e1cf6 100644
--- a/sys/arch/i386/include/vmparam.h
+++ b/sys/arch/i386/include/vmparam.h
@@ -1,4 +1,4 @@
-/* $OpenBSD: vmparam.h,v 1.35 2006/04/27 15:37:53 mickey Exp $ */
+/* $OpenBSD: vmparam.h,v 1.36 2007/02/20 21:15:01 tom Exp $ */
/* $NetBSD: vmparam.h,v 1.15 1994/10/27 04:16:34 cgd Exp $ */
/*-
@@ -91,23 +91,22 @@
/* user/kernel map constants */
#define VM_MIN_ADDRESS ((vaddr_t)0)
-#define VM_MAXUSER_ADDRESS ((vaddr_t)0xcf800000)
-#define VM_MAX_ADDRESS (vm_max_address)
-extern vaddr_t vm_max_address;
+#define VM_MAXUSER_ADDRESS ((vaddr_t)((PDSLOT_PTE<<PDSHIFT) - USPACE))
+#define VM_MAX_ADDRESS ((vaddr_t)((PDSLOT_PTE<<PDSHIFT) + \
+ (PDSLOT_PTE<<PGSHIFT)))
#define VM_MIN_KERNEL_ADDRESS ((vaddr_t)KERNBASE)
-#define VM_MAX_KERNEL_ADDRESS ((vaddr_t)0xff800000)
+#define VM_MAX_KERNEL_ADDRESS ((vaddr_t)(PDSLOT_APTE<<PDSHIFT))
/* virtual sizes (bytes) for various kernel submaps */
#define VM_PHYS_SIZE (USRIOSIZE*PAGE_SIZE)
-#define VM_PHYSSEG_MAX 8 /* actually we could have this many segments */
+#define VM_PHYSSEG_MAX 5 /* actually we could have this many segments */
#define VM_PHYSSEG_STRAT VM_PSTRAT_BSEARCH
#define VM_PHYSSEG_NOADD /* can't add RAM after vm_mem_init */
-#define VM_NFREELIST 3
+#define VM_NFREELIST 2
#define VM_FREELIST_DEFAULT 0
#define VM_FREELIST_FIRST16 1
-#define VM_FREELIST_ABOVE4G 2
/*
* pmap specific data stored in the vm_physmem[] array
diff --git a/sys/arch/i386/pci/pci_addr_fixup.c b/sys/arch/i386/pci/pci_addr_fixup.c
index 6f12fe5618a..294454e7a1a 100644
--- a/sys/arch/i386/pci/pci_addr_fixup.c
+++ b/sys/arch/i386/pci/pci_addr_fixup.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: pci_addr_fixup.c,v 1.20 2006/09/19 11:06:34 jsg Exp $ */
+/* $OpenBSD: pci_addr_fixup.c,v 1.21 2007/02/20 21:15:01 tom Exp $ */
/* $NetBSD: pci_addr_fixup.c,v 1.7 2000/08/03 20:10:45 nathanw Exp $ */
/*-
@@ -45,22 +45,24 @@
#include <i386/pci/pcibiosvar.h>
typedef int (*pciaddr_resource_manage_func_t)(struct pcibios_softc *, pci_chipset_tag_t, pcitag_t, int,
- struct extent *, int, u_long *, bus_size_t);
+ struct extent *, int, bus_addr_t *, bus_size_t);
void pciaddr_resource_manage(struct pcibios_softc *,
pci_chipset_tag_t, pcitag_t, pciaddr_resource_manage_func_t);
void pciaddr_resource_reserve(struct pcibios_softc *,
pci_chipset_tag_t, pcitag_t);
void pciaddr_resource_reserve_disabled(struct pcibios_softc *,
pci_chipset_tag_t, pcitag_t);
-int pciaddr_do_resource_reserve(struct pcibios_softc *, pci_chipset_tag_t,
- pcitag_t, int, struct extent *, int, u_long *, bus_size_t);
+int pciaddr_do_resource_reserve(struct pcibios_softc *,
+ pci_chipset_tag_t, pcitag_t, int, struct extent *, int,
+ bus_addr_t *, bus_size_t);
int pciaddr_do_resource_reserve_disabled(struct pcibios_softc *,
pci_chipset_tag_t, pcitag_t, int, struct extent *, int, u_long *,
bus_size_t);
void pciaddr_resource_allocate(struct pcibios_softc *,
pci_chipset_tag_t, pcitag_t);
-int pciaddr_do_resource_allocate(struct pcibios_softc *, pci_chipset_tag_t,
- pcitag_t, int, struct extent *, int, u_long *, bus_size_t);
+int pciaddr_do_resource_allocate(struct pcibios_softc *,
+ pci_chipset_tag_t, pcitag_t, int, struct extent *, int, bus_addr_t *,
+ bus_size_t);
bus_addr_t pciaddr_ioaddr(u_int32_t);
void pciaddr_print_devid(pci_chipset_tag_t, pcitag_t);
@@ -181,7 +183,7 @@ pciaddr_resource_manage(struct pcibios_softc *sc, pci_chipset_tag_t pc,
{
struct extent *ex;
pcireg_t val, mask;
- u_long addr;
+ bus_addr_t addr;
bus_size_t size;
int error, mapreg, type, reg_start, reg_end, width;
@@ -260,7 +262,7 @@ pciaddr_resource_manage(struct pcibios_softc *sc, pci_chipset_tag_t pc,
int
pciaddr_do_resource_allocate(struct pcibios_softc *sc, pci_chipset_tag_t pc,
- pcitag_t tag, int mapreg, struct extent *ex, int type, u_long *addr,
+ pcitag_t tag, int mapreg, struct extent *ex, int type, bus_addr_t *addr,
bus_size_t size)
{
bus_addr_t start;
@@ -306,7 +308,7 @@ pciaddr_do_resource_allocate(struct pcibios_softc *sc, pci_chipset_tag_t pc,
int
pciaddr_do_resource_reserve(struct pcibios_softc *sc, pci_chipset_tag_t pc,
- pcitag_t tag, int mapreg, struct extent *ex, int type, u_long *addr,
+ pcitag_t tag, int mapreg, struct extent *ex, int type, bus_addr_t *addr,
bus_size_t size)
{
pcireg_t val;
diff --git a/sys/arch/i386/pci/pci_machdep.c b/sys/arch/i386/pci/pci_machdep.c
index 397f3661a41..6353b135729 100644
--- a/sys/arch/i386/pci/pci_machdep.c
+++ b/sys/arch/i386/pci/pci_machdep.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: pci_machdep.c,v 1.37 2006/12/20 18:59:50 kettenis Exp $ */
+/* $OpenBSD: pci_machdep.c,v 1.38 2007/02/20 21:15:01 tom Exp $ */
/* $NetBSD: pci_machdep.c,v 1.28 1997/06/06 23:29:17 thorpej Exp $ */
/*-
@@ -113,7 +113,9 @@ extern bios_pciinfo_t *bios_pciinfo;
#endif
#include "pcibios.h"
+#if NPCIBIOS > 0
#include <i386/pci/pcibiosvar.h>
+#endif
int pci_mode = -1;
diff --git a/sys/arch/i386/pci/pcibios.c b/sys/arch/i386/pci/pcibios.c
index d58c361e319..44f11ec98a5 100644
--- a/sys/arch/i386/pci/pcibios.c
+++ b/sys/arch/i386/pci/pcibios.c
@@ -1,4 +1,4 @@
-/* $OpenBSD: pcibios.c,v 1.35 2006/11/29 22:40:13 miod Exp $ */
+/* $OpenBSD: pcibios.c,v 1.36 2007/02/20 21:15:01 tom Exp $ */
/* $NetBSD: pcibios.c,v 1.5 2000/08/01 05:23:59 uch Exp $ */
/*
@@ -266,7 +266,7 @@ pcibios_pir_init(struct pcibios_softc *sc)
for (i = 0; i < pirh->tablesize; i++)
cksum += p[i];
- printf("%s: PCI IRQ Routing Table rev %d.%d @ 0x%llx/%d "
+ printf("%s: PCI IRQ Routing Table rev %d.%d @ 0x%lx/%d "
"(%d entries)\n", sc->sc_dev.dv_xname,
pirh->version >> 8, pirh->version & 0xff, pa,
pirh->tablesize, (pirh->tablesize - sizeof(*pirh)) / 16);