/* $OpenBSD: pmap.h,v 1.16 2004/05/20 09:20:41 kettenis Exp $ */ /* $NetBSD: pmap.h,v 1.37 2000/11/19 03:16:35 thorpej Exp $ */ /*- * Copyright (c) 1998, 1999, 2000 The NetBSD Foundation, Inc. * All rights reserved. * * This code is derived from software contributed to The NetBSD Foundation * by Jason R. Thorpe of the Numerical Aerospace Simulation Facility, * NASA Ames Research Center and by Chris G. Demetriou. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. All advertising materials mentioning features or use of this software * must display the following acknowledgement: * This product includes software developed by the NetBSD * Foundation, Inc. and its contributors. * 4. Neither the name of The NetBSD Foundation nor the names of its * contributors may be used to endorse or promote products derived * from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE * POSSIBILITY OF SUCH DAMAGE. */ /* * Copyright (c) 1987 Carnegie-Mellon University * Copyright (c) 1991, 1993 * The Regents of the University of California. All rights reserved. * * This code is derived from software contributed to Berkeley by * the Systems Programming Group of the University of Utah Computer * Science Department. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. Neither the name of the University nor the names of its contributors * may be used to endorse or promote products derived from this software * without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * * @(#)pmap.h 8.1 (Berkeley) 6/10/93 */ #ifndef _PMAP_MACHINE_ #define _PMAP_MACHINE_ #include #include #include /* * Machine-dependent virtual memory state. * * If we ever support processor numbers higher than 63, we'll have to * rethink the CPU mask. * * Note pm_asn and pm_asngen are arrays allocated in pmap_create(). * Their size is based on the PCS count from the HWRPB, and indexed * by processor ID (from `whami'). * * The kernel pmap is a special case; it gets statically-allocated * arrays which hold enough for ALPHA_MAXPROCS. */ struct pmap { TAILQ_ENTRY(pmap) pm_list; /* list of all pmaps */ pt_entry_t *pm_lev1map; /* level 1 map */ int pm_count; /* pmap reference count */ struct simplelock pm_slock; /* lock on pmap */ struct pmap_statistics pm_stats; /* pmap statistics */ long pm_nlev2; /* level 2 pt page count */ long pm_nlev3; /* level 3 pt page count */ unsigned int *pm_asn; /* address space number */ unsigned long *pm_asngen; /* ASN generation number */ unsigned long pm_cpus; /* mask of CPUs using pmap */ unsigned long pm_needisync; /* mask of CPUs needing isync */ }; typedef struct pmap *pmap_t; #define PMAP_ASN_RESERVED 0 /* reserved for Lev1map users */ extern struct pmap kernel_pmap_store; #define pmap_kernel() (&kernel_pmap_store) /* * For each vm_page_t, there is a list of all currently valid virtual * mappings of that page. An entry is a pv_entry_t, the list is pv_table. */ typedef struct pv_entry { LIST_ENTRY(pv_entry) pv_list; /* pv_entry list */ struct pmap *pv_pmap; /* pmap where mapping lies */ vaddr_t pv_va; /* virtual address for mapping */ pt_entry_t *pv_pte; /* PTE that maps the VA */ } *pv_entry_t; /* * The head of the list of pv_entry_t's, also contains page attributes. */ struct pv_head { LIST_HEAD(, pv_entry) pvh_list; /* pv_entry list */ struct simplelock pvh_slock; /* lock on this head */ int pvh_attrs; /* page attributes */ int pvh_usage; /* page usage */ int pvh_refcnt; /* special use ref count */ }; /* pvh_attrs */ #define PGA_MODIFIED 0x01 /* modified */ #define PGA_REFERENCED 0x02 /* referenced */ /* pvh_usage */ #define PGU_NORMAL 0 /* free or normal use */ #define PGU_PVENT 1 /* PV entries */ #define PGU_L1PT 2 /* level 1 page table */ #define PGU_L2PT 3 /* level 2 page table */ #define PGU_L3PT 4 /* level 3 page table */ #define PGU_ISPTPAGE(pgu) ((pgu) >= PGU_L1PT) #define PGU_STRINGS \ { \ "normal", \ "pvent", \ "l1pt", \ "l2pt", \ "l3pt", \ } #ifdef _KERNEL #ifndef _LKM #if defined(NEW_SCC_DRIVER) #if defined(DEC_KN8AE) #define _PMAP_MAY_USE_PROM_CONSOLE #endif #else /* ! NEW_SCC_DRIVER */ #if defined(DEC_3000_300) \ || defined(DEC_3000_500) \ || defined(DEC_KN8AE) /* XXX */ #define _PMAP_MAY_USE_PROM_CONSOLE /* XXX */ #endif /* XXX */ #endif /* NEW_SCC_DRIVER */ #if defined(MULTIPROCESSOR) void pmap_tlb_shootdown(pmap_t, vaddr_t, pt_entry_t); void pmap_do_tlb_shootdown(void); void pmap_tlb_shootdown_q_drain(u_long, boolean_t); #define PMAP_TLB_SHOOTDOWN(pm, va, pte) \ pmap_tlb_shootdown((pm), (va), (pte)) #else #define PMAP_TLB_SHOOTDOWN(pm, va, pte) /* nothing */ #endif /* MULTIPROCESSOR */ #endif /* _LKM */ #define pmap_resident_count(pmap) ((pmap)->pm_stats.resident_count) #define pmap_wired_count(pmap) ((pmap)->pm_stats.wired_count) #define pmap_update(pmap) /* nothing (yet) */ #define pmap_proc_iflush(p, va, len) /* nothing */ extern pt_entry_t *VPT; /* Virtual Page Table */ #define PMAP_STEAL_MEMORY /* enable pmap_steal_memory() */ #define PMAP_GROWKERNEL /* enable pmap_growkernel() */ /* * Alternate mapping hooks for pool pages. Avoids thrashing the TLB. */ #define PMAP_MAP_POOLPAGE(pg) ALPHA_PHYS_TO_K0SEG(VM_PAGE_TO_PHYS(pg)) #define PMAP_UNMAP_POOLPAGE(va) PHYS_TO_VM_PAGE(ALPHA_K0SEG_TO_PHYS((va))) paddr_t vtophys(vaddr_t); /* Machine-specific functions. */ void pmap_bootstrap(paddr_t ptaddr, u_int maxasn, u_long ncpuids); int pmap_emulate_reference(struct proc *p, vaddr_t v, int user, int type); #ifdef _PMAP_MAY_USE_PROM_CONSOLE int pmap_uses_prom_console(void); #endif #define pmap_pte_pa(pte) (PG_PFNUM(*(pte)) << PGSHIFT) #define pmap_pte_prot(pte) (*(pte) & PG_PROT) #define pmap_pte_w(pte) (*(pte) & PG_WIRED) #define pmap_pte_v(pte) (*(pte) & PG_V) #define pmap_pte_pv(pte) (*(pte) & PG_PVLIST) #define pmap_pte_asm(pte) (*(pte) & PG_ASM) #define pmap_pte_exec(pte) (*(pte) & PG_EXEC) #define pmap_pte_set_w(pte, v) \ do { \ if (v) \ *(pte) |= PG_WIRED; \ else \ *(pte) &= ~PG_WIRED; \ } while (0) #define pmap_pte_w_chg(pte, nw) ((nw) ^ pmap_pte_w(pte)) #define pmap_pte_set_prot(pte, np) \ do { \ *(pte) &= ~PG_PROT; \ *(pte) |= (np); \ } while (0) #define pmap_pte_prot_chg(pte, np) ((np) ^ pmap_pte_prot(pte)) static __inline pt_entry_t *pmap_l2pte(pmap_t, vaddr_t, pt_entry_t *); static __inline pt_entry_t *pmap_l3pte(pmap_t, vaddr_t, pt_entry_t *); #define pmap_l1pte(pmap, v) \ (&(pmap)->pm_lev1map[l1pte_index((vaddr_t)(v))]) static __inline pt_entry_t * pmap_l2pte(pmap, v, l1pte) pmap_t pmap; vaddr_t v; pt_entry_t *l1pte; { pt_entry_t *lev2map; if (l1pte == NULL) { l1pte = pmap_l1pte(pmap, v); if (pmap_pte_v(l1pte) == 0) return (NULL); } lev2map = (pt_entry_t *)ALPHA_PHYS_TO_K0SEG(pmap_pte_pa(l1pte)); return (&lev2map[l2pte_index(v)]); } static __inline pt_entry_t * pmap_l3pte(pmap, v, l2pte) pmap_t pmap; vaddr_t v; pt_entry_t *l2pte; { pt_entry_t *l1pte, *lev2map, *lev3map; if (l2pte == NULL) { l1pte = pmap_l1pte(pmap, v); if (pmap_pte_v(l1pte) == 0) return (NULL); lev2map = (pt_entry_t *)ALPHA_PHYS_TO_K0SEG(pmap_pte_pa(l1pte)); l2pte = &lev2map[l2pte_index(v)]; if (pmap_pte_v(l2pte) == 0) return (NULL); } lev3map = (pt_entry_t *)ALPHA_PHYS_TO_K0SEG(pmap_pte_pa(l2pte)); return (&lev3map[l3pte_index(v)]); } /* * Macros for locking pmap structures. * * Note that we if we access the kernel pmap in interrupt context, it * is only to update statistics. Since stats are updated using atomic * operations, locking the kernel pmap is not necessary. Therefore, * it is not necessary to block interrupts when locking pmap structures. */ #define PMAP_LOCK(pmap) simple_lock(&(pmap)->pm_slock) #define PMAP_UNLOCK(pmap) simple_unlock(&(pmap)->pm_slock) /* * Macro for processing deferred I-stream synchronization. * * The pmap module may defer syncing the user I-stream until the * return to userspace, since the IMB PALcode op can be quite * expensive. Since user instructions won't be executed until * the return to userspace, this can be deferred until userret(). */ #define PMAP_USERRET(pmap) \ do { \ u_long cpu_mask = (1UL << cpu_number()); \ \ if ((pmap)->pm_needisync & cpu_mask) { \ atomic_clearbits_ulong(&(pmap)->pm_needisync, \ cpu_mask); \ alpha_pal_imb(); \ } \ } while (0) #endif /* _KERNEL */ #endif /* _PMAP_MACHINE_ */