diff options
author | Dale S. Rahn <rahnds@cvs.openbsd.org> | 2000-08-03 03:02:51 +0000 |
---|---|---|
committer | Dale S. Rahn <rahnds@cvs.openbsd.org> | 2000-08-03 03:02:51 +0000 |
commit | 512236a12d50067ded4dd94ac1aa8c45ebf12d90 (patch) | |
tree | a07f5ce87693a78cb8f67830ba37310536da70a3 /sys/arch/powerpc | |
parent | a2a41a863b468fc461b2c55e7f8cab074afa7da4 (diff) |
Fixes to bus dmamem code.
(Fixed by replacing the code with the version from the i386 port.)
Diffstat (limited to 'sys/arch/powerpc')
-rw-r--r-- | sys/arch/powerpc/include/bus.h | 54 | ||||
-rw-r--r-- | sys/arch/powerpc/powerpc/dma.c | 388 |
2 files changed, 316 insertions, 126 deletions
diff --git a/sys/arch/powerpc/include/bus.h b/sys/arch/powerpc/include/bus.h index 7e99a8785fc..cd469104d7f 100644 --- a/sys/arch/powerpc/include/bus.h +++ b/sys/arch/powerpc/include/bus.h @@ -1,4 +1,4 @@ -/* $OpenBSD: bus.h,v 1.7 2000/07/07 13:11:04 rahnds Exp $ */ +/* $OpenBSD: bus.h,v 1.8 2000/08/03 03:02:50 rahnds Exp $ */ /* * Copyright (c) 1997 Per Fogelstrom. All rights reserved. @@ -214,30 +214,30 @@ struct powerpc_bus_dma_tag { /* * DMA mapping methods. */ - int (*_dmamap_create) __P((void *, bus_size_t, int, + int (*_dmamap_create) __P((bus_dma_tag_t , bus_size_t, int, bus_size_t, bus_size_t, int, bus_dmamap_t *)); - void (*_dmamap_destroy) __P((void *, bus_dmamap_t)); - int (*_dmamap_load) __P((void *, bus_dmamap_t, void *, + void (*_dmamap_destroy) __P((bus_dma_tag_t , bus_dmamap_t)); + int (*_dmamap_load) __P((bus_dma_tag_t , bus_dmamap_t, void *, bus_size_t, struct proc *, int)); - int (*_dmamap_load_mbuf) __P((void *, bus_dmamap_t, + int (*_dmamap_load_mbuf) __P((bus_dma_tag_t , bus_dmamap_t, struct mbuf *, int)); - int (*_dmamap_load_uio) __P((void *, bus_dmamap_t, + int (*_dmamap_load_uio) __P((bus_dma_tag_t , bus_dmamap_t, struct uio *, int)); - int (*_dmamap_load_raw) __P((void *, bus_dmamap_t, + int (*_dmamap_load_raw) __P((bus_dma_tag_t , bus_dmamap_t, bus_dma_segment_t *, int, bus_size_t, int)); - void (*_dmamap_unload) __P((void *, bus_dmamap_t)); - void (*_dmamap_sync) __P((void *, bus_dmamap_t, bus_dmasync_op_t)); + void (*_dmamap_unload) __P((bus_dma_tag_t , bus_dmamap_t)); + void (*_dmamap_sync) __P((bus_dma_tag_t , bus_dmamap_t, bus_dmasync_op_t)); /* * DMA memory utility functions. */ - int (*_dmamem_alloc) __P((void *, bus_size_t, bus_size_t, + int (*_dmamem_alloc) __P((bus_dma_tag_t, bus_size_t, bus_size_t, bus_size_t, bus_dma_segment_t *, int, int *, int)); - void (*_dmamem_free) __P((void *, bus_dma_segment_t *, int)); - int (*_dmamem_map) __P((void *, bus_dma_segment_t *, + void (*_dmamem_free) __P((bus_dma_tag_t, bus_dma_segment_t *, int)); + int (*_dmamem_map) __P((bus_dma_tag_t, bus_dma_segment_t *, int, size_t, caddr_t *, int)); - void (*_dmamem_unmap) __P((void *, caddr_t, size_t)); - int (*_dmamem_mmap) __P((void *, bus_dma_segment_t *, + void (*_dmamem_unmap) __P((bus_dma_tag_t, caddr_t, size_t)); + int (*_dmamem_mmap) __P((bus_dma_tag_t, bus_dma_segment_t *, int, int, int, int)); }; @@ -270,25 +270,25 @@ struct powerpc_bus_dma_tag { #define bus_dmamem_mmap(t, sg, n, o, p, f) \ (*(t)->_dmamem_mmap)((t)->_cookie, (sg), (n), (o), (p), (f)) -int _dmamap_create __P((void *, bus_size_t, int, +int _dmamap_create __P((bus_dma_tag_t, bus_size_t, int, bus_size_t, bus_size_t, int, bus_dmamap_t *)); -void _dmamap_destroy __P((void *, bus_dmamap_t)); -int _dmamap_load __P((void *, bus_dmamap_t, void *, +void _dmamap_destroy __P((bus_dma_tag_t, bus_dmamap_t)); +int _dmamap_load __P((bus_dma_tag_t, bus_dmamap_t, void *, bus_size_t, struct proc *, int)); -int _dmamap_load_mbuf __P((void *, bus_dmamap_t, struct mbuf *, int)); -int _dmamap_load_uio __P((void *, bus_dmamap_t, struct uio *, int)); -int _dmamap_load_raw __P((void *, bus_dmamap_t, +int _dmamap_load_mbuf __P((bus_dma_tag_t, bus_dmamap_t, struct mbuf *, int)); +int _dmamap_load_uio __P((bus_dma_tag_t, bus_dmamap_t, struct uio *, int)); +int _dmamap_load_raw __P((bus_dma_tag_t, bus_dmamap_t, bus_dma_segment_t *, int, bus_size_t, int)); -void _dmamap_unload __P((void *, bus_dmamap_t)); -void _dmamap_sync __P((void *, bus_dmamap_t, bus_dmasync_op_t)); +void _dmamap_unload __P((bus_dma_tag_t, bus_dmamap_t)); +void _dmamap_sync __P((bus_dma_tag_t, bus_dmamap_t, bus_dmasync_op_t)); -int _dmamem_alloc __P((void *, bus_size_t, bus_size_t, +int _dmamem_alloc __P((bus_dma_tag_t, bus_size_t, bus_size_t, bus_size_t, bus_dma_segment_t *, int, int *, int)); -void _dmamem_free __P((void *, bus_dma_segment_t *, int)); -int _dmamem_map __P((void *, bus_dma_segment_t *, +void _dmamem_free __P((bus_dma_tag_t, bus_dma_segment_t *, int)); +int _dmamem_map __P((bus_dma_tag_t, bus_dma_segment_t *, int, size_t, caddr_t *, int)); -void _dmamem_unmap __P((void *, caddr_t, size_t)); -int _dmamem_mmap __P((void *, bus_dma_segment_t *, int, int, int, int)); +void _dmamem_unmap __P((bus_dma_tag_t, caddr_t, size_t)); +int _dmamem_mmap __P((bus_dma_tag_t, bus_dma_segment_t *, int, int, int, int)); /* * bus_dmamap_t diff --git a/sys/arch/powerpc/powerpc/dma.c b/sys/arch/powerpc/powerpc/dma.c index cb7d5c74f37..e441e55062d 100644 --- a/sys/arch/powerpc/powerpc/dma.c +++ b/sys/arch/powerpc/powerpc/dma.c @@ -1,9 +1,14 @@ -/* $OpenBSD: dma.c,v 1.2 2000/03/31 04:12:58 rahnds Exp $ */ +/* $OpenBSD: dma.c,v 1.3 2000/08/03 03:02:50 rahnds Exp $ */ +/* $NetBSD: machdep.c,v 1.214 1996/11/10 03:16:17 thorpej Exp $ */ -/* - * Copyright (c) 1998 Michael Shalayeff +/*- + * Copyright (c) 1996, 1997 The NetBSD Foundation, Inc. * All rights reserved. * + * This code is derived from software contributed to The NetBSD Foundation + * by Jason R. Thorpe of the Numerical Aerospace Simulation Facility, + * NASA Ames Research Center. + * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: @@ -14,20 +19,23 @@ * documentation and/or other materials provided with the distribution. * 3. All advertising materials mentioning features or use of this software * must display the following acknowledgement: - * This product includes software developed by Michael Shalayeff. - * 4. The name of the author may not be used to endorse or promote products - * derived from this software without specific prior written permission. + * This product includes software developed by the NetBSD + * Foundation, Inc. and its contributors. + * 4. Neither the name of The NetBSD Foundation nor the names of its + * contributors may be used to endorse or promote products derived + * from this software without specific prior written permission. * - * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR - * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES - * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. - * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, - * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT - * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, - * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY - * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT - * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF - * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS + * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED + * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR + * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS + * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS + * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN + * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) + * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE + * POSSIBILITY OF SUCH DAMAGE. */ #include <sys/param.h> @@ -53,9 +61,13 @@ #include <machine/bus.h> +/* + * Common function for DMA map creation. May be called by bus-specific + * DMA map creation functions. + */ int -_dmamap_create(v, size, nsegments, maxsegsz, boundary, flags, dmamp) - void *v; +_dmamap_create(t, size, nsegments, maxsegsz, boundary, flags, dmamp) + bus_dma_tag_t t; bus_size_t size; int nsegments; bus_size_t maxsegsz; @@ -63,32 +75,51 @@ _dmamap_create(v, size, nsegments, maxsegsz, boundary, flags, dmamp) int flags; bus_dmamap_t *dmamp; { - register struct powerpc_bus_dmamap *map; - register size_t mapsize; + struct powerpc_bus_dmamap *map; + void *mapstore; + size_t mapsize; + /* + * Allocate and initialize the DMA map. The end of the map + * is a variable-sized array of segments, so we allocate enough + * room for them in one shot. + * + * Note we don't preserve the WAITOK or NOWAIT flags. Preservation + * of ALLOCNOW notifies others that we've reserved these resources, + * and they are not to be freed. + * + * The bus_dmamap_t includes one bus_dma_segment_t, hence + * the (nsegments - 1). + */ mapsize = sizeof(struct powerpc_bus_dmamap) + (sizeof(bus_dma_segment_t) * (nsegments - 1)); - MALLOC(map, struct powerpc_bus_dmamap *, mapsize, M_DEVBUF, - (flags & BUS_DMA_NOWAIT) ? M_NOWAIT : M_WAITOK); - if (!map) + if ((mapstore = malloc(mapsize, M_DEVBUF, + (flags & BUS_DMA_NOWAIT) ? M_NOWAIT : M_WAITOK)) == NULL) return (ENOMEM); - bzero(map, mapsize); + bzero(mapstore, mapsize); + map = (struct powerpc_bus_dmamap *)mapstore; map->_dm_size = size; map->_dm_segcnt = nsegments; map->_dm_maxsegsz = maxsegsz; map->_dm_boundary = boundary; map->_dm_flags = flags & ~(BUS_DMA_WAITOK|BUS_DMA_NOWAIT); + map->dm_nsegs = 0; /* no valid mappings */ *dmamp = map; return (0); } +/* + * Common function for DMA map destruction. May be called by bus-specific + * DMA map destruction functions. + */ void -_dmamap_destroy(v, map) - void *v; +_dmamap_destroy(t, map) + bus_dma_tag_t t; bus_dmamap_t map; { + free(map, M_DEVBUF); } @@ -98,7 +129,7 @@ _dmamap_destroy(v, map) */ int _dmamap_load(t, map, buf, buflen, p, flags) - void * t; + bus_dma_tag_t t; bus_dmamap_t map; void *buf; bus_size_t buflen; @@ -115,6 +146,7 @@ _dmamap_load(t, map, buf, buflen, p, flags) * Make sure that on error condition we return "no valid mappings". */ map->dm_nsegs = 0; + if (buflen > map->_dm_size) return (EINVAL); @@ -187,151 +219,309 @@ _dmamap_load(t, map, buf, buflen, p, flags) return (0); } +/* + * Like _bus_dmamap_load(), but for mbufs. + */ int -_dmamap_load_mbuf(v, map, m, flags) - void *v; +_dmamap_load_mbuf(t, map, m, flags) + bus_dma_tag_t t; bus_dmamap_t map; struct mbuf *m; int flags; { - panic("_dmamap_load_mbuf: not implemented"); + + panic("_bus_dmamap_load: not implemented"); } +/* + * Like _bus_dmamap_load(), but for uios. + */ int -_dmamap_load_uio(v, map, uio, flags) - void *v; +_dmamap_load_uio(t, map, uio, flags) + bus_dma_tag_t t; bus_dmamap_t map; struct uio *uio; int flags; { - panic("_dmamap_load_uio: not implemented"); + + panic("_bus_dmamap_load_uio: not implemented"); } +/* + * Like _bus_dmamap_load(), but for raw memory allocated with + * bus_dmamem_alloc(). + */ int -_dmamap_load_raw(v, map, segs, nsegs, size, flags) - void *v; +_dmamap_load_raw(t, map, segs, nsegs, size, flags) + bus_dma_tag_t t; bus_dmamap_t map; bus_dma_segment_t *segs; int nsegs; bus_size_t size; int flags; { - panic("_dmamap_load_raw: not implemented"); + + panic("_bus_dmamap_load_raw: not implemented"); } +/* + * Common function for unloading a DMA map. May be called by + * bus-specific DMA map unload functions. + */ void -_dmamap_unload(v, map) - void *v; +_dmamap_unload(t, map) + bus_dma_tag_t t; bus_dmamap_t map; { - panic("_dmamap_unload: not implemented"); + + /* + * No resources to free; just mark the mappings as + * invalid. + */ + map->dm_nsegs = 0; } +/* + * Common function for DMA map synchronization. May be called + * by bus-specific DMA map synchronization functions. + */ void -_dmamap_sync(v, map, ops) - void *v; +_dmamap_sync(t, map, op) + bus_dma_tag_t t; bus_dmamap_t map; - bus_dmasync_op_t ops; + bus_dmasync_op_t op; { -#if 0 - __asm __volatile ("syncdma"); -#endif + + /* Nothing to do here. */ } +/* + * Common function for DMA-safe memory allocation. May be called + * by bus-specific DMA memory allocation functions. + */ int -_dmamem_alloc(v, size, alignment, boundary, segs, nsegs, rsegs, flags) - void *v; +_dmamem_alloc(t, size, alignment, boundary, segs, nsegs, rsegs, flags) + bus_dma_tag_t t; bus_size_t size, alignment, boundary; bus_dma_segment_t *segs; int nsegs; int *rsegs; int flags; { - vaddr_t va; - paddr_t spa, epa; - - size = round_page(size); - -#if defined(UVM) - va = uvm_pagealloc_contig(size, VM_MIN_KERNEL_ADDRESS, - VM_MAX_KERNEL_ADDRESS, NBPG); -#else -# if 0 - vm_page_alloc_memory(size, VM_MIN_KERNEL_ADDRESS, - VM_MAX_KERNEL_ADDRESS, - alignment, boundary, (void *)&va, nsegs, (flags & BUS_DMA_NOWAIT)); -# else - va = kmem_alloc_wait(phys_map, NBPG); /* XXX */ -# endif -#endif - if (va == NULL) - return (ENOMEM); - - segs[0].ds_addr = va; - segs[0].ds_len = size; - *rsegs = nsegs; - -#if 0 - /* XXX for now */ - for (epa = size + (spa = kvtop((caddr_t)va)); spa < epa; spa += NBPG) - pmap_changebit(spa, TLB_UNCACHEABLE, 0); -#endif - - return 0; + extern int avail_end; + return (_dmamem_alloc_range(t, size, alignment, boundary, + segs, nsegs, rsegs, flags, 0, 0xf0000000)); } +/* + * Common function for freeing DMA-safe memory. May be called by + * bus-specific DMA memory free functions. + */ void -_dmamem_free(v, segs, nsegs) - void *v; +_dmamem_free(t, segs, nsegs) + bus_dma_tag_t t; bus_dma_segment_t *segs; int nsegs; { -#if defined (UVM) - uvm_km_free(kmem_map, segs[0].ds_addr, M_DEVBUF); + vm_page_t m; + bus_addr_t addr; + struct pglist mlist; + int curseg; + + /* + * Build a list of pages to free back to the VM system. + */ + TAILQ_INIT(&mlist); + for (curseg = 0; curseg < nsegs; curseg++) { + for (addr = segs[curseg].ds_addr; + addr < (segs[curseg].ds_addr + segs[curseg].ds_len); + addr += PAGE_SIZE) { + m = PHYS_TO_VM_PAGE(addr); + TAILQ_INSERT_TAIL(&mlist, m, pageq); + } + } + +#if defined(UVM) + uvm_pglistfree(&mlist); #else - kmem_free(kernel_map, segs[0].ds_addr, segs[0].ds_len); + vm_page_free_memory(&mlist); #endif } +/* + * Common function for mapping DMA-safe memory. May be called by + * bus-specific DMA memory map functions. + */ int -_dmamem_map(v, segs, nsegs, size, kvap, flags) - void *v; +_dmamem_map(t, segs, nsegs, size, kvap, flags) + bus_dma_tag_t t; bus_dma_segment_t *segs; int nsegs; size_t size; caddr_t *kvap; int flags; { - *kvap = (caddr_t)segs[0].ds_addr; - return 0; + vm_offset_t va; + bus_addr_t addr; + int curseg; + + size = round_page(size); +#if defined(UVM) + va = uvm_km_valloc(kmem_map, size); +#else + va = kmem_alloc_pageable(kmem_map, size); +#endif + if (va == 0) + return (ENOMEM); + + *kvap = (caddr_t)va; + + for (curseg = 0; curseg < nsegs; curseg++) { + for (addr = segs[curseg].ds_addr; + addr < (segs[curseg].ds_addr + segs[curseg].ds_len); + addr += NBPG, va += NBPG, size -= NBPG) { + if (size == 0) + panic("_bus_dmamem_map: size botch"); + pmap_enter(pmap_kernel(), va, addr, + VM_PROT_READ | VM_PROT_WRITE, TRUE, + VM_PROT_READ | VM_PROT_WRITE); + } + } + + return (0); } +/* + * Common function for unmapping DMA-safe memory. May be called by + * bus-specific DMA memory unmapping functions. + */ void -_dmamem_unmap(v, kva, size) - void *v; +_dmamem_unmap(t, kva, size) + bus_dma_tag_t t; caddr_t kva; size_t size; { + +#ifdef DIAGNOSTIC + if ((u_long)kva & PGOFSET) + panic("_bus_dmamem_unmap"); +#endif + + size = round_page(size); +#if defined(UVM) + uvm_km_free(kmem_map, (vm_offset_t)kva, size); +#else + kmem_free(kmem_map, (vm_offset_t)kva, size); +#endif } +/* + * Common functin for mmap(2)'ing DMA-safe memory. May be called by + * bus-specific DMA mmap(2)'ing functions. + */ int -_dmamem_mmap(v, segs, nsegs, off, prot, flags) - void *v; +_dmamem_mmap(t, segs, nsegs, off, prot, flags) + bus_dma_tag_t t; bus_dma_segment_t *segs; int nsegs, off, prot, flags; { - panic("_dmamem_mmap: not implemented"); + int i; + + for (i = 0; i < nsegs; i++) { +#ifdef DIAGNOSTIC + if (off & PGOFSET) + panic("_bus_dmamem_mmap: offset unaligned"); + if (segs[i].ds_addr & PGOFSET) + panic("_bus_dmamem_mmap: segment unaligned"); + if (segs[i].ds_len & PGOFSET) + panic("_bus_dmamem_mmap: segment size not multiple" + " of page size"); +#endif + if (off >= segs[i].ds_len) { + off -= segs[i].ds_len; + continue; + } + + return (powerpc_btop((caddr_t)segs[i].ds_addr + off)); + } + + /* Page not found. */ + return (-1); } +/********************************************************************** + * DMA utility functions + **********************************************************************/ + +/* + * Allocate physical memory from the given physical address range. + * Called by DMA-safe memory allocation methods. + */ int -dma_cachectl(p, size) - caddr_t p; - int size; +_dmamem_alloc_range(t, size, alignment, boundary, segs, nsegs, rsegs, + flags, low, high) + bus_dma_tag_t t; + bus_size_t size, alignment, boundary; + bus_dma_segment_t *segs; + int nsegs; + int *rsegs; + int flags; + vm_offset_t low; + vm_offset_t high; { -#if 0 - fdcache(HPPA_SID_KERNEL, (vaddr_t)p, size); - sync_caches(); + vm_offset_t curaddr, lastaddr; + vm_page_t m; + struct pglist mlist; + int curseg, error; + + /* Always round the size. */ + size = round_page(size); + + /* + * Allocate pages from the VM system. + */ + TAILQ_INIT(&mlist); +#if defined(UVM) + error = uvm_pglistalloc(size, low, high, + alignment, boundary, &mlist, nsegs, (flags & BUS_DMA_NOWAIT) == 0); +#else + error = vm_page_alloc_memory(size, low, high, + alignment, boundary, &mlist, nsegs, (flags & BUS_DMA_NOWAIT) == 0); #endif - return 0; + if (error) + return (error); + + /* + * Compute the location, size, and number of segments actually + * returned by the VM code. + */ + m = mlist.tqh_first; + curseg = 0; + lastaddr = segs[curseg].ds_addr = VM_PAGE_TO_PHYS(m); + segs[curseg].ds_len = PAGE_SIZE; + m = m->pageq.tqe_next; + + for (; m != NULL; m = m->pageq.tqe_next) { + curaddr = VM_PAGE_TO_PHYS(m); +#ifdef DIAGNOSTIC + if (curaddr < low || curaddr >= high) { + printf("vm_page_alloc_memory returned non-sensical" + " address 0x%lx\n", curaddr); + panic("dmamem_alloc_range"); + } +#endif + if (curaddr == (lastaddr + PAGE_SIZE)) + segs[curseg].ds_len += PAGE_SIZE; + else { + curseg++; + segs[curseg].ds_addr = curaddr; + segs[curseg].ds_len = PAGE_SIZE; + } + lastaddr = curaddr; + } + + *rsegs = curseg + 1; + + return (0); } |