summaryrefslogtreecommitdiff
path: root/sys/arch/i386/pci
diff options
context:
space:
mode:
Diffstat (limited to 'sys/arch/i386/pci')
-rw-r--r--sys/arch/i386/pci/agp_machdep.c180
1 files changed, 179 insertions, 1 deletions
diff --git a/sys/arch/i386/pci/agp_machdep.c b/sys/arch/i386/pci/agp_machdep.c
index 507428a94d2..87bedfc6cd7 100644
--- a/sys/arch/i386/pci/agp_machdep.c
+++ b/sys/arch/i386/pci/agp_machdep.c
@@ -1,6 +1,21 @@
-/* $OpenBSD: agp_machdep.c,v 1.9 2007/11/25 17:11:12 oga Exp $ */
+/* $OpenBSD: agp_machdep.c,v 1.10 2009/06/06 06:02:44 oga Exp $ */
/*
+ * Copyright (c) 2008 - 2009 Owain G. Ainsworth <oga@openbsd.org>
+ *
+ * Permission to use, copy, modify, and distribute this software for any
+ * purpose with or without fee is hereby granted, provided that the above
+ * copyright notice and this permission notice appear in all copies.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
+ * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
+ * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
+ * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
+ * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
+ * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
+ * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
+ */
+/*
* Copyright (c) 2002 Michael Shalayeff
* All rights reserved.
*
@@ -27,7 +42,9 @@
*/
#include <sys/param.h>
+#include <sys/systm.h>
#include <sys/device.h>
+#include <sys/malloc.h>
#include <dev/pci/pcivar.h>
#include <dev/pci/pcireg.h>
@@ -35,9 +52,170 @@
#include <dev/pci/agpvar.h>
#include <machine/cpufunc.h>
+#include <machine/bus.h>
+
+#include "intagp.h"
+
+/* bus_dma functions */
+
+#if NINTAGP > 0
+void intagp_dma_sync(bus_dma_tag_t, bus_dmamap_t, bus_addr_t,
+ bus_size_t, int);
+#endif
void
agp_flush_cache(void)
{
wbinvd();
}
+
+/*
+ * functions for bus_dma used by drm for GEM
+ *
+ * We use the sg_dma backend (also used by iommu) to provide the actual
+ * implementation, so all we need provide is the magic to create the tag, and
+ * the appropriate callbacks.
+ *
+ * We give the backend drivers a chance to honour the bus_dma flags, some of
+ * these may be used, for example to provide snooped mappings (intagp).
+ * For intagp at least, we honour the BUS_DMA_COHERENT flag, though it is not
+ * used often, and is * technically to be used for dmamem_map, we use it for
+ * dmamap_load since adding coherency involes flags to the gtt pagetables.
+ * We only use it for very special circumstances since when a GTT mapping is
+ * set to coherent, the cpu can't read or write through the gtt aperture.
+ *
+ * Currently, since the userland agp driver still needs to access the gart, we
+ * only do bus_dma for a section that we've been told is ours, hence the need
+ * for the init function at present.
+ */
+
+int
+agp_bus_dma_init(struct agp_softc *sc, bus_addr_t start, bus_addr_t end,
+ bus_dma_tag_t *dmat)
+{
+ struct bus_dma_tag *tag;
+ struct sg_cookie *cookie;
+
+ /*
+ * XXX add agp map into the main queue that takes up our chunk of
+ * GTT space to prevent the userland api stealing any of it.
+ */
+ if ((tag = malloc(sizeof(*tag), M_DEVBUF,
+ M_WAITOK | M_CANFAIL)) == NULL)
+ return (ENOMEM);
+
+ if ((cookie = sg_dmatag_init("agpgtt", sc->sc_chipc, start, end - start,
+ sc->sc_methods->bind_page, sc->sc_methods->unbind_page,
+ sc->sc_methods->flush_tlb)) == NULL) {
+ free(tag, M_DEVBUF);
+ return (ENOMEM);
+ }
+
+ tag->_cookie = cookie;
+ tag->_dmamap_create = sg_dmamap_create;
+ tag->_dmamap_destroy = sg_dmamap_destroy;
+ tag->_dmamap_load = sg_dmamap_load;
+ tag->_dmamap_load_mbuf = sg_dmamap_load_mbuf;
+ tag->_dmamap_load_uio = sg_dmamap_load_uio;
+ tag->_dmamap_load_raw = sg_dmamap_load_raw;
+ tag->_dmamap_unload = sg_dmamap_unload;
+ tag->_dmamem_alloc = sg_dmamem_alloc;
+ tag->_dmamem_free = _bus_dmamem_free;
+ tag->_dmamem_map = _bus_dmamem_map;
+ tag->_dmamem_unmap = _bus_dmamem_unmap;
+ tag->_dmamem_mmap = _bus_dmamem_mmap;
+
+ /* Driver may need special sync handling */
+ if (sc->sc_methods->dma_sync != NULL) {
+ tag->_dmamap_sync = sc->sc_methods->dma_sync;
+ } else {
+ tag->_dmamap_sync = NULL;
+ }
+
+ *dmat = tag;
+ return (0);
+}
+
+void
+agp_bus_dma_destroy(struct agp_softc *sc, bus_dma_tag_t dmat)
+{
+ struct sg_cookie *cookie = dmat->_cookie;
+
+
+ /*
+ * XXX clear up blocker queue
+ */
+
+ sg_dmatag_destroy(cookie);
+ free(dmat, M_DEVBUF);
+}
+
+void
+agp_bus_dma_set_alignment(bus_dma_tag_t tag, bus_dmamap_t dmam,
+ u_long alignment)
+{
+ sg_dmamap_set_alignment(tag, dmam, alignment);
+}
+
+
+/*
+ * ick ick ick. However, the rest of this driver is supposedly MI (though
+ * they only exist on x86), so this can't be in dev/pci.
+ */
+
+#if NINTAGP > 0
+
+/*
+ * bus_dmamap_sync routine for intagp.
+ *
+ * This is tailored to the usage that drm with the GEM memory manager
+ * will be using, since intagp is for intel IGD, and thus shouldn't be
+ * used for anything other than gpu-based work. Essentially for the intel GEM
+ * driver we use bus_dma as an abstraction to convert our memory into a gtt
+ * address and deal with any cache incoherencies that we create.
+ *
+ * We use the cflush instruction to deal with clearing the caches, since our
+ * cache is physically indexed, we can even map then clear the page and it'll
+ * work. on i386 we need to check for the presence of cflush() in cpuid,
+ * however, all cpus that have a new enough intel GMCH should be suitable.
+ */
+void
+intagp_dma_sync(bus_dma_tag_t tag, bus_dmamap_t dmam,
+ bus_addr_t offset, bus_size_t size, int ops)
+{
+#ifdef DIAGNOSTIC
+ if ((ops & (BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE)) != 0 &&
+ (ops & (BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE)) != 0)
+ panic("agp_dmamap_sync: mix PRE and POST");
+ if (offset >= dmam->dm_mapsize)
+ panic("_intagp_dma_sync: bad offset %lu (size = %lu)",
+ offset, dmam->dm_mapsize);
+ if (size == 0 || (offset + size) > dmam->dm_mapsize)
+ panic("intagp_dma_sync: bad length");
+#endif /* DIAGNOSTIC */
+
+ /* Coherent mappings need no sync. */
+ if (dmam->_dm_flags & BUS_DMA_COHERENT)
+ return;
+
+ /*
+ * We need to clflush the object cache in all cases but postwrite.
+ *
+ * - Due to gpu incoherency, postread we need to flush speculative
+ * reads (which are not written back on intel cpus).
+ *
+ * - preread we need to flush data which will very soon be stale from
+ * the caches
+ *
+ * - prewrite we need to make sure our data hits the memory before the
+ * gpu hoovers it up.
+ *
+ * The chipset also may need flushing, but that fits badly into
+ * bus_dma and it done in the driver.
+ */
+ if (ops & BUS_DMASYNC_POSTREAD || ops & BUS_DMASYNC_PREREAD ||
+ ops & BUS_DMASYNC_PREWRITE) {
+ wbinvd();
+ }
+}
+#endif