1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
|
/* $OpenBSD: agp_machdep.c,v 1.6 2010/05/10 22:06:04 oga Exp $ */
/*
* Copyright (c) 2008 - 2009 Owain G. Ainsworth <oga@openbsd.org>
*
* Permission to use, copy, modify, and distribute this software for any
* purpose with or without fee is hereby granted, provided that the above
* copyright notice and this permission notice appear in all copies.
*
* THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
* WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
* MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
* ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
* WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
* ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
* OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
*/
/*
* Copyright (c) 2002 Michael Shalayeff
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
* OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
* IN NO EVENT SHALL THE AUTHOR OR HIS RELATIVES BE LIABLE FOR ANY DIRECT,
* INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
* SERVICES; LOSS OF MIND, USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
* STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING
* IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
* THE POSSIBILITY OF SUCH DAMAGE.
*/
#include <sys/param.h>
#include <sys/systm.h>
#include <sys/device.h>
#include <sys/malloc.h>
#include <dev/pci/pcivar.h>
#include <dev/pci/pcireg.h>
#include <dev/pci/pcidevs.h>
#include <dev/pci/agpvar.h>
#include <machine/cpufunc.h>
#include <machine/bus.h>
#include <uvm/uvm.h>
#include "intagp.h"
/* bus_dma functions */
#if NINTAGP > 0
void intagp_dma_sync(bus_dma_tag_t, bus_dmamap_t, bus_addr_t,
bus_size_t, int);
#endif
void
agp_flush_cache(void)
{
wbinvd();
}
void
agp_flush_cache_range(vaddr_t va, vsize_t sz)
{
pmap_flush_cache(va, sz);
}
/*
* functions for bus_dma used by drm for GEM
*
* We use the sg_dma backend (also used by iommu) to provide the actual
* implementation, so all we need provide is the magic to create the tag, and
* the appropriate callbacks.
*
* We give the backend drivers a chance to honour the bus_dma flags, some of
* these may be used, for example to provide snooped mappings (intagp).
* For intagp at least, we honour the BUS_DMA_COHERENT flag, though it is not
* used often, and is * technically to be used for dmamem_map, we use it for
* dmamap_load since adding coherency involes flags to the gtt pagetables.
* We only use it for very special circumstances since when a GTT mapping is
* set to coherent, the cpu can't read or write through the gtt aperture.
*
* Currently, since the userland agp driver still needs to access the gart, we
* only do bus_dma for a section that we've been told is ours, hence the need
* for the init function at present.
*/
int
agp_bus_dma_init(struct agp_softc *sc, bus_addr_t start, bus_addr_t end,
bus_dma_tag_t *dmat)
{
struct bus_dma_tag *tag;
struct sg_cookie *cookie;
/*
* XXX add agp map into the main queue that takes up our chunk of
* GTT space to prevent the userland api stealing any of it.
*/
if ((tag = malloc(sizeof(*tag), M_DEVBUF,
M_WAITOK | M_CANFAIL)) == NULL)
return (ENOMEM);
if ((cookie = sg_dmatag_init("agpgtt", sc->sc_chipc, start, end - start,
sc->sc_methods->bind_page, sc->sc_methods->unbind_page,
sc->sc_methods->flush_tlb)) == NULL) {
free(tag, M_DEVBUF);
return (ENOMEM);
}
tag->_cookie = cookie;
tag->_dmamap_create = sg_dmamap_create;
tag->_dmamap_destroy = sg_dmamap_destroy;
tag->_dmamap_load = sg_dmamap_load;
tag->_dmamap_load_mbuf = sg_dmamap_load_mbuf;
tag->_dmamap_load_uio = sg_dmamap_load_uio;
tag->_dmamap_load_raw = sg_dmamap_load_raw;
tag->_dmamap_unload = sg_dmamap_unload;
tag->_dmamem_alloc = sg_dmamem_alloc;
tag->_dmamem_free = _bus_dmamem_free;
tag->_dmamem_map = _bus_dmamem_map;
tag->_dmamem_unmap = _bus_dmamem_unmap;
tag->_dmamem_mmap = _bus_dmamem_mmap;
/* Driver may need special sync handling */
if (sc->sc_methods->dma_sync != NULL) {
tag->_dmamap_sync = sc->sc_methods->dma_sync;
} else {
tag->_dmamap_sync = _bus_dmamap_sync;
}
*dmat = tag;
return (0);
}
void
agp_bus_dma_destroy(struct agp_softc *sc, bus_dma_tag_t dmat)
{
struct sg_cookie *cookie = dmat->_cookie;
bus_addr_t offset;
/*
* XXX clear up blocker queue
*/
/*
* some backends use a dummy page to avoid errors on prefetching, etc.
* make sure that all of them are clean.
*/
for (offset = cookie->sg_ex->ex_start;
offset < cookie->sg_ex->ex_end; offset += PAGE_SIZE)
sc->sc_methods->unbind_page(sc->sc_chipc, offset);
sg_dmatag_destroy(cookie);
free(dmat, M_DEVBUF);
}
void
agp_bus_dma_set_alignment(bus_dma_tag_t tag, bus_dmamap_t dmam,
u_long alignment)
{
sg_dmamap_set_alignment(tag, dmam, alignment);
}
struct agp_map {
bus_space_tag_t bst;
bus_space_handle_t bsh;
bus_size_t size;
};
int
agp_init_map(bus_space_tag_t tag, bus_addr_t address, bus_size_t size,
int flags, struct agp_map **mapp)
{
struct agp_map *map;
int err;
map = malloc(sizeof(*map), M_AGP, M_WAITOK | M_CANFAIL);
if (map == NULL)
return (ENOMEM);
map->bst = tag;
map->size = size;
if ((err = bus_space_map(tag, address, size, flags, &map->bsh)) != 0) {
free(map, M_AGP);
return (err);
}
*mapp = map;
return (0);
}
void
agp_destroy_map(struct agp_map *map)
{
bus_space_unmap(map->bst, map->bsh, map->size);
free(map, M_AGP);
}
int
agp_map_subregion(struct agp_map *map, bus_size_t offset, bus_size_t size,
bus_space_handle_t *bshp)
{
if (offset > map->size || size > map->size || offset + size > map->size)
return (EINVAL);
return (bus_space_subregion(map->bst, map->bsh, offset, size, bshp));
}
void
agp_unmap_subregion(struct agp_map *map, bus_space_handle_t bsh,
bus_size_t size)
{
/* subregion doesn't need unmapping, do nothing */
}
/*
* ick ick ick. However, the rest of this driver is supposedly MI (though
* they only exist on x86), so this can't be in dev/pci.
*/
#if NINTAGP > 0
/*
* bus_dmamap_sync routine for intagp.
*
* This is tailored to the usage that drm with the GEM memory manager
* will be using, since intagp is for intel IGD, and thus shouldn't be
* used for anything other than gpu-based work. Essentially for the intel GEM
* driver we use bus_dma as an abstraction to convert our memory into a gtt
* address and deal with any cache incoherencies that we create.
*
* We use the cflush instruction to deal with clearing the caches, since our
* cache is physically indexed, we can even map then clear the page and it'll
* work. on i386 we need to check for the presence of cflush() in cpuid,
* however, all cpus that have a new enough intel GMCH should be suitable.
*/
void
intagp_dma_sync(bus_dma_tag_t tag, bus_dmamap_t dmam,
bus_addr_t offset, bus_size_t size, int ops)
{
bus_dma_segment_t *segp;
struct sg_page_map *spm;
void *addr;
paddr_t pa;
bus_addr_t poff, endoff, soff;
#ifdef DIAGNOSTIC
if ((ops & (BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE)) != 0 &&
(ops & (BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE)) != 0)
panic("agp_dmamap_sync: mix PRE and POST");
if (offset >= dmam->dm_mapsize)
panic("_intagp_dma_sync: bad offset %lu (size = %lu)",
offset, dmam->dm_mapsize);
if (size == 0 || (offset + size) > dmam->dm_mapsize)
panic("intagp_dma_sync: bad length");
#endif /* DIAGNOSTIC */
/* Coherent mappings need no sync. */
if (dmam->_dm_flags & BUS_DMA_COHERENT)
return;
/*
* We need to clflush the object cache in all cases but postwrite.
*
* - Due to gpu incoherency, postread we need to flush speculative
* reads (which are not written back on intel cpus).
*
* - preread we need to flush data which will very soon be stale from
* the caches
*
* - prewrite we need to make sure our data hits the memory before the
* gpu hoovers it up.
*
* The chipset also may need flushing, but that fits badly into
* bus_dma and it done in the driver.
*/
soff = trunc_page(offset);
endoff = round_page(offset + size);
if (ops & BUS_DMASYNC_POSTREAD || ops & BUS_DMASYNC_PREREAD ||
ops & BUS_DMASYNC_PREWRITE) {
if (curcpu()->ci_cflushsz == 0) {
/* save some wbinvd()s. we're MD anyway so it's ok */
wbinvd();
return;
}
mfence();
spm = dmam->_dm_cookie;
switch (spm->spm_buftype) {
case BUS_BUFTYPE_LINEAR:
addr = spm->spm_origbuf + soff;
while (soff < endoff) {
pmap_flush_cache((vaddr_t)addr, PAGE_SIZE);
soff += PAGE_SIZE;
addr += PAGE_SIZE;
} break;
case BUS_BUFTYPE_RAW:
segp = (bus_dma_segment_t *)spm->spm_origbuf;
poff = 0;
while (poff < soff) {
if (poff + segp->ds_len > soff)
break;
poff += segp->ds_len;
segp++;
}
/* first time round may not start at seg beginning */
pa = segp->ds_addr + (soff - poff);
while (poff < endoff) {
for (; pa < segp->ds_addr + segp->ds_len &&
poff < endoff; pa += PAGE_SIZE) {
pmap_flush_page(pa);
poff += PAGE_SIZE;
}
segp++;
if (poff < endoff)
pa = segp->ds_addr;
}
break;
/* You do not want to load mbufs or uios onto a graphics card */
case BUS_BUFTYPE_MBUF:
/* FALLTHROUGH */
case BUS_BUFTYPE_UIO:
/* FALLTHROUGH */
default:
panic("intagp_dmamap_sync: bad buftype %d",
spm->spm_buftype);
}
mfence();
}
}
#endif /* NINTAGP > 0 */
|