/************************************************************************** Copyright 1998-1999 Precision Insight, Inc., Cedar Park, Texas. All Rights Reserved. Copyright (c) 2005 Jesse Barnes Based on code from i830_xaa.c. Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sub license, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice (including the next paragraph) shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL PRECISION INSIGHT AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. **************************************************************************/ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include "xorg-server.h" #include #include #include #include #include #include #include "intel.h" #include "intel_uxa.h" #include "i830_reg.h" #include "i915_drm.h" #include "brw_defines.h" static const int I830CopyROP[16] = { ROP_0, /* GXclear */ ROP_DSa, /* GXand */ ROP_SDna, /* GXandReverse */ ROP_S, /* GXcopy */ ROP_DSna, /* GXandInverted */ ROP_D, /* GXnoop */ ROP_DSx, /* GXxor */ ROP_DSo, /* GXor */ ROP_DSon, /* GXnor */ ROP_DSxn, /* GXequiv */ ROP_Dn, /* GXinvert */ ROP_SDno, /* GXorReverse */ ROP_Sn, /* GXcopyInverted */ ROP_DSno, /* GXorInverted */ ROP_DSan, /* GXnand */ ROP_1 /* GXset */ }; static const int I830PatternROP[16] = { ROP_0, ROP_DPa, ROP_PDna, ROP_P, ROP_DPna, ROP_D, ROP_DPx, ROP_DPo, ROP_DPon, ROP_PDxn, ROP_Dn, ROP_PDno, ROP_Pn, ROP_DPno, ROP_DPan, ROP_1 }; #if HAS_DEVPRIVATEKEYREC DevPrivateKeyRec uxa_pixmap_index; #else int uxa_pixmap_index; #endif static void gen6_context_switch(intel_screen_private *intel, int new_mode) { intel_batch_submit(intel->scrn); } static void gen5_context_switch(intel_screen_private *intel, int new_mode) { /* Ironlake has a limitation that a 3D or Media command can't * be the first command after a BLT, unless it's * non-pipelined. Instead of trying to track it and emit a * command at the right time, we just emit a dummy * non-pipelined 3D instruction after each blit. */ if (new_mode == I915_EXEC_BLT) { OUT_BATCH(MI_FLUSH | MI_STATE_INSTRUCTION_CACHE_FLUSH | MI_INHIBIT_RENDER_CACHE_FLUSH); } else { OUT_BATCH(CMD_POLY_STIPPLE_OFFSET << 16); OUT_BATCH(0); } } static void gen4_context_switch(intel_screen_private *intel, int new_mode) { if (new_mode == I915_EXEC_BLT) { OUT_BATCH(MI_FLUSH | MI_STATE_INSTRUCTION_CACHE_FLUSH | MI_INHIBIT_RENDER_CACHE_FLUSH); } } Bool intel_uxa_get_aperture_space(ScrnInfoPtr scrn, drm_intel_bo ** bo_table, int num_bos) { intel_screen_private *intel = intel_get_screen_private(scrn); if (intel->batch_bo == NULL) { intel_uxa_debug_fallback(scrn, "VT inactive\n"); return FALSE; } bo_table[0] = intel->batch_bo; if (drm_intel_bufmgr_check_aperture_space(bo_table, num_bos) != 0) { intel_batch_submit(scrn); bo_table[0] = intel->batch_bo; if (drm_intel_bufmgr_check_aperture_space(bo_table, num_bos) != 0) { intel_uxa_debug_fallback(scrn, "Couldn't get aperture " "space for BOs\n"); return FALSE; } } return TRUE; } static Bool intel_uxa_check_solid(DrawablePtr drawable, int alu, Pixel planemask) { ScrnInfoPtr scrn = xf86ScreenToScrn(drawable->pScreen); if (!UXA_PM_IS_SOLID(drawable, planemask)) { intel_uxa_debug_fallback(scrn, "planemask is not solid\n"); return FALSE; } switch (drawable->bitsPerPixel) { case 8: case 16: case 32: break; default: return FALSE; } return TRUE; } /** * Sets up hardware state for a series of solid fills. */ static Bool intel_uxa_prepare_solid(PixmapPtr pixmap, int alu, Pixel planemask, Pixel fg) { ScrnInfoPtr scrn = xf86ScreenToScrn(pixmap->drawable.pScreen); intel_screen_private *intel = intel_get_screen_private(scrn); drm_intel_bo *bo_table[] = { NULL, /* batch_bo */ intel_uxa_get_pixmap_bo(pixmap), }; if (!intel_uxa_check_pitch_2d(pixmap)) return FALSE; if (!intel_uxa_get_aperture_space(scrn, bo_table, ARRAY_SIZE(bo_table))) return FALSE; intel->BR[13] = (I830PatternROP[alu] & 0xff) << 16; switch (pixmap->drawable.bitsPerPixel) { case 8: break; case 16: /* RGB565 */ intel->BR[13] |= (1 << 24); break; case 32: /* RGB8888 */ intel->BR[13] |= ((1 << 24) | (1 << 25)); break; } intel->BR[16] = fg; return TRUE; } static void intel_uxa_solid(PixmapPtr pixmap, int x1, int y1, int x2, int y2) { ScrnInfoPtr scrn = xf86ScreenToScrn(pixmap->drawable.pScreen); intel_screen_private *intel = intel_get_screen_private(scrn); unsigned long pitch; uint32_t cmd; if (x1 < 0) x1 = 0; if (y1 < 0) y1 = 0; if (x2 > pixmap->drawable.width) x2 = pixmap->drawable.width; if (y2 > pixmap->drawable.height) y2 = pixmap->drawable.height; if (x2 <= x1 || y2 <= y1) return; pitch = intel_pixmap_pitch(pixmap); { int len = INTEL_INFO(intel)->gen >= 0100 ? 7 : 6; BEGIN_BATCH_BLT(len); cmd = XY_COLOR_BLT_CMD | (len - 2); if (pixmap->drawable.bitsPerPixel == 32) cmd |= XY_COLOR_BLT_WRITE_ALPHA | XY_COLOR_BLT_WRITE_RGB; if (INTEL_INFO(intel)->gen >= 040 && intel_uxa_pixmap_tiled(pixmap)) { assert((pitch % 512) == 0); pitch >>= 2; cmd |= XY_COLOR_BLT_TILED; } OUT_BATCH(cmd); OUT_BATCH(intel->BR[13] | pitch); OUT_BATCH((y1 << 16) | (x1 & 0xffff)); OUT_BATCH((y2 << 16) | (x2 & 0xffff)); OUT_RELOC_PIXMAP_FENCED(pixmap, I915_GEM_DOMAIN_RENDER, I915_GEM_DOMAIN_RENDER, 0); OUT_BATCH(intel->BR[16]); ADVANCE_BATCH(); } } /** * TODO: * - support planemask using FULL_BLT_CMD? */ static Bool intel_uxa_check_copy(PixmapPtr source, PixmapPtr dest, int alu, Pixel planemask) { ScrnInfoPtr scrn = xf86ScreenToScrn(dest->drawable.pScreen); if (!UXA_PM_IS_SOLID(&source->drawable, planemask)) { intel_uxa_debug_fallback(scrn, "planemask is not solid"); return FALSE; } if (source->drawable.bitsPerPixel != dest->drawable.bitsPerPixel) { intel_uxa_debug_fallback(scrn, "mixed bpp copies unsupported\n"); return FALSE; } switch (source->drawable.bitsPerPixel) { case 8: case 16: case 32: break; default: return FALSE; } if (!intel_uxa_check_pitch_2d(source)) return FALSE; if (!intel_uxa_check_pitch_2d(dest)) return FALSE; return TRUE; } static Bool intel_uxa_prepare_copy(PixmapPtr source, PixmapPtr dest, int xdir, int ydir, int alu, Pixel planemask) { ScrnInfoPtr scrn = xf86ScreenToScrn(dest->drawable.pScreen); intel_screen_private *intel = intel_get_screen_private(scrn); drm_intel_bo *bo_table[] = { NULL, /* batch_bo */ intel_uxa_get_pixmap_bo(source), intel_uxa_get_pixmap_bo(dest), }; if (!intel_uxa_get_aperture_space(scrn, bo_table, ARRAY_SIZE(bo_table))) return FALSE; intel->render_source = source; intel->BR[13] = I830CopyROP[alu] << 16; switch (source->drawable.bitsPerPixel) { case 8: break; case 16: intel->BR[13] |= (1 << 24); break; case 32: intel->BR[13] |= ((1 << 25) | (1 << 24)); break; } return TRUE; } static void intel_uxa_copy(PixmapPtr dest, int src_x1, int src_y1, int dst_x1, int dst_y1, int w, int h) { ScrnInfoPtr scrn = xf86ScreenToScrn(dest->drawable.pScreen); intel_screen_private *intel = intel_get_screen_private(scrn); uint32_t cmd; int dst_x2, dst_y2, src_x2, src_y2; unsigned int dst_pitch, src_pitch; dst_x2 = dst_x1 + w; dst_y2 = dst_y1 + h; /* XXX Fixup extents as a lamentable workaround for missing * source clipping in the upper layers. */ if (dst_x1 < 0) src_x1 -= dst_x1, dst_x1 = 0; if (dst_y1 < 0) src_y1 -= dst_y1, dst_y1 = 0; if (dst_x2 > dest->drawable.width) dst_x2 = dest->drawable.width; if (dst_y2 > dest->drawable.height) dst_y2 = dest->drawable.height; src_x2 = src_x1 + (dst_x2 - dst_x1); src_y2 = src_y1 + (dst_y2 - dst_y1); if (src_x1 < 0) dst_x1 -= src_x1, src_x1 = 0; if (src_y1 < 0) dst_y1 -= src_y1, src_y1 = 0; if (src_x2 > intel->render_source->drawable.width) dst_x2 -= src_x2 - intel->render_source->drawable.width; if (src_y2 > intel->render_source->drawable.height) dst_y2 -= src_y2 - intel->render_source->drawable.height; if (dst_x2 <= dst_x1 || dst_y2 <= dst_y1) return; dst_pitch = intel_pixmap_pitch(dest); src_pitch = intel_pixmap_pitch(intel->render_source); { int len = INTEL_INFO(intel)->gen >= 0100 ? 10 : 8; BEGIN_BATCH_BLT(len); cmd = XY_SRC_COPY_BLT_CMD | (len - 2); if (dest->drawable.bitsPerPixel == 32) cmd |= XY_SRC_COPY_BLT_WRITE_ALPHA | XY_SRC_COPY_BLT_WRITE_RGB; if (INTEL_INFO(intel)->gen >= 040) { if (intel_uxa_pixmap_tiled(dest)) { assert((dst_pitch % 512) == 0); dst_pitch >>= 2; cmd |= XY_SRC_COPY_BLT_DST_TILED; } if (intel_uxa_pixmap_tiled(intel->render_source)) { assert((src_pitch % 512) == 0); src_pitch >>= 2; cmd |= XY_SRC_COPY_BLT_SRC_TILED; } } OUT_BATCH(cmd); OUT_BATCH(intel->BR[13] | dst_pitch); OUT_BATCH((dst_y1 << 16) | (dst_x1 & 0xffff)); OUT_BATCH((dst_y2 << 16) | (dst_x2 & 0xffff)); OUT_RELOC_PIXMAP_FENCED(dest, I915_GEM_DOMAIN_RENDER, I915_GEM_DOMAIN_RENDER, 0); OUT_BATCH((src_y1 << 16) | (src_x1 & 0xffff)); OUT_BATCH(src_pitch); OUT_RELOC_PIXMAP_FENCED(intel->render_source, I915_GEM_DOMAIN_RENDER, 0, 0); ADVANCE_BATCH(); } } static void intel_uxa_done(PixmapPtr pixmap) { ScrnInfoPtr scrn = xf86ScreenToScrn(pixmap->drawable.pScreen); intel_screen_private *intel = intel_get_screen_private(scrn); if (INTEL_INFO(intel)->gen >= 060) { /* workaround a random BLT hang */ BEGIN_BATCH_BLT(3); OUT_BATCH(XY_SETUP_CLIP_BLT_CMD | (3 - 2)); OUT_BATCH(0); OUT_BATCH(0); ADVANCE_BATCH(); } intel_uxa_debug_flush(scrn); } /** * Do any cleanup from the Composite operation. * * This is shared between i830 through i965. */ static void i830_done_composite(PixmapPtr dest) { ScrnInfoPtr scrn = xf86ScreenToScrn(dest->drawable.pScreen); intel_screen_private *intel = intel_get_screen_private(scrn); if (intel->vertex_flush) intel->vertex_flush(intel); intel_uxa_debug_flush(scrn); } #define xFixedToFloat(val) \ ((float)xFixedToInt(val) + ((float)xFixedFrac(val) / 65536.0)) static Bool _intel_transform_point(PictTransformPtr transform, float x, float y, float result[3]) { int j; for (j = 0; j < 3; j++) { result[j] = (xFixedToFloat(transform->matrix[j][0]) * x + xFixedToFloat(transform->matrix[j][1]) * y + xFixedToFloat(transform->matrix[j][2])); } if (!result[2]) return FALSE; return TRUE; } /** * Returns the floating-point coordinates transformed by the given transform. * * transform may be null. */ Bool intel_uxa_get_transformed_coordinates(int x, int y, PictTransformPtr transform, float *x_out, float *y_out) { if (transform == NULL) { *x_out = x; *y_out = y; } else { float result[3]; if (!_intel_transform_point(transform, x, y, result)) return FALSE; *x_out = result[0] / result[2]; *y_out = result[1] / result[2]; } return TRUE; } /** * Returns the un-normalized floating-point coordinates transformed by the given transform. * * transform may be null. */ Bool intel_uxa_get_transformed_coordinates_3d(int x, int y, PictTransformPtr transform, float *x_out, float *y_out, float *w_out) { if (transform == NULL) { *x_out = x; *y_out = y; *w_out = 1; } else { float result[3]; if (!_intel_transform_point(transform, x, y, result)) return FALSE; *x_out = result[0]; *y_out = result[1]; *w_out = result[2]; } return TRUE; } /** * Returns whether the provided transform is affine. * * transform may be null. */ Bool intel_uxa_transform_is_affine(PictTransformPtr t) { if (t == NULL) return TRUE; return t->matrix[2][0] == 0 && t->matrix[2][1] == 0; } dri_bo *intel_uxa_get_pixmap_bo(PixmapPtr pixmap) { struct intel_uxa_pixmap *intel; intel = intel_uxa_get_pixmap_private(pixmap); if (intel == NULL) return NULL; return intel->bo; } static unsigned intel_get_tile_width(intel_screen_private *intel, int tiling, int pitch) { unsigned long tile_width; if (tiling == I915_TILING_NONE) return 4; tile_width = (tiling == I915_TILING_Y) ? 128 : 512; if (INTEL_INFO(intel)->gen >= 040) return tile_width; while (tile_width < pitch) tile_width <<= 1; return tile_width; } void intel_uxa_set_pixmap_bo(PixmapPtr pixmap, dri_bo * bo) { ScrnInfoPtr scrn = xf86ScreenToScrn(pixmap->drawable.pScreen); intel_screen_private *intel = intel_get_screen_private(scrn); struct intel_uxa_pixmap *priv; priv = intel_uxa_get_pixmap_private(pixmap); if (priv == NULL && bo == NULL) return; if (priv != NULL) { if (priv->bo == bo) return; free_priv: dri_bo_unreference(priv->bo); list_del(&priv->batch); free(priv); priv = NULL; } if (bo != NULL) { uint32_t tiling, swizzle_mode; unsigned tile_width; int size, stride; priv = calloc(1, sizeof (struct intel_uxa_pixmap)); if (priv == NULL) goto BAIL; list_init(&priv->batch); dri_bo_reference(bo); priv->bo = bo; if (drm_intel_bo_get_tiling(bo, &tiling, &swizzle_mode)) { bo = NULL; goto free_priv; } priv->tiling = tiling; priv->busy = -1; priv->offscreen = 1; stride = (pixmap->drawable.width * pixmap->drawable.bitsPerPixel + 7) / 8; tile_width = intel_get_tile_width(intel, tiling, stride); stride = ALIGN(stride, tile_width); if (intel_pixmap_pitch(pixmap) < stride || intel_pixmap_pitch(pixmap) & (tile_width - 1) || intel_pixmap_pitch(pixmap) >= KB(32)) { xf86DrvMsg(scrn->scrnIndex, X_ERROR, "%s: stride on buffer object does not match constraints: stride=%d, must be greater than %d, but less than %d, and have alignment at least %d\n", __FUNCTION__, intel_pixmap_pitch(pixmap), stride, KB(32), tile_width); bo = NULL; goto free_priv; } if (tiling != I915_TILING_NONE) { int height; if (IS_GEN2(intel)) height = 16; else if (tiling == I915_TILING_X) height = 8; else height = 32; height = ALIGN(pixmap->drawable.height, height); size = intel_get_fence_size(intel, intel_pixmap_pitch(pixmap) * height); } else size = intel_pixmap_pitch(pixmap) * pixmap->drawable.height; if (bo->size < size || bo->size > intel->max_bo_size) { xf86DrvMsg(scrn->scrnIndex, X_ERROR, "%s: size of buffer object does not match constraints: size=%ld, must be greater than %d, but less than %d\n", __FUNCTION__, (long)bo->size, size, intel->max_bo_size); bo = NULL; goto free_priv; } } BAIL: intel_uxa_set_pixmap_private(pixmap, priv); } static Bool intel_uxa_prepare_access(PixmapPtr pixmap, uxa_access_t access) { ScrnInfoPtr scrn = xf86ScreenToScrn(pixmap->drawable.pScreen); intel_screen_private *intel = intel_get_screen_private(scrn); struct intel_uxa_pixmap *priv = intel_uxa_get_pixmap_private(pixmap); dri_bo *bo = priv->bo; int ret; /* When falling back to swrast, flush all pending operations */ if (access == UXA_ACCESS_RW || priv->dirty) intel_batch_submit(scrn); assert(bo->size <= intel->max_gtt_map_size); ret = drm_intel_gem_bo_map_gtt(bo); if (ret) { xf86DrvMsg(scrn->scrnIndex, X_WARNING, "%s: bo map (use gtt? %d, access %d) failed: %s\n", __FUNCTION__, priv->tiling || bo->size <= intel->max_gtt_map_size, access, strerror(-ret)); return FALSE; } pixmap->devPrivate.ptr = bo->virtual; priv->busy = 0; return TRUE; } static void intel_uxa_finish_access(PixmapPtr pixmap, uxa_access_t access) { struct intel_uxa_pixmap *priv; priv = intel_uxa_get_pixmap_private(pixmap); if (priv == NULL) return; drm_intel_gem_bo_unmap_gtt(priv->bo); pixmap->devPrivate.ptr = NULL; } static Bool intel_uxa_pixmap_put_image(PixmapPtr pixmap, char *src, int src_pitch, int x, int y, int w, int h) { struct intel_uxa_pixmap *priv = intel_uxa_get_pixmap_private(pixmap); int stride = intel_pixmap_pitch(pixmap); int cpp = pixmap->drawable.bitsPerPixel/8; int ret = FALSE; if (priv == NULL || priv->bo == NULL) return FALSE; if (priv->tiling == I915_TILING_NONE && (h == 1 || (src_pitch == stride && w == pixmap->drawable.width))) { return drm_intel_bo_subdata(priv->bo, y*stride + x*cpp, stride*(h-1) + w*cpp, src) == 0; } else if (drm_intel_gem_bo_map_gtt(priv->bo) == 0) { char *dst = priv->bo->virtual; int row_length = w * cpp; int num_rows = h; if (row_length == src_pitch && src_pitch == stride) num_rows = 1, row_length *= h; dst += y * stride + x * cpp; do { memcpy (dst, src, row_length); src += src_pitch; dst += stride; } while (--num_rows); drm_intel_gem_bo_unmap_gtt(priv->bo); ret = TRUE; } return ret; } static Bool intel_uxa_put_image(PixmapPtr pixmap, int x, int y, int w, int h, char *src, int src_pitch) { struct intel_uxa_pixmap *priv; priv = intel_uxa_get_pixmap_private(pixmap); if (!intel_uxa_pixmap_is_busy(priv)) { /* bo is not busy so can be replaced without a stall, upload in-place. */ return intel_uxa_pixmap_put_image(pixmap, src, src_pitch, x, y, w, h); } else { ScreenPtr screen = pixmap->drawable.pScreen; if (!priv->pinned && x == 0 && y == 0 && w == pixmap->drawable.width && h == pixmap->drawable.height) { intel_screen_private *intel = intel_get_screen_private(xf86ScreenToScrn(screen)); uint32_t tiling = priv->tiling; int size, stride; dri_bo *bo; /* Replace busy bo. */ size = intel_compute_size(intel, w, h, pixmap->drawable.bitsPerPixel, pixmap->usage_hint, &tiling, &stride); if (size > intel->max_gtt_map_size) return FALSE; bo = drm_intel_bo_alloc(intel->bufmgr, "pixmap", size, 0); if (bo == NULL) return FALSE; if (tiling != I915_TILING_NONE) drm_intel_bo_set_tiling(bo, &tiling, stride); priv->tiling = tiling; screen->ModifyPixmapHeader(pixmap, w, h, 0, 0, stride, NULL); intel_uxa_set_pixmap_bo(pixmap, bo); dri_bo_unreference(bo); return intel_uxa_pixmap_put_image(pixmap, src, src_pitch, 0, 0, w, h); } else { PixmapPtr scratch; Bool ret; /* Upload to a linear buffer and queue a blit. */ scratch = (*screen->CreatePixmap)(screen, w, h, pixmap->drawable.depth, UXA_CREATE_PIXMAP_FOR_MAP); if (!scratch) return FALSE; if (!intel_uxa_pixmap_is_offscreen(scratch)) { screen->DestroyPixmap(scratch); return FALSE; } ret = intel_uxa_pixmap_put_image(scratch, src, src_pitch, 0, 0, w, h); if (ret) { GCPtr gc = GetScratchGC(pixmap->drawable.depth, screen); if (gc) { ValidateGC(&pixmap->drawable, gc); (*gc->ops->CopyArea)(&scratch->drawable, &pixmap->drawable, gc, 0, 0, w, h, x, y); FreeScratchGC(gc); } else ret = FALSE; } (*screen->DestroyPixmap)(scratch); return ret; } } } static Bool intel_uxa_pixmap_get_image(PixmapPtr pixmap, int x, int y, int w, int h, char *dst, int dst_pitch) { struct intel_uxa_pixmap *priv = intel_uxa_get_pixmap_private(pixmap); int stride = intel_pixmap_pitch(pixmap); int cpp = pixmap->drawable.bitsPerPixel/8; /* assert(priv->tiling == I915_TILING_NONE); */ if (h == 1 || (dst_pitch == stride && w == pixmap->drawable.width)) { return drm_intel_bo_get_subdata(priv->bo, y*stride + x*cpp, (h-1)*stride + w*cpp, dst) == 0; } else { char *src; if (drm_intel_gem_bo_map_gtt(priv->bo)) return FALSE; src = (char *) priv->bo->virtual + y * stride + x * cpp; w *= cpp; do { memcpy(dst, src, w); src += stride; dst += dst_pitch; } while (--h); drm_intel_gem_bo_unmap_gtt(priv->bo); return TRUE; } } static Bool intel_uxa_get_image(PixmapPtr pixmap, int x, int y, int w, int h, char *dst, int dst_pitch) { struct intel_uxa_pixmap *priv; PixmapPtr scratch = NULL; Bool ret; /* The presumption is that we wish to keep the target hot, so * copy to a new bo and move that to the CPU in preference to * causing ping-pong of the original. * * Also the gpu is much faster at detiling. */ priv = intel_uxa_get_pixmap_private(pixmap); if (intel_uxa_pixmap_is_busy(priv) || priv->tiling != I915_TILING_NONE) { ScreenPtr screen = pixmap->drawable.pScreen; GCPtr gc; /* Copy to a linear buffer and pull. */ scratch = screen->CreatePixmap(screen, w, h, pixmap->drawable.depth, INTEL_CREATE_PIXMAP_TILING_NONE); if (!scratch) return FALSE; if (!intel_uxa_pixmap_is_offscreen(scratch)) { screen->DestroyPixmap(scratch); return FALSE; } gc = GetScratchGC(pixmap->drawable.depth, screen); if (!gc) { screen->DestroyPixmap(scratch); return FALSE; } ValidateGC(&pixmap->drawable, gc); gc->ops->CopyArea(&pixmap->drawable, &scratch->drawable, gc, x, y, w, h, 0, 0); FreeScratchGC(gc); intel_batch_submit(xf86ScreenToScrn(screen)); x = y = 0; pixmap = scratch; } ret = intel_uxa_pixmap_get_image(pixmap, x, y, w, h, dst, dst_pitch); if (scratch) scratch->drawable.pScreen->DestroyPixmap(scratch); return ret; } static CARD32 intel_cache_expire(OsTimerPtr timer, CARD32 now, pointer data) { intel_screen_private *intel = data; /* We just want to create and destroy a bo as this causes libdrm * to reap its caches. However, since we can't remove that buffer * from the cache due to its own activity, we want to use something * that we know we will reuse later. The most frequently reused buffer * we have is the batchbuffer, and the best way to trigger its * reallocation is to submit a flush. */ intel_batch_emit_flush(intel->scrn); intel_batch_submit(intel->scrn); return 0; } static void intel_flush_rendering(intel_screen_private *intel) { if (intel->needs_flush == 0) return; if (intel->has_kernel_flush) { intel_batch_submit(intel->scrn); drm_intel_bo_busy(intel->front_buffer); } else { intel_batch_emit_flush(intel->scrn); intel_batch_submit(intel->scrn); } intel->cache_expire = TimerSet(intel->cache_expire, 0, 3000, intel_cache_expire, intel); intel->needs_flush = 0; } static void intel_throttle(intel_screen_private *intel) { drmCommandNone(intel->drmSubFD, DRM_I915_GEM_THROTTLE); } void intel_uxa_block_handler(intel_screen_private *intel) { /* Emit a flush of the rendering cache, or on the 965 * and beyond rendering results may not hit the * framebuffer until significantly later. */ intel_flush_rendering(intel); intel_throttle(intel); } static PixmapPtr intel_uxa_create_pixmap(ScreenPtr screen, int w, int h, int depth, unsigned usage) { ScrnInfoPtr scrn = xf86ScreenToScrn(screen); intel_screen_private *intel = intel_get_screen_private(scrn); struct intel_uxa_pixmap *priv; PixmapPtr pixmap, new_pixmap = NULL; if (w > 32767 || h > 32767) return NullPixmap; if (depth == 1 || intel->force_fallback) return fbCreatePixmap(screen, w, h, depth, usage); if (usage == CREATE_PIXMAP_USAGE_GLYPH_PICTURE && w <= 32 && h <= 32) return fbCreatePixmap(screen, w, h, depth, usage); pixmap = fbCreatePixmap(screen, 0, 0, depth, usage); if (pixmap == NullPixmap) return pixmap; if (w && h) { unsigned int size, tiling; int stride; /* Always attempt to tile, compute_size() will remove the * tiling for pixmaps that are either too large or too small * to be effectively tiled. */ tiling = I915_TILING_X; if (usage & INTEL_CREATE_PIXMAP_TILING_Y) tiling = I915_TILING_Y; if (usage == UXA_CREATE_PIXMAP_FOR_MAP || usage & INTEL_CREATE_PIXMAP_TILING_NONE) tiling = I915_TILING_NONE; #ifdef CREATE_PIXMAP_USAGE_SHARED if (usage == CREATE_PIXMAP_USAGE_SHARED) tiling = I915_TILING_NONE; #endif /* if tiling is off force to none */ if (!intel->tiling) tiling = I915_TILING_NONE; if (tiling != I915_TILING_NONE && !(usage & INTEL_CREATE_PIXMAP_DRI2)) { if (h <= 4) tiling = I915_TILING_NONE; if (h <= 16 && tiling == I915_TILING_Y) tiling = I915_TILING_X; } size = intel_compute_size(intel, w, h, pixmap->drawable.bitsPerPixel, usage, &tiling, &stride); /* Fail very large allocations. Large BOs will tend to hit SW fallbacks * frequently, and also will tend to fail to successfully map when doing * SW fallbacks because we overcommit address space for BO access. */ if (size > intel->max_bo_size || stride >= KB(32)) goto fallback_pixmap; priv = calloc(1, sizeof (struct intel_uxa_pixmap)); if (priv == NULL) goto fallback_pixmap; if (usage == UXA_CREATE_PIXMAP_FOR_MAP) { priv->busy = 0; priv->bo = drm_intel_bo_alloc(intel->bufmgr, "pixmap", size, 0); } else { priv->busy = -1; priv->bo = drm_intel_bo_alloc_for_render(intel->bufmgr, "pixmap", size, 0); } if (!priv->bo) goto fallback_priv; if (tiling != I915_TILING_NONE) drm_intel_bo_set_tiling(priv->bo, &tiling, stride); priv->tiling = tiling; priv->offscreen = 1; list_init(&priv->batch); intel_uxa_set_pixmap_private(pixmap, priv); screen->ModifyPixmapHeader(pixmap, w, h, 0, 0, stride, NULL); } return pixmap; fallback_priv: free(priv); fallback_pixmap: fbDestroyPixmap(pixmap); if (new_pixmap) return new_pixmap; else return fbCreatePixmap(screen, w, h, depth, usage); } static Bool intel_uxa_destroy_pixmap(PixmapPtr pixmap) { if (pixmap->refcnt == 1) intel_uxa_set_pixmap_bo(pixmap, NULL); fbDestroyPixmap(pixmap); return TRUE; } Bool intel_uxa_create_screen_resources(ScreenPtr screen) { ScrnInfoPtr scrn = xf86ScreenToScrn(screen); PixmapPtr pixmap; intel_screen_private *intel = intel_get_screen_private(scrn); dri_bo *bo = intel->front_buffer, *old_bo; int old_width, old_height, old_pitch; if (!uxa_resources_init(screen)) return FALSE; if (drm_intel_gem_bo_map_gtt(bo)) return FALSE; pixmap = screen->GetScreenPixmap(screen); old_width = pixmap->drawable.width; old_height = pixmap->drawable.height; old_pitch = pixmap->devKind; old_bo = intel_uxa_get_pixmap_bo(pixmap); if (!screen->ModifyPixmapHeader(pixmap, scrn->virtualX, scrn->virtualY, -1, -1, intel->front_pitch, NULL)) return FALSE; intel_uxa_set_pixmap_bo(pixmap, bo); if (intel_uxa_get_pixmap_private(pixmap) == NULL) goto err; intel_uxa_get_pixmap_private(pixmap)->pinned |= PIN_SCANOUT; scrn->displayWidth = intel->front_pitch / intel->cpp; return TRUE; err: screen->ModifyPixmapHeader(pixmap, old_width, old_height, -1, -1, old_pitch, NULL); if (old_bo) intel_uxa_set_pixmap_bo(pixmap, old_bo); return FALSE; } #ifdef CREATE_PIXMAP_USAGE_SHARED static Bool intel_uxa_share_pixmap_backing(PixmapPtr ppix, ScreenPtr slave, void **fd_handle) { ScrnInfoPtr scrn = xf86ScreenToScrn(ppix->drawable.pScreen); intel_screen_private *intel = intel_get_screen_private(scrn); struct intel_uxa_pixmap *priv = intel_uxa_get_pixmap_private(ppix); unsigned int size, tiling, swizzle; dri_bo *bo = intel_uxa_get_pixmap_bo(ppix), *newbo; int stride; int handle; if (drm_intel_bo_references(intel->batch_bo, bo)) intel_batch_submit(intel->scrn); drm_intel_bo_get_tiling(bo, &tiling, &swizzle); if (tiling == I915_TILING_X) { if (priv->pinned) return FALSE; tiling = I915_TILING_NONE; size = intel_compute_size(intel, ppix->drawable.width, ppix->drawable.height, ppix->drawable.bitsPerPixel, INTEL_CREATE_PIXMAP_DRI2, &tiling, &stride); newbo = drm_intel_bo_alloc_for_render(intel->bufmgr, "pixmap", size, 0); if (tiling != I915_TILING_NONE) drm_intel_bo_set_tiling(newbo, &tiling, stride); priv->tiling = tiling; intel_uxa_set_pixmap_bo(ppix, newbo); ppix->drawable.pScreen->ModifyPixmapHeader(ppix, ppix->drawable.width, ppix->drawable.height, 0, 0, stride, NULL); bo = newbo; } drm_intel_bo_get_tiling(bo, &tiling, &swizzle); drm_intel_bo_gem_export_to_prime(bo, &handle); priv->pinned |= PIN_PRIME; *fd_handle = (void *)(long)handle; return TRUE; } static Bool intel_uxa_set_shared_pixmap_backing(PixmapPtr ppix, void *fd_handle) { ScrnInfoPtr scrn = xf86ScreenToScrn(ppix->drawable.pScreen); intel_screen_private *intel = intel_get_screen_private(scrn); dri_bo *bo; int ihandle = (int)(long)fd_handle; /* force untiled for now */ bo = drm_intel_bo_gem_create_from_prime(intel->bufmgr, ihandle, 0); if (!bo) return FALSE; intel_uxa_set_pixmap_bo(ppix, bo); close(ihandle); return TRUE; } #endif static void intel_limits_init(intel_screen_private *intel) { /* Limits are described in the BLT engine chapter under Graphics Data Size * Limitations, and the descriptions of SURFACE_STATE, 3DSTATE_BUFFER_INFO, * 3DSTATE_DRAWING_RECTANGLE, 3DSTATE_MAP_INFO, and 3DSTATE_MAP_INFO. * * i845 through i965 limits 2D rendering to 65536 lines and pitch of 32768. * * i965 limits 3D surface to (2*element size)-aligned offset if un-tiled. * i965 limits 3D surface to 4kB-aligned offset if tiled. * i965 limits 3D surfaces to w,h of ?,8192. * i965 limits 3D surface to pitch of 1B - 128kB. * i965 limits 3D surface pitch alignment to 1 or 2 times the element size. * i965 limits 3D surface pitch alignment to 512B if tiled. * i965 limits 3D destination drawing rect to w,h of 8192,8192. * * i915 limits 3D textures to 4B-aligned offset if un-tiled. * i915 limits 3D textures to ~4kB-aligned offset if tiled. * i915 limits 3D textures to width,height of 2048,2048. * i915 limits 3D textures to pitch of 16B - 8kB, in dwords. * i915 limits 3D destination to ~4kB-aligned offset if tiled. * i915 limits 3D destination to pitch of 16B - 8kB, in dwords, if un-tiled. * i915 limits 3D destination to pitch 64B-aligned if used with depth. * i915 limits 3D destination to pitch of 512B - 8kB, in tiles, if tiled. * i915 limits 3D destination to POT aligned pitch if tiled. * i915 limits 3D destination drawing rect to w,h of 2048,2048. * * i845 limits 3D textures to 4B-aligned offset if un-tiled. * i845 limits 3D textures to ~4kB-aligned offset if tiled. * i845 limits 3D textures to width,height of 2048,2048. * i845 limits 3D textures to pitch of 4B - 8kB, in dwords. * i845 limits 3D destination to 4B-aligned offset if un-tiled. * i845 limits 3D destination to ~4kB-aligned offset if tiled. * i845 limits 3D destination to pitch of 8B - 8kB, in dwords. * i845 limits 3D destination drawing rect to w,h of 2048,2048. * * For the tiled issues, the only tiled buffer we draw to should be * the front, which will have an appropriate pitch/offset already set up, * so UXA doesn't need to worry. */ if (INTEL_INFO(intel)->gen >= 040) { intel->accel_pixmap_offset_alignment = 4 * 2; intel->accel_max_x = 8192; intel->accel_max_y = 8192; } else { intel->accel_pixmap_offset_alignment = 4; intel->accel_max_x = 2048; intel->accel_max_y = 2048; } } static Bool intel_option_accel_none(intel_screen_private *intel) { const char *s; s = xf86GetOptValString(intel->Options, OPTION_ACCEL_METHOD); if (s == NULL) return IS_DEFAULT_ACCEL_METHOD(NOACCEL); return strcasecmp(s, "none") == 0; } static Bool intel_option_accel_blt(intel_screen_private *intel) { const char *s; s = xf86GetOptValString(intel->Options, OPTION_ACCEL_METHOD); if (s == NULL) return FALSE; return strcasecmp(s, "blt") == 0; } /** * Intialiazes the hardware for the 3D pipeline use in the 2D driver. * * Some state caching is performed to avoid redundant state emits. This * function is also responsible for marking the state as clobbered for DRI * clients. */ void IntelEmitInvarientState(ScrnInfoPtr scrn) { intel_screen_private *intel = intel_get_screen_private(scrn); /* If we've emitted our state since the last clobber by another client, * skip it. */ if (intel->last_3d != LAST_3D_OTHER) return; if (IS_GEN2(intel)) I830EmitInvarientState(scrn); else if IS_GEN3(intel) I915EmitInvarientState(scrn); } Bool intel_uxa_init(ScreenPtr screen) { ScrnInfoPtr scrn = xf86ScreenToScrn(screen); intel_screen_private *intel = intel_get_screen_private(scrn); intel_batch_init(scrn); if (INTEL_INFO(intel)->gen >= 040 && INTEL_INFO(intel)->gen < 0100) gen4_render_state_init(scrn); #if HAS_DIXREGISTERPRIVATEKEY if (!dixRegisterPrivateKey(&uxa_pixmap_index, PRIVATE_PIXMAP, 0)) #else if (!dixRequestPrivate(&uxa_pixmap_index, 0)) #endif return FALSE; intel_limits_init(intel); intel->uxa_driver = uxa_driver_alloc(); if (intel->uxa_driver == NULL) return FALSE; memset(intel->uxa_driver, 0, sizeof(*intel->uxa_driver)); intel->uxa_driver->uxa_major = 1; intel->uxa_driver->uxa_minor = 0; intel->prim_offset = 0; intel->vertex_count = 0; intel->vertex_offset = 0; intel->vertex_used = 0; intel->floats_per_vertex = 0; intel->last_floats_per_vertex = 0; intel->vertex_bo = NULL; intel->surface_used = 0; intel->surface_reloc = 0; /* Solid fill */ intel->uxa_driver->check_solid = intel_uxa_check_solid; intel->uxa_driver->prepare_solid = intel_uxa_prepare_solid; intel->uxa_driver->solid = intel_uxa_solid; intel->uxa_driver->done_solid = intel_uxa_done; /* Copy */ intel->uxa_driver->check_copy = intel_uxa_check_copy; intel->uxa_driver->prepare_copy = intel_uxa_prepare_copy; intel->uxa_driver->copy = intel_uxa_copy; intel->uxa_driver->done_copy = intel_uxa_done; /* Composite */ if (intel_option_accel_blt(intel)) { } else if (INTEL_INFO(intel)->gen < 030) { intel->uxa_driver->check_composite = i830_check_composite; intel->uxa_driver->check_composite_target = i830_check_composite_target; intel->uxa_driver->check_composite_texture = i830_check_composite_texture; intel->uxa_driver->prepare_composite = i830_prepare_composite; intel->uxa_driver->composite = i830_composite; intel->uxa_driver->done_composite = i830_done_composite; intel->vertex_flush = i830_vertex_flush; intel->batch_commit_notify = i830_batch_commit_notify; } else if (INTEL_INFO(intel)->gen < 040) { intel->uxa_driver->check_composite = i915_check_composite; intel->uxa_driver->check_composite_target = i915_check_composite_target; intel->uxa_driver->check_composite_texture = i915_check_composite_texture; intel->uxa_driver->prepare_composite = i915_prepare_composite; intel->uxa_driver->composite = i915_composite; intel->uxa_driver->done_composite = i830_done_composite; intel->vertex_flush = i915_vertex_flush; intel->batch_commit_notify = i915_batch_commit_notify; } else if (INTEL_INFO(intel)->gen < 0100) { intel->uxa_driver->check_composite = i965_check_composite; intel->uxa_driver->check_composite_texture = i965_check_composite_texture; intel->uxa_driver->prepare_composite = i965_prepare_composite; intel->uxa_driver->composite = i965_composite; intel->uxa_driver->done_composite = i830_done_composite; intel->vertex_flush = i965_vertex_flush; intel->batch_flush = i965_batch_flush; intel->batch_commit_notify = i965_batch_commit_notify; if (INTEL_INFO(intel)->gen < 050) { intel->context_switch = gen4_context_switch; } else if (INTEL_INFO(intel)->gen < 060) { intel->context_switch = gen5_context_switch; } else { intel->context_switch = gen6_context_switch; } } /* PutImage */ intel->uxa_driver->put_image = intel_uxa_put_image; intel->uxa_driver->get_image = intel_uxa_get_image; intel->uxa_driver->prepare_access = intel_uxa_prepare_access; intel->uxa_driver->finish_access = intel_uxa_finish_access; intel->uxa_driver->pixmap_is_offscreen = intel_uxa_pixmap_is_offscreen; screen->CreatePixmap = intel_uxa_create_pixmap; screen->DestroyPixmap = intel_uxa_destroy_pixmap; #ifdef CREATE_PIXMAP_USAGE_SHARED screen->SharePixmapBacking = intel_uxa_share_pixmap_backing; screen->SetSharedPixmapBacking = intel_uxa_set_shared_pixmap_backing; #endif if (!uxa_driver_init(screen, intel->uxa_driver)) { xf86DrvMsg(scrn->scrnIndex, X_ERROR, "UXA initialization failed\n"); free(intel->uxa_driver); return FALSE; } if (intel_option_accel_none(intel)) intel->force_fallback = 1; uxa_set_fallback_debug(screen, intel->fallback_debug); uxa_set_force_fallback(screen, intel->force_fallback); intel->flush_rendering = intel_flush_rendering; return TRUE; }