/* * Copyright (C) 2019 Alyssa Rosenzweig * Copyright (C) 2014-2017 Broadcom * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. * */ #ifndef __PAN_JOB_H__ #define __PAN_JOB_H__ #include "util/u_dynarray.h" #include "pipe/p_state.h" #include "pan_cs.h" #include "pan_pool.h" #include "pan_resource.h" #include "pan_scoreboard.h" /* panfrost_batch_fence is the out fence of a batch that users or other batches * might want to wait on. The batch fence lifetime is different from the batch * one as want will certainly want to wait upon the fence after the batch has * been submitted (which is when panfrost_batch objects are freed). */ struct panfrost_batch_fence { /* Refcounting object for the fence. */ struct pipe_reference reference; /* Batch that created this fence object. Will become NULL at batch * submission time. This field is mainly here to know whether the * batch has been flushed or not. */ struct panfrost_batch *batch; }; /* A panfrost_batch corresponds to a bound FBO we're rendering to, * collecting over multiple draws. */ struct panfrost_batch { struct panfrost_context *ctx; struct pipe_framebuffer_state key; /* Buffers cleared (PIPE_CLEAR_* bitmask) */ unsigned clear; /* Buffers drawn */ unsigned draws; /* Buffers read */ unsigned read; /* Packed clear values, indexed by both render target as well as word. * Essentially, a single pixel is packed, with some padding to bring it * up to a 32-bit interval; that pixel is then duplicated over to fill * all 16-bytes */ uint32_t clear_color[PIPE_MAX_COLOR_BUFS][4]; float clear_depth; unsigned clear_stencil; /* Amount of thread local storage required per thread */ unsigned stack_size; /* Amount of shared memory needed per workgroup (for compute) */ unsigned shared_size; /* The bounding box covered by this job, taking scissors into account. * Basically, the bounding box we have to run fragment shaders for */ unsigned minx, miny; unsigned maxx, maxy; /* BOs referenced not in the pool */ struct hash_table *bos; /* Pool owned by this batch (released when the batch is released) used for temporary descriptors */ struct pan_pool pool; /* Pool also owned by this batch that is not CPU mapped (created as * INVISIBLE) used for private GPU-internal structures, particularly * varyings */ struct pan_pool invisible_pool; /* Job scoreboarding state */ struct pan_scoreboard scoreboard; /* Polygon list bound to the batch, or NULL if none bound yet */ struct panfrost_bo *polygon_list; /* Scratchpad BO bound to the batch, or NULL if none bound yet */ struct panfrost_bo *scratchpad; /* Shared memory BO bound to the batch, or NULL if none bound yet */ struct panfrost_bo *shared_memory; /* Framebuffer descriptor. */ struct panfrost_ptr framebuffer; /* Thread local storage descriptor. */ struct panfrost_ptr tls; /* Tiler context */ struct pan_tiler_context tiler_ctx; /* Output sync object. Only valid when submitted is true. */ struct panfrost_batch_fence *out_sync; /* Batch dependencies */ struct util_dynarray dependencies; /* Indirect draw data */ struct panfrost_ptr indirect_draw_ctx; unsigned indirect_draw_job_id; }; /* Functions for managing the above */ void panfrost_batch_fence_unreference(struct panfrost_batch_fence *fence); void panfrost_batch_fence_reference(struct panfrost_batch_fence *batch); struct panfrost_batch * panfrost_get_batch_for_fbo(struct panfrost_context *ctx); struct panfrost_batch * panfrost_get_fresh_batch_for_fbo(struct panfrost_context *ctx); void panfrost_batch_init(struct panfrost_context *ctx); void panfrost_batch_add_bo(struct panfrost_batch *batch, struct panfrost_bo *bo, uint32_t flags); struct panfrost_bo * panfrost_batch_create_bo(struct panfrost_batch *batch, size_t size, uint32_t create_flags, uint32_t access_flags); void panfrost_flush_all_batches(struct panfrost_context *ctx); bool panfrost_pending_batches_access_bo(struct panfrost_context *ctx, const struct panfrost_bo *bo); void panfrost_flush_batches_accessing_bo(struct panfrost_context *ctx, struct panfrost_bo *bo, bool flush_readers); void panfrost_batch_set_requirements(struct panfrost_batch *batch); void panfrost_batch_adjust_stack_size(struct panfrost_batch *batch); struct panfrost_bo * panfrost_batch_get_scratchpad(struct panfrost_batch *batch, unsigned size, unsigned thread_tls_alloc, unsigned core_count); struct panfrost_bo * panfrost_batch_get_shared_memory(struct panfrost_batch *batch, unsigned size, unsigned workgroup_count); void panfrost_batch_clear(struct panfrost_batch *batch, unsigned buffers, const union pipe_color_union *color, double depth, unsigned stencil); void panfrost_batch_union_scissor(struct panfrost_batch *batch, unsigned minx, unsigned miny, unsigned maxx, unsigned maxy); void panfrost_batch_intersection_scissor(struct panfrost_batch *batch, unsigned minx, unsigned miny, unsigned maxx, unsigned maxy); mali_ptr panfrost_batch_get_bifrost_tiler(struct panfrost_batch *batch, unsigned vertex_count); mali_ptr panfrost_batch_reserve_framebuffer(struct panfrost_batch *batch); mali_ptr panfrost_batch_reserve_tls(struct panfrost_batch *batch, bool compute); #endif