/* * Copyright (C) 2019 Alyssa Rosenzweig * Copyright (C) 2014-2017 Broadcom * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. * */ #ifndef __PAN_JOB_H__ #define __PAN_JOB_H__ #include "util/u_dynarray.h" #include "pipe/p_state.h" #include "pan_allocate.h" #include "pan_resource.h" /* panfrost_batch_fence is the out fence of a batch that users or other batches * might want to wait on. The batch fence lifetime is different from the batch * one as want will certainly want to wait upon the fence after the batch has * been submitted (which is when panfrost_batch objects are freed). */ struct panfrost_batch_fence { /* Refcounting object for the fence. */ struct pipe_reference reference; /* Batch that created this fence object. Will become NULL at batch * submission time. This field is mainly here to know whether the * batch has been flushed or not. */ struct panfrost_batch *batch; /* Context this fence is attached to. We need both ctx and batch, as * the batch will go away after it's been submitted, but the fence * will stay a bit longer. */ struct panfrost_context *ctx; /* Sync object backing this fence. */ uint32_t syncobj; /* Cached value of the signaled state to avoid calling WAIT_SYNCOBJs * when we know the fence has already been signaled. */ bool signaled; }; #define PAN_REQ_MSAA (1 << 0) #define PAN_REQ_DEPTH_WRITE (1 << 1) /* A panfrost_batch corresponds to a bound FBO we're rendering to, * collecting over multiple draws. */ struct panfrost_batch { struct panfrost_context *ctx; struct pipe_framebuffer_state key; /* Buffers cleared (PIPE_CLEAR_* bitmask) */ unsigned clear; /* Packed clear values, indexed by both render target as well as word. * Essentially, a single pixel is packed, with some padding to bring it * up to a 32-bit interval; that pixel is then duplicated over to fill * all 16-bytes */ uint32_t clear_color[PIPE_MAX_COLOR_BUFS][4]; float clear_depth; unsigned clear_stencil; /* Amount of thread local storage required per thread */ unsigned stack_size; /* Amount of shared memory needed per workgroup (for compute) */ unsigned shared_size; /* Whether this job uses the corresponding requirement (PAN_REQ_* * bitmask) */ unsigned requirements; /* The bounding box covered by this job, taking scissors into account. * Basically, the bounding box we have to run fragment shaders for */ unsigned minx, miny; unsigned maxx, maxy; /* The first job in the batch */ mali_ptr first_job; /* The number of jobs in the primary batch, essentially */ unsigned job_index; /* A CPU-side pointer to the previous job for next_job linking */ struct mali_job_descriptor_header *prev_job; /* The dependency for tiler jobs (i.e. the index of the last emitted * tiler job, or zero if none have been emitted) */ unsigned tiler_dep; /* The job index of the WRITE_VALUE job (before it has been created) */ unsigned write_value_index; /* BOs referenced -- will be used for flushing logic */ struct hash_table *bos; /* Current transient BO */ struct panfrost_bo *transient_bo; /* Within the topmost transient BO, how much has been used? */ unsigned transient_offset; /* Polygon list bound to the batch, or NULL if none bound yet */ struct panfrost_bo *polygon_list; /* Scratchpad BO bound to the batch, or NULL if none bound yet */ struct panfrost_bo *scratchpad; /* Shared memory BO bound to the batch, or NULL if none bound yet */ struct panfrost_bo *shared_memory; /* Tiler heap BO bound to the batch, or NULL if none bound yet */ struct panfrost_bo *tiler_heap; /* Dummy tiler BO bound to the batch, or NULL if none bound yet */ struct panfrost_bo *tiler_dummy; /* Framebuffer descriptor. */ struct panfrost_transfer framebuffer; /* Bifrost tiler meta descriptor. */ mali_ptr tiler_meta; /* Output sync object. Only valid when submitted is true. */ struct panfrost_batch_fence *out_sync; /* Batch dependencies */ struct util_dynarray dependencies; }; /* Functions for managing the above */ void panfrost_batch_fence_unreference(struct panfrost_batch_fence *fence); void panfrost_batch_fence_reference(struct panfrost_batch_fence *batch); struct panfrost_batch * panfrost_get_batch_for_fbo(struct panfrost_context *ctx); struct panfrost_batch * panfrost_get_fresh_batch_for_fbo(struct panfrost_context *ctx); void panfrost_batch_init(struct panfrost_context *ctx); void panfrost_batch_add_bo(struct panfrost_batch *batch, struct panfrost_bo *bo, uint32_t flags); void panfrost_batch_add_fbo_bos(struct panfrost_batch *batch); struct panfrost_bo * panfrost_batch_create_bo(struct panfrost_batch *batch, size_t size, uint32_t create_flags, uint32_t access_flags); void panfrost_flush_all_batches(struct panfrost_context *ctx, bool wait); bool panfrost_pending_batches_access_bo(struct panfrost_context *ctx, const struct panfrost_bo *bo); void panfrost_flush_batches_accessing_bo(struct panfrost_context *ctx, struct panfrost_bo *bo, uint32_t flags); void panfrost_batch_set_requirements(struct panfrost_batch *batch); void panfrost_batch_adjust_stack_size(struct panfrost_batch *batch); struct panfrost_bo * panfrost_batch_get_scratchpad(struct panfrost_batch *batch, unsigned shift, unsigned thread_tls_alloc, unsigned core_count); struct panfrost_bo * panfrost_batch_get_shared_memory(struct panfrost_batch *batch, unsigned size, unsigned workgroup_count); mali_ptr panfrost_batch_get_polygon_list(struct panfrost_batch *batch, unsigned size); struct panfrost_bo * panfrost_batch_get_tiler_heap(struct panfrost_batch *batch); struct panfrost_bo * panfrost_batch_get_tiler_dummy(struct panfrost_batch *batch); void panfrost_batch_clear(struct panfrost_batch *batch, unsigned buffers, const union pipe_color_union *color, double depth, unsigned stencil); void panfrost_batch_union_scissor(struct panfrost_batch *batch, unsigned minx, unsigned miny, unsigned maxx, unsigned maxy); void panfrost_batch_intersection_scissor(struct panfrost_batch *batch, unsigned minx, unsigned miny, unsigned maxx, unsigned maxy); /* Scoreboarding */ unsigned panfrost_new_job( struct panfrost_batch *batch, enum mali_job_type type, bool barrier, unsigned local_dep, void *payload, size_t payload_size, bool inject); void panfrost_scoreboard_initialize_tiler(struct panfrost_batch *batch); bool panfrost_batch_is_scanout(struct panfrost_batch *batch); mali_ptr panfrost_batch_get_tiler_meta(struct panfrost_batch *batch, unsigned vertex_count); #endif