/* * Copyright © 2017 Intel Corporation * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS * IN THE SOFTWARE. */ #ifndef IRIS_BATCH_DOT_H #define IRIS_BATCH_DOT_H #include #include #include "i915_drm.h" #include "common/gen_decoder.h" /* The kernel assumes batchbuffers are smaller than 256kB. */ #define MAX_BATCH_SIZE (256 * 1024) struct iris_address { struct iris_bo *bo; uint64_t offset; bool write; }; struct iris_batch_buffer { struct iris_bo *bo; void *map; void *map_next; struct iris_bo *partial_bo; unsigned partial_bytes; }; struct iris_batch { struct iris_screen *screen; struct iris_vtable *vtbl; struct pipe_debug_callback *dbg; /** Current batchbuffer being queued up. */ struct iris_batch_buffer cmdbuf; /** Last BO submitted to the hardware. Used for glFinish(). */ struct iris_bo *last_cmd_bo; uint32_t hw_ctx_id; /** Which ring this batch targets - a I915_EXEC_RING_MASK value */ uint8_t ring; bool no_wrap; /** The validation list */ struct drm_i915_gem_exec_object2 *validation_list; struct iris_bo **exec_bos; int exec_count; int exec_array_size; /** The amount of aperture space (in bytes) used by all exec_bos */ int aperture_space; struct { /** * Set of struct brw_bo * that have been rendered to within this * batchbuffer and would need flushing before being used from another * cache domain that isn't coherent with it (i.e. the sampler). */ struct hash_table *render; /** * Set of struct brw_bo * that have been used as a depth buffer within * this batchbuffer and would need flushing before being used from * another cache domain that isn't coherent with it (i.e. the sampler). */ struct set *depth; } cache; #if DEBUG /** Map from batch offset to iris_alloc_state data (with DEBUG_BATCH) */ // XXX: unused struct hash_table *state_sizes; struct gen_batch_decode_ctx decoder; #endif }; void iris_init_batch(struct iris_batch *batch, struct iris_screen *screen, struct iris_vtable *vtbl, struct pipe_debug_callback *dbg, uint8_t ring); void iris_batch_free(struct iris_batch *batch); void iris_require_command_space(struct iris_batch *batch, unsigned size); void *iris_get_command_space(struct iris_batch *batch, unsigned bytes); void iris_batch_emit(struct iris_batch *batch, const void *data, unsigned size); int _iris_batch_flush_fence(struct iris_batch *batch, int in_fence_fd, int *out_fence_fd, const char *file, int line); #define iris_batch_flush_fence(batch, in_fence_fd, out_fence_fd) \ _iris_batch_flush_fence((batch), (in_fence_fd), (out_fence_fd), \ __FILE__, __LINE__) #define iris_batch_flush(batch) iris_batch_flush_fence((batch), -1, NULL) bool iris_batch_references(struct iris_batch *batch, struct iris_bo *bo); #define RELOC_WRITE EXEC_OBJECT_WRITE void iris_use_pinned_bo(struct iris_batch *batch, struct iris_bo *bo, bool writable); #endif