/************************************************************************** * * Copyright 2007 VMware, Inc. * All Rights Reserved. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the * "Software"), to deal in the Software without restriction, including * without limitation the rights to use, copy, modify, merge, publish, * distribute, sub license, and/or sell copies of the Software, and to * permit persons to whom the Software is furnished to do so, subject to * the following conditions: * * The above copyright notice and this permission notice (including the * next paragraph) shall be included in all copies or substantial portions * of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. * IN NO EVENT SHALL VMWARE AND/OR ITS SUPPLIERS BE LIABLE FOR * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. * **************************************************************************/ #ifndef U_INLINES_H #define U_INLINES_H #include "pipe/p_context.h" #include "pipe/p_defines.h" #include "pipe/p_shader_tokens.h" #include "pipe/p_state.h" #include "pipe/p_screen.h" #include "util/u_debug.h" #include "util/u_debug_describe.h" #include "util/u_debug_refcnt.h" #include "util/u_atomic.h" #include "util/u_box.h" #include "util/u_math.h" #ifdef __cplusplus extern "C" { #endif /* * Reference counting helper functions. */ static inline void pipe_reference_init(struct pipe_reference *dst, unsigned count) { p_atomic_set(&dst->count, count); } static inline boolean pipe_is_referenced(struct pipe_reference *src) { return p_atomic_read(&src->count) != 0; } /** * Update reference counting. * The old thing pointed to, if any, will be unreferenced. * Both 'dst' and 'src' may be NULL. * \return TRUE if the object's refcount hits zero and should be destroyed. */ static inline boolean pipe_reference_described(struct pipe_reference *dst, struct pipe_reference *src, debug_reference_descriptor get_desc) { if (dst != src) { /* bump the src.count first */ if (src) { ASSERTED int count = p_atomic_inc_return(&src->count); assert(count != 1); /* src had to be referenced */ debug_reference(src, get_desc, 1); } if (dst) { int count = p_atomic_dec_return(&dst->count); assert(count != -1); /* dst had to be referenced */ debug_reference(dst, get_desc, -1); if (!count) return true; } } return false; } static inline boolean pipe_reference(struct pipe_reference *dst, struct pipe_reference *src) { return pipe_reference_described(dst, src, (debug_reference_descriptor) debug_describe_reference); } static inline void pipe_surface_reference(struct pipe_surface **dst, struct pipe_surface *src) { struct pipe_surface *old_dst = *dst; if (pipe_reference_described(old_dst ? &old_dst->reference : NULL, src ? &src->reference : NULL, (debug_reference_descriptor) debug_describe_surface)) old_dst->context->surface_destroy(old_dst->context, old_dst); *dst = src; } /** * Similar to pipe_surface_reference() but always set the pointer to NULL * and pass in an explicit context. The explicit context avoids the problem * of using a deleted context's surface_destroy() method when freeing a surface * that's shared by multiple contexts. */ static inline void pipe_surface_release(struct pipe_context *pipe, struct pipe_surface **ptr) { struct pipe_surface *old = *ptr; if (pipe_reference_described(&old->reference, NULL, (debug_reference_descriptor) debug_describe_surface)) pipe->surface_destroy(pipe, old); *ptr = NULL; } static inline void pipe_resource_reference(struct pipe_resource **dst, struct pipe_resource *src) { struct pipe_resource *old_dst = *dst; if (pipe_reference_described(old_dst ? &old_dst->reference : NULL, src ? &src->reference : NULL, (debug_reference_descriptor) debug_describe_resource)) { /* Avoid recursion, which would prevent inlining this function */ do { struct pipe_resource *next = old_dst->next; old_dst->screen->resource_destroy(old_dst->screen, old_dst); old_dst = next; } while (pipe_reference_described(old_dst ? &old_dst->reference : NULL, NULL, (debug_reference_descriptor) debug_describe_resource)); } *dst = src; } /** * Same as pipe_surface_release, but used when pipe_context doesn't exist * anymore. */ static inline void pipe_surface_release_no_context(struct pipe_surface **ptr) { struct pipe_surface *surf = *ptr; if (pipe_reference_described(&surf->reference, NULL, (debug_reference_descriptor) debug_describe_surface)) { /* trivially destroy pipe_surface */ pipe_resource_reference(&surf->texture, NULL); free(surf); } *ptr = NULL; } /** * Set *dst to \p src with proper reference counting. * * The caller must guarantee that \p src and *dst were created in * the same context (if they exist), and that this must be the current context. */ static inline void pipe_sampler_view_reference(struct pipe_sampler_view **dst, struct pipe_sampler_view *src) { struct pipe_sampler_view *old_dst = *dst; if (pipe_reference_described(old_dst ? &old_dst->reference : NULL, src ? &src->reference : NULL, (debug_reference_descriptor) debug_describe_sampler_view)) old_dst->context->sampler_view_destroy(old_dst->context, old_dst); *dst = src; } static inline void pipe_so_target_reference(struct pipe_stream_output_target **dst, struct pipe_stream_output_target *src) { struct pipe_stream_output_target *old_dst = *dst; if (pipe_reference_described(old_dst ? &old_dst->reference : NULL, src ? &src->reference : NULL, (debug_reference_descriptor)debug_describe_so_target)) old_dst->context->stream_output_target_destroy(old_dst->context, old_dst); *dst = src; } static inline void pipe_vertex_buffer_unreference(struct pipe_vertex_buffer *dst) { if (dst->is_user_buffer) dst->buffer.user = NULL; else pipe_resource_reference(&dst->buffer.resource, NULL); } static inline void pipe_vertex_buffer_reference(struct pipe_vertex_buffer *dst, const struct pipe_vertex_buffer *src) { pipe_vertex_buffer_unreference(dst); if (!src->is_user_buffer) pipe_resource_reference(&dst->buffer.resource, src->buffer.resource); memcpy(dst, src, sizeof(*src)); } static inline void pipe_surface_reset(struct pipe_context *ctx, struct pipe_surface* ps, struct pipe_resource *pt, unsigned level, unsigned layer) { pipe_resource_reference(&ps->texture, pt); ps->format = pt->format; ps->width = u_minify(pt->width0, level); ps->height = u_minify(pt->height0, level); ps->u.tex.level = level; ps->u.tex.first_layer = ps->u.tex.last_layer = layer; ps->context = ctx; } static inline void pipe_surface_init(struct pipe_context *ctx, struct pipe_surface* ps, struct pipe_resource *pt, unsigned level, unsigned layer) { ps->texture = 0; pipe_reference_init(&ps->reference, 1); pipe_surface_reset(ctx, ps, pt, level, layer); } /* Return true if the surfaces are equal. */ static inline boolean pipe_surface_equal(struct pipe_surface *s1, struct pipe_surface *s2) { return s1->texture == s2->texture && s1->format == s2->format && (s1->texture->target != PIPE_BUFFER || (s1->u.buf.first_element == s2->u.buf.first_element && s1->u.buf.last_element == s2->u.buf.last_element)) && (s1->texture->target == PIPE_BUFFER || (s1->u.tex.level == s2->u.tex.level && s1->u.tex.first_layer == s2->u.tex.first_layer && s1->u.tex.last_layer == s2->u.tex.last_layer)); } /* * Convenience wrappers for screen buffer functions. */ /** * Create a new resource. * \param bind bitmask of PIPE_BIND_x flags * \param usage a PIPE_USAGE_x value */ static inline struct pipe_resource * pipe_buffer_create(struct pipe_screen *screen, unsigned bind, enum pipe_resource_usage usage, unsigned size) { struct pipe_resource buffer; memset(&buffer, 0, sizeof buffer); buffer.target = PIPE_BUFFER; buffer.format = PIPE_FORMAT_R8_UNORM; /* want TYPELESS or similar */ buffer.bind = bind; buffer.usage = usage; buffer.flags = 0; buffer.width0 = size; buffer.height0 = 1; buffer.depth0 = 1; buffer.array_size = 1; return screen->resource_create(screen, &buffer); } static inline struct pipe_resource * pipe_buffer_create_const0(struct pipe_screen *screen, unsigned bind, enum pipe_resource_usage usage, unsigned size) { struct pipe_resource buffer; memset(&buffer, 0, sizeof buffer); buffer.target = PIPE_BUFFER; buffer.format = PIPE_FORMAT_R8_UNORM; buffer.bind = bind; buffer.usage = usage; buffer.flags = screen->get_param(screen, PIPE_CAP_CONSTBUF0_FLAGS); buffer.width0 = size; buffer.height0 = 1; buffer.depth0 = 1; buffer.array_size = 1; return screen->resource_create(screen, &buffer); } /** * Map a range of a resource. * \param offset start of region, in bytes * \param length size of region, in bytes * \param access bitmask of PIPE_TRANSFER_x flags * \param transfer returns a transfer object */ static inline void * pipe_buffer_map_range(struct pipe_context *pipe, struct pipe_resource *buffer, unsigned offset, unsigned length, unsigned access, struct pipe_transfer **transfer) { struct pipe_box box; void *map; assert(offset < buffer->width0); assert(offset + length <= buffer->width0); assert(length); u_box_1d(offset, length, &box); map = pipe->transfer_map(pipe, buffer, 0, access, &box, transfer); if (!map) { return NULL; } return map; } /** * Map whole resource. * \param access bitmask of PIPE_TRANSFER_x flags * \param transfer returns a transfer object */ static inline void * pipe_buffer_map(struct pipe_context *pipe, struct pipe_resource *buffer, unsigned access, struct pipe_transfer **transfer) { return pipe_buffer_map_range(pipe, buffer, 0, buffer->width0, access, transfer); } static inline void pipe_buffer_unmap(struct pipe_context *pipe, struct pipe_transfer *transfer) { pipe->transfer_unmap(pipe, transfer); } static inline void pipe_buffer_flush_mapped_range(struct pipe_context *pipe, struct pipe_transfer *transfer, unsigned offset, unsigned length) { struct pipe_box box; int transfer_offset; assert(length); assert(transfer->box.x <= (int) offset); assert((int) (offset + length) <= transfer->box.x + transfer->box.width); /* Match old screen->buffer_flush_mapped_range() behaviour, where * offset parameter is relative to the start of the buffer, not the * mapped range. */ transfer_offset = offset - transfer->box.x; u_box_1d(transfer_offset, length, &box); pipe->transfer_flush_region(pipe, transfer, &box); } static inline void pipe_buffer_write(struct pipe_context *pipe, struct pipe_resource *buf, unsigned offset, unsigned size, const void *data) { /* Don't set any other usage bits. Drivers should derive them. */ pipe->buffer_subdata(pipe, buf, PIPE_TRANSFER_WRITE, offset, size, data); } /** * Special case for writing non-overlapping ranges. * * We can avoid GPU/CPU synchronization when writing range that has never * been written before. */ static inline void pipe_buffer_write_nooverlap(struct pipe_context *pipe, struct pipe_resource *buf, unsigned offset, unsigned size, const void *data) { pipe->buffer_subdata(pipe, buf, (PIPE_TRANSFER_WRITE | PIPE_TRANSFER_UNSYNCHRONIZED), offset, size, data); } /** * Create a new resource and immediately put data into it * \param bind bitmask of PIPE_BIND_x flags * \param usage bitmask of PIPE_USAGE_x flags */ static inline struct pipe_resource * pipe_buffer_create_with_data(struct pipe_context *pipe, unsigned bind, enum pipe_resource_usage usage, unsigned size, const void *ptr) { struct pipe_resource *res = pipe_buffer_create(pipe->screen, bind, usage, size); pipe_buffer_write_nooverlap(pipe, res, 0, size, ptr); return res; } static inline void pipe_buffer_read(struct pipe_context *pipe, struct pipe_resource *buf, unsigned offset, unsigned size, void *data) { struct pipe_transfer *src_transfer; ubyte *map; map = (ubyte *) pipe_buffer_map_range(pipe, buf, offset, size, PIPE_TRANSFER_READ, &src_transfer); if (!map) return; memcpy(data, map, size); pipe_buffer_unmap(pipe, src_transfer); } /** * Map a resource for reading/writing. * \param access bitmask of PIPE_TRANSFER_x flags */ static inline void * pipe_transfer_map(struct pipe_context *context, struct pipe_resource *resource, unsigned level, unsigned layer, unsigned access, unsigned x, unsigned y, unsigned w, unsigned h, struct pipe_transfer **transfer) { struct pipe_box box; u_box_2d_zslice(x, y, layer, w, h, &box); return context->transfer_map(context, resource, level, access, &box, transfer); } /** * Map a 3D (texture) resource for reading/writing. * \param access bitmask of PIPE_TRANSFER_x flags */ static inline void * pipe_transfer_map_3d(struct pipe_context *context, struct pipe_resource *resource, unsigned level, unsigned access, unsigned x, unsigned y, unsigned z, unsigned w, unsigned h, unsigned d, struct pipe_transfer **transfer) { struct pipe_box box; u_box_3d(x, y, z, w, h, d, &box); return context->transfer_map(context, resource, level, access, &box, transfer); } static inline void pipe_transfer_unmap(struct pipe_context *context, struct pipe_transfer *transfer) { context->transfer_unmap(context, transfer); } static inline void pipe_set_constant_buffer(struct pipe_context *pipe, enum pipe_shader_type shader, uint index, struct pipe_resource *buf) { if (buf) { struct pipe_constant_buffer cb; cb.buffer = buf; cb.buffer_offset = 0; cb.buffer_size = buf->width0; cb.user_buffer = NULL; pipe->set_constant_buffer(pipe, shader, index, &cb); } else { pipe->set_constant_buffer(pipe, shader, index, NULL); } } /** * Get the polygon offset enable/disable flag for the given polygon fill mode. * \param fill_mode one of PIPE_POLYGON_MODE_POINT/LINE/FILL */ static inline boolean util_get_offset(const struct pipe_rasterizer_state *templ, unsigned fill_mode) { switch(fill_mode) { case PIPE_POLYGON_MODE_POINT: return templ->offset_point; case PIPE_POLYGON_MODE_LINE: return templ->offset_line; case PIPE_POLYGON_MODE_FILL: return templ->offset_tri; default: assert(0); return FALSE; } } static inline float util_get_min_point_size(const struct pipe_rasterizer_state *state) { /* The point size should be clamped to this value at the rasterizer stage. */ return !state->point_quad_rasterization && !state->point_smooth && !state->multisample ? 1.0f : 0.0f; } static inline void util_query_clear_result(union pipe_query_result *result, unsigned type) { switch (type) { case PIPE_QUERY_OCCLUSION_PREDICATE: case PIPE_QUERY_OCCLUSION_PREDICATE_CONSERVATIVE: case PIPE_QUERY_SO_OVERFLOW_PREDICATE: case PIPE_QUERY_SO_OVERFLOW_ANY_PREDICATE: case PIPE_QUERY_GPU_FINISHED: result->b = FALSE; break; case PIPE_QUERY_OCCLUSION_COUNTER: case PIPE_QUERY_TIMESTAMP: case PIPE_QUERY_TIME_ELAPSED: case PIPE_QUERY_PRIMITIVES_GENERATED: case PIPE_QUERY_PRIMITIVES_EMITTED: result->u64 = 0; break; case PIPE_QUERY_SO_STATISTICS: memset(&result->so_statistics, 0, sizeof(result->so_statistics)); break; case PIPE_QUERY_TIMESTAMP_DISJOINT: memset(&result->timestamp_disjoint, 0, sizeof(result->timestamp_disjoint)); break; case PIPE_QUERY_PIPELINE_STATISTICS: memset(&result->pipeline_statistics, 0, sizeof(result->pipeline_statistics)); break; default: memset(result, 0, sizeof(*result)); } } /** Convert PIPE_TEXTURE_x to TGSI_TEXTURE_x */ static inline enum tgsi_texture_type util_pipe_tex_to_tgsi_tex(enum pipe_texture_target pipe_tex_target, unsigned nr_samples) { switch (pipe_tex_target) { case PIPE_BUFFER: return TGSI_TEXTURE_BUFFER; case PIPE_TEXTURE_1D: assert(nr_samples <= 1); return TGSI_TEXTURE_1D; case PIPE_TEXTURE_2D: return nr_samples > 1 ? TGSI_TEXTURE_2D_MSAA : TGSI_TEXTURE_2D; case PIPE_TEXTURE_RECT: assert(nr_samples <= 1); return TGSI_TEXTURE_RECT; case PIPE_TEXTURE_3D: assert(nr_samples <= 1); return TGSI_TEXTURE_3D; case PIPE_TEXTURE_CUBE: assert(nr_samples <= 1); return TGSI_TEXTURE_CUBE; case PIPE_TEXTURE_1D_ARRAY: assert(nr_samples <= 1); return TGSI_TEXTURE_1D_ARRAY; case PIPE_TEXTURE_2D_ARRAY: return nr_samples > 1 ? TGSI_TEXTURE_2D_ARRAY_MSAA : TGSI_TEXTURE_2D_ARRAY; case PIPE_TEXTURE_CUBE_ARRAY: return TGSI_TEXTURE_CUBE_ARRAY; default: assert(0 && "unexpected texture target"); return TGSI_TEXTURE_UNKNOWN; } } static inline void util_copy_constant_buffer(struct pipe_constant_buffer *dst, const struct pipe_constant_buffer *src) { if (src) { pipe_resource_reference(&dst->buffer, src->buffer); dst->buffer_offset = src->buffer_offset; dst->buffer_size = src->buffer_size; dst->user_buffer = src->user_buffer; } else { pipe_resource_reference(&dst->buffer, NULL); dst->buffer_offset = 0; dst->buffer_size = 0; dst->user_buffer = NULL; } } static inline void util_copy_shader_buffer(struct pipe_shader_buffer *dst, const struct pipe_shader_buffer *src) { if (src) { pipe_resource_reference(&dst->buffer, src->buffer); dst->buffer_offset = src->buffer_offset; dst->buffer_size = src->buffer_size; } else { pipe_resource_reference(&dst->buffer, NULL); dst->buffer_offset = 0; dst->buffer_size = 0; } } static inline void util_copy_image_view(struct pipe_image_view *dst, const struct pipe_image_view *src) { if (src) { pipe_resource_reference(&dst->resource, src->resource); dst->format = src->format; dst->access = src->access; dst->shader_access = src->shader_access; dst->u = src->u; } else { pipe_resource_reference(&dst->resource, NULL); dst->format = PIPE_FORMAT_NONE; dst->access = 0; dst->shader_access = 0; memset(&dst->u, 0, sizeof(dst->u)); } } static inline unsigned util_max_layer(const struct pipe_resource *r, unsigned level) { switch (r->target) { case PIPE_TEXTURE_3D: return u_minify(r->depth0, level) - 1; case PIPE_TEXTURE_CUBE: assert(r->array_size == 6); /* fall-through */ case PIPE_TEXTURE_1D_ARRAY: case PIPE_TEXTURE_2D_ARRAY: case PIPE_TEXTURE_CUBE_ARRAY: return r->array_size - 1; default: return 0; } } static inline unsigned util_num_layers(const struct pipe_resource *r, unsigned level) { return util_max_layer(r, level) + 1; } static inline bool util_texrange_covers_whole_level(const struct pipe_resource *tex, unsigned level, unsigned x, unsigned y, unsigned z, unsigned width, unsigned height, unsigned depth) { return x == 0 && y == 0 && z == 0 && width == u_minify(tex->width0, level) && height == u_minify(tex->height0, level) && depth == util_num_layers(tex, level); } static inline bool util_logicop_reads_dest(enum pipe_logicop op) { switch (op) { case PIPE_LOGICOP_NOR: case PIPE_LOGICOP_AND_INVERTED: case PIPE_LOGICOP_AND_REVERSE: case PIPE_LOGICOP_INVERT: case PIPE_LOGICOP_XOR: case PIPE_LOGICOP_NAND: case PIPE_LOGICOP_AND: case PIPE_LOGICOP_EQUIV: case PIPE_LOGICOP_NOOP: case PIPE_LOGICOP_OR_INVERTED: case PIPE_LOGICOP_OR_REVERSE: case PIPE_LOGICOP_OR: return true; case PIPE_LOGICOP_CLEAR: case PIPE_LOGICOP_COPY_INVERTED: case PIPE_LOGICOP_COPY: case PIPE_LOGICOP_SET: return false; } unreachable("bad logicop"); } static inline struct pipe_context * pipe_create_multimedia_context(struct pipe_screen *screen) { unsigned flags = 0; if (!screen->get_param(screen, PIPE_CAP_GRAPHICS)) flags |= PIPE_CONTEXT_COMPUTE_ONLY; return screen->context_create(screen, NULL, flags); } static inline unsigned util_res_sample_count(struct pipe_resource *res) { return res->nr_samples > 0 ? res->nr_samples : 1; } #ifdef __cplusplus } #endif #endif /* U_INLINES_H */