/* * Copyright 2010 Jerome Glisse * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * on the rights to use, copy, modify, merge, publish, distribute, sub * license, and/or sell copies of the Software, and to permit persons to whom * the Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL * THE AUTHOR(S) AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM, * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE * USE OR OTHER DEALINGS IN THE SOFTWARE. */ #include "r600_pipe.h" #include "r600_public.h" #include "r600_isa.h" #include "evergreen_compute.h" #include "r600d.h" #include "sb/sb_public.h" #include #include "pipe/p_shader_tokens.h" #include "util/u_blitter.h" #include "util/u_debug.h" #include "util/u_format_s3tc.h" #include "util/u_memory.h" #include "util/u_simple_shaders.h" #include "util/u_upload_mgr.h" #include "util/u_math.h" #include "vl/vl_decoder.h" #include "vl/vl_video_buffer.h" #include "radeon/radeon_uvd.h" #include "os/os_time.h" static const struct debug_named_value debug_options[] = { /* logging */ { "texdepth", DBG_TEX_DEPTH, "Print texture depth info" }, { "compute", DBG_COMPUTE, "Print compute info" }, { "vm", DBG_VM, "Print virtual addresses when creating resources" }, { "trace_cs", DBG_TRACE_CS, "Trace cs and write rlockup_.c file with faulty cs" }, /* shaders */ { "fs", DBG_FS, "Print fetch shaders" }, { "vs", DBG_VS, "Print vertex shaders" }, { "gs", DBG_GS, "Print geometry shaders" }, { "ps", DBG_PS, "Print pixel shaders" }, { "cs", DBG_CS, "Print compute shaders" }, /* features */ { "nohyperz", DBG_NO_HYPERZ, "Disable Hyper-Z" }, #if defined(R600_USE_LLVM) { "nollvm", DBG_NO_LLVM, "Disable the LLVM shader compiler" }, #endif { "nocpdma", DBG_NO_CP_DMA, "Disable CP DMA" }, { "nodma", DBG_NO_ASYNC_DMA, "Disable asynchronous DMA" }, /* GL uses the word INVALIDATE, gallium uses the word DISCARD */ { "noinvalrange", DBG_NO_DISCARD_RANGE, "Disable handling of INVALIDATE_RANGE map flags" }, /* shader backend */ { "sb", DBG_SB, "Enable optimization of graphics shaders" }, { "sbcl", DBG_SB_CS, "Enable optimization of compute shaders" }, { "sbdry", DBG_SB_DRY_RUN, "Don't use optimized bytecode (just print the dumps)" }, { "sbstat", DBG_SB_STAT, "Print optimization statistics for shaders" }, { "sbdump", DBG_SB_DUMP, "Print IR dumps after some optimization passes" }, { "sbnofallback", DBG_SB_NO_FALLBACK, "Abort on errors instead of fallback" }, { "sbdisasm", DBG_SB_DISASM, "Use sb disassembler for shader dumps" }, { "sbsafemath", DBG_SB_SAFEMATH, "Disable unsafe math optimizations" }, DEBUG_NAMED_VALUE_END /* must be last */ }; /* * pipe_context */ static struct r600_fence *r600_create_fence(struct r600_context *rctx) { struct r600_screen *rscreen = rctx->screen; struct r600_fence *fence = NULL; pipe_mutex_lock(rscreen->fences.mutex); if (!rscreen->fences.bo) { /* Create the shared buffer object */ rscreen->fences.bo = (struct r600_resource*) pipe_buffer_create(&rscreen->screen, PIPE_BIND_CUSTOM, PIPE_USAGE_STAGING, 4096); if (!rscreen->fences.bo) { R600_ERR("r600: failed to create bo for fence objects\n"); goto out; } rscreen->fences.data = r600_buffer_mmap_sync_with_rings(rctx, rscreen->fences.bo, PIPE_TRANSFER_READ_WRITE); } if (!LIST_IS_EMPTY(&rscreen->fences.pool)) { struct r600_fence *entry; /* Try to find a freed fence that has been signalled */ LIST_FOR_EACH_ENTRY(entry, &rscreen->fences.pool, head) { if (rscreen->fences.data[entry->index] != 0) { LIST_DELINIT(&entry->head); fence = entry; break; } } } if (!fence) { /* Allocate a new fence */ struct r600_fence_block *block; unsigned index; if ((rscreen->fences.next_index + 1) >= 1024) { R600_ERR("r600: too many concurrent fences\n"); goto out; } index = rscreen->fences.next_index++; if (!(index % FENCE_BLOCK_SIZE)) { /* Allocate a new block */ block = CALLOC_STRUCT(r600_fence_block); if (block == NULL) goto out; LIST_ADD(&block->head, &rscreen->fences.blocks); } else { block = LIST_ENTRY(struct r600_fence_block, rscreen->fences.blocks.next, head); } fence = &block->fences[index % FENCE_BLOCK_SIZE]; fence->index = index; } pipe_reference_init(&fence->reference, 1); rscreen->fences.data[fence->index] = 0; r600_context_emit_fence(rctx, rscreen->fences.bo, fence->index, 1); /* Create a dummy BO so that fence_finish without a timeout can sleep waiting for completion */ fence->sleep_bo = (struct r600_resource*) pipe_buffer_create(&rctx->screen->screen, PIPE_BIND_CUSTOM, PIPE_USAGE_STAGING, 1); /* Add the fence as a dummy relocation. */ r600_context_bo_reloc(rctx, &rctx->rings.gfx, fence->sleep_bo, RADEON_USAGE_READWRITE); out: pipe_mutex_unlock(rscreen->fences.mutex); return fence; } static void r600_flush(struct pipe_context *ctx, unsigned flags) { struct r600_context *rctx = (struct r600_context *)ctx; struct pipe_query *render_cond = NULL; unsigned render_cond_mode = 0; boolean render_cond_cond = FALSE; if (rctx->rings.gfx.cs->cdw == rctx->initial_gfx_cs_size) return; rctx->rings.gfx.flushing = true; /* Disable render condition. */ if (rctx->current_render_cond) { render_cond = rctx->current_render_cond; render_cond_cond = rctx->current_render_cond_cond; render_cond_mode = rctx->current_render_cond_mode; ctx->render_condition(ctx, NULL, FALSE, 0); } r600_context_flush(rctx, flags); rctx->rings.gfx.flushing = false; r600_begin_new_cs(rctx); /* Re-enable render condition. */ if (render_cond) { ctx->render_condition(ctx, render_cond, render_cond_cond, render_cond_mode); } rctx->initial_gfx_cs_size = rctx->rings.gfx.cs->cdw; } static void r600_flush_from_st(struct pipe_context *ctx, struct pipe_fence_handle **fence, unsigned flags) { struct r600_context *rctx = (struct r600_context *)ctx; struct r600_fence **rfence = (struct r600_fence**)fence; unsigned fflags; fflags = flags & PIPE_FLUSH_END_OF_FRAME ? RADEON_FLUSH_END_OF_FRAME : 0; if (rfence) { *rfence = r600_create_fence(rctx); } /* flush gfx & dma ring, order does not matter as only one can be live */ if (rctx->rings.dma.cs) { rctx->rings.dma.flush(rctx, fflags); } rctx->rings.gfx.flush(rctx, fflags); } static void r600_flush_gfx_ring(void *ctx, unsigned flags) { r600_flush((struct pipe_context*)ctx, flags); } static void r600_flush_dma_ring(void *ctx, unsigned flags) { struct r600_context *rctx = (struct r600_context *)ctx; struct radeon_winsys_cs *cs = rctx->rings.dma.cs; unsigned padding_dw, i; if (!cs->cdw) { return; } /* Pad the DMA CS to a multiple of 8 dwords. */ padding_dw = 8 - cs->cdw % 8; if (padding_dw < 8) { for (i = 0; i < padding_dw; i++) { cs->buf[cs->cdw++] = DMA_PACKET(DMA_PACKET_NOP, 0, 0, 0); } } rctx->rings.dma.flushing = true; rctx->ws->cs_flush(cs, flags, 0); rctx->rings.dma.flushing = false; } boolean r600_rings_is_buffer_referenced(struct r600_context *ctx, struct radeon_winsys_cs_handle *buf, enum radeon_bo_usage usage) { if (ctx->ws->cs_is_buffer_referenced(ctx->rings.gfx.cs, buf, usage)) { return TRUE; } if (ctx->rings.dma.cs) { if (ctx->ws->cs_is_buffer_referenced(ctx->rings.dma.cs, buf, usage)) { return TRUE; } } return FALSE; } void *r600_buffer_mmap_sync_with_rings(struct r600_context *ctx, struct r600_resource *resource, unsigned usage) { enum radeon_bo_usage rusage = RADEON_USAGE_READWRITE; unsigned flags = 0; bool sync_flush = TRUE; if (usage & PIPE_TRANSFER_UNSYNCHRONIZED) { return ctx->ws->buffer_map(resource->cs_buf, NULL, usage); } if (!(usage & PIPE_TRANSFER_WRITE)) { /* have to wait for pending read */ rusage = RADEON_USAGE_WRITE; } if (usage & PIPE_TRANSFER_DONTBLOCK) { flags |= RADEON_FLUSH_ASYNC; } if (ctx->ws->cs_is_buffer_referenced(ctx->rings.gfx.cs, resource->cs_buf, rusage) && ctx->rings.gfx.cs->cdw) { ctx->rings.gfx.flush(ctx, flags); if (usage & PIPE_TRANSFER_DONTBLOCK) { return NULL; } } if (ctx->rings.dma.cs) { if (ctx->ws->cs_is_buffer_referenced(ctx->rings.dma.cs, resource->cs_buf, rusage) && ctx->rings.dma.cs->cdw) { ctx->rings.dma.flush(ctx, flags); if (usage & PIPE_TRANSFER_DONTBLOCK) { return NULL; } } } if (usage & PIPE_TRANSFER_DONTBLOCK) { if (ctx->ws->buffer_is_busy(resource->buf, rusage)) { return NULL; } } if (sync_flush) { /* Try to avoid busy-waiting in radeon_bo_wait. */ ctx->ws->cs_sync_flush(ctx->rings.gfx.cs); if (ctx->rings.dma.cs) { ctx->ws->cs_sync_flush(ctx->rings.dma.cs); } } /* at this point everything is synchronized */ return ctx->ws->buffer_map(resource->cs_buf, NULL, usage); } static void r600_flush_from_winsys(void *ctx, unsigned flags) { struct r600_context *rctx = (struct r600_context *)ctx; rctx->rings.gfx.flush(rctx, flags); } static void r600_flush_dma_from_winsys(void *ctx, unsigned flags) { struct r600_context *rctx = (struct r600_context *)ctx; rctx->rings.dma.flush(rctx, flags); } static void r600_destroy_context(struct pipe_context *context) { struct r600_context *rctx = (struct r600_context *)context; r600_isa_destroy(rctx->isa); r600_sb_context_destroy(rctx->sb_context); pipe_resource_reference((struct pipe_resource**)&rctx->dummy_cmask, NULL); pipe_resource_reference((struct pipe_resource**)&rctx->dummy_fmask, NULL); if (rctx->dummy_pixel_shader) { rctx->context.delete_fs_state(&rctx->context, rctx->dummy_pixel_shader); } if (rctx->custom_dsa_flush) { rctx->context.delete_depth_stencil_alpha_state(&rctx->context, rctx->custom_dsa_flush); } if (rctx->custom_blend_resolve) { rctx->context.delete_blend_state(&rctx->context, rctx->custom_blend_resolve); } if (rctx->custom_blend_decompress) { rctx->context.delete_blend_state(&rctx->context, rctx->custom_blend_decompress); } util_unreference_framebuffer_state(&rctx->framebuffer.state); if (rctx->blitter) { util_blitter_destroy(rctx->blitter); } if (rctx->uploader) { u_upload_destroy(rctx->uploader); } if (rctx->allocator_so_filled_size) { u_suballocator_destroy(rctx->allocator_so_filled_size); } if (rctx->allocator_fetch_shader) { u_suballocator_destroy(rctx->allocator_fetch_shader); } util_slab_destroy(&rctx->pool_transfers); r600_release_command_buffer(&rctx->start_cs_cmd); if (rctx->rings.gfx.cs) { rctx->ws->cs_destroy(rctx->rings.gfx.cs); } if (rctx->rings.dma.cs) { rctx->ws->cs_destroy(rctx->rings.dma.cs); } FREE(rctx); } static struct pipe_context *r600_create_context(struct pipe_screen *screen, void *priv) { struct r600_context *rctx = CALLOC_STRUCT(r600_context); struct r600_screen* rscreen = (struct r600_screen *)screen; if (rctx == NULL) return NULL; util_slab_create(&rctx->pool_transfers, sizeof(struct r600_transfer), 64, UTIL_SLAB_SINGLETHREADED); rctx->context.screen = screen; rctx->context.priv = priv; rctx->context.destroy = r600_destroy_context; rctx->context.flush = r600_flush_from_st; /* Easy accessing of screen/winsys. */ rctx->screen = rscreen; rctx->ws = rscreen->ws; rctx->family = rscreen->family; rctx->chip_class = rscreen->chip_class; rctx->keep_tiling_flags = rscreen->info.drm_minor >= 12; LIST_INITHEAD(&rctx->active_nontimer_queries); r600_init_blit_functions(rctx); r600_init_query_functions(rctx); r600_init_context_resource_functions(rctx); r600_init_surface_functions(rctx); if (rscreen->info.has_uvd) { rctx->context.create_video_codec = r600_uvd_create_decoder; rctx->context.create_video_buffer = r600_video_buffer_create; } else { rctx->context.create_video_codec = vl_create_decoder; rctx->context.create_video_buffer = vl_video_buffer_create; } r600_init_common_state_functions(rctx); switch (rctx->chip_class) { case R600: case R700: r600_init_state_functions(rctx); r600_init_atom_start_cs(rctx); rctx->max_db = 4; rctx->custom_dsa_flush = r600_create_db_flush_dsa(rctx); rctx->custom_blend_resolve = rctx->chip_class == R700 ? r700_create_resolve_blend(rctx) : r600_create_resolve_blend(rctx); rctx->custom_blend_decompress = r600_create_decompress_blend(rctx); rctx->has_vertex_cache = !(rctx->family == CHIP_RV610 || rctx->family == CHIP_RV620 || rctx->family == CHIP_RS780 || rctx->family == CHIP_RS880 || rctx->family == CHIP_RV710); break; case EVERGREEN: case CAYMAN: evergreen_init_state_functions(rctx); evergreen_init_atom_start_cs(rctx); evergreen_init_atom_start_compute_cs(rctx); rctx->max_db = 8; rctx->custom_dsa_flush = evergreen_create_db_flush_dsa(rctx); rctx->custom_blend_resolve = evergreen_create_resolve_blend(rctx); rctx->custom_blend_decompress = evergreen_create_decompress_blend(rctx); rctx->has_vertex_cache = !(rctx->family == CHIP_CEDAR || rctx->family == CHIP_PALM || rctx->family == CHIP_SUMO || rctx->family == CHIP_SUMO2 || rctx->family == CHIP_CAICOS || rctx->family == CHIP_CAYMAN || rctx->family == CHIP_ARUBA); break; default: R600_ERR("Unsupported chip class %d.\n", rctx->chip_class); goto fail; } if (rscreen->trace_bo) { rctx->rings.gfx.cs = rctx->ws->cs_create(rctx->ws, RING_GFX, rscreen->trace_bo->cs_buf); } else { rctx->rings.gfx.cs = rctx->ws->cs_create(rctx->ws, RING_GFX, NULL); } rctx->rings.gfx.flush = r600_flush_gfx_ring; rctx->ws->cs_set_flush_callback(rctx->rings.gfx.cs, r600_flush_from_winsys, rctx); rctx->rings.gfx.flushing = false; rctx->rings.dma.cs = NULL; if (rscreen->info.r600_has_dma && !(rscreen->debug_flags & DBG_NO_ASYNC_DMA)) { rctx->rings.dma.cs = rctx->ws->cs_create(rctx->ws, RING_DMA, NULL); rctx->rings.dma.flush = r600_flush_dma_ring; rctx->ws->cs_set_flush_callback(rctx->rings.dma.cs, r600_flush_dma_from_winsys, rctx); rctx->rings.dma.flushing = false; } rctx->uploader = u_upload_create(&rctx->context, 1024 * 1024, 256, PIPE_BIND_INDEX_BUFFER | PIPE_BIND_CONSTANT_BUFFER); if (!rctx->uploader) goto fail; rctx->allocator_fetch_shader = u_suballocator_create(&rctx->context, 64 * 1024, 256, 0, PIPE_USAGE_STATIC, FALSE); if (!rctx->allocator_fetch_shader) goto fail; rctx->allocator_so_filled_size = u_suballocator_create(&rctx->context, 4096, 4, 0, PIPE_USAGE_STATIC, TRUE); if (!rctx->allocator_so_filled_size) goto fail; rctx->isa = calloc(1, sizeof(struct r600_isa)); if (!rctx->isa || r600_isa_init(rctx, rctx->isa)) goto fail; rctx->blitter = util_blitter_create(&rctx->context); if (rctx->blitter == NULL) goto fail; util_blitter_set_texture_multisample(rctx->blitter, rscreen->has_msaa); rctx->blitter->draw_rectangle = r600_draw_rectangle; r600_begin_new_cs(rctx); r600_get_backend_mask(rctx); /* this emits commands and must be last */ rctx->dummy_pixel_shader = util_make_fragment_cloneinput_shader(&rctx->context, 0, TGSI_SEMANTIC_GENERIC, TGSI_INTERPOLATE_CONSTANT); rctx->context.bind_fs_state(&rctx->context, rctx->dummy_pixel_shader); return &rctx->context; fail: r600_destroy_context(&rctx->context); return NULL; } /* * pipe_screen */ static const char* r600_get_vendor(struct pipe_screen* pscreen) { return "X.Org"; } static const char *r600_get_family_name(enum radeon_family family) { switch(family) { case CHIP_R600: return "AMD R600"; case CHIP_RV610: return "AMD RV610"; case CHIP_RV630: return "AMD RV630"; case CHIP_RV670: return "AMD RV670"; case CHIP_RV620: return "AMD RV620"; case CHIP_RV635: return "AMD RV635"; case CHIP_RS780: return "AMD RS780"; case CHIP_RS880: return "AMD RS880"; case CHIP_RV770: return "AMD RV770"; case CHIP_RV730: return "AMD RV730"; case CHIP_RV710: return "AMD RV710"; case CHIP_RV740: return "AMD RV740"; case CHIP_CEDAR: return "AMD CEDAR"; case CHIP_REDWOOD: return "AMD REDWOOD"; case CHIP_JUNIPER: return "AMD JUNIPER"; case CHIP_CYPRESS: return "AMD CYPRESS"; case CHIP_HEMLOCK: return "AMD HEMLOCK"; case CHIP_PALM: return "AMD PALM"; case CHIP_SUMO: return "AMD SUMO"; case CHIP_SUMO2: return "AMD SUMO2"; case CHIP_BARTS: return "AMD BARTS"; case CHIP_TURKS: return "AMD TURKS"; case CHIP_CAICOS: return "AMD CAICOS"; case CHIP_CAYMAN: return "AMD CAYMAN"; case CHIP_ARUBA: return "AMD ARUBA"; default: return "AMD unknown"; } } static const char* r600_get_name(struct pipe_screen* pscreen) { struct r600_screen *rscreen = (struct r600_screen *)pscreen; return r600_get_family_name(rscreen->family); } static int r600_get_param(struct pipe_screen* pscreen, enum pipe_cap param) { struct r600_screen *rscreen = (struct r600_screen *)pscreen; enum radeon_family family = rscreen->family; switch (param) { /* Supported features (boolean caps). */ case PIPE_CAP_NPOT_TEXTURES: case PIPE_CAP_TWO_SIDED_STENCIL: case PIPE_CAP_ANISOTROPIC_FILTER: case PIPE_CAP_POINT_SPRITE: case PIPE_CAP_OCCLUSION_QUERY: case PIPE_CAP_TEXTURE_SHADOW_MAP: case PIPE_CAP_TEXTURE_MIRROR_CLAMP: case PIPE_CAP_BLEND_EQUATION_SEPARATE: case PIPE_CAP_TEXTURE_SWIZZLE: case PIPE_CAP_DEPTH_CLIP_DISABLE: case PIPE_CAP_SHADER_STENCIL_EXPORT: case PIPE_CAP_VERTEX_ELEMENT_INSTANCE_DIVISOR: case PIPE_CAP_MIXED_COLORBUFFER_FORMATS: case PIPE_CAP_TGSI_FS_COORD_ORIGIN_UPPER_LEFT: case PIPE_CAP_TGSI_FS_COORD_PIXEL_CENTER_HALF_INTEGER: case PIPE_CAP_SM3: case PIPE_CAP_SEAMLESS_CUBE_MAP: case PIPE_CAP_PRIMITIVE_RESTART: case PIPE_CAP_CONDITIONAL_RENDER: case PIPE_CAP_TEXTURE_BARRIER: case PIPE_CAP_VERTEX_COLOR_UNCLAMPED: case PIPE_CAP_QUADS_FOLLOW_PROVOKING_VERTEX_CONVENTION: case PIPE_CAP_TGSI_INSTANCEID: case PIPE_CAP_VERTEX_BUFFER_OFFSET_4BYTE_ALIGNED_ONLY: case PIPE_CAP_VERTEX_BUFFER_STRIDE_4BYTE_ALIGNED_ONLY: case PIPE_CAP_VERTEX_ELEMENT_SRC_OFFSET_4BYTE_ALIGNED_ONLY: case PIPE_CAP_USER_INDEX_BUFFERS: case PIPE_CAP_USER_CONSTANT_BUFFERS: case PIPE_CAP_COMPUTE: case PIPE_CAP_START_INSTANCE: case PIPE_CAP_MAX_DUAL_SOURCE_RENDER_TARGETS: case PIPE_CAP_TEXTURE_BUFFER_OBJECTS: case PIPE_CAP_PREFER_BLIT_BASED_TEXTURE_TRANSFER: case PIPE_CAP_QUERY_PIPELINE_STATISTICS: case PIPE_CAP_TEXTURE_MULTISAMPLE: return 1; case PIPE_CAP_TGSI_TEXCOORD: return 0; case PIPE_CAP_MAX_TEXTURE_BUFFER_SIZE: return MIN2(rscreen->info.vram_size, 0xFFFFFFFF); case PIPE_CAP_MIN_MAP_BUFFER_ALIGNMENT: return R600_MAP_BUFFER_ALIGNMENT; case PIPE_CAP_CONSTANT_BUFFER_OFFSET_ALIGNMENT: return 256; case PIPE_CAP_TEXTURE_BUFFER_OFFSET_ALIGNMENT: return 1; case PIPE_CAP_GLSL_FEATURE_LEVEL: return 140; /* Supported except the original R600. */ case PIPE_CAP_INDEP_BLEND_ENABLE: case PIPE_CAP_INDEP_BLEND_FUNC: /* R600 doesn't support per-MRT blends */ return family == CHIP_R600 ? 0 : 1; /* Supported on Evergreen. */ case PIPE_CAP_SEAMLESS_CUBE_MAP_PER_TEXTURE: case PIPE_CAP_CUBE_MAP_ARRAY: return family >= CHIP_CEDAR ? 1 : 0; /* Unsupported features. */ case PIPE_CAP_TGSI_FS_COORD_ORIGIN_LOWER_LEFT: case PIPE_CAP_TGSI_FS_COORD_PIXEL_CENTER_INTEGER: case PIPE_CAP_SCALED_RESOLVE: case PIPE_CAP_TGSI_CAN_COMPACT_CONSTANTS: case PIPE_CAP_FRAGMENT_COLOR_CLAMPED: case PIPE_CAP_VERTEX_COLOR_CLAMPED: case PIPE_CAP_USER_VERTEX_BUFFERS: return 0; /* Stream output. */ case PIPE_CAP_MAX_STREAM_OUTPUT_BUFFERS: return rscreen->has_streamout ? 4 : 0; case PIPE_CAP_STREAM_OUTPUT_PAUSE_RESUME: return rscreen->has_streamout ? 1 : 0; case PIPE_CAP_MAX_STREAM_OUTPUT_SEPARATE_COMPONENTS: case PIPE_CAP_MAX_STREAM_OUTPUT_INTERLEAVED_COMPONENTS: return 32*4; /* Texturing. */ case PIPE_CAP_MAX_TEXTURE_2D_LEVELS: case PIPE_CAP_MAX_TEXTURE_3D_LEVELS: case PIPE_CAP_MAX_TEXTURE_CUBE_LEVELS: if (family >= CHIP_CEDAR) return 15; else return 14; case PIPE_CAP_MAX_TEXTURE_ARRAY_LAYERS: return rscreen->info.drm_minor >= 9 ? (family >= CHIP_CEDAR ? 16384 : 8192) : 0; case PIPE_CAP_MAX_COMBINED_SAMPLERS: return 32; /* Render targets. */ case PIPE_CAP_MAX_RENDER_TARGETS: /* XXX some r6xx are buggy and can only do 4 */ return 8; /* Timer queries, present when the clock frequency is non zero. */ case PIPE_CAP_QUERY_TIME_ELAPSED: return rscreen->info.r600_clock_crystal_freq != 0; case PIPE_CAP_QUERY_TIMESTAMP: return rscreen->info.drm_minor >= 20 && rscreen->info.r600_clock_crystal_freq != 0; case PIPE_CAP_MIN_TEXEL_OFFSET: return -8; case PIPE_CAP_MAX_TEXEL_OFFSET: return 7; case PIPE_CAP_TEXTURE_BORDER_COLOR_QUIRK: return PIPE_QUIRK_TEXTURE_BORDER_COLOR_SWIZZLE_R600; case PIPE_CAP_ENDIANNESS: return PIPE_ENDIAN_LITTLE; } return 0; } static float r600_get_paramf(struct pipe_screen* pscreen, enum pipe_capf param) { struct r600_screen *rscreen = (struct r600_screen *)pscreen; enum radeon_family family = rscreen->family; switch (param) { case PIPE_CAPF_MAX_LINE_WIDTH: case PIPE_CAPF_MAX_LINE_WIDTH_AA: case PIPE_CAPF_MAX_POINT_WIDTH: case PIPE_CAPF_MAX_POINT_WIDTH_AA: if (family >= CHIP_CEDAR) return 16384.0f; else return 8192.0f; case PIPE_CAPF_MAX_TEXTURE_ANISOTROPY: return 16.0f; case PIPE_CAPF_MAX_TEXTURE_LOD_BIAS: return 16.0f; case PIPE_CAPF_GUARD_BAND_LEFT: case PIPE_CAPF_GUARD_BAND_TOP: case PIPE_CAPF_GUARD_BAND_RIGHT: case PIPE_CAPF_GUARD_BAND_BOTTOM: return 0.0f; } return 0.0f; } static int r600_get_shader_param(struct pipe_screen* pscreen, unsigned shader, enum pipe_shader_cap param) { switch(shader) { case PIPE_SHADER_FRAGMENT: case PIPE_SHADER_VERTEX: case PIPE_SHADER_COMPUTE: break; case PIPE_SHADER_GEOMETRY: /* XXX: support and enable geometry programs */ return 0; default: /* XXX: support tessellation on Evergreen */ return 0; } switch (param) { case PIPE_SHADER_CAP_MAX_INSTRUCTIONS: case PIPE_SHADER_CAP_MAX_ALU_INSTRUCTIONS: case PIPE_SHADER_CAP_MAX_TEX_INSTRUCTIONS: case PIPE_SHADER_CAP_MAX_TEX_INDIRECTIONS: return 16384; case PIPE_SHADER_CAP_MAX_CONTROL_FLOW_DEPTH: return 32; case PIPE_SHADER_CAP_MAX_INPUTS: return 32; case PIPE_SHADER_CAP_MAX_TEMPS: return 256; /* Max native temporaries. */ case PIPE_SHADER_CAP_MAX_ADDRS: /* XXX Isn't this equal to TEMPS? */ return 1; /* Max native address registers */ case PIPE_SHADER_CAP_MAX_CONSTS: return R600_MAX_CONST_BUFFER_SIZE; case PIPE_SHADER_CAP_MAX_CONST_BUFFERS: return R600_MAX_USER_CONST_BUFFERS; case PIPE_SHADER_CAP_MAX_PREDS: return 0; /* nothing uses this */ case PIPE_SHADER_CAP_TGSI_CONT_SUPPORTED: return 1; case PIPE_SHADER_CAP_TGSI_SQRT_SUPPORTED: return 0; case PIPE_SHADER_CAP_INDIRECT_INPUT_ADDR: case PIPE_SHADER_CAP_INDIRECT_OUTPUT_ADDR: case PIPE_SHADER_CAP_INDIRECT_TEMP_ADDR: case PIPE_SHADER_CAP_INDIRECT_CONST_ADDR: return 1; case PIPE_SHADER_CAP_SUBROUTINES: return 0; case PIPE_SHADER_CAP_INTEGERS: return 1; case PIPE_SHADER_CAP_MAX_TEXTURE_SAMPLERS: return 16; case PIPE_SHADER_CAP_PREFERRED_IR: if (shader == PIPE_SHADER_COMPUTE) { return PIPE_SHADER_IR_LLVM; } else { return PIPE_SHADER_IR_TGSI; } } return 0; } static int r600_get_video_param(struct pipe_screen *screen, enum pipe_video_profile profile, enum pipe_video_entrypoint entrypoint, enum pipe_video_cap param) { switch (param) { case PIPE_VIDEO_CAP_SUPPORTED: return vl_profile_supported(screen, profile, entrypoint); case PIPE_VIDEO_CAP_NPOT_TEXTURES: return 1; case PIPE_VIDEO_CAP_MAX_WIDTH: case PIPE_VIDEO_CAP_MAX_HEIGHT: return vl_video_buffer_max_size(screen); case PIPE_VIDEO_CAP_PREFERED_FORMAT: return PIPE_FORMAT_NV12; case PIPE_VIDEO_CAP_PREFERS_INTERLACED: return false; case PIPE_VIDEO_CAP_SUPPORTS_INTERLACED: return false; case PIPE_VIDEO_CAP_SUPPORTS_PROGRESSIVE: return true; case PIPE_VIDEO_CAP_MAX_LEVEL: return vl_level_supported(screen, profile); default: return 0; } } const char * r600_llvm_gpu_string(enum radeon_family family) { const char * gpu_family; switch (family) { case CHIP_R600: case CHIP_RV630: case CHIP_RV635: case CHIP_RV670: gpu_family = "r600"; break; case CHIP_RV610: case CHIP_RV620: case CHIP_RS780: case CHIP_RS880: gpu_family = "rs880"; break; case CHIP_RV710: gpu_family = "rv710"; break; case CHIP_RV730: gpu_family = "rv730"; break; case CHIP_RV740: case CHIP_RV770: gpu_family = "rv770"; break; case CHIP_PALM: case CHIP_CEDAR: gpu_family = "cedar"; break; case CHIP_SUMO: case CHIP_SUMO2: gpu_family = "sumo"; break; case CHIP_REDWOOD: gpu_family = "redwood"; break; case CHIP_JUNIPER: gpu_family = "juniper"; break; case CHIP_HEMLOCK: case CHIP_CYPRESS: gpu_family = "cypress"; break; case CHIP_BARTS: gpu_family = "barts"; break; case CHIP_TURKS: gpu_family = "turks"; break; case CHIP_CAICOS: gpu_family = "caicos"; break; case CHIP_CAYMAN: case CHIP_ARUBA: gpu_family = "cayman"; break; default: gpu_family = ""; fprintf(stderr, "Chip not supported by r600 llvm " "backend, please file a bug at " PACKAGE_BUGREPORT "\n"); break; } return gpu_family; } static int r600_get_compute_param(struct pipe_screen *screen, enum pipe_compute_cap param, void *ret) { struct r600_screen *rscreen = (struct r600_screen *)screen; //TODO: select these params by asic switch (param) { case PIPE_COMPUTE_CAP_IR_TARGET: { const char *gpu = r600_llvm_gpu_string(rscreen->family); if (ret) { sprintf(ret, "%s-r600--", gpu); } return (8 + strlen(gpu)) * sizeof(char); } case PIPE_COMPUTE_CAP_GRID_DIMENSION: if (ret) { uint64_t * grid_dimension = ret; grid_dimension[0] = 3; } return 1 * sizeof(uint64_t); case PIPE_COMPUTE_CAP_MAX_GRID_SIZE: if (ret) { uint64_t * grid_size = ret; grid_size[0] = 65535; grid_size[1] = 65535; grid_size[2] = 1; } return 3 * sizeof(uint64_t) ; case PIPE_COMPUTE_CAP_MAX_BLOCK_SIZE: if (ret) { uint64_t * block_size = ret; block_size[0] = 256; block_size[1] = 256; block_size[2] = 256; } return 3 * sizeof(uint64_t); case PIPE_COMPUTE_CAP_MAX_THREADS_PER_BLOCK: if (ret) { uint64_t * max_threads_per_block = ret; *max_threads_per_block = 256; } return sizeof(uint64_t); case PIPE_COMPUTE_CAP_MAX_GLOBAL_SIZE: if (ret) { uint64_t * max_global_size = ret; /* XXX: This is what the proprietary driver reports, we * may want to use a different value. */ *max_global_size = 201326592; } return sizeof(uint64_t); case PIPE_COMPUTE_CAP_MAX_INPUT_SIZE: if (ret) { uint64_t * max_input_size = ret; *max_input_size = 1024; } return sizeof(uint64_t); case PIPE_COMPUTE_CAP_MAX_LOCAL_SIZE: if (ret) { uint64_t * max_local_size = ret; /* XXX: This is what the proprietary driver reports, we * may want to use a different value. */ *max_local_size = 32768; } return sizeof(uint64_t); case PIPE_COMPUTE_CAP_MAX_MEM_ALLOC_SIZE: if (ret) { uint64_t max_global_size; uint64_t * max_mem_alloc_size = ret; r600_get_compute_param(screen, PIPE_COMPUTE_CAP_MAX_GLOBAL_SIZE, &max_global_size); /* OpenCL requres this value be at least * max(MAX_GLOBAL_SIZE / 4, 128 * 1024 *1024) * I'm really not sure what value to report here, but * MAX_GLOBAL_SIZE / 4 seems resonable. */ *max_mem_alloc_size = max_global_size / 4; } return sizeof(uint64_t); default: fprintf(stderr, "unknown PIPE_COMPUTE_CAP %d\n", param); return 0; } } static void r600_destroy_screen(struct pipe_screen* pscreen) { struct r600_screen *rscreen = (struct r600_screen *)pscreen; if (rscreen == NULL) return; pipe_mutex_destroy(rscreen->aux_context_lock); rscreen->aux_context->destroy(rscreen->aux_context); if (rscreen->global_pool) { compute_memory_pool_delete(rscreen->global_pool); } if (rscreen->fences.bo) { struct r600_fence_block *entry, *tmp; LIST_FOR_EACH_ENTRY_SAFE(entry, tmp, &rscreen->fences.blocks, head) { LIST_DEL(&entry->head); FREE(entry); } rscreen->ws->buffer_unmap(rscreen->fences.bo->cs_buf); pipe_resource_reference((struct pipe_resource**)&rscreen->fences.bo, NULL); } if (rscreen->trace_bo) { rscreen->ws->buffer_unmap(rscreen->trace_bo->cs_buf); pipe_resource_reference((struct pipe_resource**)&rscreen->trace_bo, NULL); } pipe_mutex_destroy(rscreen->fences.mutex); rscreen->ws->destroy(rscreen->ws); FREE(rscreen); } static void r600_fence_reference(struct pipe_screen *pscreen, struct pipe_fence_handle **ptr, struct pipe_fence_handle *fence) { struct r600_fence **oldf = (struct r600_fence**)ptr; struct r600_fence *newf = (struct r600_fence*)fence; if (pipe_reference(&(*oldf)->reference, &newf->reference)) { struct r600_screen *rscreen = (struct r600_screen *)pscreen; pipe_mutex_lock(rscreen->fences.mutex); pipe_resource_reference((struct pipe_resource**)&(*oldf)->sleep_bo, NULL); LIST_ADDTAIL(&(*oldf)->head, &rscreen->fences.pool); pipe_mutex_unlock(rscreen->fences.mutex); } *ptr = fence; } static boolean r600_fence_signalled(struct pipe_screen *pscreen, struct pipe_fence_handle *fence) { struct r600_screen *rscreen = (struct r600_screen *)pscreen; struct r600_fence *rfence = (struct r600_fence*)fence; return rscreen->fences.data[rfence->index] != 0; } static boolean r600_fence_finish(struct pipe_screen *pscreen, struct pipe_fence_handle *fence, uint64_t timeout) { struct r600_screen *rscreen = (struct r600_screen *)pscreen; struct r600_fence *rfence = (struct r600_fence*)fence; int64_t start_time = 0; unsigned spins = 0; if (timeout != PIPE_TIMEOUT_INFINITE) { start_time = os_time_get(); /* Convert to microseconds. */ timeout /= 1000; } while (rscreen->fences.data[rfence->index] == 0) { /* Special-case infinite timeout - wait for the dummy BO to become idle */ if (timeout == PIPE_TIMEOUT_INFINITE) { rscreen->ws->buffer_wait(rfence->sleep_bo->buf, RADEON_USAGE_READWRITE); break; } /* The dummy BO will be busy until the CS including the fence has completed, or * the GPU is reset. Don't bother continuing to spin when the BO is idle. */ if (!rscreen->ws->buffer_is_busy(rfence->sleep_bo->buf, RADEON_USAGE_READWRITE)) break; if (++spins % 256) continue; #ifdef PIPE_OS_UNIX sched_yield(); #else os_time_sleep(10); #endif if (timeout != PIPE_TIMEOUT_INFINITE && os_time_get() - start_time >= timeout) { break; } } return rscreen->fences.data[rfence->index] != 0; } static int r600_interpret_tiling(struct r600_screen *rscreen, uint32_t tiling_config) { switch ((tiling_config & 0xe) >> 1) { case 0: rscreen->tiling_info.num_channels = 1; break; case 1: rscreen->tiling_info.num_channels = 2; break; case 2: rscreen->tiling_info.num_channels = 4; break; case 3: rscreen->tiling_info.num_channels = 8; break; default: return -EINVAL; } switch ((tiling_config & 0x30) >> 4) { case 0: rscreen->tiling_info.num_banks = 4; break; case 1: rscreen->tiling_info.num_banks = 8; break; default: return -EINVAL; } switch ((tiling_config & 0xc0) >> 6) { case 0: rscreen->tiling_info.group_bytes = 256; break; case 1: rscreen->tiling_info.group_bytes = 512; break; default: return -EINVAL; } return 0; } static int evergreen_interpret_tiling(struct r600_screen *rscreen, uint32_t tiling_config) { switch (tiling_config & 0xf) { case 0: rscreen->tiling_info.num_channels = 1; break; case 1: rscreen->tiling_info.num_channels = 2; break; case 2: rscreen->tiling_info.num_channels = 4; break; case 3: rscreen->tiling_info.num_channels = 8; break; default: return -EINVAL; } switch ((tiling_config & 0xf0) >> 4) { case 0: rscreen->tiling_info.num_banks = 4; break; case 1: rscreen->tiling_info.num_banks = 8; break; case 2: rscreen->tiling_info.num_banks = 16; break; default: return -EINVAL; } switch ((tiling_config & 0xf00) >> 8) { case 0: rscreen->tiling_info.group_bytes = 256; break; case 1: rscreen->tiling_info.group_bytes = 512; break; default: return -EINVAL; } return 0; } static int r600_init_tiling(struct r600_screen *rscreen) { uint32_t tiling_config = rscreen->info.r600_tiling_config; /* set default group bytes, overridden by tiling info ioctl */ if (rscreen->chip_class <= R700) { rscreen->tiling_info.group_bytes = 256; } else { rscreen->tiling_info.group_bytes = 512; } if (!tiling_config) return 0; if (rscreen->chip_class <= R700) { return r600_interpret_tiling(rscreen, tiling_config); } else { return evergreen_interpret_tiling(rscreen, tiling_config); } } static uint64_t r600_get_timestamp(struct pipe_screen *screen) { struct r600_screen *rscreen = (struct r600_screen*)screen; return 1000000 * rscreen->ws->query_value(rscreen->ws, RADEON_TIMESTAMP) / rscreen->info.r600_clock_crystal_freq; } static int r600_get_driver_query_info(struct pipe_screen *screen, unsigned index, struct pipe_driver_query_info *info) { struct r600_screen *rscreen = (struct r600_screen*)screen; struct pipe_driver_query_info list[] = { {"draw-calls", R600_QUERY_DRAW_CALLS, 0}, {"requested-VRAM", R600_QUERY_REQUESTED_VRAM, rscreen->info.vram_size, TRUE}, {"requested-GTT", R600_QUERY_REQUESTED_GTT, rscreen->info.gart_size, TRUE}, {"buffer-wait-time", R600_QUERY_BUFFER_WAIT_TIME, 0, FALSE} }; if (!info) return Elements(list); if (index >= Elements(list)) return 0; *info = list[index]; return 1; } struct pipe_screen *r600_screen_create(struct radeon_winsys *ws) { struct r600_screen *rscreen = CALLOC_STRUCT(r600_screen); if (rscreen == NULL) { return NULL; } rscreen->ws = ws; ws->query_info(ws, &rscreen->info); rscreen->debug_flags = debug_get_flags_option("R600_DEBUG", debug_options, 0); if (debug_get_bool_option("R600_DEBUG_COMPUTE", FALSE)) rscreen->debug_flags |= DBG_COMPUTE; if (debug_get_bool_option("R600_DUMP_SHADERS", FALSE)) rscreen->debug_flags |= DBG_FS | DBG_VS | DBG_GS | DBG_PS | DBG_CS; if (!debug_get_bool_option("R600_HYPERZ", TRUE)) rscreen->debug_flags |= DBG_NO_HYPERZ; if (!debug_get_bool_option("R600_LLVM", TRUE)) rscreen->debug_flags |= DBG_NO_LLVM; if (debug_get_bool_option("R600_PRINT_TEXDEPTH", FALSE)) rscreen->debug_flags |= DBG_TEX_DEPTH; rscreen->family = rscreen->info.family; rscreen->chip_class = rscreen->info.chip_class; if (rscreen->family == CHIP_UNKNOWN) { fprintf(stderr, "r600: Unknown chipset 0x%04X\n", rscreen->info.pci_id); FREE(rscreen); return NULL; } /* Figure out streamout kernel support. */ switch (rscreen->chip_class) { case R600: if (rscreen->family < CHIP_RS780) { rscreen->has_streamout = rscreen->info.drm_minor >= 14; } else { rscreen->has_streamout = rscreen->info.drm_minor >= 23; } break; case R700: rscreen->has_streamout = rscreen->info.drm_minor >= 17; break; case EVERGREEN: case CAYMAN: rscreen->has_streamout = rscreen->info.drm_minor >= 14; break; default: rscreen->has_streamout = FALSE; break; } /* MSAA support. */ switch (rscreen->chip_class) { case R600: case R700: rscreen->has_msaa = rscreen->info.drm_minor >= 22; rscreen->has_compressed_msaa_texturing = false; break; case EVERGREEN: rscreen->has_msaa = rscreen->info.drm_minor >= 19; rscreen->has_compressed_msaa_texturing = rscreen->info.drm_minor >= 24; break; case CAYMAN: rscreen->has_msaa = rscreen->info.drm_minor >= 19; rscreen->has_compressed_msaa_texturing = true; break; default: rscreen->has_msaa = FALSE; rscreen->has_compressed_msaa_texturing = false; } rscreen->has_cp_dma = rscreen->info.drm_minor >= 27 && !(rscreen->debug_flags & DBG_NO_CP_DMA); if (r600_init_tiling(rscreen)) { FREE(rscreen); return NULL; } rscreen->screen.destroy = r600_destroy_screen; rscreen->screen.get_name = r600_get_name; rscreen->screen.get_vendor = r600_get_vendor; rscreen->screen.get_param = r600_get_param; rscreen->screen.get_shader_param = r600_get_shader_param; rscreen->screen.get_paramf = r600_get_paramf; rscreen->screen.get_compute_param = r600_get_compute_param; rscreen->screen.get_timestamp = r600_get_timestamp; if (rscreen->chip_class >= EVERGREEN) { rscreen->screen.is_format_supported = evergreen_is_format_supported; rscreen->dma_blit = &evergreen_dma_blit; } else { rscreen->screen.is_format_supported = r600_is_format_supported; rscreen->dma_blit = &r600_dma_blit; } rscreen->screen.context_create = r600_create_context; rscreen->screen.fence_reference = r600_fence_reference; rscreen->screen.fence_signalled = r600_fence_signalled; rscreen->screen.fence_finish = r600_fence_finish; rscreen->screen.get_driver_query_info = r600_get_driver_query_info; if (rscreen->info.has_uvd) { rscreen->screen.get_video_param = r600_uvd_get_video_param; rscreen->screen.is_video_format_supported = ruvd_is_format_supported; } else { rscreen->screen.get_video_param = r600_get_video_param; rscreen->screen.is_video_format_supported = vl_video_buffer_is_format_supported; } r600_init_screen_resource_functions(&rscreen->screen); util_format_s3tc_init(); rscreen->fences.bo = NULL; rscreen->fences.data = NULL; rscreen->fences.next_index = 0; LIST_INITHEAD(&rscreen->fences.pool); LIST_INITHEAD(&rscreen->fences.blocks); pipe_mutex_init(rscreen->fences.mutex); rscreen->global_pool = compute_memory_pool_new(rscreen); rscreen->cs_count = 0; if (rscreen->info.drm_minor >= 28 && (rscreen->debug_flags & DBG_TRACE_CS)) { rscreen->trace_bo = (struct r600_resource*)pipe_buffer_create(&rscreen->screen, PIPE_BIND_CUSTOM, PIPE_USAGE_STAGING, 4096); if (rscreen->trace_bo) { rscreen->trace_ptr = rscreen->ws->buffer_map(rscreen->trace_bo->cs_buf, NULL, PIPE_TRANSFER_UNSYNCHRONIZED); } } /* Create the auxiliary context. */ pipe_mutex_init(rscreen->aux_context_lock); rscreen->aux_context = rscreen->screen.context_create(&rscreen->screen, NULL); #if 0 /* This is for testing whether aux_context and buffer clearing work correctly. */ struct pipe_resource templ = {}; templ.width0 = 4; templ.height0 = 2048; templ.depth0 = 1; templ.array_size = 1; templ.target = PIPE_TEXTURE_2D; templ.format = PIPE_FORMAT_R8G8B8A8_UNORM; templ.usage = PIPE_USAGE_STATIC; struct r600_resource *res = r600_resource(rscreen->screen.resource_create(&rscreen->screen, &templ)); unsigned char *map = ws->buffer_map(res->cs_buf, NULL, PIPE_TRANSFER_WRITE); memset(map, 0, 256); r600_screen_clear_buffer(rscreen, &res->b.b, 4, 4, 0xCC); r600_screen_clear_buffer(rscreen, &res->b.b, 8, 4, 0xDD); r600_screen_clear_buffer(rscreen, &res->b.b, 12, 4, 0xEE); r600_screen_clear_buffer(rscreen, &res->b.b, 20, 4, 0xFF); r600_screen_clear_buffer(rscreen, &res->b.b, 32, 20, 0x87); ws->buffer_wait(res->buf, RADEON_USAGE_WRITE); int i; for (i = 0; i < 256; i++) { printf("%02X", map[i]); if (i % 16 == 15) printf("\n"); } #endif return &rscreen->screen; }