/* * Copyright 2010 Dave Airlie * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * on the rights to use, copy, modify, merge, publish, distribute, sub * license, and/or sell copies of the Software, and to permit persons to whom * the Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL * THE AUTHOR(S) AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM, * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE * USE OR OTHER DEALINGS IN THE SOFTWARE. * * Authors: * Dave Airlie */ #include #include #include #include "state_tracker/drm_driver.h" #include "r600_priv.h" #include "r600d.h" #include "drm.h" #include "radeon_drm.h" struct r600_bo *r600_bo(struct radeon *radeon, unsigned size, unsigned alignment, unsigned binding, unsigned usage) { struct r600_bo *bo; struct radeon_bo *rbo; uint32_t initial_domain, domains; /* Staging resources particpate in transfers and blits only * and are used for uploads and downloads from regular * resources. We generate them internally for some transfers. */ if (usage == PIPE_USAGE_STAGING) domains = RADEON_GEM_DOMAIN_CPU | RADEON_GEM_DOMAIN_GTT; else domains = (RADEON_GEM_DOMAIN_CPU | RADEON_GEM_DOMAIN_GTT | RADEON_GEM_DOMAIN_VRAM); if (binding & (PIPE_BIND_CONSTANT_BUFFER | PIPE_BIND_VERTEX_BUFFER | PIPE_BIND_INDEX_BUFFER)) { bo = r600_bomgr_bo_create(radeon->bomgr, size, alignment, *radeon->cfence); if (bo) { bo->domains = domains; return bo; } } switch(usage) { case PIPE_USAGE_DYNAMIC: case PIPE_USAGE_STREAM: case PIPE_USAGE_STAGING: initial_domain = RADEON_GEM_DOMAIN_GTT; break; case PIPE_USAGE_DEFAULT: case PIPE_USAGE_STATIC: case PIPE_USAGE_IMMUTABLE: default: initial_domain = RADEON_GEM_DOMAIN_VRAM; break; } rbo = radeon_bo(radeon, 0, size, alignment, initial_domain); if (rbo == NULL) { return NULL; } bo = calloc(1, sizeof(struct r600_bo)); bo->size = size; bo->alignment = alignment; bo->domains = domains; bo->bo = rbo; if (binding & (PIPE_BIND_CONSTANT_BUFFER | PIPE_BIND_VERTEX_BUFFER | PIPE_BIND_INDEX_BUFFER)) { r600_bomgr_bo_init(radeon->bomgr, bo); } pipe_reference_init(&bo->reference, 1); return bo; } struct r600_bo *r600_bo_handle(struct radeon *radeon, unsigned handle, unsigned *array_mode) { struct r600_bo *bo = calloc(1, sizeof(struct r600_bo)); struct radeon_bo *rbo; rbo = bo->bo = radeon_bo(radeon, handle, 0, 0, 0); if (rbo == NULL) { free(bo); return NULL; } bo->size = rbo->size; bo->domains = (RADEON_GEM_DOMAIN_CPU | RADEON_GEM_DOMAIN_GTT | RADEON_GEM_DOMAIN_VRAM); pipe_reference_init(&bo->reference, 1); radeon_bo_get_tiling_flags(radeon, rbo, &bo->tiling_flags, &bo->kernel_pitch); if (array_mode) { if (bo->tiling_flags) { if (bo->tiling_flags & RADEON_TILING_MACRO) *array_mode = V_0280A0_ARRAY_2D_TILED_THIN1; else if (bo->tiling_flags & RADEON_TILING_MICRO) *array_mode = V_0280A0_ARRAY_1D_TILED_THIN1; } else { *array_mode = 0; } } return bo; } void *r600_bo_map(struct radeon *radeon, struct r600_bo *bo, unsigned usage, void *ctx) { struct pipe_context *pctx = ctx; if (usage & PB_USAGE_UNSYNCHRONIZED) { radeon_bo_map(radeon, bo->bo); return (uint8_t *) bo->bo->data + bo->offset; } if (p_atomic_read(&bo->bo->reference.count) > 1) { if (usage & PB_USAGE_DONTBLOCK) { return NULL; } if (ctx) { pctx->flush(pctx, NULL); } } if (usage & PB_USAGE_DONTBLOCK) { uint32_t domain; if (radeon_bo_busy(radeon, bo->bo, &domain)) return NULL; if (radeon_bo_map(radeon, bo->bo)) { return NULL; } goto out; } radeon_bo_map(radeon, bo->bo); if (radeon_bo_wait(radeon, bo->bo)) { radeon_bo_unmap(radeon, bo->bo); return NULL; } out: return (uint8_t *) bo->bo->data + bo->offset; } void r600_bo_unmap(struct radeon *radeon, struct r600_bo *bo) { radeon_bo_unmap(radeon, bo->bo); } void r600_bo_destroy(struct radeon *radeon, struct r600_bo *bo) { if (bo->manager_id) { if (!r600_bomgr_bo_destroy(radeon->bomgr, bo)) { /* destroy is delayed by buffer manager */ return; } } radeon_bo_reference(radeon, &bo->bo, NULL); free(bo); } boolean r600_bo_get_winsys_handle(struct radeon *radeon, struct r600_bo *bo, unsigned stride, struct winsys_handle *whandle) { whandle->stride = stride; switch(whandle->type) { case DRM_API_HANDLE_TYPE_KMS: whandle->handle = bo->bo->handle; break; case DRM_API_HANDLE_TYPE_SHARED: if (radeon_bo_get_name(radeon, bo->bo, &whandle->handle)) return FALSE; break; default: return FALSE; } return TRUE; }