summaryrefslogtreecommitdiffstats
path: root/src/gallium/auxiliary/draw/draw_llvm.c
diff options
context:
space:
mode:
authorRoland Scheidegger <[email protected]>2016-11-09 05:46:12 +0100
committerRoland Scheidegger <[email protected]>2016-11-09 05:53:16 +0100
commit4d5346aaac292213c6875aeb2814ca94b9b8e06e (patch)
tree6ffea844b957ebde402c51b4fb4e4397735db25b /src/gallium/auxiliary/draw/draw_llvm.c
parentf037afb701a8df8a646414e986b3561c8a798e05 (diff)
Revert "draw: use vectorized calculations for fetch"
Trivial. There's some regressions internally, related to overflow behavior. I'll have to look at it at another time, some interactions with vsplit/vcache are actually mind-blowing. This reverts commit 3fa10ffb496cc4e6d1003891cf0381bb5bec2a74.
Diffstat (limited to 'src/gallium/auxiliary/draw/draw_llvm.c')
-rw-r--r--src/gallium/auxiliary/draw/draw_llvm.c436
1 files changed, 159 insertions, 277 deletions
diff --git a/src/gallium/auxiliary/draw/draw_llvm.c b/src/gallium/auxiliary/draw/draw_llvm.c
index cc98d261960..5b4e2a14aef 100644
--- a/src/gallium/auxiliary/draw/draw_llvm.c
+++ b/src/gallium/auxiliary/draw/draw_llvm.c
@@ -33,8 +33,6 @@
#include "gallivm/lp_bld_arit.h"
#include "gallivm/lp_bld_arit_overflow.h"
-#include "gallivm/lp_bld_bitarit.h"
-#include "gallivm/lp_bld_gather.h"
#include "gallivm/lp_bld_logic.h"
#include "gallivm/lp_bld_const.h"
#include "gallivm/lp_bld_swizzle.h"
@@ -658,140 +656,28 @@ generate_vs(struct draw_llvm_variant *variant,
}
}
-
-static void
-convert_to_soa(struct gallivm_state *gallivm,
- LLVMValueRef src_aos[LP_MAX_VECTOR_WIDTH / 32],
- LLVMValueRef dst_soa[TGSI_NUM_CHANNELS],
- const struct lp_type soa_type)
-{
- unsigned j, k;
- struct lp_type aos_channel_type = soa_type;
-
- LLVMValueRef aos_channels[TGSI_NUM_CHANNELS];
- unsigned pixels_per_channel = soa_type.length / TGSI_NUM_CHANNELS;
-
- debug_assert(TGSI_NUM_CHANNELS == 4);
- debug_assert((soa_type.length % TGSI_NUM_CHANNELS) == 0);
-
- aos_channel_type.length >>= 1;
-
- for (j = 0; j < TGSI_NUM_CHANNELS; ++j) {
- LLVMValueRef channel[LP_MAX_VECTOR_LENGTH] = { 0 };
-
- assert(pixels_per_channel <= LP_MAX_VECTOR_LENGTH);
-
- for (k = 0; k < pixels_per_channel; ++k) {
- channel[k] = src_aos[j + TGSI_NUM_CHANNELS * k];
- }
-
- aos_channels[j] = lp_build_concat(gallivm, channel, aos_channel_type, pixels_per_channel);
- }
-
- lp_build_transpose_aos(gallivm, soa_type, aos_channels, dst_soa);
-}
-
-
static void
-fetch_vector(struct gallivm_state *gallivm,
- const struct util_format_description *format_desc,
- struct lp_type vs_type,
- LLVMValueRef vb_stride,
- LLVMValueRef map_ptr,
- LLVMValueRef buffer_size_adj,
- LLVMValueRef *inputs,
- LLVMValueRef indices,
- LLVMValueRef valid_mask)
+generate_fetch(struct gallivm_state *gallivm,
+ const struct util_format_description *format_desc,
+ LLVMValueRef vb_stride,
+ LLVMValueRef stride_fixed,
+ LLVMValueRef map_ptr,
+ LLVMValueRef buffer_size_adj,
+ LLVMValueRef ofbit,
+ LLVMValueRef *res,
+ LLVMValueRef index)
{
LLVMValueRef zero = LLVMConstNull(LLVMInt32TypeInContext(gallivm->context));
LLVMBuilderRef builder = gallivm->builder;
- struct lp_build_context blduivec;
- LLVMValueRef offset, tmp;
- LLVMValueRef aos_fetch[LP_MAX_VECTOR_WIDTH / 32];
- unsigned i;
-
- lp_build_context_init(&blduivec, gallivm, lp_uint_type(vs_type));
-
- vb_stride = lp_build_broadcast_scalar(&blduivec, vb_stride);
- buffer_size_adj = lp_build_broadcast_scalar(&blduivec, buffer_size_adj);
-
- /*
- * Sort of interestingly, with interleaved attribs, llvm 3.7+ will
- * recognize these calculations to be constant with different attribs
- * (the different offset has been added to map_ptr).
- * llvm 3.3, however, will not (I can't get llvm 3.4-3.6 to link...)
- *
- * XXX: could actually avoid this altogether (replacing by simple
- * non-widening mul) by precalculating the max index instead outside
- * the loop (at the cost of one scalar udiv per vertex element).
- */
- offset = lp_build_mul_32_lohi(&blduivec, vb_stride, indices, &tmp);
-
- tmp = lp_build_compare(gallivm, blduivec.type,
- PIPE_FUNC_EQUAL, tmp, blduivec.zero);
- valid_mask = LLVMBuildAnd(builder, tmp, valid_mask, "");
-
- tmp = lp_build_compare(gallivm, blduivec.type,
- PIPE_FUNC_LEQUAL, offset, buffer_size_adj);
- valid_mask = LLVMBuildAnd(builder, tmp, valid_mask, "");
-
- /* not valid elements use offset 0 */
- offset = LLVMBuildAnd(builder, offset, valid_mask, "");
-
- if (0) {
- lp_build_print_value(gallivm, " indices = ", indices);
- lp_build_print_value(gallivm, " offsets = ", offset);
- lp_build_print_value(gallivm, " valid_mask = ", valid_mask);
- }
-
- /*
- * Note: we probably really want to use SoA fetch, not AoS one (albeit
- * for most formats it will amount to the same as this isn't very
- * optimized). But looks dangerous since it assumes alignment.
- */
- for (i = 0; i < vs_type.length; i++) {
- LLVMValueRef offset1, elem;
- elem = lp_build_const_int32(gallivm, i);
- offset1 = LLVMBuildExtractElement(builder, offset, elem, "");
-
- aos_fetch[i] = lp_build_fetch_rgba_aos(gallivm, format_desc,
- lp_float32_vec4_type(),
- FALSE, map_ptr, offset1,
- zero, zero, NULL);
- }
- convert_to_soa(gallivm, aos_fetch, inputs, vs_type);
-
- for (i = 0; i < TGSI_NUM_CHANNELS; i++) {
- inputs[i] = LLVMBuildBitCast(builder, inputs[i], blduivec.vec_type, "");
- inputs[i] = LLVMBuildAnd(builder, inputs[i], valid_mask, "");
- inputs[i] = LLVMBuildBitCast(builder, inputs[i],
- lp_build_vec_type(gallivm, vs_type), "");
-
- }
-}
-
-
-static void
-fetch_instanced(struct gallivm_state *gallivm,
- const struct util_format_description *format_desc,
- struct lp_type vs_type,
- LLVMValueRef vb_stride,
- LLVMValueRef map_ptr,
- LLVMValueRef buffer_size_adj,
- LLVMValueRef ofbit,
- LLVMValueRef *inputs,
- LLVMValueRef index)
-{
- LLVMValueRef zero = LLVMConstNull(LLVMInt32TypeInContext(gallivm->context));
- LLVMBuilderRef builder = gallivm->builder;
- LLVMValueRef stride, buffer_overflowed, aos;
+ LLVMValueRef stride;
+ LLVMValueRef buffer_overflowed;
LLVMValueRef temp_ptr =
lp_build_alloca(gallivm,
lp_build_vec_type(gallivm, lp_float32_vec4_type()), "");
struct lp_build_if_state if_ctx;
- unsigned i;
stride = lp_build_umul_overflow(gallivm, vb_stride, index, &ofbit);
+ stride = lp_build_uadd_overflow(gallivm, stride, stride_fixed, &ofbit);
buffer_overflowed = LLVMBuildICmp(builder, LLVMIntUGT,
stride, buffer_size_adj,
@@ -799,7 +685,8 @@ fetch_instanced(struct gallivm_state *gallivm,
buffer_overflowed = LLVMBuildOr(builder, buffer_overflowed, ofbit, "");
if (0) {
- lp_build_print_value(gallivm, " instance index = ", index);
+ lp_build_printf(gallivm, " stride = %u\n", stride);
+ lp_build_printf(gallivm, " buffer size adj = %u\n", buffer_size_adj);
lp_build_print_value(gallivm, " buffer overflowed = ", buffer_overflowed);
}
@@ -812,26 +699,52 @@ fetch_instanced(struct gallivm_state *gallivm,
lp_build_else(&if_ctx);
{
LLVMValueRef val;
+ map_ptr = LLVMBuildGEP(builder, map_ptr, &stride, 1, "");
val = lp_build_fetch_rgba_aos(gallivm,
format_desc,
lp_float32_vec4_type(),
FALSE,
map_ptr,
- stride, zero, zero,
+ zero, zero, zero,
NULL);
LLVMBuildStore(builder, val, temp_ptr);
}
lp_build_endif(&if_ctx);
- aos = LLVMBuildLoad(builder, temp_ptr, "aos");
+ *res = LLVMBuildLoad(builder, temp_ptr, "aos");
+}
- for (i = 0; i < TGSI_NUM_CHANNELS; i++) {
- LLVMValueRef index = lp_build_const_int32(gallivm, i);
- inputs[i] = lp_build_extract_broadcast(gallivm,
- lp_float32_vec4_type(),
- vs_type, aos, index);
+static void
+convert_to_soa(struct gallivm_state *gallivm,
+ LLVMValueRef src_aos[LP_MAX_VECTOR_WIDTH / 32],
+ LLVMValueRef (*dst_soa)[TGSI_NUM_CHANNELS],
+ unsigned attrib, const struct lp_type soa_type)
+{
+ unsigned j, k;
+ struct lp_type aos_channel_type = soa_type;
+
+ LLVMValueRef aos_channels[TGSI_NUM_CHANNELS];
+ unsigned pixels_per_channel = soa_type.length / TGSI_NUM_CHANNELS;
+
+ debug_assert(TGSI_NUM_CHANNELS == 4);
+ debug_assert((soa_type.length % TGSI_NUM_CHANNELS) == 0);
+
+ aos_channel_type.length >>= 1;
+
+ for (j = 0; j < TGSI_NUM_CHANNELS; ++j) {
+ LLVMValueRef channel[LP_MAX_VECTOR_LENGTH] = { 0 };
+
+ assert(pixels_per_channel <= LP_MAX_VECTOR_LENGTH);
+
+ for (k = 0; k < pixels_per_channel; ++k) {
+ channel[k] = src_aos[j + TGSI_NUM_CHANNELS * k];
+ }
+
+ aos_channels[j] = lp_build_concat(gallivm, channel, aos_channel_type, pixels_per_channel);
}
+
+ lp_build_transpose_aos(gallivm, soa_type, aos_channels, dst_soa[attrib]);
}
@@ -861,7 +774,6 @@ store_aos(struct gallivm_state *gallivm,
LLVMSetAlignment(LLVMBuildStore(builder, value, data_ptr), sizeof(float));
}
-
/**
* Adjust the mask to architecture endianess. The mask will the store in struct:
*
@@ -1579,22 +1491,24 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant,
LLVMBuilderRef builder;
char func_name[64];
struct lp_type vs_type;
+ LLVMValueRef end, start;
LLVMValueRef count, fetch_elts, fetch_elt_max, fetch_count;
- LLVMValueRef vertex_id_offset, start_instance, start;
+ LLVMValueRef vertex_id_offset, start_instance;
LLVMValueRef stride, step, io_itr;
- LLVMValueRef ind_vec;
LLVMValueRef io_ptr, vbuffers_ptr, vb_ptr;
- LLVMValueRef vb_stride[PIPE_MAX_ATTRIBS];
- LLVMValueRef map_ptr[PIPE_MAX_ATTRIBS];
- LLVMValueRef buffer_size_adj[PIPE_MAX_ATTRIBS];
- LLVMValueRef ofmask[PIPE_MAX_ATTRIBS];
- LLVMValueRef instance_index[PIPE_MAX_ATTRIBS];
- LLVMValueRef fake_buf_ptr, fake_buf;
+ LLVMValueRef zero = lp_build_const_int32(gallivm, 0);
+ LLVMValueRef one = lp_build_const_int32(gallivm, 1);
+ LLVMValueRef vb_stride[PIPE_MAX_SHADER_INPUTS];
+ LLVMValueRef map_ptr[PIPE_MAX_SHADER_INPUTS];
+ LLVMValueRef buffer_size_adj[PIPE_MAX_SHADER_INPUTS];
+ LLVMValueRef stride_fixed[PIPE_MAX_SHADER_INPUTS];
+ LLVMValueRef ofbit[PIPE_MAX_SHADER_INPUTS];
+ LLVMValueRef instance_index[PIPE_MAX_SHADER_INPUTS];
struct draw_context *draw = llvm->draw;
const struct tgsi_shader_info *vs_info = &draw->vs.vertex_shader->info;
unsigned i, j;
- struct lp_build_context bld, bldivec;
+ struct lp_build_context bld;
struct lp_build_loop_state lp_loop;
const int vector_length = lp_native_vector_width / 32;
LLVMValueRef outputs[PIPE_MAX_SHADER_OUTPUTS][TGSI_NUM_CHANNELS];
@@ -1706,6 +1620,8 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant,
builder = gallivm->builder;
LLVMPositionBuilderAtEnd(builder, block);
+ lp_build_context_init(&bld, gallivm, lp_type_int(32));
+
memset(&vs_type, 0, sizeof vs_type);
vs_type.floating = TRUE; /* floating point values */
vs_type.sign = TRUE; /* values are signed */
@@ -1713,49 +1629,32 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant,
vs_type.width = 32; /* 32-bit float */
vs_type.length = vector_length;
- lp_build_context_init(&bld, gallivm, lp_type_uint(32));
- lp_build_context_init(&bldivec, gallivm, lp_int_type(vs_type));
-
/* hold temporary "bool" clipmask */
- clipmask_bool_ptr = lp_build_alloca(gallivm, bldivec.vec_type, "");
-
- fake_buf = lp_build_alloca_undef(gallivm,
- LLVMVectorType(LLVMInt64TypeInContext(context), 4), "");
- fake_buf = LLVMBuildBitCast(builder, fake_buf,
- LLVMPointerType(LLVMInt8TypeInContext(context), 0), "");
- fake_buf_ptr = LLVMBuildGEP(builder, fake_buf, &bld.zero, 1, "");
+ clipmask_bool_ptr = lp_build_alloca(gallivm, lp_build_int_vec_type(gallivm, vs_type), "");
+ LLVMBuildStore(builder, lp_build_zero(gallivm, lp_int_type(vs_type)), clipmask_bool_ptr);
/* code generated texture sampling */
sampler = draw_llvm_sampler_soa_create(draw_llvm_variant_key_samplers(key));
- step = lp_build_const_int32(gallivm, vector_length);
-
- ind_vec = bldivec.undef;
- for (i = 0; i < vs_type.length; i++) {
- LLVMValueRef index = lp_build_const_int32(gallivm, i);
- ind_vec = LLVMBuildInsertElement(builder, ind_vec, index, index, "");
- }
-
-
if (elts) {
- fetch_max = lp_build_min(&bld, fetch_count, fetch_elt_max);
+ start = zero;
+ end = fetch_count;
count = fetch_count;
- start = bldivec.zero;
}
else {
- fetch_max = lp_build_add(&bld, start, count);
- start = lp_build_broadcast_scalar(&bldivec, start);
- ind_vec = lp_build_add(&bldivec, start, ind_vec);
+ end = lp_build_add(&bld, start, count);
}
- fetch_max = lp_build_broadcast_scalar(&bldivec, fetch_max);
+ step = lp_build_const_int32(gallivm, vector_length);
+
+ fetch_max = LLVMBuildSub(builder, end, one, "fetch_max");
/*
* Pre-calculate everything which is constant per shader invocation.
*/
for (j = 0; j < key->nr_vertex_elements; ++j) {
- LLVMValueRef vb_buffer_offset, buffer_size, temp_ptr;
- LLVMValueRef vb_info, vbuffer_ptr, buf_offset, ofbit;
+ LLVMValueRef vb_buffer_offset, buffer_size;
+ LLVMValueRef vb_info, vbuffer_ptr;
struct pipe_vertex_element *velem = &key->vertex_element[j];
LLVMValueRef vb_index =
lp_build_const_int32(gallivm, velem->vertex_buffer_index);
@@ -1763,7 +1662,6 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant,
util_format_get_blocksize(velem->src_format));
LLVMValueRef src_offset = lp_build_const_int32(gallivm,
velem->src_offset);
- struct lp_build_if_state if_ctx;
if (velem->src_format != PIPE_FORMAT_NONE) {
vbuffer_ptr = LLVMBuildGEP(builder, vbuffers_ptr, &vb_index, 1, "");
@@ -1773,23 +1671,14 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant,
map_ptr[j] = draw_jit_dvbuffer_map(gallivm, vbuffer_ptr);
buffer_size = draw_jit_dvbuffer_size(gallivm, vbuffer_ptr);
- ofbit = NULL;
- buf_offset = lp_build_uadd_overflow(gallivm, vb_buffer_offset,
- src_offset, &ofbit);
+ ofbit[j] = NULL;
+ stride_fixed[j] = lp_build_uadd_overflow(gallivm, vb_buffer_offset,
+ src_offset, &ofbit[j]);
buffer_size_adj[j] = lp_build_usub_overflow(gallivm, buffer_size, bsize,
- &ofbit);
- buffer_size_adj[j] = lp_build_usub_overflow(gallivm, buffer_size_adj[j],
- buf_offset, &ofbit);
-
- /*
- * Unlike elts, we cannot easily set fake vertex buffers outside
- * the generated code. Hence, set fake vertex buffers here instead
- * basically, so fetch code can always fetch using offset 0, eliminating
- * all control flow (for instanced elements, we keep the control flow
- * for now as it's a scalar fetch, so don't need this).
- */
+ &ofbit[j]);
+
if (velem->instance_divisor) {
- /* Index is equal to the start instance plus the number of current
+ /* Index is equal to the start instance plus the number of current
* instance divided by the divisor. In this case we compute it as:
* index = start_instance + (instance_id / divisor)
*/
@@ -1799,55 +1688,32 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant,
velem->instance_divisor),
"instance_divisor");
instance_index[j] = lp_build_uadd_overflow(gallivm, start_instance,
- current_instance, &ofbit);
- map_ptr[j] = LLVMBuildGEP(builder, map_ptr[j], &buf_offset, 1, "");
-
- /* This is a scalar fetch, just keep the of bit */
- ofmask[j] = ofbit;
- }
- else {
- temp_ptr = lp_build_alloca_undef(gallivm,
- LLVMPointerType(LLVMInt8TypeInContext(context), 0), "");
-
- lp_build_if(&if_ctx, gallivm, ofbit);
- {
- LLVMBuildStore(builder, fake_buf_ptr, temp_ptr);
- }
- lp_build_else(&if_ctx);
- {
- map_ptr[j] = LLVMBuildGEP(builder, map_ptr[j], &buf_offset, 1, "");
- LLVMBuildStore(builder, map_ptr[j], temp_ptr);
- }
- lp_build_endif(&if_ctx);
- map_ptr[j] = LLVMBuildLoad(builder, temp_ptr, "map_ptr");
-
- /* expand to vector mask */
- ofmask[j] = LLVMBuildZExt(builder, ofbit, int32_type, "");
- ofmask[j] = lp_build_broadcast_scalar(&bldivec, ofmask[j]);
+ current_instance, &ofbit[j]);
}
if (0) {
- lp_build_printf(gallivm, "velem %d, vbuf index = %u, vb_stride = %u\n",
- lp_build_const_int32(gallivm, j),
+ lp_build_printf(gallivm, "vbuf index = %u, vb_stride is %u\n",
vb_index, vb_stride[j]);
- lp_build_printf(gallivm,
- " vb_buffer_offset = %u, src_offset = %u, buf_offset = %u\n",
- vb_buffer_offset, src_offset, buf_offset);
- lp_build_printf(gallivm, " buffer size = %u, blocksize = %u\n",
- buffer_size, bsize);
- lp_build_printf(gallivm, " instance_id = %u\n", system_values.instance_id);
+ lp_build_printf(gallivm, " vb_buffer_offset = %u, src_offset is %u\n",
+ vb_buffer_offset, src_offset);
+ lp_build_print_value(gallivm, " blocksize = ", bsize);
+ lp_build_printf(gallivm, " instance_id = %u\n",
+ system_values.instance_id);
+ lp_build_printf(gallivm, " buffer size = %u\n", buffer_size);
}
}
}
- lp_build_loop_begin(&lp_loop, gallivm, bld.zero);
+ lp_build_loop_begin(&lp_loop, gallivm, zero);
{
LLVMValueRef inputs[PIPE_MAX_SHADER_INPUTS][TGSI_NUM_CHANNELS];
+ LLVMValueRef aos_attribs[LP_MAX_VECTOR_WIDTH / 32] = { 0 };
LLVMValueRef io;
LLVMValueRef clipmask; /* holds the clipmask value */
- LLVMValueRef true_index_array;
+ LLVMValueRef true_index_array = lp_build_zero(gallivm,
+ lp_type_uint_vec(32, 32*vector_length));
+ LLVMValueRef true_indices[LP_MAX_VECTOR_WIDTH / 32];
const LLVMValueRef (*ptr_aos)[TGSI_NUM_CHANNELS];
- LLVMValueRef indices_valid;
io_itr = lp_loop.counter;
@@ -1857,40 +1723,59 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant,
io_itr, io, lp_loop.counter);
#endif
- true_index_array = lp_build_broadcast_scalar(&bldivec, lp_loop.counter);
- true_index_array = LLVMBuildAdd(builder, true_index_array, ind_vec, "");
-
- indices_valid = lp_build_compare(gallivm, bldivec.type,
- PIPE_FUNC_LESS,
- true_index_array, fetch_max);
-
- if (elts) {
-#if 0
- /*
- * AVX2 gather would not need zeroing offsets.
- * And it would not need dummy index buffer neither.
- * But for now don't bother.
- */
- true_index_array = lp_build_gather_masked(gallivm, vs_type.length,
- 32, 32, TRUE,
- fetch_elts, true_index_array,
- FALSE, indices_valid);
-#else
- true_index_array = LLVMBuildAnd(builder, true_index_array,
- indices_valid, "");
- /*
- * XXX should not have to do this, as scale can be handled
- * natively by loads (hits asserts though).
- */
- true_index_array = lp_build_shl_imm(&bldivec, true_index_array, 2);
- fetch_elts = LLVMBuildBitCast(builder, fetch_elts,
- LLVMPointerType(LLVMInt8TypeInContext(context),
- 0), "");
- true_index_array = lp_build_gather(gallivm, vs_type.length,
- 32, 32, TRUE,
- fetch_elts, true_index_array,
- FALSE);
-#endif
+ for (i = 0; i < vector_length; ++i) {
+ LLVMValueRef vert_index =
+ LLVMBuildAdd(builder,
+ lp_loop.counter,
+ lp_build_const_int32(gallivm, i), "");
+ LLVMValueRef true_index =
+ LLVMBuildAdd(builder, start, vert_index, "");
+
+ /* make sure we're not out of bounds which can happen
+ * if fetch_count % 4 != 0, because on the last iteration
+ * a few of the 4 vertex fetches will be out of bounds */
+ true_index = lp_build_min(&bld, true_index, fetch_max);
+
+ if (elts) {
+ LLVMValueRef fetch_ptr;
+ LLVMValueRef index_overflowed;
+ LLVMValueRef index_ptr =
+ lp_build_alloca(
+ gallivm,
+ lp_build_vec_type(gallivm, lp_type_int(32)), "");
+ struct lp_build_if_state if_ctx;
+ index_overflowed = LLVMBuildICmp(builder, LLVMIntUGT,
+ true_index, fetch_elt_max,
+ "index_overflowed");
+
+ lp_build_if(&if_ctx, gallivm, index_overflowed);
+ {
+ /* Generate maximum possible index so that
+ * generate_fetch can treat it just like
+ * any other overflow and return zeros.
+ * We don't have to worry about the restart
+ * primitive index because it has already been
+ * handled
+ */
+ LLVMValueRef val =
+ lp_build_const_int32(gallivm, 0xffffffff);
+ LLVMBuildStore(builder, val, index_ptr);
+ }
+ lp_build_else(&if_ctx);
+ {
+ LLVMValueRef val;
+ fetch_ptr = LLVMBuildGEP(builder, fetch_elts,
+ &true_index, 1, "");
+ val = LLVMBuildLoad(builder, fetch_ptr, "");
+ LLVMBuildStore(builder, val, index_ptr);
+ }
+ lp_build_endif(&if_ctx);
+ true_index = LLVMBuildLoad(builder, index_ptr, "true_index");
+ }
+ true_indices[i] = true_index;
+ true_index_array = LLVMBuildInsertElement(
+ gallivm->builder, true_index_array, true_index,
+ lp_build_const_int32(gallivm, i), "");
}
for (j = 0; j < key->nr_vertex_elements; ++j) {
@@ -1898,24 +1783,20 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant,
const struct util_format_description *format_desc =
util_format_description(velem->src_format);
- if (format_desc->format == PIPE_FORMAT_NONE) {
- for (i = 0; i < TGSI_NUM_CHANNELS; i++) {
- inputs[j][i] = lp_build_zero(gallivm, vs_type);
+ for (i = 0; i < vector_length; ++i) {
+ if (format_desc->format == PIPE_FORMAT_NONE) {
+ aos_attribs[i] = lp_build_const_vec(gallivm,
+ lp_float32_vec4_type(), 0);
+ }
+ else {
+ generate_fetch(gallivm, format_desc,
+ vb_stride[j], stride_fixed[j], map_ptr[j],
+ buffer_size_adj[j], ofbit[j], &aos_attribs[i],
+ velem->instance_divisor ?
+ instance_index[j] : true_indices[i]);
}
}
- else if (velem->instance_divisor) {
- fetch_instanced(gallivm, format_desc, vs_type,
- vb_stride[j], map_ptr[j],
- buffer_size_adj[j], ofmask[j],
- inputs[j], instance_index[j]);
- }
- else {
- indices_valid = lp_build_andnot(&bldivec, indices_valid, ofmask[j]);
- fetch_vector(gallivm, format_desc, vs_type,
- vb_stride[j], map_ptr[j],
- buffer_size_adj[j], inputs[j],
- true_index_array, indices_valid);
- }
+ convert_to_soa(gallivm, aos_attribs, inputs, j, vs_type);
}
/* In the paths with elts vertex id has to be unaffected by the
@@ -1929,8 +1810,9 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant,
* most 4095-vertices) we need to back out the original start
* index out of our vertex id here.
*/
- system_values.basevertex = lp_build_broadcast_scalar(&bldivec,
- vertex_id_offset);
+ system_values.basevertex = lp_build_broadcast(gallivm, lp_build_vec_type(gallivm,
+ lp_type_uint_vec(32, 32*vector_length)),
+ vertex_id_offset);
system_values.vertex_id = true_index_array;
system_values.vertex_id_nobase = LLVMBuildSub(builder, true_index_array,
system_values.basevertex, "");
@@ -1965,7 +1847,7 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant,
LLVMBuildStore(builder, temp, clipmask_bool_ptr);
}
else {
- clipmask = bldivec.zero;
+ clipmask = lp_build_const_int_vec(gallivm, lp_int_type(vs_type), 0);
}
/* do viewport mapping */
@@ -1974,7 +1856,7 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant,
}
}
else {
- clipmask = bldivec.zero;
+ clipmask = lp_build_const_int_vec(gallivm, lp_int_type(vs_type), 0);
}
/* store clipmask in vertex header,