From 3fa10ffb496cc4e6d1003891cf0381bb5bec2a74 Mon Sep 17 00:00:00 2001 From: Roland Scheidegger Date: Fri, 4 Nov 2016 05:13:03 +0100 Subject: draw: use vectorized calculations for fetch Instead of doing all the math with scalars, use vectors. This means the overflow math needs to be done manually, albeit that's only really problematic for the stride/index mul, the rest has been pretty much moved outside the shader loop (albeit the mul could actually be optimized away too), where things are still scalar. Because llvm is complete fail with the zero-extend widening mul, roll our own even... To eliminate control flow in the main shader loop fetch, provide fake buffers (so index 0 is always valid to fetch). Still uses aos fetch though in the end - mostly because some more code would be needed to handle unaligned fetches in that path, and because for most formats it won't make a difference anyway (we generate some truly horrendous code for things like R16G16_something for instance). Instanced fetch however stays roughly the same as before, except that no longer the same element is fetched multiple times (I've seen a reduction of ~3 times in main shader loop size due to apparently llvm not being able to deduce it's really all the same with a couple instanced elements). Also, for elts gathering, use vectorized code as well - provide a fake elt buffer if there's no valid one bound. The generated shaders are smaller and faster to compile (not entirely sure about execution speed, but generally unless there's just single vertices to handle I would expect it to be faster - there's more opportunities for future improvements by using soa fetch). No piglit change. Reviewed-by: Jose Fonseca --- src/gallium/auxiliary/draw/draw_context.c | 5 + src/gallium/auxiliary/draw/draw_llvm.c | 436 +++++++++++++++++++----------- 2 files changed, 282 insertions(+), 159 deletions(-) (limited to 'src/gallium/auxiliary/draw') diff --git a/src/gallium/auxiliary/draw/draw_context.c b/src/gallium/auxiliary/draw/draw_context.c index 56abcff5a24..0271fe348ea 100644 --- a/src/gallium/auxiliary/draw/draw_context.c +++ b/src/gallium/auxiliary/draw/draw_context.c @@ -787,6 +787,7 @@ draw_set_indexes(struct draw_context *draw, const void *elements, unsigned elem_size, unsigned elem_buffer_space) { + static const unsigned fake_index_buf = 0; assert(elem_size == 0 || elem_size == 1 || elem_size == 2 || @@ -797,6 +798,10 @@ draw_set_indexes(struct draw_context *draw, draw->pt.user.eltMax = elem_buffer_space / elem_size; else draw->pt.user.eltMax = 0; + + if (draw->pt.user.eltMax == 0) { + draw->pt.user.elts = &fake_index_buf; + } } diff --git a/src/gallium/auxiliary/draw/draw_llvm.c b/src/gallium/auxiliary/draw/draw_llvm.c index 5b4e2a14aef..cc98d261960 100644 --- a/src/gallium/auxiliary/draw/draw_llvm.c +++ b/src/gallium/auxiliary/draw/draw_llvm.c @@ -33,6 +33,8 @@ #include "gallivm/lp_bld_arit.h" #include "gallivm/lp_bld_arit_overflow.h" +#include "gallivm/lp_bld_bitarit.h" +#include "gallivm/lp_bld_gather.h" #include "gallivm/lp_bld_logic.h" #include "gallivm/lp_bld_const.h" #include "gallivm/lp_bld_swizzle.h" @@ -656,28 +658,140 @@ generate_vs(struct draw_llvm_variant *variant, } } + +static void +convert_to_soa(struct gallivm_state *gallivm, + LLVMValueRef src_aos[LP_MAX_VECTOR_WIDTH / 32], + LLVMValueRef dst_soa[TGSI_NUM_CHANNELS], + const struct lp_type soa_type) +{ + unsigned j, k; + struct lp_type aos_channel_type = soa_type; + + LLVMValueRef aos_channels[TGSI_NUM_CHANNELS]; + unsigned pixels_per_channel = soa_type.length / TGSI_NUM_CHANNELS; + + debug_assert(TGSI_NUM_CHANNELS == 4); + debug_assert((soa_type.length % TGSI_NUM_CHANNELS) == 0); + + aos_channel_type.length >>= 1; + + for (j = 0; j < TGSI_NUM_CHANNELS; ++j) { + LLVMValueRef channel[LP_MAX_VECTOR_LENGTH] = { 0 }; + + assert(pixels_per_channel <= LP_MAX_VECTOR_LENGTH); + + for (k = 0; k < pixels_per_channel; ++k) { + channel[k] = src_aos[j + TGSI_NUM_CHANNELS * k]; + } + + aos_channels[j] = lp_build_concat(gallivm, channel, aos_channel_type, pixels_per_channel); + } + + lp_build_transpose_aos(gallivm, soa_type, aos_channels, dst_soa); +} + + static void -generate_fetch(struct gallivm_state *gallivm, - const struct util_format_description *format_desc, - LLVMValueRef vb_stride, - LLVMValueRef stride_fixed, - LLVMValueRef map_ptr, - LLVMValueRef buffer_size_adj, - LLVMValueRef ofbit, - LLVMValueRef *res, - LLVMValueRef index) +fetch_vector(struct gallivm_state *gallivm, + const struct util_format_description *format_desc, + struct lp_type vs_type, + LLVMValueRef vb_stride, + LLVMValueRef map_ptr, + LLVMValueRef buffer_size_adj, + LLVMValueRef *inputs, + LLVMValueRef indices, + LLVMValueRef valid_mask) { LLVMValueRef zero = LLVMConstNull(LLVMInt32TypeInContext(gallivm->context)); LLVMBuilderRef builder = gallivm->builder; - LLVMValueRef stride; - LLVMValueRef buffer_overflowed; + struct lp_build_context blduivec; + LLVMValueRef offset, tmp; + LLVMValueRef aos_fetch[LP_MAX_VECTOR_WIDTH / 32]; + unsigned i; + + lp_build_context_init(&blduivec, gallivm, lp_uint_type(vs_type)); + + vb_stride = lp_build_broadcast_scalar(&blduivec, vb_stride); + buffer_size_adj = lp_build_broadcast_scalar(&blduivec, buffer_size_adj); + + /* + * Sort of interestingly, with interleaved attribs, llvm 3.7+ will + * recognize these calculations to be constant with different attribs + * (the different offset has been added to map_ptr). + * llvm 3.3, however, will not (I can't get llvm 3.4-3.6 to link...) + * + * XXX: could actually avoid this altogether (replacing by simple + * non-widening mul) by precalculating the max index instead outside + * the loop (at the cost of one scalar udiv per vertex element). + */ + offset = lp_build_mul_32_lohi(&blduivec, vb_stride, indices, &tmp); + + tmp = lp_build_compare(gallivm, blduivec.type, + PIPE_FUNC_EQUAL, tmp, blduivec.zero); + valid_mask = LLVMBuildAnd(builder, tmp, valid_mask, ""); + + tmp = lp_build_compare(gallivm, blduivec.type, + PIPE_FUNC_LEQUAL, offset, buffer_size_adj); + valid_mask = LLVMBuildAnd(builder, tmp, valid_mask, ""); + + /* not valid elements use offset 0 */ + offset = LLVMBuildAnd(builder, offset, valid_mask, ""); + + if (0) { + lp_build_print_value(gallivm, " indices = ", indices); + lp_build_print_value(gallivm, " offsets = ", offset); + lp_build_print_value(gallivm, " valid_mask = ", valid_mask); + } + + /* + * Note: we probably really want to use SoA fetch, not AoS one (albeit + * for most formats it will amount to the same as this isn't very + * optimized). But looks dangerous since it assumes alignment. + */ + for (i = 0; i < vs_type.length; i++) { + LLVMValueRef offset1, elem; + elem = lp_build_const_int32(gallivm, i); + offset1 = LLVMBuildExtractElement(builder, offset, elem, ""); + + aos_fetch[i] = lp_build_fetch_rgba_aos(gallivm, format_desc, + lp_float32_vec4_type(), + FALSE, map_ptr, offset1, + zero, zero, NULL); + } + convert_to_soa(gallivm, aos_fetch, inputs, vs_type); + + for (i = 0; i < TGSI_NUM_CHANNELS; i++) { + inputs[i] = LLVMBuildBitCast(builder, inputs[i], blduivec.vec_type, ""); + inputs[i] = LLVMBuildAnd(builder, inputs[i], valid_mask, ""); + inputs[i] = LLVMBuildBitCast(builder, inputs[i], + lp_build_vec_type(gallivm, vs_type), ""); + + } +} + + +static void +fetch_instanced(struct gallivm_state *gallivm, + const struct util_format_description *format_desc, + struct lp_type vs_type, + LLVMValueRef vb_stride, + LLVMValueRef map_ptr, + LLVMValueRef buffer_size_adj, + LLVMValueRef ofbit, + LLVMValueRef *inputs, + LLVMValueRef index) +{ + LLVMValueRef zero = LLVMConstNull(LLVMInt32TypeInContext(gallivm->context)); + LLVMBuilderRef builder = gallivm->builder; + LLVMValueRef stride, buffer_overflowed, aos; LLVMValueRef temp_ptr = lp_build_alloca(gallivm, lp_build_vec_type(gallivm, lp_float32_vec4_type()), ""); struct lp_build_if_state if_ctx; + unsigned i; stride = lp_build_umul_overflow(gallivm, vb_stride, index, &ofbit); - stride = lp_build_uadd_overflow(gallivm, stride, stride_fixed, &ofbit); buffer_overflowed = LLVMBuildICmp(builder, LLVMIntUGT, stride, buffer_size_adj, @@ -685,8 +799,7 @@ generate_fetch(struct gallivm_state *gallivm, buffer_overflowed = LLVMBuildOr(builder, buffer_overflowed, ofbit, ""); if (0) { - lp_build_printf(gallivm, " stride = %u\n", stride); - lp_build_printf(gallivm, " buffer size adj = %u\n", buffer_size_adj); + lp_build_print_value(gallivm, " instance index = ", index); lp_build_print_value(gallivm, " buffer overflowed = ", buffer_overflowed); } @@ -699,52 +812,26 @@ generate_fetch(struct gallivm_state *gallivm, lp_build_else(&if_ctx); { LLVMValueRef val; - map_ptr = LLVMBuildGEP(builder, map_ptr, &stride, 1, ""); val = lp_build_fetch_rgba_aos(gallivm, format_desc, lp_float32_vec4_type(), FALSE, map_ptr, - zero, zero, zero, + stride, zero, zero, NULL); LLVMBuildStore(builder, val, temp_ptr); } lp_build_endif(&if_ctx); - *res = LLVMBuildLoad(builder, temp_ptr, "aos"); -} + aos = LLVMBuildLoad(builder, temp_ptr, "aos"); -static void -convert_to_soa(struct gallivm_state *gallivm, - LLVMValueRef src_aos[LP_MAX_VECTOR_WIDTH / 32], - LLVMValueRef (*dst_soa)[TGSI_NUM_CHANNELS], - unsigned attrib, const struct lp_type soa_type) -{ - unsigned j, k; - struct lp_type aos_channel_type = soa_type; - - LLVMValueRef aos_channels[TGSI_NUM_CHANNELS]; - unsigned pixels_per_channel = soa_type.length / TGSI_NUM_CHANNELS; - - debug_assert(TGSI_NUM_CHANNELS == 4); - debug_assert((soa_type.length % TGSI_NUM_CHANNELS) == 0); - - aos_channel_type.length >>= 1; - - for (j = 0; j < TGSI_NUM_CHANNELS; ++j) { - LLVMValueRef channel[LP_MAX_VECTOR_LENGTH] = { 0 }; - - assert(pixels_per_channel <= LP_MAX_VECTOR_LENGTH); - - for (k = 0; k < pixels_per_channel; ++k) { - channel[k] = src_aos[j + TGSI_NUM_CHANNELS * k]; - } - - aos_channels[j] = lp_build_concat(gallivm, channel, aos_channel_type, pixels_per_channel); + for (i = 0; i < TGSI_NUM_CHANNELS; i++) { + LLVMValueRef index = lp_build_const_int32(gallivm, i); + inputs[i] = lp_build_extract_broadcast(gallivm, + lp_float32_vec4_type(), + vs_type, aos, index); } - - lp_build_transpose_aos(gallivm, soa_type, aos_channels, dst_soa[attrib]); } @@ -774,6 +861,7 @@ store_aos(struct gallivm_state *gallivm, LLVMSetAlignment(LLVMBuildStore(builder, value, data_ptr), sizeof(float)); } + /** * Adjust the mask to architecture endianess. The mask will the store in struct: * @@ -1491,24 +1579,22 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, LLVMBuilderRef builder; char func_name[64]; struct lp_type vs_type; - LLVMValueRef end, start; LLVMValueRef count, fetch_elts, fetch_elt_max, fetch_count; - LLVMValueRef vertex_id_offset, start_instance; + LLVMValueRef vertex_id_offset, start_instance, start; LLVMValueRef stride, step, io_itr; + LLVMValueRef ind_vec; LLVMValueRef io_ptr, vbuffers_ptr, vb_ptr; - LLVMValueRef zero = lp_build_const_int32(gallivm, 0); - LLVMValueRef one = lp_build_const_int32(gallivm, 1); - LLVMValueRef vb_stride[PIPE_MAX_SHADER_INPUTS]; - LLVMValueRef map_ptr[PIPE_MAX_SHADER_INPUTS]; - LLVMValueRef buffer_size_adj[PIPE_MAX_SHADER_INPUTS]; - LLVMValueRef stride_fixed[PIPE_MAX_SHADER_INPUTS]; - LLVMValueRef ofbit[PIPE_MAX_SHADER_INPUTS]; - LLVMValueRef instance_index[PIPE_MAX_SHADER_INPUTS]; + LLVMValueRef vb_stride[PIPE_MAX_ATTRIBS]; + LLVMValueRef map_ptr[PIPE_MAX_ATTRIBS]; + LLVMValueRef buffer_size_adj[PIPE_MAX_ATTRIBS]; + LLVMValueRef ofmask[PIPE_MAX_ATTRIBS]; + LLVMValueRef instance_index[PIPE_MAX_ATTRIBS]; + LLVMValueRef fake_buf_ptr, fake_buf; struct draw_context *draw = llvm->draw; const struct tgsi_shader_info *vs_info = &draw->vs.vertex_shader->info; unsigned i, j; - struct lp_build_context bld; + struct lp_build_context bld, bldivec; struct lp_build_loop_state lp_loop; const int vector_length = lp_native_vector_width / 32; LLVMValueRef outputs[PIPE_MAX_SHADER_OUTPUTS][TGSI_NUM_CHANNELS]; @@ -1620,8 +1706,6 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, builder = gallivm->builder; LLVMPositionBuilderAtEnd(builder, block); - lp_build_context_init(&bld, gallivm, lp_type_int(32)); - memset(&vs_type, 0, sizeof vs_type); vs_type.floating = TRUE; /* floating point values */ vs_type.sign = TRUE; /* values are signed */ @@ -1629,32 +1713,49 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, vs_type.width = 32; /* 32-bit float */ vs_type.length = vector_length; + lp_build_context_init(&bld, gallivm, lp_type_uint(32)); + lp_build_context_init(&bldivec, gallivm, lp_int_type(vs_type)); + /* hold temporary "bool" clipmask */ - clipmask_bool_ptr = lp_build_alloca(gallivm, lp_build_int_vec_type(gallivm, vs_type), ""); - LLVMBuildStore(builder, lp_build_zero(gallivm, lp_int_type(vs_type)), clipmask_bool_ptr); + clipmask_bool_ptr = lp_build_alloca(gallivm, bldivec.vec_type, ""); + + fake_buf = lp_build_alloca_undef(gallivm, + LLVMVectorType(LLVMInt64TypeInContext(context), 4), ""); + fake_buf = LLVMBuildBitCast(builder, fake_buf, + LLVMPointerType(LLVMInt8TypeInContext(context), 0), ""); + fake_buf_ptr = LLVMBuildGEP(builder, fake_buf, &bld.zero, 1, ""); /* code generated texture sampling */ sampler = draw_llvm_sampler_soa_create(draw_llvm_variant_key_samplers(key)); + step = lp_build_const_int32(gallivm, vector_length); + + ind_vec = bldivec.undef; + for (i = 0; i < vs_type.length; i++) { + LLVMValueRef index = lp_build_const_int32(gallivm, i); + ind_vec = LLVMBuildInsertElement(builder, ind_vec, index, index, ""); + } + + if (elts) { - start = zero; - end = fetch_count; + fetch_max = lp_build_min(&bld, fetch_count, fetch_elt_max); count = fetch_count; + start = bldivec.zero; } else { - end = lp_build_add(&bld, start, count); + fetch_max = lp_build_add(&bld, start, count); + start = lp_build_broadcast_scalar(&bldivec, start); + ind_vec = lp_build_add(&bldivec, start, ind_vec); } - step = lp_build_const_int32(gallivm, vector_length); - - fetch_max = LLVMBuildSub(builder, end, one, "fetch_max"); + fetch_max = lp_build_broadcast_scalar(&bldivec, fetch_max); /* * Pre-calculate everything which is constant per shader invocation. */ for (j = 0; j < key->nr_vertex_elements; ++j) { - LLVMValueRef vb_buffer_offset, buffer_size; - LLVMValueRef vb_info, vbuffer_ptr; + LLVMValueRef vb_buffer_offset, buffer_size, temp_ptr; + LLVMValueRef vb_info, vbuffer_ptr, buf_offset, ofbit; struct pipe_vertex_element *velem = &key->vertex_element[j]; LLVMValueRef vb_index = lp_build_const_int32(gallivm, velem->vertex_buffer_index); @@ -1662,6 +1763,7 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, util_format_get_blocksize(velem->src_format)); LLVMValueRef src_offset = lp_build_const_int32(gallivm, velem->src_offset); + struct lp_build_if_state if_ctx; if (velem->src_format != PIPE_FORMAT_NONE) { vbuffer_ptr = LLVMBuildGEP(builder, vbuffers_ptr, &vb_index, 1, ""); @@ -1671,14 +1773,23 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, map_ptr[j] = draw_jit_dvbuffer_map(gallivm, vbuffer_ptr); buffer_size = draw_jit_dvbuffer_size(gallivm, vbuffer_ptr); - ofbit[j] = NULL; - stride_fixed[j] = lp_build_uadd_overflow(gallivm, vb_buffer_offset, - src_offset, &ofbit[j]); + ofbit = NULL; + buf_offset = lp_build_uadd_overflow(gallivm, vb_buffer_offset, + src_offset, &ofbit); buffer_size_adj[j] = lp_build_usub_overflow(gallivm, buffer_size, bsize, - &ofbit[j]); - + &ofbit); + buffer_size_adj[j] = lp_build_usub_overflow(gallivm, buffer_size_adj[j], + buf_offset, &ofbit); + + /* + * Unlike elts, we cannot easily set fake vertex buffers outside + * the generated code. Hence, set fake vertex buffers here instead + * basically, so fetch code can always fetch using offset 0, eliminating + * all control flow (for instanced elements, we keep the control flow + * for now as it's a scalar fetch, so don't need this). + */ if (velem->instance_divisor) { - /* Index is equal to the start instance plus the number of current + /* Index is equal to the start instance plus the number of current * instance divided by the divisor. In this case we compute it as: * index = start_instance + (instance_id / divisor) */ @@ -1688,32 +1799,55 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, velem->instance_divisor), "instance_divisor"); instance_index[j] = lp_build_uadd_overflow(gallivm, start_instance, - current_instance, &ofbit[j]); + current_instance, &ofbit); + map_ptr[j] = LLVMBuildGEP(builder, map_ptr[j], &buf_offset, 1, ""); + + /* This is a scalar fetch, just keep the of bit */ + ofmask[j] = ofbit; + } + else { + temp_ptr = lp_build_alloca_undef(gallivm, + LLVMPointerType(LLVMInt8TypeInContext(context), 0), ""); + + lp_build_if(&if_ctx, gallivm, ofbit); + { + LLVMBuildStore(builder, fake_buf_ptr, temp_ptr); + } + lp_build_else(&if_ctx); + { + map_ptr[j] = LLVMBuildGEP(builder, map_ptr[j], &buf_offset, 1, ""); + LLVMBuildStore(builder, map_ptr[j], temp_ptr); + } + lp_build_endif(&if_ctx); + map_ptr[j] = LLVMBuildLoad(builder, temp_ptr, "map_ptr"); + + /* expand to vector mask */ + ofmask[j] = LLVMBuildZExt(builder, ofbit, int32_type, ""); + ofmask[j] = lp_build_broadcast_scalar(&bldivec, ofmask[j]); } if (0) { - lp_build_printf(gallivm, "vbuf index = %u, vb_stride is %u\n", + lp_build_printf(gallivm, "velem %d, vbuf index = %u, vb_stride = %u\n", + lp_build_const_int32(gallivm, j), vb_index, vb_stride[j]); - lp_build_printf(gallivm, " vb_buffer_offset = %u, src_offset is %u\n", - vb_buffer_offset, src_offset); - lp_build_print_value(gallivm, " blocksize = ", bsize); - lp_build_printf(gallivm, " instance_id = %u\n", - system_values.instance_id); - lp_build_printf(gallivm, " buffer size = %u\n", buffer_size); + lp_build_printf(gallivm, + " vb_buffer_offset = %u, src_offset = %u, buf_offset = %u\n", + vb_buffer_offset, src_offset, buf_offset); + lp_build_printf(gallivm, " buffer size = %u, blocksize = %u\n", + buffer_size, bsize); + lp_build_printf(gallivm, " instance_id = %u\n", system_values.instance_id); } } } - lp_build_loop_begin(&lp_loop, gallivm, zero); + lp_build_loop_begin(&lp_loop, gallivm, bld.zero); { LLVMValueRef inputs[PIPE_MAX_SHADER_INPUTS][TGSI_NUM_CHANNELS]; - LLVMValueRef aos_attribs[LP_MAX_VECTOR_WIDTH / 32] = { 0 }; LLVMValueRef io; LLVMValueRef clipmask; /* holds the clipmask value */ - LLVMValueRef true_index_array = lp_build_zero(gallivm, - lp_type_uint_vec(32, 32*vector_length)); - LLVMValueRef true_indices[LP_MAX_VECTOR_WIDTH / 32]; + LLVMValueRef true_index_array; const LLVMValueRef (*ptr_aos)[TGSI_NUM_CHANNELS]; + LLVMValueRef indices_valid; io_itr = lp_loop.counter; @@ -1723,59 +1857,40 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, io_itr, io, lp_loop.counter); #endif - for (i = 0; i < vector_length; ++i) { - LLVMValueRef vert_index = - LLVMBuildAdd(builder, - lp_loop.counter, - lp_build_const_int32(gallivm, i), ""); - LLVMValueRef true_index = - LLVMBuildAdd(builder, start, vert_index, ""); - - /* make sure we're not out of bounds which can happen - * if fetch_count % 4 != 0, because on the last iteration - * a few of the 4 vertex fetches will be out of bounds */ - true_index = lp_build_min(&bld, true_index, fetch_max); - - if (elts) { - LLVMValueRef fetch_ptr; - LLVMValueRef index_overflowed; - LLVMValueRef index_ptr = - lp_build_alloca( - gallivm, - lp_build_vec_type(gallivm, lp_type_int(32)), ""); - struct lp_build_if_state if_ctx; - index_overflowed = LLVMBuildICmp(builder, LLVMIntUGT, - true_index, fetch_elt_max, - "index_overflowed"); - - lp_build_if(&if_ctx, gallivm, index_overflowed); - { - /* Generate maximum possible index so that - * generate_fetch can treat it just like - * any other overflow and return zeros. - * We don't have to worry about the restart - * primitive index because it has already been - * handled - */ - LLVMValueRef val = - lp_build_const_int32(gallivm, 0xffffffff); - LLVMBuildStore(builder, val, index_ptr); - } - lp_build_else(&if_ctx); - { - LLVMValueRef val; - fetch_ptr = LLVMBuildGEP(builder, fetch_elts, - &true_index, 1, ""); - val = LLVMBuildLoad(builder, fetch_ptr, ""); - LLVMBuildStore(builder, val, index_ptr); - } - lp_build_endif(&if_ctx); - true_index = LLVMBuildLoad(builder, index_ptr, "true_index"); - } - true_indices[i] = true_index; - true_index_array = LLVMBuildInsertElement( - gallivm->builder, true_index_array, true_index, - lp_build_const_int32(gallivm, i), ""); + true_index_array = lp_build_broadcast_scalar(&bldivec, lp_loop.counter); + true_index_array = LLVMBuildAdd(builder, true_index_array, ind_vec, ""); + + indices_valid = lp_build_compare(gallivm, bldivec.type, + PIPE_FUNC_LESS, + true_index_array, fetch_max); + + if (elts) { +#if 0 + /* + * AVX2 gather would not need zeroing offsets. + * And it would not need dummy index buffer neither. + * But for now don't bother. + */ + true_index_array = lp_build_gather_masked(gallivm, vs_type.length, + 32, 32, TRUE, + fetch_elts, true_index_array, + FALSE, indices_valid); +#else + true_index_array = LLVMBuildAnd(builder, true_index_array, + indices_valid, ""); + /* + * XXX should not have to do this, as scale can be handled + * natively by loads (hits asserts though). + */ + true_index_array = lp_build_shl_imm(&bldivec, true_index_array, 2); + fetch_elts = LLVMBuildBitCast(builder, fetch_elts, + LLVMPointerType(LLVMInt8TypeInContext(context), + 0), ""); + true_index_array = lp_build_gather(gallivm, vs_type.length, + 32, 32, TRUE, + fetch_elts, true_index_array, + FALSE); +#endif } for (j = 0; j < key->nr_vertex_elements; ++j) { @@ -1783,20 +1898,24 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, const struct util_format_description *format_desc = util_format_description(velem->src_format); - for (i = 0; i < vector_length; ++i) { - if (format_desc->format == PIPE_FORMAT_NONE) { - aos_attribs[i] = lp_build_const_vec(gallivm, - lp_float32_vec4_type(), 0); - } - else { - generate_fetch(gallivm, format_desc, - vb_stride[j], stride_fixed[j], map_ptr[j], - buffer_size_adj[j], ofbit[j], &aos_attribs[i], - velem->instance_divisor ? - instance_index[j] : true_indices[i]); + if (format_desc->format == PIPE_FORMAT_NONE) { + for (i = 0; i < TGSI_NUM_CHANNELS; i++) { + inputs[j][i] = lp_build_zero(gallivm, vs_type); } } - convert_to_soa(gallivm, aos_attribs, inputs, j, vs_type); + else if (velem->instance_divisor) { + fetch_instanced(gallivm, format_desc, vs_type, + vb_stride[j], map_ptr[j], + buffer_size_adj[j], ofmask[j], + inputs[j], instance_index[j]); + } + else { + indices_valid = lp_build_andnot(&bldivec, indices_valid, ofmask[j]); + fetch_vector(gallivm, format_desc, vs_type, + vb_stride[j], map_ptr[j], + buffer_size_adj[j], inputs[j], + true_index_array, indices_valid); + } } /* In the paths with elts vertex id has to be unaffected by the @@ -1810,9 +1929,8 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, * most 4095-vertices) we need to back out the original start * index out of our vertex id here. */ - system_values.basevertex = lp_build_broadcast(gallivm, lp_build_vec_type(gallivm, - lp_type_uint_vec(32, 32*vector_length)), - vertex_id_offset); + system_values.basevertex = lp_build_broadcast_scalar(&bldivec, + vertex_id_offset); system_values.vertex_id = true_index_array; system_values.vertex_id_nobase = LLVMBuildSub(builder, true_index_array, system_values.basevertex, ""); @@ -1847,7 +1965,7 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, LLVMBuildStore(builder, temp, clipmask_bool_ptr); } else { - clipmask = lp_build_const_int_vec(gallivm, lp_int_type(vs_type), 0); + clipmask = bldivec.zero; } /* do viewport mapping */ @@ -1856,7 +1974,7 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, } } else { - clipmask = lp_build_const_int_vec(gallivm, lp_int_type(vs_type), 0); + clipmask = bldivec.zero; } /* store clipmask in vertex header, -- cgit v1.2.3