diff options
author | Roland Scheidegger <[email protected]> | 2016-11-09 05:46:12 +0100 |
---|---|---|
committer | Roland Scheidegger <[email protected]> | 2016-11-09 05:53:16 +0100 |
commit | 4d5346aaac292213c6875aeb2814ca94b9b8e06e (patch) | |
tree | 6ffea844b957ebde402c51b4fb4e4397735db25b /src/gallium/auxiliary/draw/draw_llvm.c | |
parent | f037afb701a8df8a646414e986b3561c8a798e05 (diff) |
Revert "draw: use vectorized calculations for fetch"
Trivial. There's some regressions internally, related to overflow
behavior. I'll have to look at it at another time, some interactions
with vsplit/vcache are actually mind-blowing.
This reverts commit 3fa10ffb496cc4e6d1003891cf0381bb5bec2a74.
Diffstat (limited to 'src/gallium/auxiliary/draw/draw_llvm.c')
-rw-r--r-- | src/gallium/auxiliary/draw/draw_llvm.c | 436 |
1 files changed, 159 insertions, 277 deletions
diff --git a/src/gallium/auxiliary/draw/draw_llvm.c b/src/gallium/auxiliary/draw/draw_llvm.c index cc98d261960..5b4e2a14aef 100644 --- a/src/gallium/auxiliary/draw/draw_llvm.c +++ b/src/gallium/auxiliary/draw/draw_llvm.c @@ -33,8 +33,6 @@ #include "gallivm/lp_bld_arit.h" #include "gallivm/lp_bld_arit_overflow.h" -#include "gallivm/lp_bld_bitarit.h" -#include "gallivm/lp_bld_gather.h" #include "gallivm/lp_bld_logic.h" #include "gallivm/lp_bld_const.h" #include "gallivm/lp_bld_swizzle.h" @@ -658,140 +656,28 @@ generate_vs(struct draw_llvm_variant *variant, } } - -static void -convert_to_soa(struct gallivm_state *gallivm, - LLVMValueRef src_aos[LP_MAX_VECTOR_WIDTH / 32], - LLVMValueRef dst_soa[TGSI_NUM_CHANNELS], - const struct lp_type soa_type) -{ - unsigned j, k; - struct lp_type aos_channel_type = soa_type; - - LLVMValueRef aos_channels[TGSI_NUM_CHANNELS]; - unsigned pixels_per_channel = soa_type.length / TGSI_NUM_CHANNELS; - - debug_assert(TGSI_NUM_CHANNELS == 4); - debug_assert((soa_type.length % TGSI_NUM_CHANNELS) == 0); - - aos_channel_type.length >>= 1; - - for (j = 0; j < TGSI_NUM_CHANNELS; ++j) { - LLVMValueRef channel[LP_MAX_VECTOR_LENGTH] = { 0 }; - - assert(pixels_per_channel <= LP_MAX_VECTOR_LENGTH); - - for (k = 0; k < pixels_per_channel; ++k) { - channel[k] = src_aos[j + TGSI_NUM_CHANNELS * k]; - } - - aos_channels[j] = lp_build_concat(gallivm, channel, aos_channel_type, pixels_per_channel); - } - - lp_build_transpose_aos(gallivm, soa_type, aos_channels, dst_soa); -} - - static void -fetch_vector(struct gallivm_state *gallivm, - const struct util_format_description *format_desc, - struct lp_type vs_type, - LLVMValueRef vb_stride, - LLVMValueRef map_ptr, - LLVMValueRef buffer_size_adj, - LLVMValueRef *inputs, - LLVMValueRef indices, - LLVMValueRef valid_mask) +generate_fetch(struct gallivm_state *gallivm, + const struct util_format_description *format_desc, + LLVMValueRef vb_stride, + LLVMValueRef stride_fixed, + LLVMValueRef map_ptr, + LLVMValueRef buffer_size_adj, + LLVMValueRef ofbit, + LLVMValueRef *res, + LLVMValueRef index) { LLVMValueRef zero = LLVMConstNull(LLVMInt32TypeInContext(gallivm->context)); LLVMBuilderRef builder = gallivm->builder; - struct lp_build_context blduivec; - LLVMValueRef offset, tmp; - LLVMValueRef aos_fetch[LP_MAX_VECTOR_WIDTH / 32]; - unsigned i; - - lp_build_context_init(&blduivec, gallivm, lp_uint_type(vs_type)); - - vb_stride = lp_build_broadcast_scalar(&blduivec, vb_stride); - buffer_size_adj = lp_build_broadcast_scalar(&blduivec, buffer_size_adj); - - /* - * Sort of interestingly, with interleaved attribs, llvm 3.7+ will - * recognize these calculations to be constant with different attribs - * (the different offset has been added to map_ptr). - * llvm 3.3, however, will not (I can't get llvm 3.4-3.6 to link...) - * - * XXX: could actually avoid this altogether (replacing by simple - * non-widening mul) by precalculating the max index instead outside - * the loop (at the cost of one scalar udiv per vertex element). - */ - offset = lp_build_mul_32_lohi(&blduivec, vb_stride, indices, &tmp); - - tmp = lp_build_compare(gallivm, blduivec.type, - PIPE_FUNC_EQUAL, tmp, blduivec.zero); - valid_mask = LLVMBuildAnd(builder, tmp, valid_mask, ""); - - tmp = lp_build_compare(gallivm, blduivec.type, - PIPE_FUNC_LEQUAL, offset, buffer_size_adj); - valid_mask = LLVMBuildAnd(builder, tmp, valid_mask, ""); - - /* not valid elements use offset 0 */ - offset = LLVMBuildAnd(builder, offset, valid_mask, ""); - - if (0) { - lp_build_print_value(gallivm, " indices = ", indices); - lp_build_print_value(gallivm, " offsets = ", offset); - lp_build_print_value(gallivm, " valid_mask = ", valid_mask); - } - - /* - * Note: we probably really want to use SoA fetch, not AoS one (albeit - * for most formats it will amount to the same as this isn't very - * optimized). But looks dangerous since it assumes alignment. - */ - for (i = 0; i < vs_type.length; i++) { - LLVMValueRef offset1, elem; - elem = lp_build_const_int32(gallivm, i); - offset1 = LLVMBuildExtractElement(builder, offset, elem, ""); - - aos_fetch[i] = lp_build_fetch_rgba_aos(gallivm, format_desc, - lp_float32_vec4_type(), - FALSE, map_ptr, offset1, - zero, zero, NULL); - } - convert_to_soa(gallivm, aos_fetch, inputs, vs_type); - - for (i = 0; i < TGSI_NUM_CHANNELS; i++) { - inputs[i] = LLVMBuildBitCast(builder, inputs[i], blduivec.vec_type, ""); - inputs[i] = LLVMBuildAnd(builder, inputs[i], valid_mask, ""); - inputs[i] = LLVMBuildBitCast(builder, inputs[i], - lp_build_vec_type(gallivm, vs_type), ""); - - } -} - - -static void -fetch_instanced(struct gallivm_state *gallivm, - const struct util_format_description *format_desc, - struct lp_type vs_type, - LLVMValueRef vb_stride, - LLVMValueRef map_ptr, - LLVMValueRef buffer_size_adj, - LLVMValueRef ofbit, - LLVMValueRef *inputs, - LLVMValueRef index) -{ - LLVMValueRef zero = LLVMConstNull(LLVMInt32TypeInContext(gallivm->context)); - LLVMBuilderRef builder = gallivm->builder; - LLVMValueRef stride, buffer_overflowed, aos; + LLVMValueRef stride; + LLVMValueRef buffer_overflowed; LLVMValueRef temp_ptr = lp_build_alloca(gallivm, lp_build_vec_type(gallivm, lp_float32_vec4_type()), ""); struct lp_build_if_state if_ctx; - unsigned i; stride = lp_build_umul_overflow(gallivm, vb_stride, index, &ofbit); + stride = lp_build_uadd_overflow(gallivm, stride, stride_fixed, &ofbit); buffer_overflowed = LLVMBuildICmp(builder, LLVMIntUGT, stride, buffer_size_adj, @@ -799,7 +685,8 @@ fetch_instanced(struct gallivm_state *gallivm, buffer_overflowed = LLVMBuildOr(builder, buffer_overflowed, ofbit, ""); if (0) { - lp_build_print_value(gallivm, " instance index = ", index); + lp_build_printf(gallivm, " stride = %u\n", stride); + lp_build_printf(gallivm, " buffer size adj = %u\n", buffer_size_adj); lp_build_print_value(gallivm, " buffer overflowed = ", buffer_overflowed); } @@ -812,26 +699,52 @@ fetch_instanced(struct gallivm_state *gallivm, lp_build_else(&if_ctx); { LLVMValueRef val; + map_ptr = LLVMBuildGEP(builder, map_ptr, &stride, 1, ""); val = lp_build_fetch_rgba_aos(gallivm, format_desc, lp_float32_vec4_type(), FALSE, map_ptr, - stride, zero, zero, + zero, zero, zero, NULL); LLVMBuildStore(builder, val, temp_ptr); } lp_build_endif(&if_ctx); - aos = LLVMBuildLoad(builder, temp_ptr, "aos"); + *res = LLVMBuildLoad(builder, temp_ptr, "aos"); +} - for (i = 0; i < TGSI_NUM_CHANNELS; i++) { - LLVMValueRef index = lp_build_const_int32(gallivm, i); - inputs[i] = lp_build_extract_broadcast(gallivm, - lp_float32_vec4_type(), - vs_type, aos, index); +static void +convert_to_soa(struct gallivm_state *gallivm, + LLVMValueRef src_aos[LP_MAX_VECTOR_WIDTH / 32], + LLVMValueRef (*dst_soa)[TGSI_NUM_CHANNELS], + unsigned attrib, const struct lp_type soa_type) +{ + unsigned j, k; + struct lp_type aos_channel_type = soa_type; + + LLVMValueRef aos_channels[TGSI_NUM_CHANNELS]; + unsigned pixels_per_channel = soa_type.length / TGSI_NUM_CHANNELS; + + debug_assert(TGSI_NUM_CHANNELS == 4); + debug_assert((soa_type.length % TGSI_NUM_CHANNELS) == 0); + + aos_channel_type.length >>= 1; + + for (j = 0; j < TGSI_NUM_CHANNELS; ++j) { + LLVMValueRef channel[LP_MAX_VECTOR_LENGTH] = { 0 }; + + assert(pixels_per_channel <= LP_MAX_VECTOR_LENGTH); + + for (k = 0; k < pixels_per_channel; ++k) { + channel[k] = src_aos[j + TGSI_NUM_CHANNELS * k]; + } + + aos_channels[j] = lp_build_concat(gallivm, channel, aos_channel_type, pixels_per_channel); } + + lp_build_transpose_aos(gallivm, soa_type, aos_channels, dst_soa[attrib]); } @@ -861,7 +774,6 @@ store_aos(struct gallivm_state *gallivm, LLVMSetAlignment(LLVMBuildStore(builder, value, data_ptr), sizeof(float)); } - /** * Adjust the mask to architecture endianess. The mask will the store in struct: * @@ -1579,22 +1491,24 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, LLVMBuilderRef builder; char func_name[64]; struct lp_type vs_type; + LLVMValueRef end, start; LLVMValueRef count, fetch_elts, fetch_elt_max, fetch_count; - LLVMValueRef vertex_id_offset, start_instance, start; + LLVMValueRef vertex_id_offset, start_instance; LLVMValueRef stride, step, io_itr; - LLVMValueRef ind_vec; LLVMValueRef io_ptr, vbuffers_ptr, vb_ptr; - LLVMValueRef vb_stride[PIPE_MAX_ATTRIBS]; - LLVMValueRef map_ptr[PIPE_MAX_ATTRIBS]; - LLVMValueRef buffer_size_adj[PIPE_MAX_ATTRIBS]; - LLVMValueRef ofmask[PIPE_MAX_ATTRIBS]; - LLVMValueRef instance_index[PIPE_MAX_ATTRIBS]; - LLVMValueRef fake_buf_ptr, fake_buf; + LLVMValueRef zero = lp_build_const_int32(gallivm, 0); + LLVMValueRef one = lp_build_const_int32(gallivm, 1); + LLVMValueRef vb_stride[PIPE_MAX_SHADER_INPUTS]; + LLVMValueRef map_ptr[PIPE_MAX_SHADER_INPUTS]; + LLVMValueRef buffer_size_adj[PIPE_MAX_SHADER_INPUTS]; + LLVMValueRef stride_fixed[PIPE_MAX_SHADER_INPUTS]; + LLVMValueRef ofbit[PIPE_MAX_SHADER_INPUTS]; + LLVMValueRef instance_index[PIPE_MAX_SHADER_INPUTS]; struct draw_context *draw = llvm->draw; const struct tgsi_shader_info *vs_info = &draw->vs.vertex_shader->info; unsigned i, j; - struct lp_build_context bld, bldivec; + struct lp_build_context bld; struct lp_build_loop_state lp_loop; const int vector_length = lp_native_vector_width / 32; LLVMValueRef outputs[PIPE_MAX_SHADER_OUTPUTS][TGSI_NUM_CHANNELS]; @@ -1706,6 +1620,8 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, builder = gallivm->builder; LLVMPositionBuilderAtEnd(builder, block); + lp_build_context_init(&bld, gallivm, lp_type_int(32)); + memset(&vs_type, 0, sizeof vs_type); vs_type.floating = TRUE; /* floating point values */ vs_type.sign = TRUE; /* values are signed */ @@ -1713,49 +1629,32 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, vs_type.width = 32; /* 32-bit float */ vs_type.length = vector_length; - lp_build_context_init(&bld, gallivm, lp_type_uint(32)); - lp_build_context_init(&bldivec, gallivm, lp_int_type(vs_type)); - /* hold temporary "bool" clipmask */ - clipmask_bool_ptr = lp_build_alloca(gallivm, bldivec.vec_type, ""); - - fake_buf = lp_build_alloca_undef(gallivm, - LLVMVectorType(LLVMInt64TypeInContext(context), 4), ""); - fake_buf = LLVMBuildBitCast(builder, fake_buf, - LLVMPointerType(LLVMInt8TypeInContext(context), 0), ""); - fake_buf_ptr = LLVMBuildGEP(builder, fake_buf, &bld.zero, 1, ""); + clipmask_bool_ptr = lp_build_alloca(gallivm, lp_build_int_vec_type(gallivm, vs_type), ""); + LLVMBuildStore(builder, lp_build_zero(gallivm, lp_int_type(vs_type)), clipmask_bool_ptr); /* code generated texture sampling */ sampler = draw_llvm_sampler_soa_create(draw_llvm_variant_key_samplers(key)); - step = lp_build_const_int32(gallivm, vector_length); - - ind_vec = bldivec.undef; - for (i = 0; i < vs_type.length; i++) { - LLVMValueRef index = lp_build_const_int32(gallivm, i); - ind_vec = LLVMBuildInsertElement(builder, ind_vec, index, index, ""); - } - - if (elts) { - fetch_max = lp_build_min(&bld, fetch_count, fetch_elt_max); + start = zero; + end = fetch_count; count = fetch_count; - start = bldivec.zero; } else { - fetch_max = lp_build_add(&bld, start, count); - start = lp_build_broadcast_scalar(&bldivec, start); - ind_vec = lp_build_add(&bldivec, start, ind_vec); + end = lp_build_add(&bld, start, count); } - fetch_max = lp_build_broadcast_scalar(&bldivec, fetch_max); + step = lp_build_const_int32(gallivm, vector_length); + + fetch_max = LLVMBuildSub(builder, end, one, "fetch_max"); /* * Pre-calculate everything which is constant per shader invocation. */ for (j = 0; j < key->nr_vertex_elements; ++j) { - LLVMValueRef vb_buffer_offset, buffer_size, temp_ptr; - LLVMValueRef vb_info, vbuffer_ptr, buf_offset, ofbit; + LLVMValueRef vb_buffer_offset, buffer_size; + LLVMValueRef vb_info, vbuffer_ptr; struct pipe_vertex_element *velem = &key->vertex_element[j]; LLVMValueRef vb_index = lp_build_const_int32(gallivm, velem->vertex_buffer_index); @@ -1763,7 +1662,6 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, util_format_get_blocksize(velem->src_format)); LLVMValueRef src_offset = lp_build_const_int32(gallivm, velem->src_offset); - struct lp_build_if_state if_ctx; if (velem->src_format != PIPE_FORMAT_NONE) { vbuffer_ptr = LLVMBuildGEP(builder, vbuffers_ptr, &vb_index, 1, ""); @@ -1773,23 +1671,14 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, map_ptr[j] = draw_jit_dvbuffer_map(gallivm, vbuffer_ptr); buffer_size = draw_jit_dvbuffer_size(gallivm, vbuffer_ptr); - ofbit = NULL; - buf_offset = lp_build_uadd_overflow(gallivm, vb_buffer_offset, - src_offset, &ofbit); + ofbit[j] = NULL; + stride_fixed[j] = lp_build_uadd_overflow(gallivm, vb_buffer_offset, + src_offset, &ofbit[j]); buffer_size_adj[j] = lp_build_usub_overflow(gallivm, buffer_size, bsize, - &ofbit); - buffer_size_adj[j] = lp_build_usub_overflow(gallivm, buffer_size_adj[j], - buf_offset, &ofbit); - - /* - * Unlike elts, we cannot easily set fake vertex buffers outside - * the generated code. Hence, set fake vertex buffers here instead - * basically, so fetch code can always fetch using offset 0, eliminating - * all control flow (for instanced elements, we keep the control flow - * for now as it's a scalar fetch, so don't need this). - */ + &ofbit[j]); + if (velem->instance_divisor) { - /* Index is equal to the start instance plus the number of current + /* Index is equal to the start instance plus the number of current * instance divided by the divisor. In this case we compute it as: * index = start_instance + (instance_id / divisor) */ @@ -1799,55 +1688,32 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, velem->instance_divisor), "instance_divisor"); instance_index[j] = lp_build_uadd_overflow(gallivm, start_instance, - current_instance, &ofbit); - map_ptr[j] = LLVMBuildGEP(builder, map_ptr[j], &buf_offset, 1, ""); - - /* This is a scalar fetch, just keep the of bit */ - ofmask[j] = ofbit; - } - else { - temp_ptr = lp_build_alloca_undef(gallivm, - LLVMPointerType(LLVMInt8TypeInContext(context), 0), ""); - - lp_build_if(&if_ctx, gallivm, ofbit); - { - LLVMBuildStore(builder, fake_buf_ptr, temp_ptr); - } - lp_build_else(&if_ctx); - { - map_ptr[j] = LLVMBuildGEP(builder, map_ptr[j], &buf_offset, 1, ""); - LLVMBuildStore(builder, map_ptr[j], temp_ptr); - } - lp_build_endif(&if_ctx); - map_ptr[j] = LLVMBuildLoad(builder, temp_ptr, "map_ptr"); - - /* expand to vector mask */ - ofmask[j] = LLVMBuildZExt(builder, ofbit, int32_type, ""); - ofmask[j] = lp_build_broadcast_scalar(&bldivec, ofmask[j]); + current_instance, &ofbit[j]); } if (0) { - lp_build_printf(gallivm, "velem %d, vbuf index = %u, vb_stride = %u\n", - lp_build_const_int32(gallivm, j), + lp_build_printf(gallivm, "vbuf index = %u, vb_stride is %u\n", vb_index, vb_stride[j]); - lp_build_printf(gallivm, - " vb_buffer_offset = %u, src_offset = %u, buf_offset = %u\n", - vb_buffer_offset, src_offset, buf_offset); - lp_build_printf(gallivm, " buffer size = %u, blocksize = %u\n", - buffer_size, bsize); - lp_build_printf(gallivm, " instance_id = %u\n", system_values.instance_id); + lp_build_printf(gallivm, " vb_buffer_offset = %u, src_offset is %u\n", + vb_buffer_offset, src_offset); + lp_build_print_value(gallivm, " blocksize = ", bsize); + lp_build_printf(gallivm, " instance_id = %u\n", + system_values.instance_id); + lp_build_printf(gallivm, " buffer size = %u\n", buffer_size); } } } - lp_build_loop_begin(&lp_loop, gallivm, bld.zero); + lp_build_loop_begin(&lp_loop, gallivm, zero); { LLVMValueRef inputs[PIPE_MAX_SHADER_INPUTS][TGSI_NUM_CHANNELS]; + LLVMValueRef aos_attribs[LP_MAX_VECTOR_WIDTH / 32] = { 0 }; LLVMValueRef io; LLVMValueRef clipmask; /* holds the clipmask value */ - LLVMValueRef true_index_array; + LLVMValueRef true_index_array = lp_build_zero(gallivm, + lp_type_uint_vec(32, 32*vector_length)); + LLVMValueRef true_indices[LP_MAX_VECTOR_WIDTH / 32]; const LLVMValueRef (*ptr_aos)[TGSI_NUM_CHANNELS]; - LLVMValueRef indices_valid; io_itr = lp_loop.counter; @@ -1857,40 +1723,59 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, io_itr, io, lp_loop.counter); #endif - true_index_array = lp_build_broadcast_scalar(&bldivec, lp_loop.counter); - true_index_array = LLVMBuildAdd(builder, true_index_array, ind_vec, ""); - - indices_valid = lp_build_compare(gallivm, bldivec.type, - PIPE_FUNC_LESS, - true_index_array, fetch_max); - - if (elts) { -#if 0 - /* - * AVX2 gather would not need zeroing offsets. - * And it would not need dummy index buffer neither. - * But for now don't bother. - */ - true_index_array = lp_build_gather_masked(gallivm, vs_type.length, - 32, 32, TRUE, - fetch_elts, true_index_array, - FALSE, indices_valid); -#else - true_index_array = LLVMBuildAnd(builder, true_index_array, - indices_valid, ""); - /* - * XXX should not have to do this, as scale can be handled - * natively by loads (hits asserts though). - */ - true_index_array = lp_build_shl_imm(&bldivec, true_index_array, 2); - fetch_elts = LLVMBuildBitCast(builder, fetch_elts, - LLVMPointerType(LLVMInt8TypeInContext(context), - 0), ""); - true_index_array = lp_build_gather(gallivm, vs_type.length, - 32, 32, TRUE, - fetch_elts, true_index_array, - FALSE); -#endif + for (i = 0; i < vector_length; ++i) { + LLVMValueRef vert_index = + LLVMBuildAdd(builder, + lp_loop.counter, + lp_build_const_int32(gallivm, i), ""); + LLVMValueRef true_index = + LLVMBuildAdd(builder, start, vert_index, ""); + + /* make sure we're not out of bounds which can happen + * if fetch_count % 4 != 0, because on the last iteration + * a few of the 4 vertex fetches will be out of bounds */ + true_index = lp_build_min(&bld, true_index, fetch_max); + + if (elts) { + LLVMValueRef fetch_ptr; + LLVMValueRef index_overflowed; + LLVMValueRef index_ptr = + lp_build_alloca( + gallivm, + lp_build_vec_type(gallivm, lp_type_int(32)), ""); + struct lp_build_if_state if_ctx; + index_overflowed = LLVMBuildICmp(builder, LLVMIntUGT, + true_index, fetch_elt_max, + "index_overflowed"); + + lp_build_if(&if_ctx, gallivm, index_overflowed); + { + /* Generate maximum possible index so that + * generate_fetch can treat it just like + * any other overflow and return zeros. + * We don't have to worry about the restart + * primitive index because it has already been + * handled + */ + LLVMValueRef val = + lp_build_const_int32(gallivm, 0xffffffff); + LLVMBuildStore(builder, val, index_ptr); + } + lp_build_else(&if_ctx); + { + LLVMValueRef val; + fetch_ptr = LLVMBuildGEP(builder, fetch_elts, + &true_index, 1, ""); + val = LLVMBuildLoad(builder, fetch_ptr, ""); + LLVMBuildStore(builder, val, index_ptr); + } + lp_build_endif(&if_ctx); + true_index = LLVMBuildLoad(builder, index_ptr, "true_index"); + } + true_indices[i] = true_index; + true_index_array = LLVMBuildInsertElement( + gallivm->builder, true_index_array, true_index, + lp_build_const_int32(gallivm, i), ""); } for (j = 0; j < key->nr_vertex_elements; ++j) { @@ -1898,24 +1783,20 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, const struct util_format_description *format_desc = util_format_description(velem->src_format); - if (format_desc->format == PIPE_FORMAT_NONE) { - for (i = 0; i < TGSI_NUM_CHANNELS; i++) { - inputs[j][i] = lp_build_zero(gallivm, vs_type); + for (i = 0; i < vector_length; ++i) { + if (format_desc->format == PIPE_FORMAT_NONE) { + aos_attribs[i] = lp_build_const_vec(gallivm, + lp_float32_vec4_type(), 0); + } + else { + generate_fetch(gallivm, format_desc, + vb_stride[j], stride_fixed[j], map_ptr[j], + buffer_size_adj[j], ofbit[j], &aos_attribs[i], + velem->instance_divisor ? + instance_index[j] : true_indices[i]); } } - else if (velem->instance_divisor) { - fetch_instanced(gallivm, format_desc, vs_type, - vb_stride[j], map_ptr[j], - buffer_size_adj[j], ofmask[j], - inputs[j], instance_index[j]); - } - else { - indices_valid = lp_build_andnot(&bldivec, indices_valid, ofmask[j]); - fetch_vector(gallivm, format_desc, vs_type, - vb_stride[j], map_ptr[j], - buffer_size_adj[j], inputs[j], - true_index_array, indices_valid); - } + convert_to_soa(gallivm, aos_attribs, inputs, j, vs_type); } /* In the paths with elts vertex id has to be unaffected by the @@ -1929,8 +1810,9 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, * most 4095-vertices) we need to back out the original start * index out of our vertex id here. */ - system_values.basevertex = lp_build_broadcast_scalar(&bldivec, - vertex_id_offset); + system_values.basevertex = lp_build_broadcast(gallivm, lp_build_vec_type(gallivm, + lp_type_uint_vec(32, 32*vector_length)), + vertex_id_offset); system_values.vertex_id = true_index_array; system_values.vertex_id_nobase = LLVMBuildSub(builder, true_index_array, system_values.basevertex, ""); @@ -1965,7 +1847,7 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, LLVMBuildStore(builder, temp, clipmask_bool_ptr); } else { - clipmask = bldivec.zero; + clipmask = lp_build_const_int_vec(gallivm, lp_int_type(vs_type), 0); } /* do viewport mapping */ @@ -1974,7 +1856,7 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, } } else { - clipmask = bldivec.zero; + clipmask = lp_build_const_int_vec(gallivm, lp_int_type(vs_type), 0); } /* store clipmask in vertex header, |