diff options
Diffstat (limited to 'src/gallium/auxiliary/draw')
-rw-r--r-- | src/gallium/auxiliary/draw/draw_llvm.c | 396 |
1 files changed, 265 insertions, 131 deletions
diff --git a/src/gallium/auxiliary/draw/draw_llvm.c b/src/gallium/auxiliary/draw/draw_llvm.c index ba86b11b178..115cb79197c 100644 --- a/src/gallium/auxiliary/draw/draw_llvm.c +++ b/src/gallium/auxiliary/draw/draw_llvm.c @@ -33,6 +33,8 @@ #include "gallivm/lp_bld_arit.h" #include "gallivm/lp_bld_arit_overflow.h" +#include "gallivm/lp_bld_bitarit.h" +#include "gallivm/lp_bld_gather.h" #include "gallivm/lp_bld_logic.h" #include "gallivm/lp_bld_const.h" #include "gallivm/lp_bld_swizzle.h" @@ -656,28 +658,28 @@ generate_vs(struct draw_llvm_variant *variant, } } + static void -generate_fetch(struct gallivm_state *gallivm, - const struct util_format_description *format_desc, - LLVMValueRef vb_stride, - LLVMValueRef stride_fixed, - LLVMValueRef map_ptr, - LLVMValueRef buffer_size_adj, - LLVMValueRef ofbit, - LLVMValueRef *res, - LLVMValueRef index) +fetch_instanced(struct gallivm_state *gallivm, + const struct util_format_description *format_desc, + struct lp_type vs_type, + LLVMValueRef vb_stride, + LLVMValueRef map_ptr, + LLVMValueRef buffer_size_adj, + LLVMValueRef ofbit, + LLVMValueRef *inputs, + LLVMValueRef index) { LLVMValueRef zero = LLVMConstNull(LLVMInt32TypeInContext(gallivm->context)); LLVMBuilderRef builder = gallivm->builder; - LLVMValueRef stride; - LLVMValueRef buffer_overflowed; + LLVMValueRef stride, buffer_overflowed, aos; LLVMValueRef temp_ptr = lp_build_alloca(gallivm, lp_build_vec_type(gallivm, lp_float32_vec4_type()), ""); struct lp_build_if_state if_ctx; + unsigned i; stride = lp_build_umul_overflow(gallivm, vb_stride, index, &ofbit); - stride = lp_build_uadd_overflow(gallivm, stride, stride_fixed, &ofbit); buffer_overflowed = LLVMBuildICmp(builder, LLVMIntUGT, stride, buffer_size_adj, @@ -685,8 +687,7 @@ generate_fetch(struct gallivm_state *gallivm, buffer_overflowed = LLVMBuildOr(builder, buffer_overflowed, ofbit, ""); if (0) { - lp_build_printf(gallivm, " stride = %u\n", stride); - lp_build_printf(gallivm, " buffer size adj = %u\n", buffer_size_adj); + lp_build_print_value(gallivm, " instance index = ", index); lp_build_print_value(gallivm, " buffer overflowed = ", buffer_overflowed); } @@ -699,27 +700,34 @@ generate_fetch(struct gallivm_state *gallivm, lp_build_else(&if_ctx); { LLVMValueRef val; - map_ptr = LLVMBuildGEP(builder, map_ptr, &stride, 1, ""); val = lp_build_fetch_rgba_aos(gallivm, format_desc, lp_float32_vec4_type(), FALSE, map_ptr, - zero, zero, zero, + stride, zero, zero, NULL); LLVMBuildStore(builder, val, temp_ptr); } lp_build_endif(&if_ctx); - *res = LLVMBuildLoad(builder, temp_ptr, "aos"); + aos = LLVMBuildLoad(builder, temp_ptr, "aos"); + + for (i = 0; i < TGSI_NUM_CHANNELS; i++) { + LLVMValueRef index = lp_build_const_int32(gallivm, i); + inputs[i] = lp_build_extract_broadcast(gallivm, + lp_float32_vec4_type(), + vs_type, aos, index); + } } + static void convert_to_soa(struct gallivm_state *gallivm, LLVMValueRef src_aos[LP_MAX_VECTOR_WIDTH / 32], - LLVMValueRef (*dst_soa)[TGSI_NUM_CHANNELS], - unsigned attrib, const struct lp_type soa_type) + LLVMValueRef dst_soa[TGSI_NUM_CHANNELS], + const struct lp_type soa_type) { unsigned j, k; struct lp_type aos_channel_type = soa_type; @@ -744,7 +752,86 @@ convert_to_soa(struct gallivm_state *gallivm, aos_channels[j] = lp_build_concat(gallivm, channel, aos_channel_type, pixels_per_channel); } - lp_build_transpose_aos(gallivm, soa_type, aos_channels, dst_soa[attrib]); + lp_build_transpose_aos(gallivm, soa_type, aos_channels, dst_soa); +} + + +static void +fetch_vector(struct gallivm_state *gallivm, + const struct util_format_description *format_desc, + struct lp_type vs_type, + LLVMValueRef vb_stride, + LLVMValueRef map_ptr, + LLVMValueRef buffer_size_adj, + LLVMValueRef ofmask, + LLVMValueRef *inputs, + LLVMValueRef indices) +{ + LLVMValueRef zero = LLVMConstNull(LLVMInt32TypeInContext(gallivm->context)); + LLVMBuilderRef builder = gallivm->builder; + struct lp_build_context blduivec; + LLVMValueRef offset, tmp, valid_mask; + LLVMValueRef aos_fetch[LP_MAX_VECTOR_WIDTH / 32]; + unsigned i; + + lp_build_context_init(&blduivec, gallivm, lp_uint_type(vs_type)); + + vb_stride = lp_build_broadcast_scalar(&blduivec, vb_stride); + buffer_size_adj = lp_build_broadcast_scalar(&blduivec, buffer_size_adj); + + /* + * Sort of interestingly, with interleaved attribs, llvm 3.7+ will + * recognize these calculations to be constant with different attribs + * (the different offset has been added to map_ptr). + * llvm 3.3, however, will not (I can't get llvm 3.4-3.6 to link...) + * + * XXX: could actually avoid this altogether (replacing by simple + * non-widening mul) by precalculating the max index instead outside + * the loop (at the cost of one scalar udiv per vertex element). + */ + offset = lp_build_mul_32_lohi_cpu(&blduivec, vb_stride, indices, &tmp); + + tmp = lp_build_compare(gallivm, blduivec.type, + PIPE_FUNC_EQUAL, tmp, blduivec.zero); + valid_mask = lp_build_andnot(&blduivec, tmp, ofmask); + + tmp = lp_build_compare(gallivm, blduivec.type, + PIPE_FUNC_LEQUAL, offset, buffer_size_adj); + valid_mask = LLVMBuildAnd(builder, tmp, valid_mask, ""); + + /* not valid elements use offset 0 */ + offset = LLVMBuildAnd(builder, offset, valid_mask, ""); + + if (0) { + lp_build_print_value(gallivm, " indices = ", indices); + lp_build_print_value(gallivm, " offsets = ", offset); + lp_build_print_value(gallivm, " valid_mask = ", valid_mask); + } + + /* + * Note: we probably really want to use SoA fetch, not AoS one (albeit + * for most formats it will amount to the same as this isn't very + * optimized). But looks dangerous since it assumes alignment. + */ + for (i = 0; i < vs_type.length; i++) { + LLVMValueRef offset1, elem; + elem = lp_build_const_int32(gallivm, i); + offset1 = LLVMBuildExtractElement(builder, offset, elem, ""); + + aos_fetch[i] = lp_build_fetch_rgba_aos(gallivm, format_desc, + lp_float32_vec4_type(), + FALSE, map_ptr, offset1, + zero, zero, NULL); + } + convert_to_soa(gallivm, aos_fetch, inputs, vs_type); + + for (i = 0; i < TGSI_NUM_CHANNELS; i++) { + inputs[i] = LLVMBuildBitCast(builder, inputs[i], blduivec.vec_type, ""); + inputs[i] = LLVMBuildAnd(builder, inputs[i], valid_mask, ""); + inputs[i] = LLVMBuildBitCast(builder, inputs[i], + lp_build_vec_type(gallivm, vs_type), ""); + + } } @@ -1491,24 +1578,22 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, LLVMBuilderRef builder; char func_name[64]; struct lp_type vs_type; - LLVMValueRef end, start; LLVMValueRef count, fetch_elts, fetch_elt_max, fetch_count; - LLVMValueRef vertex_id_offset, start_instance; + LLVMValueRef vertex_id_offset, start_instance, start; LLVMValueRef stride, step, io_itr; + LLVMValueRef ind_vec; LLVMValueRef io_ptr, vbuffers_ptr, vb_ptr; - LLVMValueRef zero = lp_build_const_int32(gallivm, 0); - LLVMValueRef one = lp_build_const_int32(gallivm, 1); - LLVMValueRef vb_stride[PIPE_MAX_SHADER_INPUTS]; - LLVMValueRef map_ptr[PIPE_MAX_SHADER_INPUTS]; - LLVMValueRef buffer_size_adj[PIPE_MAX_SHADER_INPUTS]; - LLVMValueRef stride_fixed[PIPE_MAX_SHADER_INPUTS]; - LLVMValueRef ofbit[PIPE_MAX_SHADER_INPUTS]; - LLVMValueRef instance_index[PIPE_MAX_SHADER_INPUTS]; + LLVMValueRef vb_stride[PIPE_MAX_ATTRIBS]; + LLVMValueRef map_ptr[PIPE_MAX_ATTRIBS]; + LLVMValueRef buffer_size_adj[PIPE_MAX_ATTRIBS]; + LLVMValueRef ofmask[PIPE_MAX_ATTRIBS]; + LLVMValueRef instance_index[PIPE_MAX_ATTRIBS]; + LLVMValueRef fake_buf_ptr, fake_buf; struct draw_context *draw = llvm->draw; const struct tgsi_shader_info *vs_info = &draw->vs.vertex_shader->info; unsigned i, j; - struct lp_build_context bld; + struct lp_build_context bld, bldivec, blduivec; struct lp_build_loop_state lp_loop; const int vector_length = lp_native_vector_width / 32; LLVMValueRef outputs[PIPE_MAX_SHADER_OUTPUTS][TGSI_NUM_CHANNELS]; @@ -1619,8 +1704,6 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, builder = gallivm->builder; LLVMPositionBuilderAtEnd(builder, block); - lp_build_context_init(&bld, gallivm, lp_type_int(32)); - memset(&vs_type, 0, sizeof vs_type); vs_type.floating = TRUE; /* floating point values */ vs_type.sign = TRUE; /* values are signed */ @@ -1628,32 +1711,51 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, vs_type.width = 32; /* 32-bit float */ vs_type.length = vector_length; + lp_build_context_init(&bld, gallivm, lp_type_uint(32)); + lp_build_context_init(&bldivec, gallivm, lp_int_type(vs_type)); + lp_build_context_init(&blduivec, gallivm, lp_uint_type(vs_type)); + /* hold temporary "bool" clipmask */ - clipmask_bool_ptr = lp_build_alloca(gallivm, lp_build_int_vec_type(gallivm, vs_type), ""); - LLVMBuildStore(builder, lp_build_zero(gallivm, lp_int_type(vs_type)), clipmask_bool_ptr); + clipmask_bool_ptr = lp_build_alloca(gallivm, blduivec.vec_type, ""); + + fake_buf = lp_build_alloca_undef(gallivm, + LLVMVectorType(LLVMInt64TypeInContext(context), 4), ""); + fake_buf = LLVMBuildBitCast(builder, fake_buf, + LLVMPointerType(LLVMInt8TypeInContext(context), 0), ""); + fake_buf_ptr = LLVMBuildGEP(builder, fake_buf, &bld.zero, 1, ""); /* code generated texture sampling */ sampler = draw_llvm_sampler_soa_create(draw_llvm_variant_key_samplers(key)); + step = lp_build_const_int32(gallivm, vector_length); + + ind_vec = blduivec.undef; + for (i = 0; i < vs_type.length; i++) { + LLVMValueRef index = lp_build_const_int32(gallivm, i); + ind_vec = LLVMBuildInsertElement(builder, ind_vec, index, index, ""); + } + + if (elts) { - start = zero; - end = fetch_count; + fetch_max = fetch_count; count = fetch_count; + start = blduivec.zero; } else { - end = lp_build_add(&bld, start, count); + fetch_max = lp_build_add(&bld, start, count); + start = lp_build_broadcast_scalar(&blduivec, start); + ind_vec = lp_build_add(&blduivec, start, ind_vec); } - step = lp_build_const_int32(gallivm, vector_length); - - fetch_max = LLVMBuildSub(builder, end, one, "fetch_max"); + fetch_max = LLVMBuildSub(builder, fetch_max, bld.one, "fetch_max"); + fetch_max = lp_build_broadcast_scalar(&blduivec, fetch_max); /* * Pre-calculate everything which is constant per shader invocation. */ for (j = 0; j < key->nr_vertex_elements; ++j) { - LLVMValueRef vb_buffer_offset, buffer_size; - LLVMValueRef vb_info, vbuffer_ptr; + LLVMValueRef vb_buffer_offset, buffer_size, temp_ptr; + LLVMValueRef vb_info, vbuffer_ptr, buf_offset, ofbit; struct pipe_vertex_element *velem = &key->vertex_element[j]; LLVMValueRef vb_index = lp_build_const_int32(gallivm, velem->vertex_buffer_index); @@ -1661,6 +1763,7 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, util_format_get_blocksize(velem->src_format)); LLVMValueRef src_offset = lp_build_const_int32(gallivm, velem->src_offset); + struct lp_build_if_state if_ctx; if (velem->src_format != PIPE_FORMAT_NONE) { vbuffer_ptr = LLVMBuildGEP(builder, vbuffers_ptr, &vb_index, 1, ""); @@ -1670,12 +1773,24 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, map_ptr[j] = draw_jit_dvbuffer_map(gallivm, vbuffer_ptr); buffer_size = draw_jit_dvbuffer_size(gallivm, vbuffer_ptr); - ofbit[j] = NULL; - stride_fixed[j] = lp_build_uadd_overflow(gallivm, vb_buffer_offset, - src_offset, &ofbit[j]); + ofbit = NULL; + buf_offset = lp_build_uadd_overflow(gallivm, vb_buffer_offset, + src_offset, &ofbit); buffer_size_adj[j] = lp_build_usub_overflow(gallivm, buffer_size, bsize, - &ofbit[j]); - + &ofbit); + buffer_size_adj[j] = lp_build_usub_overflow(gallivm, buffer_size_adj[j], + buf_offset, &ofbit); + + /* + * We can't easily set fake vertex buffers outside the generated code. + * Hence, set fake vertex buffers here instead basically, so fetch + * code can always fetch using offset 0, eliminating all control flow + * inside the main loop. + * (Alternatively, could have control flow per vector skipping fetch + * if ofbit is true.) + * For instanced elements, we keep the control flow for now as it's a + * scalar fetch, making things easier. + */ if (velem->instance_divisor) { /* Index is equal to the start instance plus the number of current * instance divided by the divisor. In this case we compute it as: @@ -1687,31 +1802,53 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, velem->instance_divisor), "instance_divisor"); instance_index[j] = lp_build_uadd_overflow(gallivm, start_instance, - current_instance, &ofbit[j]); + current_instance, &ofbit); + map_ptr[j] = LLVMBuildGEP(builder, map_ptr[j], &buf_offset, 1, ""); + + /* This is a scalar fetch, just keep the of bit */ + ofmask[j] = ofbit; + } + else { + temp_ptr = lp_build_alloca_undef(gallivm, + LLVMPointerType(LLVMInt8TypeInContext(context), 0), ""); + + lp_build_if(&if_ctx, gallivm, ofbit); + { + LLVMBuildStore(builder, fake_buf_ptr, temp_ptr); + } + lp_build_else(&if_ctx); + { + map_ptr[j] = LLVMBuildGEP(builder, map_ptr[j], &buf_offset, 1, ""); + LLVMBuildStore(builder, map_ptr[j], temp_ptr); + } + lp_build_endif(&if_ctx); + map_ptr[j] = LLVMBuildLoad(builder, temp_ptr, "map_ptr"); + + /* Expand to vector mask */ + ofmask[j] = LLVMBuildSExt(builder, ofbit, int32_type, ""); + ofmask[j] = lp_build_broadcast_scalar(&blduivec, ofmask[j]); } if (0) { - lp_build_printf(gallivm, "vbuf index = %u, vb_stride is %u\n", + lp_build_printf(gallivm, "velem %d, vbuf index = %u, vb_stride = %u\n", + lp_build_const_int32(gallivm, j), vb_index, vb_stride[j]); - lp_build_printf(gallivm, " vb_buffer_offset = %u, src_offset is %u\n", - vb_buffer_offset, src_offset); - lp_build_print_value(gallivm, " blocksize = ", bsize); - lp_build_printf(gallivm, " instance_id = %u\n", - system_values.instance_id); - lp_build_printf(gallivm, " buffer size = %u\n", buffer_size); + lp_build_printf(gallivm, + " vb_buffer_offset = %u, src_offset = %u, buf_offset = %u\n", + vb_buffer_offset, src_offset, buf_offset); + lp_build_printf(gallivm, " buffer size = %u, blocksize = %u\n", + buffer_size, bsize); + lp_build_printf(gallivm, " instance_id = %u\n", system_values.instance_id); } } } - lp_build_loop_begin(&lp_loop, gallivm, zero); + lp_build_loop_begin(&lp_loop, gallivm, bld.zero); { LLVMValueRef inputs[PIPE_MAX_SHADER_INPUTS][TGSI_NUM_CHANNELS]; - LLVMValueRef aos_attribs[LP_MAX_VECTOR_WIDTH / 32] = { 0 }; LLVMValueRef io; LLVMValueRef clipmask; /* holds the clipmask value */ - LLVMValueRef true_index_array = lp_build_zero(gallivm, - lp_type_uint_vec(32, 32*vector_length)); - LLVMValueRef true_indices[LP_MAX_VECTOR_WIDTH / 32]; + LLVMValueRef true_index_array; const LLVMValueRef (*ptr_aos)[TGSI_NUM_CHANNELS]; io_itr = lp_loop.counter; @@ -1722,59 +1859,54 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, io_itr, io, lp_loop.counter); #endif - for (i = 0; i < vector_length; ++i) { - LLVMValueRef vert_index = - LLVMBuildAdd(builder, - lp_loop.counter, - lp_build_const_int32(gallivm, i), ""); - LLVMValueRef true_index = - LLVMBuildAdd(builder, start, vert_index, ""); - - /* make sure we're not out of bounds which can happen - * if fetch_count % 4 != 0, because on the last iteration - * a few of the 4 vertex fetches will be out of bounds */ - true_index = lp_build_min(&bld, true_index, fetch_max); - - if (elts) { - LLVMValueRef fetch_ptr; - LLVMValueRef index_overflowed; - LLVMValueRef index_ptr = - lp_build_alloca( - gallivm, - lp_build_vec_type(gallivm, lp_type_int(32)), ""); - struct lp_build_if_state if_ctx; - index_overflowed = LLVMBuildICmp(builder, LLVMIntUGT, - true_index, fetch_elt_max, - "index_overflowed"); - - lp_build_if(&if_ctx, gallivm, index_overflowed); - { - /* Generate maximum possible index so that - * generate_fetch can treat it just like - * any other overflow and return zeros. - * We don't have to worry about the restart - * primitive index because it has already been - * handled - */ - LLVMValueRef val = - lp_build_const_int32(gallivm, 0xffffffff); - LLVMBuildStore(builder, val, index_ptr); - } - lp_build_else(&if_ctx); - { - LLVMValueRef val; - fetch_ptr = LLVMBuildGEP(builder, fetch_elts, - &true_index, 1, ""); - val = LLVMBuildLoad(builder, fetch_ptr, ""); - LLVMBuildStore(builder, val, index_ptr); - } - lp_build_endif(&if_ctx); - true_index = LLVMBuildLoad(builder, index_ptr, "true_index"); - } - true_indices[i] = true_index; - true_index_array = LLVMBuildInsertElement( - gallivm->builder, true_index_array, true_index, - lp_build_const_int32(gallivm, i), ""); + true_index_array = lp_build_broadcast_scalar(&blduivec, lp_loop.counter); + true_index_array = LLVMBuildAdd(builder, true_index_array, ind_vec, ""); + + /* + * XXX: This code is really fishy. We are required to use a int min + * here, not uint. The reason is that for some non-indexed draws, we + * might get something like MAX_UINT - 3 as start value (due to start + * vertex). So, the first 3 elements in the vector are huge, and + * limiting them to fetch_max is incorrect. By using int min, we'll + * pick that huge value - we rely on this creating an overflow (which + * is guaranteed) in the stride mul later (using (signed) cmp and + * incorporating the result into ofmask would also work). + * For the later elements, this just wraps around the indices, which + * is apparently ok... + */ + true_index_array = lp_build_min(&bldivec, true_index_array, fetch_max); + + if (elts) { + + /* + * Note: you'd expect some comparison/clamp against fetch_elt_max + * here. + * There used to be one here but it was incorrect: overflow was + * detected if index > fetch_elt_max - but the correct condition + * would be index >= fetch_elt_max (since this is just size of elts + * buffer / element size). + * Using the correct condition however will cause failures - due to + * vsplit/vcache code which rebases indices. So, as an example, if + * fetch_elt_max is just 1 and fetch_count 2, vsplit cache will + * replace all invalid indices with 0 - which in case of elt_bias + * not being zero will get a different fetch index than the valid + * index 0. So, just rely on vsplit code preventing out-of-bounds + * fetches. This is also why it's safe to do elts fetch even if there + * was no index buffer bound - the real buffer is never seen here. + */ + + /* + * XXX should not have to do this, as scale can be handled + * natively by loads (hits asserts though). + */ + true_index_array = lp_build_shl_imm(&blduivec, true_index_array, 2); + fetch_elts = LLVMBuildBitCast(builder, fetch_elts, + LLVMPointerType(LLVMInt8TypeInContext(context), + 0), ""); + true_index_array = lp_build_gather(gallivm, vs_type.length, + 32, 32, TRUE, + fetch_elts, true_index_array, + FALSE); } for (j = 0; j < key->nr_vertex_elements; ++j) { @@ -1782,20 +1914,23 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, const struct util_format_description *format_desc = util_format_description(velem->src_format); - for (i = 0; i < vector_length; ++i) { - if (format_desc->format == PIPE_FORMAT_NONE) { - aos_attribs[i] = lp_build_const_vec(gallivm, - lp_float32_vec4_type(), 0); - } - else { - generate_fetch(gallivm, format_desc, - vb_stride[j], stride_fixed[j], map_ptr[j], - buffer_size_adj[j], ofbit[j], &aos_attribs[i], - velem->instance_divisor ? - instance_index[j] : true_indices[i]); + if (format_desc->format == PIPE_FORMAT_NONE) { + for (i = 0; i < TGSI_NUM_CHANNELS; i++) { + inputs[j][i] = lp_build_zero(gallivm, vs_type); } } - convert_to_soa(gallivm, aos_attribs, inputs, j, vs_type); + else if (velem->instance_divisor) { + fetch_instanced(gallivm, format_desc, vs_type, + vb_stride[j], map_ptr[j], + buffer_size_adj[j], ofmask[j], + inputs[j], instance_index[j]); + } + else { + fetch_vector(gallivm, format_desc, vs_type, + vb_stride[j], map_ptr[j], + buffer_size_adj[j], ofmask[j], + inputs[j], true_index_array); + } } /* In the paths with elts vertex id has to be unaffected by the @@ -1809,9 +1944,8 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, * most 4095-vertices) we need to back out the original start * index out of our vertex id here. */ - system_values.basevertex = lp_build_broadcast(gallivm, lp_build_vec_type(gallivm, - lp_type_uint_vec(32, 32*vector_length)), - vertex_id_offset); + system_values.basevertex = lp_build_broadcast_scalar(&blduivec, + vertex_id_offset); system_values.vertex_id = true_index_array; system_values.vertex_id_nobase = LLVMBuildSub(builder, true_index_array, system_values.basevertex, ""); @@ -1846,7 +1980,7 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, LLVMBuildStore(builder, temp, clipmask_bool_ptr); } else { - clipmask = lp_build_const_int_vec(gallivm, lp_int_type(vs_type), 0); + clipmask = blduivec.zero; } /* do viewport mapping */ @@ -1855,7 +1989,7 @@ draw_llvm_generate(struct draw_llvm *llvm, struct draw_llvm_variant *variant, } } else { - clipmask = lp_build_const_int_vec(gallivm, lp_int_type(vs_type), 0); + clipmask = blduivec.zero; } /* store clipmask in vertex header, |