diff options
author | Rob Clark <[email protected]> | 2019-09-27 10:15:02 -0700 |
---|---|---|
committer | Rob Clark <[email protected]> | 2019-10-18 15:08:54 -0700 |
commit | 5e08f070f0fc99246fa8aab027feeee33b774177 (patch) | |
tree | 973a6e49abedd052345cc0690bf3e70219779b72 | |
parent | 1bdde313926b698d9959294f1563264ead532b46 (diff) |
nir: add nir_lower_amul pass
Lower amul to either imul or imul24, depending on whether 24b is enough
bits to calculate an offset within the thing being dereferenced.
Signed-off-by: Rob Clark <[email protected]>
-rw-r--r-- | src/compiler/Makefile.sources | 1 | ||||
-rw-r--r-- | src/compiler/nir/meson.build | 1 | ||||
-rw-r--r-- | src/compiler/nir/nir.h | 10 | ||||
-rw-r--r-- | src/compiler/nir/nir_lower_amul.c | 316 | ||||
-rw-r--r-- | src/compiler/nir/nir_opt_algebraic.py | 6 | ||||
-rw-r--r-- | src/freedreno/ir3/ir3_nir.c | 2 | ||||
-rw-r--r-- | src/freedreno/ir3/ir3_shader.c | 2 |
7 files changed, 335 insertions, 3 deletions
diff --git a/src/compiler/Makefile.sources b/src/compiler/Makefile.sources index 5f86868792c..3a3c29e587f 100644 --- a/src/compiler/Makefile.sources +++ b/src/compiler/Makefile.sources @@ -232,6 +232,7 @@ NIR_FILES = \ nir/nir_lower_alpha_test.c \ nir/nir_lower_alu.c \ nir/nir_lower_alu_to_scalar.c \ + nir/nir_lower_amul.c \ nir/nir_lower_array_deref_of_vec.c \ nir/nir_lower_atomics_to_ssbo.c \ nir/nir_lower_bitmap.c \ diff --git a/src/compiler/nir/meson.build b/src/compiler/nir/meson.build index a485f82cf0f..56bab469b78 100644 --- a/src/compiler/nir/meson.build +++ b/src/compiler/nir/meson.build @@ -114,6 +114,7 @@ files_libnir = files( 'nir_lower_alu.c', 'nir_lower_alu_to_scalar.c', 'nir_lower_alpha_test.c', + 'nir_lower_amul.c', 'nir_lower_array_deref_of_vec.c', 'nir_lower_atomics_to_ssbo.c', 'nir_lower_bitmap.c', diff --git a/src/compiler/nir/nir.h b/src/compiler/nir/nir.h index cb572ce05ee..d002102cad8 100644 --- a/src/compiler/nir/nir.h +++ b/src/compiler/nir/nir.h @@ -2758,6 +2758,14 @@ typedef struct nir_shader_compiler_options { bool lower_rotate; /** + * Backend supports imul24, and would like to use it (when possible) + * for address/offset calculation. If true, driver should call + * nir_lower_amul(). (If not set, amul will automatically be lowered + * to imul.) + */ + bool has_imul24; + + /** * Is this the Intel vec4 backend? * * Used to inhibit algebraic optimizations that are known to be harmful on @@ -3540,6 +3548,8 @@ void nir_compact_varyings(nir_shader *producer, nir_shader *consumer, void nir_link_xfb_varyings(nir_shader *producer, nir_shader *consumer); bool nir_link_opt_varyings(nir_shader *producer, nir_shader *consumer); +bool nir_lower_amul(nir_shader *shader, + int (*type_size)(const struct glsl_type *, bool)); void nir_assign_io_var_locations(struct exec_list *var_list, unsigned *size, diff --git a/src/compiler/nir/nir_lower_amul.c b/src/compiler/nir/nir_lower_amul.c new file mode 100644 index 00000000000..4c7f10859cb --- /dev/null +++ b/src/compiler/nir/nir_lower_amul.c @@ -0,0 +1,316 @@ +/* + * Copyright © 2019 Google, Inc. + * + * Permission is hereby granted, free of charge, to any person obtaining a + * copy of this software and associated documentation files (the "Software"), + * to deal in the Software without restriction, including without limitation + * the rights to use, copy, modify, merge, publish, distribute, sublicense, + * and/or sell copies of the Software, and to permit persons to whom the + * Software is furnished to do so, subject to the following conditions: + * + * The above copyright notice and this permission notice (including the next + * paragraph) shall be included in all copies or substantial portions of the + * Software. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE + * SOFTWARE. + */ + +#include "nir.h" +#include "nir_vla.h" + +/* Lowering for amul instructions, for drivers that support imul24. + * This pass will analyze indirect derefs, and convert corresponding + * amul instructions to either imul or imul24, depending on the + * required range. + * + * 1) Analyze the uniform variables and build a table of UBOs and SSBOs + * that are either too large, or might be too large (unknown size) + * for imul24 + * + * 2) Loop thru looking at all the intrinsics, finding dereferences of + * large variables, and recursively replacing all amul instructions + * used with imul + * + * 3) Finally loop again thru all instructions replacing any remaining + * amul with imul24. At this point any remaining amul instructions + * are not involved in calculating an offset into a large variable, + * thanks to the 2nd step, so they can be safely replace with imul24. + * + * Using two passes over all the instructions lets us handle the case + * where, due to CSE, an amul is used to calculate an offset into both + * a large and small variable. + */ + +typedef struct { + int (*type_size)(const struct glsl_type *, bool); + + /* Tables of UBOs and SSBOs mapping driver_location/base whether + * they are too large to use imul24: + */ + bool *large_ubos; + bool *large_ssbos; + + /* for cases that we cannot determine UBO/SSBO index, track if *any* + * UBO/SSBO is too large for imul24: + */ + bool has_large_ubo; + bool has_large_ssbo; +} lower_state; + +/* Lower 'amul's in offset src of large variables to 'imul': */ +static bool +lower_large_src(nir_src *src, void *s) +{ + lower_state *state = s; + + assert(src->is_ssa); + + nir_instr *parent = src->ssa->parent_instr; + + /* No need to visit instructions we've already visited.. this also + * avoids infinite recursion when phi's are involved: + */ + if (parent->pass_flags) + return false; + + bool progress = nir_foreach_src(parent, lower_large_src, state); + + if (parent->type == nir_instr_type_alu) { + nir_alu_instr *alu = nir_instr_as_alu(parent); + if (alu->op == nir_op_amul) { + alu->op = nir_op_imul; + progress = true; + } + } + + parent->pass_flags = 1; + + return progress; +} + +static bool +large_ubo(lower_state *state, nir_src src) +{ + if (!nir_src_is_const(src)) + return state->has_large_ubo; + return state->large_ubos[nir_src_as_uint(src)]; +} + +static bool +large_ssbo(lower_state *state, nir_src src) +{ + if (!nir_src_is_const(src)) + return state->has_large_ssbo; + return state->large_ssbos[nir_src_as_uint(src)]; +} + +static bool +lower_intrinsic(lower_state *state, nir_intrinsic_instr *intr) +{ + switch (intr->intrinsic) { + case nir_intrinsic_load_ubo: + //# src[] = { buffer_index, offset }. + if (large_ubo(state, intr->src[0])) + return lower_large_src(&intr->src[1], state); + return false; + + case nir_intrinsic_load_ssbo: + //# src[] = { buffer_index, offset }. + if (large_ssbo(state, intr->src[0])) + return lower_large_src(&intr->src[1], state); + return false; + + case nir_intrinsic_store_ssbo: + //# src[] = { value, block_index, offset } + if (large_ssbo(state, intr->src[1])) + return lower_large_src(&intr->src[2], state); + return false; + + case nir_intrinsic_ssbo_atomic_add: + case nir_intrinsic_ssbo_atomic_imin: + case nir_intrinsic_ssbo_atomic_umin: + case nir_intrinsic_ssbo_atomic_imax: + case nir_intrinsic_ssbo_atomic_umax: + case nir_intrinsic_ssbo_atomic_and: + case nir_intrinsic_ssbo_atomic_or: + case nir_intrinsic_ssbo_atomic_xor: + case nir_intrinsic_ssbo_atomic_exchange: + case nir_intrinsic_ssbo_atomic_comp_swap: + case nir_intrinsic_ssbo_atomic_fadd: + case nir_intrinsic_ssbo_atomic_fmin: + case nir_intrinsic_ssbo_atomic_fmax: + case nir_intrinsic_ssbo_atomic_fcomp_swap: + /* 0: SSBO index + * 1: offset + */ + if (large_ssbo(state, intr->src[0])) + return lower_large_src(&intr->src[1], state); + return false; + + case nir_intrinsic_global_atomic_add: + case nir_intrinsic_global_atomic_imin: + case nir_intrinsic_global_atomic_umin: + case nir_intrinsic_global_atomic_imax: + case nir_intrinsic_global_atomic_umax: + case nir_intrinsic_global_atomic_and: + case nir_intrinsic_global_atomic_or: + case nir_intrinsic_global_atomic_xor: + case nir_intrinsic_global_atomic_exchange: + case nir_intrinsic_global_atomic_comp_swap: + case nir_intrinsic_global_atomic_fadd: + case nir_intrinsic_global_atomic_fmin: + case nir_intrinsic_global_atomic_fmax: + case nir_intrinsic_global_atomic_fcomp_swap: + /* just assume we that 24b is not sufficient: */ + return lower_large_src(&intr->src[0], state); + + /* These should all be small enough to unconditionally use imul24: */ + case nir_intrinsic_shared_atomic_add: + case nir_intrinsic_shared_atomic_imin: + case nir_intrinsic_shared_atomic_umin: + case nir_intrinsic_shared_atomic_imax: + case nir_intrinsic_shared_atomic_umax: + case nir_intrinsic_shared_atomic_and: + case nir_intrinsic_shared_atomic_or: + case nir_intrinsic_shared_atomic_xor: + case nir_intrinsic_shared_atomic_exchange: + case nir_intrinsic_shared_atomic_comp_swap: + case nir_intrinsic_shared_atomic_fadd: + case nir_intrinsic_shared_atomic_fmin: + case nir_intrinsic_shared_atomic_fmax: + case nir_intrinsic_shared_atomic_fcomp_swap: + case nir_intrinsic_load_uniform: + case nir_intrinsic_load_input: + case nir_intrinsic_load_output: + case nir_intrinsic_store_output: + default: + return false; + } +} + +static bool +lower_instr(lower_state *state, nir_instr *instr) +{ + bool progress = false; + + if (instr->type == nir_instr_type_intrinsic) { + progress |= lower_intrinsic(state, nir_instr_as_intrinsic(instr)); + } + + return progress; +} + +static bool +is_large(lower_state *state, nir_variable *var) +{ + unsigned size = state->type_size(var->type, false); + + /* if size is not known (ie. VLA) then assume the worst: */ + if (!size) + return true; + + return size >= (1 << 23); +} + +bool +nir_lower_amul(nir_shader *shader, + int (*type_size)(const struct glsl_type *, bool)) +{ + assert(shader->options->has_imul24); + assert(type_size); + + /* uniforms list actually includes ubo's and ssbo's: */ + int num_uniforms = exec_list_length(&shader->uniforms); + + NIR_VLA_FILL(bool, large_ubos, num_uniforms, 0); + NIR_VLA_FILL(bool, large_ssbos, num_uniforms, 0); + + lower_state state = { + .type_size = type_size, + .large_ubos = large_ubos, + .large_ssbos = large_ssbos, + }; + + /* Figure out which UBOs or SSBOs are large enough to be + * disqualified from imul24: + */ + nir_foreach_variable(var, &shader->uniforms) { + if (var->data.mode == nir_var_mem_ubo) { + assert(var->data.driver_location < num_uniforms); + if (is_large(&state, var)) { + state.has_large_ubo = true; + state.large_ubos[var->data.driver_location] = true; + } + } else if (var->data.mode == nir_var_mem_ssbo) { + assert(var->data.driver_location < num_uniforms); + if (is_large(&state, var)) { + state.has_large_ssbo = true; + state.large_ssbos[var->data.driver_location] = true; + } + } + } + + /* clear pass flags: */ + nir_foreach_function(function, shader) { + nir_function_impl *impl = function->impl; + if (!impl) + continue; + + nir_foreach_block(block, impl) { + nir_foreach_instr(instr, block) { + instr->pass_flags = 0; + } + } + } + + bool progress = false; + nir_foreach_function(function, shader) { + nir_function_impl *impl = function->impl; + + if (!impl) + continue; + + nir_foreach_block(block, impl) { + nir_foreach_instr(instr, block) { + progress |= lower_instr(&state, instr); + } + } + } + + /* At this point, all 'amul's used in calculating an offset into + * a large variable have been replaced with 'imul'. So remaining + * 'amul's can be replaced with 'imul24': + */ + nir_foreach_function(function, shader) { + nir_function_impl *impl = function->impl; + + if (!impl) + continue; + + nir_foreach_block(block, impl) { + nir_foreach_instr(instr, block) { + if (instr->type != nir_instr_type_alu) + continue; + + nir_alu_instr *alu = nir_instr_as_alu(instr); + if (alu->op != nir_op_amul) + continue; + + alu->op = nir_op_imul24; + progress |= true; + } + } + + nir_metadata_preserve(impl, nir_metadata_block_index | + nir_metadata_dominance); + + } + + return progress; +} diff --git a/src/compiler/nir/nir_opt_algebraic.py b/src/compiler/nir/nir_opt_algebraic.py index fa9ae87b7d7..aad0d0056e1 100644 --- a/src/compiler/nir/nir_opt_algebraic.py +++ b/src/compiler/nir/nir_opt_algebraic.py @@ -1114,9 +1114,9 @@ optimizations.extend([ (('fsign', a), ('fsub', ('b2f', ('flt', 0.0, a)), ('b2f', ('flt', a, 0.0))), 'options->lower_fsign'), # Address/offset calculations: - # for now, unconditionally convert amul to imul, this will - # change in the following patch - (('amul', a, b), ('imul', a, b)), + # Drivers supporting imul24 should use the nir_lower_amul() pass, this + # rule converts everyone else to imul: + (('amul', a, b), ('imul', a, b), '!options->has_imul24'), (('imad24_ir3', a, b, 0), ('imul24', a, b)), (('imad24_ir3', a, 0, c), (c)), diff --git a/src/freedreno/ir3/ir3_nir.c b/src/freedreno/ir3/ir3_nir.c index f9d059254fd..f2fc46db7cc 100644 --- a/src/freedreno/ir3/ir3_nir.c +++ b/src/freedreno/ir3/ir3_nir.c @@ -57,6 +57,7 @@ static const nir_shader_compiler_options options = { .use_interpolated_input_intrinsics = true, .lower_rotate = true, .lower_to_scalar = true, + .has_imul24 = true, }; /* we don't want to lower vertex_id to _zero_based on newer gpus: */ @@ -84,6 +85,7 @@ static const nir_shader_compiler_options options_a6xx = { .lower_rotate = true, .vectorize_io = true, .lower_to_scalar = true, + .has_imul24 = true, }; const nir_shader_compiler_options * diff --git a/src/freedreno/ir3/ir3_shader.c b/src/freedreno/ir3/ir3_shader.c index 344e15f38f8..4e1b1081764 100644 --- a/src/freedreno/ir3/ir3_shader.c +++ b/src/freedreno/ir3/ir3_shader.c @@ -317,6 +317,8 @@ ir3_shader_from_nir(struct ir3_compiler *compiler, nir_shader *nir) NIR_PASS_V(nir, nir_lower_io_arrays_to_elements_no_indirects, false); + NIR_PASS_V(nir, nir_lower_amul, ir3_glsl_type_size); + /* do first pass optimization, ignoring the key: */ ir3_optimize_nir(shader, nir, NULL); |