summaryrefslogtreecommitdiffstats
path: root/src/compiler
diff options
context:
space:
mode:
authorSamuel Iglesias Gonsálvez <[email protected]>2018-06-20 09:11:14 +0200
committerAndres Gomez <[email protected]>2019-09-17 23:39:18 +0300
commitf7d73db353e5b6416c98a8c05c585ea79b3eada2 (patch)
tree955c1f81406e37115b6aaa019a52b66298b36267 /src/compiler
parent45668a8be112878a2d34a6ba11684b0ff60df4ed (diff)
nir: add support for flushing to zero denorm constants
v2: - Refactor conditions and shared function (Connor). - Move code to nir_eval_const_opcode() (Connor). - Don't flush to zero on fquantize2f16 From Vulkan spec, VK_KHR_shader_float_controls section: "3) Do denorm and rounding mode controls apply to OpSpecConstantOp? RESOLVED: Yes, except when the opcode is OpQuantizeToF16." v3: - Fix bit size (Connor). - Fix execution mode on nir_loop_analize (Connor). v4: - Adapt after API changes to nir_eval_const_opcode (Andres). v5: - Simplify constant_denorm_flush_to_zero (Caio). v6: - Adapt after API changes and to use the new constant constructors (Andres). - Replace MAYBE_UNUSED with UNUSED as the first is going away (Andres). v7: - Adapt to newly added calls (Andres). - Simplified the auxiliary to flush denorms to zero (Caio). - Updated to renamed supported capabilities member (Andres). Signed-off-by: Samuel Iglesias Gonsálvez <[email protected]> Signed-off-by: Andres Gomez <[email protected]> Reviewed-by: Connor Abbott <[email protected]> [v4] Reviewed-by: Caio Marcelo de Oliveira Filho <[email protected]>
Diffstat (limited to 'src/compiler')
-rw-r--r--src/compiler/nir/nir_constant_expressions.h3
-rw-r--r--src/compiler/nir/nir_constant_expressions.py61
-rw-r--r--src/compiler/nir/nir_loop_analyze.c67
-rw-r--r--src/compiler/nir/nir_opt_constant_folding.c15
-rw-r--r--src/compiler/spirv/spirv_to_nir.c4
5 files changed, 109 insertions, 41 deletions
diff --git a/src/compiler/nir/nir_constant_expressions.h b/src/compiler/nir/nir_constant_expressions.h
index 087663f7480..6450d5c3d03 100644
--- a/src/compiler/nir/nir_constant_expressions.h
+++ b/src/compiler/nir/nir_constant_expressions.h
@@ -32,6 +32,7 @@
void nir_eval_const_opcode(nir_op op, nir_const_value *dest,
unsigned num_components, unsigned bit_size,
- nir_const_value **src);
+ nir_const_value **src,
+ unsigned float_controls_execution_mode);
#endif /* NIR_CONSTANT_EXPRESSIONS_H */
diff --git a/src/compiler/nir/nir_constant_expressions.py b/src/compiler/nir/nir_constant_expressions.py
index fca14602855..1df97aa1000 100644
--- a/src/compiler/nir/nir_constant_expressions.py
+++ b/src/compiler/nir/nir_constant_expressions.py
@@ -69,6 +69,27 @@ template = """\
#define MAX_UINT_FOR_SIZE(bits) (UINT64_MAX >> (64 - (bits)))
/**
+ * \brief Checks if the provided value is a denorm and flushes it to zero.
+ */
+static void
+constant_denorm_flush_to_zero(nir_const_value *value, unsigned bit_size)
+{
+ switch(bit_size) {
+ case 64:
+ if (0 == (value->u64 & 0x7ff0000000000000))
+ value->u64 &= 0x8000000000000000;
+ break;
+ case 32:
+ if (0 == (value->u32 & 0x7f800000))
+ value->u32 &= 0x80000000;
+ break;
+ case 16:
+ if (0 == (value->u16 & 0x7c00))
+ value->u16 &= 0x8000;
+ }
+}
+
+/**
* Evaluate one component of packSnorm4x8.
*/
static uint8_t
@@ -262,7 +283,7 @@ struct ${type}${width}_vec {
% endfor
% endfor
-<%def name="evaluate_op(op, bit_size)">
+<%def name="evaluate_op(op, bit_size, execution_mode)">
<%
output_type = type_add_size(op.output_type, bit_size)
input_types = [type_add_size(type_, bit_size) for type_ in op.input_types]
@@ -345,6 +366,18 @@ struct ${type}${width}_vec {
% else:
_dst_val[_i].${get_const_field(output_type)} = dst;
% endif
+
+ % if op.name != "fquantize2f16" and type_base_type(output_type) == "float":
+ % if type_has_size(output_type):
+ if (nir_is_denorm_flush_to_zero(execution_mode, ${type_size(output_type)})) {
+ constant_denorm_flush_to_zero(&_dst_val[_i], ${type_size(output_type)});
+ }
+ % else:
+ if (nir_is_denorm_flush_to_zero(execution_mode, ${bit_size})) {
+ constant_denorm_flush_to_zero(&_dst_val[i], bit_size);
+ }
+ %endif
+ % endif
}
% else:
## In the non-per-component case, create a struct dst with
@@ -377,6 +410,18 @@ struct ${type}${width}_vec {
% else:
_dst_val[${k}].${get_const_field(output_type)} = dst.${"xyzw"[k]};
% endif
+
+ % if op.name != "fquantize2f16" and type_base_type(output_type) == "float":
+ % if type_has_size(output_type):
+ if (nir_is_denorm_flush_to_zero(execution_mode, ${type_size(output_type)})) {
+ constant_denorm_flush_to_zero(&_dst_val[${k}], ${type_size(output_type)});
+ }
+ % else:
+ if (nir_is_denorm_flush_to_zero(execution_mode, ${bit_size})) {
+ constant_denorm_flush_to_zero(&_dst_val[${k}], bit_size);
+ }
+ % endif
+ % endif
% endfor
% endif
</%def>
@@ -386,13 +431,14 @@ static void
evaluate_${name}(nir_const_value *_dst_val,
UNUSED unsigned num_components,
${"UNUSED" if op_bit_sizes(op) is None else ""} unsigned bit_size,
- UNUSED nir_const_value **_src)
+ UNUSED nir_const_value **_src,
+ UNUSED unsigned execution_mode)
{
% if op_bit_sizes(op) is not None:
switch (bit_size) {
% for bit_size in op_bit_sizes(op):
case ${bit_size}: {
- ${evaluate_op(op, bit_size)}
+ ${evaluate_op(op, bit_size, execution_mode)}
break;
}
% endfor
@@ -401,7 +447,7 @@ evaluate_${name}(nir_const_value *_dst_val,
unreachable("unknown bit width");
}
% else:
- ${evaluate_op(op, 0)}
+ ${evaluate_op(op, 0, execution_mode)}
% endif
}
% endfor
@@ -409,12 +455,13 @@ evaluate_${name}(nir_const_value *_dst_val,
void
nir_eval_const_opcode(nir_op op, nir_const_value *dest,
unsigned num_components, unsigned bit_width,
- nir_const_value **src)
+ nir_const_value **src,
+ unsigned float_controls_execution_mode)
{
switch (op) {
% for name in sorted(opcodes.keys()):
case nir_op_${name}:
- evaluate_${name}(dest, num_components, bit_width, src);
+ evaluate_${name}(dest, num_components, bit_width, src, float_controls_execution_mode);
return;
% endfor
default:
@@ -425,6 +472,8 @@ nir_eval_const_opcode(nir_op op, nir_const_value *dest,
from mako.template import Template
print(Template(template).render(opcodes=opcodes, type_sizes=type_sizes,
+ type_base_type=type_base_type,
+ type_size=type_size,
type_has_size=type_has_size,
type_add_size=type_add_size,
op_bit_sizes=op_bit_sizes,
diff --git a/src/compiler/nir/nir_loop_analyze.c b/src/compiler/nir/nir_loop_analyze.c
index 4689d2230af..c2473421215 100644
--- a/src/compiler/nir/nir_loop_analyze.c
+++ b/src/compiler/nir/nir_loop_analyze.c
@@ -589,29 +589,32 @@ try_find_limit_of_alu(nir_ssa_scalar limit, nir_const_value *limit_val,
}
static nir_const_value
-eval_const_unop(nir_op op, unsigned bit_size, nir_const_value src0)
+eval_const_unop(nir_op op, unsigned bit_size, nir_const_value src0,
+ unsigned execution_mode)
{
assert(nir_op_infos[op].num_inputs == 1);
nir_const_value dest;
nir_const_value *src[1] = { &src0 };
- nir_eval_const_opcode(op, &dest, 1, bit_size, src);
+ nir_eval_const_opcode(op, &dest, 1, bit_size, src, execution_mode);
return dest;
}
static nir_const_value
eval_const_binop(nir_op op, unsigned bit_size,
- nir_const_value src0, nir_const_value src1)
+ nir_const_value src0, nir_const_value src1,
+ unsigned execution_mode)
{
assert(nir_op_infos[op].num_inputs == 2);
nir_const_value dest;
nir_const_value *src[2] = { &src0, &src1 };
- nir_eval_const_opcode(op, &dest, 1, bit_size, src);
+ nir_eval_const_opcode(op, &dest, 1, bit_size, src, execution_mode);
return dest;
}
static int32_t
get_iteration(nir_op cond_op, nir_const_value initial, nir_const_value step,
- nir_const_value limit, unsigned bit_size)
+ nir_const_value limit, unsigned bit_size,
+ unsigned execution_mode)
{
nir_const_value span, iter;
@@ -620,23 +623,29 @@ get_iteration(nir_op cond_op, nir_const_value initial, nir_const_value step,
case nir_op_ilt:
case nir_op_ieq:
case nir_op_ine:
- span = eval_const_binop(nir_op_isub, bit_size, limit, initial);
- iter = eval_const_binop(nir_op_idiv, bit_size, span, step);
+ span = eval_const_binop(nir_op_isub, bit_size, limit, initial,
+ execution_mode);
+ iter = eval_const_binop(nir_op_idiv, bit_size, span, step,
+ execution_mode);
break;
case nir_op_uge:
case nir_op_ult:
- span = eval_const_binop(nir_op_isub, bit_size, limit, initial);
- iter = eval_const_binop(nir_op_udiv, bit_size, span, step);
+ span = eval_const_binop(nir_op_isub, bit_size, limit, initial,
+ execution_mode);
+ iter = eval_const_binop(nir_op_udiv, bit_size, span, step,
+ execution_mode);
break;
case nir_op_fge:
case nir_op_flt:
case nir_op_feq:
case nir_op_fne:
- span = eval_const_binop(nir_op_fsub, bit_size, limit, initial);
- iter = eval_const_binop(nir_op_fdiv, bit_size, span, step);
- iter = eval_const_unop(nir_op_f2i64, bit_size, iter);
+ span = eval_const_binop(nir_op_fsub, bit_size, limit, initial,
+ execution_mode);
+ iter = eval_const_binop(nir_op_fdiv, bit_size, span,
+ step, execution_mode);
+ iter = eval_const_unop(nir_op_f2i64, bit_size, iter, execution_mode);
break;
default:
@@ -654,7 +663,8 @@ will_break_on_first_iteration(nir_const_value step,
nir_op cond_op, unsigned bit_size,
nir_const_value initial,
nir_const_value limit,
- bool limit_rhs, bool invert_cond)
+ bool limit_rhs, bool invert_cond,
+ unsigned execution_mode)
{
if (trip_offset == 1) {
nir_op add_op;
@@ -670,7 +680,8 @@ will_break_on_first_iteration(nir_const_value step,
unreachable("Unhandled induction variable base type!");
}
- initial = eval_const_binop(add_op, bit_size, initial, step);
+ initial = eval_const_binop(add_op, bit_size, initial, step,
+ execution_mode);
}
nir_const_value *src[2];
@@ -679,7 +690,7 @@ will_break_on_first_iteration(nir_const_value step,
/* Evaluate the loop exit condition */
nir_const_value result;
- nir_eval_const_opcode(cond_op, &result, 1, bit_size, src);
+ nir_eval_const_opcode(cond_op, &result, 1, bit_size, src, execution_mode);
return invert_cond ? !result.b : result.b;
}
@@ -688,7 +699,8 @@ static bool
test_iterations(int32_t iter_int, nir_const_value step,
nir_const_value limit, nir_op cond_op, unsigned bit_size,
nir_alu_type induction_base_type,
- nir_const_value initial, bool limit_rhs, bool invert_cond)
+ nir_const_value initial, bool limit_rhs, bool invert_cond,
+ unsigned execution_mode)
{
assert(nir_op_infos[cond_op].num_inputs == 2);
@@ -715,11 +727,11 @@ test_iterations(int32_t iter_int, nir_const_value step,
* step the induction variable each iteration.
*/
nir_const_value mul_result =
- eval_const_binop(mul_op, bit_size, iter_src, step);
+ eval_const_binop(mul_op, bit_size, iter_src, step, execution_mode);
/* Add the initial value to the accumulated induction variable total */
nir_const_value add_result =
- eval_const_binop(add_op, bit_size, mul_result, initial);
+ eval_const_binop(add_op, bit_size, mul_result, initial, execution_mode);
nir_const_value *src[2];
src[limit_rhs ? 0 : 1] = &add_result;
@@ -727,7 +739,7 @@ test_iterations(int32_t iter_int, nir_const_value step,
/* Evaluate the loop exit condition */
nir_const_value result;
- nir_eval_const_opcode(cond_op, &result, 1, bit_size, src);
+ nir_eval_const_opcode(cond_op, &result, 1, bit_size, src, execution_mode);
return invert_cond ? !result.b : result.b;
}
@@ -736,7 +748,7 @@ static int
calculate_iterations(nir_const_value initial, nir_const_value step,
nir_const_value limit, nir_alu_instr *alu,
nir_ssa_scalar cond, nir_op alu_op, bool limit_rhs,
- bool invert_cond)
+ bool invert_cond, unsigned execution_mode)
{
/* nir_op_isub should have been lowered away by this point */
assert(alu->op != nir_op_isub);
@@ -786,11 +798,13 @@ calculate_iterations(nir_const_value initial, nir_const_value step,
*/
if (will_break_on_first_iteration(step, induction_base_type, trip_offset,
alu_op, bit_size, initial,
- limit, limit_rhs, invert_cond)) {
+ limit, limit_rhs, invert_cond,
+ execution_mode)) {
return 0;
}
- int iter_int = get_iteration(alu_op, initial, step, limit, bit_size);
+ int iter_int = get_iteration(alu_op, initial, step, limit, bit_size,
+ execution_mode);
/* If iter_int is negative the loop is ill-formed or is the conditional is
* unsigned with a huge iteration count so don't bother going any further.
@@ -812,7 +826,7 @@ calculate_iterations(nir_const_value initial, nir_const_value step,
if (test_iterations(iter_bias, step, limit, alu_op, bit_size,
induction_base_type, initial,
- limit_rhs, invert_cond)) {
+ limit_rhs, invert_cond, execution_mode)) {
return iter_bias > 0 ? iter_bias - trip_offset : iter_bias;
}
}
@@ -950,7 +964,7 @@ try_find_trip_count_vars_in_iand(nir_ssa_scalar *cond,
* loop.
*/
static void
-find_trip_count(loop_info_state *state)
+find_trip_count(loop_info_state *state, unsigned execution_mode)
{
bool trip_count_known = true;
bool guessed_trip_count = false;
@@ -1063,7 +1077,8 @@ find_trip_count(loop_info_state *state)
int iterations = calculate_iterations(initial_val, step_val, limit_val,
ind_var->alu, cond,
alu_op, limit_rhs,
- terminator->continue_from_then);
+ terminator->continue_from_then,
+ execution_mode);
/* Where we not able to calculate the iteration count */
if (iterations == -1) {
@@ -1203,7 +1218,7 @@ get_loop_info(loop_info_state *state, nir_function_impl *impl)
return;
/* Run through each of the terminators and try to compute a trip-count */
- find_trip_count(state);
+ find_trip_count(state, impl->function->shader->info.float_controls_execution_mode);
nir_foreach_block_in_cf_node(block, &state->loop->cf_node) {
if (force_unroll_heuristics(state, block)) {
diff --git a/src/compiler/nir/nir_opt_constant_folding.c b/src/compiler/nir/nir_opt_constant_folding.c
index 4123bc8af6d..80d9c613e5b 100644
--- a/src/compiler/nir/nir_opt_constant_folding.c
+++ b/src/compiler/nir/nir_opt_constant_folding.c
@@ -33,7 +33,7 @@
*/
static bool
-constant_fold_alu_instr(nir_alu_instr *instr, void *mem_ctx)
+constant_fold_alu_instr(nir_alu_instr *instr, void *mem_ctx, unsigned execution_mode)
{
nir_const_value src[NIR_MAX_VEC_COMPONENTS][NIR_MAX_VEC_COMPONENTS];
@@ -88,7 +88,7 @@ constant_fold_alu_instr(nir_alu_instr *instr, void *mem_ctx)
for (unsigned i = 0; i < nir_op_infos[instr->op].num_inputs; ++i)
srcs[i] = src[i];
nir_eval_const_opcode(instr->op, dest, instr->dest.dest.ssa.num_components,
- bit_size, srcs);
+ bit_size, srcs, execution_mode);
nir_load_const_instr *new_instr =
nir_load_const_instr_create(mem_ctx,
@@ -144,14 +144,14 @@ constant_fold_intrinsic_instr(nir_intrinsic_instr *instr)
}
static bool
-constant_fold_block(nir_block *block, void *mem_ctx)
+constant_fold_block(nir_block *block, void *mem_ctx, unsigned execution_mode)
{
bool progress = false;
nir_foreach_instr_safe(instr, block) {
switch (instr->type) {
case nir_instr_type_alu:
- progress |= constant_fold_alu_instr(nir_instr_as_alu(instr), mem_ctx);
+ progress |= constant_fold_alu_instr(nir_instr_as_alu(instr), mem_ctx, execution_mode);
break;
case nir_instr_type_intrinsic:
progress |=
@@ -167,13 +167,13 @@ constant_fold_block(nir_block *block, void *mem_ctx)
}
static bool
-nir_opt_constant_folding_impl(nir_function_impl *impl)
+nir_opt_constant_folding_impl(nir_function_impl *impl, unsigned execution_mode)
{
void *mem_ctx = ralloc_parent(impl);
bool progress = false;
nir_foreach_block(block, impl) {
- progress |= constant_fold_block(block, mem_ctx);
+ progress |= constant_fold_block(block, mem_ctx, execution_mode);
}
if (progress) {
@@ -192,10 +192,11 @@ bool
nir_opt_constant_folding(nir_shader *shader)
{
bool progress = false;
+ unsigned execution_mode = shader->info.float_controls_execution_mode;
nir_foreach_function(function, shader) {
if (function->impl)
- progress |= nir_opt_constant_folding_impl(function->impl);
+ progress |= nir_opt_constant_folding_impl(function->impl, execution_mode);
}
return progress;
diff --git a/src/compiler/spirv/spirv_to_nir.c b/src/compiler/spirv/spirv_to_nir.c
index f46b0a9bb02..54052062392 100644
--- a/src/compiler/spirv/spirv_to_nir.c
+++ b/src/compiler/spirv/spirv_to_nir.c
@@ -1896,7 +1896,9 @@ vtn_handle_constant(struct vtn_builder *b, SpvOp opcode,
nir_const_value *srcs[3] = {
src[0], src[1], src[2],
};
- nir_eval_const_opcode(op, val->constant->values, num_components, bit_size, srcs);
+ nir_eval_const_opcode(op, val->constant->values,
+ num_components, bit_size, srcs,
+ b->shader->info.float_controls_execution_mode);
break;
} /* default */
}