nir: Report progress from nir_lower_phis_to_scalar.
Signed-off-by: Kenneth Graunke <kenneth@whitecape.org> Reviewed-by: Eric Anholt <eric@anholt.net>
This commit is contained in:
@@ -2413,7 +2413,7 @@ bool nir_lower_vec_to_movs(nir_shader *shader);
|
|||||||
bool nir_lower_alu_to_scalar(nir_shader *shader);
|
bool nir_lower_alu_to_scalar(nir_shader *shader);
|
||||||
void nir_lower_load_const_to_scalar(nir_shader *shader);
|
void nir_lower_load_const_to_scalar(nir_shader *shader);
|
||||||
|
|
||||||
void nir_lower_phis_to_scalar(nir_shader *shader);
|
bool nir_lower_phis_to_scalar(nir_shader *shader);
|
||||||
void nir_lower_io_to_scalar(nir_shader *shader, nir_variable_mode mask);
|
void nir_lower_io_to_scalar(nir_shader *shader, nir_variable_mode mask);
|
||||||
|
|
||||||
void nir_lower_samplers(nir_shader *shader,
|
void nir_lower_samplers(nir_shader *shader,
|
||||||
|
@@ -166,6 +166,8 @@ static bool
|
|||||||
lower_phis_to_scalar_block(nir_block *block,
|
lower_phis_to_scalar_block(nir_block *block,
|
||||||
struct lower_phis_to_scalar_state *state)
|
struct lower_phis_to_scalar_state *state)
|
||||||
{
|
{
|
||||||
|
bool progress = false;
|
||||||
|
|
||||||
/* Find the last phi node in the block */
|
/* Find the last phi node in the block */
|
||||||
nir_phi_instr *last_phi = NULL;
|
nir_phi_instr *last_phi = NULL;
|
||||||
nir_foreach_instr(instr, block) {
|
nir_foreach_instr(instr, block) {
|
||||||
@@ -248,6 +250,8 @@ lower_phis_to_scalar_block(nir_block *block,
|
|||||||
ralloc_steal(state->dead_ctx, phi);
|
ralloc_steal(state->dead_ctx, phi);
|
||||||
nir_instr_remove(&phi->instr);
|
nir_instr_remove(&phi->instr);
|
||||||
|
|
||||||
|
progress = true;
|
||||||
|
|
||||||
/* We're using the safe iterator and inserting all the newly
|
/* We're using the safe iterator and inserting all the newly
|
||||||
* scalarized phi nodes before their non-scalarized version so that's
|
* scalarized phi nodes before their non-scalarized version so that's
|
||||||
* ok. However, we are also inserting vec operations after all of
|
* ok. However, we are also inserting vec operations after all of
|
||||||
@@ -258,13 +262,14 @@ lower_phis_to_scalar_block(nir_block *block,
|
|||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
|
||||||
return true;
|
return progress;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void
|
static bool
|
||||||
lower_phis_to_scalar_impl(nir_function_impl *impl)
|
lower_phis_to_scalar_impl(nir_function_impl *impl)
|
||||||
{
|
{
|
||||||
struct lower_phis_to_scalar_state state;
|
struct lower_phis_to_scalar_state state;
|
||||||
|
bool progress = false;
|
||||||
|
|
||||||
state.mem_ctx = ralloc_parent(impl);
|
state.mem_ctx = ralloc_parent(impl);
|
||||||
state.dead_ctx = ralloc_context(NULL);
|
state.dead_ctx = ralloc_context(NULL);
|
||||||
@@ -272,13 +277,14 @@ lower_phis_to_scalar_impl(nir_function_impl *impl)
|
|||||||
_mesa_key_pointer_equal);
|
_mesa_key_pointer_equal);
|
||||||
|
|
||||||
nir_foreach_block(block, impl) {
|
nir_foreach_block(block, impl) {
|
||||||
lower_phis_to_scalar_block(block, &state);
|
progress = lower_phis_to_scalar_block(block, &state) || progress;
|
||||||
}
|
}
|
||||||
|
|
||||||
nir_metadata_preserve(impl, nir_metadata_block_index |
|
nir_metadata_preserve(impl, nir_metadata_block_index |
|
||||||
nir_metadata_dominance);
|
nir_metadata_dominance);
|
||||||
|
|
||||||
ralloc_free(state.dead_ctx);
|
ralloc_free(state.dead_ctx);
|
||||||
|
return progress;
|
||||||
}
|
}
|
||||||
|
|
||||||
/** A pass that lowers vector phi nodes to scalar
|
/** A pass that lowers vector phi nodes to scalar
|
||||||
@@ -288,11 +294,15 @@ lower_phis_to_scalar_impl(nir_function_impl *impl)
|
|||||||
* instance, if one of the sources is a non-scalarizable vector, then we
|
* instance, if one of the sources is a non-scalarizable vector, then we
|
||||||
* don't bother lowering because that would generate hard-to-coalesce movs.
|
* don't bother lowering because that would generate hard-to-coalesce movs.
|
||||||
*/
|
*/
|
||||||
void
|
bool
|
||||||
nir_lower_phis_to_scalar(nir_shader *shader)
|
nir_lower_phis_to_scalar(nir_shader *shader)
|
||||||
{
|
{
|
||||||
|
bool progress = false;
|
||||||
|
|
||||||
nir_foreach_function(function, shader) {
|
nir_foreach_function(function, shader) {
|
||||||
if (function->impl)
|
if (function->impl)
|
||||||
lower_phis_to_scalar_impl(function->impl);
|
progress = lower_phis_to_scalar_impl(function->impl) || progress;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
return progress;
|
||||||
}
|
}
|
||||||
|
@@ -91,7 +91,7 @@ ir3_optimize_loop(nir_shader *s)
|
|||||||
|
|
||||||
OPT_V(s, nir_lower_vars_to_ssa);
|
OPT_V(s, nir_lower_vars_to_ssa);
|
||||||
progress |= OPT(s, nir_lower_alu_to_scalar);
|
progress |= OPT(s, nir_lower_alu_to_scalar);
|
||||||
OPT_V(s, nir_lower_phis_to_scalar);
|
progress |= OPT(s, nir_lower_phis_to_scalar);
|
||||||
|
|
||||||
progress |= OPT(s, nir_copy_prop);
|
progress |= OPT(s, nir_copy_prop);
|
||||||
progress |= OPT(s, nir_opt_dce);
|
progress |= OPT(s, nir_opt_dce);
|
||||||
|
@@ -1424,8 +1424,7 @@ vc4_optimize_nir(struct nir_shader *s)
|
|||||||
|
|
||||||
NIR_PASS_V(s, nir_lower_vars_to_ssa);
|
NIR_PASS_V(s, nir_lower_vars_to_ssa);
|
||||||
NIR_PASS(progress, s, nir_lower_alu_to_scalar);
|
NIR_PASS(progress, s, nir_lower_alu_to_scalar);
|
||||||
NIR_PASS_V(s, nir_lower_phis_to_scalar);
|
NIR_PASS(progress, s, nir_lower_phis_to_scalar);
|
||||||
|
|
||||||
NIR_PASS(progress, s, nir_copy_prop);
|
NIR_PASS(progress, s, nir_copy_prop);
|
||||||
NIR_PASS(progress, s, nir_opt_remove_phis);
|
NIR_PASS(progress, s, nir_opt_remove_phis);
|
||||||
NIR_PASS(progress, s, nir_opt_dce);
|
NIR_PASS(progress, s, nir_opt_dce);
|
||||||
|
@@ -381,7 +381,7 @@ nir_optimize(nir_shader *nir, bool is_scalar)
|
|||||||
OPT(nir_copy_prop);
|
OPT(nir_copy_prop);
|
||||||
|
|
||||||
if (is_scalar) {
|
if (is_scalar) {
|
||||||
OPT_V(nir_lower_phis_to_scalar);
|
OPT(nir_lower_phis_to_scalar);
|
||||||
}
|
}
|
||||||
|
|
||||||
OPT(nir_copy_prop);
|
OPT(nir_copy_prop);
|
||||||
|
Reference in New Issue
Block a user