v3d/compiler: add a lowering pass for robust buffer access
Reviewed-by: Alejandro Piñeiro <apinheiro@igalia.com> Part-of: <https://gitlab.freedesktop.org/mesa/mesa/-/merge_requests/6766>
This commit is contained in:

committed by
Marge Bot

parent
4401dde0e9
commit
23c727dd67
@@ -39,6 +39,7 @@ libbroadcom_compiler_files = files(
|
||||
'v3d_nir_lower_image_load_store.c',
|
||||
'v3d_nir_lower_line_smooth.c',
|
||||
'v3d_nir_lower_logic_ops.c',
|
||||
'v3d_nir_lower_robust_buffer_access.c',
|
||||
'v3d_nir_lower_scratch.c',
|
||||
'v3d_nir_lower_txf_ms.c',
|
||||
)
|
||||
|
@@ -353,6 +353,7 @@ struct v3d_key {
|
||||
uint8_t num_tex_used;
|
||||
uint8_t ucp_enables;
|
||||
bool is_last_geometry_stage;
|
||||
bool robust_buffer_access;
|
||||
|
||||
enum v3d_execution_environment environment;
|
||||
};
|
||||
@@ -911,6 +912,7 @@ void v3d_nir_lower_blend(nir_shader *s, struct v3d_compile *c);
|
||||
void v3d_nir_lower_io(nir_shader *s, struct v3d_compile *c);
|
||||
void v3d_nir_lower_line_smooth(nir_shader *shader);
|
||||
void v3d_nir_lower_logic_ops(nir_shader *s, struct v3d_compile *c);
|
||||
void v3d_nir_lower_robust_buffer_access(nir_shader *shader, struct v3d_compile *c);
|
||||
void v3d_nir_lower_scratch(nir_shader *s);
|
||||
void v3d_nir_lower_txf_ms(nir_shader *s, struct v3d_compile *c);
|
||||
void v3d_nir_lower_image_load_store(nir_shader *s);
|
||||
|
167
src/broadcom/compiler/v3d_nir_lower_robust_buffer_access.c
Normal file
167
src/broadcom/compiler/v3d_nir_lower_robust_buffer_access.c
Normal file
@@ -0,0 +1,167 @@
|
||||
/*
|
||||
* Copyright © 2020 Raspberry Pi
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
* to deal in the Software without restriction, including without limitation
|
||||
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
||||
* and/or sell copies of the Software, and to permit persons to whom the
|
||||
* Software is furnished to do so, subject to the following conditions:
|
||||
*
|
||||
* The above copyright notice and this permission notice (including the next
|
||||
* paragraph) shall be included in all copies or substantial portions of the
|
||||
* Software.
|
||||
*
|
||||
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
||||
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
||||
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
|
||||
* IN THE SOFTWARE.
|
||||
*/
|
||||
|
||||
#include "compiler/v3d_compiler.h"
|
||||
#include "compiler/nir/nir_builder.h"
|
||||
|
||||
static void
|
||||
rewrite_offset(nir_builder *b,
|
||||
nir_intrinsic_instr *instr,
|
||||
uint32_t buffer_idx,
|
||||
uint32_t offset_src,
|
||||
nir_intrinsic_op buffer_size_op)
|
||||
{
|
||||
b->cursor = nir_before_instr(&instr->instr);
|
||||
|
||||
/* Get size of the buffer */
|
||||
nir_intrinsic_instr *size =
|
||||
nir_intrinsic_instr_create(b->shader, buffer_size_op);
|
||||
size->src[0] = nir_src_for_ssa(nir_imm_int(b, buffer_idx));
|
||||
nir_ssa_dest_init(&size->instr, &size->dest, 1, 32, NULL);
|
||||
nir_builder_instr_insert(b, &size->instr);
|
||||
|
||||
/* All out TMU accesses are 32-bit aligned */
|
||||
nir_ssa_def *aligned_buffer_size =
|
||||
nir_iand(b, &size->dest.ssa, nir_imm_int(b, 0xfffffffc));
|
||||
|
||||
/* Rewrite offset */
|
||||
nir_ssa_def *offset =
|
||||
nir_umin(b, instr->src[offset_src].ssa, aligned_buffer_size);
|
||||
nir_instr_rewrite_src(&instr->instr, &instr->src[offset_src],
|
||||
nir_src_for_ssa(offset));
|
||||
}
|
||||
|
||||
static void
|
||||
lower_load(struct v3d_compile *c,
|
||||
nir_builder *b,
|
||||
nir_intrinsic_instr *instr)
|
||||
{
|
||||
uint32_t index = nir_src_as_uint(instr->src[0]);
|
||||
|
||||
nir_intrinsic_op op;
|
||||
if (instr->intrinsic == nir_intrinsic_load_ubo) {
|
||||
op = nir_intrinsic_get_ubo_size;
|
||||
if (c->key->environment == V3D_ENVIRONMENT_VULKAN)
|
||||
index--;
|
||||
} else {
|
||||
op = nir_intrinsic_get_ssbo_size;
|
||||
}
|
||||
|
||||
rewrite_offset(b, instr, index, 1, op);
|
||||
}
|
||||
|
||||
static void
|
||||
lower_store(struct v3d_compile *c,
|
||||
nir_builder *b,
|
||||
nir_intrinsic_instr *instr)
|
||||
{
|
||||
uint32_t index = nir_src_as_uint(instr->src[1]);
|
||||
rewrite_offset(b, instr, index, 2, nir_intrinsic_get_ssbo_size);
|
||||
}
|
||||
|
||||
static void
|
||||
lower_atomic(struct v3d_compile *c,
|
||||
nir_builder *b,
|
||||
nir_intrinsic_instr *instr)
|
||||
{
|
||||
uint32_t index = nir_src_as_uint(instr->src[0]);
|
||||
rewrite_offset(b, instr, index, 1, nir_intrinsic_get_ssbo_size);
|
||||
}
|
||||
|
||||
static void
|
||||
lower_shared(struct v3d_compile *c,
|
||||
nir_builder *b,
|
||||
nir_intrinsic_instr *instr)
|
||||
{
|
||||
b->cursor = nir_before_instr(&instr->instr);
|
||||
nir_ssa_def *aligned_size =
|
||||
nir_imm_int(b, c->s->info.cs.shared_size & 0xfffffffc);
|
||||
nir_ssa_def *offset = nir_umin(b, instr->src[0].ssa, aligned_size);
|
||||
nir_instr_rewrite_src(&instr->instr, &instr->src[0],
|
||||
nir_src_for_ssa(offset));
|
||||
}
|
||||
|
||||
static void
|
||||
lower_instr(struct v3d_compile *c, nir_builder *b, struct nir_instr *instr)
|
||||
{
|
||||
if (instr->type != nir_instr_type_intrinsic)
|
||||
return;
|
||||
nir_intrinsic_instr *intr = nir_instr_as_intrinsic(instr);
|
||||
|
||||
switch (intr->intrinsic) {
|
||||
case nir_intrinsic_load_ubo:
|
||||
case nir_intrinsic_load_ssbo:
|
||||
lower_load(c, b, intr);
|
||||
break;
|
||||
case nir_intrinsic_store_ssbo:
|
||||
lower_store(c, b, intr);
|
||||
break;
|
||||
case nir_intrinsic_ssbo_atomic_add:
|
||||
case nir_intrinsic_ssbo_atomic_imin:
|
||||
case nir_intrinsic_ssbo_atomic_umin:
|
||||
case nir_intrinsic_ssbo_atomic_imax:
|
||||
case nir_intrinsic_ssbo_atomic_umax:
|
||||
case nir_intrinsic_ssbo_atomic_and:
|
||||
case nir_intrinsic_ssbo_atomic_or:
|
||||
case nir_intrinsic_ssbo_atomic_xor:
|
||||
case nir_intrinsic_ssbo_atomic_exchange:
|
||||
case nir_intrinsic_ssbo_atomic_comp_swap:
|
||||
lower_atomic(c, b, intr);
|
||||
break;
|
||||
case nir_intrinsic_load_shared:
|
||||
case nir_intrinsic_shared_atomic_add:
|
||||
case nir_intrinsic_shared_atomic_imin:
|
||||
case nir_intrinsic_shared_atomic_umin:
|
||||
case nir_intrinsic_shared_atomic_imax:
|
||||
case nir_intrinsic_shared_atomic_umax:
|
||||
case nir_intrinsic_shared_atomic_and:
|
||||
case nir_intrinsic_shared_atomic_or:
|
||||
case nir_intrinsic_shared_atomic_xor:
|
||||
case nir_intrinsic_shared_atomic_exchange:
|
||||
case nir_intrinsic_shared_atomic_comp_swap:
|
||||
lower_shared(c, b, intr);
|
||||
break;
|
||||
default:
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
void
|
||||
v3d_nir_lower_robust_buffer_access(nir_shader *s, struct v3d_compile *c)
|
||||
{
|
||||
nir_foreach_function(function, s) {
|
||||
if (function->impl) {
|
||||
nir_builder b;
|
||||
nir_builder_init(&b, function->impl);
|
||||
|
||||
nir_foreach_block(block, function->impl) {
|
||||
nir_foreach_instr_safe(instr, block)
|
||||
lower_instr(c, &b, instr);
|
||||
}
|
||||
|
||||
nir_metadata_preserve(function->impl,
|
||||
nir_metadata_block_index |
|
||||
nir_metadata_dominance);
|
||||
}
|
||||
}
|
||||
}
|
@@ -1126,6 +1126,17 @@ v3d_attempt_compile(struct v3d_compile *c)
|
||||
NIR_PASS_V(c->s, v3d_nir_lower_image_load_store);
|
||||
NIR_PASS_V(c->s, nir_lower_idiv, nir_lower_idiv_fast);
|
||||
|
||||
if (c->key->robust_buffer_access) {
|
||||
/* v3d_nir_lower_robust_buffer_access assumes constant buffer
|
||||
* indices on ubo/ssbo intrinsics so run a copy propagation pass
|
||||
* before we run the lowering to warrant this. We also want to run
|
||||
* the lowering before v3d_optimize to clean-up redundant
|
||||
* get_buffer_size calls produced in the pass.
|
||||
*/
|
||||
NIR_PASS_V(c->s, nir_copy_prop);
|
||||
NIR_PASS_V(c->s, v3d_nir_lower_robust_buffer_access, c);
|
||||
}
|
||||
|
||||
v3d_optimize_nir(c->s);
|
||||
|
||||
/* Do late algebraic optimization to turn add(a, neg(b)) back into
|
||||
|
Reference in New Issue
Block a user