nir/load_store_vectorize: add data as callback args
Signed-off-by: Rhys Perry <pendingchaos02@gmail.com> Reviewed-by: Jason Ekstrand <jason@jlekstrand.net> Part-of: <https://gitlab.freedesktop.org/mesa/mesa/-/merge_requests/4202>
This commit is contained in:
@@ -3032,7 +3032,8 @@ static bool
|
|||||||
mem_vectorize_callback(unsigned align_mul, unsigned align_offset,
|
mem_vectorize_callback(unsigned align_mul, unsigned align_offset,
|
||||||
unsigned bit_size,
|
unsigned bit_size,
|
||||||
unsigned num_components,
|
unsigned num_components,
|
||||||
nir_intrinsic_instr *low, nir_intrinsic_instr *high)
|
nir_intrinsic_instr *low, nir_intrinsic_instr *high,
|
||||||
|
void *data)
|
||||||
{
|
{
|
||||||
if (num_components > 4)
|
if (num_components > 4)
|
||||||
return false;
|
return false;
|
||||||
|
@@ -5010,12 +5010,14 @@ typedef bool (*nir_should_vectorize_mem_func)(unsigned align_mul,
|
|||||||
unsigned align_offset,
|
unsigned align_offset,
|
||||||
unsigned bit_size,
|
unsigned bit_size,
|
||||||
unsigned num_components,
|
unsigned num_components,
|
||||||
nir_intrinsic_instr *low, nir_intrinsic_instr *high);
|
nir_intrinsic_instr *low, nir_intrinsic_instr *high,
|
||||||
|
void *data);
|
||||||
|
|
||||||
typedef struct {
|
typedef struct {
|
||||||
nir_should_vectorize_mem_func callback;
|
nir_should_vectorize_mem_func callback;
|
||||||
nir_variable_mode modes;
|
nir_variable_mode modes;
|
||||||
nir_variable_mode robust_modes;
|
nir_variable_mode robust_modes;
|
||||||
|
void *cb_data;
|
||||||
} nir_load_store_vectorize_options;
|
} nir_load_store_vectorize_options;
|
||||||
|
|
||||||
bool nir_opt_load_store_vectorize(nir_shader *shader, const nir_load_store_vectorize_options *options);
|
bool nir_opt_load_store_vectorize(nir_shader *shader, const nir_load_store_vectorize_options *options);
|
||||||
|
@@ -649,7 +649,8 @@ new_bitsize_acceptable(struct vectorize_ctx *ctx, unsigned new_bit_size,
|
|||||||
if (!ctx->options->callback(low->align_mul,
|
if (!ctx->options->callback(low->align_mul,
|
||||||
low->align_offset,
|
low->align_offset,
|
||||||
new_bit_size, new_num_components,
|
new_bit_size, new_num_components,
|
||||||
low->intrin, high->intrin))
|
low->intrin, high->intrin,
|
||||||
|
ctx->options->cb_data))
|
||||||
return false;
|
return false;
|
||||||
|
|
||||||
if (low->is_store) {
|
if (low->is_store) {
|
||||||
|
@@ -73,7 +73,8 @@ protected:
|
|||||||
static bool mem_vectorize_callback(unsigned align_mul, unsigned align_offset,
|
static bool mem_vectorize_callback(unsigned align_mul, unsigned align_offset,
|
||||||
unsigned bit_size,
|
unsigned bit_size,
|
||||||
unsigned num_components,
|
unsigned num_components,
|
||||||
nir_intrinsic_instr *low, nir_intrinsic_instr *high);
|
nir_intrinsic_instr *low, nir_intrinsic_instr *high,
|
||||||
|
void *data);
|
||||||
static void shared_type_info(const struct glsl_type *type, unsigned *size, unsigned *align);
|
static void shared_type_info(const struct glsl_type *type, unsigned *size, unsigned *align);
|
||||||
|
|
||||||
std::string swizzle(nir_alu_instr *instr, int src);
|
std::string swizzle(nir_alu_instr *instr, int src);
|
||||||
@@ -359,7 +360,8 @@ bool nir_load_store_vectorize_test::test_alu_def(
|
|||||||
bool nir_load_store_vectorize_test::mem_vectorize_callback(
|
bool nir_load_store_vectorize_test::mem_vectorize_callback(
|
||||||
unsigned align_mul, unsigned align_offset, unsigned bit_size,
|
unsigned align_mul, unsigned align_offset, unsigned bit_size,
|
||||||
unsigned num_components,
|
unsigned num_components,
|
||||||
nir_intrinsic_instr *low, nir_intrinsic_instr *high)
|
nir_intrinsic_instr *low, nir_intrinsic_instr *high,
|
||||||
|
void *data)
|
||||||
{
|
{
|
||||||
/* Calculate a simple alignment, like how nir_intrinsic_align() does. */
|
/* Calculate a simple alignment, like how nir_intrinsic_align() does. */
|
||||||
uint32_t align = align_mul;
|
uint32_t align = align_mul;
|
||||||
|
@@ -150,7 +150,8 @@ ir3_nir_should_vectorize_mem(unsigned align_mul, unsigned align_offset,
|
|||||||
unsigned bit_size,
|
unsigned bit_size,
|
||||||
unsigned num_components,
|
unsigned num_components,
|
||||||
nir_intrinsic_instr *low,
|
nir_intrinsic_instr *low,
|
||||||
nir_intrinsic_instr *high)
|
nir_intrinsic_instr *high,
|
||||||
|
void *data)
|
||||||
{
|
{
|
||||||
assert(bit_size >= 8);
|
assert(bit_size >= 8);
|
||||||
if (bit_size != 32)
|
if (bit_size != 32)
|
||||||
|
@@ -2204,7 +2204,8 @@ ntt_should_vectorize_instr(const nir_instr *instr, void *data)
|
|||||||
static bool
|
static bool
|
||||||
ntt_should_vectorize_io(unsigned align, unsigned bit_size,
|
ntt_should_vectorize_io(unsigned align, unsigned bit_size,
|
||||||
unsigned num_components, unsigned high_offset,
|
unsigned num_components, unsigned high_offset,
|
||||||
nir_intrinsic_instr *low, nir_intrinsic_instr *high)
|
nir_intrinsic_instr *low, nir_intrinsic_instr *high,
|
||||||
|
void *data)
|
||||||
{
|
{
|
||||||
if (bit_size != 32)
|
if (bit_size != 32)
|
||||||
return false;
|
return false;
|
||||||
|
@@ -969,7 +969,8 @@ brw_nir_should_vectorize_mem(unsigned align_mul, unsigned align_offset,
|
|||||||
unsigned bit_size,
|
unsigned bit_size,
|
||||||
unsigned num_components,
|
unsigned num_components,
|
||||||
nir_intrinsic_instr *low,
|
nir_intrinsic_instr *low,
|
||||||
nir_intrinsic_instr *high)
|
nir_intrinsic_instr *high,
|
||||||
|
void *data)
|
||||||
{
|
{
|
||||||
/* Don't combine things to generate 64-bit loads/stores. We have to split
|
/* Don't combine things to generate 64-bit loads/stores. We have to split
|
||||||
* those back into 32-bit ones anyway and UBO loads aren't split in NIR so
|
* those back into 32-bit ones anyway and UBO loads aren't split in NIR so
|
||||||
|
Reference in New Issue
Block a user