nir/nir_lower_io: Add vec4 support
The current implementation operates in scalar mode only, so add a vec4 mode where types are padded to vec4 sizes. This will be useful in the i965 driver for its vec4 nir backend (and possbly other drivers that have vec4-based shaders). Reviewed-by: Jason Ekstrand <jason.ekstrand@intel.com>
This commit is contained in:

committed by
Jason Ekstrand

parent
5e839727ed
commit
01f6235020
@@ -29,19 +29,56 @@
|
||||
/*
|
||||
* This lowering pass converts references to input/output variables with
|
||||
* loads/stores to actual input/output intrinsics.
|
||||
*
|
||||
* NOTE: This pass really only works for scalar backends at the moment due
|
||||
* to the way it packes the input/output data.
|
||||
*/
|
||||
|
||||
#include "nir.h"
|
||||
|
||||
struct lower_io_state {
|
||||
void *mem_ctx;
|
||||
bool is_scalar;
|
||||
};
|
||||
|
||||
static int
|
||||
type_size_vec4(const struct glsl_type *type)
|
||||
{
|
||||
unsigned int i;
|
||||
int size;
|
||||
|
||||
switch (glsl_get_base_type(type)) {
|
||||
case GLSL_TYPE_UINT:
|
||||
case GLSL_TYPE_INT:
|
||||
case GLSL_TYPE_FLOAT:
|
||||
case GLSL_TYPE_BOOL:
|
||||
if (glsl_type_is_matrix(type)) {
|
||||
return glsl_get_matrix_columns(type);
|
||||
} else {
|
||||
return 1;
|
||||
}
|
||||
case GLSL_TYPE_ARRAY:
|
||||
return type_size_vec4(glsl_get_array_element(type)) * glsl_get_length(type);
|
||||
case GLSL_TYPE_STRUCT:
|
||||
size = 0;
|
||||
for (i = 0; i < glsl_get_length(type); i++) {
|
||||
size += type_size_vec4(glsl_get_struct_field(type, i));
|
||||
}
|
||||
return size;
|
||||
case GLSL_TYPE_SAMPLER:
|
||||
return 0;
|
||||
case GLSL_TYPE_ATOMIC_UINT:
|
||||
return 0;
|
||||
case GLSL_TYPE_IMAGE:
|
||||
case GLSL_TYPE_VOID:
|
||||
case GLSL_TYPE_DOUBLE:
|
||||
case GLSL_TYPE_ERROR:
|
||||
case GLSL_TYPE_INTERFACE:
|
||||
unreachable("not reached");
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static unsigned
|
||||
type_size(const struct glsl_type *type)
|
||||
type_size_scalar(const struct glsl_type *type)
|
||||
{
|
||||
unsigned int size, i;
|
||||
|
||||
@@ -52,11 +89,11 @@ type_size(const struct glsl_type *type)
|
||||
case GLSL_TYPE_BOOL:
|
||||
return glsl_get_components(type);
|
||||
case GLSL_TYPE_ARRAY:
|
||||
return type_size(glsl_get_array_element(type)) * glsl_get_length(type);
|
||||
return type_size_scalar(glsl_get_array_element(type)) * glsl_get_length(type);
|
||||
case GLSL_TYPE_STRUCT:
|
||||
size = 0;
|
||||
for (i = 0; i < glsl_get_length(type); i++) {
|
||||
size += type_size(glsl_get_struct_field(type, i));
|
||||
size += type_size_scalar(glsl_get_struct_field(type, i));
|
||||
}
|
||||
return size;
|
||||
case GLSL_TYPE_SUBROUTINE:
|
||||
@@ -78,8 +115,17 @@ type_size(const struct glsl_type *type)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static unsigned
|
||||
type_size(const struct glsl_type *type, bool is_scalar)
|
||||
{
|
||||
if (is_scalar)
|
||||
return type_size_scalar(type);
|
||||
else
|
||||
return type_size_vec4(type);
|
||||
}
|
||||
|
||||
void
|
||||
nir_assign_var_locations_scalar(struct exec_list *var_list, unsigned *size)
|
||||
nir_assign_var_locations(struct exec_list *var_list, unsigned *size, bool is_scalar)
|
||||
{
|
||||
unsigned location = 0;
|
||||
|
||||
@@ -93,7 +139,7 @@ nir_assign_var_locations_scalar(struct exec_list *var_list, unsigned *size)
|
||||
continue;
|
||||
|
||||
var->data.driver_location = location;
|
||||
location += type_size(var->type);
|
||||
location += type_size(var->type, is_scalar);
|
||||
}
|
||||
|
||||
*size = location;
|
||||
@@ -139,10 +185,11 @@ mark_indirect_uses_block(nir_block *block, void *void_state)
|
||||
* assigns locations to variables that are used indirectly.
|
||||
*/
|
||||
void
|
||||
nir_assign_var_locations_scalar_direct_first(nir_shader *shader,
|
||||
struct exec_list *var_list,
|
||||
unsigned *direct_size,
|
||||
unsigned *size)
|
||||
nir_assign_var_locations_direct_first(nir_shader *shader,
|
||||
struct exec_list *var_list,
|
||||
unsigned *direct_size,
|
||||
unsigned *size,
|
||||
bool is_scalar)
|
||||
{
|
||||
struct set *indirect_set = _mesa_set_create(NULL, _mesa_hash_pointer,
|
||||
_mesa_key_pointer_equal);
|
||||
@@ -164,7 +211,7 @@ nir_assign_var_locations_scalar_direct_first(nir_shader *shader,
|
||||
continue;
|
||||
|
||||
var->data.driver_location = location;
|
||||
location += type_size(var->type);
|
||||
location += type_size(var->type, is_scalar);
|
||||
}
|
||||
|
||||
*direct_size = location;
|
||||
@@ -178,7 +225,7 @@ nir_assign_var_locations_scalar_direct_first(nir_shader *shader,
|
||||
continue;
|
||||
|
||||
var->data.driver_location = location;
|
||||
location += type_size(var->type);
|
||||
location += type_size(var->type, is_scalar);
|
||||
}
|
||||
|
||||
*size = location;
|
||||
@@ -200,7 +247,7 @@ get_io_offset(nir_deref_var *deref, nir_instr *instr, nir_src *indirect,
|
||||
|
||||
if (tail->deref_type == nir_deref_type_array) {
|
||||
nir_deref_array *deref_array = nir_deref_as_array(tail);
|
||||
unsigned size = type_size(tail->type);
|
||||
unsigned size = type_size(tail->type, state->is_scalar);
|
||||
|
||||
base_offset += size * deref_array->base_offset;
|
||||
|
||||
@@ -242,7 +289,8 @@ get_io_offset(nir_deref_var *deref, nir_instr *instr, nir_src *indirect,
|
||||
nir_deref_struct *deref_struct = nir_deref_as_struct(tail);
|
||||
|
||||
for (unsigned i = 0; i < deref_struct->index; i++)
|
||||
base_offset += type_size(glsl_get_struct_field(parent_type, i));
|
||||
base_offset += type_size(glsl_get_struct_field(parent_type, i),
|
||||
state->is_scalar);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -355,11 +403,12 @@ nir_lower_io_block(nir_block *block, void *void_state)
|
||||
}
|
||||
|
||||
static void
|
||||
nir_lower_io_impl(nir_function_impl *impl)
|
||||
nir_lower_io_impl(nir_function_impl *impl, bool is_scalar)
|
||||
{
|
||||
struct lower_io_state state;
|
||||
|
||||
state.mem_ctx = ralloc_parent(impl);
|
||||
state.is_scalar = is_scalar;
|
||||
|
||||
nir_foreach_block(impl, nir_lower_io_block, &state);
|
||||
|
||||
@@ -368,10 +417,10 @@ nir_lower_io_impl(nir_function_impl *impl)
|
||||
}
|
||||
|
||||
void
|
||||
nir_lower_io(nir_shader *shader)
|
||||
nir_lower_io(nir_shader *shader, bool is_scalar)
|
||||
{
|
||||
nir_foreach_overload(shader, overload) {
|
||||
if (overload->impl)
|
||||
nir_lower_io_impl(overload->impl);
|
||||
nir_lower_io_impl(overload->impl, is_scalar);
|
||||
}
|
||||
}
|
||||
|
Reference in New Issue
Block a user