turnip: improve emit_textures

Two things:
* Texture/sampler pointers aligned to the size of texture/sampler state
* Returning errors instead of crashing on OOM

Signed-off-by: Jonathan Marek <jonathan@marek.ca>
Reviewed-by: Kristian H. Kristensen <hoegsberg@google.com>
This commit is contained in:
Jonathan Marek
2019-11-15 15:12:25 -05:00
parent 3ab4f99461
commit 095d35eff8

View File

@@ -2410,19 +2410,48 @@ tu6_emit_consts(struct tu_cmd_buffer *cmd,
return tu_cs_end_sub_stream(&cmd->draw_state, &cs); return tu_cs_end_sub_stream(&cmd->draw_state, &cs);
} }
static struct tu_cs_entry static VkResult
tu6_emit_textures(struct tu_device *device, struct tu_cs *draw_state, tu6_emit_textures(struct tu_cmd_buffer *cmd,
const struct tu_pipeline *pipeline, gl_shader_stage type,
struct tu_descriptor_state *descriptors_state, struct tu_cs_entry *entry,
gl_shader_stage type, bool *needs_border) bool *needs_border)
{ {
struct tu_device *device = cmd->device;
struct tu_cs *draw_state = &cmd->draw_state;
struct tu_descriptor_state *descriptors_state =
&cmd->descriptors[VK_PIPELINE_BIND_POINT_GRAPHICS];
const struct tu_program_descriptor_linkage *link = const struct tu_program_descriptor_linkage *link =
&pipeline->program.link[type]; &cmd->state.pipeline->program.link[type];
VkResult result;
uint32_t size = link->texture_map.num * A6XX_TEX_CONST_DWORDS + if (link->texture_map.num == 0 && link->sampler_map.num == 0) {
link->sampler_map.num * A6XX_TEX_SAMP_DWORDS; *entry = (struct tu_cs_entry) {};
if (!size) return VK_SUCCESS;
return (struct tu_cs_entry) {}; }
/* allocate and fill texture state */
struct ts_cs_memory tex_const;
result = tu_cs_alloc(device, draw_state, link->texture_map.num, A6XX_TEX_CONST_DWORDS, &tex_const);
if (result != VK_SUCCESS)
return result;
for (unsigned i = 0; i < link->texture_map.num; i++) {
memcpy(&tex_const.map[A6XX_TEX_CONST_DWORDS*i],
texture_ptr(descriptors_state, &link->texture_map, i),
A6XX_TEX_CONST_DWORDS*4);
}
/* allocate and fill sampler state */
struct ts_cs_memory tex_samp;
result = tu_cs_alloc(device, draw_state, link->sampler_map.num, A6XX_TEX_SAMP_DWORDS, &tex_samp);
if (result != VK_SUCCESS)
return result;
for (unsigned i = 0; i < link->sampler_map.num; i++) {
struct tu_sampler *sampler = sampler_ptr(descriptors_state, &link->sampler_map, i);
memcpy(&tex_samp.map[A6XX_TEX_SAMP_DWORDS*i], sampler->state, sizeof(sampler->state));
*needs_border |= sampler->needs_border;
}
unsigned tex_samp_reg, tex_const_reg, tex_count_reg; unsigned tex_samp_reg, tex_const_reg, tex_count_reg;
enum a6xx_state_block sb; enum a6xx_state_block sb;
@@ -2451,30 +2480,9 @@ tu6_emit_textures(struct tu_device *device, struct tu_cs *draw_state,
} }
struct tu_cs cs; struct tu_cs cs;
tu_cs_begin_sub_stream(device, draw_state, size, &cs); result = tu_cs_begin_sub_stream(device, draw_state, 16, &cs);
if (result != VK_SUCCESS)
for (unsigned i = 0; i < link->texture_map.num; i++) { return result;
uint32_t *ptr = texture_ptr(descriptors_state, &link->texture_map, i);
for (unsigned j = 0; j < A6XX_TEX_CONST_DWORDS; j++)
tu_cs_emit(&cs, ptr[j]);
}
for (unsigned i = 0; i < link->sampler_map.num; i++) {
struct tu_sampler *sampler = sampler_ptr(descriptors_state, &link->sampler_map, i);
for (unsigned j = 0; j < A6XX_TEX_SAMP_DWORDS; j++)
tu_cs_emit(&cs, sampler->state[j]);
*needs_border |= sampler->needs_border;
}
struct tu_cs_entry entry = tu_cs_end_sub_stream(draw_state, &cs);
uint64_t tex_addr = entry.bo->iova + entry.offset;
uint64_t samp_addr = tex_addr + link->texture_map.num * A6XX_TEX_CONST_DWORDS*4;
tu_cs_begin_sub_stream(device, draw_state, 64, &cs);
/* output sampler state: */ /* output sampler state: */
tu_cs_emit_pkt7(&cs, tu6_stage2opcode(type), 3); tu_cs_emit_pkt7(&cs, tu6_stage2opcode(type), 3);
@@ -2483,10 +2491,10 @@ tu6_emit_textures(struct tu_device *device, struct tu_cs *draw_state,
CP_LOAD_STATE6_0_STATE_SRC(SS6_INDIRECT) | CP_LOAD_STATE6_0_STATE_SRC(SS6_INDIRECT) |
CP_LOAD_STATE6_0_STATE_BLOCK(sb) | CP_LOAD_STATE6_0_STATE_BLOCK(sb) |
CP_LOAD_STATE6_0_NUM_UNIT(link->sampler_map.num)); CP_LOAD_STATE6_0_NUM_UNIT(link->sampler_map.num));
tu_cs_emit_qw(&cs, samp_addr); /* SRC_ADDR_LO/HI */ tu_cs_emit_qw(&cs, tex_samp.iova); /* SRC_ADDR_LO/HI */
tu_cs_emit_pkt4(&cs, tex_samp_reg, 2); tu_cs_emit_pkt4(&cs, tex_samp_reg, 2);
tu_cs_emit_qw(&cs, samp_addr); /* SRC_ADDR_LO/HI */ tu_cs_emit_qw(&cs, tex_samp.iova); /* SRC_ADDR_LO/HI */
/* emit texture state: */ /* emit texture state: */
tu_cs_emit_pkt7(&cs, tu6_stage2opcode(type), 3); tu_cs_emit_pkt7(&cs, tu6_stage2opcode(type), 3);
@@ -2495,15 +2503,16 @@ tu6_emit_textures(struct tu_device *device, struct tu_cs *draw_state,
CP_LOAD_STATE6_0_STATE_SRC(SS6_INDIRECT) | CP_LOAD_STATE6_0_STATE_SRC(SS6_INDIRECT) |
CP_LOAD_STATE6_0_STATE_BLOCK(sb) | CP_LOAD_STATE6_0_STATE_BLOCK(sb) |
CP_LOAD_STATE6_0_NUM_UNIT(link->texture_map.num)); CP_LOAD_STATE6_0_NUM_UNIT(link->texture_map.num));
tu_cs_emit_qw(&cs, tex_addr); /* SRC_ADDR_LO/HI */ tu_cs_emit_qw(&cs, tex_const.iova); /* SRC_ADDR_LO/HI */
tu_cs_emit_pkt4(&cs, tex_const_reg, 2); tu_cs_emit_pkt4(&cs, tex_const_reg, 2);
tu_cs_emit_qw(&cs, tex_addr); /* SRC_ADDR_LO/HI */ tu_cs_emit_qw(&cs, tex_const.iova); /* SRC_ADDR_LO/HI */
tu_cs_emit_pkt4(&cs, tex_count_reg, 1); tu_cs_emit_pkt4(&cs, tex_count_reg, 1);
tu_cs_emit(&cs, link->texture_map.num); tu_cs_emit(&cs, link->texture_map.num);
return tu_cs_end_sub_stream(draw_state, &cs); *entry = tu_cs_end_sub_stream(draw_state, &cs);
return VK_SUCCESS;
} }
static struct tu_cs_entry static struct tu_cs_entry
@@ -2599,7 +2608,7 @@ tu6_emit_border_color(struct tu_cmd_buffer *cmd,
tu_cs_emit_qw(cs, align(entry.bo->iova + entry.offset, 128)); tu_cs_emit_qw(cs, align(entry.bo->iova + entry.offset, 128));
} }
static void static VkResult
tu6_bind_draw_states(struct tu_cmd_buffer *cmd, tu6_bind_draw_states(struct tu_cmd_buffer *cmd,
struct tu_cs *cs, struct tu_cs *cs,
const struct tu_draw_info *draw) const struct tu_draw_info *draw)
@@ -2613,10 +2622,8 @@ tu6_bind_draw_states(struct tu_cmd_buffer *cmd,
&cmd->descriptors[VK_PIPELINE_BIND_POINT_GRAPHICS]; &cmd->descriptors[VK_PIPELINE_BIND_POINT_GRAPHICS];
VkResult result = tu_cs_reserve_space(cmd->device, cs, 256); VkResult result = tu_cs_reserve_space(cmd->device, cs, 256);
if (result != VK_SUCCESS) { if (result != VK_SUCCESS)
cmd->record_result = result; return result;
return;
}
/* TODO lrz */ /* TODO lrz */
@@ -2729,6 +2736,15 @@ tu6_bind_draw_states(struct tu_cmd_buffer *cmd,
if (cmd->state.dirty & if (cmd->state.dirty &
(TU_CMD_DIRTY_PIPELINE | TU_CMD_DIRTY_DESCRIPTOR_SETS)) { (TU_CMD_DIRTY_PIPELINE | TU_CMD_DIRTY_DESCRIPTOR_SETS)) {
bool needs_border = false; bool needs_border = false;
struct tu_cs_entry vs_tex, fs_tex;
result = tu6_emit_textures(cmd, MESA_SHADER_VERTEX, &vs_tex, &needs_border);
if (result != VK_SUCCESS)
return result;
result = tu6_emit_textures(cmd, MESA_SHADER_FRAGMENT, &fs_tex, &needs_border);
if (result != VK_SUCCESS)
return result;
draw_state_groups[draw_state_group_count++] = draw_state_groups[draw_state_group_count++] =
(struct tu_draw_state_group) { (struct tu_draw_state_group) {
@@ -2746,17 +2762,13 @@ tu6_bind_draw_states(struct tu_cmd_buffer *cmd,
(struct tu_draw_state_group) { (struct tu_draw_state_group) {
.id = TU_DRAW_STATE_VS_TEX, .id = TU_DRAW_STATE_VS_TEX,
.enable_mask = 0x7, .enable_mask = 0x7,
.ib = tu6_emit_textures(cmd->device, &cmd->draw_state, pipeline, .ib = vs_tex,
descriptors_state, MESA_SHADER_VERTEX,
&needs_border)
}; };
draw_state_groups[draw_state_group_count++] = draw_state_groups[draw_state_group_count++] =
(struct tu_draw_state_group) { (struct tu_draw_state_group) {
.id = TU_DRAW_STATE_FS_TEX, .id = TU_DRAW_STATE_FS_TEX,
.enable_mask = 0x6, .enable_mask = 0x6,
.ib = tu6_emit_textures(cmd->device, &cmd->draw_state, pipeline, .ib = fs_tex,
descriptors_state, MESA_SHADER_FRAGMENT,
&needs_border)
}; };
draw_state_groups[draw_state_group_count++] = draw_state_groups[draw_state_group_count++] =
(struct tu_draw_state_group) { (struct tu_draw_state_group) {
@@ -2816,6 +2828,7 @@ tu6_bind_draw_states(struct tu_cmd_buffer *cmd,
* compute pipeline for re-emit. * compute pipeline for re-emit.
*/ */
cmd->state.dirty = TU_CMD_DIRTY_COMPUTE_PIPELINE; cmd->state.dirty = TU_CMD_DIRTY_COMPUTE_PIPELINE;
return VK_SUCCESS;
} }
static void static void
@@ -2871,10 +2884,15 @@ static void
tu_draw(struct tu_cmd_buffer *cmd, const struct tu_draw_info *draw) tu_draw(struct tu_cmd_buffer *cmd, const struct tu_draw_info *draw)
{ {
struct tu_cs *cs = &cmd->draw_cs; struct tu_cs *cs = &cmd->draw_cs;
VkResult result;
tu6_bind_draw_states(cmd, cs, draw); result = tu6_bind_draw_states(cmd, cs, draw);
if (result != VK_SUCCESS) {
cmd->record_result = result;
return;
}
VkResult result = tu_cs_reserve_space(cmd->device, cs, 32); result = tu_cs_reserve_space(cmd->device, cs, 32);
if (result != VK_SUCCESS) { if (result != VK_SUCCESS) {
cmd->record_result = result; cmd->record_result = result;
return; return;
@@ -3069,9 +3087,12 @@ tu_dispatch(struct tu_cmd_buffer *cmd,
tu_emit_compute_driver_params(cs, pipeline, info); tu_emit_compute_driver_params(cs, pipeline, info);
bool needs_border; bool needs_border;
ib = tu6_emit_textures(cmd->device, &cmd->draw_state, pipeline, result = tu6_emit_textures(cmd, MESA_SHADER_COMPUTE, &ib, &needs_border);
descriptors_state, MESA_SHADER_COMPUTE, if (result != VK_SUCCESS) {
&needs_border); cmd->record_result = result;
return;
}
if (ib.size) if (ib.size)
tu_cs_emit_ib(cs, &ib); tu_cs_emit_ib(cs, &ib);