anv/query: Use anv_address everywhere
Instead of passing around BOs and offsets, use addresses which are anv's GPU equivalent of pointers. Reviewed-by: Lionel Landwerlin <lionel.g.landwerlin@intel.com>
This commit is contained in:
@@ -139,6 +139,15 @@ void genX(DestroyQueryPool)(
|
|||||||
vk_free2(&device->alloc, pAllocator, pool);
|
vk_free2(&device->alloc, pAllocator, pool);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static struct anv_address
|
||||||
|
anv_query_address(struct anv_query_pool *pool, uint32_t query)
|
||||||
|
{
|
||||||
|
return (struct anv_address) {
|
||||||
|
.bo = &pool->bo,
|
||||||
|
.offset = query * pool->stride,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
static void
|
static void
|
||||||
cpu_write_query_result(void *dst_slot, VkQueryResultFlags flags,
|
cpu_write_query_result(void *dst_slot, VkQueryResultFlags flags,
|
||||||
uint32_t value_index, uint64_t result)
|
uint32_t value_index, uint64_t result)
|
||||||
@@ -303,13 +312,13 @@ VkResult genX(GetQueryPoolResults)(
|
|||||||
|
|
||||||
static void
|
static void
|
||||||
emit_ps_depth_count(struct anv_cmd_buffer *cmd_buffer,
|
emit_ps_depth_count(struct anv_cmd_buffer *cmd_buffer,
|
||||||
struct anv_bo *bo, uint32_t offset)
|
struct anv_address addr)
|
||||||
{
|
{
|
||||||
anv_batch_emit(&cmd_buffer->batch, GENX(PIPE_CONTROL), pc) {
|
anv_batch_emit(&cmd_buffer->batch, GENX(PIPE_CONTROL), pc) {
|
||||||
pc.DestinationAddressType = DAT_PPGTT;
|
pc.DestinationAddressType = DAT_PPGTT;
|
||||||
pc.PostSyncOperation = WritePSDepthCount;
|
pc.PostSyncOperation = WritePSDepthCount;
|
||||||
pc.DepthStallEnable = true;
|
pc.DepthStallEnable = true;
|
||||||
pc.Address = (struct anv_address) { bo, offset };
|
pc.Address = addr;
|
||||||
|
|
||||||
if (GEN_GEN == 9 && cmd_buffer->device->info.gt == 4)
|
if (GEN_GEN == 9 && cmd_buffer->device->info.gt == 4)
|
||||||
pc.CommandStreamerStallEnable = true;
|
pc.CommandStreamerStallEnable = true;
|
||||||
@@ -318,12 +327,12 @@ emit_ps_depth_count(struct anv_cmd_buffer *cmd_buffer,
|
|||||||
|
|
||||||
static void
|
static void
|
||||||
emit_query_availability(struct anv_cmd_buffer *cmd_buffer,
|
emit_query_availability(struct anv_cmd_buffer *cmd_buffer,
|
||||||
struct anv_bo *bo, uint32_t offset)
|
struct anv_address addr)
|
||||||
{
|
{
|
||||||
anv_batch_emit(&cmd_buffer->batch, GENX(PIPE_CONTROL), pc) {
|
anv_batch_emit(&cmd_buffer->batch, GENX(PIPE_CONTROL), pc) {
|
||||||
pc.DestinationAddressType = DAT_PPGTT;
|
pc.DestinationAddressType = DAT_PPGTT;
|
||||||
pc.PostSyncOperation = WriteImmediateData;
|
pc.PostSyncOperation = WriteImmediateData;
|
||||||
pc.Address = (struct anv_address) { bo, offset };
|
pc.Address = addr;
|
||||||
pc.ImmediateData = 1;
|
pc.ImmediateData = 1;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -340,20 +349,19 @@ emit_zero_queries(struct anv_cmd_buffer *cmd_buffer,
|
|||||||
const uint32_t num_elements = pool->stride / sizeof(uint64_t);
|
const uint32_t num_elements = pool->stride / sizeof(uint64_t);
|
||||||
|
|
||||||
for (uint32_t i = 0; i < num_queries; i++) {
|
for (uint32_t i = 0; i < num_queries; i++) {
|
||||||
uint32_t slot_offset = (first_index + i) * pool->stride;
|
struct anv_address slot_addr =
|
||||||
|
anv_query_address(pool, first_index + i);
|
||||||
for (uint32_t j = 1; j < num_elements; j++) {
|
for (uint32_t j = 1; j < num_elements; j++) {
|
||||||
anv_batch_emit(&cmd_buffer->batch, GENX(MI_STORE_DATA_IMM), sdi) {
|
anv_batch_emit(&cmd_buffer->batch, GENX(MI_STORE_DATA_IMM), sdi) {
|
||||||
sdi.Address.bo = &pool->bo;
|
sdi.Address = anv_address_add(slot_addr, j * sizeof(uint64_t));
|
||||||
sdi.Address.offset = slot_offset + j * sizeof(uint64_t);
|
|
||||||
sdi.ImmediateData = 0ull;
|
sdi.ImmediateData = 0ull;
|
||||||
}
|
}
|
||||||
anv_batch_emit(&cmd_buffer->batch, GENX(MI_STORE_DATA_IMM), sdi) {
|
anv_batch_emit(&cmd_buffer->batch, GENX(MI_STORE_DATA_IMM), sdi) {
|
||||||
sdi.Address.bo = &pool->bo;
|
sdi.Address = anv_address_add(slot_addr, j * sizeof(uint64_t) + 4);
|
||||||
sdi.Address.offset = slot_offset + j * sizeof(uint64_t) + 4;
|
|
||||||
sdi.ImmediateData = 0ull;
|
sdi.ImmediateData = 0ull;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
emit_query_availability(cmd_buffer, &pool->bo, slot_offset);
|
emit_query_availability(cmd_buffer, slot_addr);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -368,10 +376,7 @@ void genX(CmdResetQueryPool)(
|
|||||||
|
|
||||||
for (uint32_t i = 0; i < queryCount; i++) {
|
for (uint32_t i = 0; i < queryCount; i++) {
|
||||||
anv_batch_emit(&cmd_buffer->batch, GENX(MI_STORE_DATA_IMM), sdm) {
|
anv_batch_emit(&cmd_buffer->batch, GENX(MI_STORE_DATA_IMM), sdm) {
|
||||||
sdm.Address = (struct anv_address) {
|
sdm.Address = anv_query_address(pool, firstQuery + i);
|
||||||
.bo = &pool->bo,
|
|
||||||
.offset = (firstQuery + i) * pool->stride,
|
|
||||||
};
|
|
||||||
sdm.ImmediateData = 0;
|
sdm.ImmediateData = 0;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -393,7 +398,7 @@ static const uint32_t vk_pipeline_stat_to_reg[] = {
|
|||||||
|
|
||||||
static void
|
static void
|
||||||
emit_pipeline_stat(struct anv_cmd_buffer *cmd_buffer, uint32_t stat,
|
emit_pipeline_stat(struct anv_cmd_buffer *cmd_buffer, uint32_t stat,
|
||||||
struct anv_bo *bo, uint32_t offset)
|
struct anv_address addr)
|
||||||
{
|
{
|
||||||
STATIC_ASSERT(ANV_PIPELINE_STATISTICS_MASK ==
|
STATIC_ASSERT(ANV_PIPELINE_STATISTICS_MASK ==
|
||||||
(1 << ARRAY_SIZE(vk_pipeline_stat_to_reg)) - 1);
|
(1 << ARRAY_SIZE(vk_pipeline_stat_to_reg)) - 1);
|
||||||
@@ -402,12 +407,12 @@ emit_pipeline_stat(struct anv_cmd_buffer *cmd_buffer, uint32_t stat,
|
|||||||
uint32_t reg = vk_pipeline_stat_to_reg[stat];
|
uint32_t reg = vk_pipeline_stat_to_reg[stat];
|
||||||
|
|
||||||
anv_batch_emit(&cmd_buffer->batch, GENX(MI_STORE_REGISTER_MEM), lrm) {
|
anv_batch_emit(&cmd_buffer->batch, GENX(MI_STORE_REGISTER_MEM), lrm) {
|
||||||
lrm.RegisterAddress = reg,
|
lrm.RegisterAddress = reg;
|
||||||
lrm.MemoryAddress = (struct anv_address) { bo, offset };
|
lrm.MemoryAddress = anv_address_add(addr, 0);
|
||||||
}
|
}
|
||||||
anv_batch_emit(&cmd_buffer->batch, GENX(MI_STORE_REGISTER_MEM), lrm) {
|
anv_batch_emit(&cmd_buffer->batch, GENX(MI_STORE_REGISTER_MEM), lrm) {
|
||||||
lrm.RegisterAddress = reg + 4,
|
lrm.RegisterAddress = reg + 4;
|
||||||
lrm.MemoryAddress = (struct anv_address) { bo, offset + 4 };
|
lrm.MemoryAddress = anv_address_add(addr, 4);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -419,10 +424,11 @@ void genX(CmdBeginQuery)(
|
|||||||
{
|
{
|
||||||
ANV_FROM_HANDLE(anv_cmd_buffer, cmd_buffer, commandBuffer);
|
ANV_FROM_HANDLE(anv_cmd_buffer, cmd_buffer, commandBuffer);
|
||||||
ANV_FROM_HANDLE(anv_query_pool, pool, queryPool);
|
ANV_FROM_HANDLE(anv_query_pool, pool, queryPool);
|
||||||
|
struct anv_address query_addr = anv_query_address(pool, query);
|
||||||
|
|
||||||
switch (pool->type) {
|
switch (pool->type) {
|
||||||
case VK_QUERY_TYPE_OCCLUSION:
|
case VK_QUERY_TYPE_OCCLUSION:
|
||||||
emit_ps_depth_count(cmd_buffer, &pool->bo, query * pool->stride + 8);
|
emit_ps_depth_count(cmd_buffer, anv_address_add(query_addr, 8));
|
||||||
break;
|
break;
|
||||||
|
|
||||||
case VK_QUERY_TYPE_PIPELINE_STATISTICS: {
|
case VK_QUERY_TYPE_PIPELINE_STATISTICS: {
|
||||||
@@ -433,10 +439,11 @@ void genX(CmdBeginQuery)(
|
|||||||
}
|
}
|
||||||
|
|
||||||
uint32_t statistics = pool->pipeline_statistics;
|
uint32_t statistics = pool->pipeline_statistics;
|
||||||
uint32_t offset = query * pool->stride + 8;
|
uint32_t offset = 8;
|
||||||
while (statistics) {
|
while (statistics) {
|
||||||
uint32_t stat = u_bit_scan(&statistics);
|
uint32_t stat = u_bit_scan(&statistics);
|
||||||
emit_pipeline_stat(cmd_buffer, stat, &pool->bo, offset);
|
emit_pipeline_stat(cmd_buffer, stat,
|
||||||
|
anv_address_add(query_addr, offset));
|
||||||
offset += 16;
|
offset += 16;
|
||||||
}
|
}
|
||||||
break;
|
break;
|
||||||
@@ -454,11 +461,12 @@ void genX(CmdEndQuery)(
|
|||||||
{
|
{
|
||||||
ANV_FROM_HANDLE(anv_cmd_buffer, cmd_buffer, commandBuffer);
|
ANV_FROM_HANDLE(anv_cmd_buffer, cmd_buffer, commandBuffer);
|
||||||
ANV_FROM_HANDLE(anv_query_pool, pool, queryPool);
|
ANV_FROM_HANDLE(anv_query_pool, pool, queryPool);
|
||||||
|
struct anv_address query_addr = anv_query_address(pool, query);
|
||||||
|
|
||||||
switch (pool->type) {
|
switch (pool->type) {
|
||||||
case VK_QUERY_TYPE_OCCLUSION:
|
case VK_QUERY_TYPE_OCCLUSION:
|
||||||
emit_ps_depth_count(cmd_buffer, &pool->bo, query * pool->stride + 16);
|
emit_ps_depth_count(cmd_buffer, anv_address_add(query_addr, 16));
|
||||||
emit_query_availability(cmd_buffer, &pool->bo, query * pool->stride);
|
emit_query_availability(cmd_buffer, query_addr);
|
||||||
break;
|
break;
|
||||||
|
|
||||||
case VK_QUERY_TYPE_PIPELINE_STATISTICS: {
|
case VK_QUERY_TYPE_PIPELINE_STATISTICS: {
|
||||||
@@ -469,14 +477,15 @@ void genX(CmdEndQuery)(
|
|||||||
}
|
}
|
||||||
|
|
||||||
uint32_t statistics = pool->pipeline_statistics;
|
uint32_t statistics = pool->pipeline_statistics;
|
||||||
uint32_t offset = query * pool->stride + 16;
|
uint32_t offset = 16;
|
||||||
while (statistics) {
|
while (statistics) {
|
||||||
uint32_t stat = u_bit_scan(&statistics);
|
uint32_t stat = u_bit_scan(&statistics);
|
||||||
emit_pipeline_stat(cmd_buffer, stat, &pool->bo, offset);
|
emit_pipeline_stat(cmd_buffer, stat,
|
||||||
|
anv_address_add(query_addr, offset));
|
||||||
offset += 16;
|
offset += 16;
|
||||||
}
|
}
|
||||||
|
|
||||||
emit_query_availability(cmd_buffer, &pool->bo, query * pool->stride);
|
emit_query_availability(cmd_buffer, query_addr);
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -510,7 +519,7 @@ void genX(CmdWriteTimestamp)(
|
|||||||
{
|
{
|
||||||
ANV_FROM_HANDLE(anv_cmd_buffer, cmd_buffer, commandBuffer);
|
ANV_FROM_HANDLE(anv_cmd_buffer, cmd_buffer, commandBuffer);
|
||||||
ANV_FROM_HANDLE(anv_query_pool, pool, queryPool);
|
ANV_FROM_HANDLE(anv_query_pool, pool, queryPool);
|
||||||
uint32_t offset = query * pool->stride;
|
struct anv_address query_addr = anv_query_address(pool, query);
|
||||||
|
|
||||||
assert(pool->type == VK_QUERY_TYPE_TIMESTAMP);
|
assert(pool->type == VK_QUERY_TYPE_TIMESTAMP);
|
||||||
|
|
||||||
@@ -518,11 +527,11 @@ void genX(CmdWriteTimestamp)(
|
|||||||
case VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT:
|
case VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT:
|
||||||
anv_batch_emit(&cmd_buffer->batch, GENX(MI_STORE_REGISTER_MEM), srm) {
|
anv_batch_emit(&cmd_buffer->batch, GENX(MI_STORE_REGISTER_MEM), srm) {
|
||||||
srm.RegisterAddress = TIMESTAMP;
|
srm.RegisterAddress = TIMESTAMP;
|
||||||
srm.MemoryAddress = (struct anv_address) { &pool->bo, offset + 8 };
|
srm.MemoryAddress = anv_address_add(query_addr, 8);
|
||||||
}
|
}
|
||||||
anv_batch_emit(&cmd_buffer->batch, GENX(MI_STORE_REGISTER_MEM), srm) {
|
anv_batch_emit(&cmd_buffer->batch, GENX(MI_STORE_REGISTER_MEM), srm) {
|
||||||
srm.RegisterAddress = TIMESTAMP + 4;
|
srm.RegisterAddress = TIMESTAMP + 4;
|
||||||
srm.MemoryAddress = (struct anv_address) { &pool->bo, offset + 12 };
|
srm.MemoryAddress = anv_address_add(query_addr, 12);
|
||||||
}
|
}
|
||||||
break;
|
break;
|
||||||
|
|
||||||
@@ -531,7 +540,7 @@ void genX(CmdWriteTimestamp)(
|
|||||||
anv_batch_emit(&cmd_buffer->batch, GENX(PIPE_CONTROL), pc) {
|
anv_batch_emit(&cmd_buffer->batch, GENX(PIPE_CONTROL), pc) {
|
||||||
pc.DestinationAddressType = DAT_PPGTT;
|
pc.DestinationAddressType = DAT_PPGTT;
|
||||||
pc.PostSyncOperation = WriteTimestamp;
|
pc.PostSyncOperation = WriteTimestamp;
|
||||||
pc.Address = (struct anv_address) { &pool->bo, offset + 8 };
|
pc.Address = anv_address_add(query_addr, 8);
|
||||||
|
|
||||||
if (GEN_GEN == 9 && cmd_buffer->device->info.gt == 4)
|
if (GEN_GEN == 9 && cmd_buffer->device->info.gt == 4)
|
||||||
pc.CommandStreamerStallEnable = true;
|
pc.CommandStreamerStallEnable = true;
|
||||||
@@ -539,7 +548,7 @@ void genX(CmdWriteTimestamp)(
|
|||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
|
||||||
emit_query_availability(cmd_buffer, &pool->bo, offset);
|
emit_query_availability(cmd_buffer, query_addr);
|
||||||
|
|
||||||
/* When multiview is active the spec requires that N consecutive query
|
/* When multiview is active the spec requires that N consecutive query
|
||||||
* indices are used, where N is the number of active views in the subpass.
|
* indices are used, where N is the number of active views in the subpass.
|
||||||
@@ -578,15 +587,15 @@ mi_alu(uint32_t opcode, uint32_t operand1, uint32_t operand2)
|
|||||||
|
|
||||||
static void
|
static void
|
||||||
emit_load_alu_reg_u64(struct anv_batch *batch, uint32_t reg,
|
emit_load_alu_reg_u64(struct anv_batch *batch, uint32_t reg,
|
||||||
struct anv_bo *bo, uint32_t offset)
|
struct anv_address addr)
|
||||||
{
|
{
|
||||||
anv_batch_emit(batch, GENX(MI_LOAD_REGISTER_MEM), lrm) {
|
anv_batch_emit(batch, GENX(MI_LOAD_REGISTER_MEM), lrm) {
|
||||||
lrm.RegisterAddress = reg,
|
lrm.RegisterAddress = reg;
|
||||||
lrm.MemoryAddress = (struct anv_address) { bo, offset };
|
lrm.MemoryAddress = anv_address_add(addr, 0);
|
||||||
}
|
}
|
||||||
anv_batch_emit(batch, GENX(MI_LOAD_REGISTER_MEM), lrm) {
|
anv_batch_emit(batch, GENX(MI_LOAD_REGISTER_MEM), lrm) {
|
||||||
lrm.RegisterAddress = reg + 4;
|
lrm.RegisterAddress = reg + 4;
|
||||||
lrm.MemoryAddress = (struct anv_address) { bo, offset + 4 };
|
lrm.MemoryAddress = anv_address_add(addr, 4);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -686,35 +695,34 @@ shr_gpr0_by_2_bits(struct anv_batch *batch)
|
|||||||
|
|
||||||
static void
|
static void
|
||||||
gpu_write_query_result(struct anv_batch *batch,
|
gpu_write_query_result(struct anv_batch *batch,
|
||||||
struct anv_buffer *dst_buffer, uint32_t dst_offset,
|
struct anv_address dst_addr,
|
||||||
VkQueryResultFlags flags,
|
VkQueryResultFlags flags,
|
||||||
uint32_t value_index, uint32_t reg)
|
uint32_t value_index, uint32_t reg)
|
||||||
{
|
{
|
||||||
if (flags & VK_QUERY_RESULT_64_BIT)
|
if (flags & VK_QUERY_RESULT_64_BIT)
|
||||||
dst_offset += value_index * 8;
|
dst_addr = anv_address_add(dst_addr, value_index * 8);
|
||||||
else
|
else
|
||||||
dst_offset += value_index * 4;
|
dst_addr = anv_address_add(dst_addr, value_index * 4);
|
||||||
|
|
||||||
anv_batch_emit(batch, GENX(MI_STORE_REGISTER_MEM), srm) {
|
anv_batch_emit(batch, GENX(MI_STORE_REGISTER_MEM), srm) {
|
||||||
srm.RegisterAddress = reg;
|
srm.RegisterAddress = reg;
|
||||||
srm.MemoryAddress = anv_address_add(dst_buffer->address, dst_offset);
|
srm.MemoryAddress = anv_address_add(dst_addr, 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (flags & VK_QUERY_RESULT_64_BIT) {
|
if (flags & VK_QUERY_RESULT_64_BIT) {
|
||||||
anv_batch_emit(batch, GENX(MI_STORE_REGISTER_MEM), srm) {
|
anv_batch_emit(batch, GENX(MI_STORE_REGISTER_MEM), srm) {
|
||||||
srm.RegisterAddress = reg + 4;
|
srm.RegisterAddress = reg + 4;
|
||||||
srm.MemoryAddress = anv_address_add(dst_buffer->address,
|
srm.MemoryAddress = anv_address_add(dst_addr, 4);
|
||||||
dst_offset + 4);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static void
|
static void
|
||||||
compute_query_result(struct anv_batch *batch, uint32_t dst_reg,
|
compute_query_result(struct anv_batch *batch, uint32_t dst_reg,
|
||||||
struct anv_bo *bo, uint32_t offset)
|
struct anv_address addr)
|
||||||
{
|
{
|
||||||
emit_load_alu_reg_u64(batch, CS_GPR(0), bo, offset);
|
emit_load_alu_reg_u64(batch, CS_GPR(0), anv_address_add(addr, 0));
|
||||||
emit_load_alu_reg_u64(batch, CS_GPR(1), bo, offset + 8);
|
emit_load_alu_reg_u64(batch, CS_GPR(1), anv_address_add(addr, 8));
|
||||||
|
|
||||||
/* FIXME: We need to clamp the result for 32 bit. */
|
/* FIXME: We need to clamp the result for 32 bit. */
|
||||||
|
|
||||||
@@ -743,7 +751,6 @@ void genX(CmdCopyQueryPoolResults)(
|
|||||||
ANV_FROM_HANDLE(anv_cmd_buffer, cmd_buffer, commandBuffer);
|
ANV_FROM_HANDLE(anv_cmd_buffer, cmd_buffer, commandBuffer);
|
||||||
ANV_FROM_HANDLE(anv_query_pool, pool, queryPool);
|
ANV_FROM_HANDLE(anv_query_pool, pool, queryPool);
|
||||||
ANV_FROM_HANDLE(anv_buffer, buffer, destBuffer);
|
ANV_FROM_HANDLE(anv_buffer, buffer, destBuffer);
|
||||||
uint32_t slot_offset;
|
|
||||||
|
|
||||||
if (flags & VK_QUERY_RESULT_WAIT_BIT) {
|
if (flags & VK_QUERY_RESULT_WAIT_BIT) {
|
||||||
anv_batch_emit(&cmd_buffer->batch, GENX(PIPE_CONTROL), pc) {
|
anv_batch_emit(&cmd_buffer->batch, GENX(PIPE_CONTROL), pc) {
|
||||||
@@ -752,14 +759,15 @@ void genX(CmdCopyQueryPoolResults)(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
struct anv_address dest_addr = anv_address_add(buffer->address, destOffset);
|
||||||
for (uint32_t i = 0; i < queryCount; i++) {
|
for (uint32_t i = 0; i < queryCount; i++) {
|
||||||
slot_offset = (firstQuery + i) * pool->stride;
|
struct anv_address query_addr = anv_query_address(pool, firstQuery + i);
|
||||||
uint32_t idx = 0;
|
uint32_t idx = 0;
|
||||||
switch (pool->type) {
|
switch (pool->type) {
|
||||||
case VK_QUERY_TYPE_OCCLUSION:
|
case VK_QUERY_TYPE_OCCLUSION:
|
||||||
compute_query_result(&cmd_buffer->batch, MI_ALU_REG2,
|
compute_query_result(&cmd_buffer->batch, MI_ALU_REG2,
|
||||||
&pool->bo, slot_offset + 8);
|
anv_address_add(query_addr, 8));
|
||||||
gpu_write_query_result(&cmd_buffer->batch, buffer, destOffset,
|
gpu_write_query_result(&cmd_buffer->batch, dest_addr,
|
||||||
flags, idx++, CS_GPR(2));
|
flags, idx++, CS_GPR(2));
|
||||||
break;
|
break;
|
||||||
|
|
||||||
@@ -769,7 +777,7 @@ void genX(CmdCopyQueryPoolResults)(
|
|||||||
uint32_t stat = u_bit_scan(&statistics);
|
uint32_t stat = u_bit_scan(&statistics);
|
||||||
|
|
||||||
compute_query_result(&cmd_buffer->batch, MI_ALU_REG0,
|
compute_query_result(&cmd_buffer->batch, MI_ALU_REG0,
|
||||||
&pool->bo, slot_offset + idx * 16 + 8);
|
anv_address_add(query_addr, idx * 16 + 8));
|
||||||
|
|
||||||
/* WaDividePSInvocationCountBy4:HSW,BDW */
|
/* WaDividePSInvocationCountBy4:HSW,BDW */
|
||||||
if ((cmd_buffer->device->info.gen == 8 ||
|
if ((cmd_buffer->device->info.gen == 8 ||
|
||||||
@@ -778,7 +786,7 @@ void genX(CmdCopyQueryPoolResults)(
|
|||||||
shr_gpr0_by_2_bits(&cmd_buffer->batch);
|
shr_gpr0_by_2_bits(&cmd_buffer->batch);
|
||||||
}
|
}
|
||||||
|
|
||||||
gpu_write_query_result(&cmd_buffer->batch, buffer, destOffset,
|
gpu_write_query_result(&cmd_buffer->batch, dest_addr,
|
||||||
flags, idx++, CS_GPR(0));
|
flags, idx++, CS_GPR(0));
|
||||||
}
|
}
|
||||||
assert(idx == util_bitcount(pool->pipeline_statistics));
|
assert(idx == util_bitcount(pool->pipeline_statistics));
|
||||||
@@ -787,8 +795,8 @@ void genX(CmdCopyQueryPoolResults)(
|
|||||||
|
|
||||||
case VK_QUERY_TYPE_TIMESTAMP:
|
case VK_QUERY_TYPE_TIMESTAMP:
|
||||||
emit_load_alu_reg_u64(&cmd_buffer->batch,
|
emit_load_alu_reg_u64(&cmd_buffer->batch,
|
||||||
CS_GPR(2), &pool->bo, slot_offset + 8);
|
CS_GPR(2), anv_address_add(query_addr, 8));
|
||||||
gpu_write_query_result(&cmd_buffer->batch, buffer, destOffset,
|
gpu_write_query_result(&cmd_buffer->batch, dest_addr,
|
||||||
flags, 0, CS_GPR(2));
|
flags, 0, CS_GPR(2));
|
||||||
break;
|
break;
|
||||||
|
|
||||||
@@ -797,13 +805,12 @@ void genX(CmdCopyQueryPoolResults)(
|
|||||||
}
|
}
|
||||||
|
|
||||||
if (flags & VK_QUERY_RESULT_WITH_AVAILABILITY_BIT) {
|
if (flags & VK_QUERY_RESULT_WITH_AVAILABILITY_BIT) {
|
||||||
emit_load_alu_reg_u64(&cmd_buffer->batch, CS_GPR(0),
|
emit_load_alu_reg_u64(&cmd_buffer->batch, CS_GPR(0), query_addr);
|
||||||
&pool->bo, slot_offset);
|
gpu_write_query_result(&cmd_buffer->batch, dest_addr,
|
||||||
gpu_write_query_result(&cmd_buffer->batch, buffer, destOffset,
|
|
||||||
flags, idx, CS_GPR(0));
|
flags, idx, CS_GPR(0));
|
||||||
}
|
}
|
||||||
|
|
||||||
destOffset += destStride;
|
dest_addr = anv_address_add(dest_addr, destStride);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Reference in New Issue
Block a user