radeonsi: move remaining perfcounter code into si_perfcounter.c
Reviewed-by: Marek Olšák <marek.olsak@amd.com>
This commit is contained in:
@@ -1,639 +0,0 @@
|
|||||||
/*
|
|
||||||
* Copyright 2015 Advanced Micro Devices, Inc.
|
|
||||||
* All Rights Reserved.
|
|
||||||
*
|
|
||||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
||||||
* copy of this software and associated documentation files (the "Software"),
|
|
||||||
* to deal in the Software without restriction, including without limitation
|
|
||||||
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
||||||
* and/or sell copies of the Software, and to permit persons to whom the
|
|
||||||
* Software is furnished to do so, subject to the following conditions:
|
|
||||||
*
|
|
||||||
* The above copyright notice and this permission notice (including the next
|
|
||||||
* paragraph) shall be included in all copies or substantial portions of the
|
|
||||||
* Software.
|
|
||||||
*
|
|
||||||
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
||||||
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
||||||
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
||||||
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
||||||
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
|
||||||
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
||||||
* SOFTWARE.
|
|
||||||
*/
|
|
||||||
|
|
||||||
#include "util/u_memory.h"
|
|
||||||
#include "radeonsi/si_query.h"
|
|
||||||
#include "radeonsi/si_pipe.h"
|
|
||||||
#include "amd/common/sid.h"
|
|
||||||
|
|
||||||
/* Max counters per HW block */
|
|
||||||
#define SI_QUERY_MAX_COUNTERS 16
|
|
||||||
|
|
||||||
static struct si_perfcounter_block *
|
|
||||||
lookup_counter(struct si_perfcounters *pc, unsigned index,
|
|
||||||
unsigned *base_gid, unsigned *sub_index)
|
|
||||||
{
|
|
||||||
struct si_perfcounter_block *block = pc->blocks;
|
|
||||||
unsigned bid;
|
|
||||||
|
|
||||||
*base_gid = 0;
|
|
||||||
for (bid = 0; bid < pc->num_blocks; ++bid, ++block) {
|
|
||||||
unsigned total = block->num_groups * block->num_selectors;
|
|
||||||
|
|
||||||
if (index < total) {
|
|
||||||
*sub_index = index;
|
|
||||||
return block;
|
|
||||||
}
|
|
||||||
|
|
||||||
index -= total;
|
|
||||||
*base_gid += block->num_groups;
|
|
||||||
}
|
|
||||||
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
static struct si_perfcounter_block *
|
|
||||||
lookup_group(struct si_perfcounters *pc, unsigned *index)
|
|
||||||
{
|
|
||||||
unsigned bid;
|
|
||||||
struct si_perfcounter_block *block = pc->blocks;
|
|
||||||
|
|
||||||
for (bid = 0; bid < pc->num_blocks; ++bid, ++block) {
|
|
||||||
if (*index < block->num_groups)
|
|
||||||
return block;
|
|
||||||
*index -= block->num_groups;
|
|
||||||
}
|
|
||||||
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
struct si_pc_group {
|
|
||||||
struct si_pc_group *next;
|
|
||||||
struct si_perfcounter_block *block;
|
|
||||||
unsigned sub_gid; /* only used during init */
|
|
||||||
unsigned result_base; /* only used during init */
|
|
||||||
int se;
|
|
||||||
int instance;
|
|
||||||
unsigned num_counters;
|
|
||||||
unsigned selectors[SI_QUERY_MAX_COUNTERS];
|
|
||||||
};
|
|
||||||
|
|
||||||
struct si_pc_counter {
|
|
||||||
unsigned base;
|
|
||||||
unsigned qwords;
|
|
||||||
unsigned stride; /* in uint64s */
|
|
||||||
};
|
|
||||||
|
|
||||||
#define SI_PC_SHADERS_WINDOWING (1 << 31)
|
|
||||||
|
|
||||||
struct si_query_pc {
|
|
||||||
struct si_query_hw b;
|
|
||||||
|
|
||||||
unsigned shaders;
|
|
||||||
unsigned num_counters;
|
|
||||||
struct si_pc_counter *counters;
|
|
||||||
struct si_pc_group *groups;
|
|
||||||
};
|
|
||||||
|
|
||||||
static void si_pc_query_destroy(struct si_screen *sscreen,
|
|
||||||
struct si_query *rquery)
|
|
||||||
{
|
|
||||||
struct si_query_pc *query = (struct si_query_pc *)rquery;
|
|
||||||
|
|
||||||
while (query->groups) {
|
|
||||||
struct si_pc_group *group = query->groups;
|
|
||||||
query->groups = group->next;
|
|
||||||
FREE(group);
|
|
||||||
}
|
|
||||||
|
|
||||||
FREE(query->counters);
|
|
||||||
|
|
||||||
si_query_hw_destroy(sscreen, rquery);
|
|
||||||
}
|
|
||||||
|
|
||||||
static bool si_pc_query_prepare_buffer(struct si_screen *screen,
|
|
||||||
struct si_query_hw *hwquery,
|
|
||||||
struct r600_resource *buffer)
|
|
||||||
{
|
|
||||||
/* no-op */
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
static void si_pc_query_emit_start(struct si_context *sctx,
|
|
||||||
struct si_query_hw *hwquery,
|
|
||||||
struct r600_resource *buffer, uint64_t va)
|
|
||||||
{
|
|
||||||
struct si_perfcounters *pc = sctx->screen->perfcounters;
|
|
||||||
struct si_query_pc *query = (struct si_query_pc *)hwquery;
|
|
||||||
struct si_pc_group *group;
|
|
||||||
int current_se = -1;
|
|
||||||
int current_instance = -1;
|
|
||||||
|
|
||||||
if (query->shaders)
|
|
||||||
pc->emit_shaders(sctx, query->shaders);
|
|
||||||
|
|
||||||
for (group = query->groups; group; group = group->next) {
|
|
||||||
struct si_perfcounter_block *block = group->block;
|
|
||||||
|
|
||||||
if (group->se != current_se || group->instance != current_instance) {
|
|
||||||
current_se = group->se;
|
|
||||||
current_instance = group->instance;
|
|
||||||
pc->emit_instance(sctx, group->se, group->instance);
|
|
||||||
}
|
|
||||||
|
|
||||||
pc->emit_select(sctx, block, group->num_counters, group->selectors);
|
|
||||||
}
|
|
||||||
|
|
||||||
if (current_se != -1 || current_instance != -1)
|
|
||||||
pc->emit_instance(sctx, -1, -1);
|
|
||||||
|
|
||||||
pc->emit_start(sctx, buffer, va);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void si_pc_query_emit_stop(struct si_context *sctx,
|
|
||||||
struct si_query_hw *hwquery,
|
|
||||||
struct r600_resource *buffer, uint64_t va)
|
|
||||||
{
|
|
||||||
struct si_perfcounters *pc = sctx->screen->perfcounters;
|
|
||||||
struct si_query_pc *query = (struct si_query_pc *)hwquery;
|
|
||||||
struct si_pc_group *group;
|
|
||||||
|
|
||||||
pc->emit_stop(sctx, buffer, va);
|
|
||||||
|
|
||||||
for (group = query->groups; group; group = group->next) {
|
|
||||||
struct si_perfcounter_block *block = group->block;
|
|
||||||
unsigned se = group->se >= 0 ? group->se : 0;
|
|
||||||
unsigned se_end = se + 1;
|
|
||||||
|
|
||||||
if ((block->flags & SI_PC_BLOCK_SE) && (group->se < 0))
|
|
||||||
se_end = sctx->screen->info.max_se;
|
|
||||||
|
|
||||||
do {
|
|
||||||
unsigned instance = group->instance >= 0 ? group->instance : 0;
|
|
||||||
|
|
||||||
do {
|
|
||||||
pc->emit_instance(sctx, se, instance);
|
|
||||||
pc->emit_read(sctx, block,
|
|
||||||
group->num_counters, group->selectors,
|
|
||||||
buffer, va);
|
|
||||||
va += sizeof(uint64_t) * group->num_counters;
|
|
||||||
} while (group->instance < 0 && ++instance < block->num_instances);
|
|
||||||
} while (++se < se_end);
|
|
||||||
}
|
|
||||||
|
|
||||||
pc->emit_instance(sctx, -1, -1);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void si_pc_query_clear_result(struct si_query_hw *hwquery,
|
|
||||||
union pipe_query_result *result)
|
|
||||||
{
|
|
||||||
struct si_query_pc *query = (struct si_query_pc *)hwquery;
|
|
||||||
|
|
||||||
memset(result, 0, sizeof(result->batch[0]) * query->num_counters);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void si_pc_query_add_result(struct si_screen *sscreen,
|
|
||||||
struct si_query_hw *hwquery,
|
|
||||||
void *buffer,
|
|
||||||
union pipe_query_result *result)
|
|
||||||
{
|
|
||||||
struct si_query_pc *query = (struct si_query_pc *)hwquery;
|
|
||||||
uint64_t *results = buffer;
|
|
||||||
unsigned i, j;
|
|
||||||
|
|
||||||
for (i = 0; i < query->num_counters; ++i) {
|
|
||||||
struct si_pc_counter *counter = &query->counters[i];
|
|
||||||
|
|
||||||
for (j = 0; j < counter->qwords; ++j) {
|
|
||||||
uint32_t value = results[counter->base + j * counter->stride];
|
|
||||||
result->batch[i].u64 += value;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
static struct si_query_ops batch_query_ops = {
|
|
||||||
.destroy = si_pc_query_destroy,
|
|
||||||
.begin = si_query_hw_begin,
|
|
||||||
.end = si_query_hw_end,
|
|
||||||
.get_result = si_query_hw_get_result
|
|
||||||
};
|
|
||||||
|
|
||||||
static struct si_query_hw_ops batch_query_hw_ops = {
|
|
||||||
.prepare_buffer = si_pc_query_prepare_buffer,
|
|
||||||
.emit_start = si_pc_query_emit_start,
|
|
||||||
.emit_stop = si_pc_query_emit_stop,
|
|
||||||
.clear_result = si_pc_query_clear_result,
|
|
||||||
.add_result = si_pc_query_add_result,
|
|
||||||
};
|
|
||||||
|
|
||||||
static struct si_pc_group *get_group_state(struct si_screen *screen,
|
|
||||||
struct si_query_pc *query,
|
|
||||||
struct si_perfcounter_block *block,
|
|
||||||
unsigned sub_gid)
|
|
||||||
{
|
|
||||||
struct si_pc_group *group = query->groups;
|
|
||||||
|
|
||||||
while (group) {
|
|
||||||
if (group->block == block && group->sub_gid == sub_gid)
|
|
||||||
return group;
|
|
||||||
group = group->next;
|
|
||||||
}
|
|
||||||
|
|
||||||
group = CALLOC_STRUCT(si_pc_group);
|
|
||||||
if (!group)
|
|
||||||
return NULL;
|
|
||||||
|
|
||||||
group->block = block;
|
|
||||||
group->sub_gid = sub_gid;
|
|
||||||
|
|
||||||
if (block->flags & SI_PC_BLOCK_SHADER) {
|
|
||||||
unsigned sub_gids = block->num_instances;
|
|
||||||
unsigned shader_id;
|
|
||||||
unsigned shaders;
|
|
||||||
unsigned query_shaders;
|
|
||||||
|
|
||||||
if (block->flags & SI_PC_BLOCK_SE_GROUPS)
|
|
||||||
sub_gids = sub_gids * screen->info.max_se;
|
|
||||||
shader_id = sub_gid / sub_gids;
|
|
||||||
sub_gid = sub_gid % sub_gids;
|
|
||||||
|
|
||||||
shaders = screen->perfcounters->shader_type_bits[shader_id];
|
|
||||||
|
|
||||||
query_shaders = query->shaders & ~SI_PC_SHADERS_WINDOWING;
|
|
||||||
if (query_shaders && query_shaders != shaders) {
|
|
||||||
fprintf(stderr, "si_perfcounter: incompatible shader groups\n");
|
|
||||||
FREE(group);
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
query->shaders = shaders;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (block->flags & SI_PC_BLOCK_SHADER_WINDOWED && !query->shaders) {
|
|
||||||
// A non-zero value in query->shaders ensures that the shader
|
|
||||||
// masking is reset unless the user explicitly requests one.
|
|
||||||
query->shaders = SI_PC_SHADERS_WINDOWING;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (block->flags & SI_PC_BLOCK_SE_GROUPS) {
|
|
||||||
group->se = sub_gid / block->num_instances;
|
|
||||||
sub_gid = sub_gid % block->num_instances;
|
|
||||||
} else {
|
|
||||||
group->se = -1;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (block->flags & SI_PC_BLOCK_INSTANCE_GROUPS) {
|
|
||||||
group->instance = sub_gid;
|
|
||||||
} else {
|
|
||||||
group->instance = -1;
|
|
||||||
}
|
|
||||||
|
|
||||||
group->next = query->groups;
|
|
||||||
query->groups = group;
|
|
||||||
|
|
||||||
return group;
|
|
||||||
}
|
|
||||||
|
|
||||||
struct pipe_query *si_create_batch_query(struct pipe_context *ctx,
|
|
||||||
unsigned num_queries,
|
|
||||||
unsigned *query_types)
|
|
||||||
{
|
|
||||||
struct si_screen *screen =
|
|
||||||
(struct si_screen *)ctx->screen;
|
|
||||||
struct si_perfcounters *pc = screen->perfcounters;
|
|
||||||
struct si_perfcounter_block *block;
|
|
||||||
struct si_pc_group *group;
|
|
||||||
struct si_query_pc *query;
|
|
||||||
unsigned base_gid, sub_gid, sub_index;
|
|
||||||
unsigned i, j;
|
|
||||||
|
|
||||||
if (!pc)
|
|
||||||
return NULL;
|
|
||||||
|
|
||||||
query = CALLOC_STRUCT(si_query_pc);
|
|
||||||
if (!query)
|
|
||||||
return NULL;
|
|
||||||
|
|
||||||
query->b.b.ops = &batch_query_ops;
|
|
||||||
query->b.ops = &batch_query_hw_ops;
|
|
||||||
|
|
||||||
query->num_counters = num_queries;
|
|
||||||
|
|
||||||
/* Collect selectors per group */
|
|
||||||
for (i = 0; i < num_queries; ++i) {
|
|
||||||
unsigned sub_gid;
|
|
||||||
|
|
||||||
if (query_types[i] < SI_QUERY_FIRST_PERFCOUNTER)
|
|
||||||
goto error;
|
|
||||||
|
|
||||||
block = lookup_counter(pc, query_types[i] - SI_QUERY_FIRST_PERFCOUNTER,
|
|
||||||
&base_gid, &sub_index);
|
|
||||||
if (!block)
|
|
||||||
goto error;
|
|
||||||
|
|
||||||
sub_gid = sub_index / block->num_selectors;
|
|
||||||
sub_index = sub_index % block->num_selectors;
|
|
||||||
|
|
||||||
group = get_group_state(screen, query, block, sub_gid);
|
|
||||||
if (!group)
|
|
||||||
goto error;
|
|
||||||
|
|
||||||
if (group->num_counters >= block->num_counters) {
|
|
||||||
fprintf(stderr,
|
|
||||||
"perfcounter group %s: too many selected\n",
|
|
||||||
block->basename);
|
|
||||||
goto error;
|
|
||||||
}
|
|
||||||
group->selectors[group->num_counters] = sub_index;
|
|
||||||
++group->num_counters;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Compute result bases and CS size per group */
|
|
||||||
query->b.num_cs_dw_end = pc->num_stop_cs_dwords;
|
|
||||||
query->b.num_cs_dw_end += pc->num_instance_cs_dwords;
|
|
||||||
|
|
||||||
i = 0;
|
|
||||||
for (group = query->groups; group; group = group->next) {
|
|
||||||
struct si_perfcounter_block *block = group->block;
|
|
||||||
unsigned read_dw;
|
|
||||||
unsigned instances = 1;
|
|
||||||
|
|
||||||
if ((block->flags & SI_PC_BLOCK_SE) && group->se < 0)
|
|
||||||
instances = screen->info.max_se;
|
|
||||||
if (group->instance < 0)
|
|
||||||
instances *= block->num_instances;
|
|
||||||
|
|
||||||
group->result_base = i;
|
|
||||||
query->b.result_size += sizeof(uint64_t) * instances * group->num_counters;
|
|
||||||
i += instances * group->num_counters;
|
|
||||||
|
|
||||||
read_dw = 6 * group->num_counters;
|
|
||||||
query->b.num_cs_dw_end += instances * read_dw;
|
|
||||||
query->b.num_cs_dw_end += instances * pc->num_instance_cs_dwords;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (query->shaders) {
|
|
||||||
if (query->shaders == SI_PC_SHADERS_WINDOWING)
|
|
||||||
query->shaders = 0xffffffff;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Map user-supplied query array to result indices */
|
|
||||||
query->counters = CALLOC(num_queries, sizeof(*query->counters));
|
|
||||||
for (i = 0; i < num_queries; ++i) {
|
|
||||||
struct si_pc_counter *counter = &query->counters[i];
|
|
||||||
struct si_perfcounter_block *block;
|
|
||||||
|
|
||||||
block = lookup_counter(pc, query_types[i] - SI_QUERY_FIRST_PERFCOUNTER,
|
|
||||||
&base_gid, &sub_index);
|
|
||||||
|
|
||||||
sub_gid = sub_index / block->num_selectors;
|
|
||||||
sub_index = sub_index % block->num_selectors;
|
|
||||||
|
|
||||||
group = get_group_state(screen, query, block, sub_gid);
|
|
||||||
assert(group != NULL);
|
|
||||||
|
|
||||||
for (j = 0; j < group->num_counters; ++j) {
|
|
||||||
if (group->selectors[j] == sub_index)
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
|
|
||||||
counter->base = group->result_base + j;
|
|
||||||
counter->stride = group->num_counters;
|
|
||||||
|
|
||||||
counter->qwords = 1;
|
|
||||||
if ((block->flags & SI_PC_BLOCK_SE) && group->se < 0)
|
|
||||||
counter->qwords = screen->info.max_se;
|
|
||||||
if (group->instance < 0)
|
|
||||||
counter->qwords *= block->num_instances;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!si_query_hw_init(screen, &query->b))
|
|
||||||
goto error;
|
|
||||||
|
|
||||||
return (struct pipe_query *)query;
|
|
||||||
|
|
||||||
error:
|
|
||||||
si_pc_query_destroy(screen, &query->b.b);
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
static bool si_init_block_names(struct si_screen *screen,
|
|
||||||
struct si_perfcounter_block *block)
|
|
||||||
{
|
|
||||||
unsigned i, j, k;
|
|
||||||
unsigned groups_shader = 1, groups_se = 1, groups_instance = 1;
|
|
||||||
unsigned namelen;
|
|
||||||
char *groupname;
|
|
||||||
char *p;
|
|
||||||
|
|
||||||
if (block->flags & SI_PC_BLOCK_INSTANCE_GROUPS)
|
|
||||||
groups_instance = block->num_instances;
|
|
||||||
if (block->flags & SI_PC_BLOCK_SE_GROUPS)
|
|
||||||
groups_se = screen->info.max_se;
|
|
||||||
if (block->flags & SI_PC_BLOCK_SHADER)
|
|
||||||
groups_shader = screen->perfcounters->num_shader_types;
|
|
||||||
|
|
||||||
namelen = strlen(block->basename);
|
|
||||||
block->group_name_stride = namelen + 1;
|
|
||||||
if (block->flags & SI_PC_BLOCK_SHADER)
|
|
||||||
block->group_name_stride += 3;
|
|
||||||
if (block->flags & SI_PC_BLOCK_SE_GROUPS) {
|
|
||||||
assert(groups_se <= 10);
|
|
||||||
block->group_name_stride += 1;
|
|
||||||
|
|
||||||
if (block->flags & SI_PC_BLOCK_INSTANCE_GROUPS)
|
|
||||||
block->group_name_stride += 1;
|
|
||||||
}
|
|
||||||
if (block->flags & SI_PC_BLOCK_INSTANCE_GROUPS) {
|
|
||||||
assert(groups_instance <= 100);
|
|
||||||
block->group_name_stride += 2;
|
|
||||||
}
|
|
||||||
|
|
||||||
block->group_names = MALLOC(block->num_groups * block->group_name_stride);
|
|
||||||
if (!block->group_names)
|
|
||||||
return false;
|
|
||||||
|
|
||||||
groupname = block->group_names;
|
|
||||||
for (i = 0; i < groups_shader; ++i) {
|
|
||||||
const char *shader_suffix = screen->perfcounters->shader_type_suffixes[i];
|
|
||||||
unsigned shaderlen = strlen(shader_suffix);
|
|
||||||
for (j = 0; j < groups_se; ++j) {
|
|
||||||
for (k = 0; k < groups_instance; ++k) {
|
|
||||||
strcpy(groupname, block->basename);
|
|
||||||
p = groupname + namelen;
|
|
||||||
|
|
||||||
if (block->flags & SI_PC_BLOCK_SHADER) {
|
|
||||||
strcpy(p, shader_suffix);
|
|
||||||
p += shaderlen;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (block->flags & SI_PC_BLOCK_SE_GROUPS) {
|
|
||||||
p += sprintf(p, "%d", j);
|
|
||||||
if (block->flags & SI_PC_BLOCK_INSTANCE_GROUPS)
|
|
||||||
*p++ = '_';
|
|
||||||
}
|
|
||||||
|
|
||||||
if (block->flags & SI_PC_BLOCK_INSTANCE_GROUPS)
|
|
||||||
p += sprintf(p, "%d", k);
|
|
||||||
|
|
||||||
groupname += block->group_name_stride;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
assert(block->num_selectors <= 1000);
|
|
||||||
block->selector_name_stride = block->group_name_stride + 4;
|
|
||||||
block->selector_names = MALLOC(block->num_groups * block->num_selectors *
|
|
||||||
block->selector_name_stride);
|
|
||||||
if (!block->selector_names)
|
|
||||||
return false;
|
|
||||||
|
|
||||||
groupname = block->group_names;
|
|
||||||
p = block->selector_names;
|
|
||||||
for (i = 0; i < block->num_groups; ++i) {
|
|
||||||
for (j = 0; j < block->num_selectors; ++j) {
|
|
||||||
sprintf(p, "%s_%03d", groupname, j);
|
|
||||||
p += block->selector_name_stride;
|
|
||||||
}
|
|
||||||
groupname += block->group_name_stride;
|
|
||||||
}
|
|
||||||
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
int si_get_perfcounter_info(struct si_screen *screen,
|
|
||||||
unsigned index,
|
|
||||||
struct pipe_driver_query_info *info)
|
|
||||||
{
|
|
||||||
struct si_perfcounters *pc = screen->perfcounters;
|
|
||||||
struct si_perfcounter_block *block;
|
|
||||||
unsigned base_gid, sub;
|
|
||||||
|
|
||||||
if (!pc)
|
|
||||||
return 0;
|
|
||||||
|
|
||||||
if (!info) {
|
|
||||||
unsigned bid, num_queries = 0;
|
|
||||||
|
|
||||||
for (bid = 0; bid < pc->num_blocks; ++bid) {
|
|
||||||
num_queries += pc->blocks[bid].num_selectors *
|
|
||||||
pc->blocks[bid].num_groups;
|
|
||||||
}
|
|
||||||
|
|
||||||
return num_queries;
|
|
||||||
}
|
|
||||||
|
|
||||||
block = lookup_counter(pc, index, &base_gid, &sub);
|
|
||||||
if (!block)
|
|
||||||
return 0;
|
|
||||||
|
|
||||||
if (!block->selector_names) {
|
|
||||||
if (!si_init_block_names(screen, block))
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
info->name = block->selector_names + sub * block->selector_name_stride;
|
|
||||||
info->query_type = SI_QUERY_FIRST_PERFCOUNTER + index;
|
|
||||||
info->max_value.u64 = 0;
|
|
||||||
info->type = PIPE_DRIVER_QUERY_TYPE_UINT64;
|
|
||||||
info->result_type = PIPE_DRIVER_QUERY_RESULT_TYPE_AVERAGE;
|
|
||||||
info->group_id = base_gid + sub / block->num_selectors;
|
|
||||||
info->flags = PIPE_DRIVER_QUERY_FLAG_BATCH;
|
|
||||||
if (sub > 0 && sub + 1 < block->num_selectors * block->num_groups)
|
|
||||||
info->flags |= PIPE_DRIVER_QUERY_FLAG_DONT_LIST;
|
|
||||||
return 1;
|
|
||||||
}
|
|
||||||
|
|
||||||
int si_get_perfcounter_group_info(struct si_screen *screen,
|
|
||||||
unsigned index,
|
|
||||||
struct pipe_driver_query_group_info *info)
|
|
||||||
{
|
|
||||||
struct si_perfcounters *pc = screen->perfcounters;
|
|
||||||
struct si_perfcounter_block *block;
|
|
||||||
|
|
||||||
if (!pc)
|
|
||||||
return 0;
|
|
||||||
|
|
||||||
if (!info)
|
|
||||||
return pc->num_groups;
|
|
||||||
|
|
||||||
block = lookup_group(pc, &index);
|
|
||||||
if (!block)
|
|
||||||
return 0;
|
|
||||||
|
|
||||||
if (!block->group_names) {
|
|
||||||
if (!si_init_block_names(screen, block))
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
info->name = block->group_names + index * block->group_name_stride;
|
|
||||||
info->num_queries = block->num_selectors;
|
|
||||||
info->max_active_queries = block->num_counters;
|
|
||||||
return 1;
|
|
||||||
}
|
|
||||||
|
|
||||||
void si_perfcounters_destroy(struct si_screen *sscreen)
|
|
||||||
{
|
|
||||||
if (sscreen->perfcounters)
|
|
||||||
sscreen->perfcounters->cleanup(sscreen);
|
|
||||||
}
|
|
||||||
|
|
||||||
bool si_perfcounters_init(struct si_perfcounters *pc,
|
|
||||||
unsigned num_blocks)
|
|
||||||
{
|
|
||||||
pc->blocks = CALLOC(num_blocks, sizeof(struct si_perfcounter_block));
|
|
||||||
if (!pc->blocks)
|
|
||||||
return false;
|
|
||||||
|
|
||||||
pc->separate_se = debug_get_bool_option("RADEON_PC_SEPARATE_SE", false);
|
|
||||||
pc->separate_instance = debug_get_bool_option("RADEON_PC_SEPARATE_INSTANCE", false);
|
|
||||||
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
void si_perfcounters_add_block(struct si_screen *sscreen,
|
|
||||||
struct si_perfcounters *pc,
|
|
||||||
const char *name, unsigned flags,
|
|
||||||
unsigned counters, unsigned selectors,
|
|
||||||
unsigned instances, void *data)
|
|
||||||
{
|
|
||||||
struct si_perfcounter_block *block = &pc->blocks[pc->num_blocks];
|
|
||||||
|
|
||||||
assert(counters <= SI_QUERY_MAX_COUNTERS);
|
|
||||||
|
|
||||||
block->basename = name;
|
|
||||||
block->flags = flags;
|
|
||||||
block->num_counters = counters;
|
|
||||||
block->num_selectors = selectors;
|
|
||||||
block->num_instances = MAX2(instances, 1);
|
|
||||||
block->data = data;
|
|
||||||
|
|
||||||
if (pc->separate_se && (block->flags & SI_PC_BLOCK_SE))
|
|
||||||
block->flags |= SI_PC_BLOCK_SE_GROUPS;
|
|
||||||
if (pc->separate_instance && block->num_instances > 1)
|
|
||||||
block->flags |= SI_PC_BLOCK_INSTANCE_GROUPS;
|
|
||||||
|
|
||||||
if (block->flags & SI_PC_BLOCK_INSTANCE_GROUPS) {
|
|
||||||
block->num_groups = block->num_instances;
|
|
||||||
} else {
|
|
||||||
block->num_groups = 1;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (block->flags & SI_PC_BLOCK_SE_GROUPS)
|
|
||||||
block->num_groups *= sscreen->info.max_se;
|
|
||||||
if (block->flags & SI_PC_BLOCK_SHADER)
|
|
||||||
block->num_groups *= pc->num_shader_types;
|
|
||||||
|
|
||||||
++pc->num_blocks;
|
|
||||||
pc->num_groups += block->num_groups;
|
|
||||||
}
|
|
||||||
|
|
||||||
void si_perfcounters_do_destroy(struct si_perfcounters *pc)
|
|
||||||
{
|
|
||||||
unsigned i;
|
|
||||||
|
|
||||||
for (i = 0; i < pc->num_blocks; ++i) {
|
|
||||||
FREE(pc->blocks[i].group_names);
|
|
||||||
FREE(pc->blocks[i].selector_names);
|
|
||||||
}
|
|
||||||
FREE(pc->blocks);
|
|
||||||
FREE(pc);
|
|
||||||
}
|
|
@@ -49,7 +49,6 @@ C_SOURCES := \
|
|||||||
si_test_dma_perf.c \
|
si_test_dma_perf.c \
|
||||||
si_texture.c \
|
si_texture.c \
|
||||||
si_uvd.c \
|
si_uvd.c \
|
||||||
../radeon/r600_perfcounter.c \
|
|
||||||
../radeon/radeon_uvd.c \
|
../radeon/radeon_uvd.c \
|
||||||
../radeon/radeon_uvd.h \
|
../radeon/radeon_uvd.h \
|
||||||
../radeon/radeon_vcn_dec_jpeg.c \
|
../radeon/radeon_vcn_dec_jpeg.c \
|
||||||
|
@@ -65,7 +65,6 @@ files_libradeonsi = files(
|
|||||||
'si_test_dma_perf.c',
|
'si_test_dma_perf.c',
|
||||||
'si_texture.c',
|
'si_texture.c',
|
||||||
'si_uvd.c',
|
'si_uvd.c',
|
||||||
'../radeon/r600_perfcounter.c',
|
|
||||||
'../radeon/radeon_uvd.c',
|
'../radeon/radeon_uvd.c',
|
||||||
'../radeon/radeon_uvd.h',
|
'../radeon/radeon_uvd.h',
|
||||||
'../radeon/radeon_vcn_enc_1_2.c',
|
'../radeon/radeon_vcn_enc_1_2.c',
|
||||||
|
@@ -27,6 +27,24 @@
|
|||||||
#include "util/u_memory.h"
|
#include "util/u_memory.h"
|
||||||
|
|
||||||
|
|
||||||
|
enum si_pc_block_flags {
|
||||||
|
/* This block is part of the shader engine */
|
||||||
|
SI_PC_BLOCK_SE = (1 << 0),
|
||||||
|
|
||||||
|
/* Expose per-instance groups instead of summing all instances (within
|
||||||
|
* an SE). */
|
||||||
|
SI_PC_BLOCK_INSTANCE_GROUPS = (1 << 1),
|
||||||
|
|
||||||
|
/* Expose per-SE groups instead of summing instances across SEs. */
|
||||||
|
SI_PC_BLOCK_SE_GROUPS = (1 << 2),
|
||||||
|
|
||||||
|
/* Shader block */
|
||||||
|
SI_PC_BLOCK_SHADER = (1 << 3),
|
||||||
|
|
||||||
|
/* Non-shader block with perfcounters windowed by shaders. */
|
||||||
|
SI_PC_BLOCK_SHADER_WINDOWED = (1 << 4),
|
||||||
|
};
|
||||||
|
|
||||||
enum si_pc_reg_layout {
|
enum si_pc_reg_layout {
|
||||||
/* All secondary selector dwords follow as one block after the primary
|
/* All secondary selector dwords follow as one block after the primary
|
||||||
* selector dwords for the counters that have secondary selectors.
|
* selector dwords for the counters that have secondary selectors.
|
||||||
@@ -69,12 +87,24 @@ struct si_pc_block_base {
|
|||||||
unsigned layout;
|
unsigned layout;
|
||||||
};
|
};
|
||||||
|
|
||||||
struct si_pc_block {
|
struct si_pc_block_gfxdescr {
|
||||||
struct si_pc_block_base *b;
|
struct si_pc_block_base *b;
|
||||||
unsigned selectors;
|
unsigned selectors;
|
||||||
unsigned instances;
|
unsigned instances;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
struct si_pc_block {
|
||||||
|
const struct si_pc_block_gfxdescr *b;
|
||||||
|
unsigned num_instances;
|
||||||
|
|
||||||
|
unsigned num_groups;
|
||||||
|
char *group_names;
|
||||||
|
unsigned group_name_stride;
|
||||||
|
|
||||||
|
char *selector_names;
|
||||||
|
unsigned selector_name_stride;
|
||||||
|
};
|
||||||
|
|
||||||
/* The order is chosen to be compatible with GPUPerfStudio's hardcoding of
|
/* The order is chosen to be compatible with GPUPerfStudio's hardcoding of
|
||||||
* performance counter group IDs.
|
* performance counter group IDs.
|
||||||
*/
|
*/
|
||||||
@@ -93,6 +123,38 @@ static const unsigned si_pc_shader_type_bits[] = {
|
|||||||
S_036780_CS_EN(1),
|
S_036780_CS_EN(1),
|
||||||
};
|
};
|
||||||
|
|
||||||
|
/* Max counters per HW block */
|
||||||
|
#define SI_QUERY_MAX_COUNTERS 16
|
||||||
|
|
||||||
|
#define SI_PC_SHADERS_WINDOWING (1 << 31)
|
||||||
|
|
||||||
|
struct si_query_group {
|
||||||
|
struct si_query_group *next;
|
||||||
|
struct si_pc_block *block;
|
||||||
|
unsigned sub_gid; /* only used during init */
|
||||||
|
unsigned result_base; /* only used during init */
|
||||||
|
int se;
|
||||||
|
int instance;
|
||||||
|
unsigned num_counters;
|
||||||
|
unsigned selectors[SI_QUERY_MAX_COUNTERS];
|
||||||
|
};
|
||||||
|
|
||||||
|
struct si_query_counter {
|
||||||
|
unsigned base;
|
||||||
|
unsigned qwords;
|
||||||
|
unsigned stride; /* in uint64s */
|
||||||
|
};
|
||||||
|
|
||||||
|
struct si_query_pc {
|
||||||
|
struct si_query_hw b;
|
||||||
|
|
||||||
|
unsigned shaders;
|
||||||
|
unsigned num_counters;
|
||||||
|
struct si_query_counter *counters;
|
||||||
|
struct si_query_group *groups;
|
||||||
|
};
|
||||||
|
|
||||||
|
|
||||||
static struct si_pc_block_base cik_CB = {
|
static struct si_pc_block_base cik_CB = {
|
||||||
.name = "CB",
|
.name = "CB",
|
||||||
.num_counters = 4,
|
.num_counters = 4,
|
||||||
@@ -344,7 +406,7 @@ static struct si_pc_block_base cik_SRBM = {
|
|||||||
* blindly once it believes it has identified the hardware, so the order of
|
* blindly once it believes it has identified the hardware, so the order of
|
||||||
* blocks here matters.
|
* blocks here matters.
|
||||||
*/
|
*/
|
||||||
static struct si_pc_block groups_CIK[] = {
|
static struct si_pc_block_gfxdescr groups_CIK[] = {
|
||||||
{ &cik_CB, 226},
|
{ &cik_CB, 226},
|
||||||
{ &cik_CPF, 17 },
|
{ &cik_CPF, 17 },
|
||||||
{ &cik_DB, 257},
|
{ &cik_DB, 257},
|
||||||
@@ -371,7 +433,7 @@ static struct si_pc_block groups_CIK[] = {
|
|||||||
|
|
||||||
};
|
};
|
||||||
|
|
||||||
static struct si_pc_block groups_VI[] = {
|
static struct si_pc_block_gfxdescr groups_VI[] = {
|
||||||
{ &cik_CB, 405},
|
{ &cik_CB, 405},
|
||||||
{ &cik_CPF, 19 },
|
{ &cik_CPF, 19 },
|
||||||
{ &cik_DB, 257},
|
{ &cik_DB, 257},
|
||||||
@@ -398,7 +460,7 @@ static struct si_pc_block groups_VI[] = {
|
|||||||
|
|
||||||
};
|
};
|
||||||
|
|
||||||
static struct si_pc_block groups_gfx9[] = {
|
static struct si_pc_block_gfxdescr groups_gfx9[] = {
|
||||||
{ &cik_CB, 438},
|
{ &cik_CB, 438},
|
||||||
{ &cik_CPF, 32 },
|
{ &cik_CPF, 32 },
|
||||||
{ &cik_DB, 328},
|
{ &cik_DB, 328},
|
||||||
@@ -422,6 +484,58 @@ static struct si_pc_block groups_gfx9[] = {
|
|||||||
{ &cik_CPC, 35 },
|
{ &cik_CPC, 35 },
|
||||||
};
|
};
|
||||||
|
|
||||||
|
static bool si_pc_block_has_per_se_groups(const struct si_perfcounters *pc,
|
||||||
|
const struct si_pc_block *block)
|
||||||
|
{
|
||||||
|
return block->b->b->flags & SI_PC_BLOCK_SE_GROUPS ||
|
||||||
|
(block->b->b->flags & SI_PC_BLOCK_SE && pc->separate_se);
|
||||||
|
}
|
||||||
|
|
||||||
|
static bool si_pc_block_has_per_instance_groups(const struct si_perfcounters *pc,
|
||||||
|
const struct si_pc_block *block)
|
||||||
|
{
|
||||||
|
return block->b->b->flags & SI_PC_BLOCK_INSTANCE_GROUPS ||
|
||||||
|
(block->num_instances > 1 && pc->separate_instance);
|
||||||
|
}
|
||||||
|
|
||||||
|
static struct si_pc_block *
|
||||||
|
lookup_counter(struct si_perfcounters *pc, unsigned index,
|
||||||
|
unsigned *base_gid, unsigned *sub_index)
|
||||||
|
{
|
||||||
|
struct si_pc_block *block = pc->blocks;
|
||||||
|
unsigned bid;
|
||||||
|
|
||||||
|
*base_gid = 0;
|
||||||
|
for (bid = 0; bid < pc->num_blocks; ++bid, ++block) {
|
||||||
|
unsigned total = block->num_groups * block->b->selectors;
|
||||||
|
|
||||||
|
if (index < total) {
|
||||||
|
*sub_index = index;
|
||||||
|
return block;
|
||||||
|
}
|
||||||
|
|
||||||
|
index -= total;
|
||||||
|
*base_gid += block->num_groups;
|
||||||
|
}
|
||||||
|
|
||||||
|
return NULL;
|
||||||
|
}
|
||||||
|
|
||||||
|
static struct si_pc_block *
|
||||||
|
lookup_group(struct si_perfcounters *pc, unsigned *index)
|
||||||
|
{
|
||||||
|
unsigned bid;
|
||||||
|
struct si_pc_block *block = pc->blocks;
|
||||||
|
|
||||||
|
for (bid = 0; bid < pc->num_blocks; ++bid, ++block) {
|
||||||
|
if (*index < block->num_groups)
|
||||||
|
return block;
|
||||||
|
*index -= block->num_groups;
|
||||||
|
}
|
||||||
|
|
||||||
|
return NULL;
|
||||||
|
}
|
||||||
|
|
||||||
static void si_pc_emit_instance(struct si_context *sctx,
|
static void si_pc_emit_instance(struct si_context *sctx,
|
||||||
int se, int instance)
|
int se, int instance)
|
||||||
{
|
{
|
||||||
@@ -454,11 +568,10 @@ static void si_pc_emit_shaders(struct si_context *sctx,
|
|||||||
}
|
}
|
||||||
|
|
||||||
static void si_pc_emit_select(struct si_context *sctx,
|
static void si_pc_emit_select(struct si_context *sctx,
|
||||||
struct si_perfcounter_block *group,
|
struct si_pc_block *block,
|
||||||
unsigned count, unsigned *selectors)
|
unsigned count, unsigned *selectors)
|
||||||
{
|
{
|
||||||
struct si_pc_block *sigroup = (struct si_pc_block *)group->data;
|
struct si_pc_block_base *regs = block->b->b;
|
||||||
struct si_pc_block_base *regs = sigroup->b;
|
|
||||||
struct radeon_cmdbuf *cs = sctx->gfx_cs;
|
struct radeon_cmdbuf *cs = sctx->gfx_cs;
|
||||||
unsigned idx;
|
unsigned idx;
|
||||||
unsigned layout_multi = regs->layout & SI_PC_MULTI_MASK;
|
unsigned layout_multi = regs->layout & SI_PC_MULTI_MASK;
|
||||||
@@ -596,12 +709,10 @@ static void si_pc_emit_stop(struct si_context *sctx,
|
|||||||
}
|
}
|
||||||
|
|
||||||
static void si_pc_emit_read(struct si_context *sctx,
|
static void si_pc_emit_read(struct si_context *sctx,
|
||||||
struct si_perfcounter_block *group,
|
struct si_pc_block *block,
|
||||||
unsigned count, unsigned *selectors,
|
unsigned count, uint64_t va)
|
||||||
struct r600_resource *buffer, uint64_t va)
|
|
||||||
{
|
{
|
||||||
struct si_pc_block *sigroup = (struct si_pc_block *)group->data;
|
struct si_pc_block_base *regs = block->b->b;
|
||||||
struct si_pc_block_base *regs = sigroup->b;
|
|
||||||
struct radeon_cmdbuf *cs = sctx->gfx_cs;
|
struct radeon_cmdbuf *cs = sctx->gfx_cs;
|
||||||
unsigned idx;
|
unsigned idx;
|
||||||
unsigned reg = regs->counter0_lo;
|
unsigned reg = regs->counter0_lo;
|
||||||
@@ -641,16 +752,499 @@ static void si_pc_emit_read(struct si_context *sctx,
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static void si_pc_cleanup(struct si_screen *sscreen)
|
static void si_pc_query_destroy(struct si_screen *sscreen,
|
||||||
|
struct si_query *rquery)
|
||||||
{
|
{
|
||||||
si_perfcounters_do_destroy(sscreen->perfcounters);
|
struct si_query_pc *query = (struct si_query_pc *)rquery;
|
||||||
sscreen->perfcounters = NULL;
|
|
||||||
|
while (query->groups) {
|
||||||
|
struct si_query_group *group = query->groups;
|
||||||
|
query->groups = group->next;
|
||||||
|
FREE(group);
|
||||||
|
}
|
||||||
|
|
||||||
|
FREE(query->counters);
|
||||||
|
|
||||||
|
si_query_hw_destroy(sscreen, rquery);
|
||||||
|
}
|
||||||
|
|
||||||
|
static bool si_pc_query_prepare_buffer(struct si_screen *screen,
|
||||||
|
struct si_query_hw *hwquery,
|
||||||
|
struct r600_resource *buffer)
|
||||||
|
{
|
||||||
|
/* no-op */
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
static void si_pc_query_emit_start(struct si_context *sctx,
|
||||||
|
struct si_query_hw *hwquery,
|
||||||
|
struct r600_resource *buffer, uint64_t va)
|
||||||
|
{
|
||||||
|
struct si_query_pc *query = (struct si_query_pc *)hwquery;
|
||||||
|
struct si_query_group *group;
|
||||||
|
int current_se = -1;
|
||||||
|
int current_instance = -1;
|
||||||
|
|
||||||
|
if (query->shaders)
|
||||||
|
si_pc_emit_shaders(sctx, query->shaders);
|
||||||
|
|
||||||
|
for (group = query->groups; group; group = group->next) {
|
||||||
|
struct si_pc_block *block = group->block;
|
||||||
|
|
||||||
|
if (group->se != current_se || group->instance != current_instance) {
|
||||||
|
current_se = group->se;
|
||||||
|
current_instance = group->instance;
|
||||||
|
si_pc_emit_instance(sctx, group->se, group->instance);
|
||||||
|
}
|
||||||
|
|
||||||
|
si_pc_emit_select(sctx, block, group->num_counters, group->selectors);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (current_se != -1 || current_instance != -1)
|
||||||
|
si_pc_emit_instance(sctx, -1, -1);
|
||||||
|
|
||||||
|
si_pc_emit_start(sctx, buffer, va);
|
||||||
|
}
|
||||||
|
|
||||||
|
static void si_pc_query_emit_stop(struct si_context *sctx,
|
||||||
|
struct si_query_hw *hwquery,
|
||||||
|
struct r600_resource *buffer, uint64_t va)
|
||||||
|
{
|
||||||
|
struct si_query_pc *query = (struct si_query_pc *)hwquery;
|
||||||
|
struct si_query_group *group;
|
||||||
|
|
||||||
|
si_pc_emit_stop(sctx, buffer, va);
|
||||||
|
|
||||||
|
for (group = query->groups; group; group = group->next) {
|
||||||
|
struct si_pc_block *block = group->block;
|
||||||
|
unsigned se = group->se >= 0 ? group->se : 0;
|
||||||
|
unsigned se_end = se + 1;
|
||||||
|
|
||||||
|
if ((block->b->b->flags & SI_PC_BLOCK_SE) && (group->se < 0))
|
||||||
|
se_end = sctx->screen->info.max_se;
|
||||||
|
|
||||||
|
do {
|
||||||
|
unsigned instance = group->instance >= 0 ? group->instance : 0;
|
||||||
|
|
||||||
|
do {
|
||||||
|
si_pc_emit_instance(sctx, se, instance);
|
||||||
|
si_pc_emit_read(sctx, block, group->num_counters, va);
|
||||||
|
va += sizeof(uint64_t) * group->num_counters;
|
||||||
|
} while (group->instance < 0 && ++instance < block->num_instances);
|
||||||
|
} while (++se < se_end);
|
||||||
|
}
|
||||||
|
|
||||||
|
si_pc_emit_instance(sctx, -1, -1);
|
||||||
|
}
|
||||||
|
|
||||||
|
static void si_pc_query_clear_result(struct si_query_hw *hwquery,
|
||||||
|
union pipe_query_result *result)
|
||||||
|
{
|
||||||
|
struct si_query_pc *query = (struct si_query_pc *)hwquery;
|
||||||
|
|
||||||
|
memset(result, 0, sizeof(result->batch[0]) * query->num_counters);
|
||||||
|
}
|
||||||
|
|
||||||
|
static void si_pc_query_add_result(struct si_screen *screen,
|
||||||
|
struct si_query_hw *hwquery,
|
||||||
|
void *buffer,
|
||||||
|
union pipe_query_result *result)
|
||||||
|
{
|
||||||
|
struct si_query_pc *query = (struct si_query_pc *)hwquery;
|
||||||
|
uint64_t *results = buffer;
|
||||||
|
unsigned i, j;
|
||||||
|
|
||||||
|
for (i = 0; i < query->num_counters; ++i) {
|
||||||
|
struct si_query_counter *counter = &query->counters[i];
|
||||||
|
|
||||||
|
for (j = 0; j < counter->qwords; ++j) {
|
||||||
|
uint32_t value = results[counter->base + j * counter->stride];
|
||||||
|
result->batch[i].u64 += value;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
static struct si_query_ops batch_query_ops = {
|
||||||
|
.destroy = si_pc_query_destroy,
|
||||||
|
.begin = si_query_hw_begin,
|
||||||
|
.end = si_query_hw_end,
|
||||||
|
.get_result = si_query_hw_get_result
|
||||||
|
};
|
||||||
|
|
||||||
|
static struct si_query_hw_ops batch_query_hw_ops = {
|
||||||
|
.prepare_buffer = si_pc_query_prepare_buffer,
|
||||||
|
.emit_start = si_pc_query_emit_start,
|
||||||
|
.emit_stop = si_pc_query_emit_stop,
|
||||||
|
.clear_result = si_pc_query_clear_result,
|
||||||
|
.add_result = si_pc_query_add_result,
|
||||||
|
};
|
||||||
|
|
||||||
|
static struct si_query_group *get_group_state(struct si_screen *screen,
|
||||||
|
struct si_query_pc *query,
|
||||||
|
struct si_pc_block *block,
|
||||||
|
unsigned sub_gid)
|
||||||
|
{
|
||||||
|
struct si_query_group *group = query->groups;
|
||||||
|
|
||||||
|
while (group) {
|
||||||
|
if (group->block == block && group->sub_gid == sub_gid)
|
||||||
|
return group;
|
||||||
|
group = group->next;
|
||||||
|
}
|
||||||
|
|
||||||
|
group = CALLOC_STRUCT(si_query_group);
|
||||||
|
if (!group)
|
||||||
|
return NULL;
|
||||||
|
|
||||||
|
group->block = block;
|
||||||
|
group->sub_gid = sub_gid;
|
||||||
|
|
||||||
|
if (block->b->b->flags & SI_PC_BLOCK_SHADER) {
|
||||||
|
unsigned sub_gids = block->num_instances;
|
||||||
|
unsigned shader_id;
|
||||||
|
unsigned shaders;
|
||||||
|
unsigned query_shaders;
|
||||||
|
|
||||||
|
if (si_pc_block_has_per_se_groups(screen->perfcounters, block))
|
||||||
|
sub_gids = sub_gids * screen->info.max_se;
|
||||||
|
shader_id = sub_gid / sub_gids;
|
||||||
|
sub_gid = sub_gid % sub_gids;
|
||||||
|
|
||||||
|
shaders = si_pc_shader_type_bits[shader_id];
|
||||||
|
|
||||||
|
query_shaders = query->shaders & ~SI_PC_SHADERS_WINDOWING;
|
||||||
|
if (query_shaders && query_shaders != shaders) {
|
||||||
|
fprintf(stderr, "si_perfcounter: incompatible shader groups\n");
|
||||||
|
FREE(group);
|
||||||
|
return NULL;
|
||||||
|
}
|
||||||
|
query->shaders = shaders;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (block->b->b->flags & SI_PC_BLOCK_SHADER_WINDOWED && !query->shaders) {
|
||||||
|
// A non-zero value in query->shaders ensures that the shader
|
||||||
|
// masking is reset unless the user explicitly requests one.
|
||||||
|
query->shaders = SI_PC_SHADERS_WINDOWING;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (si_pc_block_has_per_se_groups(screen->perfcounters, block)) {
|
||||||
|
group->se = sub_gid / block->num_instances;
|
||||||
|
sub_gid = sub_gid % block->num_instances;
|
||||||
|
} else {
|
||||||
|
group->se = -1;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (si_pc_block_has_per_instance_groups(screen->perfcounters, block)) {
|
||||||
|
group->instance = sub_gid;
|
||||||
|
} else {
|
||||||
|
group->instance = -1;
|
||||||
|
}
|
||||||
|
|
||||||
|
group->next = query->groups;
|
||||||
|
query->groups = group;
|
||||||
|
|
||||||
|
return group;
|
||||||
|
}
|
||||||
|
|
||||||
|
struct pipe_query *si_create_batch_query(struct pipe_context *ctx,
|
||||||
|
unsigned num_queries,
|
||||||
|
unsigned *query_types)
|
||||||
|
{
|
||||||
|
struct si_screen *screen =
|
||||||
|
(struct si_screen *)ctx->screen;
|
||||||
|
struct si_perfcounters *pc = screen->perfcounters;
|
||||||
|
struct si_pc_block *block;
|
||||||
|
struct si_query_group *group;
|
||||||
|
struct si_query_pc *query;
|
||||||
|
unsigned base_gid, sub_gid, sub_index;
|
||||||
|
unsigned i, j;
|
||||||
|
|
||||||
|
if (!pc)
|
||||||
|
return NULL;
|
||||||
|
|
||||||
|
query = CALLOC_STRUCT(si_query_pc);
|
||||||
|
if (!query)
|
||||||
|
return NULL;
|
||||||
|
|
||||||
|
query->b.b.ops = &batch_query_ops;
|
||||||
|
query->b.ops = &batch_query_hw_ops;
|
||||||
|
|
||||||
|
query->num_counters = num_queries;
|
||||||
|
|
||||||
|
/* Collect selectors per group */
|
||||||
|
for (i = 0; i < num_queries; ++i) {
|
||||||
|
unsigned sub_gid;
|
||||||
|
|
||||||
|
if (query_types[i] < SI_QUERY_FIRST_PERFCOUNTER)
|
||||||
|
goto error;
|
||||||
|
|
||||||
|
block = lookup_counter(pc, query_types[i] - SI_QUERY_FIRST_PERFCOUNTER,
|
||||||
|
&base_gid, &sub_index);
|
||||||
|
if (!block)
|
||||||
|
goto error;
|
||||||
|
|
||||||
|
sub_gid = sub_index / block->b->selectors;
|
||||||
|
sub_index = sub_index % block->b->selectors;
|
||||||
|
|
||||||
|
group = get_group_state(screen, query, block, sub_gid);
|
||||||
|
if (!group)
|
||||||
|
goto error;
|
||||||
|
|
||||||
|
if (group->num_counters >= block->b->b->num_counters) {
|
||||||
|
fprintf(stderr,
|
||||||
|
"perfcounter group %s: too many selected\n",
|
||||||
|
block->b->b->name);
|
||||||
|
goto error;
|
||||||
|
}
|
||||||
|
group->selectors[group->num_counters] = sub_index;
|
||||||
|
++group->num_counters;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Compute result bases and CS size per group */
|
||||||
|
query->b.num_cs_dw_end = pc->num_stop_cs_dwords;
|
||||||
|
query->b.num_cs_dw_end += pc->num_instance_cs_dwords;
|
||||||
|
|
||||||
|
i = 0;
|
||||||
|
for (group = query->groups; group; group = group->next) {
|
||||||
|
struct si_pc_block *block = group->block;
|
||||||
|
unsigned read_dw;
|
||||||
|
unsigned instances = 1;
|
||||||
|
|
||||||
|
if ((block->b->b->flags & SI_PC_BLOCK_SE) && group->se < 0)
|
||||||
|
instances = screen->info.max_se;
|
||||||
|
if (group->instance < 0)
|
||||||
|
instances *= block->num_instances;
|
||||||
|
|
||||||
|
group->result_base = i;
|
||||||
|
query->b.result_size += sizeof(uint64_t) * instances * group->num_counters;
|
||||||
|
i += instances * group->num_counters;
|
||||||
|
|
||||||
|
read_dw = 6 * group->num_counters;
|
||||||
|
query->b.num_cs_dw_end += instances * read_dw;
|
||||||
|
query->b.num_cs_dw_end += instances * pc->num_instance_cs_dwords;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (query->shaders) {
|
||||||
|
if (query->shaders == SI_PC_SHADERS_WINDOWING)
|
||||||
|
query->shaders = 0xffffffff;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Map user-supplied query array to result indices */
|
||||||
|
query->counters = CALLOC(num_queries, sizeof(*query->counters));
|
||||||
|
for (i = 0; i < num_queries; ++i) {
|
||||||
|
struct si_query_counter *counter = &query->counters[i];
|
||||||
|
struct si_pc_block *block;
|
||||||
|
|
||||||
|
block = lookup_counter(pc, query_types[i] - SI_QUERY_FIRST_PERFCOUNTER,
|
||||||
|
&base_gid, &sub_index);
|
||||||
|
|
||||||
|
sub_gid = sub_index / block->b->selectors;
|
||||||
|
sub_index = sub_index % block->b->selectors;
|
||||||
|
|
||||||
|
group = get_group_state(screen, query, block, sub_gid);
|
||||||
|
assert(group != NULL);
|
||||||
|
|
||||||
|
for (j = 0; j < group->num_counters; ++j) {
|
||||||
|
if (group->selectors[j] == sub_index)
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
counter->base = group->result_base + j;
|
||||||
|
counter->stride = group->num_counters;
|
||||||
|
|
||||||
|
counter->qwords = 1;
|
||||||
|
if ((block->b->b->flags & SI_PC_BLOCK_SE) && group->se < 0)
|
||||||
|
counter->qwords = screen->info.max_se;
|
||||||
|
if (group->instance < 0)
|
||||||
|
counter->qwords *= block->num_instances;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!si_query_hw_init(screen, &query->b))
|
||||||
|
goto error;
|
||||||
|
|
||||||
|
return (struct pipe_query *)query;
|
||||||
|
|
||||||
|
error:
|
||||||
|
si_pc_query_destroy(screen, &query->b.b);
|
||||||
|
return NULL;
|
||||||
|
}
|
||||||
|
|
||||||
|
static bool si_init_block_names(struct si_screen *screen,
|
||||||
|
struct si_pc_block *block)
|
||||||
|
{
|
||||||
|
bool per_instance_groups = si_pc_block_has_per_instance_groups(screen->perfcounters, block);
|
||||||
|
bool per_se_groups = si_pc_block_has_per_se_groups(screen->perfcounters, block);
|
||||||
|
unsigned i, j, k;
|
||||||
|
unsigned groups_shader = 1, groups_se = 1, groups_instance = 1;
|
||||||
|
unsigned namelen;
|
||||||
|
char *groupname;
|
||||||
|
char *p;
|
||||||
|
|
||||||
|
if (per_instance_groups)
|
||||||
|
groups_instance = block->num_instances;
|
||||||
|
if (per_se_groups)
|
||||||
|
groups_se = screen->info.max_se;
|
||||||
|
if (block->b->b->flags & SI_PC_BLOCK_SHADER)
|
||||||
|
groups_shader = ARRAY_SIZE(si_pc_shader_type_bits);
|
||||||
|
|
||||||
|
namelen = strlen(block->b->b->name);
|
||||||
|
block->group_name_stride = namelen + 1;
|
||||||
|
if (block->b->b->flags & SI_PC_BLOCK_SHADER)
|
||||||
|
block->group_name_stride += 3;
|
||||||
|
if (per_se_groups) {
|
||||||
|
assert(groups_se <= 10);
|
||||||
|
block->group_name_stride += 1;
|
||||||
|
|
||||||
|
if (per_instance_groups)
|
||||||
|
block->group_name_stride += 1;
|
||||||
|
}
|
||||||
|
if (per_instance_groups) {
|
||||||
|
assert(groups_instance <= 100);
|
||||||
|
block->group_name_stride += 2;
|
||||||
|
}
|
||||||
|
|
||||||
|
block->group_names = MALLOC(block->num_groups * block->group_name_stride);
|
||||||
|
if (!block->group_names)
|
||||||
|
return false;
|
||||||
|
|
||||||
|
groupname = block->group_names;
|
||||||
|
for (i = 0; i < groups_shader; ++i) {
|
||||||
|
const char *shader_suffix = si_pc_shader_type_suffixes[i];
|
||||||
|
unsigned shaderlen = strlen(shader_suffix);
|
||||||
|
for (j = 0; j < groups_se; ++j) {
|
||||||
|
for (k = 0; k < groups_instance; ++k) {
|
||||||
|
strcpy(groupname, block->b->b->name);
|
||||||
|
p = groupname + namelen;
|
||||||
|
|
||||||
|
if (block->b->b->flags & SI_PC_BLOCK_SHADER) {
|
||||||
|
strcpy(p, shader_suffix);
|
||||||
|
p += shaderlen;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (per_se_groups) {
|
||||||
|
p += sprintf(p, "%d", j);
|
||||||
|
if (per_instance_groups)
|
||||||
|
*p++ = '_';
|
||||||
|
}
|
||||||
|
|
||||||
|
if (per_instance_groups)
|
||||||
|
p += sprintf(p, "%d", k);
|
||||||
|
|
||||||
|
groupname += block->group_name_stride;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
assert(block->b->selectors <= 1000);
|
||||||
|
block->selector_name_stride = block->group_name_stride + 4;
|
||||||
|
block->selector_names = MALLOC(block->num_groups * block->b->selectors *
|
||||||
|
block->selector_name_stride);
|
||||||
|
if (!block->selector_names)
|
||||||
|
return false;
|
||||||
|
|
||||||
|
groupname = block->group_names;
|
||||||
|
p = block->selector_names;
|
||||||
|
for (i = 0; i < block->num_groups; ++i) {
|
||||||
|
for (j = 0; j < block->b->selectors; ++j) {
|
||||||
|
sprintf(p, "%s_%03d", groupname, j);
|
||||||
|
p += block->selector_name_stride;
|
||||||
|
}
|
||||||
|
groupname += block->group_name_stride;
|
||||||
|
}
|
||||||
|
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
int si_get_perfcounter_info(struct si_screen *screen,
|
||||||
|
unsigned index,
|
||||||
|
struct pipe_driver_query_info *info)
|
||||||
|
{
|
||||||
|
struct si_perfcounters *pc = screen->perfcounters;
|
||||||
|
struct si_pc_block *block;
|
||||||
|
unsigned base_gid, sub;
|
||||||
|
|
||||||
|
if (!pc)
|
||||||
|
return 0;
|
||||||
|
|
||||||
|
if (!info) {
|
||||||
|
unsigned bid, num_queries = 0;
|
||||||
|
|
||||||
|
for (bid = 0; bid < pc->num_blocks; ++bid) {
|
||||||
|
num_queries += pc->blocks[bid].b->selectors *
|
||||||
|
pc->blocks[bid].num_groups;
|
||||||
|
}
|
||||||
|
|
||||||
|
return num_queries;
|
||||||
|
}
|
||||||
|
|
||||||
|
block = lookup_counter(pc, index, &base_gid, &sub);
|
||||||
|
if (!block)
|
||||||
|
return 0;
|
||||||
|
|
||||||
|
if (!block->selector_names) {
|
||||||
|
if (!si_init_block_names(screen, block))
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
info->name = block->selector_names + sub * block->selector_name_stride;
|
||||||
|
info->query_type = SI_QUERY_FIRST_PERFCOUNTER + index;
|
||||||
|
info->max_value.u64 = 0;
|
||||||
|
info->type = PIPE_DRIVER_QUERY_TYPE_UINT64;
|
||||||
|
info->result_type = PIPE_DRIVER_QUERY_RESULT_TYPE_AVERAGE;
|
||||||
|
info->group_id = base_gid + sub / block->b->selectors;
|
||||||
|
info->flags = PIPE_DRIVER_QUERY_FLAG_BATCH;
|
||||||
|
if (sub > 0 && sub + 1 < block->b->selectors * block->num_groups)
|
||||||
|
info->flags |= PIPE_DRIVER_QUERY_FLAG_DONT_LIST;
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
int si_get_perfcounter_group_info(struct si_screen *screen,
|
||||||
|
unsigned index,
|
||||||
|
struct pipe_driver_query_group_info *info)
|
||||||
|
{
|
||||||
|
struct si_perfcounters *pc = screen->perfcounters;
|
||||||
|
struct si_pc_block *block;
|
||||||
|
|
||||||
|
if (!pc)
|
||||||
|
return 0;
|
||||||
|
|
||||||
|
if (!info)
|
||||||
|
return pc->num_groups;
|
||||||
|
|
||||||
|
block = lookup_group(pc, &index);
|
||||||
|
if (!block)
|
||||||
|
return 0;
|
||||||
|
|
||||||
|
if (!block->group_names) {
|
||||||
|
if (!si_init_block_names(screen, block))
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
info->name = block->group_names + index * block->group_name_stride;
|
||||||
|
info->num_queries = block->b->selectors;
|
||||||
|
info->max_active_queries = block->b->b->num_counters;
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
void si_destroy_perfcounters(struct si_screen *screen)
|
||||||
|
{
|
||||||
|
struct si_perfcounters *pc = screen->perfcounters;
|
||||||
|
unsigned i;
|
||||||
|
|
||||||
|
if (!pc)
|
||||||
|
return;
|
||||||
|
|
||||||
|
for (i = 0; i < pc->num_blocks; ++i) {
|
||||||
|
FREE(pc->blocks[i].group_names);
|
||||||
|
FREE(pc->blocks[i].selector_names);
|
||||||
|
}
|
||||||
|
FREE(pc->blocks);
|
||||||
|
FREE(pc);
|
||||||
|
screen->perfcounters = NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
void si_init_perfcounters(struct si_screen *screen)
|
void si_init_perfcounters(struct si_screen *screen)
|
||||||
{
|
{
|
||||||
struct si_perfcounters *pc;
|
struct si_perfcounters *pc;
|
||||||
struct si_pc_block *blocks;
|
const struct si_pc_block_gfxdescr *blocks;
|
||||||
unsigned num_blocks;
|
unsigned num_blocks;
|
||||||
unsigned i;
|
unsigned i;
|
||||||
|
|
||||||
@@ -679,52 +1273,50 @@ void si_init_perfcounters(struct si_screen *screen)
|
|||||||
screen->info.max_sh_per_se);
|
screen->info.max_sh_per_se);
|
||||||
}
|
}
|
||||||
|
|
||||||
pc = CALLOC_STRUCT(si_perfcounters);
|
screen->perfcounters = pc = CALLOC_STRUCT(si_perfcounters);
|
||||||
if (!pc)
|
if (!pc)
|
||||||
return;
|
return;
|
||||||
|
|
||||||
pc->num_stop_cs_dwords = 14 + si_cp_write_fence_dwords(screen);
|
pc->num_stop_cs_dwords = 14 + si_cp_write_fence_dwords(screen);
|
||||||
pc->num_instance_cs_dwords = 3;
|
pc->num_instance_cs_dwords = 3;
|
||||||
|
|
||||||
pc->num_shader_types = ARRAY_SIZE(si_pc_shader_type_bits);
|
pc->separate_se = debug_get_bool_option("RADEON_PC_SEPARATE_SE", false);
|
||||||
pc->shader_type_suffixes = si_pc_shader_type_suffixes;
|
pc->separate_instance = debug_get_bool_option("RADEON_PC_SEPARATE_INSTANCE", false);
|
||||||
pc->shader_type_bits = si_pc_shader_type_bits;
|
|
||||||
|
|
||||||
pc->emit_instance = si_pc_emit_instance;
|
pc->blocks = CALLOC(num_blocks, sizeof(struct si_pc_block));
|
||||||
pc->emit_shaders = si_pc_emit_shaders;
|
if (!pc->blocks)
|
||||||
pc->emit_select = si_pc_emit_select;
|
|
||||||
pc->emit_start = si_pc_emit_start;
|
|
||||||
pc->emit_stop = si_pc_emit_stop;
|
|
||||||
pc->emit_read = si_pc_emit_read;
|
|
||||||
pc->cleanup = si_pc_cleanup;
|
|
||||||
|
|
||||||
if (!si_perfcounters_init(pc, num_blocks))
|
|
||||||
goto error;
|
goto error;
|
||||||
|
pc->num_blocks = num_blocks;
|
||||||
|
|
||||||
for (i = 0; i < num_blocks; ++i) {
|
for (i = 0; i < num_blocks; ++i) {
|
||||||
struct si_pc_block *block = &blocks[i];
|
struct si_pc_block *block = &pc->blocks[i];
|
||||||
unsigned instances = block->instances;
|
block->b = &blocks[i];
|
||||||
|
block->num_instances = block->b->instances;
|
||||||
|
|
||||||
if (!strcmp(block->b->name, "CB") ||
|
if (!strcmp(block->b->b->name, "CB") ||
|
||||||
!strcmp(block->b->name, "DB"))
|
!strcmp(block->b->b->name, "DB"))
|
||||||
instances = screen->info.max_se;
|
block->num_instances = screen->info.max_se;
|
||||||
else if (!strcmp(block->b->name, "TCC"))
|
else if (!strcmp(block->b->b->name, "TCC"))
|
||||||
instances = screen->info.num_tcc_blocks;
|
block->num_instances = screen->info.num_tcc_blocks;
|
||||||
else if (!strcmp(block->b->name, "IA"))
|
else if (!strcmp(block->b->b->name, "IA"))
|
||||||
instances = MAX2(1, screen->info.max_se / 2);
|
block->num_instances = MAX2(1, screen->info.max_se / 2);
|
||||||
|
|
||||||
si_perfcounters_add_block(screen, pc,
|
if (si_pc_block_has_per_instance_groups(pc, block)) {
|
||||||
block->b->name,
|
block->num_groups = block->num_instances;
|
||||||
block->b->flags,
|
} else {
|
||||||
block->b->num_counters,
|
block->num_groups = 1;
|
||||||
block->selectors,
|
}
|
||||||
instances,
|
|
||||||
block);
|
if (si_pc_block_has_per_se_groups(pc, block))
|
||||||
|
block->num_groups *= screen->info.max_se;
|
||||||
|
if (block->b->b->flags & SI_PC_BLOCK_SHADER)
|
||||||
|
block->num_groups *= ARRAY_SIZE(si_pc_shader_type_bits);
|
||||||
|
|
||||||
|
pc->num_groups += block->num_groups;
|
||||||
}
|
}
|
||||||
|
|
||||||
screen->perfcounters = pc;
|
|
||||||
return;
|
return;
|
||||||
|
|
||||||
error:
|
error:
|
||||||
si_perfcounters_do_destroy(pc);
|
si_destroy_perfcounters(screen);
|
||||||
}
|
}
|
||||||
|
@@ -705,7 +705,7 @@ static void si_destroy_screen(struct pipe_screen* pscreen)
|
|||||||
mtx_destroy(&sscreen->shader_parts_mutex);
|
mtx_destroy(&sscreen->shader_parts_mutex);
|
||||||
si_destroy_shader_cache(sscreen);
|
si_destroy_shader_cache(sscreen);
|
||||||
|
|
||||||
si_perfcounters_destroy(sscreen);
|
si_destroy_perfcounters(sscreen);
|
||||||
si_gpu_load_kill_thread(sscreen);
|
si_gpu_load_kill_thread(sscreen);
|
||||||
|
|
||||||
mtx_destroy(&sscreen->gpu_load_mutex);
|
mtx_destroy(&sscreen->gpu_load_mutex);
|
||||||
|
@@ -1236,11 +1236,9 @@ unsigned si_end_counter(struct si_screen *sscreen, unsigned type,
|
|||||||
/* si_compute.c */
|
/* si_compute.c */
|
||||||
void si_init_compute_functions(struct si_context *sctx);
|
void si_init_compute_functions(struct si_context *sctx);
|
||||||
|
|
||||||
/* r600_perfcounters.c */
|
|
||||||
void si_perfcounters_destroy(struct si_screen *sscreen);
|
|
||||||
|
|
||||||
/* si_perfcounters.c */
|
/* si_perfcounters.c */
|
||||||
void si_init_perfcounters(struct si_screen *screen);
|
void si_init_perfcounters(struct si_screen *screen);
|
||||||
|
void si_destroy_perfcounters(struct si_screen *screen);
|
||||||
|
|
||||||
/* si_pipe.c */
|
/* si_pipe.c */
|
||||||
bool si_check_device_reset(struct si_context *sctx);
|
bool si_check_device_reset(struct si_context *sctx);
|
||||||
|
@@ -213,79 +213,14 @@ bool si_query_hw_get_result(struct si_context *sctx,
|
|||||||
union pipe_query_result *result);
|
union pipe_query_result *result);
|
||||||
|
|
||||||
/* Performance counters */
|
/* Performance counters */
|
||||||
enum {
|
|
||||||
/* This block is part of the shader engine */
|
|
||||||
SI_PC_BLOCK_SE = (1 << 0),
|
|
||||||
|
|
||||||
/* Expose per-instance groups instead of summing all instances (within
|
|
||||||
* an SE). */
|
|
||||||
SI_PC_BLOCK_INSTANCE_GROUPS = (1 << 1),
|
|
||||||
|
|
||||||
/* Expose per-SE groups instead of summing instances across SEs. */
|
|
||||||
SI_PC_BLOCK_SE_GROUPS = (1 << 2),
|
|
||||||
|
|
||||||
/* Shader block */
|
|
||||||
SI_PC_BLOCK_SHADER = (1 << 3),
|
|
||||||
|
|
||||||
/* Non-shader block with perfcounters windowed by shaders. */
|
|
||||||
SI_PC_BLOCK_SHADER_WINDOWED = (1 << 4),
|
|
||||||
};
|
|
||||||
|
|
||||||
/* Describes a hardware block with performance counters. Multiple instances of
|
|
||||||
* each block, possibly per-SE, may exist on the chip. Depending on the block
|
|
||||||
* and on the user's configuration, we either
|
|
||||||
* (a) expose every instance as a performance counter group,
|
|
||||||
* (b) expose a single performance counter group that reports the sum over all
|
|
||||||
* instances, or
|
|
||||||
* (c) expose one performance counter group per instance, but summed over all
|
|
||||||
* shader engines.
|
|
||||||
*/
|
|
||||||
struct si_perfcounter_block {
|
|
||||||
const char *basename;
|
|
||||||
unsigned flags;
|
|
||||||
unsigned num_counters;
|
|
||||||
unsigned num_selectors;
|
|
||||||
unsigned num_instances;
|
|
||||||
|
|
||||||
unsigned num_groups;
|
|
||||||
char *group_names;
|
|
||||||
unsigned group_name_stride;
|
|
||||||
|
|
||||||
char *selector_names;
|
|
||||||
unsigned selector_name_stride;
|
|
||||||
|
|
||||||
void *data;
|
|
||||||
};
|
|
||||||
|
|
||||||
struct si_perfcounters {
|
struct si_perfcounters {
|
||||||
unsigned num_groups;
|
unsigned num_groups;
|
||||||
unsigned num_blocks;
|
unsigned num_blocks;
|
||||||
struct si_perfcounter_block *blocks;
|
struct si_pc_block *blocks;
|
||||||
|
|
||||||
unsigned num_stop_cs_dwords;
|
unsigned num_stop_cs_dwords;
|
||||||
unsigned num_instance_cs_dwords;
|
unsigned num_instance_cs_dwords;
|
||||||
|
|
||||||
unsigned num_shader_types;
|
|
||||||
const char * const *shader_type_suffixes;
|
|
||||||
const unsigned *shader_type_bits;
|
|
||||||
|
|
||||||
void (*emit_instance)(struct si_context *,
|
|
||||||
int se, int instance);
|
|
||||||
void (*emit_shaders)(struct si_context *, unsigned shaders);
|
|
||||||
void (*emit_select)(struct si_context *,
|
|
||||||
struct si_perfcounter_block *,
|
|
||||||
unsigned count, unsigned *selectors);
|
|
||||||
void (*emit_start)(struct si_context *,
|
|
||||||
struct r600_resource *buffer, uint64_t va);
|
|
||||||
void (*emit_stop)(struct si_context *,
|
|
||||||
struct r600_resource *buffer, uint64_t va);
|
|
||||||
void (*emit_read)(struct si_context *,
|
|
||||||
struct si_perfcounter_block *,
|
|
||||||
unsigned count, unsigned *selectors,
|
|
||||||
struct r600_resource *buffer, uint64_t va);
|
|
||||||
|
|
||||||
void (*cleanup)(struct si_screen *);
|
|
||||||
|
|
||||||
bool separate_se;
|
bool separate_se;
|
||||||
bool separate_instance;
|
bool separate_instance;
|
||||||
};
|
};
|
||||||
@@ -301,13 +236,6 @@ int si_get_perfcounter_group_info(struct si_screen *,
|
|||||||
unsigned index,
|
unsigned index,
|
||||||
struct pipe_driver_query_group_info *info);
|
struct pipe_driver_query_group_info *info);
|
||||||
|
|
||||||
bool si_perfcounters_init(struct si_perfcounters *, unsigned num_blocks);
|
|
||||||
void si_perfcounters_add_block(struct si_screen *,
|
|
||||||
struct si_perfcounters *,
|
|
||||||
const char *name, unsigned flags,
|
|
||||||
unsigned counters, unsigned selectors,
|
|
||||||
unsigned instances, void *data);
|
|
||||||
void si_perfcounters_do_destroy(struct si_perfcounters *);
|
|
||||||
void si_query_hw_reset_buffers(struct si_context *sctx,
|
void si_query_hw_reset_buffers(struct si_context *sctx,
|
||||||
struct si_query_hw *query);
|
struct si_query_hw *query);
|
||||||
|
|
||||||
|
Reference in New Issue
Block a user