
This has proven to be incredibly useful for debugging CMA allocation failures and driving memory management improvements. However, we don't want to burden entry and exit from the BO cache with the labeling ioctl's overhead on release builds.
1229 lines
48 KiB
C
1229 lines
48 KiB
C
/*
|
|
* Copyright © 2014 Broadcom
|
|
* Copyright (C) 2012 Rob Clark <robclark@freedesktop.org>
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice (including the next
|
|
* paragraph) shall be included in all copies or substantial portions of the
|
|
* Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
|
|
* IN THE SOFTWARE.
|
|
*/
|
|
|
|
#include "util/u_blit.h"
|
|
#include "util/u_memory.h"
|
|
#include "util/u_format.h"
|
|
#include "util/u_inlines.h"
|
|
#include "util/u_surface.h"
|
|
#include "util/u_upload_mgr.h"
|
|
|
|
#include "drm_fourcc.h"
|
|
#include "vc4_drm.h"
|
|
#include "vc4_screen.h"
|
|
#include "vc4_context.h"
|
|
#include "vc4_resource.h"
|
|
#include "vc4_tiling.h"
|
|
|
|
#ifndef DRM_FORMAT_MOD_INVALID
|
|
#define DRM_FORMAT_MOD_INVALID ((1ULL << 56) - 1)
|
|
#endif
|
|
|
|
static bool
|
|
vc4_resource_bo_alloc(struct vc4_resource *rsc)
|
|
{
|
|
struct pipe_resource *prsc = &rsc->base;
|
|
struct pipe_screen *pscreen = prsc->screen;
|
|
struct vc4_bo *bo;
|
|
|
|
if (vc4_debug & VC4_DEBUG_SURFACE) {
|
|
fprintf(stderr, "alloc %p: size %d + offset %d -> %d\n",
|
|
rsc,
|
|
rsc->slices[0].size,
|
|
rsc->slices[0].offset,
|
|
rsc->slices[0].offset +
|
|
rsc->slices[0].size +
|
|
rsc->cube_map_stride * (prsc->array_size - 1));
|
|
}
|
|
|
|
bo = vc4_bo_alloc(vc4_screen(pscreen),
|
|
rsc->slices[0].offset +
|
|
rsc->slices[0].size +
|
|
rsc->cube_map_stride * (prsc->array_size - 1),
|
|
"resource");
|
|
if (bo) {
|
|
vc4_bo_unreference(&rsc->bo);
|
|
rsc->bo = bo;
|
|
return true;
|
|
} else {
|
|
return false;
|
|
}
|
|
}
|
|
|
|
static void
|
|
vc4_resource_transfer_unmap(struct pipe_context *pctx,
|
|
struct pipe_transfer *ptrans)
|
|
{
|
|
struct vc4_context *vc4 = vc4_context(pctx);
|
|
struct vc4_transfer *trans = vc4_transfer(ptrans);
|
|
|
|
if (trans->map) {
|
|
struct vc4_resource *rsc;
|
|
struct vc4_resource_slice *slice;
|
|
if (trans->ss_resource) {
|
|
rsc = vc4_resource(trans->ss_resource);
|
|
slice = &rsc->slices[0];
|
|
} else {
|
|
rsc = vc4_resource(ptrans->resource);
|
|
slice = &rsc->slices[ptrans->level];
|
|
}
|
|
|
|
if (ptrans->usage & PIPE_TRANSFER_WRITE) {
|
|
vc4_store_tiled_image(rsc->bo->map + slice->offset +
|
|
ptrans->box.z * rsc->cube_map_stride,
|
|
slice->stride,
|
|
trans->map, ptrans->stride,
|
|
slice->tiling, rsc->cpp,
|
|
&ptrans->box);
|
|
}
|
|
free(trans->map);
|
|
}
|
|
|
|
if (trans->ss_resource && (ptrans->usage & PIPE_TRANSFER_WRITE)) {
|
|
struct pipe_blit_info blit;
|
|
memset(&blit, 0, sizeof(blit));
|
|
|
|
blit.src.resource = trans->ss_resource;
|
|
blit.src.format = trans->ss_resource->format;
|
|
blit.src.box.width = trans->ss_box.width;
|
|
blit.src.box.height = trans->ss_box.height;
|
|
blit.src.box.depth = 1;
|
|
|
|
blit.dst.resource = ptrans->resource;
|
|
blit.dst.format = ptrans->resource->format;
|
|
blit.dst.level = ptrans->level;
|
|
blit.dst.box = trans->ss_box;
|
|
|
|
blit.mask = util_format_get_mask(ptrans->resource->format);
|
|
blit.filter = PIPE_TEX_FILTER_NEAREST;
|
|
|
|
pctx->blit(pctx, &blit);
|
|
|
|
pipe_resource_reference(&trans->ss_resource, NULL);
|
|
}
|
|
|
|
pipe_resource_reference(&ptrans->resource, NULL);
|
|
slab_free(&vc4->transfer_pool, ptrans);
|
|
}
|
|
|
|
static struct pipe_resource *
|
|
vc4_get_temp_resource(struct pipe_context *pctx,
|
|
struct pipe_resource *prsc,
|
|
const struct pipe_box *box)
|
|
{
|
|
struct pipe_resource temp_setup;
|
|
|
|
memset(&temp_setup, 0, sizeof(temp_setup));
|
|
temp_setup.target = prsc->target;
|
|
temp_setup.format = prsc->format;
|
|
temp_setup.width0 = box->width;
|
|
temp_setup.height0 = box->height;
|
|
temp_setup.depth0 = 1;
|
|
temp_setup.array_size = 1;
|
|
|
|
return pctx->screen->resource_create(pctx->screen, &temp_setup);
|
|
}
|
|
|
|
static void *
|
|
vc4_resource_transfer_map(struct pipe_context *pctx,
|
|
struct pipe_resource *prsc,
|
|
unsigned level, unsigned usage,
|
|
const struct pipe_box *box,
|
|
struct pipe_transfer **pptrans)
|
|
{
|
|
struct vc4_context *vc4 = vc4_context(pctx);
|
|
struct vc4_resource *rsc = vc4_resource(prsc);
|
|
struct vc4_transfer *trans;
|
|
struct pipe_transfer *ptrans;
|
|
enum pipe_format format = prsc->format;
|
|
char *buf;
|
|
|
|
/* Upgrade DISCARD_RANGE to WHOLE_RESOURCE if the whole resource is
|
|
* being mapped.
|
|
*/
|
|
if ((usage & PIPE_TRANSFER_DISCARD_RANGE) &&
|
|
!(usage & PIPE_TRANSFER_UNSYNCHRONIZED) &&
|
|
!(prsc->flags & PIPE_RESOURCE_FLAG_MAP_COHERENT) &&
|
|
prsc->last_level == 0 &&
|
|
prsc->width0 == box->width &&
|
|
prsc->height0 == box->height &&
|
|
prsc->depth0 == box->depth &&
|
|
prsc->array_size == 1 &&
|
|
rsc->bo->private) {
|
|
usage |= PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE;
|
|
}
|
|
|
|
if (usage & PIPE_TRANSFER_DISCARD_WHOLE_RESOURCE) {
|
|
if (vc4_resource_bo_alloc(rsc)) {
|
|
/* If it might be bound as one of our vertex buffers,
|
|
* make sure we re-emit vertex buffer state.
|
|
*/
|
|
if (prsc->bind & PIPE_BIND_VERTEX_BUFFER)
|
|
vc4->dirty |= VC4_DIRTY_VTXBUF;
|
|
} else {
|
|
/* If we failed to reallocate, flush users so that we
|
|
* don't violate any syncing requirements.
|
|
*/
|
|
vc4_flush_jobs_reading_resource(vc4, prsc);
|
|
}
|
|
} else if (!(usage & PIPE_TRANSFER_UNSYNCHRONIZED)) {
|
|
/* If we're writing and the buffer is being used by the CL, we
|
|
* have to flush the CL first. If we're only reading, we need
|
|
* to flush if the CL has written our buffer.
|
|
*/
|
|
if (usage & PIPE_TRANSFER_WRITE)
|
|
vc4_flush_jobs_reading_resource(vc4, prsc);
|
|
else
|
|
vc4_flush_jobs_writing_resource(vc4, prsc);
|
|
}
|
|
|
|
if (usage & PIPE_TRANSFER_WRITE) {
|
|
rsc->writes++;
|
|
rsc->initialized_buffers = ~0;
|
|
}
|
|
|
|
trans = slab_alloc(&vc4->transfer_pool);
|
|
if (!trans)
|
|
return NULL;
|
|
|
|
/* XXX: Handle DONTBLOCK, DISCARD_RANGE, PERSISTENT, COHERENT. */
|
|
|
|
/* slab_alloc_st() doesn't zero: */
|
|
memset(trans, 0, sizeof(*trans));
|
|
ptrans = &trans->base;
|
|
|
|
pipe_resource_reference(&ptrans->resource, prsc);
|
|
ptrans->level = level;
|
|
ptrans->usage = usage;
|
|
ptrans->box = *box;
|
|
|
|
/* If the resource is multisampled, we need to resolve to single
|
|
* sample. This seems like it should be handled at a higher layer.
|
|
*/
|
|
if (prsc->nr_samples > 1) {
|
|
trans->ss_resource = vc4_get_temp_resource(pctx, prsc, box);
|
|
if (!trans->ss_resource)
|
|
goto fail;
|
|
assert(!trans->ss_resource->nr_samples);
|
|
|
|
/* The ptrans->box gets modified for tile alignment, so save
|
|
* the original box for unmap time.
|
|
*/
|
|
trans->ss_box = *box;
|
|
|
|
if (usage & PIPE_TRANSFER_READ) {
|
|
struct pipe_blit_info blit;
|
|
memset(&blit, 0, sizeof(blit));
|
|
|
|
blit.src.resource = ptrans->resource;
|
|
blit.src.format = ptrans->resource->format;
|
|
blit.src.level = ptrans->level;
|
|
blit.src.box = trans->ss_box;
|
|
|
|
blit.dst.resource = trans->ss_resource;
|
|
blit.dst.format = trans->ss_resource->format;
|
|
blit.dst.box.width = trans->ss_box.width;
|
|
blit.dst.box.height = trans->ss_box.height;
|
|
blit.dst.box.depth = 1;
|
|
|
|
blit.mask = util_format_get_mask(prsc->format);
|
|
blit.filter = PIPE_TEX_FILTER_NEAREST;
|
|
|
|
pctx->blit(pctx, &blit);
|
|
vc4_flush_jobs_writing_resource(vc4, blit.dst.resource);
|
|
}
|
|
|
|
/* The rest of the mapping process should use our temporary. */
|
|
prsc = trans->ss_resource;
|
|
rsc = vc4_resource(prsc);
|
|
ptrans->box.x = 0;
|
|
ptrans->box.y = 0;
|
|
ptrans->box.z = 0;
|
|
}
|
|
|
|
if (usage & PIPE_TRANSFER_UNSYNCHRONIZED)
|
|
buf = vc4_bo_map_unsynchronized(rsc->bo);
|
|
else
|
|
buf = vc4_bo_map(rsc->bo);
|
|
if (!buf) {
|
|
fprintf(stderr, "Failed to map bo\n");
|
|
goto fail;
|
|
}
|
|
|
|
*pptrans = ptrans;
|
|
|
|
struct vc4_resource_slice *slice = &rsc->slices[level];
|
|
if (rsc->tiled) {
|
|
uint32_t utile_w = vc4_utile_width(rsc->cpp);
|
|
uint32_t utile_h = vc4_utile_height(rsc->cpp);
|
|
|
|
/* No direct mappings of tiled, since we need to manually
|
|
* tile/untile.
|
|
*/
|
|
if (usage & PIPE_TRANSFER_MAP_DIRECTLY)
|
|
return NULL;
|
|
|
|
if (format == PIPE_FORMAT_ETC1_RGB8) {
|
|
/* ETC1 is arranged as 64-bit blocks, where each block
|
|
* is 4x4 pixels. Texture tiling operates on the
|
|
* 64-bit block the way it would an uncompressed
|
|
* pixels.
|
|
*/
|
|
assert(!(ptrans->box.x & 3));
|
|
assert(!(ptrans->box.y & 3));
|
|
ptrans->box.x >>= 2;
|
|
ptrans->box.y >>= 2;
|
|
ptrans->box.width = (ptrans->box.width + 3) >> 2;
|
|
ptrans->box.height = (ptrans->box.height + 3) >> 2;
|
|
}
|
|
|
|
/* We need to align the box to utile boundaries, since that's
|
|
* what load/store operates on. This may cause us to need to
|
|
* read out the original contents in that border area. Right
|
|
* now we just read out the entire contents, including the
|
|
* middle area that will just get overwritten.
|
|
*/
|
|
uint32_t box_start_x = ptrans->box.x & (utile_w - 1);
|
|
uint32_t box_start_y = ptrans->box.y & (utile_h - 1);
|
|
bool needs_load = (usage & PIPE_TRANSFER_READ) != 0;
|
|
|
|
if (box_start_x) {
|
|
ptrans->box.width += box_start_x;
|
|
ptrans->box.x -= box_start_x;
|
|
needs_load = true;
|
|
}
|
|
if (box_start_y) {
|
|
ptrans->box.height += box_start_y;
|
|
ptrans->box.y -= box_start_y;
|
|
needs_load = true;
|
|
}
|
|
if (ptrans->box.width & (utile_w - 1)) {
|
|
/* We only need to force a load if our border region
|
|
* we're extending into is actually part of the
|
|
* texture.
|
|
*/
|
|
uint32_t slice_width = u_minify(prsc->width0, level);
|
|
if (ptrans->box.x + ptrans->box.width != slice_width)
|
|
needs_load = true;
|
|
ptrans->box.width = align(ptrans->box.width, utile_w);
|
|
}
|
|
if (ptrans->box.height & (utile_h - 1)) {
|
|
uint32_t slice_height = u_minify(prsc->height0, level);
|
|
if (ptrans->box.y + ptrans->box.height != slice_height)
|
|
needs_load = true;
|
|
ptrans->box.height = align(ptrans->box.height, utile_h);
|
|
}
|
|
|
|
ptrans->stride = ptrans->box.width * rsc->cpp;
|
|
ptrans->layer_stride = ptrans->stride * ptrans->box.height;
|
|
|
|
trans->map = malloc(ptrans->layer_stride * ptrans->box.depth);
|
|
|
|
if (needs_load) {
|
|
vc4_load_tiled_image(trans->map, ptrans->stride,
|
|
buf + slice->offset +
|
|
ptrans->box.z * rsc->cube_map_stride,
|
|
slice->stride,
|
|
slice->tiling, rsc->cpp,
|
|
&ptrans->box);
|
|
}
|
|
return (trans->map +
|
|
box_start_x * rsc->cpp +
|
|
box_start_y * ptrans->stride);
|
|
} else {
|
|
ptrans->stride = slice->stride;
|
|
ptrans->layer_stride = ptrans->stride;
|
|
|
|
return buf + slice->offset +
|
|
ptrans->box.y / util_format_get_blockheight(format) * ptrans->stride +
|
|
ptrans->box.x / util_format_get_blockwidth(format) * rsc->cpp +
|
|
ptrans->box.z * rsc->cube_map_stride;
|
|
}
|
|
|
|
|
|
fail:
|
|
vc4_resource_transfer_unmap(pctx, ptrans);
|
|
return NULL;
|
|
}
|
|
|
|
static void
|
|
vc4_resource_destroy(struct pipe_screen *pscreen,
|
|
struct pipe_resource *prsc)
|
|
{
|
|
struct vc4_screen *screen = vc4_screen(pscreen);
|
|
struct vc4_resource *rsc = vc4_resource(prsc);
|
|
vc4_bo_unreference(&rsc->bo);
|
|
|
|
if (rsc->scanout)
|
|
renderonly_scanout_destroy(rsc->scanout, screen->ro);
|
|
|
|
free(rsc);
|
|
}
|
|
|
|
static boolean
|
|
vc4_resource_get_handle(struct pipe_screen *pscreen,
|
|
struct pipe_context *pctx,
|
|
struct pipe_resource *prsc,
|
|
struct winsys_handle *whandle,
|
|
unsigned usage)
|
|
{
|
|
struct vc4_screen *screen = vc4_screen(pscreen);
|
|
struct vc4_resource *rsc = vc4_resource(prsc);
|
|
|
|
whandle->stride = rsc->slices[0].stride;
|
|
whandle->offset = 0;
|
|
|
|
/* If we're passing some reference to our BO out to some other part of
|
|
* the system, then we can't do any optimizations about only us being
|
|
* the ones seeing it (like BO caching or shadow update avoidance).
|
|
*/
|
|
rsc->bo->private = false;
|
|
|
|
if (rsc->tiled)
|
|
whandle->modifier = DRM_FORMAT_MOD_BROADCOM_VC4_T_TILED;
|
|
else
|
|
whandle->modifier = DRM_FORMAT_MOD_LINEAR;
|
|
|
|
switch (whandle->type) {
|
|
case DRM_API_HANDLE_TYPE_SHARED:
|
|
if (screen->ro) {
|
|
/* This could probably be supported, assuming that a
|
|
* control node was used for pl111.
|
|
*/
|
|
fprintf(stderr, "flink unsupported with pl111\n");
|
|
return FALSE;
|
|
}
|
|
|
|
return vc4_bo_flink(rsc->bo, &whandle->handle);
|
|
case DRM_API_HANDLE_TYPE_KMS:
|
|
if (screen->ro && renderonly_get_handle(rsc->scanout, whandle))
|
|
return TRUE;
|
|
whandle->handle = rsc->bo->handle;
|
|
return TRUE;
|
|
case DRM_API_HANDLE_TYPE_FD:
|
|
/* FDs are cross-device, so we can export directly from vc4.
|
|
*/
|
|
whandle->handle = vc4_bo_get_dmabuf(rsc->bo);
|
|
return whandle->handle != -1;
|
|
}
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
static void
|
|
vc4_setup_slices(struct vc4_resource *rsc, const char *caller)
|
|
{
|
|
struct pipe_resource *prsc = &rsc->base;
|
|
uint32_t width = prsc->width0;
|
|
uint32_t height = prsc->height0;
|
|
if (prsc->format == PIPE_FORMAT_ETC1_RGB8) {
|
|
width = (width + 3) >> 2;
|
|
height = (height + 3) >> 2;
|
|
}
|
|
|
|
uint32_t pot_width = util_next_power_of_two(width);
|
|
uint32_t pot_height = util_next_power_of_two(height);
|
|
uint32_t offset = 0;
|
|
uint32_t utile_w = vc4_utile_width(rsc->cpp);
|
|
uint32_t utile_h = vc4_utile_height(rsc->cpp);
|
|
|
|
for (int i = prsc->last_level; i >= 0; i--) {
|
|
struct vc4_resource_slice *slice = &rsc->slices[i];
|
|
|
|
uint32_t level_width, level_height;
|
|
if (i == 0) {
|
|
level_width = width;
|
|
level_height = height;
|
|
} else {
|
|
level_width = u_minify(pot_width, i);
|
|
level_height = u_minify(pot_height, i);
|
|
}
|
|
|
|
if (!rsc->tiled) {
|
|
slice->tiling = VC4_TILING_FORMAT_LINEAR;
|
|
if (prsc->nr_samples > 1) {
|
|
/* MSAA (4x) surfaces are stored as raw tile buffer contents. */
|
|
level_width = align(level_width, 32);
|
|
level_height = align(level_height, 32);
|
|
} else {
|
|
level_width = align(level_width, utile_w);
|
|
}
|
|
} else {
|
|
if (vc4_size_is_lt(level_width, level_height,
|
|
rsc->cpp)) {
|
|
slice->tiling = VC4_TILING_FORMAT_LT;
|
|
level_width = align(level_width, utile_w);
|
|
level_height = align(level_height, utile_h);
|
|
} else {
|
|
slice->tiling = VC4_TILING_FORMAT_T;
|
|
level_width = align(level_width,
|
|
4 * 2 * utile_w);
|
|
level_height = align(level_height,
|
|
4 * 2 * utile_h);
|
|
}
|
|
}
|
|
|
|
slice->offset = offset;
|
|
slice->stride = (level_width * rsc->cpp *
|
|
MAX2(prsc->nr_samples, 1));
|
|
slice->size = level_height * slice->stride;
|
|
|
|
offset += slice->size;
|
|
|
|
if (vc4_debug & VC4_DEBUG_SURFACE) {
|
|
static const char tiling_chars[] = {
|
|
[VC4_TILING_FORMAT_LINEAR] = 'R',
|
|
[VC4_TILING_FORMAT_LT] = 'L',
|
|
[VC4_TILING_FORMAT_T] = 'T'
|
|
};
|
|
fprintf(stderr,
|
|
"rsc %s %p (format %s: vc4 %d), %dx%d: "
|
|
"level %d (%c) -> %dx%d, stride %d@0x%08x\n",
|
|
caller, rsc,
|
|
util_format_short_name(prsc->format),
|
|
rsc->vc4_format,
|
|
prsc->width0, prsc->height0,
|
|
i, tiling_chars[slice->tiling],
|
|
level_width, level_height,
|
|
slice->stride, slice->offset);
|
|
}
|
|
}
|
|
|
|
/* The texture base pointer that has to point to level 0 doesn't have
|
|
* intra-page bits, so we have to align it, and thus shift up all the
|
|
* smaller slices.
|
|
*/
|
|
uint32_t page_align_offset = (align(rsc->slices[0].offset, 4096) -
|
|
rsc->slices[0].offset);
|
|
if (page_align_offset) {
|
|
for (int i = 0; i <= prsc->last_level; i++)
|
|
rsc->slices[i].offset += page_align_offset;
|
|
}
|
|
|
|
/* Cube map faces appear as whole miptrees at a page-aligned offset
|
|
* from the first face's miptree.
|
|
*/
|
|
if (prsc->target == PIPE_TEXTURE_CUBE) {
|
|
rsc->cube_map_stride = align(rsc->slices[0].offset +
|
|
rsc->slices[0].size, 4096);
|
|
}
|
|
}
|
|
|
|
static struct vc4_resource *
|
|
vc4_resource_setup(struct pipe_screen *pscreen,
|
|
const struct pipe_resource *tmpl)
|
|
{
|
|
struct vc4_resource *rsc = CALLOC_STRUCT(vc4_resource);
|
|
if (!rsc)
|
|
return NULL;
|
|
struct pipe_resource *prsc = &rsc->base;
|
|
|
|
*prsc = *tmpl;
|
|
|
|
pipe_reference_init(&prsc->reference, 1);
|
|
prsc->screen = pscreen;
|
|
|
|
if (prsc->nr_samples <= 1)
|
|
rsc->cpp = util_format_get_blocksize(tmpl->format);
|
|
else
|
|
rsc->cpp = sizeof(uint32_t);
|
|
|
|
assert(rsc->cpp);
|
|
|
|
return rsc;
|
|
}
|
|
|
|
static enum vc4_texture_data_type
|
|
get_resource_texture_format(struct pipe_resource *prsc)
|
|
{
|
|
struct vc4_resource *rsc = vc4_resource(prsc);
|
|
uint8_t format = vc4_get_tex_format(prsc->format);
|
|
|
|
if (!rsc->tiled) {
|
|
if (prsc->nr_samples > 1) {
|
|
return ~0;
|
|
} else {
|
|
assert(format == VC4_TEXTURE_TYPE_RGBA8888);
|
|
return VC4_TEXTURE_TYPE_RGBA32R;
|
|
}
|
|
}
|
|
|
|
return format;
|
|
}
|
|
|
|
static bool
|
|
find_modifier(uint64_t needle, const uint64_t *haystack, int count)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < count; i++) {
|
|
if (haystack[i] == needle)
|
|
return true;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
static struct pipe_resource *
|
|
vc4_resource_create_with_modifiers(struct pipe_screen *pscreen,
|
|
const struct pipe_resource *tmpl,
|
|
const uint64_t *modifiers,
|
|
int count)
|
|
{
|
|
struct vc4_screen *screen = vc4_screen(pscreen);
|
|
struct vc4_resource *rsc = vc4_resource_setup(pscreen, tmpl);
|
|
struct pipe_resource *prsc = &rsc->base;
|
|
bool linear_ok = find_modifier(DRM_FORMAT_MOD_LINEAR, modifiers, count);
|
|
/* Use a tiled layout if we can, for better 3D performance. */
|
|
bool should_tile = true;
|
|
|
|
/* VBOs/PBOs are untiled (and 1 height). */
|
|
if (tmpl->target == PIPE_BUFFER)
|
|
should_tile = false;
|
|
|
|
/* MSAA buffers are linear. */
|
|
if (tmpl->nr_samples > 1)
|
|
should_tile = false;
|
|
|
|
/* No tiling when we're sharing with another device (pl111). */
|
|
if (screen->ro && (tmpl->bind & PIPE_BIND_SCANOUT))
|
|
should_tile = false;
|
|
|
|
/* Cursors are always linear, and the user can request linear as well.
|
|
*/
|
|
if (tmpl->bind & (PIPE_BIND_LINEAR | PIPE_BIND_CURSOR))
|
|
should_tile = false;
|
|
|
|
/* No shared objects with LT format -- the kernel only has T-format
|
|
* metadata. LT objects are small enough it's not worth the trouble to
|
|
* give them metadata to tile.
|
|
*/
|
|
if ((tmpl->bind & (PIPE_BIND_SHARED | PIPE_BIND_SCANOUT)) &&
|
|
vc4_size_is_lt(prsc->width0, prsc->height0, rsc->cpp))
|
|
should_tile = false;
|
|
|
|
/* If we're sharing or scanning out, we need the ioctl present to
|
|
* inform the kernel or the other side.
|
|
*/
|
|
if ((tmpl->bind & (PIPE_BIND_SHARED |
|
|
PIPE_BIND_SCANOUT)) && !screen->has_tiling_ioctl)
|
|
should_tile = false;
|
|
|
|
/* No user-specified modifier; determine our own. */
|
|
if (count == 1 && modifiers[0] == DRM_FORMAT_MOD_INVALID) {
|
|
linear_ok = true;
|
|
rsc->tiled = should_tile;
|
|
} else if (should_tile &&
|
|
find_modifier(DRM_FORMAT_MOD_BROADCOM_VC4_T_TILED,
|
|
modifiers, count)) {
|
|
rsc->tiled = true;
|
|
} else if (linear_ok) {
|
|
rsc->tiled = false;
|
|
} else {
|
|
fprintf(stderr, "Unsupported modifier requested\n");
|
|
return NULL;
|
|
}
|
|
|
|
if (tmpl->target != PIPE_BUFFER)
|
|
rsc->vc4_format = get_resource_texture_format(prsc);
|
|
|
|
vc4_setup_slices(rsc, "create");
|
|
if (!vc4_resource_bo_alloc(rsc))
|
|
goto fail;
|
|
|
|
if (screen->has_tiling_ioctl) {
|
|
uint64_t modifier;
|
|
if (rsc->tiled)
|
|
modifier = DRM_FORMAT_MOD_BROADCOM_VC4_T_TILED;
|
|
else
|
|
modifier = DRM_FORMAT_MOD_LINEAR;
|
|
struct drm_vc4_set_tiling set_tiling = {
|
|
.handle = rsc->bo->handle,
|
|
.modifier = modifier,
|
|
};
|
|
int ret = vc4_ioctl(screen->fd, DRM_IOCTL_VC4_SET_TILING,
|
|
&set_tiling);
|
|
if (ret != 0)
|
|
goto fail;
|
|
}
|
|
|
|
if (screen->ro && tmpl->bind & PIPE_BIND_SCANOUT) {
|
|
rsc->scanout =
|
|
renderonly_scanout_for_resource(prsc, screen->ro, NULL);
|
|
if (!rsc->scanout)
|
|
goto fail;
|
|
}
|
|
|
|
vc4_bo_label(screen, rsc->bo, "%sresource %dx%d@%d/%d",
|
|
(tmpl->bind & PIPE_BIND_SCANOUT) ? "scanout " : "",
|
|
tmpl->width0, tmpl->height0,
|
|
rsc->cpp * 8, prsc->last_level);
|
|
|
|
return prsc;
|
|
fail:
|
|
vc4_resource_destroy(pscreen, prsc);
|
|
return NULL;
|
|
}
|
|
|
|
struct pipe_resource *
|
|
vc4_resource_create(struct pipe_screen *pscreen,
|
|
const struct pipe_resource *tmpl)
|
|
{
|
|
const uint64_t mod = DRM_FORMAT_MOD_INVALID;
|
|
return vc4_resource_create_with_modifiers(pscreen, tmpl, &mod, 1);
|
|
}
|
|
|
|
static struct pipe_resource *
|
|
vc4_resource_from_handle(struct pipe_screen *pscreen,
|
|
const struct pipe_resource *tmpl,
|
|
struct winsys_handle *whandle,
|
|
unsigned usage)
|
|
{
|
|
struct vc4_screen *screen = vc4_screen(pscreen);
|
|
struct vc4_resource *rsc = vc4_resource_setup(pscreen, tmpl);
|
|
struct pipe_resource *prsc = &rsc->base;
|
|
struct vc4_resource_slice *slice = &rsc->slices[0];
|
|
|
|
if (!rsc)
|
|
return NULL;
|
|
|
|
if (whandle->offset != 0) {
|
|
fprintf(stderr,
|
|
"Attempt to import unsupported winsys offset %u\n",
|
|
whandle->offset);
|
|
return NULL;
|
|
}
|
|
|
|
switch (whandle->type) {
|
|
case DRM_API_HANDLE_TYPE_SHARED:
|
|
rsc->bo = vc4_bo_open_name(screen,
|
|
whandle->handle, whandle->stride);
|
|
break;
|
|
case DRM_API_HANDLE_TYPE_FD:
|
|
rsc->bo = vc4_bo_open_dmabuf(screen,
|
|
whandle->handle, whandle->stride);
|
|
break;
|
|
default:
|
|
fprintf(stderr,
|
|
"Attempt to import unsupported handle type %d\n",
|
|
whandle->type);
|
|
}
|
|
|
|
if (!rsc->bo)
|
|
goto fail;
|
|
|
|
struct drm_vc4_get_tiling get_tiling = {
|
|
.handle = rsc->bo->handle,
|
|
};
|
|
int ret = vc4_ioctl(screen->fd, DRM_IOCTL_VC4_GET_TILING, &get_tiling);
|
|
|
|
if (ret != 0) {
|
|
whandle->modifier = DRM_FORMAT_MOD_LINEAR;
|
|
} else if (whandle->modifier == DRM_FORMAT_MOD_INVALID) {
|
|
whandle->modifier = get_tiling.modifier;
|
|
} else if (whandle->modifier != get_tiling.modifier) {
|
|
fprintf(stderr,
|
|
"Modifier 0x%llx vs. tiling (0x%llx) mismatch\n",
|
|
(long long)whandle->modifier, get_tiling.modifier);
|
|
goto fail;
|
|
}
|
|
|
|
switch (whandle->modifier) {
|
|
case DRM_FORMAT_MOD_LINEAR:
|
|
rsc->tiled = false;
|
|
break;
|
|
case DRM_FORMAT_MOD_BROADCOM_VC4_T_TILED:
|
|
rsc->tiled = true;
|
|
break;
|
|
default:
|
|
fprintf(stderr,
|
|
"Attempt to import unsupported modifier 0x%llx\n",
|
|
(long long)whandle->modifier);
|
|
goto fail;
|
|
}
|
|
|
|
rsc->vc4_format = get_resource_texture_format(prsc);
|
|
vc4_setup_slices(rsc, "import");
|
|
|
|
if (screen->ro) {
|
|
/* Make sure that renderonly has a handle to our buffer in the
|
|
* display's fd, so that a later renderonly_get_handle()
|
|
* returns correct handles or GEM names.
|
|
*/
|
|
rsc->scanout =
|
|
renderonly_create_gpu_import_for_resource(prsc,
|
|
screen->ro,
|
|
NULL);
|
|
if (!rsc->scanout)
|
|
goto fail;
|
|
}
|
|
|
|
if (whandle->stride != slice->stride) {
|
|
static bool warned = false;
|
|
if (!warned) {
|
|
warned = true;
|
|
fprintf(stderr,
|
|
"Attempting to import %dx%d %s with "
|
|
"unsupported stride %d instead of %d\n",
|
|
prsc->width0, prsc->height0,
|
|
util_format_short_name(prsc->format),
|
|
whandle->stride,
|
|
slice->stride);
|
|
}
|
|
goto fail;
|
|
}
|
|
|
|
return prsc;
|
|
|
|
fail:
|
|
vc4_resource_destroy(pscreen, prsc);
|
|
return NULL;
|
|
}
|
|
|
|
static struct pipe_surface *
|
|
vc4_create_surface(struct pipe_context *pctx,
|
|
struct pipe_resource *ptex,
|
|
const struct pipe_surface *surf_tmpl)
|
|
{
|
|
struct vc4_surface *surface = CALLOC_STRUCT(vc4_surface);
|
|
struct vc4_resource *rsc = vc4_resource(ptex);
|
|
|
|
if (!surface)
|
|
return NULL;
|
|
|
|
assert(surf_tmpl->u.tex.first_layer == surf_tmpl->u.tex.last_layer);
|
|
|
|
struct pipe_surface *psurf = &surface->base;
|
|
unsigned level = surf_tmpl->u.tex.level;
|
|
|
|
pipe_reference_init(&psurf->reference, 1);
|
|
pipe_resource_reference(&psurf->texture, ptex);
|
|
|
|
psurf->context = pctx;
|
|
psurf->format = surf_tmpl->format;
|
|
psurf->width = u_minify(ptex->width0, level);
|
|
psurf->height = u_minify(ptex->height0, level);
|
|
psurf->u.tex.level = level;
|
|
psurf->u.tex.first_layer = surf_tmpl->u.tex.first_layer;
|
|
psurf->u.tex.last_layer = surf_tmpl->u.tex.last_layer;
|
|
surface->offset = (rsc->slices[level].offset +
|
|
psurf->u.tex.first_layer * rsc->cube_map_stride);
|
|
surface->tiling = rsc->slices[level].tiling;
|
|
|
|
return &surface->base;
|
|
}
|
|
|
|
static void
|
|
vc4_surface_destroy(struct pipe_context *pctx, struct pipe_surface *psurf)
|
|
{
|
|
pipe_resource_reference(&psurf->texture, NULL);
|
|
FREE(psurf);
|
|
}
|
|
|
|
static void
|
|
vc4_dump_surface_non_msaa(struct pipe_surface *psurf)
|
|
{
|
|
struct pipe_resource *prsc = psurf->texture;
|
|
struct vc4_resource *rsc = vc4_resource(prsc);
|
|
uint32_t *map = vc4_bo_map(rsc->bo);
|
|
uint32_t stride = rsc->slices[0].stride / 4;
|
|
uint32_t width = psurf->width;
|
|
uint32_t height = psurf->height;
|
|
uint32_t chunk_w = width / 79;
|
|
uint32_t chunk_h = height / 40;
|
|
uint32_t found_colors[10];
|
|
uint32_t num_found_colors = 0;
|
|
|
|
if (rsc->vc4_format != VC4_TEXTURE_TYPE_RGBA32R) {
|
|
fprintf(stderr, "%s: Unsupported format %s\n",
|
|
__func__, util_format_short_name(psurf->format));
|
|
return;
|
|
}
|
|
|
|
for (int by = 0; by < height; by += chunk_h) {
|
|
for (int bx = 0; bx < width; bx += chunk_w) {
|
|
int all_found_color = -1; /* nothing found */
|
|
|
|
for (int y = by; y < MIN2(height, by + chunk_h); y++) {
|
|
for (int x = bx; x < MIN2(width, bx + chunk_w); x++) {
|
|
uint32_t pix = map[y * stride + x];
|
|
|
|
int i;
|
|
for (i = 0; i < num_found_colors; i++) {
|
|
if (pix == found_colors[i])
|
|
break;
|
|
}
|
|
if (i == num_found_colors &&
|
|
num_found_colors <
|
|
ARRAY_SIZE(found_colors)) {
|
|
found_colors[num_found_colors++] = pix;
|
|
}
|
|
|
|
if (i < num_found_colors) {
|
|
if (all_found_color == -1)
|
|
all_found_color = i;
|
|
else if (i != all_found_color)
|
|
all_found_color = ARRAY_SIZE(found_colors);
|
|
}
|
|
}
|
|
}
|
|
/* If all pixels for this chunk have a consistent
|
|
* value, then print a character for it. Either a
|
|
* fixed name (particularly common for piglit tests),
|
|
* or a runtime-generated number.
|
|
*/
|
|
if (all_found_color >= 0 &&
|
|
all_found_color < ARRAY_SIZE(found_colors)) {
|
|
static const struct {
|
|
uint32_t val;
|
|
const char *c;
|
|
} named_colors[] = {
|
|
{ 0xff000000, "█" },
|
|
{ 0x00000000, "█" },
|
|
{ 0xffff0000, "r" },
|
|
{ 0xff00ff00, "g" },
|
|
{ 0xff0000ff, "b" },
|
|
{ 0xffffffff, "w" },
|
|
};
|
|
int i;
|
|
for (i = 0; i < ARRAY_SIZE(named_colors); i++) {
|
|
if (named_colors[i].val ==
|
|
found_colors[all_found_color]) {
|
|
fprintf(stderr, "%s",
|
|
named_colors[i].c);
|
|
break;
|
|
}
|
|
}
|
|
/* For unnamed colors, print a number and the
|
|
* numbers will have values printed at the
|
|
* end.
|
|
*/
|
|
if (i == ARRAY_SIZE(named_colors)) {
|
|
fprintf(stderr, "%c",
|
|
'0' + all_found_color);
|
|
}
|
|
} else {
|
|
/* If there's no consistent color, print this.
|
|
*/
|
|
fprintf(stderr, ".");
|
|
}
|
|
}
|
|
fprintf(stderr, "\n");
|
|
}
|
|
|
|
for (int i = 0; i < num_found_colors; i++) {
|
|
fprintf(stderr, "color %d: 0x%08x\n", i, found_colors[i]);
|
|
}
|
|
}
|
|
|
|
static uint32_t
|
|
vc4_surface_msaa_get_sample(struct pipe_surface *psurf,
|
|
uint32_t x, uint32_t y, uint32_t sample)
|
|
{
|
|
struct pipe_resource *prsc = psurf->texture;
|
|
struct vc4_resource *rsc = vc4_resource(prsc);
|
|
uint32_t tile_w = 32, tile_h = 32;
|
|
uint32_t tiles_w = DIV_ROUND_UP(psurf->width, 32);
|
|
|
|
uint32_t tile_x = x / tile_w;
|
|
uint32_t tile_y = y / tile_h;
|
|
uint32_t *tile = (vc4_bo_map(rsc->bo) +
|
|
VC4_TILE_BUFFER_SIZE * (tile_y * tiles_w + tile_x));
|
|
uint32_t subtile_x = x % tile_w;
|
|
uint32_t subtile_y = y % tile_h;
|
|
|
|
uint32_t quad_samples = VC4_MAX_SAMPLES * 4;
|
|
uint32_t tile_stride = quad_samples * tile_w / 2;
|
|
|
|
return *((uint32_t *)tile +
|
|
(subtile_y >> 1) * tile_stride +
|
|
(subtile_x >> 1) * quad_samples +
|
|
((subtile_y & 1) << 1) +
|
|
(subtile_x & 1) +
|
|
sample);
|
|
}
|
|
|
|
static void
|
|
vc4_dump_surface_msaa_char(struct pipe_surface *psurf,
|
|
uint32_t start_x, uint32_t start_y,
|
|
uint32_t w, uint32_t h)
|
|
{
|
|
bool all_same_color = true;
|
|
uint32_t all_pix = 0;
|
|
|
|
for (int y = start_y; y < start_y + h; y++) {
|
|
for (int x = start_x; x < start_x + w; x++) {
|
|
for (int s = 0; s < VC4_MAX_SAMPLES; s++) {
|
|
uint32_t pix = vc4_surface_msaa_get_sample(psurf,
|
|
x, y,
|
|
s);
|
|
if (x == start_x && y == start_y)
|
|
all_pix = pix;
|
|
else if (all_pix != pix)
|
|
all_same_color = false;
|
|
}
|
|
}
|
|
}
|
|
if (all_same_color) {
|
|
static const struct {
|
|
uint32_t val;
|
|
const char *c;
|
|
} named_colors[] = {
|
|
{ 0xff000000, "█" },
|
|
{ 0x00000000, "█" },
|
|
{ 0xffff0000, "r" },
|
|
{ 0xff00ff00, "g" },
|
|
{ 0xff0000ff, "b" },
|
|
{ 0xffffffff, "w" },
|
|
};
|
|
int i;
|
|
for (i = 0; i < ARRAY_SIZE(named_colors); i++) {
|
|
if (named_colors[i].val == all_pix) {
|
|
fprintf(stderr, "%s",
|
|
named_colors[i].c);
|
|
return;
|
|
}
|
|
}
|
|
fprintf(stderr, "x");
|
|
} else {
|
|
fprintf(stderr, ".");
|
|
}
|
|
}
|
|
|
|
static void
|
|
vc4_dump_surface_msaa(struct pipe_surface *psurf)
|
|
{
|
|
uint32_t tile_w = 32, tile_h = 32;
|
|
uint32_t tiles_w = DIV_ROUND_UP(psurf->width, tile_w);
|
|
uint32_t tiles_h = DIV_ROUND_UP(psurf->height, tile_h);
|
|
uint32_t char_w = 140, char_h = 60;
|
|
uint32_t char_w_per_tile = char_w / tiles_w - 1;
|
|
uint32_t char_h_per_tile = char_h / tiles_h - 1;
|
|
|
|
fprintf(stderr, "Surface: %dx%d (%dx MSAA)\n",
|
|
psurf->width, psurf->height, psurf->texture->nr_samples);
|
|
|
|
for (int x = 0; x < (char_w_per_tile + 1) * tiles_w; x++)
|
|
fprintf(stderr, "-");
|
|
fprintf(stderr, "\n");
|
|
|
|
for (int ty = 0; ty < psurf->height; ty += tile_h) {
|
|
for (int y = 0; y < char_h_per_tile; y++) {
|
|
|
|
for (int tx = 0; tx < psurf->width; tx += tile_w) {
|
|
for (int x = 0; x < char_w_per_tile; x++) {
|
|
uint32_t bx1 = (x * tile_w /
|
|
char_w_per_tile);
|
|
uint32_t bx2 = ((x + 1) * tile_w /
|
|
char_w_per_tile);
|
|
uint32_t by1 = (y * tile_h /
|
|
char_h_per_tile);
|
|
uint32_t by2 = ((y + 1) * tile_h /
|
|
char_h_per_tile);
|
|
|
|
vc4_dump_surface_msaa_char(psurf,
|
|
tx + bx1,
|
|
ty + by1,
|
|
bx2 - bx1,
|
|
by2 - by1);
|
|
}
|
|
fprintf(stderr, "|");
|
|
}
|
|
fprintf(stderr, "\n");
|
|
}
|
|
|
|
for (int x = 0; x < (char_w_per_tile + 1) * tiles_w; x++)
|
|
fprintf(stderr, "-");
|
|
fprintf(stderr, "\n");
|
|
}
|
|
}
|
|
|
|
/** Debug routine to dump the contents of an 8888 surface to the console */
|
|
void
|
|
vc4_dump_surface(struct pipe_surface *psurf)
|
|
{
|
|
if (!psurf)
|
|
return;
|
|
|
|
if (psurf->texture->nr_samples > 1)
|
|
vc4_dump_surface_msaa(psurf);
|
|
else
|
|
vc4_dump_surface_non_msaa(psurf);
|
|
}
|
|
|
|
static void
|
|
vc4_flush_resource(struct pipe_context *pctx, struct pipe_resource *resource)
|
|
{
|
|
/* All calls to flush_resource are followed by a flush of the context,
|
|
* so there's nothing to do.
|
|
*/
|
|
}
|
|
|
|
void
|
|
vc4_update_shadow_baselevel_texture(struct pipe_context *pctx,
|
|
struct pipe_sampler_view *pview)
|
|
{
|
|
struct vc4_sampler_view *view = vc4_sampler_view(pview);
|
|
struct vc4_resource *shadow = vc4_resource(view->texture);
|
|
struct vc4_resource *orig = vc4_resource(pview->texture);
|
|
|
|
assert(view->texture != pview->texture);
|
|
|
|
if (shadow->writes == orig->writes && orig->bo->private)
|
|
return;
|
|
|
|
perf_debug("Updating %dx%d@%d shadow texture due to %s\n",
|
|
orig->base.width0, orig->base.height0,
|
|
pview->u.tex.first_level,
|
|
pview->u.tex.first_level ? "base level" : "raster layout");
|
|
|
|
for (int i = 0; i <= shadow->base.last_level; i++) {
|
|
unsigned width = u_minify(shadow->base.width0, i);
|
|
unsigned height = u_minify(shadow->base.height0, i);
|
|
struct pipe_blit_info info = {
|
|
.dst = {
|
|
.resource = &shadow->base,
|
|
.level = i,
|
|
.box = {
|
|
.x = 0,
|
|
.y = 0,
|
|
.z = 0,
|
|
.width = width,
|
|
.height = height,
|
|
.depth = 1,
|
|
},
|
|
.format = shadow->base.format,
|
|
},
|
|
.src = {
|
|
.resource = &orig->base,
|
|
.level = pview->u.tex.first_level + i,
|
|
.box = {
|
|
.x = 0,
|
|
.y = 0,
|
|
.z = 0,
|
|
.width = width,
|
|
.height = height,
|
|
.depth = 1,
|
|
},
|
|
.format = orig->base.format,
|
|
},
|
|
.mask = ~0,
|
|
};
|
|
pctx->blit(pctx, &info);
|
|
}
|
|
|
|
shadow->writes = orig->writes;
|
|
}
|
|
|
|
/**
|
|
* Converts a 4-byte index buffer to 2 bytes.
|
|
*
|
|
* Since GLES2 only has support for 1 and 2-byte indices, the hardware doesn't
|
|
* include 4-byte index support, and we have to shrink it down.
|
|
*
|
|
* There's no fallback support for when indices end up being larger than 2^16,
|
|
* though it will at least assertion fail. Also, if the original index data
|
|
* was in user memory, it would be nice to not have uploaded it to a VBO
|
|
* before translating.
|
|
*/
|
|
struct pipe_resource *
|
|
vc4_get_shadow_index_buffer(struct pipe_context *pctx,
|
|
const struct pipe_draw_info *info,
|
|
uint32_t offset,
|
|
uint32_t count,
|
|
uint32_t *shadow_offset)
|
|
{
|
|
struct vc4_context *vc4 = vc4_context(pctx);
|
|
struct vc4_resource *orig = vc4_resource(info->index.resource);
|
|
perf_debug("Fallback conversion for %d uint indices\n", count);
|
|
|
|
void *data;
|
|
struct pipe_resource *shadow_rsc = NULL;
|
|
u_upload_alloc(vc4->uploader, 0, count * 2, 4,
|
|
shadow_offset, &shadow_rsc, &data);
|
|
uint16_t *dst = data;
|
|
|
|
struct pipe_transfer *src_transfer = NULL;
|
|
const uint32_t *src;
|
|
if (info->has_user_indices) {
|
|
src = info->index.user;
|
|
} else {
|
|
src = pipe_buffer_map_range(pctx, &orig->base,
|
|
offset,
|
|
count * 4,
|
|
PIPE_TRANSFER_READ, &src_transfer);
|
|
}
|
|
|
|
for (int i = 0; i < count; i++) {
|
|
uint32_t src_index = src[i];
|
|
assert(src_index <= 0xffff);
|
|
dst[i] = src_index;
|
|
}
|
|
|
|
if (src_transfer)
|
|
pctx->transfer_unmap(pctx, src_transfer);
|
|
|
|
return shadow_rsc;
|
|
}
|
|
|
|
void
|
|
vc4_resource_screen_init(struct pipe_screen *pscreen)
|
|
{
|
|
struct vc4_screen *screen = vc4_screen(pscreen);
|
|
|
|
pscreen->resource_create = vc4_resource_create;
|
|
pscreen->resource_create_with_modifiers =
|
|
vc4_resource_create_with_modifiers;
|
|
pscreen->resource_from_handle = vc4_resource_from_handle;
|
|
pscreen->resource_destroy = u_resource_destroy_vtbl;
|
|
pscreen->resource_get_handle = vc4_resource_get_handle;
|
|
pscreen->resource_destroy = vc4_resource_destroy;
|
|
|
|
/* Test if the kernel has GET_TILING; it will return -EINVAL if the
|
|
* ioctl does not exist, but -ENOENT if we pass an impossible handle.
|
|
* 0 cannot be a valid GEM object, so use that.
|
|
*/
|
|
struct drm_vc4_get_tiling get_tiling = {
|
|
.handle = 0x0,
|
|
};
|
|
int ret = vc4_ioctl(screen->fd, DRM_IOCTL_VC4_GET_TILING, &get_tiling);
|
|
if (ret == -1 && errno == ENOENT)
|
|
screen->has_tiling_ioctl = true;
|
|
}
|
|
|
|
void
|
|
vc4_resource_context_init(struct pipe_context *pctx)
|
|
{
|
|
pctx->transfer_map = vc4_resource_transfer_map;
|
|
pctx->transfer_flush_region = u_default_transfer_flush_region;
|
|
pctx->transfer_unmap = vc4_resource_transfer_unmap;
|
|
pctx->buffer_subdata = u_default_buffer_subdata;
|
|
pctx->texture_subdata = u_default_texture_subdata;
|
|
pctx->create_surface = vc4_create_surface;
|
|
pctx->surface_destroy = vc4_surface_destroy;
|
|
pctx->resource_copy_region = util_resource_copy_region;
|
|
pctx->blit = vc4_blit;
|
|
pctx->flush_resource = vc4_flush_resource;
|
|
}
|