mirror of
https://github.com/raspberrypi/linux.git
synced 2025-12-06 01:49:46 +00:00
drm/vc4: Synchronize validation code for v2 submission upstream.
Signed-off-by: Eric Anholt <eric@anholt.net>
This commit is contained in:
@@ -189,17 +189,6 @@ to_vc4_encoder(struct drm_encoder *encoder)
|
||||
#define HVS_READ(offset) readl(vc4->hvs->regs + offset)
|
||||
#define HVS_WRITE(offset, val) writel(val, vc4->hvs->regs + offset)
|
||||
|
||||
enum vc4_bo_mode {
|
||||
VC4_MODE_UNDECIDED,
|
||||
VC4_MODE_RENDER,
|
||||
VC4_MODE_SHADER,
|
||||
};
|
||||
|
||||
struct vc4_bo_exec_state {
|
||||
struct drm_gem_cma_object *bo;
|
||||
enum vc4_bo_mode mode;
|
||||
};
|
||||
|
||||
struct vc4_exec_info {
|
||||
/* Sequence number for this bin/render job. */
|
||||
uint64_t seqno;
|
||||
@@ -210,7 +199,7 @@ struct vc4_exec_info {
|
||||
/* This is the array of BOs that were looked up at the start of exec.
|
||||
* Command validation will use indices into this array.
|
||||
*/
|
||||
struct vc4_bo_exec_state *bo;
|
||||
struct drm_gem_cma_object **bo;
|
||||
uint32_t bo_count;
|
||||
|
||||
/* Pointers for our position in vc4->job_list */
|
||||
@@ -238,7 +227,6 @@ struct vc4_exec_info {
|
||||
* command lists.
|
||||
*/
|
||||
struct vc4_shader_state {
|
||||
uint8_t packet;
|
||||
uint32_t addr;
|
||||
/* Maximum vertex index referenced by any primitive using this
|
||||
* shader state.
|
||||
@@ -254,6 +242,7 @@ struct vc4_exec_info {
|
||||
bool found_tile_binning_mode_config_packet;
|
||||
bool found_start_tile_binning_packet;
|
||||
bool found_increment_semaphore_packet;
|
||||
bool found_flush;
|
||||
uint8_t bin_tiles_x, bin_tiles_y;
|
||||
struct drm_gem_cma_object *tile_bo;
|
||||
uint32_t tile_alloc_offset;
|
||||
@@ -265,6 +254,9 @@ struct vc4_exec_info {
|
||||
uint32_t ct0ca, ct0ea;
|
||||
uint32_t ct1ca, ct1ea;
|
||||
|
||||
/* Pointer to the unvalidated bin CL (if present). */
|
||||
void *bin_u;
|
||||
|
||||
/* Pointers to the shader recs. These paddr gets incremented as CL
|
||||
* packets are relocated in validate_gl_shader_state, and the vaddrs
|
||||
* (u and v) get incremented and size decremented as the shader recs
|
||||
@@ -455,10 +447,8 @@ vc4_validate_bin_cl(struct drm_device *dev,
|
||||
int
|
||||
vc4_validate_shader_recs(struct drm_device *dev, struct vc4_exec_info *exec);
|
||||
|
||||
bool vc4_use_bo(struct vc4_exec_info *exec,
|
||||
uint32_t hindex,
|
||||
enum vc4_bo_mode mode,
|
||||
struct drm_gem_cma_object **obj);
|
||||
struct drm_gem_cma_object *vc4_use_bo(struct vc4_exec_info *exec,
|
||||
uint32_t hindex);
|
||||
|
||||
int vc4_get_rcl(struct drm_device *dev, struct vc4_exec_info *exec);
|
||||
|
||||
|
||||
@@ -169,8 +169,8 @@ vc4_save_hang_state(struct drm_device *dev)
|
||||
}
|
||||
|
||||
for (i = 0; i < exec->bo_count; i++) {
|
||||
drm_gem_object_reference(&exec->bo[i].bo->base);
|
||||
kernel_state->bo[i] = &exec->bo[i].bo->base;
|
||||
drm_gem_object_reference(&exec->bo[i]->base);
|
||||
kernel_state->bo[i] = &exec->bo[i]->base;
|
||||
}
|
||||
|
||||
list_for_each_entry(bo, &exec->unref_list, unref_head) {
|
||||
@@ -397,7 +397,7 @@ vc4_update_bo_seqnos(struct vc4_exec_info *exec, uint64_t seqno)
|
||||
unsigned i;
|
||||
|
||||
for (i = 0; i < exec->bo_count; i++) {
|
||||
bo = to_vc4_bo(&exec->bo[i].bo->base);
|
||||
bo = to_vc4_bo(&exec->bo[i]->base);
|
||||
bo->seqno = seqno;
|
||||
}
|
||||
|
||||
@@ -467,7 +467,7 @@ vc4_cl_lookup_bos(struct drm_device *dev,
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
exec->bo = kcalloc(exec->bo_count, sizeof(struct vc4_bo_exec_state),
|
||||
exec->bo = kcalloc(exec->bo_count, sizeof(struct drm_gem_cma_object *),
|
||||
GFP_KERNEL);
|
||||
if (!exec->bo) {
|
||||
DRM_ERROR("Failed to allocate validated BO pointers\n");
|
||||
@@ -500,7 +500,7 @@ vc4_cl_lookup_bos(struct drm_device *dev,
|
||||
goto fail;
|
||||
}
|
||||
drm_gem_object_reference(bo);
|
||||
exec->bo[i].bo = (struct drm_gem_cma_object *)bo;
|
||||
exec->bo[i] = (struct drm_gem_cma_object *)bo;
|
||||
}
|
||||
spin_unlock(&file_priv->table_lock);
|
||||
|
||||
@@ -591,6 +591,8 @@ vc4_get_bcl(struct drm_device *dev, struct vc4_exec_info *exec)
|
||||
|
||||
exec->ct0ca = exec->exec_bo->paddr + bin_offset;
|
||||
|
||||
exec->bin_u = bin;
|
||||
|
||||
exec->shader_rec_v = exec->exec_bo->vaddr + shader_rec_offset;
|
||||
exec->shader_rec_p = exec->exec_bo->paddr + shader_rec_offset;
|
||||
exec->shader_rec_size = args->shader_rec_size;
|
||||
@@ -622,7 +624,7 @@ vc4_complete_exec(struct drm_device *dev, struct vc4_exec_info *exec)
|
||||
mutex_lock(&dev->struct_mutex);
|
||||
if (exec->bo) {
|
||||
for (i = 0; i < exec->bo_count; i++)
|
||||
drm_gem_object_unreference(&exec->bo[i].bo->base);
|
||||
drm_gem_object_unreference(&exec->bo[i]->base);
|
||||
kfree(exec->bo);
|
||||
}
|
||||
|
||||
|
||||
@@ -436,7 +436,8 @@ static int vc4_rcl_surface_setup(struct vc4_exec_info *exec,
|
||||
if (surf->hindex == ~0)
|
||||
return 0;
|
||||
|
||||
if (!vc4_use_bo(exec, surf->hindex, VC4_MODE_RENDER, obj))
|
||||
*obj = vc4_use_bo(exec, surf->hindex);
|
||||
if (!*obj)
|
||||
return -EINVAL;
|
||||
|
||||
if (surf->flags & VC4_SUBMIT_RCL_SURFACE_READ_IS_FULL_RES) {
|
||||
@@ -537,7 +538,8 @@ vc4_rcl_render_config_surface_setup(struct vc4_exec_info *exec,
|
||||
if (surf->hindex == ~0)
|
||||
return 0;
|
||||
|
||||
if (!vc4_use_bo(exec, surf->hindex, VC4_MODE_RENDER, obj))
|
||||
*obj = vc4_use_bo(exec, surf->hindex);
|
||||
if (!*obj)
|
||||
return -EINVAL;
|
||||
|
||||
if (tiling > VC4_TILING_FORMAT_LT) {
|
||||
|
||||
@@ -94,42 +94,42 @@ size_is_lt(uint32_t width, uint32_t height, int cpp)
|
||||
height <= 4 * utile_height(cpp));
|
||||
}
|
||||
|
||||
bool
|
||||
vc4_use_bo(struct vc4_exec_info *exec,
|
||||
uint32_t hindex,
|
||||
enum vc4_bo_mode mode,
|
||||
struct drm_gem_cma_object **obj)
|
||||
struct drm_gem_cma_object *
|
||||
vc4_use_bo(struct vc4_exec_info *exec, uint32_t hindex)
|
||||
{
|
||||
*obj = NULL;
|
||||
struct drm_gem_cma_object *obj;
|
||||
struct vc4_bo *bo;
|
||||
|
||||
if (hindex >= exec->bo_count) {
|
||||
DRM_ERROR("BO index %d greater than BO count %d\n",
|
||||
hindex, exec->bo_count);
|
||||
return false;
|
||||
return NULL;
|
||||
}
|
||||
obj = exec->bo[hindex];
|
||||
bo = to_vc4_bo(&obj->base);
|
||||
|
||||
if (bo->validated_shader) {
|
||||
DRM_ERROR("Trying to use shader BO as something other than "
|
||||
"a shader\n");
|
||||
return NULL;
|
||||
}
|
||||
|
||||
if (exec->bo[hindex].mode != mode) {
|
||||
if (exec->bo[hindex].mode == VC4_MODE_UNDECIDED) {
|
||||
exec->bo[hindex].mode = mode;
|
||||
} else {
|
||||
DRM_ERROR("BO index %d reused with mode %d vs %d\n",
|
||||
hindex, exec->bo[hindex].mode, mode);
|
||||
return false;
|
||||
}
|
||||
}
|
||||
return obj;
|
||||
}
|
||||
|
||||
*obj = exec->bo[hindex].bo;
|
||||
return true;
|
||||
static struct drm_gem_cma_object *
|
||||
vc4_use_handle(struct vc4_exec_info *exec, uint32_t gem_handles_packet_index)
|
||||
{
|
||||
return vc4_use_bo(exec, exec->bo_index[gem_handles_packet_index]);
|
||||
}
|
||||
|
||||
static bool
|
||||
vc4_use_handle(struct vc4_exec_info *exec,
|
||||
uint32_t gem_handles_packet_index,
|
||||
enum vc4_bo_mode mode,
|
||||
struct drm_gem_cma_object **obj)
|
||||
validate_bin_pos(struct vc4_exec_info *exec, void *untrusted, uint32_t pos)
|
||||
{
|
||||
return vc4_use_bo(exec, exec->bo_index[gem_handles_packet_index],
|
||||
mode, obj);
|
||||
/* Note that the untrusted pointer passed to these functions is
|
||||
* incremented past the packet byte.
|
||||
*/
|
||||
return (untrusted - 1 == exec->bin_u + pos);
|
||||
}
|
||||
|
||||
static uint32_t
|
||||
@@ -202,13 +202,13 @@ vc4_check_tex_size(struct vc4_exec_info *exec, struct drm_gem_cma_object *fbo,
|
||||
}
|
||||
|
||||
static int
|
||||
validate_flush_all(VALIDATE_ARGS)
|
||||
validate_flush(VALIDATE_ARGS)
|
||||
{
|
||||
if (exec->found_increment_semaphore_packet) {
|
||||
DRM_ERROR("VC4_PACKET_FLUSH_ALL after "
|
||||
"VC4_PACKET_INCREMENT_SEMAPHORE\n");
|
||||
if (!validate_bin_pos(exec, untrusted, exec->args->bin_cl_size - 1)) {
|
||||
DRM_ERROR("Bin CL must end with VC4_PACKET_FLUSH\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
exec->found_flush = true;
|
||||
|
||||
return 0;
|
||||
}
|
||||
@@ -233,17 +233,13 @@ validate_start_tile_binning(VALIDATE_ARGS)
|
||||
static int
|
||||
validate_increment_semaphore(VALIDATE_ARGS)
|
||||
{
|
||||
if (exec->found_increment_semaphore_packet) {
|
||||
DRM_ERROR("Duplicate VC4_PACKET_INCREMENT_SEMAPHORE\n");
|
||||
if (!validate_bin_pos(exec, untrusted, exec->args->bin_cl_size - 2)) {
|
||||
DRM_ERROR("Bin CL must end with "
|
||||
"VC4_PACKET_INCREMENT_SEMAPHORE\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
exec->found_increment_semaphore_packet = true;
|
||||
|
||||
/* Once we've found the semaphore increment, there should be one FLUSH
|
||||
* then the end of the command list. The FLUSH actually triggers the
|
||||
* increment, so we only need to make sure there
|
||||
*/
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
@@ -257,11 +253,6 @@ validate_indexed_prim_list(VALIDATE_ARGS)
|
||||
uint32_t index_size = (*(uint8_t *)(untrusted + 0) >> 4) ? 2 : 1;
|
||||
struct vc4_shader_state *shader_state;
|
||||
|
||||
if (exec->found_increment_semaphore_packet) {
|
||||
DRM_ERROR("Drawing after VC4_PACKET_INCREMENT_SEMAPHORE\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
/* Check overflow condition */
|
||||
if (exec->shader_state_count == 0) {
|
||||
DRM_ERROR("shader state must precede primitives\n");
|
||||
@@ -272,7 +263,8 @@ validate_indexed_prim_list(VALIDATE_ARGS)
|
||||
if (max_index > shader_state->max_index)
|
||||
shader_state->max_index = max_index;
|
||||
|
||||
if (!vc4_use_handle(exec, 0, VC4_MODE_RENDER, &ib))
|
||||
ib = vc4_use_handle(exec, 0);
|
||||
if (!ib)
|
||||
return -EINVAL;
|
||||
|
||||
if (offset > ib->base.size ||
|
||||
@@ -295,11 +287,6 @@ validate_gl_array_primitive(VALIDATE_ARGS)
|
||||
uint32_t max_index;
|
||||
struct vc4_shader_state *shader_state;
|
||||
|
||||
if (exec->found_increment_semaphore_packet) {
|
||||
DRM_ERROR("Drawing after VC4_PACKET_INCREMENT_SEMAPHORE\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
/* Check overflow condition */
|
||||
if (exec->shader_state_count == 0) {
|
||||
DRM_ERROR("shader state must precede primitives\n");
|
||||
@@ -329,7 +316,6 @@ validate_gl_shader_state(VALIDATE_ARGS)
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
exec->shader_state[i].packet = VC4_PACKET_GL_SHADER_STATE;
|
||||
exec->shader_state[i].addr = *(uint32_t *)untrusted;
|
||||
exec->shader_state[i].max_index = 0;
|
||||
|
||||
@@ -347,31 +333,6 @@ validate_gl_shader_state(VALIDATE_ARGS)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int
|
||||
validate_nv_shader_state(VALIDATE_ARGS)
|
||||
{
|
||||
uint32_t i = exec->shader_state_count++;
|
||||
|
||||
if (i >= exec->shader_state_size) {
|
||||
DRM_ERROR("More requests for shader states than declared\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
exec->shader_state[i].packet = VC4_PACKET_NV_SHADER_STATE;
|
||||
exec->shader_state[i].addr = *(uint32_t *)untrusted;
|
||||
|
||||
if (exec->shader_state[i].addr & 15) {
|
||||
DRM_ERROR("NV shader state address 0x%08x misaligned\n",
|
||||
exec->shader_state[i].addr);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
*(uint32_t *)validated = (exec->shader_state[i].addr +
|
||||
exec->shader_rec_p);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int
|
||||
validate_tile_binning_config(VALIDATE_ARGS)
|
||||
{
|
||||
@@ -473,8 +434,8 @@ static const struct cmd_info {
|
||||
} cmd_info[] = {
|
||||
VC4_DEFINE_PACKET(VC4_PACKET_HALT, NULL),
|
||||
VC4_DEFINE_PACKET(VC4_PACKET_NOP, NULL),
|
||||
VC4_DEFINE_PACKET(VC4_PACKET_FLUSH, NULL),
|
||||
VC4_DEFINE_PACKET(VC4_PACKET_FLUSH_ALL, validate_flush_all),
|
||||
VC4_DEFINE_PACKET(VC4_PACKET_FLUSH, validate_flush),
|
||||
VC4_DEFINE_PACKET(VC4_PACKET_FLUSH_ALL, NULL),
|
||||
VC4_DEFINE_PACKET(VC4_PACKET_START_TILE_BINNING,
|
||||
validate_start_tile_binning),
|
||||
VC4_DEFINE_PACKET(VC4_PACKET_INCREMENT_SEMAPHORE,
|
||||
@@ -488,7 +449,6 @@ static const struct cmd_info {
|
||||
VC4_DEFINE_PACKET(VC4_PACKET_PRIMITIVE_LIST_FORMAT, NULL),
|
||||
|
||||
VC4_DEFINE_PACKET(VC4_PACKET_GL_SHADER_STATE, validate_gl_shader_state),
|
||||
VC4_DEFINE_PACKET(VC4_PACKET_NV_SHADER_STATE, validate_nv_shader_state),
|
||||
|
||||
VC4_DEFINE_PACKET(VC4_PACKET_CONFIGURATION_BITS, NULL),
|
||||
VC4_DEFINE_PACKET(VC4_PACKET_FLAT_SHADE_FLAGS, NULL),
|
||||
@@ -575,8 +535,16 @@ vc4_validate_bin_cl(struct drm_device *dev,
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
if (!exec->found_increment_semaphore_packet) {
|
||||
DRM_ERROR("Bin CL missing VC4_PACKET_INCREMENT_SEMAPHORE\n");
|
||||
/* The bin CL must be ended with INCREMENT_SEMAPHORE and FLUSH. The
|
||||
* semaphore is used to trigger the render CL to start up, and the
|
||||
* FLUSH is what caps the bin lists with
|
||||
* VC4_PACKET_RETURN_FROM_SUB_LIST (so they jump back to the main
|
||||
* render CL when they get called to) and actually triggers the queued
|
||||
* semaphore increment.
|
||||
*/
|
||||
if (!exec->found_increment_semaphore_packet || !exec->found_flush) {
|
||||
DRM_ERROR("Bin CL missing VC4_PACKET_INCREMENT_SEMAPHORE + "
|
||||
"VC4_PACKET_FLUSH\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
@@ -607,7 +575,8 @@ reloc_tex(struct vc4_exec_info *exec,
|
||||
uint32_t cube_map_stride = 0;
|
||||
enum vc4_texture_data_type type;
|
||||
|
||||
if (!vc4_use_bo(exec, texture_handle_index, VC4_MODE_RENDER, &tex))
|
||||
tex = vc4_use_bo(exec, texture_handle_index);
|
||||
if (!tex)
|
||||
return false;
|
||||
|
||||
if (sample->is_direct) {
|
||||
@@ -755,51 +724,28 @@ reloc_tex(struct vc4_exec_info *exec,
|
||||
}
|
||||
|
||||
static int
|
||||
validate_shader_rec(struct drm_device *dev,
|
||||
struct vc4_exec_info *exec,
|
||||
struct vc4_shader_state *state)
|
||||
validate_gl_shader_rec(struct drm_device *dev,
|
||||
struct vc4_exec_info *exec,
|
||||
struct vc4_shader_state *state)
|
||||
{
|
||||
uint32_t *src_handles;
|
||||
void *pkt_u, *pkt_v;
|
||||
enum shader_rec_reloc_type {
|
||||
RELOC_CODE,
|
||||
RELOC_VBO,
|
||||
static const uint32_t shader_reloc_offsets[] = {
|
||||
4, /* fs */
|
||||
16, /* vs */
|
||||
28, /* cs */
|
||||
};
|
||||
struct shader_rec_reloc {
|
||||
enum shader_rec_reloc_type type;
|
||||
uint32_t offset;
|
||||
};
|
||||
static const struct shader_rec_reloc gl_relocs[] = {
|
||||
{ RELOC_CODE, 4 }, /* fs */
|
||||
{ RELOC_CODE, 16 }, /* vs */
|
||||
{ RELOC_CODE, 28 }, /* cs */
|
||||
};
|
||||
static const struct shader_rec_reloc nv_relocs[] = {
|
||||
{ RELOC_CODE, 4 }, /* fs */
|
||||
{ RELOC_VBO, 12 }
|
||||
};
|
||||
const struct shader_rec_reloc *relocs;
|
||||
struct drm_gem_cma_object *bo[ARRAY_SIZE(gl_relocs) + 8];
|
||||
uint32_t nr_attributes = 0, nr_fixed_relocs, nr_relocs, packet_size;
|
||||
uint32_t shader_reloc_count = ARRAY_SIZE(shader_reloc_offsets);
|
||||
struct drm_gem_cma_object *bo[shader_reloc_count + 8];
|
||||
uint32_t nr_attributes, nr_relocs, packet_size;
|
||||
int i;
|
||||
struct vc4_validated_shader_info *shader;
|
||||
|
||||
if (state->packet == VC4_PACKET_NV_SHADER_STATE) {
|
||||
relocs = nv_relocs;
|
||||
nr_fixed_relocs = ARRAY_SIZE(nv_relocs);
|
||||
|
||||
packet_size = 16;
|
||||
} else {
|
||||
relocs = gl_relocs;
|
||||
nr_fixed_relocs = ARRAY_SIZE(gl_relocs);
|
||||
|
||||
nr_attributes = state->addr & 0x7;
|
||||
if (nr_attributes == 0)
|
||||
nr_attributes = 8;
|
||||
packet_size = gl_shader_rec_size(state->addr);
|
||||
}
|
||||
nr_relocs = nr_fixed_relocs + nr_attributes;
|
||||
nr_attributes = state->addr & 0x7;
|
||||
if (nr_attributes == 0)
|
||||
nr_attributes = 8;
|
||||
packet_size = gl_shader_rec_size(state->addr);
|
||||
|
||||
nr_relocs = ARRAY_SIZE(shader_reloc_offsets) + nr_attributes;
|
||||
if (nr_relocs * 4 > exec->shader_rec_size) {
|
||||
DRM_ERROR("overflowed shader recs reading %d handles "
|
||||
"from %d bytes left\n",
|
||||
@@ -829,21 +775,30 @@ validate_shader_rec(struct drm_device *dev,
|
||||
exec->shader_rec_v += roundup(packet_size, 16);
|
||||
exec->shader_rec_size -= packet_size;
|
||||
|
||||
for (i = 0; i < nr_relocs; i++) {
|
||||
enum vc4_bo_mode mode;
|
||||
|
||||
if (i < nr_fixed_relocs && relocs[i].type == RELOC_CODE)
|
||||
mode = VC4_MODE_SHADER;
|
||||
else
|
||||
mode = VC4_MODE_RENDER;
|
||||
|
||||
if (!vc4_use_bo(exec, src_handles[i], mode, &bo[i]))
|
||||
return false;
|
||||
if (!(*(uint16_t *)pkt_u & VC4_SHADER_FLAG_FS_SINGLE_THREAD)) {
|
||||
DRM_ERROR("Multi-threaded fragment shaders not supported.\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
for (i = 0; i < nr_fixed_relocs; i++) {
|
||||
struct vc4_bo *vc4_bo;
|
||||
uint32_t o = relocs[i].offset;
|
||||
for (i = 0; i < shader_reloc_count; i++) {
|
||||
if (src_handles[i] > exec->bo_count) {
|
||||
DRM_ERROR("Shader handle %d too big\n", src_handles[i]);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
bo[i] = exec->bo[src_handles[i]];
|
||||
if (!bo[i])
|
||||
return -EINVAL;
|
||||
}
|
||||
for (i = shader_reloc_count; i < nr_relocs; i++) {
|
||||
bo[i] = vc4_use_bo(exec, src_handles[i]);
|
||||
if (!bo[i])
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
for (i = 0; i < shader_reloc_count; i++) {
|
||||
struct vc4_validated_shader_info *validated_shader;
|
||||
uint32_t o = shader_reloc_offsets[i];
|
||||
uint32_t src_offset = *(uint32_t *)(pkt_u + o);
|
||||
uint32_t *texture_handles_u;
|
||||
void *uniform_data_u;
|
||||
@@ -851,57 +806,50 @@ validate_shader_rec(struct drm_device *dev,
|
||||
|
||||
*(uint32_t *)(pkt_v + o) = bo[i]->paddr + src_offset;
|
||||
|
||||
switch (relocs[i].type) {
|
||||
case RELOC_CODE:
|
||||
if (src_offset != 0) {
|
||||
DRM_ERROR("Shaders must be at offset 0 "
|
||||
"of the BO.\n");
|
||||
goto fail;
|
||||
}
|
||||
|
||||
vc4_bo = to_vc4_bo(&bo[i]->base);
|
||||
shader = vc4_bo->validated_shader;
|
||||
if (!shader)
|
||||
goto fail;
|
||||
|
||||
if (shader->uniforms_src_size > exec->uniforms_size) {
|
||||
DRM_ERROR("Uniforms src buffer overflow\n");
|
||||
goto fail;
|
||||
}
|
||||
|
||||
texture_handles_u = exec->uniforms_u;
|
||||
uniform_data_u = (texture_handles_u +
|
||||
shader->num_texture_samples);
|
||||
|
||||
memcpy(exec->uniforms_v, uniform_data_u,
|
||||
shader->uniforms_size);
|
||||
|
||||
for (tex = 0;
|
||||
tex < shader->num_texture_samples;
|
||||
tex++) {
|
||||
if (!reloc_tex(exec,
|
||||
uniform_data_u,
|
||||
&shader->texture_samples[tex],
|
||||
texture_handles_u[tex])) {
|
||||
goto fail;
|
||||
}
|
||||
}
|
||||
|
||||
*(uint32_t *)(pkt_v + o + 4) = exec->uniforms_p;
|
||||
|
||||
exec->uniforms_u += shader->uniforms_src_size;
|
||||
exec->uniforms_v += shader->uniforms_size;
|
||||
exec->uniforms_p += shader->uniforms_size;
|
||||
|
||||
break;
|
||||
|
||||
case RELOC_VBO:
|
||||
break;
|
||||
if (src_offset != 0) {
|
||||
DRM_ERROR("Shaders must be at offset 0 of "
|
||||
"the BO.\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
validated_shader = to_vc4_bo(&bo[i]->base)->validated_shader;
|
||||
if (!validated_shader)
|
||||
return -EINVAL;
|
||||
|
||||
if (validated_shader->uniforms_src_size >
|
||||
exec->uniforms_size) {
|
||||
DRM_ERROR("Uniforms src buffer overflow\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
texture_handles_u = exec->uniforms_u;
|
||||
uniform_data_u = (texture_handles_u +
|
||||
validated_shader->num_texture_samples);
|
||||
|
||||
memcpy(exec->uniforms_v, uniform_data_u,
|
||||
validated_shader->uniforms_size);
|
||||
|
||||
for (tex = 0;
|
||||
tex < validated_shader->num_texture_samples;
|
||||
tex++) {
|
||||
if (!reloc_tex(exec,
|
||||
uniform_data_u,
|
||||
&validated_shader->texture_samples[tex],
|
||||
texture_handles_u[tex])) {
|
||||
return -EINVAL;
|
||||
}
|
||||
}
|
||||
|
||||
*(uint32_t *)(pkt_v + o + 4) = exec->uniforms_p;
|
||||
|
||||
exec->uniforms_u += validated_shader->uniforms_src_size;
|
||||
exec->uniforms_v += validated_shader->uniforms_size;
|
||||
exec->uniforms_p += validated_shader->uniforms_size;
|
||||
}
|
||||
|
||||
for (i = 0; i < nr_attributes; i++) {
|
||||
struct drm_gem_cma_object *vbo = bo[nr_fixed_relocs + i];
|
||||
struct drm_gem_cma_object *vbo =
|
||||
bo[ARRAY_SIZE(shader_reloc_offsets) + i];
|
||||
uint32_t o = 36 + i * 8;
|
||||
uint32_t offset = *(uint32_t *)(pkt_u + o + 0);
|
||||
uint32_t attr_size = *(uint8_t *)(pkt_u + o + 4) + 1;
|
||||
@@ -933,9 +881,6 @@ validate_shader_rec(struct drm_device *dev,
|
||||
}
|
||||
|
||||
return 0;
|
||||
|
||||
fail:
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
int
|
||||
@@ -946,7 +891,7 @@ vc4_validate_shader_recs(struct drm_device *dev,
|
||||
int ret = 0;
|
||||
|
||||
for (i = 0; i < exec->shader_state_count; i++) {
|
||||
ret = validate_shader_rec(dev, exec, &exec->shader_state[i]);
|
||||
ret = validate_gl_shader_rec(dev, exec, &exec->shader_state[i]);
|
||||
if (ret)
|
||||
return ret;
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user