mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-23 01:50:07 +03:00
Disabling NVGPU_SUPPORT_MIG in suspend path leads to inconsistencies. During driver removal without the flag set, the driver still tries to free structures that might not have been allocated in the first place. e.g. nvgpu_gr_zbc_deinit, nvgpu_gr_zcull_deinit. Added NULL checks for ZBC and ZCULL structures before freeing them as a solution. Jira NVGPU-6832 Change-Id: I8a0c64ca982d11fee55542abd3c5bce5a51b4007 Signed-off-by: Debarshi Dutta <ddutta@nvidia.com> Reviewed-on: https://git-master.nvidia.com/r/c/linux-nvgpu/+/2535101 Reviewed-by: svc-mobile-coverity <svc-mobile-coverity@nvidia.com> Reviewed-by: svc-mobile-misra <svc-mobile-misra@nvidia.com> Reviewed-by: svc-mobile-cert <svc-mobile-cert@nvidia.com> Reviewed-by: svc_kernel_abi <svc_kernel_abi@nvidia.com> Reviewed-by: Lakshmanan M <lm@nvidia.com> Reviewed-by: Dinesh T <dt@nvidia.com> Reviewed-by: Deepak Nibade <dnibade@nvidia.com> Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com> GVS: Gerrit_Virtual_Submit Tested-by: mobile promotions <svcmobile_promotions@nvidia.com>
691 lines
19 KiB
C
691 lines
19 KiB
C
/*
|
|
* Copyright (c) 2019-2021, NVIDIA CORPORATION. All rights reserved.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
|
* DEALINGS IN THE SOFTWARE.
|
|
*/
|
|
|
|
#include <nvgpu/gk20a.h>
|
|
#include <nvgpu/io.h>
|
|
#include <nvgpu/bug.h>
|
|
#include <nvgpu/string.h>
|
|
#include <nvgpu/power_features/pg.h>
|
|
#ifdef CONFIG_NVGPU_LS_PMU
|
|
#include <nvgpu/pmu/pmu_pg.h>
|
|
#endif
|
|
|
|
#include "zbc_priv.h"
|
|
|
|
#define ZBC_ENTRY_UPDATED 1
|
|
#define ZBC_ENTRY_ADDED 2
|
|
|
|
static void nvgpu_gr_zbc_update_stencil_reg(struct gk20a *g,
|
|
struct nvgpu_gr_zbc_entry *stencil_val, u32 index)
|
|
{
|
|
/* update l2 table */
|
|
if (g->ops.ltc.set_zbc_s_entry != NULL) {
|
|
g->ops.ltc.set_zbc_s_entry(g, stencil_val->stencil, index);
|
|
}
|
|
|
|
/* update zbc stencil registers */
|
|
g->ops.gr.zbc.add_stencil(g, stencil_val, index);
|
|
}
|
|
|
|
static int nvgpu_gr_zbc_add_stencil(struct gk20a *g, struct nvgpu_gr_zbc *zbc,
|
|
struct nvgpu_gr_zbc_entry *stencil_val)
|
|
{
|
|
struct zbc_stencil_table *s_tbl;
|
|
u32 i;
|
|
int entry_added = -ENOSPC;
|
|
bool entry_exist = false;
|
|
|
|
/* search existing tables */
|
|
for (i = zbc->min_stencil_index; i <= zbc->max_used_stencil_index;
|
|
i++) {
|
|
|
|
s_tbl = &zbc->zbc_s_tbl[i];
|
|
|
|
if ((s_tbl->ref_cnt != 0U) &&
|
|
(s_tbl->stencil == stencil_val->stencil) &&
|
|
(s_tbl->format == stencil_val->format)) {
|
|
s_tbl->ref_cnt = nvgpu_safe_add_u32(s_tbl->ref_cnt, 1U);
|
|
entry_exist = true;
|
|
entry_added = ZBC_ENTRY_UPDATED;
|
|
break;
|
|
}
|
|
}
|
|
/* add new table */
|
|
if (!entry_exist &&
|
|
(zbc->max_used_stencil_index < zbc->max_stencil_index)) {
|
|
|
|
/* Increment used index and add new entry at that index */
|
|
zbc->max_used_stencil_index =
|
|
nvgpu_safe_add_u32(zbc->max_used_stencil_index, 1U);
|
|
|
|
s_tbl = &zbc->zbc_s_tbl[zbc->max_used_stencil_index];
|
|
WARN_ON(s_tbl->ref_cnt != 0U);
|
|
|
|
/* update sw copy */
|
|
s_tbl->stencil = stencil_val->stencil;
|
|
s_tbl->format = stencil_val->format;
|
|
s_tbl->ref_cnt = nvgpu_safe_add_u32(s_tbl->ref_cnt, 1U);
|
|
|
|
nvgpu_gr_zbc_update_stencil_reg(g, stencil_val,
|
|
zbc->max_used_stencil_index);
|
|
|
|
entry_added = ZBC_ENTRY_ADDED;
|
|
}
|
|
return entry_added;
|
|
}
|
|
|
|
static void nvgpu_gr_zbc_update_depth_reg(struct gk20a *g,
|
|
struct nvgpu_gr_zbc_entry *depth_val, u32 index)
|
|
{
|
|
/* update l2 table */
|
|
g->ops.ltc.set_zbc_depth_entry(g, depth_val->depth, index);
|
|
|
|
/* update zbc registers */
|
|
g->ops.gr.zbc.add_depth(g, depth_val, index);
|
|
}
|
|
|
|
static int nvgpu_gr_zbc_add_depth(struct gk20a *g, struct nvgpu_gr_zbc *zbc,
|
|
struct nvgpu_gr_zbc_entry *depth_val)
|
|
{
|
|
struct zbc_depth_table *d_tbl;
|
|
u32 i;
|
|
int entry_added = -ENOSPC;
|
|
bool entry_exist = false;
|
|
|
|
/* search existing tables */
|
|
for (i = zbc->min_depth_index; i <= zbc->max_used_depth_index; i++) {
|
|
|
|
d_tbl = &zbc->zbc_dep_tbl[i];
|
|
|
|
if ((d_tbl->ref_cnt != 0U) &&
|
|
(d_tbl->depth == depth_val->depth) &&
|
|
(d_tbl->format == depth_val->format)) {
|
|
d_tbl->ref_cnt = nvgpu_safe_add_u32(d_tbl->ref_cnt, 1U);
|
|
entry_exist = true;
|
|
entry_added = ZBC_ENTRY_UPDATED;
|
|
break;
|
|
}
|
|
}
|
|
/* add new table */
|
|
if (!entry_exist &&
|
|
(zbc->max_used_depth_index < zbc->max_depth_index)) {
|
|
|
|
/* Increment used index and add new entry at that index */
|
|
zbc->max_used_depth_index =
|
|
nvgpu_safe_add_u32(zbc->max_used_depth_index, 1U);
|
|
|
|
d_tbl = &zbc->zbc_dep_tbl[zbc->max_used_depth_index];
|
|
WARN_ON(d_tbl->ref_cnt != 0U);
|
|
|
|
/* update sw copy */
|
|
d_tbl->depth = depth_val->depth;
|
|
d_tbl->format = depth_val->format;
|
|
d_tbl->ref_cnt = nvgpu_safe_add_u32(d_tbl->ref_cnt, 1U);
|
|
|
|
nvgpu_gr_zbc_update_depth_reg(g, depth_val,
|
|
zbc->max_used_depth_index);
|
|
|
|
entry_added = ZBC_ENTRY_ADDED;
|
|
}
|
|
|
|
return entry_added;
|
|
}
|
|
|
|
static void nvgpu_gr_zbc_update_color_reg(struct gk20a *g,
|
|
struct nvgpu_gr_zbc_entry *color_val, u32 index)
|
|
{
|
|
/* update l2 table */
|
|
g->ops.ltc.set_zbc_color_entry(g, color_val->color_l2, index);
|
|
|
|
/* update zbc registers */
|
|
g->ops.gr.zbc.add_color(g, color_val, index);
|
|
}
|
|
|
|
static int nvgpu_gr_zbc_add_color(struct gk20a *g, struct nvgpu_gr_zbc *zbc,
|
|
struct nvgpu_gr_zbc_entry *color_val)
|
|
{
|
|
struct zbc_color_table *c_tbl;
|
|
u32 i;
|
|
int entry_added = -ENOSPC;
|
|
bool entry_exist = false;
|
|
|
|
/* search existing table */
|
|
for (i = zbc->min_color_index; i <= zbc->max_used_color_index; i++) {
|
|
|
|
c_tbl = &zbc->zbc_col_tbl[i];
|
|
|
|
if ((c_tbl->ref_cnt != 0U) &&
|
|
(c_tbl->format == color_val->format) &&
|
|
(nvgpu_memcmp((u8 *)c_tbl->color_ds,
|
|
(u8 *)color_val->color_ds,
|
|
sizeof(color_val->color_ds)) == 0) &&
|
|
(nvgpu_memcmp((u8 *)c_tbl->color_l2,
|
|
(u8 *)color_val->color_l2,
|
|
sizeof(color_val->color_l2)) == 0)) {
|
|
|
|
c_tbl->ref_cnt = nvgpu_safe_add_u32(c_tbl->ref_cnt, 1U);
|
|
entry_exist = true;
|
|
entry_added = ZBC_ENTRY_UPDATED;
|
|
break;
|
|
}
|
|
}
|
|
|
|
/* add new entry */
|
|
if (!entry_exist &&
|
|
(zbc->max_used_color_index < zbc->max_color_index)) {
|
|
|
|
/* Increment used index and add new entry at that index */
|
|
zbc->max_used_color_index =
|
|
nvgpu_safe_add_u32(zbc->max_used_color_index, 1U);
|
|
|
|
c_tbl = &zbc->zbc_col_tbl[zbc->max_used_color_index];
|
|
WARN_ON(c_tbl->ref_cnt != 0U);
|
|
|
|
/* update local copy */
|
|
for (i = 0; i < NVGPU_GR_ZBC_COLOR_VALUE_SIZE; i++) {
|
|
c_tbl->color_l2[i] = color_val->color_l2[i];
|
|
c_tbl->color_ds[i] = color_val->color_ds[i];
|
|
}
|
|
c_tbl->format = color_val->format;
|
|
c_tbl->ref_cnt = nvgpu_safe_add_u32(c_tbl->ref_cnt, 1U);
|
|
|
|
nvgpu_gr_zbc_update_color_reg(g, color_val,
|
|
zbc->max_used_color_index);
|
|
|
|
entry_added = ZBC_ENTRY_ADDED;
|
|
}
|
|
|
|
return entry_added;
|
|
}
|
|
|
|
static int nvgpu_gr_zbc_add(struct gk20a *g, struct nvgpu_gr_zbc *zbc,
|
|
struct nvgpu_gr_zbc_entry *zbc_val)
|
|
{
|
|
int added = false;
|
|
#if defined(CONFIG_NVGPU_LS_PMU) && defined(CONFIG_NVGPU_POWER_PG)
|
|
u32 entries;
|
|
#endif
|
|
|
|
/* no endian swap ? */
|
|
nvgpu_mutex_acquire(&zbc->zbc_lock);
|
|
nvgpu_speculation_barrier();
|
|
switch (zbc_val->type) {
|
|
case NVGPU_GR_ZBC_TYPE_COLOR:
|
|
added = nvgpu_gr_zbc_add_color(g, zbc, zbc_val);
|
|
break;
|
|
case NVGPU_GR_ZBC_TYPE_DEPTH:
|
|
added = nvgpu_gr_zbc_add_depth(g, zbc, zbc_val);
|
|
break;
|
|
case NVGPU_GR_ZBC_TYPE_STENCIL:
|
|
if (nvgpu_is_enabled(g, NVGPU_SUPPORT_ZBC_STENCIL)) {
|
|
added = nvgpu_gr_zbc_add_stencil(g, zbc, zbc_val);
|
|
} else {
|
|
nvgpu_err(g,
|
|
"invalid zbc table type %d", zbc_val->type);
|
|
added = -EINVAL;
|
|
goto err_mutex;
|
|
}
|
|
break;
|
|
default:
|
|
nvgpu_err(g,
|
|
"invalid zbc table type %d", zbc_val->type);
|
|
added = -EINVAL;
|
|
goto err_mutex;
|
|
}
|
|
|
|
#if defined(CONFIG_NVGPU_LS_PMU) && defined(CONFIG_NVGPU_POWER_PG)
|
|
if (added == ZBC_ENTRY_ADDED) {
|
|
/* update zbc for elpg only when new entry is added */
|
|
entries = max(
|
|
nvgpu_safe_sub_u32(zbc->max_used_color_index,
|
|
zbc->min_color_index),
|
|
nvgpu_safe_sub_u32(zbc->max_used_depth_index,
|
|
zbc->min_depth_index));
|
|
if (g->elpg_enabled) {
|
|
nvgpu_pmu_save_zbc(g, entries);
|
|
}
|
|
}
|
|
#endif
|
|
|
|
err_mutex:
|
|
nvgpu_mutex_release(&zbc->zbc_lock);
|
|
if (added < 0) {
|
|
return added;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int nvgpu_gr_zbc_set_table(struct gk20a *g, struct nvgpu_gr_zbc *zbc,
|
|
struct nvgpu_gr_zbc_entry *zbc_val)
|
|
{
|
|
nvgpu_log(g, gpu_dbg_zbc, " zbc_val->type %u", zbc_val->type);
|
|
|
|
return nvgpu_pg_elpg_protected_call(g,
|
|
nvgpu_gr_zbc_add(g, zbc, zbc_val));
|
|
}
|
|
|
|
/* get a zbc table entry specified by index
|
|
* return table size when type is invalid */
|
|
int nvgpu_gr_zbc_query_table(struct gk20a *g, struct nvgpu_gr_zbc *zbc,
|
|
struct nvgpu_gr_zbc_query_params *query_params)
|
|
{
|
|
u32 index = query_params->index_size;
|
|
u32 i;
|
|
|
|
nvgpu_speculation_barrier();
|
|
switch (query_params->type) {
|
|
case NVGPU_GR_ZBC_TYPE_INVALID:
|
|
nvgpu_log(g, gpu_dbg_zbc, "Query zbc size");
|
|
query_params->index_size = nvgpu_safe_add_u32(
|
|
nvgpu_safe_sub_u32(zbc->max_color_index,
|
|
zbc->min_color_index), 1U);
|
|
break;
|
|
case NVGPU_GR_ZBC_TYPE_COLOR:
|
|
if ((index < zbc->min_color_index) ||
|
|
(index > zbc->max_color_index)) {
|
|
nvgpu_err(g, "invalid zbc color table index %u", index);
|
|
return -EINVAL;
|
|
}
|
|
nvgpu_log(g, gpu_dbg_zbc, "Query zbc color at index %u", index);
|
|
|
|
nvgpu_speculation_barrier();
|
|
for (i = 0; i < NVGPU_GR_ZBC_COLOR_VALUE_SIZE; i++) {
|
|
query_params->color_l2[i] =
|
|
zbc->zbc_col_tbl[index].color_l2[i];
|
|
query_params->color_ds[i] =
|
|
zbc->zbc_col_tbl[index].color_ds[i];
|
|
}
|
|
query_params->format = zbc->zbc_col_tbl[index].format;
|
|
query_params->ref_cnt = zbc->zbc_col_tbl[index].ref_cnt;
|
|
|
|
break;
|
|
case NVGPU_GR_ZBC_TYPE_DEPTH:
|
|
if ((index < zbc->min_depth_index) ||
|
|
(index > zbc->max_depth_index)) {
|
|
nvgpu_err(g, "invalid zbc depth table index %u", index);
|
|
return -EINVAL;
|
|
}
|
|
nvgpu_log(g, gpu_dbg_zbc, "Query zbc depth at index %u", index);
|
|
|
|
nvgpu_speculation_barrier();
|
|
query_params->depth = zbc->zbc_dep_tbl[index].depth;
|
|
query_params->format = zbc->zbc_dep_tbl[index].format;
|
|
query_params->ref_cnt = zbc->zbc_dep_tbl[index].ref_cnt;
|
|
break;
|
|
case NVGPU_GR_ZBC_TYPE_STENCIL:
|
|
if (nvgpu_is_enabled(g, NVGPU_SUPPORT_ZBC_STENCIL)) {
|
|
if ((index < zbc->min_stencil_index) ||
|
|
(index > zbc->max_stencil_index)) {
|
|
nvgpu_err(g,
|
|
"invalid zbc stencil table index %u",
|
|
index);
|
|
return -EINVAL;
|
|
}
|
|
nvgpu_log(g, gpu_dbg_zbc,
|
|
"Query zbc stencil at index %u", index);
|
|
|
|
nvgpu_speculation_barrier();
|
|
query_params->stencil = zbc->zbc_s_tbl[index].stencil;
|
|
query_params->format = zbc->zbc_s_tbl[index].format;
|
|
query_params->ref_cnt = zbc->zbc_s_tbl[index].ref_cnt;
|
|
} else {
|
|
nvgpu_err(g, "invalid zbc table type");
|
|
return -EINVAL;
|
|
}
|
|
break;
|
|
default:
|
|
nvgpu_err(g, "invalid zbc table type");
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Update zbc table registers as per sw copy of zbc tables
|
|
*/
|
|
void nvgpu_gr_zbc_load_table(struct gk20a *g, struct nvgpu_gr_zbc *zbc)
|
|
{
|
|
unsigned int i;
|
|
|
|
for (i = zbc->min_color_index; i <= zbc->max_used_color_index; i++) {
|
|
struct zbc_color_table *c_tbl = &zbc->zbc_col_tbl[i];
|
|
struct nvgpu_gr_zbc_entry zbc_val;
|
|
|
|
zbc_val.type = NVGPU_GR_ZBC_TYPE_COLOR;
|
|
nvgpu_memcpy((u8 *)zbc_val.color_ds,
|
|
(u8 *)c_tbl->color_ds, sizeof(zbc_val.color_ds));
|
|
nvgpu_memcpy((u8 *)zbc_val.color_l2,
|
|
(u8 *)c_tbl->color_l2, sizeof(zbc_val.color_l2));
|
|
zbc_val.format = c_tbl->format;
|
|
|
|
nvgpu_gr_zbc_update_color_reg(g, &zbc_val, i);
|
|
}
|
|
|
|
for (i = zbc->min_depth_index; i <= zbc->max_used_depth_index; i++) {
|
|
struct zbc_depth_table *d_tbl = &zbc->zbc_dep_tbl[i];
|
|
struct nvgpu_gr_zbc_entry zbc_val;
|
|
|
|
zbc_val.type = NVGPU_GR_ZBC_TYPE_DEPTH;
|
|
zbc_val.depth = d_tbl->depth;
|
|
zbc_val.format = d_tbl->format;
|
|
|
|
nvgpu_gr_zbc_update_depth_reg(g, &zbc_val, i);
|
|
}
|
|
|
|
if (nvgpu_is_enabled(g, NVGPU_SUPPORT_ZBC_STENCIL)) {
|
|
for (i = zbc->min_stencil_index;
|
|
i <= zbc->max_used_stencil_index; i++) {
|
|
struct zbc_stencil_table *s_tbl = &zbc->zbc_s_tbl[i];
|
|
struct nvgpu_gr_zbc_entry zbc_val;
|
|
|
|
zbc_val.type = NVGPU_GR_ZBC_TYPE_STENCIL;
|
|
zbc_val.stencil = s_tbl->stencil;
|
|
zbc_val.format = s_tbl->format;
|
|
|
|
nvgpu_gr_zbc_update_stencil_reg(g, &zbc_val, i);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void nvgpu_gr_zbc_load_default_sw_stencil_table(struct gk20a *g,
|
|
struct nvgpu_gr_zbc *zbc)
|
|
{
|
|
u32 index = zbc->min_stencil_index;
|
|
|
|
zbc->zbc_s_tbl[index].stencil = 0x0;
|
|
zbc->zbc_s_tbl[index].format = GR_ZBC_STENCIL_CLEAR_FMT_U8;
|
|
zbc->zbc_s_tbl[index].ref_cnt =
|
|
nvgpu_safe_add_u32(zbc->zbc_s_tbl[index].ref_cnt, 1U);
|
|
index = nvgpu_safe_add_u32(index, 1U);
|
|
|
|
zbc->zbc_s_tbl[index].stencil = 0x1;
|
|
zbc->zbc_s_tbl[index].format = GR_ZBC_STENCIL_CLEAR_FMT_U8;
|
|
zbc->zbc_s_tbl[index].ref_cnt =
|
|
nvgpu_safe_add_u32(zbc->zbc_s_tbl[index].ref_cnt, 1U);
|
|
index = nvgpu_safe_add_u32(index, 1U);
|
|
|
|
zbc->zbc_s_tbl[index].stencil = 0xff;
|
|
zbc->zbc_s_tbl[index].format = GR_ZBC_STENCIL_CLEAR_FMT_U8;
|
|
zbc->zbc_s_tbl[index].ref_cnt =
|
|
nvgpu_safe_add_u32(zbc->zbc_s_tbl[index].ref_cnt, 1U);
|
|
|
|
zbc->max_used_stencil_index = index;
|
|
}
|
|
|
|
static void nvgpu_gr_zbc_load_default_sw_depth_table(struct gk20a *g,
|
|
struct nvgpu_gr_zbc *zbc)
|
|
{
|
|
u32 index = zbc->min_depth_index;
|
|
|
|
zbc->zbc_dep_tbl[index].format = GR_ZBC_Z_FMT_VAL_FP32;
|
|
zbc->zbc_dep_tbl[index].depth = 0x3f800000;
|
|
zbc->zbc_dep_tbl[index].ref_cnt =
|
|
nvgpu_safe_add_u32(zbc->zbc_dep_tbl[index].ref_cnt, 1U);
|
|
index = nvgpu_safe_add_u32(index, 1U);
|
|
|
|
zbc->zbc_dep_tbl[index].format = GR_ZBC_Z_FMT_VAL_FP32;
|
|
zbc->zbc_dep_tbl[index].depth = 0;
|
|
zbc->zbc_dep_tbl[index].ref_cnt =
|
|
nvgpu_safe_add_u32(zbc->zbc_dep_tbl[index].ref_cnt, 1U);
|
|
|
|
zbc->max_used_depth_index = index;
|
|
}
|
|
|
|
static void nvgpu_gr_zbc_load_default_sw_color_table(struct gk20a *g,
|
|
struct nvgpu_gr_zbc *zbc)
|
|
{
|
|
u32 i;
|
|
u32 index = zbc->min_color_index;
|
|
|
|
/* Opaque black (i.e. solid black, fmt 0x28 = A8B8G8R8) */
|
|
zbc->zbc_col_tbl[index].format = GR_ZBC_SOLID_BLACK_COLOR_FMT;
|
|
for (i = 0U; i < NVGPU_GR_ZBC_COLOR_VALUE_SIZE; i++) {
|
|
zbc->zbc_col_tbl[index].color_ds[i] = 0U;
|
|
zbc->zbc_col_tbl[index].color_l2[i] = 0xff000000U;
|
|
}
|
|
zbc->zbc_col_tbl[index].color_ds[3] = 0x3f800000U;
|
|
zbc->zbc_col_tbl[index].ref_cnt =
|
|
nvgpu_safe_add_u32(zbc->zbc_col_tbl[index].ref_cnt, 1U);
|
|
index = nvgpu_safe_add_u32(index, 1U);
|
|
|
|
/* Transparent black = (fmt 1 = zero) */
|
|
zbc->zbc_col_tbl[index].format = GR_ZBC_TRANSPARENT_BLACK_COLOR_FMT;
|
|
for (i = 0; i < NVGPU_GR_ZBC_COLOR_VALUE_SIZE; i++) {
|
|
zbc->zbc_col_tbl[index].color_ds[i] = 0U;
|
|
zbc->zbc_col_tbl[index].color_l2[i] = 0U;
|
|
}
|
|
zbc->zbc_col_tbl[index].ref_cnt =
|
|
nvgpu_safe_add_u32(zbc->zbc_col_tbl[index].ref_cnt, 1U);
|
|
index = nvgpu_safe_add_u32(index, 1U);
|
|
|
|
/* Opaque white (i.e. solid white) = (fmt 2 = uniform 1) */
|
|
zbc->zbc_col_tbl[index].format = GR_ZBC_SOLID_WHITE_COLOR_FMT;
|
|
for (i = 0; i < NVGPU_GR_ZBC_COLOR_VALUE_SIZE; i++) {
|
|
zbc->zbc_col_tbl[index].color_ds[i] = 0x3f800000U;
|
|
zbc->zbc_col_tbl[index].color_l2[i] = 0xffffffffU;
|
|
}
|
|
zbc->zbc_col_tbl[index].ref_cnt =
|
|
nvgpu_safe_add_u32(zbc->zbc_col_tbl[index].ref_cnt, 1U);
|
|
|
|
zbc->max_used_color_index = index;
|
|
}
|
|
|
|
static void nvgpu_gr_zbc_init_indices(struct gk20a *g, struct nvgpu_gr_zbc *zbc)
|
|
{
|
|
struct nvgpu_gr_zbc_table_indices zbc_indices;
|
|
|
|
g->ops.gr.zbc.init_table_indices(g, &zbc_indices);
|
|
|
|
zbc->min_color_index = zbc_indices.min_color_index;
|
|
zbc->max_color_index = zbc_indices.max_color_index;
|
|
zbc->min_depth_index = zbc_indices.min_depth_index;
|
|
zbc->max_depth_index = zbc_indices.max_depth_index;
|
|
zbc->min_stencil_index = zbc_indices.min_stencil_index;
|
|
zbc->max_stencil_index = zbc_indices.max_stencil_index;
|
|
|
|
nvgpu_log(g, gpu_dbg_zbc, "zbc->min_color_index %u",
|
|
zbc->min_color_index);
|
|
nvgpu_log(g, gpu_dbg_zbc, "zbc->max_color_index %u",
|
|
zbc->max_color_index);
|
|
nvgpu_log(g, gpu_dbg_zbc, "zbc->min_depth_index %u",
|
|
zbc->min_depth_index);
|
|
nvgpu_log(g, gpu_dbg_zbc, "zbc->max_depth_index %u",
|
|
zbc->max_depth_index);
|
|
nvgpu_log(g, gpu_dbg_zbc, "zbc->min_stencil_index %u",
|
|
zbc->min_stencil_index);
|
|
nvgpu_log(g, gpu_dbg_zbc, "zbc->max_stencil_index %u",
|
|
zbc->max_stencil_index);
|
|
}
|
|
|
|
static void nvgpu_gr_zbc_load_default_sw_table(struct gk20a *g,
|
|
struct nvgpu_gr_zbc *zbc)
|
|
{
|
|
nvgpu_mutex_init(&zbc->zbc_lock);
|
|
|
|
nvgpu_gr_zbc_load_default_sw_color_table(g, zbc);
|
|
|
|
nvgpu_gr_zbc_load_default_sw_depth_table(g, zbc);
|
|
|
|
if (nvgpu_is_enabled(g, NVGPU_SUPPORT_ZBC_STENCIL)) {
|
|
nvgpu_gr_zbc_load_default_sw_stencil_table(g, zbc);
|
|
}
|
|
}
|
|
|
|
static int gr_zbc_allocate_local_tbls(struct gk20a *g, struct nvgpu_gr_zbc *zbc)
|
|
{
|
|
u32 zbc_col_size = nvgpu_safe_add_u32(zbc->max_color_index,
|
|
zbc->min_color_index);
|
|
u32 zbc_dep_size = nvgpu_safe_add_u32(zbc->max_depth_index,
|
|
zbc->min_depth_index);
|
|
u32 zbc_s_size = nvgpu_safe_add_u32(zbc->max_stencil_index,
|
|
zbc->min_stencil_index);
|
|
|
|
zbc->zbc_col_tbl = nvgpu_kzalloc(g,
|
|
sizeof(struct zbc_color_table) * zbc_col_size);
|
|
if (zbc->zbc_col_tbl == NULL) {
|
|
goto alloc_col_tbl_err;
|
|
}
|
|
|
|
zbc->zbc_dep_tbl = nvgpu_kzalloc(g,
|
|
sizeof(struct zbc_depth_table) * zbc_dep_size);
|
|
|
|
if (zbc->zbc_dep_tbl == NULL) {
|
|
goto alloc_dep_tbl_err;
|
|
}
|
|
|
|
zbc->zbc_s_tbl = nvgpu_kzalloc(g,
|
|
sizeof(struct zbc_stencil_table) * zbc_s_size);
|
|
if (zbc->zbc_s_tbl == NULL) {
|
|
goto alloc_s_tbl_err;
|
|
}
|
|
|
|
return 0;
|
|
|
|
alloc_s_tbl_err:
|
|
nvgpu_kfree(g, zbc->zbc_dep_tbl);
|
|
alloc_dep_tbl_err:
|
|
nvgpu_kfree(g, zbc->zbc_col_tbl);
|
|
alloc_col_tbl_err:
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/* allocate the struct and load the table */
|
|
int nvgpu_gr_zbc_init(struct gk20a *g, struct nvgpu_gr_zbc **zbc)
|
|
{
|
|
int ret = -ENOMEM;
|
|
struct nvgpu_gr_zbc *gr_zbc = NULL;
|
|
|
|
*zbc = NULL;
|
|
|
|
gr_zbc = nvgpu_kzalloc(g, sizeof(*gr_zbc));
|
|
if (gr_zbc == NULL) {
|
|
return ret;
|
|
}
|
|
|
|
nvgpu_gr_zbc_init_indices(g, gr_zbc);
|
|
|
|
ret = gr_zbc_allocate_local_tbls(g, gr_zbc);
|
|
if (ret != 0) {
|
|
goto alloc_err;
|
|
}
|
|
|
|
nvgpu_gr_zbc_load_default_sw_table(g, gr_zbc);
|
|
|
|
*zbc = gr_zbc;
|
|
return ret;
|
|
|
|
alloc_err:
|
|
nvgpu_kfree(g, gr_zbc);
|
|
return ret;
|
|
}
|
|
|
|
/* deallocate the memory for the struct */
|
|
void nvgpu_gr_zbc_deinit(struct gk20a *g, struct nvgpu_gr_zbc *zbc)
|
|
{
|
|
if (zbc == NULL) {
|
|
return;
|
|
}
|
|
|
|
nvgpu_kfree(g, zbc->zbc_col_tbl);
|
|
nvgpu_kfree(g, zbc->zbc_dep_tbl);
|
|
nvgpu_kfree(g, zbc->zbc_s_tbl);
|
|
nvgpu_kfree(g, zbc);
|
|
}
|
|
|
|
struct nvgpu_gr_zbc_entry *nvgpu_gr_zbc_entry_alloc(struct gk20a *g)
|
|
{
|
|
return nvgpu_kzalloc(g, sizeof(struct nvgpu_gr_zbc_entry));
|
|
}
|
|
void nvgpu_gr_zbc_entry_free(struct gk20a *g, struct nvgpu_gr_zbc_entry *entry)
|
|
{
|
|
nvgpu_kfree(g, entry);
|
|
}
|
|
|
|
u32 nvgpu_gr_zbc_get_entry_color_ds(struct nvgpu_gr_zbc_entry *entry,
|
|
int idx)
|
|
{
|
|
return entry->color_ds[idx];
|
|
}
|
|
|
|
void nvgpu_gr_zbc_set_entry_color_ds(struct nvgpu_gr_zbc_entry *entry,
|
|
int idx, u32 ds)
|
|
{
|
|
entry->color_ds[idx] = ds;
|
|
}
|
|
|
|
u32 nvgpu_gr_zbc_get_entry_color_l2(struct nvgpu_gr_zbc_entry *entry,
|
|
int idx)
|
|
{
|
|
return entry->color_l2[idx];
|
|
}
|
|
|
|
void nvgpu_gr_zbc_set_entry_color_l2(struct nvgpu_gr_zbc_entry *entry,
|
|
int idx, u32 l2)
|
|
{
|
|
entry->color_l2[idx] = l2;
|
|
}
|
|
|
|
u32 nvgpu_gr_zbc_get_entry_depth(struct nvgpu_gr_zbc_entry *entry)
|
|
{
|
|
return entry->depth;
|
|
}
|
|
|
|
void nvgpu_gr_zbc_set_entry_depth(struct nvgpu_gr_zbc_entry *entry,
|
|
u32 depth)
|
|
{
|
|
entry->depth = depth;
|
|
}
|
|
|
|
u32 nvgpu_gr_zbc_get_entry_stencil(struct nvgpu_gr_zbc_entry *entry)
|
|
{
|
|
return entry->stencil;
|
|
}
|
|
|
|
void nvgpu_gr_zbc_set_entry_stencil(struct nvgpu_gr_zbc_entry *entry,
|
|
u32 stencil)
|
|
{
|
|
entry->stencil = stencil;
|
|
}
|
|
|
|
u32 nvgpu_gr_zbc_get_entry_type(struct nvgpu_gr_zbc_entry *entry)
|
|
{
|
|
return entry->type;
|
|
}
|
|
|
|
void nvgpu_gr_zbc_set_entry_type(struct nvgpu_gr_zbc_entry *entry,
|
|
u32 type)
|
|
{
|
|
entry->type = type;
|
|
}
|
|
|
|
u32 nvgpu_gr_zbc_get_entry_format(struct nvgpu_gr_zbc_entry *entry)
|
|
{
|
|
return entry->format;
|
|
}
|
|
|
|
void nvgpu_gr_zbc_set_entry_format(struct nvgpu_gr_zbc_entry *entry,
|
|
u32 format)
|
|
{
|
|
entry->format = format;
|
|
}
|