/* * TU104 FBPA * * Copyright (c) 2018-2021, NVIDIA CORPORATION. All rights reserved. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included in * all copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER * DEALINGS IN THE SOFTWARE. */ #include #include #include #include #include #include #include #include #include "hal/fbpa/fbpa_tu104.h" int tu104_fbpa_init(struct gk20a *g) { u32 val; val = nvgpu_readl(g, fbpa_ecc_intr_ctrl_r()); val |= fbpa_ecc_intr_ctrl_sec_intr_en_enabled_f() | fbpa_ecc_intr_ctrl_ded_intr_en_enabled_f(); nvgpu_cic_mon_intr_stall_unit_config(g, NVGPU_CIC_INTR_UNIT_FBPA, NVGPU_CIC_INTR_ENABLE); nvgpu_writel(g, fbpa_ecc_intr_ctrl_r(), val); /* read back broadcast register */ (void) nvgpu_readl(g, fbpa_ecc_intr_ctrl_r()); return 0; } static void tu104_fbpa_handle_ecc_intr(struct gk20a *g, u32 fbpa_id, u32 subp_id) { u32 status, sec_cnt, ded_cnt; u32 offset = nvgpu_get_litter_value(g, GPU_LIT_FBPA_STRIDE) * fbpa_id; u32 cnt_idx = fbpa_id * 2U + subp_id; status = nvgpu_readl(g, offset + fbpa_0_ecc_status_r(subp_id)); if ((status & fbpa_0_ecc_status_sec_counter_overflow_pending_f()) != 0U) { nvgpu_err(g, "fbpa %u subp %u ecc sec counter overflow", fbpa_id, subp_id); } if ((status & fbpa_0_ecc_status_ded_counter_overflow_pending_f()) != 0U) { nvgpu_err(g, "fbpa %u subp %u ecc ded counter overflow", fbpa_id, subp_id); } if ((status & fbpa_0_ecc_status_sec_intr_pending_f()) != 0U) { sec_cnt = nvgpu_readl(g, offset + fbpa_0_ecc_sec_count_r(subp_id)); nvgpu_writel(g, offset + fbpa_0_ecc_sec_count_r(subp_id), 0u); g->ecc.fbpa.fbpa_ecc_sec_err_count[cnt_idx].counter += sec_cnt; } if ((status & fbpa_0_ecc_status_ded_intr_pending_f()) != 0U) { ded_cnt = nvgpu_readl(g, offset + fbpa_0_ecc_ded_count_r(subp_id)); nvgpu_writel(g, offset + fbpa_0_ecc_ded_count_r(subp_id), 0u); g->ecc.fbpa.fbpa_ecc_ded_err_count[cnt_idx].counter += ded_cnt; } nvgpu_writel(g, offset + fbpa_0_ecc_status_r(subp_id), status); } void tu104_fbpa_handle_intr(struct gk20a *g, u32 fbpa_id) { u32 offset, status; u32 ecc_subp0_mask = fbpa_0_intr_status_sec_subp0_pending_f() | fbpa_0_intr_status_ded_subp0_pending_f(); u32 ecc_subp1_mask = fbpa_0_intr_status_sec_subp1_pending_f() | fbpa_0_intr_status_ded_subp1_pending_f(); offset = nvgpu_get_litter_value(g, GPU_LIT_FBPA_STRIDE) * fbpa_id; status = nvgpu_readl(g, offset + fbpa_0_intr_status_r()); if ((status & (ecc_subp0_mask | ecc_subp1_mask)) == 0U) { nvgpu_err(g, "Unknown interrupt fbpa %u status %08x", fbpa_id, status); nvgpu_err(g, "Suspected unrecoverable EDC interrupt;" " HW no longer reliable"); nvgpu_sw_quiesce(g); return; } if ((status & ecc_subp0_mask) != 0U) { tu104_fbpa_handle_ecc_intr(g, fbpa_id, 0u); } if ((status & ecc_subp1_mask) != 0U) { tu104_fbpa_handle_ecc_intr(g, fbpa_id, 1u); } } int nvgpu_ecc_counter_init_per_fbpa(struct gk20a *g, struct nvgpu_ecc_stat **stat, const char *name) { u32 i; u32 num_fbpa = nvgpu_get_litter_value(g, GPU_LIT_NUM_FBPAS); struct nvgpu_ecc_stat *stats; char fbpa_str[10] = {0}; stats = nvgpu_kzalloc(g, nvgpu_safe_mult_u64(sizeof(*stats), (size_t)num_fbpa)); if (stats == NULL) { return -ENOMEM; } for (i = 0; i < num_fbpa; i++) { /** * Store stats name as below: * fbpa_ */ (void)strcpy(stats[i].name, "fbpa"); (void)nvgpu_strnadd_u32(fbpa_str, i, sizeof(fbpa_str), 10U); (void)strncat(stats[i].name, fbpa_str, NVGPU_ECC_STAT_NAME_MAX_SIZE - strlen(stats[i].name)); (void)strncat(stats[i].name, "_", NVGPU_ECC_STAT_NAME_MAX_SIZE - strlen(stats[i].name)); (void)strncat(stats[i].name, name, NVGPU_ECC_STAT_NAME_MAX_SIZE - strlen(stats[i].name)); nvgpu_ecc_stat_add(g, &stats[i]); } *stat = stats; return 0; } static void free_fbpa_ecc_stat_count_array(struct gk20a *g, struct nvgpu_ecc_stat **stats_p) { u32 num_fbpa = nvgpu_get_litter_value(g, GPU_LIT_NUM_FBPAS); struct nvgpu_ecc_stat *stats; u32 i; if (*stats_p != NULL) { stats = *stats_p; for (i = 0; i < num_fbpa; i++) { nvgpu_ecc_stat_del(g, &stats[i]); } nvgpu_kfree(g, stats); *stats_p = NULL; } } int tu104_fbpa_ecc_init(struct gk20a *g) { int err; err = NVGPU_ECC_COUNTER_INIT_PER_FBPA(fbpa_ecc_sec_err_count); if (err != 0) { goto done; } err = NVGPU_ECC_COUNTER_INIT_PER_FBPA(fbpa_ecc_ded_err_count); if (err != 0) { goto done; } done: if (err != 0) { nvgpu_err(g, "ecc counter allocate failed, err=%d", err); tu104_fbpa_ecc_free(g); } return err; } void tu104_fbpa_ecc_free(struct gk20a *g) { struct nvgpu_ecc *ecc = &g->ecc; free_fbpa_ecc_stat_count_array(g, &ecc->fbpa.fbpa_ecc_sec_err_count); free_fbpa_ecc_stat_count_array(g, &ecc->fbpa.fbpa_ecc_ded_err_count); }