mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-22 17:36:20 +03:00
gpu: nvgpu: create separate file for common.fb hals
Moved common.fb hals from gk20a.h to newly created file gops_fb.h. Jira NVGPU-4141 Change-Id: Ic1f96c942257d7f2c4bb437231626c7518c133b5 Signed-off-by: Seshendra Gadagottu <sgadagottu@nvidia.com> Reviewed-on: https://git-master.nvidia.com/r/2224601 Reviewed-by: mobile promotions <svcmobile_promotions@nvidia.com> Tested-by: mobile promotions <svcmobile_promotions@nvidia.com>
This commit is contained in:
committed by
Alex Waterman
parent
56359946aa
commit
83077ef4cd
@@ -628,7 +628,8 @@ mc:
|
||||
fb_fusa:
|
||||
safe: yes
|
||||
owner: Seshendra G
|
||||
sources: [ hal/fb/fb_gm20b_fusa.c,
|
||||
sources: [ include/nvgpu/gops_fb.h,
|
||||
hal/fb/fb_gm20b_fusa.c,
|
||||
hal/fb/fb_gm20b.h,
|
||||
hal/fb/fb_gv11b_fusa.c,
|
||||
hal/fb/fb_gv11b.h,
|
||||
|
||||
@@ -36,6 +36,7 @@
|
||||
* - @ref unit-mm
|
||||
* - @ref unit-fifo
|
||||
* - @ref unit-gr
|
||||
* - @ref unit-fb
|
||||
* - @ref unit-devctl
|
||||
* - @ref unit-sdl
|
||||
* - @ref unit-init
|
||||
@@ -153,6 +154,7 @@ enum nvgpu_unit;
|
||||
#include <nvgpu/gops_mm.h>
|
||||
#include <nvgpu/gops_priv_ring.h>
|
||||
#include <nvgpu/gops_therm.h>
|
||||
#include <nvgpu/gops_fb.h>
|
||||
|
||||
#include "hal/clk/clk_gk20a.h"
|
||||
|
||||
@@ -290,91 +292,8 @@ struct gpu_ops {
|
||||
bool (*is_valid_compute)(u32 class_num);
|
||||
} gpu_class;
|
||||
|
||||
struct {
|
||||
struct nvgpu_hw_err_inject_info_desc * (*get_hubmmu_err_desc)
|
||||
(struct gk20a *g);
|
||||
void (*init_hw)(struct gk20a *g);
|
||||
void (*init_fs_state)(struct gk20a *g);
|
||||
void (*init_uncompressed_kind_map)(struct gk20a *g);
|
||||
void (*init_kind_attr)(struct gk20a *g);
|
||||
void (*set_mmu_page_size)(struct gk20a *g);
|
||||
u32 (*mmu_ctrl)(struct gk20a *g);
|
||||
u32 (*mmu_debug_ctrl)(struct gk20a *g);
|
||||
u32 (*mmu_debug_wr)(struct gk20a *g);
|
||||
u32 (*mmu_debug_rd)(struct gk20a *g);
|
||||
struct gops_fb fb;
|
||||
|
||||
#ifdef CONFIG_NVGPU_COMPRESSION
|
||||
void (*cbc_configure)(struct gk20a *g, struct nvgpu_cbc *cbc);
|
||||
bool (*set_use_full_comp_tag_line)(struct gk20a *g);
|
||||
|
||||
/*
|
||||
* Compression tag line coverage. When mapping a compressible
|
||||
* buffer, ctagline is increased when the virtual address
|
||||
* crosses over the compression page boundary.
|
||||
*/
|
||||
u64 (*compression_page_size)(struct gk20a *g);
|
||||
|
||||
/*
|
||||
* Minimum page size that can be used for compressible kinds.
|
||||
*/
|
||||
unsigned int (*compressible_page_size)(struct gk20a *g);
|
||||
|
||||
/*
|
||||
* Compressible kind mappings: Mask for the virtual and physical
|
||||
* address bits that must match.
|
||||
*/
|
||||
u64 (*compression_align_mask)(struct gk20a *g);
|
||||
#endif
|
||||
|
||||
void (*dump_vpr_info)(struct gk20a *g);
|
||||
void (*dump_wpr_info)(struct gk20a *g);
|
||||
int (*vpr_info_fetch)(struct gk20a *g);
|
||||
void (*read_wpr_info)(struct gk20a *g, u64 *wpr_base, u64 *wpr_size);
|
||||
#ifdef CONFIG_NVGPU_DEBUGGER
|
||||
bool (*is_debug_mode_enabled)(struct gk20a *g);
|
||||
void (*set_debug_mode)(struct gk20a *g, bool enable);
|
||||
void (*set_mmu_debug_mode)(struct gk20a *g, bool enable);
|
||||
#endif
|
||||
int (*tlb_invalidate)(struct gk20a *g, struct nvgpu_mem *pdb);
|
||||
void (*handle_replayable_fault)(struct gk20a *g);
|
||||
int (*mem_unlock)(struct gk20a *g);
|
||||
int (*init_nvlink)(struct gk20a *g);
|
||||
int (*enable_nvlink)(struct gk20a *g);
|
||||
int (*init_fbpa)(struct gk20a *g);
|
||||
void (*handle_fbpa_intr)(struct gk20a *g, u32 fbpa_id);
|
||||
void (*write_mmu_fault_buffer_lo_hi)(struct gk20a *g, u32 index,
|
||||
u32 addr_lo, u32 addr_hi);
|
||||
void (*write_mmu_fault_buffer_get)(struct gk20a *g, u32 index,
|
||||
u32 reg_val);
|
||||
void (*write_mmu_fault_buffer_size)(struct gk20a *g, u32 index,
|
||||
u32 reg_val);
|
||||
void (*write_mmu_fault_status)(struct gk20a *g, u32 reg_val);
|
||||
u32 (*read_mmu_fault_buffer_get)(struct gk20a *g, u32 index);
|
||||
u32 (*read_mmu_fault_buffer_put)(struct gk20a *g, u32 index);
|
||||
u32 (*read_mmu_fault_buffer_size)(struct gk20a *g, u32 index);
|
||||
void (*read_mmu_fault_addr_lo_hi)(struct gk20a *g,
|
||||
u32 *addr_lo, u32 *addr_hi);
|
||||
void (*read_mmu_fault_inst_lo_hi)(struct gk20a *g,
|
||||
u32 *inst_lo, u32 *inst_hi);
|
||||
u32 (*read_mmu_fault_info)(struct gk20a *g);
|
||||
u32 (*read_mmu_fault_status)(struct gk20a *g);
|
||||
int (*mmu_invalidate_replay)(struct gk20a *g,
|
||||
u32 invalidate_replay_val);
|
||||
bool (*is_fault_buf_enabled)(struct gk20a *g, u32 index);
|
||||
void (*fault_buf_set_state_hw)(struct gk20a *g,
|
||||
u32 index, u32 state);
|
||||
void (*fault_buf_configure_hw)(struct gk20a *g, u32 index);
|
||||
#ifdef CONFIG_NVGPU_DGPU
|
||||
size_t (*get_vidmem_size)(struct gk20a *g);
|
||||
#endif
|
||||
int (*apply_pdb_cache_war)(struct gk20a *g);
|
||||
struct {
|
||||
void (*enable)(struct gk20a *g);
|
||||
void (*disable)(struct gk20a *g);
|
||||
void (*isr)(struct gk20a *g);
|
||||
bool (*is_mmu_fault_pending)(struct gk20a *g);
|
||||
} intr;
|
||||
} fb;
|
||||
struct {
|
||||
u32 (*falcon_base_addr)(void);
|
||||
} nvdec;
|
||||
|
||||
143
drivers/gpu/nvgpu/include/nvgpu/gops_fb.h
Normal file
143
drivers/gpu/nvgpu/include/nvgpu/gops_fb.h
Normal file
@@ -0,0 +1,143 @@
|
||||
/*
|
||||
* Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||||
* copy of this software and associated documentation files (the "Software"),
|
||||
* to deal in the Software without restriction, including without limitation
|
||||
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
||||
* and/or sell copies of the Software, and to permit persons to whom the
|
||||
* Software is furnished to do so, subject to the following conditions:
|
||||
*
|
||||
* The above copyright notice and this permission notice shall be included in
|
||||
* all copies or substantial portions of the Software.
|
||||
*
|
||||
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
||||
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
||||
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
||||
* DEALINGS IN THE SOFTWARE.
|
||||
*/
|
||||
#ifndef NVGPU_GOPS_FB_H
|
||||
#define NVGPU_GOPS_FB_H
|
||||
|
||||
#include <nvgpu/types.h>
|
||||
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* common.fb interface.
|
||||
*/
|
||||
struct gk20a;
|
||||
|
||||
/**
|
||||
* common.fb intr subunit hal operations.
|
||||
*
|
||||
* This structure stores common.fb interrupt subunit hal pointers.
|
||||
*
|
||||
* @see gpu_ops
|
||||
*/
|
||||
struct gops_fb_intr {
|
||||
void (*enable)(struct gk20a *g);
|
||||
void (*disable)(struct gk20a *g);
|
||||
void (*isr)(struct gk20a *g);
|
||||
bool (*is_mmu_fault_pending)(struct gk20a *g);
|
||||
};
|
||||
|
||||
/**
|
||||
* common.fb unit hal operations.
|
||||
*
|
||||
* This structure stores common.fb unit hal pointers.
|
||||
*
|
||||
* @see gpu_ops
|
||||
*/
|
||||
struct gops_fb {
|
||||
|
||||
void (*init_hw)(struct gk20a *g);
|
||||
void (*init_fs_state)(struct gk20a *g);
|
||||
void (*set_mmu_page_size)(struct gk20a *g);
|
||||
u32 (*mmu_ctrl)(struct gk20a *g);
|
||||
u32 (*mmu_debug_ctrl)(struct gk20a *g);
|
||||
u32 (*mmu_debug_wr)(struct gk20a *g);
|
||||
u32 (*mmu_debug_rd)(struct gk20a *g);
|
||||
void (*dump_vpr_info)(struct gk20a *g);
|
||||
void (*dump_wpr_info)(struct gk20a *g);
|
||||
int (*vpr_info_fetch)(struct gk20a *g);
|
||||
void (*read_wpr_info)(struct gk20a *g, u64 *wpr_base, u64 *wpr_size);
|
||||
int (*tlb_invalidate)(struct gk20a *g, struct nvgpu_mem *pdb);
|
||||
void (*fault_buf_configure_hw)(struct gk20a *g, u32 index);
|
||||
bool (*is_fault_buf_enabled)(struct gk20a *g, u32 index);
|
||||
void (*fault_buf_set_state_hw)(struct gk20a *g, u32 index, u32 state);
|
||||
u32 (*read_mmu_fault_buffer_get)(struct gk20a *g, u32 index);
|
||||
u32 (*read_mmu_fault_buffer_put)(struct gk20a *g, u32 index);
|
||||
u32 (*read_mmu_fault_buffer_size)(struct gk20a *g, u32 index);
|
||||
u32 (*read_mmu_fault_info)(struct gk20a *g);
|
||||
u32 (*read_mmu_fault_status)(struct gk20a *g);
|
||||
void (*write_mmu_fault_buffer_lo_hi)(struct gk20a *g, u32 index,
|
||||
u32 addr_lo, u32 addr_hi);
|
||||
void (*write_mmu_fault_buffer_get)(struct gk20a *g, u32 index,
|
||||
u32 reg_val);
|
||||
void (*write_mmu_fault_buffer_size)(struct gk20a *g, u32 index,
|
||||
u32 reg_val);
|
||||
void (*read_mmu_fault_addr_lo_hi)(struct gk20a *g,
|
||||
u32 *addr_lo, u32 *addr_hi);
|
||||
void (*read_mmu_fault_inst_lo_hi)(struct gk20a *g,
|
||||
u32 *inst_lo, u32 *inst_hi);
|
||||
void (*write_mmu_fault_status)(struct gk20a *g, u32 reg_val);
|
||||
int (*mmu_invalidate_replay)(struct gk20a *g,
|
||||
u32 invalidate_replay_val);
|
||||
|
||||
struct gops_fb_intr intr;
|
||||
|
||||
/** @cond DOXYGEN_SHOULD_SKIP_THIS */
|
||||
struct nvgpu_hw_err_inject_info_desc * (*get_hubmmu_err_desc)
|
||||
(struct gk20a *g);
|
||||
void (*init_uncompressed_kind_map)(struct gk20a *g);
|
||||
void (*init_kind_attr)(struct gk20a *g);
|
||||
|
||||
#ifdef CONFIG_NVGPU_COMPRESSION
|
||||
void (*cbc_configure)(struct gk20a *g, struct nvgpu_cbc *cbc);
|
||||
bool (*set_use_full_comp_tag_line)(struct gk20a *g);
|
||||
|
||||
/*
|
||||
* Compression tag line coverage. When mapping a compressible
|
||||
* buffer, ctagline is increased when the virtual address
|
||||
* crosses over the compression page boundary.
|
||||
*/
|
||||
u64 (*compression_page_size)(struct gk20a *g);
|
||||
|
||||
/*
|
||||
* Minimum page size that can be used for compressible kinds.
|
||||
*/
|
||||
unsigned int (*compressible_page_size)(struct gk20a *g);
|
||||
|
||||
/*
|
||||
* Compressible kind mappings: Mask for the virtual and physical
|
||||
* address bits that must match.
|
||||
*/
|
||||
u64 (*compression_align_mask)(struct gk20a *g);
|
||||
#endif
|
||||
|
||||
#ifdef CONFIG_NVGPU_DEBUGGER
|
||||
bool (*is_debug_mode_enabled)(struct gk20a *g);
|
||||
void (*set_debug_mode)(struct gk20a *g, bool enable);
|
||||
void (*set_mmu_debug_mode)(struct gk20a *g, bool enable);
|
||||
#endif
|
||||
|
||||
void (*handle_replayable_fault)(struct gk20a *g);
|
||||
int (*mem_unlock)(struct gk20a *g);
|
||||
int (*init_nvlink)(struct gk20a *g);
|
||||
int (*enable_nvlink)(struct gk20a *g);
|
||||
|
||||
#ifdef CONFIG_NVGPU_DGPU
|
||||
size_t (*get_vidmem_size)(struct gk20a *g);
|
||||
#endif
|
||||
int (*apply_pdb_cache_war)(struct gk20a *g);
|
||||
int (*init_fbpa)(struct gk20a *g);
|
||||
void (*handle_fbpa_intr)(struct gk20a *g, u32 fbpa_id);
|
||||
/** @endcond DOXYGEN_SHOULD_SKIP_THIS */
|
||||
};
|
||||
|
||||
#endif /* NVGPU_GOPS_FB_H */
|
||||
|
||||
Reference in New Issue
Block a user