mirror of
https://github.com/NVIDIA/open-gpu-kernel-modules.git
synced 2026-03-02 11:49:55 +00:00
545.23.06
This commit is contained in:
@@ -64,6 +64,10 @@ extern "C" {
|
||||
|
||||
#define kbifIsSnoopDmaCapable(pGpu, pKernelBif) ((REF_VAL(BIF_DMA_CAPS_SNOOP, \
|
||||
kbifGetDmaCaps(pGpu, pKernelBif))))
|
||||
// Number of BAR entries: BAR0_LO, BAR0_HI, BAR1_LO, BAR1_HI, BAR2_LO, BAR2_HI and BAR3
|
||||
#define KBIF_NUM_BAR_OFFSET_ENTRIES 0x7
|
||||
// Indicates a non existent BAR offset
|
||||
#define KBIF_INVALID_BAR_REG_OFFSET 0xFFFF
|
||||
|
||||
// XVE bus options
|
||||
typedef enum BUS_OPTIONS
|
||||
@@ -74,6 +78,44 @@ typedef enum BUS_OPTIONS
|
||||
|
||||
} BUS_OPTIONS;
|
||||
|
||||
//
|
||||
// References to save/restore PCIe Config space using registry_map
|
||||
//
|
||||
typedef struct KBIF_XVE_REGMAP_REF
|
||||
{
|
||||
// XVE Valid register map array
|
||||
const NvU32 *xveRegMapValid;
|
||||
|
||||
// XVE Writable register map array
|
||||
const NvU32 *xveRegMapWrite;
|
||||
|
||||
// XVE Valid register map array count
|
||||
NvU16 numXveRegMapValid;
|
||||
|
||||
// XVE Writable register map array count
|
||||
NvU16 numXveRegMapWrite;
|
||||
|
||||
// PCIe function number
|
||||
NvU8 nFunc;
|
||||
|
||||
// Buffer to store boot PCIe config space
|
||||
NvU32 *bufBootConfigSpace;
|
||||
|
||||
// Buffer to store boot PCIe MSIX table (GH100+)
|
||||
NvU32 *bufMsixTable;
|
||||
|
||||
} KBIF_XVE_REGMAP_REF, *PKBIF_XVE_REGMAP_REF;
|
||||
|
||||
typedef struct
|
||||
{
|
||||
// Used to save/restore config space after a hot reset
|
||||
NvU32 gpuBootConfigSpace[PCIE_CONFIG_SPACE_SIZE/sizeof(NvU32)];
|
||||
NvU32 azaliaBootConfigSpace[PCIE_CONFIG_SPACE_SIZE/sizeof(NvU32)];
|
||||
|
||||
// Saved device control register value (Kepler+)
|
||||
NvU32 xveDevCtrl;
|
||||
} KBIF_CACHE_DATA;
|
||||
|
||||
typedef struct KERNEL_HOST_VGPU_DEVICE KERNEL_HOST_VGPU_DEVICE;
|
||||
|
||||
#ifdef NVOC_KERNEL_BIF_H_PRIVATE_ACCESS_ALLOWED
|
||||
@@ -112,10 +154,22 @@ struct KernelBif {
|
||||
void (*__kbifApplyWARBug3208922__)(struct OBJGPU *, struct KernelBif *);
|
||||
void (*__kbifProbePcieReqAtomicCaps__)(struct OBJGPU *, struct KernelBif *);
|
||||
void (*__kbifEnablePcieAtomics__)(struct OBJGPU *, struct KernelBif *);
|
||||
NV_STATUS (*__kbifInitXveRegMap__)(struct OBJGPU *, struct KernelBif *, NvU8);
|
||||
NvU32 (*__kbifGetMSIXTableVectorControlSize__)(struct OBJGPU *, struct KernelBif *);
|
||||
NV_STATUS (*__kbifGetPciConfigSpacePriMirror__)(struct OBJGPU *, struct KernelBif *, NvU32 *, NvU32 *);
|
||||
NV_STATUS (*__kbifGetBusOptionsAddr__)(struct OBJGPU *, struct KernelBif *, BUS_OPTIONS, NvU32 *);
|
||||
NV_STATUS (*__kbifPreOsGlobalErotGrantRequest__)(struct OBJGPU *, struct KernelBif *);
|
||||
NV_STATUS (*__kbifStopSysMemRequests__)(struct OBJGPU *, struct KernelBif *, NvBool);
|
||||
NV_STATUS (*__kbifWaitForTransactionsComplete__)(struct OBJGPU *, struct KernelBif *);
|
||||
NV_STATUS (*__kbifTriggerFlr__)(struct OBJGPU *, struct KernelBif *);
|
||||
void (*__kbifCacheFlrSupport__)(struct OBJGPU *, struct KernelBif *);
|
||||
void (*__kbifCache64bBar0Support__)(struct OBJGPU *, struct KernelBif *);
|
||||
void (*__kbifCacheVFInfo__)(struct OBJGPU *, struct KernelBif *);
|
||||
void (*__kbifRestoreBar0__)(struct OBJGPU *, struct KernelBif *, void *, NvU32 *);
|
||||
NvBool (*__kbifAnyBarsAreValid__)(struct OBJGPU *, struct KernelBif *);
|
||||
NV_STATUS (*__kbifRestoreBarsAndCommand__)(struct OBJGPU *, struct KernelBif *);
|
||||
void (*__kbifStoreBarRegOffsets__)(struct OBJGPU *, struct KernelBif *, NvU32);
|
||||
NV_STATUS (*__kbifInit__)(struct OBJGPU *, struct KernelBif *);
|
||||
NV_STATUS (*__kbifStatePreLoad__)(POBJGPU, struct KernelBif *, NvU32);
|
||||
NV_STATUS (*__kbifStatePostUnload__)(POBJGPU, struct KernelBif *, NvU32);
|
||||
void (*__kbifStateDestroy__)(POBJGPU, struct KernelBif *);
|
||||
@@ -140,14 +194,21 @@ struct KernelBif {
|
||||
NvBool PDB_PROP_KBIF_PCIE_GEN4_CAPABLE;
|
||||
NvBool PDB_PROP_KBIF_PCIE_RELAXED_ORDERING_SET_IN_EMULATED_CONFIG_SPACE;
|
||||
NvBool PDB_PROP_KBIF_SYSTEM_ACCESS_DISABLED;
|
||||
NvBool PDB_PROP_KBIF_FLR_SUPPORTED;
|
||||
NvBool PDB_PROP_KBIF_64BIT_BAR0_SUPPORTED;
|
||||
NvBool PDB_PROP_KBIF_DEVICE_IS_MULTIFUNCTION;
|
||||
NvBool PDB_PROP_KBIF_GCX_PMU_CFG_SPACE_RESTORE;
|
||||
NvU32 dmaCaps;
|
||||
RmPhysAddr dmaWindowStartAddress;
|
||||
NvU32 p2pOverride;
|
||||
NvU32 forceP2PType;
|
||||
NvBool peerMappingOverride;
|
||||
NvBool EnteredRecoverySinceErrorsLastChecked;
|
||||
KBIF_CACHE_DATA cacheData;
|
||||
NvU32 osPcieAtomicsOpMask;
|
||||
KBIF_XVE_REGMAP_REF xveRegmapRef[2];
|
||||
NvBool bMnocAvailable;
|
||||
NvU32 barRegOffsets[7];
|
||||
};
|
||||
|
||||
#ifndef __NVOC_CLASS_KernelBif_TYPEDEF__
|
||||
@@ -179,6 +240,8 @@ extern const struct NVOC_CLASS_DEF __nvoc_class_def_KernelBif;
|
||||
#define PDB_PROP_KBIF_IS_MSIX_ENABLED_BASE_NAME PDB_PROP_KBIF_IS_MSIX_ENABLED
|
||||
#define PDB_PROP_KBIF_P2P_WRITES_DISABLED_BASE_CAST
|
||||
#define PDB_PROP_KBIF_P2P_WRITES_DISABLED_BASE_NAME PDB_PROP_KBIF_P2P_WRITES_DISABLED
|
||||
#define PDB_PROP_KBIF_FLR_SUPPORTED_BASE_CAST
|
||||
#define PDB_PROP_KBIF_FLR_SUPPORTED_BASE_NAME PDB_PROP_KBIF_FLR_SUPPORTED
|
||||
#define PDB_PROP_KBIF_USE_CONFIG_SPACE_TO_REARM_MSI_BASE_CAST
|
||||
#define PDB_PROP_KBIF_USE_CONFIG_SPACE_TO_REARM_MSI_BASE_NAME PDB_PROP_KBIF_USE_CONFIG_SPACE_TO_REARM_MSI
|
||||
#define PDB_PROP_KBIF_IS_MSI_ENABLED_BASE_CAST
|
||||
@@ -187,6 +250,10 @@ extern const struct NVOC_CLASS_DEF __nvoc_class_def_KernelBif;
|
||||
#define PDB_PROP_KBIF_PCIE_RELAXED_ORDERING_SET_IN_EMULATED_CONFIG_SPACE_BASE_NAME PDB_PROP_KBIF_PCIE_RELAXED_ORDERING_SET_IN_EMULATED_CONFIG_SPACE
|
||||
#define PDB_PROP_KBIF_SYSTEM_ACCESS_DISABLED_BASE_CAST
|
||||
#define PDB_PROP_KBIF_SYSTEM_ACCESS_DISABLED_BASE_NAME PDB_PROP_KBIF_SYSTEM_ACCESS_DISABLED
|
||||
#define PDB_PROP_KBIF_GCX_PMU_CFG_SPACE_RESTORE_BASE_CAST
|
||||
#define PDB_PROP_KBIF_GCX_PMU_CFG_SPACE_RESTORE_BASE_NAME PDB_PROP_KBIF_GCX_PMU_CFG_SPACE_RESTORE
|
||||
#define PDB_PROP_KBIF_64BIT_BAR0_SUPPORTED_BASE_CAST
|
||||
#define PDB_PROP_KBIF_64BIT_BAR0_SUPPORTED_BASE_NAME PDB_PROP_KBIF_64BIT_BAR0_SUPPORTED
|
||||
#define PDB_PROP_KBIF_UPSTREAM_LTR_SUPPORT_WAR_BUG_200634944_BASE_CAST
|
||||
#define PDB_PROP_KBIF_UPSTREAM_LTR_SUPPORT_WAR_BUG_200634944_BASE_NAME PDB_PROP_KBIF_UPSTREAM_LTR_SUPPORT_WAR_BUG_200634944
|
||||
#define PDB_PROP_KBIF_IS_MSIX_CACHED_BASE_CAST
|
||||
@@ -201,6 +268,8 @@ extern const struct NVOC_CLASS_DEF __nvoc_class_def_KernelBif;
|
||||
#define PDB_PROP_KBIF_IS_FMODEL_MSI_BROKEN_BASE_NAME PDB_PROP_KBIF_IS_FMODEL_MSI_BROKEN
|
||||
#define PDB_PROP_KBIF_IS_MSI_CACHED_BASE_CAST
|
||||
#define PDB_PROP_KBIF_IS_MSI_CACHED_BASE_NAME PDB_PROP_KBIF_IS_MSI_CACHED
|
||||
#define PDB_PROP_KBIF_DEVICE_IS_MULTIFUNCTION_BASE_CAST
|
||||
#define PDB_PROP_KBIF_DEVICE_IS_MULTIFUNCTION_BASE_NAME PDB_PROP_KBIF_DEVICE_IS_MULTIFUNCTION
|
||||
#define PDB_PROP_KBIF_SUPPORT_NONCOHERENT_BASE_CAST
|
||||
#define PDB_PROP_KBIF_SUPPORT_NONCOHERENT_BASE_NAME PDB_PROP_KBIF_SUPPORT_NONCOHERENT
|
||||
|
||||
@@ -258,14 +327,38 @@ NV_STATUS __nvoc_objCreate_KernelBif(KernelBif**, Dynamic*, NvU32);
|
||||
#define kbifProbePcieReqAtomicCaps_HAL(pGpu, pKernelBif) kbifProbePcieReqAtomicCaps_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifEnablePcieAtomics(pGpu, pKernelBif) kbifEnablePcieAtomics_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifEnablePcieAtomics_HAL(pGpu, pKernelBif) kbifEnablePcieAtomics_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifInitXveRegMap(pGpu, pKernelBif, arg0) kbifInitXveRegMap_DISPATCH(pGpu, pKernelBif, arg0)
|
||||
#define kbifInitXveRegMap_HAL(pGpu, pKernelBif, arg0) kbifInitXveRegMap_DISPATCH(pGpu, pKernelBif, arg0)
|
||||
#define kbifGetMSIXTableVectorControlSize(pGpu, pKernelBif) kbifGetMSIXTableVectorControlSize_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifGetMSIXTableVectorControlSize_HAL(pGpu, pKernelBif) kbifGetMSIXTableVectorControlSize_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifGetPciConfigSpacePriMirror(pGpu, pKernelBif, pMirrorBase, pMirrorSize) kbifGetPciConfigSpacePriMirror_DISPATCH(pGpu, pKernelBif, pMirrorBase, pMirrorSize)
|
||||
#define kbifGetPciConfigSpacePriMirror_HAL(pGpu, pKernelBif, pMirrorBase, pMirrorSize) kbifGetPciConfigSpacePriMirror_DISPATCH(pGpu, pKernelBif, pMirrorBase, pMirrorSize)
|
||||
#define kbifGetBusOptionsAddr(pGpu, pKernelBif, options, addrReg) kbifGetBusOptionsAddr_DISPATCH(pGpu, pKernelBif, options, addrReg)
|
||||
#define kbifGetBusOptionsAddr_HAL(pGpu, pKernelBif, options, addrReg) kbifGetBusOptionsAddr_DISPATCH(pGpu, pKernelBif, options, addrReg)
|
||||
#define kbifPreOsGlobalErotGrantRequest(pGpu, pKernelBif) kbifPreOsGlobalErotGrantRequest_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifPreOsGlobalErotGrantRequest_HAL(pGpu, pKernelBif) kbifPreOsGlobalErotGrantRequest_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifStopSysMemRequests(pGpu, pKernelBif, bStop) kbifStopSysMemRequests_DISPATCH(pGpu, pKernelBif, bStop)
|
||||
#define kbifStopSysMemRequests_HAL(pGpu, pKernelBif, bStop) kbifStopSysMemRequests_DISPATCH(pGpu, pKernelBif, bStop)
|
||||
#define kbifWaitForTransactionsComplete(pGpu, pKernelBif) kbifWaitForTransactionsComplete_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifWaitForTransactionsComplete_HAL(pGpu, pKernelBif) kbifWaitForTransactionsComplete_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifTriggerFlr(pGpu, pKernelBif) kbifTriggerFlr_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifTriggerFlr_HAL(pGpu, pKernelBif) kbifTriggerFlr_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifCacheFlrSupport(pGpu, pKernelBif) kbifCacheFlrSupport_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifCacheFlrSupport_HAL(pGpu, pKernelBif) kbifCacheFlrSupport_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifCache64bBar0Support(pGpu, pKernelBif) kbifCache64bBar0Support_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifCache64bBar0Support_HAL(pGpu, pKernelBif) kbifCache64bBar0Support_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifCacheVFInfo(pGpu, pKernelBif) kbifCacheVFInfo_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifCacheVFInfo_HAL(pGpu, pKernelBif) kbifCacheVFInfo_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifRestoreBar0(pGpu, pKernelBif, arg0, arg1) kbifRestoreBar0_DISPATCH(pGpu, pKernelBif, arg0, arg1)
|
||||
#define kbifRestoreBar0_HAL(pGpu, pKernelBif, arg0, arg1) kbifRestoreBar0_DISPATCH(pGpu, pKernelBif, arg0, arg1)
|
||||
#define kbifAnyBarsAreValid(pGpu, pKernelBif) kbifAnyBarsAreValid_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifAnyBarsAreValid_HAL(pGpu, pKernelBif) kbifAnyBarsAreValid_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifRestoreBarsAndCommand(pGpu, pKernelBif) kbifRestoreBarsAndCommand_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifRestoreBarsAndCommand_HAL(pGpu, pKernelBif) kbifRestoreBarsAndCommand_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifStoreBarRegOffsets(pGpu, pKernelBif, arg0) kbifStoreBarRegOffsets_DISPATCH(pGpu, pKernelBif, arg0)
|
||||
#define kbifStoreBarRegOffsets_HAL(pGpu, pKernelBif, arg0) kbifStoreBarRegOffsets_DISPATCH(pGpu, pKernelBif, arg0)
|
||||
#define kbifInit(pGpu, pKernelBif) kbifInit_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifInit_HAL(pGpu, pKernelBif) kbifInit_DISPATCH(pGpu, pKernelBif)
|
||||
#define kbifStatePreLoad(pGpu, pEngstate, arg0) kbifStatePreLoad_DISPATCH(pGpu, pEngstate, arg0)
|
||||
#define kbifStatePostUnload(pGpu, pEngstate, arg0) kbifStatePostUnload_DISPATCH(pGpu, pEngstate, arg0)
|
||||
#define kbifStateDestroy(pGpu, pEngstate) kbifStateDestroy_DISPATCH(pGpu, pEngstate)
|
||||
@@ -275,6 +368,10 @@ NV_STATUS __nvoc_objCreate_KernelBif(KernelBif**, Dynamic*, NvU32);
|
||||
#define kbifStatePreInitLocked(pGpu, pEngstate) kbifStatePreInitLocked_DISPATCH(pGpu, pEngstate)
|
||||
#define kbifStatePreInitUnlocked(pGpu, pEngstate) kbifStatePreInitUnlocked_DISPATCH(pGpu, pEngstate)
|
||||
#define kbifIsPresent(pGpu, pEngstate) kbifIsPresent_DISPATCH(pGpu, pEngstate)
|
||||
void kbifDestruct_GM107(struct KernelBif *pKernelBif);
|
||||
|
||||
|
||||
#define __nvoc_kbifDestruct(pKernelBif) kbifDestruct_GM107(pKernelBif)
|
||||
static inline NvU32 kbifGetBusIntfType_2f2c74(struct KernelBif *pKernelBif) {
|
||||
return (3);
|
||||
}
|
||||
@@ -358,20 +455,20 @@ static inline NV_STATUS kbifDisableSysmemAccess(struct OBJGPU *pGpu, struct Kern
|
||||
|
||||
#define kbifDisableSysmemAccess_HAL(pGpu, pKernelBif, bDisable) kbifDisableSysmemAccess(pGpu, pKernelBif, bDisable)
|
||||
|
||||
static inline void kbifIsMnocSupported_b3696a(struct OBJGPU *pGpu, struct KernelBif *pKernelBif) {
|
||||
static inline void kbifCacheMnocSupport_b3696a(struct OBJGPU *pGpu, struct KernelBif *pKernelBif) {
|
||||
return;
|
||||
}
|
||||
|
||||
|
||||
#ifdef __nvoc_kernel_bif_h_disabled
|
||||
static inline void kbifIsMnocSupported(struct OBJGPU *pGpu, struct KernelBif *pKernelBif) {
|
||||
static inline void kbifCacheMnocSupport(struct OBJGPU *pGpu, struct KernelBif *pKernelBif) {
|
||||
NV_ASSERT_FAILED_PRECOMP("KernelBif was disabled!");
|
||||
}
|
||||
#else //__nvoc_kernel_bif_h_disabled
|
||||
#define kbifIsMnocSupported(pGpu, pKernelBif) kbifIsMnocSupported_b3696a(pGpu, pKernelBif)
|
||||
#define kbifCacheMnocSupport(pGpu, pKernelBif) kbifCacheMnocSupport_b3696a(pGpu, pKernelBif)
|
||||
#endif //__nvoc_kernel_bif_h_disabled
|
||||
|
||||
#define kbifIsMnocSupported_HAL(pGpu, pKernelBif) kbifIsMnocSupported(pGpu, pKernelBif)
|
||||
#define kbifCacheMnocSupport_HAL(pGpu, pKernelBif) kbifCacheMnocSupport(pGpu, pKernelBif)
|
||||
|
||||
NV_STATUS kbifConstructEngine_IMPL(struct OBJGPU *pGpu, struct KernelBif *pKernelBif, ENGDESCRIPTOR arg0);
|
||||
|
||||
@@ -583,6 +680,22 @@ static inline void kbifEnablePcieAtomics_DISPATCH(struct OBJGPU *pGpu, struct Ke
|
||||
pKernelBif->__kbifEnablePcieAtomics__(pGpu, pKernelBif);
|
||||
}
|
||||
|
||||
NV_STATUS kbifInitXveRegMap_TU102(struct OBJGPU *pGpu, struct KernelBif *pKernelBif, NvU8 arg0);
|
||||
|
||||
NV_STATUS kbifInitXveRegMap_GA102(struct OBJGPU *pGpu, struct KernelBif *pKernelBif, NvU8 arg0);
|
||||
|
||||
static inline NV_STATUS kbifInitXveRegMap_DISPATCH(struct OBJGPU *pGpu, struct KernelBif *pKernelBif, NvU8 arg0) {
|
||||
return pKernelBif->__kbifInitXveRegMap__(pGpu, pKernelBif, arg0);
|
||||
}
|
||||
|
||||
NvU32 kbifGetMSIXTableVectorControlSize_TU102(struct OBJGPU *pGpu, struct KernelBif *pKernelBif);
|
||||
|
||||
NvU32 kbifGetMSIXTableVectorControlSize_GH100(struct OBJGPU *pGpu, struct KernelBif *pKernelBif);
|
||||
|
||||
static inline NvU32 kbifGetMSIXTableVectorControlSize_DISPATCH(struct OBJGPU *pGpu, struct KernelBif *pKernelBif) {
|
||||
return pKernelBif->__kbifGetMSIXTableVectorControlSize__(pGpu, pKernelBif);
|
||||
}
|
||||
|
||||
NV_STATUS kbifGetPciConfigSpacePriMirror_GM107(struct OBJGPU *pGpu, struct KernelBif *pKernelBif, NvU32 *pMirrorBase, NvU32 *pMirrorSize);
|
||||
|
||||
NV_STATUS kbifGetPciConfigSpacePriMirror_GH100(struct OBJGPU *pGpu, struct KernelBif *pKernelBif, NvU32 *pMirrorBase, NvU32 *pMirrorSize);
|
||||
@@ -609,6 +722,50 @@ static inline NV_STATUS kbifPreOsGlobalErotGrantRequest_DISPATCH(struct OBJGPU *
|
||||
return pKernelBif->__kbifPreOsGlobalErotGrantRequest__(pGpu, pKernelBif);
|
||||
}
|
||||
|
||||
NV_STATUS kbifStopSysMemRequests_GM107(struct OBJGPU *pGpu, struct KernelBif *pKernelBif, NvBool bStop);
|
||||
|
||||
NV_STATUS kbifStopSysMemRequests_GH100(struct OBJGPU *pGpu, struct KernelBif *pKernelBif, NvBool bStop);
|
||||
|
||||
static inline NV_STATUS kbifStopSysMemRequests_DISPATCH(struct OBJGPU *pGpu, struct KernelBif *pKernelBif, NvBool bStop) {
|
||||
return pKernelBif->__kbifStopSysMemRequests__(pGpu, pKernelBif, bStop);
|
||||
}
|
||||
|
||||
NV_STATUS kbifWaitForTransactionsComplete_TU102(struct OBJGPU *pGpu, struct KernelBif *pKernelBif);
|
||||
|
||||
NV_STATUS kbifWaitForTransactionsComplete_GH100(struct OBJGPU *pGpu, struct KernelBif *pKernelBif);
|
||||
|
||||
static inline NV_STATUS kbifWaitForTransactionsComplete_DISPATCH(struct OBJGPU *pGpu, struct KernelBif *pKernelBif) {
|
||||
return pKernelBif->__kbifWaitForTransactionsComplete__(pGpu, pKernelBif);
|
||||
}
|
||||
|
||||
NV_STATUS kbifTriggerFlr_TU102(struct OBJGPU *pGpu, struct KernelBif *pKernelBif);
|
||||
|
||||
NV_STATUS kbifTriggerFlr_GH100(struct OBJGPU *pGpu, struct KernelBif *pKernelBif);
|
||||
|
||||
static inline NV_STATUS kbifTriggerFlr_DISPATCH(struct OBJGPU *pGpu, struct KernelBif *pKernelBif) {
|
||||
return pKernelBif->__kbifTriggerFlr__(pGpu, pKernelBif);
|
||||
}
|
||||
|
||||
void kbifCacheFlrSupport_TU102(struct OBJGPU *pGpu, struct KernelBif *pKernelBif);
|
||||
|
||||
void kbifCacheFlrSupport_GH100(struct OBJGPU *pGpu, struct KernelBif *pKernelBif);
|
||||
|
||||
static inline void kbifCacheFlrSupport_DISPATCH(struct OBJGPU *pGpu, struct KernelBif *pKernelBif) {
|
||||
pKernelBif->__kbifCacheFlrSupport__(pGpu, pKernelBif);
|
||||
}
|
||||
|
||||
static inline void kbifCache64bBar0Support_b3696a(struct OBJGPU *pGpu, struct KernelBif *pKernelBif) {
|
||||
return;
|
||||
}
|
||||
|
||||
void kbifCache64bBar0Support_GA100(struct OBJGPU *pGpu, struct KernelBif *pKernelBif);
|
||||
|
||||
void kbifCache64bBar0Support_GH100(struct OBJGPU *pGpu, struct KernelBif *pKernelBif);
|
||||
|
||||
static inline void kbifCache64bBar0Support_DISPATCH(struct OBJGPU *pGpu, struct KernelBif *pKernelBif) {
|
||||
pKernelBif->__kbifCache64bBar0Support__(pGpu, pKernelBif);
|
||||
}
|
||||
|
||||
void kbifCacheVFInfo_TU102(struct OBJGPU *pGpu, struct KernelBif *pKernelBif);
|
||||
|
||||
void kbifCacheVFInfo_GH100(struct OBJGPU *pGpu, struct KernelBif *pKernelBif);
|
||||
@@ -617,6 +774,50 @@ static inline void kbifCacheVFInfo_DISPATCH(struct OBJGPU *pGpu, struct KernelBi
|
||||
pKernelBif->__kbifCacheVFInfo__(pGpu, pKernelBif);
|
||||
}
|
||||
|
||||
void kbifRestoreBar0_GM107(struct OBJGPU *pGpu, struct KernelBif *pKernelBif, void *arg0, NvU32 *arg1);
|
||||
|
||||
void kbifRestoreBar0_GA100(struct OBJGPU *pGpu, struct KernelBif *pKernelBif, void *arg0, NvU32 *arg1);
|
||||
|
||||
static inline void kbifRestoreBar0_DISPATCH(struct OBJGPU *pGpu, struct KernelBif *pKernelBif, void *arg0, NvU32 *arg1) {
|
||||
pKernelBif->__kbifRestoreBar0__(pGpu, pKernelBif, arg0, arg1);
|
||||
}
|
||||
|
||||
NvBool kbifAnyBarsAreValid_GM107(struct OBJGPU *pGpu, struct KernelBif *pKernelBif);
|
||||
|
||||
NvBool kbifAnyBarsAreValid_GA100(struct OBJGPU *pGpu, struct KernelBif *pKernelBif);
|
||||
|
||||
static inline NvBool kbifAnyBarsAreValid_DISPATCH(struct OBJGPU *pGpu, struct KernelBif *pKernelBif) {
|
||||
return pKernelBif->__kbifAnyBarsAreValid__(pGpu, pKernelBif);
|
||||
}
|
||||
|
||||
NV_STATUS kbifRestoreBarsAndCommand_GM107(struct OBJGPU *pGpu, struct KernelBif *pKernelBif);
|
||||
|
||||
NV_STATUS kbifRestoreBarsAndCommand_GA100(struct OBJGPU *pGpu, struct KernelBif *pKernelBif);
|
||||
|
||||
NV_STATUS kbifRestoreBarsAndCommand_GH100(struct OBJGPU *pGpu, struct KernelBif *pKernelBif);
|
||||
|
||||
static inline NV_STATUS kbifRestoreBarsAndCommand_DISPATCH(struct OBJGPU *pGpu, struct KernelBif *pKernelBif) {
|
||||
return pKernelBif->__kbifRestoreBarsAndCommand__(pGpu, pKernelBif);
|
||||
}
|
||||
|
||||
void kbifStoreBarRegOffsets_GA100(struct OBJGPU *pGpu, struct KernelBif *pKernelBif, NvU32 arg0);
|
||||
|
||||
static inline void kbifStoreBarRegOffsets_b3696a(struct OBJGPU *pGpu, struct KernelBif *pKernelBif, NvU32 arg0) {
|
||||
return;
|
||||
}
|
||||
|
||||
static inline void kbifStoreBarRegOffsets_DISPATCH(struct OBJGPU *pGpu, struct KernelBif *pKernelBif, NvU32 arg0) {
|
||||
pKernelBif->__kbifStoreBarRegOffsets__(pGpu, pKernelBif, arg0);
|
||||
}
|
||||
|
||||
NV_STATUS kbifInit_GM107(struct OBJGPU *pGpu, struct KernelBif *pKernelBif);
|
||||
|
||||
NV_STATUS kbifInit_GH100(struct OBJGPU *pGpu, struct KernelBif *pKernelBif);
|
||||
|
||||
static inline NV_STATUS kbifInit_DISPATCH(struct OBJGPU *pGpu, struct KernelBif *pKernelBif) {
|
||||
return pKernelBif->__kbifInit__(pGpu, pKernelBif);
|
||||
}
|
||||
|
||||
static inline NV_STATUS kbifStatePreLoad_DISPATCH(POBJGPU pGpu, struct KernelBif *pEngstate, NvU32 arg0) {
|
||||
return pEngstate->__kbifStatePreLoad__(pGpu, pEngstate, arg0);
|
||||
}
|
||||
@@ -780,4 +981,5 @@ static inline NvU32 kbifGetDmaCaps(struct OBJGPU *pGpu, struct KernelBif *pKerne
|
||||
#ifdef __cplusplus
|
||||
} // extern "C"
|
||||
#endif
|
||||
|
||||
#endif // _G_KERNEL_BIF_NVOC_H_
|
||||
|
||||
Reference in New Issue
Block a user