// SPDX-License-Identifier: GPL-2.0-only /* * PCIe DMA test framework for Tegra PCIe. * * Copyright (c) 2021-2024 NVIDIA CORPORATION & AFFILIATES. All rights reserved. */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #define MODULENAME "pcie_dma_host" struct ep_pvt { struct pci_dev *pdev; /* Configurable BAR0/BAR2 virt and phy base addresses */ void __iomem *bar_virt; dma_addr_t bar_phy; /* DMA register BAR virt and phy base addresses */ void __iomem *dma_virt; phys_addr_t dma_phy_base; u32 dma_phy_size; /* dma_alloc_coherent() using RP pci_dev */ void *rp_dma_virt; dma_addr_t rp_dma_phy; /* dma_alloc_coherent() using EP pci_dev */ void *ep_dma_virt; dma_addr_t ep_dma_phy; struct dentry *debugfs; u32 dma_size; u32 stress_count; u32 edma_ch; u32 prev_edma_ch; u32 msi_irq; u64 msi_addr; u32 msi_data; u32 pmsi_irq; u64 pmsi_addr; u32 pmsi_data; u8 chip_id; struct edmalib_common edma; }; static irqreturn_t ep_isr(int irq, void *arg) { struct ep_pvt *ep = (struct ep_pvt *)arg; struct pcie_epf_bar *epf_bar = (__force struct pcie_epf_bar *)ep->bar_virt; struct sanity_data *wr_data = &epf_bar->wr_data[0]; wr_data->crc = crc32_le(~0, ep->ep_dma_virt + BAR0_DMA_BUF_OFFSET + wr_data->dst_offset, wr_data->size); return IRQ_HANDLED; } static void tegra_pcie_dma_raise_irq(void *p) { pr_err("%s: donot support raise IRQ from RP. CRC test if any started may fail.\n", __func__); } /* debugfs to perform eDMA lib transfers */ static int edmalib_test(struct seq_file *s, void *data) { struct ep_pvt *ep = (struct ep_pvt *)dev_get_drvdata(s->private); struct pcie_epf_bar *epf_bar = (__force struct pcie_epf_bar *)ep->bar_virt; struct pci_dev *pdev = ep->pdev; struct edmalib_common *edma = &ep->edma; struct pci_dev *ppdev = pcie_find_root_port(pdev); ep->edma.fdev = &ep->pdev->dev; ep->edma.epf_bar = epf_bar; ep->edma.bar_phy = ep->bar_phy; ep->edma.dma_virt = ep->dma_virt; ep->edma.priv = (void *)ep; ep->edma.raise_irq = tegra_pcie_dma_raise_irq; /* RP uses "Base + SZ_16M + 1M(reserved)" offset for DMA data transfers */ if (REMOTE_EDMA_TEST_EN) { ep->edma.src_virt = ep->ep_dma_virt + SZ_16M + SZ_1M; ep->edma.src_dma_addr = ep->ep_dma_phy + SZ_16M + SZ_1M; ep->edma.dst_dma_addr = epf_bar->ep_phy_addr + SZ_16M + SZ_1M; ep->edma.msi_addr = ep->msi_addr; ep->edma.msi_data = ep->msi_data; ep->edma.msi_irq = ep->msi_irq; ep->edma.cdev = &pdev->dev; ep->edma.remote.dma_phy_base = ep->dma_phy_base; ep->edma.remote.dma_size = ep->dma_phy_size; } else { ep->edma.src_dma_addr = ep->rp_dma_phy + SZ_16M + SZ_1M; ep->edma.src_virt = ep->rp_dma_virt + SZ_16M + SZ_1M; ep->edma.dst_dma_addr = ep->bar_phy + SZ_16M + SZ_1M; ep->edma.msi_addr = ep->pmsi_addr; ep->edma.msi_data = ep->pmsi_data; ep->edma.msi_irq = ep->pmsi_irq; ep->edma.cdev = &ppdev->dev; } return edmalib_common_test(&ep->edma); } static void init_debugfs(struct ep_pvt *ep) { debugfs_create_devm_seqfile(&ep->pdev->dev, "edmalib_test", ep->debugfs, edmalib_test); debugfs_create_u32("edma_ch", 0644, ep->debugfs, &ep->edma.edma_ch); /* Enable remote dma ASYNC for ch 0 as default */ ep->edma.edma_ch = 0x80000011; ep->edma.st_as_ch = -1; debugfs_create_u32("stress_count", 0644, ep->debugfs, &ep->edma.stress_count); ep->edma.stress_count = 10; debugfs_create_u32("dma_size", 0644, ep->debugfs, &ep->edma.dma_size); ep->edma.dma_size = SZ_1M; debugfs_create_u32("nents", 0644, ep->debugfs, &ep->edma.nents); /* Set DMA_LL_DEFAULT_SIZE as default nents, Max NUM_EDMA_DESC */ ep->edma.nents = DMA_LL_DEFAULT_SIZE; } static int ep_test_dma_probe(struct pci_dev *pdev, const struct pci_device_id *id) { struct ep_pvt *ep; struct pcie_epf_bar *epf_bar; struct pci_dev *ppdev = pcie_find_root_port(pdev); int ret = 0; u32 val, i, bar, dma_bar; u16 val_16; char *name; ep = devm_kzalloc(&pdev->dev, sizeof(*ep), GFP_KERNEL); if (!ep) return -ENOMEM; ep->chip_id = __tegra_get_chip_id(); if (ep->chip_id == TEGRA234) ep->edma.chip_id = NVPCIE_DMA_SOC_T234; else ep->edma.chip_id = NVPCIE_DMA_SOC_T264; ep->edma.ll_desc = devm_kzalloc(&pdev->dev, sizeof(*ep->edma.ll_desc) * NUM_EDMA_DESC, GFP_KERNEL); if (!ep->edma.ll_desc) return -ENOMEM; ep->pdev = pdev; pci_set_drvdata(pdev, ep); ret = pci_enable_device(pdev); if (ret < 0) { dev_err(&pdev->dev, "Failed to enable PCI device\n"); return ret; } #if defined(NV_PCI_ENABLE_PCIE_ERROR_REPORTING_PRESENT) /* Linux 6.5 */ pci_enable_pcie_error_reporting(pdev); #endif pci_set_master(pdev); ret = pci_request_regions(pdev, MODULENAME); if (ret < 0) { dev_err(&pdev->dev, "Failed to request PCI regions\n"); goto fail_region_request; } if (ep->chip_id == TEGRA234) bar = 0; else bar = 2; ep->bar_phy = pci_resource_start(pdev, bar); ep->bar_virt = devm_ioremap_wc(&pdev->dev, ep->bar_phy, pci_resource_len(pdev, bar)); if (!ep->bar_virt) { dev_err(&pdev->dev, "Failed to IO remap BAR%d\n", bar); ret = -ENOMEM; goto fail_region_remap; } if (ep->chip_id == TEGRA234) dma_bar = 4; else dma_bar = 0; ep->dma_phy_base = pci_resource_start(pdev, dma_bar); ep->dma_phy_size = pci_resource_len(pdev, dma_bar); ep->dma_virt = devm_ioremap(&pdev->dev, ep->dma_phy_base, ep->dma_phy_size); if (!ep->dma_virt) { dev_err(&pdev->dev, "Failed to IO remap BAR%d\n", dma_bar); ret = -ENOMEM; goto fail_region_remap; } ret = pci_alloc_irq_vectors(pdev, 16, 16, PCI_IRQ_MSI); if (ret < 0) { dev_err(&pdev->dev, "Failed to enable MSI interrupt\n"); ret = -ENODEV; goto fail_region_remap; } ret = request_irq(pci_irq_vector(pdev, 1), ep_isr, IRQF_SHARED, "pcie_ep_isr", ep); if (ret < 0) { dev_err(&pdev->dev, "Failed to register isr\n"); goto fail_isr; } ep->rp_dma_virt = dma_alloc_coherent(&ppdev->dev, BAR0_SIZE, &ep->rp_dma_phy, GFP_KERNEL); if (!ep->rp_dma_virt) { dev_err(&pdev->dev, "Failed to allocate DMA memory\n"); ret = -ENOMEM; goto fail_rp_dma_alloc; } get_random_bytes(ep->rp_dma_virt, BAR0_SIZE); dev_info(&ppdev->dev, "DMA mem ppdev, IOVA: 0x%llx size: %d\n", ep->rp_dma_phy, BAR0_SIZE); ep->ep_dma_virt = dma_alloc_coherent(&pdev->dev, BAR0_SIZE, &ep->ep_dma_phy, GFP_KERNEL); if (!ep->ep_dma_virt) { dev_err(&pdev->dev, "Failed to allocate DMA memory for EP\n"); ret = -ENOMEM; goto fail_ep_dma_alloc; } get_random_bytes(ep->ep_dma_virt, BAR0_SIZE); dev_info(&pdev->dev, "DMA mem pdev, IOVA: 0x%llx size: %d\n", ep->ep_dma_phy, BAR0_SIZE); /* Update RP DMA system memory base address allocated with EP pci_dev in BAR0 */ epf_bar = (__force struct pcie_epf_bar *)ep->bar_virt; epf_bar->rp_phy_addr = ep->ep_dma_phy; pci_read_config_word(pdev, pdev->msi_cap + PCI_MSI_FLAGS, &val_16); if (val_16 & PCI_MSI_FLAGS_64BIT) { pci_read_config_dword(pdev, pdev->msi_cap + PCI_MSI_ADDRESS_HI, &val); ep->msi_addr = val; pci_read_config_word(pdev, pdev->msi_cap + PCI_MSI_DATA_64, &val_16); ep->msi_data = val_16; } else { pci_read_config_word(pdev, pdev->msi_cap + PCI_MSI_DATA_32, &val_16); ep->msi_data = val_16; } pci_read_config_dword(pdev, pdev->msi_cap + PCI_MSI_ADDRESS_LO, &val); ep->msi_addr = (ep->msi_addr << 32) | val; ep->msi_irq = pci_irq_vector(pdev, 0); pci_read_config_word(ppdev, ppdev->msi_cap + PCI_MSI_FLAGS, &val_16); if (val_16 & PCI_MSI_FLAGS_64BIT) { pci_read_config_dword(ppdev, ppdev->msi_cap + PCI_MSI_ADDRESS_HI, &val); ep->pmsi_addr = val; pci_read_config_word(ppdev, ppdev->msi_cap + PCI_MSI_DATA_64, &val_16); ep->pmsi_data = val_16; } else { pci_read_config_word(ppdev, ppdev->msi_cap + PCI_MSI_DATA_32, &val_16); ep->pmsi_data = val_16; } pci_read_config_dword(ppdev, ppdev->msi_cap + PCI_MSI_ADDRESS_LO, &val); ep->pmsi_addr = (ep->pmsi_addr << 32) | val; ep->pmsi_irq = pci_irq_vector(ppdev, 0); ep->pmsi_data += 0; name = devm_kasprintf(&ep->pdev->dev, GFP_KERNEL, "%s_pcie_dma_test", dev_name(&pdev->dev)); if (!name) { dev_err(&pdev->dev, "%s: Fail to set debugfs name\n", __func__); ret = -ENOMEM; goto fail_name; } for (i = 0; i < TEGRA_PCIE_DMA_WRITE; i++) init_waitqueue_head(&ep->edma.wr_wq[i]); ep->debugfs = debugfs_create_dir(name, NULL); init_debugfs(ep); return ret; fail_name: dma_free_coherent(&pdev->dev, BAR0_SIZE, ep->ep_dma_virt, ep->ep_dma_phy); fail_ep_dma_alloc: dma_free_coherent(&ppdev->dev, BAR0_SIZE, ep->rp_dma_virt, ep->rp_dma_phy); fail_rp_dma_alloc: free_irq(pci_irq_vector(pdev, 1), ep); fail_isr: pci_free_irq_vectors(pdev); fail_region_remap: pci_release_regions(pdev); fail_region_request: pci_clear_master(pdev); return ret; } static void ep_test_dma_remove(struct pci_dev *pdev) { struct ep_pvt *ep = pci_get_drvdata(pdev); struct pci_dev *ppdev = pcie_find_root_port(pdev); debugfs_remove_recursive(ep->debugfs); tegra_pcie_dma_deinit(&ep->edma.cookie); dma_free_coherent(&pdev->dev, BAR0_SIZE, ep->ep_dma_virt, ep->ep_dma_phy); dma_free_coherent(&ppdev->dev, BAR0_SIZE, ep->rp_dma_virt, ep->rp_dma_phy); free_irq(pci_irq_vector(pdev, 1), ep); pci_free_irq_vectors(pdev); pci_release_regions(pdev); pci_clear_master(pdev); } static const struct pci_device_id ep_pci_tbl[] = { { PCI_DEVICE(0x10DE, 0x1AD4)}, { PCI_DEVICE(0x10DE, 0x1AD5)}, { PCI_DEVICE(0x10DE, 0x229a)}, {}, }; MODULE_DEVICE_TABLE(pci, ep_pci_tbl); static struct pci_driver ep_pci_driver = { .name = MODULENAME, .id_table = ep_pci_tbl, .probe = ep_test_dma_probe, .remove = ep_test_dma_remove, }; module_pci_driver(ep_pci_driver); MODULE_DESCRIPTION("Tegra PCIe client driver for endpoint DMA test func"); MODULE_LICENSE("GPL v2"); MODULE_AUTHOR("Manikanta Maddireddy ");