// SPDX-License-Identifier: GPL-2.0 #include #include #include #include "nitrox_dev.h" #include "nitrox_csr.h" #include "nitrox_common.h" #define NR_RING_VECTORS 3 #define NPS_CORE_INT_ACTIVE_ENTRY 192 /** * nps_pkt_slc_isr - IRQ handler for NPS solicit port * @irq: irq number * @data: argument */ static irqreturn_t nps_pkt_slc_isr(int irq, void *data) { struct bh_data *slc = data; union nps_pkt_slc_cnts pkt_slc_cnts; pkt_slc_cnts.value = readq(slc->completion_cnt_csr_addr); /* New packet on SLC output port */ if (pkt_slc_cnts.s.slc_int) tasklet_hi_schedule(&slc->resp_handler); return IRQ_HANDLED; } static void clear_nps_core_err_intr(struct nitrox_device *ndev) { u64 value; /* Write 1 to clear */ value = nitrox_read_csr(ndev, NPS_CORE_INT); nitrox_write_csr(ndev, NPS_CORE_INT, value); dev_err_ratelimited(DEV(ndev), "NSP_CORE_INT 0x%016llx\n", value); } static void clear_nps_pkt_err_intr(struct nitrox_device *ndev) { union nps_pkt_int pkt_int; unsigned long value, offset; int i; pkt_int.value = nitrox_read_csr(ndev, NPS_PKT_INT); dev_err_ratelimited(DEV(ndev), "NPS_PKT_INT 0x%016llx\n", pkt_int.value); if (pkt_int.s.slc_err) { offset = NPS_PKT_SLC_ERR_TYPE; value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); dev_err_ratelimited(DEV(ndev), "NPS_PKT_SLC_ERR_TYPE 0x%016lx\n", value); offset = NPS_PKT_SLC_RERR_LO; value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); /* enable the solicit ports */ for_each_set_bit(i, &value, BITS_PER_LONG) enable_pkt_solicit_port(ndev, i); dev_err_ratelimited(DEV(ndev), "NPS_PKT_SLC_RERR_LO 0x%016lx\n", value); offset = NPS_PKT_SLC_RERR_HI; value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); dev_err_ratelimited(DEV(ndev), "NPS_PKT_SLC_RERR_HI 0x%016lx\n", value); } if (pkt_int.s.in_err) { offset = NPS_PKT_IN_ERR_TYPE; value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); dev_err_ratelimited(DEV(ndev), "NPS_PKT_IN_ERR_TYPE 0x%016lx\n", value); offset = NPS_PKT_IN_RERR_LO; value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); /* enable the input ring */ for_each_set_bit(i, &value, BITS_PER_LONG) enable_pkt_input_ring(ndev, i); dev_err_ratelimited(DEV(ndev), "NPS_PKT_IN_RERR_LO 0x%016lx\n", value); offset = NPS_PKT_IN_RERR_HI; value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); dev_err_ratelimited(DEV(ndev), "NPS_PKT_IN_RERR_HI 0x%016lx\n", value); } } static void clear_pom_err_intr(struct nitrox_device *ndev) { u64 value; value = nitrox_read_csr(ndev, POM_INT); nitrox_write_csr(ndev, POM_INT, value); dev_err_ratelimited(DEV(ndev), "POM_INT 0x%016llx\n", value); } static void clear_pem_err_intr(struct nitrox_device *ndev) { u64 value; value = nitrox_read_csr(ndev, PEM0_INT); nitrox_write_csr(ndev, PEM0_INT, value); dev_err_ratelimited(DEV(ndev), "PEM(0)_INT 0x%016llx\n", value); } static void clear_lbc_err_intr(struct nitrox_device *ndev) { union lbc_int lbc_int; u64 value, offset; int i; lbc_int.value = nitrox_read_csr(ndev, LBC_INT); dev_err_ratelimited(DEV(ndev), "LBC_INT 0x%016llx\n", lbc_int.value); if (lbc_int.s.dma_rd_err) { for (i = 0; i < NR_CLUSTERS; i++) { offset = EFL_CORE_VF_ERR_INT0X(i); value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); offset = EFL_CORE_VF_ERR_INT1X(i); value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); } } if (lbc_int.s.cam_soft_err) { dev_err_ratelimited(DEV(ndev), "CAM_SOFT_ERR, invalidating LBC\n"); invalidate_lbc(ndev); } if (lbc_int.s.pref_dat_len_mismatch_err) { offset = LBC_PLM_VF1_64_INT; value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); offset = LBC_PLM_VF65_128_INT; value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); } if (lbc_int.s.rd_dat_len_mismatch_err) { offset = LBC_ELM_VF1_64_INT; value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); offset = LBC_ELM_VF65_128_INT; value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); } nitrox_write_csr(ndev, LBC_INT, lbc_int.value); } static void clear_efl_err_intr(struct nitrox_device *ndev) { int i; for (i = 0; i < NR_CLUSTERS; i++) { union efl_core_int core_int; u64 value, offset; offset = EFL_CORE_INTX(i); core_int.value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, core_int.value); dev_err_ratelimited(DEV(ndev), "ELF_CORE(%d)_INT 0x%016llx\n", i, core_int.value); if (core_int.s.se_err) { offset = EFL_CORE_SE_ERR_INTX(i); value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); } } } static void clear_bmi_err_intr(struct nitrox_device *ndev) { u64 value; value = nitrox_read_csr(ndev, BMI_INT); nitrox_write_csr(ndev, BMI_INT, value); dev_err_ratelimited(DEV(ndev), "BMI_INT 0x%016llx\n", value); } /** * clear_nps_core_int_active - clear NPS_CORE_INT_ACTIVE interrupts * @ndev: NITROX device */ static void clear_nps_core_int_active(struct nitrox_device *ndev) { union nps_core_int_active core_int_active; core_int_active.value = nitrox_read_csr(ndev, NPS_CORE_INT_ACTIVE); if (core_int_active.s.nps_core) clear_nps_core_err_intr(ndev); if (core_int_active.s.nps_pkt) clear_nps_pkt_err_intr(ndev); if (core_int_active.s.pom) clear_pom_err_intr(ndev); if (core_int_active.s.pem) clear_pem_err_intr(ndev); if (core_int_active.s.lbc) clear_lbc_err_intr(ndev); if (core_int_active.s.efl) clear_efl_err_intr(ndev); if (core_int_active.s.bmi) clear_bmi_err_intr(ndev); /* If more work callback the ISR, set resend */ core_int_active.s.resend = 1; nitrox_write_csr(ndev, NPS_CORE_INT_ACTIVE, core_int_active.value); } static irqreturn_t nps_core_int_isr(int irq, void *data) { struct nitrox_device *ndev = data; clear_nps_core_int_active(ndev); return IRQ_HANDLED; } static int nitrox_enable_msix(struct nitrox_device *ndev) { struct msix_entry *entries; char **names; int i, nr_entries, ret; /* * PF MSI-X vectors * * Entry 0: NPS PKT ring 0 * Entry 1: AQMQ ring 0 * Entry 2: ZQM ring 0 * Entry 3: NPS PKT ring 1 * Entry 4: AQMQ ring 1 * Entry 5: ZQM ring 1 * .... * Entry 192: NPS_CORE_INT_ACTIVE */ nr_entries = (ndev->nr_queues * NR_RING_VECTORS) + 1; entries = kzalloc_node(nr_entries * sizeof(struct msix_entry), GFP_KERNEL, ndev->node); if (!entries) return -ENOMEM; names = kcalloc(nr_entries, sizeof(char *), GFP_KERNEL); if (!names) { kfree(entries); return -ENOMEM; } /* fill entires */ for (i = 0; i < (nr_entries - 1); i++) entries[i].entry = i; entries[i].entry = NPS_CORE_INT_ACTIVE_ENTRY; for (i = 0; i < nr_entries; i++) { *(names + i) = kzalloc(MAX_MSIX_VECTOR_NAME, GFP_KERNEL); if (!(*(names + i))) { ret = -ENOMEM; goto msix_fail; } } ndev->msix.entries = entries; ndev->msix.names = names; ndev->msix.nr_entries = nr_entries; ret = pci_enable_msix_exact(ndev->pdev, ndev->msix.entries, ndev->msix.nr_entries); if (ret) { dev_err(&ndev->pdev->dev, "Failed to enable MSI-X IRQ(s) %d\n", ret); goto msix_fail; } return 0; msix_fail: for (i = 0; i < nr_entries; i++) kfree(*(names + i)); kfree(entries); kfree(names); return ret; } static void nitrox_cleanup_pkt_slc_bh(struct nitrox_device *ndev) { int i; if (!ndev->bh.slc) return; for (i = 0; i < ndev->nr_queues; i++) { struct bh_data *bh = &ndev->bh.slc[i]; tasklet_disable(&bh->resp_handler); tasklet_kill(&bh->resp_handler); } kfree(ndev->bh.slc); ndev->bh.slc = NULL; } static int nitrox_setup_pkt_slc_bh(struct nitrox_device *ndev) { u32 size; int i; size = ndev->nr_queues * sizeof(struct bh_data); ndev->bh.slc = kzalloc(size, GFP_KERNEL); if (!ndev->bh.slc) return -ENOMEM; for (i = 0; i < ndev->nr_queues; i++) { struct bh_data *bh = &ndev->bh.slc[i]; u64 offset; offset = NPS_PKT_SLC_CNTSX(i); /* pre calculate completion count address */ bh->completion_cnt_csr_addr = NITROX_CSR_ADDR(ndev, offset); bh->cmdq = &ndev->pkt_cmdqs[i]; tasklet_init(&bh->resp_handler, pkt_slc_resp_handler, (unsigned long)bh); } return 0; } static int nitrox_request_irqs(struct nitrox_device *ndev) { struct pci_dev *pdev = ndev->pdev; struct msix_entry *msix_ent = ndev->msix.entries; int nr_ring_vectors, i = 0, ring, cpu, ret; char *name; /* * PF MSI-X vectors * * Entry 0: NPS PKT ring 0 * Entry 1: AQMQ ring 0 * Entry 2: ZQM ring 0 * Entry 3: NPS PKT ring 1 * .... * Entry 192: NPS_CORE_INT_ACTIVE */ nr_ring_vectors = ndev->nr_queues * NR_RING_VECTORS; /* request irq for pkt ring/ports only */ while (i < nr_ring_vectors) { name = *(ndev->msix.names + i); ring = (i / NR_RING_VECTORS); snprintf(name, MAX_MSIX_VECTOR_NAME, "n5(%d)-slc-ring%d", ndev->idx, ring); ret = request_irq(msix_ent[i].vector, nps_pkt_slc_isr, 0, name, &ndev->bh.slc[ring]); if (ret) { dev_err(&pdev->dev, "failed to get irq %d for %s\n", msix_ent[i].vector, name); return ret; } cpu = ring % num_online_cpus(); irq_set_affinity_hint(msix_ent[i].vector, get_cpu_mask(cpu)); set_bit(i, ndev->msix.irqs); i += NR_RING_VECTORS; } /* Request IRQ for NPS_CORE_INT_ACTIVE */ name = *(ndev->msix.names + i); snprintf(name, MAX_MSIX_VECTOR_NAME, "n5(%d)-nps-core-int", ndev->idx); ret = request_irq(msix_ent[i].vector, nps_core_int_isr, 0, name, ndev); if (ret) { dev_err(&pdev->dev, "failed to get irq %d for %s\n", msix_ent[i].vector, name); return ret; } set_bit(i, ndev->msix.irqs); return 0; } static void nitrox_disable_msix(struct nitrox_device *ndev) { struct msix_entry *msix_ent = ndev->msix.entries; char **names = ndev->msix.names; int i = 0, ring, nr_ring_vectors; nr_ring_vectors = ndev->msix.nr_entries - 1; /* clear pkt ring irqs */ while (i < nr_ring_vectors) { if (test_and_clear_bit(i, ndev->msix.irqs)) { ring = (i / NR_RING_VECTORS); irq_set_affinity_hint(msix_ent[i].vector, NULL); free_irq(msix_ent[i].vector, &ndev->bh.slc[ring]); } i += NR_RING_VECTORS; } irq_set_affinity_hint(msix_ent[i].vector, NULL); free_irq(msix_ent[i].vector, ndev); clear_bit(i, ndev->msix.irqs); kfree(ndev->msix.entries); for (i = 0; i < ndev->msix.nr_entries; i++) kfree(*(names + i)); kfree(names); pci_disable_msix(ndev->pdev); } /** * nitrox_pf_cleanup_isr: Cleanup PF MSI-X and IRQ * @ndev: NITROX device */ void nitrox_pf_cleanup_isr(struct nitrox_device *ndev) { nitrox_disable_msix(ndev); nitrox_cleanup_pkt_slc_bh(ndev); } /** * nitrox_init_isr - Initialize PF MSI-X vectors and IRQ * @ndev: NITROX device * * Return: 0 on success, a negative value on failure. */ int nitrox_pf_init_isr(struct nitrox_device *ndev) { int err; err = nitrox_setup_pkt_slc_bh(ndev); if (err) return err; err = nitrox_enable_msix(ndev); if (err) goto msix_fail; err = nitrox_request_irqs(ndev); if (err) goto irq_fail; return 0; irq_fail: nitrox_disable_msix(ndev); msix_fail: nitrox_cleanup_pkt_slc_bh(ndev); return err; }