// SPDX-License-Identifier: GPL-2.0 /* Copyright (C) 2023, Intel Corporation. */ #include "ice.h" #include "ice_lib.h" #include "ice_irq.h" /** * ice_init_irq_tracker - initialize interrupt tracker * @pf: board private structure * @max_vectors: maximum number of vectors that tracker can hold * @num_static: number of preallocated interrupts */ static void ice_init_irq_tracker(struct ice_pf *pf, unsigned int max_vectors, unsigned int num_static) { pf->irq_tracker.num_entries = max_vectors; pf->irq_tracker.num_static = num_static; xa_init_flags(&pf->irq_tracker.entries, XA_FLAGS_ALLOC); } /** * ice_deinit_irq_tracker - free xarray tracker * @pf: board private structure */ static void ice_deinit_irq_tracker(struct ice_pf *pf) { xa_destroy(&pf->irq_tracker.entries); } /** * ice_free_irq_res - free a block of resources * @pf: board private structure * @index: starting index previously returned by ice_get_res */ static void ice_free_irq_res(struct ice_pf *pf, u16 index) { struct ice_irq_entry *entry; entry = xa_erase(&pf->irq_tracker.entries, index); kfree(entry); } /** * ice_get_irq_res - get an interrupt resource * @pf: board private structure * @dyn_only: force entry to be dynamically allocated * * Allocate new irq entry in the free slot of the tracker. Since xarray * is used, always allocate new entry at the lowest possible index. Set * proper allocation limit for maximum tracker entries. * * Returns allocated irq entry or NULL on failure. */ static struct ice_irq_entry *ice_get_irq_res(struct ice_pf *pf, bool dyn_only) { struct xa_limit limit = { .max = pf->irq_tracker.num_entries, .min = 0 }; unsigned int num_static = pf->irq_tracker.num_static; struct ice_irq_entry *entry; unsigned int index; int ret; entry = kzalloc(sizeof(*entry), GFP_KERNEL); if (!entry) return NULL; /* skip preallocated entries if the caller says so */ if (dyn_only) limit.min = num_static; ret = xa_alloc(&pf->irq_tracker.entries, &index, entry, limit, GFP_KERNEL); if (ret) { kfree(entry); entry = NULL; } else { entry->index = index; entry->dynamic = index >= num_static; } return entry; } /** * ice_reduce_msix_usage - Reduce usage of MSI-X vectors * @pf: board private structure * @v_remain: number of remaining MSI-X vectors to be distributed * * Reduce the usage of MSI-X vectors when entire request cannot be fulfilled. * pf->num_lan_msix and pf->num_rdma_msix values are set based on number of * remaining vectors. */ static void ice_reduce_msix_usage(struct ice_pf *pf, int v_remain) { int v_rdma; if (!ice_is_rdma_ena(pf)) { pf->num_lan_msix = v_remain; return; } /* RDMA needs at least 1 interrupt in addition to AEQ MSIX */ v_rdma = ICE_RDMA_NUM_AEQ_MSIX + 1; if (v_remain < ICE_MIN_LAN_TXRX_MSIX + ICE_MIN_RDMA_MSIX) { dev_warn(ice_pf_to_dev(pf), "Not enough MSI-X vectors to support RDMA.\n"); clear_bit(ICE_FLAG_RDMA_ENA, pf->flags); pf->num_rdma_msix = 0; pf->num_lan_msix = ICE_MIN_LAN_TXRX_MSIX; } else if ((v_remain < ICE_MIN_LAN_TXRX_MSIX + v_rdma) || (v_remain - v_rdma < v_rdma)) { /* Support minimum RDMA and give remaining vectors to LAN MSIX */ pf->num_rdma_msix = ICE_MIN_RDMA_MSIX; pf->num_lan_msix = v_remain - ICE_MIN_RDMA_MSIX; } else { /* Split remaining MSIX with RDMA after accounting for AEQ MSIX */ pf->num_rdma_msix = (v_remain - ICE_RDMA_NUM_AEQ_MSIX) / 2 + ICE_RDMA_NUM_AEQ_MSIX; pf->num_lan_msix = v_remain - pf->num_rdma_msix; } } /** * ice_ena_msix_range - Request a range of MSIX vectors from the OS * @pf: board private structure * * Compute the number of MSIX vectors wanted and request from the OS. Adjust * device usage if there are not enough vectors. Return the number of vectors * reserved or negative on failure. */ static int ice_ena_msix_range(struct ice_pf *pf) { int num_cpus, hw_num_msix, v_other, v_wanted, v_actual; struct device *dev = ice_pf_to_dev(pf); int err; hw_num_msix = pf->hw.func_caps.common_cap.num_msix_vectors; num_cpus = num_online_cpus(); /* LAN miscellaneous handler */ v_other = ICE_MIN_LAN_OICR_MSIX; /* Flow Director */ if (test_bit(ICE_FLAG_FD_ENA, pf->flags)) v_other += ICE_FDIR_MSIX; /* switchdev */ v_other += ICE_ESWITCH_MSIX; v_wanted = v_other; /* LAN traffic */ pf->num_lan_msix = num_cpus; v_wanted += pf->num_lan_msix; /* RDMA auxiliary driver */ if (ice_is_rdma_ena(pf)) { pf->num_rdma_msix = num_cpus + ICE_RDMA_NUM_AEQ_MSIX; v_wanted += pf->num_rdma_msix; } if (v_wanted > hw_num_msix) { int v_remain; dev_warn(dev, "not enough device MSI-X vectors. wanted = %d, available = %d\n", v_wanted, hw_num_msix); if (hw_num_msix < ICE_MIN_MSIX) { err = -ERANGE; goto exit_err; } v_remain = hw_num_msix - v_other; if (v_remain < ICE_MIN_LAN_TXRX_MSIX) { v_other = ICE_MIN_MSIX - ICE_MIN_LAN_TXRX_MSIX; v_remain = ICE_MIN_LAN_TXRX_MSIX; } ice_reduce_msix_usage(pf, v_remain); v_wanted = pf->num_lan_msix + pf->num_rdma_msix + v_other; dev_notice(dev, "Reducing request to %d MSI-X vectors for LAN traffic.\n", pf->num_lan_msix); if (ice_is_rdma_ena(pf)) dev_notice(dev, "Reducing request to %d MSI-X vectors for RDMA.\n", pf->num_rdma_msix); } /* actually reserve the vectors */ v_actual = pci_alloc_irq_vectors(pf->pdev, ICE_MIN_MSIX, v_wanted, PCI_IRQ_MSIX); if (v_actual < 0) { dev_err(dev, "unable to reserve MSI-X vectors\n"); err = v_actual; goto exit_err; } if (v_actual < v_wanted) { dev_warn(dev, "not enough OS MSI-X vectors. requested = %d, obtained = %d\n", v_wanted, v_actual); if (v_actual < ICE_MIN_MSIX) { /* error if we can't get minimum vectors */ pci_free_irq_vectors(pf->pdev); err = -ERANGE; goto exit_err; } else { int v_remain = v_actual - v_other; if (v_remain < ICE_MIN_LAN_TXRX_MSIX) v_remain = ICE_MIN_LAN_TXRX_MSIX; ice_reduce_msix_usage(pf, v_remain); dev_notice(dev, "Enabled %d MSI-X vectors for LAN traffic.\n", pf->num_lan_msix); if (ice_is_rdma_ena(pf)) dev_notice(dev, "Enabled %d MSI-X vectors for RDMA.\n", pf->num_rdma_msix); } } return v_actual; exit_err: pf->num_rdma_msix = 0; pf->num_lan_msix = 0; return err; } /** * ice_clear_interrupt_scheme - Undo things done by ice_init_interrupt_scheme * @pf: board private structure */ void ice_clear_interrupt_scheme(struct ice_pf *pf) { pci_free_irq_vectors(pf->pdev); ice_deinit_irq_tracker(pf); } /** * ice_init_interrupt_scheme - Determine proper interrupt scheme * @pf: board private structure to initialize */ int ice_init_interrupt_scheme(struct ice_pf *pf) { int total_vectors = pf->hw.func_caps.common_cap.num_msix_vectors; int vectors, max_vectors; vectors = ice_ena_msix_range(pf); if (vectors < 0) return -ENOMEM; if (pci_msix_can_alloc_dyn(pf->pdev)) max_vectors = total_vectors; else max_vectors = vectors; ice_init_irq_tracker(pf, max_vectors, vectors); return 0; } /** * ice_alloc_irq - Allocate new interrupt vector * @pf: board private structure * @dyn_only: force dynamic allocation of the interrupt * * Allocate new interrupt vector for a given owner id. * return struct msi_map with interrupt details and track * allocated interrupt appropriately. * * This function reserves new irq entry from the irq_tracker. * if according to the tracker information all interrupts that * were allocated with ice_pci_alloc_irq_vectors are already used * and dynamically allocated interrupts are supported then new * interrupt will be allocated with pci_msix_alloc_irq_at. * * Some callers may only support dynamically allocated interrupts. * This is indicated with dyn_only flag. * * On failure, return map with negative .index. The caller * is expected to check returned map index. * */ struct msi_map ice_alloc_irq(struct ice_pf *pf, bool dyn_only) { int sriov_base_vector = pf->sriov_base_vector; struct msi_map map = { .index = -ENOENT }; struct device *dev = ice_pf_to_dev(pf); struct ice_irq_entry *entry; entry = ice_get_irq_res(pf, dyn_only); if (!entry) return map; /* fail if we're about to violate SRIOV vectors space */ if (sriov_base_vector && entry->index >= sriov_base_vector) goto exit_free_res; if (pci_msix_can_alloc_dyn(pf->pdev) && entry->dynamic) { map = pci_msix_alloc_irq_at(pf->pdev, entry->index, NULL); if (map.index < 0) goto exit_free_res; dev_dbg(dev, "allocated new irq at index %d\n", map.index); } else { map.index = entry->index; map.virq = pci_irq_vector(pf->pdev, map.index); } return map; exit_free_res: dev_err(dev, "Could not allocate irq at idx %d\n", entry->index); ice_free_irq_res(pf, entry->index); return map; } /** * ice_free_irq - Free interrupt vector * @pf: board private structure * @map: map with interrupt details * * Remove allocated interrupt from the interrupt tracker. If interrupt was * allocated dynamically, free respective interrupt vector. */ void ice_free_irq(struct ice_pf *pf, struct msi_map map) { struct ice_irq_entry *entry; entry = xa_load(&pf->irq_tracker.entries, map.index); if (!entry) { dev_err(ice_pf_to_dev(pf), "Failed to get MSIX interrupt entry at index %d", map.index); return; } dev_dbg(ice_pf_to_dev(pf), "Free irq at index %d\n", map.index); if (entry->dynamic) pci_msix_free_irq(pf->pdev, map); ice_free_irq_res(pf, map.index); } /** * ice_get_max_used_msix_vector - Get the max used interrupt vector * @pf: board private structure * * Return index of maximum used interrupt vectors with respect to the * beginning of the MSIX table. Take into account that some interrupts * may have been dynamically allocated after MSIX was initially enabled. */ int ice_get_max_used_msix_vector(struct ice_pf *pf) { unsigned long start, index, max_idx; void *entry; /* Treat all preallocated interrupts as used */ start = pf->irq_tracker.num_static; max_idx = start - 1; xa_for_each_start(&pf->irq_tracker.entries, index, entry, start) { if (index > max_idx) max_idx = index; } return max_idx; }