// SPDX-License-Identifier: GPL-2.0 #include #include #include "pci.h" /* * On the state of PCI's devres implementation: * * The older devres API for PCI has two significant problems: * * 1. It is very strongly tied to the statically allocated mapping table in * struct pcim_iomap_devres below. This is mostly solved in the sense of the * pcim_ functions in this file providing things like ranged mapping by * bypassing this table, whereas the functions that were present in the old * API still enter the mapping addresses into the table for users of the old * API. * * 2. The region-request-functions in pci.c do become managed IF the device has * been enabled with pcim_enable_device() instead of pci_enable_device(). * This resulted in the API becoming inconsistent: Some functions have an * obviously managed counter-part (e.g., pci_iomap() <-> pcim_iomap()), * whereas some don't and are never managed, while others don't and are * _sometimes_ managed (e.g. pci_request_region()). * * Consequently, in the new API, region requests performed by the pcim_ * functions are automatically cleaned up through the devres callback * pcim_addr_resource_release(). * * Users of pcim_enable_device() + pci_*region*() are redirected in * pci.c to the managed functions here in this file. This isn't exactly * perfect, but the only alternative way would be to port ALL drivers * using said combination to pcim_ functions. * * TODO: * Remove the legacy table entirely once all calls to pcim_iomap_table() in * the kernel have been removed. */ /* * Legacy struct storing addresses to whole mapped BARs. */ struct pcim_iomap_devres { void __iomem *table[PCI_STD_NUM_BARS]; }; /* Used to restore the old INTx state on driver detach. */ struct pcim_intx_devres { int orig_intx; }; enum pcim_addr_devres_type { /* Default initializer. */ PCIM_ADDR_DEVRES_TYPE_INVALID, /* A requested region spanning an entire BAR. */ PCIM_ADDR_DEVRES_TYPE_REGION, /* * A requested region spanning an entire BAR, and a mapping for * the entire BAR. */ PCIM_ADDR_DEVRES_TYPE_REGION_MAPPING, /* * A mapping within a BAR, either spanning the whole BAR or just a * range. Without a requested region. */ PCIM_ADDR_DEVRES_TYPE_MAPPING, }; /* * This struct envelops IO or MEM addresses, i.e., mappings and region * requests, because those are very frequently requested and released * together. */ struct pcim_addr_devres { enum pcim_addr_devres_type type; void __iomem *baseaddr; unsigned long offset; unsigned long len; int bar; }; static inline void pcim_addr_devres_clear(struct pcim_addr_devres *res) { memset(res, 0, sizeof(*res)); res->bar = -1; } /* * The following functions, __pcim_*_region*, exist as counterparts to the * versions from pci.c - which, unfortunately, can be in "hybrid mode", i.e., * sometimes managed, sometimes not. * * To separate the APIs cleanly, we define our own, simplified versions here. */ /** * __pcim_request_region_range - Request a ranged region * @pdev: PCI device the region belongs to * @bar: BAR the range is within * @offset: offset from the BAR's start address * @maxlen: length in bytes, beginning at @offset * @name: name associated with the request * @req_flags: flags for the request, e.g., for kernel-exclusive requests * * Returns: 0 on success, a negative error code on failure. * * Request a range within a device's PCI BAR. Sanity check the input. */ static int __pcim_request_region_range(struct pci_dev *pdev, int bar, unsigned long offset, unsigned long maxlen, const char *name, int req_flags) { resource_size_t start = pci_resource_start(pdev, bar); resource_size_t len = pci_resource_len(pdev, bar); unsigned long dev_flags = pci_resource_flags(pdev, bar); if (start == 0 || len == 0) /* Unused BAR. */ return 0; if (len <= offset) return -EINVAL; start += offset; len -= offset; if (len > maxlen && maxlen != 0) len = maxlen; if (dev_flags & IORESOURCE_IO) { if (!request_region(start, len, name)) return -EBUSY; } else if (dev_flags & IORESOURCE_MEM) { if (!__request_mem_region(start, len, name, req_flags)) return -EBUSY; } else { /* That's not a device we can request anything on. */ return -ENODEV; } return 0; } static void __pcim_release_region_range(struct pci_dev *pdev, int bar, unsigned long offset, unsigned long maxlen) { resource_size_t start = pci_resource_start(pdev, bar); resource_size_t len = pci_resource_len(pdev, bar); unsigned long flags = pci_resource_flags(pdev, bar); if (len <= offset || start == 0) return; if (len == 0 || maxlen == 0) /* This an unused BAR. Do nothing. */ return; start += offset; len -= offset; if (len > maxlen) len = maxlen; if (flags & IORESOURCE_IO) release_region(start, len); else if (flags & IORESOURCE_MEM) release_mem_region(start, len); } static int __pcim_request_region(struct pci_dev *pdev, int bar, const char *name, int flags) { unsigned long offset = 0; unsigned long len = pci_resource_len(pdev, bar); return __pcim_request_region_range(pdev, bar, offset, len, name, flags); } static void __pcim_release_region(struct pci_dev *pdev, int bar) { unsigned long offset = 0; unsigned long len = pci_resource_len(pdev, bar); __pcim_release_region_range(pdev, bar, offset, len); } static void pcim_addr_resource_release(struct device *dev, void *resource_raw) { struct pci_dev *pdev = to_pci_dev(dev); struct pcim_addr_devres *res = resource_raw; switch (res->type) { case PCIM_ADDR_DEVRES_TYPE_REGION: __pcim_release_region(pdev, res->bar); break; case PCIM_ADDR_DEVRES_TYPE_REGION_MAPPING: pci_iounmap(pdev, res->baseaddr); __pcim_release_region(pdev, res->bar); break; case PCIM_ADDR_DEVRES_TYPE_MAPPING: pci_iounmap(pdev, res->baseaddr); break; default: break; } } static struct pcim_addr_devres *pcim_addr_devres_alloc(struct pci_dev *pdev) { struct pcim_addr_devres *res; res = devres_alloc_node(pcim_addr_resource_release, sizeof(*res), GFP_KERNEL, dev_to_node(&pdev->dev)); if (res) pcim_addr_devres_clear(res); return res; } /* Just for consistency and readability. */ static inline void pcim_addr_devres_free(struct pcim_addr_devres *res) { devres_free(res); } /* * Used by devres to identify a pcim_addr_devres. */ static int pcim_addr_resources_match(struct device *dev, void *a_raw, void *b_raw) { struct pcim_addr_devres *a, *b; a = a_raw; b = b_raw; if (a->type != b->type) return 0; switch (a->type) { case PCIM_ADDR_DEVRES_TYPE_REGION: case PCIM_ADDR_DEVRES_TYPE_REGION_MAPPING: return a->bar == b->bar; case PCIM_ADDR_DEVRES_TYPE_MAPPING: return a->baseaddr == b->baseaddr; default: return 0; } } static void devm_pci_unmap_iospace(struct device *dev, void *ptr) { struct resource **res = ptr; pci_unmap_iospace(*res); } /** * devm_pci_remap_iospace - Managed pci_remap_iospace() * @dev: Generic device to remap IO address for * @res: Resource describing the I/O space * @phys_addr: physical address of range to be mapped * * Managed pci_remap_iospace(). Map is automatically unmapped on driver * detach. */ int devm_pci_remap_iospace(struct device *dev, const struct resource *res, phys_addr_t phys_addr) { const struct resource **ptr; int error; ptr = devres_alloc(devm_pci_unmap_iospace, sizeof(*ptr), GFP_KERNEL); if (!ptr) return -ENOMEM; error = pci_remap_iospace(res, phys_addr); if (error) { devres_free(ptr); } else { *ptr = res; devres_add(dev, ptr); } return error; } EXPORT_SYMBOL(devm_pci_remap_iospace); /** * devm_pci_remap_cfgspace - Managed pci_remap_cfgspace() * @dev: Generic device to remap IO address for * @offset: Resource address to map * @size: Size of map * * Managed pci_remap_cfgspace(). Map is automatically unmapped on driver * detach. */ void __iomem *devm_pci_remap_cfgspace(struct device *dev, resource_size_t offset, resource_size_t size) { void __iomem **ptr, *addr; ptr = devres_alloc(devm_ioremap_release, sizeof(*ptr), GFP_KERNEL); if (!ptr) return NULL; addr = pci_remap_cfgspace(offset, size); if (addr) { *ptr = addr; devres_add(dev, ptr); } else devres_free(ptr); return addr; } EXPORT_SYMBOL(devm_pci_remap_cfgspace); /** * devm_pci_remap_cfg_resource - check, request region and ioremap cfg resource * @dev: generic device to handle the resource for * @res: configuration space resource to be handled * * Checks that a resource is a valid memory region, requests the memory * region and ioremaps with pci_remap_cfgspace() API that ensures the * proper PCI configuration space memory attributes are guaranteed. * * All operations are managed and will be undone on driver detach. * * Returns a pointer to the remapped memory or an IOMEM_ERR_PTR() encoded error * code on failure. Usage example:: * * res = platform_get_resource(pdev, IORESOURCE_MEM, 0); * base = devm_pci_remap_cfg_resource(&pdev->dev, res); * if (IS_ERR(base)) * return PTR_ERR(base); */ void __iomem *devm_pci_remap_cfg_resource(struct device *dev, struct resource *res) { resource_size_t size; const char *name; void __iomem *dest_ptr; BUG_ON(!dev); if (!res || resource_type(res) != IORESOURCE_MEM) { dev_err(dev, "invalid resource\n"); return IOMEM_ERR_PTR(-EINVAL); } size = resource_size(res); if (res->name) name = devm_kasprintf(dev, GFP_KERNEL, "%s %s", dev_name(dev), res->name); else name = devm_kstrdup(dev, dev_name(dev), GFP_KERNEL); if (!name) return IOMEM_ERR_PTR(-ENOMEM); if (!devm_request_mem_region(dev, res->start, size, name)) { dev_err(dev, "can't request region for resource %pR\n", res); return IOMEM_ERR_PTR(-EBUSY); } dest_ptr = devm_pci_remap_cfgspace(dev, res->start, size); if (!dest_ptr) { dev_err(dev, "ioremap failed for resource %pR\n", res); devm_release_mem_region(dev, res->start, size); dest_ptr = IOMEM_ERR_PTR(-ENOMEM); } return dest_ptr; } EXPORT_SYMBOL(devm_pci_remap_cfg_resource); static void __pcim_clear_mwi(void *pdev_raw) { struct pci_dev *pdev = pdev_raw; pci_clear_mwi(pdev); } /** * pcim_set_mwi - a device-managed pci_set_mwi() * @pdev: the PCI device for which MWI is enabled * * Managed pci_set_mwi(). * * RETURNS: An appropriate -ERRNO error value on error, or zero for success. */ int pcim_set_mwi(struct pci_dev *pdev) { int ret; ret = devm_add_action(&pdev->dev, __pcim_clear_mwi, pdev); if (ret != 0) return ret; ret = pci_set_mwi(pdev); if (ret != 0) devm_remove_action(&pdev->dev, __pcim_clear_mwi, pdev); return ret; } EXPORT_SYMBOL(pcim_set_mwi); static inline bool mask_contains_bar(int mask, int bar) { return mask & BIT(bar); } /* * This is a copy of pci_intx() used to bypass the problem of recursive * function calls due to the hybrid nature of pci_intx(). */ static void __pcim_intx(struct pci_dev *pdev, int enable) { u16 pci_command, new; pci_read_config_word(pdev, PCI_COMMAND, &pci_command); if (enable) new = pci_command & ~PCI_COMMAND_INTX_DISABLE; else new = pci_command | PCI_COMMAND_INTX_DISABLE; if (new != pci_command) pci_write_config_word(pdev, PCI_COMMAND, new); } static void pcim_intx_restore(struct device *dev, void *data) { struct pci_dev *pdev = to_pci_dev(dev); struct pcim_intx_devres *res = data; __pcim_intx(pdev, res->orig_intx); } static struct pcim_intx_devres *get_or_create_intx_devres(struct device *dev) { struct pcim_intx_devres *res; res = devres_find(dev, pcim_intx_restore, NULL, NULL); if (res) return res; res = devres_alloc(pcim_intx_restore, sizeof(*res), GFP_KERNEL); if (res) devres_add(dev, res); return res; } /** * pcim_intx - managed pci_intx() * @pdev: the PCI device to operate on * @enable: boolean: whether to enable or disable PCI INTx * * Returns: 0 on success, -ENOMEM on error. * * Enable/disable PCI INTx for device @pdev. * Restore the original state on driver detach. */ int pcim_intx(struct pci_dev *pdev, int enable) { struct pcim_intx_devres *res; res = get_or_create_intx_devres(&pdev->dev); if (!res) return -ENOMEM; res->orig_intx = !enable; __pcim_intx(pdev, enable); return 0; } static void pcim_disable_device(void *pdev_raw) { struct pci_dev *pdev = pdev_raw; if (!pdev->pinned) pci_disable_device(pdev); } /** * pcim_enable_device - Managed pci_enable_device() * @pdev: PCI device to be initialized * * Returns: 0 on success, negative error code on failure. * * Managed pci_enable_device(). Device will automatically be disabled on * driver detach. */ int pcim_enable_device(struct pci_dev *pdev) { int ret; ret = devm_add_action(&pdev->dev, pcim_disable_device, pdev); if (ret != 0) return ret; /* * We prefer removing the action in case of an error over * devm_add_action_or_reset() because the latter could theoretically be * disturbed by users having pinned the device too soon. */ ret = pci_enable_device(pdev); if (ret != 0) { devm_remove_action(&pdev->dev, pcim_disable_device, pdev); return ret; } pdev->is_managed = true; return ret; } EXPORT_SYMBOL(pcim_enable_device); /** * pcim_pin_device - Pin managed PCI device * @pdev: PCI device to pin * * Pin managed PCI device @pdev. Pinned device won't be disabled on driver * detach. @pdev must have been enabled with pcim_enable_device(). */ void pcim_pin_device(struct pci_dev *pdev) { pdev->pinned = true; } EXPORT_SYMBOL(pcim_pin_device); static void pcim_iomap_release(struct device *gendev, void *res) { /* * Do nothing. This is legacy code. * * Cleanup of the mappings is now done directly through the callbacks * registered when creating them. */ } /** * pcim_iomap_table - access iomap allocation table (DEPRECATED) * @pdev: PCI device to access iomap table for * * Returns: * Const pointer to array of __iomem pointers on success, NULL on failure. * * Access iomap allocation table for @dev. If iomap table doesn't * exist and @pdev is managed, it will be allocated. All iomaps * recorded in the iomap table are automatically unmapped on driver * detach. * * This function might sleep when the table is first allocated but can * be safely called without context and guaranteed to succeed once * allocated. * * This function is DEPRECATED. Do not use it in new code. Instead, obtain a * mapping's address directly from one of the pcim_* mapping functions. For * example: * void __iomem \*mappy = pcim_iomap(pdev, bar, length); */ void __iomem * const *pcim_iomap_table(struct pci_dev *pdev) { struct pcim_iomap_devres *dr, *new_dr; dr = devres_find(&pdev->dev, pcim_iomap_release, NULL, NULL); if (dr) return dr->table; new_dr = devres_alloc_node(pcim_iomap_release, sizeof(*new_dr), GFP_KERNEL, dev_to_node(&pdev->dev)); if (!new_dr) return NULL; dr = devres_get(&pdev->dev, new_dr, NULL, NULL); return dr->table; } EXPORT_SYMBOL(pcim_iomap_table); /* * Fill the legacy mapping-table, so that drivers using the old API can * still get a BAR's mapping address through pcim_iomap_table(). */ static int pcim_add_mapping_to_legacy_table(struct pci_dev *pdev, void __iomem *mapping, int bar) { void __iomem **legacy_iomap_table; if (bar >= PCI_STD_NUM_BARS) return -EINVAL; legacy_iomap_table = (void __iomem **)pcim_iomap_table(pdev); if (!legacy_iomap_table) return -ENOMEM; /* The legacy mechanism doesn't allow for duplicate mappings. */ WARN_ON(legacy_iomap_table[bar]); legacy_iomap_table[bar] = mapping; return 0; } /* * Remove a mapping. The table only contains whole-BAR mappings, so this will * never interfere with ranged mappings. */ static void pcim_remove_mapping_from_legacy_table(struct pci_dev *pdev, void __iomem *addr) { int bar; void __iomem **legacy_iomap_table; legacy_iomap_table = (void __iomem **)pcim_iomap_table(pdev); if (!legacy_iomap_table) return; for (bar = 0; bar < PCI_STD_NUM_BARS; bar++) { if (legacy_iomap_table[bar] == addr) { legacy_iomap_table[bar] = NULL; return; } } } /* * The same as pcim_remove_mapping_from_legacy_table(), but identifies the * mapping by its BAR index. */ static void pcim_remove_bar_from_legacy_table(struct pci_dev *pdev, int bar) { void __iomem **legacy_iomap_table; if (bar >= PCI_STD_NUM_BARS) return; legacy_iomap_table = (void __iomem **)pcim_iomap_table(pdev); if (!legacy_iomap_table) return; legacy_iomap_table[bar] = NULL; } /** * pcim_iomap - Managed pcim_iomap() * @pdev: PCI device to iomap for * @bar: BAR to iomap * @maxlen: Maximum length of iomap * * Returns: __iomem pointer on success, NULL on failure. * * Managed pci_iomap(). Map is automatically unmapped on driver detach. If * desired, unmap manually only with pcim_iounmap(). * * This SHOULD only be used once per BAR. * * NOTE: * Contrary to the other pcim_* functions, this function does not return an * IOMEM_ERR_PTR() on failure, but a simple NULL. This is done for backwards * compatibility. */ void __iomem *pcim_iomap(struct pci_dev *pdev, int bar, unsigned long maxlen) { void __iomem *mapping; struct pcim_addr_devres *res; res = pcim_addr_devres_alloc(pdev); if (!res) return NULL; res->type = PCIM_ADDR_DEVRES_TYPE_MAPPING; mapping = pci_iomap(pdev, bar, maxlen); if (!mapping) goto err_iomap; res->baseaddr = mapping; if (pcim_add_mapping_to_legacy_table(pdev, mapping, bar) != 0) goto err_table; devres_add(&pdev->dev, res); return mapping; err_table: pci_iounmap(pdev, mapping); err_iomap: pcim_addr_devres_free(res); return NULL; } EXPORT_SYMBOL(pcim_iomap); /** * pcim_iounmap - Managed pci_iounmap() * @pdev: PCI device to iounmap for * @addr: Address to unmap * * Managed pci_iounmap(). @addr must have been mapped using a pcim_* mapping * function. */ void pcim_iounmap(struct pci_dev *pdev, void __iomem *addr) { struct pcim_addr_devres res_searched; pcim_addr_devres_clear(&res_searched); res_searched.type = PCIM_ADDR_DEVRES_TYPE_MAPPING; res_searched.baseaddr = addr; if (devres_release(&pdev->dev, pcim_addr_resource_release, pcim_addr_resources_match, &res_searched) != 0) { /* Doesn't exist. User passed nonsense. */ return; } pcim_remove_mapping_from_legacy_table(pdev, addr); } EXPORT_SYMBOL(pcim_iounmap); /** * pcim_iomap_region - Request and iomap a PCI BAR * @pdev: PCI device to map IO resources for * @bar: Index of a BAR to map * @name: Name associated with the request * * Returns: __iomem pointer on success, an IOMEM_ERR_PTR on failure. * * Mapping and region will get automatically released on driver detach. If * desired, release manually only with pcim_iounmap_region(). */ static void __iomem *pcim_iomap_region(struct pci_dev *pdev, int bar, const char *name) { int ret; struct pcim_addr_devres *res; res = pcim_addr_devres_alloc(pdev); if (!res) return IOMEM_ERR_PTR(-ENOMEM); res->type = PCIM_ADDR_DEVRES_TYPE_REGION_MAPPING; res->bar = bar; ret = __pcim_request_region(pdev, bar, name, 0); if (ret != 0) goto err_region; res->baseaddr = pci_iomap(pdev, bar, 0); if (!res->baseaddr) { ret = -EINVAL; goto err_iomap; } devres_add(&pdev->dev, res); return res->baseaddr; err_iomap: __pcim_release_region(pdev, bar); err_region: pcim_addr_devres_free(res); return IOMEM_ERR_PTR(ret); } /** * pcim_iounmap_region - Unmap and release a PCI BAR * @pdev: PCI device to operate on * @bar: Index of BAR to unmap and release * * Unmap a BAR and release its region manually. Only pass BARs that were * previously mapped by pcim_iomap_region(). */ static void pcim_iounmap_region(struct pci_dev *pdev, int bar) { struct pcim_addr_devres res_searched; pcim_addr_devres_clear(&res_searched); res_searched.type = PCIM_ADDR_DEVRES_TYPE_REGION_MAPPING; res_searched.bar = bar; devres_release(&pdev->dev, pcim_addr_resource_release, pcim_addr_resources_match, &res_searched); } /** * pcim_iomap_regions - Request and iomap PCI BARs * @pdev: PCI device to map IO resources for * @mask: Mask of BARs to request and iomap * @name: Name associated with the requests * * Returns: 0 on success, negative error code on failure. * * Request and iomap regions specified by @mask. */ int pcim_iomap_regions(struct pci_dev *pdev, int mask, const char *name) { int ret; int bar; void __iomem *mapping; for (bar = 0; bar < DEVICE_COUNT_RESOURCE; bar++) { if (!mask_contains_bar(mask, bar)) continue; mapping = pcim_iomap_region(pdev, bar, name); if (IS_ERR(mapping)) { ret = PTR_ERR(mapping); goto err; } ret = pcim_add_mapping_to_legacy_table(pdev, mapping, bar); if (ret != 0) goto err; } return 0; err: while (--bar >= 0) { pcim_iounmap_region(pdev, bar); pcim_remove_bar_from_legacy_table(pdev, bar); } return ret; } EXPORT_SYMBOL(pcim_iomap_regions); static int _pcim_request_region(struct pci_dev *pdev, int bar, const char *name, int request_flags) { int ret; struct pcim_addr_devres *res; res = pcim_addr_devres_alloc(pdev); if (!res) return -ENOMEM; res->type = PCIM_ADDR_DEVRES_TYPE_REGION; res->bar = bar; ret = __pcim_request_region(pdev, bar, name, request_flags); if (ret != 0) { pcim_addr_devres_free(res); return ret; } devres_add(&pdev->dev, res); return 0; } /** * pcim_request_region - Request a PCI BAR * @pdev: PCI device to requestion region for * @bar: Index of BAR to request * @name: Name associated with the request * * Returns: 0 on success, a negative error code on failure. * * Request region specified by @bar. * * The region will automatically be released on driver detach. If desired, * release manually only with pcim_release_region(). */ int pcim_request_region(struct pci_dev *pdev, int bar, const char *name) { return _pcim_request_region(pdev, bar, name, 0); } /** * pcim_request_region_exclusive - Request a PCI BAR exclusively * @pdev: PCI device to requestion region for * @bar: Index of BAR to request * @name: Name associated with the request * * Returns: 0 on success, a negative error code on failure. * * Request region specified by @bar exclusively. * * The region will automatically be released on driver detach. If desired, * release manually only with pcim_release_region(). */ int pcim_request_region_exclusive(struct pci_dev *pdev, int bar, const char *name) { return _pcim_request_region(pdev, bar, name, IORESOURCE_EXCLUSIVE); } /** * pcim_release_region - Release a PCI BAR * @pdev: PCI device to operate on * @bar: Index of BAR to release * * Release a region manually that was previously requested by * pcim_request_region(). */ void pcim_release_region(struct pci_dev *pdev, int bar) { struct pcim_addr_devres res_searched; pcim_addr_devres_clear(&res_searched); res_searched.type = PCIM_ADDR_DEVRES_TYPE_REGION; res_searched.bar = bar; devres_release(&pdev->dev, pcim_addr_resource_release, pcim_addr_resources_match, &res_searched); } /** * pcim_release_all_regions - Release all regions of a PCI-device * @pdev: the PCI device * * Release all regions previously requested through pcim_request_region() * or pcim_request_all_regions(). * * Can be called from any context, i.e., not necessarily as a counterpart to * pcim_request_all_regions(). */ static void pcim_release_all_regions(struct pci_dev *pdev) { int bar; for (bar = 0; bar < PCI_STD_NUM_BARS; bar++) pcim_release_region(pdev, bar); } /** * pcim_request_all_regions - Request all regions * @pdev: PCI device to map IO resources for * @name: name associated with the request * * Returns: 0 on success, negative error code on failure. * * Requested regions will automatically be released at driver detach. If * desired, release individual regions with pcim_release_region() or all of * them at once with pcim_release_all_regions(). */ static int pcim_request_all_regions(struct pci_dev *pdev, const char *name) { int ret; int bar; for (bar = 0; bar < PCI_STD_NUM_BARS; bar++) { ret = pcim_request_region(pdev, bar, name); if (ret != 0) goto err; } return 0; err: pcim_release_all_regions(pdev); return ret; } /** * pcim_iomap_regions_request_all - Request all BARs and iomap specified ones * (DEPRECATED) * @pdev: PCI device to map IO resources for * @mask: Mask of BARs to iomap * @name: Name associated with the requests * * Returns: 0 on success, negative error code on failure. * * Request all PCI BARs and iomap regions specified by @mask. * * To release these resources manually, call pcim_release_region() for the * regions and pcim_iounmap() for the mappings. * * This function is DEPRECATED. Don't use it in new code. Instead, use one * of the pcim_* region request functions in combination with a pcim_* * mapping function. */ int pcim_iomap_regions_request_all(struct pci_dev *pdev, int mask, const char *name) { int bar; int ret; void __iomem **legacy_iomap_table; ret = pcim_request_all_regions(pdev, name); if (ret != 0) return ret; for (bar = 0; bar < PCI_STD_NUM_BARS; bar++) { if (!mask_contains_bar(mask, bar)) continue; if (!pcim_iomap(pdev, bar, 0)) goto err; } return 0; err: /* * If bar is larger than 0, then pcim_iomap() above has most likely * failed because of -EINVAL. If it is equal 0, most likely the table * couldn't be created, indicating -ENOMEM. */ ret = bar > 0 ? -EINVAL : -ENOMEM; legacy_iomap_table = (void __iomem **)pcim_iomap_table(pdev); while (--bar >= 0) pcim_iounmap(pdev, legacy_iomap_table[bar]); pcim_release_all_regions(pdev); return ret; } EXPORT_SYMBOL(pcim_iomap_regions_request_all); /** * pcim_iounmap_regions - Unmap and release PCI BARs * @pdev: PCI device to map IO resources for * @mask: Mask of BARs to unmap and release * * Unmap and release regions specified by @mask. */ void pcim_iounmap_regions(struct pci_dev *pdev, int mask) { int i; for (i = 0; i < PCI_STD_NUM_BARS; i++) { if (!mask_contains_bar(mask, i)) continue; pcim_iounmap_region(pdev, i); pcim_remove_bar_from_legacy_table(pdev, i); } } EXPORT_SYMBOL(pcim_iounmap_regions); /** * pcim_iomap_range - Create a ranged __iomap mapping within a PCI BAR * @pdev: PCI device to map IO resources for * @bar: Index of the BAR * @offset: Offset from the begin of the BAR * @len: Length in bytes for the mapping * * Returns: __iomem pointer on success, an IOMEM_ERR_PTR on failure. * * Creates a new IO-Mapping within the specified @bar, ranging from @offset to * @offset + @len. * * The mapping will automatically get unmapped on driver detach. If desired, * release manually only with pcim_iounmap(). */ void __iomem *pcim_iomap_range(struct pci_dev *pdev, int bar, unsigned long offset, unsigned long len) { void __iomem *mapping; struct pcim_addr_devres *res; res = pcim_addr_devres_alloc(pdev); if (!res) return IOMEM_ERR_PTR(-ENOMEM); mapping = pci_iomap_range(pdev, bar, offset, len); if (!mapping) { pcim_addr_devres_free(res); return IOMEM_ERR_PTR(-EINVAL); } res->type = PCIM_ADDR_DEVRES_TYPE_MAPPING; res->baseaddr = mapping; /* * Ranged mappings don't get added to the legacy-table, since the table * only ever keeps track of whole BARs. */ devres_add(&pdev->dev, res); return mapping; } EXPORT_SYMBOL(pcim_iomap_range);