mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-26 14:14:01 +08:00
Merge branch 'pci/yijing-dev_is_pci' into next
* pci/yijing-dev_is_pci: alpha/PCI: Use dev_is_pci() to identify PCI devices arm/PCI: Use dev_is_pci() to identify PCI devices arm/PCI: Use dev_is_pci() to identify PCI devices parisc/PCI: Use dev_is_pci() to identify PCI devices sparc/PCI: Use dev_is_pci() to identify PCI devices ia64/PCI: Use dev_is_pci() to identify PCI devices x86/PCI: Use dev_is_pci() to identify PCI devices PCI: Use dev_is_pci() to identify PCI devices
This commit is contained in:
commit
1d72e71d45
@ -325,7 +325,7 @@ pci_map_single_1(struct pci_dev *pdev, void *cpu_addr, size_t size,
|
||||
/* Helper for generic DMA-mapping functions. */
|
||||
static struct pci_dev *alpha_gendev_to_pci(struct device *dev)
|
||||
{
|
||||
if (dev && dev->bus == &pci_bus_type)
|
||||
if (dev && dev_is_pci(dev))
|
||||
return to_pci_dev(dev);
|
||||
|
||||
/* Assume that non-PCI devices asking for DMA are either ISA or EISA,
|
||||
|
@ -257,7 +257,7 @@ static int it8152_needs_bounce(struct device *dev, dma_addr_t dma_addr, size_t s
|
||||
*/
|
||||
static int it8152_pci_platform_notify(struct device *dev)
|
||||
{
|
||||
if (dev->bus == &pci_bus_type) {
|
||||
if (dev_is_pci(dev)) {
|
||||
if (dev->dma_mask)
|
||||
*dev->dma_mask = (SZ_64M - 1) | PHYS_OFFSET;
|
||||
dev->coherent_dma_mask = (SZ_64M - 1) | PHYS_OFFSET;
|
||||
@ -268,7 +268,7 @@ static int it8152_pci_platform_notify(struct device *dev)
|
||||
|
||||
static int it8152_pci_platform_notify_remove(struct device *dev)
|
||||
{
|
||||
if (dev->bus == &pci_bus_type)
|
||||
if (dev_is_pci(dev))
|
||||
dmabounce_unregister_dev(dev);
|
||||
|
||||
return 0;
|
||||
|
@ -326,7 +326,7 @@ static int ixp4xx_needs_bounce(struct device *dev, dma_addr_t dma_addr, size_t s
|
||||
*/
|
||||
static int ixp4xx_pci_platform_notify(struct device *dev)
|
||||
{
|
||||
if(dev->bus == &pci_bus_type) {
|
||||
if (dev_is_pci(dev)) {
|
||||
*dev->dma_mask = SZ_64M - 1;
|
||||
dev->coherent_dma_mask = SZ_64M - 1;
|
||||
dmabounce_register_dev(dev, 2048, 4096, ixp4xx_needs_bounce);
|
||||
@ -336,9 +336,9 @@ static int ixp4xx_pci_platform_notify(struct device *dev)
|
||||
|
||||
static int ixp4xx_pci_platform_notify_remove(struct device *dev)
|
||||
{
|
||||
if(dev->bus == &pci_bus_type) {
|
||||
if (dev_is_pci(dev))
|
||||
dmabounce_unregister_dev(dev);
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
@ -255,7 +255,7 @@ static u64 prefetch_spill_page;
|
||||
#endif
|
||||
|
||||
#ifdef CONFIG_PCI
|
||||
# define GET_IOC(dev) (((dev)->bus == &pci_bus_type) \
|
||||
# define GET_IOC(dev) ((dev_is_pci(dev)) \
|
||||
? ((struct ioc *) PCI_CONTROLLER(to_pci_dev(dev))->iommu) : NULL)
|
||||
#else
|
||||
# define GET_IOC(dev) NULL
|
||||
|
@ -34,7 +34,7 @@
|
||||
*/
|
||||
static int sn_dma_supported(struct device *dev, u64 mask)
|
||||
{
|
||||
BUG_ON(dev->bus != &pci_bus_type);
|
||||
BUG_ON(!dev_is_pci(dev));
|
||||
|
||||
if (mask < 0x7fffffff)
|
||||
return 0;
|
||||
@ -50,7 +50,7 @@ static int sn_dma_supported(struct device *dev, u64 mask)
|
||||
*/
|
||||
int sn_dma_set_mask(struct device *dev, u64 dma_mask)
|
||||
{
|
||||
BUG_ON(dev->bus != &pci_bus_type);
|
||||
BUG_ON(!dev_is_pci(dev));
|
||||
|
||||
if (!sn_dma_supported(dev, dma_mask))
|
||||
return 0;
|
||||
@ -85,7 +85,7 @@ static void *sn_dma_alloc_coherent(struct device *dev, size_t size,
|
||||
struct pci_dev *pdev = to_pci_dev(dev);
|
||||
struct sn_pcibus_provider *provider = SN_PCIDEV_BUSPROVIDER(pdev);
|
||||
|
||||
BUG_ON(dev->bus != &pci_bus_type);
|
||||
BUG_ON(!dev_is_pci(dev));
|
||||
|
||||
/*
|
||||
* Allocate the memory.
|
||||
@ -143,7 +143,7 @@ static void sn_dma_free_coherent(struct device *dev, size_t size, void *cpu_addr
|
||||
struct pci_dev *pdev = to_pci_dev(dev);
|
||||
struct sn_pcibus_provider *provider = SN_PCIDEV_BUSPROVIDER(pdev);
|
||||
|
||||
BUG_ON(dev->bus != &pci_bus_type);
|
||||
BUG_ON(!dev_is_pci(dev));
|
||||
|
||||
provider->dma_unmap(pdev, dma_handle, 0);
|
||||
free_pages((unsigned long)cpu_addr, get_order(size));
|
||||
@ -187,7 +187,7 @@ static dma_addr_t sn_dma_map_page(struct device *dev, struct page *page,
|
||||
|
||||
dmabarr = dma_get_attr(DMA_ATTR_WRITE_BARRIER, attrs);
|
||||
|
||||
BUG_ON(dev->bus != &pci_bus_type);
|
||||
BUG_ON(!dev_is_pci(dev));
|
||||
|
||||
phys_addr = __pa(cpu_addr);
|
||||
if (dmabarr)
|
||||
@ -223,7 +223,7 @@ static void sn_dma_unmap_page(struct device *dev, dma_addr_t dma_addr,
|
||||
struct pci_dev *pdev = to_pci_dev(dev);
|
||||
struct sn_pcibus_provider *provider = SN_PCIDEV_BUSPROVIDER(pdev);
|
||||
|
||||
BUG_ON(dev->bus != &pci_bus_type);
|
||||
BUG_ON(!dev_is_pci(dev));
|
||||
|
||||
provider->dma_unmap(pdev, dma_addr, dir);
|
||||
}
|
||||
@ -247,7 +247,7 @@ static void sn_dma_unmap_sg(struct device *dev, struct scatterlist *sgl,
|
||||
struct sn_pcibus_provider *provider = SN_PCIDEV_BUSPROVIDER(pdev);
|
||||
struct scatterlist *sg;
|
||||
|
||||
BUG_ON(dev->bus != &pci_bus_type);
|
||||
BUG_ON(!dev_is_pci(dev));
|
||||
|
||||
for_each_sg(sgl, sg, nhwentries, i) {
|
||||
provider->dma_unmap(pdev, sg->dma_address, dir);
|
||||
@ -284,7 +284,7 @@ static int sn_dma_map_sg(struct device *dev, struct scatterlist *sgl,
|
||||
|
||||
dmabarr = dma_get_attr(DMA_ATTR_WRITE_BARRIER, attrs);
|
||||
|
||||
BUG_ON(dev->bus != &pci_bus_type);
|
||||
BUG_ON(!dev_is_pci(dev));
|
||||
|
||||
/*
|
||||
* Setup a DMA address for each entry in the scatterlist.
|
||||
@ -323,26 +323,26 @@ static int sn_dma_map_sg(struct device *dev, struct scatterlist *sgl,
|
||||
static void sn_dma_sync_single_for_cpu(struct device *dev, dma_addr_t dma_handle,
|
||||
size_t size, enum dma_data_direction dir)
|
||||
{
|
||||
BUG_ON(dev->bus != &pci_bus_type);
|
||||
BUG_ON(!dev_is_pci(dev));
|
||||
}
|
||||
|
||||
static void sn_dma_sync_single_for_device(struct device *dev, dma_addr_t dma_handle,
|
||||
size_t size,
|
||||
enum dma_data_direction dir)
|
||||
{
|
||||
BUG_ON(dev->bus != &pci_bus_type);
|
||||
BUG_ON(!dev_is_pci(dev));
|
||||
}
|
||||
|
||||
static void sn_dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg,
|
||||
int nelems, enum dma_data_direction dir)
|
||||
{
|
||||
BUG_ON(dev->bus != &pci_bus_type);
|
||||
BUG_ON(!dev_is_pci(dev));
|
||||
}
|
||||
|
||||
static void sn_dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg,
|
||||
int nelems, enum dma_data_direction dir)
|
||||
{
|
||||
BUG_ON(dev->bus != &pci_bus_type);
|
||||
BUG_ON(!dev_is_pci(dev));
|
||||
}
|
||||
|
||||
static int sn_dma_mapping_error(struct device *dev, dma_addr_t dma_addr)
|
||||
|
@ -282,18 +282,6 @@ find_pa_parent_type(const struct parisc_device *padev, int type)
|
||||
return NULL;
|
||||
}
|
||||
|
||||
#ifdef CONFIG_PCI
|
||||
static inline int is_pci_dev(struct device *dev)
|
||||
{
|
||||
return dev->bus == &pci_bus_type;
|
||||
}
|
||||
#else
|
||||
static inline int is_pci_dev(struct device *dev)
|
||||
{
|
||||
return 0;
|
||||
}
|
||||
#endif
|
||||
|
||||
/*
|
||||
* get_node_path fills in @path with the firmware path to the device.
|
||||
* Note that if @node is a parisc device, we don't fill in the 'mod' field.
|
||||
@ -306,7 +294,7 @@ static void get_node_path(struct device *dev, struct hardware_path *path)
|
||||
int i = 5;
|
||||
memset(&path->bc, -1, 6);
|
||||
|
||||
if (is_pci_dev(dev)) {
|
||||
if (dev_is_pci(dev)) {
|
||||
unsigned int devfn = to_pci_dev(dev)->devfn;
|
||||
path->mod = PCI_FUNC(devfn);
|
||||
path->bc[i--] = PCI_SLOT(devfn);
|
||||
@ -314,7 +302,7 @@ static void get_node_path(struct device *dev, struct hardware_path *path)
|
||||
}
|
||||
|
||||
while (dev != &root) {
|
||||
if (is_pci_dev(dev)) {
|
||||
if (dev_is_pci(dev)) {
|
||||
unsigned int devfn = to_pci_dev(dev)->devfn;
|
||||
path->bc[i--] = PCI_SLOT(devfn) | (PCI_FUNC(devfn)<< 5);
|
||||
} else if (dev->bus == &parisc_bus_type) {
|
||||
@ -695,7 +683,7 @@ static int check_parent(struct device * dev, void * data)
|
||||
if (dev->bus == &parisc_bus_type) {
|
||||
if (match_parisc_device(dev, d->index, d->modpath))
|
||||
d->dev = dev;
|
||||
} else if (is_pci_dev(dev)) {
|
||||
} else if (dev_is_pci(dev)) {
|
||||
if (match_pci_device(dev, d->index, d->modpath))
|
||||
d->dev = dev;
|
||||
} else if (dev->bus == NULL) {
|
||||
@ -753,7 +741,7 @@ struct device *hwpath_to_device(struct hardware_path *modpath)
|
||||
if (!parent)
|
||||
return NULL;
|
||||
}
|
||||
if (is_pci_dev(parent)) /* pci devices already parse MOD */
|
||||
if (dev_is_pci(parent)) /* pci devices already parse MOD */
|
||||
return parent;
|
||||
else
|
||||
return parse_tree_node(parent, 6, modpath);
|
||||
@ -772,7 +760,7 @@ void device_to_hwpath(struct device *dev, struct hardware_path *path)
|
||||
padev = to_parisc_device(dev);
|
||||
get_node_path(dev->parent, path);
|
||||
path->mod = padev->hw_path;
|
||||
} else if (is_pci_dev(dev)) {
|
||||
} else if (dev_is_pci(dev)) {
|
||||
get_node_path(dev, path);
|
||||
}
|
||||
}
|
||||
|
@ -854,7 +854,7 @@ int dma_supported(struct device *dev, u64 device_mask)
|
||||
return 1;
|
||||
|
||||
#ifdef CONFIG_PCI
|
||||
if (dev->bus == &pci_bus_type)
|
||||
if (dev_is_pci(dev))
|
||||
return pci64_dma_supported(to_pci_dev(dev), device_mask);
|
||||
#endif
|
||||
|
||||
|
@ -666,10 +666,9 @@ EXPORT_SYMBOL(dma_ops);
|
||||
*/
|
||||
int dma_supported(struct device *dev, u64 mask)
|
||||
{
|
||||
#ifdef CONFIG_PCI
|
||||
if (dev->bus == &pci_bus_type)
|
||||
if (dev_is_pci(dev))
|
||||
return 1;
|
||||
#endif
|
||||
|
||||
return 0;
|
||||
}
|
||||
EXPORT_SYMBOL(dma_supported);
|
||||
|
@ -1034,9 +1034,7 @@ static int mp_config_acpi_gsi(struct device *dev, u32 gsi, int trigger,
|
||||
|
||||
if (!acpi_ioapic)
|
||||
return 0;
|
||||
if (!dev)
|
||||
return 0;
|
||||
if (dev->bus != &pci_bus_type)
|
||||
if (!dev || !dev_is_pci(dev))
|
||||
return 0;
|
||||
|
||||
pdev = to_pci_dev(dev);
|
||||
|
@ -358,7 +358,7 @@ static void pci_acpi_cleanup(struct device *dev)
|
||||
|
||||
static bool pci_acpi_bus_match(struct device *dev)
|
||||
{
|
||||
return dev->bus == &pci_bus_type;
|
||||
return dev_is_pci(dev);
|
||||
}
|
||||
|
||||
static struct acpi_bus_type acpi_pci_bus = {
|
||||
|
Loading…
Reference in New Issue
Block a user