mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-24 21:24:00 +08:00
ae13b7b4e0
IOMMU_INIT_POST and IOMMU_INIT_POST_FINISH pass the plain value 0 instead of NULL to __IOMMU_INIT. Fix this and make sparse happy by doing so. Signed-off-by: Mathias Krause <minipli@googlemail.com> Cc: Joerg Roedel <joerg.roedel@amd.com> Link: http://lkml.kernel.org/r/1346621506-30857-8-git-send-email-minipli@googlemail.com Signed-off-by: Ingo Molnar <mingo@kernel.org>
101 lines
3.7 KiB
C
101 lines
3.7 KiB
C
#ifndef _ASM_X86_IOMMU_TABLE_H
|
|
#define _ASM_X86_IOMMU_TABLE_H
|
|
|
|
#include <asm/swiotlb.h>
|
|
|
|
/*
|
|
* History lesson:
|
|
* The execution chain of IOMMUs in 2.6.36 looks as so:
|
|
*
|
|
* [xen-swiotlb]
|
|
* |
|
|
* +----[swiotlb *]--+
|
|
* / | \
|
|
* / | \
|
|
* [GART] [Calgary] [Intel VT-d]
|
|
* /
|
|
* /
|
|
* [AMD-Vi]
|
|
*
|
|
* *: if SWIOTLB detected 'iommu=soft'/'swiotlb=force' it would skip
|
|
* over the rest of IOMMUs and unconditionally initialize the SWIOTLB.
|
|
* Also it would surreptitiously initialize set the swiotlb=1 if there were
|
|
* more than 4GB and if the user did not pass in 'iommu=off'. The swiotlb
|
|
* flag would be turned off by all IOMMUs except the Calgary one.
|
|
*
|
|
* The IOMMU_INIT* macros allow a similar tree (or more complex if desired)
|
|
* to be built by defining who we depend on.
|
|
*
|
|
* And all that needs to be done is to use one of the macros in the IOMMU
|
|
* and the pci-dma.c will take care of the rest.
|
|
*/
|
|
|
|
struct iommu_table_entry {
|
|
initcall_t detect;
|
|
initcall_t depend;
|
|
void (*early_init)(void); /* No memory allocate available. */
|
|
void (*late_init)(void); /* Yes, can allocate memory. */
|
|
#define IOMMU_FINISH_IF_DETECTED (1<<0)
|
|
#define IOMMU_DETECTED (1<<1)
|
|
int flags;
|
|
};
|
|
/*
|
|
* Macro fills out an entry in the .iommu_table that is equivalent
|
|
* to the fields that 'struct iommu_table_entry' has. The entries
|
|
* that are put in the .iommu_table section are not put in any order
|
|
* hence during boot-time we will have to resort them based on
|
|
* dependency. */
|
|
|
|
|
|
#define __IOMMU_INIT(_detect, _depend, _early_init, _late_init, _finish)\
|
|
static const struct iommu_table_entry \
|
|
__iommu_entry_##_detect __used \
|
|
__attribute__ ((unused, __section__(".iommu_table"), \
|
|
aligned((sizeof(void *))))) \
|
|
= {_detect, _depend, _early_init, _late_init, \
|
|
_finish ? IOMMU_FINISH_IF_DETECTED : 0}
|
|
/*
|
|
* The simplest IOMMU definition. Provide the detection routine
|
|
* and it will be run after the SWIOTLB and the other IOMMUs
|
|
* that utilize this macro. If the IOMMU is detected (ie, the
|
|
* detect routine returns a positive value), the other IOMMUs
|
|
* are also checked. You can use IOMMU_INIT_POST_FINISH if you prefer
|
|
* to stop detecting the other IOMMUs after yours has been detected.
|
|
*/
|
|
#define IOMMU_INIT_POST(_detect) \
|
|
__IOMMU_INIT(_detect, pci_swiotlb_detect_4gb, NULL, NULL, 0)
|
|
|
|
#define IOMMU_INIT_POST_FINISH(detect) \
|
|
__IOMMU_INIT(_detect, pci_swiotlb_detect_4gb, NULL, NULL, 1)
|
|
|
|
/*
|
|
* A more sophisticated version of IOMMU_INIT. This variant requires:
|
|
* a). A detection routine function.
|
|
* b). The name of the detection routine we depend on to get called
|
|
* before us.
|
|
* c). The init routine which gets called if the detection routine
|
|
* returns a positive value from the pci_iommu_alloc. This means
|
|
* no presence of a memory allocator.
|
|
* d). Similar to the 'init', except that this gets called from pci_iommu_init
|
|
* where we do have a memory allocator.
|
|
*
|
|
* The standard vs the _FINISH differs in that the _FINISH variant will
|
|
* continue detecting other IOMMUs in the call list after the
|
|
* the detection routine returns a positive number. The _FINISH will
|
|
* stop the execution chain. Both will still call the 'init' and
|
|
* 'late_init' functions if they are set.
|
|
*/
|
|
#define IOMMU_INIT_FINISH(_detect, _depend, _init, _late_init) \
|
|
__IOMMU_INIT(_detect, _depend, _init, _late_init, 1)
|
|
|
|
#define IOMMU_INIT(_detect, _depend, _init, _late_init) \
|
|
__IOMMU_INIT(_detect, _depend, _init, _late_init, 0)
|
|
|
|
void sort_iommu_table(struct iommu_table_entry *start,
|
|
struct iommu_table_entry *finish);
|
|
|
|
void check_iommu_entries(struct iommu_table_entry *start,
|
|
struct iommu_table_entry *finish);
|
|
|
|
#endif /* _ASM_X86_IOMMU_TABLE_H */
|