2019-06-04 16:11:33 +08:00
|
|
|
/* SPDX-License-Identifier: GPL-2.0-only */
|
2017-04-05 15:54:56 +08:00
|
|
|
/*
|
|
|
|
* Copyright 2017 Benjamin Herrenschmidt, IBM Corporation
|
|
|
|
*/
|
|
|
|
|
|
|
|
#ifndef _KVM_PPC_BOOK3S_XIVE_H
|
|
|
|
#define _KVM_PPC_BOOK3S_XIVE_H
|
|
|
|
|
|
|
|
#ifdef CONFIG_KVM_XICS
|
|
|
|
#include "book3s_xics.h"
|
|
|
|
|
2019-04-18 18:39:29 +08:00
|
|
|
/*
|
|
|
|
* The XIVE Interrupt source numbers are within the range 0 to
|
|
|
|
* KVMPPC_XICS_NR_IRQS.
|
|
|
|
*/
|
|
|
|
#define KVMPPC_XIVE_FIRST_IRQ 0
|
|
|
|
#define KVMPPC_XIVE_NR_IRQS KVMPPC_XICS_NR_IRQS
|
|
|
|
|
2017-04-05 15:54:56 +08:00
|
|
|
/*
|
|
|
|
* State for one guest irq source.
|
|
|
|
*
|
|
|
|
* For each guest source we allocate a HW interrupt in the XIVE
|
|
|
|
* which we use for all SW triggers. It will be unused for
|
|
|
|
* pass-through but it's easier to keep around as the same
|
|
|
|
* guest interrupt can alternatively be emulated or pass-through
|
|
|
|
* if a physical device is hot unplugged and replaced with an
|
|
|
|
* emulated one.
|
|
|
|
*
|
|
|
|
* This state structure is very similar to the XICS one with
|
|
|
|
* additional XIVE specific tracking.
|
|
|
|
*/
|
|
|
|
struct kvmppc_xive_irq_state {
|
|
|
|
bool valid; /* Interrupt entry is valid */
|
|
|
|
|
|
|
|
u32 number; /* Guest IRQ number */
|
|
|
|
u32 ipi_number; /* XIVE IPI HW number */
|
|
|
|
struct xive_irq_data ipi_data; /* XIVE IPI associated data */
|
|
|
|
u32 pt_number; /* XIVE Pass-through number if any */
|
|
|
|
struct xive_irq_data *pt_data; /* XIVE Pass-through associated data */
|
|
|
|
|
|
|
|
/* Targetting as set by guest */
|
|
|
|
u8 guest_priority; /* Guest set priority */
|
|
|
|
u8 saved_priority; /* Saved priority when masking */
|
|
|
|
|
|
|
|
/* Actual targetting */
|
|
|
|
u32 act_server; /* Actual server */
|
|
|
|
u8 act_priority; /* Actual priority */
|
|
|
|
|
|
|
|
/* Various state bits */
|
|
|
|
bool in_eoi; /* Synchronize with H_EOI */
|
|
|
|
bool old_p; /* P bit state when masking */
|
|
|
|
bool old_q; /* Q bit state when masking */
|
|
|
|
bool lsi; /* level-sensitive interrupt */
|
|
|
|
bool asserted; /* Only for emulated LSI: current state */
|
|
|
|
|
|
|
|
/* Saved for migration state */
|
|
|
|
bool in_queue;
|
|
|
|
bool saved_p;
|
|
|
|
bool saved_q;
|
|
|
|
u8 saved_scan_prio;
|
2019-04-18 18:39:30 +08:00
|
|
|
|
|
|
|
/* Xive native */
|
|
|
|
u32 eisn; /* Guest Effective IRQ number */
|
2017-04-05 15:54:56 +08:00
|
|
|
};
|
|
|
|
|
|
|
|
/* Select the "right" interrupt (IPI vs. passthrough) */
|
|
|
|
static inline void kvmppc_xive_select_irq(struct kvmppc_xive_irq_state *state,
|
|
|
|
u32 *out_hw_irq,
|
|
|
|
struct xive_irq_data **out_xd)
|
|
|
|
{
|
|
|
|
if (state->pt_number) {
|
|
|
|
if (out_hw_irq)
|
|
|
|
*out_hw_irq = state->pt_number;
|
|
|
|
if (out_xd)
|
|
|
|
*out_xd = state->pt_data;
|
|
|
|
} else {
|
|
|
|
if (out_hw_irq)
|
|
|
|
*out_hw_irq = state->ipi_number;
|
|
|
|
if (out_xd)
|
|
|
|
*out_xd = &state->ipi_data;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* This corresponds to an "ICS" in XICS terminology, we use it
|
|
|
|
* as a mean to break up source information into multiple structures.
|
|
|
|
*/
|
|
|
|
struct kvmppc_xive_src_block {
|
|
|
|
arch_spinlock_t lock;
|
|
|
|
u16 id;
|
|
|
|
struct kvmppc_xive_irq_state irq_state[KVMPPC_XICS_IRQ_PER_ICS];
|
|
|
|
};
|
|
|
|
|
KVM: PPC: Book3S HV: XIVE: Add passthrough support
The KVM XICS-over-XIVE device and the proposed KVM XIVE native device
implement an IRQ space for the guest using the generic IPI interrupts
of the XIVE IC controller. These interrupts are allocated at the OPAL
level and "mapped" into the guest IRQ number space in the range 0-0x1FFF.
Interrupt management is performed in the XIVE way: using loads and
stores on the addresses of the XIVE IPI interrupt ESB pages.
Both KVM devices share the same internal structure caching information
on the interrupts, among which the xive_irq_data struct containing the
addresses of the IPI ESB pages and an extra one in case of pass-through.
The later contains the addresses of the ESB pages of the underlying HW
controller interrupts, PHB4 in all cases for now.
A guest, when running in the XICS legacy interrupt mode, lets the KVM
XICS-over-XIVE device "handle" interrupt management, that is to
perform the loads and stores on the addresses of the ESB pages of the
guest interrupts. However, when running in XIVE native exploitation
mode, the KVM XIVE native device exposes the interrupt ESB pages to
the guest and lets the guest perform directly the loads and stores.
The VMA exposing the ESB pages make use of a custom VM fault handler
which role is to populate the VMA with appropriate pages. When a fault
occurs, the guest IRQ number is deduced from the offset, and the ESB
pages of associated XIVE IPI interrupt are inserted in the VMA (using
the internal structure caching information on the interrupts).
Supporting device passthrough in the guest running in XIVE native
exploitation mode adds some extra refinements because the ESB pages
of a different HW controller (PHB4) need to be exposed to the guest
along with the initial IPI ESB pages of the XIVE IC controller. But
the overall mechanic is the same.
When the device HW irqs are mapped into or unmapped from the guest
IRQ number space, the passthru_irq helpers, kvmppc_xive_set_mapped()
and kvmppc_xive_clr_mapped(), are called to record or clear the
passthrough interrupt information and to perform the switch.
The approach taken by this patch is to clear the ESB pages of the
guest IRQ number being mapped and let the VM fault handler repopulate.
The handler will insert the ESB page corresponding to the HW interrupt
of the device being passed-through or the initial IPI ESB page if the
device is being removed.
Signed-off-by: Cédric Le Goater <clg@kaod.org>
Reviewed-by: David Gibson <david@gibson.dropbear.id.au>
Signed-off-by: Paul Mackerras <paulus@ozlabs.org>
2019-04-18 18:39:39 +08:00
|
|
|
struct kvmppc_xive;
|
|
|
|
|
|
|
|
struct kvmppc_xive_ops {
|
|
|
|
int (*reset_mapped)(struct kvm *kvm, unsigned long guest_irq);
|
|
|
|
};
|
2017-04-05 15:54:56 +08:00
|
|
|
|
2021-07-20 21:42:08 +08:00
|
|
|
#define KVMPPC_XIVE_FLAG_SINGLE_ESCALATION 0x1
|
2021-07-20 21:42:09 +08:00
|
|
|
#define KVMPPC_XIVE_FLAG_SAVE_RESTORE 0x2
|
2021-07-20 21:42:08 +08:00
|
|
|
|
2017-04-05 15:54:56 +08:00
|
|
|
struct kvmppc_xive {
|
|
|
|
struct kvm *kvm;
|
|
|
|
struct kvm_device *dev;
|
|
|
|
struct dentry *dentry;
|
|
|
|
|
|
|
|
/* VP block associated with the VM */
|
|
|
|
u32 vp_base;
|
|
|
|
|
|
|
|
/* Blocks of sources */
|
|
|
|
struct kvmppc_xive_src_block *src_blocks[KVMPPC_XICS_MAX_ICS_ID + 1];
|
|
|
|
u32 max_sbid;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* For state save, we lazily scan the queues on the first interrupt
|
|
|
|
* being migrated. We don't have a clean way to reset that flags
|
|
|
|
* so we keep track of the number of valid sources and how many of
|
|
|
|
* them were migrated so we can reset when all of them have been
|
|
|
|
* processed.
|
|
|
|
*/
|
|
|
|
u32 src_count;
|
|
|
|
u32 saved_src_count;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Some irqs are delayed on restore until the source is created,
|
|
|
|
* keep track here of how many of them
|
|
|
|
*/
|
|
|
|
u32 delayed_irqs;
|
|
|
|
|
|
|
|
/* Which queues (priorities) are in use by the guest */
|
|
|
|
u8 qmap;
|
|
|
|
|
|
|
|
/* Queue orders */
|
|
|
|
u32 q_order;
|
|
|
|
u32 q_page_order;
|
|
|
|
|
2018-01-12 10:37:12 +08:00
|
|
|
/* Flags */
|
2021-07-20 21:42:08 +08:00
|
|
|
u8 flags;
|
KVM: PPC: Book3S HV: XIVE: Add passthrough support
The KVM XICS-over-XIVE device and the proposed KVM XIVE native device
implement an IRQ space for the guest using the generic IPI interrupts
of the XIVE IC controller. These interrupts are allocated at the OPAL
level and "mapped" into the guest IRQ number space in the range 0-0x1FFF.
Interrupt management is performed in the XIVE way: using loads and
stores on the addresses of the XIVE IPI interrupt ESB pages.
Both KVM devices share the same internal structure caching information
on the interrupts, among which the xive_irq_data struct containing the
addresses of the IPI ESB pages and an extra one in case of pass-through.
The later contains the addresses of the ESB pages of the underlying HW
controller interrupts, PHB4 in all cases for now.
A guest, when running in the XICS legacy interrupt mode, lets the KVM
XICS-over-XIVE device "handle" interrupt management, that is to
perform the loads and stores on the addresses of the ESB pages of the
guest interrupts. However, when running in XIVE native exploitation
mode, the KVM XIVE native device exposes the interrupt ESB pages to
the guest and lets the guest perform directly the loads and stores.
The VMA exposing the ESB pages make use of a custom VM fault handler
which role is to populate the VMA with appropriate pages. When a fault
occurs, the guest IRQ number is deduced from the offset, and the ESB
pages of associated XIVE IPI interrupt are inserted in the VMA (using
the internal structure caching information on the interrupts).
Supporting device passthrough in the guest running in XIVE native
exploitation mode adds some extra refinements because the ESB pages
of a different HW controller (PHB4) need to be exposed to the guest
along with the initial IPI ESB pages of the XIVE IC controller. But
the overall mechanic is the same.
When the device HW irqs are mapped into or unmapped from the guest
IRQ number space, the passthru_irq helpers, kvmppc_xive_set_mapped()
and kvmppc_xive_clr_mapped(), are called to record or clear the
passthrough interrupt information and to perform the switch.
The approach taken by this patch is to clear the ESB pages of the
guest IRQ number being mapped and let the VM fault handler repopulate.
The handler will insert the ESB page corresponding to the HW interrupt
of the device being passed-through or the initial IPI ESB page if the
device is being removed.
Signed-off-by: Cédric Le Goater <clg@kaod.org>
Reviewed-by: David Gibson <david@gibson.dropbear.id.au>
Signed-off-by: Paul Mackerras <paulus@ozlabs.org>
2019-04-18 18:39:39 +08:00
|
|
|
|
2019-09-27 19:54:01 +08:00
|
|
|
/* Number of entries in the VP block */
|
|
|
|
u32 nr_servers;
|
|
|
|
|
KVM: PPC: Book3S HV: XIVE: Add passthrough support
The KVM XICS-over-XIVE device and the proposed KVM XIVE native device
implement an IRQ space for the guest using the generic IPI interrupts
of the XIVE IC controller. These interrupts are allocated at the OPAL
level and "mapped" into the guest IRQ number space in the range 0-0x1FFF.
Interrupt management is performed in the XIVE way: using loads and
stores on the addresses of the XIVE IPI interrupt ESB pages.
Both KVM devices share the same internal structure caching information
on the interrupts, among which the xive_irq_data struct containing the
addresses of the IPI ESB pages and an extra one in case of pass-through.
The later contains the addresses of the ESB pages of the underlying HW
controller interrupts, PHB4 in all cases for now.
A guest, when running in the XICS legacy interrupt mode, lets the KVM
XICS-over-XIVE device "handle" interrupt management, that is to
perform the loads and stores on the addresses of the ESB pages of the
guest interrupts. However, when running in XIVE native exploitation
mode, the KVM XIVE native device exposes the interrupt ESB pages to
the guest and lets the guest perform directly the loads and stores.
The VMA exposing the ESB pages make use of a custom VM fault handler
which role is to populate the VMA with appropriate pages. When a fault
occurs, the guest IRQ number is deduced from the offset, and the ESB
pages of associated XIVE IPI interrupt are inserted in the VMA (using
the internal structure caching information on the interrupts).
Supporting device passthrough in the guest running in XIVE native
exploitation mode adds some extra refinements because the ESB pages
of a different HW controller (PHB4) need to be exposed to the guest
along with the initial IPI ESB pages of the XIVE IC controller. But
the overall mechanic is the same.
When the device HW irqs are mapped into or unmapped from the guest
IRQ number space, the passthru_irq helpers, kvmppc_xive_set_mapped()
and kvmppc_xive_clr_mapped(), are called to record or clear the
passthrough interrupt information and to perform the switch.
The approach taken by this patch is to clear the ESB pages of the
guest IRQ number being mapped and let the VM fault handler repopulate.
The handler will insert the ESB page corresponding to the HW interrupt
of the device being passed-through or the initial IPI ESB page if the
device is being removed.
Signed-off-by: Cédric Le Goater <clg@kaod.org>
Reviewed-by: David Gibson <david@gibson.dropbear.id.au>
Signed-off-by: Paul Mackerras <paulus@ozlabs.org>
2019-04-18 18:39:39 +08:00
|
|
|
struct kvmppc_xive_ops *ops;
|
|
|
|
struct address_space *mapping;
|
|
|
|
struct mutex mapping_lock;
|
2019-05-24 21:20:30 +08:00
|
|
|
struct mutex lock;
|
2017-04-05 15:54:56 +08:00
|
|
|
};
|
|
|
|
|
|
|
|
#define KVMPPC_XIVE_Q_COUNT 8
|
|
|
|
|
|
|
|
struct kvmppc_xive_vcpu {
|
|
|
|
struct kvmppc_xive *xive;
|
|
|
|
struct kvm_vcpu *vcpu;
|
|
|
|
bool valid;
|
|
|
|
|
|
|
|
/* Server number. This is the HW CPU ID from a guest perspective */
|
|
|
|
u32 server_num;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* HW VP corresponding to this VCPU. This is the base of the VP
|
|
|
|
* block plus the server number.
|
|
|
|
*/
|
|
|
|
u32 vp_id;
|
|
|
|
u32 vp_chip_id;
|
|
|
|
u32 vp_cam;
|
|
|
|
|
|
|
|
/* IPI used for sending ... IPIs */
|
|
|
|
u32 vp_ipi;
|
|
|
|
struct xive_irq_data vp_ipi_data;
|
|
|
|
|
|
|
|
/* Local emulation state */
|
|
|
|
uint8_t cppr; /* guest CPPR */
|
|
|
|
uint8_t hw_cppr;/* Hardware CPPR */
|
|
|
|
uint8_t mfrr;
|
|
|
|
uint8_t pending;
|
|
|
|
|
|
|
|
/* Each VP has 8 queues though we only provision some */
|
|
|
|
struct xive_q queues[KVMPPC_XIVE_Q_COUNT];
|
|
|
|
u32 esc_virq[KVMPPC_XIVE_Q_COUNT];
|
|
|
|
char *esc_virq_names[KVMPPC_XIVE_Q_COUNT];
|
|
|
|
|
|
|
|
/* Stash a delayed irq on restore from migration (see set_icp) */
|
|
|
|
u32 delayed_irq;
|
|
|
|
|
|
|
|
/* Stats */
|
|
|
|
u64 stat_rm_h_xirr;
|
|
|
|
u64 stat_rm_h_ipoll;
|
|
|
|
u64 stat_rm_h_cppr;
|
|
|
|
u64 stat_rm_h_eoi;
|
|
|
|
u64 stat_rm_h_ipi;
|
|
|
|
u64 stat_vm_h_xirr;
|
|
|
|
u64 stat_vm_h_ipoll;
|
|
|
|
u64 stat_vm_h_cppr;
|
|
|
|
u64 stat_vm_h_eoi;
|
|
|
|
u64 stat_vm_h_ipi;
|
|
|
|
};
|
|
|
|
|
|
|
|
static inline struct kvm_vcpu *kvmppc_xive_find_server(struct kvm *kvm, u32 nr)
|
|
|
|
{
|
|
|
|
struct kvm_vcpu *vcpu = NULL;
|
2021-11-17 00:04:02 +08:00
|
|
|
unsigned long i;
|
2017-04-05 15:54:56 +08:00
|
|
|
|
|
|
|
kvm_for_each_vcpu(i, vcpu, kvm) {
|
|
|
|
if (vcpu->arch.xive_vcpu && nr == vcpu->arch.xive_vcpu->server_num)
|
|
|
|
return vcpu;
|
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline struct kvmppc_xive_src_block *kvmppc_xive_find_source(struct kvmppc_xive *xive,
|
|
|
|
u32 irq, u16 *source)
|
|
|
|
{
|
|
|
|
u32 bid = irq >> KVMPPC_XICS_ICS_SHIFT;
|
|
|
|
u16 src = irq & KVMPPC_XICS_SRC_MASK;
|
|
|
|
|
|
|
|
if (source)
|
|
|
|
*source = src;
|
|
|
|
if (bid > KVMPPC_XICS_MAX_ICS_ID)
|
|
|
|
return NULL;
|
|
|
|
return xive->src_blocks[bid];
|
|
|
|
}
|
|
|
|
|
2020-12-11 01:14:50 +08:00
|
|
|
/*
|
|
|
|
* When the XIVE resources are allocated at the HW level, the VP
|
|
|
|
* structures describing the vCPUs of a guest are distributed among
|
|
|
|
* the chips to optimize the PowerBUS usage. For best performance, the
|
|
|
|
* guest vCPUs can be pinned to match the VP structure distribution.
|
|
|
|
*
|
|
|
|
* Currently, the VP identifiers are deduced from the vCPU id using
|
|
|
|
* the kvmppc_pack_vcpu_id() routine which is not incorrect but not
|
|
|
|
* optimal either. It VSMT is used, the result is not continuous and
|
|
|
|
* the constraints on HW resources described above can not be met.
|
|
|
|
*/
|
2019-04-18 18:39:28 +08:00
|
|
|
static inline u32 kvmppc_xive_vp(struct kvmppc_xive *xive, u32 server)
|
|
|
|
{
|
|
|
|
return xive->vp_base + kvmppc_pack_vcpu_id(xive->kvm, server);
|
|
|
|
}
|
|
|
|
|
KVM: PPC: Book3S HV: XIVE: Ensure VP isn't already in use
Connecting a vCPU to a XIVE KVM device means establishing a 1:1
association between a vCPU id and the offset (VP id) of a VP
structure within a fixed size block of VPs. We currently try to
enforce the 1:1 relationship by checking that a vCPU with the
same id isn't already connected. This is good but unfortunately
not enough because we don't map VP ids to raw vCPU ids but to
packed vCPU ids, and the packing function kvmppc_pack_vcpu_id()
isn't bijective by design. We got away with it because QEMU passes
vCPU ids that fit well in the packing pattern. But nothing prevents
userspace to come up with a forged vCPU id resulting in a packed id
collision which causes the KVM device to associate two vCPUs to the
same VP. This greatly confuses the irq layer and ultimately crashes
the kernel, as shown below.
Example: a guest with 1 guest thread per core, a core stride of
8 and 300 vCPUs has vCPU ids 0,8,16...2392. If QEMU is patched to
inject at some point an invalid vCPU id 348, which is the packed
version of itself and 2392, we get:
genirq: Flags mismatch irq 199. 00010000 (kvm-2-2392) vs. 00010000 (kvm-2-348)
CPU: 24 PID: 88176 Comm: qemu-system-ppc Not tainted 5.3.0-xive-nr-servers-5.3-gku+ #38
Call Trace:
[c000003f7f9937e0] [c000000000c0110c] dump_stack+0xb0/0xf4 (unreliable)
[c000003f7f993820] [c0000000001cb480] __setup_irq+0xa70/0xad0
[c000003f7f9938d0] [c0000000001cb75c] request_threaded_irq+0x13c/0x260
[c000003f7f993940] [c00800000d44e7ac] kvmppc_xive_attach_escalation+0x104/0x270 [kvm]
[c000003f7f9939d0] [c00800000d45013c] kvmppc_xive_connect_vcpu+0x424/0x620 [kvm]
[c000003f7f993ac0] [c00800000d444428] kvm_arch_vcpu_ioctl+0x260/0x448 [kvm]
[c000003f7f993b90] [c00800000d43593c] kvm_vcpu_ioctl+0x154/0x7c8 [kvm]
[c000003f7f993d00] [c0000000004840f0] do_vfs_ioctl+0xe0/0xc30
[c000003f7f993db0] [c000000000484d44] ksys_ioctl+0x104/0x120
[c000003f7f993e00] [c000000000484d88] sys_ioctl+0x28/0x80
[c000003f7f993e20] [c00000000000b278] system_call+0x5c/0x68
xive-kvm: Failed to request escalation interrupt for queue 0 of VCPU 2392
------------[ cut here ]------------
remove_proc_entry: removing non-empty directory 'irq/199', leaking at least 'kvm-2-348'
WARNING: CPU: 24 PID: 88176 at /home/greg/Work/linux/kernel-kvm-ppc/fs/proc/generic.c:684 remove_proc_entry+0x1ec/0x200
Modules linked in: kvm_hv kvm dm_mod vhost_net vhost tap xt_CHECKSUM iptable_mangle xt_MASQUERADE iptable_nat nf_nat xt_conntrack nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 ipt_REJECT nf_reject_ipv4 tun bridge stp llc ebtable_filter ebtables ip6table_filter ip6_tables iptable_filter squashfs loop fuse i2c_dev sg ofpart ocxl powernv_flash at24 xts mtd uio_pdrv_genirq vmx_crypto opal_prd ipmi_powernv uio ipmi_devintf ipmi_msghandler ibmpowernv ib_iser rdma_cm iw_cm ib_cm ib_core iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi ip_tables ext4 mbcache jbd2 raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor xor async_tx raid6_pq libcrc32c raid1 raid0 linear sd_mod ast i2c_algo_bit drm_vram_helper ttm drm_kms_helper syscopyarea sysfillrect sysimgblt fb_sys_fops drm ahci libahci libata tg3 drm_panel_orientation_quirks [last unloaded: kvm]
CPU: 24 PID: 88176 Comm: qemu-system-ppc Not tainted 5.3.0-xive-nr-servers-5.3-gku+ #38
NIP: c00000000053b0cc LR: c00000000053b0c8 CTR: c0000000000ba3b0
REGS: c000003f7f9934b0 TRAP: 0700 Not tainted (5.3.0-xive-nr-servers-5.3-gku+)
MSR: 9000000000029033 <SF,HV,EE,ME,IR,DR,RI,LE> CR: 48228222 XER: 20040000
CFAR: c000000000131a50 IRQMASK: 0
GPR00: c00000000053b0c8 c000003f7f993740 c0000000015ec500 0000000000000057
GPR04: 0000000000000001 0000000000000000 000049fb98484262 0000000000001bcf
GPR08: 0000000000000007 0000000000000007 0000000000000001 9000000000001033
GPR12: 0000000000008000 c000003ffffeb800 0000000000000000 000000012f4ce5a1
GPR16: 000000012ef5a0c8 0000000000000000 000000012f113bb0 0000000000000000
GPR20: 000000012f45d918 c000003f863758b0 c000003f86375870 0000000000000006
GPR24: c000003f86375a30 0000000000000007 c0002039373d9020 c0000000014c4a48
GPR28: 0000000000000001 c000003fe62a4f6b c00020394b2e9fab c000003fe62a4ec0
NIP [c00000000053b0cc] remove_proc_entry+0x1ec/0x200
LR [c00000000053b0c8] remove_proc_entry+0x1e8/0x200
Call Trace:
[c000003f7f993740] [c00000000053b0c8] remove_proc_entry+0x1e8/0x200 (unreliable)
[c000003f7f9937e0] [c0000000001d3654] unregister_irq_proc+0x114/0x150
[c000003f7f993880] [c0000000001c6284] free_desc+0x54/0xb0
[c000003f7f9938c0] [c0000000001c65ec] irq_free_descs+0xac/0x100
[c000003f7f993910] [c0000000001d1ff8] irq_dispose_mapping+0x68/0x80
[c000003f7f993940] [c00800000d44e8a4] kvmppc_xive_attach_escalation+0x1fc/0x270 [kvm]
[c000003f7f9939d0] [c00800000d45013c] kvmppc_xive_connect_vcpu+0x424/0x620 [kvm]
[c000003f7f993ac0] [c00800000d444428] kvm_arch_vcpu_ioctl+0x260/0x448 [kvm]
[c000003f7f993b90] [c00800000d43593c] kvm_vcpu_ioctl+0x154/0x7c8 [kvm]
[c000003f7f993d00] [c0000000004840f0] do_vfs_ioctl+0xe0/0xc30
[c000003f7f993db0] [c000000000484d44] ksys_ioctl+0x104/0x120
[c000003f7f993e00] [c000000000484d88] sys_ioctl+0x28/0x80
[c000003f7f993e20] [c00000000000b278] system_call+0x5c/0x68
Instruction dump:
2c230000 41820008 3923ff78 e8e900a0 3c82ff69 3c62ff8d 7fa6eb78 7fc5f378
3884f080 3863b948 4bbf6925 60000000 <0fe00000> 4bffff7c fba10088 4bbf6e41
---[ end trace b925b67a74a1d8d1 ]---
BUG: Kernel NULL pointer dereference at 0x00000010
Faulting instruction address: 0xc00800000d44fc04
Oops: Kernel access of bad area, sig: 11 [#1]
LE PAGE_SIZE=64K MMU=Radix MMU=Hash SMP NR_CPUS=2048 NUMA PowerNV
Modules linked in: kvm_hv kvm dm_mod vhost_net vhost tap xt_CHECKSUM iptable_mangle xt_MASQUERADE iptable_nat nf_nat xt_conntrack nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 ipt_REJECT nf_reject_ipv4 tun bridge stp llc ebtable_filter ebtables ip6table_filter ip6_tables iptable_filter squashfs loop fuse i2c_dev sg ofpart ocxl powernv_flash at24 xts mtd uio_pdrv_genirq vmx_crypto opal_prd ipmi_powernv uio ipmi_devintf ipmi_msghandler ibmpowernv ib_iser rdma_cm iw_cm ib_cm ib_core iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi ip_tables ext4 mbcache jbd2 raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor xor async_tx raid6_pq libcrc32c raid1 raid0 linear sd_mod ast i2c_algo_bit drm_vram_helper ttm drm_kms_helper syscopyarea sysfillrect sysimgblt fb_sys_fops drm ahci libahci libata tg3 drm_panel_orientation_quirks [last unloaded: kvm]
CPU: 24 PID: 88176 Comm: qemu-system-ppc Tainted: G W 5.3.0-xive-nr-servers-5.3-gku+ #38
NIP: c00800000d44fc04 LR: c00800000d44fc00 CTR: c0000000001cd970
REGS: c000003f7f9938e0 TRAP: 0300 Tainted: G W (5.3.0-xive-nr-servers-5.3-gku+)
MSR: 9000000000009033 <SF,HV,EE,ME,IR,DR,RI,LE> CR: 24228882 XER: 20040000
CFAR: c0000000001cd9ac DAR: 0000000000000010 DSISR: 40000000 IRQMASK: 0
GPR00: c00800000d44fc00 c000003f7f993b70 c00800000d468300 0000000000000000
GPR04: 00000000000000c7 0000000000000000 0000000000000000 c000003ffacd06d8
GPR08: 0000000000000000 c000003ffacd0738 0000000000000000 fffffffffffffffd
GPR12: 0000000000000040 c000003ffffeb800 0000000000000000 000000012f4ce5a1
GPR16: 000000012ef5a0c8 0000000000000000 000000012f113bb0 0000000000000000
GPR20: 000000012f45d918 00007ffffe0d9a80 000000012f4f5df0 000000012ef8c9f8
GPR24: 0000000000000001 0000000000000000 c000003fe4501ed0 c000003f8b1d0000
GPR28: c0000033314689c0 c000003fe4501c00 c000003fe4501e70 c000003fe4501e90
NIP [c00800000d44fc04] kvmppc_xive_cleanup_vcpu+0xfc/0x210 [kvm]
LR [c00800000d44fc00] kvmppc_xive_cleanup_vcpu+0xf8/0x210 [kvm]
Call Trace:
[c000003f7f993b70] [c00800000d44fc00] kvmppc_xive_cleanup_vcpu+0xf8/0x210 [kvm] (unreliable)
[c000003f7f993bd0] [c00800000d450bd4] kvmppc_xive_release+0xdc/0x1b0 [kvm]
[c000003f7f993c30] [c00800000d436a98] kvm_device_release+0xb0/0x110 [kvm]
[c000003f7f993c70] [c00000000046730c] __fput+0xec/0x320
[c000003f7f993cd0] [c000000000164ae0] task_work_run+0x150/0x1c0
[c000003f7f993d30] [c000000000025034] do_notify_resume+0x304/0x440
[c000003f7f993e20] [c00000000000dcc4] ret_from_except_lite+0x70/0x74
Instruction dump:
3bff0008 7fbfd040 419e0054 847e0004 2fa30000 419effec e93d0000 8929203c
2f890000 419effb8 4800821d e8410018 <e9230010> e9490008 9b2a0039 7c0004ac
---[ end trace b925b67a74a1d8d2 ]---
Kernel panic - not syncing: Fatal exception
This affects both XIVE and XICS-on-XIVE devices since the beginning.
Check the VP id instead of the vCPU id when a new vCPU is connected.
The allocation of the XIVE CPU structure in kvmppc_xive_connect_vcpu()
is moved after the check to avoid the need for rollback.
Cc: stable@vger.kernel.org # v4.12+
Signed-off-by: Greg Kurz <groug@kaod.org>
Reviewed-by: Cédric Le Goater <clg@kaod.org>
Signed-off-by: Paul Mackerras <paulus@ozlabs.org>
2019-09-27 19:53:43 +08:00
|
|
|
static inline bool kvmppc_xive_vp_in_use(struct kvm *kvm, u32 vp_id)
|
|
|
|
{
|
|
|
|
struct kvm_vcpu *vcpu = NULL;
|
2021-11-17 00:04:02 +08:00
|
|
|
unsigned long i;
|
KVM: PPC: Book3S HV: XIVE: Ensure VP isn't already in use
Connecting a vCPU to a XIVE KVM device means establishing a 1:1
association between a vCPU id and the offset (VP id) of a VP
structure within a fixed size block of VPs. We currently try to
enforce the 1:1 relationship by checking that a vCPU with the
same id isn't already connected. This is good but unfortunately
not enough because we don't map VP ids to raw vCPU ids but to
packed vCPU ids, and the packing function kvmppc_pack_vcpu_id()
isn't bijective by design. We got away with it because QEMU passes
vCPU ids that fit well in the packing pattern. But nothing prevents
userspace to come up with a forged vCPU id resulting in a packed id
collision which causes the KVM device to associate two vCPUs to the
same VP. This greatly confuses the irq layer and ultimately crashes
the kernel, as shown below.
Example: a guest with 1 guest thread per core, a core stride of
8 and 300 vCPUs has vCPU ids 0,8,16...2392. If QEMU is patched to
inject at some point an invalid vCPU id 348, which is the packed
version of itself and 2392, we get:
genirq: Flags mismatch irq 199. 00010000 (kvm-2-2392) vs. 00010000 (kvm-2-348)
CPU: 24 PID: 88176 Comm: qemu-system-ppc Not tainted 5.3.0-xive-nr-servers-5.3-gku+ #38
Call Trace:
[c000003f7f9937e0] [c000000000c0110c] dump_stack+0xb0/0xf4 (unreliable)
[c000003f7f993820] [c0000000001cb480] __setup_irq+0xa70/0xad0
[c000003f7f9938d0] [c0000000001cb75c] request_threaded_irq+0x13c/0x260
[c000003f7f993940] [c00800000d44e7ac] kvmppc_xive_attach_escalation+0x104/0x270 [kvm]
[c000003f7f9939d0] [c00800000d45013c] kvmppc_xive_connect_vcpu+0x424/0x620 [kvm]
[c000003f7f993ac0] [c00800000d444428] kvm_arch_vcpu_ioctl+0x260/0x448 [kvm]
[c000003f7f993b90] [c00800000d43593c] kvm_vcpu_ioctl+0x154/0x7c8 [kvm]
[c000003f7f993d00] [c0000000004840f0] do_vfs_ioctl+0xe0/0xc30
[c000003f7f993db0] [c000000000484d44] ksys_ioctl+0x104/0x120
[c000003f7f993e00] [c000000000484d88] sys_ioctl+0x28/0x80
[c000003f7f993e20] [c00000000000b278] system_call+0x5c/0x68
xive-kvm: Failed to request escalation interrupt for queue 0 of VCPU 2392
------------[ cut here ]------------
remove_proc_entry: removing non-empty directory 'irq/199', leaking at least 'kvm-2-348'
WARNING: CPU: 24 PID: 88176 at /home/greg/Work/linux/kernel-kvm-ppc/fs/proc/generic.c:684 remove_proc_entry+0x1ec/0x200
Modules linked in: kvm_hv kvm dm_mod vhost_net vhost tap xt_CHECKSUM iptable_mangle xt_MASQUERADE iptable_nat nf_nat xt_conntrack nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 ipt_REJECT nf_reject_ipv4 tun bridge stp llc ebtable_filter ebtables ip6table_filter ip6_tables iptable_filter squashfs loop fuse i2c_dev sg ofpart ocxl powernv_flash at24 xts mtd uio_pdrv_genirq vmx_crypto opal_prd ipmi_powernv uio ipmi_devintf ipmi_msghandler ibmpowernv ib_iser rdma_cm iw_cm ib_cm ib_core iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi ip_tables ext4 mbcache jbd2 raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor xor async_tx raid6_pq libcrc32c raid1 raid0 linear sd_mod ast i2c_algo_bit drm_vram_helper ttm drm_kms_helper syscopyarea sysfillrect sysimgblt fb_sys_fops drm ahci libahci libata tg3 drm_panel_orientation_quirks [last unloaded: kvm]
CPU: 24 PID: 88176 Comm: qemu-system-ppc Not tainted 5.3.0-xive-nr-servers-5.3-gku+ #38
NIP: c00000000053b0cc LR: c00000000053b0c8 CTR: c0000000000ba3b0
REGS: c000003f7f9934b0 TRAP: 0700 Not tainted (5.3.0-xive-nr-servers-5.3-gku+)
MSR: 9000000000029033 <SF,HV,EE,ME,IR,DR,RI,LE> CR: 48228222 XER: 20040000
CFAR: c000000000131a50 IRQMASK: 0
GPR00: c00000000053b0c8 c000003f7f993740 c0000000015ec500 0000000000000057
GPR04: 0000000000000001 0000000000000000 000049fb98484262 0000000000001bcf
GPR08: 0000000000000007 0000000000000007 0000000000000001 9000000000001033
GPR12: 0000000000008000 c000003ffffeb800 0000000000000000 000000012f4ce5a1
GPR16: 000000012ef5a0c8 0000000000000000 000000012f113bb0 0000000000000000
GPR20: 000000012f45d918 c000003f863758b0 c000003f86375870 0000000000000006
GPR24: c000003f86375a30 0000000000000007 c0002039373d9020 c0000000014c4a48
GPR28: 0000000000000001 c000003fe62a4f6b c00020394b2e9fab c000003fe62a4ec0
NIP [c00000000053b0cc] remove_proc_entry+0x1ec/0x200
LR [c00000000053b0c8] remove_proc_entry+0x1e8/0x200
Call Trace:
[c000003f7f993740] [c00000000053b0c8] remove_proc_entry+0x1e8/0x200 (unreliable)
[c000003f7f9937e0] [c0000000001d3654] unregister_irq_proc+0x114/0x150
[c000003f7f993880] [c0000000001c6284] free_desc+0x54/0xb0
[c000003f7f9938c0] [c0000000001c65ec] irq_free_descs+0xac/0x100
[c000003f7f993910] [c0000000001d1ff8] irq_dispose_mapping+0x68/0x80
[c000003f7f993940] [c00800000d44e8a4] kvmppc_xive_attach_escalation+0x1fc/0x270 [kvm]
[c000003f7f9939d0] [c00800000d45013c] kvmppc_xive_connect_vcpu+0x424/0x620 [kvm]
[c000003f7f993ac0] [c00800000d444428] kvm_arch_vcpu_ioctl+0x260/0x448 [kvm]
[c000003f7f993b90] [c00800000d43593c] kvm_vcpu_ioctl+0x154/0x7c8 [kvm]
[c000003f7f993d00] [c0000000004840f0] do_vfs_ioctl+0xe0/0xc30
[c000003f7f993db0] [c000000000484d44] ksys_ioctl+0x104/0x120
[c000003f7f993e00] [c000000000484d88] sys_ioctl+0x28/0x80
[c000003f7f993e20] [c00000000000b278] system_call+0x5c/0x68
Instruction dump:
2c230000 41820008 3923ff78 e8e900a0 3c82ff69 3c62ff8d 7fa6eb78 7fc5f378
3884f080 3863b948 4bbf6925 60000000 <0fe00000> 4bffff7c fba10088 4bbf6e41
---[ end trace b925b67a74a1d8d1 ]---
BUG: Kernel NULL pointer dereference at 0x00000010
Faulting instruction address: 0xc00800000d44fc04
Oops: Kernel access of bad area, sig: 11 [#1]
LE PAGE_SIZE=64K MMU=Radix MMU=Hash SMP NR_CPUS=2048 NUMA PowerNV
Modules linked in: kvm_hv kvm dm_mod vhost_net vhost tap xt_CHECKSUM iptable_mangle xt_MASQUERADE iptable_nat nf_nat xt_conntrack nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 ipt_REJECT nf_reject_ipv4 tun bridge stp llc ebtable_filter ebtables ip6table_filter ip6_tables iptable_filter squashfs loop fuse i2c_dev sg ofpart ocxl powernv_flash at24 xts mtd uio_pdrv_genirq vmx_crypto opal_prd ipmi_powernv uio ipmi_devintf ipmi_msghandler ibmpowernv ib_iser rdma_cm iw_cm ib_cm ib_core iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi ip_tables ext4 mbcache jbd2 raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor xor async_tx raid6_pq libcrc32c raid1 raid0 linear sd_mod ast i2c_algo_bit drm_vram_helper ttm drm_kms_helper syscopyarea sysfillrect sysimgblt fb_sys_fops drm ahci libahci libata tg3 drm_panel_orientation_quirks [last unloaded: kvm]
CPU: 24 PID: 88176 Comm: qemu-system-ppc Tainted: G W 5.3.0-xive-nr-servers-5.3-gku+ #38
NIP: c00800000d44fc04 LR: c00800000d44fc00 CTR: c0000000001cd970
REGS: c000003f7f9938e0 TRAP: 0300 Tainted: G W (5.3.0-xive-nr-servers-5.3-gku+)
MSR: 9000000000009033 <SF,HV,EE,ME,IR,DR,RI,LE> CR: 24228882 XER: 20040000
CFAR: c0000000001cd9ac DAR: 0000000000000010 DSISR: 40000000 IRQMASK: 0
GPR00: c00800000d44fc00 c000003f7f993b70 c00800000d468300 0000000000000000
GPR04: 00000000000000c7 0000000000000000 0000000000000000 c000003ffacd06d8
GPR08: 0000000000000000 c000003ffacd0738 0000000000000000 fffffffffffffffd
GPR12: 0000000000000040 c000003ffffeb800 0000000000000000 000000012f4ce5a1
GPR16: 000000012ef5a0c8 0000000000000000 000000012f113bb0 0000000000000000
GPR20: 000000012f45d918 00007ffffe0d9a80 000000012f4f5df0 000000012ef8c9f8
GPR24: 0000000000000001 0000000000000000 c000003fe4501ed0 c000003f8b1d0000
GPR28: c0000033314689c0 c000003fe4501c00 c000003fe4501e70 c000003fe4501e90
NIP [c00800000d44fc04] kvmppc_xive_cleanup_vcpu+0xfc/0x210 [kvm]
LR [c00800000d44fc00] kvmppc_xive_cleanup_vcpu+0xf8/0x210 [kvm]
Call Trace:
[c000003f7f993b70] [c00800000d44fc00] kvmppc_xive_cleanup_vcpu+0xf8/0x210 [kvm] (unreliable)
[c000003f7f993bd0] [c00800000d450bd4] kvmppc_xive_release+0xdc/0x1b0 [kvm]
[c000003f7f993c30] [c00800000d436a98] kvm_device_release+0xb0/0x110 [kvm]
[c000003f7f993c70] [c00000000046730c] __fput+0xec/0x320
[c000003f7f993cd0] [c000000000164ae0] task_work_run+0x150/0x1c0
[c000003f7f993d30] [c000000000025034] do_notify_resume+0x304/0x440
[c000003f7f993e20] [c00000000000dcc4] ret_from_except_lite+0x70/0x74
Instruction dump:
3bff0008 7fbfd040 419e0054 847e0004 2fa30000 419effec e93d0000 8929203c
2f890000 419effb8 4800821d e8410018 <e9230010> e9490008 9b2a0039 7c0004ac
---[ end trace b925b67a74a1d8d2 ]---
Kernel panic - not syncing: Fatal exception
This affects both XIVE and XICS-on-XIVE devices since the beginning.
Check the VP id instead of the vCPU id when a new vCPU is connected.
The allocation of the XIVE CPU structure in kvmppc_xive_connect_vcpu()
is moved after the check to avoid the need for rollback.
Cc: stable@vger.kernel.org # v4.12+
Signed-off-by: Greg Kurz <groug@kaod.org>
Reviewed-by: Cédric Le Goater <clg@kaod.org>
Signed-off-by: Paul Mackerras <paulus@ozlabs.org>
2019-09-27 19:53:43 +08:00
|
|
|
|
|
|
|
kvm_for_each_vcpu(i, vcpu, kvm) {
|
|
|
|
if (vcpu->arch.xive_vcpu && vp_id == vcpu->arch.xive_vcpu->vp_id)
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2017-04-05 15:54:56 +08:00
|
|
|
/*
|
|
|
|
* Mapping between guest priorities and host priorities
|
|
|
|
* is as follow.
|
|
|
|
*
|
|
|
|
* Guest request for 0...6 are honored. Guest request for anything
|
2018-01-12 10:37:12 +08:00
|
|
|
* higher results in a priority of 6 being applied.
|
2017-04-05 15:54:56 +08:00
|
|
|
*
|
|
|
|
* Similar mapping is done for CPPR values
|
|
|
|
*/
|
|
|
|
static inline u8 xive_prio_from_guest(u8 prio)
|
|
|
|
{
|
2018-01-12 10:37:12 +08:00
|
|
|
if (prio == 0xff || prio < 6)
|
2017-04-05 15:54:56 +08:00
|
|
|
return prio;
|
2018-01-12 10:37:12 +08:00
|
|
|
return 6;
|
2017-04-05 15:54:56 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline u8 xive_prio_to_guest(u8 prio)
|
|
|
|
{
|
2018-01-12 10:37:12 +08:00
|
|
|
return prio;
|
2017-04-05 15:54:56 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 __xive_read_eq(__be32 *qpage, u32 msk, u32 *idx, u32 *toggle)
|
|
|
|
{
|
|
|
|
u32 cur;
|
|
|
|
|
|
|
|
if (!qpage)
|
|
|
|
return 0;
|
|
|
|
cur = be32_to_cpup(qpage + *idx);
|
|
|
|
if ((cur >> 31) == *toggle)
|
|
|
|
return 0;
|
|
|
|
*idx = (*idx + 1) & msk;
|
|
|
|
if (*idx == 0)
|
|
|
|
(*toggle) ^= 1;
|
|
|
|
return cur & 0x7fffffff;
|
|
|
|
}
|
|
|
|
|
2019-04-18 18:39:28 +08:00
|
|
|
/*
|
|
|
|
* Common Xive routines for XICS-over-XIVE and XIVE native
|
|
|
|
*/
|
|
|
|
void kvmppc_xive_disable_vcpu_interrupts(struct kvm_vcpu *vcpu);
|
|
|
|
int kvmppc_xive_debug_show_queues(struct seq_file *m, struct kvm_vcpu *vcpu);
|
2020-12-11 01:14:38 +08:00
|
|
|
void kvmppc_xive_debug_show_sources(struct seq_file *m,
|
|
|
|
struct kvmppc_xive_src_block *sb);
|
2019-04-18 18:39:29 +08:00
|
|
|
struct kvmppc_xive_src_block *kvmppc_xive_create_src_block(
|
|
|
|
struct kvmppc_xive *xive, int irq);
|
|
|
|
void kvmppc_xive_free_sources(struct kvmppc_xive_src_block *sb);
|
2019-04-18 18:39:30 +08:00
|
|
|
int kvmppc_xive_select_target(struct kvm *kvm, u32 *server, u8 prio);
|
2019-04-18 18:39:31 +08:00
|
|
|
int kvmppc_xive_attach_escalation(struct kvm_vcpu *vcpu, u8 prio,
|
|
|
|
bool single_escalation);
|
2019-04-18 18:39:42 +08:00
|
|
|
struct kvmppc_xive *kvmppc_xive_get_device(struct kvm *kvm, u32 type);
|
2022-04-13 18:55:07 +08:00
|
|
|
void xive_cleanup_single_escalation(struct kvm_vcpu *vcpu, int irq);
|
2019-09-27 19:53:55 +08:00
|
|
|
int kvmppc_xive_compute_vp_id(struct kvmppc_xive *xive, u32 cpu, u32 *vp);
|
2019-09-27 19:54:01 +08:00
|
|
|
int kvmppc_xive_set_nr_servers(struct kvmppc_xive *xive, u64 addr);
|
2021-07-20 21:42:09 +08:00
|
|
|
bool kvmppc_xive_check_save_restore(struct kvm_vcpu *vcpu);
|
2019-04-18 18:39:28 +08:00
|
|
|
|
2021-07-20 21:42:08 +08:00
|
|
|
static inline bool kvmppc_xive_has_single_escalation(struct kvmppc_xive *xive)
|
|
|
|
{
|
|
|
|
return xive->flags & KVMPPC_XIVE_FLAG_SINGLE_ESCALATION;
|
|
|
|
}
|
|
|
|
|
2017-04-05 15:54:56 +08:00
|
|
|
#endif /* CONFIG_KVM_XICS */
|
|
|
|
#endif /* _KVM_PPC_BOOK3S_XICS_H */
|