arm64 fixes for -rc7
- Fix GICv2 emulation bug (KVM) - Fix deadlock in virtual GIC interrupt injection code (KVM) - Fix kprobes blacklist init failure due to broken kallsyms lookup -----BEGIN PGP SIGNATURE----- iQEzBAABCgAdFiEEPxTL6PPUbjXGY88ct6xw3ITBYzQFAl1mW/0ACgkQt6xw3ITB YzQWEQgAnG+20c1uzPqIEME3o0bCXsci0I1w4yPYPH/NVrfmJdAjfway/cLZlMCh JYKBTc392kfYXgzAvMqvjNe+9O/qg9MnMxV7Afjfk9e0k+41Kcw/7U39/B6q4yOg zIz4YW6SdbbbUfY9VgR8SjsZ3HwJoP8YVNEv3W259funtFHkKDO35h1iG1f57GWL UELjEnag9dmCH35ykt7+SC4woBGjOQnic2XLSS7VZuU/26TfnUyo2HCdDcLcViCZ DHDQgaHAscb+tgJjeNtNEeRu+GPBG9LOhPw4EpBjGHeoJla7z74GOJPMEt4Ufoqd qG9TPfOFAFM30/PAdlWF3SjfzqK3PA== =WYzY -----END PGP SIGNATURE----- Merge tag 'arm64-fixes' of git://git.kernel.org/pub/scm/linux/kernel/git/arm64/linux Pull arm64 fixes from Will Deacon: "Hot on the heels of our last set of fixes are a few more for -rc7. Two of them are fixing issues with our virtual interrupt controller implementation in KVM/arm, while the other is a longstanding but straightforward kallsyms fix which was been acked by Masami and resolves an initialisation failure in kprobes observed on arm64. - Fix GICv2 emulation bug (KVM) - Fix deadlock in virtual GIC interrupt injection code (KVM) - Fix kprobes blacklist init failure due to broken kallsyms lookup" * tag 'arm64-fixes' of git://git.kernel.org/pub/scm/linux/kernel/git/arm64/linux: KVM: arm/arm64: vgic-v2: Handle SGI bits in GICD_I{S,C}PENDR0 as WI KVM: arm/arm64: vgic: Fix potential deadlock when ap_list is long kallsyms: Don't let kallsyms_lookup_size_offset() fail on retrieving the first symbol
This commit is contained in:
commit
9cf6b756cd
@ -263,8 +263,10 @@ int kallsyms_lookup_size_offset(unsigned long addr, unsigned long *symbolsize,
|
|||||||
{
|
{
|
||||||
char namebuf[KSYM_NAME_LEN];
|
char namebuf[KSYM_NAME_LEN];
|
||||||
|
|
||||||
if (is_ksym_addr(addr))
|
if (is_ksym_addr(addr)) {
|
||||||
return !!get_symbol_pos(addr, symbolsize, offset);
|
get_symbol_pos(addr, symbolsize, offset);
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
return !!module_address_lookup(addr, symbolsize, offset, NULL, namebuf) ||
|
return !!module_address_lookup(addr, symbolsize, offset, NULL, namebuf) ||
|
||||||
!!__bpf_address_lookup(addr, symbolsize, offset, namebuf);
|
!!__bpf_address_lookup(addr, symbolsize, offset, namebuf);
|
||||||
}
|
}
|
||||||
|
@ -211,6 +211,12 @@ static void vgic_hw_irq_spending(struct kvm_vcpu *vcpu, struct vgic_irq *irq,
|
|||||||
vgic_irq_set_phys_active(irq, true);
|
vgic_irq_set_phys_active(irq, true);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static bool is_vgic_v2_sgi(struct kvm_vcpu *vcpu, struct vgic_irq *irq)
|
||||||
|
{
|
||||||
|
return (vgic_irq_is_sgi(irq->intid) &&
|
||||||
|
vcpu->kvm->arch.vgic.vgic_model == KVM_DEV_TYPE_ARM_VGIC_V2);
|
||||||
|
}
|
||||||
|
|
||||||
void vgic_mmio_write_spending(struct kvm_vcpu *vcpu,
|
void vgic_mmio_write_spending(struct kvm_vcpu *vcpu,
|
||||||
gpa_t addr, unsigned int len,
|
gpa_t addr, unsigned int len,
|
||||||
unsigned long val)
|
unsigned long val)
|
||||||
@ -223,6 +229,12 @@ void vgic_mmio_write_spending(struct kvm_vcpu *vcpu,
|
|||||||
for_each_set_bit(i, &val, len * 8) {
|
for_each_set_bit(i, &val, len * 8) {
|
||||||
struct vgic_irq *irq = vgic_get_irq(vcpu->kvm, vcpu, intid + i);
|
struct vgic_irq *irq = vgic_get_irq(vcpu->kvm, vcpu, intid + i);
|
||||||
|
|
||||||
|
/* GICD_ISPENDR0 SGI bits are WI */
|
||||||
|
if (is_vgic_v2_sgi(vcpu, irq)) {
|
||||||
|
vgic_put_irq(vcpu->kvm, irq);
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
raw_spin_lock_irqsave(&irq->irq_lock, flags);
|
raw_spin_lock_irqsave(&irq->irq_lock, flags);
|
||||||
if (irq->hw)
|
if (irq->hw)
|
||||||
vgic_hw_irq_spending(vcpu, irq, is_uaccess);
|
vgic_hw_irq_spending(vcpu, irq, is_uaccess);
|
||||||
@ -270,6 +282,12 @@ void vgic_mmio_write_cpending(struct kvm_vcpu *vcpu,
|
|||||||
for_each_set_bit(i, &val, len * 8) {
|
for_each_set_bit(i, &val, len * 8) {
|
||||||
struct vgic_irq *irq = vgic_get_irq(vcpu->kvm, vcpu, intid + i);
|
struct vgic_irq *irq = vgic_get_irq(vcpu->kvm, vcpu, intid + i);
|
||||||
|
|
||||||
|
/* GICD_ICPENDR0 SGI bits are WI */
|
||||||
|
if (is_vgic_v2_sgi(vcpu, irq)) {
|
||||||
|
vgic_put_irq(vcpu->kvm, irq);
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
raw_spin_lock_irqsave(&irq->irq_lock, flags);
|
raw_spin_lock_irqsave(&irq->irq_lock, flags);
|
||||||
|
|
||||||
if (irq->hw)
|
if (irq->hw)
|
||||||
|
@ -184,7 +184,10 @@ void vgic_v2_populate_lr(struct kvm_vcpu *vcpu, struct vgic_irq *irq, int lr)
|
|||||||
if (vgic_irq_is_sgi(irq->intid)) {
|
if (vgic_irq_is_sgi(irq->intid)) {
|
||||||
u32 src = ffs(irq->source);
|
u32 src = ffs(irq->source);
|
||||||
|
|
||||||
BUG_ON(!src);
|
if (WARN_RATELIMIT(!src, "No SGI source for INTID %d\n",
|
||||||
|
irq->intid))
|
||||||
|
return;
|
||||||
|
|
||||||
val |= (src - 1) << GICH_LR_PHYSID_CPUID_SHIFT;
|
val |= (src - 1) << GICH_LR_PHYSID_CPUID_SHIFT;
|
||||||
irq->source &= ~(1 << (src - 1));
|
irq->source &= ~(1 << (src - 1));
|
||||||
if (irq->source) {
|
if (irq->source) {
|
||||||
|
@ -167,7 +167,10 @@ void vgic_v3_populate_lr(struct kvm_vcpu *vcpu, struct vgic_irq *irq, int lr)
|
|||||||
model == KVM_DEV_TYPE_ARM_VGIC_V2) {
|
model == KVM_DEV_TYPE_ARM_VGIC_V2) {
|
||||||
u32 src = ffs(irq->source);
|
u32 src = ffs(irq->source);
|
||||||
|
|
||||||
BUG_ON(!src);
|
if (WARN_RATELIMIT(!src, "No SGI source for INTID %d\n",
|
||||||
|
irq->intid))
|
||||||
|
return;
|
||||||
|
|
||||||
val |= (src - 1) << GICH_LR_PHYSID_CPUID_SHIFT;
|
val |= (src - 1) << GICH_LR_PHYSID_CPUID_SHIFT;
|
||||||
irq->source &= ~(1 << (src - 1));
|
irq->source &= ~(1 << (src - 1));
|
||||||
if (irq->source) {
|
if (irq->source) {
|
||||||
|
@ -254,6 +254,13 @@ static int vgic_irq_cmp(void *priv, struct list_head *a, struct list_head *b)
|
|||||||
bool penda, pendb;
|
bool penda, pendb;
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* list_sort may call this function with the same element when
|
||||||
|
* the list is fairly long.
|
||||||
|
*/
|
||||||
|
if (unlikely(irqa == irqb))
|
||||||
|
return 0;
|
||||||
|
|
||||||
raw_spin_lock(&irqa->irq_lock);
|
raw_spin_lock(&irqa->irq_lock);
|
||||||
raw_spin_lock_nested(&irqb->irq_lock, SINGLE_DEPTH_NESTING);
|
raw_spin_lock_nested(&irqb->irq_lock, SINGLE_DEPTH_NESTING);
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user