This is the 5.4.173 stable release
-----BEGIN PGP SIGNATURE----- iQIzBAABCAAdFiEEZH8oZUiU471FcZm+ONu9yGCSaT4FAmHpGw4ACgkQONu9yGCS aT5sgg//YM/5m2nAXX1AiJra494E/2gqWS6jBG+XHVIOoJfz+rrqey7rFHYkL/+r nt8kSHH5Ts84AhZ8EioJLmhaabPDjj/uZbiqhilVk6RBYbnMU+beYdETT/RVZn55 9Jrankoe7ZqCV9oVpIK30o+/+Izqfxn9F298SXVzwAjjI5QWabN74pprZqJkdGCn eFW1EzEIMak7h+E5joYCF1fZj2lMry60dBCLKYry+jprufUks3cdxgBLmuuvFjr6 P9gYJliEor/nIT+dpDfJDItVwONBAE/YxU5k6hpui0piOkPljuxNDQCWE3XR9PGa 8KbUaikD8vpsxHUhjKuObq4dqWujFGukDgiHAMZwipWbS/4TOj4zDayHNsZgu25n D/T0LCfbjoAboJwSQeeniD+uS/C57ev6uj3JVBj4TovhIlwUaavc9t4KXuWsAI8n WQglrfNmbPSD9H5Y1wwdotxrNuoSViT+5lOSEQ4g4TwVEEaYi4u/pQhil2woslnE g+mwL9WmyjMrnzIZtibDCyVYm5f8aLIQGiNn8LNdEiLHq510+eOSHT+pOhZ5om4v uPp7PI1F63N5O4TG49PFbS1ztVLG9TYOOpivRSQXY/++ly+gNSVz79KftX++gQRJ RGAPMVZX0iZsPzGp6e6DVbS9Y0i91JvLfSbcxyHGAjnlSr5bd6I= =SjmA -----END PGP SIGNATURE----- Merge 5.4.173 into android11-5.4-lts Changes in 5.4.173 kbuild: Add $(KBUILD_HOSTLDFLAGS) to 'has_libelf' test devtmpfs regression fix: reconfigure on each mount orangefs: Fix the size of a memory allocation in orangefs_bufmap_alloc() vfs: fs_context: fix up param length parsing in legacy_parse_param perf: Protect perf_guest_cbs with RCU KVM: s390: Clarify SIGP orders versus STOP/RESTART media: uvcvideo: fix division by zero at stream start rtlwifi: rtl8192cu: Fix WARNING when calling local_irq_restore() with interrupts enabled firmware: qemu_fw_cfg: fix sysfs information leak firmware: qemu_fw_cfg: fix NULL-pointer deref on duplicate entries firmware: qemu_fw_cfg: fix kobject leak in probe error path KVM: x86: remove PMU FIXED_CTR3 from msrs_to_save_all ALSA: hda/realtek - Fix silent output on Gigabyte X570 Aorus Master after reboot from Windows mtd: fixup CFI on ixp4xx ARM: 9025/1: Kconfig: CPU_BIG_ENDIAN depends on !LD_IS_LLD Linux 5.4.173 Signed-off-by: Greg Kroah-Hartman <gregkh@google.com> Change-Id: I1127aba3d2970fc47a45fa127abed48255bd53b7
This commit is contained in:
commit
5d5005448e
4
Makefile
4
Makefile
@ -1,7 +1,7 @@
|
||||
# SPDX-License-Identifier: GPL-2.0
|
||||
VERSION = 5
|
||||
PATCHLEVEL = 4
|
||||
SUBLEVEL = 172
|
||||
SUBLEVEL = 173
|
||||
EXTRAVERSION =
|
||||
NAME = Kleptomaniac Octopus
|
||||
|
||||
@ -1095,7 +1095,7 @@ HOST_LIBELF_LIBS = $(shell pkg-config libelf --libs 2>/dev/null || echo -lelf)
|
||||
|
||||
ifdef CONFIG_STACK_VALIDATION
|
||||
has_libelf := $(call try-run,\
|
||||
echo "int main() {}" | $(HOSTCC) -xc -o /dev/null $(HOST_LIBELF_LIBS) -,1,0)
|
||||
echo "int main() {}" | $(HOSTCC) $(KBUILD_HOSTLDFLAGS) -xc -o /dev/null $(HOST_LIBELF_LIBS) -,1,0)
|
||||
ifeq ($(has_libelf),1)
|
||||
objtool_target := tools/objtool FORCE
|
||||
else
|
||||
|
@ -62,9 +62,10 @@ user_backtrace(struct frame_tail __user *tail,
|
||||
void
|
||||
perf_callchain_user(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs)
|
||||
{
|
||||
struct perf_guest_info_callbacks *guest_cbs = perf_get_guest_cbs();
|
||||
struct frame_tail __user *tail;
|
||||
|
||||
if (perf_guest_cbs && perf_guest_cbs->is_in_guest()) {
|
||||
if (guest_cbs && guest_cbs->is_in_guest()) {
|
||||
/* We don't support guest os callchain now */
|
||||
return;
|
||||
}
|
||||
@ -98,9 +99,10 @@ callchain_trace(struct stackframe *fr,
|
||||
void
|
||||
perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs)
|
||||
{
|
||||
struct perf_guest_info_callbacks *guest_cbs = perf_get_guest_cbs();
|
||||
struct stackframe fr;
|
||||
|
||||
if (perf_guest_cbs && perf_guest_cbs->is_in_guest()) {
|
||||
if (guest_cbs && guest_cbs->is_in_guest()) {
|
||||
/* We don't support guest os callchain now */
|
||||
return;
|
||||
}
|
||||
@ -111,18 +113,21 @@ perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, struct pt_regs *re
|
||||
|
||||
unsigned long perf_instruction_pointer(struct pt_regs *regs)
|
||||
{
|
||||
if (perf_guest_cbs && perf_guest_cbs->is_in_guest())
|
||||
return perf_guest_cbs->get_guest_ip();
|
||||
struct perf_guest_info_callbacks *guest_cbs = perf_get_guest_cbs();
|
||||
|
||||
if (guest_cbs && guest_cbs->is_in_guest())
|
||||
return guest_cbs->get_guest_ip();
|
||||
|
||||
return instruction_pointer(regs);
|
||||
}
|
||||
|
||||
unsigned long perf_misc_flags(struct pt_regs *regs)
|
||||
{
|
||||
struct perf_guest_info_callbacks *guest_cbs = perf_get_guest_cbs();
|
||||
int misc = 0;
|
||||
|
||||
if (perf_guest_cbs && perf_guest_cbs->is_in_guest()) {
|
||||
if (perf_guest_cbs->is_user_mode())
|
||||
if (guest_cbs && guest_cbs->is_in_guest()) {
|
||||
if (guest_cbs->is_user_mode())
|
||||
misc |= PERF_RECORD_MISC_GUEST_USER;
|
||||
else
|
||||
misc |= PERF_RECORD_MISC_GUEST_KERNEL;
|
||||
|
@ -743,6 +743,7 @@ config SWP_EMULATE
|
||||
config CPU_BIG_ENDIAN
|
||||
bool "Build big-endian kernel"
|
||||
depends on ARCH_SUPPORTS_BIG_ENDIAN
|
||||
depends on !LD_IS_LLD
|
||||
help
|
||||
Say Y if you plan on running a kernel in big-endian mode.
|
||||
Note that your board must be properly built and your board
|
||||
|
@ -102,7 +102,9 @@ compat_user_backtrace(struct compat_frame_tail __user *tail,
|
||||
void perf_callchain_user(struct perf_callchain_entry_ctx *entry,
|
||||
struct pt_regs *regs)
|
||||
{
|
||||
if (perf_guest_cbs && perf_guest_cbs->is_in_guest()) {
|
||||
struct perf_guest_info_callbacks *guest_cbs = perf_get_guest_cbs();
|
||||
|
||||
if (guest_cbs && guest_cbs->is_in_guest()) {
|
||||
/* We don't support guest os callchain now */
|
||||
return;
|
||||
}
|
||||
@ -147,9 +149,10 @@ static int callchain_trace(struct stackframe *frame, void *data)
|
||||
void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry,
|
||||
struct pt_regs *regs)
|
||||
{
|
||||
struct perf_guest_info_callbacks *guest_cbs = perf_get_guest_cbs();
|
||||
struct stackframe frame;
|
||||
|
||||
if (perf_guest_cbs && perf_guest_cbs->is_in_guest()) {
|
||||
if (guest_cbs && guest_cbs->is_in_guest()) {
|
||||
/* We don't support guest os callchain now */
|
||||
return;
|
||||
}
|
||||
@ -160,18 +163,21 @@ void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry,
|
||||
|
||||
unsigned long perf_instruction_pointer(struct pt_regs *regs)
|
||||
{
|
||||
if (perf_guest_cbs && perf_guest_cbs->is_in_guest())
|
||||
return perf_guest_cbs->get_guest_ip();
|
||||
struct perf_guest_info_callbacks *guest_cbs = perf_get_guest_cbs();
|
||||
|
||||
if (guest_cbs && guest_cbs->is_in_guest())
|
||||
return guest_cbs->get_guest_ip();
|
||||
|
||||
return instruction_pointer(regs);
|
||||
}
|
||||
|
||||
unsigned long perf_misc_flags(struct pt_regs *regs)
|
||||
{
|
||||
struct perf_guest_info_callbacks *guest_cbs = perf_get_guest_cbs();
|
||||
int misc = 0;
|
||||
|
||||
if (perf_guest_cbs && perf_guest_cbs->is_in_guest()) {
|
||||
if (perf_guest_cbs->is_user_mode())
|
||||
if (guest_cbs && guest_cbs->is_in_guest()) {
|
||||
if (guest_cbs->is_user_mode())
|
||||
misc |= PERF_RECORD_MISC_GUEST_USER;
|
||||
else
|
||||
misc |= PERF_RECORD_MISC_GUEST_KERNEL;
|
||||
|
@ -86,10 +86,11 @@ static unsigned long user_backtrace(struct perf_callchain_entry_ctx *entry,
|
||||
void perf_callchain_user(struct perf_callchain_entry_ctx *entry,
|
||||
struct pt_regs *regs)
|
||||
{
|
||||
struct perf_guest_info_callbacks *guest_cbs = perf_get_guest_cbs();
|
||||
unsigned long fp = 0;
|
||||
|
||||
/* C-SKY does not support virtualization. */
|
||||
if (perf_guest_cbs && perf_guest_cbs->is_in_guest())
|
||||
if (guest_cbs && guest_cbs->is_in_guest())
|
||||
return;
|
||||
|
||||
fp = regs->regs[4];
|
||||
@ -110,10 +111,11 @@ void perf_callchain_user(struct perf_callchain_entry_ctx *entry,
|
||||
void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry,
|
||||
struct pt_regs *regs)
|
||||
{
|
||||
struct perf_guest_info_callbacks *guest_cbs = perf_get_guest_cbs();
|
||||
struct stackframe fr;
|
||||
|
||||
/* C-SKY does not support virtualization. */
|
||||
if (perf_guest_cbs && perf_guest_cbs->is_in_guest()) {
|
||||
if (guest_cbs && guest_cbs->is_in_guest()) {
|
||||
pr_warn("C-SKY does not support perf in guest mode!");
|
||||
return;
|
||||
}
|
||||
|
@ -1363,6 +1363,7 @@ void
|
||||
perf_callchain_user(struct perf_callchain_entry_ctx *entry,
|
||||
struct pt_regs *regs)
|
||||
{
|
||||
struct perf_guest_info_callbacks *guest_cbs = perf_get_guest_cbs();
|
||||
unsigned long fp = 0;
|
||||
unsigned long gp = 0;
|
||||
unsigned long lp = 0;
|
||||
@ -1371,7 +1372,7 @@ perf_callchain_user(struct perf_callchain_entry_ctx *entry,
|
||||
|
||||
leaf_fp = 0;
|
||||
|
||||
if (perf_guest_cbs && perf_guest_cbs->is_in_guest()) {
|
||||
if (guest_cbs && guest_cbs->is_in_guest()) {
|
||||
/* We don't support guest os callchain now */
|
||||
return;
|
||||
}
|
||||
@ -1479,9 +1480,10 @@ void
|
||||
perf_callchain_kernel(struct perf_callchain_entry_ctx *entry,
|
||||
struct pt_regs *regs)
|
||||
{
|
||||
struct perf_guest_info_callbacks *guest_cbs = perf_get_guest_cbs();
|
||||
struct stackframe fr;
|
||||
|
||||
if (perf_guest_cbs && perf_guest_cbs->is_in_guest()) {
|
||||
if (guest_cbs && guest_cbs->is_in_guest()) {
|
||||
/* We don't support guest os callchain now */
|
||||
return;
|
||||
}
|
||||
@ -1493,20 +1495,23 @@ perf_callchain_kernel(struct perf_callchain_entry_ctx *entry,
|
||||
|
||||
unsigned long perf_instruction_pointer(struct pt_regs *regs)
|
||||
{
|
||||
struct perf_guest_info_callbacks *guest_cbs = perf_get_guest_cbs();
|
||||
|
||||
/* However, NDS32 does not support virtualization */
|
||||
if (perf_guest_cbs && perf_guest_cbs->is_in_guest())
|
||||
return perf_guest_cbs->get_guest_ip();
|
||||
if (guest_cbs && guest_cbs->is_in_guest())
|
||||
return guest_cbs->get_guest_ip();
|
||||
|
||||
return instruction_pointer(regs);
|
||||
}
|
||||
|
||||
unsigned long perf_misc_flags(struct pt_regs *regs)
|
||||
{
|
||||
struct perf_guest_info_callbacks *guest_cbs = perf_get_guest_cbs();
|
||||
int misc = 0;
|
||||
|
||||
/* However, NDS32 does not support virtualization */
|
||||
if (perf_guest_cbs && perf_guest_cbs->is_in_guest()) {
|
||||
if (perf_guest_cbs->is_user_mode())
|
||||
if (guest_cbs && guest_cbs->is_in_guest()) {
|
||||
if (guest_cbs->is_user_mode())
|
||||
misc |= PERF_RECORD_MISC_GUEST_USER;
|
||||
else
|
||||
misc |= PERF_RECORD_MISC_GUEST_KERNEL;
|
||||
|
@ -60,10 +60,11 @@ static unsigned long user_backtrace(struct perf_callchain_entry_ctx *entry,
|
||||
void perf_callchain_user(struct perf_callchain_entry_ctx *entry,
|
||||
struct pt_regs *regs)
|
||||
{
|
||||
struct perf_guest_info_callbacks *guest_cbs = perf_get_guest_cbs();
|
||||
unsigned long fp = 0;
|
||||
|
||||
/* RISC-V does not support perf in guest mode. */
|
||||
if (perf_guest_cbs && perf_guest_cbs->is_in_guest())
|
||||
if (guest_cbs && guest_cbs->is_in_guest())
|
||||
return;
|
||||
|
||||
fp = regs->s0;
|
||||
@ -84,8 +85,10 @@ void notrace walk_stackframe(struct task_struct *task,
|
||||
void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry,
|
||||
struct pt_regs *regs)
|
||||
{
|
||||
struct perf_guest_info_callbacks *guest_cbs = perf_get_guest_cbs();
|
||||
|
||||
/* RISC-V does not support perf in guest mode. */
|
||||
if (perf_guest_cbs && perf_guest_cbs->is_in_guest()) {
|
||||
if (guest_cbs && guest_cbs->is_in_guest()) {
|
||||
pr_warn("RISC-V does not support perf in guest mode!");
|
||||
return;
|
||||
}
|
||||
|
@ -1982,6 +1982,13 @@ int kvm_s390_is_stop_irq_pending(struct kvm_vcpu *vcpu)
|
||||
return test_bit(IRQ_PEND_SIGP_STOP, &li->pending_irqs);
|
||||
}
|
||||
|
||||
int kvm_s390_is_restart_irq_pending(struct kvm_vcpu *vcpu)
|
||||
{
|
||||
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
|
||||
|
||||
return test_bit(IRQ_PEND_RESTART, &li->pending_irqs);
|
||||
}
|
||||
|
||||
void kvm_s390_clear_stop_irq(struct kvm_vcpu *vcpu)
|
||||
{
|
||||
struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int;
|
||||
|
@ -4205,10 +4205,15 @@ void kvm_s390_vcpu_stop(struct kvm_vcpu *vcpu)
|
||||
spin_lock(&vcpu->kvm->arch.start_stop_lock);
|
||||
online_vcpus = atomic_read(&vcpu->kvm->online_vcpus);
|
||||
|
||||
/* SIGP STOP and SIGP STOP AND STORE STATUS has been fully processed */
|
||||
/*
|
||||
* Set the VCPU to STOPPED and THEN clear the interrupt flag,
|
||||
* now that the SIGP STOP and SIGP STOP AND STORE STATUS orders
|
||||
* have been fully processed. This will ensure that the VCPU
|
||||
* is kept BUSY if another VCPU is inquiring with SIGP SENSE.
|
||||
*/
|
||||
kvm_s390_set_cpuflags(vcpu, CPUSTAT_STOPPED);
|
||||
kvm_s390_clear_stop_irq(vcpu);
|
||||
|
||||
kvm_s390_set_cpuflags(vcpu, CPUSTAT_STOPPED);
|
||||
__disable_ibs_on_vcpu(vcpu);
|
||||
|
||||
for (i = 0; i < online_vcpus; i++) {
|
||||
|
@ -373,6 +373,7 @@ void kvm_s390_destroy_adapters(struct kvm *kvm);
|
||||
int kvm_s390_ext_call_pending(struct kvm_vcpu *vcpu);
|
||||
extern struct kvm_device_ops kvm_flic_ops;
|
||||
int kvm_s390_is_stop_irq_pending(struct kvm_vcpu *vcpu);
|
||||
int kvm_s390_is_restart_irq_pending(struct kvm_vcpu *vcpu);
|
||||
void kvm_s390_clear_stop_irq(struct kvm_vcpu *vcpu);
|
||||
int kvm_s390_set_irq_state(struct kvm_vcpu *vcpu,
|
||||
void __user *buf, int len);
|
||||
|
@ -288,6 +288,34 @@ static int handle_sigp_dst(struct kvm_vcpu *vcpu, u8 order_code,
|
||||
if (!dst_vcpu)
|
||||
return SIGP_CC_NOT_OPERATIONAL;
|
||||
|
||||
/*
|
||||
* SIGP RESTART, SIGP STOP, and SIGP STOP AND STORE STATUS orders
|
||||
* are processed asynchronously. Until the affected VCPU finishes
|
||||
* its work and calls back into KVM to clear the (RESTART or STOP)
|
||||
* interrupt, we need to return any new non-reset orders "busy".
|
||||
*
|
||||
* This is important because a single VCPU could issue:
|
||||
* 1) SIGP STOP $DESTINATION
|
||||
* 2) SIGP SENSE $DESTINATION
|
||||
*
|
||||
* If the SIGP SENSE would not be rejected as "busy", it could
|
||||
* return an incorrect answer as to whether the VCPU is STOPPED
|
||||
* or OPERATING.
|
||||
*/
|
||||
if (order_code != SIGP_INITIAL_CPU_RESET &&
|
||||
order_code != SIGP_CPU_RESET) {
|
||||
/*
|
||||
* Lockless check. Both SIGP STOP and SIGP (RE)START
|
||||
* properly synchronize everything while processing
|
||||
* their orders, while the guest cannot observe a
|
||||
* difference when issuing other orders from two
|
||||
* different VCPUs.
|
||||
*/
|
||||
if (kvm_s390_is_stop_irq_pending(dst_vcpu) ||
|
||||
kvm_s390_is_restart_irq_pending(dst_vcpu))
|
||||
return SIGP_CC_BUSY;
|
||||
}
|
||||
|
||||
switch (order_code) {
|
||||
case SIGP_SENSE:
|
||||
vcpu->stat.instruction_sigp_sense++;
|
||||
|
@ -2366,10 +2366,11 @@ static bool perf_hw_regs(struct pt_regs *regs)
|
||||
void
|
||||
perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs)
|
||||
{
|
||||
struct perf_guest_info_callbacks *guest_cbs = perf_get_guest_cbs();
|
||||
struct unwind_state state;
|
||||
unsigned long addr;
|
||||
|
||||
if (perf_guest_cbs && perf_guest_cbs->is_in_guest()) {
|
||||
if (guest_cbs && guest_cbs->is_in_guest()) {
|
||||
/* TODO: We don't support guest os callchain now */
|
||||
return;
|
||||
}
|
||||
@ -2475,10 +2476,11 @@ perf_callchain_user32(struct pt_regs *regs, struct perf_callchain_entry_ctx *ent
|
||||
void
|
||||
perf_callchain_user(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs)
|
||||
{
|
||||
struct perf_guest_info_callbacks *guest_cbs = perf_get_guest_cbs();
|
||||
struct stack_frame frame;
|
||||
const unsigned long __user *fp;
|
||||
|
||||
if (perf_guest_cbs && perf_guest_cbs->is_in_guest()) {
|
||||
if (guest_cbs && guest_cbs->is_in_guest()) {
|
||||
/* TODO: We don't support guest os callchain now */
|
||||
return;
|
||||
}
|
||||
@ -2562,18 +2564,21 @@ static unsigned long code_segment_base(struct pt_regs *regs)
|
||||
|
||||
unsigned long perf_instruction_pointer(struct pt_regs *regs)
|
||||
{
|
||||
if (perf_guest_cbs && perf_guest_cbs->is_in_guest())
|
||||
return perf_guest_cbs->get_guest_ip();
|
||||
struct perf_guest_info_callbacks *guest_cbs = perf_get_guest_cbs();
|
||||
|
||||
if (guest_cbs && guest_cbs->is_in_guest())
|
||||
return guest_cbs->get_guest_ip();
|
||||
|
||||
return regs->ip + code_segment_base(regs);
|
||||
}
|
||||
|
||||
unsigned long perf_misc_flags(struct pt_regs *regs)
|
||||
{
|
||||
struct perf_guest_info_callbacks *guest_cbs = perf_get_guest_cbs();
|
||||
int misc = 0;
|
||||
|
||||
if (perf_guest_cbs && perf_guest_cbs->is_in_guest()) {
|
||||
if (perf_guest_cbs->is_user_mode())
|
||||
if (guest_cbs && guest_cbs->is_in_guest()) {
|
||||
if (guest_cbs->is_user_mode())
|
||||
misc |= PERF_RECORD_MISC_GUEST_USER;
|
||||
else
|
||||
misc |= PERF_RECORD_MISC_GUEST_KERNEL;
|
||||
|
@ -2333,6 +2333,7 @@ static int handle_pmi_common(struct pt_regs *regs, u64 status)
|
||||
{
|
||||
struct perf_sample_data data;
|
||||
struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events);
|
||||
struct perf_guest_info_callbacks *guest_cbs;
|
||||
int bit;
|
||||
int handled = 0;
|
||||
|
||||
@ -2386,9 +2387,11 @@ static int handle_pmi_common(struct pt_regs *regs, u64 status)
|
||||
*/
|
||||
if (__test_and_clear_bit(55, (unsigned long *)&status)) {
|
||||
handled++;
|
||||
if (unlikely(perf_guest_cbs && perf_guest_cbs->is_in_guest() &&
|
||||
perf_guest_cbs->handle_intel_pt_intr))
|
||||
perf_guest_cbs->handle_intel_pt_intr();
|
||||
|
||||
guest_cbs = perf_get_guest_cbs();
|
||||
if (unlikely(guest_cbs && guest_cbs->is_in_guest() &&
|
||||
guest_cbs->handle_intel_pt_intr))
|
||||
guest_cbs->handle_intel_pt_intr();
|
||||
else
|
||||
intel_pt_interrupt();
|
||||
}
|
||||
|
@ -1218,7 +1218,7 @@ static const u32 msrs_to_save_all[] = {
|
||||
MSR_IA32_UMWAIT_CONTROL,
|
||||
|
||||
MSR_ARCH_PERFMON_FIXED_CTR0, MSR_ARCH_PERFMON_FIXED_CTR1,
|
||||
MSR_ARCH_PERFMON_FIXED_CTR0 + 2, MSR_ARCH_PERFMON_FIXED_CTR0 + 3,
|
||||
MSR_ARCH_PERFMON_FIXED_CTR0 + 2,
|
||||
MSR_CORE_PERF_FIXED_CTR_CTRL, MSR_CORE_PERF_GLOBAL_STATUS,
|
||||
MSR_CORE_PERF_GLOBAL_CTRL, MSR_CORE_PERF_GLOBAL_OVF_CTRL,
|
||||
MSR_ARCH_PERFMON_PERFCTR0, MSR_ARCH_PERFMON_PERFCTR1,
|
||||
|
@ -25,6 +25,7 @@
|
||||
#include <linux/sched.h>
|
||||
#include <linux/slab.h>
|
||||
#include <linux/kthread.h>
|
||||
#include <linux/fs_context.h>
|
||||
#include <uapi/linux/mount.h>
|
||||
#include "base.h"
|
||||
|
||||
@ -62,8 +63,15 @@ static struct dentry *public_dev_mount(struct file_system_type *fs_type, int fla
|
||||
const char *dev_name, void *data)
|
||||
{
|
||||
struct super_block *s = mnt->mnt_sb;
|
||||
int err;
|
||||
|
||||
atomic_inc(&s->s_active);
|
||||
down_write(&s->s_umount);
|
||||
err = reconfigure_single(s, flags, data);
|
||||
if (err < 0) {
|
||||
deactivate_locked_super(s);
|
||||
return ERR_PTR(err);
|
||||
}
|
||||
return dget(s->s_root);
|
||||
}
|
||||
|
||||
|
@ -385,9 +385,7 @@ static void fw_cfg_sysfs_cache_cleanup(void)
|
||||
struct fw_cfg_sysfs_entry *entry, *next;
|
||||
|
||||
list_for_each_entry_safe(entry, next, &fw_cfg_entry_cache, list) {
|
||||
/* will end up invoking fw_cfg_sysfs_cache_delist()
|
||||
* via each object's release() method (i.e. destructor)
|
||||
*/
|
||||
fw_cfg_sysfs_cache_delist(entry);
|
||||
kobject_put(&entry->kobj);
|
||||
}
|
||||
}
|
||||
@ -445,7 +443,6 @@ static void fw_cfg_sysfs_release_entry(struct kobject *kobj)
|
||||
{
|
||||
struct fw_cfg_sysfs_entry *entry = to_entry(kobj);
|
||||
|
||||
fw_cfg_sysfs_cache_delist(entry);
|
||||
kfree(entry);
|
||||
}
|
||||
|
||||
@ -598,20 +595,18 @@ static int fw_cfg_register_file(const struct fw_cfg_file *f)
|
||||
/* set file entry information */
|
||||
entry->size = be32_to_cpu(f->size);
|
||||
entry->select = be16_to_cpu(f->select);
|
||||
memcpy(entry->name, f->name, FW_CFG_MAX_FILE_PATH);
|
||||
strscpy(entry->name, f->name, FW_CFG_MAX_FILE_PATH);
|
||||
|
||||
/* register entry under "/sys/firmware/qemu_fw_cfg/by_key/" */
|
||||
err = kobject_init_and_add(&entry->kobj, &fw_cfg_sysfs_entry_ktype,
|
||||
fw_cfg_sel_ko, "%d", entry->select);
|
||||
if (err) {
|
||||
kobject_put(&entry->kobj);
|
||||
return err;
|
||||
}
|
||||
if (err)
|
||||
goto err_put_entry;
|
||||
|
||||
/* add raw binary content access */
|
||||
err = sysfs_create_bin_file(&entry->kobj, &fw_cfg_sysfs_attr_raw);
|
||||
if (err)
|
||||
goto err_add_raw;
|
||||
goto err_del_entry;
|
||||
|
||||
/* try adding "/sys/firmware/qemu_fw_cfg/by_name/" symlink */
|
||||
fw_cfg_build_symlink(fw_cfg_fname_kset, &entry->kobj, entry->name);
|
||||
@ -620,9 +615,10 @@ static int fw_cfg_register_file(const struct fw_cfg_file *f)
|
||||
fw_cfg_sysfs_cache_enlist(entry);
|
||||
return 0;
|
||||
|
||||
err_add_raw:
|
||||
err_del_entry:
|
||||
kobject_del(&entry->kobj);
|
||||
kfree(entry);
|
||||
err_put_entry:
|
||||
kobject_put(&entry->kobj);
|
||||
return err;
|
||||
}
|
||||
|
||||
|
@ -1915,6 +1915,10 @@ static int uvc_video_start_transfer(struct uvc_streaming *stream,
|
||||
if (ep == NULL)
|
||||
return -EIO;
|
||||
|
||||
/* Reject broken descriptors. */
|
||||
if (usb_endpoint_maxp(&ep->desc) == 0)
|
||||
return -EIO;
|
||||
|
||||
ret = uvc_init_video_bulk(stream, ep, gfp_flags);
|
||||
}
|
||||
|
||||
|
@ -55,12 +55,14 @@ choice
|
||||
LITTLE_ENDIAN_BYTE, if the bytes are reversed.
|
||||
|
||||
config MTD_CFI_NOSWAP
|
||||
depends on !ARCH_IXP4XX || CPU_BIG_ENDIAN
|
||||
bool "NO"
|
||||
|
||||
config MTD_CFI_BE_BYTE_SWAP
|
||||
bool "BIG_ENDIAN_BYTE"
|
||||
|
||||
config MTD_CFI_LE_BYTE_SWAP
|
||||
depends on !ARCH_IXP4XX
|
||||
bool "LITTLE_ENDIAN_BYTE"
|
||||
|
||||
endchoice
|
||||
|
@ -303,7 +303,7 @@ config MTD_DC21285
|
||||
|
||||
config MTD_IXP4XX
|
||||
tristate "CFI Flash device mapped on Intel IXP4xx based systems"
|
||||
depends on MTD_CFI && MTD_COMPLEX_MAPPINGS && ARCH_IXP4XX
|
||||
depends on MTD_CFI && MTD_COMPLEX_MAPPINGS && ARCH_IXP4XX && MTD_CFI_ADV_OPTIONS
|
||||
help
|
||||
This enables MTD access to flash devices on platforms based
|
||||
on Intel's IXP4xx family of network processors such as the
|
||||
|
@ -1000,6 +1000,7 @@ int rtl92cu_hw_init(struct ieee80211_hw *hw)
|
||||
_initpabias(hw);
|
||||
rtl92c_dm_init(hw);
|
||||
exit:
|
||||
local_irq_disable();
|
||||
local_irq_restore(flags);
|
||||
return err;
|
||||
}
|
||||
|
@ -585,7 +585,7 @@ static int legacy_parse_param(struct fs_context *fc, struct fs_parameter *param)
|
||||
param->key);
|
||||
}
|
||||
|
||||
if (len > PAGE_SIZE - 2 - size)
|
||||
if (size + len + 2 > PAGE_SIZE)
|
||||
return invalf(fc, "VFS: Legacy: Cumulative options too large");
|
||||
if (strchr(param->key, ',') ||
|
||||
(param->type == fs_value_is_string &&
|
||||
|
@ -179,7 +179,7 @@ orangefs_bufmap_free(struct orangefs_bufmap *bufmap)
|
||||
{
|
||||
kfree(bufmap->page_array);
|
||||
kfree(bufmap->desc_array);
|
||||
kfree(bufmap->buffer_index_array);
|
||||
bitmap_free(bufmap->buffer_index_array);
|
||||
kfree(bufmap);
|
||||
}
|
||||
|
||||
@ -229,8 +229,7 @@ orangefs_bufmap_alloc(struct ORANGEFS_dev_map_desc *user_desc)
|
||||
bufmap->desc_size = user_desc->size;
|
||||
bufmap->desc_shift = ilog2(bufmap->desc_size);
|
||||
|
||||
bufmap->buffer_index_array =
|
||||
kzalloc(DIV_ROUND_UP(bufmap->desc_count, BITS_PER_LONG), GFP_KERNEL);
|
||||
bufmap->buffer_index_array = bitmap_zalloc(bufmap->desc_count, GFP_KERNEL);
|
||||
if (!bufmap->buffer_index_array)
|
||||
goto out_free_bufmap;
|
||||
|
||||
@ -253,7 +252,7 @@ orangefs_bufmap_alloc(struct ORANGEFS_dev_map_desc *user_desc)
|
||||
out_free_desc_array:
|
||||
kfree(bufmap->desc_array);
|
||||
out_free_index_array:
|
||||
kfree(bufmap->buffer_index_array);
|
||||
bitmap_free(bufmap->buffer_index_array);
|
||||
out_free_bufmap:
|
||||
kfree(bufmap);
|
||||
out:
|
||||
|
@ -1470,8 +1470,8 @@ struct dentry *mount_nodev(struct file_system_type *fs_type,
|
||||
}
|
||||
EXPORT_SYMBOL(mount_nodev);
|
||||
|
||||
static int reconfigure_single(struct super_block *s,
|
||||
int flags, void *data)
|
||||
int reconfigure_single(struct super_block *s,
|
||||
int flags, void *data)
|
||||
{
|
||||
struct fs_context *fc;
|
||||
int ret;
|
||||
|
@ -135,6 +135,8 @@ extern int generic_parse_monolithic(struct fs_context *fc, void *data);
|
||||
extern int vfs_get_tree(struct fs_context *fc);
|
||||
extern void put_fs_context(struct fs_context *fc);
|
||||
extern void fc_drop_locked(struct fs_context *fc);
|
||||
int reconfigure_single(struct super_block *s,
|
||||
int flags, void *data);
|
||||
|
||||
/*
|
||||
* sget() wrappers to be called from the ->get_tree() op.
|
||||
|
@ -1179,7 +1179,18 @@ extern void perf_event_bpf_event(struct bpf_prog *prog,
|
||||
enum perf_bpf_event_type type,
|
||||
u16 flags);
|
||||
|
||||
extern struct perf_guest_info_callbacks *perf_guest_cbs;
|
||||
extern struct perf_guest_info_callbacks __rcu *perf_guest_cbs;
|
||||
static inline struct perf_guest_info_callbacks *perf_get_guest_cbs(void)
|
||||
{
|
||||
/*
|
||||
* Callbacks are RCU-protected and must be READ_ONCE to avoid reloading
|
||||
* the callbacks between a !NULL check and dereferences, to ensure
|
||||
* pending stores/changes to the callback pointers are visible before a
|
||||
* non-NULL perf_guest_cbs is visible to readers, and to prevent a
|
||||
* module from unloading callbacks while readers are active.
|
||||
*/
|
||||
return rcu_dereference(perf_guest_cbs);
|
||||
}
|
||||
extern int perf_register_guest_info_callbacks(struct perf_guest_info_callbacks *callbacks);
|
||||
extern int perf_unregister_guest_info_callbacks(struct perf_guest_info_callbacks *callbacks);
|
||||
|
||||
|
@ -6061,18 +6061,25 @@ static void perf_pending_event(struct irq_work *entry)
|
||||
* Later on, we might change it to a list if there is
|
||||
* another virtualization implementation supporting the callbacks.
|
||||
*/
|
||||
struct perf_guest_info_callbacks *perf_guest_cbs;
|
||||
struct perf_guest_info_callbacks __rcu *perf_guest_cbs;
|
||||
|
||||
int perf_register_guest_info_callbacks(struct perf_guest_info_callbacks *cbs)
|
||||
{
|
||||
perf_guest_cbs = cbs;
|
||||
if (WARN_ON_ONCE(rcu_access_pointer(perf_guest_cbs)))
|
||||
return -EBUSY;
|
||||
|
||||
rcu_assign_pointer(perf_guest_cbs, cbs);
|
||||
return 0;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(perf_register_guest_info_callbacks);
|
||||
|
||||
int perf_unregister_guest_info_callbacks(struct perf_guest_info_callbacks *cbs)
|
||||
{
|
||||
perf_guest_cbs = NULL;
|
||||
if (WARN_ON_ONCE(rcu_access_pointer(perf_guest_cbs) != cbs))
|
||||
return -EINVAL;
|
||||
|
||||
rcu_assign_pointer(perf_guest_cbs, NULL);
|
||||
synchronize_rcu();
|
||||
return 0;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(perf_unregister_guest_info_callbacks);
|
||||
|
@ -1926,6 +1926,7 @@ enum {
|
||||
ALC887_FIXUP_ASUS_BASS,
|
||||
ALC887_FIXUP_BASS_CHMAP,
|
||||
ALC1220_FIXUP_GB_DUAL_CODECS,
|
||||
ALC1220_FIXUP_GB_X570,
|
||||
ALC1220_FIXUP_CLEVO_P950,
|
||||
ALC1220_FIXUP_CLEVO_PB51ED,
|
||||
ALC1220_FIXUP_CLEVO_PB51ED_PINS,
|
||||
@ -2115,6 +2116,29 @@ static void alc1220_fixup_gb_dual_codecs(struct hda_codec *codec,
|
||||
}
|
||||
}
|
||||
|
||||
static void alc1220_fixup_gb_x570(struct hda_codec *codec,
|
||||
const struct hda_fixup *fix,
|
||||
int action)
|
||||
{
|
||||
static const hda_nid_t conn1[] = { 0x0c };
|
||||
static const struct coef_fw gb_x570_coefs[] = {
|
||||
WRITE_COEF(0x1a, 0x01c1),
|
||||
WRITE_COEF(0x1b, 0x0202),
|
||||
WRITE_COEF(0x43, 0x3005),
|
||||
{}
|
||||
};
|
||||
|
||||
switch (action) {
|
||||
case HDA_FIXUP_ACT_PRE_PROBE:
|
||||
snd_hda_override_conn_list(codec, 0x14, ARRAY_SIZE(conn1), conn1);
|
||||
snd_hda_override_conn_list(codec, 0x1b, ARRAY_SIZE(conn1), conn1);
|
||||
break;
|
||||
case HDA_FIXUP_ACT_INIT:
|
||||
alc_process_coef_fw(codec, gb_x570_coefs);
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
static void alc1220_fixup_clevo_p950(struct hda_codec *codec,
|
||||
const struct hda_fixup *fix,
|
||||
int action)
|
||||
@ -2417,6 +2441,10 @@ static const struct hda_fixup alc882_fixups[] = {
|
||||
.type = HDA_FIXUP_FUNC,
|
||||
.v.func = alc1220_fixup_gb_dual_codecs,
|
||||
},
|
||||
[ALC1220_FIXUP_GB_X570] = {
|
||||
.type = HDA_FIXUP_FUNC,
|
||||
.v.func = alc1220_fixup_gb_x570,
|
||||
},
|
||||
[ALC1220_FIXUP_CLEVO_P950] = {
|
||||
.type = HDA_FIXUP_FUNC,
|
||||
.v.func = alc1220_fixup_clevo_p950,
|
||||
@ -2519,7 +2547,7 @@ static const struct snd_pci_quirk alc882_fixup_tbl[] = {
|
||||
SND_PCI_QUIRK(0x13fe, 0x1009, "Advantech MIT-W101", ALC886_FIXUP_EAPD),
|
||||
SND_PCI_QUIRK(0x1458, 0xa002, "Gigabyte EP45-DS3/Z87X-UD3H", ALC889_FIXUP_FRONT_HP_NO_PRESENCE),
|
||||
SND_PCI_QUIRK(0x1458, 0xa0b8, "Gigabyte AZ370-Gaming", ALC1220_FIXUP_GB_DUAL_CODECS),
|
||||
SND_PCI_QUIRK(0x1458, 0xa0cd, "Gigabyte X570 Aorus Master", ALC1220_FIXUP_CLEVO_P950),
|
||||
SND_PCI_QUIRK(0x1458, 0xa0cd, "Gigabyte X570 Aorus Master", ALC1220_FIXUP_GB_X570),
|
||||
SND_PCI_QUIRK(0x1458, 0xa0ce, "Gigabyte X570 Aorus Xtreme", ALC1220_FIXUP_CLEVO_P950),
|
||||
SND_PCI_QUIRK(0x1462, 0x11f7, "MSI-GE63", ALC1220_FIXUP_CLEVO_P950),
|
||||
SND_PCI_QUIRK(0x1462, 0x1228, "MSI-GP63", ALC1220_FIXUP_CLEVO_P950),
|
||||
|
Loading…
Reference in New Issue
Block a user