[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <205522b8-a0f2-9baf-7365-782be0668d74@redhat.com>
Date: Wed, 11 Mar 2020 10:43:33 +0100
From: Paolo Bonzini <pbonzini@...hat.com>
To: Joe Perches <joe@...ches.com>,
Sean Christopherson <sean.j.christopherson@...el.com>,
Vitaly Kuznetsov <vkuznets@...hat.com>,
Wanpeng Li <wanpengli@...cent.com>,
Jim Mattson <jmattson@...gle.com>,
Joerg Roedel <joro@...tes.org>
Cc: Thomas Gleixner <tglx@...utronix.de>,
Borislav Petkov <bp@...en8.de>,
"H. Peter Anvin" <hpa@...or.com>, x86@...nel.org,
kvm@...r.kernel.org, linux-kernel@...r.kernel.org
Subject: Re: [PATCH -next 021/491] KERNEL VIRTUAL MACHINE FOR X86 (KVM/x86):
Use fallthrough;
On 11/03/20 05:51, Joe Perches wrote:
> Convert the various uses of fallthrough comments to fallthrough;
>
> Done via script
> Link: https://lore.kernel.org/lkml/b56602fcf79f849e733e7b521bb0e17895d390fa.1582230379.git.joe.com/
>
> Signed-off-by: Joe Perches <joe@...ches.com>
> ---
> arch/x86/kvm/emulate.c | 2 +-
> arch/x86/kvm/hyperv.c | 2 +-
> arch/x86/kvm/irq_comm.c | 2 +-
> arch/x86/kvm/lapic.c | 6 +++---
> arch/x86/kvm/mmu/mmu.c | 2 +-
> arch/x86/kvm/svm.c | 2 +-
> arch/x86/kvm/vmx/vmx.c | 15 +++++++--------
> arch/x86/kvm/x86.c | 12 ++++--------
> 8 files changed, 19 insertions(+), 24 deletions(-)
>
> diff --git a/arch/x86/kvm/emulate.c b/arch/x86/kvm/emulate.c
> index bc00642..bae4d8 100644
> --- a/arch/x86/kvm/emulate.c
> +++ b/arch/x86/kvm/emulate.c
> @@ -3025,7 +3025,7 @@ static void string_registers_quirk(struct x86_emulate_ctxt *ctxt)
> case 0xa4: /* movsb */
> case 0xa5: /* movsd/w */
> *reg_rmw(ctxt, VCPU_REGS_RSI) &= (u32)-1;
> - /* fall through */
> + fallthrough;
> case 0xaa: /* stosb */
> case 0xab: /* stosd/w */
> *reg_rmw(ctxt, VCPU_REGS_RDI) &= (u32)-1;
> diff --git a/arch/x86/kvm/hyperv.c b/arch/x86/kvm/hyperv.c
> index a86fda7..934bfb4 100644
> --- a/arch/x86/kvm/hyperv.c
> +++ b/arch/x86/kvm/hyperv.c
> @@ -1652,7 +1652,7 @@ int kvm_hv_hypercall(struct kvm_vcpu *vcpu)
> ret = kvm_hvcall_signal_event(vcpu, fast, ingpa);
> if (ret != HV_STATUS_INVALID_PORT_ID)
> break;
> - /* fall through - maybe userspace knows this conn_id. */
> + fallthrough; /* maybe userspace knows this conn_id */
> case HVCALL_POST_MESSAGE:
> /* don't bother userspace if it has no way to handle it */
> if (unlikely(rep || !vcpu_to_synic(vcpu)->active)) {
> diff --git a/arch/x86/kvm/irq_comm.c b/arch/x86/kvm/irq_comm.c
> index c47d2a..4aa1c2e 100644
> --- a/arch/x86/kvm/irq_comm.c
> +++ b/arch/x86/kvm/irq_comm.c
> @@ -285,7 +285,7 @@ int kvm_set_routing_entry(struct kvm *kvm,
> switch (ue->u.irqchip.irqchip) {
> case KVM_IRQCHIP_PIC_SLAVE:
> e->irqchip.pin += PIC_NUM_PINS / 2;
> - /* fall through */
> + fallthrough;
> case KVM_IRQCHIP_PIC_MASTER:
> if (ue->u.irqchip.pin >= PIC_NUM_PINS / 2)
> return -EINVAL;
> diff --git a/arch/x86/kvm/lapic.c b/arch/x86/kvm/lapic.c
> index e3099c..64b7a9c 100644
> --- a/arch/x86/kvm/lapic.c
> +++ b/arch/x86/kvm/lapic.c
> @@ -1023,7 +1023,7 @@ static int __apic_accept_irq(struct kvm_lapic *apic, int delivery_mode,
> switch (delivery_mode) {
> case APIC_DM_LOWEST:
> vcpu->arch.apic_arb_prio++;
> - /* fall through */
> + fallthrough;
> case APIC_DM_FIXED:
> if (unlikely(trig_mode && !level))
> break;
> @@ -1311,7 +1311,7 @@ static u32 __apic_read(struct kvm_lapic *apic, unsigned int offset)
> break;
> case APIC_TASKPRI:
> report_tpr_access(apic, false);
> - /* fall thru */
> + fallthrough;
> default:
> val = kvm_lapic_get_reg(apic, offset);
> break;
> @@ -1952,7 +1952,7 @@ int kvm_lapic_reg_write(struct kvm_lapic *apic, u32 reg, u32 val)
>
> case APIC_LVT0:
> apic_manage_nmi_watchdog(apic, val);
> - /* fall through */
> + fallthrough;
> case APIC_LVTTHMR:
> case APIC_LVTPC:
> case APIC_LVT1:
> diff --git a/arch/x86/kvm/mmu/mmu.c b/arch/x86/kvm/mmu/mmu.c
> index 87e9ba..8593cd 100644
> --- a/arch/x86/kvm/mmu/mmu.c
> +++ b/arch/x86/kvm/mmu/mmu.c
> @@ -4475,7 +4475,7 @@ __reset_rsvds_bits_mask(struct kvm_vcpu *vcpu,
> rsvd_bits(maxphyaddr, 51);
> rsvd_check->rsvd_bits_mask[1][4] =
> rsvd_check->rsvd_bits_mask[0][4];
> - /* fall through */
> + fallthrough;
> case PT64_ROOT_4LEVEL:
> rsvd_check->rsvd_bits_mask[0][3] = exb_bit_rsvd |
> nonleaf_bit8_rsvd | rsvd_bits(7, 7) |
> diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c
> index 910005..73fa903 100644
> --- a/arch/x86/kvm/svm.c
> +++ b/arch/x86/kvm/svm.c
> @@ -4449,7 +4449,7 @@ static int svm_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr)
> case MSR_IA32_APICBASE:
> if (kvm_vcpu_apicv_active(vcpu))
> avic_update_vapic_bar(to_svm(vcpu), data);
> - /* Fall through */
> + fallthrough;
> default:
> return kvm_set_msr_common(vcpu, msr);
> }
> diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c
> index 69c5bd..1577cd 100644
> --- a/arch/x86/kvm/vmx/vmx.c
> +++ b/arch/x86/kvm/vmx/vmx.c
> @@ -4538,12 +4538,12 @@ static bool rmode_exception(struct kvm_vcpu *vcpu, int vec)
> vmcs_read32(VM_EXIT_INSTRUCTION_LEN);
> if (vcpu->guest_debug & KVM_GUESTDBG_USE_SW_BP)
> return false;
> - /* fall through */
> + fallthrough;
> case DB_VECTOR:
> if (vcpu->guest_debug &
> (KVM_GUESTDBG_SINGLESTEP | KVM_GUESTDBG_USE_HW_BP))
> return false;
> - /* fall through */
> + fallthrough;
> case DE_VECTOR:
> case OF_VECTOR:
> case BR_VECTOR:
> @@ -4692,7 +4692,7 @@ static int handle_exception_nmi(struct kvm_vcpu *vcpu)
> }
> kvm_run->debug.arch.dr6 = dr6 | DR6_FIXED_1;
> kvm_run->debug.arch.dr7 = vmcs_readl(GUEST_DR7);
> - /* fall through */
> + fallthrough;
> case BP_VECTOR:
> /*
> * Update instruction length as we may reinject #BP from
> @@ -5119,7 +5119,7 @@ static int handle_task_switch(struct kvm_vcpu *vcpu)
> error_code =
> vmcs_read32(IDT_VECTORING_ERROR_CODE);
> }
> - /* fall through */
> + fallthrough;
> case INTR_TYPE_SOFT_EXCEPTION:
> kvm_clear_exception_queue(vcpu);
> break;
> @@ -5469,8 +5469,7 @@ static int handle_invpcid(struct kvm_vcpu *vcpu)
> * global flush. If needed, we could optimize this later by
> * keeping track of global entries in shadow page tables.
> */
> -
> - /* fall-through */
> + fallthrough;
> case INVPCID_TYPE_ALL_INCL_GLOBAL:
> kvm_mmu_unload(vcpu);
> return kvm_skip_emulated_instruction(vcpu);
> @@ -6401,7 +6400,7 @@ static void __vmx_complete_interrupts(struct kvm_vcpu *vcpu,
> break;
> case INTR_TYPE_SOFT_EXCEPTION:
> vcpu->arch.event_exit_inst_len = vmcs_read32(instr_len_field);
> - /* fall through */
> + fallthrough;
> case INTR_TYPE_HARD_EXCEPTION:
> if (idt_vectoring_info & VECTORING_INFO_DELIVER_CODE_MASK) {
> u32 err = vmcs_read32(error_code_field);
> @@ -6411,7 +6410,7 @@ static void __vmx_complete_interrupts(struct kvm_vcpu *vcpu,
> break;
> case INTR_TYPE_SOFT_INTR:
> vcpu->arch.event_exit_inst_len = vmcs_read32(instr_len_field);
> - /* fall through */
> + fallthrough;
> case INTR_TYPE_EXT_INTR:
> kvm_queue_interrupt(vcpu, vector, type == INTR_TYPE_SOFT_INTR);
> break;
> diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
> index 2bbc0e0..e6280e 100644
> --- a/arch/x86/kvm/x86.c
> +++ b/arch/x86/kvm/x86.c
> @@ -1071,7 +1071,6 @@ static int __kvm_set_dr(struct kvm_vcpu *vcpu, int dr, unsigned long val)
> vcpu->arch.eff_db[dr] = val;
> break;
> case 4:
> - /* fall through */
> case 6:
> if (val & 0xffffffff00000000ULL)
> return -1; /* #GP */
> @@ -1079,7 +1078,6 @@ static int __kvm_set_dr(struct kvm_vcpu *vcpu, int dr, unsigned long val)
> kvm_update_dr6(vcpu);
> break;
> case 5:
> - /* fall through */
> default: /* 7 */
> if (!kvm_dr7_valid(val))
> return -1; /* #GP */
> @@ -1110,7 +1108,6 @@ int kvm_get_dr(struct kvm_vcpu *vcpu, int dr, unsigned long *val)
> *val = vcpu->arch.db[array_index_nospec(dr, size)];
> break;
> case 4:
> - /* fall through */
> case 6:
> if (vcpu->guest_debug & KVM_GUESTDBG_USE_HW_BP)
> *val = vcpu->arch.dr6;
> @@ -1118,7 +1115,6 @@ int kvm_get_dr(struct kvm_vcpu *vcpu, int dr, unsigned long *val)
> *val = kvm_x86_ops->get_dr6(vcpu);
> break;
> case 5:
> - /* fall through */
> default: /* 7 */
> *val = vcpu->arch.dr7;
> break;
> @@ -2885,7 +2881,8 @@ int kvm_set_msr_common(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
>
> case MSR_K7_PERFCTR0 ... MSR_K7_PERFCTR3:
> case MSR_P6_PERFCTR0 ... MSR_P6_PERFCTR1:
> - pr = true; /* fall through */
> + pr = true;
> + fallthrough;
> case MSR_K7_EVNTSEL0 ... MSR_K7_EVNTSEL3:
> case MSR_P6_EVNTSEL0 ... MSR_P6_EVNTSEL1:
> if (kvm_pmu_is_valid_msr(vcpu, msr))
> @@ -4181,8 +4178,7 @@ static int kvm_vcpu_ioctl_enable_cap(struct kvm_vcpu *vcpu,
> case KVM_CAP_HYPERV_SYNIC2:
> if (cap->args[0])
> return -EINVAL;
> - /* fall through */
> -
> + fallthrough;
> case KVM_CAP_HYPERV_SYNIC:
> if (!irqchip_in_kernel(vcpu->kvm))
> return -EINVAL;
> @@ -8478,7 +8474,7 @@ static inline int vcpu_block(struct kvm *kvm, struct kvm_vcpu *vcpu)
> vcpu->arch.pv.pv_unhalted = false;
> vcpu->arch.mp_state =
> KVM_MP_STATE_RUNNABLE;
> - /* fall through */
> + fallthrough;
> case KVM_MP_STATE_RUNNABLE:
> vcpu->arch.apf.halted = false;
> break;
>
Acked-by: Paolo Bonzini <pbonzini@...hat.com>
Powered by blists - more mailing lists