[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAK9=C2WJv4PX5FAAPUcttK4UDDNWdyNWmLG_zXwx60WrMgXO8Q@mail.gmail.com>
Date: Wed, 13 Dec 2023 11:46:19 +0530
From: Anup Patel <apatel@...tanamicro.com>
To: Clément Léger <cleger@...osinc.com>
Cc: Paul Walmsley <paul.walmsley@...ive.com>,
Palmer Dabbelt <palmer@...belt.com>,
Albert Ou <aou@...s.berkeley.edu>,
Anup Patel <anup@...infault.org>,
Atish Patra <atishp@...shpatra.org>,
Andrew Jones <ajones@...tanamicro.com>,
linux-riscv@...ts.infradead.org, linux-kernel@...r.kernel.org,
kvm@...r.kernel.org, kvm-riscv@...ts.infradead.org
Subject: Re: [PATCH v2 4/5] riscv: kvm: Use SYM_*() assembly macros instead of
deprecated ones
On Tue, Oct 24, 2023 at 6:58 PM Clément Léger <cleger@...osinc.com> wrote:
>
> ENTRY()/END()/WEAK() macros are deprecated and we should make use of the
> new SYM_*() macros [1] for better annotation of symbols. Replace the
> deprecated ones with the new ones and fix wrong usage of END()/ENDPROC()
> to correctly describe the symbols.
>
> [1] https://docs.kernel.org/core-api/asm-annotations.html
>
> Signed-off-by: Clément Léger <cleger@...osinc.com>
> Reviewed-by: Andrew Jones <ajones@...tanamicro.com>
Queued this patch for Linux-6.8
Thanks,
Anup
> ---
> arch/riscv/kvm/vcpu_switch.S | 28 ++++++++++++----------------
> 1 file changed, 12 insertions(+), 16 deletions(-)
>
> diff --git a/arch/riscv/kvm/vcpu_switch.S b/arch/riscv/kvm/vcpu_switch.S
> index d74df8eb4d71..8b18473780ac 100644
> --- a/arch/riscv/kvm/vcpu_switch.S
> +++ b/arch/riscv/kvm/vcpu_switch.S
> @@ -15,7 +15,7 @@
> .altmacro
> .option norelax
>
> -ENTRY(__kvm_riscv_switch_to)
> +SYM_FUNC_START(__kvm_riscv_switch_to)
> /* Save Host GPRs (except A0 and T0-T6) */
> REG_S ra, (KVM_ARCH_HOST_RA)(a0)
> REG_S sp, (KVM_ARCH_HOST_SP)(a0)
> @@ -208,9 +208,9 @@ __kvm_switch_return:
>
> /* Return to C code */
> ret
> -ENDPROC(__kvm_riscv_switch_to)
> +SYM_FUNC_END(__kvm_riscv_switch_to)
>
> -ENTRY(__kvm_riscv_unpriv_trap)
> +SYM_CODE_START(__kvm_riscv_unpriv_trap)
> /*
> * We assume that faulting unpriv load/store instruction is
> * 4-byte long and blindly increment SEPC by 4.
> @@ -231,12 +231,10 @@ ENTRY(__kvm_riscv_unpriv_trap)
> csrr a1, CSR_HTINST
> REG_S a1, (KVM_ARCH_TRAP_HTINST)(a0)
> sret
> -ENDPROC(__kvm_riscv_unpriv_trap)
> +SYM_CODE_END(__kvm_riscv_unpriv_trap)
>
> #ifdef CONFIG_FPU
> - .align 3
> - .global __kvm_riscv_fp_f_save
> -__kvm_riscv_fp_f_save:
> +SYM_FUNC_START(__kvm_riscv_fp_f_save)
> csrr t2, CSR_SSTATUS
> li t1, SR_FS
> csrs CSR_SSTATUS, t1
> @@ -276,10 +274,9 @@ __kvm_riscv_fp_f_save:
> sw t0, KVM_ARCH_FP_F_FCSR(a0)
> csrw CSR_SSTATUS, t2
> ret
> +SYM_FUNC_END(__kvm_riscv_fp_f_save)
>
> - .align 3
> - .global __kvm_riscv_fp_d_save
> -__kvm_riscv_fp_d_save:
> +SYM_FUNC_START(__kvm_riscv_fp_d_save)
> csrr t2, CSR_SSTATUS
> li t1, SR_FS
> csrs CSR_SSTATUS, t1
> @@ -319,10 +316,9 @@ __kvm_riscv_fp_d_save:
> sw t0, KVM_ARCH_FP_D_FCSR(a0)
> csrw CSR_SSTATUS, t2
> ret
> +SYM_FUNC_END(__kvm_riscv_fp_d_save)
>
> - .align 3
> - .global __kvm_riscv_fp_f_restore
> -__kvm_riscv_fp_f_restore:
> +SYM_FUNC_START(__kvm_riscv_fp_f_restore)
> csrr t2, CSR_SSTATUS
> li t1, SR_FS
> lw t0, KVM_ARCH_FP_F_FCSR(a0)
> @@ -362,10 +358,9 @@ __kvm_riscv_fp_f_restore:
> fscsr t0
> csrw CSR_SSTATUS, t2
> ret
> +SYM_FUNC_END(__kvm_riscv_fp_f_restore)
>
> - .align 3
> - .global __kvm_riscv_fp_d_restore
> -__kvm_riscv_fp_d_restore:
> +SYM_FUNC_START(__kvm_riscv_fp_d_restore)
> csrr t2, CSR_SSTATUS
> li t1, SR_FS
> lw t0, KVM_ARCH_FP_D_FCSR(a0)
> @@ -405,4 +400,5 @@ __kvm_riscv_fp_d_restore:
> fscsr t0
> csrw CSR_SSTATUS, t2
> ret
> +SYM_FUNC_END(__kvm_riscv_fp_d_restore)
> #endif
> --
> 2.42.0
>
>
> --
> kvm-riscv mailing list
> kvm-riscv@...ts.infradead.org
> http://lists.infradead.org/mailman/listinfo/kvm-riscv
Powered by blists - more mailing lists