lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date: Tue, 7 May 2024 11:06:02 +0800
From: maobibo <maobibo@...ngson.cn>
To: Huacai Chen <chenhuacai@...nel.org>
Cc: Tianrui Zhao <zhaotianrui@...ngson.cn>, Juergen Gross <jgross@...e.com>,
 Paolo Bonzini <pbonzini@...hat.com>, Jonathan Corbet <corbet@....net>,
 loongarch@...ts.linux.dev, linux-kernel@...r.kernel.org,
 virtualization@...ts.linux.dev, kvm@...r.kernel.org
Subject: Re: [PATCH v8 4/6] LoongArch: KVM: Add vcpu search support from
 physical cpuid



On 2024/5/7 上午10:05, Huacai Chen wrote:
> On Tue, May 7, 2024 at 9:40 AM maobibo <maobibo@...ngson.cn> wrote:
>>
>>
>>
>> On 2024/5/6 下午10:17, Huacai Chen wrote:
>>> On Mon, May 6, 2024 at 6:05 PM maobibo <maobibo@...ngson.cn> wrote:
>>>>
>>>>
>>>>
>>>> On 2024/5/6 下午5:40, Huacai Chen wrote:
>>>>> On Mon, May 6, 2024 at 5:35 PM maobibo <maobibo@...ngson.cn> wrote:
>>>>>>
>>>>>>
>>>>>>
>>>>>> On 2024/5/6 下午4:59, Huacai Chen wrote:
>>>>>>> On Mon, May 6, 2024 at 4:18 PM maobibo <maobibo@...ngson.cn> wrote:
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On 2024/5/6 下午3:06, Huacai Chen wrote:
>>>>>>>>> Hi, Bibo,
>>>>>>>>>
>>>>>>>>> On Mon, May 6, 2024 at 2:36 PM maobibo <maobibo@...ngson.cn> wrote:
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On 2024/5/6 上午9:49, Huacai Chen wrote:
>>>>>>>>>>> Hi, Bibo,
>>>>>>>>>>>
>>>>>>>>>>> On Sun, Apr 28, 2024 at 6:05 PM Bibo Mao <maobibo@...ngson.cn> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>> Physical cpuid is used for interrupt routing for irqchips such as
>>>>>>>>>>>> ipi/msi/extioi interrupt controller. And physical cpuid is stored
>>>>>>>>>>>> at CSR register LOONGARCH_CSR_CPUID, it can not be changed once vcpu
>>>>>>>>>>>> is created and physical cpuid of two vcpus cannot be the same.
>>>>>>>>>>>>
>>>>>>>>>>>> Different irqchips have different size declaration about physical cpuid,
>>>>>>>>>>>> max cpuid value for CSR LOONGARCH_CSR_CPUID on 3A5000 is 512, max cpuid
>>>>>>>>>>>> supported by IPI hardware is 1024, 256 for extioi irqchip, and 65536
>>>>>>>>>>>> for MSI irqchip.
>>>>>>>>>>>>
>>>>>>>>>>>> The smallest value from all interrupt controllers is selected now,
>>>>>>>>>>>> and the max cpuid size is defines as 256 by KVM which comes from
>>>>>>>>>>>> extioi irqchip.
>>>>>>>>>>>>
>>>>>>>>>>>> Signed-off-by: Bibo Mao <maobibo@...ngson.cn>
>>>>>>>>>>>> ---
>>>>>>>>>>>>        arch/loongarch/include/asm/kvm_host.h | 26 ++++++++
>>>>>>>>>>>>        arch/loongarch/include/asm/kvm_vcpu.h |  1 +
>>>>>>>>>>>>        arch/loongarch/kvm/vcpu.c             | 93 ++++++++++++++++++++++++++-
>>>>>>>>>>>>        arch/loongarch/kvm/vm.c               | 11 ++++
>>>>>>>>>>>>        4 files changed, 130 insertions(+), 1 deletion(-)
>>>>>>>>>>>>
>>>>>>>>>>>> diff --git a/arch/loongarch/include/asm/kvm_host.h b/arch/loongarch/include/asm/kvm_host.h
>>>>>>>>>>>> index 2d62f7b0d377..3ba16ef1fe69 100644
>>>>>>>>>>>> --- a/arch/loongarch/include/asm/kvm_host.h
>>>>>>>>>>>> +++ b/arch/loongarch/include/asm/kvm_host.h
>>>>>>>>>>>> @@ -64,6 +64,30 @@ struct kvm_world_switch {
>>>>>>>>>>>>
>>>>>>>>>>>>        #define MAX_PGTABLE_LEVELS     4
>>>>>>>>>>>>
>>>>>>>>>>>> +/*
>>>>>>>>>>>> + * Physical cpu id is used for interrupt routing, there are different
>>>>>>>>>>>> + * definitions about physical cpuid on different hardwares.
>>>>>>>>>>>> + *  For LOONGARCH_CSR_CPUID register, max cpuid size if 512
>>>>>>>>>>>> + *  For IPI HW, max dest CPUID size 1024
>>>>>>>>>>>> + *  For extioi interrupt controller, max dest CPUID size is 256
>>>>>>>>>>>> + *  For MSI interrupt controller, max supported CPUID size is 65536
>>>>>>>>>>>> + *
>>>>>>>>>>>> + * Currently max CPUID is defined as 256 for KVM hypervisor, in future
>>>>>>>>>>>> + * it will be expanded to 4096, including 16 packages at most. And every
>>>>>>>>>>>> + * package supports at most 256 vcpus
>>>>>>>>>>>> + */
>>>>>>>>>>>> +#define KVM_MAX_PHYID          256
>>>>>>>>>>>> +
>>>>>>>>>>>> +struct kvm_phyid_info {
>>>>>>>>>>>> +       struct kvm_vcpu *vcpu;
>>>>>>>>>>>> +       bool            enabled;
>>>>>>>>>>>> +};
>>>>>>>>>>>> +
>>>>>>>>>>>> +struct kvm_phyid_map {
>>>>>>>>>>>> +       int max_phyid;
>>>>>>>>>>>> +       struct kvm_phyid_info phys_map[KVM_MAX_PHYID];
>>>>>>>>>>>> +};
>>>>>>>>>>>> +
>>>>>>>>>>>>        struct kvm_arch {
>>>>>>>>>>>>               /* Guest physical mm */
>>>>>>>>>>>>               kvm_pte_t *pgd;
>>>>>>>>>>>> @@ -71,6 +95,8 @@ struct kvm_arch {
>>>>>>>>>>>>               unsigned long invalid_ptes[MAX_PGTABLE_LEVELS];
>>>>>>>>>>>>               unsigned int  pte_shifts[MAX_PGTABLE_LEVELS];
>>>>>>>>>>>>               unsigned int  root_level;
>>>>>>>>>>>> +       spinlock_t    phyid_map_lock;
>>>>>>>>>>>> +       struct kvm_phyid_map  *phyid_map;
>>>>>>>>>>>>
>>>>>>>>>>>>               s64 time_offset;
>>>>>>>>>>>>               struct kvm_context __percpu *vmcs;
>>>>>>>>>>>> diff --git a/arch/loongarch/include/asm/kvm_vcpu.h b/arch/loongarch/include/asm/kvm_vcpu.h
>>>>>>>>>>>> index 0cb4fdb8a9b5..9f53950959da 100644
>>>>>>>>>>>> --- a/arch/loongarch/include/asm/kvm_vcpu.h
>>>>>>>>>>>> +++ b/arch/loongarch/include/asm/kvm_vcpu.h
>>>>>>>>>>>> @@ -81,6 +81,7 @@ void kvm_save_timer(struct kvm_vcpu *vcpu);
>>>>>>>>>>>>        void kvm_restore_timer(struct kvm_vcpu *vcpu);
>>>>>>>>>>>>
>>>>>>>>>>>>        int kvm_vcpu_ioctl_interrupt(struct kvm_vcpu *vcpu, struct kvm_interrupt *irq);
>>>>>>>>>>>> +struct kvm_vcpu *kvm_get_vcpu_by_cpuid(struct kvm *kvm, int cpuid);
>>>>>>>>>>>>
>>>>>>>>>>>>        /*
>>>>>>>>>>>>         * Loongarch KVM guest interrupt handling
>>>>>>>>>>>> diff --git a/arch/loongarch/kvm/vcpu.c b/arch/loongarch/kvm/vcpu.c
>>>>>>>>>>>> index 3a8779065f73..b633fd28b8db 100644
>>>>>>>>>>>> --- a/arch/loongarch/kvm/vcpu.c
>>>>>>>>>>>> +++ b/arch/loongarch/kvm/vcpu.c
>>>>>>>>>>>> @@ -274,6 +274,95 @@ static int _kvm_getcsr(struct kvm_vcpu *vcpu, unsigned int id, u64 *val)
>>>>>>>>>>>>               return 0;
>>>>>>>>>>>>        }
>>>>>>>>>>>>
>>>>>>>>>>>> +static inline int kvm_set_cpuid(struct kvm_vcpu *vcpu, u64 val)
>>>>>>>>>>>> +{
>>>>>>>>>>>> +       int cpuid;
>>>>>>>>>>>> +       struct loongarch_csrs *csr = vcpu->arch.csr;
>>>>>>>>>>>> +       struct kvm_phyid_map  *map;
>>>>>>>>>>>> +
>>>>>>>>>>>> +       if (val >= KVM_MAX_PHYID)
>>>>>>>>>>>> +               return -EINVAL;
>>>>>>>>>>>> +
>>>>>>>>>>>> +       cpuid = kvm_read_sw_gcsr(csr, LOONGARCH_CSR_ESTAT);
>>>>>>>>>>>> +       map = vcpu->kvm->arch.phyid_map;
>>>>>>>>>>>> +       spin_lock(&vcpu->kvm->arch.phyid_map_lock);
>>>>>>>>>>>> +       if (map->phys_map[cpuid].enabled) {
>>>>>>>>>>>> +               /*
>>>>>>>>>>>> +                * Cpuid is already set before
>>>>>>>>>>>> +                * Forbid changing different cpuid at runtime
>>>>>>>>>>>> +                */
>>>>>>>>>>>> +               if (cpuid != val) {
>>>>>>>>>>>> +                       /*
>>>>>>>>>>>> +                        * Cpuid 0 is initial value for vcpu, maybe invalid
>>>>>>>>>>>> +                        * unset value for vcpu
>>>>>>>>>>>> +                        */
>>>>>>>>>>>> +                       if (cpuid) {
>>>>>>>>>>>> +                               spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
>>>>>>>>>>>> +                               return -EINVAL;
>>>>>>>>>>>> +                       }
>>>>>>>>>>>> +               } else {
>>>>>>>>>>>> +                        /* Discard duplicated cpuid set */
>>>>>>>>>>>> +                       spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
>>>>>>>>>>>> +                       return 0;
>>>>>>>>>>>> +               }
>>>>>>>>>>>> +       }
>>>>>>>>>>> I have changed the logic and comments when I apply, you can double
>>>>>>>>>>> check whether it is correct.
>>>>>>>>>> I checkout the latest version, the modification in function
>>>>>>>>>> kvm_set_cpuid() is good for me.
>>>>>>>>> Now the modified version is like this:
>>>>>>>>>
>>>>>>>>> + if (map->phys_map[cpuid].enabled) {
>>>>>>>>> + /* Discard duplicated CPUID set operation */
>>>>>>>>> + if (cpuid == val) {
>>>>>>>>> + spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
>>>>>>>>> + return 0;
>>>>>>>>> + }
>>>>>>>>> +
>>>>>>>>> + /*
>>>>>>>>> + * CPUID is already set before
>>>>>>>>> + * Forbid changing different CPUID at runtime
>>>>>>>>> + * But CPUID 0 is the initial value for vcpu, so allow
>>>>>>>>> + * changing from 0 to others
>>>>>>>>> + */
>>>>>>>>> + if (cpuid) {
>>>>>>>>> + spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
>>>>>>>>> + return -EINVAL;
>>>>>>>>> + }
>>>>>>>>> + }
>>>>>>>>> But I still doubt whether we should allow changing from 0 to others
>>>>>>>>> while map->phys_map[cpuid].enabled is 1.
>>>>>>>> It is necessary since the default sw cpuid is zero :-( And we can
>>>>>>>> optimize it in later, such as set INVALID cpuid in function
>>>>>>>> kvm_arch_vcpu_create() and logic will be simple in function kvm_set_cpuid().
>>>>>>> In my opinion, if a vcpu with a uninitialized default physid=0, then
>>>>>>> map->phys_map[cpuid].enabled should be 0, then code won't come here.
>>>>>>> And if a vcpu with a real physid=0, then map->phys_map[cpuid].enabled
>>>>>>> is 1, but we shouldn't allow it to change physid in this case.
>>>>>> yes, that is actually a problem.
>>>>>>
>>>>>> vcpu0 firstly set physid=0, and vcpu0 set physid=1 again is not allowed.
>>>>>> vcpu0 firstly set physid=0, and vcpu1 set physid=1 is allowed.
>>>>>
>>>>> So can we simply drop the if (cpuid) checking? That means:
>>>>> + if (map->phys_map[cpuid].enabled) {
>>>>> + /* Discard duplicated CPUID set operation */
>>>>> + if (cpuid == val) {
>>>>> + spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
>>>>> + return 0;
>>>>> + }
>>>>> +
>>>>> + spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
>>>>> + return -EINVAL;
>>>>> + }
>>>> yes, the similar modification such as following, since the secondary
>>>> scenario should be allowed.
>>>>     "vcpu0 firstly set physid=0, and vcpu1 set physid=1 is allowed though
>>>> default sw cpuid is zero"
>>>>
>>>> --- a/arch/loongarch/kvm/vcpu.c
>>>> +++ b/arch/loongarch/kvm/vcpu.c
>>>> @@ -272,7 +272,7 @@ static inline int kvm_set_cpuid(struct kvm_vcpu
>>>> *vcpu, u64 val)
>>>>            cpuid = kvm_read_sw_gcsr(csr, LOONGARCH_CSR_CPUID);
>>>>
>>>>            spin_lock(&vcpu->kvm->arch.phyid_map_lock);
>>>> -       if (map->phys_map[cpuid].enabled) {
>>>> +       if ((cpuid != KVM_MAX_PHYID) && map->phys_map[cpuid].enabled) {
>>>>                    /* Discard duplicated CPUID set operation */
>>>>                    if (cpuid == val) {
>>>>                            spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
>>>> @@ -282,13 +282,9 @@ static inline int kvm_set_cpuid(struct kvm_vcpu
>>>> *vcpu, u64 val)
>>>>                    /*
>>>>                     * CPUID is already set before
>>>>                     * Forbid changing different CPUID at runtime
>>>> -                * But CPUID 0 is the initial value for vcpu, so allow
>>>> -                * changing from 0 to others
>>>>                     */
>>>> -               if (cpuid) {
>>>> -                       spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
>>>> -                       return -EINVAL;
>>>> -               }
>>>> +               spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
>>>> +               return -EINVAL;
>>>>            }
>>>>
>>>>            if (map->phys_map[val].enabled) {
>>>> @@ -1029,6 +1025,7 @@ int kvm_arch_vcpu_create(struct kvm_vcpu *vcpu)
>>>>
>>>>            /* Set cpuid */
>>>>            kvm_write_sw_gcsr(csr, LOONGARCH_CSR_TMID, vcpu->vcpu_id);
>>>> +       kvm_write_sw_gcsr(csr, LOONGARCH_CSR_CPUID, KVM_MAX_PHYID);
>>>>
>>>>            /* Start with no pending virtual guest interrupts */
>>>>            csr->csrs[LOONGARCH_CSR_GINTC] = 0;
>>> Very nice, but I think kvm_drop_cpuid() should also set to KVM_MAX_PHYID.
>>> Now I update my loongarch-kvm branch, you can test it again, and hope
>>> it is in the perfect status.
>> I sync and test the latest code from loongarch-kvm, pv ipi works well
>> with 256 vcpus. And the code looks good to me, thanks for your review in
>> short time.
> OK, if SWDBG also works well, I will send PR to Paolo tomorrow.
yes, sw debug works well with patch from qemu. And I will refresh patch 
to qemu after it is merged.

https://lore.kernel.org/all/20240218070025.218680-1-maobibo@loongson.cn/

--- a/configs/targets/loongarch64-softmmu.mak
+++ b/configs/targets/loongarch64-softmmu.mak
@@ -1,5 +1,6 @@
  TARGET_ARCH=loongarch64
  TARGET_BASE_ARCH=loongarch
  TARGET_SUPPORTS_MTTCG=y
+TARGET_KVM_HAVE_GUEST_DEBUG=y
  TARGET_XML_FILES= gdb-xml/loongarch-base32.xml 
gdb-xml/loongarch-base64.xml gdb-xml/loongarch-fpu.xml
  TARGET_NEED_FDT=y

Regards
Bibo Mao
> 
> Huacai
> 
>>
>> Regards
>> Bibo Mao
>>>
>>> Huacai
>>>>
>>>>
>>>>>
>>>>> Huacai
>>>>>
>>>>>>
>>>>>>
>>>>>>>
>>>>>>> Huacai
>>>>>>>
>>>>>>>>
>>>>>>>> Regards
>>>>>>>> Bibo Mao
>>>>>>>>
>>>>>>>>>
>>>>>>>>> Huacai
>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>> +
>>>>>>>>>>>> +       if (map->phys_map[val].enabled) {
>>>>>>>>>>>> +               /*
>>>>>>>>>>>> +                * New cpuid is already set with other vcpu
>>>>>>>>>>>> +                * Forbid sharing the same cpuid between different vcpus
>>>>>>>>>>>> +                */
>>>>>>>>>>>> +               if (map->phys_map[val].vcpu != vcpu) {
>>>>>>>>>>>> +                       spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
>>>>>>>>>>>> +                       return -EINVAL;
>>>>>>>>>>>> +               }
>>>>>>>>>>>> +
>>>>>>>>>>>> +               /* Discard duplicated cpuid set operation*/
>>>>>>>>>>>> +               spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
>>>>>>>>>>>> +               return 0;
>>>>>>>>>>>> +       }
>>>>>>>>>>>> +
>>>>>>>>>>>> +       kvm_write_sw_gcsr(csr, LOONGARCH_CSR_CPUID, val);
>>>>>>>>>>>> +       map->phys_map[val].enabled      = true;
>>>>>>>>>>>> +       map->phys_map[val].vcpu         = vcpu;
>>>>>>>>>>>> +       if (map->max_phyid < val)
>>>>>>>>>>>> +               map->max_phyid = val;
>>>>>>>>>>>> +       spin_unlock(&vcpu->kvm->arch.phyid_map_lock);
>>>>>>>>>>>> +       return 0;
>>>>>>>>>>>> +}
>>>>>>>>>>>> +
>>>>>>>>>>>> +struct kvm_vcpu *kvm_get_vcpu_by_cpuid(struct kvm *kvm, int cpuid)
>>>>>>>>>>>> +{
>>>>>>>>>>>> +       struct kvm_phyid_map  *map;
>>>>>>>>>>>> +
>>>>>>>>>>>> +       if (cpuid >= KVM_MAX_PHYID)
>>>>>>>>>>>> +               return NULL;
>>>>>>>>>>>> +
>>>>>>>>>>>> +       map = kvm->arch.phyid_map;
>>>>>>>>>>>> +       if (map->phys_map[cpuid].enabled)
>>>>>>>>>>>> +               return map->phys_map[cpuid].vcpu;
>>>>>>>>>>>> +
>>>>>>>>>>>> +       return NULL;
>>>>>>>>>>>> +}
>>>>>>>>>>>> +
>>>>>>>>>>>> +static inline void kvm_drop_cpuid(struct kvm_vcpu *vcpu)
>>>>>>>>>>>> +{
>>>>>>>>>>>> +       int cpuid;
>>>>>>>>>>>> +       struct loongarch_csrs *csr = vcpu->arch.csr;
>>>>>>>>>>>> +       struct kvm_phyid_map  *map;
>>>>>>>>>>>> +
>>>>>>>>>>>> +       map = vcpu->kvm->arch.phyid_map;
>>>>>>>>>>>> +       cpuid = kvm_read_sw_gcsr(csr, LOONGARCH_CSR_ESTAT);
>>>>>>>>>>>> +       if (cpuid >= KVM_MAX_PHYID)
>>>>>>>>>>>> +               return;
>>>>>>>>>>>> +
>>>>>>>>>>>> +       if (map->phys_map[cpuid].enabled) {
>>>>>>>>>>>> +               map->phys_map[cpuid].vcpu = NULL;
>>>>>>>>>>>> +               map->phys_map[cpuid].enabled = false;
>>>>>>>>>>>> +               kvm_write_sw_gcsr(csr, LOONGARCH_CSR_CPUID, 0);
>>>>>>>>>>>> +       }
>>>>>>>>>>>> +}
>>>>>>>>>>> While kvm_set_cpuid() is protected by a spinlock, do kvm_drop_cpuid()
>>>>>>>>>>> and kvm_get_vcpu_by_cpuid() also need it?
>>>>>>>>>>>
>>>>>>>>>> It is good to me that spinlock is added in function kvm_drop_cpuid().
>>>>>>>>>> And thinks for the efforts.
>>>>>>>>>>
>>>>>>>>>> Regards
>>>>>>>>>> Bibo Mao
>>>>>>>>>>>> +
>>>>>>>>>>>>        static int _kvm_setcsr(struct kvm_vcpu *vcpu, unsigned int id, u64 val)
>>>>>>>>>>>>        {
>>>>>>>>>>>>               int ret = 0, gintc;
>>>>>>>>>>>> @@ -291,7 +380,8 @@ static int _kvm_setcsr(struct kvm_vcpu *vcpu, unsigned int id, u64 val)
>>>>>>>>>>>>                       kvm_set_sw_gcsr(csr, LOONGARCH_CSR_ESTAT, gintc);
>>>>>>>>>>>>
>>>>>>>>>>>>                       return ret;
>>>>>>>>>>>> -       }
>>>>>>>>>>>> +       } else if (id == LOONGARCH_CSR_CPUID)
>>>>>>>>>>>> +               return kvm_set_cpuid(vcpu, val);
>>>>>>>>>>>>
>>>>>>>>>>>>               kvm_write_sw_gcsr(csr, id, val);
>>>>>>>>>>>>
>>>>>>>>>>>> @@ -943,6 +1033,7 @@ void kvm_arch_vcpu_destroy(struct kvm_vcpu *vcpu)
>>>>>>>>>>>>               hrtimer_cancel(&vcpu->arch.swtimer);
>>>>>>>>>>>>               kvm_mmu_free_memory_cache(&vcpu->arch.mmu_page_cache);
>>>>>>>>>>>>               kfree(vcpu->arch.csr);
>>>>>>>>>>>> +       kvm_drop_cpuid(vcpu);
>>>>>>>>>>> I think this line should be before the above kfree(), otherwise you
>>>>>>>>>>> get a "use after free".
>>>>>>>>>>>
>>>>>>>>>>> Huacai
>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>               /*
>>>>>>>>>>>>                * If the vCPU is freed and reused as another vCPU, we don't want the
>>>>>>>>>>>> diff --git a/arch/loongarch/kvm/vm.c b/arch/loongarch/kvm/vm.c
>>>>>>>>>>>> index 0a37f6fa8f2d..6006a28653ad 100644
>>>>>>>>>>>> --- a/arch/loongarch/kvm/vm.c
>>>>>>>>>>>> +++ b/arch/loongarch/kvm/vm.c
>>>>>>>>>>>> @@ -30,6 +30,14 @@ int kvm_arch_init_vm(struct kvm *kvm, unsigned long type)
>>>>>>>>>>>>               if (!kvm->arch.pgd)
>>>>>>>>>>>>                       return -ENOMEM;
>>>>>>>>>>>>
>>>>>>>>>>>> +       kvm->arch.phyid_map = kvzalloc(sizeof(struct kvm_phyid_map),
>>>>>>>>>>>> +                               GFP_KERNEL_ACCOUNT);
>>>>>>>>>>>> +       if (!kvm->arch.phyid_map) {
>>>>>>>>>>>> +               free_page((unsigned long)kvm->arch.pgd);
>>>>>>>>>>>> +               kvm->arch.pgd = NULL;
>>>>>>>>>>>> +               return -ENOMEM;
>>>>>>>>>>>> +       }
>>>>>>>>>>>> +
>>>>>>>>>>>>               kvm_init_vmcs(kvm);
>>>>>>>>>>>>               kvm->arch.gpa_size = BIT(cpu_vabits - 1);
>>>>>>>>>>>>               kvm->arch.root_level = CONFIG_PGTABLE_LEVELS - 1;
>>>>>>>>>>>> @@ -44,6 +52,7 @@ int kvm_arch_init_vm(struct kvm *kvm, unsigned long type)
>>>>>>>>>>>>               for (i = 0; i <= kvm->arch.root_level; i++)
>>>>>>>>>>>>                       kvm->arch.pte_shifts[i] = PAGE_SHIFT + i * (PAGE_SHIFT - 3);
>>>>>>>>>>>>
>>>>>>>>>>>> +       spin_lock_init(&kvm->arch.phyid_map_lock);
>>>>>>>>>>>>               return 0;
>>>>>>>>>>>>        }
>>>>>>>>>>>>
>>>>>>>>>>>> @@ -51,7 +60,9 @@ void kvm_arch_destroy_vm(struct kvm *kvm)
>>>>>>>>>>>>        {
>>>>>>>>>>>>               kvm_destroy_vcpus(kvm);
>>>>>>>>>>>>               free_page((unsigned long)kvm->arch.pgd);
>>>>>>>>>>>> +       kvfree(kvm->arch.phyid_map);
>>>>>>>>>>>>               kvm->arch.pgd = NULL;
>>>>>>>>>>>> +       kvm->arch.phyid_map = NULL;
>>>>>>>>>>>>        }
>>>>>>>>>>>>
>>>>>>>>>>>>        int kvm_vm_ioctl_check_extension(struct kvm *kvm, long ext)
>>>>>>>>>>>> --
>>>>>>>>>>>> 2.39.3
>>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>
>>>>>>
>>>>
>>
>>


Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ