lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [day] [month] [year] [list]
Message-ID: <45da8ea8.a92.181105d91bc.Coremail.chenxuebing@jari.cn>
Date:   Sun, 29 May 2022 23:09:36 +0800 (GMT+08:00)
From:   "XueBing Chen" <chenxuebing@...i.cn>
To:     mpe@...erman.id.au, benh@...nel.crashing.org, paulus@...ba.org,
        npiggin@...il.com, farosas@...ux.ibm.com, bharata@...ux.ibm.com,
        aneesh.kumar@...ux.ibm.com
Cc:     aik@...abs.ru, nathan@...nel.org, linuxppc-dev@...ts.ozlabs.org,
        linux-kernel@...r.kernel.org
Subject: [PATCH] KVM: PPC: Book3S HV Nested: clean up warnings found by
 checkpatch


Clean up a handful of checkpatch warnings:
- WARNING: Prefer 'long' over 'long int' as the int is unnecessary
- WARNING: Missing a blank line after declarations
- WARNING: Prefer 'unsigned int *' to bare use of 'unsigned *'
- WARNING: unnecessary whitespace before a quoted newline
- WARNING: please, no spaces at the start of a line
- WARNING: Comparisons should place the constant on the right
side of the test

Signed-off-by: XueBing Chen <chenxuebing@...i.cn>
---
 arch/powerpc/kvm/book3s_hv_nested.c | 21 ++++++++++++---------
 1 file changed, 12 insertions(+), 9 deletions(-)

diff --git a/arch/powerpc/kvm/book3s_hv_nested.c b/arch/powerpc/kvm/book3s_hv_nested.c
index 9d373f8963ee..af56958b0a28 100644
--- a/arch/powerpc/kvm/book3s_hv_nested.c
+++ b/arch/powerpc/kvm/book3s_hv_nested.c
@@ -288,7 +288,7 @@ static void load_l2_hv_regs(struct kvm_vcpu *vcpu,
 
 long kvmhv_enter_nested_guest(struct kvm_vcpu *vcpu)
 {
-	long int err, r;
+	long err, r;
 	struct kvm_nested_guest *l2;
 	struct pt_regs l2_regs, saved_l1_regs;
 	struct hv_guest_state l2_hv = {0}, saved_l1_hv;
@@ -430,7 +430,7 @@ long kvmhv_enter_nested_guest(struct kvm_vcpu *vcpu)
 
 long kvmhv_nested_init(void)
 {
-	long int ptb_order;
+	long ptb_order;
 	unsigned long ptcr;
 	long rc;
 
@@ -646,6 +646,7 @@ static void kvmhv_update_ptbl_cache(struct kvm_nested_guest *gp)
 	ptbl_addr = (kvm->arch.l1_ptcr & PRTB_MASK) + (gp->l1_lpid << 4);
 	if (gp->l1_lpid < (1ul << ((kvm->arch.l1_ptcr & PRTS_MASK) + 8))) {
 		int srcu_idx = srcu_read_lock(&kvm->srcu);
+
 		ret = kvm_read_guest(kvm, ptbl_addr,
 				     &ptbl_entry, sizeof(ptbl_entry));
 		srcu_read_unlock(&kvm->srcu, srcu_idx);
@@ -849,7 +850,7 @@ static struct kvm_nested_guest *kvmhv_find_nested(struct kvm *kvm, int lpid)
 }
 
 pte_t *find_kvm_nested_guest_pte(struct kvm *kvm, unsigned long lpid,
-				 unsigned long ea, unsigned *hshift)
+				 unsigned long ea, unsigned int *hshift)
 {
 	struct kvm_nested_guest *gp;
 	pte_t *pte;
@@ -859,7 +860,7 @@ pte_t *find_kvm_nested_guest_pte(struct kvm *kvm, unsigned long lpid,
 		return NULL;
 
 	VM_WARN(!spin_is_locked(&kvm->mmu_lock),
-		"%s called with kvm mmu_lock not held \n", __func__);
+		"%s called with kvm mmu_lock not held\n", __func__);
 	pte = __find_linux_pte(gp->shadow_pgtable, ea, NULL, hshift);
 
 	return pte;
@@ -1003,6 +1004,7 @@ void kvmhv_remove_nest_rmap_range(struct kvm *kvm,
 
 	for (; gfn < end_gfn; gfn++) {
 		unsigned long *rmap = &memslot->arch.rmap[gfn];
+
 		kvmhv_remove_nest_rmap_list(kvm, rmap, hpa, addr_mask);
 	}
 }
@@ -1475,7 +1477,7 @@ static inline int kvmppc_radix_shift_to_level(int shift)
 }
 
 /* called with gp->tlb_lock held */
-static long int __kvmhv_nested_page_fault(struct kvm_vcpu *vcpu,
+static long __kvmhv_nested_page_fault(struct kvm_vcpu *vcpu,
 					  struct kvm_nested_guest *gp)
 {
 	struct kvm *kvm = vcpu->kvm;
@@ -1491,7 +1493,7 @@ static long int __kvmhv_nested_page_fault(struct kvm_vcpu *vcpu,
 	unsigned int shift, l1_shift, level;
 	bool writing = !!(dsisr & DSISR_ISSTORE);
 	bool kvm_ro = false;
-	long int ret;
+	long ret;
 
 	if (!gp->l1_gr_to_hr) {
 		kvmhv_update_ptbl_cache(gp);
@@ -1614,7 +1616,8 @@ static long int __kvmhv_nested_page_fault(struct kvm_vcpu *vcpu,
 	if (shift > l1_shift) {
 		u64 mask;
 		unsigned int actual_shift = PAGE_SHIFT;
-		if (PMD_SHIFT < l1_shift)
+
+		if (l1_shift > PMD_SHIFT)
 			actual_shift = PMD_SHIFT;
 		mask = (1UL << shift) - (1UL << actual_shift);
 		pte = __pte(pte_val(pte) | (gpa & mask));
@@ -1644,10 +1647,10 @@ static long int __kvmhv_nested_page_fault(struct kvm_vcpu *vcpu,
 	return RESUME_GUEST;
 }
 
-long int kvmhv_nested_page_fault(struct kvm_vcpu *vcpu)
+long kvmhv_nested_page_fault(struct kvm_vcpu *vcpu)
 {
 	struct kvm_nested_guest *gp = vcpu->arch.nested;
-	long int ret;
+	long ret;
 
 	mutex_lock(&gp->tlb_lock);
 	ret = __kvmhv_nested_page_fault(vcpu, gp);
-- 
2.36.1

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ