[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20250815015404.468511-12-zhao.xichao@vivo.com>
Date: Fri, 15 Aug 2025 09:54:00 +0800
From: Xichao Zhao <zhao.xichao@...o.com>
To: Song Liu <song@...nel.org>,
Yu Kuai <yukuai3@...wei.com>,
linux-raid@...r.kernel.org (open list:SOFTWARE RAID (Multiple Disks) SUPPORT),
linux-kernel@...r.kernel.org (open list)
Cc: Xichao Zhao <zhao.xichao@...o.com>
Subject: [PATCH v1 11/13] lib/raid6: Clean up code style in avx512.c
Reduce or add spaces to clean up code style.
No functional changes here.
Signed-off-by: Xichao Zhao <zhao.xichao@...o.com>
---
lib/raid6/avx512.c | 94 +++++++++++++++++++++++-----------------------
1 file changed, 47 insertions(+), 47 deletions(-)
diff --git a/lib/raid6/avx512.c b/lib/raid6/avx512.c
index 009bd0adeebf..18707cbb2bf1 100644
--- a/lib/raid6/avx512.c
+++ b/lib/raid6/avx512.c
@@ -46,8 +46,8 @@ static void raid6_avx5121_gen_syndrome(int disks, size_t bytes, void **ptrs)
int d, z, z0;
z0 = disks - 3; /* Highest data disk */
- p = dptr[z0+1]; /* XOR parity */
- q = dptr[z0+2]; /* RS syndrome */
+ p = dptr[z0 + 1]; /* XOR parity */
+ q = dptr[z0 + 2]; /* RS syndrome */
kernel_fpu_begin();
@@ -64,7 +64,7 @@ static void raid6_avx5121_gen_syndrome(int disks, size_t bytes, void **ptrs)
"vmovdqa64 %1,%%zmm6"
:
: "m" (dptr[z0][d]), "m" (dptr[z0-1][d]));
- for (z = z0-2; z >= 0; z--) {
+ for (z = z0 - 2; z >= 0; z--) {
asm volatile("prefetchnta %0\n\t"
"vpcmpgtb %%zmm4,%%zmm1,%%k1\n\t"
"vpmovm2b %%k1,%%zmm5\n\t"
@@ -104,22 +104,22 @@ static void raid6_avx5121_xor_syndrome(int disks, int start, int stop,
int d, z, z0;
z0 = stop; /* P/Q right side optimization */
- p = dptr[disks-2]; /* XOR parity */
- q = dptr[disks-1]; /* RS syndrome */
+ p = dptr[disks - 2]; /* XOR parity */
+ q = dptr[disks - 1]; /* RS syndrome */
kernel_fpu_begin();
asm volatile("vmovdqa64 %0,%%zmm0"
: : "m" (raid6_avx512_constants.x1d[0]));
- for (d = 0 ; d < bytes ; d += 64) {
+ for (d = 0; d < bytes; d += 64) {
asm volatile("vmovdqa64 %0,%%zmm4\n\t"
"vmovdqa64 %1,%%zmm2\n\t"
"vpxorq %%zmm4,%%zmm2,%%zmm2"
:
: "m" (dptr[z0][d]), "m" (p[d]));
/* P/Q data pages */
- for (z = z0-1 ; z >= start ; z--) {
+ for (z = z0 - 1; z >= start; z--) {
asm volatile("vpxorq %%zmm5,%%zmm5,%%zmm5\n\t"
"vpcmpgtb %%zmm4,%%zmm5,%%k1\n\t"
"vpmovm2b %%k1,%%zmm5\n\t"
@@ -133,7 +133,7 @@ static void raid6_avx5121_xor_syndrome(int disks, int start, int stop,
: "m" (dptr[z][d]));
}
/* P/Q left side optimization */
- for (z = start-1 ; z >= 0 ; z--) {
+ for (z = start - 1; z >= 0; z--) {
asm volatile("vpxorq %%zmm5,%%zmm5,%%zmm5\n\t"
"vpcmpgtb %%zmm4,%%zmm5,%%k1\n\t"
"vpmovm2b %%k1,%%zmm5\n\t"
@@ -173,8 +173,8 @@ static void raid6_avx5122_gen_syndrome(int disks, size_t bytes, void **ptrs)
int d, z, z0;
z0 = disks - 3; /* Highest data disk */
- p = dptr[z0+1]; /* XOR parity */
- q = dptr[z0+2]; /* RS syndrome */
+ p = dptr[z0 + 1]; /* XOR parity */
+ q = dptr[z0 + 2]; /* RS syndrome */
kernel_fpu_begin();
@@ -192,8 +192,8 @@ static void raid6_avx5122_gen_syndrome(int disks, size_t bytes, void **ptrs)
"vmovdqa64 %%zmm2,%%zmm4\n\t" /* Q[0] */
"vmovdqa64 %%zmm3,%%zmm6" /* Q[1] */
:
- : "m" (dptr[z0][d]), "m" (dptr[z0][d+64]));
- for (z = z0-1; z >= 0; z--) {
+ : "m" (dptr[z0][d]), "m" (dptr[z0][d + 64]));
+ for (z = z0 - 1; z >= 0; z--) {
asm volatile("prefetchnta %0\n\t"
"prefetchnta %1\n\t"
"vpcmpgtb %%zmm4,%%zmm1,%%k1\n\t"
@@ -213,7 +213,7 @@ static void raid6_avx5122_gen_syndrome(int disks, size_t bytes, void **ptrs)
"vpxorq %%zmm5,%%zmm4,%%zmm4\n\t"
"vpxorq %%zmm7,%%zmm6,%%zmm6"
:
- : "m" (dptr[z][d]), "m" (dptr[z][d+64]));
+ : "m" (dptr[z][d]), "m" (dptr[z][d + 64]));
}
asm volatile("vmovntdq %%zmm2,%0\n\t"
"vmovntdq %%zmm3,%1\n\t"
@@ -221,7 +221,7 @@ static void raid6_avx5122_gen_syndrome(int disks, size_t bytes, void **ptrs)
"vmovntdq %%zmm6,%3"
:
: "m" (p[d]), "m" (p[d+64]), "m" (q[d]),
- "m" (q[d+64]));
+ "m" (q[d + 64]));
}
asm volatile("sfence" : : : "memory");
@@ -236,15 +236,15 @@ static void raid6_avx5122_xor_syndrome(int disks, int start, int stop,
int d, z, z0;
z0 = stop; /* P/Q right side optimization */
- p = dptr[disks-2]; /* XOR parity */
- q = dptr[disks-1]; /* RS syndrome */
+ p = dptr[disks - 2]; /* XOR parity */
+ q = dptr[disks - 1]; /* RS syndrome */
kernel_fpu_begin();
asm volatile("vmovdqa64 %0,%%zmm0"
: : "m" (raid6_avx512_constants.x1d[0]));
- for (d = 0 ; d < bytes ; d += 128) {
+ for (d = 0; d < bytes; d += 128) {
asm volatile("vmovdqa64 %0,%%zmm4\n\t"
"vmovdqa64 %1,%%zmm6\n\t"
"vmovdqa64 %2,%%zmm2\n\t"
@@ -252,10 +252,10 @@ static void raid6_avx5122_xor_syndrome(int disks, int start, int stop,
"vpxorq %%zmm4,%%zmm2,%%zmm2\n\t"
"vpxorq %%zmm6,%%zmm3,%%zmm3"
:
- : "m" (dptr[z0][d]), "m" (dptr[z0][d+64]),
- "m" (p[d]), "m" (p[d+64]));
+ : "m" (dptr[z0][d]), "m" (dptr[z0][d + 64]),
+ "m" (p[d]), "m" (p[d + 64]));
/* P/Q data pages */
- for (z = z0-1 ; z >= start ; z--) {
+ for (z = z0 - 1; z >= start; z--) {
asm volatile("vpxorq %%zmm5,%%zmm5,%%zmm5\n\t"
"vpxorq %%zmm7,%%zmm7,%%zmm7\n\t"
"vpcmpgtb %%zmm4,%%zmm5,%%k1\n\t"
@@ -275,10 +275,10 @@ static void raid6_avx5122_xor_syndrome(int disks, int start, int stop,
"vpxorq %%zmm5,%%zmm4,%%zmm4\n\t"
"vpxorq %%zmm7,%%zmm6,%%zmm6"
:
- : "m" (dptr[z][d]), "m" (dptr[z][d+64]));
+ : "m" (dptr[z][d]), "m" (dptr[z][d + 64]));
}
/* P/Q left side optimization */
- for (z = start-1 ; z >= 0 ; z--) {
+ for (z = start - 1; z >= 0; z--) {
asm volatile("vpxorq %%zmm5,%%zmm5,%%zmm5\n\t"
"vpxorq %%zmm7,%%zmm7,%%zmm7\n\t"
"vpcmpgtb %%zmm4,%%zmm5,%%k1\n\t"
@@ -304,8 +304,8 @@ static void raid6_avx5122_xor_syndrome(int disks, int start, int stop,
"vmovdqa64 %%zmm2,%2\n\t"
"vmovdqa64 %%zmm3,%3"
:
- : "m" (q[d]), "m" (q[d+64]), "m" (p[d]),
- "m" (p[d+64]));
+ : "m" (q[d]), "m" (q[d + 64]), "m" (p[d]),
+ "m" (p[d + 64]));
}
asm volatile("sfence" : : : "memory");
@@ -332,8 +332,8 @@ static void raid6_avx5124_gen_syndrome(int disks, size_t bytes, void **ptrs)
int d, z, z0;
z0 = disks - 3; /* Highest data disk */
- p = dptr[z0+1]; /* XOR parity */
- q = dptr[z0+2]; /* RS syndrome */
+ p = dptr[z0 + 1]; /* XOR parity */
+ q = dptr[z0 + 2]; /* RS syndrome */
kernel_fpu_begin();
@@ -389,8 +389,8 @@ static void raid6_avx5124_gen_syndrome(int disks, size_t bytes, void **ptrs)
"vpxorq %%zmm13,%%zmm12,%%zmm12\n\t"
"vpxorq %%zmm15,%%zmm14,%%zmm14"
:
- : "m" (dptr[z][d]), "m" (dptr[z][d+64]),
- "m" (dptr[z][d+128]), "m" (dptr[z][d+192]));
+ : "m" (dptr[z][d]), "m" (dptr[z][d + 64]),
+ "m" (dptr[z][d + 128]), "m" (dptr[z][d + 192]));
}
asm volatile("vmovntdq %%zmm2,%0\n\t"
"vpxorq %%zmm2,%%zmm2,%%zmm2\n\t"
@@ -409,9 +409,9 @@ static void raid6_avx5124_gen_syndrome(int disks, size_t bytes, void **ptrs)
"vmovntdq %%zmm14,%7\n\t"
"vpxorq %%zmm14,%%zmm14,%%zmm14"
:
- : "m" (p[d]), "m" (p[d+64]), "m" (p[d+128]),
- "m" (p[d+192]), "m" (q[d]), "m" (q[d+64]),
- "m" (q[d+128]), "m" (q[d+192]));
+ : "m" (p[d]), "m" (p[d + 64]), "m" (p[d + 128]),
+ "m" (p[d + 192]), "m" (q[d]), "m" (q[d + 64]),
+ "m" (q[d + 128]), "m" (q[d + 192]));
}
asm volatile("sfence" : : : "memory");
@@ -426,15 +426,15 @@ static void raid6_avx5124_xor_syndrome(int disks, int start, int stop,
int d, z, z0;
z0 = stop; /* P/Q right side optimization */
- p = dptr[disks-2]; /* XOR parity */
- q = dptr[disks-1]; /* RS syndrome */
+ p = dptr[disks - 2]; /* XOR parity */
+ q = dptr[disks - 1]; /* RS syndrome */
kernel_fpu_begin();
asm volatile("vmovdqa64 %0,%%zmm0"
:: "m" (raid6_avx512_constants.x1d[0]));
- for (d = 0 ; d < bytes ; d += 256) {
+ for (d = 0; d < bytes; d += 256) {
asm volatile("vmovdqa64 %0,%%zmm4\n\t"
"vmovdqa64 %1,%%zmm6\n\t"
"vmovdqa64 %2,%%zmm12\n\t"
@@ -448,12 +448,12 @@ static void raid6_avx5124_xor_syndrome(int disks, int start, int stop,
"vpxorq %%zmm12,%%zmm10,%%zmm10\n\t"
"vpxorq %%zmm14,%%zmm11,%%zmm11"
:
- : "m" (dptr[z0][d]), "m" (dptr[z0][d+64]),
- "m" (dptr[z0][d+128]), "m" (dptr[z0][d+192]),
- "m" (p[d]), "m" (p[d+64]), "m" (p[d+128]),
- "m" (p[d+192]));
+ : "m" (dptr[z0][d]), "m" (dptr[z0][d + 64]),
+ "m" (dptr[z0][d + 128]), "m" (dptr[z0][d + 192]),
+ "m" (p[d]), "m" (p[d + 64]), "m" (p[d + 128]),
+ "m" (p[d + 192]));
/* P/Q data pages */
- for (z = z0-1 ; z >= start ; z--) {
+ for (z = z0 - 1; z >= start; z--) {
asm volatile("vpxorq %%zmm5,%%zmm5,%%zmm5\n\t"
"vpxorq %%zmm7,%%zmm7,%%zmm7\n\t"
"vpxorq %%zmm13,%%zmm13,%%zmm13\n\t"
@@ -493,16 +493,16 @@ static void raid6_avx5124_xor_syndrome(int disks, int start, int stop,
"vpxorq %%zmm13,%%zmm12,%%zmm12\n\t"
"vpxorq %%zmm15,%%zmm14,%%zmm14"
:
- : "m" (dptr[z][d]), "m" (dptr[z][d+64]),
- "m" (dptr[z][d+128]),
- "m" (dptr[z][d+192]));
+ : "m" (dptr[z][d]), "m" (dptr[z][d + 64]),
+ "m" (dptr[z][d + 128]),
+ "m" (dptr[z][d + 192]));
}
asm volatile("prefetchnta %0\n\t"
"prefetchnta %1\n\t"
:
- : "m" (q[d]), "m" (q[d+128]));
+ : "m" (q[d]), "m" (q[d + 128]));
/* P/Q left side optimization */
- for (z = start-1 ; z >= 0 ; z--) {
+ for (z = start - 1; z >= 0; z--) {
asm volatile("vpxorq %%zmm5,%%zmm5,%%zmm5\n\t"
"vpxorq %%zmm7,%%zmm7,%%zmm7\n\t"
"vpxorq %%zmm13,%%zmm13,%%zmm13\n\t"
@@ -543,9 +543,9 @@ static void raid6_avx5124_xor_syndrome(int disks, int start, int stop,
"vmovntdq %%zmm12,%6\n\t"
"vmovntdq %%zmm14,%7"
:
- : "m" (p[d]), "m" (p[d+64]), "m" (p[d+128]),
- "m" (p[d+192]), "m" (q[d]), "m" (q[d+64]),
- "m" (q[d+128]), "m" (q[d+192]));
+ : "m" (p[d]), "m" (p[d + 64]), "m" (p[d + 128]),
+ "m" (p[d + 192]), "m" (q[d]), "m" (q[d + 64]),
+ "m" (q[d + 128]), "m" (q[d + 192]));
}
asm volatile("sfence" : : : "memory");
kernel_fpu_end();
--
2.34.1
Powered by blists - more mailing lists