--- arch/x86/crypto/aes-x86_64-asm_64.S | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) --- a/arch/x86/crypto/aes-x86_64-asm_64.S +++ b/arch/x86/crypto/aes-x86_64-asm_64.S @@ -99,14 +99,14 @@ FUNC: subq $16, R16; \ shrl $16,r2 ## E; \ xorl OFFSET(r8),ra ## E; \ xorl OFFSET+4(r8),rb ## E; \ + movq r3,R9; \ xorl TAB+3072(,r7,4),r5 ## E;\ xorl TAB+2048(,r4,4),r6 ## E;\ - movzbl r1 ## L,r7 ## E; \ movzbl r1 ## H,r4 ## E; \ - movl TAB+1024(,r4,4),r4 ## E;\ - movw r3 ## X,r1 ## X; \ - roll $16,r1 ## E; \ + movzbl r1 ## L,r7 ## E; \ shrl $16,r3 ## E; \ + shrl $16,r1 ## E; \ + movl TAB+1024(,r4,4),r4 ## E;\ xorl TAB(,r7,4),r5 ## E; \ movzbl r3 ## H,r7 ## E; \ movzbl r3 ## L,r3 ## E; \ @@ -114,7 +114,7 @@ FUNC: subq $16, R16; \ xorl TAB+2048(,r3,4),r5 ## E;\ movzbl r1 ## H,r7 ## E; \ movzbl r1 ## L,r3 ## E; \ - shrl $16,r1 ## E; \ + movq R9,r1; \ xorl TAB+3072(,r7,4),r6 ## E;\ movl TAB+2048(,r3,4),r3 ## E;\ movzbl r1 ## H,r7 ## E; \