mirror of
https://github.com/acidanthera/audk.git
synced 2025-09-03 07:58:11 +02:00
2827 lines
76 KiB
NASM
2827 lines
76 KiB
NASM
default rel
|
|
%define XMMWORD
|
|
%define YMMWORD
|
|
%define ZMMWORD
|
|
section .text code align=64
|
|
|
|
|
|
EXTERN asm_AES_encrypt
|
|
EXTERN asm_AES_decrypt
|
|
|
|
|
|
ALIGN 64
|
|
_bsaes_encrypt8:
|
|
|
|
lea r11,[$L$BS0]
|
|
|
|
movdqa xmm8,XMMWORD[rax]
|
|
lea rax,[16+rax]
|
|
movdqa xmm7,XMMWORD[80+r11]
|
|
pxor xmm15,xmm8
|
|
pxor xmm0,xmm8
|
|
pxor xmm1,xmm8
|
|
pxor xmm2,xmm8
|
|
DB 102,68,15,56,0,255
|
|
DB 102,15,56,0,199
|
|
pxor xmm3,xmm8
|
|
pxor xmm4,xmm8
|
|
DB 102,15,56,0,207
|
|
DB 102,15,56,0,215
|
|
pxor xmm5,xmm8
|
|
pxor xmm6,xmm8
|
|
DB 102,15,56,0,223
|
|
DB 102,15,56,0,231
|
|
DB 102,15,56,0,239
|
|
DB 102,15,56,0,247
|
|
_bsaes_encrypt8_bitslice:
|
|
movdqa xmm7,XMMWORD[r11]
|
|
movdqa xmm8,XMMWORD[16+r11]
|
|
movdqa xmm9,xmm5
|
|
psrlq xmm5,1
|
|
movdqa xmm10,xmm3
|
|
psrlq xmm3,1
|
|
pxor xmm5,xmm6
|
|
pxor xmm3,xmm4
|
|
pand xmm5,xmm7
|
|
pand xmm3,xmm7
|
|
pxor xmm6,xmm5
|
|
psllq xmm5,1
|
|
pxor xmm4,xmm3
|
|
psllq xmm3,1
|
|
pxor xmm5,xmm9
|
|
pxor xmm3,xmm10
|
|
movdqa xmm9,xmm1
|
|
psrlq xmm1,1
|
|
movdqa xmm10,xmm15
|
|
psrlq xmm15,1
|
|
pxor xmm1,xmm2
|
|
pxor xmm15,xmm0
|
|
pand xmm1,xmm7
|
|
pand xmm15,xmm7
|
|
pxor xmm2,xmm1
|
|
psllq xmm1,1
|
|
pxor xmm0,xmm15
|
|
psllq xmm15,1
|
|
pxor xmm1,xmm9
|
|
pxor xmm15,xmm10
|
|
movdqa xmm7,XMMWORD[32+r11]
|
|
movdqa xmm9,xmm4
|
|
psrlq xmm4,2
|
|
movdqa xmm10,xmm3
|
|
psrlq xmm3,2
|
|
pxor xmm4,xmm6
|
|
pxor xmm3,xmm5
|
|
pand xmm4,xmm8
|
|
pand xmm3,xmm8
|
|
pxor xmm6,xmm4
|
|
psllq xmm4,2
|
|
pxor xmm5,xmm3
|
|
psllq xmm3,2
|
|
pxor xmm4,xmm9
|
|
pxor xmm3,xmm10
|
|
movdqa xmm9,xmm0
|
|
psrlq xmm0,2
|
|
movdqa xmm10,xmm15
|
|
psrlq xmm15,2
|
|
pxor xmm0,xmm2
|
|
pxor xmm15,xmm1
|
|
pand xmm0,xmm8
|
|
pand xmm15,xmm8
|
|
pxor xmm2,xmm0
|
|
psllq xmm0,2
|
|
pxor xmm1,xmm15
|
|
psllq xmm15,2
|
|
pxor xmm0,xmm9
|
|
pxor xmm15,xmm10
|
|
movdqa xmm9,xmm2
|
|
psrlq xmm2,4
|
|
movdqa xmm10,xmm1
|
|
psrlq xmm1,4
|
|
pxor xmm2,xmm6
|
|
pxor xmm1,xmm5
|
|
pand xmm2,xmm7
|
|
pand xmm1,xmm7
|
|
pxor xmm6,xmm2
|
|
psllq xmm2,4
|
|
pxor xmm5,xmm1
|
|
psllq xmm1,4
|
|
pxor xmm2,xmm9
|
|
pxor xmm1,xmm10
|
|
movdqa xmm9,xmm0
|
|
psrlq xmm0,4
|
|
movdqa xmm10,xmm15
|
|
psrlq xmm15,4
|
|
pxor xmm0,xmm4
|
|
pxor xmm15,xmm3
|
|
pand xmm0,xmm7
|
|
pand xmm15,xmm7
|
|
pxor xmm4,xmm0
|
|
psllq xmm0,4
|
|
pxor xmm3,xmm15
|
|
psllq xmm15,4
|
|
pxor xmm0,xmm9
|
|
pxor xmm15,xmm10
|
|
dec r10d
|
|
jmp NEAR $L$enc_sbox
|
|
ALIGN 16
|
|
$L$enc_loop:
|
|
pxor xmm15,XMMWORD[rax]
|
|
pxor xmm0,XMMWORD[16+rax]
|
|
pxor xmm1,XMMWORD[32+rax]
|
|
pxor xmm2,XMMWORD[48+rax]
|
|
DB 102,68,15,56,0,255
|
|
DB 102,15,56,0,199
|
|
pxor xmm3,XMMWORD[64+rax]
|
|
pxor xmm4,XMMWORD[80+rax]
|
|
DB 102,15,56,0,207
|
|
DB 102,15,56,0,215
|
|
pxor xmm5,XMMWORD[96+rax]
|
|
pxor xmm6,XMMWORD[112+rax]
|
|
DB 102,15,56,0,223
|
|
DB 102,15,56,0,231
|
|
DB 102,15,56,0,239
|
|
DB 102,15,56,0,247
|
|
lea rax,[128+rax]
|
|
$L$enc_sbox:
|
|
pxor xmm4,xmm5
|
|
pxor xmm1,xmm0
|
|
pxor xmm2,xmm15
|
|
pxor xmm5,xmm1
|
|
pxor xmm4,xmm15
|
|
|
|
pxor xmm5,xmm2
|
|
pxor xmm2,xmm6
|
|
pxor xmm6,xmm4
|
|
pxor xmm2,xmm3
|
|
pxor xmm3,xmm4
|
|
pxor xmm2,xmm0
|
|
|
|
pxor xmm1,xmm6
|
|
pxor xmm0,xmm4
|
|
movdqa xmm10,xmm6
|
|
movdqa xmm9,xmm0
|
|
movdqa xmm8,xmm4
|
|
movdqa xmm12,xmm1
|
|
movdqa xmm11,xmm5
|
|
|
|
pxor xmm10,xmm3
|
|
pxor xmm9,xmm1
|
|
pxor xmm8,xmm2
|
|
movdqa xmm13,xmm10
|
|
pxor xmm12,xmm3
|
|
movdqa xmm7,xmm9
|
|
pxor xmm11,xmm15
|
|
movdqa xmm14,xmm10
|
|
|
|
por xmm9,xmm8
|
|
por xmm10,xmm11
|
|
pxor xmm14,xmm7
|
|
pand xmm13,xmm11
|
|
pxor xmm11,xmm8
|
|
pand xmm7,xmm8
|
|
pand xmm14,xmm11
|
|
movdqa xmm11,xmm2
|
|
pxor xmm11,xmm15
|
|
pand xmm12,xmm11
|
|
pxor xmm10,xmm12
|
|
pxor xmm9,xmm12
|
|
movdqa xmm12,xmm6
|
|
movdqa xmm11,xmm4
|
|
pxor xmm12,xmm0
|
|
pxor xmm11,xmm5
|
|
movdqa xmm8,xmm12
|
|
pand xmm12,xmm11
|
|
por xmm8,xmm11
|
|
pxor xmm7,xmm12
|
|
pxor xmm10,xmm14
|
|
pxor xmm9,xmm13
|
|
pxor xmm8,xmm14
|
|
movdqa xmm11,xmm1
|
|
pxor xmm7,xmm13
|
|
movdqa xmm12,xmm3
|
|
pxor xmm8,xmm13
|
|
movdqa xmm13,xmm0
|
|
pand xmm11,xmm2
|
|
movdqa xmm14,xmm6
|
|
pand xmm12,xmm15
|
|
pand xmm13,xmm4
|
|
por xmm14,xmm5
|
|
pxor xmm10,xmm11
|
|
pxor xmm9,xmm12
|
|
pxor xmm8,xmm13
|
|
pxor xmm7,xmm14
|
|
|
|
|
|
|
|
|
|
|
|
movdqa xmm11,xmm10
|
|
pand xmm10,xmm8
|
|
pxor xmm11,xmm9
|
|
|
|
movdqa xmm13,xmm7
|
|
movdqa xmm14,xmm11
|
|
pxor xmm13,xmm10
|
|
pand xmm14,xmm13
|
|
|
|
movdqa xmm12,xmm8
|
|
pxor xmm14,xmm9
|
|
pxor xmm12,xmm7
|
|
|
|
pxor xmm10,xmm9
|
|
|
|
pand xmm12,xmm10
|
|
|
|
movdqa xmm9,xmm13
|
|
pxor xmm12,xmm7
|
|
|
|
pxor xmm9,xmm12
|
|
pxor xmm8,xmm12
|
|
|
|
pand xmm9,xmm7
|
|
|
|
pxor xmm13,xmm9
|
|
pxor xmm8,xmm9
|
|
|
|
pand xmm13,xmm14
|
|
|
|
pxor xmm13,xmm11
|
|
movdqa xmm11,xmm5
|
|
movdqa xmm7,xmm4
|
|
movdqa xmm9,xmm14
|
|
pxor xmm9,xmm13
|
|
pand xmm9,xmm5
|
|
pxor xmm5,xmm4
|
|
pand xmm4,xmm14
|
|
pand xmm5,xmm13
|
|
pxor xmm5,xmm4
|
|
pxor xmm4,xmm9
|
|
pxor xmm11,xmm15
|
|
pxor xmm7,xmm2
|
|
pxor xmm14,xmm12
|
|
pxor xmm13,xmm8
|
|
movdqa xmm10,xmm14
|
|
movdqa xmm9,xmm12
|
|
pxor xmm10,xmm13
|
|
pxor xmm9,xmm8
|
|
pand xmm10,xmm11
|
|
pand xmm9,xmm15
|
|
pxor xmm11,xmm7
|
|
pxor xmm15,xmm2
|
|
pand xmm7,xmm14
|
|
pand xmm2,xmm12
|
|
pand xmm11,xmm13
|
|
pand xmm15,xmm8
|
|
pxor xmm7,xmm11
|
|
pxor xmm15,xmm2
|
|
pxor xmm11,xmm10
|
|
pxor xmm2,xmm9
|
|
pxor xmm5,xmm11
|
|
pxor xmm15,xmm11
|
|
pxor xmm4,xmm7
|
|
pxor xmm2,xmm7
|
|
|
|
movdqa xmm11,xmm6
|
|
movdqa xmm7,xmm0
|
|
pxor xmm11,xmm3
|
|
pxor xmm7,xmm1
|
|
movdqa xmm10,xmm14
|
|
movdqa xmm9,xmm12
|
|
pxor xmm10,xmm13
|
|
pxor xmm9,xmm8
|
|
pand xmm10,xmm11
|
|
pand xmm9,xmm3
|
|
pxor xmm11,xmm7
|
|
pxor xmm3,xmm1
|
|
pand xmm7,xmm14
|
|
pand xmm1,xmm12
|
|
pand xmm11,xmm13
|
|
pand xmm3,xmm8
|
|
pxor xmm7,xmm11
|
|
pxor xmm3,xmm1
|
|
pxor xmm11,xmm10
|
|
pxor xmm1,xmm9
|
|
pxor xmm14,xmm12
|
|
pxor xmm13,xmm8
|
|
movdqa xmm10,xmm14
|
|
pxor xmm10,xmm13
|
|
pand xmm10,xmm6
|
|
pxor xmm6,xmm0
|
|
pand xmm0,xmm14
|
|
pand xmm6,xmm13
|
|
pxor xmm6,xmm0
|
|
pxor xmm0,xmm10
|
|
pxor xmm6,xmm11
|
|
pxor xmm3,xmm11
|
|
pxor xmm0,xmm7
|
|
pxor xmm1,xmm7
|
|
pxor xmm6,xmm15
|
|
pxor xmm0,xmm5
|
|
pxor xmm3,xmm6
|
|
pxor xmm5,xmm15
|
|
pxor xmm15,xmm0
|
|
|
|
pxor xmm0,xmm4
|
|
pxor xmm4,xmm1
|
|
pxor xmm1,xmm2
|
|
pxor xmm2,xmm4
|
|
pxor xmm3,xmm4
|
|
|
|
pxor xmm5,xmm2
|
|
dec r10d
|
|
jl NEAR $L$enc_done
|
|
pshufd xmm7,xmm15,0x93
|
|
pshufd xmm8,xmm0,0x93
|
|
pxor xmm15,xmm7
|
|
pshufd xmm9,xmm3,0x93
|
|
pxor xmm0,xmm8
|
|
pshufd xmm10,xmm5,0x93
|
|
pxor xmm3,xmm9
|
|
pshufd xmm11,xmm2,0x93
|
|
pxor xmm5,xmm10
|
|
pshufd xmm12,xmm6,0x93
|
|
pxor xmm2,xmm11
|
|
pshufd xmm13,xmm1,0x93
|
|
pxor xmm6,xmm12
|
|
pshufd xmm14,xmm4,0x93
|
|
pxor xmm1,xmm13
|
|
pxor xmm4,xmm14
|
|
|
|
pxor xmm8,xmm15
|
|
pxor xmm7,xmm4
|
|
pxor xmm8,xmm4
|
|
pshufd xmm15,xmm15,0x4E
|
|
pxor xmm9,xmm0
|
|
pshufd xmm0,xmm0,0x4E
|
|
pxor xmm12,xmm2
|
|
pxor xmm15,xmm7
|
|
pxor xmm13,xmm6
|
|
pxor xmm0,xmm8
|
|
pxor xmm11,xmm5
|
|
pshufd xmm7,xmm2,0x4E
|
|
pxor xmm14,xmm1
|
|
pshufd xmm8,xmm6,0x4E
|
|
pxor xmm10,xmm3
|
|
pshufd xmm2,xmm5,0x4E
|
|
pxor xmm10,xmm4
|
|
pshufd xmm6,xmm4,0x4E
|
|
pxor xmm11,xmm4
|
|
pshufd xmm5,xmm1,0x4E
|
|
pxor xmm7,xmm11
|
|
pshufd xmm1,xmm3,0x4E
|
|
pxor xmm8,xmm12
|
|
pxor xmm2,xmm10
|
|
pxor xmm6,xmm14
|
|
pxor xmm5,xmm13
|
|
movdqa xmm3,xmm7
|
|
pxor xmm1,xmm9
|
|
movdqa xmm4,xmm8
|
|
movdqa xmm7,XMMWORD[48+r11]
|
|
jnz NEAR $L$enc_loop
|
|
movdqa xmm7,XMMWORD[64+r11]
|
|
jmp NEAR $L$enc_loop
|
|
ALIGN 16
|
|
$L$enc_done:
|
|
movdqa xmm7,XMMWORD[r11]
|
|
movdqa xmm8,XMMWORD[16+r11]
|
|
movdqa xmm9,xmm1
|
|
psrlq xmm1,1
|
|
movdqa xmm10,xmm2
|
|
psrlq xmm2,1
|
|
pxor xmm1,xmm4
|
|
pxor xmm2,xmm6
|
|
pand xmm1,xmm7
|
|
pand xmm2,xmm7
|
|
pxor xmm4,xmm1
|
|
psllq xmm1,1
|
|
pxor xmm6,xmm2
|
|
psllq xmm2,1
|
|
pxor xmm1,xmm9
|
|
pxor xmm2,xmm10
|
|
movdqa xmm9,xmm3
|
|
psrlq xmm3,1
|
|
movdqa xmm10,xmm15
|
|
psrlq xmm15,1
|
|
pxor xmm3,xmm5
|
|
pxor xmm15,xmm0
|
|
pand xmm3,xmm7
|
|
pand xmm15,xmm7
|
|
pxor xmm5,xmm3
|
|
psllq xmm3,1
|
|
pxor xmm0,xmm15
|
|
psllq xmm15,1
|
|
pxor xmm3,xmm9
|
|
pxor xmm15,xmm10
|
|
movdqa xmm7,XMMWORD[32+r11]
|
|
movdqa xmm9,xmm6
|
|
psrlq xmm6,2
|
|
movdqa xmm10,xmm2
|
|
psrlq xmm2,2
|
|
pxor xmm6,xmm4
|
|
pxor xmm2,xmm1
|
|
pand xmm6,xmm8
|
|
pand xmm2,xmm8
|
|
pxor xmm4,xmm6
|
|
psllq xmm6,2
|
|
pxor xmm1,xmm2
|
|
psllq xmm2,2
|
|
pxor xmm6,xmm9
|
|
pxor xmm2,xmm10
|
|
movdqa xmm9,xmm0
|
|
psrlq xmm0,2
|
|
movdqa xmm10,xmm15
|
|
psrlq xmm15,2
|
|
pxor xmm0,xmm5
|
|
pxor xmm15,xmm3
|
|
pand xmm0,xmm8
|
|
pand xmm15,xmm8
|
|
pxor xmm5,xmm0
|
|
psllq xmm0,2
|
|
pxor xmm3,xmm15
|
|
psllq xmm15,2
|
|
pxor xmm0,xmm9
|
|
pxor xmm15,xmm10
|
|
movdqa xmm9,xmm5
|
|
psrlq xmm5,4
|
|
movdqa xmm10,xmm3
|
|
psrlq xmm3,4
|
|
pxor xmm5,xmm4
|
|
pxor xmm3,xmm1
|
|
pand xmm5,xmm7
|
|
pand xmm3,xmm7
|
|
pxor xmm4,xmm5
|
|
psllq xmm5,4
|
|
pxor xmm1,xmm3
|
|
psllq xmm3,4
|
|
pxor xmm5,xmm9
|
|
pxor xmm3,xmm10
|
|
movdqa xmm9,xmm0
|
|
psrlq xmm0,4
|
|
movdqa xmm10,xmm15
|
|
psrlq xmm15,4
|
|
pxor xmm0,xmm6
|
|
pxor xmm15,xmm2
|
|
pand xmm0,xmm7
|
|
pand xmm15,xmm7
|
|
pxor xmm6,xmm0
|
|
psllq xmm0,4
|
|
pxor xmm2,xmm15
|
|
psllq xmm15,4
|
|
pxor xmm0,xmm9
|
|
pxor xmm15,xmm10
|
|
movdqa xmm7,XMMWORD[rax]
|
|
pxor xmm3,xmm7
|
|
pxor xmm5,xmm7
|
|
pxor xmm2,xmm7
|
|
pxor xmm6,xmm7
|
|
pxor xmm1,xmm7
|
|
pxor xmm4,xmm7
|
|
pxor xmm15,xmm7
|
|
pxor xmm0,xmm7
|
|
DB 0F3h,0C3h ;repret
|
|
|
|
|
|
|
|
|
|
ALIGN 64
|
|
_bsaes_decrypt8:
|
|
|
|
lea r11,[$L$BS0]
|
|
|
|
movdqa xmm8,XMMWORD[rax]
|
|
lea rax,[16+rax]
|
|
movdqa xmm7,XMMWORD[((-48))+r11]
|
|
pxor xmm15,xmm8
|
|
pxor xmm0,xmm8
|
|
pxor xmm1,xmm8
|
|
pxor xmm2,xmm8
|
|
DB 102,68,15,56,0,255
|
|
DB 102,15,56,0,199
|
|
pxor xmm3,xmm8
|
|
pxor xmm4,xmm8
|
|
DB 102,15,56,0,207
|
|
DB 102,15,56,0,215
|
|
pxor xmm5,xmm8
|
|
pxor xmm6,xmm8
|
|
DB 102,15,56,0,223
|
|
DB 102,15,56,0,231
|
|
DB 102,15,56,0,239
|
|
DB 102,15,56,0,247
|
|
movdqa xmm7,XMMWORD[r11]
|
|
movdqa xmm8,XMMWORD[16+r11]
|
|
movdqa xmm9,xmm5
|
|
psrlq xmm5,1
|
|
movdqa xmm10,xmm3
|
|
psrlq xmm3,1
|
|
pxor xmm5,xmm6
|
|
pxor xmm3,xmm4
|
|
pand xmm5,xmm7
|
|
pand xmm3,xmm7
|
|
pxor xmm6,xmm5
|
|
psllq xmm5,1
|
|
pxor xmm4,xmm3
|
|
psllq xmm3,1
|
|
pxor xmm5,xmm9
|
|
pxor xmm3,xmm10
|
|
movdqa xmm9,xmm1
|
|
psrlq xmm1,1
|
|
movdqa xmm10,xmm15
|
|
psrlq xmm15,1
|
|
pxor xmm1,xmm2
|
|
pxor xmm15,xmm0
|
|
pand xmm1,xmm7
|
|
pand xmm15,xmm7
|
|
pxor xmm2,xmm1
|
|
psllq xmm1,1
|
|
pxor xmm0,xmm15
|
|
psllq xmm15,1
|
|
pxor xmm1,xmm9
|
|
pxor xmm15,xmm10
|
|
movdqa xmm7,XMMWORD[32+r11]
|
|
movdqa xmm9,xmm4
|
|
psrlq xmm4,2
|
|
movdqa xmm10,xmm3
|
|
psrlq xmm3,2
|
|
pxor xmm4,xmm6
|
|
pxor xmm3,xmm5
|
|
pand xmm4,xmm8
|
|
pand xmm3,xmm8
|
|
pxor xmm6,xmm4
|
|
psllq xmm4,2
|
|
pxor xmm5,xmm3
|
|
psllq xmm3,2
|
|
pxor xmm4,xmm9
|
|
pxor xmm3,xmm10
|
|
movdqa xmm9,xmm0
|
|
psrlq xmm0,2
|
|
movdqa xmm10,xmm15
|
|
psrlq xmm15,2
|
|
pxor xmm0,xmm2
|
|
pxor xmm15,xmm1
|
|
pand xmm0,xmm8
|
|
pand xmm15,xmm8
|
|
pxor xmm2,xmm0
|
|
psllq xmm0,2
|
|
pxor xmm1,xmm15
|
|
psllq xmm15,2
|
|
pxor xmm0,xmm9
|
|
pxor xmm15,xmm10
|
|
movdqa xmm9,xmm2
|
|
psrlq xmm2,4
|
|
movdqa xmm10,xmm1
|
|
psrlq xmm1,4
|
|
pxor xmm2,xmm6
|
|
pxor xmm1,xmm5
|
|
pand xmm2,xmm7
|
|
pand xmm1,xmm7
|
|
pxor xmm6,xmm2
|
|
psllq xmm2,4
|
|
pxor xmm5,xmm1
|
|
psllq xmm1,4
|
|
pxor xmm2,xmm9
|
|
pxor xmm1,xmm10
|
|
movdqa xmm9,xmm0
|
|
psrlq xmm0,4
|
|
movdqa xmm10,xmm15
|
|
psrlq xmm15,4
|
|
pxor xmm0,xmm4
|
|
pxor xmm15,xmm3
|
|
pand xmm0,xmm7
|
|
pand xmm15,xmm7
|
|
pxor xmm4,xmm0
|
|
psllq xmm0,4
|
|
pxor xmm3,xmm15
|
|
psllq xmm15,4
|
|
pxor xmm0,xmm9
|
|
pxor xmm15,xmm10
|
|
dec r10d
|
|
jmp NEAR $L$dec_sbox
|
|
ALIGN 16
|
|
$L$dec_loop:
|
|
pxor xmm15,XMMWORD[rax]
|
|
pxor xmm0,XMMWORD[16+rax]
|
|
pxor xmm1,XMMWORD[32+rax]
|
|
pxor xmm2,XMMWORD[48+rax]
|
|
DB 102,68,15,56,0,255
|
|
DB 102,15,56,0,199
|
|
pxor xmm3,XMMWORD[64+rax]
|
|
pxor xmm4,XMMWORD[80+rax]
|
|
DB 102,15,56,0,207
|
|
DB 102,15,56,0,215
|
|
pxor xmm5,XMMWORD[96+rax]
|
|
pxor xmm6,XMMWORD[112+rax]
|
|
DB 102,15,56,0,223
|
|
DB 102,15,56,0,231
|
|
DB 102,15,56,0,239
|
|
DB 102,15,56,0,247
|
|
lea rax,[128+rax]
|
|
$L$dec_sbox:
|
|
pxor xmm2,xmm3
|
|
|
|
pxor xmm3,xmm6
|
|
pxor xmm1,xmm6
|
|
pxor xmm5,xmm3
|
|
pxor xmm6,xmm5
|
|
pxor xmm0,xmm6
|
|
|
|
pxor xmm15,xmm0
|
|
pxor xmm1,xmm4
|
|
pxor xmm2,xmm15
|
|
pxor xmm4,xmm15
|
|
pxor xmm0,xmm2
|
|
movdqa xmm10,xmm2
|
|
movdqa xmm9,xmm6
|
|
movdqa xmm8,xmm0
|
|
movdqa xmm12,xmm3
|
|
movdqa xmm11,xmm4
|
|
|
|
pxor xmm10,xmm15
|
|
pxor xmm9,xmm3
|
|
pxor xmm8,xmm5
|
|
movdqa xmm13,xmm10
|
|
pxor xmm12,xmm15
|
|
movdqa xmm7,xmm9
|
|
pxor xmm11,xmm1
|
|
movdqa xmm14,xmm10
|
|
|
|
por xmm9,xmm8
|
|
por xmm10,xmm11
|
|
pxor xmm14,xmm7
|
|
pand xmm13,xmm11
|
|
pxor xmm11,xmm8
|
|
pand xmm7,xmm8
|
|
pand xmm14,xmm11
|
|
movdqa xmm11,xmm5
|
|
pxor xmm11,xmm1
|
|
pand xmm12,xmm11
|
|
pxor xmm10,xmm12
|
|
pxor xmm9,xmm12
|
|
movdqa xmm12,xmm2
|
|
movdqa xmm11,xmm0
|
|
pxor xmm12,xmm6
|
|
pxor xmm11,xmm4
|
|
movdqa xmm8,xmm12
|
|
pand xmm12,xmm11
|
|
por xmm8,xmm11
|
|
pxor xmm7,xmm12
|
|
pxor xmm10,xmm14
|
|
pxor xmm9,xmm13
|
|
pxor xmm8,xmm14
|
|
movdqa xmm11,xmm3
|
|
pxor xmm7,xmm13
|
|
movdqa xmm12,xmm15
|
|
pxor xmm8,xmm13
|
|
movdqa xmm13,xmm6
|
|
pand xmm11,xmm5
|
|
movdqa xmm14,xmm2
|
|
pand xmm12,xmm1
|
|
pand xmm13,xmm0
|
|
por xmm14,xmm4
|
|
pxor xmm10,xmm11
|
|
pxor xmm9,xmm12
|
|
pxor xmm8,xmm13
|
|
pxor xmm7,xmm14
|
|
|
|
|
|
|
|
|
|
|
|
movdqa xmm11,xmm10
|
|
pand xmm10,xmm8
|
|
pxor xmm11,xmm9
|
|
|
|
movdqa xmm13,xmm7
|
|
movdqa xmm14,xmm11
|
|
pxor xmm13,xmm10
|
|
pand xmm14,xmm13
|
|
|
|
movdqa xmm12,xmm8
|
|
pxor xmm14,xmm9
|
|
pxor xmm12,xmm7
|
|
|
|
pxor xmm10,xmm9
|
|
|
|
pand xmm12,xmm10
|
|
|
|
movdqa xmm9,xmm13
|
|
pxor xmm12,xmm7
|
|
|
|
pxor xmm9,xmm12
|
|
pxor xmm8,xmm12
|
|
|
|
pand xmm9,xmm7
|
|
|
|
pxor xmm13,xmm9
|
|
pxor xmm8,xmm9
|
|
|
|
pand xmm13,xmm14
|
|
|
|
pxor xmm13,xmm11
|
|
movdqa xmm11,xmm4
|
|
movdqa xmm7,xmm0
|
|
movdqa xmm9,xmm14
|
|
pxor xmm9,xmm13
|
|
pand xmm9,xmm4
|
|
pxor xmm4,xmm0
|
|
pand xmm0,xmm14
|
|
pand xmm4,xmm13
|
|
pxor xmm4,xmm0
|
|
pxor xmm0,xmm9
|
|
pxor xmm11,xmm1
|
|
pxor xmm7,xmm5
|
|
pxor xmm14,xmm12
|
|
pxor xmm13,xmm8
|
|
movdqa xmm10,xmm14
|
|
movdqa xmm9,xmm12
|
|
pxor xmm10,xmm13
|
|
pxor xmm9,xmm8
|
|
pand xmm10,xmm11
|
|
pand xmm9,xmm1
|
|
pxor xmm11,xmm7
|
|
pxor xmm1,xmm5
|
|
pand xmm7,xmm14
|
|
pand xmm5,xmm12
|
|
pand xmm11,xmm13
|
|
pand xmm1,xmm8
|
|
pxor xmm7,xmm11
|
|
pxor xmm1,xmm5
|
|
pxor xmm11,xmm10
|
|
pxor xmm5,xmm9
|
|
pxor xmm4,xmm11
|
|
pxor xmm1,xmm11
|
|
pxor xmm0,xmm7
|
|
pxor xmm5,xmm7
|
|
|
|
movdqa xmm11,xmm2
|
|
movdqa xmm7,xmm6
|
|
pxor xmm11,xmm15
|
|
pxor xmm7,xmm3
|
|
movdqa xmm10,xmm14
|
|
movdqa xmm9,xmm12
|
|
pxor xmm10,xmm13
|
|
pxor xmm9,xmm8
|
|
pand xmm10,xmm11
|
|
pand xmm9,xmm15
|
|
pxor xmm11,xmm7
|
|
pxor xmm15,xmm3
|
|
pand xmm7,xmm14
|
|
pand xmm3,xmm12
|
|
pand xmm11,xmm13
|
|
pand xmm15,xmm8
|
|
pxor xmm7,xmm11
|
|
pxor xmm15,xmm3
|
|
pxor xmm11,xmm10
|
|
pxor xmm3,xmm9
|
|
pxor xmm14,xmm12
|
|
pxor xmm13,xmm8
|
|
movdqa xmm10,xmm14
|
|
pxor xmm10,xmm13
|
|
pand xmm10,xmm2
|
|
pxor xmm2,xmm6
|
|
pand xmm6,xmm14
|
|
pand xmm2,xmm13
|
|
pxor xmm2,xmm6
|
|
pxor xmm6,xmm10
|
|
pxor xmm2,xmm11
|
|
pxor xmm15,xmm11
|
|
pxor xmm6,xmm7
|
|
pxor xmm3,xmm7
|
|
pxor xmm0,xmm6
|
|
pxor xmm5,xmm4
|
|
|
|
pxor xmm3,xmm0
|
|
pxor xmm1,xmm6
|
|
pxor xmm4,xmm6
|
|
pxor xmm3,xmm1
|
|
pxor xmm6,xmm15
|
|
pxor xmm3,xmm4
|
|
pxor xmm2,xmm5
|
|
pxor xmm5,xmm0
|
|
pxor xmm2,xmm3
|
|
|
|
pxor xmm3,xmm15
|
|
pxor xmm6,xmm2
|
|
dec r10d
|
|
jl NEAR $L$dec_done
|
|
|
|
pshufd xmm7,xmm15,0x4E
|
|
pshufd xmm13,xmm2,0x4E
|
|
pxor xmm7,xmm15
|
|
pshufd xmm14,xmm4,0x4E
|
|
pxor xmm13,xmm2
|
|
pshufd xmm8,xmm0,0x4E
|
|
pxor xmm14,xmm4
|
|
pshufd xmm9,xmm5,0x4E
|
|
pxor xmm8,xmm0
|
|
pshufd xmm10,xmm3,0x4E
|
|
pxor xmm9,xmm5
|
|
pxor xmm15,xmm13
|
|
pxor xmm0,xmm13
|
|
pshufd xmm11,xmm1,0x4E
|
|
pxor xmm10,xmm3
|
|
pxor xmm5,xmm7
|
|
pxor xmm3,xmm8
|
|
pshufd xmm12,xmm6,0x4E
|
|
pxor xmm11,xmm1
|
|
pxor xmm0,xmm14
|
|
pxor xmm1,xmm9
|
|
pxor xmm12,xmm6
|
|
|
|
pxor xmm5,xmm14
|
|
pxor xmm3,xmm13
|
|
pxor xmm1,xmm13
|
|
pxor xmm6,xmm10
|
|
pxor xmm2,xmm11
|
|
pxor xmm1,xmm14
|
|
pxor xmm6,xmm14
|
|
pxor xmm4,xmm12
|
|
pshufd xmm7,xmm15,0x93
|
|
pshufd xmm8,xmm0,0x93
|
|
pxor xmm15,xmm7
|
|
pshufd xmm9,xmm5,0x93
|
|
pxor xmm0,xmm8
|
|
pshufd xmm10,xmm3,0x93
|
|
pxor xmm5,xmm9
|
|
pshufd xmm11,xmm1,0x93
|
|
pxor xmm3,xmm10
|
|
pshufd xmm12,xmm6,0x93
|
|
pxor xmm1,xmm11
|
|
pshufd xmm13,xmm2,0x93
|
|
pxor xmm6,xmm12
|
|
pshufd xmm14,xmm4,0x93
|
|
pxor xmm2,xmm13
|
|
pxor xmm4,xmm14
|
|
|
|
pxor xmm8,xmm15
|
|
pxor xmm7,xmm4
|
|
pxor xmm8,xmm4
|
|
pshufd xmm15,xmm15,0x4E
|
|
pxor xmm9,xmm0
|
|
pshufd xmm0,xmm0,0x4E
|
|
pxor xmm12,xmm1
|
|
pxor xmm15,xmm7
|
|
pxor xmm13,xmm6
|
|
pxor xmm0,xmm8
|
|
pxor xmm11,xmm3
|
|
pshufd xmm7,xmm1,0x4E
|
|
pxor xmm14,xmm2
|
|
pshufd xmm8,xmm6,0x4E
|
|
pxor xmm10,xmm5
|
|
pshufd xmm1,xmm3,0x4E
|
|
pxor xmm10,xmm4
|
|
pshufd xmm6,xmm4,0x4E
|
|
pxor xmm11,xmm4
|
|
pshufd xmm3,xmm2,0x4E
|
|
pxor xmm7,xmm11
|
|
pshufd xmm2,xmm5,0x4E
|
|
pxor xmm8,xmm12
|
|
pxor xmm10,xmm1
|
|
pxor xmm6,xmm14
|
|
pxor xmm13,xmm3
|
|
movdqa xmm3,xmm7
|
|
pxor xmm2,xmm9
|
|
movdqa xmm5,xmm13
|
|
movdqa xmm4,xmm8
|
|
movdqa xmm1,xmm2
|
|
movdqa xmm2,xmm10
|
|
movdqa xmm7,XMMWORD[((-16))+r11]
|
|
jnz NEAR $L$dec_loop
|
|
movdqa xmm7,XMMWORD[((-32))+r11]
|
|
jmp NEAR $L$dec_loop
|
|
ALIGN 16
|
|
$L$dec_done:
|
|
movdqa xmm7,XMMWORD[r11]
|
|
movdqa xmm8,XMMWORD[16+r11]
|
|
movdqa xmm9,xmm2
|
|
psrlq xmm2,1
|
|
movdqa xmm10,xmm1
|
|
psrlq xmm1,1
|
|
pxor xmm2,xmm4
|
|
pxor xmm1,xmm6
|
|
pand xmm2,xmm7
|
|
pand xmm1,xmm7
|
|
pxor xmm4,xmm2
|
|
psllq xmm2,1
|
|
pxor xmm6,xmm1
|
|
psllq xmm1,1
|
|
pxor xmm2,xmm9
|
|
pxor xmm1,xmm10
|
|
movdqa xmm9,xmm5
|
|
psrlq xmm5,1
|
|
movdqa xmm10,xmm15
|
|
psrlq xmm15,1
|
|
pxor xmm5,xmm3
|
|
pxor xmm15,xmm0
|
|
pand xmm5,xmm7
|
|
pand xmm15,xmm7
|
|
pxor xmm3,xmm5
|
|
psllq xmm5,1
|
|
pxor xmm0,xmm15
|
|
psllq xmm15,1
|
|
pxor xmm5,xmm9
|
|
pxor xmm15,xmm10
|
|
movdqa xmm7,XMMWORD[32+r11]
|
|
movdqa xmm9,xmm6
|
|
psrlq xmm6,2
|
|
movdqa xmm10,xmm1
|
|
psrlq xmm1,2
|
|
pxor xmm6,xmm4
|
|
pxor xmm1,xmm2
|
|
pand xmm6,xmm8
|
|
pand xmm1,xmm8
|
|
pxor xmm4,xmm6
|
|
psllq xmm6,2
|
|
pxor xmm2,xmm1
|
|
psllq xmm1,2
|
|
pxor xmm6,xmm9
|
|
pxor xmm1,xmm10
|
|
movdqa xmm9,xmm0
|
|
psrlq xmm0,2
|
|
movdqa xmm10,xmm15
|
|
psrlq xmm15,2
|
|
pxor xmm0,xmm3
|
|
pxor xmm15,xmm5
|
|
pand xmm0,xmm8
|
|
pand xmm15,xmm8
|
|
pxor xmm3,xmm0
|
|
psllq xmm0,2
|
|
pxor xmm5,xmm15
|
|
psllq xmm15,2
|
|
pxor xmm0,xmm9
|
|
pxor xmm15,xmm10
|
|
movdqa xmm9,xmm3
|
|
psrlq xmm3,4
|
|
movdqa xmm10,xmm5
|
|
psrlq xmm5,4
|
|
pxor xmm3,xmm4
|
|
pxor xmm5,xmm2
|
|
pand xmm3,xmm7
|
|
pand xmm5,xmm7
|
|
pxor xmm4,xmm3
|
|
psllq xmm3,4
|
|
pxor xmm2,xmm5
|
|
psllq xmm5,4
|
|
pxor xmm3,xmm9
|
|
pxor xmm5,xmm10
|
|
movdqa xmm9,xmm0
|
|
psrlq xmm0,4
|
|
movdqa xmm10,xmm15
|
|
psrlq xmm15,4
|
|
pxor xmm0,xmm6
|
|
pxor xmm15,xmm1
|
|
pand xmm0,xmm7
|
|
pand xmm15,xmm7
|
|
pxor xmm6,xmm0
|
|
psllq xmm0,4
|
|
pxor xmm1,xmm15
|
|
psllq xmm15,4
|
|
pxor xmm0,xmm9
|
|
pxor xmm15,xmm10
|
|
movdqa xmm7,XMMWORD[rax]
|
|
pxor xmm5,xmm7
|
|
pxor xmm3,xmm7
|
|
pxor xmm1,xmm7
|
|
pxor xmm6,xmm7
|
|
pxor xmm2,xmm7
|
|
pxor xmm4,xmm7
|
|
pxor xmm15,xmm7
|
|
pxor xmm0,xmm7
|
|
DB 0F3h,0C3h ;repret
|
|
|
|
|
|
|
|
ALIGN 16
|
|
_bsaes_key_convert:
|
|
|
|
lea r11,[$L$masks]
|
|
movdqu xmm7,XMMWORD[rcx]
|
|
lea rcx,[16+rcx]
|
|
movdqa xmm0,XMMWORD[r11]
|
|
movdqa xmm1,XMMWORD[16+r11]
|
|
movdqa xmm2,XMMWORD[32+r11]
|
|
movdqa xmm3,XMMWORD[48+r11]
|
|
movdqa xmm4,XMMWORD[64+r11]
|
|
pcmpeqd xmm5,xmm5
|
|
|
|
movdqu xmm6,XMMWORD[rcx]
|
|
movdqa XMMWORD[rax],xmm7
|
|
lea rax,[16+rax]
|
|
dec r10d
|
|
jmp NEAR $L$key_loop
|
|
ALIGN 16
|
|
$L$key_loop:
|
|
DB 102,15,56,0,244
|
|
|
|
movdqa xmm8,xmm0
|
|
movdqa xmm9,xmm1
|
|
|
|
pand xmm8,xmm6
|
|
pand xmm9,xmm6
|
|
movdqa xmm10,xmm2
|
|
pcmpeqb xmm8,xmm0
|
|
psllq xmm0,4
|
|
movdqa xmm11,xmm3
|
|
pcmpeqb xmm9,xmm1
|
|
psllq xmm1,4
|
|
|
|
pand xmm10,xmm6
|
|
pand xmm11,xmm6
|
|
movdqa xmm12,xmm0
|
|
pcmpeqb xmm10,xmm2
|
|
psllq xmm2,4
|
|
movdqa xmm13,xmm1
|
|
pcmpeqb xmm11,xmm3
|
|
psllq xmm3,4
|
|
|
|
movdqa xmm14,xmm2
|
|
movdqa xmm15,xmm3
|
|
pxor xmm8,xmm5
|
|
pxor xmm9,xmm5
|
|
|
|
pand xmm12,xmm6
|
|
pand xmm13,xmm6
|
|
movdqa XMMWORD[rax],xmm8
|
|
pcmpeqb xmm12,xmm0
|
|
psrlq xmm0,4
|
|
movdqa XMMWORD[16+rax],xmm9
|
|
pcmpeqb xmm13,xmm1
|
|
psrlq xmm1,4
|
|
lea rcx,[16+rcx]
|
|
|
|
pand xmm14,xmm6
|
|
pand xmm15,xmm6
|
|
movdqa XMMWORD[32+rax],xmm10
|
|
pcmpeqb xmm14,xmm2
|
|
psrlq xmm2,4
|
|
movdqa XMMWORD[48+rax],xmm11
|
|
pcmpeqb xmm15,xmm3
|
|
psrlq xmm3,4
|
|
movdqu xmm6,XMMWORD[rcx]
|
|
|
|
pxor xmm13,xmm5
|
|
pxor xmm14,xmm5
|
|
movdqa XMMWORD[64+rax],xmm12
|
|
movdqa XMMWORD[80+rax],xmm13
|
|
movdqa XMMWORD[96+rax],xmm14
|
|
movdqa XMMWORD[112+rax],xmm15
|
|
lea rax,[128+rax]
|
|
dec r10d
|
|
jnz NEAR $L$key_loop
|
|
|
|
movdqa xmm7,XMMWORD[80+r11]
|
|
|
|
DB 0F3h,0C3h ;repret
|
|
|
|
|
|
EXTERN asm_AES_cbc_encrypt
|
|
global ossl_bsaes_cbc_encrypt
|
|
|
|
ALIGN 16
|
|
ossl_bsaes_cbc_encrypt:
|
|
|
|
DB 243,15,30,250
|
|
mov r11d,DWORD[48+rsp]
|
|
cmp r11d,0
|
|
jne NEAR asm_AES_cbc_encrypt
|
|
cmp r8,128
|
|
jb NEAR asm_AES_cbc_encrypt
|
|
|
|
mov rax,rsp
|
|
$L$cbc_dec_prologue:
|
|
push rbp
|
|
|
|
push rbx
|
|
|
|
push r12
|
|
|
|
push r13
|
|
|
|
push r14
|
|
|
|
push r15
|
|
|
|
lea rsp,[((-72))+rsp]
|
|
|
|
mov r10,QWORD[160+rsp]
|
|
lea rsp,[((-160))+rsp]
|
|
movaps XMMWORD[64+rsp],xmm6
|
|
movaps XMMWORD[80+rsp],xmm7
|
|
movaps XMMWORD[96+rsp],xmm8
|
|
movaps XMMWORD[112+rsp],xmm9
|
|
movaps XMMWORD[128+rsp],xmm10
|
|
movaps XMMWORD[144+rsp],xmm11
|
|
movaps XMMWORD[160+rsp],xmm12
|
|
movaps XMMWORD[176+rsp],xmm13
|
|
movaps XMMWORD[192+rsp],xmm14
|
|
movaps XMMWORD[208+rsp],xmm15
|
|
$L$cbc_dec_body:
|
|
mov rbp,rsp
|
|
|
|
mov eax,DWORD[240+r9]
|
|
mov r12,rcx
|
|
mov r13,rdx
|
|
mov r14,r8
|
|
mov r15,r9
|
|
mov rbx,r10
|
|
shr r14,4
|
|
|
|
mov edx,eax
|
|
shl rax,7
|
|
sub rax,96
|
|
sub rsp,rax
|
|
|
|
mov rax,rsp
|
|
mov rcx,r15
|
|
mov r10d,edx
|
|
call _bsaes_key_convert
|
|
pxor xmm7,XMMWORD[rsp]
|
|
movdqa XMMWORD[rax],xmm6
|
|
movdqa XMMWORD[rsp],xmm7
|
|
|
|
movdqu xmm14,XMMWORD[rbx]
|
|
sub r14,8
|
|
$L$cbc_dec_loop:
|
|
movdqu xmm15,XMMWORD[r12]
|
|
movdqu xmm0,XMMWORD[16+r12]
|
|
movdqu xmm1,XMMWORD[32+r12]
|
|
movdqu xmm2,XMMWORD[48+r12]
|
|
movdqu xmm3,XMMWORD[64+r12]
|
|
movdqu xmm4,XMMWORD[80+r12]
|
|
mov rax,rsp
|
|
movdqu xmm5,XMMWORD[96+r12]
|
|
mov r10d,edx
|
|
movdqu xmm6,XMMWORD[112+r12]
|
|
movdqa XMMWORD[32+rbp],xmm14
|
|
|
|
call _bsaes_decrypt8
|
|
|
|
pxor xmm15,XMMWORD[32+rbp]
|
|
movdqu xmm7,XMMWORD[r12]
|
|
movdqu xmm8,XMMWORD[16+r12]
|
|
pxor xmm0,xmm7
|
|
movdqu xmm9,XMMWORD[32+r12]
|
|
pxor xmm5,xmm8
|
|
movdqu xmm10,XMMWORD[48+r12]
|
|
pxor xmm3,xmm9
|
|
movdqu xmm11,XMMWORD[64+r12]
|
|
pxor xmm1,xmm10
|
|
movdqu xmm12,XMMWORD[80+r12]
|
|
pxor xmm6,xmm11
|
|
movdqu xmm13,XMMWORD[96+r12]
|
|
pxor xmm2,xmm12
|
|
movdqu xmm14,XMMWORD[112+r12]
|
|
pxor xmm4,xmm13
|
|
movdqu XMMWORD[r13],xmm15
|
|
lea r12,[128+r12]
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
movdqu XMMWORD[32+r13],xmm5
|
|
movdqu XMMWORD[48+r13],xmm3
|
|
movdqu XMMWORD[64+r13],xmm1
|
|
movdqu XMMWORD[80+r13],xmm6
|
|
movdqu XMMWORD[96+r13],xmm2
|
|
movdqu XMMWORD[112+r13],xmm4
|
|
lea r13,[128+r13]
|
|
sub r14,8
|
|
jnc NEAR $L$cbc_dec_loop
|
|
|
|
add r14,8
|
|
jz NEAR $L$cbc_dec_done
|
|
|
|
movdqu xmm15,XMMWORD[r12]
|
|
mov rax,rsp
|
|
mov r10d,edx
|
|
cmp r14,2
|
|
jb NEAR $L$cbc_dec_one
|
|
movdqu xmm0,XMMWORD[16+r12]
|
|
je NEAR $L$cbc_dec_two
|
|
movdqu xmm1,XMMWORD[32+r12]
|
|
cmp r14,4
|
|
jb NEAR $L$cbc_dec_three
|
|
movdqu xmm2,XMMWORD[48+r12]
|
|
je NEAR $L$cbc_dec_four
|
|
movdqu xmm3,XMMWORD[64+r12]
|
|
cmp r14,6
|
|
jb NEAR $L$cbc_dec_five
|
|
movdqu xmm4,XMMWORD[80+r12]
|
|
je NEAR $L$cbc_dec_six
|
|
movdqu xmm5,XMMWORD[96+r12]
|
|
movdqa XMMWORD[32+rbp],xmm14
|
|
call _bsaes_decrypt8
|
|
pxor xmm15,XMMWORD[32+rbp]
|
|
movdqu xmm7,XMMWORD[r12]
|
|
movdqu xmm8,XMMWORD[16+r12]
|
|
pxor xmm0,xmm7
|
|
movdqu xmm9,XMMWORD[32+r12]
|
|
pxor xmm5,xmm8
|
|
movdqu xmm10,XMMWORD[48+r12]
|
|
pxor xmm3,xmm9
|
|
movdqu xmm11,XMMWORD[64+r12]
|
|
pxor xmm1,xmm10
|
|
movdqu xmm12,XMMWORD[80+r12]
|
|
pxor xmm6,xmm11
|
|
movdqu xmm14,XMMWORD[96+r12]
|
|
pxor xmm2,xmm12
|
|
movdqu XMMWORD[r13],xmm15
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
movdqu XMMWORD[32+r13],xmm5
|
|
movdqu XMMWORD[48+r13],xmm3
|
|
movdqu XMMWORD[64+r13],xmm1
|
|
movdqu XMMWORD[80+r13],xmm6
|
|
movdqu XMMWORD[96+r13],xmm2
|
|
jmp NEAR $L$cbc_dec_done
|
|
ALIGN 16
|
|
$L$cbc_dec_six:
|
|
movdqa XMMWORD[32+rbp],xmm14
|
|
call _bsaes_decrypt8
|
|
pxor xmm15,XMMWORD[32+rbp]
|
|
movdqu xmm7,XMMWORD[r12]
|
|
movdqu xmm8,XMMWORD[16+r12]
|
|
pxor xmm0,xmm7
|
|
movdqu xmm9,XMMWORD[32+r12]
|
|
pxor xmm5,xmm8
|
|
movdqu xmm10,XMMWORD[48+r12]
|
|
pxor xmm3,xmm9
|
|
movdqu xmm11,XMMWORD[64+r12]
|
|
pxor xmm1,xmm10
|
|
movdqu xmm14,XMMWORD[80+r12]
|
|
pxor xmm6,xmm11
|
|
movdqu XMMWORD[r13],xmm15
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
movdqu XMMWORD[32+r13],xmm5
|
|
movdqu XMMWORD[48+r13],xmm3
|
|
movdqu XMMWORD[64+r13],xmm1
|
|
movdqu XMMWORD[80+r13],xmm6
|
|
jmp NEAR $L$cbc_dec_done
|
|
ALIGN 16
|
|
$L$cbc_dec_five:
|
|
movdqa XMMWORD[32+rbp],xmm14
|
|
call _bsaes_decrypt8
|
|
pxor xmm15,XMMWORD[32+rbp]
|
|
movdqu xmm7,XMMWORD[r12]
|
|
movdqu xmm8,XMMWORD[16+r12]
|
|
pxor xmm0,xmm7
|
|
movdqu xmm9,XMMWORD[32+r12]
|
|
pxor xmm5,xmm8
|
|
movdqu xmm10,XMMWORD[48+r12]
|
|
pxor xmm3,xmm9
|
|
movdqu xmm14,XMMWORD[64+r12]
|
|
pxor xmm1,xmm10
|
|
movdqu XMMWORD[r13],xmm15
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
movdqu XMMWORD[32+r13],xmm5
|
|
movdqu XMMWORD[48+r13],xmm3
|
|
movdqu XMMWORD[64+r13],xmm1
|
|
jmp NEAR $L$cbc_dec_done
|
|
ALIGN 16
|
|
$L$cbc_dec_four:
|
|
movdqa XMMWORD[32+rbp],xmm14
|
|
call _bsaes_decrypt8
|
|
pxor xmm15,XMMWORD[32+rbp]
|
|
movdqu xmm7,XMMWORD[r12]
|
|
movdqu xmm8,XMMWORD[16+r12]
|
|
pxor xmm0,xmm7
|
|
movdqu xmm9,XMMWORD[32+r12]
|
|
pxor xmm5,xmm8
|
|
movdqu xmm14,XMMWORD[48+r12]
|
|
pxor xmm3,xmm9
|
|
movdqu XMMWORD[r13],xmm15
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
movdqu XMMWORD[32+r13],xmm5
|
|
movdqu XMMWORD[48+r13],xmm3
|
|
jmp NEAR $L$cbc_dec_done
|
|
ALIGN 16
|
|
$L$cbc_dec_three:
|
|
movdqa XMMWORD[32+rbp],xmm14
|
|
call _bsaes_decrypt8
|
|
pxor xmm15,XMMWORD[32+rbp]
|
|
movdqu xmm7,XMMWORD[r12]
|
|
movdqu xmm8,XMMWORD[16+r12]
|
|
pxor xmm0,xmm7
|
|
movdqu xmm14,XMMWORD[32+r12]
|
|
pxor xmm5,xmm8
|
|
movdqu XMMWORD[r13],xmm15
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
movdqu XMMWORD[32+r13],xmm5
|
|
jmp NEAR $L$cbc_dec_done
|
|
ALIGN 16
|
|
$L$cbc_dec_two:
|
|
movdqa XMMWORD[32+rbp],xmm14
|
|
call _bsaes_decrypt8
|
|
pxor xmm15,XMMWORD[32+rbp]
|
|
movdqu xmm7,XMMWORD[r12]
|
|
movdqu xmm14,XMMWORD[16+r12]
|
|
pxor xmm0,xmm7
|
|
movdqu XMMWORD[r13],xmm15
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
jmp NEAR $L$cbc_dec_done
|
|
ALIGN 16
|
|
$L$cbc_dec_one:
|
|
lea rcx,[r12]
|
|
lea rdx,[32+rbp]
|
|
lea r8,[r15]
|
|
call asm_AES_decrypt
|
|
pxor xmm14,XMMWORD[32+rbp]
|
|
movdqu XMMWORD[r13],xmm14
|
|
movdqa xmm14,xmm15
|
|
|
|
$L$cbc_dec_done:
|
|
movdqu XMMWORD[rbx],xmm14
|
|
lea rax,[rsp]
|
|
pxor xmm0,xmm0
|
|
$L$cbc_dec_bzero:
|
|
movdqa XMMWORD[rax],xmm0
|
|
movdqa XMMWORD[16+rax],xmm0
|
|
lea rax,[32+rax]
|
|
cmp rbp,rax
|
|
ja NEAR $L$cbc_dec_bzero
|
|
|
|
lea rax,[120+rbp]
|
|
|
|
movaps xmm6,XMMWORD[64+rbp]
|
|
movaps xmm7,XMMWORD[80+rbp]
|
|
movaps xmm8,XMMWORD[96+rbp]
|
|
movaps xmm9,XMMWORD[112+rbp]
|
|
movaps xmm10,XMMWORD[128+rbp]
|
|
movaps xmm11,XMMWORD[144+rbp]
|
|
movaps xmm12,XMMWORD[160+rbp]
|
|
movaps xmm13,XMMWORD[176+rbp]
|
|
movaps xmm14,XMMWORD[192+rbp]
|
|
movaps xmm15,XMMWORD[208+rbp]
|
|
lea rax,[160+rax]
|
|
$L$cbc_dec_tail:
|
|
mov r15,QWORD[((-48))+rax]
|
|
|
|
mov r14,QWORD[((-40))+rax]
|
|
|
|
mov r13,QWORD[((-32))+rax]
|
|
|
|
mov r12,QWORD[((-24))+rax]
|
|
|
|
mov rbx,QWORD[((-16))+rax]
|
|
|
|
mov rbp,QWORD[((-8))+rax]
|
|
|
|
lea rsp,[rax]
|
|
|
|
$L$cbc_dec_epilogue:
|
|
DB 0F3h,0C3h ;repret
|
|
|
|
|
|
|
|
global ossl_bsaes_ctr32_encrypt_blocks
|
|
|
|
ALIGN 16
|
|
ossl_bsaes_ctr32_encrypt_blocks:
|
|
|
|
DB 243,15,30,250
|
|
mov rax,rsp
|
|
$L$ctr_enc_prologue:
|
|
push rbp
|
|
|
|
push rbx
|
|
|
|
push r12
|
|
|
|
push r13
|
|
|
|
push r14
|
|
|
|
push r15
|
|
|
|
lea rsp,[((-72))+rsp]
|
|
|
|
mov r10,QWORD[160+rsp]
|
|
lea rsp,[((-160))+rsp]
|
|
movaps XMMWORD[64+rsp],xmm6
|
|
movaps XMMWORD[80+rsp],xmm7
|
|
movaps XMMWORD[96+rsp],xmm8
|
|
movaps XMMWORD[112+rsp],xmm9
|
|
movaps XMMWORD[128+rsp],xmm10
|
|
movaps XMMWORD[144+rsp],xmm11
|
|
movaps XMMWORD[160+rsp],xmm12
|
|
movaps XMMWORD[176+rsp],xmm13
|
|
movaps XMMWORD[192+rsp],xmm14
|
|
movaps XMMWORD[208+rsp],xmm15
|
|
$L$ctr_enc_body:
|
|
mov rbp,rsp
|
|
|
|
movdqu xmm0,XMMWORD[r10]
|
|
mov eax,DWORD[240+r9]
|
|
mov r12,rcx
|
|
mov r13,rdx
|
|
mov r14,r8
|
|
mov r15,r9
|
|
movdqa XMMWORD[32+rbp],xmm0
|
|
cmp r8,8
|
|
jb NEAR $L$ctr_enc_short
|
|
|
|
mov ebx,eax
|
|
shl rax,7
|
|
sub rax,96
|
|
sub rsp,rax
|
|
|
|
mov rax,rsp
|
|
mov rcx,r15
|
|
mov r10d,ebx
|
|
call _bsaes_key_convert
|
|
pxor xmm7,xmm6
|
|
movdqa XMMWORD[rax],xmm7
|
|
|
|
movdqa xmm8,XMMWORD[rsp]
|
|
lea r11,[$L$ADD1]
|
|
movdqa xmm15,XMMWORD[32+rbp]
|
|
movdqa xmm7,XMMWORD[((-32))+r11]
|
|
DB 102,68,15,56,0,199
|
|
DB 102,68,15,56,0,255
|
|
movdqa XMMWORD[rsp],xmm8
|
|
jmp NEAR $L$ctr_enc_loop
|
|
ALIGN 16
|
|
$L$ctr_enc_loop:
|
|
movdqa XMMWORD[32+rbp],xmm15
|
|
movdqa xmm0,xmm15
|
|
movdqa xmm1,xmm15
|
|
paddd xmm0,XMMWORD[r11]
|
|
movdqa xmm2,xmm15
|
|
paddd xmm1,XMMWORD[16+r11]
|
|
movdqa xmm3,xmm15
|
|
paddd xmm2,XMMWORD[32+r11]
|
|
movdqa xmm4,xmm15
|
|
paddd xmm3,XMMWORD[48+r11]
|
|
movdqa xmm5,xmm15
|
|
paddd xmm4,XMMWORD[64+r11]
|
|
movdqa xmm6,xmm15
|
|
paddd xmm5,XMMWORD[80+r11]
|
|
paddd xmm6,XMMWORD[96+r11]
|
|
|
|
|
|
|
|
movdqa xmm8,XMMWORD[rsp]
|
|
lea rax,[16+rsp]
|
|
movdqa xmm7,XMMWORD[((-16))+r11]
|
|
pxor xmm15,xmm8
|
|
pxor xmm0,xmm8
|
|
pxor xmm1,xmm8
|
|
pxor xmm2,xmm8
|
|
DB 102,68,15,56,0,255
|
|
DB 102,15,56,0,199
|
|
pxor xmm3,xmm8
|
|
pxor xmm4,xmm8
|
|
DB 102,15,56,0,207
|
|
DB 102,15,56,0,215
|
|
pxor xmm5,xmm8
|
|
pxor xmm6,xmm8
|
|
DB 102,15,56,0,223
|
|
DB 102,15,56,0,231
|
|
DB 102,15,56,0,239
|
|
DB 102,15,56,0,247
|
|
lea r11,[$L$BS0]
|
|
mov r10d,ebx
|
|
|
|
call _bsaes_encrypt8_bitslice
|
|
|
|
sub r14,8
|
|
jc NEAR $L$ctr_enc_loop_done
|
|
|
|
movdqu xmm7,XMMWORD[r12]
|
|
movdqu xmm8,XMMWORD[16+r12]
|
|
movdqu xmm9,XMMWORD[32+r12]
|
|
movdqu xmm10,XMMWORD[48+r12]
|
|
movdqu xmm11,XMMWORD[64+r12]
|
|
movdqu xmm12,XMMWORD[80+r12]
|
|
movdqu xmm13,XMMWORD[96+r12]
|
|
movdqu xmm14,XMMWORD[112+r12]
|
|
lea r12,[128+r12]
|
|
pxor xmm7,xmm15
|
|
movdqa xmm15,XMMWORD[32+rbp]
|
|
pxor xmm0,xmm8
|
|
movdqu XMMWORD[r13],xmm7
|
|
pxor xmm3,xmm9
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
pxor xmm5,xmm10
|
|
movdqu XMMWORD[32+r13],xmm3
|
|
pxor xmm2,xmm11
|
|
movdqu XMMWORD[48+r13],xmm5
|
|
pxor xmm6,xmm12
|
|
movdqu XMMWORD[64+r13],xmm2
|
|
pxor xmm1,xmm13
|
|
movdqu XMMWORD[80+r13],xmm6
|
|
pxor xmm4,xmm14
|
|
movdqu XMMWORD[96+r13],xmm1
|
|
lea r11,[$L$ADD1]
|
|
movdqu XMMWORD[112+r13],xmm4
|
|
lea r13,[128+r13]
|
|
paddd xmm15,XMMWORD[112+r11]
|
|
jnz NEAR $L$ctr_enc_loop
|
|
|
|
jmp NEAR $L$ctr_enc_done
|
|
ALIGN 16
|
|
$L$ctr_enc_loop_done:
|
|
add r14,8
|
|
movdqu xmm7,XMMWORD[r12]
|
|
pxor xmm15,xmm7
|
|
movdqu XMMWORD[r13],xmm15
|
|
cmp r14,2
|
|
jb NEAR $L$ctr_enc_done
|
|
movdqu xmm8,XMMWORD[16+r12]
|
|
pxor xmm0,xmm8
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
je NEAR $L$ctr_enc_done
|
|
movdqu xmm9,XMMWORD[32+r12]
|
|
pxor xmm3,xmm9
|
|
movdqu XMMWORD[32+r13],xmm3
|
|
cmp r14,4
|
|
jb NEAR $L$ctr_enc_done
|
|
movdqu xmm10,XMMWORD[48+r12]
|
|
pxor xmm5,xmm10
|
|
movdqu XMMWORD[48+r13],xmm5
|
|
je NEAR $L$ctr_enc_done
|
|
movdqu xmm11,XMMWORD[64+r12]
|
|
pxor xmm2,xmm11
|
|
movdqu XMMWORD[64+r13],xmm2
|
|
cmp r14,6
|
|
jb NEAR $L$ctr_enc_done
|
|
movdqu xmm12,XMMWORD[80+r12]
|
|
pxor xmm6,xmm12
|
|
movdqu XMMWORD[80+r13],xmm6
|
|
je NEAR $L$ctr_enc_done
|
|
movdqu xmm13,XMMWORD[96+r12]
|
|
pxor xmm1,xmm13
|
|
movdqu XMMWORD[96+r13],xmm1
|
|
jmp NEAR $L$ctr_enc_done
|
|
|
|
ALIGN 16
|
|
$L$ctr_enc_short:
|
|
lea rcx,[32+rbp]
|
|
lea rdx,[48+rbp]
|
|
lea r8,[r15]
|
|
call asm_AES_encrypt
|
|
movdqu xmm0,XMMWORD[r12]
|
|
lea r12,[16+r12]
|
|
mov eax,DWORD[44+rbp]
|
|
bswap eax
|
|
pxor xmm0,XMMWORD[48+rbp]
|
|
inc eax
|
|
movdqu XMMWORD[r13],xmm0
|
|
bswap eax
|
|
lea r13,[16+r13]
|
|
mov DWORD[44+rsp],eax
|
|
dec r14
|
|
jnz NEAR $L$ctr_enc_short
|
|
|
|
$L$ctr_enc_done:
|
|
lea rax,[rsp]
|
|
pxor xmm0,xmm0
|
|
$L$ctr_enc_bzero:
|
|
movdqa XMMWORD[rax],xmm0
|
|
movdqa XMMWORD[16+rax],xmm0
|
|
lea rax,[32+rax]
|
|
cmp rbp,rax
|
|
ja NEAR $L$ctr_enc_bzero
|
|
|
|
lea rax,[120+rbp]
|
|
|
|
movaps xmm6,XMMWORD[64+rbp]
|
|
movaps xmm7,XMMWORD[80+rbp]
|
|
movaps xmm8,XMMWORD[96+rbp]
|
|
movaps xmm9,XMMWORD[112+rbp]
|
|
movaps xmm10,XMMWORD[128+rbp]
|
|
movaps xmm11,XMMWORD[144+rbp]
|
|
movaps xmm12,XMMWORD[160+rbp]
|
|
movaps xmm13,XMMWORD[176+rbp]
|
|
movaps xmm14,XMMWORD[192+rbp]
|
|
movaps xmm15,XMMWORD[208+rbp]
|
|
lea rax,[160+rax]
|
|
$L$ctr_enc_tail:
|
|
mov r15,QWORD[((-48))+rax]
|
|
|
|
mov r14,QWORD[((-40))+rax]
|
|
|
|
mov r13,QWORD[((-32))+rax]
|
|
|
|
mov r12,QWORD[((-24))+rax]
|
|
|
|
mov rbx,QWORD[((-16))+rax]
|
|
|
|
mov rbp,QWORD[((-8))+rax]
|
|
|
|
lea rsp,[rax]
|
|
|
|
$L$ctr_enc_epilogue:
|
|
DB 0F3h,0C3h ;repret
|
|
|
|
|
|
global ossl_bsaes_xts_encrypt
|
|
|
|
ALIGN 16
|
|
ossl_bsaes_xts_encrypt:
|
|
|
|
DB 243,15,30,250
|
|
mov rax,rsp
|
|
$L$xts_enc_prologue:
|
|
push rbp
|
|
|
|
push rbx
|
|
|
|
push r12
|
|
|
|
push r13
|
|
|
|
push r14
|
|
|
|
push r15
|
|
|
|
lea rsp,[((-72))+rsp]
|
|
|
|
mov r10,QWORD[160+rsp]
|
|
mov r11,QWORD[168+rsp]
|
|
lea rsp,[((-160))+rsp]
|
|
movaps XMMWORD[64+rsp],xmm6
|
|
movaps XMMWORD[80+rsp],xmm7
|
|
movaps XMMWORD[96+rsp],xmm8
|
|
movaps XMMWORD[112+rsp],xmm9
|
|
movaps XMMWORD[128+rsp],xmm10
|
|
movaps XMMWORD[144+rsp],xmm11
|
|
movaps XMMWORD[160+rsp],xmm12
|
|
movaps XMMWORD[176+rsp],xmm13
|
|
movaps XMMWORD[192+rsp],xmm14
|
|
movaps XMMWORD[208+rsp],xmm15
|
|
$L$xts_enc_body:
|
|
mov rbp,rsp
|
|
|
|
mov r12,rcx
|
|
mov r13,rdx
|
|
mov r14,r8
|
|
mov r15,r9
|
|
|
|
lea rcx,[r11]
|
|
lea rdx,[32+rbp]
|
|
lea r8,[r10]
|
|
call asm_AES_encrypt
|
|
|
|
mov eax,DWORD[240+r15]
|
|
mov rbx,r14
|
|
|
|
mov edx,eax
|
|
shl rax,7
|
|
sub rax,96
|
|
sub rsp,rax
|
|
|
|
mov rax,rsp
|
|
mov rcx,r15
|
|
mov r10d,edx
|
|
call _bsaes_key_convert
|
|
pxor xmm7,xmm6
|
|
movdqa XMMWORD[rax],xmm7
|
|
|
|
and r14,-16
|
|
sub rsp,0x80
|
|
movdqa xmm6,XMMWORD[32+rbp]
|
|
|
|
pxor xmm14,xmm14
|
|
movdqa xmm12,XMMWORD[$L$xts_magic]
|
|
pcmpgtd xmm14,xmm6
|
|
|
|
sub r14,0x80
|
|
jc NEAR $L$xts_enc_short
|
|
jmp NEAR $L$xts_enc_loop
|
|
|
|
ALIGN 16
|
|
$L$xts_enc_loop:
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm15,xmm6
|
|
movdqa XMMWORD[rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm0,xmm6
|
|
movdqa XMMWORD[16+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm7,XMMWORD[r12]
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm1,xmm6
|
|
movdqa XMMWORD[32+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm8,XMMWORD[16+r12]
|
|
pxor xmm15,xmm7
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm2,xmm6
|
|
movdqa XMMWORD[48+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm9,XMMWORD[32+r12]
|
|
pxor xmm0,xmm8
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm3,xmm6
|
|
movdqa XMMWORD[64+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm10,XMMWORD[48+r12]
|
|
pxor xmm1,xmm9
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm4,xmm6
|
|
movdqa XMMWORD[80+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm11,XMMWORD[64+r12]
|
|
pxor xmm2,xmm10
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm5,xmm6
|
|
movdqa XMMWORD[96+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm12,XMMWORD[80+r12]
|
|
pxor xmm3,xmm11
|
|
movdqu xmm13,XMMWORD[96+r12]
|
|
pxor xmm4,xmm12
|
|
movdqu xmm14,XMMWORD[112+r12]
|
|
lea r12,[128+r12]
|
|
movdqa XMMWORD[112+rsp],xmm6
|
|
pxor xmm5,xmm13
|
|
lea rax,[128+rsp]
|
|
pxor xmm6,xmm14
|
|
mov r10d,edx
|
|
|
|
call _bsaes_encrypt8
|
|
|
|
pxor xmm15,XMMWORD[rsp]
|
|
pxor xmm0,XMMWORD[16+rsp]
|
|
movdqu XMMWORD[r13],xmm15
|
|
pxor xmm3,XMMWORD[32+rsp]
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
pxor xmm5,XMMWORD[48+rsp]
|
|
movdqu XMMWORD[32+r13],xmm3
|
|
pxor xmm2,XMMWORD[64+rsp]
|
|
movdqu XMMWORD[48+r13],xmm5
|
|
pxor xmm6,XMMWORD[80+rsp]
|
|
movdqu XMMWORD[64+r13],xmm2
|
|
pxor xmm1,XMMWORD[96+rsp]
|
|
movdqu XMMWORD[80+r13],xmm6
|
|
pxor xmm4,XMMWORD[112+rsp]
|
|
movdqu XMMWORD[96+r13],xmm1
|
|
movdqu XMMWORD[112+r13],xmm4
|
|
lea r13,[128+r13]
|
|
|
|
movdqa xmm6,XMMWORD[112+rsp]
|
|
pxor xmm14,xmm14
|
|
movdqa xmm12,XMMWORD[$L$xts_magic]
|
|
pcmpgtd xmm14,xmm6
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
|
|
sub r14,0x80
|
|
jnc NEAR $L$xts_enc_loop
|
|
|
|
$L$xts_enc_short:
|
|
add r14,0x80
|
|
jz NEAR $L$xts_enc_done
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm15,xmm6
|
|
movdqa XMMWORD[rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm0,xmm6
|
|
movdqa XMMWORD[16+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm7,XMMWORD[r12]
|
|
cmp r14,16
|
|
je NEAR $L$xts_enc_1
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm1,xmm6
|
|
movdqa XMMWORD[32+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm8,XMMWORD[16+r12]
|
|
cmp r14,32
|
|
je NEAR $L$xts_enc_2
|
|
pxor xmm15,xmm7
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm2,xmm6
|
|
movdqa XMMWORD[48+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm9,XMMWORD[32+r12]
|
|
cmp r14,48
|
|
je NEAR $L$xts_enc_3
|
|
pxor xmm0,xmm8
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm3,xmm6
|
|
movdqa XMMWORD[64+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm10,XMMWORD[48+r12]
|
|
cmp r14,64
|
|
je NEAR $L$xts_enc_4
|
|
pxor xmm1,xmm9
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm4,xmm6
|
|
movdqa XMMWORD[80+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm11,XMMWORD[64+r12]
|
|
cmp r14,80
|
|
je NEAR $L$xts_enc_5
|
|
pxor xmm2,xmm10
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm5,xmm6
|
|
movdqa XMMWORD[96+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm12,XMMWORD[80+r12]
|
|
cmp r14,96
|
|
je NEAR $L$xts_enc_6
|
|
pxor xmm3,xmm11
|
|
movdqu xmm13,XMMWORD[96+r12]
|
|
pxor xmm4,xmm12
|
|
movdqa XMMWORD[112+rsp],xmm6
|
|
lea r12,[112+r12]
|
|
pxor xmm5,xmm13
|
|
lea rax,[128+rsp]
|
|
mov r10d,edx
|
|
|
|
call _bsaes_encrypt8
|
|
|
|
pxor xmm15,XMMWORD[rsp]
|
|
pxor xmm0,XMMWORD[16+rsp]
|
|
movdqu XMMWORD[r13],xmm15
|
|
pxor xmm3,XMMWORD[32+rsp]
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
pxor xmm5,XMMWORD[48+rsp]
|
|
movdqu XMMWORD[32+r13],xmm3
|
|
pxor xmm2,XMMWORD[64+rsp]
|
|
movdqu XMMWORD[48+r13],xmm5
|
|
pxor xmm6,XMMWORD[80+rsp]
|
|
movdqu XMMWORD[64+r13],xmm2
|
|
pxor xmm1,XMMWORD[96+rsp]
|
|
movdqu XMMWORD[80+r13],xmm6
|
|
movdqu XMMWORD[96+r13],xmm1
|
|
lea r13,[112+r13]
|
|
|
|
movdqa xmm6,XMMWORD[112+rsp]
|
|
jmp NEAR $L$xts_enc_done
|
|
ALIGN 16
|
|
$L$xts_enc_6:
|
|
pxor xmm3,xmm11
|
|
lea r12,[96+r12]
|
|
pxor xmm4,xmm12
|
|
lea rax,[128+rsp]
|
|
mov r10d,edx
|
|
|
|
call _bsaes_encrypt8
|
|
|
|
pxor xmm15,XMMWORD[rsp]
|
|
pxor xmm0,XMMWORD[16+rsp]
|
|
movdqu XMMWORD[r13],xmm15
|
|
pxor xmm3,XMMWORD[32+rsp]
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
pxor xmm5,XMMWORD[48+rsp]
|
|
movdqu XMMWORD[32+r13],xmm3
|
|
pxor xmm2,XMMWORD[64+rsp]
|
|
movdqu XMMWORD[48+r13],xmm5
|
|
pxor xmm6,XMMWORD[80+rsp]
|
|
movdqu XMMWORD[64+r13],xmm2
|
|
movdqu XMMWORD[80+r13],xmm6
|
|
lea r13,[96+r13]
|
|
|
|
movdqa xmm6,XMMWORD[96+rsp]
|
|
jmp NEAR $L$xts_enc_done
|
|
ALIGN 16
|
|
$L$xts_enc_5:
|
|
pxor xmm2,xmm10
|
|
lea r12,[80+r12]
|
|
pxor xmm3,xmm11
|
|
lea rax,[128+rsp]
|
|
mov r10d,edx
|
|
|
|
call _bsaes_encrypt8
|
|
|
|
pxor xmm15,XMMWORD[rsp]
|
|
pxor xmm0,XMMWORD[16+rsp]
|
|
movdqu XMMWORD[r13],xmm15
|
|
pxor xmm3,XMMWORD[32+rsp]
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
pxor xmm5,XMMWORD[48+rsp]
|
|
movdqu XMMWORD[32+r13],xmm3
|
|
pxor xmm2,XMMWORD[64+rsp]
|
|
movdqu XMMWORD[48+r13],xmm5
|
|
movdqu XMMWORD[64+r13],xmm2
|
|
lea r13,[80+r13]
|
|
|
|
movdqa xmm6,XMMWORD[80+rsp]
|
|
jmp NEAR $L$xts_enc_done
|
|
ALIGN 16
|
|
$L$xts_enc_4:
|
|
pxor xmm1,xmm9
|
|
lea r12,[64+r12]
|
|
pxor xmm2,xmm10
|
|
lea rax,[128+rsp]
|
|
mov r10d,edx
|
|
|
|
call _bsaes_encrypt8
|
|
|
|
pxor xmm15,XMMWORD[rsp]
|
|
pxor xmm0,XMMWORD[16+rsp]
|
|
movdqu XMMWORD[r13],xmm15
|
|
pxor xmm3,XMMWORD[32+rsp]
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
pxor xmm5,XMMWORD[48+rsp]
|
|
movdqu XMMWORD[32+r13],xmm3
|
|
movdqu XMMWORD[48+r13],xmm5
|
|
lea r13,[64+r13]
|
|
|
|
movdqa xmm6,XMMWORD[64+rsp]
|
|
jmp NEAR $L$xts_enc_done
|
|
ALIGN 16
|
|
$L$xts_enc_3:
|
|
pxor xmm0,xmm8
|
|
lea r12,[48+r12]
|
|
pxor xmm1,xmm9
|
|
lea rax,[128+rsp]
|
|
mov r10d,edx
|
|
|
|
call _bsaes_encrypt8
|
|
|
|
pxor xmm15,XMMWORD[rsp]
|
|
pxor xmm0,XMMWORD[16+rsp]
|
|
movdqu XMMWORD[r13],xmm15
|
|
pxor xmm3,XMMWORD[32+rsp]
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
movdqu XMMWORD[32+r13],xmm3
|
|
lea r13,[48+r13]
|
|
|
|
movdqa xmm6,XMMWORD[48+rsp]
|
|
jmp NEAR $L$xts_enc_done
|
|
ALIGN 16
|
|
$L$xts_enc_2:
|
|
pxor xmm15,xmm7
|
|
lea r12,[32+r12]
|
|
pxor xmm0,xmm8
|
|
lea rax,[128+rsp]
|
|
mov r10d,edx
|
|
|
|
call _bsaes_encrypt8
|
|
|
|
pxor xmm15,XMMWORD[rsp]
|
|
pxor xmm0,XMMWORD[16+rsp]
|
|
movdqu XMMWORD[r13],xmm15
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
lea r13,[32+r13]
|
|
|
|
movdqa xmm6,XMMWORD[32+rsp]
|
|
jmp NEAR $L$xts_enc_done
|
|
ALIGN 16
|
|
$L$xts_enc_1:
|
|
pxor xmm7,xmm15
|
|
lea r12,[16+r12]
|
|
movdqa XMMWORD[32+rbp],xmm7
|
|
lea rcx,[32+rbp]
|
|
lea rdx,[32+rbp]
|
|
lea r8,[r15]
|
|
call asm_AES_encrypt
|
|
pxor xmm15,XMMWORD[32+rbp]
|
|
|
|
|
|
|
|
|
|
|
|
movdqu XMMWORD[r13],xmm15
|
|
lea r13,[16+r13]
|
|
|
|
movdqa xmm6,XMMWORD[16+rsp]
|
|
|
|
$L$xts_enc_done:
|
|
and ebx,15
|
|
jz NEAR $L$xts_enc_ret
|
|
mov rdx,r13
|
|
|
|
$L$xts_enc_steal:
|
|
movzx eax,BYTE[r12]
|
|
movzx ecx,BYTE[((-16))+rdx]
|
|
lea r12,[1+r12]
|
|
mov BYTE[((-16))+rdx],al
|
|
mov BYTE[rdx],cl
|
|
lea rdx,[1+rdx]
|
|
sub ebx,1
|
|
jnz NEAR $L$xts_enc_steal
|
|
|
|
movdqu xmm15,XMMWORD[((-16))+r13]
|
|
lea rcx,[32+rbp]
|
|
pxor xmm15,xmm6
|
|
lea rdx,[32+rbp]
|
|
movdqa XMMWORD[32+rbp],xmm15
|
|
lea r8,[r15]
|
|
call asm_AES_encrypt
|
|
pxor xmm6,XMMWORD[32+rbp]
|
|
movdqu XMMWORD[(-16)+r13],xmm6
|
|
|
|
$L$xts_enc_ret:
|
|
lea rax,[rsp]
|
|
pxor xmm0,xmm0
|
|
$L$xts_enc_bzero:
|
|
movdqa XMMWORD[rax],xmm0
|
|
movdqa XMMWORD[16+rax],xmm0
|
|
lea rax,[32+rax]
|
|
cmp rbp,rax
|
|
ja NEAR $L$xts_enc_bzero
|
|
|
|
lea rax,[120+rbp]
|
|
|
|
movaps xmm6,XMMWORD[64+rbp]
|
|
movaps xmm7,XMMWORD[80+rbp]
|
|
movaps xmm8,XMMWORD[96+rbp]
|
|
movaps xmm9,XMMWORD[112+rbp]
|
|
movaps xmm10,XMMWORD[128+rbp]
|
|
movaps xmm11,XMMWORD[144+rbp]
|
|
movaps xmm12,XMMWORD[160+rbp]
|
|
movaps xmm13,XMMWORD[176+rbp]
|
|
movaps xmm14,XMMWORD[192+rbp]
|
|
movaps xmm15,XMMWORD[208+rbp]
|
|
lea rax,[160+rax]
|
|
$L$xts_enc_tail:
|
|
mov r15,QWORD[((-48))+rax]
|
|
|
|
mov r14,QWORD[((-40))+rax]
|
|
|
|
mov r13,QWORD[((-32))+rax]
|
|
|
|
mov r12,QWORD[((-24))+rax]
|
|
|
|
mov rbx,QWORD[((-16))+rax]
|
|
|
|
mov rbp,QWORD[((-8))+rax]
|
|
|
|
lea rsp,[rax]
|
|
|
|
$L$xts_enc_epilogue:
|
|
DB 0F3h,0C3h ;repret
|
|
|
|
|
|
|
|
global ossl_bsaes_xts_decrypt
|
|
|
|
ALIGN 16
|
|
ossl_bsaes_xts_decrypt:
|
|
|
|
DB 243,15,30,250
|
|
mov rax,rsp
|
|
$L$xts_dec_prologue:
|
|
push rbp
|
|
|
|
push rbx
|
|
|
|
push r12
|
|
|
|
push r13
|
|
|
|
push r14
|
|
|
|
push r15
|
|
|
|
lea rsp,[((-72))+rsp]
|
|
|
|
mov r10,QWORD[160+rsp]
|
|
mov r11,QWORD[168+rsp]
|
|
lea rsp,[((-160))+rsp]
|
|
movaps XMMWORD[64+rsp],xmm6
|
|
movaps XMMWORD[80+rsp],xmm7
|
|
movaps XMMWORD[96+rsp],xmm8
|
|
movaps XMMWORD[112+rsp],xmm9
|
|
movaps XMMWORD[128+rsp],xmm10
|
|
movaps XMMWORD[144+rsp],xmm11
|
|
movaps XMMWORD[160+rsp],xmm12
|
|
movaps XMMWORD[176+rsp],xmm13
|
|
movaps XMMWORD[192+rsp],xmm14
|
|
movaps XMMWORD[208+rsp],xmm15
|
|
$L$xts_dec_body:
|
|
mov rbp,rsp
|
|
mov r12,rcx
|
|
mov r13,rdx
|
|
mov r14,r8
|
|
mov r15,r9
|
|
|
|
lea rcx,[r11]
|
|
lea rdx,[32+rbp]
|
|
lea r8,[r10]
|
|
call asm_AES_encrypt
|
|
|
|
mov eax,DWORD[240+r15]
|
|
mov rbx,r14
|
|
|
|
mov edx,eax
|
|
shl rax,7
|
|
sub rax,96
|
|
sub rsp,rax
|
|
|
|
mov rax,rsp
|
|
mov rcx,r15
|
|
mov r10d,edx
|
|
call _bsaes_key_convert
|
|
pxor xmm7,XMMWORD[rsp]
|
|
movdqa XMMWORD[rax],xmm6
|
|
movdqa XMMWORD[rsp],xmm7
|
|
|
|
xor eax,eax
|
|
and r14,-16
|
|
test ebx,15
|
|
setnz al
|
|
shl rax,4
|
|
sub r14,rax
|
|
|
|
sub rsp,0x80
|
|
movdqa xmm6,XMMWORD[32+rbp]
|
|
|
|
pxor xmm14,xmm14
|
|
movdqa xmm12,XMMWORD[$L$xts_magic]
|
|
pcmpgtd xmm14,xmm6
|
|
|
|
sub r14,0x80
|
|
jc NEAR $L$xts_dec_short
|
|
jmp NEAR $L$xts_dec_loop
|
|
|
|
ALIGN 16
|
|
$L$xts_dec_loop:
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm15,xmm6
|
|
movdqa XMMWORD[rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm0,xmm6
|
|
movdqa XMMWORD[16+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm7,XMMWORD[r12]
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm1,xmm6
|
|
movdqa XMMWORD[32+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm8,XMMWORD[16+r12]
|
|
pxor xmm15,xmm7
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm2,xmm6
|
|
movdqa XMMWORD[48+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm9,XMMWORD[32+r12]
|
|
pxor xmm0,xmm8
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm3,xmm6
|
|
movdqa XMMWORD[64+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm10,XMMWORD[48+r12]
|
|
pxor xmm1,xmm9
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm4,xmm6
|
|
movdqa XMMWORD[80+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm11,XMMWORD[64+r12]
|
|
pxor xmm2,xmm10
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm5,xmm6
|
|
movdqa XMMWORD[96+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm12,XMMWORD[80+r12]
|
|
pxor xmm3,xmm11
|
|
movdqu xmm13,XMMWORD[96+r12]
|
|
pxor xmm4,xmm12
|
|
movdqu xmm14,XMMWORD[112+r12]
|
|
lea r12,[128+r12]
|
|
movdqa XMMWORD[112+rsp],xmm6
|
|
pxor xmm5,xmm13
|
|
lea rax,[128+rsp]
|
|
pxor xmm6,xmm14
|
|
mov r10d,edx
|
|
|
|
call _bsaes_decrypt8
|
|
|
|
pxor xmm15,XMMWORD[rsp]
|
|
pxor xmm0,XMMWORD[16+rsp]
|
|
movdqu XMMWORD[r13],xmm15
|
|
pxor xmm5,XMMWORD[32+rsp]
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
pxor xmm3,XMMWORD[48+rsp]
|
|
movdqu XMMWORD[32+r13],xmm5
|
|
pxor xmm1,XMMWORD[64+rsp]
|
|
movdqu XMMWORD[48+r13],xmm3
|
|
pxor xmm6,XMMWORD[80+rsp]
|
|
movdqu XMMWORD[64+r13],xmm1
|
|
pxor xmm2,XMMWORD[96+rsp]
|
|
movdqu XMMWORD[80+r13],xmm6
|
|
pxor xmm4,XMMWORD[112+rsp]
|
|
movdqu XMMWORD[96+r13],xmm2
|
|
movdqu XMMWORD[112+r13],xmm4
|
|
lea r13,[128+r13]
|
|
|
|
movdqa xmm6,XMMWORD[112+rsp]
|
|
pxor xmm14,xmm14
|
|
movdqa xmm12,XMMWORD[$L$xts_magic]
|
|
pcmpgtd xmm14,xmm6
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
|
|
sub r14,0x80
|
|
jnc NEAR $L$xts_dec_loop
|
|
|
|
$L$xts_dec_short:
|
|
add r14,0x80
|
|
jz NEAR $L$xts_dec_done
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm15,xmm6
|
|
movdqa XMMWORD[rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm0,xmm6
|
|
movdqa XMMWORD[16+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm7,XMMWORD[r12]
|
|
cmp r14,16
|
|
je NEAR $L$xts_dec_1
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm1,xmm6
|
|
movdqa XMMWORD[32+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm8,XMMWORD[16+r12]
|
|
cmp r14,32
|
|
je NEAR $L$xts_dec_2
|
|
pxor xmm15,xmm7
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm2,xmm6
|
|
movdqa XMMWORD[48+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm9,XMMWORD[32+r12]
|
|
cmp r14,48
|
|
je NEAR $L$xts_dec_3
|
|
pxor xmm0,xmm8
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm3,xmm6
|
|
movdqa XMMWORD[64+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm10,XMMWORD[48+r12]
|
|
cmp r14,64
|
|
je NEAR $L$xts_dec_4
|
|
pxor xmm1,xmm9
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm4,xmm6
|
|
movdqa XMMWORD[80+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm11,XMMWORD[64+r12]
|
|
cmp r14,80
|
|
je NEAR $L$xts_dec_5
|
|
pxor xmm2,xmm10
|
|
pshufd xmm13,xmm14,0x13
|
|
pxor xmm14,xmm14
|
|
movdqa xmm5,xmm6
|
|
movdqa XMMWORD[96+rsp],xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
pcmpgtd xmm14,xmm6
|
|
pxor xmm6,xmm13
|
|
movdqu xmm12,XMMWORD[80+r12]
|
|
cmp r14,96
|
|
je NEAR $L$xts_dec_6
|
|
pxor xmm3,xmm11
|
|
movdqu xmm13,XMMWORD[96+r12]
|
|
pxor xmm4,xmm12
|
|
movdqa XMMWORD[112+rsp],xmm6
|
|
lea r12,[112+r12]
|
|
pxor xmm5,xmm13
|
|
lea rax,[128+rsp]
|
|
mov r10d,edx
|
|
|
|
call _bsaes_decrypt8
|
|
|
|
pxor xmm15,XMMWORD[rsp]
|
|
pxor xmm0,XMMWORD[16+rsp]
|
|
movdqu XMMWORD[r13],xmm15
|
|
pxor xmm5,XMMWORD[32+rsp]
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
pxor xmm3,XMMWORD[48+rsp]
|
|
movdqu XMMWORD[32+r13],xmm5
|
|
pxor xmm1,XMMWORD[64+rsp]
|
|
movdqu XMMWORD[48+r13],xmm3
|
|
pxor xmm6,XMMWORD[80+rsp]
|
|
movdqu XMMWORD[64+r13],xmm1
|
|
pxor xmm2,XMMWORD[96+rsp]
|
|
movdqu XMMWORD[80+r13],xmm6
|
|
movdqu XMMWORD[96+r13],xmm2
|
|
lea r13,[112+r13]
|
|
|
|
movdqa xmm6,XMMWORD[112+rsp]
|
|
jmp NEAR $L$xts_dec_done
|
|
ALIGN 16
|
|
$L$xts_dec_6:
|
|
pxor xmm3,xmm11
|
|
lea r12,[96+r12]
|
|
pxor xmm4,xmm12
|
|
lea rax,[128+rsp]
|
|
mov r10d,edx
|
|
|
|
call _bsaes_decrypt8
|
|
|
|
pxor xmm15,XMMWORD[rsp]
|
|
pxor xmm0,XMMWORD[16+rsp]
|
|
movdqu XMMWORD[r13],xmm15
|
|
pxor xmm5,XMMWORD[32+rsp]
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
pxor xmm3,XMMWORD[48+rsp]
|
|
movdqu XMMWORD[32+r13],xmm5
|
|
pxor xmm1,XMMWORD[64+rsp]
|
|
movdqu XMMWORD[48+r13],xmm3
|
|
pxor xmm6,XMMWORD[80+rsp]
|
|
movdqu XMMWORD[64+r13],xmm1
|
|
movdqu XMMWORD[80+r13],xmm6
|
|
lea r13,[96+r13]
|
|
|
|
movdqa xmm6,XMMWORD[96+rsp]
|
|
jmp NEAR $L$xts_dec_done
|
|
ALIGN 16
|
|
$L$xts_dec_5:
|
|
pxor xmm2,xmm10
|
|
lea r12,[80+r12]
|
|
pxor xmm3,xmm11
|
|
lea rax,[128+rsp]
|
|
mov r10d,edx
|
|
|
|
call _bsaes_decrypt8
|
|
|
|
pxor xmm15,XMMWORD[rsp]
|
|
pxor xmm0,XMMWORD[16+rsp]
|
|
movdqu XMMWORD[r13],xmm15
|
|
pxor xmm5,XMMWORD[32+rsp]
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
pxor xmm3,XMMWORD[48+rsp]
|
|
movdqu XMMWORD[32+r13],xmm5
|
|
pxor xmm1,XMMWORD[64+rsp]
|
|
movdqu XMMWORD[48+r13],xmm3
|
|
movdqu XMMWORD[64+r13],xmm1
|
|
lea r13,[80+r13]
|
|
|
|
movdqa xmm6,XMMWORD[80+rsp]
|
|
jmp NEAR $L$xts_dec_done
|
|
ALIGN 16
|
|
$L$xts_dec_4:
|
|
pxor xmm1,xmm9
|
|
lea r12,[64+r12]
|
|
pxor xmm2,xmm10
|
|
lea rax,[128+rsp]
|
|
mov r10d,edx
|
|
|
|
call _bsaes_decrypt8
|
|
|
|
pxor xmm15,XMMWORD[rsp]
|
|
pxor xmm0,XMMWORD[16+rsp]
|
|
movdqu XMMWORD[r13],xmm15
|
|
pxor xmm5,XMMWORD[32+rsp]
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
pxor xmm3,XMMWORD[48+rsp]
|
|
movdqu XMMWORD[32+r13],xmm5
|
|
movdqu XMMWORD[48+r13],xmm3
|
|
lea r13,[64+r13]
|
|
|
|
movdqa xmm6,XMMWORD[64+rsp]
|
|
jmp NEAR $L$xts_dec_done
|
|
ALIGN 16
|
|
$L$xts_dec_3:
|
|
pxor xmm0,xmm8
|
|
lea r12,[48+r12]
|
|
pxor xmm1,xmm9
|
|
lea rax,[128+rsp]
|
|
mov r10d,edx
|
|
|
|
call _bsaes_decrypt8
|
|
|
|
pxor xmm15,XMMWORD[rsp]
|
|
pxor xmm0,XMMWORD[16+rsp]
|
|
movdqu XMMWORD[r13],xmm15
|
|
pxor xmm5,XMMWORD[32+rsp]
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
movdqu XMMWORD[32+r13],xmm5
|
|
lea r13,[48+r13]
|
|
|
|
movdqa xmm6,XMMWORD[48+rsp]
|
|
jmp NEAR $L$xts_dec_done
|
|
ALIGN 16
|
|
$L$xts_dec_2:
|
|
pxor xmm15,xmm7
|
|
lea r12,[32+r12]
|
|
pxor xmm0,xmm8
|
|
lea rax,[128+rsp]
|
|
mov r10d,edx
|
|
|
|
call _bsaes_decrypt8
|
|
|
|
pxor xmm15,XMMWORD[rsp]
|
|
pxor xmm0,XMMWORD[16+rsp]
|
|
movdqu XMMWORD[r13],xmm15
|
|
movdqu XMMWORD[16+r13],xmm0
|
|
lea r13,[32+r13]
|
|
|
|
movdqa xmm6,XMMWORD[32+rsp]
|
|
jmp NEAR $L$xts_dec_done
|
|
ALIGN 16
|
|
$L$xts_dec_1:
|
|
pxor xmm7,xmm15
|
|
lea r12,[16+r12]
|
|
movdqa XMMWORD[32+rbp],xmm7
|
|
lea rcx,[32+rbp]
|
|
lea rdx,[32+rbp]
|
|
lea r8,[r15]
|
|
call asm_AES_decrypt
|
|
pxor xmm15,XMMWORD[32+rbp]
|
|
|
|
|
|
|
|
|
|
|
|
movdqu XMMWORD[r13],xmm15
|
|
lea r13,[16+r13]
|
|
|
|
movdqa xmm6,XMMWORD[16+rsp]
|
|
|
|
$L$xts_dec_done:
|
|
and ebx,15
|
|
jz NEAR $L$xts_dec_ret
|
|
|
|
pxor xmm14,xmm14
|
|
movdqa xmm12,XMMWORD[$L$xts_magic]
|
|
pcmpgtd xmm14,xmm6
|
|
pshufd xmm13,xmm14,0x13
|
|
movdqa xmm5,xmm6
|
|
paddq xmm6,xmm6
|
|
pand xmm13,xmm12
|
|
movdqu xmm15,XMMWORD[r12]
|
|
pxor xmm6,xmm13
|
|
|
|
lea rcx,[32+rbp]
|
|
pxor xmm15,xmm6
|
|
lea rdx,[32+rbp]
|
|
movdqa XMMWORD[32+rbp],xmm15
|
|
lea r8,[r15]
|
|
call asm_AES_decrypt
|
|
pxor xmm6,XMMWORD[32+rbp]
|
|
mov rdx,r13
|
|
movdqu XMMWORD[r13],xmm6
|
|
|
|
$L$xts_dec_steal:
|
|
movzx eax,BYTE[16+r12]
|
|
movzx ecx,BYTE[rdx]
|
|
lea r12,[1+r12]
|
|
mov BYTE[rdx],al
|
|
mov BYTE[16+rdx],cl
|
|
lea rdx,[1+rdx]
|
|
sub ebx,1
|
|
jnz NEAR $L$xts_dec_steal
|
|
|
|
movdqu xmm15,XMMWORD[r13]
|
|
lea rcx,[32+rbp]
|
|
pxor xmm15,xmm5
|
|
lea rdx,[32+rbp]
|
|
movdqa XMMWORD[32+rbp],xmm15
|
|
lea r8,[r15]
|
|
call asm_AES_decrypt
|
|
pxor xmm5,XMMWORD[32+rbp]
|
|
movdqu XMMWORD[r13],xmm5
|
|
|
|
$L$xts_dec_ret:
|
|
lea rax,[rsp]
|
|
pxor xmm0,xmm0
|
|
$L$xts_dec_bzero:
|
|
movdqa XMMWORD[rax],xmm0
|
|
movdqa XMMWORD[16+rax],xmm0
|
|
lea rax,[32+rax]
|
|
cmp rbp,rax
|
|
ja NEAR $L$xts_dec_bzero
|
|
|
|
lea rax,[120+rbp]
|
|
|
|
movaps xmm6,XMMWORD[64+rbp]
|
|
movaps xmm7,XMMWORD[80+rbp]
|
|
movaps xmm8,XMMWORD[96+rbp]
|
|
movaps xmm9,XMMWORD[112+rbp]
|
|
movaps xmm10,XMMWORD[128+rbp]
|
|
movaps xmm11,XMMWORD[144+rbp]
|
|
movaps xmm12,XMMWORD[160+rbp]
|
|
movaps xmm13,XMMWORD[176+rbp]
|
|
movaps xmm14,XMMWORD[192+rbp]
|
|
movaps xmm15,XMMWORD[208+rbp]
|
|
lea rax,[160+rax]
|
|
$L$xts_dec_tail:
|
|
mov r15,QWORD[((-48))+rax]
|
|
|
|
mov r14,QWORD[((-40))+rax]
|
|
|
|
mov r13,QWORD[((-32))+rax]
|
|
|
|
mov r12,QWORD[((-24))+rax]
|
|
|
|
mov rbx,QWORD[((-16))+rax]
|
|
|
|
mov rbp,QWORD[((-8))+rax]
|
|
|
|
lea rsp,[rax]
|
|
|
|
$L$xts_dec_epilogue:
|
|
DB 0F3h,0C3h ;repret
|
|
|
|
|
|
|
|
section .rdata rdata align=64
|
|
ALIGN 64
|
|
_bsaes_const:
|
|
$L$M0ISR:
|
|
DQ 0x0a0e0206070b0f03,0x0004080c0d010509
|
|
$L$ISRM0:
|
|
DQ 0x01040b0e0205080f,0x0306090c00070a0d
|
|
$L$ISR:
|
|
DQ 0x0504070602010003,0x0f0e0d0c080b0a09
|
|
$L$BS0:
|
|
DQ 0x5555555555555555,0x5555555555555555
|
|
$L$BS1:
|
|
DQ 0x3333333333333333,0x3333333333333333
|
|
$L$BS2:
|
|
DQ 0x0f0f0f0f0f0f0f0f,0x0f0f0f0f0f0f0f0f
|
|
$L$SR:
|
|
DQ 0x0504070600030201,0x0f0e0d0c0a09080b
|
|
$L$SRM0:
|
|
DQ 0x0304090e00050a0f,0x01060b0c0207080d
|
|
$L$M0SR:
|
|
DQ 0x0a0e02060f03070b,0x0004080c05090d01
|
|
$L$SWPUP:
|
|
DQ 0x0706050403020100,0x0c0d0e0f0b0a0908
|
|
$L$SWPUPM0SR:
|
|
DQ 0x0a0d02060c03070b,0x0004080f05090e01
|
|
$L$ADD1:
|
|
DQ 0x0000000000000000,0x0000000100000000
|
|
$L$ADD2:
|
|
DQ 0x0000000000000000,0x0000000200000000
|
|
$L$ADD3:
|
|
DQ 0x0000000000000000,0x0000000300000000
|
|
$L$ADD4:
|
|
DQ 0x0000000000000000,0x0000000400000000
|
|
$L$ADD5:
|
|
DQ 0x0000000000000000,0x0000000500000000
|
|
$L$ADD6:
|
|
DQ 0x0000000000000000,0x0000000600000000
|
|
$L$ADD7:
|
|
DQ 0x0000000000000000,0x0000000700000000
|
|
$L$ADD8:
|
|
DQ 0x0000000000000000,0x0000000800000000
|
|
$L$xts_magic:
|
|
DD 0x87,0,1,0
|
|
$L$masks:
|
|
DQ 0x0101010101010101,0x0101010101010101
|
|
DQ 0x0202020202020202,0x0202020202020202
|
|
DQ 0x0404040404040404,0x0404040404040404
|
|
DQ 0x0808080808080808,0x0808080808080808
|
|
$L$M0:
|
|
DQ 0x02060a0e03070b0f,0x0004080c0105090d
|
|
$L$63:
|
|
DQ 0x6363636363636363,0x6363636363636363
|
|
ALIGN 64
|
|
|
|
DB 66,105,116,45,115,108,105,99,101,100,32,65,69,83,32,102
|
|
DB 111,114,32,120,56,54,95,54,52,47,83,83,83,69,51,44
|
|
DB 32,69,109,105,108,105,97,32,75,195,164,115,112,101,114,44
|
|
DB 32,80,101,116,101,114,32,83,99,104,119,97,98,101,44,32
|
|
DB 65,110,100,121,32,80,111,108,121,97,107,111,118,0
|
|
EXTERN __imp_RtlVirtualUnwind
|
|
|
|
ALIGN 16
|
|
se_handler:
|
|
push rsi
|
|
push rdi
|
|
push rbx
|
|
push rbp
|
|
push r12
|
|
push r13
|
|
push r14
|
|
push r15
|
|
pushfq
|
|
sub rsp,64
|
|
|
|
mov rax,QWORD[120+r8]
|
|
mov rbx,QWORD[248+r8]
|
|
|
|
mov rsi,QWORD[8+r9]
|
|
mov r11,QWORD[56+r9]
|
|
|
|
mov r10d,DWORD[r11]
|
|
lea r10,[r10*1+rsi]
|
|
cmp rbx,r10
|
|
jbe NEAR $L$in_prologue
|
|
|
|
mov r10d,DWORD[4+r11]
|
|
lea r10,[r10*1+rsi]
|
|
cmp rbx,r10
|
|
jae NEAR $L$in_prologue
|
|
|
|
mov r10d,DWORD[8+r11]
|
|
lea r10,[r10*1+rsi]
|
|
cmp rbx,r10
|
|
jae NEAR $L$in_tail
|
|
|
|
mov rax,QWORD[160+r8]
|
|
|
|
lea rsi,[64+rax]
|
|
lea rdi,[512+r8]
|
|
mov ecx,20
|
|
DD 0xa548f3fc
|
|
lea rax,[((160+120))+rax]
|
|
|
|
$L$in_tail:
|
|
mov rbp,QWORD[((-48))+rax]
|
|
mov rbx,QWORD[((-40))+rax]
|
|
mov r12,QWORD[((-32))+rax]
|
|
mov r13,QWORD[((-24))+rax]
|
|
mov r14,QWORD[((-16))+rax]
|
|
mov r15,QWORD[((-8))+rax]
|
|
mov QWORD[144+r8],rbx
|
|
mov QWORD[160+r8],rbp
|
|
mov QWORD[216+r8],r12
|
|
mov QWORD[224+r8],r13
|
|
mov QWORD[232+r8],r14
|
|
mov QWORD[240+r8],r15
|
|
|
|
$L$in_prologue:
|
|
mov QWORD[152+r8],rax
|
|
|
|
mov rdi,QWORD[40+r9]
|
|
mov rsi,r8
|
|
mov ecx,154
|
|
DD 0xa548f3fc
|
|
|
|
mov rsi,r9
|
|
xor rcx,rcx
|
|
mov rdx,QWORD[8+rsi]
|
|
mov r8,QWORD[rsi]
|
|
mov r9,QWORD[16+rsi]
|
|
mov r10,QWORD[40+rsi]
|
|
lea r11,[56+rsi]
|
|
lea r12,[24+rsi]
|
|
mov QWORD[32+rsp],r10
|
|
mov QWORD[40+rsp],r11
|
|
mov QWORD[48+rsp],r12
|
|
mov QWORD[56+rsp],rcx
|
|
call QWORD[__imp_RtlVirtualUnwind]
|
|
|
|
mov eax,1
|
|
add rsp,64
|
|
popfq
|
|
pop r15
|
|
pop r14
|
|
pop r13
|
|
pop r12
|
|
pop rbp
|
|
pop rbx
|
|
pop rdi
|
|
pop rsi
|
|
DB 0F3h,0C3h ;repret
|
|
|
|
|
|
section .pdata rdata align=4
|
|
ALIGN 4
|
|
DD $L$cbc_dec_prologue wrt ..imagebase
|
|
DD $L$cbc_dec_epilogue wrt ..imagebase
|
|
DD $L$cbc_dec_info wrt ..imagebase
|
|
|
|
DD $L$ctr_enc_prologue wrt ..imagebase
|
|
DD $L$ctr_enc_epilogue wrt ..imagebase
|
|
DD $L$ctr_enc_info wrt ..imagebase
|
|
|
|
DD $L$xts_enc_prologue wrt ..imagebase
|
|
DD $L$xts_enc_epilogue wrt ..imagebase
|
|
DD $L$xts_enc_info wrt ..imagebase
|
|
|
|
DD $L$xts_dec_prologue wrt ..imagebase
|
|
DD $L$xts_dec_epilogue wrt ..imagebase
|
|
DD $L$xts_dec_info wrt ..imagebase
|
|
|
|
section .xdata rdata align=8
|
|
ALIGN 8
|
|
$L$cbc_dec_info:
|
|
DB 9,0,0,0
|
|
DD se_handler wrt ..imagebase
|
|
DD $L$cbc_dec_body wrt ..imagebase,$L$cbc_dec_epilogue wrt ..imagebase
|
|
DD $L$cbc_dec_tail wrt ..imagebase
|
|
DD 0
|
|
$L$ctr_enc_info:
|
|
DB 9,0,0,0
|
|
DD se_handler wrt ..imagebase
|
|
DD $L$ctr_enc_body wrt ..imagebase,$L$ctr_enc_epilogue wrt ..imagebase
|
|
DD $L$ctr_enc_tail wrt ..imagebase
|
|
DD 0
|
|
$L$xts_enc_info:
|
|
DB 9,0,0,0
|
|
DD se_handler wrt ..imagebase
|
|
DD $L$xts_enc_body wrt ..imagebase,$L$xts_enc_epilogue wrt ..imagebase
|
|
DD $L$xts_enc_tail wrt ..imagebase
|
|
DD 0
|
|
$L$xts_dec_info:
|
|
DB 9,0,0,0
|
|
DD se_handler wrt ..imagebase
|
|
DD $L$xts_dec_body wrt ..imagebase,$L$xts_dec_epilogue wrt ..imagebase
|
|
DD $L$xts_dec_tail wrt ..imagebase
|
|
DD 0
|