| ; This file is generated from a similarly-named Perl script in the BoringSSL |
| ; source tree. Do not edit by hand. |
| |
| default rel |
| %define XMMWORD |
| %define YMMWORD |
| %define ZMMWORD |
| |
| %ifdef BORINGSSL_PREFIX |
| %include "boringssl_prefix_symbols_nasm.inc" |
| %endif |
| section .text code align=64 |
| |
| EXTERN OPENSSL_ia32cap_P |
| global gcm_init_clmul |
| |
| ALIGN 16 |
| gcm_init_clmul: |
| |
| $L$_init_clmul: |
| $L$SEH_begin_gcm_init_clmul: |
| |
| DB 0x48,0x83,0xec,0x18 |
| DB 0x0f,0x29,0x34,0x24 |
| movdqu xmm2,XMMWORD[rdx] |
| pshufd xmm2,xmm2,78 |
| |
| |
| pshufd xmm4,xmm2,255 |
| movdqa xmm3,xmm2 |
| psllq xmm2,1 |
| pxor xmm5,xmm5 |
| psrlq xmm3,63 |
| pcmpgtd xmm5,xmm4 |
| pslldq xmm3,8 |
| por xmm2,xmm3 |
| |
| |
| pand xmm5,XMMWORD[$L$0x1c2_polynomial] |
| pxor xmm2,xmm5 |
| |
| |
| pshufd xmm6,xmm2,78 |
| movdqa xmm0,xmm2 |
| pxor xmm6,xmm2 |
| movdqa xmm1,xmm0 |
| pshufd xmm3,xmm0,78 |
| pxor xmm3,xmm0 |
| DB 102,15,58,68,194,0 |
| DB 102,15,58,68,202,17 |
| DB 102,15,58,68,222,0 |
| pxor xmm3,xmm0 |
| pxor xmm3,xmm1 |
| |
| movdqa xmm4,xmm3 |
| psrldq xmm3,8 |
| pslldq xmm4,8 |
| pxor xmm1,xmm3 |
| pxor xmm0,xmm4 |
| |
| movdqa xmm4,xmm0 |
| movdqa xmm3,xmm0 |
| psllq xmm0,5 |
| pxor xmm3,xmm0 |
| psllq xmm0,1 |
| pxor xmm0,xmm3 |
| psllq xmm0,57 |
| movdqa xmm3,xmm0 |
| pslldq xmm0,8 |
| psrldq xmm3,8 |
| pxor xmm0,xmm4 |
| pxor xmm1,xmm3 |
| |
| |
| movdqa xmm4,xmm0 |
| psrlq xmm0,1 |
| pxor xmm1,xmm4 |
| pxor xmm4,xmm0 |
| psrlq xmm0,5 |
| pxor xmm0,xmm4 |
| psrlq xmm0,1 |
| pxor xmm0,xmm1 |
| pshufd xmm3,xmm2,78 |
| pshufd xmm4,xmm0,78 |
| pxor xmm3,xmm2 |
| movdqu XMMWORD[rcx],xmm2 |
| pxor xmm4,xmm0 |
| movdqu XMMWORD[16+rcx],xmm0 |
| DB 102,15,58,15,227,8 |
| movdqu XMMWORD[32+rcx],xmm4 |
| movdqa xmm1,xmm0 |
| pshufd xmm3,xmm0,78 |
| pxor xmm3,xmm0 |
| DB 102,15,58,68,194,0 |
| DB 102,15,58,68,202,17 |
| DB 102,15,58,68,222,0 |
| pxor xmm3,xmm0 |
| pxor xmm3,xmm1 |
| |
| movdqa xmm4,xmm3 |
| psrldq xmm3,8 |
| pslldq xmm4,8 |
| pxor xmm1,xmm3 |
| pxor xmm0,xmm4 |
| |
| movdqa xmm4,xmm0 |
| movdqa xmm3,xmm0 |
| psllq xmm0,5 |
| pxor xmm3,xmm0 |
| psllq xmm0,1 |
| pxor xmm0,xmm3 |
| psllq xmm0,57 |
| movdqa xmm3,xmm0 |
| pslldq xmm0,8 |
| psrldq xmm3,8 |
| pxor xmm0,xmm4 |
| pxor xmm1,xmm3 |
| |
| |
| movdqa xmm4,xmm0 |
| psrlq xmm0,1 |
| pxor xmm1,xmm4 |
| pxor xmm4,xmm0 |
| psrlq xmm0,5 |
| pxor xmm0,xmm4 |
| psrlq xmm0,1 |
| pxor xmm0,xmm1 |
| movdqa xmm5,xmm0 |
| movdqa xmm1,xmm0 |
| pshufd xmm3,xmm0,78 |
| pxor xmm3,xmm0 |
| DB 102,15,58,68,194,0 |
| DB 102,15,58,68,202,17 |
| DB 102,15,58,68,222,0 |
| pxor xmm3,xmm0 |
| pxor xmm3,xmm1 |
| |
| movdqa xmm4,xmm3 |
| psrldq xmm3,8 |
| pslldq xmm4,8 |
| pxor xmm1,xmm3 |
| pxor xmm0,xmm4 |
| |
| movdqa xmm4,xmm0 |
| movdqa xmm3,xmm0 |
| psllq xmm0,5 |
| pxor xmm3,xmm0 |
| psllq xmm0,1 |
| pxor xmm0,xmm3 |
| psllq xmm0,57 |
| movdqa xmm3,xmm0 |
| pslldq xmm0,8 |
| psrldq xmm3,8 |
| pxor xmm0,xmm4 |
| pxor xmm1,xmm3 |
| |
| |
| movdqa xmm4,xmm0 |
| psrlq xmm0,1 |
| pxor xmm1,xmm4 |
| pxor xmm4,xmm0 |
| psrlq xmm0,5 |
| pxor xmm0,xmm4 |
| psrlq xmm0,1 |
| pxor xmm0,xmm1 |
| pshufd xmm3,xmm5,78 |
| pshufd xmm4,xmm0,78 |
| pxor xmm3,xmm5 |
| movdqu XMMWORD[48+rcx],xmm5 |
| pxor xmm4,xmm0 |
| movdqu XMMWORD[64+rcx],xmm0 |
| DB 102,15,58,15,227,8 |
| movdqu XMMWORD[80+rcx],xmm4 |
| movaps xmm6,XMMWORD[rsp] |
| lea rsp,[24+rsp] |
| $L$SEH_end_gcm_init_clmul: |
| DB 0F3h,0C3h ;repret |
| |
| |
| global gcm_gmult_clmul |
| |
| ALIGN 16 |
| gcm_gmult_clmul: |
| |
| $L$_gmult_clmul: |
| movdqu xmm0,XMMWORD[rcx] |
| movdqa xmm5,XMMWORD[$L$bswap_mask] |
| movdqu xmm2,XMMWORD[rdx] |
| movdqu xmm4,XMMWORD[32+rdx] |
| DB 102,15,56,0,197 |
| movdqa xmm1,xmm0 |
| pshufd xmm3,xmm0,78 |
| pxor xmm3,xmm0 |
| DB 102,15,58,68,194,0 |
| DB 102,15,58,68,202,17 |
| DB 102,15,58,68,220,0 |
| pxor xmm3,xmm0 |
| pxor xmm3,xmm1 |
| |
| movdqa xmm4,xmm3 |
| psrldq xmm3,8 |
| pslldq xmm4,8 |
| pxor xmm1,xmm3 |
| pxor xmm0,xmm4 |
| |
| movdqa xmm4,xmm0 |
| movdqa xmm3,xmm0 |
| psllq xmm0,5 |
| pxor xmm3,xmm0 |
| psllq xmm0,1 |
| pxor xmm0,xmm3 |
| psllq xmm0,57 |
| movdqa xmm3,xmm0 |
| pslldq xmm0,8 |
| psrldq xmm3,8 |
| pxor xmm0,xmm4 |
| pxor xmm1,xmm3 |
| |
| |
| movdqa xmm4,xmm0 |
| psrlq xmm0,1 |
| pxor xmm1,xmm4 |
| pxor xmm4,xmm0 |
| psrlq xmm0,5 |
| pxor xmm0,xmm4 |
| psrlq xmm0,1 |
| pxor xmm0,xmm1 |
| DB 102,15,56,0,197 |
| movdqu XMMWORD[rcx],xmm0 |
| DB 0F3h,0C3h ;repret |
| |
| |
| global gcm_ghash_clmul |
| |
| ALIGN 32 |
| gcm_ghash_clmul: |
| |
| $L$_ghash_clmul: |
| lea rax,[((-136))+rsp] |
| $L$SEH_begin_gcm_ghash_clmul: |
| |
| DB 0x48,0x8d,0x60,0xe0 |
| DB 0x0f,0x29,0x70,0xe0 |
| DB 0x0f,0x29,0x78,0xf0 |
| DB 0x44,0x0f,0x29,0x00 |
| DB 0x44,0x0f,0x29,0x48,0x10 |
| DB 0x44,0x0f,0x29,0x50,0x20 |
| DB 0x44,0x0f,0x29,0x58,0x30 |
| DB 0x44,0x0f,0x29,0x60,0x40 |
| DB 0x44,0x0f,0x29,0x68,0x50 |
| DB 0x44,0x0f,0x29,0x70,0x60 |
| DB 0x44,0x0f,0x29,0x78,0x70 |
| movdqa xmm10,XMMWORD[$L$bswap_mask] |
| |
| movdqu xmm0,XMMWORD[rcx] |
| movdqu xmm2,XMMWORD[rdx] |
| movdqu xmm7,XMMWORD[32+rdx] |
| DB 102,65,15,56,0,194 |
| |
| sub r9,0x10 |
| jz NEAR $L$odd_tail |
| |
| movdqu xmm6,XMMWORD[16+rdx] |
| lea rax,[OPENSSL_ia32cap_P] |
| mov eax,DWORD[4+rax] |
| cmp r9,0x30 |
| jb NEAR $L$skip4x |
| |
| and eax,71303168 |
| cmp eax,4194304 |
| je NEAR $L$skip4x |
| |
| sub r9,0x30 |
| mov rax,0xA040608020C0E000 |
| movdqu xmm14,XMMWORD[48+rdx] |
| movdqu xmm15,XMMWORD[64+rdx] |
| |
| |
| |
| |
| movdqu xmm3,XMMWORD[48+r8] |
| movdqu xmm11,XMMWORD[32+r8] |
| DB 102,65,15,56,0,218 |
| DB 102,69,15,56,0,218 |
| movdqa xmm5,xmm3 |
| pshufd xmm4,xmm3,78 |
| pxor xmm4,xmm3 |
| DB 102,15,58,68,218,0 |
| DB 102,15,58,68,234,17 |
| DB 102,15,58,68,231,0 |
| |
| movdqa xmm13,xmm11 |
| pshufd xmm12,xmm11,78 |
| pxor xmm12,xmm11 |
| DB 102,68,15,58,68,222,0 |
| DB 102,68,15,58,68,238,17 |
| DB 102,68,15,58,68,231,16 |
| xorps xmm3,xmm11 |
| xorps xmm5,xmm13 |
| movups xmm7,XMMWORD[80+rdx] |
| xorps xmm4,xmm12 |
| |
| movdqu xmm11,XMMWORD[16+r8] |
| movdqu xmm8,XMMWORD[r8] |
| DB 102,69,15,56,0,218 |
| DB 102,69,15,56,0,194 |
| movdqa xmm13,xmm11 |
| pshufd xmm12,xmm11,78 |
| pxor xmm0,xmm8 |
| pxor xmm12,xmm11 |
| DB 102,69,15,58,68,222,0 |
| movdqa xmm1,xmm0 |
| pshufd xmm8,xmm0,78 |
| pxor xmm8,xmm0 |
| DB 102,69,15,58,68,238,17 |
| DB 102,68,15,58,68,231,0 |
| xorps xmm3,xmm11 |
| xorps xmm5,xmm13 |
| |
| lea r8,[64+r8] |
| sub r9,0x40 |
| jc NEAR $L$tail4x |
| |
| jmp NEAR $L$mod4_loop |
| ALIGN 32 |
| $L$mod4_loop: |
| DB 102,65,15,58,68,199,0 |
| xorps xmm4,xmm12 |
| movdqu xmm11,XMMWORD[48+r8] |
| DB 102,69,15,56,0,218 |
| DB 102,65,15,58,68,207,17 |
| xorps xmm0,xmm3 |
| movdqu xmm3,XMMWORD[32+r8] |
| movdqa xmm13,xmm11 |
| DB 102,68,15,58,68,199,16 |
| pshufd xmm12,xmm11,78 |
| xorps xmm1,xmm5 |
| pxor xmm12,xmm11 |
| DB 102,65,15,56,0,218 |
| movups xmm7,XMMWORD[32+rdx] |
| xorps xmm8,xmm4 |
| DB 102,68,15,58,68,218,0 |
| pshufd xmm4,xmm3,78 |
| |
| pxor xmm8,xmm0 |
| movdqa xmm5,xmm3 |
| pxor xmm8,xmm1 |
| pxor xmm4,xmm3 |
| movdqa xmm9,xmm8 |
| DB 102,68,15,58,68,234,17 |
| pslldq xmm8,8 |
| psrldq xmm9,8 |
| pxor xmm0,xmm8 |
| movdqa xmm8,XMMWORD[$L$7_mask] |
| pxor xmm1,xmm9 |
| DB 102,76,15,110,200 |
| |
| pand xmm8,xmm0 |
| DB 102,69,15,56,0,200 |
| pxor xmm9,xmm0 |
| DB 102,68,15,58,68,231,0 |
| psllq xmm9,57 |
| movdqa xmm8,xmm9 |
| pslldq xmm9,8 |
| DB 102,15,58,68,222,0 |
| psrldq xmm8,8 |
| pxor xmm0,xmm9 |
| pxor xmm1,xmm8 |
| movdqu xmm8,XMMWORD[r8] |
| |
| movdqa xmm9,xmm0 |
| psrlq xmm0,1 |
| DB 102,15,58,68,238,17 |
| xorps xmm3,xmm11 |
| movdqu xmm11,XMMWORD[16+r8] |
| DB 102,69,15,56,0,218 |
| DB 102,15,58,68,231,16 |
| xorps xmm5,xmm13 |
| movups xmm7,XMMWORD[80+rdx] |
| DB 102,69,15,56,0,194 |
| pxor xmm1,xmm9 |
| pxor xmm9,xmm0 |
| psrlq xmm0,5 |
| |
| movdqa xmm13,xmm11 |
| pxor xmm4,xmm12 |
| pshufd xmm12,xmm11,78 |
| pxor xmm0,xmm9 |
| pxor xmm1,xmm8 |
| pxor xmm12,xmm11 |
| DB 102,69,15,58,68,222,0 |
| psrlq xmm0,1 |
| pxor xmm0,xmm1 |
| movdqa xmm1,xmm0 |
| DB 102,69,15,58,68,238,17 |
| xorps xmm3,xmm11 |
| pshufd xmm8,xmm0,78 |
| pxor xmm8,xmm0 |
| |
| DB 102,68,15,58,68,231,0 |
| xorps xmm5,xmm13 |
| |
| lea r8,[64+r8] |
| sub r9,0x40 |
| jnc NEAR $L$mod4_loop |
| |
| $L$tail4x: |
| DB 102,65,15,58,68,199,0 |
| DB 102,65,15,58,68,207,17 |
| DB 102,68,15,58,68,199,16 |
| xorps xmm4,xmm12 |
| xorps xmm0,xmm3 |
| xorps xmm1,xmm5 |
| pxor xmm1,xmm0 |
| pxor xmm8,xmm4 |
| |
| pxor xmm8,xmm1 |
| pxor xmm1,xmm0 |
| |
| movdqa xmm9,xmm8 |
| psrldq xmm8,8 |
| pslldq xmm9,8 |
| pxor xmm1,xmm8 |
| pxor xmm0,xmm9 |
| |
| movdqa xmm4,xmm0 |
| movdqa xmm3,xmm0 |
| psllq xmm0,5 |
| pxor xmm3,xmm0 |
| psllq xmm0,1 |
| pxor xmm0,xmm3 |
| psllq xmm0,57 |
| movdqa xmm3,xmm0 |
| pslldq xmm0,8 |
| psrldq xmm3,8 |
| pxor xmm0,xmm4 |
| pxor xmm1,xmm3 |
| |
| |
| movdqa xmm4,xmm0 |
| psrlq xmm0,1 |
| pxor xmm1,xmm4 |
| pxor xmm4,xmm0 |
| psrlq xmm0,5 |
| pxor xmm0,xmm4 |
| psrlq xmm0,1 |
| pxor xmm0,xmm1 |
| add r9,0x40 |
| jz NEAR $L$done |
| movdqu xmm7,XMMWORD[32+rdx] |
| sub r9,0x10 |
| jz NEAR $L$odd_tail |
| $L$skip4x: |
| |
| |
| |
| |
| |
| movdqu xmm8,XMMWORD[r8] |
| movdqu xmm3,XMMWORD[16+r8] |
| DB 102,69,15,56,0,194 |
| DB 102,65,15,56,0,218 |
| pxor xmm0,xmm8 |
| |
| movdqa xmm5,xmm3 |
| pshufd xmm4,xmm3,78 |
| pxor xmm4,xmm3 |
| DB 102,15,58,68,218,0 |
| DB 102,15,58,68,234,17 |
| DB 102,15,58,68,231,0 |
| |
| lea r8,[32+r8] |
| nop |
| sub r9,0x20 |
| jbe NEAR $L$even_tail |
| nop |
| jmp NEAR $L$mod_loop |
| |
| ALIGN 32 |
| $L$mod_loop: |
| movdqa xmm1,xmm0 |
| movdqa xmm8,xmm4 |
| pshufd xmm4,xmm0,78 |
| pxor xmm4,xmm0 |
| |
| DB 102,15,58,68,198,0 |
| DB 102,15,58,68,206,17 |
| DB 102,15,58,68,231,16 |
| |
| pxor xmm0,xmm3 |
| pxor xmm1,xmm5 |
| movdqu xmm9,XMMWORD[r8] |
| pxor xmm8,xmm0 |
| DB 102,69,15,56,0,202 |
| movdqu xmm3,XMMWORD[16+r8] |
| |
| pxor xmm8,xmm1 |
| pxor xmm1,xmm9 |
| pxor xmm4,xmm8 |
| DB 102,65,15,56,0,218 |
| movdqa xmm8,xmm4 |
| psrldq xmm8,8 |
| pslldq xmm4,8 |
| pxor xmm1,xmm8 |
| pxor xmm0,xmm4 |
| |
| movdqa xmm5,xmm3 |
| |
| movdqa xmm9,xmm0 |
| movdqa xmm8,xmm0 |
| psllq xmm0,5 |
| pxor xmm8,xmm0 |
| DB 102,15,58,68,218,0 |
| psllq xmm0,1 |
| pxor xmm0,xmm8 |
| psllq xmm0,57 |
| movdqa xmm8,xmm0 |
| pslldq xmm0,8 |
| psrldq xmm8,8 |
| pxor xmm0,xmm9 |
| pshufd xmm4,xmm5,78 |
| pxor xmm1,xmm8 |
| pxor xmm4,xmm5 |
| |
| movdqa xmm9,xmm0 |
| psrlq xmm0,1 |
| DB 102,15,58,68,234,17 |
| pxor xmm1,xmm9 |
| pxor xmm9,xmm0 |
| psrlq xmm0,5 |
| pxor xmm0,xmm9 |
| lea r8,[32+r8] |
| psrlq xmm0,1 |
| DB 102,15,58,68,231,0 |
| pxor xmm0,xmm1 |
| |
| sub r9,0x20 |
| ja NEAR $L$mod_loop |
| |
| $L$even_tail: |
| movdqa xmm1,xmm0 |
| movdqa xmm8,xmm4 |
| pshufd xmm4,xmm0,78 |
| pxor xmm4,xmm0 |
| |
| DB 102,15,58,68,198,0 |
| DB 102,15,58,68,206,17 |
| DB 102,15,58,68,231,16 |
| |
| pxor xmm0,xmm3 |
| pxor xmm1,xmm5 |
| pxor xmm8,xmm0 |
| pxor xmm8,xmm1 |
| pxor xmm4,xmm8 |
| movdqa xmm8,xmm4 |
| psrldq xmm8,8 |
| pslldq xmm4,8 |
| pxor xmm1,xmm8 |
| pxor xmm0,xmm4 |
| |
| movdqa xmm4,xmm0 |
| movdqa xmm3,xmm0 |
| psllq xmm0,5 |
| pxor xmm3,xmm0 |
| psllq xmm0,1 |
| pxor xmm0,xmm3 |
| psllq xmm0,57 |
| movdqa xmm3,xmm0 |
| pslldq xmm0,8 |
| psrldq xmm3,8 |
| pxor xmm0,xmm4 |
| pxor xmm1,xmm3 |
| |
| |
| movdqa xmm4,xmm0 |
| psrlq xmm0,1 |
| pxor xmm1,xmm4 |
| pxor xmm4,xmm0 |
| psrlq xmm0,5 |
| pxor xmm0,xmm4 |
| psrlq xmm0,1 |
| pxor xmm0,xmm1 |
| test r9,r9 |
| jnz NEAR $L$done |
| |
| $L$odd_tail: |
| movdqu xmm8,XMMWORD[r8] |
| DB 102,69,15,56,0,194 |
| pxor xmm0,xmm8 |
| movdqa xmm1,xmm0 |
| pshufd xmm3,xmm0,78 |
| pxor xmm3,xmm0 |
| DB 102,15,58,68,194,0 |
| DB 102,15,58,68,202,17 |
| DB 102,15,58,68,223,0 |
| pxor xmm3,xmm0 |
| pxor xmm3,xmm1 |
| |
| movdqa xmm4,xmm3 |
| psrldq xmm3,8 |
| pslldq xmm4,8 |
| pxor xmm1,xmm3 |
| pxor xmm0,xmm4 |
| |
| movdqa xmm4,xmm0 |
| movdqa xmm3,xmm0 |
| psllq xmm0,5 |
| pxor xmm3,xmm0 |
| psllq xmm0,1 |
| pxor xmm0,xmm3 |
| psllq xmm0,57 |
| movdqa xmm3,xmm0 |
| pslldq xmm0,8 |
| psrldq xmm3,8 |
| pxor xmm0,xmm4 |
| pxor xmm1,xmm3 |
| |
| |
| movdqa xmm4,xmm0 |
| psrlq xmm0,1 |
| pxor xmm1,xmm4 |
| pxor xmm4,xmm0 |
| psrlq xmm0,5 |
| pxor xmm0,xmm4 |
| psrlq xmm0,1 |
| pxor xmm0,xmm1 |
| $L$done: |
| DB 102,65,15,56,0,194 |
| movdqu XMMWORD[rcx],xmm0 |
| movaps xmm6,XMMWORD[rsp] |
| movaps xmm7,XMMWORD[16+rsp] |
| movaps xmm8,XMMWORD[32+rsp] |
| movaps xmm9,XMMWORD[48+rsp] |
| movaps xmm10,XMMWORD[64+rsp] |
| movaps xmm11,XMMWORD[80+rsp] |
| movaps xmm12,XMMWORD[96+rsp] |
| movaps xmm13,XMMWORD[112+rsp] |
| movaps xmm14,XMMWORD[128+rsp] |
| movaps xmm15,XMMWORD[144+rsp] |
| lea rsp,[168+rsp] |
| $L$SEH_end_gcm_ghash_clmul: |
| DB 0F3h,0C3h ;repret |
| |
| |
| global gcm_init_avx |
| |
| ALIGN 32 |
| gcm_init_avx: |
| |
| $L$SEH_begin_gcm_init_avx: |
| |
| DB 0x48,0x83,0xec,0x18 |
| DB 0x0f,0x29,0x34,0x24 |
| vzeroupper |
| |
| vmovdqu xmm2,XMMWORD[rdx] |
| vpshufd xmm2,xmm2,78 |
| |
| |
| vpshufd xmm4,xmm2,255 |
| vpsrlq xmm3,xmm2,63 |
| vpsllq xmm2,xmm2,1 |
| vpxor xmm5,xmm5,xmm5 |
| vpcmpgtd xmm5,xmm5,xmm4 |
| vpslldq xmm3,xmm3,8 |
| vpor xmm2,xmm2,xmm3 |
| |
| |
| vpand xmm5,xmm5,XMMWORD[$L$0x1c2_polynomial] |
| vpxor xmm2,xmm2,xmm5 |
| |
| vpunpckhqdq xmm6,xmm2,xmm2 |
| vmovdqa xmm0,xmm2 |
| vpxor xmm6,xmm6,xmm2 |
| mov r10,4 |
| jmp NEAR $L$init_start_avx |
| ALIGN 32 |
| $L$init_loop_avx: |
| vpalignr xmm5,xmm4,xmm3,8 |
| vmovdqu XMMWORD[(-16)+rcx],xmm5 |
| vpunpckhqdq xmm3,xmm0,xmm0 |
| vpxor xmm3,xmm3,xmm0 |
| vpclmulqdq xmm1,xmm0,xmm2,0x11 |
| vpclmulqdq xmm0,xmm0,xmm2,0x00 |
| vpclmulqdq xmm3,xmm3,xmm6,0x00 |
| vpxor xmm4,xmm1,xmm0 |
| vpxor xmm3,xmm3,xmm4 |
| |
| vpslldq xmm4,xmm3,8 |
| vpsrldq xmm3,xmm3,8 |
| vpxor xmm0,xmm0,xmm4 |
| vpxor xmm1,xmm1,xmm3 |
| vpsllq xmm3,xmm0,57 |
| vpsllq xmm4,xmm0,62 |
| vpxor xmm4,xmm4,xmm3 |
| vpsllq xmm3,xmm0,63 |
| vpxor xmm4,xmm4,xmm3 |
| vpslldq xmm3,xmm4,8 |
| vpsrldq xmm4,xmm4,8 |
| vpxor xmm0,xmm0,xmm3 |
| vpxor xmm1,xmm1,xmm4 |
| |
| vpsrlq xmm4,xmm0,1 |
| vpxor xmm1,xmm1,xmm0 |
| vpxor xmm0,xmm0,xmm4 |
| vpsrlq xmm4,xmm4,5 |
| vpxor xmm0,xmm0,xmm4 |
| vpsrlq xmm0,xmm0,1 |
| vpxor xmm0,xmm0,xmm1 |
| $L$init_start_avx: |
| vmovdqa xmm5,xmm0 |
| vpunpckhqdq xmm3,xmm0,xmm0 |
| vpxor xmm3,xmm3,xmm0 |
| vpclmulqdq xmm1,xmm0,xmm2,0x11 |
| vpclmulqdq xmm0,xmm0,xmm2,0x00 |
| vpclmulqdq xmm3,xmm3,xmm6,0x00 |
| vpxor xmm4,xmm1,xmm0 |
| vpxor xmm3,xmm3,xmm4 |
| |
| vpslldq xmm4,xmm3,8 |
| vpsrldq xmm3,xmm3,8 |
| vpxor xmm0,xmm0,xmm4 |
| vpxor xmm1,xmm1,xmm3 |
| vpsllq xmm3,xmm0,57 |
| vpsllq xmm4,xmm0,62 |
| vpxor xmm4,xmm4,xmm3 |
| vpsllq xmm3,xmm0,63 |
| vpxor xmm4,xmm4,xmm3 |
| vpslldq xmm3,xmm4,8 |
| vpsrldq xmm4,xmm4,8 |
| vpxor xmm0,xmm0,xmm3 |
| vpxor xmm1,xmm1,xmm4 |
| |
| vpsrlq xmm4,xmm0,1 |
| vpxor xmm1,xmm1,xmm0 |
| vpxor xmm0,xmm0,xmm4 |
| vpsrlq xmm4,xmm4,5 |
| vpxor xmm0,xmm0,xmm4 |
| vpsrlq xmm0,xmm0,1 |
| vpxor xmm0,xmm0,xmm1 |
| vpshufd xmm3,xmm5,78 |
| vpshufd xmm4,xmm0,78 |
| vpxor xmm3,xmm3,xmm5 |
| vmovdqu XMMWORD[rcx],xmm5 |
| vpxor xmm4,xmm4,xmm0 |
| vmovdqu XMMWORD[16+rcx],xmm0 |
| lea rcx,[48+rcx] |
| sub r10,1 |
| jnz NEAR $L$init_loop_avx |
| |
| vpalignr xmm5,xmm3,xmm4,8 |
| vmovdqu XMMWORD[(-16)+rcx],xmm5 |
| |
| vzeroupper |
| movaps xmm6,XMMWORD[rsp] |
| lea rsp,[24+rsp] |
| $L$SEH_end_gcm_init_avx: |
| DB 0F3h,0C3h ;repret |
| |
| |
| global gcm_gmult_avx |
| |
| ALIGN 32 |
| gcm_gmult_avx: |
| |
| jmp NEAR $L$_gmult_clmul |
| |
| |
| global gcm_ghash_avx |
| |
| ALIGN 32 |
| gcm_ghash_avx: |
| |
| lea rax,[((-136))+rsp] |
| $L$SEH_begin_gcm_ghash_avx: |
| |
| DB 0x48,0x8d,0x60,0xe0 |
| DB 0x0f,0x29,0x70,0xe0 |
| DB 0x0f,0x29,0x78,0xf0 |
| DB 0x44,0x0f,0x29,0x00 |
| DB 0x44,0x0f,0x29,0x48,0x10 |
| DB 0x44,0x0f,0x29,0x50,0x20 |
| DB 0x44,0x0f,0x29,0x58,0x30 |
| DB 0x44,0x0f,0x29,0x60,0x40 |
| DB 0x44,0x0f,0x29,0x68,0x50 |
| DB 0x44,0x0f,0x29,0x70,0x60 |
| DB 0x44,0x0f,0x29,0x78,0x70 |
| vzeroupper |
| |
| vmovdqu xmm10,XMMWORD[rcx] |
| lea r10,[$L$0x1c2_polynomial] |
| lea rdx,[64+rdx] |
| vmovdqu xmm13,XMMWORD[$L$bswap_mask] |
| vpshufb xmm10,xmm10,xmm13 |
| cmp r9,0x80 |
| jb NEAR $L$short_avx |
| sub r9,0x80 |
| |
| vmovdqu xmm14,XMMWORD[112+r8] |
| vmovdqu xmm6,XMMWORD[((0-64))+rdx] |
| vpshufb xmm14,xmm14,xmm13 |
| vmovdqu xmm7,XMMWORD[((32-64))+rdx] |
| |
| vpunpckhqdq xmm9,xmm14,xmm14 |
| vmovdqu xmm15,XMMWORD[96+r8] |
| vpclmulqdq xmm0,xmm14,xmm6,0x00 |
| vpxor xmm9,xmm9,xmm14 |
| vpshufb xmm15,xmm15,xmm13 |
| vpclmulqdq xmm1,xmm14,xmm6,0x11 |
| vmovdqu xmm6,XMMWORD[((16-64))+rdx] |
| vpunpckhqdq xmm8,xmm15,xmm15 |
| vmovdqu xmm14,XMMWORD[80+r8] |
| vpclmulqdq xmm2,xmm9,xmm7,0x00 |
| vpxor xmm8,xmm8,xmm15 |
| |
| vpshufb xmm14,xmm14,xmm13 |
| vpclmulqdq xmm3,xmm15,xmm6,0x00 |
| vpunpckhqdq xmm9,xmm14,xmm14 |
| vpclmulqdq xmm4,xmm15,xmm6,0x11 |
| vmovdqu xmm6,XMMWORD[((48-64))+rdx] |
| vpxor xmm9,xmm9,xmm14 |
| vmovdqu xmm15,XMMWORD[64+r8] |
| vpclmulqdq xmm5,xmm8,xmm7,0x10 |
| vmovdqu xmm7,XMMWORD[((80-64))+rdx] |
| |
| vpshufb xmm15,xmm15,xmm13 |
| vpxor xmm3,xmm3,xmm0 |
| vpclmulqdq xmm0,xmm14,xmm6,0x00 |
| vpxor xmm4,xmm4,xmm1 |
| vpunpckhqdq xmm8,xmm15,xmm15 |
| vpclmulqdq xmm1,xmm14,xmm6,0x11 |
| vmovdqu xmm6,XMMWORD[((64-64))+rdx] |
| vpxor xmm5,xmm5,xmm2 |
| vpclmulqdq xmm2,xmm9,xmm7,0x00 |
| vpxor xmm8,xmm8,xmm15 |
| |
| vmovdqu xmm14,XMMWORD[48+r8] |
| vpxor xmm0,xmm0,xmm3 |
| vpclmulqdq xmm3,xmm15,xmm6,0x00 |
| vpxor xmm1,xmm1,xmm4 |
| vpshufb xmm14,xmm14,xmm13 |
| vpclmulqdq xmm4,xmm15,xmm6,0x11 |
| vmovdqu xmm6,XMMWORD[((96-64))+rdx] |
| vpxor xmm2,xmm2,xmm5 |
| vpunpckhqdq xmm9,xmm14,xmm14 |
| vpclmulqdq xmm5,xmm8,xmm7,0x10 |
| vmovdqu xmm7,XMMWORD[((128-64))+rdx] |
| vpxor xmm9,xmm9,xmm14 |
| |
| vmovdqu xmm15,XMMWORD[32+r8] |
| vpxor xmm3,xmm3,xmm0 |
| vpclmulqdq xmm0,xmm14,xmm6,0x00 |
| vpxor xmm4,xmm4,xmm1 |
| vpshufb xmm15,xmm15,xmm13 |
| vpclmulqdq xmm1,xmm14,xmm6,0x11 |
| vmovdqu xmm6,XMMWORD[((112-64))+rdx] |
| vpxor xmm5,xmm5,xmm2 |
| vpunpckhqdq xmm8,xmm15,xmm15 |
| vpclmulqdq xmm2,xmm9,xmm7,0x00 |
| vpxor xmm8,xmm8,xmm15 |
| |
| vmovdqu xmm14,XMMWORD[16+r8] |
| vpxor xmm0,xmm0,xmm3 |
| vpclmulqdq xmm3,xmm15,xmm6,0x00 |
| vpxor xmm1,xmm1,xmm4 |
| vpshufb xmm14,xmm14,xmm13 |
| vpclmulqdq xmm4,xmm15,xmm6,0x11 |
| vmovdqu xmm6,XMMWORD[((144-64))+rdx] |
| vpxor xmm2,xmm2,xmm5 |
| vpunpckhqdq xmm9,xmm14,xmm14 |
| vpclmulqdq xmm5,xmm8,xmm7,0x10 |
| vmovdqu xmm7,XMMWORD[((176-64))+rdx] |
| vpxor xmm9,xmm9,xmm14 |
| |
| vmovdqu xmm15,XMMWORD[r8] |
| vpxor xmm3,xmm3,xmm0 |
| vpclmulqdq xmm0,xmm14,xmm6,0x00 |
| vpxor xmm4,xmm4,xmm1 |
| vpshufb xmm15,xmm15,xmm13 |
| vpclmulqdq xmm1,xmm14,xmm6,0x11 |
| vmovdqu xmm6,XMMWORD[((160-64))+rdx] |
| vpxor xmm5,xmm5,xmm2 |
| vpclmulqdq xmm2,xmm9,xmm7,0x10 |
| |
| lea r8,[128+r8] |
| cmp r9,0x80 |
| jb NEAR $L$tail_avx |
| |
| vpxor xmm15,xmm15,xmm10 |
| sub r9,0x80 |
| jmp NEAR $L$oop8x_avx |
| |
| ALIGN 32 |
| $L$oop8x_avx: |
| vpunpckhqdq xmm8,xmm15,xmm15 |
| vmovdqu xmm14,XMMWORD[112+r8] |
| vpxor xmm3,xmm3,xmm0 |
| vpxor xmm8,xmm8,xmm15 |
| vpclmulqdq xmm10,xmm15,xmm6,0x00 |
| vpshufb xmm14,xmm14,xmm13 |
| vpxor xmm4,xmm4,xmm1 |
| vpclmulqdq xmm11,xmm15,xmm6,0x11 |
| vmovdqu xmm6,XMMWORD[((0-64))+rdx] |
| vpunpckhqdq xmm9,xmm14,xmm14 |
| vpxor xmm5,xmm5,xmm2 |
| vpclmulqdq xmm12,xmm8,xmm7,0x00 |
| vmovdqu xmm7,XMMWORD[((32-64))+rdx] |
| vpxor xmm9,xmm9,xmm14 |
| |
| vmovdqu xmm15,XMMWORD[96+r8] |
| vpclmulqdq xmm0,xmm14,xmm6,0x00 |
| vpxor xmm10,xmm10,xmm3 |
| vpshufb xmm15,xmm15,xmm13 |
| vpclmulqdq xmm1,xmm14,xmm6,0x11 |
| vxorps xmm11,xmm11,xmm4 |
| vmovdqu xmm6,XMMWORD[((16-64))+rdx] |
| vpunpckhqdq xmm8,xmm15,xmm15 |
| vpclmulqdq xmm2,xmm9,xmm7,0x00 |
| vpxor xmm12,xmm12,xmm5 |
| vxorps xmm8,xmm8,xmm15 |
| |
| vmovdqu xmm14,XMMWORD[80+r8] |
| vpxor xmm12,xmm12,xmm10 |
| vpclmulqdq xmm3,xmm15,xmm6,0x00 |
| vpxor xmm12,xmm12,xmm11 |
| vpslldq xmm9,xmm12,8 |
| vpxor xmm3,xmm3,xmm0 |
| vpclmulqdq xmm4,xmm15,xmm6,0x11 |
| vpsrldq xmm12,xmm12,8 |
| vpxor xmm10,xmm10,xmm9 |
| vmovdqu xmm6,XMMWORD[((48-64))+rdx] |
| vpshufb xmm14,xmm14,xmm13 |
| vxorps xmm11,xmm11,xmm12 |
| vpxor xmm4,xmm4,xmm1 |
| vpunpckhqdq xmm9,xmm14,xmm14 |
| vpclmulqdq xmm5,xmm8,xmm7,0x10 |
| vmovdqu xmm7,XMMWORD[((80-64))+rdx] |
| vpxor xmm9,xmm9,xmm14 |
| vpxor xmm5,xmm5,xmm2 |
| |
| vmovdqu xmm15,XMMWORD[64+r8] |
| vpalignr xmm12,xmm10,xmm10,8 |
| vpclmulqdq xmm0,xmm14,xmm6,0x00 |
| vpshufb xmm15,xmm15,xmm13 |
| vpxor xmm0,xmm0,xmm3 |
| vpclmulqdq xmm1,xmm14,xmm6,0x11 |
| vmovdqu xmm6,XMMWORD[((64-64))+rdx] |
| vpunpckhqdq xmm8,xmm15,xmm15 |
| vpxor xmm1,xmm1,xmm4 |
| vpclmulqdq xmm2,xmm9,xmm7,0x00 |
| vxorps xmm8,xmm8,xmm15 |
| vpxor xmm2,xmm2,xmm5 |
| |
| vmovdqu xmm14,XMMWORD[48+r8] |
| vpclmulqdq xmm10,xmm10,XMMWORD[r10],0x10 |
| vpclmulqdq xmm3,xmm15,xmm6,0x00 |
| vpshufb xmm14,xmm14,xmm13 |
| vpxor xmm3,xmm3,xmm0 |
| vpclmulqdq xmm4,xmm15,xmm6,0x11 |
| vmovdqu xmm6,XMMWORD[((96-64))+rdx] |
| vpunpckhqdq xmm9,xmm14,xmm14 |
| vpxor xmm4,xmm4,xmm1 |
| vpclmulqdq xmm5,xmm8,xmm7,0x10 |
| vmovdqu xmm7,XMMWORD[((128-64))+rdx] |
| vpxor xmm9,xmm9,xmm14 |
| vpxor xmm5,xmm5,xmm2 |
| |
| vmovdqu xmm15,XMMWORD[32+r8] |
| vpclmulqdq xmm0,xmm14,xmm6,0x00 |
| vpshufb xmm15,xmm15,xmm13 |
| vpxor xmm0,xmm0,xmm3 |
| vpclmulqdq xmm1,xmm14,xmm6,0x11 |
| vmovdqu xmm6,XMMWORD[((112-64))+rdx] |
| vpunpckhqdq xmm8,xmm15,xmm15 |
| vpxor xmm1,xmm1,xmm4 |
| vpclmulqdq xmm2,xmm9,xmm7,0x00 |
| vpxor xmm8,xmm8,xmm15 |
| vpxor xmm2,xmm2,xmm5 |
| vxorps xmm10,xmm10,xmm12 |
| |
| vmovdqu xmm14,XMMWORD[16+r8] |
| vpalignr xmm12,xmm10,xmm10,8 |
| vpclmulqdq xmm3,xmm15,xmm6,0x00 |
| vpshufb xmm14,xmm14,xmm13 |
| vpxor xmm3,xmm3,xmm0 |
| vpclmulqdq xmm4,xmm15,xmm6,0x11 |
| vmovdqu xmm6,XMMWORD[((144-64))+rdx] |
| vpclmulqdq xmm10,xmm10,XMMWORD[r10],0x10 |
| vxorps xmm12,xmm12,xmm11 |
| vpunpckhqdq xmm9,xmm14,xmm14 |
| vpxor xmm4,xmm4,xmm1 |
| vpclmulqdq xmm5,xmm8,xmm7,0x10 |
| vmovdqu xmm7,XMMWORD[((176-64))+rdx] |
| vpxor xmm9,xmm9,xmm14 |
| vpxor xmm5,xmm5,xmm2 |
| |
| vmovdqu xmm15,XMMWORD[r8] |
| vpclmulqdq xmm0,xmm14,xmm6,0x00 |
| vpshufb xmm15,xmm15,xmm13 |
| vpclmulqdq xmm1,xmm14,xmm6,0x11 |
| vmovdqu xmm6,XMMWORD[((160-64))+rdx] |
| vpxor xmm15,xmm15,xmm12 |
| vpclmulqdq xmm2,xmm9,xmm7,0x10 |
| vpxor xmm15,xmm15,xmm10 |
| |
| lea r8,[128+r8] |
| sub r9,0x80 |
| jnc NEAR $L$oop8x_avx |
| |
| add r9,0x80 |
| jmp NEAR $L$tail_no_xor_avx |
| |
| ALIGN 32 |
| $L$short_avx: |
| vmovdqu xmm14,XMMWORD[((-16))+r9*1+r8] |
| lea r8,[r9*1+r8] |
| vmovdqu xmm6,XMMWORD[((0-64))+rdx] |
| vmovdqu xmm7,XMMWORD[((32-64))+rdx] |
| vpshufb xmm15,xmm14,xmm13 |
| |
| vmovdqa xmm3,xmm0 |
| vmovdqa xmm4,xmm1 |
| vmovdqa xmm5,xmm2 |
| sub r9,0x10 |
| jz NEAR $L$tail_avx |
| |
| vpunpckhqdq xmm8,xmm15,xmm15 |
| vpxor xmm3,xmm3,xmm0 |
| vpclmulqdq xmm0,xmm15,xmm6,0x00 |
| vpxor xmm8,xmm8,xmm15 |
| vmovdqu xmm14,XMMWORD[((-32))+r8] |
| vpxor xmm4,xmm4,xmm1 |
| vpclmulqdq xmm1,xmm15,xmm6,0x11 |
| vmovdqu xmm6,XMMWORD[((16-64))+rdx] |
| vpshufb xmm15,xmm14,xmm13 |
| vpxor xmm5,xmm5,xmm2 |
| vpclmulqdq xmm2,xmm8,xmm7,0x00 |
| vpsrldq xmm7,xmm7,8 |
| sub r9,0x10 |
| jz NEAR $L$tail_avx |
| |
| vpunpckhqdq xmm8,xmm15,xmm15 |
| vpxor xmm3,xmm3,xmm0 |
| vpclmulqdq xmm0,xmm15,xmm6,0x00 |
| vpxor xmm8,xmm8,xmm15 |
| vmovdqu xmm14,XMMWORD[((-48))+r8] |
| vpxor xmm4,xmm4,xmm1 |
| vpclmulqdq xmm1,xmm15,xmm6,0x11 |
| vmovdqu xmm6,XMMWORD[((48-64))+rdx] |
| vpshufb xmm15,xmm14,xmm13 |
| vpxor xmm5,xmm5,xmm2 |
| vpclmulqdq xmm2,xmm8,xmm7,0x00 |
| vmovdqu xmm7,XMMWORD[((80-64))+rdx] |
| sub r9,0x10 |
| jz NEAR $L$tail_avx |
| |
| vpunpckhqdq xmm8,xmm15,xmm15 |
| vpxor xmm3,xmm3,xmm0 |
| vpclmulqdq xmm0,xmm15,xmm6,0x00 |
| vpxor xmm8,xmm8,xmm15 |
| vmovdqu xmm14,XMMWORD[((-64))+r8] |
| vpxor xmm4,xmm4,xmm1 |
| vpclmulqdq xmm1,xmm15,xmm6,0x11 |
| vmovdqu xmm6,XMMWORD[((64-64))+rdx] |
| vpshufb xmm15,xmm14,xmm13 |
| vpxor xmm5,xmm5,xmm2 |
| vpclmulqdq xmm2,xmm8,xmm7,0x00 |
| vpsrldq xmm7,xmm7,8 |
| sub r9,0x10 |
| jz NEAR $L$tail_avx |
| |
| vpunpckhqdq xmm8,xmm15,xmm15 |
| vpxor xmm3,xmm3,xmm0 |
| vpclmulqdq xmm0,xmm15,xmm6,0x00 |
| vpxor xmm8,xmm8,xmm15 |
| vmovdqu xmm14,XMMWORD[((-80))+r8] |
| vpxor xmm4,xmm4,xmm1 |
| vpclmulqdq xmm1,xmm15,xmm6,0x11 |
| vmovdqu xmm6,XMMWORD[((96-64))+rdx] |
| vpshufb xmm15,xmm14,xmm13 |
| vpxor xmm5,xmm5,xmm2 |
| vpclmulqdq xmm2,xmm8,xmm7,0x00 |
| vmovdqu xmm7,XMMWORD[((128-64))+rdx] |
| sub r9,0x10 |
| jz NEAR $L$tail_avx |
| |
| vpunpckhqdq xmm8,xmm15,xmm15 |
| vpxor xmm3,xmm3,xmm0 |
| vpclmulqdq xmm0,xmm15,xmm6,0x00 |
| vpxor xmm8,xmm8,xmm15 |
| vmovdqu xmm14,XMMWORD[((-96))+r8] |
| vpxor xmm4,xmm4,xmm1 |
| vpclmulqdq xmm1,xmm15,xmm6,0x11 |
| vmovdqu xmm6,XMMWORD[((112-64))+rdx] |
| vpshufb xmm15,xmm14,xmm13 |
| vpxor xmm5,xmm5,xmm2 |
| vpclmulqdq xmm2,xmm8,xmm7,0x00 |
| vpsrldq xmm7,xmm7,8 |
| sub r9,0x10 |
| jz NEAR $L$tail_avx |
| |
| vpunpckhqdq xmm8,xmm15,xmm15 |
| vpxor xmm3,xmm3,xmm0 |
| vpclmulqdq xmm0,xmm15,xmm6,0x00 |
| vpxor xmm8,xmm8,xmm15 |
| vmovdqu xmm14,XMMWORD[((-112))+r8] |
| vpxor xmm4,xmm4,xmm1 |
| vpclmulqdq xmm1,xmm15,xmm6,0x11 |
| vmovdqu xmm6,XMMWORD[((144-64))+rdx] |
| vpshufb xmm15,xmm14,xmm13 |
| vpxor xmm5,xmm5,xmm2 |
| vpclmulqdq xmm2,xmm8,xmm7,0x00 |
| vmovq xmm7,QWORD[((184-64))+rdx] |
| sub r9,0x10 |
| jmp NEAR $L$tail_avx |
| |
| ALIGN 32 |
| $L$tail_avx: |
| vpxor xmm15,xmm15,xmm10 |
| $L$tail_no_xor_avx: |
| vpunpckhqdq xmm8,xmm15,xmm15 |
| vpxor xmm3,xmm3,xmm0 |
| vpclmulqdq xmm0,xmm15,xmm6,0x00 |
| vpxor xmm8,xmm8,xmm15 |
| vpxor xmm4,xmm4,xmm1 |
| vpclmulqdq xmm1,xmm15,xmm6,0x11 |
| vpxor xmm5,xmm5,xmm2 |
| vpclmulqdq xmm2,xmm8,xmm7,0x00 |
| |
| vmovdqu xmm12,XMMWORD[r10] |
| |
| vpxor xmm10,xmm3,xmm0 |
| vpxor xmm11,xmm4,xmm1 |
| vpxor xmm5,xmm5,xmm2 |
| |
| vpxor xmm5,xmm5,xmm10 |
| vpxor xmm5,xmm5,xmm11 |
| vpslldq xmm9,xmm5,8 |
| vpsrldq xmm5,xmm5,8 |
| vpxor xmm10,xmm10,xmm9 |
| vpxor xmm11,xmm11,xmm5 |
| |
| vpclmulqdq xmm9,xmm10,xmm12,0x10 |
| vpalignr xmm10,xmm10,xmm10,8 |
| vpxor xmm10,xmm10,xmm9 |
| |
| vpclmulqdq xmm9,xmm10,xmm12,0x10 |
| vpalignr xmm10,xmm10,xmm10,8 |
| vpxor xmm10,xmm10,xmm11 |
| vpxor xmm10,xmm10,xmm9 |
| |
| cmp r9,0 |
| jne NEAR $L$short_avx |
| |
| vpshufb xmm10,xmm10,xmm13 |
| vmovdqu XMMWORD[rcx],xmm10 |
| vzeroupper |
| movaps xmm6,XMMWORD[rsp] |
| movaps xmm7,XMMWORD[16+rsp] |
| movaps xmm8,XMMWORD[32+rsp] |
| movaps xmm9,XMMWORD[48+rsp] |
| movaps xmm10,XMMWORD[64+rsp] |
| movaps xmm11,XMMWORD[80+rsp] |
| movaps xmm12,XMMWORD[96+rsp] |
| movaps xmm13,XMMWORD[112+rsp] |
| movaps xmm14,XMMWORD[128+rsp] |
| movaps xmm15,XMMWORD[144+rsp] |
| lea rsp,[168+rsp] |
| $L$SEH_end_gcm_ghash_avx: |
| DB 0F3h,0C3h ;repret |
| |
| |
| ALIGN 64 |
| $L$bswap_mask: |
| DB 15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0 |
| $L$0x1c2_polynomial: |
| DB 1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0xc2 |
| $L$7_mask: |
| DD 7,0,7,0 |
| ALIGN 64 |
| |
| DB 71,72,65,83,72,32,102,111,114,32,120,56,54,95,54,52 |
| DB 44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32 |
| DB 60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111 |
| DB 114,103,62,0 |
| ALIGN 64 |
| section .pdata rdata align=4 |
| ALIGN 4 |
| DD $L$SEH_begin_gcm_init_clmul wrt ..imagebase |
| DD $L$SEH_end_gcm_init_clmul wrt ..imagebase |
| DD $L$SEH_info_gcm_init_clmul wrt ..imagebase |
| |
| DD $L$SEH_begin_gcm_ghash_clmul wrt ..imagebase |
| DD $L$SEH_end_gcm_ghash_clmul wrt ..imagebase |
| DD $L$SEH_info_gcm_ghash_clmul wrt ..imagebase |
| DD $L$SEH_begin_gcm_init_avx wrt ..imagebase |
| DD $L$SEH_end_gcm_init_avx wrt ..imagebase |
| DD $L$SEH_info_gcm_init_clmul wrt ..imagebase |
| |
| DD $L$SEH_begin_gcm_ghash_avx wrt ..imagebase |
| DD $L$SEH_end_gcm_ghash_avx wrt ..imagebase |
| DD $L$SEH_info_gcm_ghash_clmul wrt ..imagebase |
| section .xdata rdata align=8 |
| ALIGN 8 |
| $L$SEH_info_gcm_init_clmul: |
| DB 0x01,0x08,0x03,0x00 |
| DB 0x08,0x68,0x00,0x00 |
| DB 0x04,0x22,0x00,0x00 |
| $L$SEH_info_gcm_ghash_clmul: |
| DB 0x01,0x33,0x16,0x00 |
| DB 0x33,0xf8,0x09,0x00 |
| DB 0x2e,0xe8,0x08,0x00 |
| DB 0x29,0xd8,0x07,0x00 |
| DB 0x24,0xc8,0x06,0x00 |
| DB 0x1f,0xb8,0x05,0x00 |
| DB 0x1a,0xa8,0x04,0x00 |
| DB 0x15,0x98,0x03,0x00 |
| DB 0x10,0x88,0x02,0x00 |
| DB 0x0c,0x78,0x01,0x00 |
| DB 0x08,0x68,0x00,0x00 |
| DB 0x04,0x01,0x15,0x00 |