|  | // This file is generated from a similarly-named Perl script in the BoringSSL | 
|  | // source tree. Do not edit by hand. | 
|  |  | 
|  | #include <openssl/asm_base.h> | 
|  |  | 
|  | #if !defined(OPENSSL_NO_ASM) && defined(OPENSSL_X86_64) && defined(__ELF__) | 
|  | .text | 
|  |  | 
|  | .globl	sha256_block_data_order_nohw | 
|  | .hidden sha256_block_data_order_nohw | 
|  | .type	sha256_block_data_order_nohw,@function | 
|  | .align	16 | 
|  | sha256_block_data_order_nohw: | 
|  | .cfi_startproc | 
|  | _CET_ENDBR | 
|  | movq	%rsp,%rax | 
|  | .cfi_def_cfa_register	%rax | 
|  | pushq	%rbx | 
|  | .cfi_offset	%rbx,-16 | 
|  | pushq	%rbp | 
|  | .cfi_offset	%rbp,-24 | 
|  | pushq	%r12 | 
|  | .cfi_offset	%r12,-32 | 
|  | pushq	%r13 | 
|  | .cfi_offset	%r13,-40 | 
|  | pushq	%r14 | 
|  | .cfi_offset	%r14,-48 | 
|  | pushq	%r15 | 
|  | .cfi_offset	%r15,-56 | 
|  | shlq	$4,%rdx | 
|  | subq	$64+32,%rsp | 
|  | leaq	(%rsi,%rdx,4),%rdx | 
|  | andq	$-64,%rsp | 
|  | movq	%rdi,64+0(%rsp) | 
|  | movq	%rsi,64+8(%rsp) | 
|  | movq	%rdx,64+16(%rsp) | 
|  | movq	%rax,88(%rsp) | 
|  | .cfi_escape	0x0f,0x06,0x77,0xd8,0x00,0x06,0x23,0x08 | 
|  | .Lprologue: | 
|  |  | 
|  | movl	0(%rdi),%eax | 
|  | movl	4(%rdi),%ebx | 
|  | movl	8(%rdi),%ecx | 
|  | movl	12(%rdi),%edx | 
|  | movl	16(%rdi),%r8d | 
|  | movl	20(%rdi),%r9d | 
|  | movl	24(%rdi),%r10d | 
|  | movl	28(%rdi),%r11d | 
|  | jmp	.Lloop | 
|  |  | 
|  | .align	16 | 
|  | .Lloop: | 
|  | movl	%ebx,%edi | 
|  | leaq	K256(%rip),%rbp | 
|  | xorl	%ecx,%edi | 
|  | movl	0(%rsi),%r12d | 
|  | movl	%r8d,%r13d | 
|  | movl	%eax,%r14d | 
|  | bswapl	%r12d | 
|  | rorl	$14,%r13d | 
|  | movl	%r9d,%r15d | 
|  |  | 
|  | xorl	%r8d,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r10d,%r15d | 
|  |  | 
|  | movl	%r12d,0(%rsp) | 
|  | xorl	%eax,%r14d | 
|  | andl	%r8d,%r15d | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%r11d,%r12d | 
|  | xorl	%r10d,%r15d | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%r8d,%r13d | 
|  | addl	%r15d,%r12d | 
|  |  | 
|  | movl	%eax,%r15d | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%eax,%r14d | 
|  |  | 
|  | xorl	%ebx,%r15d | 
|  | rorl	$6,%r13d | 
|  | movl	%ebx,%r11d | 
|  |  | 
|  | andl	%r15d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%edi,%r11d | 
|  | addl	%r12d,%edx | 
|  | addl	%r12d,%r11d | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | addl	%r14d,%r11d | 
|  | movl	4(%rsi),%r12d | 
|  | movl	%edx,%r13d | 
|  | movl	%r11d,%r14d | 
|  | bswapl	%r12d | 
|  | rorl	$14,%r13d | 
|  | movl	%r8d,%edi | 
|  |  | 
|  | xorl	%edx,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r9d,%edi | 
|  |  | 
|  | movl	%r12d,4(%rsp) | 
|  | xorl	%r11d,%r14d | 
|  | andl	%edx,%edi | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%r10d,%r12d | 
|  | xorl	%r9d,%edi | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%edx,%r13d | 
|  | addl	%edi,%r12d | 
|  |  | 
|  | movl	%r11d,%edi | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%r11d,%r14d | 
|  |  | 
|  | xorl	%eax,%edi | 
|  | rorl	$6,%r13d | 
|  | movl	%eax,%r10d | 
|  |  | 
|  | andl	%edi,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%r15d,%r10d | 
|  | addl	%r12d,%ecx | 
|  | addl	%r12d,%r10d | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | addl	%r14d,%r10d | 
|  | movl	8(%rsi),%r12d | 
|  | movl	%ecx,%r13d | 
|  | movl	%r10d,%r14d | 
|  | bswapl	%r12d | 
|  | rorl	$14,%r13d | 
|  | movl	%edx,%r15d | 
|  |  | 
|  | xorl	%ecx,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r8d,%r15d | 
|  |  | 
|  | movl	%r12d,8(%rsp) | 
|  | xorl	%r10d,%r14d | 
|  | andl	%ecx,%r15d | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%r9d,%r12d | 
|  | xorl	%r8d,%r15d | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%ecx,%r13d | 
|  | addl	%r15d,%r12d | 
|  |  | 
|  | movl	%r10d,%r15d | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%r10d,%r14d | 
|  |  | 
|  | xorl	%r11d,%r15d | 
|  | rorl	$6,%r13d | 
|  | movl	%r11d,%r9d | 
|  |  | 
|  | andl	%r15d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%edi,%r9d | 
|  | addl	%r12d,%ebx | 
|  | addl	%r12d,%r9d | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | addl	%r14d,%r9d | 
|  | movl	12(%rsi),%r12d | 
|  | movl	%ebx,%r13d | 
|  | movl	%r9d,%r14d | 
|  | bswapl	%r12d | 
|  | rorl	$14,%r13d | 
|  | movl	%ecx,%edi | 
|  |  | 
|  | xorl	%ebx,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%edx,%edi | 
|  |  | 
|  | movl	%r12d,12(%rsp) | 
|  | xorl	%r9d,%r14d | 
|  | andl	%ebx,%edi | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%r8d,%r12d | 
|  | xorl	%edx,%edi | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%ebx,%r13d | 
|  | addl	%edi,%r12d | 
|  |  | 
|  | movl	%r9d,%edi | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%r9d,%r14d | 
|  |  | 
|  | xorl	%r10d,%edi | 
|  | rorl	$6,%r13d | 
|  | movl	%r10d,%r8d | 
|  |  | 
|  | andl	%edi,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%r15d,%r8d | 
|  | addl	%r12d,%eax | 
|  | addl	%r12d,%r8d | 
|  |  | 
|  | leaq	20(%rbp),%rbp | 
|  | addl	%r14d,%r8d | 
|  | movl	16(%rsi),%r12d | 
|  | movl	%eax,%r13d | 
|  | movl	%r8d,%r14d | 
|  | bswapl	%r12d | 
|  | rorl	$14,%r13d | 
|  | movl	%ebx,%r15d | 
|  |  | 
|  | xorl	%eax,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%ecx,%r15d | 
|  |  | 
|  | movl	%r12d,16(%rsp) | 
|  | xorl	%r8d,%r14d | 
|  | andl	%eax,%r15d | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%edx,%r12d | 
|  | xorl	%ecx,%r15d | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%eax,%r13d | 
|  | addl	%r15d,%r12d | 
|  |  | 
|  | movl	%r8d,%r15d | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%r8d,%r14d | 
|  |  | 
|  | xorl	%r9d,%r15d | 
|  | rorl	$6,%r13d | 
|  | movl	%r9d,%edx | 
|  |  | 
|  | andl	%r15d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%edi,%edx | 
|  | addl	%r12d,%r11d | 
|  | addl	%r12d,%edx | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | addl	%r14d,%edx | 
|  | movl	20(%rsi),%r12d | 
|  | movl	%r11d,%r13d | 
|  | movl	%edx,%r14d | 
|  | bswapl	%r12d | 
|  | rorl	$14,%r13d | 
|  | movl	%eax,%edi | 
|  |  | 
|  | xorl	%r11d,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%ebx,%edi | 
|  |  | 
|  | movl	%r12d,20(%rsp) | 
|  | xorl	%edx,%r14d | 
|  | andl	%r11d,%edi | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%ecx,%r12d | 
|  | xorl	%ebx,%edi | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%r11d,%r13d | 
|  | addl	%edi,%r12d | 
|  |  | 
|  | movl	%edx,%edi | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%edx,%r14d | 
|  |  | 
|  | xorl	%r8d,%edi | 
|  | rorl	$6,%r13d | 
|  | movl	%r8d,%ecx | 
|  |  | 
|  | andl	%edi,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%r15d,%ecx | 
|  | addl	%r12d,%r10d | 
|  | addl	%r12d,%ecx | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | addl	%r14d,%ecx | 
|  | movl	24(%rsi),%r12d | 
|  | movl	%r10d,%r13d | 
|  | movl	%ecx,%r14d | 
|  | bswapl	%r12d | 
|  | rorl	$14,%r13d | 
|  | movl	%r11d,%r15d | 
|  |  | 
|  | xorl	%r10d,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%eax,%r15d | 
|  |  | 
|  | movl	%r12d,24(%rsp) | 
|  | xorl	%ecx,%r14d | 
|  | andl	%r10d,%r15d | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%ebx,%r12d | 
|  | xorl	%eax,%r15d | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%r10d,%r13d | 
|  | addl	%r15d,%r12d | 
|  |  | 
|  | movl	%ecx,%r15d | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%ecx,%r14d | 
|  |  | 
|  | xorl	%edx,%r15d | 
|  | rorl	$6,%r13d | 
|  | movl	%edx,%ebx | 
|  |  | 
|  | andl	%r15d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%edi,%ebx | 
|  | addl	%r12d,%r9d | 
|  | addl	%r12d,%ebx | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | addl	%r14d,%ebx | 
|  | movl	28(%rsi),%r12d | 
|  | movl	%r9d,%r13d | 
|  | movl	%ebx,%r14d | 
|  | bswapl	%r12d | 
|  | rorl	$14,%r13d | 
|  | movl	%r10d,%edi | 
|  |  | 
|  | xorl	%r9d,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r11d,%edi | 
|  |  | 
|  | movl	%r12d,28(%rsp) | 
|  | xorl	%ebx,%r14d | 
|  | andl	%r9d,%edi | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%eax,%r12d | 
|  | xorl	%r11d,%edi | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%r9d,%r13d | 
|  | addl	%edi,%r12d | 
|  |  | 
|  | movl	%ebx,%edi | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%ebx,%r14d | 
|  |  | 
|  | xorl	%ecx,%edi | 
|  | rorl	$6,%r13d | 
|  | movl	%ecx,%eax | 
|  |  | 
|  | andl	%edi,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%r15d,%eax | 
|  | addl	%r12d,%r8d | 
|  | addl	%r12d,%eax | 
|  |  | 
|  | leaq	20(%rbp),%rbp | 
|  | addl	%r14d,%eax | 
|  | movl	32(%rsi),%r12d | 
|  | movl	%r8d,%r13d | 
|  | movl	%eax,%r14d | 
|  | bswapl	%r12d | 
|  | rorl	$14,%r13d | 
|  | movl	%r9d,%r15d | 
|  |  | 
|  | xorl	%r8d,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r10d,%r15d | 
|  |  | 
|  | movl	%r12d,32(%rsp) | 
|  | xorl	%eax,%r14d | 
|  | andl	%r8d,%r15d | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%r11d,%r12d | 
|  | xorl	%r10d,%r15d | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%r8d,%r13d | 
|  | addl	%r15d,%r12d | 
|  |  | 
|  | movl	%eax,%r15d | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%eax,%r14d | 
|  |  | 
|  | xorl	%ebx,%r15d | 
|  | rorl	$6,%r13d | 
|  | movl	%ebx,%r11d | 
|  |  | 
|  | andl	%r15d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%edi,%r11d | 
|  | addl	%r12d,%edx | 
|  | addl	%r12d,%r11d | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | addl	%r14d,%r11d | 
|  | movl	36(%rsi),%r12d | 
|  | movl	%edx,%r13d | 
|  | movl	%r11d,%r14d | 
|  | bswapl	%r12d | 
|  | rorl	$14,%r13d | 
|  | movl	%r8d,%edi | 
|  |  | 
|  | xorl	%edx,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r9d,%edi | 
|  |  | 
|  | movl	%r12d,36(%rsp) | 
|  | xorl	%r11d,%r14d | 
|  | andl	%edx,%edi | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%r10d,%r12d | 
|  | xorl	%r9d,%edi | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%edx,%r13d | 
|  | addl	%edi,%r12d | 
|  |  | 
|  | movl	%r11d,%edi | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%r11d,%r14d | 
|  |  | 
|  | xorl	%eax,%edi | 
|  | rorl	$6,%r13d | 
|  | movl	%eax,%r10d | 
|  |  | 
|  | andl	%edi,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%r15d,%r10d | 
|  | addl	%r12d,%ecx | 
|  | addl	%r12d,%r10d | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | addl	%r14d,%r10d | 
|  | movl	40(%rsi),%r12d | 
|  | movl	%ecx,%r13d | 
|  | movl	%r10d,%r14d | 
|  | bswapl	%r12d | 
|  | rorl	$14,%r13d | 
|  | movl	%edx,%r15d | 
|  |  | 
|  | xorl	%ecx,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r8d,%r15d | 
|  |  | 
|  | movl	%r12d,40(%rsp) | 
|  | xorl	%r10d,%r14d | 
|  | andl	%ecx,%r15d | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%r9d,%r12d | 
|  | xorl	%r8d,%r15d | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%ecx,%r13d | 
|  | addl	%r15d,%r12d | 
|  |  | 
|  | movl	%r10d,%r15d | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%r10d,%r14d | 
|  |  | 
|  | xorl	%r11d,%r15d | 
|  | rorl	$6,%r13d | 
|  | movl	%r11d,%r9d | 
|  |  | 
|  | andl	%r15d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%edi,%r9d | 
|  | addl	%r12d,%ebx | 
|  | addl	%r12d,%r9d | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | addl	%r14d,%r9d | 
|  | movl	44(%rsi),%r12d | 
|  | movl	%ebx,%r13d | 
|  | movl	%r9d,%r14d | 
|  | bswapl	%r12d | 
|  | rorl	$14,%r13d | 
|  | movl	%ecx,%edi | 
|  |  | 
|  | xorl	%ebx,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%edx,%edi | 
|  |  | 
|  | movl	%r12d,44(%rsp) | 
|  | xorl	%r9d,%r14d | 
|  | andl	%ebx,%edi | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%r8d,%r12d | 
|  | xorl	%edx,%edi | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%ebx,%r13d | 
|  | addl	%edi,%r12d | 
|  |  | 
|  | movl	%r9d,%edi | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%r9d,%r14d | 
|  |  | 
|  | xorl	%r10d,%edi | 
|  | rorl	$6,%r13d | 
|  | movl	%r10d,%r8d | 
|  |  | 
|  | andl	%edi,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%r15d,%r8d | 
|  | addl	%r12d,%eax | 
|  | addl	%r12d,%r8d | 
|  |  | 
|  | leaq	20(%rbp),%rbp | 
|  | addl	%r14d,%r8d | 
|  | movl	48(%rsi),%r12d | 
|  | movl	%eax,%r13d | 
|  | movl	%r8d,%r14d | 
|  | bswapl	%r12d | 
|  | rorl	$14,%r13d | 
|  | movl	%ebx,%r15d | 
|  |  | 
|  | xorl	%eax,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%ecx,%r15d | 
|  |  | 
|  | movl	%r12d,48(%rsp) | 
|  | xorl	%r8d,%r14d | 
|  | andl	%eax,%r15d | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%edx,%r12d | 
|  | xorl	%ecx,%r15d | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%eax,%r13d | 
|  | addl	%r15d,%r12d | 
|  |  | 
|  | movl	%r8d,%r15d | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%r8d,%r14d | 
|  |  | 
|  | xorl	%r9d,%r15d | 
|  | rorl	$6,%r13d | 
|  | movl	%r9d,%edx | 
|  |  | 
|  | andl	%r15d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%edi,%edx | 
|  | addl	%r12d,%r11d | 
|  | addl	%r12d,%edx | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | addl	%r14d,%edx | 
|  | movl	52(%rsi),%r12d | 
|  | movl	%r11d,%r13d | 
|  | movl	%edx,%r14d | 
|  | bswapl	%r12d | 
|  | rorl	$14,%r13d | 
|  | movl	%eax,%edi | 
|  |  | 
|  | xorl	%r11d,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%ebx,%edi | 
|  |  | 
|  | movl	%r12d,52(%rsp) | 
|  | xorl	%edx,%r14d | 
|  | andl	%r11d,%edi | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%ecx,%r12d | 
|  | xorl	%ebx,%edi | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%r11d,%r13d | 
|  | addl	%edi,%r12d | 
|  |  | 
|  | movl	%edx,%edi | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%edx,%r14d | 
|  |  | 
|  | xorl	%r8d,%edi | 
|  | rorl	$6,%r13d | 
|  | movl	%r8d,%ecx | 
|  |  | 
|  | andl	%edi,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%r15d,%ecx | 
|  | addl	%r12d,%r10d | 
|  | addl	%r12d,%ecx | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | addl	%r14d,%ecx | 
|  | movl	56(%rsi),%r12d | 
|  | movl	%r10d,%r13d | 
|  | movl	%ecx,%r14d | 
|  | bswapl	%r12d | 
|  | rorl	$14,%r13d | 
|  | movl	%r11d,%r15d | 
|  |  | 
|  | xorl	%r10d,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%eax,%r15d | 
|  |  | 
|  | movl	%r12d,56(%rsp) | 
|  | xorl	%ecx,%r14d | 
|  | andl	%r10d,%r15d | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%ebx,%r12d | 
|  | xorl	%eax,%r15d | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%r10d,%r13d | 
|  | addl	%r15d,%r12d | 
|  |  | 
|  | movl	%ecx,%r15d | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%ecx,%r14d | 
|  |  | 
|  | xorl	%edx,%r15d | 
|  | rorl	$6,%r13d | 
|  | movl	%edx,%ebx | 
|  |  | 
|  | andl	%r15d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%edi,%ebx | 
|  | addl	%r12d,%r9d | 
|  | addl	%r12d,%ebx | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | addl	%r14d,%ebx | 
|  | movl	60(%rsi),%r12d | 
|  | movl	%r9d,%r13d | 
|  | movl	%ebx,%r14d | 
|  | bswapl	%r12d | 
|  | rorl	$14,%r13d | 
|  | movl	%r10d,%edi | 
|  |  | 
|  | xorl	%r9d,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r11d,%edi | 
|  |  | 
|  | movl	%r12d,60(%rsp) | 
|  | xorl	%ebx,%r14d | 
|  | andl	%r9d,%edi | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%eax,%r12d | 
|  | xorl	%r11d,%edi | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%r9d,%r13d | 
|  | addl	%edi,%r12d | 
|  |  | 
|  | movl	%ebx,%edi | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%ebx,%r14d | 
|  |  | 
|  | xorl	%ecx,%edi | 
|  | rorl	$6,%r13d | 
|  | movl	%ecx,%eax | 
|  |  | 
|  | andl	%edi,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%r15d,%eax | 
|  | addl	%r12d,%r8d | 
|  | addl	%r12d,%eax | 
|  |  | 
|  | leaq	20(%rbp),%rbp | 
|  | jmp	.Lrounds_16_xx | 
|  | .align	16 | 
|  | .Lrounds_16_xx: | 
|  | movl	4(%rsp),%r13d | 
|  | movl	56(%rsp),%r15d | 
|  |  | 
|  | movl	%r13d,%r12d | 
|  | rorl	$11,%r13d | 
|  | addl	%r14d,%eax | 
|  | movl	%r15d,%r14d | 
|  | rorl	$2,%r15d | 
|  |  | 
|  | xorl	%r12d,%r13d | 
|  | shrl	$3,%r12d | 
|  | rorl	$7,%r13d | 
|  | xorl	%r14d,%r15d | 
|  | shrl	$10,%r14d | 
|  |  | 
|  | rorl	$17,%r15d | 
|  | xorl	%r13d,%r12d | 
|  | xorl	%r14d,%r15d | 
|  | addl	36(%rsp),%r12d | 
|  |  | 
|  | addl	0(%rsp),%r12d | 
|  | movl	%r8d,%r13d | 
|  | addl	%r15d,%r12d | 
|  | movl	%eax,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r9d,%r15d | 
|  |  | 
|  | xorl	%r8d,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r10d,%r15d | 
|  |  | 
|  | movl	%r12d,0(%rsp) | 
|  | xorl	%eax,%r14d | 
|  | andl	%r8d,%r15d | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%r11d,%r12d | 
|  | xorl	%r10d,%r15d | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%r8d,%r13d | 
|  | addl	%r15d,%r12d | 
|  |  | 
|  | movl	%eax,%r15d | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%eax,%r14d | 
|  |  | 
|  | xorl	%ebx,%r15d | 
|  | rorl	$6,%r13d | 
|  | movl	%ebx,%r11d | 
|  |  | 
|  | andl	%r15d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%edi,%r11d | 
|  | addl	%r12d,%edx | 
|  | addl	%r12d,%r11d | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | movl	8(%rsp),%r13d | 
|  | movl	60(%rsp),%edi | 
|  |  | 
|  | movl	%r13d,%r12d | 
|  | rorl	$11,%r13d | 
|  | addl	%r14d,%r11d | 
|  | movl	%edi,%r14d | 
|  | rorl	$2,%edi | 
|  |  | 
|  | xorl	%r12d,%r13d | 
|  | shrl	$3,%r12d | 
|  | rorl	$7,%r13d | 
|  | xorl	%r14d,%edi | 
|  | shrl	$10,%r14d | 
|  |  | 
|  | rorl	$17,%edi | 
|  | xorl	%r13d,%r12d | 
|  | xorl	%r14d,%edi | 
|  | addl	40(%rsp),%r12d | 
|  |  | 
|  | addl	4(%rsp),%r12d | 
|  | movl	%edx,%r13d | 
|  | addl	%edi,%r12d | 
|  | movl	%r11d,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r8d,%edi | 
|  |  | 
|  | xorl	%edx,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r9d,%edi | 
|  |  | 
|  | movl	%r12d,4(%rsp) | 
|  | xorl	%r11d,%r14d | 
|  | andl	%edx,%edi | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%r10d,%r12d | 
|  | xorl	%r9d,%edi | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%edx,%r13d | 
|  | addl	%edi,%r12d | 
|  |  | 
|  | movl	%r11d,%edi | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%r11d,%r14d | 
|  |  | 
|  | xorl	%eax,%edi | 
|  | rorl	$6,%r13d | 
|  | movl	%eax,%r10d | 
|  |  | 
|  | andl	%edi,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%r15d,%r10d | 
|  | addl	%r12d,%ecx | 
|  | addl	%r12d,%r10d | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | movl	12(%rsp),%r13d | 
|  | movl	0(%rsp),%r15d | 
|  |  | 
|  | movl	%r13d,%r12d | 
|  | rorl	$11,%r13d | 
|  | addl	%r14d,%r10d | 
|  | movl	%r15d,%r14d | 
|  | rorl	$2,%r15d | 
|  |  | 
|  | xorl	%r12d,%r13d | 
|  | shrl	$3,%r12d | 
|  | rorl	$7,%r13d | 
|  | xorl	%r14d,%r15d | 
|  | shrl	$10,%r14d | 
|  |  | 
|  | rorl	$17,%r15d | 
|  | xorl	%r13d,%r12d | 
|  | xorl	%r14d,%r15d | 
|  | addl	44(%rsp),%r12d | 
|  |  | 
|  | addl	8(%rsp),%r12d | 
|  | movl	%ecx,%r13d | 
|  | addl	%r15d,%r12d | 
|  | movl	%r10d,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%edx,%r15d | 
|  |  | 
|  | xorl	%ecx,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r8d,%r15d | 
|  |  | 
|  | movl	%r12d,8(%rsp) | 
|  | xorl	%r10d,%r14d | 
|  | andl	%ecx,%r15d | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%r9d,%r12d | 
|  | xorl	%r8d,%r15d | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%ecx,%r13d | 
|  | addl	%r15d,%r12d | 
|  |  | 
|  | movl	%r10d,%r15d | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%r10d,%r14d | 
|  |  | 
|  | xorl	%r11d,%r15d | 
|  | rorl	$6,%r13d | 
|  | movl	%r11d,%r9d | 
|  |  | 
|  | andl	%r15d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%edi,%r9d | 
|  | addl	%r12d,%ebx | 
|  | addl	%r12d,%r9d | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | movl	16(%rsp),%r13d | 
|  | movl	4(%rsp),%edi | 
|  |  | 
|  | movl	%r13d,%r12d | 
|  | rorl	$11,%r13d | 
|  | addl	%r14d,%r9d | 
|  | movl	%edi,%r14d | 
|  | rorl	$2,%edi | 
|  |  | 
|  | xorl	%r12d,%r13d | 
|  | shrl	$3,%r12d | 
|  | rorl	$7,%r13d | 
|  | xorl	%r14d,%edi | 
|  | shrl	$10,%r14d | 
|  |  | 
|  | rorl	$17,%edi | 
|  | xorl	%r13d,%r12d | 
|  | xorl	%r14d,%edi | 
|  | addl	48(%rsp),%r12d | 
|  |  | 
|  | addl	12(%rsp),%r12d | 
|  | movl	%ebx,%r13d | 
|  | addl	%edi,%r12d | 
|  | movl	%r9d,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%ecx,%edi | 
|  |  | 
|  | xorl	%ebx,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%edx,%edi | 
|  |  | 
|  | movl	%r12d,12(%rsp) | 
|  | xorl	%r9d,%r14d | 
|  | andl	%ebx,%edi | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%r8d,%r12d | 
|  | xorl	%edx,%edi | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%ebx,%r13d | 
|  | addl	%edi,%r12d | 
|  |  | 
|  | movl	%r9d,%edi | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%r9d,%r14d | 
|  |  | 
|  | xorl	%r10d,%edi | 
|  | rorl	$6,%r13d | 
|  | movl	%r10d,%r8d | 
|  |  | 
|  | andl	%edi,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%r15d,%r8d | 
|  | addl	%r12d,%eax | 
|  | addl	%r12d,%r8d | 
|  |  | 
|  | leaq	20(%rbp),%rbp | 
|  | movl	20(%rsp),%r13d | 
|  | movl	8(%rsp),%r15d | 
|  |  | 
|  | movl	%r13d,%r12d | 
|  | rorl	$11,%r13d | 
|  | addl	%r14d,%r8d | 
|  | movl	%r15d,%r14d | 
|  | rorl	$2,%r15d | 
|  |  | 
|  | xorl	%r12d,%r13d | 
|  | shrl	$3,%r12d | 
|  | rorl	$7,%r13d | 
|  | xorl	%r14d,%r15d | 
|  | shrl	$10,%r14d | 
|  |  | 
|  | rorl	$17,%r15d | 
|  | xorl	%r13d,%r12d | 
|  | xorl	%r14d,%r15d | 
|  | addl	52(%rsp),%r12d | 
|  |  | 
|  | addl	16(%rsp),%r12d | 
|  | movl	%eax,%r13d | 
|  | addl	%r15d,%r12d | 
|  | movl	%r8d,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%ebx,%r15d | 
|  |  | 
|  | xorl	%eax,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%ecx,%r15d | 
|  |  | 
|  | movl	%r12d,16(%rsp) | 
|  | xorl	%r8d,%r14d | 
|  | andl	%eax,%r15d | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%edx,%r12d | 
|  | xorl	%ecx,%r15d | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%eax,%r13d | 
|  | addl	%r15d,%r12d | 
|  |  | 
|  | movl	%r8d,%r15d | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%r8d,%r14d | 
|  |  | 
|  | xorl	%r9d,%r15d | 
|  | rorl	$6,%r13d | 
|  | movl	%r9d,%edx | 
|  |  | 
|  | andl	%r15d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%edi,%edx | 
|  | addl	%r12d,%r11d | 
|  | addl	%r12d,%edx | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | movl	24(%rsp),%r13d | 
|  | movl	12(%rsp),%edi | 
|  |  | 
|  | movl	%r13d,%r12d | 
|  | rorl	$11,%r13d | 
|  | addl	%r14d,%edx | 
|  | movl	%edi,%r14d | 
|  | rorl	$2,%edi | 
|  |  | 
|  | xorl	%r12d,%r13d | 
|  | shrl	$3,%r12d | 
|  | rorl	$7,%r13d | 
|  | xorl	%r14d,%edi | 
|  | shrl	$10,%r14d | 
|  |  | 
|  | rorl	$17,%edi | 
|  | xorl	%r13d,%r12d | 
|  | xorl	%r14d,%edi | 
|  | addl	56(%rsp),%r12d | 
|  |  | 
|  | addl	20(%rsp),%r12d | 
|  | movl	%r11d,%r13d | 
|  | addl	%edi,%r12d | 
|  | movl	%edx,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%eax,%edi | 
|  |  | 
|  | xorl	%r11d,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%ebx,%edi | 
|  |  | 
|  | movl	%r12d,20(%rsp) | 
|  | xorl	%edx,%r14d | 
|  | andl	%r11d,%edi | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%ecx,%r12d | 
|  | xorl	%ebx,%edi | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%r11d,%r13d | 
|  | addl	%edi,%r12d | 
|  |  | 
|  | movl	%edx,%edi | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%edx,%r14d | 
|  |  | 
|  | xorl	%r8d,%edi | 
|  | rorl	$6,%r13d | 
|  | movl	%r8d,%ecx | 
|  |  | 
|  | andl	%edi,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%r15d,%ecx | 
|  | addl	%r12d,%r10d | 
|  | addl	%r12d,%ecx | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | movl	28(%rsp),%r13d | 
|  | movl	16(%rsp),%r15d | 
|  |  | 
|  | movl	%r13d,%r12d | 
|  | rorl	$11,%r13d | 
|  | addl	%r14d,%ecx | 
|  | movl	%r15d,%r14d | 
|  | rorl	$2,%r15d | 
|  |  | 
|  | xorl	%r12d,%r13d | 
|  | shrl	$3,%r12d | 
|  | rorl	$7,%r13d | 
|  | xorl	%r14d,%r15d | 
|  | shrl	$10,%r14d | 
|  |  | 
|  | rorl	$17,%r15d | 
|  | xorl	%r13d,%r12d | 
|  | xorl	%r14d,%r15d | 
|  | addl	60(%rsp),%r12d | 
|  |  | 
|  | addl	24(%rsp),%r12d | 
|  | movl	%r10d,%r13d | 
|  | addl	%r15d,%r12d | 
|  | movl	%ecx,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r11d,%r15d | 
|  |  | 
|  | xorl	%r10d,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%eax,%r15d | 
|  |  | 
|  | movl	%r12d,24(%rsp) | 
|  | xorl	%ecx,%r14d | 
|  | andl	%r10d,%r15d | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%ebx,%r12d | 
|  | xorl	%eax,%r15d | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%r10d,%r13d | 
|  | addl	%r15d,%r12d | 
|  |  | 
|  | movl	%ecx,%r15d | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%ecx,%r14d | 
|  |  | 
|  | xorl	%edx,%r15d | 
|  | rorl	$6,%r13d | 
|  | movl	%edx,%ebx | 
|  |  | 
|  | andl	%r15d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%edi,%ebx | 
|  | addl	%r12d,%r9d | 
|  | addl	%r12d,%ebx | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | movl	32(%rsp),%r13d | 
|  | movl	20(%rsp),%edi | 
|  |  | 
|  | movl	%r13d,%r12d | 
|  | rorl	$11,%r13d | 
|  | addl	%r14d,%ebx | 
|  | movl	%edi,%r14d | 
|  | rorl	$2,%edi | 
|  |  | 
|  | xorl	%r12d,%r13d | 
|  | shrl	$3,%r12d | 
|  | rorl	$7,%r13d | 
|  | xorl	%r14d,%edi | 
|  | shrl	$10,%r14d | 
|  |  | 
|  | rorl	$17,%edi | 
|  | xorl	%r13d,%r12d | 
|  | xorl	%r14d,%edi | 
|  | addl	0(%rsp),%r12d | 
|  |  | 
|  | addl	28(%rsp),%r12d | 
|  | movl	%r9d,%r13d | 
|  | addl	%edi,%r12d | 
|  | movl	%ebx,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r10d,%edi | 
|  |  | 
|  | xorl	%r9d,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r11d,%edi | 
|  |  | 
|  | movl	%r12d,28(%rsp) | 
|  | xorl	%ebx,%r14d | 
|  | andl	%r9d,%edi | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%eax,%r12d | 
|  | xorl	%r11d,%edi | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%r9d,%r13d | 
|  | addl	%edi,%r12d | 
|  |  | 
|  | movl	%ebx,%edi | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%ebx,%r14d | 
|  |  | 
|  | xorl	%ecx,%edi | 
|  | rorl	$6,%r13d | 
|  | movl	%ecx,%eax | 
|  |  | 
|  | andl	%edi,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%r15d,%eax | 
|  | addl	%r12d,%r8d | 
|  | addl	%r12d,%eax | 
|  |  | 
|  | leaq	20(%rbp),%rbp | 
|  | movl	36(%rsp),%r13d | 
|  | movl	24(%rsp),%r15d | 
|  |  | 
|  | movl	%r13d,%r12d | 
|  | rorl	$11,%r13d | 
|  | addl	%r14d,%eax | 
|  | movl	%r15d,%r14d | 
|  | rorl	$2,%r15d | 
|  |  | 
|  | xorl	%r12d,%r13d | 
|  | shrl	$3,%r12d | 
|  | rorl	$7,%r13d | 
|  | xorl	%r14d,%r15d | 
|  | shrl	$10,%r14d | 
|  |  | 
|  | rorl	$17,%r15d | 
|  | xorl	%r13d,%r12d | 
|  | xorl	%r14d,%r15d | 
|  | addl	4(%rsp),%r12d | 
|  |  | 
|  | addl	32(%rsp),%r12d | 
|  | movl	%r8d,%r13d | 
|  | addl	%r15d,%r12d | 
|  | movl	%eax,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r9d,%r15d | 
|  |  | 
|  | xorl	%r8d,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r10d,%r15d | 
|  |  | 
|  | movl	%r12d,32(%rsp) | 
|  | xorl	%eax,%r14d | 
|  | andl	%r8d,%r15d | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%r11d,%r12d | 
|  | xorl	%r10d,%r15d | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%r8d,%r13d | 
|  | addl	%r15d,%r12d | 
|  |  | 
|  | movl	%eax,%r15d | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%eax,%r14d | 
|  |  | 
|  | xorl	%ebx,%r15d | 
|  | rorl	$6,%r13d | 
|  | movl	%ebx,%r11d | 
|  |  | 
|  | andl	%r15d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%edi,%r11d | 
|  | addl	%r12d,%edx | 
|  | addl	%r12d,%r11d | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | movl	40(%rsp),%r13d | 
|  | movl	28(%rsp),%edi | 
|  |  | 
|  | movl	%r13d,%r12d | 
|  | rorl	$11,%r13d | 
|  | addl	%r14d,%r11d | 
|  | movl	%edi,%r14d | 
|  | rorl	$2,%edi | 
|  |  | 
|  | xorl	%r12d,%r13d | 
|  | shrl	$3,%r12d | 
|  | rorl	$7,%r13d | 
|  | xorl	%r14d,%edi | 
|  | shrl	$10,%r14d | 
|  |  | 
|  | rorl	$17,%edi | 
|  | xorl	%r13d,%r12d | 
|  | xorl	%r14d,%edi | 
|  | addl	8(%rsp),%r12d | 
|  |  | 
|  | addl	36(%rsp),%r12d | 
|  | movl	%edx,%r13d | 
|  | addl	%edi,%r12d | 
|  | movl	%r11d,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r8d,%edi | 
|  |  | 
|  | xorl	%edx,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r9d,%edi | 
|  |  | 
|  | movl	%r12d,36(%rsp) | 
|  | xorl	%r11d,%r14d | 
|  | andl	%edx,%edi | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%r10d,%r12d | 
|  | xorl	%r9d,%edi | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%edx,%r13d | 
|  | addl	%edi,%r12d | 
|  |  | 
|  | movl	%r11d,%edi | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%r11d,%r14d | 
|  |  | 
|  | xorl	%eax,%edi | 
|  | rorl	$6,%r13d | 
|  | movl	%eax,%r10d | 
|  |  | 
|  | andl	%edi,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%r15d,%r10d | 
|  | addl	%r12d,%ecx | 
|  | addl	%r12d,%r10d | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | movl	44(%rsp),%r13d | 
|  | movl	32(%rsp),%r15d | 
|  |  | 
|  | movl	%r13d,%r12d | 
|  | rorl	$11,%r13d | 
|  | addl	%r14d,%r10d | 
|  | movl	%r15d,%r14d | 
|  | rorl	$2,%r15d | 
|  |  | 
|  | xorl	%r12d,%r13d | 
|  | shrl	$3,%r12d | 
|  | rorl	$7,%r13d | 
|  | xorl	%r14d,%r15d | 
|  | shrl	$10,%r14d | 
|  |  | 
|  | rorl	$17,%r15d | 
|  | xorl	%r13d,%r12d | 
|  | xorl	%r14d,%r15d | 
|  | addl	12(%rsp),%r12d | 
|  |  | 
|  | addl	40(%rsp),%r12d | 
|  | movl	%ecx,%r13d | 
|  | addl	%r15d,%r12d | 
|  | movl	%r10d,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%edx,%r15d | 
|  |  | 
|  | xorl	%ecx,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r8d,%r15d | 
|  |  | 
|  | movl	%r12d,40(%rsp) | 
|  | xorl	%r10d,%r14d | 
|  | andl	%ecx,%r15d | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%r9d,%r12d | 
|  | xorl	%r8d,%r15d | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%ecx,%r13d | 
|  | addl	%r15d,%r12d | 
|  |  | 
|  | movl	%r10d,%r15d | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%r10d,%r14d | 
|  |  | 
|  | xorl	%r11d,%r15d | 
|  | rorl	$6,%r13d | 
|  | movl	%r11d,%r9d | 
|  |  | 
|  | andl	%r15d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%edi,%r9d | 
|  | addl	%r12d,%ebx | 
|  | addl	%r12d,%r9d | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | movl	48(%rsp),%r13d | 
|  | movl	36(%rsp),%edi | 
|  |  | 
|  | movl	%r13d,%r12d | 
|  | rorl	$11,%r13d | 
|  | addl	%r14d,%r9d | 
|  | movl	%edi,%r14d | 
|  | rorl	$2,%edi | 
|  |  | 
|  | xorl	%r12d,%r13d | 
|  | shrl	$3,%r12d | 
|  | rorl	$7,%r13d | 
|  | xorl	%r14d,%edi | 
|  | shrl	$10,%r14d | 
|  |  | 
|  | rorl	$17,%edi | 
|  | xorl	%r13d,%r12d | 
|  | xorl	%r14d,%edi | 
|  | addl	16(%rsp),%r12d | 
|  |  | 
|  | addl	44(%rsp),%r12d | 
|  | movl	%ebx,%r13d | 
|  | addl	%edi,%r12d | 
|  | movl	%r9d,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%ecx,%edi | 
|  |  | 
|  | xorl	%ebx,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%edx,%edi | 
|  |  | 
|  | movl	%r12d,44(%rsp) | 
|  | xorl	%r9d,%r14d | 
|  | andl	%ebx,%edi | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%r8d,%r12d | 
|  | xorl	%edx,%edi | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%ebx,%r13d | 
|  | addl	%edi,%r12d | 
|  |  | 
|  | movl	%r9d,%edi | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%r9d,%r14d | 
|  |  | 
|  | xorl	%r10d,%edi | 
|  | rorl	$6,%r13d | 
|  | movl	%r10d,%r8d | 
|  |  | 
|  | andl	%edi,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%r15d,%r8d | 
|  | addl	%r12d,%eax | 
|  | addl	%r12d,%r8d | 
|  |  | 
|  | leaq	20(%rbp),%rbp | 
|  | movl	52(%rsp),%r13d | 
|  | movl	40(%rsp),%r15d | 
|  |  | 
|  | movl	%r13d,%r12d | 
|  | rorl	$11,%r13d | 
|  | addl	%r14d,%r8d | 
|  | movl	%r15d,%r14d | 
|  | rorl	$2,%r15d | 
|  |  | 
|  | xorl	%r12d,%r13d | 
|  | shrl	$3,%r12d | 
|  | rorl	$7,%r13d | 
|  | xorl	%r14d,%r15d | 
|  | shrl	$10,%r14d | 
|  |  | 
|  | rorl	$17,%r15d | 
|  | xorl	%r13d,%r12d | 
|  | xorl	%r14d,%r15d | 
|  | addl	20(%rsp),%r12d | 
|  |  | 
|  | addl	48(%rsp),%r12d | 
|  | movl	%eax,%r13d | 
|  | addl	%r15d,%r12d | 
|  | movl	%r8d,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%ebx,%r15d | 
|  |  | 
|  | xorl	%eax,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%ecx,%r15d | 
|  |  | 
|  | movl	%r12d,48(%rsp) | 
|  | xorl	%r8d,%r14d | 
|  | andl	%eax,%r15d | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%edx,%r12d | 
|  | xorl	%ecx,%r15d | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%eax,%r13d | 
|  | addl	%r15d,%r12d | 
|  |  | 
|  | movl	%r8d,%r15d | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%r8d,%r14d | 
|  |  | 
|  | xorl	%r9d,%r15d | 
|  | rorl	$6,%r13d | 
|  | movl	%r9d,%edx | 
|  |  | 
|  | andl	%r15d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%edi,%edx | 
|  | addl	%r12d,%r11d | 
|  | addl	%r12d,%edx | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | movl	56(%rsp),%r13d | 
|  | movl	44(%rsp),%edi | 
|  |  | 
|  | movl	%r13d,%r12d | 
|  | rorl	$11,%r13d | 
|  | addl	%r14d,%edx | 
|  | movl	%edi,%r14d | 
|  | rorl	$2,%edi | 
|  |  | 
|  | xorl	%r12d,%r13d | 
|  | shrl	$3,%r12d | 
|  | rorl	$7,%r13d | 
|  | xorl	%r14d,%edi | 
|  | shrl	$10,%r14d | 
|  |  | 
|  | rorl	$17,%edi | 
|  | xorl	%r13d,%r12d | 
|  | xorl	%r14d,%edi | 
|  | addl	24(%rsp),%r12d | 
|  |  | 
|  | addl	52(%rsp),%r12d | 
|  | movl	%r11d,%r13d | 
|  | addl	%edi,%r12d | 
|  | movl	%edx,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%eax,%edi | 
|  |  | 
|  | xorl	%r11d,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%ebx,%edi | 
|  |  | 
|  | movl	%r12d,52(%rsp) | 
|  | xorl	%edx,%r14d | 
|  | andl	%r11d,%edi | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%ecx,%r12d | 
|  | xorl	%ebx,%edi | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%r11d,%r13d | 
|  | addl	%edi,%r12d | 
|  |  | 
|  | movl	%edx,%edi | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%edx,%r14d | 
|  |  | 
|  | xorl	%r8d,%edi | 
|  | rorl	$6,%r13d | 
|  | movl	%r8d,%ecx | 
|  |  | 
|  | andl	%edi,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%r15d,%ecx | 
|  | addl	%r12d,%r10d | 
|  | addl	%r12d,%ecx | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | movl	60(%rsp),%r13d | 
|  | movl	48(%rsp),%r15d | 
|  |  | 
|  | movl	%r13d,%r12d | 
|  | rorl	$11,%r13d | 
|  | addl	%r14d,%ecx | 
|  | movl	%r15d,%r14d | 
|  | rorl	$2,%r15d | 
|  |  | 
|  | xorl	%r12d,%r13d | 
|  | shrl	$3,%r12d | 
|  | rorl	$7,%r13d | 
|  | xorl	%r14d,%r15d | 
|  | shrl	$10,%r14d | 
|  |  | 
|  | rorl	$17,%r15d | 
|  | xorl	%r13d,%r12d | 
|  | xorl	%r14d,%r15d | 
|  | addl	28(%rsp),%r12d | 
|  |  | 
|  | addl	56(%rsp),%r12d | 
|  | movl	%r10d,%r13d | 
|  | addl	%r15d,%r12d | 
|  | movl	%ecx,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r11d,%r15d | 
|  |  | 
|  | xorl	%r10d,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%eax,%r15d | 
|  |  | 
|  | movl	%r12d,56(%rsp) | 
|  | xorl	%ecx,%r14d | 
|  | andl	%r10d,%r15d | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%ebx,%r12d | 
|  | xorl	%eax,%r15d | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%r10d,%r13d | 
|  | addl	%r15d,%r12d | 
|  |  | 
|  | movl	%ecx,%r15d | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%ecx,%r14d | 
|  |  | 
|  | xorl	%edx,%r15d | 
|  | rorl	$6,%r13d | 
|  | movl	%edx,%ebx | 
|  |  | 
|  | andl	%r15d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%edi,%ebx | 
|  | addl	%r12d,%r9d | 
|  | addl	%r12d,%ebx | 
|  |  | 
|  | leaq	4(%rbp),%rbp | 
|  | movl	0(%rsp),%r13d | 
|  | movl	52(%rsp),%edi | 
|  |  | 
|  | movl	%r13d,%r12d | 
|  | rorl	$11,%r13d | 
|  | addl	%r14d,%ebx | 
|  | movl	%edi,%r14d | 
|  | rorl	$2,%edi | 
|  |  | 
|  | xorl	%r12d,%r13d | 
|  | shrl	$3,%r12d | 
|  | rorl	$7,%r13d | 
|  | xorl	%r14d,%edi | 
|  | shrl	$10,%r14d | 
|  |  | 
|  | rorl	$17,%edi | 
|  | xorl	%r13d,%r12d | 
|  | xorl	%r14d,%edi | 
|  | addl	32(%rsp),%r12d | 
|  |  | 
|  | addl	60(%rsp),%r12d | 
|  | movl	%r9d,%r13d | 
|  | addl	%edi,%r12d | 
|  | movl	%ebx,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r10d,%edi | 
|  |  | 
|  | xorl	%r9d,%r13d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r11d,%edi | 
|  |  | 
|  | movl	%r12d,60(%rsp) | 
|  | xorl	%ebx,%r14d | 
|  | andl	%r9d,%edi | 
|  |  | 
|  | rorl	$5,%r13d | 
|  | addl	%eax,%r12d | 
|  | xorl	%r11d,%edi | 
|  |  | 
|  | rorl	$11,%r14d | 
|  | xorl	%r9d,%r13d | 
|  | addl	%edi,%r12d | 
|  |  | 
|  | movl	%ebx,%edi | 
|  | addl	(%rbp),%r12d | 
|  | xorl	%ebx,%r14d | 
|  |  | 
|  | xorl	%ecx,%edi | 
|  | rorl	$6,%r13d | 
|  | movl	%ecx,%eax | 
|  |  | 
|  | andl	%edi,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%r13d,%r12d | 
|  |  | 
|  | xorl	%r15d,%eax | 
|  | addl	%r12d,%r8d | 
|  | addl	%r12d,%eax | 
|  |  | 
|  | leaq	20(%rbp),%rbp | 
|  | cmpb	$0,3(%rbp) | 
|  | jnz	.Lrounds_16_xx | 
|  |  | 
|  | movq	64+0(%rsp),%rdi | 
|  | addl	%r14d,%eax | 
|  | leaq	64(%rsi),%rsi | 
|  |  | 
|  | addl	0(%rdi),%eax | 
|  | addl	4(%rdi),%ebx | 
|  | addl	8(%rdi),%ecx | 
|  | addl	12(%rdi),%edx | 
|  | addl	16(%rdi),%r8d | 
|  | addl	20(%rdi),%r9d | 
|  | addl	24(%rdi),%r10d | 
|  | addl	28(%rdi),%r11d | 
|  |  | 
|  | cmpq	64+16(%rsp),%rsi | 
|  |  | 
|  | movl	%eax,0(%rdi) | 
|  | movl	%ebx,4(%rdi) | 
|  | movl	%ecx,8(%rdi) | 
|  | movl	%edx,12(%rdi) | 
|  | movl	%r8d,16(%rdi) | 
|  | movl	%r9d,20(%rdi) | 
|  | movl	%r10d,24(%rdi) | 
|  | movl	%r11d,28(%rdi) | 
|  | jb	.Lloop | 
|  |  | 
|  | movq	88(%rsp),%rsi | 
|  | .cfi_def_cfa	%rsi,8 | 
|  | movq	-48(%rsi),%r15 | 
|  | .cfi_restore	%r15 | 
|  | movq	-40(%rsi),%r14 | 
|  | .cfi_restore	%r14 | 
|  | movq	-32(%rsi),%r13 | 
|  | .cfi_restore	%r13 | 
|  | movq	-24(%rsi),%r12 | 
|  | .cfi_restore	%r12 | 
|  | movq	-16(%rsi),%rbp | 
|  | .cfi_restore	%rbp | 
|  | movq	-8(%rsi),%rbx | 
|  | .cfi_restore	%rbx | 
|  | leaq	(%rsi),%rsp | 
|  | .cfi_def_cfa_register	%rsp | 
|  | .Lepilogue: | 
|  | ret | 
|  | .cfi_endproc | 
|  | .size	sha256_block_data_order_nohw,.-sha256_block_data_order_nohw | 
|  | .section	.rodata | 
|  | .align	64 | 
|  | .type	K256,@object | 
|  | K256: | 
|  | .long	0x428a2f98,0x71374491,0xb5c0fbcf,0xe9b5dba5 | 
|  | .long	0x428a2f98,0x71374491,0xb5c0fbcf,0xe9b5dba5 | 
|  | .long	0x3956c25b,0x59f111f1,0x923f82a4,0xab1c5ed5 | 
|  | .long	0x3956c25b,0x59f111f1,0x923f82a4,0xab1c5ed5 | 
|  | .long	0xd807aa98,0x12835b01,0x243185be,0x550c7dc3 | 
|  | .long	0xd807aa98,0x12835b01,0x243185be,0x550c7dc3 | 
|  | .long	0x72be5d74,0x80deb1fe,0x9bdc06a7,0xc19bf174 | 
|  | .long	0x72be5d74,0x80deb1fe,0x9bdc06a7,0xc19bf174 | 
|  | .long	0xe49b69c1,0xefbe4786,0x0fc19dc6,0x240ca1cc | 
|  | .long	0xe49b69c1,0xefbe4786,0x0fc19dc6,0x240ca1cc | 
|  | .long	0x2de92c6f,0x4a7484aa,0x5cb0a9dc,0x76f988da | 
|  | .long	0x2de92c6f,0x4a7484aa,0x5cb0a9dc,0x76f988da | 
|  | .long	0x983e5152,0xa831c66d,0xb00327c8,0xbf597fc7 | 
|  | .long	0x983e5152,0xa831c66d,0xb00327c8,0xbf597fc7 | 
|  | .long	0xc6e00bf3,0xd5a79147,0x06ca6351,0x14292967 | 
|  | .long	0xc6e00bf3,0xd5a79147,0x06ca6351,0x14292967 | 
|  | .long	0x27b70a85,0x2e1b2138,0x4d2c6dfc,0x53380d13 | 
|  | .long	0x27b70a85,0x2e1b2138,0x4d2c6dfc,0x53380d13 | 
|  | .long	0x650a7354,0x766a0abb,0x81c2c92e,0x92722c85 | 
|  | .long	0x650a7354,0x766a0abb,0x81c2c92e,0x92722c85 | 
|  | .long	0xa2bfe8a1,0xa81a664b,0xc24b8b70,0xc76c51a3 | 
|  | .long	0xa2bfe8a1,0xa81a664b,0xc24b8b70,0xc76c51a3 | 
|  | .long	0xd192e819,0xd6990624,0xf40e3585,0x106aa070 | 
|  | .long	0xd192e819,0xd6990624,0xf40e3585,0x106aa070 | 
|  | .long	0x19a4c116,0x1e376c08,0x2748774c,0x34b0bcb5 | 
|  | .long	0x19a4c116,0x1e376c08,0x2748774c,0x34b0bcb5 | 
|  | .long	0x391c0cb3,0x4ed8aa4a,0x5b9cca4f,0x682e6ff3 | 
|  | .long	0x391c0cb3,0x4ed8aa4a,0x5b9cca4f,0x682e6ff3 | 
|  | .long	0x748f82ee,0x78a5636f,0x84c87814,0x8cc70208 | 
|  | .long	0x748f82ee,0x78a5636f,0x84c87814,0x8cc70208 | 
|  | .long	0x90befffa,0xa4506ceb,0xbef9a3f7,0xc67178f2 | 
|  | .long	0x90befffa,0xa4506ceb,0xbef9a3f7,0xc67178f2 | 
|  |  | 
|  | .long	0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f | 
|  | .long	0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f | 
|  | .long	0x03020100,0x0b0a0908,0xffffffff,0xffffffff | 
|  | .long	0x03020100,0x0b0a0908,0xffffffff,0xffffffff | 
|  | .long	0xffffffff,0xffffffff,0x03020100,0x0b0a0908 | 
|  | .long	0xffffffff,0xffffffff,0x03020100,0x0b0a0908 | 
|  | .byte	83,72,65,50,53,54,32,98,108,111,99,107,32,116,114,97,110,115,102,111,114,109,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62,0 | 
|  | .text | 
|  | .globl	sha256_block_data_order_hw | 
|  | .hidden sha256_block_data_order_hw | 
|  | .type	sha256_block_data_order_hw,@function | 
|  | .align	64 | 
|  | sha256_block_data_order_hw: | 
|  | .cfi_startproc | 
|  | _CET_ENDBR | 
|  | leaq	K256+128(%rip),%rcx | 
|  | movdqu	(%rdi),%xmm1 | 
|  | movdqu	16(%rdi),%xmm2 | 
|  | movdqa	512-128(%rcx),%xmm7 | 
|  |  | 
|  | pshufd	$0x1b,%xmm1,%xmm0 | 
|  | pshufd	$0xb1,%xmm1,%xmm1 | 
|  | pshufd	$0x1b,%xmm2,%xmm2 | 
|  | movdqa	%xmm7,%xmm8 | 
|  | palignr	$8,%xmm2,%xmm1 | 
|  | punpcklqdq	%xmm0,%xmm2 | 
|  | jmp	.Loop_shaext | 
|  |  | 
|  | .align	16 | 
|  | .Loop_shaext: | 
|  | movdqu	(%rsi),%xmm3 | 
|  | movdqu	16(%rsi),%xmm4 | 
|  | movdqu	32(%rsi),%xmm5 | 
|  | pshufb	%xmm7,%xmm3 | 
|  | movdqu	48(%rsi),%xmm6 | 
|  |  | 
|  | movdqa	0-128(%rcx),%xmm0 | 
|  | paddd	%xmm3,%xmm0 | 
|  | pshufb	%xmm7,%xmm4 | 
|  | movdqa	%xmm2,%xmm10 | 
|  | sha256rnds2	%xmm1,%xmm2 | 
|  | pshufd	$0x0e,%xmm0,%xmm0 | 
|  | nop | 
|  | movdqa	%xmm1,%xmm9 | 
|  | sha256rnds2	%xmm2,%xmm1 | 
|  |  | 
|  | movdqa	32-128(%rcx),%xmm0 | 
|  | paddd	%xmm4,%xmm0 | 
|  | pshufb	%xmm7,%xmm5 | 
|  | sha256rnds2	%xmm1,%xmm2 | 
|  | pshufd	$0x0e,%xmm0,%xmm0 | 
|  | leaq	64(%rsi),%rsi | 
|  | sha256msg1	%xmm4,%xmm3 | 
|  | sha256rnds2	%xmm2,%xmm1 | 
|  |  | 
|  | movdqa	64-128(%rcx),%xmm0 | 
|  | paddd	%xmm5,%xmm0 | 
|  | pshufb	%xmm7,%xmm6 | 
|  | sha256rnds2	%xmm1,%xmm2 | 
|  | pshufd	$0x0e,%xmm0,%xmm0 | 
|  | movdqa	%xmm6,%xmm7 | 
|  | palignr	$4,%xmm5,%xmm7 | 
|  | nop | 
|  | paddd	%xmm7,%xmm3 | 
|  | sha256msg1	%xmm5,%xmm4 | 
|  | sha256rnds2	%xmm2,%xmm1 | 
|  |  | 
|  | movdqa	96-128(%rcx),%xmm0 | 
|  | paddd	%xmm6,%xmm0 | 
|  | sha256msg2	%xmm6,%xmm3 | 
|  | sha256rnds2	%xmm1,%xmm2 | 
|  | pshufd	$0x0e,%xmm0,%xmm0 | 
|  | movdqa	%xmm3,%xmm7 | 
|  | palignr	$4,%xmm6,%xmm7 | 
|  | nop | 
|  | paddd	%xmm7,%xmm4 | 
|  | sha256msg1	%xmm6,%xmm5 | 
|  | sha256rnds2	%xmm2,%xmm1 | 
|  | movdqa	128-128(%rcx),%xmm0 | 
|  | paddd	%xmm3,%xmm0 | 
|  | sha256msg2	%xmm3,%xmm4 | 
|  | sha256rnds2	%xmm1,%xmm2 | 
|  | pshufd	$0x0e,%xmm0,%xmm0 | 
|  | movdqa	%xmm4,%xmm7 | 
|  | palignr	$4,%xmm3,%xmm7 | 
|  | nop | 
|  | paddd	%xmm7,%xmm5 | 
|  | sha256msg1	%xmm3,%xmm6 | 
|  | sha256rnds2	%xmm2,%xmm1 | 
|  | movdqa	160-128(%rcx),%xmm0 | 
|  | paddd	%xmm4,%xmm0 | 
|  | sha256msg2	%xmm4,%xmm5 | 
|  | sha256rnds2	%xmm1,%xmm2 | 
|  | pshufd	$0x0e,%xmm0,%xmm0 | 
|  | movdqa	%xmm5,%xmm7 | 
|  | palignr	$4,%xmm4,%xmm7 | 
|  | nop | 
|  | paddd	%xmm7,%xmm6 | 
|  | sha256msg1	%xmm4,%xmm3 | 
|  | sha256rnds2	%xmm2,%xmm1 | 
|  | movdqa	192-128(%rcx),%xmm0 | 
|  | paddd	%xmm5,%xmm0 | 
|  | sha256msg2	%xmm5,%xmm6 | 
|  | sha256rnds2	%xmm1,%xmm2 | 
|  | pshufd	$0x0e,%xmm0,%xmm0 | 
|  | movdqa	%xmm6,%xmm7 | 
|  | palignr	$4,%xmm5,%xmm7 | 
|  | nop | 
|  | paddd	%xmm7,%xmm3 | 
|  | sha256msg1	%xmm5,%xmm4 | 
|  | sha256rnds2	%xmm2,%xmm1 | 
|  | movdqa	224-128(%rcx),%xmm0 | 
|  | paddd	%xmm6,%xmm0 | 
|  | sha256msg2	%xmm6,%xmm3 | 
|  | sha256rnds2	%xmm1,%xmm2 | 
|  | pshufd	$0x0e,%xmm0,%xmm0 | 
|  | movdqa	%xmm3,%xmm7 | 
|  | palignr	$4,%xmm6,%xmm7 | 
|  | nop | 
|  | paddd	%xmm7,%xmm4 | 
|  | sha256msg1	%xmm6,%xmm5 | 
|  | sha256rnds2	%xmm2,%xmm1 | 
|  | movdqa	256-128(%rcx),%xmm0 | 
|  | paddd	%xmm3,%xmm0 | 
|  | sha256msg2	%xmm3,%xmm4 | 
|  | sha256rnds2	%xmm1,%xmm2 | 
|  | pshufd	$0x0e,%xmm0,%xmm0 | 
|  | movdqa	%xmm4,%xmm7 | 
|  | palignr	$4,%xmm3,%xmm7 | 
|  | nop | 
|  | paddd	%xmm7,%xmm5 | 
|  | sha256msg1	%xmm3,%xmm6 | 
|  | sha256rnds2	%xmm2,%xmm1 | 
|  | movdqa	288-128(%rcx),%xmm0 | 
|  | paddd	%xmm4,%xmm0 | 
|  | sha256msg2	%xmm4,%xmm5 | 
|  | sha256rnds2	%xmm1,%xmm2 | 
|  | pshufd	$0x0e,%xmm0,%xmm0 | 
|  | movdqa	%xmm5,%xmm7 | 
|  | palignr	$4,%xmm4,%xmm7 | 
|  | nop | 
|  | paddd	%xmm7,%xmm6 | 
|  | sha256msg1	%xmm4,%xmm3 | 
|  | sha256rnds2	%xmm2,%xmm1 | 
|  | movdqa	320-128(%rcx),%xmm0 | 
|  | paddd	%xmm5,%xmm0 | 
|  | sha256msg2	%xmm5,%xmm6 | 
|  | sha256rnds2	%xmm1,%xmm2 | 
|  | pshufd	$0x0e,%xmm0,%xmm0 | 
|  | movdqa	%xmm6,%xmm7 | 
|  | palignr	$4,%xmm5,%xmm7 | 
|  | nop | 
|  | paddd	%xmm7,%xmm3 | 
|  | sha256msg1	%xmm5,%xmm4 | 
|  | sha256rnds2	%xmm2,%xmm1 | 
|  | movdqa	352-128(%rcx),%xmm0 | 
|  | paddd	%xmm6,%xmm0 | 
|  | sha256msg2	%xmm6,%xmm3 | 
|  | sha256rnds2	%xmm1,%xmm2 | 
|  | pshufd	$0x0e,%xmm0,%xmm0 | 
|  | movdqa	%xmm3,%xmm7 | 
|  | palignr	$4,%xmm6,%xmm7 | 
|  | nop | 
|  | paddd	%xmm7,%xmm4 | 
|  | sha256msg1	%xmm6,%xmm5 | 
|  | sha256rnds2	%xmm2,%xmm1 | 
|  | movdqa	384-128(%rcx),%xmm0 | 
|  | paddd	%xmm3,%xmm0 | 
|  | sha256msg2	%xmm3,%xmm4 | 
|  | sha256rnds2	%xmm1,%xmm2 | 
|  | pshufd	$0x0e,%xmm0,%xmm0 | 
|  | movdqa	%xmm4,%xmm7 | 
|  | palignr	$4,%xmm3,%xmm7 | 
|  | nop | 
|  | paddd	%xmm7,%xmm5 | 
|  | sha256msg1	%xmm3,%xmm6 | 
|  | sha256rnds2	%xmm2,%xmm1 | 
|  | movdqa	416-128(%rcx),%xmm0 | 
|  | paddd	%xmm4,%xmm0 | 
|  | sha256msg2	%xmm4,%xmm5 | 
|  | sha256rnds2	%xmm1,%xmm2 | 
|  | pshufd	$0x0e,%xmm0,%xmm0 | 
|  | movdqa	%xmm5,%xmm7 | 
|  | palignr	$4,%xmm4,%xmm7 | 
|  | sha256rnds2	%xmm2,%xmm1 | 
|  | paddd	%xmm7,%xmm6 | 
|  |  | 
|  | movdqa	448-128(%rcx),%xmm0 | 
|  | paddd	%xmm5,%xmm0 | 
|  | sha256rnds2	%xmm1,%xmm2 | 
|  | pshufd	$0x0e,%xmm0,%xmm0 | 
|  | sha256msg2	%xmm5,%xmm6 | 
|  | movdqa	%xmm8,%xmm7 | 
|  | sha256rnds2	%xmm2,%xmm1 | 
|  |  | 
|  | movdqa	480-128(%rcx),%xmm0 | 
|  | paddd	%xmm6,%xmm0 | 
|  | nop | 
|  | sha256rnds2	%xmm1,%xmm2 | 
|  | pshufd	$0x0e,%xmm0,%xmm0 | 
|  | decq	%rdx | 
|  | nop | 
|  | sha256rnds2	%xmm2,%xmm1 | 
|  |  | 
|  | paddd	%xmm10,%xmm2 | 
|  | paddd	%xmm9,%xmm1 | 
|  | jnz	.Loop_shaext | 
|  |  | 
|  | pshufd	$0xb1,%xmm2,%xmm2 | 
|  | pshufd	$0x1b,%xmm1,%xmm7 | 
|  | pshufd	$0xb1,%xmm1,%xmm1 | 
|  | punpckhqdq	%xmm2,%xmm1 | 
|  | palignr	$8,%xmm7,%xmm2 | 
|  |  | 
|  | movdqu	%xmm1,(%rdi) | 
|  | movdqu	%xmm2,16(%rdi) | 
|  | ret | 
|  | .cfi_endproc | 
|  | .size	sha256_block_data_order_hw,.-sha256_block_data_order_hw | 
|  | .globl	sha256_block_data_order_ssse3 | 
|  | .hidden sha256_block_data_order_ssse3 | 
|  | .type	sha256_block_data_order_ssse3,@function | 
|  | .align	64 | 
|  | sha256_block_data_order_ssse3: | 
|  | .cfi_startproc | 
|  | _CET_ENDBR | 
|  | movq	%rsp,%rax | 
|  | .cfi_def_cfa_register	%rax | 
|  | pushq	%rbx | 
|  | .cfi_offset	%rbx,-16 | 
|  | pushq	%rbp | 
|  | .cfi_offset	%rbp,-24 | 
|  | pushq	%r12 | 
|  | .cfi_offset	%r12,-32 | 
|  | pushq	%r13 | 
|  | .cfi_offset	%r13,-40 | 
|  | pushq	%r14 | 
|  | .cfi_offset	%r14,-48 | 
|  | pushq	%r15 | 
|  | .cfi_offset	%r15,-56 | 
|  | shlq	$4,%rdx | 
|  | subq	$96,%rsp | 
|  | leaq	(%rsi,%rdx,4),%rdx | 
|  | andq	$-64,%rsp | 
|  | movq	%rdi,64+0(%rsp) | 
|  | movq	%rsi,64+8(%rsp) | 
|  | movq	%rdx,64+16(%rsp) | 
|  | movq	%rax,88(%rsp) | 
|  | .cfi_escape	0x0f,0x06,0x77,0xd8,0x00,0x06,0x23,0x08 | 
|  | .Lprologue_ssse3: | 
|  |  | 
|  | movl	0(%rdi),%eax | 
|  | movl	4(%rdi),%ebx | 
|  | movl	8(%rdi),%ecx | 
|  | movl	12(%rdi),%edx | 
|  | movl	16(%rdi),%r8d | 
|  | movl	20(%rdi),%r9d | 
|  | movl	24(%rdi),%r10d | 
|  | movl	28(%rdi),%r11d | 
|  |  | 
|  |  | 
|  | jmp	.Lloop_ssse3 | 
|  | .align	16 | 
|  | .Lloop_ssse3: | 
|  | movdqa	K256+512(%rip),%xmm7 | 
|  | movdqu	0(%rsi),%xmm0 | 
|  | movdqu	16(%rsi),%xmm1 | 
|  | movdqu	32(%rsi),%xmm2 | 
|  | pshufb	%xmm7,%xmm0 | 
|  | movdqu	48(%rsi),%xmm3 | 
|  | leaq	K256(%rip),%rbp | 
|  | pshufb	%xmm7,%xmm1 | 
|  | movdqa	0(%rbp),%xmm4 | 
|  | movdqa	32(%rbp),%xmm5 | 
|  | pshufb	%xmm7,%xmm2 | 
|  | paddd	%xmm0,%xmm4 | 
|  | movdqa	64(%rbp),%xmm6 | 
|  | pshufb	%xmm7,%xmm3 | 
|  | movdqa	96(%rbp),%xmm7 | 
|  | paddd	%xmm1,%xmm5 | 
|  | paddd	%xmm2,%xmm6 | 
|  | paddd	%xmm3,%xmm7 | 
|  | movdqa	%xmm4,0(%rsp) | 
|  | movl	%eax,%r14d | 
|  | movdqa	%xmm5,16(%rsp) | 
|  | movl	%ebx,%edi | 
|  | movdqa	%xmm6,32(%rsp) | 
|  | xorl	%ecx,%edi | 
|  | movdqa	%xmm7,48(%rsp) | 
|  | movl	%r8d,%r13d | 
|  | jmp	.Lssse3_00_47 | 
|  |  | 
|  | .align	16 | 
|  | .Lssse3_00_47: | 
|  | subq	$-128,%rbp | 
|  | rorl	$14,%r13d | 
|  | movdqa	%xmm1,%xmm4 | 
|  | movl	%r14d,%eax | 
|  | movl	%r9d,%r12d | 
|  | movdqa	%xmm3,%xmm7 | 
|  | rorl	$9,%r14d | 
|  | xorl	%r8d,%r13d | 
|  | xorl	%r10d,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%eax,%r14d | 
|  | palignr	$4,%xmm0,%xmm4 | 
|  | andl	%r8d,%r12d | 
|  | xorl	%r8d,%r13d | 
|  | palignr	$4,%xmm2,%xmm7 | 
|  | addl	0(%rsp),%r11d | 
|  | movl	%eax,%r15d | 
|  | xorl	%r10d,%r12d | 
|  | rorl	$11,%r14d | 
|  | movdqa	%xmm4,%xmm5 | 
|  | xorl	%ebx,%r15d | 
|  | addl	%r12d,%r11d | 
|  | movdqa	%xmm4,%xmm6 | 
|  | rorl	$6,%r13d | 
|  | andl	%r15d,%edi | 
|  | psrld	$3,%xmm4 | 
|  | xorl	%eax,%r14d | 
|  | addl	%r13d,%r11d | 
|  | xorl	%ebx,%edi | 
|  | paddd	%xmm7,%xmm0 | 
|  | rorl	$2,%r14d | 
|  | addl	%r11d,%edx | 
|  | psrld	$7,%xmm6 | 
|  | addl	%edi,%r11d | 
|  | movl	%edx,%r13d | 
|  | pshufd	$250,%xmm3,%xmm7 | 
|  | addl	%r11d,%r14d | 
|  | rorl	$14,%r13d | 
|  | pslld	$14,%xmm5 | 
|  | movl	%r14d,%r11d | 
|  | movl	%r8d,%r12d | 
|  | pxor	%xmm6,%xmm4 | 
|  | rorl	$9,%r14d | 
|  | xorl	%edx,%r13d | 
|  | xorl	%r9d,%r12d | 
|  | rorl	$5,%r13d | 
|  | psrld	$11,%xmm6 | 
|  | xorl	%r11d,%r14d | 
|  | pxor	%xmm5,%xmm4 | 
|  | andl	%edx,%r12d | 
|  | xorl	%edx,%r13d | 
|  | pslld	$11,%xmm5 | 
|  | addl	4(%rsp),%r10d | 
|  | movl	%r11d,%edi | 
|  | pxor	%xmm6,%xmm4 | 
|  | xorl	%r9d,%r12d | 
|  | rorl	$11,%r14d | 
|  | movdqa	%xmm7,%xmm6 | 
|  | xorl	%eax,%edi | 
|  | addl	%r12d,%r10d | 
|  | pxor	%xmm5,%xmm4 | 
|  | rorl	$6,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%r11d,%r14d | 
|  | psrld	$10,%xmm7 | 
|  | addl	%r13d,%r10d | 
|  | xorl	%eax,%r15d | 
|  | paddd	%xmm4,%xmm0 | 
|  | rorl	$2,%r14d | 
|  | addl	%r10d,%ecx | 
|  | psrlq	$17,%xmm6 | 
|  | addl	%r15d,%r10d | 
|  | movl	%ecx,%r13d | 
|  | addl	%r10d,%r14d | 
|  | pxor	%xmm6,%xmm7 | 
|  | rorl	$14,%r13d | 
|  | movl	%r14d,%r10d | 
|  | movl	%edx,%r12d | 
|  | rorl	$9,%r14d | 
|  | psrlq	$2,%xmm6 | 
|  | xorl	%ecx,%r13d | 
|  | xorl	%r8d,%r12d | 
|  | pxor	%xmm6,%xmm7 | 
|  | rorl	$5,%r13d | 
|  | xorl	%r10d,%r14d | 
|  | andl	%ecx,%r12d | 
|  | pshufd	$128,%xmm7,%xmm7 | 
|  | xorl	%ecx,%r13d | 
|  | addl	8(%rsp),%r9d | 
|  | movl	%r10d,%r15d | 
|  | psrldq	$8,%xmm7 | 
|  | xorl	%r8d,%r12d | 
|  | rorl	$11,%r14d | 
|  | xorl	%r11d,%r15d | 
|  | addl	%r12d,%r9d | 
|  | rorl	$6,%r13d | 
|  | paddd	%xmm7,%xmm0 | 
|  | andl	%r15d,%edi | 
|  | xorl	%r10d,%r14d | 
|  | addl	%r13d,%r9d | 
|  | pshufd	$80,%xmm0,%xmm7 | 
|  | xorl	%r11d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r9d,%ebx | 
|  | movdqa	%xmm7,%xmm6 | 
|  | addl	%edi,%r9d | 
|  | movl	%ebx,%r13d | 
|  | psrld	$10,%xmm7 | 
|  | addl	%r9d,%r14d | 
|  | rorl	$14,%r13d | 
|  | psrlq	$17,%xmm6 | 
|  | movl	%r14d,%r9d | 
|  | movl	%ecx,%r12d | 
|  | pxor	%xmm6,%xmm7 | 
|  | rorl	$9,%r14d | 
|  | xorl	%ebx,%r13d | 
|  | xorl	%edx,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%r9d,%r14d | 
|  | psrlq	$2,%xmm6 | 
|  | andl	%ebx,%r12d | 
|  | xorl	%ebx,%r13d | 
|  | addl	12(%rsp),%r8d | 
|  | pxor	%xmm6,%xmm7 | 
|  | movl	%r9d,%edi | 
|  | xorl	%edx,%r12d | 
|  | rorl	$11,%r14d | 
|  | pshufd	$8,%xmm7,%xmm7 | 
|  | xorl	%r10d,%edi | 
|  | addl	%r12d,%r8d | 
|  | movdqa	0(%rbp),%xmm6 | 
|  | rorl	$6,%r13d | 
|  | andl	%edi,%r15d | 
|  | pslldq	$8,%xmm7 | 
|  | xorl	%r9d,%r14d | 
|  | addl	%r13d,%r8d | 
|  | xorl	%r10d,%r15d | 
|  | paddd	%xmm7,%xmm0 | 
|  | rorl	$2,%r14d | 
|  | addl	%r8d,%eax | 
|  | addl	%r15d,%r8d | 
|  | paddd	%xmm0,%xmm6 | 
|  | movl	%eax,%r13d | 
|  | addl	%r8d,%r14d | 
|  | movdqa	%xmm6,0(%rsp) | 
|  | rorl	$14,%r13d | 
|  | movdqa	%xmm2,%xmm4 | 
|  | movl	%r14d,%r8d | 
|  | movl	%ebx,%r12d | 
|  | movdqa	%xmm0,%xmm7 | 
|  | rorl	$9,%r14d | 
|  | xorl	%eax,%r13d | 
|  | xorl	%ecx,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%r8d,%r14d | 
|  | palignr	$4,%xmm1,%xmm4 | 
|  | andl	%eax,%r12d | 
|  | xorl	%eax,%r13d | 
|  | palignr	$4,%xmm3,%xmm7 | 
|  | addl	16(%rsp),%edx | 
|  | movl	%r8d,%r15d | 
|  | xorl	%ecx,%r12d | 
|  | rorl	$11,%r14d | 
|  | movdqa	%xmm4,%xmm5 | 
|  | xorl	%r9d,%r15d | 
|  | addl	%r12d,%edx | 
|  | movdqa	%xmm4,%xmm6 | 
|  | rorl	$6,%r13d | 
|  | andl	%r15d,%edi | 
|  | psrld	$3,%xmm4 | 
|  | xorl	%r8d,%r14d | 
|  | addl	%r13d,%edx | 
|  | xorl	%r9d,%edi | 
|  | paddd	%xmm7,%xmm1 | 
|  | rorl	$2,%r14d | 
|  | addl	%edx,%r11d | 
|  | psrld	$7,%xmm6 | 
|  | addl	%edi,%edx | 
|  | movl	%r11d,%r13d | 
|  | pshufd	$250,%xmm0,%xmm7 | 
|  | addl	%edx,%r14d | 
|  | rorl	$14,%r13d | 
|  | pslld	$14,%xmm5 | 
|  | movl	%r14d,%edx | 
|  | movl	%eax,%r12d | 
|  | pxor	%xmm6,%xmm4 | 
|  | rorl	$9,%r14d | 
|  | xorl	%r11d,%r13d | 
|  | xorl	%ebx,%r12d | 
|  | rorl	$5,%r13d | 
|  | psrld	$11,%xmm6 | 
|  | xorl	%edx,%r14d | 
|  | pxor	%xmm5,%xmm4 | 
|  | andl	%r11d,%r12d | 
|  | xorl	%r11d,%r13d | 
|  | pslld	$11,%xmm5 | 
|  | addl	20(%rsp),%ecx | 
|  | movl	%edx,%edi | 
|  | pxor	%xmm6,%xmm4 | 
|  | xorl	%ebx,%r12d | 
|  | rorl	$11,%r14d | 
|  | movdqa	%xmm7,%xmm6 | 
|  | xorl	%r8d,%edi | 
|  | addl	%r12d,%ecx | 
|  | pxor	%xmm5,%xmm4 | 
|  | rorl	$6,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%edx,%r14d | 
|  | psrld	$10,%xmm7 | 
|  | addl	%r13d,%ecx | 
|  | xorl	%r8d,%r15d | 
|  | paddd	%xmm4,%xmm1 | 
|  | rorl	$2,%r14d | 
|  | addl	%ecx,%r10d | 
|  | psrlq	$17,%xmm6 | 
|  | addl	%r15d,%ecx | 
|  | movl	%r10d,%r13d | 
|  | addl	%ecx,%r14d | 
|  | pxor	%xmm6,%xmm7 | 
|  | rorl	$14,%r13d | 
|  | movl	%r14d,%ecx | 
|  | movl	%r11d,%r12d | 
|  | rorl	$9,%r14d | 
|  | psrlq	$2,%xmm6 | 
|  | xorl	%r10d,%r13d | 
|  | xorl	%eax,%r12d | 
|  | pxor	%xmm6,%xmm7 | 
|  | rorl	$5,%r13d | 
|  | xorl	%ecx,%r14d | 
|  | andl	%r10d,%r12d | 
|  | pshufd	$128,%xmm7,%xmm7 | 
|  | xorl	%r10d,%r13d | 
|  | addl	24(%rsp),%ebx | 
|  | movl	%ecx,%r15d | 
|  | psrldq	$8,%xmm7 | 
|  | xorl	%eax,%r12d | 
|  | rorl	$11,%r14d | 
|  | xorl	%edx,%r15d | 
|  | addl	%r12d,%ebx | 
|  | rorl	$6,%r13d | 
|  | paddd	%xmm7,%xmm1 | 
|  | andl	%r15d,%edi | 
|  | xorl	%ecx,%r14d | 
|  | addl	%r13d,%ebx | 
|  | pshufd	$80,%xmm1,%xmm7 | 
|  | xorl	%edx,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%ebx,%r9d | 
|  | movdqa	%xmm7,%xmm6 | 
|  | addl	%edi,%ebx | 
|  | movl	%r9d,%r13d | 
|  | psrld	$10,%xmm7 | 
|  | addl	%ebx,%r14d | 
|  | rorl	$14,%r13d | 
|  | psrlq	$17,%xmm6 | 
|  | movl	%r14d,%ebx | 
|  | movl	%r10d,%r12d | 
|  | pxor	%xmm6,%xmm7 | 
|  | rorl	$9,%r14d | 
|  | xorl	%r9d,%r13d | 
|  | xorl	%r11d,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%ebx,%r14d | 
|  | psrlq	$2,%xmm6 | 
|  | andl	%r9d,%r12d | 
|  | xorl	%r9d,%r13d | 
|  | addl	28(%rsp),%eax | 
|  | pxor	%xmm6,%xmm7 | 
|  | movl	%ebx,%edi | 
|  | xorl	%r11d,%r12d | 
|  | rorl	$11,%r14d | 
|  | pshufd	$8,%xmm7,%xmm7 | 
|  | xorl	%ecx,%edi | 
|  | addl	%r12d,%eax | 
|  | movdqa	32(%rbp),%xmm6 | 
|  | rorl	$6,%r13d | 
|  | andl	%edi,%r15d | 
|  | pslldq	$8,%xmm7 | 
|  | xorl	%ebx,%r14d | 
|  | addl	%r13d,%eax | 
|  | xorl	%ecx,%r15d | 
|  | paddd	%xmm7,%xmm1 | 
|  | rorl	$2,%r14d | 
|  | addl	%eax,%r8d | 
|  | addl	%r15d,%eax | 
|  | paddd	%xmm1,%xmm6 | 
|  | movl	%r8d,%r13d | 
|  | addl	%eax,%r14d | 
|  | movdqa	%xmm6,16(%rsp) | 
|  | rorl	$14,%r13d | 
|  | movdqa	%xmm3,%xmm4 | 
|  | movl	%r14d,%eax | 
|  | movl	%r9d,%r12d | 
|  | movdqa	%xmm1,%xmm7 | 
|  | rorl	$9,%r14d | 
|  | xorl	%r8d,%r13d | 
|  | xorl	%r10d,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%eax,%r14d | 
|  | palignr	$4,%xmm2,%xmm4 | 
|  | andl	%r8d,%r12d | 
|  | xorl	%r8d,%r13d | 
|  | palignr	$4,%xmm0,%xmm7 | 
|  | addl	32(%rsp),%r11d | 
|  | movl	%eax,%r15d | 
|  | xorl	%r10d,%r12d | 
|  | rorl	$11,%r14d | 
|  | movdqa	%xmm4,%xmm5 | 
|  | xorl	%ebx,%r15d | 
|  | addl	%r12d,%r11d | 
|  | movdqa	%xmm4,%xmm6 | 
|  | rorl	$6,%r13d | 
|  | andl	%r15d,%edi | 
|  | psrld	$3,%xmm4 | 
|  | xorl	%eax,%r14d | 
|  | addl	%r13d,%r11d | 
|  | xorl	%ebx,%edi | 
|  | paddd	%xmm7,%xmm2 | 
|  | rorl	$2,%r14d | 
|  | addl	%r11d,%edx | 
|  | psrld	$7,%xmm6 | 
|  | addl	%edi,%r11d | 
|  | movl	%edx,%r13d | 
|  | pshufd	$250,%xmm1,%xmm7 | 
|  | addl	%r11d,%r14d | 
|  | rorl	$14,%r13d | 
|  | pslld	$14,%xmm5 | 
|  | movl	%r14d,%r11d | 
|  | movl	%r8d,%r12d | 
|  | pxor	%xmm6,%xmm4 | 
|  | rorl	$9,%r14d | 
|  | xorl	%edx,%r13d | 
|  | xorl	%r9d,%r12d | 
|  | rorl	$5,%r13d | 
|  | psrld	$11,%xmm6 | 
|  | xorl	%r11d,%r14d | 
|  | pxor	%xmm5,%xmm4 | 
|  | andl	%edx,%r12d | 
|  | xorl	%edx,%r13d | 
|  | pslld	$11,%xmm5 | 
|  | addl	36(%rsp),%r10d | 
|  | movl	%r11d,%edi | 
|  | pxor	%xmm6,%xmm4 | 
|  | xorl	%r9d,%r12d | 
|  | rorl	$11,%r14d | 
|  | movdqa	%xmm7,%xmm6 | 
|  | xorl	%eax,%edi | 
|  | addl	%r12d,%r10d | 
|  | pxor	%xmm5,%xmm4 | 
|  | rorl	$6,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%r11d,%r14d | 
|  | psrld	$10,%xmm7 | 
|  | addl	%r13d,%r10d | 
|  | xorl	%eax,%r15d | 
|  | paddd	%xmm4,%xmm2 | 
|  | rorl	$2,%r14d | 
|  | addl	%r10d,%ecx | 
|  | psrlq	$17,%xmm6 | 
|  | addl	%r15d,%r10d | 
|  | movl	%ecx,%r13d | 
|  | addl	%r10d,%r14d | 
|  | pxor	%xmm6,%xmm7 | 
|  | rorl	$14,%r13d | 
|  | movl	%r14d,%r10d | 
|  | movl	%edx,%r12d | 
|  | rorl	$9,%r14d | 
|  | psrlq	$2,%xmm6 | 
|  | xorl	%ecx,%r13d | 
|  | xorl	%r8d,%r12d | 
|  | pxor	%xmm6,%xmm7 | 
|  | rorl	$5,%r13d | 
|  | xorl	%r10d,%r14d | 
|  | andl	%ecx,%r12d | 
|  | pshufd	$128,%xmm7,%xmm7 | 
|  | xorl	%ecx,%r13d | 
|  | addl	40(%rsp),%r9d | 
|  | movl	%r10d,%r15d | 
|  | psrldq	$8,%xmm7 | 
|  | xorl	%r8d,%r12d | 
|  | rorl	$11,%r14d | 
|  | xorl	%r11d,%r15d | 
|  | addl	%r12d,%r9d | 
|  | rorl	$6,%r13d | 
|  | paddd	%xmm7,%xmm2 | 
|  | andl	%r15d,%edi | 
|  | xorl	%r10d,%r14d | 
|  | addl	%r13d,%r9d | 
|  | pshufd	$80,%xmm2,%xmm7 | 
|  | xorl	%r11d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r9d,%ebx | 
|  | movdqa	%xmm7,%xmm6 | 
|  | addl	%edi,%r9d | 
|  | movl	%ebx,%r13d | 
|  | psrld	$10,%xmm7 | 
|  | addl	%r9d,%r14d | 
|  | rorl	$14,%r13d | 
|  | psrlq	$17,%xmm6 | 
|  | movl	%r14d,%r9d | 
|  | movl	%ecx,%r12d | 
|  | pxor	%xmm6,%xmm7 | 
|  | rorl	$9,%r14d | 
|  | xorl	%ebx,%r13d | 
|  | xorl	%edx,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%r9d,%r14d | 
|  | psrlq	$2,%xmm6 | 
|  | andl	%ebx,%r12d | 
|  | xorl	%ebx,%r13d | 
|  | addl	44(%rsp),%r8d | 
|  | pxor	%xmm6,%xmm7 | 
|  | movl	%r9d,%edi | 
|  | xorl	%edx,%r12d | 
|  | rorl	$11,%r14d | 
|  | pshufd	$8,%xmm7,%xmm7 | 
|  | xorl	%r10d,%edi | 
|  | addl	%r12d,%r8d | 
|  | movdqa	64(%rbp),%xmm6 | 
|  | rorl	$6,%r13d | 
|  | andl	%edi,%r15d | 
|  | pslldq	$8,%xmm7 | 
|  | xorl	%r9d,%r14d | 
|  | addl	%r13d,%r8d | 
|  | xorl	%r10d,%r15d | 
|  | paddd	%xmm7,%xmm2 | 
|  | rorl	$2,%r14d | 
|  | addl	%r8d,%eax | 
|  | addl	%r15d,%r8d | 
|  | paddd	%xmm2,%xmm6 | 
|  | movl	%eax,%r13d | 
|  | addl	%r8d,%r14d | 
|  | movdqa	%xmm6,32(%rsp) | 
|  | rorl	$14,%r13d | 
|  | movdqa	%xmm0,%xmm4 | 
|  | movl	%r14d,%r8d | 
|  | movl	%ebx,%r12d | 
|  | movdqa	%xmm2,%xmm7 | 
|  | rorl	$9,%r14d | 
|  | xorl	%eax,%r13d | 
|  | xorl	%ecx,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%r8d,%r14d | 
|  | palignr	$4,%xmm3,%xmm4 | 
|  | andl	%eax,%r12d | 
|  | xorl	%eax,%r13d | 
|  | palignr	$4,%xmm1,%xmm7 | 
|  | addl	48(%rsp),%edx | 
|  | movl	%r8d,%r15d | 
|  | xorl	%ecx,%r12d | 
|  | rorl	$11,%r14d | 
|  | movdqa	%xmm4,%xmm5 | 
|  | xorl	%r9d,%r15d | 
|  | addl	%r12d,%edx | 
|  | movdqa	%xmm4,%xmm6 | 
|  | rorl	$6,%r13d | 
|  | andl	%r15d,%edi | 
|  | psrld	$3,%xmm4 | 
|  | xorl	%r8d,%r14d | 
|  | addl	%r13d,%edx | 
|  | xorl	%r9d,%edi | 
|  | paddd	%xmm7,%xmm3 | 
|  | rorl	$2,%r14d | 
|  | addl	%edx,%r11d | 
|  | psrld	$7,%xmm6 | 
|  | addl	%edi,%edx | 
|  | movl	%r11d,%r13d | 
|  | pshufd	$250,%xmm2,%xmm7 | 
|  | addl	%edx,%r14d | 
|  | rorl	$14,%r13d | 
|  | pslld	$14,%xmm5 | 
|  | movl	%r14d,%edx | 
|  | movl	%eax,%r12d | 
|  | pxor	%xmm6,%xmm4 | 
|  | rorl	$9,%r14d | 
|  | xorl	%r11d,%r13d | 
|  | xorl	%ebx,%r12d | 
|  | rorl	$5,%r13d | 
|  | psrld	$11,%xmm6 | 
|  | xorl	%edx,%r14d | 
|  | pxor	%xmm5,%xmm4 | 
|  | andl	%r11d,%r12d | 
|  | xorl	%r11d,%r13d | 
|  | pslld	$11,%xmm5 | 
|  | addl	52(%rsp),%ecx | 
|  | movl	%edx,%edi | 
|  | pxor	%xmm6,%xmm4 | 
|  | xorl	%ebx,%r12d | 
|  | rorl	$11,%r14d | 
|  | movdqa	%xmm7,%xmm6 | 
|  | xorl	%r8d,%edi | 
|  | addl	%r12d,%ecx | 
|  | pxor	%xmm5,%xmm4 | 
|  | rorl	$6,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%edx,%r14d | 
|  | psrld	$10,%xmm7 | 
|  | addl	%r13d,%ecx | 
|  | xorl	%r8d,%r15d | 
|  | paddd	%xmm4,%xmm3 | 
|  | rorl	$2,%r14d | 
|  | addl	%ecx,%r10d | 
|  | psrlq	$17,%xmm6 | 
|  | addl	%r15d,%ecx | 
|  | movl	%r10d,%r13d | 
|  | addl	%ecx,%r14d | 
|  | pxor	%xmm6,%xmm7 | 
|  | rorl	$14,%r13d | 
|  | movl	%r14d,%ecx | 
|  | movl	%r11d,%r12d | 
|  | rorl	$9,%r14d | 
|  | psrlq	$2,%xmm6 | 
|  | xorl	%r10d,%r13d | 
|  | xorl	%eax,%r12d | 
|  | pxor	%xmm6,%xmm7 | 
|  | rorl	$5,%r13d | 
|  | xorl	%ecx,%r14d | 
|  | andl	%r10d,%r12d | 
|  | pshufd	$128,%xmm7,%xmm7 | 
|  | xorl	%r10d,%r13d | 
|  | addl	56(%rsp),%ebx | 
|  | movl	%ecx,%r15d | 
|  | psrldq	$8,%xmm7 | 
|  | xorl	%eax,%r12d | 
|  | rorl	$11,%r14d | 
|  | xorl	%edx,%r15d | 
|  | addl	%r12d,%ebx | 
|  | rorl	$6,%r13d | 
|  | paddd	%xmm7,%xmm3 | 
|  | andl	%r15d,%edi | 
|  | xorl	%ecx,%r14d | 
|  | addl	%r13d,%ebx | 
|  | pshufd	$80,%xmm3,%xmm7 | 
|  | xorl	%edx,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%ebx,%r9d | 
|  | movdqa	%xmm7,%xmm6 | 
|  | addl	%edi,%ebx | 
|  | movl	%r9d,%r13d | 
|  | psrld	$10,%xmm7 | 
|  | addl	%ebx,%r14d | 
|  | rorl	$14,%r13d | 
|  | psrlq	$17,%xmm6 | 
|  | movl	%r14d,%ebx | 
|  | movl	%r10d,%r12d | 
|  | pxor	%xmm6,%xmm7 | 
|  | rorl	$9,%r14d | 
|  | xorl	%r9d,%r13d | 
|  | xorl	%r11d,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%ebx,%r14d | 
|  | psrlq	$2,%xmm6 | 
|  | andl	%r9d,%r12d | 
|  | xorl	%r9d,%r13d | 
|  | addl	60(%rsp),%eax | 
|  | pxor	%xmm6,%xmm7 | 
|  | movl	%ebx,%edi | 
|  | xorl	%r11d,%r12d | 
|  | rorl	$11,%r14d | 
|  | pshufd	$8,%xmm7,%xmm7 | 
|  | xorl	%ecx,%edi | 
|  | addl	%r12d,%eax | 
|  | movdqa	96(%rbp),%xmm6 | 
|  | rorl	$6,%r13d | 
|  | andl	%edi,%r15d | 
|  | pslldq	$8,%xmm7 | 
|  | xorl	%ebx,%r14d | 
|  | addl	%r13d,%eax | 
|  | xorl	%ecx,%r15d | 
|  | paddd	%xmm7,%xmm3 | 
|  | rorl	$2,%r14d | 
|  | addl	%eax,%r8d | 
|  | addl	%r15d,%eax | 
|  | paddd	%xmm3,%xmm6 | 
|  | movl	%r8d,%r13d | 
|  | addl	%eax,%r14d | 
|  | movdqa	%xmm6,48(%rsp) | 
|  | cmpb	$0,131(%rbp) | 
|  | jne	.Lssse3_00_47 | 
|  | rorl	$14,%r13d | 
|  | movl	%r14d,%eax | 
|  | movl	%r9d,%r12d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r8d,%r13d | 
|  | xorl	%r10d,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%eax,%r14d | 
|  | andl	%r8d,%r12d | 
|  | xorl	%r8d,%r13d | 
|  | addl	0(%rsp),%r11d | 
|  | movl	%eax,%r15d | 
|  | xorl	%r10d,%r12d | 
|  | rorl	$11,%r14d | 
|  | xorl	%ebx,%r15d | 
|  | addl	%r12d,%r11d | 
|  | rorl	$6,%r13d | 
|  | andl	%r15d,%edi | 
|  | xorl	%eax,%r14d | 
|  | addl	%r13d,%r11d | 
|  | xorl	%ebx,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r11d,%edx | 
|  | addl	%edi,%r11d | 
|  | movl	%edx,%r13d | 
|  | addl	%r11d,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r14d,%r11d | 
|  | movl	%r8d,%r12d | 
|  | rorl	$9,%r14d | 
|  | xorl	%edx,%r13d | 
|  | xorl	%r9d,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%r11d,%r14d | 
|  | andl	%edx,%r12d | 
|  | xorl	%edx,%r13d | 
|  | addl	4(%rsp),%r10d | 
|  | movl	%r11d,%edi | 
|  | xorl	%r9d,%r12d | 
|  | rorl	$11,%r14d | 
|  | xorl	%eax,%edi | 
|  | addl	%r12d,%r10d | 
|  | rorl	$6,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%r11d,%r14d | 
|  | addl	%r13d,%r10d | 
|  | xorl	%eax,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%r10d,%ecx | 
|  | addl	%r15d,%r10d | 
|  | movl	%ecx,%r13d | 
|  | addl	%r10d,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r14d,%r10d | 
|  | movl	%edx,%r12d | 
|  | rorl	$9,%r14d | 
|  | xorl	%ecx,%r13d | 
|  | xorl	%r8d,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%r10d,%r14d | 
|  | andl	%ecx,%r12d | 
|  | xorl	%ecx,%r13d | 
|  | addl	8(%rsp),%r9d | 
|  | movl	%r10d,%r15d | 
|  | xorl	%r8d,%r12d | 
|  | rorl	$11,%r14d | 
|  | xorl	%r11d,%r15d | 
|  | addl	%r12d,%r9d | 
|  | rorl	$6,%r13d | 
|  | andl	%r15d,%edi | 
|  | xorl	%r10d,%r14d | 
|  | addl	%r13d,%r9d | 
|  | xorl	%r11d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r9d,%ebx | 
|  | addl	%edi,%r9d | 
|  | movl	%ebx,%r13d | 
|  | addl	%r9d,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r14d,%r9d | 
|  | movl	%ecx,%r12d | 
|  | rorl	$9,%r14d | 
|  | xorl	%ebx,%r13d | 
|  | xorl	%edx,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%r9d,%r14d | 
|  | andl	%ebx,%r12d | 
|  | xorl	%ebx,%r13d | 
|  | addl	12(%rsp),%r8d | 
|  | movl	%r9d,%edi | 
|  | xorl	%edx,%r12d | 
|  | rorl	$11,%r14d | 
|  | xorl	%r10d,%edi | 
|  | addl	%r12d,%r8d | 
|  | rorl	$6,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%r9d,%r14d | 
|  | addl	%r13d,%r8d | 
|  | xorl	%r10d,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%r8d,%eax | 
|  | addl	%r15d,%r8d | 
|  | movl	%eax,%r13d | 
|  | addl	%r8d,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r14d,%r8d | 
|  | movl	%ebx,%r12d | 
|  | rorl	$9,%r14d | 
|  | xorl	%eax,%r13d | 
|  | xorl	%ecx,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%r8d,%r14d | 
|  | andl	%eax,%r12d | 
|  | xorl	%eax,%r13d | 
|  | addl	16(%rsp),%edx | 
|  | movl	%r8d,%r15d | 
|  | xorl	%ecx,%r12d | 
|  | rorl	$11,%r14d | 
|  | xorl	%r9d,%r15d | 
|  | addl	%r12d,%edx | 
|  | rorl	$6,%r13d | 
|  | andl	%r15d,%edi | 
|  | xorl	%r8d,%r14d | 
|  | addl	%r13d,%edx | 
|  | xorl	%r9d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%edx,%r11d | 
|  | addl	%edi,%edx | 
|  | movl	%r11d,%r13d | 
|  | addl	%edx,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r14d,%edx | 
|  | movl	%eax,%r12d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r11d,%r13d | 
|  | xorl	%ebx,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%edx,%r14d | 
|  | andl	%r11d,%r12d | 
|  | xorl	%r11d,%r13d | 
|  | addl	20(%rsp),%ecx | 
|  | movl	%edx,%edi | 
|  | xorl	%ebx,%r12d | 
|  | rorl	$11,%r14d | 
|  | xorl	%r8d,%edi | 
|  | addl	%r12d,%ecx | 
|  | rorl	$6,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%edx,%r14d | 
|  | addl	%r13d,%ecx | 
|  | xorl	%r8d,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%ecx,%r10d | 
|  | addl	%r15d,%ecx | 
|  | movl	%r10d,%r13d | 
|  | addl	%ecx,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r14d,%ecx | 
|  | movl	%r11d,%r12d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r10d,%r13d | 
|  | xorl	%eax,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%ecx,%r14d | 
|  | andl	%r10d,%r12d | 
|  | xorl	%r10d,%r13d | 
|  | addl	24(%rsp),%ebx | 
|  | movl	%ecx,%r15d | 
|  | xorl	%eax,%r12d | 
|  | rorl	$11,%r14d | 
|  | xorl	%edx,%r15d | 
|  | addl	%r12d,%ebx | 
|  | rorl	$6,%r13d | 
|  | andl	%r15d,%edi | 
|  | xorl	%ecx,%r14d | 
|  | addl	%r13d,%ebx | 
|  | xorl	%edx,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%ebx,%r9d | 
|  | addl	%edi,%ebx | 
|  | movl	%r9d,%r13d | 
|  | addl	%ebx,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r14d,%ebx | 
|  | movl	%r10d,%r12d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r9d,%r13d | 
|  | xorl	%r11d,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%ebx,%r14d | 
|  | andl	%r9d,%r12d | 
|  | xorl	%r9d,%r13d | 
|  | addl	28(%rsp),%eax | 
|  | movl	%ebx,%edi | 
|  | xorl	%r11d,%r12d | 
|  | rorl	$11,%r14d | 
|  | xorl	%ecx,%edi | 
|  | addl	%r12d,%eax | 
|  | rorl	$6,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%ebx,%r14d | 
|  | addl	%r13d,%eax | 
|  | xorl	%ecx,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%eax,%r8d | 
|  | addl	%r15d,%eax | 
|  | movl	%r8d,%r13d | 
|  | addl	%eax,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r14d,%eax | 
|  | movl	%r9d,%r12d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r8d,%r13d | 
|  | xorl	%r10d,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%eax,%r14d | 
|  | andl	%r8d,%r12d | 
|  | xorl	%r8d,%r13d | 
|  | addl	32(%rsp),%r11d | 
|  | movl	%eax,%r15d | 
|  | xorl	%r10d,%r12d | 
|  | rorl	$11,%r14d | 
|  | xorl	%ebx,%r15d | 
|  | addl	%r12d,%r11d | 
|  | rorl	$6,%r13d | 
|  | andl	%r15d,%edi | 
|  | xorl	%eax,%r14d | 
|  | addl	%r13d,%r11d | 
|  | xorl	%ebx,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r11d,%edx | 
|  | addl	%edi,%r11d | 
|  | movl	%edx,%r13d | 
|  | addl	%r11d,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r14d,%r11d | 
|  | movl	%r8d,%r12d | 
|  | rorl	$9,%r14d | 
|  | xorl	%edx,%r13d | 
|  | xorl	%r9d,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%r11d,%r14d | 
|  | andl	%edx,%r12d | 
|  | xorl	%edx,%r13d | 
|  | addl	36(%rsp),%r10d | 
|  | movl	%r11d,%edi | 
|  | xorl	%r9d,%r12d | 
|  | rorl	$11,%r14d | 
|  | xorl	%eax,%edi | 
|  | addl	%r12d,%r10d | 
|  | rorl	$6,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%r11d,%r14d | 
|  | addl	%r13d,%r10d | 
|  | xorl	%eax,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%r10d,%ecx | 
|  | addl	%r15d,%r10d | 
|  | movl	%ecx,%r13d | 
|  | addl	%r10d,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r14d,%r10d | 
|  | movl	%edx,%r12d | 
|  | rorl	$9,%r14d | 
|  | xorl	%ecx,%r13d | 
|  | xorl	%r8d,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%r10d,%r14d | 
|  | andl	%ecx,%r12d | 
|  | xorl	%ecx,%r13d | 
|  | addl	40(%rsp),%r9d | 
|  | movl	%r10d,%r15d | 
|  | xorl	%r8d,%r12d | 
|  | rorl	$11,%r14d | 
|  | xorl	%r11d,%r15d | 
|  | addl	%r12d,%r9d | 
|  | rorl	$6,%r13d | 
|  | andl	%r15d,%edi | 
|  | xorl	%r10d,%r14d | 
|  | addl	%r13d,%r9d | 
|  | xorl	%r11d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%r9d,%ebx | 
|  | addl	%edi,%r9d | 
|  | movl	%ebx,%r13d | 
|  | addl	%r9d,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r14d,%r9d | 
|  | movl	%ecx,%r12d | 
|  | rorl	$9,%r14d | 
|  | xorl	%ebx,%r13d | 
|  | xorl	%edx,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%r9d,%r14d | 
|  | andl	%ebx,%r12d | 
|  | xorl	%ebx,%r13d | 
|  | addl	44(%rsp),%r8d | 
|  | movl	%r9d,%edi | 
|  | xorl	%edx,%r12d | 
|  | rorl	$11,%r14d | 
|  | xorl	%r10d,%edi | 
|  | addl	%r12d,%r8d | 
|  | rorl	$6,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%r9d,%r14d | 
|  | addl	%r13d,%r8d | 
|  | xorl	%r10d,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%r8d,%eax | 
|  | addl	%r15d,%r8d | 
|  | movl	%eax,%r13d | 
|  | addl	%r8d,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r14d,%r8d | 
|  | movl	%ebx,%r12d | 
|  | rorl	$9,%r14d | 
|  | xorl	%eax,%r13d | 
|  | xorl	%ecx,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%r8d,%r14d | 
|  | andl	%eax,%r12d | 
|  | xorl	%eax,%r13d | 
|  | addl	48(%rsp),%edx | 
|  | movl	%r8d,%r15d | 
|  | xorl	%ecx,%r12d | 
|  | rorl	$11,%r14d | 
|  | xorl	%r9d,%r15d | 
|  | addl	%r12d,%edx | 
|  | rorl	$6,%r13d | 
|  | andl	%r15d,%edi | 
|  | xorl	%r8d,%r14d | 
|  | addl	%r13d,%edx | 
|  | xorl	%r9d,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%edx,%r11d | 
|  | addl	%edi,%edx | 
|  | movl	%r11d,%r13d | 
|  | addl	%edx,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r14d,%edx | 
|  | movl	%eax,%r12d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r11d,%r13d | 
|  | xorl	%ebx,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%edx,%r14d | 
|  | andl	%r11d,%r12d | 
|  | xorl	%r11d,%r13d | 
|  | addl	52(%rsp),%ecx | 
|  | movl	%edx,%edi | 
|  | xorl	%ebx,%r12d | 
|  | rorl	$11,%r14d | 
|  | xorl	%r8d,%edi | 
|  | addl	%r12d,%ecx | 
|  | rorl	$6,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%edx,%r14d | 
|  | addl	%r13d,%ecx | 
|  | xorl	%r8d,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%ecx,%r10d | 
|  | addl	%r15d,%ecx | 
|  | movl	%r10d,%r13d | 
|  | addl	%ecx,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r14d,%ecx | 
|  | movl	%r11d,%r12d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r10d,%r13d | 
|  | xorl	%eax,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%ecx,%r14d | 
|  | andl	%r10d,%r12d | 
|  | xorl	%r10d,%r13d | 
|  | addl	56(%rsp),%ebx | 
|  | movl	%ecx,%r15d | 
|  | xorl	%eax,%r12d | 
|  | rorl	$11,%r14d | 
|  | xorl	%edx,%r15d | 
|  | addl	%r12d,%ebx | 
|  | rorl	$6,%r13d | 
|  | andl	%r15d,%edi | 
|  | xorl	%ecx,%r14d | 
|  | addl	%r13d,%ebx | 
|  | xorl	%edx,%edi | 
|  | rorl	$2,%r14d | 
|  | addl	%ebx,%r9d | 
|  | addl	%edi,%ebx | 
|  | movl	%r9d,%r13d | 
|  | addl	%ebx,%r14d | 
|  | rorl	$14,%r13d | 
|  | movl	%r14d,%ebx | 
|  | movl	%r10d,%r12d | 
|  | rorl	$9,%r14d | 
|  | xorl	%r9d,%r13d | 
|  | xorl	%r11d,%r12d | 
|  | rorl	$5,%r13d | 
|  | xorl	%ebx,%r14d | 
|  | andl	%r9d,%r12d | 
|  | xorl	%r9d,%r13d | 
|  | addl	60(%rsp),%eax | 
|  | movl	%ebx,%edi | 
|  | xorl	%r11d,%r12d | 
|  | rorl	$11,%r14d | 
|  | xorl	%ecx,%edi | 
|  | addl	%r12d,%eax | 
|  | rorl	$6,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%ebx,%r14d | 
|  | addl	%r13d,%eax | 
|  | xorl	%ecx,%r15d | 
|  | rorl	$2,%r14d | 
|  | addl	%eax,%r8d | 
|  | addl	%r15d,%eax | 
|  | movl	%r8d,%r13d | 
|  | addl	%eax,%r14d | 
|  | movq	64+0(%rsp),%rdi | 
|  | movl	%r14d,%eax | 
|  |  | 
|  | addl	0(%rdi),%eax | 
|  | leaq	64(%rsi),%rsi | 
|  | addl	4(%rdi),%ebx | 
|  | addl	8(%rdi),%ecx | 
|  | addl	12(%rdi),%edx | 
|  | addl	16(%rdi),%r8d | 
|  | addl	20(%rdi),%r9d | 
|  | addl	24(%rdi),%r10d | 
|  | addl	28(%rdi),%r11d | 
|  |  | 
|  | cmpq	64+16(%rsp),%rsi | 
|  |  | 
|  | movl	%eax,0(%rdi) | 
|  | movl	%ebx,4(%rdi) | 
|  | movl	%ecx,8(%rdi) | 
|  | movl	%edx,12(%rdi) | 
|  | movl	%r8d,16(%rdi) | 
|  | movl	%r9d,20(%rdi) | 
|  | movl	%r10d,24(%rdi) | 
|  | movl	%r11d,28(%rdi) | 
|  | jb	.Lloop_ssse3 | 
|  |  | 
|  | movq	88(%rsp),%rsi | 
|  | .cfi_def_cfa	%rsi,8 | 
|  | movq	-48(%rsi),%r15 | 
|  | .cfi_restore	%r15 | 
|  | movq	-40(%rsi),%r14 | 
|  | .cfi_restore	%r14 | 
|  | movq	-32(%rsi),%r13 | 
|  | .cfi_restore	%r13 | 
|  | movq	-24(%rsi),%r12 | 
|  | .cfi_restore	%r12 | 
|  | movq	-16(%rsi),%rbp | 
|  | .cfi_restore	%rbp | 
|  | movq	-8(%rsi),%rbx | 
|  | .cfi_restore	%rbx | 
|  | leaq	(%rsi),%rsp | 
|  | .cfi_def_cfa_register	%rsp | 
|  | .Lepilogue_ssse3: | 
|  | ret | 
|  | .cfi_endproc | 
|  | .size	sha256_block_data_order_ssse3,.-sha256_block_data_order_ssse3 | 
|  | .globl	sha256_block_data_order_avx | 
|  | .hidden sha256_block_data_order_avx | 
|  | .type	sha256_block_data_order_avx,@function | 
|  | .align	64 | 
|  | sha256_block_data_order_avx: | 
|  | .cfi_startproc | 
|  | _CET_ENDBR | 
|  | movq	%rsp,%rax | 
|  | .cfi_def_cfa_register	%rax | 
|  | pushq	%rbx | 
|  | .cfi_offset	%rbx,-16 | 
|  | pushq	%rbp | 
|  | .cfi_offset	%rbp,-24 | 
|  | pushq	%r12 | 
|  | .cfi_offset	%r12,-32 | 
|  | pushq	%r13 | 
|  | .cfi_offset	%r13,-40 | 
|  | pushq	%r14 | 
|  | .cfi_offset	%r14,-48 | 
|  | pushq	%r15 | 
|  | .cfi_offset	%r15,-56 | 
|  | shlq	$4,%rdx | 
|  | subq	$96,%rsp | 
|  | leaq	(%rsi,%rdx,4),%rdx | 
|  | andq	$-64,%rsp | 
|  | movq	%rdi,64+0(%rsp) | 
|  | movq	%rsi,64+8(%rsp) | 
|  | movq	%rdx,64+16(%rsp) | 
|  | movq	%rax,88(%rsp) | 
|  | .cfi_escape	0x0f,0x06,0x77,0xd8,0x00,0x06,0x23,0x08 | 
|  | .Lprologue_avx: | 
|  |  | 
|  | vzeroupper | 
|  | movl	0(%rdi),%eax | 
|  | movl	4(%rdi),%ebx | 
|  | movl	8(%rdi),%ecx | 
|  | movl	12(%rdi),%edx | 
|  | movl	16(%rdi),%r8d | 
|  | movl	20(%rdi),%r9d | 
|  | movl	24(%rdi),%r10d | 
|  | movl	28(%rdi),%r11d | 
|  | vmovdqa	K256+512+32(%rip),%xmm8 | 
|  | vmovdqa	K256+512+64(%rip),%xmm9 | 
|  | jmp	.Lloop_avx | 
|  | .align	16 | 
|  | .Lloop_avx: | 
|  | vmovdqa	K256+512(%rip),%xmm7 | 
|  | vmovdqu	0(%rsi),%xmm0 | 
|  | vmovdqu	16(%rsi),%xmm1 | 
|  | vmovdqu	32(%rsi),%xmm2 | 
|  | vmovdqu	48(%rsi),%xmm3 | 
|  | vpshufb	%xmm7,%xmm0,%xmm0 | 
|  | leaq	K256(%rip),%rbp | 
|  | vpshufb	%xmm7,%xmm1,%xmm1 | 
|  | vpshufb	%xmm7,%xmm2,%xmm2 | 
|  | vpaddd	0(%rbp),%xmm0,%xmm4 | 
|  | vpshufb	%xmm7,%xmm3,%xmm3 | 
|  | vpaddd	32(%rbp),%xmm1,%xmm5 | 
|  | vpaddd	64(%rbp),%xmm2,%xmm6 | 
|  | vpaddd	96(%rbp),%xmm3,%xmm7 | 
|  | vmovdqa	%xmm4,0(%rsp) | 
|  | movl	%eax,%r14d | 
|  | vmovdqa	%xmm5,16(%rsp) | 
|  | movl	%ebx,%edi | 
|  | vmovdqa	%xmm6,32(%rsp) | 
|  | xorl	%ecx,%edi | 
|  | vmovdqa	%xmm7,48(%rsp) | 
|  | movl	%r8d,%r13d | 
|  | jmp	.Lavx_00_47 | 
|  |  | 
|  | .align	16 | 
|  | .Lavx_00_47: | 
|  | subq	$-128,%rbp | 
|  | vpalignr	$4,%xmm0,%xmm1,%xmm4 | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%eax | 
|  | movl	%r9d,%r12d | 
|  | vpalignr	$4,%xmm2,%xmm3,%xmm7 | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%r8d,%r13d | 
|  | xorl	%r10d,%r12d | 
|  | vpsrld	$7,%xmm4,%xmm6 | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%eax,%r14d | 
|  | andl	%r8d,%r12d | 
|  | vpaddd	%xmm7,%xmm0,%xmm0 | 
|  | xorl	%r8d,%r13d | 
|  | addl	0(%rsp),%r11d | 
|  | movl	%eax,%r15d | 
|  | vpsrld	$3,%xmm4,%xmm7 | 
|  | xorl	%r10d,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%ebx,%r15d | 
|  | vpslld	$14,%xmm4,%xmm5 | 
|  | addl	%r12d,%r11d | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%r15d,%edi | 
|  | vpxor	%xmm6,%xmm7,%xmm4 | 
|  | xorl	%eax,%r14d | 
|  | addl	%r13d,%r11d | 
|  | xorl	%ebx,%edi | 
|  | vpshufd	$250,%xmm3,%xmm7 | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%r11d,%edx | 
|  | addl	%edi,%r11d | 
|  | vpsrld	$11,%xmm6,%xmm6 | 
|  | movl	%edx,%r13d | 
|  | addl	%r11d,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | vpxor	%xmm5,%xmm4,%xmm4 | 
|  | movl	%r14d,%r11d | 
|  | movl	%r8d,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | vpslld	$11,%xmm5,%xmm5 | 
|  | xorl	%edx,%r13d | 
|  | xorl	%r9d,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | vpxor	%xmm6,%xmm4,%xmm4 | 
|  | xorl	%r11d,%r14d | 
|  | andl	%edx,%r12d | 
|  | xorl	%edx,%r13d | 
|  | vpsrld	$10,%xmm7,%xmm6 | 
|  | addl	4(%rsp),%r10d | 
|  | movl	%r11d,%edi | 
|  | xorl	%r9d,%r12d | 
|  | vpxor	%xmm5,%xmm4,%xmm4 | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%eax,%edi | 
|  | addl	%r12d,%r10d | 
|  | vpsrlq	$17,%xmm7,%xmm7 | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%r11d,%r14d | 
|  | vpaddd	%xmm4,%xmm0,%xmm0 | 
|  | addl	%r13d,%r10d | 
|  | xorl	%eax,%r15d | 
|  | shrdl	$2,%r14d,%r14d | 
|  | vpxor	%xmm7,%xmm6,%xmm6 | 
|  | addl	%r10d,%ecx | 
|  | addl	%r15d,%r10d | 
|  | movl	%ecx,%r13d | 
|  | vpsrlq	$2,%xmm7,%xmm7 | 
|  | addl	%r10d,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%r10d | 
|  | vpxor	%xmm7,%xmm6,%xmm6 | 
|  | movl	%edx,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%ecx,%r13d | 
|  | vpshufb	%xmm8,%xmm6,%xmm6 | 
|  | xorl	%r8d,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%r10d,%r14d | 
|  | vpaddd	%xmm6,%xmm0,%xmm0 | 
|  | andl	%ecx,%r12d | 
|  | xorl	%ecx,%r13d | 
|  | addl	8(%rsp),%r9d | 
|  | vpshufd	$80,%xmm0,%xmm7 | 
|  | movl	%r10d,%r15d | 
|  | xorl	%r8d,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | vpsrld	$10,%xmm7,%xmm6 | 
|  | xorl	%r11d,%r15d | 
|  | addl	%r12d,%r9d | 
|  | shrdl	$6,%r13d,%r13d | 
|  | vpsrlq	$17,%xmm7,%xmm7 | 
|  | andl	%r15d,%edi | 
|  | xorl	%r10d,%r14d | 
|  | addl	%r13d,%r9d | 
|  | vpxor	%xmm7,%xmm6,%xmm6 | 
|  | xorl	%r11d,%edi | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%r9d,%ebx | 
|  | vpsrlq	$2,%xmm7,%xmm7 | 
|  | addl	%edi,%r9d | 
|  | movl	%ebx,%r13d | 
|  | addl	%r9d,%r14d | 
|  | vpxor	%xmm7,%xmm6,%xmm6 | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%r9d | 
|  | movl	%ecx,%r12d | 
|  | vpshufb	%xmm9,%xmm6,%xmm6 | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%ebx,%r13d | 
|  | xorl	%edx,%r12d | 
|  | vpaddd	%xmm6,%xmm0,%xmm0 | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%r9d,%r14d | 
|  | andl	%ebx,%r12d | 
|  | vpaddd	0(%rbp),%xmm0,%xmm6 | 
|  | xorl	%ebx,%r13d | 
|  | addl	12(%rsp),%r8d | 
|  | movl	%r9d,%edi | 
|  | xorl	%edx,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%r10d,%edi | 
|  | addl	%r12d,%r8d | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%r9d,%r14d | 
|  | addl	%r13d,%r8d | 
|  | xorl	%r10d,%r15d | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%r8d,%eax | 
|  | addl	%r15d,%r8d | 
|  | movl	%eax,%r13d | 
|  | addl	%r8d,%r14d | 
|  | vmovdqa	%xmm6,0(%rsp) | 
|  | vpalignr	$4,%xmm1,%xmm2,%xmm4 | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%r8d | 
|  | movl	%ebx,%r12d | 
|  | vpalignr	$4,%xmm3,%xmm0,%xmm7 | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%eax,%r13d | 
|  | xorl	%ecx,%r12d | 
|  | vpsrld	$7,%xmm4,%xmm6 | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%r8d,%r14d | 
|  | andl	%eax,%r12d | 
|  | vpaddd	%xmm7,%xmm1,%xmm1 | 
|  | xorl	%eax,%r13d | 
|  | addl	16(%rsp),%edx | 
|  | movl	%r8d,%r15d | 
|  | vpsrld	$3,%xmm4,%xmm7 | 
|  | xorl	%ecx,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%r9d,%r15d | 
|  | vpslld	$14,%xmm4,%xmm5 | 
|  | addl	%r12d,%edx | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%r15d,%edi | 
|  | vpxor	%xmm6,%xmm7,%xmm4 | 
|  | xorl	%r8d,%r14d | 
|  | addl	%r13d,%edx | 
|  | xorl	%r9d,%edi | 
|  | vpshufd	$250,%xmm0,%xmm7 | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%edx,%r11d | 
|  | addl	%edi,%edx | 
|  | vpsrld	$11,%xmm6,%xmm6 | 
|  | movl	%r11d,%r13d | 
|  | addl	%edx,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | vpxor	%xmm5,%xmm4,%xmm4 | 
|  | movl	%r14d,%edx | 
|  | movl	%eax,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | vpslld	$11,%xmm5,%xmm5 | 
|  | xorl	%r11d,%r13d | 
|  | xorl	%ebx,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | vpxor	%xmm6,%xmm4,%xmm4 | 
|  | xorl	%edx,%r14d | 
|  | andl	%r11d,%r12d | 
|  | xorl	%r11d,%r13d | 
|  | vpsrld	$10,%xmm7,%xmm6 | 
|  | addl	20(%rsp),%ecx | 
|  | movl	%edx,%edi | 
|  | xorl	%ebx,%r12d | 
|  | vpxor	%xmm5,%xmm4,%xmm4 | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%r8d,%edi | 
|  | addl	%r12d,%ecx | 
|  | vpsrlq	$17,%xmm7,%xmm7 | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%edx,%r14d | 
|  | vpaddd	%xmm4,%xmm1,%xmm1 | 
|  | addl	%r13d,%ecx | 
|  | xorl	%r8d,%r15d | 
|  | shrdl	$2,%r14d,%r14d | 
|  | vpxor	%xmm7,%xmm6,%xmm6 | 
|  | addl	%ecx,%r10d | 
|  | addl	%r15d,%ecx | 
|  | movl	%r10d,%r13d | 
|  | vpsrlq	$2,%xmm7,%xmm7 | 
|  | addl	%ecx,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%ecx | 
|  | vpxor	%xmm7,%xmm6,%xmm6 | 
|  | movl	%r11d,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%r10d,%r13d | 
|  | vpshufb	%xmm8,%xmm6,%xmm6 | 
|  | xorl	%eax,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%ecx,%r14d | 
|  | vpaddd	%xmm6,%xmm1,%xmm1 | 
|  | andl	%r10d,%r12d | 
|  | xorl	%r10d,%r13d | 
|  | addl	24(%rsp),%ebx | 
|  | vpshufd	$80,%xmm1,%xmm7 | 
|  | movl	%ecx,%r15d | 
|  | xorl	%eax,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | vpsrld	$10,%xmm7,%xmm6 | 
|  | xorl	%edx,%r15d | 
|  | addl	%r12d,%ebx | 
|  | shrdl	$6,%r13d,%r13d | 
|  | vpsrlq	$17,%xmm7,%xmm7 | 
|  | andl	%r15d,%edi | 
|  | xorl	%ecx,%r14d | 
|  | addl	%r13d,%ebx | 
|  | vpxor	%xmm7,%xmm6,%xmm6 | 
|  | xorl	%edx,%edi | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%ebx,%r9d | 
|  | vpsrlq	$2,%xmm7,%xmm7 | 
|  | addl	%edi,%ebx | 
|  | movl	%r9d,%r13d | 
|  | addl	%ebx,%r14d | 
|  | vpxor	%xmm7,%xmm6,%xmm6 | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%ebx | 
|  | movl	%r10d,%r12d | 
|  | vpshufb	%xmm9,%xmm6,%xmm6 | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%r9d,%r13d | 
|  | xorl	%r11d,%r12d | 
|  | vpaddd	%xmm6,%xmm1,%xmm1 | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%ebx,%r14d | 
|  | andl	%r9d,%r12d | 
|  | vpaddd	32(%rbp),%xmm1,%xmm6 | 
|  | xorl	%r9d,%r13d | 
|  | addl	28(%rsp),%eax | 
|  | movl	%ebx,%edi | 
|  | xorl	%r11d,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%ecx,%edi | 
|  | addl	%r12d,%eax | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%ebx,%r14d | 
|  | addl	%r13d,%eax | 
|  | xorl	%ecx,%r15d | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%eax,%r8d | 
|  | addl	%r15d,%eax | 
|  | movl	%r8d,%r13d | 
|  | addl	%eax,%r14d | 
|  | vmovdqa	%xmm6,16(%rsp) | 
|  | vpalignr	$4,%xmm2,%xmm3,%xmm4 | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%eax | 
|  | movl	%r9d,%r12d | 
|  | vpalignr	$4,%xmm0,%xmm1,%xmm7 | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%r8d,%r13d | 
|  | xorl	%r10d,%r12d | 
|  | vpsrld	$7,%xmm4,%xmm6 | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%eax,%r14d | 
|  | andl	%r8d,%r12d | 
|  | vpaddd	%xmm7,%xmm2,%xmm2 | 
|  | xorl	%r8d,%r13d | 
|  | addl	32(%rsp),%r11d | 
|  | movl	%eax,%r15d | 
|  | vpsrld	$3,%xmm4,%xmm7 | 
|  | xorl	%r10d,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%ebx,%r15d | 
|  | vpslld	$14,%xmm4,%xmm5 | 
|  | addl	%r12d,%r11d | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%r15d,%edi | 
|  | vpxor	%xmm6,%xmm7,%xmm4 | 
|  | xorl	%eax,%r14d | 
|  | addl	%r13d,%r11d | 
|  | xorl	%ebx,%edi | 
|  | vpshufd	$250,%xmm1,%xmm7 | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%r11d,%edx | 
|  | addl	%edi,%r11d | 
|  | vpsrld	$11,%xmm6,%xmm6 | 
|  | movl	%edx,%r13d | 
|  | addl	%r11d,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | vpxor	%xmm5,%xmm4,%xmm4 | 
|  | movl	%r14d,%r11d | 
|  | movl	%r8d,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | vpslld	$11,%xmm5,%xmm5 | 
|  | xorl	%edx,%r13d | 
|  | xorl	%r9d,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | vpxor	%xmm6,%xmm4,%xmm4 | 
|  | xorl	%r11d,%r14d | 
|  | andl	%edx,%r12d | 
|  | xorl	%edx,%r13d | 
|  | vpsrld	$10,%xmm7,%xmm6 | 
|  | addl	36(%rsp),%r10d | 
|  | movl	%r11d,%edi | 
|  | xorl	%r9d,%r12d | 
|  | vpxor	%xmm5,%xmm4,%xmm4 | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%eax,%edi | 
|  | addl	%r12d,%r10d | 
|  | vpsrlq	$17,%xmm7,%xmm7 | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%r11d,%r14d | 
|  | vpaddd	%xmm4,%xmm2,%xmm2 | 
|  | addl	%r13d,%r10d | 
|  | xorl	%eax,%r15d | 
|  | shrdl	$2,%r14d,%r14d | 
|  | vpxor	%xmm7,%xmm6,%xmm6 | 
|  | addl	%r10d,%ecx | 
|  | addl	%r15d,%r10d | 
|  | movl	%ecx,%r13d | 
|  | vpsrlq	$2,%xmm7,%xmm7 | 
|  | addl	%r10d,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%r10d | 
|  | vpxor	%xmm7,%xmm6,%xmm6 | 
|  | movl	%edx,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%ecx,%r13d | 
|  | vpshufb	%xmm8,%xmm6,%xmm6 | 
|  | xorl	%r8d,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%r10d,%r14d | 
|  | vpaddd	%xmm6,%xmm2,%xmm2 | 
|  | andl	%ecx,%r12d | 
|  | xorl	%ecx,%r13d | 
|  | addl	40(%rsp),%r9d | 
|  | vpshufd	$80,%xmm2,%xmm7 | 
|  | movl	%r10d,%r15d | 
|  | xorl	%r8d,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | vpsrld	$10,%xmm7,%xmm6 | 
|  | xorl	%r11d,%r15d | 
|  | addl	%r12d,%r9d | 
|  | shrdl	$6,%r13d,%r13d | 
|  | vpsrlq	$17,%xmm7,%xmm7 | 
|  | andl	%r15d,%edi | 
|  | xorl	%r10d,%r14d | 
|  | addl	%r13d,%r9d | 
|  | vpxor	%xmm7,%xmm6,%xmm6 | 
|  | xorl	%r11d,%edi | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%r9d,%ebx | 
|  | vpsrlq	$2,%xmm7,%xmm7 | 
|  | addl	%edi,%r9d | 
|  | movl	%ebx,%r13d | 
|  | addl	%r9d,%r14d | 
|  | vpxor	%xmm7,%xmm6,%xmm6 | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%r9d | 
|  | movl	%ecx,%r12d | 
|  | vpshufb	%xmm9,%xmm6,%xmm6 | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%ebx,%r13d | 
|  | xorl	%edx,%r12d | 
|  | vpaddd	%xmm6,%xmm2,%xmm2 | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%r9d,%r14d | 
|  | andl	%ebx,%r12d | 
|  | vpaddd	64(%rbp),%xmm2,%xmm6 | 
|  | xorl	%ebx,%r13d | 
|  | addl	44(%rsp),%r8d | 
|  | movl	%r9d,%edi | 
|  | xorl	%edx,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%r10d,%edi | 
|  | addl	%r12d,%r8d | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%r9d,%r14d | 
|  | addl	%r13d,%r8d | 
|  | xorl	%r10d,%r15d | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%r8d,%eax | 
|  | addl	%r15d,%r8d | 
|  | movl	%eax,%r13d | 
|  | addl	%r8d,%r14d | 
|  | vmovdqa	%xmm6,32(%rsp) | 
|  | vpalignr	$4,%xmm3,%xmm0,%xmm4 | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%r8d | 
|  | movl	%ebx,%r12d | 
|  | vpalignr	$4,%xmm1,%xmm2,%xmm7 | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%eax,%r13d | 
|  | xorl	%ecx,%r12d | 
|  | vpsrld	$7,%xmm4,%xmm6 | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%r8d,%r14d | 
|  | andl	%eax,%r12d | 
|  | vpaddd	%xmm7,%xmm3,%xmm3 | 
|  | xorl	%eax,%r13d | 
|  | addl	48(%rsp),%edx | 
|  | movl	%r8d,%r15d | 
|  | vpsrld	$3,%xmm4,%xmm7 | 
|  | xorl	%ecx,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%r9d,%r15d | 
|  | vpslld	$14,%xmm4,%xmm5 | 
|  | addl	%r12d,%edx | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%r15d,%edi | 
|  | vpxor	%xmm6,%xmm7,%xmm4 | 
|  | xorl	%r8d,%r14d | 
|  | addl	%r13d,%edx | 
|  | xorl	%r9d,%edi | 
|  | vpshufd	$250,%xmm2,%xmm7 | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%edx,%r11d | 
|  | addl	%edi,%edx | 
|  | vpsrld	$11,%xmm6,%xmm6 | 
|  | movl	%r11d,%r13d | 
|  | addl	%edx,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | vpxor	%xmm5,%xmm4,%xmm4 | 
|  | movl	%r14d,%edx | 
|  | movl	%eax,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | vpslld	$11,%xmm5,%xmm5 | 
|  | xorl	%r11d,%r13d | 
|  | xorl	%ebx,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | vpxor	%xmm6,%xmm4,%xmm4 | 
|  | xorl	%edx,%r14d | 
|  | andl	%r11d,%r12d | 
|  | xorl	%r11d,%r13d | 
|  | vpsrld	$10,%xmm7,%xmm6 | 
|  | addl	52(%rsp),%ecx | 
|  | movl	%edx,%edi | 
|  | xorl	%ebx,%r12d | 
|  | vpxor	%xmm5,%xmm4,%xmm4 | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%r8d,%edi | 
|  | addl	%r12d,%ecx | 
|  | vpsrlq	$17,%xmm7,%xmm7 | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%edx,%r14d | 
|  | vpaddd	%xmm4,%xmm3,%xmm3 | 
|  | addl	%r13d,%ecx | 
|  | xorl	%r8d,%r15d | 
|  | shrdl	$2,%r14d,%r14d | 
|  | vpxor	%xmm7,%xmm6,%xmm6 | 
|  | addl	%ecx,%r10d | 
|  | addl	%r15d,%ecx | 
|  | movl	%r10d,%r13d | 
|  | vpsrlq	$2,%xmm7,%xmm7 | 
|  | addl	%ecx,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%ecx | 
|  | vpxor	%xmm7,%xmm6,%xmm6 | 
|  | movl	%r11d,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%r10d,%r13d | 
|  | vpshufb	%xmm8,%xmm6,%xmm6 | 
|  | xorl	%eax,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%ecx,%r14d | 
|  | vpaddd	%xmm6,%xmm3,%xmm3 | 
|  | andl	%r10d,%r12d | 
|  | xorl	%r10d,%r13d | 
|  | addl	56(%rsp),%ebx | 
|  | vpshufd	$80,%xmm3,%xmm7 | 
|  | movl	%ecx,%r15d | 
|  | xorl	%eax,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | vpsrld	$10,%xmm7,%xmm6 | 
|  | xorl	%edx,%r15d | 
|  | addl	%r12d,%ebx | 
|  | shrdl	$6,%r13d,%r13d | 
|  | vpsrlq	$17,%xmm7,%xmm7 | 
|  | andl	%r15d,%edi | 
|  | xorl	%ecx,%r14d | 
|  | addl	%r13d,%ebx | 
|  | vpxor	%xmm7,%xmm6,%xmm6 | 
|  | xorl	%edx,%edi | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%ebx,%r9d | 
|  | vpsrlq	$2,%xmm7,%xmm7 | 
|  | addl	%edi,%ebx | 
|  | movl	%r9d,%r13d | 
|  | addl	%ebx,%r14d | 
|  | vpxor	%xmm7,%xmm6,%xmm6 | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%ebx | 
|  | movl	%r10d,%r12d | 
|  | vpshufb	%xmm9,%xmm6,%xmm6 | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%r9d,%r13d | 
|  | xorl	%r11d,%r12d | 
|  | vpaddd	%xmm6,%xmm3,%xmm3 | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%ebx,%r14d | 
|  | andl	%r9d,%r12d | 
|  | vpaddd	96(%rbp),%xmm3,%xmm6 | 
|  | xorl	%r9d,%r13d | 
|  | addl	60(%rsp),%eax | 
|  | movl	%ebx,%edi | 
|  | xorl	%r11d,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%ecx,%edi | 
|  | addl	%r12d,%eax | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%ebx,%r14d | 
|  | addl	%r13d,%eax | 
|  | xorl	%ecx,%r15d | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%eax,%r8d | 
|  | addl	%r15d,%eax | 
|  | movl	%r8d,%r13d | 
|  | addl	%eax,%r14d | 
|  | vmovdqa	%xmm6,48(%rsp) | 
|  | cmpb	$0,131(%rbp) | 
|  | jne	.Lavx_00_47 | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%eax | 
|  | movl	%r9d,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%r8d,%r13d | 
|  | xorl	%r10d,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%eax,%r14d | 
|  | andl	%r8d,%r12d | 
|  | xorl	%r8d,%r13d | 
|  | addl	0(%rsp),%r11d | 
|  | movl	%eax,%r15d | 
|  | xorl	%r10d,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%ebx,%r15d | 
|  | addl	%r12d,%r11d | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%r15d,%edi | 
|  | xorl	%eax,%r14d | 
|  | addl	%r13d,%r11d | 
|  | xorl	%ebx,%edi | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%r11d,%edx | 
|  | addl	%edi,%r11d | 
|  | movl	%edx,%r13d | 
|  | addl	%r11d,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%r11d | 
|  | movl	%r8d,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%edx,%r13d | 
|  | xorl	%r9d,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%r11d,%r14d | 
|  | andl	%edx,%r12d | 
|  | xorl	%edx,%r13d | 
|  | addl	4(%rsp),%r10d | 
|  | movl	%r11d,%edi | 
|  | xorl	%r9d,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%eax,%edi | 
|  | addl	%r12d,%r10d | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%r11d,%r14d | 
|  | addl	%r13d,%r10d | 
|  | xorl	%eax,%r15d | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%r10d,%ecx | 
|  | addl	%r15d,%r10d | 
|  | movl	%ecx,%r13d | 
|  | addl	%r10d,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%r10d | 
|  | movl	%edx,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%ecx,%r13d | 
|  | xorl	%r8d,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%r10d,%r14d | 
|  | andl	%ecx,%r12d | 
|  | xorl	%ecx,%r13d | 
|  | addl	8(%rsp),%r9d | 
|  | movl	%r10d,%r15d | 
|  | xorl	%r8d,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%r11d,%r15d | 
|  | addl	%r12d,%r9d | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%r15d,%edi | 
|  | xorl	%r10d,%r14d | 
|  | addl	%r13d,%r9d | 
|  | xorl	%r11d,%edi | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%r9d,%ebx | 
|  | addl	%edi,%r9d | 
|  | movl	%ebx,%r13d | 
|  | addl	%r9d,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%r9d | 
|  | movl	%ecx,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%ebx,%r13d | 
|  | xorl	%edx,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%r9d,%r14d | 
|  | andl	%ebx,%r12d | 
|  | xorl	%ebx,%r13d | 
|  | addl	12(%rsp),%r8d | 
|  | movl	%r9d,%edi | 
|  | xorl	%edx,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%r10d,%edi | 
|  | addl	%r12d,%r8d | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%r9d,%r14d | 
|  | addl	%r13d,%r8d | 
|  | xorl	%r10d,%r15d | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%r8d,%eax | 
|  | addl	%r15d,%r8d | 
|  | movl	%eax,%r13d | 
|  | addl	%r8d,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%r8d | 
|  | movl	%ebx,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%eax,%r13d | 
|  | xorl	%ecx,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%r8d,%r14d | 
|  | andl	%eax,%r12d | 
|  | xorl	%eax,%r13d | 
|  | addl	16(%rsp),%edx | 
|  | movl	%r8d,%r15d | 
|  | xorl	%ecx,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%r9d,%r15d | 
|  | addl	%r12d,%edx | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%r15d,%edi | 
|  | xorl	%r8d,%r14d | 
|  | addl	%r13d,%edx | 
|  | xorl	%r9d,%edi | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%edx,%r11d | 
|  | addl	%edi,%edx | 
|  | movl	%r11d,%r13d | 
|  | addl	%edx,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%edx | 
|  | movl	%eax,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%r11d,%r13d | 
|  | xorl	%ebx,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%edx,%r14d | 
|  | andl	%r11d,%r12d | 
|  | xorl	%r11d,%r13d | 
|  | addl	20(%rsp),%ecx | 
|  | movl	%edx,%edi | 
|  | xorl	%ebx,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%r8d,%edi | 
|  | addl	%r12d,%ecx | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%edx,%r14d | 
|  | addl	%r13d,%ecx | 
|  | xorl	%r8d,%r15d | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%ecx,%r10d | 
|  | addl	%r15d,%ecx | 
|  | movl	%r10d,%r13d | 
|  | addl	%ecx,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%ecx | 
|  | movl	%r11d,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%r10d,%r13d | 
|  | xorl	%eax,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%ecx,%r14d | 
|  | andl	%r10d,%r12d | 
|  | xorl	%r10d,%r13d | 
|  | addl	24(%rsp),%ebx | 
|  | movl	%ecx,%r15d | 
|  | xorl	%eax,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%edx,%r15d | 
|  | addl	%r12d,%ebx | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%r15d,%edi | 
|  | xorl	%ecx,%r14d | 
|  | addl	%r13d,%ebx | 
|  | xorl	%edx,%edi | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%ebx,%r9d | 
|  | addl	%edi,%ebx | 
|  | movl	%r9d,%r13d | 
|  | addl	%ebx,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%ebx | 
|  | movl	%r10d,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%r9d,%r13d | 
|  | xorl	%r11d,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%ebx,%r14d | 
|  | andl	%r9d,%r12d | 
|  | xorl	%r9d,%r13d | 
|  | addl	28(%rsp),%eax | 
|  | movl	%ebx,%edi | 
|  | xorl	%r11d,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%ecx,%edi | 
|  | addl	%r12d,%eax | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%ebx,%r14d | 
|  | addl	%r13d,%eax | 
|  | xorl	%ecx,%r15d | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%eax,%r8d | 
|  | addl	%r15d,%eax | 
|  | movl	%r8d,%r13d | 
|  | addl	%eax,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%eax | 
|  | movl	%r9d,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%r8d,%r13d | 
|  | xorl	%r10d,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%eax,%r14d | 
|  | andl	%r8d,%r12d | 
|  | xorl	%r8d,%r13d | 
|  | addl	32(%rsp),%r11d | 
|  | movl	%eax,%r15d | 
|  | xorl	%r10d,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%ebx,%r15d | 
|  | addl	%r12d,%r11d | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%r15d,%edi | 
|  | xorl	%eax,%r14d | 
|  | addl	%r13d,%r11d | 
|  | xorl	%ebx,%edi | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%r11d,%edx | 
|  | addl	%edi,%r11d | 
|  | movl	%edx,%r13d | 
|  | addl	%r11d,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%r11d | 
|  | movl	%r8d,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%edx,%r13d | 
|  | xorl	%r9d,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%r11d,%r14d | 
|  | andl	%edx,%r12d | 
|  | xorl	%edx,%r13d | 
|  | addl	36(%rsp),%r10d | 
|  | movl	%r11d,%edi | 
|  | xorl	%r9d,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%eax,%edi | 
|  | addl	%r12d,%r10d | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%r11d,%r14d | 
|  | addl	%r13d,%r10d | 
|  | xorl	%eax,%r15d | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%r10d,%ecx | 
|  | addl	%r15d,%r10d | 
|  | movl	%ecx,%r13d | 
|  | addl	%r10d,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%r10d | 
|  | movl	%edx,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%ecx,%r13d | 
|  | xorl	%r8d,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%r10d,%r14d | 
|  | andl	%ecx,%r12d | 
|  | xorl	%ecx,%r13d | 
|  | addl	40(%rsp),%r9d | 
|  | movl	%r10d,%r15d | 
|  | xorl	%r8d,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%r11d,%r15d | 
|  | addl	%r12d,%r9d | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%r15d,%edi | 
|  | xorl	%r10d,%r14d | 
|  | addl	%r13d,%r9d | 
|  | xorl	%r11d,%edi | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%r9d,%ebx | 
|  | addl	%edi,%r9d | 
|  | movl	%ebx,%r13d | 
|  | addl	%r9d,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%r9d | 
|  | movl	%ecx,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%ebx,%r13d | 
|  | xorl	%edx,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%r9d,%r14d | 
|  | andl	%ebx,%r12d | 
|  | xorl	%ebx,%r13d | 
|  | addl	44(%rsp),%r8d | 
|  | movl	%r9d,%edi | 
|  | xorl	%edx,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%r10d,%edi | 
|  | addl	%r12d,%r8d | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%r9d,%r14d | 
|  | addl	%r13d,%r8d | 
|  | xorl	%r10d,%r15d | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%r8d,%eax | 
|  | addl	%r15d,%r8d | 
|  | movl	%eax,%r13d | 
|  | addl	%r8d,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%r8d | 
|  | movl	%ebx,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%eax,%r13d | 
|  | xorl	%ecx,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%r8d,%r14d | 
|  | andl	%eax,%r12d | 
|  | xorl	%eax,%r13d | 
|  | addl	48(%rsp),%edx | 
|  | movl	%r8d,%r15d | 
|  | xorl	%ecx,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%r9d,%r15d | 
|  | addl	%r12d,%edx | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%r15d,%edi | 
|  | xorl	%r8d,%r14d | 
|  | addl	%r13d,%edx | 
|  | xorl	%r9d,%edi | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%edx,%r11d | 
|  | addl	%edi,%edx | 
|  | movl	%r11d,%r13d | 
|  | addl	%edx,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%edx | 
|  | movl	%eax,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%r11d,%r13d | 
|  | xorl	%ebx,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%edx,%r14d | 
|  | andl	%r11d,%r12d | 
|  | xorl	%r11d,%r13d | 
|  | addl	52(%rsp),%ecx | 
|  | movl	%edx,%edi | 
|  | xorl	%ebx,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%r8d,%edi | 
|  | addl	%r12d,%ecx | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%edx,%r14d | 
|  | addl	%r13d,%ecx | 
|  | xorl	%r8d,%r15d | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%ecx,%r10d | 
|  | addl	%r15d,%ecx | 
|  | movl	%r10d,%r13d | 
|  | addl	%ecx,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%ecx | 
|  | movl	%r11d,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%r10d,%r13d | 
|  | xorl	%eax,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%ecx,%r14d | 
|  | andl	%r10d,%r12d | 
|  | xorl	%r10d,%r13d | 
|  | addl	56(%rsp),%ebx | 
|  | movl	%ecx,%r15d | 
|  | xorl	%eax,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%edx,%r15d | 
|  | addl	%r12d,%ebx | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%r15d,%edi | 
|  | xorl	%ecx,%r14d | 
|  | addl	%r13d,%ebx | 
|  | xorl	%edx,%edi | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%ebx,%r9d | 
|  | addl	%edi,%ebx | 
|  | movl	%r9d,%r13d | 
|  | addl	%ebx,%r14d | 
|  | shrdl	$14,%r13d,%r13d | 
|  | movl	%r14d,%ebx | 
|  | movl	%r10d,%r12d | 
|  | shrdl	$9,%r14d,%r14d | 
|  | xorl	%r9d,%r13d | 
|  | xorl	%r11d,%r12d | 
|  | shrdl	$5,%r13d,%r13d | 
|  | xorl	%ebx,%r14d | 
|  | andl	%r9d,%r12d | 
|  | xorl	%r9d,%r13d | 
|  | addl	60(%rsp),%eax | 
|  | movl	%ebx,%edi | 
|  | xorl	%r11d,%r12d | 
|  | shrdl	$11,%r14d,%r14d | 
|  | xorl	%ecx,%edi | 
|  | addl	%r12d,%eax | 
|  | shrdl	$6,%r13d,%r13d | 
|  | andl	%edi,%r15d | 
|  | xorl	%ebx,%r14d | 
|  | addl	%r13d,%eax | 
|  | xorl	%ecx,%r15d | 
|  | shrdl	$2,%r14d,%r14d | 
|  | addl	%eax,%r8d | 
|  | addl	%r15d,%eax | 
|  | movl	%r8d,%r13d | 
|  | addl	%eax,%r14d | 
|  | movq	64+0(%rsp),%rdi | 
|  | movl	%r14d,%eax | 
|  |  | 
|  | addl	0(%rdi),%eax | 
|  | leaq	64(%rsi),%rsi | 
|  | addl	4(%rdi),%ebx | 
|  | addl	8(%rdi),%ecx | 
|  | addl	12(%rdi),%edx | 
|  | addl	16(%rdi),%r8d | 
|  | addl	20(%rdi),%r9d | 
|  | addl	24(%rdi),%r10d | 
|  | addl	28(%rdi),%r11d | 
|  |  | 
|  | cmpq	64+16(%rsp),%rsi | 
|  |  | 
|  | movl	%eax,0(%rdi) | 
|  | movl	%ebx,4(%rdi) | 
|  | movl	%ecx,8(%rdi) | 
|  | movl	%edx,12(%rdi) | 
|  | movl	%r8d,16(%rdi) | 
|  | movl	%r9d,20(%rdi) | 
|  | movl	%r10d,24(%rdi) | 
|  | movl	%r11d,28(%rdi) | 
|  | jb	.Lloop_avx | 
|  |  | 
|  | movq	88(%rsp),%rsi | 
|  | .cfi_def_cfa	%rsi,8 | 
|  | vzeroupper | 
|  | movq	-48(%rsi),%r15 | 
|  | .cfi_restore	%r15 | 
|  | movq	-40(%rsi),%r14 | 
|  | .cfi_restore	%r14 | 
|  | movq	-32(%rsi),%r13 | 
|  | .cfi_restore	%r13 | 
|  | movq	-24(%rsi),%r12 | 
|  | .cfi_restore	%r12 | 
|  | movq	-16(%rsi),%rbp | 
|  | .cfi_restore	%rbp | 
|  | movq	-8(%rsi),%rbx | 
|  | .cfi_restore	%rbx | 
|  | leaq	(%rsi),%rsp | 
|  | .cfi_def_cfa_register	%rsp | 
|  | .Lepilogue_avx: | 
|  | ret | 
|  | .cfi_endproc | 
|  | .size	sha256_block_data_order_avx,.-sha256_block_data_order_avx | 
|  | #endif |