| #if defined(__x86_64__) && !defined(OPENSSL_NO_ASM) |
| .text |
| .type _x86_64_AES_encrypt,@function |
| .align 16 |
| _x86_64_AES_encrypt: |
| xorl 0(%r15),%eax |
| xorl 4(%r15),%ebx |
| xorl 8(%r15),%ecx |
| xorl 12(%r15),%edx |
| |
| movl 240(%r15),%r13d |
| subl $1,%r13d |
| jmp .Lenc_loop |
| .align 16 |
| .Lenc_loop: |
| |
| movzbl %al,%esi |
| movzbl %bl,%edi |
| movzbl %cl,%ebp |
| movl 0(%r14,%rsi,8),%r10d |
| movl 0(%r14,%rdi,8),%r11d |
| movl 0(%r14,%rbp,8),%r12d |
| |
| movzbl %bh,%esi |
| movzbl %ch,%edi |
| movzbl %dl,%ebp |
| xorl 3(%r14,%rsi,8),%r10d |
| xorl 3(%r14,%rdi,8),%r11d |
| movl 0(%r14,%rbp,8),%r8d |
| |
| movzbl %dh,%esi |
| shrl $16,%ecx |
| movzbl %ah,%ebp |
| xorl 3(%r14,%rsi,8),%r12d |
| shrl $16,%edx |
| xorl 3(%r14,%rbp,8),%r8d |
| |
| shrl $16,%ebx |
| leaq 16(%r15),%r15 |
| shrl $16,%eax |
| |
| movzbl %cl,%esi |
| movzbl %dl,%edi |
| movzbl %al,%ebp |
| xorl 2(%r14,%rsi,8),%r10d |
| xorl 2(%r14,%rdi,8),%r11d |
| xorl 2(%r14,%rbp,8),%r12d |
| |
| movzbl %dh,%esi |
| movzbl %ah,%edi |
| movzbl %bl,%ebp |
| xorl 1(%r14,%rsi,8),%r10d |
| xorl 1(%r14,%rdi,8),%r11d |
| xorl 2(%r14,%rbp,8),%r8d |
| |
| movl 12(%r15),%edx |
| movzbl %bh,%edi |
| movzbl %ch,%ebp |
| movl 0(%r15),%eax |
| xorl 1(%r14,%rdi,8),%r12d |
| xorl 1(%r14,%rbp,8),%r8d |
| |
| movl 4(%r15),%ebx |
| movl 8(%r15),%ecx |
| xorl %r10d,%eax |
| xorl %r11d,%ebx |
| xorl %r12d,%ecx |
| xorl %r8d,%edx |
| subl $1,%r13d |
| jnz .Lenc_loop |
| movzbl %al,%esi |
| movzbl %bl,%edi |
| movzbl %cl,%ebp |
| movzbl 2(%r14,%rsi,8),%r10d |
| movzbl 2(%r14,%rdi,8),%r11d |
| movzbl 2(%r14,%rbp,8),%r12d |
| |
| movzbl %dl,%esi |
| movzbl %bh,%edi |
| movzbl %ch,%ebp |
| movzbl 2(%r14,%rsi,8),%r8d |
| movl 0(%r14,%rdi,8),%edi |
| movl 0(%r14,%rbp,8),%ebp |
| |
| andl $0x0000ff00,%edi |
| andl $0x0000ff00,%ebp |
| |
| xorl %edi,%r10d |
| xorl %ebp,%r11d |
| shrl $16,%ecx |
| |
| movzbl %dh,%esi |
| movzbl %ah,%edi |
| shrl $16,%edx |
| movl 0(%r14,%rsi,8),%esi |
| movl 0(%r14,%rdi,8),%edi |
| |
| andl $0x0000ff00,%esi |
| andl $0x0000ff00,%edi |
| shrl $16,%ebx |
| xorl %esi,%r12d |
| xorl %edi,%r8d |
| shrl $16,%eax |
| |
| movzbl %cl,%esi |
| movzbl %dl,%edi |
| movzbl %al,%ebp |
| movl 0(%r14,%rsi,8),%esi |
| movl 0(%r14,%rdi,8),%edi |
| movl 0(%r14,%rbp,8),%ebp |
| |
| andl $0x00ff0000,%esi |
| andl $0x00ff0000,%edi |
| andl $0x00ff0000,%ebp |
| |
| xorl %esi,%r10d |
| xorl %edi,%r11d |
| xorl %ebp,%r12d |
| |
| movzbl %bl,%esi |
| movzbl %dh,%edi |
| movzbl %ah,%ebp |
| movl 0(%r14,%rsi,8),%esi |
| movl 2(%r14,%rdi,8),%edi |
| movl 2(%r14,%rbp,8),%ebp |
| |
| andl $0x00ff0000,%esi |
| andl $0xff000000,%edi |
| andl $0xff000000,%ebp |
| |
| xorl %esi,%r8d |
| xorl %edi,%r10d |
| xorl %ebp,%r11d |
| |
| movzbl %bh,%esi |
| movzbl %ch,%edi |
| movl 16+12(%r15),%edx |
| movl 2(%r14,%rsi,8),%esi |
| movl 2(%r14,%rdi,8),%edi |
| movl 16+0(%r15),%eax |
| |
| andl $0xff000000,%esi |
| andl $0xff000000,%edi |
| |
| xorl %esi,%r12d |
| xorl %edi,%r8d |
| |
| movl 16+4(%r15),%ebx |
| movl 16+8(%r15),%ecx |
| xorl %r10d,%eax |
| xorl %r11d,%ebx |
| xorl %r12d,%ecx |
| xorl %r8d,%edx |
| .byte 0xf3,0xc3 |
| .size _x86_64_AES_encrypt,.-_x86_64_AES_encrypt |
| .type _x86_64_AES_encrypt_compact,@function |
| .align 16 |
| _x86_64_AES_encrypt_compact: |
| leaq 128(%r14),%r8 |
| movl 0-128(%r8),%edi |
| movl 32-128(%r8),%ebp |
| movl 64-128(%r8),%r10d |
| movl 96-128(%r8),%r11d |
| movl 128-128(%r8),%edi |
| movl 160-128(%r8),%ebp |
| movl 192-128(%r8),%r10d |
| movl 224-128(%r8),%r11d |
| jmp .Lenc_loop_compact |
| .align 16 |
| .Lenc_loop_compact: |
| xorl 0(%r15),%eax |
| xorl 4(%r15),%ebx |
| xorl 8(%r15),%ecx |
| xorl 12(%r15),%edx |
| leaq 16(%r15),%r15 |
| movzbl %al,%r10d |
| movzbl %bl,%r11d |
| movzbl %cl,%r12d |
| movzbl %dl,%r8d |
| movzbl %bh,%esi |
| movzbl %ch,%edi |
| shrl $16,%ecx |
| movzbl %dh,%ebp |
| movzbl (%r14,%r10,1),%r10d |
| movzbl (%r14,%r11,1),%r11d |
| movzbl (%r14,%r12,1),%r12d |
| movzbl (%r14,%r8,1),%r8d |
| |
| movzbl (%r14,%rsi,1),%r9d |
| movzbl %ah,%esi |
| movzbl (%r14,%rdi,1),%r13d |
| movzbl %cl,%edi |
| movzbl (%r14,%rbp,1),%ebp |
| movzbl (%r14,%rsi,1),%esi |
| |
| shll $8,%r9d |
| shrl $16,%edx |
| shll $8,%r13d |
| xorl %r9d,%r10d |
| shrl $16,%eax |
| movzbl %dl,%r9d |
| shrl $16,%ebx |
| xorl %r13d,%r11d |
| shll $8,%ebp |
| movzbl %al,%r13d |
| movzbl (%r14,%rdi,1),%edi |
| xorl %ebp,%r12d |
| |
| shll $8,%esi |
| movzbl %bl,%ebp |
| shll $16,%edi |
| xorl %esi,%r8d |
| movzbl (%r14,%r9,1),%r9d |
| movzbl %dh,%esi |
| movzbl (%r14,%r13,1),%r13d |
| xorl %edi,%r10d |
| |
| shrl $8,%ecx |
| movzbl %ah,%edi |
| shll $16,%r9d |
| shrl $8,%ebx |
| shll $16,%r13d |
| xorl %r9d,%r11d |
| movzbl (%r14,%rbp,1),%ebp |
| movzbl (%r14,%rsi,1),%esi |
| movzbl (%r14,%rdi,1),%edi |
| movzbl (%r14,%rcx,1),%edx |
| movzbl (%r14,%rbx,1),%ecx |
| |
| shll $16,%ebp |
| xorl %r13d,%r12d |
| shll $24,%esi |
| xorl %ebp,%r8d |
| shll $24,%edi |
| xorl %esi,%r10d |
| shll $24,%edx |
| xorl %edi,%r11d |
| shll $24,%ecx |
| movl %r10d,%eax |
| movl %r11d,%ebx |
| xorl %r12d,%ecx |
| xorl %r8d,%edx |
| cmpq 16(%rsp),%r15 |
| je .Lenc_compact_done |
| movl $0x80808080,%r10d |
| movl $0x80808080,%r11d |
| andl %eax,%r10d |
| andl %ebx,%r11d |
| movl %r10d,%esi |
| movl %r11d,%edi |
| shrl $7,%r10d |
| leal (%rax,%rax,1),%r8d |
| shrl $7,%r11d |
| leal (%rbx,%rbx,1),%r9d |
| subl %r10d,%esi |
| subl %r11d,%edi |
| andl $0xfefefefe,%r8d |
| andl $0xfefefefe,%r9d |
| andl $0x1b1b1b1b,%esi |
| andl $0x1b1b1b1b,%edi |
| movl %eax,%r10d |
| movl %ebx,%r11d |
| xorl %esi,%r8d |
| xorl %edi,%r9d |
| |
| xorl %r8d,%eax |
| xorl %r9d,%ebx |
| movl $0x80808080,%r12d |
| roll $24,%eax |
| movl $0x80808080,%ebp |
| roll $24,%ebx |
| andl %ecx,%r12d |
| andl %edx,%ebp |
| xorl %r8d,%eax |
| xorl %r9d,%ebx |
| movl %r12d,%esi |
| rorl $16,%r10d |
| movl %ebp,%edi |
| rorl $16,%r11d |
| leal (%rcx,%rcx,1),%r8d |
| shrl $7,%r12d |
| xorl %r10d,%eax |
| shrl $7,%ebp |
| xorl %r11d,%ebx |
| rorl $8,%r10d |
| leal (%rdx,%rdx,1),%r9d |
| rorl $8,%r11d |
| subl %r12d,%esi |
| subl %ebp,%edi |
| xorl %r10d,%eax |
| xorl %r11d,%ebx |
| |
| andl $0xfefefefe,%r8d |
| andl $0xfefefefe,%r9d |
| andl $0x1b1b1b1b,%esi |
| andl $0x1b1b1b1b,%edi |
| movl %ecx,%r12d |
| movl %edx,%ebp |
| xorl %esi,%r8d |
| xorl %edi,%r9d |
| |
| rorl $16,%r12d |
| xorl %r8d,%ecx |
| rorl $16,%ebp |
| xorl %r9d,%edx |
| roll $24,%ecx |
| movl 0(%r14),%esi |
| roll $24,%edx |
| xorl %r8d,%ecx |
| movl 64(%r14),%edi |
| xorl %r9d,%edx |
| movl 128(%r14),%r8d |
| xorl %r12d,%ecx |
| rorl $8,%r12d |
| xorl %ebp,%edx |
| rorl $8,%ebp |
| xorl %r12d,%ecx |
| movl 192(%r14),%r9d |
| xorl %ebp,%edx |
| jmp .Lenc_loop_compact |
| .align 16 |
| .Lenc_compact_done: |
| xorl 0(%r15),%eax |
| xorl 4(%r15),%ebx |
| xorl 8(%r15),%ecx |
| xorl 12(%r15),%edx |
| .byte 0xf3,0xc3 |
| .size _x86_64_AES_encrypt_compact,.-_x86_64_AES_encrypt_compact |
| .align 16 |
| .globl aes_nohw_encrypt |
| .hidden aes_nohw_encrypt |
| .type aes_nohw_encrypt,@function |
| .hidden aes_nohw_encrypt |
| aes_nohw_encrypt: |
| .cfi_startproc |
| movq %rsp,%rax |
| .cfi_def_cfa_register %rax |
| pushq %rbx |
| .cfi_offset %rbx,-16 |
| pushq %rbp |
| .cfi_offset %rbp,-24 |
| pushq %r12 |
| .cfi_offset %r12,-32 |
| pushq %r13 |
| .cfi_offset %r13,-40 |
| pushq %r14 |
| .cfi_offset %r14,-48 |
| pushq %r15 |
| .cfi_offset %r15,-56 |
| |
| |
| leaq -63(%rdx),%rcx |
| andq $-64,%rsp |
| subq %rsp,%rcx |
| negq %rcx |
| andq $0x3c0,%rcx |
| subq %rcx,%rsp |
| subq $32,%rsp |
| |
| movq %rsi,16(%rsp) |
| movq %rax,24(%rsp) |
| .cfi_escape 0x0f,0x05,0x77,0x18,0x06,0x23,0x08 |
| .Lenc_prologue: |
| |
| movq %rdx,%r15 |
| movl 240(%r15),%r13d |
| |
| movl 0(%rdi),%eax |
| movl 4(%rdi),%ebx |
| movl 8(%rdi),%ecx |
| movl 12(%rdi),%edx |
| |
| shll $4,%r13d |
| leaq (%r15,%r13,1),%rbp |
| movq %r15,(%rsp) |
| movq %rbp,8(%rsp) |
| |
| |
| leaq .LAES_Te+2048(%rip),%r14 |
| leaq 768(%rsp),%rbp |
| subq %r14,%rbp |
| andq $0x300,%rbp |
| leaq (%r14,%rbp,1),%r14 |
| |
| call _x86_64_AES_encrypt_compact |
| |
| movq 16(%rsp),%r9 |
| movq 24(%rsp),%rsi |
| .cfi_def_cfa %rsi,8 |
| movl %eax,0(%r9) |
| movl %ebx,4(%r9) |
| movl %ecx,8(%r9) |
| movl %edx,12(%r9) |
| |
| movq -48(%rsi),%r15 |
| .cfi_restore %r15 |
| movq -40(%rsi),%r14 |
| .cfi_restore %r14 |
| movq -32(%rsi),%r13 |
| .cfi_restore %r13 |
| movq -24(%rsi),%r12 |
| .cfi_restore %r12 |
| movq -16(%rsi),%rbp |
| .cfi_restore %rbp |
| movq -8(%rsi),%rbx |
| .cfi_restore %rbx |
| leaq (%rsi),%rsp |
| .cfi_def_cfa_register %rsp |
| .Lenc_epilogue: |
| .byte 0xf3,0xc3 |
| .cfi_endproc |
| .size aes_nohw_encrypt,.-aes_nohw_encrypt |
| .type _x86_64_AES_decrypt,@function |
| .align 16 |
| _x86_64_AES_decrypt: |
| xorl 0(%r15),%eax |
| xorl 4(%r15),%ebx |
| xorl 8(%r15),%ecx |
| xorl 12(%r15),%edx |
| |
| movl 240(%r15),%r13d |
| subl $1,%r13d |
| jmp .Ldec_loop |
| .align 16 |
| .Ldec_loop: |
| |
| movzbl %al,%esi |
| movzbl %bl,%edi |
| movzbl %cl,%ebp |
| movl 0(%r14,%rsi,8),%r10d |
| movl 0(%r14,%rdi,8),%r11d |
| movl 0(%r14,%rbp,8),%r12d |
| |
| movzbl %dh,%esi |
| movzbl %ah,%edi |
| movzbl %dl,%ebp |
| xorl 3(%r14,%rsi,8),%r10d |
| xorl 3(%r14,%rdi,8),%r11d |
| movl 0(%r14,%rbp,8),%r8d |
| |
| movzbl %bh,%esi |
| shrl $16,%eax |
| movzbl %ch,%ebp |
| xorl 3(%r14,%rsi,8),%r12d |
| shrl $16,%edx |
| xorl 3(%r14,%rbp,8),%r8d |
| |
| shrl $16,%ebx |
| leaq 16(%r15),%r15 |
| shrl $16,%ecx |
| |
| movzbl %cl,%esi |
| movzbl %dl,%edi |
| movzbl %al,%ebp |
| xorl 2(%r14,%rsi,8),%r10d |
| xorl 2(%r14,%rdi,8),%r11d |
| xorl 2(%r14,%rbp,8),%r12d |
| |
| movzbl %bh,%esi |
| movzbl %ch,%edi |
| movzbl %bl,%ebp |
| xorl 1(%r14,%rsi,8),%r10d |
| xorl 1(%r14,%rdi,8),%r11d |
| xorl 2(%r14,%rbp,8),%r8d |
| |
| movzbl %dh,%esi |
| movl 12(%r15),%edx |
| movzbl %ah,%ebp |
| xorl 1(%r14,%rsi,8),%r12d |
| movl 0(%r15),%eax |
| xorl 1(%r14,%rbp,8),%r8d |
| |
| xorl %r10d,%eax |
| movl 4(%r15),%ebx |
| movl 8(%r15),%ecx |
| xorl %r12d,%ecx |
| xorl %r11d,%ebx |
| xorl %r8d,%edx |
| subl $1,%r13d |
| jnz .Ldec_loop |
| leaq 2048(%r14),%r14 |
| movzbl %al,%esi |
| movzbl %bl,%edi |
| movzbl %cl,%ebp |
| movzbl (%r14,%rsi,1),%r10d |
| movzbl (%r14,%rdi,1),%r11d |
| movzbl (%r14,%rbp,1),%r12d |
| |
| movzbl %dl,%esi |
| movzbl %dh,%edi |
| movzbl %ah,%ebp |
| movzbl (%r14,%rsi,1),%r8d |
| movzbl (%r14,%rdi,1),%edi |
| movzbl (%r14,%rbp,1),%ebp |
| |
| shll $8,%edi |
| shll $8,%ebp |
| |
| xorl %edi,%r10d |
| xorl %ebp,%r11d |
| shrl $16,%edx |
| |
| movzbl %bh,%esi |
| movzbl %ch,%edi |
| shrl $16,%eax |
| movzbl (%r14,%rsi,1),%esi |
| movzbl (%r14,%rdi,1),%edi |
| |
| shll $8,%esi |
| shll $8,%edi |
| shrl $16,%ebx |
| xorl %esi,%r12d |
| xorl %edi,%r8d |
| shrl $16,%ecx |
| |
| movzbl %cl,%esi |
| movzbl %dl,%edi |
| movzbl %al,%ebp |
| movzbl (%r14,%rsi,1),%esi |
| movzbl (%r14,%rdi,1),%edi |
| movzbl (%r14,%rbp,1),%ebp |
| |
| shll $16,%esi |
| shll $16,%edi |
| shll $16,%ebp |
| |
| xorl %esi,%r10d |
| xorl %edi,%r11d |
| xorl %ebp,%r12d |
| |
| movzbl %bl,%esi |
| movzbl %bh,%edi |
| movzbl %ch,%ebp |
| movzbl (%r14,%rsi,1),%esi |
| movzbl (%r14,%rdi,1),%edi |
| movzbl (%r14,%rbp,1),%ebp |
| |
| shll $16,%esi |
| shll $24,%edi |
| shll $24,%ebp |
| |
| xorl %esi,%r8d |
| xorl %edi,%r10d |
| xorl %ebp,%r11d |
| |
| movzbl %dh,%esi |
| movzbl %ah,%edi |
| movl 16+12(%r15),%edx |
| movzbl (%r14,%rsi,1),%esi |
| movzbl (%r14,%rdi,1),%edi |
| movl 16+0(%r15),%eax |
| |
| shll $24,%esi |
| shll $24,%edi |
| |
| xorl %esi,%r12d |
| xorl %edi,%r8d |
| |
| movl 16+4(%r15),%ebx |
| movl 16+8(%r15),%ecx |
| leaq -2048(%r14),%r14 |
| xorl %r10d,%eax |
| xorl %r11d,%ebx |
| xorl %r12d,%ecx |
| xorl %r8d,%edx |
| .byte 0xf3,0xc3 |
| .size _x86_64_AES_decrypt,.-_x86_64_AES_decrypt |
| .type _x86_64_AES_decrypt_compact,@function |
| .align 16 |
| _x86_64_AES_decrypt_compact: |
| leaq 128(%r14),%r8 |
| movl 0-128(%r8),%edi |
| movl 32-128(%r8),%ebp |
| movl 64-128(%r8),%r10d |
| movl 96-128(%r8),%r11d |
| movl 128-128(%r8),%edi |
| movl 160-128(%r8),%ebp |
| movl 192-128(%r8),%r10d |
| movl 224-128(%r8),%r11d |
| jmp .Ldec_loop_compact |
| |
| .align 16 |
| .Ldec_loop_compact: |
| xorl 0(%r15),%eax |
| xorl 4(%r15),%ebx |
| xorl 8(%r15),%ecx |
| xorl 12(%r15),%edx |
| leaq 16(%r15),%r15 |
| movzbl %al,%r10d |
| movzbl %bl,%r11d |
| movzbl %cl,%r12d |
| movzbl %dl,%r8d |
| movzbl %dh,%esi |
| movzbl %ah,%edi |
| shrl $16,%edx |
| movzbl %bh,%ebp |
| movzbl (%r14,%r10,1),%r10d |
| movzbl (%r14,%r11,1),%r11d |
| movzbl (%r14,%r12,1),%r12d |
| movzbl (%r14,%r8,1),%r8d |
| |
| movzbl (%r14,%rsi,1),%r9d |
| movzbl %ch,%esi |
| movzbl (%r14,%rdi,1),%r13d |
| movzbl (%r14,%rbp,1),%ebp |
| movzbl (%r14,%rsi,1),%esi |
| |
| shrl $16,%ecx |
| shll $8,%r13d |
| shll $8,%r9d |
| movzbl %cl,%edi |
| shrl $16,%eax |
| xorl %r9d,%r10d |
| shrl $16,%ebx |
| movzbl %dl,%r9d |
| |
| shll $8,%ebp |
| xorl %r13d,%r11d |
| shll $8,%esi |
| movzbl %al,%r13d |
| movzbl (%r14,%rdi,1),%edi |
| xorl %ebp,%r12d |
| movzbl %bl,%ebp |
| |
| shll $16,%edi |
| xorl %esi,%r8d |
| movzbl (%r14,%r9,1),%r9d |
| movzbl %bh,%esi |
| movzbl (%r14,%rbp,1),%ebp |
| xorl %edi,%r10d |
| movzbl (%r14,%r13,1),%r13d |
| movzbl %ch,%edi |
| |
| shll $16,%ebp |
| shll $16,%r9d |
| shll $16,%r13d |
| xorl %ebp,%r8d |
| movzbl %dh,%ebp |
| xorl %r9d,%r11d |
| shrl $8,%eax |
| xorl %r13d,%r12d |
| |
| movzbl (%r14,%rsi,1),%esi |
| movzbl (%r14,%rdi,1),%ebx |
| movzbl (%r14,%rbp,1),%ecx |
| movzbl (%r14,%rax,1),%edx |
| |
| movl %r10d,%eax |
| shll $24,%esi |
| shll $24,%ebx |
| shll $24,%ecx |
| xorl %esi,%eax |
| shll $24,%edx |
| xorl %r11d,%ebx |
| xorl %r12d,%ecx |
| xorl %r8d,%edx |
| cmpq 16(%rsp),%r15 |
| je .Ldec_compact_done |
| |
| movq 256+0(%r14),%rsi |
| shlq $32,%rbx |
| shlq $32,%rdx |
| movq 256+8(%r14),%rdi |
| orq %rbx,%rax |
| orq %rdx,%rcx |
| movq 256+16(%r14),%rbp |
| movq %rsi,%r9 |
| movq %rsi,%r12 |
| andq %rax,%r9 |
| andq %rcx,%r12 |
| movq %r9,%rbx |
| movq %r12,%rdx |
| shrq $7,%r9 |
| leaq (%rax,%rax,1),%r8 |
| shrq $7,%r12 |
| leaq (%rcx,%rcx,1),%r11 |
| subq %r9,%rbx |
| subq %r12,%rdx |
| andq %rdi,%r8 |
| andq %rdi,%r11 |
| andq %rbp,%rbx |
| andq %rbp,%rdx |
| xorq %rbx,%r8 |
| xorq %rdx,%r11 |
| movq %rsi,%r10 |
| movq %rsi,%r13 |
| |
| andq %r8,%r10 |
| andq %r11,%r13 |
| movq %r10,%rbx |
| movq %r13,%rdx |
| shrq $7,%r10 |
| leaq (%r8,%r8,1),%r9 |
| shrq $7,%r13 |
| leaq (%r11,%r11,1),%r12 |
| subq %r10,%rbx |
| subq %r13,%rdx |
| andq %rdi,%r9 |
| andq %rdi,%r12 |
| andq %rbp,%rbx |
| andq %rbp,%rdx |
| xorq %rbx,%r9 |
| xorq %rdx,%r12 |
| movq %rsi,%r10 |
| movq %rsi,%r13 |
| |
| andq %r9,%r10 |
| andq %r12,%r13 |
| movq %r10,%rbx |
| movq %r13,%rdx |
| shrq $7,%r10 |
| xorq %rax,%r8 |
| shrq $7,%r13 |
| xorq %rcx,%r11 |
| subq %r10,%rbx |
| subq %r13,%rdx |
| leaq (%r9,%r9,1),%r10 |
| leaq (%r12,%r12,1),%r13 |
| xorq %rax,%r9 |
| xorq %rcx,%r12 |
| andq %rdi,%r10 |
| andq %rdi,%r13 |
| andq %rbp,%rbx |
| andq %rbp,%rdx |
| xorq %rbx,%r10 |
| xorq %rdx,%r13 |
| |
| xorq %r10,%rax |
| xorq %r13,%rcx |
| xorq %r10,%r8 |
| xorq %r13,%r11 |
| movq %rax,%rbx |
| movq %rcx,%rdx |
| xorq %r10,%r9 |
| shrq $32,%rbx |
| xorq %r13,%r12 |
| shrq $32,%rdx |
| xorq %r8,%r10 |
| roll $8,%eax |
| xorq %r11,%r13 |
| roll $8,%ecx |
| xorq %r9,%r10 |
| roll $8,%ebx |
| xorq %r12,%r13 |
| |
| roll $8,%edx |
| xorl %r10d,%eax |
| shrq $32,%r10 |
| xorl %r13d,%ecx |
| shrq $32,%r13 |
| xorl %r10d,%ebx |
| xorl %r13d,%edx |
| |
| movq %r8,%r10 |
| roll $24,%r8d |
| movq %r11,%r13 |
| roll $24,%r11d |
| shrq $32,%r10 |
| xorl %r8d,%eax |
| shrq $32,%r13 |
| xorl %r11d,%ecx |
| roll $24,%r10d |
| movq %r9,%r8 |
| roll $24,%r13d |
| movq %r12,%r11 |
| shrq $32,%r8 |
| xorl %r10d,%ebx |
| shrq $32,%r11 |
| xorl %r13d,%edx |
| |
| movq 0(%r14),%rsi |
| roll $16,%r9d |
| movq 64(%r14),%rdi |
| roll $16,%r12d |
| movq 128(%r14),%rbp |
| roll $16,%r8d |
| movq 192(%r14),%r10 |
| xorl %r9d,%eax |
| roll $16,%r11d |
| xorl %r12d,%ecx |
| movq 256(%r14),%r13 |
| xorl %r8d,%ebx |
| xorl %r11d,%edx |
| jmp .Ldec_loop_compact |
| .align 16 |
| .Ldec_compact_done: |
| xorl 0(%r15),%eax |
| xorl 4(%r15),%ebx |
| xorl 8(%r15),%ecx |
| xorl 12(%r15),%edx |
| .byte 0xf3,0xc3 |
| .size _x86_64_AES_decrypt_compact,.-_x86_64_AES_decrypt_compact |
| .align 16 |
| .globl aes_nohw_decrypt |
| .hidden aes_nohw_decrypt |
| .type aes_nohw_decrypt,@function |
| .hidden aes_nohw_decrypt |
| aes_nohw_decrypt: |
| .cfi_startproc |
| movq %rsp,%rax |
| .cfi_def_cfa_register %rax |
| pushq %rbx |
| .cfi_offset %rbx,-16 |
| pushq %rbp |
| .cfi_offset %rbp,-24 |
| pushq %r12 |
| .cfi_offset %r12,-32 |
| pushq %r13 |
| .cfi_offset %r13,-40 |
| pushq %r14 |
| .cfi_offset %r14,-48 |
| pushq %r15 |
| .cfi_offset %r15,-56 |
| |
| |
| leaq -63(%rdx),%rcx |
| andq $-64,%rsp |
| subq %rsp,%rcx |
| negq %rcx |
| andq $0x3c0,%rcx |
| subq %rcx,%rsp |
| subq $32,%rsp |
| |
| movq %rsi,16(%rsp) |
| movq %rax,24(%rsp) |
| .cfi_escape 0x0f,0x05,0x77,0x18,0x06,0x23,0x08 |
| .Ldec_prologue: |
| |
| movq %rdx,%r15 |
| movl 240(%r15),%r13d |
| |
| movl 0(%rdi),%eax |
| movl 4(%rdi),%ebx |
| movl 8(%rdi),%ecx |
| movl 12(%rdi),%edx |
| |
| shll $4,%r13d |
| leaq (%r15,%r13,1),%rbp |
| movq %r15,(%rsp) |
| movq %rbp,8(%rsp) |
| |
| |
| leaq .LAES_Td+2048(%rip),%r14 |
| leaq 768(%rsp),%rbp |
| subq %r14,%rbp |
| andq $0x300,%rbp |
| leaq (%r14,%rbp,1),%r14 |
| shrq $3,%rbp |
| addq %rbp,%r14 |
| |
| call _x86_64_AES_decrypt_compact |
| |
| movq 16(%rsp),%r9 |
| movq 24(%rsp),%rsi |
| .cfi_def_cfa %rsi,8 |
| movl %eax,0(%r9) |
| movl %ebx,4(%r9) |
| movl %ecx,8(%r9) |
| movl %edx,12(%r9) |
| |
| movq -48(%rsi),%r15 |
| .cfi_restore %r15 |
| movq -40(%rsi),%r14 |
| .cfi_restore %r14 |
| movq -32(%rsi),%r13 |
| .cfi_restore %r13 |
| movq -24(%rsi),%r12 |
| .cfi_restore %r12 |
| movq -16(%rsi),%rbp |
| .cfi_restore %rbp |
| movq -8(%rsi),%rbx |
| .cfi_restore %rbx |
| leaq (%rsi),%rsp |
| .cfi_def_cfa_register %rsp |
| .Ldec_epilogue: |
| .byte 0xf3,0xc3 |
| .cfi_endproc |
| .size aes_nohw_decrypt,.-aes_nohw_decrypt |
| .align 16 |
| .globl aes_nohw_set_encrypt_key |
| .hidden aes_nohw_set_encrypt_key |
| .type aes_nohw_set_encrypt_key,@function |
| aes_nohw_set_encrypt_key: |
| .cfi_startproc |
| pushq %rbx |
| .cfi_adjust_cfa_offset 8 |
| .cfi_offset %rbx,-16 |
| pushq %rbp |
| .cfi_adjust_cfa_offset 8 |
| .cfi_offset %rbp,-24 |
| pushq %r12 |
| .cfi_adjust_cfa_offset 8 |
| .cfi_offset %r12,-32 |
| pushq %r13 |
| .cfi_adjust_cfa_offset 8 |
| .cfi_offset %r13,-40 |
| pushq %r14 |
| .cfi_adjust_cfa_offset 8 |
| .cfi_offset %r14,-48 |
| pushq %r15 |
| .cfi_adjust_cfa_offset 8 |
| .cfi_offset %r15,-56 |
| subq $8,%rsp |
| .cfi_adjust_cfa_offset 8 |
| .Lenc_key_prologue: |
| |
| call _x86_64_AES_set_encrypt_key |
| |
| movq 40(%rsp),%rbp |
| .cfi_restore %rbp |
| movq 48(%rsp),%rbx |
| .cfi_restore %rbx |
| addq $56,%rsp |
| .cfi_adjust_cfa_offset -56 |
| .Lenc_key_epilogue: |
| .byte 0xf3,0xc3 |
| .cfi_endproc |
| .size aes_nohw_set_encrypt_key,.-aes_nohw_set_encrypt_key |
| |
| .type _x86_64_AES_set_encrypt_key,@function |
| .align 16 |
| _x86_64_AES_set_encrypt_key: |
| movl %esi,%ecx |
| movq %rdi,%rsi |
| movq %rdx,%rdi |
| |
| testq $-1,%rsi |
| jz .Lbadpointer |
| testq $-1,%rdi |
| jz .Lbadpointer |
| |
| leaq .LAES_Te(%rip),%rbp |
| leaq 2048+128(%rbp),%rbp |
| |
| |
| movl 0-128(%rbp),%eax |
| movl 32-128(%rbp),%ebx |
| movl 64-128(%rbp),%r8d |
| movl 96-128(%rbp),%edx |
| movl 128-128(%rbp),%eax |
| movl 160-128(%rbp),%ebx |
| movl 192-128(%rbp),%r8d |
| movl 224-128(%rbp),%edx |
| |
| cmpl $128,%ecx |
| je .L10rounds |
| cmpl $192,%ecx |
| je .L12rounds |
| cmpl $256,%ecx |
| je .L14rounds |
| movq $-2,%rax |
| jmp .Lexit |
| |
| .L10rounds: |
| movq 0(%rsi),%rax |
| movq 8(%rsi),%rdx |
| movq %rax,0(%rdi) |
| movq %rdx,8(%rdi) |
| |
| shrq $32,%rdx |
| xorl %ecx,%ecx |
| jmp .L10shortcut |
| .align 4 |
| .L10loop: |
| movl 0(%rdi),%eax |
| movl 12(%rdi),%edx |
| .L10shortcut: |
| movzbl %dl,%esi |
| movzbl -128(%rbp,%rsi,1),%ebx |
| movzbl %dh,%esi |
| shll $24,%ebx |
| xorl %ebx,%eax |
| |
| movzbl -128(%rbp,%rsi,1),%ebx |
| shrl $16,%edx |
| movzbl %dl,%esi |
| xorl %ebx,%eax |
| |
| movzbl -128(%rbp,%rsi,1),%ebx |
| movzbl %dh,%esi |
| shll $8,%ebx |
| xorl %ebx,%eax |
| |
| movzbl -128(%rbp,%rsi,1),%ebx |
| shll $16,%ebx |
| xorl %ebx,%eax |
| |
| xorl 1024-128(%rbp,%rcx,4),%eax |
| movl %eax,16(%rdi) |
| xorl 4(%rdi),%eax |
| movl %eax,20(%rdi) |
| xorl 8(%rdi),%eax |
| movl %eax,24(%rdi) |
| xorl 12(%rdi),%eax |
| movl %eax,28(%rdi) |
| addl $1,%ecx |
| leaq 16(%rdi),%rdi |
| cmpl $10,%ecx |
| jl .L10loop |
| |
| movl $10,80(%rdi) |
| xorq %rax,%rax |
| jmp .Lexit |
| |
| .L12rounds: |
| movq 0(%rsi),%rax |
| movq 8(%rsi),%rbx |
| movq 16(%rsi),%rdx |
| movq %rax,0(%rdi) |
| movq %rbx,8(%rdi) |
| movq %rdx,16(%rdi) |
| |
| shrq $32,%rdx |
| xorl %ecx,%ecx |
| jmp .L12shortcut |
| .align 4 |
| .L12loop: |
| movl 0(%rdi),%eax |
| movl 20(%rdi),%edx |
| .L12shortcut: |
| movzbl %dl,%esi |
| movzbl -128(%rbp,%rsi,1),%ebx |
| movzbl %dh,%esi |
| shll $24,%ebx |
| xorl %ebx,%eax |
| |
| movzbl -128(%rbp,%rsi,1),%ebx |
| shrl $16,%edx |
| movzbl %dl,%esi |
| xorl %ebx,%eax |
| |
| movzbl -128(%rbp,%rsi,1),%ebx |
| movzbl %dh,%esi |
| shll $8,%ebx |
| xorl %ebx,%eax |
| |
| movzbl -128(%rbp,%rsi,1),%ebx |
| shll $16,%ebx |
| xorl %ebx,%eax |
| |
| xorl 1024-128(%rbp,%rcx,4),%eax |
| movl %eax,24(%rdi) |
| xorl 4(%rdi),%eax |
| movl %eax,28(%rdi) |
| xorl 8(%rdi),%eax |
| movl %eax,32(%rdi) |
| xorl 12(%rdi),%eax |
| movl %eax,36(%rdi) |
| |
| cmpl $7,%ecx |
| je .L12break |
| addl $1,%ecx |
| |
| xorl 16(%rdi),%eax |
| movl %eax,40(%rdi) |
| xorl 20(%rdi),%eax |
| movl %eax,44(%rdi) |
| |
| leaq 24(%rdi),%rdi |
| jmp .L12loop |
| .L12break: |
| movl $12,72(%rdi) |
| xorq %rax,%rax |
| jmp .Lexit |
| |
| .L14rounds: |
| movq 0(%rsi),%rax |
| movq 8(%rsi),%rbx |
| movq 16(%rsi),%rcx |
| movq 24(%rsi),%rdx |
| movq %rax,0(%rdi) |
| movq %rbx,8(%rdi) |
| movq %rcx,16(%rdi) |
| movq %rdx,24(%rdi) |
| |
| shrq $32,%rdx |
| xorl %ecx,%ecx |
| jmp .L14shortcut |
| .align 4 |
| .L14loop: |
| movl 0(%rdi),%eax |
| movl 28(%rdi),%edx |
| .L14shortcut: |
| movzbl %dl,%esi |
| movzbl -128(%rbp,%rsi,1),%ebx |
| movzbl %dh,%esi |
| shll $24,%ebx |
| xorl %ebx,%eax |
| |
| movzbl -128(%rbp,%rsi,1),%ebx |
| shrl $16,%edx |
| movzbl %dl,%esi |
| xorl %ebx,%eax |
| |
| movzbl -128(%rbp,%rsi,1),%ebx |
| movzbl %dh,%esi |
| shll $8,%ebx |
| xorl %ebx,%eax |
| |
| movzbl -128(%rbp,%rsi,1),%ebx |
| shll $16,%ebx |
| xorl %ebx,%eax |
| |
| xorl 1024-128(%rbp,%rcx,4),%eax |
| movl %eax,32(%rdi) |
| xorl 4(%rdi),%eax |
| movl %eax,36(%rdi) |
| xorl 8(%rdi),%eax |
| movl %eax,40(%rdi) |
| xorl 12(%rdi),%eax |
| movl %eax,44(%rdi) |
| |
| cmpl $6,%ecx |
| je .L14break |
| addl $1,%ecx |
| |
| movl %eax,%edx |
| movl 16(%rdi),%eax |
| movzbl %dl,%esi |
| movzbl -128(%rbp,%rsi,1),%ebx |
| movzbl %dh,%esi |
| xorl %ebx,%eax |
| |
| movzbl -128(%rbp,%rsi,1),%ebx |
| shrl $16,%edx |
| shll $8,%ebx |
| movzbl %dl,%esi |
| xorl %ebx,%eax |
| |
| movzbl -128(%rbp,%rsi,1),%ebx |
| movzbl %dh,%esi |
| shll $16,%ebx |
| xorl %ebx,%eax |
| |
| movzbl -128(%rbp,%rsi,1),%ebx |
| shll $24,%ebx |
| xorl %ebx,%eax |
| |
| movl %eax,48(%rdi) |
| xorl 20(%rdi),%eax |
| movl %eax,52(%rdi) |
| xorl 24(%rdi),%eax |
| movl %eax,56(%rdi) |
| xorl 28(%rdi),%eax |
| movl %eax,60(%rdi) |
| |
| leaq 32(%rdi),%rdi |
| jmp .L14loop |
| .L14break: |
| movl $14,48(%rdi) |
| xorq %rax,%rax |
| jmp .Lexit |
| |
| .Lbadpointer: |
| movq $-1,%rax |
| .Lexit: |
| .byte 0xf3,0xc3 |
| .size _x86_64_AES_set_encrypt_key,.-_x86_64_AES_set_encrypt_key |
| .align 16 |
| .globl aes_nohw_set_decrypt_key |
| .hidden aes_nohw_set_decrypt_key |
| .type aes_nohw_set_decrypt_key,@function |
| aes_nohw_set_decrypt_key: |
| .cfi_startproc |
| pushq %rbx |
| .cfi_adjust_cfa_offset 8 |
| .cfi_offset %rbx,-16 |
| pushq %rbp |
| .cfi_adjust_cfa_offset 8 |
| .cfi_offset %rbp,-24 |
| pushq %r12 |
| .cfi_adjust_cfa_offset 8 |
| .cfi_offset %r12,-32 |
| pushq %r13 |
| .cfi_adjust_cfa_offset 8 |
| .cfi_offset %r13,-40 |
| pushq %r14 |
| .cfi_adjust_cfa_offset 8 |
| .cfi_offset %r14,-48 |
| pushq %r15 |
| .cfi_adjust_cfa_offset 8 |
| .cfi_offset %r15,-56 |
| pushq %rdx |
| .cfi_adjust_cfa_offset 8 |
| .Ldec_key_prologue: |
| |
| call _x86_64_AES_set_encrypt_key |
| movq (%rsp),%r8 |
| cmpl $0,%eax |
| jne .Labort |
| |
| movl 240(%r8),%r14d |
| xorq %rdi,%rdi |
| leaq (%rdi,%r14,4),%rcx |
| movq %r8,%rsi |
| leaq (%r8,%rcx,4),%rdi |
| .align 4 |
| .Linvert: |
| movq 0(%rsi),%rax |
| movq 8(%rsi),%rbx |
| movq 0(%rdi),%rcx |
| movq 8(%rdi),%rdx |
| movq %rax,0(%rdi) |
| movq %rbx,8(%rdi) |
| movq %rcx,0(%rsi) |
| movq %rdx,8(%rsi) |
| leaq 16(%rsi),%rsi |
| leaq -16(%rdi),%rdi |
| cmpq %rsi,%rdi |
| jne .Linvert |
| |
| leaq .LAES_Te+2048+1024(%rip),%rax |
| |
| movq 40(%rax),%rsi |
| movq 48(%rax),%rdi |
| movq 56(%rax),%rbp |
| |
| movq %r8,%r15 |
| subl $1,%r14d |
| .align 4 |
| .Lpermute: |
| leaq 16(%r15),%r15 |
| movq 0(%r15),%rax |
| movq 8(%r15),%rcx |
| movq %rsi,%r9 |
| movq %rsi,%r12 |
| andq %rax,%r9 |
| andq %rcx,%r12 |
| movq %r9,%rbx |
| movq %r12,%rdx |
| shrq $7,%r9 |
| leaq (%rax,%rax,1),%r8 |
| shrq $7,%r12 |
| leaq (%rcx,%rcx,1),%r11 |
| subq %r9,%rbx |
| subq %r12,%rdx |
| andq %rdi,%r8 |
| andq %rdi,%r11 |
| andq %rbp,%rbx |
| andq %rbp,%rdx |
| xorq %rbx,%r8 |
| xorq %rdx,%r11 |
| movq %rsi,%r10 |
| movq %rsi,%r13 |
| |
| andq %r8,%r10 |
| andq %r11,%r13 |
| movq %r10,%rbx |
| movq %r13,%rdx |
| shrq $7,%r10 |
| leaq (%r8,%r8,1),%r9 |
| shrq $7,%r13 |
| leaq (%r11,%r11,1),%r12 |
| subq %r10,%rbx |
| subq %r13,%rdx |
| andq %rdi,%r9 |
| andq %rdi,%r12 |
| andq %rbp,%rbx |
| andq %rbp,%rdx |
| xorq %rbx,%r9 |
| xorq %rdx,%r12 |
| movq %rsi,%r10 |
| movq %rsi,%r13 |
| |
| andq %r9,%r10 |
| andq %r12,%r13 |
| movq %r10,%rbx |
| movq %r13,%rdx |
| shrq $7,%r10 |
| xorq %rax,%r8 |
| shrq $7,%r13 |
| xorq %rcx,%r11 |
| subq %r10,%rbx |
| subq %r13,%rdx |
| leaq (%r9,%r9,1),%r10 |
| leaq (%r12,%r12,1),%r13 |
| xorq %rax,%r9 |
| xorq %rcx,%r12 |
| andq %rdi,%r10 |
| andq %rdi,%r13 |
| andq %rbp,%rbx |
| andq %rbp,%rdx |
| xorq %rbx,%r10 |
| xorq %rdx,%r13 |
| |
| xorq %r10,%rax |
| xorq %r13,%rcx |
| xorq %r10,%r8 |
| xorq %r13,%r11 |
| movq %rax,%rbx |
| movq %rcx,%rdx |
| xorq %r10,%r9 |
| shrq $32,%rbx |
| xorq %r13,%r12 |
| shrq $32,%rdx |
| xorq %r8,%r10 |
| roll $8,%eax |
| xorq %r11,%r13 |
| roll $8,%ecx |
| xorq %r9,%r10 |
| roll $8,%ebx |
| xorq %r12,%r13 |
| |
| roll $8,%edx |
| xorl %r10d,%eax |
| shrq $32,%r10 |
| xorl %r13d,%ecx |
| shrq $32,%r13 |
| xorl %r10d,%ebx |
| xorl %r13d,%edx |
| |
| movq %r8,%r10 |
| roll $24,%r8d |
| movq %r11,%r13 |
| roll $24,%r11d |
| shrq $32,%r10 |
| xorl %r8d,%eax |
| shrq $32,%r13 |
| xorl %r11d,%ecx |
| roll $24,%r10d |
| movq %r9,%r8 |
| roll $24,%r13d |
| movq %r12,%r11 |
| shrq $32,%r8 |
| xorl %r10d,%ebx |
| shrq $32,%r11 |
| xorl %r13d,%edx |
| |
| |
| roll $16,%r9d |
| |
| roll $16,%r12d |
| |
| roll $16,%r8d |
| |
| xorl %r9d,%eax |
| roll $16,%r11d |
| xorl %r12d,%ecx |
| |
| xorl %r8d,%ebx |
| xorl %r11d,%edx |
| movl %eax,0(%r15) |
| movl %ebx,4(%r15) |
| movl %ecx,8(%r15) |
| movl %edx,12(%r15) |
| subl $1,%r14d |
| jnz .Lpermute |
| |
| xorq %rax,%rax |
| .Labort: |
| movq 8(%rsp),%r15 |
| .cfi_restore %r15 |
| movq 16(%rsp),%r14 |
| .cfi_restore %r14 |
| movq 24(%rsp),%r13 |
| .cfi_restore %r13 |
| movq 32(%rsp),%r12 |
| .cfi_restore %r12 |
| movq 40(%rsp),%rbp |
| .cfi_restore %rbp |
| movq 48(%rsp),%rbx |
| .cfi_restore %rbx |
| addq $56,%rsp |
| .cfi_adjust_cfa_offset -56 |
| .Ldec_key_epilogue: |
| .byte 0xf3,0xc3 |
| .cfi_endproc |
| .size aes_nohw_set_decrypt_key,.-aes_nohw_set_decrypt_key |
| .align 16 |
| .globl aes_nohw_cbc_encrypt |
| .hidden aes_nohw_cbc_encrypt |
| .type aes_nohw_cbc_encrypt,@function |
| .extern OPENSSL_ia32cap_P |
| .hidden OPENSSL_ia32cap_P |
| .hidden aes_nohw_cbc_encrypt |
| aes_nohw_cbc_encrypt: |
| .cfi_startproc |
| cmpq $0,%rdx |
| je .Lcbc_epilogue |
| pushfq |
| .cfi_adjust_cfa_offset 8 |
| .cfi_offset 49,-16 |
| pushq %rbx |
| .cfi_adjust_cfa_offset 8 |
| .cfi_offset %rbx,-24 |
| pushq %rbp |
| .cfi_adjust_cfa_offset 8 |
| .cfi_offset %rbp,-32 |
| pushq %r12 |
| .cfi_adjust_cfa_offset 8 |
| .cfi_offset %r12,-40 |
| pushq %r13 |
| .cfi_adjust_cfa_offset 8 |
| .cfi_offset %r13,-48 |
| pushq %r14 |
| .cfi_adjust_cfa_offset 8 |
| .cfi_offset %r14,-56 |
| pushq %r15 |
| .cfi_adjust_cfa_offset 8 |
| .cfi_offset %r15,-64 |
| .Lcbc_prologue: |
| |
| cld |
| movl %r9d,%r9d |
| |
| leaq .LAES_Te(%rip),%r14 |
| leaq .LAES_Td(%rip),%r10 |
| cmpq $0,%r9 |
| cmoveq %r10,%r14 |
| |
| leaq OPENSSL_ia32cap_P(%rip),%r10 |
| movl (%r10),%r10d |
| cmpq $512,%rdx |
| jb .Lcbc_slow_prologue |
| testq $15,%rdx |
| jnz .Lcbc_slow_prologue |
| btl $28,%r10d |
| jc .Lcbc_slow_prologue |
| |
| |
| leaq -88-248(%rsp),%r15 |
| andq $-64,%r15 |
| |
| |
| movq %r14,%r10 |
| leaq 2304(%r14),%r11 |
| movq %r15,%r12 |
| andq $0xFFF,%r10 |
| andq $0xFFF,%r11 |
| andq $0xFFF,%r12 |
| |
| cmpq %r11,%r12 |
| jb .Lcbc_te_break_out |
| subq %r11,%r12 |
| subq %r12,%r15 |
| jmp .Lcbc_te_ok |
| .Lcbc_te_break_out: |
| subq %r10,%r12 |
| andq $0xFFF,%r12 |
| addq $320,%r12 |
| subq %r12,%r15 |
| .align 4 |
| .Lcbc_te_ok: |
| |
| xchgq %rsp,%r15 |
| .cfi_def_cfa_register %r15 |
| |
| movq %r15,16(%rsp) |
| .cfi_escape 0x0f,0x05,0x77,0x10,0x06,0x23,0x40 |
| .Lcbc_fast_body: |
| movq %rdi,24(%rsp) |
| movq %rsi,32(%rsp) |
| movq %rdx,40(%rsp) |
| movq %rcx,48(%rsp) |
| movq %r8,56(%rsp) |
| movl $0,80+240(%rsp) |
| movq %r8,%rbp |
| movq %r9,%rbx |
| movq %rsi,%r9 |
| movq %rdi,%r8 |
| movq %rcx,%r15 |
| |
| movl 240(%r15),%eax |
| |
| movq %r15,%r10 |
| subq %r14,%r10 |
| andq $0xfff,%r10 |
| cmpq $2304,%r10 |
| jb .Lcbc_do_ecopy |
| cmpq $4096-248,%r10 |
| jb .Lcbc_skip_ecopy |
| .align 4 |
| .Lcbc_do_ecopy: |
| movq %r15,%rsi |
| leaq 80(%rsp),%rdi |
| leaq 80(%rsp),%r15 |
| movl $30,%ecx |
| .long 0x90A548F3 |
| movl %eax,(%rdi) |
| .Lcbc_skip_ecopy: |
| movq %r15,0(%rsp) |
| |
| movl $18,%ecx |
| .align 4 |
| .Lcbc_prefetch_te: |
| movq 0(%r14),%r10 |
| movq 32(%r14),%r11 |
| movq 64(%r14),%r12 |
| movq 96(%r14),%r13 |
| leaq 128(%r14),%r14 |
| subl $1,%ecx |
| jnz .Lcbc_prefetch_te |
| leaq -2304(%r14),%r14 |
| |
| cmpq $0,%rbx |
| je .LFAST_DECRYPT |
| |
| |
| movl 0(%rbp),%eax |
| movl 4(%rbp),%ebx |
| movl 8(%rbp),%ecx |
| movl 12(%rbp),%edx |
| |
| .align 4 |
| .Lcbc_fast_enc_loop: |
| xorl 0(%r8),%eax |
| xorl 4(%r8),%ebx |
| xorl 8(%r8),%ecx |
| xorl 12(%r8),%edx |
| movq 0(%rsp),%r15 |
| movq %r8,24(%rsp) |
| |
| call _x86_64_AES_encrypt |
| |
| movq 24(%rsp),%r8 |
| movq 40(%rsp),%r10 |
| movl %eax,0(%r9) |
| movl %ebx,4(%r9) |
| movl %ecx,8(%r9) |
| movl %edx,12(%r9) |
| |
| leaq 16(%r8),%r8 |
| leaq 16(%r9),%r9 |
| subq $16,%r10 |
| testq $-16,%r10 |
| movq %r10,40(%rsp) |
| jnz .Lcbc_fast_enc_loop |
| movq 56(%rsp),%rbp |
| movl %eax,0(%rbp) |
| movl %ebx,4(%rbp) |
| movl %ecx,8(%rbp) |
| movl %edx,12(%rbp) |
| |
| jmp .Lcbc_fast_cleanup |
| |
| |
| .align 16 |
| .LFAST_DECRYPT: |
| cmpq %r8,%r9 |
| je .Lcbc_fast_dec_in_place |
| |
| movq %rbp,64(%rsp) |
| .align 4 |
| .Lcbc_fast_dec_loop: |
| movl 0(%r8),%eax |
| movl 4(%r8),%ebx |
| movl 8(%r8),%ecx |
| movl 12(%r8),%edx |
| movq 0(%rsp),%r15 |
| movq %r8,24(%rsp) |
| |
| call _x86_64_AES_decrypt |
| |
| movq 64(%rsp),%rbp |
| movq 24(%rsp),%r8 |
| movq 40(%rsp),%r10 |
| xorl 0(%rbp),%eax |
| xorl 4(%rbp),%ebx |
| xorl 8(%rbp),%ecx |
| xorl 12(%rbp),%edx |
| movq %r8,%rbp |
| |
| subq $16,%r10 |
| movq %r10,40(%rsp) |
| movq %rbp,64(%rsp) |
| |
| movl %eax,0(%r9) |
| movl %ebx,4(%r9) |
| movl %ecx,8(%r9) |
| movl %edx,12(%r9) |
| |
| leaq 16(%r8),%r8 |
| leaq 16(%r9),%r9 |
| jnz .Lcbc_fast_dec_loop |
| movq 56(%rsp),%r12 |
| movq 0(%rbp),%r10 |
| movq 8(%rbp),%r11 |
| movq %r10,0(%r12) |
| movq %r11,8(%r12) |
| jmp .Lcbc_fast_cleanup |
| |
| .align 16 |
| .Lcbc_fast_dec_in_place: |
| movq 0(%rbp),%r10 |
| movq 8(%rbp),%r11 |
| movq %r10,0+64(%rsp) |
| movq %r11,8+64(%rsp) |
| .align 4 |
| .Lcbc_fast_dec_in_place_loop: |
| movl 0(%r8),%eax |
| movl 4(%r8),%ebx |
| movl 8(%r8),%ecx |
| movl 12(%r8),%edx |
| movq 0(%rsp),%r15 |
| movq %r8,24(%rsp) |
| |
| call _x86_64_AES_decrypt |
| |
| movq 24(%rsp),%r8 |
| movq 40(%rsp),%r10 |
| xorl 0+64(%rsp),%eax |
| xorl 4+64(%rsp),%ebx |
| xorl 8+64(%rsp),%ecx |
| xorl 12+64(%rsp),%edx |
| |
| movq 0(%r8),%r11 |
| movq 8(%r8),%r12 |
| subq $16,%r10 |
| jz .Lcbc_fast_dec_in_place_done |
| |
| movq %r11,0+64(%rsp) |
| movq %r12,8+64(%rsp) |
| |
| movl %eax,0(%r9) |
| movl %ebx,4(%r9) |
| movl %ecx,8(%r9) |
| movl %edx,12(%r9) |
| |
| leaq 16(%r8),%r8 |
| leaq 16(%r9),%r9 |
| movq %r10,40(%rsp) |
| jmp .Lcbc_fast_dec_in_place_loop |
| .Lcbc_fast_dec_in_place_done: |
| movq 56(%rsp),%rdi |
| movq %r11,0(%rdi) |
| movq %r12,8(%rdi) |
| |
| movl %eax,0(%r9) |
| movl %ebx,4(%r9) |
| movl %ecx,8(%r9) |
| movl %edx,12(%r9) |
| |
| .align 4 |
| .Lcbc_fast_cleanup: |
| cmpl $0,80+240(%rsp) |
| leaq 80(%rsp),%rdi |
| je .Lcbc_exit |
| movl $30,%ecx |
| xorq %rax,%rax |
| .long 0x90AB48F3 |
| |
| jmp .Lcbc_exit |
| |
| |
| .align 16 |
| .Lcbc_slow_prologue: |
| |
| leaq -88(%rsp),%rbp |
| andq $-64,%rbp |
| |
| leaq -88-63(%rcx),%r10 |
| subq %rbp,%r10 |
| negq %r10 |
| andq $0x3c0,%r10 |
| subq %r10,%rbp |
| |
| xchgq %rsp,%rbp |
| |
| movq %rbp,16(%rsp) |
| .Lcbc_slow_body: |
| |
| |
| |
| |
| movq %r8,56(%rsp) |
| movq %r8,%rbp |
| movq %r9,%rbx |
| movq %rsi,%r9 |
| movq %rdi,%r8 |
| movq %rcx,%r15 |
| movq %rdx,%r10 |
| |
| movl 240(%r15),%eax |
| movq %r15,0(%rsp) |
| shll $4,%eax |
| leaq (%r15,%rax,1),%rax |
| movq %rax,8(%rsp) |
| |
| |
| leaq 2048(%r14),%r14 |
| leaq 768-8(%rsp),%rax |
| subq %r14,%rax |
| andq $0x300,%rax |
| leaq (%r14,%rax,1),%r14 |
| |
| cmpq $0,%rbx |
| je .LSLOW_DECRYPT |
| |
| |
| testq $-16,%r10 |
| movl 0(%rbp),%eax |
| movl 4(%rbp),%ebx |
| movl 8(%rbp),%ecx |
| movl 12(%rbp),%edx |
| jz .Lcbc_slow_enc_tail |
| |
| .align 4 |
| .Lcbc_slow_enc_loop: |
| xorl 0(%r8),%eax |
| xorl 4(%r8),%ebx |
| xorl 8(%r8),%ecx |
| xorl 12(%r8),%edx |
| movq 0(%rsp),%r15 |
| movq %r8,24(%rsp) |
| movq %r9,32(%rsp) |
| movq %r10,40(%rsp) |
| |
| call _x86_64_AES_encrypt_compact |
| |
| movq 24(%rsp),%r8 |
| movq 32(%rsp),%r9 |
| movq 40(%rsp),%r10 |
| movl %eax,0(%r9) |
| movl %ebx,4(%r9) |
| movl %ecx,8(%r9) |
| movl %edx,12(%r9) |
| |
| leaq 16(%r8),%r8 |
| leaq 16(%r9),%r9 |
| subq $16,%r10 |
| testq $-16,%r10 |
| jnz .Lcbc_slow_enc_loop |
| testq $15,%r10 |
| jnz .Lcbc_slow_enc_tail |
| movq 56(%rsp),%rbp |
| movl %eax,0(%rbp) |
| movl %ebx,4(%rbp) |
| movl %ecx,8(%rbp) |
| movl %edx,12(%rbp) |
| |
| jmp .Lcbc_exit |
| |
| .align 4 |
| .Lcbc_slow_enc_tail: |
| movq %rax,%r11 |
| movq %rcx,%r12 |
| movq %r10,%rcx |
| movq %r8,%rsi |
| movq %r9,%rdi |
| .long 0x9066A4F3 |
| movq $16,%rcx |
| subq %r10,%rcx |
| xorq %rax,%rax |
| .long 0x9066AAF3 |
| movq %r9,%r8 |
| movq $16,%r10 |
| movq %r11,%rax |
| movq %r12,%rcx |
| jmp .Lcbc_slow_enc_loop |
| |
| .align 16 |
| .LSLOW_DECRYPT: |
| shrq $3,%rax |
| addq %rax,%r14 |
| |
| movq 0(%rbp),%r11 |
| movq 8(%rbp),%r12 |
| movq %r11,0+64(%rsp) |
| movq %r12,8+64(%rsp) |
| |
| .align 4 |
| .Lcbc_slow_dec_loop: |
| movl 0(%r8),%eax |
| movl 4(%r8),%ebx |
| movl 8(%r8),%ecx |
| movl 12(%r8),%edx |
| movq 0(%rsp),%r15 |
| movq %r8,24(%rsp) |
| movq %r9,32(%rsp) |
| movq %r10,40(%rsp) |
| |
| call _x86_64_AES_decrypt_compact |
| |
| movq 24(%rsp),%r8 |
| movq 32(%rsp),%r9 |
| movq 40(%rsp),%r10 |
| xorl 0+64(%rsp),%eax |
| xorl 4+64(%rsp),%ebx |
| xorl 8+64(%rsp),%ecx |
| xorl 12+64(%rsp),%edx |
| |
| movq 0(%r8),%r11 |
| movq 8(%r8),%r12 |
| subq $16,%r10 |
| jc .Lcbc_slow_dec_partial |
| jz .Lcbc_slow_dec_done |
| |
| movq %r11,0+64(%rsp) |
| movq %r12,8+64(%rsp) |
| |
| movl %eax,0(%r9) |
| movl %ebx,4(%r9) |
| movl %ecx,8(%r9) |
| movl %edx,12(%r9) |
| |
| leaq 16(%r8),%r8 |
| leaq 16(%r9),%r9 |
| jmp .Lcbc_slow_dec_loop |
| .Lcbc_slow_dec_done: |
| movq 56(%rsp),%rdi |
| movq %r11,0(%rdi) |
| movq %r12,8(%rdi) |
| |
| movl %eax,0(%r9) |
| movl %ebx,4(%r9) |
| movl %ecx,8(%r9) |
| movl %edx,12(%r9) |
| |
| jmp .Lcbc_exit |
| |
| .align 4 |
| .Lcbc_slow_dec_partial: |
| movq 56(%rsp),%rdi |
| movq %r11,0(%rdi) |
| movq %r12,8(%rdi) |
| |
| movl %eax,0+64(%rsp) |
| movl %ebx,4+64(%rsp) |
| movl %ecx,8+64(%rsp) |
| movl %edx,12+64(%rsp) |
| |
| movq %r9,%rdi |
| leaq 64(%rsp),%rsi |
| leaq 16(%r10),%rcx |
| .long 0x9066A4F3 |
| jmp .Lcbc_exit |
| |
| .align 16 |
| .Lcbc_exit: |
| movq 16(%rsp),%rsi |
| .cfi_def_cfa %rsi,64 |
| movq (%rsi),%r15 |
| .cfi_restore %r15 |
| movq 8(%rsi),%r14 |
| .cfi_restore %r14 |
| movq 16(%rsi),%r13 |
| .cfi_restore %r13 |
| movq 24(%rsi),%r12 |
| .cfi_restore %r12 |
| movq 32(%rsi),%rbp |
| .cfi_restore %rbp |
| movq 40(%rsi),%rbx |
| .cfi_restore %rbx |
| leaq 48(%rsi),%rsp |
| .cfi_def_cfa %rsp,16 |
| .Lcbc_popfq: |
| popfq |
| .cfi_adjust_cfa_offset -8 |
| .cfi_restore 49 |
| .Lcbc_epilogue: |
| .byte 0xf3,0xc3 |
| .cfi_endproc |
| .size aes_nohw_cbc_encrypt,.-aes_nohw_cbc_encrypt |
| .align 64 |
| .LAES_Te: |
| .long 0xa56363c6,0xa56363c6 |
| .long 0x847c7cf8,0x847c7cf8 |
| .long 0x997777ee,0x997777ee |
| .long 0x8d7b7bf6,0x8d7b7bf6 |
| .long 0x0df2f2ff,0x0df2f2ff |
| .long 0xbd6b6bd6,0xbd6b6bd6 |
| .long 0xb16f6fde,0xb16f6fde |
| .long 0x54c5c591,0x54c5c591 |
| .long 0x50303060,0x50303060 |
| .long 0x03010102,0x03010102 |
| .long 0xa96767ce,0xa96767ce |
| .long 0x7d2b2b56,0x7d2b2b56 |
| .long 0x19fefee7,0x19fefee7 |
| .long 0x62d7d7b5,0x62d7d7b5 |
| .long 0xe6abab4d,0xe6abab4d |
| .long 0x9a7676ec,0x9a7676ec |
| .long 0x45caca8f,0x45caca8f |
| .long 0x9d82821f,0x9d82821f |
| .long 0x40c9c989,0x40c9c989 |
| .long 0x877d7dfa,0x877d7dfa |
| .long 0x15fafaef,0x15fafaef |
| .long 0xeb5959b2,0xeb5959b2 |
| .long 0xc947478e,0xc947478e |
| .long 0x0bf0f0fb,0x0bf0f0fb |
| .long 0xecadad41,0xecadad41 |
| .long 0x67d4d4b3,0x67d4d4b3 |
| .long 0xfda2a25f,0xfda2a25f |
| .long 0xeaafaf45,0xeaafaf45 |
| .long 0xbf9c9c23,0xbf9c9c23 |
| .long 0xf7a4a453,0xf7a4a453 |
| .long 0x967272e4,0x967272e4 |
| .long 0x5bc0c09b,0x5bc0c09b |
| .long 0xc2b7b775,0xc2b7b775 |
| .long 0x1cfdfde1,0x1cfdfde1 |
| .long 0xae93933d,0xae93933d |
| .long 0x6a26264c,0x6a26264c |
| .long 0x5a36366c,0x5a36366c |
| .long 0x413f3f7e,0x413f3f7e |
| .long 0x02f7f7f5,0x02f7f7f5 |
| .long 0x4fcccc83,0x4fcccc83 |
| .long 0x5c343468,0x5c343468 |
| .long 0xf4a5a551,0xf4a5a551 |
| .long 0x34e5e5d1,0x34e5e5d1 |
| .long 0x08f1f1f9,0x08f1f1f9 |
| .long 0x937171e2,0x937171e2 |
| .long 0x73d8d8ab,0x73d8d8ab |
| .long 0x53313162,0x53313162 |
| .long 0x3f15152a,0x3f15152a |
| .long 0x0c040408,0x0c040408 |
| .long 0x52c7c795,0x52c7c795 |
| .long 0x65232346,0x65232346 |
| .long 0x5ec3c39d,0x5ec3c39d |
| .long 0x28181830,0x28181830 |
| .long 0xa1969637,0xa1969637 |
| .long 0x0f05050a,0x0f05050a |
| .long 0xb59a9a2f,0xb59a9a2f |
| .long 0x0907070e,0x0907070e |
| .long 0x36121224,0x36121224 |
| .long 0x9b80801b,0x9b80801b |
| .long 0x3de2e2df,0x3de2e2df |
| .long 0x26ebebcd,0x26ebebcd |
| .long 0x6927274e,0x6927274e |
| .long 0xcdb2b27f,0xcdb2b27f |
| .long 0x9f7575ea,0x9f7575ea |
| .long 0x1b090912,0x1b090912 |
| .long 0x9e83831d,0x9e83831d |
| .long 0x742c2c58,0x742c2c58 |
| .long 0x2e1a1a34,0x2e1a1a34 |
| .long 0x2d1b1b36,0x2d1b1b36 |
| .long 0xb26e6edc,0xb26e6edc |
| .long 0xee5a5ab4,0xee5a5ab4 |
| .long 0xfba0a05b,0xfba0a05b |
| .long 0xf65252a4,0xf65252a4 |
| .long 0x4d3b3b76,0x4d3b3b76 |
| .long 0x61d6d6b7,0x61d6d6b7 |
| .long 0xceb3b37d,0xceb3b37d |
| .long 0x7b292952,0x7b292952 |
| .long 0x3ee3e3dd,0x3ee3e3dd |
| .long 0x712f2f5e,0x712f2f5e |
| .long 0x97848413,0x97848413 |
| .long 0xf55353a6,0xf55353a6 |
| .long 0x68d1d1b9,0x68d1d1b9 |
| .long 0x00000000,0x00000000 |
| .long 0x2cededc1,0x2cededc1 |
| .long 0x60202040,0x60202040 |
| .long 0x1ffcfce3,0x1ffcfce3 |
| .long 0xc8b1b179,0xc8b1b179 |
| .long 0xed5b5bb6,0xed5b5bb6 |
| .long 0xbe6a6ad4,0xbe6a6ad4 |
| .long 0x46cbcb8d,0x46cbcb8d |
| .long 0xd9bebe67,0xd9bebe67 |
| .long 0x4b393972,0x4b393972 |
| .long 0xde4a4a94,0xde4a4a94 |
| .long 0xd44c4c98,0xd44c4c98 |
| .long 0xe85858b0,0xe85858b0 |
| .long 0x4acfcf85,0x4acfcf85 |
| .long 0x6bd0d0bb,0x6bd0d0bb |
| .long 0x2aefefc5,0x2aefefc5 |
| .long 0xe5aaaa4f,0xe5aaaa4f |
| .long 0x16fbfbed,0x16fbfbed |
| .long 0xc5434386,0xc5434386 |
| .long 0xd74d4d9a,0xd74d4d9a |
| .long 0x55333366,0x55333366 |
| .long 0x94858511,0x94858511 |
| .long 0xcf45458a,0xcf45458a |
| .long 0x10f9f9e9,0x10f9f9e9 |
| .long 0x06020204,0x06020204 |
| .long 0x817f7ffe,0x817f7ffe |
| .long 0xf05050a0,0xf05050a0 |
| .long 0x443c3c78,0x443c3c78 |
| .long 0xba9f9f25,0xba9f9f25 |
| .long 0xe3a8a84b,0xe3a8a84b |
| .long 0xf35151a2,0xf35151a2 |
| .long 0xfea3a35d,0xfea3a35d |
| .long 0xc0404080,0xc0404080 |
| .long 0x8a8f8f05,0x8a8f8f05 |
| .long 0xad92923f,0xad92923f |
| .long 0xbc9d9d21,0xbc9d9d21 |
| .long 0x48383870,0x48383870 |
| .long 0x04f5f5f1,0x04f5f5f1 |
| .long 0xdfbcbc63,0xdfbcbc63 |
| .long 0xc1b6b677,0xc1b6b677 |
| .long 0x75dadaaf,0x75dadaaf |
| .long 0x63212142,0x63212142 |
| .long 0x30101020,0x30101020 |
| .long 0x1affffe5,0x1affffe5 |
| .long 0x0ef3f3fd,0x0ef3f3fd |
| .long 0x6dd2d2bf,0x6dd2d2bf |
| .long 0x4ccdcd81,0x4ccdcd81 |
| .long 0x140c0c18,0x140c0c18 |
| .long 0x35131326,0x35131326 |
| .long 0x2fececc3,0x2fececc3 |
| .long 0xe15f5fbe,0xe15f5fbe |
| .long 0xa2979735,0xa2979735 |
| .long 0xcc444488,0xcc444488 |
| .long 0x3917172e,0x3917172e |
| .long 0x57c4c493,0x57c4c493 |
| .long 0xf2a7a755,0xf2a7a755 |
| .long 0x827e7efc,0x827e7efc |
| .long 0x473d3d7a,0x473d3d7a |
| .long 0xac6464c8,0xac6464c8 |
| .long 0xe75d5dba,0xe75d5dba |
| .long 0x2b191932,0x2b191932 |
| .long 0x957373e6,0x957373e6 |
| .long 0xa06060c0,0xa06060c0 |
| .long 0x98818119,0x98818119 |
| .long 0xd14f4f9e,0xd14f4f9e |
| .long 0x7fdcdca3,0x7fdcdca3 |
| .long 0x66222244,0x66222244 |
| .long 0x7e2a2a54,0x7e2a2a54 |
| .long 0xab90903b,0xab90903b |
| .long 0x8388880b,0x8388880b |
| .long 0xca46468c,0xca46468c |
| .long 0x29eeeec7,0x29eeeec7 |
| .long 0xd3b8b86b,0xd3b8b86b |
| .long 0x3c141428,0x3c141428 |
| .long 0x79dedea7,0x79dedea7 |
| .long 0xe25e5ebc,0xe25e5ebc |
| .long 0x1d0b0b16,0x1d0b0b16 |
| .long 0x76dbdbad,0x76dbdbad |
| .long 0x3be0e0db,0x3be0e0db |
| .long 0x56323264,0x56323264 |
| .long 0x4e3a3a74,0x4e3a3a74 |
| .long 0x1e0a0a14,0x1e0a0a14 |
| .long 0xdb494992,0xdb494992 |
| .long 0x0a06060c,0x0a06060c |
| .long 0x6c242448,0x6c242448 |
| .long 0xe45c5cb8,0xe45c5cb8 |
| .long 0x5dc2c29f,0x5dc2c29f |
| .long 0x6ed3d3bd,0x6ed3d3bd |
| .long 0xefacac43,0xefacac43 |
| .long 0xa66262c4,0xa66262c4 |
| .long 0xa8919139,0xa8919139 |
| .long 0xa4959531,0xa4959531 |
| .long 0x37e4e4d3,0x37e4e4d3 |
| .long 0x8b7979f2,0x8b7979f2 |
| .long 0x32e7e7d5,0x32e7e7d5 |
| .long 0x43c8c88b,0x43c8c88b |
| .long 0x5937376e,0x5937376e |
| .long 0xb76d6dda,0xb76d6dda |
| .long 0x8c8d8d01,0x8c8d8d01 |
| .long 0x64d5d5b1,0x64d5d5b1 |
| .long 0xd24e4e9c,0xd24e4e9c |
| .long 0xe0a9a949,0xe0a9a949 |
| .long 0xb46c6cd8,0xb46c6cd8 |
| .long 0xfa5656ac,0xfa5656ac |
| .long 0x07f4f4f3,0x07f4f4f3 |
| .long 0x25eaeacf,0x25eaeacf |
| .long 0xaf6565ca,0xaf6565ca |
| .long 0x8e7a7af4,0x8e7a7af4 |
| .long 0xe9aeae47,0xe9aeae47 |
| .long 0x18080810,0x18080810 |
| .long 0xd5baba6f,0xd5baba6f |
| .long 0x887878f0,0x887878f0 |
| .long 0x6f25254a,0x6f25254a |
| .long 0x722e2e5c,0x722e2e5c |
| .long 0x241c1c38,0x241c1c38 |
| .long 0xf1a6a657,0xf1a6a657 |
| .long 0xc7b4b473,0xc7b4b473 |
| .long 0x51c6c697,0x51c6c697 |
| .long 0x23e8e8cb,0x23e8e8cb |
| .long 0x7cdddda1,0x7cdddda1 |
| .long 0x9c7474e8,0x9c7474e8 |
| .long 0x211f1f3e,0x211f1f3e |
| .long 0xdd4b4b96,0xdd4b4b96 |
| .long 0xdcbdbd61,0xdcbdbd61 |
| .long 0x868b8b0d,0x868b8b0d |
| .long 0x858a8a0f,0x858a8a0f |
| .long 0x907070e0,0x907070e0 |
| .long 0x423e3e7c,0x423e3e7c |
| .long 0xc4b5b571,0xc4b5b571 |
| .long 0xaa6666cc,0xaa6666cc |
| .long 0xd8484890,0xd8484890 |
| .long 0x05030306,0x05030306 |
| .long 0x01f6f6f7,0x01f6f6f7 |
| .long 0x120e0e1c,0x120e0e1c |
| .long 0xa36161c2,0xa36161c2 |
| .long 0x5f35356a,0x5f35356a |
| .long 0xf95757ae,0xf95757ae |
| .long 0xd0b9b969,0xd0b9b969 |
| .long 0x91868617,0x91868617 |
| .long 0x58c1c199,0x58c1c199 |
| .long 0x271d1d3a,0x271d1d3a |
| .long 0xb99e9e27,0xb99e9e27 |
| .long 0x38e1e1d9,0x38e1e1d9 |
| .long 0x13f8f8eb,0x13f8f8eb |
| .long 0xb398982b,0xb398982b |
| .long 0x33111122,0x33111122 |
| .long 0xbb6969d2,0xbb6969d2 |
| .long 0x70d9d9a9,0x70d9d9a9 |
| .long 0x898e8e07,0x898e8e07 |
| .long 0xa7949433,0xa7949433 |
| .long 0xb69b9b2d,0xb69b9b2d |
| .long 0x221e1e3c,0x221e1e3c |
| .long 0x92878715,0x92878715 |
| .long 0x20e9e9c9,0x20e9e9c9 |
| .long 0x49cece87,0x49cece87 |
| .long 0xff5555aa,0xff5555aa |
| .long 0x78282850,0x78282850 |
| .long 0x7adfdfa5,0x7adfdfa5 |
| .long 0x8f8c8c03,0x8f8c8c03 |
| .long 0xf8a1a159,0xf8a1a159 |
| .long 0x80898909,0x80898909 |
| .long 0x170d0d1a,0x170d0d1a |
| .long 0xdabfbf65,0xdabfbf65 |
| .long 0x31e6e6d7,0x31e6e6d7 |
| .long 0xc6424284,0xc6424284 |
| .long 0xb86868d0,0xb86868d0 |
| .long 0xc3414182,0xc3414182 |
| .long 0xb0999929,0xb0999929 |
| .long 0x772d2d5a,0x772d2d5a |
| .long 0x110f0f1e,0x110f0f1e |
| .long 0xcbb0b07b,0xcbb0b07b |
| .long 0xfc5454a8,0xfc5454a8 |
| .long 0xd6bbbb6d,0xd6bbbb6d |
| .long 0x3a16162c,0x3a16162c |
| .byte 0x63,0x7c,0x77,0x7b,0xf2,0x6b,0x6f,0xc5 |
| .byte 0x30,0x01,0x67,0x2b,0xfe,0xd7,0xab,0x76 |
| .byte 0xca,0x82,0xc9,0x7d,0xfa,0x59,0x47,0xf0 |
| .byte 0xad,0xd4,0xa2,0xaf,0x9c,0xa4,0x72,0xc0 |
| .byte 0xb7,0xfd,0x93,0x26,0x36,0x3f,0xf7,0xcc |
| .byte 0x34,0xa5,0xe5,0xf1,0x71,0xd8,0x31,0x15 |
| .byte 0x04,0xc7,0x23,0xc3,0x18,0x96,0x05,0x9a |
| .byte 0x07,0x12,0x80,0xe2,0xeb,0x27,0xb2,0x75 |
| .byte 0x09,0x83,0x2c,0x1a,0x1b,0x6e,0x5a,0xa0 |
| .byte 0x52,0x3b,0xd6,0xb3,0x29,0xe3,0x2f,0x84 |
| .byte 0x53,0xd1,0x00,0xed,0x20,0xfc,0xb1,0x5b |
| .byte 0x6a,0xcb,0xbe,0x39,0x4a,0x4c,0x58,0xcf |
| .byte 0xd0,0xef,0xaa,0xfb,0x43,0x4d,0x33,0x85 |
| .byte 0x45,0xf9,0x02,0x7f,0x50,0x3c,0x9f,0xa8 |
| .byte 0x51,0xa3,0x40,0x8f,0x92,0x9d,0x38,0xf5 |
| .byte 0xbc,0xb6,0xda,0x21,0x10,0xff,0xf3,0xd2 |
| .byte 0xcd,0x0c,0x13,0xec,0x5f,0x97,0x44,0x17 |
| .byte 0xc4,0xa7,0x7e,0x3d,0x64,0x5d,0x19,0x73 |
| .byte 0x60,0x81,0x4f,0xdc,0x22,0x2a,0x90,0x88 |
| .byte 0x46,0xee,0xb8,0x14,0xde,0x5e,0x0b,0xdb |
| .byte 0xe0,0x32,0x3a,0x0a,0x49,0x06,0x24,0x5c |
| .byte 0xc2,0xd3,0xac,0x62,0x91,0x95,0xe4,0x79 |
| .byte 0xe7,0xc8,0x37,0x6d,0x8d,0xd5,0x4e,0xa9 |
| .byte 0x6c,0x56,0xf4,0xea,0x65,0x7a,0xae,0x08 |
| .byte 0xba,0x78,0x25,0x2e,0x1c,0xa6,0xb4,0xc6 |
| .byte 0xe8,0xdd,0x74,0x1f,0x4b,0xbd,0x8b,0x8a |
| .byte 0x70,0x3e,0xb5,0x66,0x48,0x03,0xf6,0x0e |
| .byte 0x61,0x35,0x57,0xb9,0x86,0xc1,0x1d,0x9e |
| .byte 0xe1,0xf8,0x98,0x11,0x69,0xd9,0x8e,0x94 |
| .byte 0x9b,0x1e,0x87,0xe9,0xce,0x55,0x28,0xdf |
| .byte 0x8c,0xa1,0x89,0x0d,0xbf,0xe6,0x42,0x68 |
| .byte 0x41,0x99,0x2d,0x0f,0xb0,0x54,0xbb,0x16 |
| .byte 0x63,0x7c,0x77,0x7b,0xf2,0x6b,0x6f,0xc5 |
| .byte 0x30,0x01,0x67,0x2b,0xfe,0xd7,0xab,0x76 |
| .byte 0xca,0x82,0xc9,0x7d,0xfa,0x59,0x47,0xf0 |
| .byte 0xad,0xd4,0xa2,0xaf,0x9c,0xa4,0x72,0xc0 |
| .byte 0xb7,0xfd,0x93,0x26,0x36,0x3f,0xf7,0xcc |
| .byte 0x34,0xa5,0xe5,0xf1,0x71,0xd8,0x31,0x15 |
| .byte 0x04,0xc7,0x23,0xc3,0x18,0x96,0x05,0x9a |
| .byte 0x07,0x12,0x80,0xe2,0xeb,0x27,0xb2,0x75 |
| .byte 0x09,0x83,0x2c,0x1a,0x1b,0x6e,0x5a,0xa0 |
| .byte 0x52,0x3b,0xd6,0xb3,0x29,0xe3,0x2f,0x84 |
| .byte 0x53,0xd1,0x00,0xed,0x20,0xfc,0xb1,0x5b |
| .byte 0x6a,0xcb,0xbe,0x39,0x4a,0x4c,0x58,0xcf |
| .byte 0xd0,0xef,0xaa,0xfb,0x43,0x4d,0x33,0x85 |
| .byte 0x45,0xf9,0x02,0x7f,0x50,0x3c,0x9f,0xa8 |
| .byte 0x51,0xa3,0x40,0x8f,0x92,0x9d,0x38,0xf5 |
| .byte 0xbc,0xb6,0xda,0x21,0x10,0xff,0xf3,0xd2 |
| .byte 0xcd,0x0c,0x13,0xec,0x5f,0x97,0x44,0x17 |
| .byte 0xc4,0xa7,0x7e,0x3d,0x64,0x5d,0x19,0x73 |
| .byte 0x60,0x81,0x4f,0xdc,0x22,0x2a,0x90,0x88 |
| .byte 0x46,0xee,0xb8,0x14,0xde,0x5e,0x0b,0xdb |
| .byte 0xe0,0x32,0x3a,0x0a,0x49,0x06,0x24,0x5c |
| .byte 0xc2,0xd3,0xac,0x62,0x91,0x95,0xe4,0x79 |
| .byte 0xe7,0xc8,0x37,0x6d,0x8d,0xd5,0x4e,0xa9 |
| .byte 0x6c,0x56,0xf4,0xea,0x65,0x7a,0xae,0x08 |
| .byte 0xba,0x78,0x25,0x2e,0x1c,0xa6,0xb4,0xc6 |
| .byte 0xe8,0xdd,0x74,0x1f,0x4b,0xbd,0x8b,0x8a |
| .byte 0x70,0x3e,0xb5,0x66,0x48,0x03,0xf6,0x0e |
| .byte 0x61,0x35,0x57,0xb9,0x86,0xc1,0x1d,0x9e |
| .byte 0xe1,0xf8,0x98,0x11,0x69,0xd9,0x8e,0x94 |
| .byte 0x9b,0x1e,0x87,0xe9,0xce,0x55,0x28,0xdf |
| .byte 0x8c,0xa1,0x89,0x0d,0xbf,0xe6,0x42,0x68 |
| .byte 0x41,0x99,0x2d,0x0f,0xb0,0x54,0xbb,0x16 |
| .byte 0x63,0x7c,0x77,0x7b,0xf2,0x6b,0x6f,0xc5 |
| .byte 0x30,0x01,0x67,0x2b,0xfe,0xd7,0xab,0x76 |
| .byte 0xca,0x82,0xc9,0x7d,0xfa,0x59,0x47,0xf0 |
| .byte 0xad,0xd4,0xa2,0xaf,0x9c,0xa4,0x72,0xc0 |
| .byte 0xb7,0xfd,0x93,0x26,0x36,0x3f,0xf7,0xcc |
| .byte 0x34,0xa5,0xe5,0xf1,0x71,0xd8,0x31,0x15 |
| .byte 0x04,0xc7,0x23,0xc3,0x18,0x96,0x05,0x9a |
| .byte 0x07,0x12,0x80,0xe2,0xeb,0x27,0xb2,0x75 |
| .byte 0x09,0x83,0x2c,0x1a,0x1b,0x6e,0x5a,0xa0 |
| .byte 0x52,0x3b,0xd6,0xb3,0x29,0xe3,0x2f,0x84 |
| .byte 0x53,0xd1,0x00,0xed,0x20,0xfc,0xb1,0x5b |
| .byte 0x6a,0xcb,0xbe,0x39,0x4a,0x4c,0x58,0xcf |
| .byte 0xd0,0xef,0xaa,0xfb,0x43,0x4d,0x33,0x85 |
| .byte 0x45,0xf9,0x02,0x7f,0x50,0x3c,0x9f,0xa8 |
| .byte 0x51,0xa3,0x40,0x8f,0x92,0x9d,0x38,0xf5 |
| .byte 0xbc,0xb6,0xda,0x21,0x10,0xff,0xf3,0xd2 |
| .byte 0xcd,0x0c,0x13,0xec,0x5f,0x97,0x44,0x17 |
| .byte 0xc4,0xa7,0x7e,0x3d,0x64,0x5d,0x19,0x73 |
| .byte 0x60,0x81,0x4f,0xdc,0x22,0x2a,0x90,0x88 |
| .byte 0x46,0xee,0xb8,0x14,0xde,0x5e,0x0b,0xdb |
| .byte 0xe0,0x32,0x3a,0x0a,0x49,0x06,0x24,0x5c |
| .byte 0xc2,0xd3,0xac,0x62,0x91,0x95,0xe4,0x79 |
| .byte 0xe7,0xc8,0x37,0x6d,0x8d,0xd5,0x4e,0xa9 |
| .byte 0x6c,0x56,0xf4,0xea,0x65,0x7a,0xae,0x08 |
| .byte 0xba,0x78,0x25,0x2e,0x1c,0xa6,0xb4,0xc6 |
| .byte 0xe8,0xdd,0x74,0x1f,0x4b,0xbd,0x8b,0x8a |
| .byte 0x70,0x3e,0xb5,0x66,0x48,0x03,0xf6,0x0e |
| .byte 0x61,0x35,0x57,0xb9,0x86,0xc1,0x1d,0x9e |
| .byte 0xe1,0xf8,0x98,0x11,0x69,0xd9,0x8e,0x94 |
| .byte 0x9b,0x1e,0x87,0xe9,0xce,0x55,0x28,0xdf |
| .byte 0x8c,0xa1,0x89,0x0d,0xbf,0xe6,0x42,0x68 |
| .byte 0x41,0x99,0x2d,0x0f,0xb0,0x54,0xbb,0x16 |
| .byte 0x63,0x7c,0x77,0x7b,0xf2,0x6b,0x6f,0xc5 |
| .byte 0x30,0x01,0x67,0x2b,0xfe,0xd7,0xab,0x76 |
| .byte 0xca,0x82,0xc9,0x7d,0xfa,0x59,0x47,0xf0 |
| .byte 0xad,0xd4,0xa2,0xaf,0x9c,0xa4,0x72,0xc0 |
| .byte 0xb7,0xfd,0x93,0x26,0x36,0x3f,0xf7,0xcc |
| .byte 0x34,0xa5,0xe5,0xf1,0x71,0xd8,0x31,0x15 |
| .byte 0x04,0xc7,0x23,0xc3,0x18,0x96,0x05,0x9a |
| .byte 0x07,0x12,0x80,0xe2,0xeb,0x27,0xb2,0x75 |
| .byte 0x09,0x83,0x2c,0x1a,0x1b,0x6e,0x5a,0xa0 |
| .byte 0x52,0x3b,0xd6,0xb3,0x29,0xe3,0x2f,0x84 |
| .byte 0x53,0xd1,0x00,0xed,0x20,0xfc,0xb1,0x5b |
| .byte 0x6a,0xcb,0xbe,0x39,0x4a,0x4c,0x58,0xcf |
| .byte 0xd0,0xef,0xaa,0xfb,0x43,0x4d,0x33,0x85 |
| .byte 0x45,0xf9,0x02,0x7f,0x50,0x3c,0x9f,0xa8 |
| .byte 0x51,0xa3,0x40,0x8f,0x92,0x9d,0x38,0xf5 |
| .byte 0xbc,0xb6,0xda,0x21,0x10,0xff,0xf3,0xd2 |
| .byte 0xcd,0x0c,0x13,0xec,0x5f,0x97,0x44,0x17 |
| .byte 0xc4,0xa7,0x7e,0x3d,0x64,0x5d,0x19,0x73 |
| .byte 0x60,0x81,0x4f,0xdc,0x22,0x2a,0x90,0x88 |
| .byte 0x46,0xee,0xb8,0x14,0xde,0x5e,0x0b,0xdb |
| .byte 0xe0,0x32,0x3a,0x0a,0x49,0x06,0x24,0x5c |
| .byte 0xc2,0xd3,0xac,0x62,0x91,0x95,0xe4,0x79 |
| .byte 0xe7,0xc8,0x37,0x6d,0x8d,0xd5,0x4e,0xa9 |
| .byte 0x6c,0x56,0xf4,0xea,0x65,0x7a,0xae,0x08 |
| .byte 0xba,0x78,0x25,0x2e,0x1c,0xa6,0xb4,0xc6 |
| .byte 0xe8,0xdd,0x74,0x1f,0x4b,0xbd,0x8b,0x8a |
| .byte 0x70,0x3e,0xb5,0x66,0x48,0x03,0xf6,0x0e |
| .byte 0x61,0x35,0x57,0xb9,0x86,0xc1,0x1d,0x9e |
| .byte 0xe1,0xf8,0x98,0x11,0x69,0xd9,0x8e,0x94 |
| .byte 0x9b,0x1e,0x87,0xe9,0xce,0x55,0x28,0xdf |
| .byte 0x8c,0xa1,0x89,0x0d,0xbf,0xe6,0x42,0x68 |
| .byte 0x41,0x99,0x2d,0x0f,0xb0,0x54,0xbb,0x16 |
| .long 0x00000001, 0x00000002, 0x00000004, 0x00000008 |
| .long 0x00000010, 0x00000020, 0x00000040, 0x00000080 |
| .long 0x0000001b, 0x00000036, 0x80808080, 0x80808080 |
| .long 0xfefefefe, 0xfefefefe, 0x1b1b1b1b, 0x1b1b1b1b |
| .align 64 |
| .LAES_Td: |
| .long 0x50a7f451,0x50a7f451 |
| .long 0x5365417e,0x5365417e |
| .long 0xc3a4171a,0xc3a4171a |
| .long 0x965e273a,0x965e273a |
| .long 0xcb6bab3b,0xcb6bab3b |
| .long 0xf1459d1f,0xf1459d1f |
| .long 0xab58faac,0xab58faac |
| .long 0x9303e34b,0x9303e34b |
| .long 0x55fa3020,0x55fa3020 |
| .long 0xf66d76ad,0xf66d76ad |
| .long 0x9176cc88,0x9176cc88 |
| .long 0x254c02f5,0x254c02f5 |
| .long 0xfcd7e54f,0xfcd7e54f |
| .long 0xd7cb2ac5,0xd7cb2ac5 |
| .long 0x80443526,0x80443526 |
| .long 0x8fa362b5,0x8fa362b5 |
| .long 0x495ab1de,0x495ab1de |
| .long 0x671bba25,0x671bba25 |
| .long 0x980eea45,0x980eea45 |
| .long 0xe1c0fe5d,0xe1c0fe5d |
| .long 0x02752fc3,0x02752fc3 |
| .long 0x12f04c81,0x12f04c81 |
| .long 0xa397468d,0xa397468d |
| .long 0xc6f9d36b,0xc6f9d36b |
| .long 0xe75f8f03,0xe75f8f03 |
| .long 0x959c9215,0x959c9215 |
| .long 0xeb7a6dbf,0xeb7a6dbf |
| .long 0xda595295,0xda595295 |
| .long 0x2d83bed4,0x2d83bed4 |
| .long 0xd3217458,0xd3217458 |
| .long 0x2969e049,0x2969e049 |
| .long 0x44c8c98e,0x44c8c98e |
| .long 0x6a89c275,0x6a89c275 |
| .long 0x78798ef4,0x78798ef4 |
| .long 0x6b3e5899,0x6b3e5899 |
| .long 0xdd71b927,0xdd71b927 |
| .long 0xb64fe1be,0xb64fe1be |
| .long 0x17ad88f0,0x17ad88f0 |
| .long 0x66ac20c9,0x66ac20c9 |
| .long 0xb43ace7d,0xb43ace7d |
| .long 0x184adf63,0x184adf63 |
| .long 0x82311ae5,0x82311ae5 |
| .long 0x60335197,0x60335197 |
| .long 0x457f5362,0x457f5362 |
| .long 0xe07764b1,0xe07764b1 |
| .long 0x84ae6bbb,0x84ae6bbb |
| .long 0x1ca081fe,0x1ca081fe |
| .long 0x942b08f9,0x942b08f9 |
| .long 0x58684870,0x58684870 |
| .long 0x19fd458f,0x19fd458f |
| .long 0x876cde94,0x876cde94 |
| .long 0xb7f87b52,0xb7f87b52 |
| .long 0x23d373ab,0x23d373ab |
| .long 0xe2024b72,0xe2024b72 |
| .long 0x578f1fe3,0x578f1fe3 |
| .long 0x2aab5566,0x2aab5566 |
| .long 0x0728ebb2,0x0728ebb2 |
| .long 0x03c2b52f,0x03c2b52f |
| .long 0x9a7bc586,0x9a7bc586 |
| .long 0xa50837d3,0xa50837d3 |
| .long 0xf2872830,0xf2872830 |
| .long 0xb2a5bf23,0xb2a5bf23 |
| .long 0xba6a0302,0xba6a0302 |
| .long 0x5c8216ed,0x5c8216ed |
| .long 0x2b1ccf8a,0x2b1ccf8a |
| .long 0x92b479a7,0x92b479a7 |
| .long 0xf0f207f3,0xf0f207f3 |
| .long 0xa1e2694e,0xa1e2694e |
| .long 0xcdf4da65,0xcdf4da65 |
| .long 0xd5be0506,0xd5be0506 |
| .long 0x1f6234d1,0x1f6234d1 |
| .long 0x8afea6c4,0x8afea6c4 |
| .long 0x9d532e34,0x9d532e34 |
| .long 0xa055f3a2,0xa055f3a2 |
| .long 0x32e18a05,0x32e18a05 |
| .long 0x75ebf6a4,0x75ebf6a4 |
| .long 0x39ec830b,0x39ec830b |
| .long 0xaaef6040,0xaaef6040 |
| .long 0x069f715e,0x069f715e |
| .long 0x51106ebd,0x51106ebd |
| .long 0xf98a213e,0xf98a213e |
| .long 0x3d06dd96,0x3d06dd96 |
| .long 0xae053edd,0xae053edd |
| .long 0x46bde64d,0x46bde64d |
| .long 0xb58d5491,0xb58d5491 |
| .long 0x055dc471,0x055dc471 |
| .long 0x6fd40604,0x6fd40604 |
| .long 0xff155060,0xff155060 |
| .long 0x24fb9819,0x24fb9819 |
| .long 0x97e9bdd6,0x97e9bdd6 |
| .long 0xcc434089,0xcc434089 |
| .long 0x779ed967,0x779ed967 |
| .long 0xbd42e8b0,0xbd42e8b0 |
| .long 0x888b8907,0x888b8907 |
| .long 0x385b19e7,0x385b19e7 |
| .long 0xdbeec879,0xdbeec879 |
| .long 0x470a7ca1,0x470a7ca1 |
| .long 0xe90f427c,0xe90f427c |
| .long 0xc91e84f8,0xc91e84f8 |
| .long 0x00000000,0x00000000 |
| .long 0x83868009,0x83868009 |
| .long 0x48ed2b32,0x48ed2b32 |
| .long 0xac70111e,0xac70111e |
| .long 0x4e725a6c,0x4e725a6c |
| .long 0xfbff0efd,0xfbff0efd |
| .long 0x5638850f,0x5638850f |
| .long 0x1ed5ae3d,0x1ed5ae3d |
| .long 0x27392d36,0x27392d36 |
| .long 0x64d90f0a,0x64d90f0a |
| .long 0x21a65c68,0x21a65c68 |
| .long 0xd1545b9b,0xd1545b9b |
| .long 0x3a2e3624,0x3a2e3624 |
| .long 0xb1670a0c,0xb1670a0c |
| .long 0x0fe75793,0x0fe75793 |
| .long 0xd296eeb4,0xd296eeb4 |
| .long 0x9e919b1b,0x9e919b1b |
| .long 0x4fc5c080,0x4fc5c080 |
| .long 0xa220dc61,0xa220dc61 |
| .long 0x694b775a,0x694b775a |
| .long 0x161a121c,0x161a121c |
| .long 0x0aba93e2,0x0aba93e2 |
| .long 0xe52aa0c0,0xe52aa0c0 |
| .long 0x43e0223c,0x43e0223c |
| .long 0x1d171b12,0x1d171b12 |
| .long 0x0b0d090e,0x0b0d090e |
| .long 0xadc78bf2,0xadc78bf2 |
| .long 0xb9a8b62d,0xb9a8b62d |
| .long 0xc8a91e14,0xc8a91e14 |
| .long 0x8519f157,0x8519f157 |
| .long 0x4c0775af,0x4c0775af |
| .long 0xbbdd99ee,0xbbdd99ee |
| .long 0xfd607fa3,0xfd607fa3 |
| .long 0x9f2601f7,0x9f2601f7 |
| .long 0xbcf5725c,0xbcf5725c |
| .long 0xc53b6644,0xc53b6644 |
| .long 0x347efb5b,0x347efb5b |
| .long 0x7629438b,0x7629438b |
| .long 0xdcc623cb,0xdcc623cb |
| .long 0x68fcedb6,0x68fcedb6 |
| .long 0x63f1e4b8,0x63f1e4b8 |
| .long 0xcadc31d7,0xcadc31d7 |
| .long 0x10856342,0x10856342 |
| .long 0x40229713,0x40229713 |
| .long 0x2011c684,0x2011c684 |
| .long 0x7d244a85,0x7d244a85 |
| .long 0xf83dbbd2,0xf83dbbd2 |
| .long 0x1132f9ae,0x1132f9ae |
| .long 0x6da129c7,0x6da129c7 |
| .long 0x4b2f9e1d,0x4b2f9e1d |
| .long 0xf330b2dc,0xf330b2dc |
| .long 0xec52860d,0xec52860d |
| .long 0xd0e3c177,0xd0e3c177 |
| .long 0x6c16b32b,0x6c16b32b |
| .long 0x99b970a9,0x99b970a9 |
| .long 0xfa489411,0xfa489411 |
| .long 0x2264e947,0x2264e947 |
| .long 0xc48cfca8,0xc48cfca8 |
| .long 0x1a3ff0a0,0x1a3ff0a0 |
| .long 0xd82c7d56,0xd82c7d56 |
| .long 0xef903322,0xef903322 |
| .long 0xc74e4987,0xc74e4987 |
| .long 0xc1d138d9,0xc1d138d9 |
| .long 0xfea2ca8c,0xfea2ca8c |
| .long 0x360bd498,0x360bd498 |
| .long 0xcf81f5a6,0xcf81f5a6 |
| .long 0x28de7aa5,0x28de7aa5 |
| .long 0x268eb7da,0x268eb7da |
| .long 0xa4bfad3f,0xa4bfad3f |
| .long 0xe49d3a2c,0xe49d3a2c |
| .long 0x0d927850,0x0d927850 |
| .long 0x9bcc5f6a,0x9bcc5f6a |
| .long 0x62467e54,0x62467e54 |
| .long 0xc2138df6,0xc2138df6 |
| .long 0xe8b8d890,0xe8b8d890 |
| .long 0x5ef7392e,0x5ef7392e |
| .long 0xf5afc382,0xf5afc382 |
| .long 0xbe805d9f,0xbe805d9f |
| .long 0x7c93d069,0x7c93d069 |
| .long 0xa92dd56f,0xa92dd56f |
| .long 0xb31225cf,0xb31225cf |
| .long 0x3b99acc8,0x3b99acc8 |
| .long 0xa77d1810,0xa77d1810 |
| .long 0x6e639ce8,0x6e639ce8 |
| .long 0x7bbb3bdb,0x7bbb3bdb |
| .long 0x097826cd,0x097826cd |
| .long 0xf418596e,0xf418596e |
| .long 0x01b79aec,0x01b79aec |
| .long 0xa89a4f83,0xa89a4f83 |
| .long 0x656e95e6,0x656e95e6 |
| .long 0x7ee6ffaa,0x7ee6ffaa |
| .long 0x08cfbc21,0x08cfbc21 |
| .long 0xe6e815ef,0xe6e815ef |
| .long 0xd99be7ba,0xd99be7ba |
| .long 0xce366f4a,0xce366f4a |
| .long 0xd4099fea,0xd4099fea |
| .long 0xd67cb029,0xd67cb029 |
| .long 0xafb2a431,0xafb2a431 |
| .long 0x31233f2a,0x31233f2a |
| .long 0x3094a5c6,0x3094a5c6 |
| .long 0xc066a235,0xc066a235 |
| .long 0x37bc4e74,0x37bc4e74 |
| .long 0xa6ca82fc,0xa6ca82fc |
| .long 0xb0d090e0,0xb0d090e0 |
| .long 0x15d8a733,0x15d8a733 |
| .long 0x4a9804f1,0x4a9804f1 |
| .long 0xf7daec41,0xf7daec41 |
| .long 0x0e50cd7f,0x0e50cd7f |
| .long 0x2ff69117,0x2ff69117 |
| .long 0x8dd64d76,0x8dd64d76 |
| .long 0x4db0ef43,0x4db0ef43 |
| .long 0x544daacc,0x544daacc |
| .long 0xdf0496e4,0xdf0496e4 |
| .long 0xe3b5d19e,0xe3b5d19e |
| .long 0x1b886a4c,0x1b886a4c |
| .long 0xb81f2cc1,0xb81f2cc1 |
| .long 0x7f516546,0x7f516546 |
| .long 0x04ea5e9d,0x04ea5e9d |
| .long 0x5d358c01,0x5d358c01 |
| .long 0x737487fa,0x737487fa |
| .long 0x2e410bfb,0x2e410bfb |
| .long 0x5a1d67b3,0x5a1d67b3 |
| .long 0x52d2db92,0x52d2db92 |
| .long 0x335610e9,0x335610e9 |
| .long 0x1347d66d,0x1347d66d |
| .long 0x8c61d79a,0x8c61d79a |
| .long 0x7a0ca137,0x7a0ca137 |
| .long 0x8e14f859,0x8e14f859 |
| .long 0x893c13eb,0x893c13eb |
| .long 0xee27a9ce,0xee27a9ce |
| .long 0x35c961b7,0x35c961b7 |
| .long 0xede51ce1,0xede51ce1 |
| .long 0x3cb1477a,0x3cb1477a |
| .long 0x59dfd29c,0x59dfd29c |
| .long 0x3f73f255,0x3f73f255 |
| .long 0x79ce1418,0x79ce1418 |
| .long 0xbf37c773,0xbf37c773 |
| .long 0xeacdf753,0xeacdf753 |
| .long 0x5baafd5f,0x5baafd5f |
| .long 0x146f3ddf,0x146f3ddf |
| .long 0x86db4478,0x86db4478 |
| .long 0x81f3afca,0x81f3afca |
| .long 0x3ec468b9,0x3ec468b9 |
| .long 0x2c342438,0x2c342438 |
| .long 0x5f40a3c2,0x5f40a3c2 |
| .long 0x72c31d16,0x72c31d16 |
| .long 0x0c25e2bc,0x0c25e2bc |
| .long 0x8b493c28,0x8b493c28 |
| .long 0x41950dff,0x41950dff |
| .long 0x7101a839,0x7101a839 |
| .long 0xdeb30c08,0xdeb30c08 |
| .long 0x9ce4b4d8,0x9ce4b4d8 |
| .long 0x90c15664,0x90c15664 |
| .long 0x6184cb7b,0x6184cb7b |
| .long 0x70b632d5,0x70b632d5 |
| .long 0x745c6c48,0x745c6c48 |
| .long 0x4257b8d0,0x4257b8d0 |
| .byte 0x52,0x09,0x6a,0xd5,0x30,0x36,0xa5,0x38 |
| .byte 0xbf,0x40,0xa3,0x9e,0x81,0xf3,0xd7,0xfb |
| .byte 0x7c,0xe3,0x39,0x82,0x9b,0x2f,0xff,0x87 |
| .byte 0x34,0x8e,0x43,0x44,0xc4,0xde,0xe9,0xcb |
| .byte 0x54,0x7b,0x94,0x32,0xa6,0xc2,0x23,0x3d |
| .byte 0xee,0x4c,0x95,0x0b,0x42,0xfa,0xc3,0x4e |
| .byte 0x08,0x2e,0xa1,0x66,0x28,0xd9,0x24,0xb2 |
| .byte 0x76,0x5b,0xa2,0x49,0x6d,0x8b,0xd1,0x25 |
| .byte 0x72,0xf8,0xf6,0x64,0x86,0x68,0x98,0x16 |
| .byte 0xd4,0xa4,0x5c,0xcc,0x5d,0x65,0xb6,0x92 |
| .byte 0x6c,0x70,0x48,0x50,0xfd,0xed,0xb9,0xda |
| .byte 0x5e,0x15,0x46,0x57,0xa7,0x8d,0x9d,0x84 |
| .byte 0x90,0xd8,0xab,0x00,0x8c,0xbc,0xd3,0x0a |
| .byte 0xf7,0xe4,0x58,0x05,0xb8,0xb3,0x45,0x06 |
| .byte 0xd0,0x2c,0x1e,0x8f,0xca,0x3f,0x0f,0x02 |
| .byte 0xc1,0xaf,0xbd,0x03,0x01,0x13,0x8a,0x6b |
| .byte 0x3a,0x91,0x11,0x41,0x4f,0x67,0xdc,0xea |
| .byte 0x97,0xf2,0xcf,0xce,0xf0,0xb4,0xe6,0x73 |
| .byte 0x96,0xac,0x74,0x22,0xe7,0xad,0x35,0x85 |
| .byte 0xe2,0xf9,0x37,0xe8,0x1c,0x75,0xdf,0x6e |
| .byte 0x47,0xf1,0x1a,0x71,0x1d,0x29,0xc5,0x89 |
| .byte 0x6f,0xb7,0x62,0x0e,0xaa,0x18,0xbe,0x1b |
| .byte 0xfc,0x56,0x3e,0x4b,0xc6,0xd2,0x79,0x20 |
| .byte 0x9a,0xdb,0xc0,0xfe,0x78,0xcd,0x5a,0xf4 |
| .byte 0x1f,0xdd,0xa8,0x33,0x88,0x07,0xc7,0x31 |
| .byte 0xb1,0x12,0x10,0x59,0x27,0x80,0xec,0x5f |
| .byte 0x60,0x51,0x7f,0xa9,0x19,0xb5,0x4a,0x0d |
| .byte 0x2d,0xe5,0x7a,0x9f,0x93,0xc9,0x9c,0xef |
| .byte 0xa0,0xe0,0x3b,0x4d,0xae,0x2a,0xf5,0xb0 |
| .byte 0xc8,0xeb,0xbb,0x3c,0x83,0x53,0x99,0x61 |
| .byte 0x17,0x2b,0x04,0x7e,0xba,0x77,0xd6,0x26 |
| .byte 0xe1,0x69,0x14,0x63,0x55,0x21,0x0c,0x7d |
| .long 0x80808080, 0x80808080, 0xfefefefe, 0xfefefefe |
| .long 0x1b1b1b1b, 0x1b1b1b1b, 0, 0 |
| .byte 0x52,0x09,0x6a,0xd5,0x30,0x36,0xa5,0x38 |
| .byte 0xbf,0x40,0xa3,0x9e,0x81,0xf3,0xd7,0xfb |
| .byte 0x7c,0xe3,0x39,0x82,0x9b,0x2f,0xff,0x87 |
| .byte 0x34,0x8e,0x43,0x44,0xc4,0xde,0xe9,0xcb |
| .byte 0x54,0x7b,0x94,0x32,0xa6,0xc2,0x23,0x3d |
| .byte 0xee,0x4c,0x95,0x0b,0x42,0xfa,0xc3,0x4e |
| .byte 0x08,0x2e,0xa1,0x66,0x28,0xd9,0x24,0xb2 |
| .byte 0x76,0x5b,0xa2,0x49,0x6d,0x8b,0xd1,0x25 |
| .byte 0x72,0xf8,0xf6,0x64,0x86,0x68,0x98,0x16 |
| .byte 0xd4,0xa4,0x5c,0xcc,0x5d,0x65,0xb6,0x92 |
| .byte 0x6c,0x70,0x48,0x50,0xfd,0xed,0xb9,0xda |
| .byte 0x5e,0x15,0x46,0x57,0xa7,0x8d,0x9d,0x84 |
| .byte 0x90,0xd8,0xab,0x00,0x8c,0xbc,0xd3,0x0a |
| .byte 0xf7,0xe4,0x58,0x05,0xb8,0xb3,0x45,0x06 |
| .byte 0xd0,0x2c,0x1e,0x8f,0xca,0x3f,0x0f,0x02 |
| .byte 0xc1,0xaf,0xbd,0x03,0x01,0x13,0x8a,0x6b |
| .byte 0x3a,0x91,0x11,0x41,0x4f,0x67,0xdc,0xea |
| .byte 0x97,0xf2,0xcf,0xce,0xf0,0xb4,0xe6,0x73 |
| .byte 0x96,0xac,0x74,0x22,0xe7,0xad,0x35,0x85 |
| .byte 0xe2,0xf9,0x37,0xe8,0x1c,0x75,0xdf,0x6e |
| .byte 0x47,0xf1,0x1a,0x71,0x1d,0x29,0xc5,0x89 |
| .byte 0x6f,0xb7,0x62,0x0e,0xaa,0x18,0xbe,0x1b |
| .byte 0xfc,0x56,0x3e,0x4b,0xc6,0xd2,0x79,0x20 |
| .byte 0x9a,0xdb,0xc0,0xfe,0x78,0xcd,0x5a,0xf4 |
| .byte 0x1f,0xdd,0xa8,0x33,0x88,0x07,0xc7,0x31 |
| .byte 0xb1,0x12,0x10,0x59,0x27,0x80,0xec,0x5f |
| .byte 0x60,0x51,0x7f,0xa9,0x19,0xb5,0x4a,0x0d |
| .byte 0x2d,0xe5,0x7a,0x9f,0x93,0xc9,0x9c,0xef |
| .byte 0xa0,0xe0,0x3b,0x4d,0xae,0x2a,0xf5,0xb0 |
| .byte 0xc8,0xeb,0xbb,0x3c,0x83,0x53,0x99,0x61 |
| .byte 0x17,0x2b,0x04,0x7e,0xba,0x77,0xd6,0x26 |
| .byte 0xe1,0x69,0x14,0x63,0x55,0x21,0x0c,0x7d |
| .long 0x80808080, 0x80808080, 0xfefefefe, 0xfefefefe |
| .long 0x1b1b1b1b, 0x1b1b1b1b, 0, 0 |
| .byte 0x52,0x09,0x6a,0xd5,0x30,0x36,0xa5,0x38 |
| .byte 0xbf,0x40,0xa3,0x9e,0x81,0xf3,0xd7,0xfb |
| .byte 0x7c,0xe3,0x39,0x82,0x9b,0x2f,0xff,0x87 |
| .byte 0x34,0x8e,0x43,0x44,0xc4,0xde,0xe9,0xcb |
| .byte 0x54,0x7b,0x94,0x32,0xa6,0xc2,0x23,0x3d |
| .byte 0xee,0x4c,0x95,0x0b,0x42,0xfa,0xc3,0x4e |
| .byte 0x08,0x2e,0xa1,0x66,0x28,0xd9,0x24,0xb2 |
| .byte 0x76,0x5b,0xa2,0x49,0x6d,0x8b,0xd1,0x25 |
| .byte 0x72,0xf8,0xf6,0x64,0x86,0x68,0x98,0x16 |
| .byte 0xd4,0xa4,0x5c,0xcc,0x5d,0x65,0xb6,0x92 |
| .byte 0x6c,0x70,0x48,0x50,0xfd,0xed,0xb9,0xda |
| .byte 0x5e,0x15,0x46,0x57,0xa7,0x8d,0x9d,0x84 |
| .byte 0x90,0xd8,0xab,0x00,0x8c,0xbc,0xd3,0x0a |
| .byte 0xf7,0xe4,0x58,0x05,0xb8,0xb3,0x45,0x06 |
| .byte 0xd0,0x2c,0x1e,0x8f,0xca,0x3f,0x0f,0x02 |
| .byte 0xc1,0xaf,0xbd,0x03,0x01,0x13,0x8a,0x6b |
| .byte 0x3a,0x91,0x11,0x41,0x4f,0x67,0xdc,0xea |
| .byte 0x97,0xf2,0xcf,0xce,0xf0,0xb4,0xe6,0x73 |
| .byte 0x96,0xac,0x74,0x22,0xe7,0xad,0x35,0x85 |
| .byte 0xe2,0xf9,0x37,0xe8,0x1c,0x75,0xdf,0x6e |
| .byte 0x47,0xf1,0x1a,0x71,0x1d,0x29,0xc5,0x89 |
| .byte 0x6f,0xb7,0x62,0x0e,0xaa,0x18,0xbe,0x1b |
| .byte 0xfc,0x56,0x3e,0x4b,0xc6,0xd2,0x79,0x20 |
| .byte 0x9a,0xdb,0xc0,0xfe,0x78,0xcd,0x5a,0xf4 |
| .byte 0x1f,0xdd,0xa8,0x33,0x88,0x07,0xc7,0x31 |
| .byte 0xb1,0x12,0x10,0x59,0x27,0x80,0xec,0x5f |
| .byte 0x60,0x51,0x7f,0xa9,0x19,0xb5,0x4a,0x0d |
| .byte 0x2d,0xe5,0x7a,0x9f,0x93,0xc9,0x9c,0xef |
| .byte 0xa0,0xe0,0x3b,0x4d,0xae,0x2a,0xf5,0xb0 |
| .byte 0xc8,0xeb,0xbb,0x3c,0x83,0x53,0x99,0x61 |
| .byte 0x17,0x2b,0x04,0x7e,0xba,0x77,0xd6,0x26 |
| .byte 0xe1,0x69,0x14,0x63,0x55,0x21,0x0c,0x7d |
| .long 0x80808080, 0x80808080, 0xfefefefe, 0xfefefefe |
| .long 0x1b1b1b1b, 0x1b1b1b1b, 0, 0 |
| .byte 0x52,0x09,0x6a,0xd5,0x30,0x36,0xa5,0x38 |
| .byte 0xbf,0x40,0xa3,0x9e,0x81,0xf3,0xd7,0xfb |
| .byte 0x7c,0xe3,0x39,0x82,0x9b,0x2f,0xff,0x87 |
| .byte 0x34,0x8e,0x43,0x44,0xc4,0xde,0xe9,0xcb |
| .byte 0x54,0x7b,0x94,0x32,0xa6,0xc2,0x23,0x3d |
| .byte 0xee,0x4c,0x95,0x0b,0x42,0xfa,0xc3,0x4e |
| .byte 0x08,0x2e,0xa1,0x66,0x28,0xd9,0x24,0xb2 |
| .byte 0x76,0x5b,0xa2,0x49,0x6d,0x8b,0xd1,0x25 |
| .byte 0x72,0xf8,0xf6,0x64,0x86,0x68,0x98,0x16 |
| .byte 0xd4,0xa4,0x5c,0xcc,0x5d,0x65,0xb6,0x92 |
| .byte 0x6c,0x70,0x48,0x50,0xfd,0xed,0xb9,0xda |
| .byte 0x5e,0x15,0x46,0x57,0xa7,0x8d,0x9d,0x84 |
| .byte 0x90,0xd8,0xab,0x00,0x8c,0xbc,0xd3,0x0a |
| .byte 0xf7,0xe4,0x58,0x05,0xb8,0xb3,0x45,0x06 |
| .byte 0xd0,0x2c,0x1e,0x8f,0xca,0x3f,0x0f,0x02 |
| .byte 0xc1,0xaf,0xbd,0x03,0x01,0x13,0x8a,0x6b |
| .byte 0x3a,0x91,0x11,0x41,0x4f,0x67,0xdc,0xea |
| .byte 0x97,0xf2,0xcf,0xce,0xf0,0xb4,0xe6,0x73 |
| .byte 0x96,0xac,0x74,0x22,0xe7,0xad,0x35,0x85 |
| .byte 0xe2,0xf9,0x37,0xe8,0x1c,0x75,0xdf,0x6e |
| .byte 0x47,0xf1,0x1a,0x71,0x1d,0x29,0xc5,0x89 |
| .byte 0x6f,0xb7,0x62,0x0e,0xaa,0x18,0xbe,0x1b |
| .byte 0xfc,0x56,0x3e,0x4b,0xc6,0xd2,0x79,0x20 |
| .byte 0x9a,0xdb,0xc0,0xfe,0x78,0xcd,0x5a,0xf4 |
| .byte 0x1f,0xdd,0xa8,0x33,0x88,0x07,0xc7,0x31 |
| .byte 0xb1,0x12,0x10,0x59,0x27,0x80,0xec,0x5f |
| .byte 0x60,0x51,0x7f,0xa9,0x19,0xb5,0x4a,0x0d |
| .byte 0x2d,0xe5,0x7a,0x9f,0x93,0xc9,0x9c,0xef |
| .byte 0xa0,0xe0,0x3b,0x4d,0xae,0x2a,0xf5,0xb0 |
| .byte 0xc8,0xeb,0xbb,0x3c,0x83,0x53,0x99,0x61 |
| .byte 0x17,0x2b,0x04,0x7e,0xba,0x77,0xd6,0x26 |
| .byte 0xe1,0x69,0x14,0x63,0x55,0x21,0x0c,0x7d |
| .long 0x80808080, 0x80808080, 0xfefefefe, 0xfefefefe |
| .long 0x1b1b1b1b, 0x1b1b1b1b, 0, 0 |
| .byte 65,69,83,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62,0 |
| .align 64 |
| #endif |