| %ifidn __OUTPUT_FORMAT__,obj |
| section code use32 class=code align=64 |
| %elifidn __OUTPUT_FORMAT__,win32 |
| %ifdef __YASM_VERSION_ID__ |
| %if __YASM_VERSION_ID__ < 01010000h |
| %error yasm version 1.1.0 or later needed. |
| %endif |
| ; Yasm automatically includes .00 and complains about redefining it. |
| ; https://www.tortall.net/projects/yasm/manual/html/objfmt-win32-safeseh.html |
| %else |
| $@feat.00 equ 1 |
| %endif |
| section .text code align=64 |
| %else |
| section .text code |
| %endif |
| global _gcm_gmult_4bit_mmx |
| align 16 |
| _gcm_gmult_4bit_mmx: |
| L$_gcm_gmult_4bit_mmx_begin: |
| push ebp |
| push ebx |
| push esi |
| push edi |
| mov edi,DWORD [20+esp] |
| mov esi,DWORD [24+esp] |
| call L$000pic_point |
| L$000pic_point: |
| pop eax |
| lea eax,[(L$rem_4bit-L$000pic_point)+eax] |
| movzx ebx,BYTE [15+edi] |
| xor ecx,ecx |
| mov edx,ebx |
| mov cl,dl |
| mov ebp,14 |
| shl cl,4 |
| and edx,240 |
| movq mm0,[8+ecx*1+esi] |
| movq mm1,[ecx*1+esi] |
| movd ebx,mm0 |
| jmp NEAR L$001mmx_loop |
| align 16 |
| L$001mmx_loop: |
| psrlq mm0,4 |
| and ebx,15 |
| movq mm2,mm1 |
| psrlq mm1,4 |
| pxor mm0,[8+edx*1+esi] |
| mov cl,BYTE [ebp*1+edi] |
| psllq mm2,60 |
| pxor mm1,[ebx*8+eax] |
| dec ebp |
| movd ebx,mm0 |
| pxor mm1,[edx*1+esi] |
| mov edx,ecx |
| pxor mm0,mm2 |
| js NEAR L$002mmx_break |
| shl cl,4 |
| and ebx,15 |
| psrlq mm0,4 |
| and edx,240 |
| movq mm2,mm1 |
| psrlq mm1,4 |
| pxor mm0,[8+ecx*1+esi] |
| psllq mm2,60 |
| pxor mm1,[ebx*8+eax] |
| movd ebx,mm0 |
| pxor mm1,[ecx*1+esi] |
| pxor mm0,mm2 |
| jmp NEAR L$001mmx_loop |
| align 16 |
| L$002mmx_break: |
| shl cl,4 |
| and ebx,15 |
| psrlq mm0,4 |
| and edx,240 |
| movq mm2,mm1 |
| psrlq mm1,4 |
| pxor mm0,[8+ecx*1+esi] |
| psllq mm2,60 |
| pxor mm1,[ebx*8+eax] |
| movd ebx,mm0 |
| pxor mm1,[ecx*1+esi] |
| pxor mm0,mm2 |
| psrlq mm0,4 |
| and ebx,15 |
| movq mm2,mm1 |
| psrlq mm1,4 |
| pxor mm0,[8+edx*1+esi] |
| psllq mm2,60 |
| pxor mm1,[ebx*8+eax] |
| movd ebx,mm0 |
| pxor mm1,[edx*1+esi] |
| pxor mm0,mm2 |
| psrlq mm0,32 |
| movd edx,mm1 |
| psrlq mm1,32 |
| movd ecx,mm0 |
| movd ebp,mm1 |
| bswap ebx |
| bswap edx |
| bswap ecx |
| bswap ebp |
| emms |
| mov DWORD [12+edi],ebx |
| mov DWORD [4+edi],edx |
| mov DWORD [8+edi],ecx |
| mov DWORD [edi],ebp |
| pop edi |
| pop esi |
| pop ebx |
| pop ebp |
| ret |
| global _gcm_ghash_4bit_mmx |
| align 16 |
| _gcm_ghash_4bit_mmx: |
| L$_gcm_ghash_4bit_mmx_begin: |
| push ebp |
| push ebx |
| push esi |
| push edi |
| mov eax,DWORD [20+esp] |
| mov ebx,DWORD [24+esp] |
| mov ecx,DWORD [28+esp] |
| mov edx,DWORD [32+esp] |
| mov ebp,esp |
| call L$003pic_point |
| L$003pic_point: |
| pop esi |
| lea esi,[(L$rem_8bit-L$003pic_point)+esi] |
| sub esp,544 |
| and esp,-64 |
| sub esp,16 |
| add edx,ecx |
| mov DWORD [544+esp],eax |
| mov DWORD [552+esp],edx |
| mov DWORD [556+esp],ebp |
| add ebx,128 |
| lea edi,[144+esp] |
| lea ebp,[400+esp] |
| mov edx,DWORD [ebx-120] |
| movq mm0,[ebx-120] |
| movq mm3,[ebx-128] |
| shl edx,4 |
| mov BYTE [esp],dl |
| mov edx,DWORD [ebx-104] |
| movq mm2,[ebx-104] |
| movq mm5,[ebx-112] |
| movq [edi-128],mm0 |
| psrlq mm0,4 |
| movq [edi],mm3 |
| movq mm7,mm3 |
| psrlq mm3,4 |
| shl edx,4 |
| mov BYTE [1+esp],dl |
| mov edx,DWORD [ebx-88] |
| movq mm1,[ebx-88] |
| psllq mm7,60 |
| movq mm4,[ebx-96] |
| por mm0,mm7 |
| movq [edi-120],mm2 |
| psrlq mm2,4 |
| movq [8+edi],mm5 |
| movq mm6,mm5 |
| movq [ebp-128],mm0 |
| psrlq mm5,4 |
| movq [ebp],mm3 |
| shl edx,4 |
| mov BYTE [2+esp],dl |
| mov edx,DWORD [ebx-72] |
| movq mm0,[ebx-72] |
| psllq mm6,60 |
| movq mm3,[ebx-80] |
| por mm2,mm6 |
| movq [edi-112],mm1 |
| psrlq mm1,4 |
| movq [16+edi],mm4 |
| movq mm7,mm4 |
| movq [ebp-120],mm2 |
| psrlq mm4,4 |
| movq [8+ebp],mm5 |
| shl edx,4 |
| mov BYTE [3+esp],dl |
| mov edx,DWORD [ebx-56] |
| movq mm2,[ebx-56] |
| psllq mm7,60 |
| movq mm5,[ebx-64] |
| por mm1,mm7 |
| movq [edi-104],mm0 |
| psrlq mm0,4 |
| movq [24+edi],mm3 |
| movq mm6,mm3 |
| movq [ebp-112],mm1 |
| psrlq mm3,4 |
| movq [16+ebp],mm4 |
| shl edx,4 |
| mov BYTE [4+esp],dl |
| mov edx,DWORD [ebx-40] |
| movq mm1,[ebx-40] |
| psllq mm6,60 |
| movq mm4,[ebx-48] |
| por mm0,mm6 |
| movq [edi-96],mm2 |
| psrlq mm2,4 |
| movq [32+edi],mm5 |
| movq mm7,mm5 |
| movq [ebp-104],mm0 |
| psrlq mm5,4 |
| movq [24+ebp],mm3 |
| shl edx,4 |
| mov BYTE [5+esp],dl |
| mov edx,DWORD [ebx-24] |
| movq mm0,[ebx-24] |
| psllq mm7,60 |
| movq mm3,[ebx-32] |
| por mm2,mm7 |
| movq [edi-88],mm1 |
| psrlq mm1,4 |
| movq [40+edi],mm4 |
| movq mm6,mm4 |
| movq [ebp-96],mm2 |
| psrlq mm4,4 |
| movq [32+ebp],mm5 |
| shl edx,4 |
| mov BYTE [6+esp],dl |
| mov edx,DWORD [ebx-8] |
| movq mm2,[ebx-8] |
| psllq mm6,60 |
| movq mm5,[ebx-16] |
| por mm1,mm6 |
| movq [edi-80],mm0 |
| psrlq mm0,4 |
| movq [48+edi],mm3 |
| movq mm7,mm3 |
| movq [ebp-88],mm1 |
| psrlq mm3,4 |
| movq [40+ebp],mm4 |
| shl edx,4 |
| mov BYTE [7+esp],dl |
| mov edx,DWORD [8+ebx] |
| movq mm1,[8+ebx] |
| psllq mm7,60 |
| movq mm4,[ebx] |
| por mm0,mm7 |
| movq [edi-72],mm2 |
| psrlq mm2,4 |
| movq [56+edi],mm5 |
| movq mm6,mm5 |
| movq [ebp-80],mm0 |
| psrlq mm5,4 |
| movq [48+ebp],mm3 |
| shl edx,4 |
| mov BYTE [8+esp],dl |
| mov edx,DWORD [24+ebx] |
| movq mm0,[24+ebx] |
| psllq mm6,60 |
| movq mm3,[16+ebx] |
| por mm2,mm6 |
| movq [edi-64],mm1 |
| psrlq mm1,4 |
| movq [64+edi],mm4 |
| movq mm7,mm4 |
| movq [ebp-72],mm2 |
| psrlq mm4,4 |
| movq [56+ebp],mm5 |
| shl edx,4 |
| mov BYTE [9+esp],dl |
| mov edx,DWORD [40+ebx] |
| movq mm2,[40+ebx] |
| psllq mm7,60 |
| movq mm5,[32+ebx] |
| por mm1,mm7 |
| movq [edi-56],mm0 |
| psrlq mm0,4 |
| movq [72+edi],mm3 |
| movq mm6,mm3 |
| movq [ebp-64],mm1 |
| psrlq mm3,4 |
| movq [64+ebp],mm4 |
| shl edx,4 |
| mov BYTE [10+esp],dl |
| mov edx,DWORD [56+ebx] |
| movq mm1,[56+ebx] |
| psllq mm6,60 |
| movq mm4,[48+ebx] |
| por mm0,mm6 |
| movq [edi-48],mm2 |
| psrlq mm2,4 |
| movq [80+edi],mm5 |
| movq mm7,mm5 |
| movq [ebp-56],mm0 |
| psrlq mm5,4 |
| movq [72+ebp],mm3 |
| shl edx,4 |
| mov BYTE [11+esp],dl |
| mov edx,DWORD [72+ebx] |
| movq mm0,[72+ebx] |
| psllq mm7,60 |
| movq mm3,[64+ebx] |
| por mm2,mm7 |
| movq [edi-40],mm1 |
| psrlq mm1,4 |
| movq [88+edi],mm4 |
| movq mm6,mm4 |
| movq [ebp-48],mm2 |
| psrlq mm4,4 |
| movq [80+ebp],mm5 |
| shl edx,4 |
| mov BYTE [12+esp],dl |
| mov edx,DWORD [88+ebx] |
| movq mm2,[88+ebx] |
| psllq mm6,60 |
| movq mm5,[80+ebx] |
| por mm1,mm6 |
| movq [edi-32],mm0 |
| psrlq mm0,4 |
| movq [96+edi],mm3 |
| movq mm7,mm3 |
| movq [ebp-40],mm1 |
| psrlq mm3,4 |
| movq [88+ebp],mm4 |
| shl edx,4 |
| mov BYTE [13+esp],dl |
| mov edx,DWORD [104+ebx] |
| movq mm1,[104+ebx] |
| psllq mm7,60 |
| movq mm4,[96+ebx] |
| por mm0,mm7 |
| movq [edi-24],mm2 |
| psrlq mm2,4 |
| movq [104+edi],mm5 |
| movq mm6,mm5 |
| movq [ebp-32],mm0 |
| psrlq mm5,4 |
| movq [96+ebp],mm3 |
| shl edx,4 |
| mov BYTE [14+esp],dl |
| mov edx,DWORD [120+ebx] |
| movq mm0,[120+ebx] |
| psllq mm6,60 |
| movq mm3,[112+ebx] |
| por mm2,mm6 |
| movq [edi-16],mm1 |
| psrlq mm1,4 |
| movq [112+edi],mm4 |
| movq mm7,mm4 |
| movq [ebp-24],mm2 |
| psrlq mm4,4 |
| movq [104+ebp],mm5 |
| shl edx,4 |
| mov BYTE [15+esp],dl |
| psllq mm7,60 |
| por mm1,mm7 |
| movq [edi-8],mm0 |
| psrlq mm0,4 |
| movq [120+edi],mm3 |
| movq mm6,mm3 |
| movq [ebp-16],mm1 |
| psrlq mm3,4 |
| movq [112+ebp],mm4 |
| psllq mm6,60 |
| por mm0,mm6 |
| movq [ebp-8],mm0 |
| movq [120+ebp],mm3 |
| movq mm6,[eax] |
| mov ebx,DWORD [8+eax] |
| mov edx,DWORD [12+eax] |
| align 16 |
| L$004outer: |
| xor edx,DWORD [12+ecx] |
| xor ebx,DWORD [8+ecx] |
| pxor mm6,[ecx] |
| lea ecx,[16+ecx] |
| mov DWORD [536+esp],ebx |
| movq [528+esp],mm6 |
| mov DWORD [548+esp],ecx |
| xor eax,eax |
| rol edx,8 |
| mov al,dl |
| mov ebp,eax |
| and al,15 |
| shr ebp,4 |
| pxor mm0,mm0 |
| rol edx,8 |
| pxor mm1,mm1 |
| pxor mm2,mm2 |
| movq mm7,[16+eax*8+esp] |
| movq mm6,[144+eax*8+esp] |
| mov al,dl |
| movd ebx,mm7 |
| psrlq mm7,8 |
| movq mm3,mm6 |
| mov edi,eax |
| psrlq mm6,8 |
| pxor mm7,[272+ebp*8+esp] |
| and al,15 |
| psllq mm3,56 |
| shr edi,4 |
| pxor mm7,[16+eax*8+esp] |
| rol edx,8 |
| pxor mm6,[144+eax*8+esp] |
| pxor mm7,mm3 |
| pxor mm6,[400+ebp*8+esp] |
| xor bl,BYTE [ebp*1+esp] |
| mov al,dl |
| movd ecx,mm7 |
| movzx ebx,bl |
| psrlq mm7,8 |
| movq mm3,mm6 |
| mov ebp,eax |
| psrlq mm6,8 |
| pxor mm7,[272+edi*8+esp] |
| and al,15 |
| psllq mm3,56 |
| shr ebp,4 |
| pinsrw mm2,WORD [ebx*2+esi],2 |
| pxor mm7,[16+eax*8+esp] |
| rol edx,8 |
| pxor mm6,[144+eax*8+esp] |
| pxor mm7,mm3 |
| pxor mm6,[400+edi*8+esp] |
| xor cl,BYTE [edi*1+esp] |
| mov al,dl |
| mov edx,DWORD [536+esp] |
| movd ebx,mm7 |
| movzx ecx,cl |
| psrlq mm7,8 |
| movq mm3,mm6 |
| mov edi,eax |
| psrlq mm6,8 |
| pxor mm7,[272+ebp*8+esp] |
| and al,15 |
| psllq mm3,56 |
| pxor mm6,mm2 |
| shr edi,4 |
| pinsrw mm1,WORD [ecx*2+esi],2 |
| pxor mm7,[16+eax*8+esp] |
| rol edx,8 |
| pxor mm6,[144+eax*8+esp] |
| pxor mm7,mm3 |
| pxor mm6,[400+ebp*8+esp] |
| xor bl,BYTE [ebp*1+esp] |
| mov al,dl |
| movd ecx,mm7 |
| movzx ebx,bl |
| psrlq mm7,8 |
| movq mm3,mm6 |
| mov ebp,eax |
| psrlq mm6,8 |
| pxor mm7,[272+edi*8+esp] |
| and al,15 |
| psllq mm3,56 |
| pxor mm6,mm1 |
| shr ebp,4 |
| pinsrw mm0,WORD [ebx*2+esi],2 |
| pxor mm7,[16+eax*8+esp] |
| rol edx,8 |
| pxor mm6,[144+eax*8+esp] |
| pxor mm7,mm3 |
| pxor mm6,[400+edi*8+esp] |
| xor cl,BYTE [edi*1+esp] |
| mov al,dl |
| movd ebx,mm7 |
| movzx ecx,cl |
| psrlq mm7,8 |
| movq mm3,mm6 |
| mov edi,eax |
| psrlq mm6,8 |
| pxor mm7,[272+ebp*8+esp] |
| and al,15 |
| psllq mm3,56 |
| pxor mm6,mm0 |
| shr edi,4 |
| pinsrw mm2,WORD [ecx*2+esi],2 |
| pxor mm7,[16+eax*8+esp] |
| rol edx,8 |
| pxor mm6,[144+eax*8+esp] |
| pxor mm7,mm3 |
| pxor mm6,[400+ebp*8+esp] |
| xor bl,BYTE [ebp*1+esp] |
| mov al,dl |
| movd ecx,mm7 |
| movzx ebx,bl |
| psrlq mm7,8 |
| movq mm3,mm6 |
| mov ebp,eax |
| psrlq mm6,8 |
| pxor mm7,[272+edi*8+esp] |
| and al,15 |
| psllq mm3,56 |
| pxor mm6,mm2 |
| shr ebp,4 |
| pinsrw mm1,WORD [ebx*2+esi],2 |
| pxor mm7,[16+eax*8+esp] |
| rol edx,8 |
| pxor mm6,[144+eax*8+esp] |
| pxor mm7,mm3 |
| pxor mm6,[400+edi*8+esp] |
| xor cl,BYTE [edi*1+esp] |
| mov al,dl |
| mov edx,DWORD [532+esp] |
| movd ebx,mm7 |
| movzx ecx,cl |
| psrlq mm7,8 |
| movq mm3,mm6 |
| mov edi,eax |
| psrlq mm6,8 |
| pxor mm7,[272+ebp*8+esp] |
| and al,15 |
| psllq mm3,56 |
| pxor mm6,mm1 |
| shr edi,4 |
| pinsrw mm0,WORD [ecx*2+esi],2 |
| pxor mm7,[16+eax*8+esp] |
| rol edx,8 |
| pxor mm6,[144+eax*8+esp] |
| pxor mm7,mm3 |
| pxor mm6,[400+ebp*8+esp] |
| xor bl,BYTE [ebp*1+esp] |
| mov al,dl |
| movd ecx,mm7 |
| movzx ebx,bl |
| psrlq mm7,8 |
| movq mm3,mm6 |
| mov ebp,eax |
| psrlq mm6,8 |
| pxor mm7,[272+edi*8+esp] |
| and al,15 |
| psllq mm3,56 |
| pxor mm6,mm0 |
| shr ebp,4 |
| pinsrw mm2,WORD [ebx*2+esi],2 |
| pxor mm7,[16+eax*8+esp] |
| rol edx,8 |
| pxor mm6,[144+eax*8+esp] |
| pxor mm7,mm3 |
| pxor mm6,[400+edi*8+esp] |
| xor cl,BYTE [edi*1+esp] |
| mov al,dl |
| movd ebx,mm7 |
| movzx ecx,cl |
| psrlq mm7,8 |
| movq mm3,mm6 |
| mov edi,eax |
| psrlq mm6,8 |
| pxor mm7,[272+ebp*8+esp] |
| and al,15 |
| psllq mm3,56 |
| pxor mm6,mm2 |
| shr edi,4 |
| pinsrw mm1,WORD [ecx*2+esi],2 |
| pxor mm7,[16+eax*8+esp] |
| rol edx,8 |
| pxor mm6,[144+eax*8+esp] |
| pxor mm7,mm3 |
| pxor mm6,[400+ebp*8+esp] |
| xor bl,BYTE [ebp*1+esp] |
| mov al,dl |
| movd ecx,mm7 |
| movzx ebx,bl |
| psrlq mm7,8 |
| movq mm3,mm6 |
| mov ebp,eax |
| psrlq mm6,8 |
| pxor mm7,[272+edi*8+esp] |
| and al,15 |
| psllq mm3,56 |
| pxor mm6,mm1 |
| shr ebp,4 |
| pinsrw mm0,WORD [ebx*2+esi],2 |
| pxor mm7,[16+eax*8+esp] |
| rol edx,8 |
| pxor mm6,[144+eax*8+esp] |
| pxor mm7,mm3 |
| pxor mm6,[400+edi*8+esp] |
| xor cl,BYTE [edi*1+esp] |
| mov al,dl |
| mov edx,DWORD [528+esp] |
| movd ebx,mm7 |
| movzx ecx,cl |
| psrlq mm7,8 |
| movq mm3,mm6 |
| mov edi,eax |
| psrlq mm6,8 |
| pxor mm7,[272+ebp*8+esp] |
| and al,15 |
| psllq mm3,56 |
| pxor mm6,mm0 |
| shr edi,4 |
| pinsrw mm2,WORD [ecx*2+esi],2 |
| pxor mm7,[16+eax*8+esp] |
| rol edx,8 |
| pxor mm6,[144+eax*8+esp] |
| pxor mm7,mm3 |
| pxor mm6,[400+ebp*8+esp] |
| xor bl,BYTE [ebp*1+esp] |
| mov al,dl |
| movd ecx,mm7 |
| movzx ebx,bl |
| psrlq mm7,8 |
| movq mm3,mm6 |
| mov ebp,eax |
| psrlq mm6,8 |
| pxor mm7,[272+edi*8+esp] |
| and al,15 |
| psllq mm3,56 |
| pxor mm6,mm2 |
| shr ebp,4 |
| pinsrw mm1,WORD [ebx*2+esi],2 |
| pxor mm7,[16+eax*8+esp] |
| rol edx,8 |
| pxor mm6,[144+eax*8+esp] |
| pxor mm7,mm3 |
| pxor mm6,[400+edi*8+esp] |
| xor cl,BYTE [edi*1+esp] |
| mov al,dl |
| movd ebx,mm7 |
| movzx ecx,cl |
| psrlq mm7,8 |
| movq mm3,mm6 |
| mov edi,eax |
| psrlq mm6,8 |
| pxor mm7,[272+ebp*8+esp] |
| and al,15 |
| psllq mm3,56 |
| pxor mm6,mm1 |
| shr edi,4 |
| pinsrw mm0,WORD [ecx*2+esi],2 |
| pxor mm7,[16+eax*8+esp] |
| rol edx,8 |
| pxor mm6,[144+eax*8+esp] |
| pxor mm7,mm3 |
| pxor mm6,[400+ebp*8+esp] |
| xor bl,BYTE [ebp*1+esp] |
| mov al,dl |
| movd ecx,mm7 |
| movzx ebx,bl |
| psrlq mm7,8 |
| movq mm3,mm6 |
| mov ebp,eax |
| psrlq mm6,8 |
| pxor mm7,[272+edi*8+esp] |
| and al,15 |
| psllq mm3,56 |
| pxor mm6,mm0 |
| shr ebp,4 |
| pinsrw mm2,WORD [ebx*2+esi],2 |
| pxor mm7,[16+eax*8+esp] |
| rol edx,8 |
| pxor mm6,[144+eax*8+esp] |
| pxor mm7,mm3 |
| pxor mm6,[400+edi*8+esp] |
| xor cl,BYTE [edi*1+esp] |
| mov al,dl |
| mov edx,DWORD [524+esp] |
| movd ebx,mm7 |
| movzx ecx,cl |
| psrlq mm7,8 |
| movq mm3,mm6 |
| mov edi,eax |
| psrlq mm6,8 |
| pxor mm7,[272+ebp*8+esp] |
| and al,15 |
| psllq mm3,56 |
| pxor mm6,mm2 |
| shr edi,4 |
| pinsrw mm1,WORD [ecx*2+esi],2 |
| pxor mm7,[16+eax*8+esp] |
| pxor mm6,[144+eax*8+esp] |
| xor bl,BYTE [ebp*1+esp] |
| pxor mm7,mm3 |
| pxor mm6,[400+ebp*8+esp] |
| movzx ebx,bl |
| pxor mm2,mm2 |
| psllq mm1,4 |
| movd ecx,mm7 |
| psrlq mm7,4 |
| movq mm3,mm6 |
| psrlq mm6,4 |
| shl ecx,4 |
| pxor mm7,[16+edi*8+esp] |
| psllq mm3,60 |
| movzx ecx,cl |
| pxor mm7,mm3 |
| pxor mm6,[144+edi*8+esp] |
| pinsrw mm0,WORD [ebx*2+esi],2 |
| pxor mm6,mm1 |
| movd edx,mm7 |
| pinsrw mm2,WORD [ecx*2+esi],3 |
| psllq mm0,12 |
| pxor mm6,mm0 |
| psrlq mm7,32 |
| pxor mm6,mm2 |
| mov ecx,DWORD [548+esp] |
| movd ebx,mm7 |
| movq mm3,mm6 |
| psllw mm6,8 |
| psrlw mm3,8 |
| por mm6,mm3 |
| bswap edx |
| pshufw mm6,mm6,27 |
| bswap ebx |
| cmp ecx,DWORD [552+esp] |
| jne NEAR L$004outer |
| mov eax,DWORD [544+esp] |
| mov DWORD [12+eax],edx |
| mov DWORD [8+eax],ebx |
| movq [eax],mm6 |
| mov esp,DWORD [556+esp] |
| emms |
| pop edi |
| pop esi |
| pop ebx |
| pop ebp |
| ret |
| global _gcm_init_clmul |
| align 16 |
| _gcm_init_clmul: |
| L$_gcm_init_clmul_begin: |
| mov edx,DWORD [4+esp] |
| mov eax,DWORD [8+esp] |
| call L$005pic |
| L$005pic: |
| pop ecx |
| lea ecx,[(L$bswap-L$005pic)+ecx] |
| movdqu xmm2,[eax] |
| pshufd xmm2,xmm2,78 |
| pshufd xmm4,xmm2,255 |
| movdqa xmm3,xmm2 |
| psllq xmm2,1 |
| pxor xmm5,xmm5 |
| psrlq xmm3,63 |
| pcmpgtd xmm5,xmm4 |
| pslldq xmm3,8 |
| por xmm2,xmm3 |
| pand xmm5,[16+ecx] |
| pxor xmm2,xmm5 |
| movdqa xmm0,xmm2 |
| movdqa xmm1,xmm0 |
| pshufd xmm3,xmm0,78 |
| pshufd xmm4,xmm2,78 |
| pxor xmm3,xmm0 |
| pxor xmm4,xmm2 |
| db 102,15,58,68,194,0 |
| db 102,15,58,68,202,17 |
| db 102,15,58,68,220,0 |
| xorps xmm3,xmm0 |
| xorps xmm3,xmm1 |
| movdqa xmm4,xmm3 |
| psrldq xmm3,8 |
| pslldq xmm4,8 |
| pxor xmm1,xmm3 |
| pxor xmm0,xmm4 |
| movdqa xmm4,xmm0 |
| movdqa xmm3,xmm0 |
| psllq xmm0,5 |
| pxor xmm3,xmm0 |
| psllq xmm0,1 |
| pxor xmm0,xmm3 |
| psllq xmm0,57 |
| movdqa xmm3,xmm0 |
| pslldq xmm0,8 |
| psrldq xmm3,8 |
| pxor xmm0,xmm4 |
| pxor xmm1,xmm3 |
| movdqa xmm4,xmm0 |
| psrlq xmm0,1 |
| pxor xmm1,xmm4 |
| pxor xmm4,xmm0 |
| psrlq xmm0,5 |
| pxor xmm0,xmm4 |
| psrlq xmm0,1 |
| pxor xmm0,xmm1 |
| pshufd xmm3,xmm2,78 |
| pshufd xmm4,xmm0,78 |
| pxor xmm3,xmm2 |
| movdqu [edx],xmm2 |
| pxor xmm4,xmm0 |
| movdqu [16+edx],xmm0 |
| db 102,15,58,15,227,8 |
| movdqu [32+edx],xmm4 |
| ret |
| global _gcm_gmult_clmul |
| align 16 |
| _gcm_gmult_clmul: |
| L$_gcm_gmult_clmul_begin: |
| mov eax,DWORD [4+esp] |
| mov edx,DWORD [8+esp] |
| call L$006pic |
| L$006pic: |
| pop ecx |
| lea ecx,[(L$bswap-L$006pic)+ecx] |
| movdqu xmm0,[eax] |
| movdqa xmm5,[ecx] |
| movups xmm2,[edx] |
| db 102,15,56,0,197 |
| movups xmm4,[32+edx] |
| movdqa xmm1,xmm0 |
| pshufd xmm3,xmm0,78 |
| pxor xmm3,xmm0 |
| db 102,15,58,68,194,0 |
| db 102,15,58,68,202,17 |
| db 102,15,58,68,220,0 |
| xorps xmm3,xmm0 |
| xorps xmm3,xmm1 |
| movdqa xmm4,xmm3 |
| psrldq xmm3,8 |
| pslldq xmm4,8 |
| pxor xmm1,xmm3 |
| pxor xmm0,xmm4 |
| movdqa xmm4,xmm0 |
| movdqa xmm3,xmm0 |
| psllq xmm0,5 |
| pxor xmm3,xmm0 |
| psllq xmm0,1 |
| pxor xmm0,xmm3 |
| psllq xmm0,57 |
| movdqa xmm3,xmm0 |
| pslldq xmm0,8 |
| psrldq xmm3,8 |
| pxor xmm0,xmm4 |
| pxor xmm1,xmm3 |
| movdqa xmm4,xmm0 |
| psrlq xmm0,1 |
| pxor xmm1,xmm4 |
| pxor xmm4,xmm0 |
| psrlq xmm0,5 |
| pxor xmm0,xmm4 |
| psrlq xmm0,1 |
| pxor xmm0,xmm1 |
| db 102,15,56,0,197 |
| movdqu [eax],xmm0 |
| ret |
| global _gcm_ghash_clmul |
| align 16 |
| _gcm_ghash_clmul: |
| L$_gcm_ghash_clmul_begin: |
| push ebp |
| push ebx |
| push esi |
| push edi |
| mov eax,DWORD [20+esp] |
| mov edx,DWORD [24+esp] |
| mov esi,DWORD [28+esp] |
| mov ebx,DWORD [32+esp] |
| call L$007pic |
| L$007pic: |
| pop ecx |
| lea ecx,[(L$bswap-L$007pic)+ecx] |
| movdqu xmm0,[eax] |
| movdqa xmm5,[ecx] |
| movdqu xmm2,[edx] |
| db 102,15,56,0,197 |
| sub ebx,16 |
| jz NEAR L$008odd_tail |
| movdqu xmm3,[esi] |
| movdqu xmm6,[16+esi] |
| db 102,15,56,0,221 |
| db 102,15,56,0,245 |
| movdqu xmm5,[32+edx] |
| pxor xmm0,xmm3 |
| pshufd xmm3,xmm6,78 |
| movdqa xmm7,xmm6 |
| pxor xmm3,xmm6 |
| lea esi,[32+esi] |
| db 102,15,58,68,242,0 |
| db 102,15,58,68,250,17 |
| db 102,15,58,68,221,0 |
| movups xmm2,[16+edx] |
| nop |
| sub ebx,32 |
| jbe NEAR L$009even_tail |
| jmp NEAR L$010mod_loop |
| align 32 |
| L$010mod_loop: |
| pshufd xmm4,xmm0,78 |
| movdqa xmm1,xmm0 |
| pxor xmm4,xmm0 |
| nop |
| db 102,15,58,68,194,0 |
| db 102,15,58,68,202,17 |
| db 102,15,58,68,229,16 |
| movups xmm2,[edx] |
| xorps xmm0,xmm6 |
| movdqa xmm5,[ecx] |
| xorps xmm1,xmm7 |
| movdqu xmm7,[esi] |
| pxor xmm3,xmm0 |
| movdqu xmm6,[16+esi] |
| pxor xmm3,xmm1 |
| db 102,15,56,0,253 |
| pxor xmm4,xmm3 |
| movdqa xmm3,xmm4 |
| psrldq xmm4,8 |
| pslldq xmm3,8 |
| pxor xmm1,xmm4 |
| pxor xmm0,xmm3 |
| db 102,15,56,0,245 |
| pxor xmm1,xmm7 |
| movdqa xmm7,xmm6 |
| movdqa xmm4,xmm0 |
| movdqa xmm3,xmm0 |
| psllq xmm0,5 |
| pxor xmm3,xmm0 |
| psllq xmm0,1 |
| pxor xmm0,xmm3 |
| db 102,15,58,68,242,0 |
| movups xmm5,[32+edx] |
| psllq xmm0,57 |
| movdqa xmm3,xmm0 |
| pslldq xmm0,8 |
| psrldq xmm3,8 |
| pxor xmm0,xmm4 |
| pxor xmm1,xmm3 |
| pshufd xmm3,xmm7,78 |
| movdqa xmm4,xmm0 |
| psrlq xmm0,1 |
| pxor xmm3,xmm7 |
| pxor xmm1,xmm4 |
| db 102,15,58,68,250,17 |
| movups xmm2,[16+edx] |
| pxor xmm4,xmm0 |
| psrlq xmm0,5 |
| pxor xmm0,xmm4 |
| psrlq xmm0,1 |
| pxor xmm0,xmm1 |
| db 102,15,58,68,221,0 |
| lea esi,[32+esi] |
| sub ebx,32 |
| ja NEAR L$010mod_loop |
| L$009even_tail: |
| pshufd xmm4,xmm0,78 |
| movdqa xmm1,xmm0 |
| pxor xmm4,xmm0 |
| db 102,15,58,68,194,0 |
| db 102,15,58,68,202,17 |
| db 102,15,58,68,229,16 |
| movdqa xmm5,[ecx] |
| xorps xmm0,xmm6 |
| xorps xmm1,xmm7 |
| pxor xmm3,xmm0 |
| pxor xmm3,xmm1 |
| pxor xmm4,xmm3 |
| movdqa xmm3,xmm4 |
| psrldq xmm4,8 |
| pslldq xmm3,8 |
| pxor xmm1,xmm4 |
| pxor xmm0,xmm3 |
| movdqa xmm4,xmm0 |
| movdqa xmm3,xmm0 |
| psllq xmm0,5 |
| pxor xmm3,xmm0 |
| psllq xmm0,1 |
| pxor xmm0,xmm3 |
| psllq xmm0,57 |
| movdqa xmm3,xmm0 |
| pslldq xmm0,8 |
| psrldq xmm3,8 |
| pxor xmm0,xmm4 |
| pxor xmm1,xmm3 |
| movdqa xmm4,xmm0 |
| psrlq xmm0,1 |
| pxor xmm1,xmm4 |
| pxor xmm4,xmm0 |
| psrlq xmm0,5 |
| pxor xmm0,xmm4 |
| psrlq xmm0,1 |
| pxor xmm0,xmm1 |
| test ebx,ebx |
| jnz NEAR L$011done |
| movups xmm2,[edx] |
| L$008odd_tail: |
| movdqu xmm3,[esi] |
| db 102,15,56,0,221 |
| pxor xmm0,xmm3 |
| movdqa xmm1,xmm0 |
| pshufd xmm3,xmm0,78 |
| pshufd xmm4,xmm2,78 |
| pxor xmm3,xmm0 |
| pxor xmm4,xmm2 |
| db 102,15,58,68,194,0 |
| db 102,15,58,68,202,17 |
| db 102,15,58,68,220,0 |
| xorps xmm3,xmm0 |
| xorps xmm3,xmm1 |
| movdqa xmm4,xmm3 |
| psrldq xmm3,8 |
| pslldq xmm4,8 |
| pxor xmm1,xmm3 |
| pxor xmm0,xmm4 |
| movdqa xmm4,xmm0 |
| movdqa xmm3,xmm0 |
| psllq xmm0,5 |
| pxor xmm3,xmm0 |
| psllq xmm0,1 |
| pxor xmm0,xmm3 |
| psllq xmm0,57 |
| movdqa xmm3,xmm0 |
| pslldq xmm0,8 |
| psrldq xmm3,8 |
| pxor xmm0,xmm4 |
| pxor xmm1,xmm3 |
| movdqa xmm4,xmm0 |
| psrlq xmm0,1 |
| pxor xmm1,xmm4 |
| pxor xmm4,xmm0 |
| psrlq xmm0,5 |
| pxor xmm0,xmm4 |
| psrlq xmm0,1 |
| pxor xmm0,xmm1 |
| L$011done: |
| db 102,15,56,0,197 |
| movdqu [eax],xmm0 |
| pop edi |
| pop esi |
| pop ebx |
| pop ebp |
| ret |
| align 64 |
| L$bswap: |
| db 15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0 |
| db 1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,194 |
| align 64 |
| L$rem_8bit: |
| dw 0,450,900,582,1800,1738,1164,1358 |
| dw 3600,4050,3476,3158,2328,2266,2716,2910 |
| dw 7200,7650,8100,7782,6952,6890,6316,6510 |
| dw 4656,5106,4532,4214,5432,5370,5820,6014 |
| dw 14400,14722,15300,14854,16200,16010,15564,15630 |
| dw 13904,14226,13780,13334,12632,12442,13020,13086 |
| dw 9312,9634,10212,9766,9064,8874,8428,8494 |
| dw 10864,11186,10740,10294,11640,11450,12028,12094 |
| dw 28800,28994,29444,29382,30600,30282,29708,30158 |
| dw 32400,32594,32020,31958,31128,30810,31260,31710 |
| dw 27808,28002,28452,28390,27560,27242,26668,27118 |
| dw 25264,25458,24884,24822,26040,25722,26172,26622 |
| dw 18624,18690,19268,19078,20424,19978,19532,19854 |
| dw 18128,18194,17748,17558,16856,16410,16988,17310 |
| dw 21728,21794,22372,22182,21480,21034,20588,20910 |
| dw 23280,23346,22900,22710,24056,23610,24188,24510 |
| dw 57600,57538,57988,58182,58888,59338,58764,58446 |
| dw 61200,61138,60564,60758,59416,59866,60316,59998 |
| dw 64800,64738,65188,65382,64040,64490,63916,63598 |
| dw 62256,62194,61620,61814,62520,62970,63420,63102 |
| dw 55616,55426,56004,56070,56904,57226,56780,56334 |
| dw 55120,54930,54484,54550,53336,53658,54236,53790 |
| dw 50528,50338,50916,50982,49768,50090,49644,49198 |
| dw 52080,51890,51444,51510,52344,52666,53244,52798 |
| dw 37248,36930,37380,37830,38536,38730,38156,38094 |
| dw 40848,40530,39956,40406,39064,39258,39708,39646 |
| dw 36256,35938,36388,36838,35496,35690,35116,35054 |
| dw 33712,33394,32820,33270,33976,34170,34620,34558 |
| dw 43456,43010,43588,43910,44744,44810,44364,44174 |
| dw 42960,42514,42068,42390,41176,41242,41820,41630 |
| dw 46560,46114,46692,47014,45800,45866,45420,45230 |
| dw 48112,47666,47220,47542,48376,48442,49020,48830 |
| align 64 |
| L$rem_4bit: |
| dd 0,0,0,471859200,0,943718400,0,610271232 |
| dd 0,1887436800,0,1822425088,0,1220542464,0,1423966208 |
| dd 0,3774873600,0,4246732800,0,3644850176,0,3311403008 |
| dd 0,2441084928,0,2376073216,0,2847932416,0,3051356160 |
| db 71,72,65,83,72,32,102,111,114,32,120,56,54,44,32,67 |
| db 82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112 |
| db 112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62 |
| db 0 |