From 626bebeedef59448117e70d62d78ea6b5dd046c8 Mon Sep 17 00:00:00 2001 From: "Dr. Stephen Henson" Date: Mon, 23 Jul 2007 11:54:47 +0000 Subject: [PATCH] Delete obsolete files. --- fips-1.0/aes/asm/fips-ax86-elf.s | 1705 ------------------- fips-1.0/des/asm/fips-dx86-elf.s | 2707 ------------------------------ fips-1.0/sha/asm/fips-sx86-elf.s | 1568 ----------------- 3 files changed, 5980 deletions(-) delete mode 100644 fips-1.0/aes/asm/fips-ax86-elf.s delete mode 100644 fips-1.0/des/asm/fips-dx86-elf.s delete mode 100644 fips-1.0/sha/asm/fips-sx86-elf.s diff --git a/fips-1.0/aes/asm/fips-ax86-elf.s b/fips-1.0/aes/asm/fips-ax86-elf.s deleted file mode 100644 index fdfb6b978c..0000000000 --- a/fips-1.0/aes/asm/fips-ax86-elf.s +++ /dev/null @@ -1,1705 +0,0 @@ - - - - - - - .file "aes-586.s" -.globl AES_Te -.text -.globl _x86_AES_encrypt -.type _x86_AES_encrypt,@function -.align 16 -_x86_AES_encrypt: - movl %edi, 12(%esp) - xorl (%edi), %eax - xorl 4(%edi), %ebx - xorl 8(%edi), %ecx - xorl 12(%edi), %edx - movl 240(%edi), %esi - leal -2(%esi,%esi), %esi - leal (%edi,%esi,8), %esi - movl %esi, 16(%esp) -.align 4 -.L000loop: - movl %eax, %esi - andl $255, %esi - movl (%ebp,%esi,8), %esi - movzbl %bh, %edi - xorl 3(%ebp,%edi,8), %esi - movl %ecx, %edi - shrl $16, %edi - andl $255, %edi - xorl 2(%ebp,%edi,8), %esi - movl %edx, %edi - shrl $24, %edi - xorl 1(%ebp,%edi,8), %esi - movl %esi, 4(%esp) - - movl %ebx, %esi - andl $255, %esi - shrl $16, %ebx - movl (%ebp,%esi,8), %esi - movzbl %ch, %edi - xorl 3(%ebp,%edi,8), %esi - movl %edx, %edi - shrl $16, %edi - andl $255, %edi - xorl 2(%ebp,%edi,8), %esi - movl %eax, %edi - shrl $24, %edi - xorl 1(%ebp,%edi,8), %esi - movl %esi, 8(%esp) - - movl %ecx, %esi - andl $255, %esi - shrl $24, %ecx - movl (%ebp,%esi,8), %esi - movzbl %dh, %edi - xorl 3(%ebp,%edi,8), %esi - movl %eax, %edi - shrl $16, %edi - andl $255, %edx - andl $255, %edi - xorl 2(%ebp,%edi,8), %esi - movzbl %bh, %edi - xorl 1(%ebp,%edi,8), %esi - - movl 12(%esp), %edi - movl (%ebp,%edx,8), %edx - movzbl %ah, %eax - xorl 3(%ebp,%eax,8), %edx - movl 4(%esp), %eax - andl $255, %ebx - xorl 2(%ebp,%ebx,8), %edx - movl 8(%esp), %ebx - xorl 1(%ebp,%ecx,8), %edx - movl %esi, %ecx - - addl $16, %edi - xorl (%edi), %eax - xorl 4(%edi), %ebx - xorl 8(%edi), %ecx - xorl 12(%edi), %edx - cmpl 16(%esp), %edi - movl %edi, 12(%esp) - jb .L000loop - movl %eax, %esi - andl $255, %esi - movl 2(%ebp,%esi,8), %esi - andl $255, %esi - movzbl %bh, %edi - movl (%ebp,%edi,8), %edi - andl $65280, %edi - xorl %edi, %esi - movl %ecx, %edi - shrl $16, %edi - andl $255, %edi - movl (%ebp,%edi,8), %edi - andl $16711680, %edi - xorl %edi, %esi - movl %edx, %edi - shrl $24, %edi - movl 2(%ebp,%edi,8), %edi - andl $4278190080, %edi - xorl %edi, %esi - movl %esi, 4(%esp) - movl %ebx, %esi - andl $255, %esi - shrl $16, %ebx - movl 2(%ebp,%esi,8), %esi - andl $255, %esi - movzbl %ch, %edi - movl (%ebp,%edi,8), %edi - andl $65280, %edi - xorl %edi, %esi - movl %edx, %edi - shrl $16, %edi - andl $255, %edi - movl (%ebp,%edi,8), %edi - andl $16711680, %edi - xorl %edi, %esi - movl %eax, %edi - shrl $24, %edi - movl 2(%ebp,%edi,8), %edi - andl $4278190080, %edi - xorl %edi, %esi - movl %esi, 8(%esp) - movl %ecx, %esi - andl $255, %esi - shrl $24, %ecx - movl 2(%ebp,%esi,8), %esi - andl $255, %esi - movzbl %dh, %edi - movl (%ebp,%edi,8), %edi - andl $65280, %edi - xorl %edi, %esi - movl %eax, %edi - shrl $16, %edi - andl $255, %edx - andl $255, %edi - movl (%ebp,%edi,8), %edi - andl $16711680, %edi - xorl %edi, %esi - movzbl %bh, %edi - movl 2(%ebp,%edi,8), %edi - andl $4278190080, %edi - xorl %edi, %esi - movl 12(%esp), %edi - andl $255, %edx - movl 2(%ebp,%edx,8), %edx - andl $255, %edx - movzbl %ah, %eax - movl (%ebp,%eax,8), %eax - andl $65280, %eax - xorl %eax, %edx - movl 4(%esp), %eax - andl $255, %ebx - movl (%ebp,%ebx,8), %ebx - andl $16711680, %ebx - xorl %ebx, %edx - movl 8(%esp), %ebx - movl 2(%ebp,%ecx,8), %ecx - andl $4278190080, %ecx - xorl %ecx, %edx - movl %esi, %ecx - addl $16, %edi - xorl (%edi), %eax - xorl 4(%edi), %ebx - xorl 8(%edi), %ecx - xorl 12(%edi), %edx - ret -.align 64 -AES_Te: - .long 2774754246,2774754246 - .long 2222750968,2222750968 - .long 2574743534,2574743534 - .long 2373680118,2373680118 - .long 234025727,234025727 - .long 3177933782,3177933782 - .long 2976870366,2976870366 - .long 1422247313,1422247313 - .long 1345335392,1345335392 - .long 50397442,50397442 - .long 2842126286,2842126286 - .long 2099981142,2099981142 - .long 436141799,436141799 - .long 1658312629,1658312629 - .long 3870010189,3870010189 - .long 2591454956,2591454956 - .long 1170918031,1170918031 - .long 2642575903,2642575903 - .long 1086966153,1086966153 - .long 2273148410,2273148410 - .long 368769775,368769775 - .long 3948501426,3948501426 - .long 3376891790,3376891790 - .long 200339707,200339707 - .long 3970805057,3970805057 - .long 1742001331,1742001331 - .long 4255294047,4255294047 - .long 3937382213,3937382213 - .long 3214711843,3214711843 - .long 4154762323,4154762323 - .long 2524082916,2524082916 - .long 1539358875,1539358875 - .long 3266819957,3266819957 - .long 486407649,486407649 - .long 2928907069,2928907069 - .long 1780885068,1780885068 - .long 1513502316,1513502316 - .long 1094664062,1094664062 - .long 49805301,49805301 - .long 1338821763,1338821763 - .long 1546925160,1546925160 - .long 4104496465,4104496465 - .long 887481809,887481809 - .long 150073849,150073849 - .long 2473685474,2473685474 - .long 1943591083,1943591083 - .long 1395732834,1395732834 - .long 1058346282,1058346282 - .long 201589768,201589768 - .long 1388824469,1388824469 - .long 1696801606,1696801606 - .long 1589887901,1589887901 - .long 672667696,672667696 - .long 2711000631,2711000631 - .long 251987210,251987210 - .long 3046808111,3046808111 - .long 151455502,151455502 - .long 907153956,907153956 - .long 2608889883,2608889883 - .long 1038279391,1038279391 - .long 652995533,652995533 - .long 1764173646,1764173646 - .long 3451040383,3451040383 - .long 2675275242,2675275242 - .long 453576978,453576978 - .long 2659418909,2659418909 - .long 1949051992,1949051992 - .long 773462580,773462580 - .long 756751158,756751158 - .long 2993581788,2993581788 - .long 3998898868,3998898868 - .long 4221608027,4221608027 - .long 4132590244,4132590244 - .long 1295727478,1295727478 - .long 1641469623,1641469623 - .long 3467883389,3467883389 - .long 2066295122,2066295122 - .long 1055122397,1055122397 - .long 1898917726,1898917726 - .long 2542044179,2542044179 - .long 4115878822,4115878822 - .long 1758581177,1758581177 - .long 0,0 - .long 753790401,753790401 - .long 1612718144,1612718144 - .long 536673507,536673507 - .long 3367088505,3367088505 - .long 3982187446,3982187446 - .long 3194645204,3194645204 - .long 1187761037,1187761037 - .long 3653156455,3653156455 - .long 1262041458,1262041458 - .long 3729410708,3729410708 - .long 3561770136,3561770136 - .long 3898103984,3898103984 - .long 1255133061,1255133061 - .long 1808847035,1808847035 - .long 720367557,720367557 - .long 3853167183,3853167183 - .long 385612781,385612781 - .long 3309519750,3309519750 - .long 3612167578,3612167578 - .long 1429418854,1429418854 - .long 2491778321,2491778321 - .long 3477423498,3477423498 - .long 284817897,284817897 - .long 100794884,100794884 - .long 2172616702,2172616702 - .long 4031795360,4031795360 - .long 1144798328,1144798328 - .long 3131023141,3131023141 - .long 3819481163,3819481163 - .long 4082192802,4082192802 - .long 4272137053,4272137053 - .long 3225436288,3225436288 - .long 2324664069,2324664069 - .long 2912064063,2912064063 - .long 3164445985,3164445985 - .long 1211644016,1211644016 - .long 83228145,83228145 - .long 3753688163,3753688163 - .long 3249976951,3249976951 - .long 1977277103,1977277103 - .long 1663115586,1663115586 - .long 806359072,806359072 - .long 452984805,452984805 - .long 250868733,250868733 - .long 1842533055,1842533055 - .long 1288555905,1288555905 - .long 336333848,336333848 - .long 890442534,890442534 - .long 804056259,804056259 - .long 3781124030,3781124030 - .long 2727843637,2727843637 - .long 3427026056,3427026056 - .long 957814574,957814574 - .long 1472513171,1472513171 - .long 4071073621,4071073621 - .long 2189328124,2189328124 - .long 1195195770,1195195770 - .long 2892260552,2892260552 - .long 3881655738,3881655738 - .long 723065138,723065138 - .long 2507371494,2507371494 - .long 2690670784,2690670784 - .long 2558624025,2558624025 - .long 3511635870,3511635870 - .long 2145180835,2145180835 - .long 1713513028,1713513028 - .long 2116692564,2116692564 - .long 2878378043,2878378043 - .long 2206763019,2206763019 - .long 3393603212,3393603212 - .long 703524551,703524551 - .long 3552098411,3552098411 - .long 1007948840,1007948840 - .long 2044649127,2044649127 - .long 3797835452,3797835452 - .long 487262998,487262998 - .long 1994120109,1994120109 - .long 1004593371,1004593371 - .long 1446130276,1446130276 - .long 1312438900,1312438900 - .long 503974420,503974420 - .long 3679013266,3679013266 - .long 168166924,168166924 - .long 1814307912,1814307912 - .long 3831258296,3831258296 - .long 1573044895,1573044895 - .long 1859376061,1859376061 - .long 4021070915,4021070915 - .long 2791465668,2791465668 - .long 2828112185,2828112185 - .long 2761266481,2761266481 - .long 937747667,937747667 - .long 2339994098,2339994098 - .long 854058965,854058965 - .long 1137232011,1137232011 - .long 1496790894,1496790894 - .long 3077402074,3077402074 - .long 2358086913,2358086913 - .long 1691735473,1691735473 - .long 3528347292,3528347292 - .long 3769215305,3769215305 - .long 3027004632,3027004632 - .long 4199962284,4199962284 - .long 133494003,133494003 - .long 636152527,636152527 - .long 2942657994,2942657994 - .long 2390391540,2390391540 - .long 3920539207,3920539207 - .long 403179536,403179536 - .long 3585784431,3585784431 - .long 2289596656,2289596656 - .long 1864705354,1864705354 - .long 1915629148,1915629148 - .long 605822008,605822008 - .long 4054230615,4054230615 - .long 3350508659,3350508659 - .long 1371981463,1371981463 - .long 602466507,602466507 - .long 2094914977,2094914977 - .long 2624877800,2624877800 - .long 555687742,555687742 - .long 3712699286,3712699286 - .long 3703422305,3703422305 - .long 2257292045,2257292045 - .long 2240449039,2240449039 - .long 2423288032,2423288032 - .long 1111375484,1111375484 - .long 3300242801,3300242801 - .long 2858837708,2858837708 - .long 3628615824,3628615824 - .long 84083462,84083462 - .long 32962295,32962295 - .long 302911004,302911004 - .long 2741068226,2741068226 - .long 1597322602,1597322602 - .long 4183250862,4183250862 - .long 3501832553,3501832553 - .long 2441512471,2441512471 - .long 1489093017,1489093017 - .long 656219450,656219450 - .long 3114180135,3114180135 - .long 954327513,954327513 - .long 335083755,335083755 - .long 3013122091,3013122091 - .long 856756514,856756514 - .long 3144247762,3144247762 - .long 1893325225,1893325225 - .long 2307821063,2307821063 - .long 2811532339,2811532339 - .long 3063651117,3063651117 - .long 572399164,572399164 - .long 2458355477,2458355477 - .long 552200649,552200649 - .long 1238290055,1238290055 - .long 4283782570,4283782570 - .long 2015897680,2015897680 - .long 2061492133,2061492133 - .long 2408352771,2408352771 - .long 4171342169,4171342169 - .long 2156497161,2156497161 - .long 386731290,386731290 - .long 3669999461,3669999461 - .long 837215959,837215959 - .long 3326231172,3326231172 - .long 3093850320,3093850320 - .long 3275833730,3275833730 - .long 2962856233,2962856233 - .long 1999449434,1999449434 - .long 286199582,286199582 - .long 3417354363,3417354363 - .long 4233385128,4233385128 - .long 3602627437,3602627437 - .long 974525996,974525996 - .long 1,2,4,8 - .long 16,32,64,128 - .long 27,54,0,0,0,0,0,0 -.L__x86_AES_encrypt_end: -.size _x86_AES_encrypt,.L__x86_AES_encrypt_end-_x86_AES_encrypt -.ident "_x86_AES_encrypt" -.globl AES_Te -.text -.globl AES_encrypt -.type AES_encrypt,@function -.align 16 -AES_encrypt: - pushl %ebp - pushl %ebx - pushl %esi - pushl %edi - - movl 20(%esp), %esi - movl 28(%esp), %edi - movl %esp, %eax - subl $24, %esp - andl $-64, %esp - addl $4, %esp - movl %eax, 16(%esp) - call .L001pic_point -.L001pic_point: - popl %ebp - leal AES_Te-.L001pic_point(%ebp),%ebp - movl (%esi), %eax - movl 4(%esi), %ebx - movl 8(%esi), %ecx - movl 12(%esi), %edx - call _x86_AES_encrypt - movl 16(%esp), %esp - movl 24(%esp), %esi - movl %eax, (%esi) - movl %ebx, 4(%esi) - movl %ecx, 8(%esi) - movl %edx, 12(%esi) - popl %edi - popl %esi - popl %ebx - popl %ebp - ret -.L_AES_encrypt_end: -.size AES_encrypt,.L_AES_encrypt_end-AES_encrypt -.ident "AES_encrypt" -.globl AES_Td -.text -.globl _x86_AES_decrypt -.type _x86_AES_decrypt,@function -.align 16 -_x86_AES_decrypt: - movl %edi, 12(%esp) - xorl (%edi), %eax - xorl 4(%edi), %ebx - xorl 8(%edi), %ecx - xorl 12(%edi), %edx - movl 240(%edi), %esi - leal -2(%esi,%esi), %esi - leal (%edi,%esi,8), %esi - movl %esi, 16(%esp) -.align 4 -.L002loop: - movl %eax, %esi - andl $255, %esi - movl (%ebp,%esi,8), %esi - movzbl %dh, %edi - xorl 3(%ebp,%edi,8), %esi - movl %ecx, %edi - shrl $16, %edi - andl $255, %edi - xorl 2(%ebp,%edi,8), %esi - movl %ebx, %edi - shrl $24, %edi - xorl 1(%ebp,%edi,8), %esi - movl %esi, 4(%esp) - - movl %ebx, %esi - andl $255, %esi - movl (%ebp,%esi,8), %esi - movzbl %ah, %edi - xorl 3(%ebp,%edi,8), %esi - movl %edx, %edi - shrl $16, %edi - andl $255, %edi - xorl 2(%ebp,%edi,8), %esi - movl %ecx, %edi - shrl $24, %edi - xorl 1(%ebp,%edi,8), %esi - movl %esi, 8(%esp) - - movl %ecx, %esi - andl $255, %esi - movl (%ebp,%esi,8), %esi - movzbl %bh, %edi - xorl 3(%ebp,%edi,8), %esi - movl %eax, %edi - shrl $16, %edi - andl $255, %edi - xorl 2(%ebp,%edi,8), %esi - movl %edx, %edi - shrl $24, %edi - xorl 1(%ebp,%edi,8), %esi - - movl 12(%esp), %edi - andl $255, %edx - movl (%ebp,%edx,8), %edx - movzbl %ch, %ecx - xorl 3(%ebp,%ecx,8), %edx - movl %esi, %ecx - shrl $16, %ebx - andl $255, %ebx - xorl 2(%ebp,%ebx,8), %edx - movl 8(%esp), %ebx - shrl $24, %eax - xorl 1(%ebp,%eax,8), %edx - movl 4(%esp), %eax - - addl $16, %edi - xorl (%edi), %eax - xorl 4(%edi), %ebx - xorl 8(%edi), %ecx - xorl 12(%edi), %edx - cmpl 16(%esp), %edi - movl %edi, 12(%esp) - jb .L002loop - movl %eax, %esi - andl $255, %esi - movzbl 2048(%ebp,%esi,1),%esi - movzbl %dh, %edi - movzbl 2048(%ebp,%edi,1),%edi - sall $8, %edi - xorl %edi, %esi - movl %ecx, %edi - shrl $16, %edi - andl $255, %edi - movzbl 2048(%ebp,%edi,1),%edi - sall $16, %edi - xorl %edi, %esi - movl %ebx, %edi - shrl $24, %edi - movzbl 2048(%ebp,%edi,1),%edi - sall $24, %edi - xorl %edi, %esi - movl %esi, 4(%esp) - movl %ebx, %esi - andl $255, %esi - movzbl 2048(%ebp,%esi,1),%esi - movzbl %ah, %edi - movzbl 2048(%ebp,%edi,1),%edi - sall $8, %edi - xorl %edi, %esi - movl %edx, %edi - shrl $16, %edi - andl $255, %edi - movzbl 2048(%ebp,%edi,1),%edi - sall $16, %edi - xorl %edi, %esi - movl %ecx, %edi - shrl $24, %edi - movzbl 2048(%ebp,%edi,1),%edi - sall $24, %edi - xorl %edi, %esi - movl %esi, 8(%esp) - movl %ecx, %esi - andl $255, %esi - movzbl 2048(%ebp,%esi,1),%esi - movzbl %bh, %edi - movzbl 2048(%ebp,%edi,1),%edi - sall $8, %edi - xorl %edi, %esi - movl %eax, %edi - shrl $16, %edi - andl $255, %edi - movzbl 2048(%ebp,%edi,1),%edi - sall $16, %edi - xorl %edi, %esi - movl %edx, %edi - shrl $24, %edi - movzbl 2048(%ebp,%edi,1),%edi - sall $24, %edi - xorl %edi, %esi - movl 12(%esp), %edi - andl $255, %edx - movzbl 2048(%ebp,%edx,1),%edx - movzbl %ch, %ecx - movzbl 2048(%ebp,%ecx,1),%ecx - sall $8, %ecx - xorl %ecx, %edx - movl %esi, %ecx - shrl $16, %ebx - andl $255, %ebx - movzbl 2048(%ebp,%ebx,1),%ebx - sall $16, %ebx - xorl %ebx, %edx - movl 8(%esp), %ebx - shrl $24, %eax - movzbl 2048(%ebp,%eax,1),%eax - sall $24, %eax - xorl %eax, %edx - movl 4(%esp), %eax - addl $16, %edi - xorl (%edi), %eax - xorl 4(%edi), %ebx - xorl 8(%edi), %ecx - xorl 12(%edi), %edx - ret -.align 64 -AES_Td: - .long 1353184337,1353184337 - .long 1399144830,1399144830 - .long 3282310938,3282310938 - .long 2522752826,2522752826 - .long 3412831035,3412831035 - .long 4047871263,4047871263 - .long 2874735276,2874735276 - .long 2466505547,2466505547 - .long 1442459680,1442459680 - .long 4134368941,4134368941 - .long 2440481928,2440481928 - .long 625738485,625738485 - .long 4242007375,4242007375 - .long 3620416197,3620416197 - .long 2151953702,2151953702 - .long 2409849525,2409849525 - .long 1230680542,1230680542 - .long 1729870373,1729870373 - .long 2551114309,2551114309 - .long 3787521629,3787521629 - .long 41234371,41234371 - .long 317738113,317738113 - .long 2744600205,2744600205 - .long 3338261355,3338261355 - .long 3881799427,3881799427 - .long 2510066197,2510066197 - .long 3950669247,3950669247 - .long 3663286933,3663286933 - .long 763608788,763608788 - .long 3542185048,3542185048 - .long 694804553,694804553 - .long 1154009486,1154009486 - .long 1787413109,1787413109 - .long 2021232372,2021232372 - .long 1799248025,1799248025 - .long 3715217703,3715217703 - .long 3058688446,3058688446 - .long 397248752,397248752 - .long 1722556617,1722556617 - .long 3023752829,3023752829 - .long 407560035,407560035 - .long 2184256229,2184256229 - .long 1613975959,1613975959 - .long 1165972322,1165972322 - .long 3765920945,3765920945 - .long 2226023355,2226023355 - .long 480281086,480281086 - .long 2485848313,2485848313 - .long 1483229296,1483229296 - .long 436028815,436028815 - .long 2272059028,2272059028 - .long 3086515026,3086515026 - .long 601060267,601060267 - .long 3791801202,3791801202 - .long 1468997603,1468997603 - .long 715871590,715871590 - .long 120122290,120122290 - .long 63092015,63092015 - .long 2591802758,2591802758 - .long 2768779219,2768779219 - .long 4068943920,4068943920 - .long 2997206819,2997206819 - .long 3127509762,3127509762 - .long 1552029421,1552029421 - .long 723308426,723308426 - .long 2461301159,2461301159 - .long 4042393587,4042393587 - .long 2715969870,2715969870 - .long 3455375973,3455375973 - .long 3586000134,3586000134 - .long 526529745,526529745 - .long 2331944644,2331944644 - .long 2639474228,2639474228 - .long 2689987490,2689987490 - .long 853641733,853641733 - .long 1978398372,1978398372 - .long 971801355,971801355 - .long 2867814464,2867814464 - .long 111112542,111112542 - .long 1360031421,1360031421 - .long 4186579262,4186579262 - .long 1023860118,1023860118 - .long 2919579357,2919579357 - .long 1186850381,1186850381 - .long 3045938321,3045938321 - .long 90031217,90031217 - .long 1876166148,1876166148 - .long 4279586912,4279586912 - .long 620468249,620468249 - .long 2548678102,2548678102 - .long 3426959497,3426959497 - .long 2006899047,2006899047 - .long 3175278768,3175278768 - .long 2290845959,2290845959 - .long 945494503,945494503 - .long 3689859193,3689859193 - .long 1191869601,1191869601 - .long 3910091388,3910091388 - .long 3374220536,3374220536 - .long 0,0 - .long 2206629897,2206629897 - .long 1223502642,1223502642 - .long 2893025566,2893025566 - .long 1316117100,1316117100 - .long 4227796733,4227796733 - .long 1446544655,1446544655 - .long 517320253,517320253 - .long 658058550,658058550 - .long 1691946762,1691946762 - .long 564550760,564550760 - .long 3511966619,3511966619 - .long 976107044,976107044 - .long 2976320012,2976320012 - .long 266819475,266819475 - .long 3533106868,3533106868 - .long 2660342555,2660342555 - .long 1338359936,1338359936 - .long 2720062561,2720062561 - .long 1766553434,1766553434 - .long 370807324,370807324 - .long 179999714,179999714 - .long 3844776128,3844776128 - .long 1138762300,1138762300 - .long 488053522,488053522 - .long 185403662,185403662 - .long 2915535858,2915535858 - .long 3114841645,3114841645 - .long 3366526484,3366526484 - .long 2233069911,2233069911 - .long 1275557295,1275557295 - .long 3151862254,3151862254 - .long 4250959779,4250959779 - .long 2670068215,2670068215 - .long 3170202204,3170202204 - .long 3309004356,3309004356 - .long 880737115,880737115 - .long 1982415755,1982415755 - .long 3703972811,3703972811 - .long 1761406390,1761406390 - .long 1676797112,1676797112 - .long 3403428311,3403428311 - .long 277177154,277177154 - .long 1076008723,1076008723 - .long 538035844,538035844 - .long 2099530373,2099530373 - .long 4164795346,4164795346 - .long 288553390,288553390 - .long 1839278535,1839278535 - .long 1261411869,1261411869 - .long 4080055004,4080055004 - .long 3964831245,3964831245 - .long 3504587127,3504587127 - .long 1813426987,1813426987 - .long 2579067049,2579067049 - .long 4199060497,4199060497 - .long 577038663,577038663 - .long 3297574056,3297574056 - .long 440397984,440397984 - .long 3626794326,3626794326 - .long 4019204898,4019204898 - .long 3343796615,3343796615 - .long 3251714265,3251714265 - .long 4272081548,4272081548 - .long 906744984,906744984 - .long 3481400742,3481400742 - .long 685669029,685669029 - .long 646887386,646887386 - .long 2764025151,2764025151 - .long 3835509292,3835509292 - .long 227702864,227702864 - .long 2613862250,2613862250 - .long 1648787028,1648787028 - .long 3256061430,3256061430 - .long 3904428176,3904428176 - .long 1593260334,1593260334 - .long 4121936770,4121936770 - .long 3196083615,3196083615 - .long 2090061929,2090061929 - .long 2838353263,2838353263 - .long 3004310991,3004310991 - .long 999926984,999926984 - .long 2809993232,2809993232 - .long 1852021992,1852021992 - .long 2075868123,2075868123 - .long 158869197,158869197 - .long 4095236462,4095236462 - .long 28809964,28809964 - .long 2828685187,2828685187 - .long 1701746150,1701746150 - .long 2129067946,2129067946 - .long 147831841,147831841 - .long 3873969647,3873969647 - .long 3650873274,3650873274 - .long 3459673930,3459673930 - .long 3557400554,3557400554 - .long 3598495785,3598495785 - .long 2947720241,2947720241 - .long 824393514,824393514 - .long 815048134,815048134 - .long 3227951669,3227951669 - .long 935087732,935087732 - .long 2798289660,2798289660 - .long 2966458592,2966458592 - .long 366520115,366520115 - .long 1251476721,1251476721 - .long 4158319681,4158319681 - .long 240176511,240176511 - .long 804688151,804688151 - .long 2379631990,2379631990 - .long 1303441219,1303441219 - .long 1414376140,1414376140 - .long 3741619940,3741619940 - .long 3820343710,3820343710 - .long 461924940,461924940 - .long 3089050817,3089050817 - .long 2136040774,2136040774 - .long 82468509,82468509 - .long 1563790337,1563790337 - .long 1937016826,1937016826 - .long 776014843,776014843 - .long 1511876531,1511876531 - .long 1389550482,1389550482 - .long 861278441,861278441 - .long 323475053,323475053 - .long 2355222426,2355222426 - .long 2047648055,2047648055 - .long 2383738969,2383738969 - .long 2302415851,2302415851 - .long 3995576782,3995576782 - .long 902390199,902390199 - .long 3991215329,3991215329 - .long 1018251130,1018251130 - .long 1507840668,1507840668 - .long 1064563285,1064563285 - .long 2043548696,2043548696 - .long 3208103795,3208103795 - .long 3939366739,3939366739 - .long 1537932639,1537932639 - .long 342834655,342834655 - .long 2262516856,2262516856 - .long 2180231114,2180231114 - .long 1053059257,1053059257 - .long 741614648,741614648 - .long 1598071746,1598071746 - .long 1925389590,1925389590 - .long 203809468,203809468 - .long 2336832552,2336832552 - .long 1100287487,1100287487 - .long 1895934009,1895934009 - .long 3736275976,3736275976 - .long 2632234200,2632234200 - .long 2428589668,2428589668 - .long 1636092795,1636092795 - .long 1890988757,1890988757 - .long 1952214088,1952214088 - .long 1113045200,1113045200 - .byte 82,9,106,213,48,54,165,56 - .byte 191,64,163,158,129,243,215,251 - .byte 124,227,57,130,155,47,255,135 - .byte 52,142,67,68,196,222,233,203 - .byte 84,123,148,50,166,194,35,61 - .byte 238,76,149,11,66,250,195,78 - .byte 8,46,161,102,40,217,36,178 - .byte 118,91,162,73,109,139,209,37 - .byte 114,248,246,100,134,104,152,22 - .byte 212,164,92,204,93,101,182,146 - .byte 108,112,72,80,253,237,185,218 - .byte 94,21,70,87,167,141,157,132 - .byte 144,216,171,0,140,188,211,10 - .byte 247,228,88,5,184,179,69,6 - .byte 208,44,30,143,202,63,15,2 - .byte 193,175,189,3,1,19,138,107 - .byte 58,145,17,65,79,103,220,234 - .byte 151,242,207,206,240,180,230,115 - .byte 150,172,116,34,231,173,53,133 - .byte 226,249,55,232,28,117,223,110 - .byte 71,241,26,113,29,41,197,137 - .byte 111,183,98,14,170,24,190,27 - .byte 252,86,62,75,198,210,121,32 - .byte 154,219,192,254,120,205,90,244 - .byte 31,221,168,51,136,7,199,49 - .byte 177,18,16,89,39,128,236,95 - .byte 96,81,127,169,25,181,74,13 - .byte 45,229,122,159,147,201,156,239 - .byte 160,224,59,77,174,42,245,176 - .byte 200,235,187,60,131,83,153,97 - .byte 23,43,4,126,186,119,214,38 - .byte 225,105,20,99,85,33,12,125 -.L__x86_AES_decrypt_end: -.size _x86_AES_decrypt,.L__x86_AES_decrypt_end-_x86_AES_decrypt -.ident "_x86_AES_decrypt" -.globl AES_Td -.text -.globl AES_decrypt -.type AES_decrypt,@function -.align 16 -AES_decrypt: - pushl %ebp - pushl %ebx - pushl %esi - pushl %edi - - movl 20(%esp), %esi - movl 28(%esp), %edi - movl %esp, %eax - subl $24, %esp - andl $-64, %esp - addl $4, %esp - movl %eax, 16(%esp) - call .L003pic_point -.L003pic_point: - popl %ebp - leal AES_Td-.L003pic_point(%ebp),%ebp - leal 2176(%ebp), %ebp - movl -128(%ebp), %eax - movl -96(%ebp), %ebx - movl -64(%ebp), %ecx - movl -32(%ebp), %edx - movl (%ebp), %eax - movl 32(%ebp), %ebx - movl 64(%ebp), %ecx - movl 96(%ebp), %edx - leal -2176(%ebp), %ebp - movl (%esi), %eax - movl 4(%esi), %ebx - movl 8(%esi), %ecx - movl 12(%esi), %edx - call _x86_AES_decrypt - movl 16(%esp), %esp - movl 24(%esp), %esi - movl %eax, (%esi) - movl %ebx, 4(%esi) - movl %ecx, 8(%esi) - movl %edx, 12(%esi) - popl %edi - popl %esi - popl %ebx - popl %ebp - ret -.L_AES_decrypt_end: -.size AES_decrypt,.L_AES_decrypt_end-AES_decrypt -.ident "AES_decrypt" -.globl AES_Te -.globl AES_Td -.text -.globl AES_cbc_encrypt -.type AES_cbc_encrypt,@function -.align 16 -AES_cbc_encrypt: - pushl %ebp - pushl %ebx - pushl %esi - pushl %edi - - movl 28(%esp), %ecx - cmpl $0, %ecx - je .L004enc_out - call .L005pic_point -.L005pic_point: - popl %ebp - pushfl - cld - cmpl $0, 44(%esp) - je .L006DECRYPT - leal AES_Te-.L005pic_point(%ebp),%ebp - leal -308(%esp), %edi - andl $-64, %edi - movl %ebp, %eax - leal 2048(%ebp), %ebx - movl %edi, %edx - andl $4095, %eax - andl $4095, %ebx - andl $4095, %edx - cmpl %ebx, %edx - jb .L007te_break_out - subl %ebx, %edx - subl %edx, %edi - jmp .L008te_ok -.L007te_break_out: - subl %eax, %edx - andl $4095, %edx - addl $320, %edx - subl %edx, %edi -.align 4 -.L008te_ok: - movl 24(%esp), %eax - movl 28(%esp), %ebx - movl 36(%esp), %edx - movl 40(%esp), %esi - xchgl %edi, %esp - addl $4, %esp - movl %edi, 16(%esp) - movl %eax, 20(%esp) - movl %ebx, 24(%esp) - movl %ecx, 28(%esp) - movl %edx, 32(%esp) - movl %esi, 36(%esp) - movl $0, 300(%esp) - movl %edx, %ebx - movl $61, %ecx - subl %ebp, %ebx - movl %edx, %esi - andl $4095, %ebx - leal 60(%esp), %edi - cmpl $2048, %ebx - jb .L009do_ecopy - cmpl $3852, %ebx - jb .L010skip_ecopy -.align 4 -.L009do_ecopy: - movl %edi, 32(%esp) - .long 2784229001 -.L010skip_ecopy: - movl %eax, %esi - movl $16, %edi -.align 4 -.L011prefetch_te: - movl (%ebp), %eax - movl 32(%ebp), %ebx - movl 64(%ebp), %ecx - movl 96(%ebp), %edx - leal 128(%ebp), %ebp - decl %edi - jnz .L011prefetch_te - subl $2048, %ebp - movl 28(%esp), %ecx - movl 36(%esp), %edi - testl $4294967280, %ecx - jz .L012enc_tail - movl (%edi), %eax - movl 4(%edi), %ebx -.align 4 -.L013enc_loop: - movl 8(%edi), %ecx - movl 12(%edi), %edx - xorl (%esi), %eax - xorl 4(%esi), %ebx - xorl 8(%esi), %ecx - xorl 12(%esi), %edx - movl 32(%esp), %edi - call _x86_AES_encrypt - movl 20(%esp), %esi - movl 24(%esp), %edi - movl %eax, (%edi) - movl %ebx, 4(%edi) - movl %ecx, 8(%edi) - movl %edx, 12(%edi) - movl 28(%esp), %ecx - leal 16(%esi), %esi - movl %esi, 20(%esp) - leal 16(%edi), %edx - movl %edx, 24(%esp) - subl $16, %ecx - testl $4294967280, %ecx - movl %ecx, 28(%esp) - jnz .L013enc_loop - testl $15, %ecx - jnz .L012enc_tail - movl 36(%esp), %esi - movl 8(%edi), %ecx - movl 12(%edi), %edx - movl %eax, (%esi) - movl %ebx, 4(%esi) - movl %ecx, 8(%esi) - movl %edx, 12(%esi) - cmpl $0, 300(%esp) - movl 32(%esp), %edi - movl 16(%esp), %esp - je .L014skip_ezero - movl $60, %ecx - xorl %eax, %eax -.align 4 - .long 2884892297 -.L014skip_ezero: - popfl -.L004enc_out: - popl %edi - popl %esi - popl %ebx - popl %ebp - ret - pushfl -.align 4 -.L012enc_tail: - pushl %edi - movl 24(%esp), %edi - movl $16, %ebx - subl %ecx, %ebx - cmpl %esi, %edi - je .L015enc_in_place -.align 4 - .long 2767451785 - jmp .L016enc_skip_in_place -.L015enc_in_place: - leal (%edi,%ecx), %edi -.L016enc_skip_in_place: - movl %ebx, %ecx - xorl %eax, %eax -.align 4 - .long 2868115081 - popl %edi - movl 24(%esp), %esi - movl (%edi), %eax - movl 4(%edi), %ebx - movl $16, 28(%esp) - jmp .L013enc_loop -.align 4 -.L006DECRYPT: - leal AES_Td-.L005pic_point(%ebp),%ebp - leal -308(%esp), %edi - andl $-64, %edi - movl %ebp, %eax - leal 2304(%ebp), %ebx - movl %edi, %edx - andl $4095, %eax - andl $4095, %ebx - andl $4095, %edx - cmpl %ebx, %edx - jb .L017td_break_out - subl %ebx, %edx - subl %edx, %edi - jmp .L018td_ok -.L017td_break_out: - subl %eax, %edx - andl $4095, %edx - addl $320, %edx - subl %edx, %edi -.align 4 -.L018td_ok: - movl 24(%esp), %eax - movl 28(%esp), %ebx - movl 36(%esp), %edx - movl 40(%esp), %esi - xchgl %edi, %esp - addl $4, %esp - movl %edi, 16(%esp) - movl %eax, 20(%esp) - movl %ebx, 24(%esp) - movl %ecx, 28(%esp) - movl %edx, 32(%esp) - movl %esi, 36(%esp) - movl $0, 300(%esp) - movl %edx, %ebx - movl $61, %ecx - subl %ebp, %ebx - movl %edx, %esi - andl $4095, %ebx - leal 60(%esp), %edi - cmpl $2304, %ebx - jb .L019do_dcopy - cmpl $3852, %ebx - jb .L020skip_dcopy -.align 4 -.L019do_dcopy: - movl %edi, 32(%esp) - .long 2784229001 -.L020skip_dcopy: - movl %eax, %esi - movl $18, %edi -.align 4 -.L021prefetch_td: - movl (%ebp), %eax - movl 32(%ebp), %ebx - movl 64(%ebp), %ecx - movl 96(%ebp), %edx - leal 128(%ebp), %ebp - decl %edi - jnz .L021prefetch_td - subl $2304, %ebp - cmpl 24(%esp), %esi - je .L022dec_in_place - movl 36(%esp), %edi - movl %edi, 40(%esp) -.align 4 -.L023dec_loop: - movl (%esi), %eax - movl 4(%esi), %ebx - movl 8(%esi), %ecx - movl 12(%esi), %edx - movl 32(%esp), %edi - call _x86_AES_decrypt - movl 40(%esp), %edi - movl 28(%esp), %esi - xorl (%edi), %eax - xorl 4(%edi), %ebx - xorl 8(%edi), %ecx - xorl 12(%edi), %edx - subl $16, %esi - jc .L024dec_partial - movl %esi, 28(%esp) - movl 20(%esp), %esi - movl 24(%esp), %edi - movl %eax, (%edi) - movl %ebx, 4(%edi) - movl %ecx, 8(%edi) - movl %edx, 12(%edi) - movl %esi, 40(%esp) - leal 16(%esi), %esi - movl %esi, 20(%esp) - leal 16(%edi), %edi - movl %edi, 24(%esp) - jnz .L023dec_loop - movl 40(%esp), %edi -.L025dec_end: - movl 36(%esp), %esi - movl (%edi), %eax - movl 4(%edi), %ebx - movl 8(%edi), %ecx - movl 12(%edi), %edx - movl %eax, (%esi) - movl %ebx, 4(%esi) - movl %ecx, 8(%esi) - movl %edx, 12(%esi) - jmp .L026dec_out -.align 4 -.L024dec_partial: - leal 44(%esp), %edi - movl %eax, (%edi) - movl %ebx, 4(%edi) - movl %ecx, 8(%edi) - movl %edx, 12(%edi) - leal 16(%esi), %ecx - movl %edi, %esi - movl 24(%esp), %edi - .long 2767451785 - movl 20(%esp), %edi - jmp .L025dec_end -.align 4 -.L022dec_in_place: -.L027dec_in_place_loop: - leal 44(%esp), %edi - movl (%esi), %eax - movl 4(%esi), %ebx - movl 8(%esi), %ecx - movl 12(%esi), %edx - movl %eax, (%edi) - movl %ebx, 4(%edi) - movl %ecx, 8(%edi) - movl %edx, 12(%edi) - movl 32(%esp), %edi - call _x86_AES_decrypt - movl 36(%esp), %edi - movl 24(%esp), %esi - xorl (%edi), %eax - xorl 4(%edi), %ebx - xorl 8(%edi), %ecx - xorl 12(%edi), %edx - movl %eax, (%esi) - movl %ebx, 4(%esi) - movl %ecx, 8(%esi) - movl %edx, 12(%esi) - leal 16(%esi), %esi - movl %esi, 24(%esp) - leal 44(%esp), %esi - movl (%esi), %eax - movl 4(%esi), %ebx - movl 8(%esi), %ecx - movl 12(%esi), %edx - movl %eax, (%edi) - movl %ebx, 4(%edi) - movl %ecx, 8(%edi) - movl %edx, 12(%edi) - movl 20(%esp), %esi - leal 16(%esi), %esi - movl %esi, 20(%esp) - movl 28(%esp), %ecx - subl $16, %ecx - jc .L028dec_in_place_partial - movl %ecx, 28(%esp) - jnz .L027dec_in_place_loop - jmp .L026dec_out -.align 4 -.L028dec_in_place_partial: - movl 24(%esp), %edi - leal 44(%esp), %esi - leal (%edi,%ecx), %edi - leal 16(%esi,%ecx), %esi - negl %ecx - .long 2767451785 -.align 4 -.L026dec_out: - cmpl $0, 300(%esp) - movl 32(%esp), %edi - movl 16(%esp), %esp - je .L029skip_dzero - movl $60, %ecx - xorl %eax, %eax -.align 4 - .long 2884892297 -.L029skip_dzero: - popfl - popl %edi - popl %esi - popl %ebx - popl %ebp - ret -.L_AES_cbc_encrypt_end: -.size AES_cbc_encrypt,.L_AES_cbc_encrypt_end-AES_cbc_encrypt -.ident "AES_cbc_encrypt" -.globl AES_Te -.text -.globl AES_set_encrypt_key -.type AES_set_encrypt_key,@function -.align 16 -AES_set_encrypt_key: - pushl %ebp - pushl %ebx - pushl %esi - pushl %edi - - movl 20(%esp), %esi - movl 28(%esp), %edi - testl $-1, %esi - jz .L030badpointer - testl $-1, %edi - jz .L030badpointer - call .L031pic_point -.L031pic_point: - popl %ebp - leal AES_Te-.L031pic_point(%ebp),%ebp - movl 24(%esp), %ecx - cmpl $128, %ecx - je .L03210rounds - cmpl $192, %ecx - je .L03312rounds - cmpl $256, %ecx - je .L03414rounds - movl $-2, %eax - jmp .L035exit -.L03210rounds: - movl (%esi), %eax - movl 4(%esi), %ebx - movl 8(%esi), %ecx - movl 12(%esi), %edx - movl %eax, (%edi) - movl %ebx, 4(%edi) - movl %ecx, 8(%edi) - movl %edx, 12(%edi) - xorl %ecx, %ecx - jmp .L03610shortcut -.align 4 -.L03710loop: - movl (%edi), %eax - movl 12(%edi), %edx -.L03610shortcut: - movzbl %dl, %esi - movl 2(%ebp,%esi,8), %ebx - movzbl %dh, %esi - andl $4278190080, %ebx - xorl %ebx, %eax - movl 2(%ebp,%esi,8), %ebx - shrl $16, %edx - andl $255, %ebx - movzbl %dl, %esi - xorl %ebx, %eax - movl (%ebp,%esi,8), %ebx - movzbl %dh, %esi - andl $65280, %ebx - xorl %ebx, %eax - movl (%ebp,%esi,8), %ebx - andl $16711680, %ebx - xorl %ebx, %eax - xorl 2048(%ebp,%ecx,4),%eax - movl %eax, 16(%edi) - xorl 4(%edi), %eax - movl %eax, 20(%edi) - xorl 8(%edi), %eax - movl %eax, 24(%edi) - xorl 12(%edi), %eax - movl %eax, 28(%edi) - incl %ecx - addl $16, %edi - cmpl $10, %ecx - jl .L03710loop - movl $10, 80(%edi) - xorl %eax, %eax - jmp .L035exit -.L03312rounds: - movl (%esi), %eax - movl 4(%esi), %ebx - movl 8(%esi), %ecx - movl 12(%esi), %edx - movl %eax, (%edi) - movl %ebx, 4(%edi) - movl %ecx, 8(%edi) - movl %edx, 12(%edi) - movl 16(%esi), %ecx - movl 20(%esi), %edx - movl %ecx, 16(%edi) - movl %edx, 20(%edi) - xorl %ecx, %ecx - jmp .L03812shortcut -.align 4 -.L03912loop: - movl (%edi), %eax - movl 20(%edi), %edx -.L03812shortcut: - movzbl %dl, %esi - movl 2(%ebp,%esi,8), %ebx - movzbl %dh, %esi - andl $4278190080, %ebx - xorl %ebx, %eax - movl 2(%ebp,%esi,8), %ebx - shrl $16, %edx - andl $255, %ebx - movzbl %dl, %esi - xorl %ebx, %eax - movl (%ebp,%esi,8), %ebx - movzbl %dh, %esi - andl $65280, %ebx - xorl %ebx, %eax - movl (%ebp,%esi,8), %ebx - andl $16711680, %ebx - xorl %ebx, %eax - xorl 2048(%ebp,%ecx,4),%eax - movl %eax, 24(%edi) - xorl 4(%edi), %eax - movl %eax, 28(%edi) - xorl 8(%edi), %eax - movl %eax, 32(%edi) - xorl 12(%edi), %eax - movl %eax, 36(%edi) - cmpl $7, %ecx - je .L04012break - incl %ecx - xorl 16(%edi), %eax - movl %eax, 40(%edi) - xorl 20(%edi), %eax - movl %eax, 44(%edi) - addl $24, %edi - jmp .L03912loop -.L04012break: - movl $12, 72(%edi) - xorl %eax, %eax - jmp .L035exit -.L03414rounds: - movl (%esi), %eax - movl 4(%esi), %ebx - movl 8(%esi), %ecx - movl 12(%esi), %edx - movl %eax, (%edi) - movl %ebx, 4(%edi) - movl %ecx, 8(%edi) - movl %edx, 12(%edi) - movl 16(%esi), %eax - movl 20(%esi), %ebx - movl 24(%esi), %ecx - movl 28(%esi), %edx - movl %eax, 16(%edi) - movl %ebx, 20(%edi) - movl %ecx, 24(%edi) - movl %edx, 28(%edi) - xorl %ecx, %ecx - jmp .L04114shortcut -.align 4 -.L04214loop: - movl 28(%edi), %edx -.L04114shortcut: - movl (%edi), %eax - movzbl %dl, %esi - movl 2(%ebp,%esi,8), %ebx - movzbl %dh, %esi - andl $4278190080, %ebx - xorl %ebx, %eax - movl 2(%ebp,%esi,8), %ebx - shrl $16, %edx - andl $255, %ebx - movzbl %dl, %esi - xorl %ebx, %eax - movl (%ebp,%esi,8), %ebx - movzbl %dh, %esi - andl $65280, %ebx - xorl %ebx, %eax - movl (%ebp,%esi,8), %ebx - andl $16711680, %ebx - xorl %ebx, %eax - xorl 2048(%ebp,%ecx,4),%eax - movl %eax, 32(%edi) - xorl 4(%edi), %eax - movl %eax, 36(%edi) - xorl 8(%edi), %eax - movl %eax, 40(%edi) - xorl 12(%edi), %eax - movl %eax, 44(%edi) - cmpl $6, %ecx - je .L04314break - incl %ecx - movl %eax, %edx - movl 16(%edi), %eax - movzbl %dl, %esi - movl 2(%ebp,%esi,8), %ebx - movzbl %dh, %esi - andl $255, %ebx - xorl %ebx, %eax - movl (%ebp,%esi,8), %ebx - shrl $16, %edx - andl $65280, %ebx - movzbl %dl, %esi - xorl %ebx, %eax - movl (%ebp,%esi,8), %ebx - movzbl %dh, %esi - andl $16711680, %ebx - xorl %ebx, %eax - movl 2(%ebp,%esi,8), %ebx - andl $4278190080, %ebx - xorl %ebx, %eax - movl %eax, 48(%edi) - xorl 20(%edi), %eax - movl %eax, 52(%edi) - xorl 24(%edi), %eax - movl %eax, 56(%edi) - xorl 28(%edi), %eax - movl %eax, 60(%edi) - addl $32, %edi - jmp .L04214loop -.L04314break: - movl $14, 48(%edi) - xorl %eax, %eax - jmp .L035exit -.L030badpointer: - movl $-1, %eax -.L035exit: - popl %edi - popl %esi - popl %ebx - popl %ebp - ret -.L_AES_set_encrypt_key_end: -.size AES_set_encrypt_key,.L_AES_set_encrypt_key_end-AES_set_encrypt_key -.ident "AES_set_encrypt_key" -.globl AES_Td -.globl AES_Te -.text -.globl AES_set_decrypt_key -.type AES_set_decrypt_key,@function -.align 16 -AES_set_decrypt_key: - movl 4(%esp), %eax - movl 8(%esp), %ecx - movl 12(%esp), %edx - subl $12, %esp - movl %eax, (%esp) - movl %ecx, 4(%esp) - movl %edx, 8(%esp) - call AES_set_encrypt_key - addl $12, %esp - cmpl $0, %eax - je .L044proceed - ret -.L044proceed: - pushl %ebp - pushl %ebx - pushl %esi - pushl %edi - movl 28(%esp), %esi - movl 240(%esi), %ecx - leal (,%ecx,4), %ecx - leal (%esi,%ecx,4), %edi -.align 4 -.L045invert: - movl (%esi), %eax - movl 4(%esi), %ebx - movl (%edi), %ecx - movl 4(%edi), %edx - movl %eax, (%edi) - movl %ebx, 4(%edi) - movl %ecx, (%esi) - movl %edx, 4(%esi) - movl 8(%esi), %eax - movl 12(%esi), %ebx - movl 8(%edi), %ecx - movl 12(%edi), %edx - movl %eax, 8(%edi) - movl %ebx, 12(%edi) - movl %ecx, 8(%esi) - movl %edx, 12(%esi) - addl $16, %esi - subl $16, %edi - cmpl %edi, %esi - jne .L045invert - call .L046pic_point -.L046pic_point: - popl %ebp - leal AES_Td-.L046pic_point(%ebp),%edi - leal AES_Te-.L046pic_point(%ebp),%ebp - movl 28(%esp), %esi - movl 240(%esi), %ecx - decl %ecx -.align 4 -.L047permute: - addl $16, %esi - movl (%esi), %eax - movl %eax, %edx - movzbl %ah, %ebx - shrl $16, %edx - andl $255, %eax - movzbl 2(%ebp,%eax,8), %eax - movzbl 2(%ebp,%ebx,8), %ebx - movl (%edi,%eax,8), %eax - xorl 3(%edi,%ebx,8), %eax - movzbl %dh, %ebx - andl $255, %edx - movzbl 2(%ebp,%edx,8), %edx - movzbl 2(%ebp,%ebx,8), %ebx - xorl 2(%edi,%edx,8), %eax - xorl 1(%edi,%ebx,8), %eax - movl %eax, (%esi) - movl 4(%esi), %eax - movl %eax, %edx - movzbl %ah, %ebx - shrl $16, %edx - andl $255, %eax - movzbl 2(%ebp,%eax,8), %eax - movzbl 2(%ebp,%ebx,8), %ebx - movl (%edi,%eax,8), %eax - xorl 3(%edi,%ebx,8), %eax - movzbl %dh, %ebx - andl $255, %edx - movzbl 2(%ebp,%edx,8), %edx - movzbl 2(%ebp,%ebx,8), %ebx - xorl 2(%edi,%edx,8), %eax - xorl 1(%edi,%ebx,8), %eax - movl %eax, 4(%esi) - movl 8(%esi), %eax - movl %eax, %edx - movzbl %ah, %ebx - shrl $16, %edx - andl $255, %eax - movzbl 2(%ebp,%eax,8), %eax - movzbl 2(%ebp,%ebx,8), %ebx - movl (%edi,%eax,8), %eax - xorl 3(%edi,%ebx,8), %eax - movzbl %dh, %ebx - andl $255, %edx - movzbl 2(%ebp,%edx,8), %edx - movzbl 2(%ebp,%ebx,8), %ebx - xorl 2(%edi,%edx,8), %eax - xorl 1(%edi,%ebx,8), %eax - movl %eax, 8(%esi) - movl 12(%esi), %eax - movl %eax, %edx - movzbl %ah, %ebx - shrl $16, %edx - andl $255, %eax - movzbl 2(%ebp,%eax,8), %eax - movzbl 2(%ebp,%ebx,8), %ebx - movl (%edi,%eax,8), %eax - xorl 3(%edi,%ebx,8), %eax - movzbl %dh, %ebx - andl $255, %edx - movzbl 2(%ebp,%edx,8), %edx - movzbl 2(%ebp,%ebx,8), %ebx - xorl 2(%edi,%edx,8), %eax - xorl 1(%edi,%ebx,8), %eax - movl %eax, 12(%esi) - decl %ecx - jnz .L047permute - xorl %eax, %eax - popl %edi - popl %esi - popl %ebx - popl %ebp - ret -.L_AES_set_decrypt_key_end: -.size AES_set_decrypt_key,.L_AES_set_decrypt_key_end-AES_set_decrypt_key -.ident "AES_set_decrypt_key" diff --git a/fips-1.0/des/asm/fips-dx86-elf.s b/fips-1.0/des/asm/fips-dx86-elf.s deleted file mode 100644 index 7b4b11f0f3..0000000000 --- a/fips-1.0/des/asm/fips-dx86-elf.s +++ /dev/null @@ -1,2707 +0,0 @@ - - - - - - - .file "des-586.s" - .version "01.01" -gcc2_compiled.: -.text - .align 16 -.globl DES_encrypt1 - .type DES_encrypt1,@function -DES_encrypt1: - pushl %esi - pushl %edi - - - movl 12(%esp), %esi - xorl %ecx, %ecx - pushl %ebx - pushl %ebp - movl (%esi), %eax - movl 28(%esp), %ebx - movl 4(%esi), %edi - - - roll $4, %eax - movl %eax, %esi - xorl %edi, %eax - andl $0xf0f0f0f0, %eax - xorl %eax, %esi - xorl %eax, %edi - - roll $20, %edi - movl %edi, %eax - xorl %esi, %edi - andl $0xfff0000f, %edi - xorl %edi, %eax - xorl %edi, %esi - - roll $14, %eax - movl %eax, %edi - xorl %esi, %eax - andl $0x33333333, %eax - xorl %eax, %edi - xorl %eax, %esi - - roll $22, %esi - movl %esi, %eax - xorl %edi, %esi - andl $0x03fc03fc, %esi - xorl %esi, %eax - xorl %esi, %edi - - roll $9, %eax - movl %eax, %esi - xorl %edi, %eax - andl $0xaaaaaaaa, %eax - xorl %eax, %esi - xorl %eax, %edi - -.byte 209 -.byte 199 - .align 8 - call .L000PIC_me_up -.L000PIC_me_up: - popl %ebp - addl $_GLOBAL_OFFSET_TABLE_+[.-.L000PIC_me_up],%ebp - movl DES_SPtrans@GOT(%ebp),%ebp - movl 24(%esp), %ecx - cmpl $0, %ebx - je .L001start_decrypt - - - movl (%ecx), %eax - xorl %ebx, %ebx - movl 4(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 8(%ecx), %eax - xorl %ebx, %ebx - movl 12(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 16(%ecx), %eax - xorl %ebx, %ebx - movl 20(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 24(%ecx), %eax - xorl %ebx, %ebx - movl 28(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 32(%ecx), %eax - xorl %ebx, %ebx - movl 36(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 40(%ecx), %eax - xorl %ebx, %ebx - movl 44(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 48(%ecx), %eax - xorl %ebx, %ebx - movl 52(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 56(%ecx), %eax - xorl %ebx, %ebx - movl 60(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 64(%ecx), %eax - xorl %ebx, %ebx - movl 68(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 72(%ecx), %eax - xorl %ebx, %ebx - movl 76(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 80(%ecx), %eax - xorl %ebx, %ebx - movl 84(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 88(%ecx), %eax - xorl %ebx, %ebx - movl 92(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 96(%ecx), %eax - xorl %ebx, %ebx - movl 100(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 104(%ecx), %eax - xorl %ebx, %ebx - movl 108(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 112(%ecx), %eax - xorl %ebx, %ebx - movl 116(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 120(%ecx), %eax - xorl %ebx, %ebx - movl 124(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - jmp .L002end -.L001start_decrypt: - - - movl 120(%ecx), %eax - xorl %ebx, %ebx - movl 124(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 112(%ecx), %eax - xorl %ebx, %ebx - movl 116(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 104(%ecx), %eax - xorl %ebx, %ebx - movl 108(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 96(%ecx), %eax - xorl %ebx, %ebx - movl 100(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 88(%ecx), %eax - xorl %ebx, %ebx - movl 92(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 80(%ecx), %eax - xorl %ebx, %ebx - movl 84(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 72(%ecx), %eax - xorl %ebx, %ebx - movl 76(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 64(%ecx), %eax - xorl %ebx, %ebx - movl 68(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 56(%ecx), %eax - xorl %ebx, %ebx - movl 60(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 48(%ecx), %eax - xorl %ebx, %ebx - movl 52(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 40(%ecx), %eax - xorl %ebx, %ebx - movl 44(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 32(%ecx), %eax - xorl %ebx, %ebx - movl 36(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 24(%ecx), %eax - xorl %ebx, %ebx - movl 28(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 16(%ecx), %eax - xorl %ebx, %ebx - movl 20(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 8(%ecx), %eax - xorl %ebx, %ebx - movl 12(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl (%ecx), %eax - xorl %ebx, %ebx - movl 4(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi -.L002end: - - - movl 20(%esp), %edx -.byte 209 -.byte 206 - movl %edi, %eax - xorl %esi, %edi - andl $0xaaaaaaaa, %edi - xorl %edi, %eax - xorl %edi, %esi - - roll $23, %eax - movl %eax, %edi - xorl %esi, %eax - andl $0x03fc03fc, %eax - xorl %eax, %edi - xorl %eax, %esi - - roll $10, %edi - movl %edi, %eax - xorl %esi, %edi - andl $0x33333333, %edi - xorl %edi, %eax - xorl %edi, %esi - - roll $18, %esi - movl %esi, %edi - xorl %eax, %esi - andl $0xfff0000f, %esi - xorl %esi, %edi - xorl %esi, %eax - - roll $12, %edi - movl %edi, %esi - xorl %eax, %edi - andl $0xf0f0f0f0, %edi - xorl %edi, %esi - xorl %edi, %eax - - rorl $4, %eax - movl %eax, (%edx) - movl %esi, 4(%edx) - popl %ebp - popl %ebx - popl %edi - popl %esi - ret -.L_DES_encrypt1_end: - .size DES_encrypt1,.L_DES_encrypt1_end-DES_encrypt1 -.ident "desasm.pl" -.text - .align 16 -.globl DES_encrypt2 - .type DES_encrypt2,@function -DES_encrypt2: - pushl %esi - pushl %edi - - - movl 12(%esp), %eax - xorl %ecx, %ecx - pushl %ebx - pushl %ebp - movl (%eax), %esi - movl 28(%esp), %ebx - roll $3, %esi - movl 4(%eax), %edi - roll $3, %edi - .align 8 - call .L003PIC_me_up -.L003PIC_me_up: - popl %ebp - addl $_GLOBAL_OFFSET_TABLE_+[.-.L003PIC_me_up],%ebp - movl DES_SPtrans@GOT(%ebp),%ebp - movl 24(%esp), %ecx - cmpl $0, %ebx - je .L004start_decrypt - - - movl (%ecx), %eax - xorl %ebx, %ebx - movl 4(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 8(%ecx), %eax - xorl %ebx, %ebx - movl 12(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 16(%ecx), %eax - xorl %ebx, %ebx - movl 20(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 24(%ecx), %eax - xorl %ebx, %ebx - movl 28(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 32(%ecx), %eax - xorl %ebx, %ebx - movl 36(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 40(%ecx), %eax - xorl %ebx, %ebx - movl 44(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 48(%ecx), %eax - xorl %ebx, %ebx - movl 52(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 56(%ecx), %eax - xorl %ebx, %ebx - movl 60(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 64(%ecx), %eax - xorl %ebx, %ebx - movl 68(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 72(%ecx), %eax - xorl %ebx, %ebx - movl 76(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 80(%ecx), %eax - xorl %ebx, %ebx - movl 84(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 88(%ecx), %eax - xorl %ebx, %ebx - movl 92(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 96(%ecx), %eax - xorl %ebx, %ebx - movl 100(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 104(%ecx), %eax - xorl %ebx, %ebx - movl 108(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 112(%ecx), %eax - xorl %ebx, %ebx - movl 116(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 120(%ecx), %eax - xorl %ebx, %ebx - movl 124(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - jmp .L005end -.L004start_decrypt: - - - movl 120(%ecx), %eax - xorl %ebx, %ebx - movl 124(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 112(%ecx), %eax - xorl %ebx, %ebx - movl 116(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 104(%ecx), %eax - xorl %ebx, %ebx - movl 108(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 96(%ecx), %eax - xorl %ebx, %ebx - movl 100(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 88(%ecx), %eax - xorl %ebx, %ebx - movl 92(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 80(%ecx), %eax - xorl %ebx, %ebx - movl 84(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 72(%ecx), %eax - xorl %ebx, %ebx - movl 76(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 64(%ecx), %eax - xorl %ebx, %ebx - movl 68(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 56(%ecx), %eax - xorl %ebx, %ebx - movl 60(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 48(%ecx), %eax - xorl %ebx, %ebx - movl 52(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 40(%ecx), %eax - xorl %ebx, %ebx - movl 44(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 32(%ecx), %eax - xorl %ebx, %ebx - movl 36(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 24(%ecx), %eax - xorl %ebx, %ebx - movl 28(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl 16(%ecx), %eax - xorl %ebx, %ebx - movl 20(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi - - - movl 8(%ecx), %eax - xorl %ebx, %ebx - movl 12(%ecx), %edx - xorl %esi, %eax - xorl %ecx, %ecx - xorl %esi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%edi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%edi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%edi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%edi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%edi - xorl 0x700(%ebp,%ecx),%edi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%edi - xorl 0x500(%ebp,%edx),%edi - - - movl (%ecx), %eax - xorl %ebx, %ebx - movl 4(%ecx), %edx - xorl %edi, %eax - xorl %ecx, %ecx - xorl %edi, %edx - andl $0xfcfcfcfc, %eax - andl $0xcfcfcfcf, %edx - movb %al, %bl - movb %ah, %cl - rorl $4, %edx - xorl (%ebp,%ebx),%esi - movb %dl, %bl - xorl 0x200(%ebp,%ecx),%esi - movb %dh, %cl - shrl $16, %eax - xorl 0x100(%ebp,%ebx),%esi - movb %ah, %bl - shrl $16, %edx - xorl 0x300(%ebp,%ecx),%esi - movb %dh, %cl - andl $0xff, %eax - andl $0xff, %edx - xorl 0x600(%ebp,%ebx),%esi - xorl 0x700(%ebp,%ecx),%esi - movl 24(%esp), %ecx - xorl 0x400(%ebp,%eax),%esi - xorl 0x500(%ebp,%edx),%esi -.L005end: - - - rorl $3, %edi - movl 20(%esp), %eax - rorl $3, %esi - movl %edi, (%eax) - movl %esi, 4(%eax) - popl %ebp - popl %ebx - popl %edi - popl %esi - ret -.L_DES_encrypt2_end: - .size DES_encrypt2,.L_DES_encrypt2_end-DES_encrypt2 -.ident "desasm.pl" -.text - .align 16 -.globl DES_encrypt3 - .type DES_encrypt3,@function -DES_encrypt3: - pushl %ebx - movl 8(%esp), %ebx - pushl %ebp - pushl %esi - pushl %edi - - - movl (%ebx), %edi - movl 4(%ebx), %esi - subl $12, %esp - - - roll $4, %edi - movl %edi, %edx - xorl %esi, %edi - andl $0xf0f0f0f0, %edi - xorl %edi, %edx - xorl %edi, %esi - - roll $20, %esi - movl %esi, %edi - xorl %edx, %esi - andl $0xfff0000f, %esi - xorl %esi, %edi - xorl %esi, %edx - - roll $14, %edi - movl %edi, %esi - xorl %edx, %edi - andl $0x33333333, %edi - xorl %edi, %esi - xorl %edi, %edx - - roll $22, %edx - movl %edx, %edi - xorl %esi, %edx - andl $0x03fc03fc, %edx - xorl %edx, %edi - xorl %edx, %esi - - roll $9, %edi - movl %edi, %edx - xorl %esi, %edi - andl $0xaaaaaaaa, %edi - xorl %edi, %edx - xorl %edi, %esi - - rorl $3, %edx - rorl $2, %esi - movl %esi, 4(%ebx) - movl 36(%esp), %eax - movl %edx, (%ebx) - movl 40(%esp), %edi - movl 44(%esp), %esi - movl $1, 8(%esp) - movl %eax, 4(%esp) - movl %ebx, (%esp) - call DES_encrypt2 - movl $0, 8(%esp) - movl %edi, 4(%esp) - movl %ebx, (%esp) - call DES_encrypt2 - movl $1, 8(%esp) - movl %esi, 4(%esp) - movl %ebx, (%esp) - call DES_encrypt2 - addl $12, %esp - movl (%ebx), %edi - movl 4(%ebx), %esi - - - roll $2, %esi - roll $3, %edi - movl %edi, %eax - xorl %esi, %edi - andl $0xaaaaaaaa, %edi - xorl %edi, %eax - xorl %edi, %esi - - roll $23, %eax - movl %eax, %edi - xorl %esi, %eax - andl $0x03fc03fc, %eax - xorl %eax, %edi - xorl %eax, %esi - - roll $10, %edi - movl %edi, %eax - xorl %esi, %edi - andl $0x33333333, %edi - xorl %edi, %eax - xorl %edi, %esi - - roll $18, %esi - movl %esi, %edi - xorl %eax, %esi - andl $0xfff0000f, %esi - xorl %esi, %edi - xorl %esi, %eax - - roll $12, %edi - movl %edi, %esi - xorl %eax, %edi - andl $0xf0f0f0f0, %edi - xorl %edi, %esi - xorl %edi, %eax - - rorl $4, %eax - movl %eax, (%ebx) - movl %esi, 4(%ebx) - popl %edi - popl %esi - popl %ebp - popl %ebx - ret -.L_DES_encrypt3_end: - .size DES_encrypt3,.L_DES_encrypt3_end-DES_encrypt3 -.ident "desasm.pl" -.text - .align 16 -.globl DES_decrypt3 - .type DES_decrypt3,@function -DES_decrypt3: - pushl %ebx - movl 8(%esp), %ebx - pushl %ebp - pushl %esi - pushl %edi - - - movl (%ebx), %edi - movl 4(%ebx), %esi - subl $12, %esp - - - roll $4, %edi - movl %edi, %edx - xorl %esi, %edi - andl $0xf0f0f0f0, %edi - xorl %edi, %edx - xorl %edi, %esi - - roll $20, %esi - movl %esi, %edi - xorl %edx, %esi - andl $0xfff0000f, %esi - xorl %esi, %edi - xorl %esi, %edx - - roll $14, %edi - movl %edi, %esi - xorl %edx, %edi - andl $0x33333333, %edi - xorl %edi, %esi - xorl %edi, %edx - - roll $22, %edx - movl %edx, %edi - xorl %esi, %edx - andl $0x03fc03fc, %edx - xorl %edx, %edi - xorl %edx, %esi - - roll $9, %edi - movl %edi, %edx - xorl %esi, %edi - andl $0xaaaaaaaa, %edi - xorl %edi, %edx - xorl %edi, %esi - - rorl $3, %edx - rorl $2, %esi - movl %esi, 4(%ebx) - movl 36(%esp), %esi - movl %edx, (%ebx) - movl 40(%esp), %edi - movl 44(%esp), %eax - movl $0, 8(%esp) - movl %eax, 4(%esp) - movl %ebx, (%esp) - call DES_encrypt2 - movl $1, 8(%esp) - movl %edi, 4(%esp) - movl %ebx, (%esp) - call DES_encrypt2 - movl $0, 8(%esp) - movl %esi, 4(%esp) - movl %ebx, (%esp) - call DES_encrypt2 - addl $12, %esp - movl (%ebx), %edi - movl 4(%ebx), %esi - - - roll $2, %esi - roll $3, %edi - movl %edi, %eax - xorl %esi, %edi - andl $0xaaaaaaaa, %edi - xorl %edi, %eax - xorl %edi, %esi - - roll $23, %eax - movl %eax, %edi - xorl %esi, %eax - andl $0x03fc03fc, %eax - xorl %eax, %edi - xorl %eax, %esi - - roll $10, %edi - movl %edi, %eax - xorl %esi, %edi - andl $0x33333333, %edi - xorl %edi, %eax - xorl %edi, %esi - - roll $18, %esi - movl %esi, %edi - xorl %eax, %esi - andl $0xfff0000f, %esi - xorl %esi, %edi - xorl %esi, %eax - - roll $12, %edi - movl %edi, %esi - xorl %eax, %edi - andl $0xf0f0f0f0, %edi - xorl %edi, %esi - xorl %edi, %eax - - rorl $4, %eax - movl %eax, (%ebx) - movl %esi, 4(%ebx) - popl %edi - popl %esi - popl %ebp - popl %ebx - ret -.L_DES_decrypt3_end: - .size DES_decrypt3,.L_DES_decrypt3_end-DES_decrypt3 -.ident "desasm.pl" -.text - .align 16 -.globl DES_ncbc_encrypt - .type DES_ncbc_encrypt,@function -DES_ncbc_encrypt: - - pushl %ebp - pushl %ebx - pushl %esi - pushl %edi - movl 28(%esp), %ebp - - movl 36(%esp), %ebx - movl (%ebx), %esi - movl 4(%ebx), %edi - pushl %edi - pushl %esi - pushl %edi - pushl %esi - movl %esp, %ebx - movl 36(%esp), %esi - movl 40(%esp), %edi - - movl 56(%esp), %ecx - - pushl %ecx - - movl 52(%esp), %eax - pushl %eax - pushl %ebx - cmpl $0, %ecx - jz .L006decrypt - andl $4294967288, %ebp - movl 12(%esp), %eax - movl 16(%esp), %ebx - jz .L007encrypt_finish -.L008encrypt_loop: - movl (%esi), %ecx - movl 4(%esi), %edx - xorl %ecx, %eax - xorl %edx, %ebx - movl %eax, 12(%esp) - movl %ebx, 16(%esp) - call DES_encrypt1 - movl 12(%esp), %eax - movl 16(%esp), %ebx - movl %eax, (%edi) - movl %ebx, 4(%edi) - addl $8, %esi - addl $8, %edi - subl $8, %ebp - jnz .L008encrypt_loop -.L007encrypt_finish: - movl 56(%esp), %ebp - andl $7, %ebp - jz .L009finish - call .L010PIC_point -.L010PIC_point: - popl %edx - leal .L011cbc_enc_jmp_table-.L010PIC_point(%edx),%ecx - movl (%ecx,%ebp,4), %ebp - addl %edx, %ebp - xorl %ecx, %ecx - xorl %edx, %edx - jmp *%ebp -.L012ej7: - movb 6(%esi), %dh - sall $8, %edx -.L013ej6: - movb 5(%esi), %dh -.L014ej5: - movb 4(%esi), %dl -.L015ej4: - movl (%esi), %ecx - jmp .L016ejend -.L017ej3: - movb 2(%esi), %ch - sall $8, %ecx -.L018ej2: - movb 1(%esi), %ch -.L019ej1: - movb (%esi), %cl -.L016ejend: - xorl %ecx, %eax - xorl %edx, %ebx - movl %eax, 12(%esp) - movl %ebx, 16(%esp) - call DES_encrypt1 - movl 12(%esp), %eax - movl 16(%esp), %ebx - movl %eax, (%edi) - movl %ebx, 4(%edi) - jmp .L009finish -.align 16 -.L006decrypt: - andl $4294967288, %ebp - movl 20(%esp), %eax - movl 24(%esp), %ebx - jz .L020decrypt_finish -.L021decrypt_loop: - movl (%esi), %eax - movl 4(%esi), %ebx - movl %eax, 12(%esp) - movl %ebx, 16(%esp) - call DES_encrypt1 - movl 12(%esp), %eax - movl 16(%esp), %ebx - movl 20(%esp), %ecx - movl 24(%esp), %edx - xorl %eax, %ecx - xorl %ebx, %edx - movl (%esi), %eax - movl 4(%esi), %ebx - movl %ecx, (%edi) - movl %edx, 4(%edi) - movl %eax, 20(%esp) - movl %ebx, 24(%esp) - addl $8, %esi - addl $8, %edi - subl $8, %ebp - jnz .L021decrypt_loop -.L020decrypt_finish: - movl 56(%esp), %ebp - andl $7, %ebp - jz .L009finish - movl (%esi), %eax - movl 4(%esi), %ebx - movl %eax, 12(%esp) - movl %ebx, 16(%esp) - call DES_encrypt1 - movl 12(%esp), %eax - movl 16(%esp), %ebx - movl 20(%esp), %ecx - movl 24(%esp), %edx - xorl %eax, %ecx - xorl %ebx, %edx - movl (%esi), %eax - movl 4(%esi), %ebx -.L022dj7: - rorl $16, %edx - movb %dl, 6(%edi) - shrl $16, %edx -.L023dj6: - movb %dh, 5(%edi) -.L024dj5: - movb %dl, 4(%edi) -.L025dj4: - movl %ecx, (%edi) - jmp .L026djend -.L027dj3: - rorl $16, %ecx - movb %cl, 2(%edi) - sall $16, %ecx -.L028dj2: - movb %ch, 1(%esi) -.L029dj1: - movb %cl, (%esi) -.L026djend: - jmp .L009finish -.align 16 -.L009finish: - movl 64(%esp), %ecx - addl $28, %esp - movl %eax, (%ecx) - movl %ebx, 4(%ecx) - popl %edi - popl %esi - popl %ebx - popl %ebp - ret -.align 16 -.L011cbc_enc_jmp_table: - .long 0 - .long .L019ej1-.L010PIC_point - .long .L018ej2-.L010PIC_point - .long .L017ej3-.L010PIC_point - .long .L015ej4-.L010PIC_point - .long .L014ej5-.L010PIC_point - .long .L013ej6-.L010PIC_point - .long .L012ej7-.L010PIC_point -.L_DES_ncbc_encrypt_end: - .size DES_ncbc_encrypt,.L_DES_ncbc_encrypt_end-DES_ncbc_encrypt -.ident "desasm.pl" -.text - .align 16 -.globl DES_ede3_cbc_encrypt - .type DES_ede3_cbc_encrypt,@function -DES_ede3_cbc_encrypt: - - pushl %ebp - pushl %ebx - pushl %esi - pushl %edi - movl 28(%esp), %ebp - - movl 44(%esp), %ebx - movl (%ebx), %esi - movl 4(%ebx), %edi - pushl %edi - pushl %esi - pushl %edi - pushl %esi - movl %esp, %ebx - movl 36(%esp), %esi - movl 40(%esp), %edi - - movl 64(%esp), %ecx - - movl 56(%esp), %eax - pushl %eax - - movl 56(%esp), %eax - pushl %eax - - movl 56(%esp), %eax - pushl %eax - pushl %ebx - cmpl $0, %ecx - jz .L030decrypt - andl $4294967288, %ebp - movl 16(%esp), %eax - movl 20(%esp), %ebx - jz .L031encrypt_finish -.L032encrypt_loop: - movl (%esi), %ecx - movl 4(%esi), %edx - xorl %ecx, %eax - xorl %edx, %ebx - movl %eax, 16(%esp) - movl %ebx, 20(%esp) - call DES_encrypt3 - movl 16(%esp), %eax - movl 20(%esp), %ebx - movl %eax, (%edi) - movl %ebx, 4(%edi) - addl $8, %esi - addl $8, %edi - subl $8, %ebp - jnz .L032encrypt_loop -.L031encrypt_finish: - movl 60(%esp), %ebp - andl $7, %ebp - jz .L033finish - call .L034PIC_point -.L034PIC_point: - popl %edx - leal .L035cbc_enc_jmp_table-.L034PIC_point(%edx),%ecx - movl (%ecx,%ebp,4), %ebp - addl %edx, %ebp - xorl %ecx, %ecx - xorl %edx, %edx - jmp *%ebp -.L036ej7: - movb 6(%esi), %dh - sall $8, %edx -.L037ej6: - movb 5(%esi), %dh -.L038ej5: - movb 4(%esi), %dl -.L039ej4: - movl (%esi), %ecx - jmp .L040ejend -.L041ej3: - movb 2(%esi), %ch - sall $8, %ecx -.L042ej2: - movb 1(%esi), %ch -.L043ej1: - movb (%esi), %cl -.L040ejend: - xorl %ecx, %eax - xorl %edx, %ebx - movl %eax, 16(%esp) - movl %ebx, 20(%esp) - call DES_encrypt3 - movl 16(%esp), %eax - movl 20(%esp), %ebx - movl %eax, (%edi) - movl %ebx, 4(%edi) - jmp .L033finish -.align 16 -.L030decrypt: - andl $4294967288, %ebp - movl 24(%esp), %eax - movl 28(%esp), %ebx - jz .L044decrypt_finish -.L045decrypt_loop: - movl (%esi), %eax - movl 4(%esi), %ebx - movl %eax, 16(%esp) - movl %ebx, 20(%esp) - call DES_decrypt3 - movl 16(%esp), %eax - movl 20(%esp), %ebx - movl 24(%esp), %ecx - movl 28(%esp), %edx - xorl %eax, %ecx - xorl %ebx, %edx - movl (%esi), %eax - movl 4(%esi), %ebx - movl %ecx, (%edi) - movl %edx, 4(%edi) - movl %eax, 24(%esp) - movl %ebx, 28(%esp) - addl $8, %esi - addl $8, %edi - subl $8, %ebp - jnz .L045decrypt_loop -.L044decrypt_finish: - movl 60(%esp), %ebp - andl $7, %ebp - jz .L033finish - movl (%esi), %eax - movl 4(%esi), %ebx - movl %eax, 16(%esp) - movl %ebx, 20(%esp) - call DES_decrypt3 - movl 16(%esp), %eax - movl 20(%esp), %ebx - movl 24(%esp), %ecx - movl 28(%esp), %edx - xorl %eax, %ecx - xorl %ebx, %edx - movl (%esi), %eax - movl 4(%esi), %ebx -.L046dj7: - rorl $16, %edx - movb %dl, 6(%edi) - shrl $16, %edx -.L047dj6: - movb %dh, 5(%edi) -.L048dj5: - movb %dl, 4(%edi) -.L049dj4: - movl %ecx, (%edi) - jmp .L050djend -.L051dj3: - rorl $16, %ecx - movb %cl, 2(%edi) - sall $16, %ecx -.L052dj2: - movb %ch, 1(%esi) -.L053dj1: - movb %cl, (%esi) -.L050djend: - jmp .L033finish -.align 16 -.L033finish: - movl 76(%esp), %ecx - addl $32, %esp - movl %eax, (%ecx) - movl %ebx, 4(%ecx) - popl %edi - popl %esi - popl %ebx - popl %ebp - ret -.align 16 -.L035cbc_enc_jmp_table: - .long 0 - .long .L043ej1-.L034PIC_point - .long .L042ej2-.L034PIC_point - .long .L041ej3-.L034PIC_point - .long .L039ej4-.L034PIC_point - .long .L038ej5-.L034PIC_point - .long .L037ej6-.L034PIC_point - .long .L036ej7-.L034PIC_point -.L_DES_ede3_cbc_encrypt_end: - .size DES_ede3_cbc_encrypt,.L_DES_ede3_cbc_encrypt_end-DES_ede3_cbc_encrypt -.ident "desasm.pl" diff --git a/fips-1.0/sha/asm/fips-sx86-elf.s b/fips-1.0/sha/asm/fips-sx86-elf.s deleted file mode 100644 index 2a4d98791d..0000000000 --- a/fips-1.0/sha/asm/fips-sx86-elf.s +++ /dev/null @@ -1,1568 +0,0 @@ - - - - - - - .file "sha1-586.s" - .version "01.01" -gcc2_compiled.: -.text - .align 16 -.globl sha1_block_asm_data_order - .type sha1_block_asm_data_order,@function -sha1_block_asm_data_order: - movl 12(%esp), %ecx - pushl %esi - sall $6, %ecx - movl 12(%esp), %esi - pushl %ebp - addl %esi, %ecx - pushl %ebx - movl 16(%esp), %ebp - pushl %edi - movl 12(%ebp), %edx - subl $108, %esp - movl 16(%ebp), %edi - movl 8(%ebp), %ebx - movl %ecx, 68(%esp) - -.L000start: - - movl (%esi), %eax - movl 4(%esi), %ecx - - xchgb %al, %ah - rorl $16, %eax - xchgb %al, %ah - - xchgb %cl, %ch - rorl $16, %ecx - xchgb %cl, %ch - movl %eax, (%esp) - movl %ecx, 4(%esp) - movl 8(%esi), %eax - movl 12(%esi), %ecx - - xchgb %al, %ah - rorl $16, %eax - xchgb %al, %ah - - xchgb %cl, %ch - rorl $16, %ecx - xchgb %cl, %ch - movl %eax, 8(%esp) - movl %ecx, 12(%esp) - movl 16(%esi), %eax - movl 20(%esi), %ecx - - xchgb %al, %ah - rorl $16, %eax - xchgb %al, %ah - - xchgb %cl, %ch - rorl $16, %ecx - xchgb %cl, %ch - movl %eax, 16(%esp) - movl %ecx, 20(%esp) - movl 24(%esi), %eax - movl 28(%esi), %ecx - - xchgb %al, %ah - rorl $16, %eax - xchgb %al, %ah - - xchgb %cl, %ch - rorl $16, %ecx - xchgb %cl, %ch - movl %eax, 24(%esp) - movl %ecx, 28(%esp) - movl 32(%esi), %eax - movl 36(%esi), %ecx - - xchgb %al, %ah - rorl $16, %eax - xchgb %al, %ah - - xchgb %cl, %ch - rorl $16, %ecx - xchgb %cl, %ch - movl %eax, 32(%esp) - movl %ecx, 36(%esp) - movl 40(%esi), %eax - movl 44(%esi), %ecx - - xchgb %al, %ah - rorl $16, %eax - xchgb %al, %ah - - xchgb %cl, %ch - rorl $16, %ecx - xchgb %cl, %ch - movl %eax, 40(%esp) - movl %ecx, 44(%esp) - movl 48(%esi), %eax - movl 52(%esi), %ecx - - xchgb %al, %ah - rorl $16, %eax - xchgb %al, %ah - - xchgb %cl, %ch - rorl $16, %ecx - xchgb %cl, %ch - movl %eax, 48(%esp) - movl %ecx, 52(%esp) - movl 56(%esi), %eax - movl 60(%esi), %ecx - - xchgb %al, %ah - rorl $16, %eax - xchgb %al, %ah - - xchgb %cl, %ch - rorl $16, %ecx - xchgb %cl, %ch - movl %eax, 56(%esp) - movl %ecx, 60(%esp) - - - movl %esi, 132(%esp) -.L001shortcut: - - - movl (%ebp), %eax - movl 4(%ebp), %ecx - - movl %eax, %ebp - movl %ebx, %esi - roll $5, %ebp - xorl %edx, %esi - andl %ecx, %esi - rorl $2, %ecx - addl %edi, %ebp - movl (%esp), %edi - xorl %edx, %esi - leal 1518500249(%ebp,%edi,1),%ebp - addl %ebp, %esi - - movl %esi, %ebp - movl %ecx, %edi - roll $5, %ebp - xorl %ebx, %edi - andl %eax, %edi - rorl $2, %eax - addl %edx, %ebp - movl 4(%esp), %edx - xorl %ebx, %edi - leal 1518500249(%ebp,%edx,1),%ebp - addl %ebp, %edi - - movl %edi, %ebp - movl %eax, %edx - roll $5, %ebp - xorl %ecx, %edx - andl %esi, %edx - rorl $2, %esi - addl %ebx, %ebp - movl 8(%esp), %ebx - xorl %ecx, %edx - leal 1518500249(%ebp,%ebx,1),%ebp - addl %ebp, %edx - - movl %edx, %ebp - movl %esi, %ebx - roll $5, %ebp - xorl %eax, %ebx - andl %edi, %ebx - rorl $2, %edi - addl %ecx, %ebp - movl 12(%esp), %ecx - xorl %eax, %ebx - leal 1518500249(%ebp,%ecx,1),%ebp - addl %ebp, %ebx - - movl %ebx, %ebp - movl %edi, %ecx - roll $5, %ebp - xorl %esi, %ecx - andl %edx, %ecx - rorl $2, %edx - addl %eax, %ebp - movl 16(%esp), %eax - xorl %esi, %ecx - leal 1518500249(%ebp,%eax,1),%ebp - addl %ebp, %ecx - - movl %ecx, %ebp - movl %edx, %eax - roll $5, %ebp - xorl %edi, %eax - andl %ebx, %eax - rorl $2, %ebx - addl %esi, %ebp - movl 20(%esp), %esi - xorl %edi, %eax - leal 1518500249(%ebp,%esi,1),%ebp - addl %ebp, %eax - - movl %eax, %ebp - movl %ebx, %esi - roll $5, %ebp - xorl %edx, %esi - andl %ecx, %esi - rorl $2, %ecx - addl %edi, %ebp - movl 24(%esp), %edi - xorl %edx, %esi - leal 1518500249(%ebp,%edi,1),%ebp - addl %ebp, %esi - - movl %esi, %ebp - movl %ecx, %edi - roll $5, %ebp - xorl %ebx, %edi - andl %eax, %edi - rorl $2, %eax - addl %edx, %ebp - movl 28(%esp), %edx - xorl %ebx, %edi - leal 1518500249(%ebp,%edx,1),%ebp - addl %ebp, %edi - - movl %edi, %ebp - movl %eax, %edx - roll $5, %ebp - xorl %ecx, %edx - andl %esi, %edx - rorl $2, %esi - addl %ebx, %ebp - movl 32(%esp), %ebx - xorl %ecx, %edx - leal 1518500249(%ebp,%ebx,1),%ebp - addl %ebp, %edx - - movl %edx, %ebp - movl %esi, %ebx - roll $5, %ebp - xorl %eax, %ebx - andl %edi, %ebx - rorl $2, %edi - addl %ecx, %ebp - movl 36(%esp), %ecx - xorl %eax, %ebx - leal 1518500249(%ebp,%ecx,1),%ebp - addl %ebp, %ebx - - movl %ebx, %ebp - movl %edi, %ecx - roll $5, %ebp - xorl %esi, %ecx - andl %edx, %ecx - rorl $2, %edx - addl %eax, %ebp - movl 40(%esp), %eax - xorl %esi, %ecx - leal 1518500249(%ebp,%eax,1),%ebp - addl %ebp, %ecx - - movl %ecx, %ebp - movl %edx, %eax - roll $5, %ebp - xorl %edi, %eax - andl %ebx, %eax - rorl $2, %ebx - addl %esi, %ebp - movl 44(%esp), %esi - xorl %edi, %eax - leal 1518500249(%ebp,%esi,1),%ebp - addl %ebp, %eax - - movl %eax, %ebp - movl %ebx, %esi - roll $5, %ebp - xorl %edx, %esi - andl %ecx, %esi - rorl $2, %ecx - addl %edi, %ebp - movl 48(%esp), %edi - xorl %edx, %esi - leal 1518500249(%ebp,%edi,1),%ebp - addl %ebp, %esi - - movl %esi, %ebp - movl %ecx, %edi - roll $5, %ebp - xorl %ebx, %edi - andl %eax, %edi - rorl $2, %eax - addl %edx, %ebp - movl 52(%esp), %edx - xorl %ebx, %edi - leal 1518500249(%ebp,%edx,1),%ebp - addl %ebp, %edi - - movl %edi, %ebp - movl %eax, %edx - roll $5, %ebp - xorl %ecx, %edx - andl %esi, %edx - rorl $2, %esi - addl %ebx, %ebp - movl 56(%esp), %ebx - xorl %ecx, %edx - leal 1518500249(%ebp,%ebx,1),%ebp - addl %ebp, %edx - - movl %edx, %ebp - movl %esi, %ebx - roll $5, %ebp - xorl %eax, %ebx - andl %edi, %ebx - rorl $2, %edi - addl %ecx, %ebp - movl 60(%esp), %ecx - xorl %eax, %ebx - leal 1518500249(%ebp,%ecx,1),%ebp - addl %ebp, %ebx - - movl 8(%esp), %ecx - movl %edi, %ebp - xorl (%esp), %ecx - xorl %esi, %ebp - xorl 32(%esp), %ecx - andl %edx, %ebp - xorl 52(%esp), %ecx - rorl $2, %edx - xorl %esi, %ebp -.byte 209 -.byte 193 - movl %ecx, (%esp) - leal 1518500249(%ecx,%eax,1),%ecx - movl %ebx, %eax - addl %ebp, %ecx - roll $5, %eax - addl %eax, %ecx - - movl 12(%esp), %eax - movl %edx, %ebp - xorl 4(%esp), %eax - xorl %edi, %ebp - xorl 36(%esp), %eax - andl %ebx, %ebp - xorl 56(%esp), %eax - rorl $2, %ebx - xorl %edi, %ebp -.byte 209 -.byte 192 - movl %eax, 4(%esp) - leal 1518500249(%eax,%esi,1),%eax - movl %ecx, %esi - addl %ebp, %eax - roll $5, %esi - addl %esi, %eax - - movl 16(%esp), %esi - movl %ebx, %ebp - xorl 8(%esp), %esi - xorl %edx, %ebp - xorl 40(%esp), %esi - andl %ecx, %ebp - xorl 60(%esp), %esi - rorl $2, %ecx - xorl %edx, %ebp -.byte 209 -.byte 198 - movl %esi, 8(%esp) - leal 1518500249(%esi,%edi,1),%esi - movl %eax, %edi - addl %ebp, %esi - roll $5, %edi - addl %edi, %esi - - movl 20(%esp), %edi - movl %ecx, %ebp - xorl 12(%esp), %edi - xorl %ebx, %ebp - xorl 44(%esp), %edi - andl %eax, %ebp - xorl (%esp), %edi - rorl $2, %eax - xorl %ebx, %ebp -.byte 209 -.byte 199 - movl %edi, 12(%esp) - leal 1518500249(%edi,%edx,1),%edi - movl %esi, %edx - addl %ebp, %edi - roll $5, %edx - addl %edx, %edi - - movl 16(%esp), %edx - movl %esi, %ebp - xorl 24(%esp), %edx - rorl $2, %esi - xorl 48(%esp), %edx - xorl %eax, %ebp - xorl 4(%esp), %edx - xorl %ecx, %ebp -.byte 209 -.byte 194 - movl %edx, 16(%esp) - leal 1859775393(%edx,%ebx,1),%edx - movl %edi, %ebx - roll $5, %ebx - addl %ebp, %edx - addl %ebx, %edx - - movl 20(%esp), %ebx - movl %edi, %ebp - xorl 28(%esp), %ebx - rorl $2, %edi - xorl 52(%esp), %ebx - xorl %esi, %ebp - xorl 8(%esp), %ebx - xorl %eax, %ebp -.byte 209 -.byte 195 - movl %ebx, 20(%esp) - leal 1859775393(%ebx,%ecx,1),%ebx - movl %edx, %ecx - roll $5, %ecx - addl %ebp, %ebx - addl %ecx, %ebx - - movl 24(%esp), %ecx - movl %edx, %ebp - xorl 32(%esp), %ecx - rorl $2, %edx - xorl 56(%esp), %ecx - xorl %edi, %ebp - xorl 12(%esp), %ecx - xorl %esi, %ebp -.byte 209 -.byte 193 - movl %ecx, 24(%esp) - leal 1859775393(%ecx,%eax,1),%ecx - movl %ebx, %eax - roll $5, %eax - addl %ebp, %ecx - addl %eax, %ecx - - movl 28(%esp), %eax - movl %ebx, %ebp - xorl 36(%esp), %eax - rorl $2, %ebx - xorl 60(%esp), %eax - xorl %edx, %ebp - xorl 16(%esp), %eax - xorl %edi, %ebp -.byte 209 -.byte 192 - movl %eax, 28(%esp) - leal 1859775393(%eax,%esi,1),%eax - movl %ecx, %esi - roll $5, %esi - addl %ebp, %eax - addl %esi, %eax - - movl 32(%esp), %esi - movl %ecx, %ebp - xorl 40(%esp), %esi - rorl $2, %ecx - xorl (%esp), %esi - xorl %ebx, %ebp - xorl 20(%esp), %esi - xorl %edx, %ebp -.byte 209 -.byte 198 - movl %esi, 32(%esp) - leal 1859775393(%esi,%edi,1),%esi - movl %eax, %edi - roll $5, %edi - addl %ebp, %esi - addl %edi, %esi - - movl 36(%esp), %edi - movl %eax, %ebp - xorl 44(%esp), %edi - rorl $2, %eax - xorl 4(%esp), %edi - xorl %ecx, %ebp - xorl 24(%esp), %edi - xorl %ebx, %ebp -.byte 209 -.byte 199 - movl %edi, 36(%esp) - leal 1859775393(%edi,%edx,1),%edi - movl %esi, %edx - roll $5, %edx - addl %ebp, %edi - addl %edx, %edi - - movl 40(%esp), %edx - movl %esi, %ebp - xorl 48(%esp), %edx - rorl $2, %esi - xorl 8(%esp), %edx - xorl %eax, %ebp - xorl 28(%esp), %edx - xorl %ecx, %ebp -.byte 209 -.byte 194 - movl %edx, 40(%esp) - leal 1859775393(%edx,%ebx,1),%edx - movl %edi, %ebx - roll $5, %ebx - addl %ebp, %edx - addl %ebx, %edx - - movl 44(%esp), %ebx - movl %edi, %ebp - xorl 52(%esp), %ebx - rorl $2, %edi - xorl 12(%esp), %ebx - xorl %esi, %ebp - xorl 32(%esp), %ebx - xorl %eax, %ebp -.byte 209 -.byte 195 - movl %ebx, 44(%esp) - leal 1859775393(%ebx,%ecx,1),%ebx - movl %edx, %ecx - roll $5, %ecx - addl %ebp, %ebx - addl %ecx, %ebx - - movl 48(%esp), %ecx - movl %edx, %ebp - xorl 56(%esp), %ecx - rorl $2, %edx - xorl 16(%esp), %ecx - xorl %edi, %ebp - xorl 36(%esp), %ecx - xorl %esi, %ebp -.byte 209 -.byte 193 - movl %ecx, 48(%esp) - leal 1859775393(%ecx,%eax,1),%ecx - movl %ebx, %eax - roll $5, %eax - addl %ebp, %ecx - addl %eax, %ecx - - movl 52(%esp), %eax - movl %ebx, %ebp - xorl 60(%esp), %eax - rorl $2, %ebx - xorl 20(%esp), %eax - xorl %edx, %ebp - xorl 40(%esp), %eax - xorl %edi, %ebp -.byte 209 -.byte 192 - movl %eax, 52(%esp) - leal 1859775393(%eax,%esi,1),%eax - movl %ecx, %esi - roll $5, %esi - addl %ebp, %eax - addl %esi, %eax - - movl 56(%esp), %esi - movl %ecx, %ebp - xorl (%esp), %esi - rorl $2, %ecx - xorl 24(%esp), %esi - xorl %ebx, %ebp - xorl 44(%esp), %esi - xorl %edx, %ebp -.byte 209 -.byte 198 - movl %esi, 56(%esp) - leal 1859775393(%esi,%edi,1),%esi - movl %eax, %edi - roll $5, %edi - addl %ebp, %esi - addl %edi, %esi - - movl 60(%esp), %edi - movl %eax, %ebp - xorl 4(%esp), %edi - rorl $2, %eax - xorl 28(%esp), %edi - xorl %ecx, %ebp - xorl 48(%esp), %edi - xorl %ebx, %ebp -.byte 209 -.byte 199 - movl %edi, 60(%esp) - leal 1859775393(%edi,%edx,1),%edi - movl %esi, %edx - roll $5, %edx - addl %ebp, %edi - addl %edx, %edi - - movl (%esp), %edx - movl %esi, %ebp - xorl 8(%esp), %edx - rorl $2, %esi - xorl 32(%esp), %edx - xorl %eax, %ebp - xorl 52(%esp), %edx - xorl %ecx, %ebp -.byte 209 -.byte 194 - movl %edx, (%esp) - leal 1859775393(%edx,%ebx,1),%edx - movl %edi, %ebx - roll $5, %ebx - addl %ebp, %edx - addl %ebx, %edx - - movl 4(%esp), %ebx - movl %edi, %ebp - xorl 12(%esp), %ebx - rorl $2, %edi - xorl 36(%esp), %ebx - xorl %esi, %ebp - xorl 56(%esp), %ebx - xorl %eax, %ebp -.byte 209 -.byte 195 - movl %ebx, 4(%esp) - leal 1859775393(%ebx,%ecx,1),%ebx - movl %edx, %ecx - roll $5, %ecx - addl %ebp, %ebx - addl %ecx, %ebx - - movl 8(%esp), %ecx - movl %edx, %ebp - xorl 16(%esp), %ecx - rorl $2, %edx - xorl 40(%esp), %ecx - xorl %edi, %ebp - xorl 60(%esp), %ecx - xorl %esi, %ebp -.byte 209 -.byte 193 - movl %ecx, 8(%esp) - leal 1859775393(%ecx,%eax,1),%ecx - movl %ebx, %eax - roll $5, %eax - addl %ebp, %ecx - addl %eax, %ecx - - movl 12(%esp), %eax - movl %ebx, %ebp - xorl 20(%esp), %eax - rorl $2, %ebx - xorl 44(%esp), %eax - xorl %edx, %ebp - xorl (%esp), %eax - xorl %edi, %ebp -.byte 209 -.byte 192 - movl %eax, 12(%esp) - leal 1859775393(%eax,%esi,1),%eax - movl %ecx, %esi - roll $5, %esi - addl %ebp, %eax - addl %esi, %eax - - movl 16(%esp), %esi - movl %ecx, %ebp - xorl 24(%esp), %esi - rorl $2, %ecx - xorl 48(%esp), %esi - xorl %ebx, %ebp - xorl 4(%esp), %esi - xorl %edx, %ebp -.byte 209 -.byte 198 - movl %esi, 16(%esp) - leal 1859775393(%esi,%edi,1),%esi - movl %eax, %edi - roll $5, %edi - addl %ebp, %esi - addl %edi, %esi - - movl 20(%esp), %edi - movl %eax, %ebp - xorl 28(%esp), %edi - rorl $2, %eax - xorl 52(%esp), %edi - xorl %ecx, %ebp - xorl 8(%esp), %edi - xorl %ebx, %ebp -.byte 209 -.byte 199 - movl %edi, 20(%esp) - leal 1859775393(%edi,%edx,1),%edi - movl %esi, %edx - roll $5, %edx - addl %ebp, %edi - addl %edx, %edi - - movl 24(%esp), %edx - movl %esi, %ebp - xorl 32(%esp), %edx - rorl $2, %esi - xorl 56(%esp), %edx - xorl %eax, %ebp - xorl 12(%esp), %edx - xorl %ecx, %ebp -.byte 209 -.byte 194 - movl %edx, 24(%esp) - leal 1859775393(%edx,%ebx,1),%edx - movl %edi, %ebx - roll $5, %ebx - addl %ebp, %edx - addl %ebx, %edx - - movl 28(%esp), %ebx - movl %edi, %ebp - xorl 36(%esp), %ebx - rorl $2, %edi - xorl 60(%esp), %ebx - xorl %esi, %ebp - xorl 16(%esp), %ebx - xorl %eax, %ebp -.byte 209 -.byte 195 - movl %ebx, 28(%esp) - leal 1859775393(%ebx,%ecx,1),%ebx - movl %edx, %ecx - roll $5, %ecx - addl %ebp, %ebx - addl %ecx, %ebx - - movl 32(%esp), %ecx - movl %edx, %ebp - xorl 40(%esp), %ecx - orl %edi, %ebp - xorl (%esp), %ecx - andl %esi, %ebp - xorl 20(%esp), %ecx -.byte 209 -.byte 193 - movl %ecx, 32(%esp) - leal 2400959708(%ecx,%eax,1),%ecx - movl %edx, %eax - rorl $2, %edx - andl %edi, %eax - orl %eax, %ebp - movl %ebx, %eax - roll $5, %eax - addl %eax, %ebp - addl %ebp, %ecx - - movl 36(%esp), %eax - movl %ebx, %ebp - xorl 44(%esp), %eax - orl %edx, %ebp - xorl 4(%esp), %eax - andl %edi, %ebp - xorl 24(%esp), %eax -.byte 209 -.byte 192 - movl %eax, 36(%esp) - leal 2400959708(%eax,%esi,1),%eax - movl %ebx, %esi - rorl $2, %ebx - andl %edx, %esi - orl %esi, %ebp - movl %ecx, %esi - roll $5, %esi - addl %esi, %ebp - addl %ebp, %eax - - movl 40(%esp), %esi - movl %ecx, %ebp - xorl 48(%esp), %esi - orl %ebx, %ebp - xorl 8(%esp), %esi - andl %edx, %ebp - xorl 28(%esp), %esi -.byte 209 -.byte 198 - movl %esi, 40(%esp) - leal 2400959708(%esi,%edi,1),%esi - movl %ecx, %edi - rorl $2, %ecx - andl %ebx, %edi - orl %edi, %ebp - movl %eax, %edi - roll $5, %edi - addl %edi, %ebp - addl %ebp, %esi - - movl 44(%esp), %edi - movl %eax, %ebp - xorl 52(%esp), %edi - orl %ecx, %ebp - xorl 12(%esp), %edi - andl %ebx, %ebp - xorl 32(%esp), %edi -.byte 209 -.byte 199 - movl %edi, 44(%esp) - leal 2400959708(%edi,%edx,1),%edi - movl %eax, %edx - rorl $2, %eax - andl %ecx, %edx - orl %edx, %ebp - movl %esi, %edx - roll $5, %edx - addl %edx, %ebp - addl %ebp, %edi - - movl 48(%esp), %edx - movl %esi, %ebp - xorl 56(%esp), %edx - orl %eax, %ebp - xorl 16(%esp), %edx - andl %ecx, %ebp - xorl 36(%esp), %edx -.byte 209 -.byte 194 - movl %edx, 48(%esp) - leal 2400959708(%edx,%ebx,1),%edx - movl %esi, %ebx - rorl $2, %esi - andl %eax, %ebx - orl %ebx, %ebp - movl %edi, %ebx - roll $5, %ebx - addl %ebx, %ebp - addl %ebp, %edx - - movl 52(%esp), %ebx - movl %edi, %ebp - xorl 60(%esp), %ebx - orl %esi, %ebp - xorl 20(%esp), %ebx - andl %eax, %ebp - xorl 40(%esp), %ebx -.byte 209 -.byte 195 - movl %ebx, 52(%esp) - leal 2400959708(%ebx,%ecx,1),%ebx - movl %edi, %ecx - rorl $2, %edi - andl %esi, %ecx - orl %ecx, %ebp - movl %edx, %ecx - roll $5, %ecx - addl %ecx, %ebp - addl %ebp, %ebx - - movl 56(%esp), %ecx - movl %edx, %ebp - xorl (%esp), %ecx - orl %edi, %ebp - xorl 24(%esp), %ecx - andl %esi, %ebp - xorl 44(%esp), %ecx -.byte 209 -.byte 193 - movl %ecx, 56(%esp) - leal 2400959708(%ecx,%eax,1),%ecx - movl %edx, %eax - rorl $2, %edx - andl %edi, %eax - orl %eax, %ebp - movl %ebx, %eax - roll $5, %eax - addl %eax, %ebp - addl %ebp, %ecx - - movl 60(%esp), %eax - movl %ebx, %ebp - xorl 4(%esp), %eax - orl %edx, %ebp - xorl 28(%esp), %eax - andl %edi, %ebp - xorl 48(%esp), %eax -.byte 209 -.byte 192 - movl %eax, 60(%esp) - leal 2400959708(%eax,%esi,1),%eax - movl %ebx, %esi - rorl $2, %ebx - andl %edx, %esi - orl %esi, %ebp - movl %ecx, %esi - roll $5, %esi - addl %esi, %ebp - addl %ebp, %eax - - movl (%esp), %esi - movl %ecx, %ebp - xorl 8(%esp), %esi - orl %ebx, %ebp - xorl 32(%esp), %esi - andl %edx, %ebp - xorl 52(%esp), %esi -.byte 209 -.byte 198 - movl %esi, (%esp) - leal 2400959708(%esi,%edi,1),%esi - movl %ecx, %edi - rorl $2, %ecx - andl %ebx, %edi - orl %edi, %ebp - movl %eax, %edi - roll $5, %edi - addl %edi, %ebp - addl %ebp, %esi - - movl 4(%esp), %edi - movl %eax, %ebp - xorl 12(%esp), %edi - orl %ecx, %ebp - xorl 36(%esp), %edi - andl %ebx, %ebp - xorl 56(%esp), %edi -.byte 209 -.byte 199 - movl %edi, 4(%esp) - leal 2400959708(%edi,%edx,1),%edi - movl %eax, %edx - rorl $2, %eax - andl %ecx, %edx - orl %edx, %ebp - movl %esi, %edx - roll $5, %edx - addl %edx, %ebp - addl %ebp, %edi - - movl 8(%esp), %edx - movl %esi, %ebp - xorl 16(%esp), %edx - orl %eax, %ebp - xorl 40(%esp), %edx - andl %ecx, %ebp - xorl 60(%esp), %edx -.byte 209 -.byte 194 - movl %edx, 8(%esp) - leal 2400959708(%edx,%ebx,1),%edx - movl %esi, %ebx - rorl $2, %esi - andl %eax, %ebx - orl %ebx, %ebp - movl %edi, %ebx - roll $5, %ebx - addl %ebx, %ebp - addl %ebp, %edx - - movl 12(%esp), %ebx - movl %edi, %ebp - xorl 20(%esp), %ebx - orl %esi, %ebp - xorl 44(%esp), %ebx - andl %eax, %ebp - xorl (%esp), %ebx -.byte 209 -.byte 195 - movl %ebx, 12(%esp) - leal 2400959708(%ebx,%ecx,1),%ebx - movl %edi, %ecx - rorl $2, %edi - andl %esi, %ecx - orl %ecx, %ebp - movl %edx, %ecx - roll $5, %ecx - addl %ecx, %ebp - addl %ebp, %ebx - - movl 16(%esp), %ecx - movl %edx, %ebp - xorl 24(%esp), %ecx - orl %edi, %ebp - xorl 48(%esp), %ecx - andl %esi, %ebp - xorl 4(%esp), %ecx -.byte 209 -.byte 193 - movl %ecx, 16(%esp) - leal 2400959708(%ecx,%eax,1),%ecx - movl %edx, %eax - rorl $2, %edx - andl %edi, %eax - orl %eax, %ebp - movl %ebx, %eax - roll $5, %eax - addl %eax, %ebp - addl %ebp, %ecx - - movl 20(%esp), %eax - movl %ebx, %ebp - xorl 28(%esp), %eax - orl %edx, %ebp - xorl 52(%esp), %eax - andl %edi, %ebp - xorl 8(%esp), %eax -.byte 209 -.byte 192 - movl %eax, 20(%esp) - leal 2400959708(%eax,%esi,1),%eax - movl %ebx, %esi - rorl $2, %ebx - andl %edx, %esi - orl %esi, %ebp - movl %ecx, %esi - roll $5, %esi - addl %esi, %ebp - addl %ebp, %eax - - movl 24(%esp), %esi - movl %ecx, %ebp - xorl 32(%esp), %esi - orl %ebx, %ebp - xorl 56(%esp), %esi - andl %edx, %ebp - xorl 12(%esp), %esi -.byte 209 -.byte 198 - movl %esi, 24(%esp) - leal 2400959708(%esi,%edi,1),%esi - movl %ecx, %edi - rorl $2, %ecx - andl %ebx, %edi - orl %edi, %ebp - movl %eax, %edi - roll $5, %edi - addl %edi, %ebp - addl %ebp, %esi - - movl 28(%esp), %edi - movl %eax, %ebp - xorl 36(%esp), %edi - orl %ecx, %ebp - xorl 60(%esp), %edi - andl %ebx, %ebp - xorl 16(%esp), %edi -.byte 209 -.byte 199 - movl %edi, 28(%esp) - leal 2400959708(%edi,%edx,1),%edi - movl %eax, %edx - rorl $2, %eax - andl %ecx, %edx - orl %edx, %ebp - movl %esi, %edx - roll $5, %edx - addl %edx, %ebp - addl %ebp, %edi - - movl 32(%esp), %edx - movl %esi, %ebp - xorl 40(%esp), %edx - orl %eax, %ebp - xorl (%esp), %edx - andl %ecx, %ebp - xorl 20(%esp), %edx -.byte 209 -.byte 194 - movl %edx, 32(%esp) - leal 2400959708(%edx,%ebx,1),%edx - movl %esi, %ebx - rorl $2, %esi - andl %eax, %ebx - orl %ebx, %ebp - movl %edi, %ebx - roll $5, %ebx - addl %ebx, %ebp - addl %ebp, %edx - - movl 36(%esp), %ebx - movl %edi, %ebp - xorl 44(%esp), %ebx - orl %esi, %ebp - xorl 4(%esp), %ebx - andl %eax, %ebp - xorl 24(%esp), %ebx -.byte 209 -.byte 195 - movl %ebx, 36(%esp) - leal 2400959708(%ebx,%ecx,1),%ebx - movl %edi, %ecx - rorl $2, %edi - andl %esi, %ecx - orl %ecx, %ebp - movl %edx, %ecx - roll $5, %ecx - addl %ecx, %ebp - addl %ebp, %ebx - - movl 40(%esp), %ecx - movl %edx, %ebp - xorl 48(%esp), %ecx - orl %edi, %ebp - xorl 8(%esp), %ecx - andl %esi, %ebp - xorl 28(%esp), %ecx -.byte 209 -.byte 193 - movl %ecx, 40(%esp) - leal 2400959708(%ecx,%eax,1),%ecx - movl %edx, %eax - rorl $2, %edx - andl %edi, %eax - orl %eax, %ebp - movl %ebx, %eax - roll $5, %eax - addl %eax, %ebp - addl %ebp, %ecx - - movl 44(%esp), %eax - movl %ebx, %ebp - xorl 52(%esp), %eax - orl %edx, %ebp - xorl 12(%esp), %eax - andl %edi, %ebp - xorl 32(%esp), %eax -.byte 209 -.byte 192 - movl %eax, 44(%esp) - leal 2400959708(%eax,%esi,1),%eax - movl %ebx, %esi - rorl $2, %ebx - andl %edx, %esi - orl %esi, %ebp - movl %ecx, %esi - roll $5, %esi - addl %esi, %ebp - addl %ebp, %eax - - movl 48(%esp), %esi - movl %ecx, %ebp - xorl 56(%esp), %esi - rorl $2, %ecx - xorl 16(%esp), %esi - xorl %ebx, %ebp - xorl 36(%esp), %esi - xorl %edx, %ebp -.byte 209 -.byte 198 - movl %esi, 48(%esp) - leal 3395469782(%esi,%edi,1),%esi - movl %eax, %edi - roll $5, %edi - addl %ebp, %esi - addl %edi, %esi - - movl 52(%esp), %edi - movl %eax, %ebp - xorl 60(%esp), %edi - rorl $2, %eax - xorl 20(%esp), %edi - xorl %ecx, %ebp - xorl 40(%esp), %edi - xorl %ebx, %ebp -.byte 209 -.byte 199 - movl %edi, 52(%esp) - leal 3395469782(%edi,%edx,1),%edi - movl %esi, %edx - roll $5, %edx - addl %ebp, %edi - addl %edx, %edi - - movl 56(%esp), %edx - movl %esi, %ebp - xorl (%esp), %edx - rorl $2, %esi - xorl 24(%esp), %edx - xorl %eax, %ebp - xorl 44(%esp), %edx - xorl %ecx, %ebp -.byte 209 -.byte 194 - movl %edx, 56(%esp) - leal 3395469782(%edx,%ebx,1),%edx - movl %edi, %ebx - roll $5, %ebx - addl %ebp, %edx - addl %ebx, %edx - - movl 60(%esp), %ebx - movl %edi, %ebp - xorl 4(%esp), %ebx - rorl $2, %edi - xorl 28(%esp), %ebx - xorl %esi, %ebp - xorl 48(%esp), %ebx - xorl %eax, %ebp -.byte 209 -.byte 195 - movl %ebx, 60(%esp) - leal 3395469782(%ebx,%ecx,1),%ebx - movl %edx, %ecx - roll $5, %ecx - addl %ebp, %ebx - addl %ecx, %ebx - - movl (%esp), %ecx - movl %edx, %ebp - xorl 8(%esp), %ecx - rorl $2, %edx - xorl 32(%esp), %ecx - xorl %edi, %ebp - xorl 52(%esp), %ecx - xorl %esi, %ebp -.byte 209 -.byte 193 - movl %ecx, (%esp) - leal 3395469782(%ecx,%eax,1),%ecx - movl %ebx, %eax - roll $5, %eax - addl %ebp, %ecx - addl %eax, %ecx - - movl 4(%esp), %eax - movl %ebx, %ebp - xorl 12(%esp), %eax - rorl $2, %ebx - xorl 36(%esp), %eax - xorl %edx, %ebp - xorl 56(%esp), %eax - xorl %edi, %ebp -.byte 209 -.byte 192 - movl %eax, 4(%esp) - leal 3395469782(%eax,%esi,1),%eax - movl %ecx, %esi - roll $5, %esi - addl %ebp, %eax - addl %esi, %eax - - movl 8(%esp), %esi - movl %ecx, %ebp - xorl 16(%esp), %esi - rorl $2, %ecx - xorl 40(%esp), %esi - xorl %ebx, %ebp - xorl 60(%esp), %esi - xorl %edx, %ebp -.byte 209 -.byte 198 - movl %esi, 8(%esp) - leal 3395469782(%esi,%edi,1),%esi - movl %eax, %edi - roll $5, %edi - addl %ebp, %esi - addl %edi, %esi - - movl 12(%esp), %edi - movl %eax, %ebp - xorl 20(%esp), %edi - rorl $2, %eax - xorl 44(%esp), %edi - xorl %ecx, %ebp - xorl (%esp), %edi - xorl %ebx, %ebp -.byte 209 -.byte 199 - movl %edi, 12(%esp) - leal 3395469782(%edi,%edx,1),%edi - movl %esi, %edx - roll $5, %edx - addl %ebp, %edi - addl %edx, %edi - - movl 16(%esp), %edx - movl %esi, %ebp - xorl 24(%esp), %edx - rorl $2, %esi - xorl 48(%esp), %edx - xorl %eax, %ebp - xorl 4(%esp), %edx - xorl %ecx, %ebp -.byte 209 -.byte 194 - movl %edx, 16(%esp) - leal 3395469782(%edx,%ebx,1),%edx - movl %edi, %ebx - roll $5, %ebx - addl %ebp, %edx - addl %ebx, %edx - - movl 20(%esp), %ebx - movl %edi, %ebp - xorl 28(%esp), %ebx - rorl $2, %edi - xorl 52(%esp), %ebx - xorl %esi, %ebp - xorl 8(%esp), %ebx - xorl %eax, %ebp -.byte 209 -.byte 195 - movl %ebx, 20(%esp) - leal 3395469782(%ebx,%ecx,1),%ebx - movl %edx, %ecx - roll $5, %ecx - addl %ebp, %ebx - addl %ecx, %ebx - - movl 24(%esp), %ecx - movl %edx, %ebp - xorl 32(%esp), %ecx - rorl $2, %edx - xorl 56(%esp), %ecx - xorl %edi, %ebp - xorl 12(%esp), %ecx - xorl %esi, %ebp -.byte 209 -.byte 193 - movl %ecx, 24(%esp) - leal 3395469782(%ecx,%eax,1),%ecx - movl %ebx, %eax - roll $5, %eax - addl %ebp, %ecx - addl %eax, %ecx - - movl 28(%esp), %eax - movl %ebx, %ebp - xorl 36(%esp), %eax - rorl $2, %ebx - xorl 60(%esp), %eax - xorl %edx, %ebp - xorl 16(%esp), %eax - xorl %edi, %ebp -.byte 209 -.byte 192 - movl %eax, 28(%esp) - leal 3395469782(%eax,%esi,1),%eax - movl %ecx, %esi - roll $5, %esi - addl %ebp, %eax - addl %esi, %eax - - movl 32(%esp), %esi - movl %ecx, %ebp - xorl 40(%esp), %esi - rorl $2, %ecx - xorl (%esp), %esi - xorl %ebx, %ebp - xorl 20(%esp), %esi - xorl %edx, %ebp -.byte 209 -.byte 198 - movl %esi, 32(%esp) - leal 3395469782(%esi,%edi,1),%esi - movl %eax, %edi - roll $5, %edi - addl %ebp, %esi - addl %edi, %esi - - movl 36(%esp), %edi - movl %eax, %ebp - xorl 44(%esp), %edi - rorl $2, %eax - xorl 4(%esp), %edi - xorl %ecx, %ebp - xorl 24(%esp), %edi - xorl %ebx, %ebp -.byte 209 -.byte 199 - movl %edi, 36(%esp) - leal 3395469782(%edi,%edx,1),%edi - movl %esi, %edx - roll $5, %edx - addl %ebp, %edi - addl %edx, %edi - - movl 40(%esp), %edx - movl %esi, %ebp - xorl 48(%esp), %edx - rorl $2, %esi - xorl 8(%esp), %edx - xorl %eax, %ebp - xorl 28(%esp), %edx - xorl %ecx, %ebp -.byte 209 -.byte 194 - movl %edx, 40(%esp) - leal 3395469782(%edx,%ebx,1),%edx - movl %edi, %ebx - roll $5, %ebx - addl %ebp, %edx - addl %ebx, %edx - - movl 44(%esp), %ebx - movl %edi, %ebp - xorl 52(%esp), %ebx - rorl $2, %edi - xorl 12(%esp), %ebx - xorl %esi, %ebp - xorl 32(%esp), %ebx - xorl %eax, %ebp -.byte 209 -.byte 195 - movl %ebx, 44(%esp) - leal 3395469782(%ebx,%ecx,1),%ebx - movl %edx, %ecx - roll $5, %ecx - addl %ebp, %ebx - addl %ecx, %ebx - - movl 48(%esp), %ecx - movl %edx, %ebp - xorl 56(%esp), %ecx - rorl $2, %edx - xorl 16(%esp), %ecx - xorl %edi, %ebp - xorl 36(%esp), %ecx - xorl %esi, %ebp -.byte 209 -.byte 193 - movl %ecx, 48(%esp) - leal 3395469782(%ecx,%eax,1),%ecx - movl %ebx, %eax - roll $5, %eax - addl %ebp, %ecx - addl %eax, %ecx - - movl 52(%esp), %eax - movl %ebx, %ebp - xorl 60(%esp), %eax - rorl $2, %ebx - xorl 20(%esp), %eax - xorl %edx, %ebp - xorl 40(%esp), %eax - xorl %edi, %ebp -.byte 209 -.byte 192 - movl %eax, 52(%esp) - leal 3395469782(%eax,%esi,1),%eax - movl %ecx, %esi - roll $5, %esi - addl %ebp, %eax - addl %esi, %eax - - movl 56(%esp), %esi - movl %ecx, %ebp - xorl (%esp), %esi - rorl $2, %ecx - xorl 24(%esp), %esi - xorl %ebx, %ebp - xorl 44(%esp), %esi - xorl %edx, %ebp -.byte 209 -.byte 198 - movl %esi, 56(%esp) - leal 3395469782(%esi,%edi,1),%esi - movl %eax, %edi - roll $5, %edi - addl %ebp, %esi - addl %edi, %esi - - movl 60(%esp), %edi - movl %eax, %ebp - xorl 4(%esp), %edi - rorl $2, %eax - xorl 28(%esp), %edi - xorl %ecx, %ebp - xorl 48(%esp), %edi - xorl %ebx, %ebp -.byte 209 -.byte 199 - movl %edi, 60(%esp) - leal 3395469782(%edi,%edx,1),%edi - movl %esi, %edx - roll $5, %edx - addl %ebp, %edi - addl %edx, %edi - - - movl 128(%esp), %ebp - movl 12(%ebp), %edx - addl %ecx, %edx - movl 4(%ebp), %ecx - addl %esi, %ecx - movl %eax, %esi - movl (%ebp), %eax - movl %edx, 12(%ebp) - addl %edi, %eax - movl 16(%ebp), %edi - addl %ebx, %edi - movl 8(%ebp), %ebx - addl %esi, %ebx - movl %eax, (%ebp) - movl 132(%esp), %esi - movl %ebx, 8(%ebp) - addl $64, %esi - movl 68(%esp), %eax - movl %edi, 16(%ebp) - cmpl %eax, %esi - movl %ecx, 4(%ebp) - jb .L000start - addl $108, %esp - popl %edi - popl %ebx - popl %ebp - popl %esi - ret -.L_sha1_block_asm_data_order_end: - .size sha1_block_asm_data_order,.L_sha1_block_asm_data_order_end-sha1_block_asm_data_order -.ident "desasm.pl" -.text - .align 16 -.globl sha1_block_asm_host_order - .type sha1_block_asm_host_order,@function -sha1_block_asm_host_order: - movl 12(%esp), %ecx - pushl %esi - sall $6, %ecx - movl 12(%esp), %esi - pushl %ebp - addl %esi, %ecx - pushl %ebx - movl 16(%esp), %ebp - pushl %edi - movl 12(%ebp), %edx - subl $108, %esp - movl 16(%ebp), %edi - movl 8(%ebp), %ebx - movl %ecx, 68(%esp) - - movl (%esi), %eax - movl 4(%esi), %ecx - movl %eax, (%esp) - movl %ecx, 4(%esp) - movl 8(%esi), %eax - movl 12(%esi), %ecx - movl %eax, 8(%esp) - movl %ecx, 12(%esp) - movl 16(%esi), %eax - movl 20(%esi), %ecx - movl %eax, 16(%esp) - movl %ecx, 20(%esp) - movl 24(%esi), %eax - movl 28(%esi), %ecx - movl %eax, 24(%esp) - movl %ecx, 28(%esp) - movl 32(%esi), %eax - movl 36(%esi), %ecx - movl %eax, 32(%esp) - movl %ecx, 36(%esp) - movl 40(%esi), %eax - movl 44(%esi), %ecx - movl %eax, 40(%esp) - movl %ecx, 44(%esp) - movl 48(%esi), %eax - movl 52(%esi), %ecx - movl %eax, 48(%esp) - movl %ecx, 52(%esp) - movl 56(%esi), %eax - movl 60(%esi), %ecx - movl %eax, 56(%esp) - movl %ecx, 60(%esp) - jmp .L001shortcut -.L_sha1_block_asm_host_order_end: - .size sha1_block_asm_host_order,.L_sha1_block_asm_host_order_end-sha1_block_asm_host_order -.ident "desasm.pl" -- 2.25.1