+++ /dev/null
-
-
-
-
-
-
- .file "aes-586.s"
-.globl AES_Te
-.text
-.globl _x86_AES_encrypt
-.type _x86_AES_encrypt,@function
-.align 16
-_x86_AES_encrypt:
- movl %edi, 12(%esp)
- xorl (%edi), %eax
- xorl 4(%edi), %ebx
- xorl 8(%edi), %ecx
- xorl 12(%edi), %edx
- movl 240(%edi), %esi
- leal -2(%esi,%esi), %esi
- leal (%edi,%esi,8), %esi
- movl %esi, 16(%esp)
-.align 4
-.L000loop:
- movl %eax, %esi
- andl $255, %esi
- movl (%ebp,%esi,8), %esi
- movzbl %bh, %edi
- xorl 3(%ebp,%edi,8), %esi
- movl %ecx, %edi
- shrl $16, %edi
- andl $255, %edi
- xorl 2(%ebp,%edi,8), %esi
- movl %edx, %edi
- shrl $24, %edi
- xorl 1(%ebp,%edi,8), %esi
- movl %esi, 4(%esp)
-
- movl %ebx, %esi
- andl $255, %esi
- shrl $16, %ebx
- movl (%ebp,%esi,8), %esi
- movzbl %ch, %edi
- xorl 3(%ebp,%edi,8), %esi
- movl %edx, %edi
- shrl $16, %edi
- andl $255, %edi
- xorl 2(%ebp,%edi,8), %esi
- movl %eax, %edi
- shrl $24, %edi
- xorl 1(%ebp,%edi,8), %esi
- movl %esi, 8(%esp)
-
- movl %ecx, %esi
- andl $255, %esi
- shrl $24, %ecx
- movl (%ebp,%esi,8), %esi
- movzbl %dh, %edi
- xorl 3(%ebp,%edi,8), %esi
- movl %eax, %edi
- shrl $16, %edi
- andl $255, %edx
- andl $255, %edi
- xorl 2(%ebp,%edi,8), %esi
- movzbl %bh, %edi
- xorl 1(%ebp,%edi,8), %esi
-
- movl 12(%esp), %edi
- movl (%ebp,%edx,8), %edx
- movzbl %ah, %eax
- xorl 3(%ebp,%eax,8), %edx
- movl 4(%esp), %eax
- andl $255, %ebx
- xorl 2(%ebp,%ebx,8), %edx
- movl 8(%esp), %ebx
- xorl 1(%ebp,%ecx,8), %edx
- movl %esi, %ecx
-
- addl $16, %edi
- xorl (%edi), %eax
- xorl 4(%edi), %ebx
- xorl 8(%edi), %ecx
- xorl 12(%edi), %edx
- cmpl 16(%esp), %edi
- movl %edi, 12(%esp)
- jb .L000loop
- movl %eax, %esi
- andl $255, %esi
- movl 2(%ebp,%esi,8), %esi
- andl $255, %esi
- movzbl %bh, %edi
- movl (%ebp,%edi,8), %edi
- andl $65280, %edi
- xorl %edi, %esi
- movl %ecx, %edi
- shrl $16, %edi
- andl $255, %edi
- movl (%ebp,%edi,8), %edi
- andl $16711680, %edi
- xorl %edi, %esi
- movl %edx, %edi
- shrl $24, %edi
- movl 2(%ebp,%edi,8), %edi
- andl $4278190080, %edi
- xorl %edi, %esi
- movl %esi, 4(%esp)
- movl %ebx, %esi
- andl $255, %esi
- shrl $16, %ebx
- movl 2(%ebp,%esi,8), %esi
- andl $255, %esi
- movzbl %ch, %edi
- movl (%ebp,%edi,8), %edi
- andl $65280, %edi
- xorl %edi, %esi
- movl %edx, %edi
- shrl $16, %edi
- andl $255, %edi
- movl (%ebp,%edi,8), %edi
- andl $16711680, %edi
- xorl %edi, %esi
- movl %eax, %edi
- shrl $24, %edi
- movl 2(%ebp,%edi,8), %edi
- andl $4278190080, %edi
- xorl %edi, %esi
- movl %esi, 8(%esp)
- movl %ecx, %esi
- andl $255, %esi
- shrl $24, %ecx
- movl 2(%ebp,%esi,8), %esi
- andl $255, %esi
- movzbl %dh, %edi
- movl (%ebp,%edi,8), %edi
- andl $65280, %edi
- xorl %edi, %esi
- movl %eax, %edi
- shrl $16, %edi
- andl $255, %edx
- andl $255, %edi
- movl (%ebp,%edi,8), %edi
- andl $16711680, %edi
- xorl %edi, %esi
- movzbl %bh, %edi
- movl 2(%ebp,%edi,8), %edi
- andl $4278190080, %edi
- xorl %edi, %esi
- movl 12(%esp), %edi
- andl $255, %edx
- movl 2(%ebp,%edx,8), %edx
- andl $255, %edx
- movzbl %ah, %eax
- movl (%ebp,%eax,8), %eax
- andl $65280, %eax
- xorl %eax, %edx
- movl 4(%esp), %eax
- andl $255, %ebx
- movl (%ebp,%ebx,8), %ebx
- andl $16711680, %ebx
- xorl %ebx, %edx
- movl 8(%esp), %ebx
- movl 2(%ebp,%ecx,8), %ecx
- andl $4278190080, %ecx
- xorl %ecx, %edx
- movl %esi, %ecx
- addl $16, %edi
- xorl (%edi), %eax
- xorl 4(%edi), %ebx
- xorl 8(%edi), %ecx
- xorl 12(%edi), %edx
- ret
-.align 64
-AES_Te:
- .long 2774754246,2774754246
- .long 2222750968,2222750968
- .long 2574743534,2574743534
- .long 2373680118,2373680118
- .long 234025727,234025727
- .long 3177933782,3177933782
- .long 2976870366,2976870366
- .long 1422247313,1422247313
- .long 1345335392,1345335392
- .long 50397442,50397442
- .long 2842126286,2842126286
- .long 2099981142,2099981142
- .long 436141799,436141799
- .long 1658312629,1658312629
- .long 3870010189,3870010189
- .long 2591454956,2591454956
- .long 1170918031,1170918031
- .long 2642575903,2642575903
- .long 1086966153,1086966153
- .long 2273148410,2273148410
- .long 368769775,368769775
- .long 3948501426,3948501426
- .long 3376891790,3376891790
- .long 200339707,200339707
- .long 3970805057,3970805057
- .long 1742001331,1742001331
- .long 4255294047,4255294047
- .long 3937382213,3937382213
- .long 3214711843,3214711843
- .long 4154762323,4154762323
- .long 2524082916,2524082916
- .long 1539358875,1539358875
- .long 3266819957,3266819957
- .long 486407649,486407649
- .long 2928907069,2928907069
- .long 1780885068,1780885068
- .long 1513502316,1513502316
- .long 1094664062,1094664062
- .long 49805301,49805301
- .long 1338821763,1338821763
- .long 1546925160,1546925160
- .long 4104496465,4104496465
- .long 887481809,887481809
- .long 150073849,150073849
- .long 2473685474,2473685474
- .long 1943591083,1943591083
- .long 1395732834,1395732834
- .long 1058346282,1058346282
- .long 201589768,201589768
- .long 1388824469,1388824469
- .long 1696801606,1696801606
- .long 1589887901,1589887901
- .long 672667696,672667696
- .long 2711000631,2711000631
- .long 251987210,251987210
- .long 3046808111,3046808111
- .long 151455502,151455502
- .long 907153956,907153956
- .long 2608889883,2608889883
- .long 1038279391,1038279391
- .long 652995533,652995533
- .long 1764173646,1764173646
- .long 3451040383,3451040383
- .long 2675275242,2675275242
- .long 453576978,453576978
- .long 2659418909,2659418909
- .long 1949051992,1949051992
- .long 773462580,773462580
- .long 756751158,756751158
- .long 2993581788,2993581788
- .long 3998898868,3998898868
- .long 4221608027,4221608027
- .long 4132590244,4132590244
- .long 1295727478,1295727478
- .long 1641469623,1641469623
- .long 3467883389,3467883389
- .long 2066295122,2066295122
- .long 1055122397,1055122397
- .long 1898917726,1898917726
- .long 2542044179,2542044179
- .long 4115878822,4115878822
- .long 1758581177,1758581177
- .long 0,0
- .long 753790401,753790401
- .long 1612718144,1612718144
- .long 536673507,536673507
- .long 3367088505,3367088505
- .long 3982187446,3982187446
- .long 3194645204,3194645204
- .long 1187761037,1187761037
- .long 3653156455,3653156455
- .long 1262041458,1262041458
- .long 3729410708,3729410708
- .long 3561770136,3561770136
- .long 3898103984,3898103984
- .long 1255133061,1255133061
- .long 1808847035,1808847035
- .long 720367557,720367557
- .long 3853167183,3853167183
- .long 385612781,385612781
- .long 3309519750,3309519750
- .long 3612167578,3612167578
- .long 1429418854,1429418854
- .long 2491778321,2491778321
- .long 3477423498,3477423498
- .long 284817897,284817897
- .long 100794884,100794884
- .long 2172616702,2172616702
- .long 4031795360,4031795360
- .long 1144798328,1144798328
- .long 3131023141,3131023141
- .long 3819481163,3819481163
- .long 4082192802,4082192802
- .long 4272137053,4272137053
- .long 3225436288,3225436288
- .long 2324664069,2324664069
- .long 2912064063,2912064063
- .long 3164445985,3164445985
- .long 1211644016,1211644016
- .long 83228145,83228145
- .long 3753688163,3753688163
- .long 3249976951,3249976951
- .long 1977277103,1977277103
- .long 1663115586,1663115586
- .long 806359072,806359072
- .long 452984805,452984805
- .long 250868733,250868733
- .long 1842533055,1842533055
- .long 1288555905,1288555905
- .long 336333848,336333848
- .long 890442534,890442534
- .long 804056259,804056259
- .long 3781124030,3781124030
- .long 2727843637,2727843637
- .long 3427026056,3427026056
- .long 957814574,957814574
- .long 1472513171,1472513171
- .long 4071073621,4071073621
- .long 2189328124,2189328124
- .long 1195195770,1195195770
- .long 2892260552,2892260552
- .long 3881655738,3881655738
- .long 723065138,723065138
- .long 2507371494,2507371494
- .long 2690670784,2690670784
- .long 2558624025,2558624025
- .long 3511635870,3511635870
- .long 2145180835,2145180835
- .long 1713513028,1713513028
- .long 2116692564,2116692564
- .long 2878378043,2878378043
- .long 2206763019,2206763019
- .long 3393603212,3393603212
- .long 703524551,703524551
- .long 3552098411,3552098411
- .long 1007948840,1007948840
- .long 2044649127,2044649127
- .long 3797835452,3797835452
- .long 487262998,487262998
- .long 1994120109,1994120109
- .long 1004593371,1004593371
- .long 1446130276,1446130276
- .long 1312438900,1312438900
- .long 503974420,503974420
- .long 3679013266,3679013266
- .long 168166924,168166924
- .long 1814307912,1814307912
- .long 3831258296,3831258296
- .long 1573044895,1573044895
- .long 1859376061,1859376061
- .long 4021070915,4021070915
- .long 2791465668,2791465668
- .long 2828112185,2828112185
- .long 2761266481,2761266481
- .long 937747667,937747667
- .long 2339994098,2339994098
- .long 854058965,854058965
- .long 1137232011,1137232011
- .long 1496790894,1496790894
- .long 3077402074,3077402074
- .long 2358086913,2358086913
- .long 1691735473,1691735473
- .long 3528347292,3528347292
- .long 3769215305,3769215305
- .long 3027004632,3027004632
- .long 4199962284,4199962284
- .long 133494003,133494003
- .long 636152527,636152527
- .long 2942657994,2942657994
- .long 2390391540,2390391540
- .long 3920539207,3920539207
- .long 403179536,403179536
- .long 3585784431,3585784431
- .long 2289596656,2289596656
- .long 1864705354,1864705354
- .long 1915629148,1915629148
- .long 605822008,605822008
- .long 4054230615,4054230615
- .long 3350508659,3350508659
- .long 1371981463,1371981463
- .long 602466507,602466507
- .long 2094914977,2094914977
- .long 2624877800,2624877800
- .long 555687742,555687742
- .long 3712699286,3712699286
- .long 3703422305,3703422305
- .long 2257292045,2257292045
- .long 2240449039,2240449039
- .long 2423288032,2423288032
- .long 1111375484,1111375484
- .long 3300242801,3300242801
- .long 2858837708,2858837708
- .long 3628615824,3628615824
- .long 84083462,84083462
- .long 32962295,32962295
- .long 302911004,302911004
- .long 2741068226,2741068226
- .long 1597322602,1597322602
- .long 4183250862,4183250862
- .long 3501832553,3501832553
- .long 2441512471,2441512471
- .long 1489093017,1489093017
- .long 656219450,656219450
- .long 3114180135,3114180135
- .long 954327513,954327513
- .long 335083755,335083755
- .long 3013122091,3013122091
- .long 856756514,856756514
- .long 3144247762,3144247762
- .long 1893325225,1893325225
- .long 2307821063,2307821063
- .long 2811532339,2811532339
- .long 3063651117,3063651117
- .long 572399164,572399164
- .long 2458355477,2458355477
- .long 552200649,552200649
- .long 1238290055,1238290055
- .long 4283782570,4283782570
- .long 2015897680,2015897680
- .long 2061492133,2061492133
- .long 2408352771,2408352771
- .long 4171342169,4171342169
- .long 2156497161,2156497161
- .long 386731290,386731290
- .long 3669999461,3669999461
- .long 837215959,837215959
- .long 3326231172,3326231172
- .long 3093850320,3093850320
- .long 3275833730,3275833730
- .long 2962856233,2962856233
- .long 1999449434,1999449434
- .long 286199582,286199582
- .long 3417354363,3417354363
- .long 4233385128,4233385128
- .long 3602627437,3602627437
- .long 974525996,974525996
- .long 1,2,4,8
- .long 16,32,64,128
- .long 27,54,0,0,0,0,0,0
-.L__x86_AES_encrypt_end:
-.size _x86_AES_encrypt,.L__x86_AES_encrypt_end-_x86_AES_encrypt
-.ident "_x86_AES_encrypt"
-.globl AES_Te
-.text
-.globl AES_encrypt
-.type AES_encrypt,@function
-.align 16
-AES_encrypt:
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
-
- movl 20(%esp), %esi
- movl 28(%esp), %edi
- movl %esp, %eax
- subl $24, %esp
- andl $-64, %esp
- addl $4, %esp
- movl %eax, 16(%esp)
- call .L001pic_point
-.L001pic_point:
- popl %ebp
- leal AES_Te-.L001pic_point(%ebp),%ebp
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl 8(%esi), %ecx
- movl 12(%esi), %edx
- call _x86_AES_encrypt
- movl 16(%esp), %esp
- movl 24(%esp), %esi
- movl %eax, (%esi)
- movl %ebx, 4(%esi)
- movl %ecx, 8(%esi)
- movl %edx, 12(%esi)
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.L_AES_encrypt_end:
-.size AES_encrypt,.L_AES_encrypt_end-AES_encrypt
-.ident "AES_encrypt"
-.globl AES_Td
-.text
-.globl _x86_AES_decrypt
-.type _x86_AES_decrypt,@function
-.align 16
-_x86_AES_decrypt:
- movl %edi, 12(%esp)
- xorl (%edi), %eax
- xorl 4(%edi), %ebx
- xorl 8(%edi), %ecx
- xorl 12(%edi), %edx
- movl 240(%edi), %esi
- leal -2(%esi,%esi), %esi
- leal (%edi,%esi,8), %esi
- movl %esi, 16(%esp)
-.align 4
-.L002loop:
- movl %eax, %esi
- andl $255, %esi
- movl (%ebp,%esi,8), %esi
- movzbl %dh, %edi
- xorl 3(%ebp,%edi,8), %esi
- movl %ecx, %edi
- shrl $16, %edi
- andl $255, %edi
- xorl 2(%ebp,%edi,8), %esi
- movl %ebx, %edi
- shrl $24, %edi
- xorl 1(%ebp,%edi,8), %esi
- movl %esi, 4(%esp)
-
- movl %ebx, %esi
- andl $255, %esi
- movl (%ebp,%esi,8), %esi
- movzbl %ah, %edi
- xorl 3(%ebp,%edi,8), %esi
- movl %edx, %edi
- shrl $16, %edi
- andl $255, %edi
- xorl 2(%ebp,%edi,8), %esi
- movl %ecx, %edi
- shrl $24, %edi
- xorl 1(%ebp,%edi,8), %esi
- movl %esi, 8(%esp)
-
- movl %ecx, %esi
- andl $255, %esi
- movl (%ebp,%esi,8), %esi
- movzbl %bh, %edi
- xorl 3(%ebp,%edi,8), %esi
- movl %eax, %edi
- shrl $16, %edi
- andl $255, %edi
- xorl 2(%ebp,%edi,8), %esi
- movl %edx, %edi
- shrl $24, %edi
- xorl 1(%ebp,%edi,8), %esi
-
- movl 12(%esp), %edi
- andl $255, %edx
- movl (%ebp,%edx,8), %edx
- movzbl %ch, %ecx
- xorl 3(%ebp,%ecx,8), %edx
- movl %esi, %ecx
- shrl $16, %ebx
- andl $255, %ebx
- xorl 2(%ebp,%ebx,8), %edx
- movl 8(%esp), %ebx
- shrl $24, %eax
- xorl 1(%ebp,%eax,8), %edx
- movl 4(%esp), %eax
-
- addl $16, %edi
- xorl (%edi), %eax
- xorl 4(%edi), %ebx
- xorl 8(%edi), %ecx
- xorl 12(%edi), %edx
- cmpl 16(%esp), %edi
- movl %edi, 12(%esp)
- jb .L002loop
- movl %eax, %esi
- andl $255, %esi
- movzbl 2048(%ebp,%esi,1),%esi
- movzbl %dh, %edi
- movzbl 2048(%ebp,%edi,1),%edi
- sall $8, %edi
- xorl %edi, %esi
- movl %ecx, %edi
- shrl $16, %edi
- andl $255, %edi
- movzbl 2048(%ebp,%edi,1),%edi
- sall $16, %edi
- xorl %edi, %esi
- movl %ebx, %edi
- shrl $24, %edi
- movzbl 2048(%ebp,%edi,1),%edi
- sall $24, %edi
- xorl %edi, %esi
- movl %esi, 4(%esp)
- movl %ebx, %esi
- andl $255, %esi
- movzbl 2048(%ebp,%esi,1),%esi
- movzbl %ah, %edi
- movzbl 2048(%ebp,%edi,1),%edi
- sall $8, %edi
- xorl %edi, %esi
- movl %edx, %edi
- shrl $16, %edi
- andl $255, %edi
- movzbl 2048(%ebp,%edi,1),%edi
- sall $16, %edi
- xorl %edi, %esi
- movl %ecx, %edi
- shrl $24, %edi
- movzbl 2048(%ebp,%edi,1),%edi
- sall $24, %edi
- xorl %edi, %esi
- movl %esi, 8(%esp)
- movl %ecx, %esi
- andl $255, %esi
- movzbl 2048(%ebp,%esi,1),%esi
- movzbl %bh, %edi
- movzbl 2048(%ebp,%edi,1),%edi
- sall $8, %edi
- xorl %edi, %esi
- movl %eax, %edi
- shrl $16, %edi
- andl $255, %edi
- movzbl 2048(%ebp,%edi,1),%edi
- sall $16, %edi
- xorl %edi, %esi
- movl %edx, %edi
- shrl $24, %edi
- movzbl 2048(%ebp,%edi,1),%edi
- sall $24, %edi
- xorl %edi, %esi
- movl 12(%esp), %edi
- andl $255, %edx
- movzbl 2048(%ebp,%edx,1),%edx
- movzbl %ch, %ecx
- movzbl 2048(%ebp,%ecx,1),%ecx
- sall $8, %ecx
- xorl %ecx, %edx
- movl %esi, %ecx
- shrl $16, %ebx
- andl $255, %ebx
- movzbl 2048(%ebp,%ebx,1),%ebx
- sall $16, %ebx
- xorl %ebx, %edx
- movl 8(%esp), %ebx
- shrl $24, %eax
- movzbl 2048(%ebp,%eax,1),%eax
- sall $24, %eax
- xorl %eax, %edx
- movl 4(%esp), %eax
- addl $16, %edi
- xorl (%edi), %eax
- xorl 4(%edi), %ebx
- xorl 8(%edi), %ecx
- xorl 12(%edi), %edx
- ret
-.align 64
-AES_Td:
- .long 1353184337,1353184337
- .long 1399144830,1399144830
- .long 3282310938,3282310938
- .long 2522752826,2522752826
- .long 3412831035,3412831035
- .long 4047871263,4047871263
- .long 2874735276,2874735276
- .long 2466505547,2466505547
- .long 1442459680,1442459680
- .long 4134368941,4134368941
- .long 2440481928,2440481928
- .long 625738485,625738485
- .long 4242007375,4242007375
- .long 3620416197,3620416197
- .long 2151953702,2151953702
- .long 2409849525,2409849525
- .long 1230680542,1230680542
- .long 1729870373,1729870373
- .long 2551114309,2551114309
- .long 3787521629,3787521629
- .long 41234371,41234371
- .long 317738113,317738113
- .long 2744600205,2744600205
- .long 3338261355,3338261355
- .long 3881799427,3881799427
- .long 2510066197,2510066197
- .long 3950669247,3950669247
- .long 3663286933,3663286933
- .long 763608788,763608788
- .long 3542185048,3542185048
- .long 694804553,694804553
- .long 1154009486,1154009486
- .long 1787413109,1787413109
- .long 2021232372,2021232372
- .long 1799248025,1799248025
- .long 3715217703,3715217703
- .long 3058688446,3058688446
- .long 397248752,397248752
- .long 1722556617,1722556617
- .long 3023752829,3023752829
- .long 407560035,407560035
- .long 2184256229,2184256229
- .long 1613975959,1613975959
- .long 1165972322,1165972322
- .long 3765920945,3765920945
- .long 2226023355,2226023355
- .long 480281086,480281086
- .long 2485848313,2485848313
- .long 1483229296,1483229296
- .long 436028815,436028815
- .long 2272059028,2272059028
- .long 3086515026,3086515026
- .long 601060267,601060267
- .long 3791801202,3791801202
- .long 1468997603,1468997603
- .long 715871590,715871590
- .long 120122290,120122290
- .long 63092015,63092015
- .long 2591802758,2591802758
- .long 2768779219,2768779219
- .long 4068943920,4068943920
- .long 2997206819,2997206819
- .long 3127509762,3127509762
- .long 1552029421,1552029421
- .long 723308426,723308426
- .long 2461301159,2461301159
- .long 4042393587,4042393587
- .long 2715969870,2715969870
- .long 3455375973,3455375973
- .long 3586000134,3586000134
- .long 526529745,526529745
- .long 2331944644,2331944644
- .long 2639474228,2639474228
- .long 2689987490,2689987490
- .long 853641733,853641733
- .long 1978398372,1978398372
- .long 971801355,971801355
- .long 2867814464,2867814464
- .long 111112542,111112542
- .long 1360031421,1360031421
- .long 4186579262,4186579262
- .long 1023860118,1023860118
- .long 2919579357,2919579357
- .long 1186850381,1186850381
- .long 3045938321,3045938321
- .long 90031217,90031217
- .long 1876166148,1876166148
- .long 4279586912,4279586912
- .long 620468249,620468249
- .long 2548678102,2548678102
- .long 3426959497,3426959497
- .long 2006899047,2006899047
- .long 3175278768,3175278768
- .long 2290845959,2290845959
- .long 945494503,945494503
- .long 3689859193,3689859193
- .long 1191869601,1191869601
- .long 3910091388,3910091388
- .long 3374220536,3374220536
- .long 0,0
- .long 2206629897,2206629897
- .long 1223502642,1223502642
- .long 2893025566,2893025566
- .long 1316117100,1316117100
- .long 4227796733,4227796733
- .long 1446544655,1446544655
- .long 517320253,517320253
- .long 658058550,658058550
- .long 1691946762,1691946762
- .long 564550760,564550760
- .long 3511966619,3511966619
- .long 976107044,976107044
- .long 2976320012,2976320012
- .long 266819475,266819475
- .long 3533106868,3533106868
- .long 2660342555,2660342555
- .long 1338359936,1338359936
- .long 2720062561,2720062561
- .long 1766553434,1766553434
- .long 370807324,370807324
- .long 179999714,179999714
- .long 3844776128,3844776128
- .long 1138762300,1138762300
- .long 488053522,488053522
- .long 185403662,185403662
- .long 2915535858,2915535858
- .long 3114841645,3114841645
- .long 3366526484,3366526484
- .long 2233069911,2233069911
- .long 1275557295,1275557295
- .long 3151862254,3151862254
- .long 4250959779,4250959779
- .long 2670068215,2670068215
- .long 3170202204,3170202204
- .long 3309004356,3309004356
- .long 880737115,880737115
- .long 1982415755,1982415755
- .long 3703972811,3703972811
- .long 1761406390,1761406390
- .long 1676797112,1676797112
- .long 3403428311,3403428311
- .long 277177154,277177154
- .long 1076008723,1076008723
- .long 538035844,538035844
- .long 2099530373,2099530373
- .long 4164795346,4164795346
- .long 288553390,288553390
- .long 1839278535,1839278535
- .long 1261411869,1261411869
- .long 4080055004,4080055004
- .long 3964831245,3964831245
- .long 3504587127,3504587127
- .long 1813426987,1813426987
- .long 2579067049,2579067049
- .long 4199060497,4199060497
- .long 577038663,577038663
- .long 3297574056,3297574056
- .long 440397984,440397984
- .long 3626794326,3626794326
- .long 4019204898,4019204898
- .long 3343796615,3343796615
- .long 3251714265,3251714265
- .long 4272081548,4272081548
- .long 906744984,906744984
- .long 3481400742,3481400742
- .long 685669029,685669029
- .long 646887386,646887386
- .long 2764025151,2764025151
- .long 3835509292,3835509292
- .long 227702864,227702864
- .long 2613862250,2613862250
- .long 1648787028,1648787028
- .long 3256061430,3256061430
- .long 3904428176,3904428176
- .long 1593260334,1593260334
- .long 4121936770,4121936770
- .long 3196083615,3196083615
- .long 2090061929,2090061929
- .long 2838353263,2838353263
- .long 3004310991,3004310991
- .long 999926984,999926984
- .long 2809993232,2809993232
- .long 1852021992,1852021992
- .long 2075868123,2075868123
- .long 158869197,158869197
- .long 4095236462,4095236462
- .long 28809964,28809964
- .long 2828685187,2828685187
- .long 1701746150,1701746150
- .long 2129067946,2129067946
- .long 147831841,147831841
- .long 3873969647,3873969647
- .long 3650873274,3650873274
- .long 3459673930,3459673930
- .long 3557400554,3557400554
- .long 3598495785,3598495785
- .long 2947720241,2947720241
- .long 824393514,824393514
- .long 815048134,815048134
- .long 3227951669,3227951669
- .long 935087732,935087732
- .long 2798289660,2798289660
- .long 2966458592,2966458592
- .long 366520115,366520115
- .long 1251476721,1251476721
- .long 4158319681,4158319681
- .long 240176511,240176511
- .long 804688151,804688151
- .long 2379631990,2379631990
- .long 1303441219,1303441219
- .long 1414376140,1414376140
- .long 3741619940,3741619940
- .long 3820343710,3820343710
- .long 461924940,461924940
- .long 3089050817,3089050817
- .long 2136040774,2136040774
- .long 82468509,82468509
- .long 1563790337,1563790337
- .long 1937016826,1937016826
- .long 776014843,776014843
- .long 1511876531,1511876531
- .long 1389550482,1389550482
- .long 861278441,861278441
- .long 323475053,323475053
- .long 2355222426,2355222426
- .long 2047648055,2047648055
- .long 2383738969,2383738969
- .long 2302415851,2302415851
- .long 3995576782,3995576782
- .long 902390199,902390199
- .long 3991215329,3991215329
- .long 1018251130,1018251130
- .long 1507840668,1507840668
- .long 1064563285,1064563285
- .long 2043548696,2043548696
- .long 3208103795,3208103795
- .long 3939366739,3939366739
- .long 1537932639,1537932639
- .long 342834655,342834655
- .long 2262516856,2262516856
- .long 2180231114,2180231114
- .long 1053059257,1053059257
- .long 741614648,741614648
- .long 1598071746,1598071746
- .long 1925389590,1925389590
- .long 203809468,203809468
- .long 2336832552,2336832552
- .long 1100287487,1100287487
- .long 1895934009,1895934009
- .long 3736275976,3736275976
- .long 2632234200,2632234200
- .long 2428589668,2428589668
- .long 1636092795,1636092795
- .long 1890988757,1890988757
- .long 1952214088,1952214088
- .long 1113045200,1113045200
- .byte 82,9,106,213,48,54,165,56
- .byte 191,64,163,158,129,243,215,251
- .byte 124,227,57,130,155,47,255,135
- .byte 52,142,67,68,196,222,233,203
- .byte 84,123,148,50,166,194,35,61
- .byte 238,76,149,11,66,250,195,78
- .byte 8,46,161,102,40,217,36,178
- .byte 118,91,162,73,109,139,209,37
- .byte 114,248,246,100,134,104,152,22
- .byte 212,164,92,204,93,101,182,146
- .byte 108,112,72,80,253,237,185,218
- .byte 94,21,70,87,167,141,157,132
- .byte 144,216,171,0,140,188,211,10
- .byte 247,228,88,5,184,179,69,6
- .byte 208,44,30,143,202,63,15,2
- .byte 193,175,189,3,1,19,138,107
- .byte 58,145,17,65,79,103,220,234
- .byte 151,242,207,206,240,180,230,115
- .byte 150,172,116,34,231,173,53,133
- .byte 226,249,55,232,28,117,223,110
- .byte 71,241,26,113,29,41,197,137
- .byte 111,183,98,14,170,24,190,27
- .byte 252,86,62,75,198,210,121,32
- .byte 154,219,192,254,120,205,90,244
- .byte 31,221,168,51,136,7,199,49
- .byte 177,18,16,89,39,128,236,95
- .byte 96,81,127,169,25,181,74,13
- .byte 45,229,122,159,147,201,156,239
- .byte 160,224,59,77,174,42,245,176
- .byte 200,235,187,60,131,83,153,97
- .byte 23,43,4,126,186,119,214,38
- .byte 225,105,20,99,85,33,12,125
-.L__x86_AES_decrypt_end:
-.size _x86_AES_decrypt,.L__x86_AES_decrypt_end-_x86_AES_decrypt
-.ident "_x86_AES_decrypt"
-.globl AES_Td
-.text
-.globl AES_decrypt
-.type AES_decrypt,@function
-.align 16
-AES_decrypt:
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
-
- movl 20(%esp), %esi
- movl 28(%esp), %edi
- movl %esp, %eax
- subl $24, %esp
- andl $-64, %esp
- addl $4, %esp
- movl %eax, 16(%esp)
- call .L003pic_point
-.L003pic_point:
- popl %ebp
- leal AES_Td-.L003pic_point(%ebp),%ebp
- leal 2176(%ebp), %ebp
- movl -128(%ebp), %eax
- movl -96(%ebp), %ebx
- movl -64(%ebp), %ecx
- movl -32(%ebp), %edx
- movl (%ebp), %eax
- movl 32(%ebp), %ebx
- movl 64(%ebp), %ecx
- movl 96(%ebp), %edx
- leal -2176(%ebp), %ebp
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl 8(%esi), %ecx
- movl 12(%esi), %edx
- call _x86_AES_decrypt
- movl 16(%esp), %esp
- movl 24(%esp), %esi
- movl %eax, (%esi)
- movl %ebx, 4(%esi)
- movl %ecx, 8(%esi)
- movl %edx, 12(%esi)
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.L_AES_decrypt_end:
-.size AES_decrypt,.L_AES_decrypt_end-AES_decrypt
-.ident "AES_decrypt"
-.globl AES_Te
-.globl AES_Td
-.text
-.globl AES_cbc_encrypt
-.type AES_cbc_encrypt,@function
-.align 16
-AES_cbc_encrypt:
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
-
- movl 28(%esp), %ecx
- cmpl $0, %ecx
- je .L004enc_out
- call .L005pic_point
-.L005pic_point:
- popl %ebp
- pushfl
- cld
- cmpl $0, 44(%esp)
- je .L006DECRYPT
- leal AES_Te-.L005pic_point(%ebp),%ebp
- leal -308(%esp), %edi
- andl $-64, %edi
- movl %ebp, %eax
- leal 2048(%ebp), %ebx
- movl %edi, %edx
- andl $4095, %eax
- andl $4095, %ebx
- andl $4095, %edx
- cmpl %ebx, %edx
- jb .L007te_break_out
- subl %ebx, %edx
- subl %edx, %edi
- jmp .L008te_ok
-.L007te_break_out:
- subl %eax, %edx
- andl $4095, %edx
- addl $320, %edx
- subl %edx, %edi
-.align 4
-.L008te_ok:
- movl 24(%esp), %eax
- movl 28(%esp), %ebx
- movl 36(%esp), %edx
- movl 40(%esp), %esi
- xchgl %edi, %esp
- addl $4, %esp
- movl %edi, 16(%esp)
- movl %eax, 20(%esp)
- movl %ebx, 24(%esp)
- movl %ecx, 28(%esp)
- movl %edx, 32(%esp)
- movl %esi, 36(%esp)
- movl $0, 300(%esp)
- movl %edx, %ebx
- movl $61, %ecx
- subl %ebp, %ebx
- movl %edx, %esi
- andl $4095, %ebx
- leal 60(%esp), %edi
- cmpl $2048, %ebx
- jb .L009do_ecopy
- cmpl $3852, %ebx
- jb .L010skip_ecopy
-.align 4
-.L009do_ecopy:
- movl %edi, 32(%esp)
- .long 2784229001
-.L010skip_ecopy:
- movl %eax, %esi
- movl $16, %edi
-.align 4
-.L011prefetch_te:
- movl (%ebp), %eax
- movl 32(%ebp), %ebx
- movl 64(%ebp), %ecx
- movl 96(%ebp), %edx
- leal 128(%ebp), %ebp
- decl %edi
- jnz .L011prefetch_te
- subl $2048, %ebp
- movl 28(%esp), %ecx
- movl 36(%esp), %edi
- testl $4294967280, %ecx
- jz .L012enc_tail
- movl (%edi), %eax
- movl 4(%edi), %ebx
-.align 4
-.L013enc_loop:
- movl 8(%edi), %ecx
- movl 12(%edi), %edx
- xorl (%esi), %eax
- xorl 4(%esi), %ebx
- xorl 8(%esi), %ecx
- xorl 12(%esi), %edx
- movl 32(%esp), %edi
- call _x86_AES_encrypt
- movl 20(%esp), %esi
- movl 24(%esp), %edi
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- movl %ecx, 8(%edi)
- movl %edx, 12(%edi)
- movl 28(%esp), %ecx
- leal 16(%esi), %esi
- movl %esi, 20(%esp)
- leal 16(%edi), %edx
- movl %edx, 24(%esp)
- subl $16, %ecx
- testl $4294967280, %ecx
- movl %ecx, 28(%esp)
- jnz .L013enc_loop
- testl $15, %ecx
- jnz .L012enc_tail
- movl 36(%esp), %esi
- movl 8(%edi), %ecx
- movl 12(%edi), %edx
- movl %eax, (%esi)
- movl %ebx, 4(%esi)
- movl %ecx, 8(%esi)
- movl %edx, 12(%esi)
- cmpl $0, 300(%esp)
- movl 32(%esp), %edi
- movl 16(%esp), %esp
- je .L014skip_ezero
- movl $60, %ecx
- xorl %eax, %eax
-.align 4
- .long 2884892297
-.L014skip_ezero:
- popfl
-.L004enc_out:
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
- pushfl
-.align 4
-.L012enc_tail:
- pushl %edi
- movl 24(%esp), %edi
- movl $16, %ebx
- subl %ecx, %ebx
- cmpl %esi, %edi
- je .L015enc_in_place
-.align 4
- .long 2767451785
- jmp .L016enc_skip_in_place
-.L015enc_in_place:
- leal (%edi,%ecx), %edi
-.L016enc_skip_in_place:
- movl %ebx, %ecx
- xorl %eax, %eax
-.align 4
- .long 2868115081
- popl %edi
- movl 24(%esp), %esi
- movl (%edi), %eax
- movl 4(%edi), %ebx
- movl $16, 28(%esp)
- jmp .L013enc_loop
-.align 4
-.L006DECRYPT:
- leal AES_Td-.L005pic_point(%ebp),%ebp
- leal -308(%esp), %edi
- andl $-64, %edi
- movl %ebp, %eax
- leal 2304(%ebp), %ebx
- movl %edi, %edx
- andl $4095, %eax
- andl $4095, %ebx
- andl $4095, %edx
- cmpl %ebx, %edx
- jb .L017td_break_out
- subl %ebx, %edx
- subl %edx, %edi
- jmp .L018td_ok
-.L017td_break_out:
- subl %eax, %edx
- andl $4095, %edx
- addl $320, %edx
- subl %edx, %edi
-.align 4
-.L018td_ok:
- movl 24(%esp), %eax
- movl 28(%esp), %ebx
- movl 36(%esp), %edx
- movl 40(%esp), %esi
- xchgl %edi, %esp
- addl $4, %esp
- movl %edi, 16(%esp)
- movl %eax, 20(%esp)
- movl %ebx, 24(%esp)
- movl %ecx, 28(%esp)
- movl %edx, 32(%esp)
- movl %esi, 36(%esp)
- movl $0, 300(%esp)
- movl %edx, %ebx
- movl $61, %ecx
- subl %ebp, %ebx
- movl %edx, %esi
- andl $4095, %ebx
- leal 60(%esp), %edi
- cmpl $2304, %ebx
- jb .L019do_dcopy
- cmpl $3852, %ebx
- jb .L020skip_dcopy
-.align 4
-.L019do_dcopy:
- movl %edi, 32(%esp)
- .long 2784229001
-.L020skip_dcopy:
- movl %eax, %esi
- movl $18, %edi
-.align 4
-.L021prefetch_td:
- movl (%ebp), %eax
- movl 32(%ebp), %ebx
- movl 64(%ebp), %ecx
- movl 96(%ebp), %edx
- leal 128(%ebp), %ebp
- decl %edi
- jnz .L021prefetch_td
- subl $2304, %ebp
- cmpl 24(%esp), %esi
- je .L022dec_in_place
- movl 36(%esp), %edi
- movl %edi, 40(%esp)
-.align 4
-.L023dec_loop:
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl 8(%esi), %ecx
- movl 12(%esi), %edx
- movl 32(%esp), %edi
- call _x86_AES_decrypt
- movl 40(%esp), %edi
- movl 28(%esp), %esi
- xorl (%edi), %eax
- xorl 4(%edi), %ebx
- xorl 8(%edi), %ecx
- xorl 12(%edi), %edx
- subl $16, %esi
- jc .L024dec_partial
- movl %esi, 28(%esp)
- movl 20(%esp), %esi
- movl 24(%esp), %edi
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- movl %ecx, 8(%edi)
- movl %edx, 12(%edi)
- movl %esi, 40(%esp)
- leal 16(%esi), %esi
- movl %esi, 20(%esp)
- leal 16(%edi), %edi
- movl %edi, 24(%esp)
- jnz .L023dec_loop
- movl 40(%esp), %edi
-.L025dec_end:
- movl 36(%esp), %esi
- movl (%edi), %eax
- movl 4(%edi), %ebx
- movl 8(%edi), %ecx
- movl 12(%edi), %edx
- movl %eax, (%esi)
- movl %ebx, 4(%esi)
- movl %ecx, 8(%esi)
- movl %edx, 12(%esi)
- jmp .L026dec_out
-.align 4
-.L024dec_partial:
- leal 44(%esp), %edi
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- movl %ecx, 8(%edi)
- movl %edx, 12(%edi)
- leal 16(%esi), %ecx
- movl %edi, %esi
- movl 24(%esp), %edi
- .long 2767451785
- movl 20(%esp), %edi
- jmp .L025dec_end
-.align 4
-.L022dec_in_place:
-.L027dec_in_place_loop:
- leal 44(%esp), %edi
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl 8(%esi), %ecx
- movl 12(%esi), %edx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- movl %ecx, 8(%edi)
- movl %edx, 12(%edi)
- movl 32(%esp), %edi
- call _x86_AES_decrypt
- movl 36(%esp), %edi
- movl 24(%esp), %esi
- xorl (%edi), %eax
- xorl 4(%edi), %ebx
- xorl 8(%edi), %ecx
- xorl 12(%edi), %edx
- movl %eax, (%esi)
- movl %ebx, 4(%esi)
- movl %ecx, 8(%esi)
- movl %edx, 12(%esi)
- leal 16(%esi), %esi
- movl %esi, 24(%esp)
- leal 44(%esp), %esi
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl 8(%esi), %ecx
- movl 12(%esi), %edx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- movl %ecx, 8(%edi)
- movl %edx, 12(%edi)
- movl 20(%esp), %esi
- leal 16(%esi), %esi
- movl %esi, 20(%esp)
- movl 28(%esp), %ecx
- subl $16, %ecx
- jc .L028dec_in_place_partial
- movl %ecx, 28(%esp)
- jnz .L027dec_in_place_loop
- jmp .L026dec_out
-.align 4
-.L028dec_in_place_partial:
- movl 24(%esp), %edi
- leal 44(%esp), %esi
- leal (%edi,%ecx), %edi
- leal 16(%esi,%ecx), %esi
- negl %ecx
- .long 2767451785
-.align 4
-.L026dec_out:
- cmpl $0, 300(%esp)
- movl 32(%esp), %edi
- movl 16(%esp), %esp
- je .L029skip_dzero
- movl $60, %ecx
- xorl %eax, %eax
-.align 4
- .long 2884892297
-.L029skip_dzero:
- popfl
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.L_AES_cbc_encrypt_end:
-.size AES_cbc_encrypt,.L_AES_cbc_encrypt_end-AES_cbc_encrypt
-.ident "AES_cbc_encrypt"
-.globl AES_Te
-.text
-.globl AES_set_encrypt_key
-.type AES_set_encrypt_key,@function
-.align 16
-AES_set_encrypt_key:
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
-
- movl 20(%esp), %esi
- movl 28(%esp), %edi
- testl $-1, %esi
- jz .L030badpointer
- testl $-1, %edi
- jz .L030badpointer
- call .L031pic_point
-.L031pic_point:
- popl %ebp
- leal AES_Te-.L031pic_point(%ebp),%ebp
- movl 24(%esp), %ecx
- cmpl $128, %ecx
- je .L03210rounds
- cmpl $192, %ecx
- je .L03312rounds
- cmpl $256, %ecx
- je .L03414rounds
- movl $-2, %eax
- jmp .L035exit
-.L03210rounds:
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl 8(%esi), %ecx
- movl 12(%esi), %edx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- movl %ecx, 8(%edi)
- movl %edx, 12(%edi)
- xorl %ecx, %ecx
- jmp .L03610shortcut
-.align 4
-.L03710loop:
- movl (%edi), %eax
- movl 12(%edi), %edx
-.L03610shortcut:
- movzbl %dl, %esi
- movl 2(%ebp,%esi,8), %ebx
- movzbl %dh, %esi
- andl $4278190080, %ebx
- xorl %ebx, %eax
- movl 2(%ebp,%esi,8), %ebx
- shrl $16, %edx
- andl $255, %ebx
- movzbl %dl, %esi
- xorl %ebx, %eax
- movl (%ebp,%esi,8), %ebx
- movzbl %dh, %esi
- andl $65280, %ebx
- xorl %ebx, %eax
- movl (%ebp,%esi,8), %ebx
- andl $16711680, %ebx
- xorl %ebx, %eax
- xorl 2048(%ebp,%ecx,4),%eax
- movl %eax, 16(%edi)
- xorl 4(%edi), %eax
- movl %eax, 20(%edi)
- xorl 8(%edi), %eax
- movl %eax, 24(%edi)
- xorl 12(%edi), %eax
- movl %eax, 28(%edi)
- incl %ecx
- addl $16, %edi
- cmpl $10, %ecx
- jl .L03710loop
- movl $10, 80(%edi)
- xorl %eax, %eax
- jmp .L035exit
-.L03312rounds:
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl 8(%esi), %ecx
- movl 12(%esi), %edx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- movl %ecx, 8(%edi)
- movl %edx, 12(%edi)
- movl 16(%esi), %ecx
- movl 20(%esi), %edx
- movl %ecx, 16(%edi)
- movl %edx, 20(%edi)
- xorl %ecx, %ecx
- jmp .L03812shortcut
-.align 4
-.L03912loop:
- movl (%edi), %eax
- movl 20(%edi), %edx
-.L03812shortcut:
- movzbl %dl, %esi
- movl 2(%ebp,%esi,8), %ebx
- movzbl %dh, %esi
- andl $4278190080, %ebx
- xorl %ebx, %eax
- movl 2(%ebp,%esi,8), %ebx
- shrl $16, %edx
- andl $255, %ebx
- movzbl %dl, %esi
- xorl %ebx, %eax
- movl (%ebp,%esi,8), %ebx
- movzbl %dh, %esi
- andl $65280, %ebx
- xorl %ebx, %eax
- movl (%ebp,%esi,8), %ebx
- andl $16711680, %ebx
- xorl %ebx, %eax
- xorl 2048(%ebp,%ecx,4),%eax
- movl %eax, 24(%edi)
- xorl 4(%edi), %eax
- movl %eax, 28(%edi)
- xorl 8(%edi), %eax
- movl %eax, 32(%edi)
- xorl 12(%edi), %eax
- movl %eax, 36(%edi)
- cmpl $7, %ecx
- je .L04012break
- incl %ecx
- xorl 16(%edi), %eax
- movl %eax, 40(%edi)
- xorl 20(%edi), %eax
- movl %eax, 44(%edi)
- addl $24, %edi
- jmp .L03912loop
-.L04012break:
- movl $12, 72(%edi)
- xorl %eax, %eax
- jmp .L035exit
-.L03414rounds:
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl 8(%esi), %ecx
- movl 12(%esi), %edx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- movl %ecx, 8(%edi)
- movl %edx, 12(%edi)
- movl 16(%esi), %eax
- movl 20(%esi), %ebx
- movl 24(%esi), %ecx
- movl 28(%esi), %edx
- movl %eax, 16(%edi)
- movl %ebx, 20(%edi)
- movl %ecx, 24(%edi)
- movl %edx, 28(%edi)
- xorl %ecx, %ecx
- jmp .L04114shortcut
-.align 4
-.L04214loop:
- movl 28(%edi), %edx
-.L04114shortcut:
- movl (%edi), %eax
- movzbl %dl, %esi
- movl 2(%ebp,%esi,8), %ebx
- movzbl %dh, %esi
- andl $4278190080, %ebx
- xorl %ebx, %eax
- movl 2(%ebp,%esi,8), %ebx
- shrl $16, %edx
- andl $255, %ebx
- movzbl %dl, %esi
- xorl %ebx, %eax
- movl (%ebp,%esi,8), %ebx
- movzbl %dh, %esi
- andl $65280, %ebx
- xorl %ebx, %eax
- movl (%ebp,%esi,8), %ebx
- andl $16711680, %ebx
- xorl %ebx, %eax
- xorl 2048(%ebp,%ecx,4),%eax
- movl %eax, 32(%edi)
- xorl 4(%edi), %eax
- movl %eax, 36(%edi)
- xorl 8(%edi), %eax
- movl %eax, 40(%edi)
- xorl 12(%edi), %eax
- movl %eax, 44(%edi)
- cmpl $6, %ecx
- je .L04314break
- incl %ecx
- movl %eax, %edx
- movl 16(%edi), %eax
- movzbl %dl, %esi
- movl 2(%ebp,%esi,8), %ebx
- movzbl %dh, %esi
- andl $255, %ebx
- xorl %ebx, %eax
- movl (%ebp,%esi,8), %ebx
- shrl $16, %edx
- andl $65280, %ebx
- movzbl %dl, %esi
- xorl %ebx, %eax
- movl (%ebp,%esi,8), %ebx
- movzbl %dh, %esi
- andl $16711680, %ebx
- xorl %ebx, %eax
- movl 2(%ebp,%esi,8), %ebx
- andl $4278190080, %ebx
- xorl %ebx, %eax
- movl %eax, 48(%edi)
- xorl 20(%edi), %eax
- movl %eax, 52(%edi)
- xorl 24(%edi), %eax
- movl %eax, 56(%edi)
- xorl 28(%edi), %eax
- movl %eax, 60(%edi)
- addl $32, %edi
- jmp .L04214loop
-.L04314break:
- movl $14, 48(%edi)
- xorl %eax, %eax
- jmp .L035exit
-.L030badpointer:
- movl $-1, %eax
-.L035exit:
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.L_AES_set_encrypt_key_end:
-.size AES_set_encrypt_key,.L_AES_set_encrypt_key_end-AES_set_encrypt_key
-.ident "AES_set_encrypt_key"
-.globl AES_Td
-.globl AES_Te
-.text
-.globl AES_set_decrypt_key
-.type AES_set_decrypt_key,@function
-.align 16
-AES_set_decrypt_key:
- movl 4(%esp), %eax
- movl 8(%esp), %ecx
- movl 12(%esp), %edx
- subl $12, %esp
- movl %eax, (%esp)
- movl %ecx, 4(%esp)
- movl %edx, 8(%esp)
- call AES_set_encrypt_key
- addl $12, %esp
- cmpl $0, %eax
- je .L044proceed
- ret
-.L044proceed:
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
- movl 28(%esp), %esi
- movl 240(%esi), %ecx
- leal (,%ecx,4), %ecx
- leal (%esi,%ecx,4), %edi
-.align 4
-.L045invert:
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl (%edi), %ecx
- movl 4(%edi), %edx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- movl %ecx, (%esi)
- movl %edx, 4(%esi)
- movl 8(%esi), %eax
- movl 12(%esi), %ebx
- movl 8(%edi), %ecx
- movl 12(%edi), %edx
- movl %eax, 8(%edi)
- movl %ebx, 12(%edi)
- movl %ecx, 8(%esi)
- movl %edx, 12(%esi)
- addl $16, %esi
- subl $16, %edi
- cmpl %edi, %esi
- jne .L045invert
- call .L046pic_point
-.L046pic_point:
- popl %ebp
- leal AES_Td-.L046pic_point(%ebp),%edi
- leal AES_Te-.L046pic_point(%ebp),%ebp
- movl 28(%esp), %esi
- movl 240(%esi), %ecx
- decl %ecx
-.align 4
-.L047permute:
- addl $16, %esi
- movl (%esi), %eax
- movl %eax, %edx
- movzbl %ah, %ebx
- shrl $16, %edx
- andl $255, %eax
- movzbl 2(%ebp,%eax,8), %eax
- movzbl 2(%ebp,%ebx,8), %ebx
- movl (%edi,%eax,8), %eax
- xorl 3(%edi,%ebx,8), %eax
- movzbl %dh, %ebx
- andl $255, %edx
- movzbl 2(%ebp,%edx,8), %edx
- movzbl 2(%ebp,%ebx,8), %ebx
- xorl 2(%edi,%edx,8), %eax
- xorl 1(%edi,%ebx,8), %eax
- movl %eax, (%esi)
- movl 4(%esi), %eax
- movl %eax, %edx
- movzbl %ah, %ebx
- shrl $16, %edx
- andl $255, %eax
- movzbl 2(%ebp,%eax,8), %eax
- movzbl 2(%ebp,%ebx,8), %ebx
- movl (%edi,%eax,8), %eax
- xorl 3(%edi,%ebx,8), %eax
- movzbl %dh, %ebx
- andl $255, %edx
- movzbl 2(%ebp,%edx,8), %edx
- movzbl 2(%ebp,%ebx,8), %ebx
- xorl 2(%edi,%edx,8), %eax
- xorl 1(%edi,%ebx,8), %eax
- movl %eax, 4(%esi)
- movl 8(%esi), %eax
- movl %eax, %edx
- movzbl %ah, %ebx
- shrl $16, %edx
- andl $255, %eax
- movzbl 2(%ebp,%eax,8), %eax
- movzbl 2(%ebp,%ebx,8), %ebx
- movl (%edi,%eax,8), %eax
- xorl 3(%edi,%ebx,8), %eax
- movzbl %dh, %ebx
- andl $255, %edx
- movzbl 2(%ebp,%edx,8), %edx
- movzbl 2(%ebp,%ebx,8), %ebx
- xorl 2(%edi,%edx,8), %eax
- xorl 1(%edi,%ebx,8), %eax
- movl %eax, 8(%esi)
- movl 12(%esi), %eax
- movl %eax, %edx
- movzbl %ah, %ebx
- shrl $16, %edx
- andl $255, %eax
- movzbl 2(%ebp,%eax,8), %eax
- movzbl 2(%ebp,%ebx,8), %ebx
- movl (%edi,%eax,8), %eax
- xorl 3(%edi,%ebx,8), %eax
- movzbl %dh, %ebx
- andl $255, %edx
- movzbl 2(%ebp,%edx,8), %edx
- movzbl 2(%ebp,%ebx,8), %ebx
- xorl 2(%edi,%edx,8), %eax
- xorl 1(%edi,%ebx,8), %eax
- movl %eax, 12(%esi)
- decl %ecx
- jnz .L047permute
- xorl %eax, %eax
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.L_AES_set_decrypt_key_end:
-.size AES_set_decrypt_key,.L_AES_set_decrypt_key_end-AES_set_decrypt_key
-.ident "AES_set_decrypt_key"
+++ /dev/null
-
-
-
-
-
-
- .file "des-586.s"
- .version "01.01"
-gcc2_compiled.:
-.text
- .align 16
-.globl DES_encrypt1
- .type DES_encrypt1,@function
-DES_encrypt1:
- pushl %esi
- pushl %edi
-
-
- movl 12(%esp), %esi
- xorl %ecx, %ecx
- pushl %ebx
- pushl %ebp
- movl (%esi), %eax
- movl 28(%esp), %ebx
- movl 4(%esi), %edi
-
-
- roll $4, %eax
- movl %eax, %esi
- xorl %edi, %eax
- andl $0xf0f0f0f0, %eax
- xorl %eax, %esi
- xorl %eax, %edi
-
- roll $20, %edi
- movl %edi, %eax
- xorl %esi, %edi
- andl $0xfff0000f, %edi
- xorl %edi, %eax
- xorl %edi, %esi
-
- roll $14, %eax
- movl %eax, %edi
- xorl %esi, %eax
- andl $0x33333333, %eax
- xorl %eax, %edi
- xorl %eax, %esi
-
- roll $22, %esi
- movl %esi, %eax
- xorl %edi, %esi
- andl $0x03fc03fc, %esi
- xorl %esi, %eax
- xorl %esi, %edi
-
- roll $9, %eax
- movl %eax, %esi
- xorl %edi, %eax
- andl $0xaaaaaaaa, %eax
- xorl %eax, %esi
- xorl %eax, %edi
-
-.byte 209
-.byte 199
- .align 8
- call .L000PIC_me_up
-.L000PIC_me_up:
- popl %ebp
- addl $_GLOBAL_OFFSET_TABLE_+[.-.L000PIC_me_up],%ebp
- movl DES_SPtrans@GOT(%ebp),%ebp
- movl 24(%esp), %ecx
- cmpl $0, %ebx
- je .L001start_decrypt
-
-
- movl (%ecx), %eax
- xorl %ebx, %ebx
- movl 4(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 8(%ecx), %eax
- xorl %ebx, %ebx
- movl 12(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 16(%ecx), %eax
- xorl %ebx, %ebx
- movl 20(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 24(%ecx), %eax
- xorl %ebx, %ebx
- movl 28(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 32(%ecx), %eax
- xorl %ebx, %ebx
- movl 36(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 40(%ecx), %eax
- xorl %ebx, %ebx
- movl 44(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 48(%ecx), %eax
- xorl %ebx, %ebx
- movl 52(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 56(%ecx), %eax
- xorl %ebx, %ebx
- movl 60(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 64(%ecx), %eax
- xorl %ebx, %ebx
- movl 68(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 72(%ecx), %eax
- xorl %ebx, %ebx
- movl 76(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 80(%ecx), %eax
- xorl %ebx, %ebx
- movl 84(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 88(%ecx), %eax
- xorl %ebx, %ebx
- movl 92(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 96(%ecx), %eax
- xorl %ebx, %ebx
- movl 100(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 104(%ecx), %eax
- xorl %ebx, %ebx
- movl 108(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 112(%ecx), %eax
- xorl %ebx, %ebx
- movl 116(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 120(%ecx), %eax
- xorl %ebx, %ebx
- movl 124(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
- jmp .L002end
-.L001start_decrypt:
-
-
- movl 120(%ecx), %eax
- xorl %ebx, %ebx
- movl 124(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 112(%ecx), %eax
- xorl %ebx, %ebx
- movl 116(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 104(%ecx), %eax
- xorl %ebx, %ebx
- movl 108(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 96(%ecx), %eax
- xorl %ebx, %ebx
- movl 100(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 88(%ecx), %eax
- xorl %ebx, %ebx
- movl 92(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 80(%ecx), %eax
- xorl %ebx, %ebx
- movl 84(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 72(%ecx), %eax
- xorl %ebx, %ebx
- movl 76(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 64(%ecx), %eax
- xorl %ebx, %ebx
- movl 68(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 56(%ecx), %eax
- xorl %ebx, %ebx
- movl 60(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 48(%ecx), %eax
- xorl %ebx, %ebx
- movl 52(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 40(%ecx), %eax
- xorl %ebx, %ebx
- movl 44(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 32(%ecx), %eax
- xorl %ebx, %ebx
- movl 36(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 24(%ecx), %eax
- xorl %ebx, %ebx
- movl 28(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 16(%ecx), %eax
- xorl %ebx, %ebx
- movl 20(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 8(%ecx), %eax
- xorl %ebx, %ebx
- movl 12(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl (%ecx), %eax
- xorl %ebx, %ebx
- movl 4(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-.L002end:
-
-
- movl 20(%esp), %edx
-.byte 209
-.byte 206
- movl %edi, %eax
- xorl %esi, %edi
- andl $0xaaaaaaaa, %edi
- xorl %edi, %eax
- xorl %edi, %esi
-
- roll $23, %eax
- movl %eax, %edi
- xorl %esi, %eax
- andl $0x03fc03fc, %eax
- xorl %eax, %edi
- xorl %eax, %esi
-
- roll $10, %edi
- movl %edi, %eax
- xorl %esi, %edi
- andl $0x33333333, %edi
- xorl %edi, %eax
- xorl %edi, %esi
-
- roll $18, %esi
- movl %esi, %edi
- xorl %eax, %esi
- andl $0xfff0000f, %esi
- xorl %esi, %edi
- xorl %esi, %eax
-
- roll $12, %edi
- movl %edi, %esi
- xorl %eax, %edi
- andl $0xf0f0f0f0, %edi
- xorl %edi, %esi
- xorl %edi, %eax
-
- rorl $4, %eax
- movl %eax, (%edx)
- movl %esi, 4(%edx)
- popl %ebp
- popl %ebx
- popl %edi
- popl %esi
- ret
-.L_DES_encrypt1_end:
- .size DES_encrypt1,.L_DES_encrypt1_end-DES_encrypt1
-.ident "desasm.pl"
-.text
- .align 16
-.globl DES_encrypt2
- .type DES_encrypt2,@function
-DES_encrypt2:
- pushl %esi
- pushl %edi
-
-
- movl 12(%esp), %eax
- xorl %ecx, %ecx
- pushl %ebx
- pushl %ebp
- movl (%eax), %esi
- movl 28(%esp), %ebx
- roll $3, %esi
- movl 4(%eax), %edi
- roll $3, %edi
- .align 8
- call .L003PIC_me_up
-.L003PIC_me_up:
- popl %ebp
- addl $_GLOBAL_OFFSET_TABLE_+[.-.L003PIC_me_up],%ebp
- movl DES_SPtrans@GOT(%ebp),%ebp
- movl 24(%esp), %ecx
- cmpl $0, %ebx
- je .L004start_decrypt
-
-
- movl (%ecx), %eax
- xorl %ebx, %ebx
- movl 4(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 8(%ecx), %eax
- xorl %ebx, %ebx
- movl 12(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 16(%ecx), %eax
- xorl %ebx, %ebx
- movl 20(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 24(%ecx), %eax
- xorl %ebx, %ebx
- movl 28(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 32(%ecx), %eax
- xorl %ebx, %ebx
- movl 36(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 40(%ecx), %eax
- xorl %ebx, %ebx
- movl 44(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 48(%ecx), %eax
- xorl %ebx, %ebx
- movl 52(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 56(%ecx), %eax
- xorl %ebx, %ebx
- movl 60(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 64(%ecx), %eax
- xorl %ebx, %ebx
- movl 68(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 72(%ecx), %eax
- xorl %ebx, %ebx
- movl 76(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 80(%ecx), %eax
- xorl %ebx, %ebx
- movl 84(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 88(%ecx), %eax
- xorl %ebx, %ebx
- movl 92(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 96(%ecx), %eax
- xorl %ebx, %ebx
- movl 100(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 104(%ecx), %eax
- xorl %ebx, %ebx
- movl 108(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 112(%ecx), %eax
- xorl %ebx, %ebx
- movl 116(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 120(%ecx), %eax
- xorl %ebx, %ebx
- movl 124(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
- jmp .L005end
-.L004start_decrypt:
-
-
- movl 120(%ecx), %eax
- xorl %ebx, %ebx
- movl 124(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 112(%ecx), %eax
- xorl %ebx, %ebx
- movl 116(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 104(%ecx), %eax
- xorl %ebx, %ebx
- movl 108(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 96(%ecx), %eax
- xorl %ebx, %ebx
- movl 100(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 88(%ecx), %eax
- xorl %ebx, %ebx
- movl 92(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 80(%ecx), %eax
- xorl %ebx, %ebx
- movl 84(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 72(%ecx), %eax
- xorl %ebx, %ebx
- movl 76(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 64(%ecx), %eax
- xorl %ebx, %ebx
- movl 68(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 56(%ecx), %eax
- xorl %ebx, %ebx
- movl 60(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 48(%ecx), %eax
- xorl %ebx, %ebx
- movl 52(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 40(%ecx), %eax
- xorl %ebx, %ebx
- movl 44(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 32(%ecx), %eax
- xorl %ebx, %ebx
- movl 36(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 24(%ecx), %eax
- xorl %ebx, %ebx
- movl 28(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl 16(%ecx), %eax
- xorl %ebx, %ebx
- movl 20(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-
-
- movl 8(%ecx), %eax
- xorl %ebx, %ebx
- movl 12(%ecx), %edx
- xorl %esi, %eax
- xorl %ecx, %ecx
- xorl %esi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%edi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%edi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%edi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%edi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%edi
- xorl 0x700(%ebp,%ecx),%edi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%edi
- xorl 0x500(%ebp,%edx),%edi
-
-
- movl (%ecx), %eax
- xorl %ebx, %ebx
- movl 4(%ecx), %edx
- xorl %edi, %eax
- xorl %ecx, %ecx
- xorl %edi, %edx
- andl $0xfcfcfcfc, %eax
- andl $0xcfcfcfcf, %edx
- movb %al, %bl
- movb %ah, %cl
- rorl $4, %edx
- xorl (%ebp,%ebx),%esi
- movb %dl, %bl
- xorl 0x200(%ebp,%ecx),%esi
- movb %dh, %cl
- shrl $16, %eax
- xorl 0x100(%ebp,%ebx),%esi
- movb %ah, %bl
- shrl $16, %edx
- xorl 0x300(%ebp,%ecx),%esi
- movb %dh, %cl
- andl $0xff, %eax
- andl $0xff, %edx
- xorl 0x600(%ebp,%ebx),%esi
- xorl 0x700(%ebp,%ecx),%esi
- movl 24(%esp), %ecx
- xorl 0x400(%ebp,%eax),%esi
- xorl 0x500(%ebp,%edx),%esi
-.L005end:
-
-
- rorl $3, %edi
- movl 20(%esp), %eax
- rorl $3, %esi
- movl %edi, (%eax)
- movl %esi, 4(%eax)
- popl %ebp
- popl %ebx
- popl %edi
- popl %esi
- ret
-.L_DES_encrypt2_end:
- .size DES_encrypt2,.L_DES_encrypt2_end-DES_encrypt2
-.ident "desasm.pl"
-.text
- .align 16
-.globl DES_encrypt3
- .type DES_encrypt3,@function
-DES_encrypt3:
- pushl %ebx
- movl 8(%esp), %ebx
- pushl %ebp
- pushl %esi
- pushl %edi
-
-
- movl (%ebx), %edi
- movl 4(%ebx), %esi
- subl $12, %esp
-
-
- roll $4, %edi
- movl %edi, %edx
- xorl %esi, %edi
- andl $0xf0f0f0f0, %edi
- xorl %edi, %edx
- xorl %edi, %esi
-
- roll $20, %esi
- movl %esi, %edi
- xorl %edx, %esi
- andl $0xfff0000f, %esi
- xorl %esi, %edi
- xorl %esi, %edx
-
- roll $14, %edi
- movl %edi, %esi
- xorl %edx, %edi
- andl $0x33333333, %edi
- xorl %edi, %esi
- xorl %edi, %edx
-
- roll $22, %edx
- movl %edx, %edi
- xorl %esi, %edx
- andl $0x03fc03fc, %edx
- xorl %edx, %edi
- xorl %edx, %esi
-
- roll $9, %edi
- movl %edi, %edx
- xorl %esi, %edi
- andl $0xaaaaaaaa, %edi
- xorl %edi, %edx
- xorl %edi, %esi
-
- rorl $3, %edx
- rorl $2, %esi
- movl %esi, 4(%ebx)
- movl 36(%esp), %eax
- movl %edx, (%ebx)
- movl 40(%esp), %edi
- movl 44(%esp), %esi
- movl $1, 8(%esp)
- movl %eax, 4(%esp)
- movl %ebx, (%esp)
- call DES_encrypt2
- movl $0, 8(%esp)
- movl %edi, 4(%esp)
- movl %ebx, (%esp)
- call DES_encrypt2
- movl $1, 8(%esp)
- movl %esi, 4(%esp)
- movl %ebx, (%esp)
- call DES_encrypt2
- addl $12, %esp
- movl (%ebx), %edi
- movl 4(%ebx), %esi
-
-
- roll $2, %esi
- roll $3, %edi
- movl %edi, %eax
- xorl %esi, %edi
- andl $0xaaaaaaaa, %edi
- xorl %edi, %eax
- xorl %edi, %esi
-
- roll $23, %eax
- movl %eax, %edi
- xorl %esi, %eax
- andl $0x03fc03fc, %eax
- xorl %eax, %edi
- xorl %eax, %esi
-
- roll $10, %edi
- movl %edi, %eax
- xorl %esi, %edi
- andl $0x33333333, %edi
- xorl %edi, %eax
- xorl %edi, %esi
-
- roll $18, %esi
- movl %esi, %edi
- xorl %eax, %esi
- andl $0xfff0000f, %esi
- xorl %esi, %edi
- xorl %esi, %eax
-
- roll $12, %edi
- movl %edi, %esi
- xorl %eax, %edi
- andl $0xf0f0f0f0, %edi
- xorl %edi, %esi
- xorl %edi, %eax
-
- rorl $4, %eax
- movl %eax, (%ebx)
- movl %esi, 4(%ebx)
- popl %edi
- popl %esi
- popl %ebp
- popl %ebx
- ret
-.L_DES_encrypt3_end:
- .size DES_encrypt3,.L_DES_encrypt3_end-DES_encrypt3
-.ident "desasm.pl"
-.text
- .align 16
-.globl DES_decrypt3
- .type DES_decrypt3,@function
-DES_decrypt3:
- pushl %ebx
- movl 8(%esp), %ebx
- pushl %ebp
- pushl %esi
- pushl %edi
-
-
- movl (%ebx), %edi
- movl 4(%ebx), %esi
- subl $12, %esp
-
-
- roll $4, %edi
- movl %edi, %edx
- xorl %esi, %edi
- andl $0xf0f0f0f0, %edi
- xorl %edi, %edx
- xorl %edi, %esi
-
- roll $20, %esi
- movl %esi, %edi
- xorl %edx, %esi
- andl $0xfff0000f, %esi
- xorl %esi, %edi
- xorl %esi, %edx
-
- roll $14, %edi
- movl %edi, %esi
- xorl %edx, %edi
- andl $0x33333333, %edi
- xorl %edi, %esi
- xorl %edi, %edx
-
- roll $22, %edx
- movl %edx, %edi
- xorl %esi, %edx
- andl $0x03fc03fc, %edx
- xorl %edx, %edi
- xorl %edx, %esi
-
- roll $9, %edi
- movl %edi, %edx
- xorl %esi, %edi
- andl $0xaaaaaaaa, %edi
- xorl %edi, %edx
- xorl %edi, %esi
-
- rorl $3, %edx
- rorl $2, %esi
- movl %esi, 4(%ebx)
- movl 36(%esp), %esi
- movl %edx, (%ebx)
- movl 40(%esp), %edi
- movl 44(%esp), %eax
- movl $0, 8(%esp)
- movl %eax, 4(%esp)
- movl %ebx, (%esp)
- call DES_encrypt2
- movl $1, 8(%esp)
- movl %edi, 4(%esp)
- movl %ebx, (%esp)
- call DES_encrypt2
- movl $0, 8(%esp)
- movl %esi, 4(%esp)
- movl %ebx, (%esp)
- call DES_encrypt2
- addl $12, %esp
- movl (%ebx), %edi
- movl 4(%ebx), %esi
-
-
- roll $2, %esi
- roll $3, %edi
- movl %edi, %eax
- xorl %esi, %edi
- andl $0xaaaaaaaa, %edi
- xorl %edi, %eax
- xorl %edi, %esi
-
- roll $23, %eax
- movl %eax, %edi
- xorl %esi, %eax
- andl $0x03fc03fc, %eax
- xorl %eax, %edi
- xorl %eax, %esi
-
- roll $10, %edi
- movl %edi, %eax
- xorl %esi, %edi
- andl $0x33333333, %edi
- xorl %edi, %eax
- xorl %edi, %esi
-
- roll $18, %esi
- movl %esi, %edi
- xorl %eax, %esi
- andl $0xfff0000f, %esi
- xorl %esi, %edi
- xorl %esi, %eax
-
- roll $12, %edi
- movl %edi, %esi
- xorl %eax, %edi
- andl $0xf0f0f0f0, %edi
- xorl %edi, %esi
- xorl %edi, %eax
-
- rorl $4, %eax
- movl %eax, (%ebx)
- movl %esi, 4(%ebx)
- popl %edi
- popl %esi
- popl %ebp
- popl %ebx
- ret
-.L_DES_decrypt3_end:
- .size DES_decrypt3,.L_DES_decrypt3_end-DES_decrypt3
-.ident "desasm.pl"
-.text
- .align 16
-.globl DES_ncbc_encrypt
- .type DES_ncbc_encrypt,@function
-DES_ncbc_encrypt:
-
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
- movl 28(%esp), %ebp
-
- movl 36(%esp), %ebx
- movl (%ebx), %esi
- movl 4(%ebx), %edi
- pushl %edi
- pushl %esi
- pushl %edi
- pushl %esi
- movl %esp, %ebx
- movl 36(%esp), %esi
- movl 40(%esp), %edi
-
- movl 56(%esp), %ecx
-
- pushl %ecx
-
- movl 52(%esp), %eax
- pushl %eax
- pushl %ebx
- cmpl $0, %ecx
- jz .L006decrypt
- andl $4294967288, %ebp
- movl 12(%esp), %eax
- movl 16(%esp), %ebx
- jz .L007encrypt_finish
-.L008encrypt_loop:
- movl (%esi), %ecx
- movl 4(%esi), %edx
- xorl %ecx, %eax
- xorl %edx, %ebx
- movl %eax, 12(%esp)
- movl %ebx, 16(%esp)
- call DES_encrypt1
- movl 12(%esp), %eax
- movl 16(%esp), %ebx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- addl $8, %esi
- addl $8, %edi
- subl $8, %ebp
- jnz .L008encrypt_loop
-.L007encrypt_finish:
- movl 56(%esp), %ebp
- andl $7, %ebp
- jz .L009finish
- call .L010PIC_point
-.L010PIC_point:
- popl %edx
- leal .L011cbc_enc_jmp_table-.L010PIC_point(%edx),%ecx
- movl (%ecx,%ebp,4), %ebp
- addl %edx, %ebp
- xorl %ecx, %ecx
- xorl %edx, %edx
- jmp *%ebp
-.L012ej7:
- movb 6(%esi), %dh
- sall $8, %edx
-.L013ej6:
- movb 5(%esi), %dh
-.L014ej5:
- movb 4(%esi), %dl
-.L015ej4:
- movl (%esi), %ecx
- jmp .L016ejend
-.L017ej3:
- movb 2(%esi), %ch
- sall $8, %ecx
-.L018ej2:
- movb 1(%esi), %ch
-.L019ej1:
- movb (%esi), %cl
-.L016ejend:
- xorl %ecx, %eax
- xorl %edx, %ebx
- movl %eax, 12(%esp)
- movl %ebx, 16(%esp)
- call DES_encrypt1
- movl 12(%esp), %eax
- movl 16(%esp), %ebx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- jmp .L009finish
-.align 16
-.L006decrypt:
- andl $4294967288, %ebp
- movl 20(%esp), %eax
- movl 24(%esp), %ebx
- jz .L020decrypt_finish
-.L021decrypt_loop:
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl %eax, 12(%esp)
- movl %ebx, 16(%esp)
- call DES_encrypt1
- movl 12(%esp), %eax
- movl 16(%esp), %ebx
- movl 20(%esp), %ecx
- movl 24(%esp), %edx
- xorl %eax, %ecx
- xorl %ebx, %edx
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl %ecx, (%edi)
- movl %edx, 4(%edi)
- movl %eax, 20(%esp)
- movl %ebx, 24(%esp)
- addl $8, %esi
- addl $8, %edi
- subl $8, %ebp
- jnz .L021decrypt_loop
-.L020decrypt_finish:
- movl 56(%esp), %ebp
- andl $7, %ebp
- jz .L009finish
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl %eax, 12(%esp)
- movl %ebx, 16(%esp)
- call DES_encrypt1
- movl 12(%esp), %eax
- movl 16(%esp), %ebx
- movl 20(%esp), %ecx
- movl 24(%esp), %edx
- xorl %eax, %ecx
- xorl %ebx, %edx
- movl (%esi), %eax
- movl 4(%esi), %ebx
-.L022dj7:
- rorl $16, %edx
- movb %dl, 6(%edi)
- shrl $16, %edx
-.L023dj6:
- movb %dh, 5(%edi)
-.L024dj5:
- movb %dl, 4(%edi)
-.L025dj4:
- movl %ecx, (%edi)
- jmp .L026djend
-.L027dj3:
- rorl $16, %ecx
- movb %cl, 2(%edi)
- sall $16, %ecx
-.L028dj2:
- movb %ch, 1(%esi)
-.L029dj1:
- movb %cl, (%esi)
-.L026djend:
- jmp .L009finish
-.align 16
-.L009finish:
- movl 64(%esp), %ecx
- addl $28, %esp
- movl %eax, (%ecx)
- movl %ebx, 4(%ecx)
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.align 16
-.L011cbc_enc_jmp_table:
- .long 0
- .long .L019ej1-.L010PIC_point
- .long .L018ej2-.L010PIC_point
- .long .L017ej3-.L010PIC_point
- .long .L015ej4-.L010PIC_point
- .long .L014ej5-.L010PIC_point
- .long .L013ej6-.L010PIC_point
- .long .L012ej7-.L010PIC_point
-.L_DES_ncbc_encrypt_end:
- .size DES_ncbc_encrypt,.L_DES_ncbc_encrypt_end-DES_ncbc_encrypt
-.ident "desasm.pl"
-.text
- .align 16
-.globl DES_ede3_cbc_encrypt
- .type DES_ede3_cbc_encrypt,@function
-DES_ede3_cbc_encrypt:
-
- pushl %ebp
- pushl %ebx
- pushl %esi
- pushl %edi
- movl 28(%esp), %ebp
-
- movl 44(%esp), %ebx
- movl (%ebx), %esi
- movl 4(%ebx), %edi
- pushl %edi
- pushl %esi
- pushl %edi
- pushl %esi
- movl %esp, %ebx
- movl 36(%esp), %esi
- movl 40(%esp), %edi
-
- movl 64(%esp), %ecx
-
- movl 56(%esp), %eax
- pushl %eax
-
- movl 56(%esp), %eax
- pushl %eax
-
- movl 56(%esp), %eax
- pushl %eax
- pushl %ebx
- cmpl $0, %ecx
- jz .L030decrypt
- andl $4294967288, %ebp
- movl 16(%esp), %eax
- movl 20(%esp), %ebx
- jz .L031encrypt_finish
-.L032encrypt_loop:
- movl (%esi), %ecx
- movl 4(%esi), %edx
- xorl %ecx, %eax
- xorl %edx, %ebx
- movl %eax, 16(%esp)
- movl %ebx, 20(%esp)
- call DES_encrypt3
- movl 16(%esp), %eax
- movl 20(%esp), %ebx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- addl $8, %esi
- addl $8, %edi
- subl $8, %ebp
- jnz .L032encrypt_loop
-.L031encrypt_finish:
- movl 60(%esp), %ebp
- andl $7, %ebp
- jz .L033finish
- call .L034PIC_point
-.L034PIC_point:
- popl %edx
- leal .L035cbc_enc_jmp_table-.L034PIC_point(%edx),%ecx
- movl (%ecx,%ebp,4), %ebp
- addl %edx, %ebp
- xorl %ecx, %ecx
- xorl %edx, %edx
- jmp *%ebp
-.L036ej7:
- movb 6(%esi), %dh
- sall $8, %edx
-.L037ej6:
- movb 5(%esi), %dh
-.L038ej5:
- movb 4(%esi), %dl
-.L039ej4:
- movl (%esi), %ecx
- jmp .L040ejend
-.L041ej3:
- movb 2(%esi), %ch
- sall $8, %ecx
-.L042ej2:
- movb 1(%esi), %ch
-.L043ej1:
- movb (%esi), %cl
-.L040ejend:
- xorl %ecx, %eax
- xorl %edx, %ebx
- movl %eax, 16(%esp)
- movl %ebx, 20(%esp)
- call DES_encrypt3
- movl 16(%esp), %eax
- movl 20(%esp), %ebx
- movl %eax, (%edi)
- movl %ebx, 4(%edi)
- jmp .L033finish
-.align 16
-.L030decrypt:
- andl $4294967288, %ebp
- movl 24(%esp), %eax
- movl 28(%esp), %ebx
- jz .L044decrypt_finish
-.L045decrypt_loop:
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl %eax, 16(%esp)
- movl %ebx, 20(%esp)
- call DES_decrypt3
- movl 16(%esp), %eax
- movl 20(%esp), %ebx
- movl 24(%esp), %ecx
- movl 28(%esp), %edx
- xorl %eax, %ecx
- xorl %ebx, %edx
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl %ecx, (%edi)
- movl %edx, 4(%edi)
- movl %eax, 24(%esp)
- movl %ebx, 28(%esp)
- addl $8, %esi
- addl $8, %edi
- subl $8, %ebp
- jnz .L045decrypt_loop
-.L044decrypt_finish:
- movl 60(%esp), %ebp
- andl $7, %ebp
- jz .L033finish
- movl (%esi), %eax
- movl 4(%esi), %ebx
- movl %eax, 16(%esp)
- movl %ebx, 20(%esp)
- call DES_decrypt3
- movl 16(%esp), %eax
- movl 20(%esp), %ebx
- movl 24(%esp), %ecx
- movl 28(%esp), %edx
- xorl %eax, %ecx
- xorl %ebx, %edx
- movl (%esi), %eax
- movl 4(%esi), %ebx
-.L046dj7:
- rorl $16, %edx
- movb %dl, 6(%edi)
- shrl $16, %edx
-.L047dj6:
- movb %dh, 5(%edi)
-.L048dj5:
- movb %dl, 4(%edi)
-.L049dj4:
- movl %ecx, (%edi)
- jmp .L050djend
-.L051dj3:
- rorl $16, %ecx
- movb %cl, 2(%edi)
- sall $16, %ecx
-.L052dj2:
- movb %ch, 1(%esi)
-.L053dj1:
- movb %cl, (%esi)
-.L050djend:
- jmp .L033finish
-.align 16
-.L033finish:
- movl 76(%esp), %ecx
- addl $32, %esp
- movl %eax, (%ecx)
- movl %ebx, 4(%ecx)
- popl %edi
- popl %esi
- popl %ebx
- popl %ebp
- ret
-.align 16
-.L035cbc_enc_jmp_table:
- .long 0
- .long .L043ej1-.L034PIC_point
- .long .L042ej2-.L034PIC_point
- .long .L041ej3-.L034PIC_point
- .long .L039ej4-.L034PIC_point
- .long .L038ej5-.L034PIC_point
- .long .L037ej6-.L034PIC_point
- .long .L036ej7-.L034PIC_point
-.L_DES_ede3_cbc_encrypt_end:
- .size DES_ede3_cbc_encrypt,.L_DES_ede3_cbc_encrypt_end-DES_ede3_cbc_encrypt
-.ident "desasm.pl"