# qhasm: int64 r11_caller # qhasm: int64 r12_caller # qhasm: int64 r13_caller # qhasm: int64 r14_caller # qhasm: int64 r15_caller # qhasm: int64 rbx_caller # qhasm: int64 rbp_caller # qhasm: caller r11_caller # qhasm: caller r12_caller # qhasm: caller r13_caller # qhasm: caller r14_caller # qhasm: caller r15_caller # qhasm: caller rbx_caller # qhasm: caller rbp_caller # qhasm: stack64 r11_stack # qhasm: stack64 r12_stack # qhasm: stack64 r13_stack # qhasm: stack64 r14_stack # qhasm: stack64 r15_stack # qhasm: stack64 rbx_stack # qhasm: stack64 rbp_stack # qhasm: int64 table # qhasm: int64 c # qhasm: int64 k # qhasm: int64 iv # qhasm: int64 x0 # qhasm: int64 x1 # qhasm: int64 x2 # qhasm: int64 x3 # qhasm: int64 e # qhasm: int64 in # qhasm: int64 out # qhasm: stack64 out_stack # qhasm: stack64 n0 # qhasm: stack64 n1 # qhasm: stack64 n2 # qhasm: stack64 n3 # qhasm: stack64 r0 # qhasm: stack64 r1 # qhasm: stack64 r2 # qhasm: stack64 r3 # qhasm: stack64 r4 # qhasm: stack64 r5 # qhasm: stack64 r6 # qhasm: stack64 r7 # qhasm: stack64 r8 # qhasm: stack64 r9 # qhasm: stack64 r10 # qhasm: stack64 r11 # qhasm: stack64 r12 # qhasm: stack64 r13 # qhasm: stack64 r14 # qhasm: stack64 r15 # qhasm: stack64 r16 # qhasm: stack64 r17 # qhasm: stack64 r18 # qhasm: stack64 r19 # qhasm: stack64 r20 # qhasm: stack64 r21 # qhasm: stack64 r22 # qhasm: stack64 r23 # qhasm: stack64 r24 # qhasm: stack64 r25 # qhasm: stack64 r26 # qhasm: stack64 r27 # qhasm: stack64 r28 # qhasm: stack64 r29 # qhasm: stack64 r30 # qhasm: stack64 r31 # qhasm: stack64 r32 # qhasm: stack64 r33 # qhasm: stack64 r34 # qhasm: stack64 r35 # qhasm: stack64 r36 # qhasm: stack64 r37 # qhasm: stack64 r38 # qhasm: stack64 r39 # qhasm: stack64 r40 # qhasm: stack64 r41 # qhasm: stack64 r42 # qhasm: stack64 r43 # qhasm: int64 y0 # qhasm: int64 y1 # qhasm: int64 y2 # qhasm: int64 y3 # qhasm: int64 z0 # qhasm: int64 z1 # qhasm: int64 z2 # qhasm: int64 z3 # qhasm: int64 p00 # qhasm: int64 p01 # qhasm: int64 p02 # qhasm: int64 p03 # qhasm: int64 p10 # qhasm: int64 p11 # qhasm: int64 p12 # qhasm: int64 p13 # qhasm: int64 p20 # qhasm: int64 p21 # qhasm: int64 p22 # qhasm: int64 p23 # qhasm: int64 p30 # qhasm: int64 p31 # qhasm: int64 p32 # qhasm: int64 p33 # qhasm: int64 q00 # qhasm: int64 q01 # qhasm: int64 q02 # qhasm: int64 q03 # qhasm: int64 q10 # qhasm: int64 q11 # qhasm: int64 q12 # qhasm: int64 q13 # qhasm: int64 q20 # qhasm: int64 q21 # qhasm: int64 q22 # qhasm: int64 q23 # qhasm: int64 q30 # qhasm: int64 q31 # qhasm: int64 q32 # qhasm: int64 q33 # qhasm: enter aes128_amd64_2_block stackaligned4096 aes128_amd64_2_constants .text .p2align 5 .globl _aes128_amd64_2_block .globl aes128_amd64_2_block _aes128_amd64_2_block: aes128_amd64_2_block: mov %rsp,%r11 sub $aes128_amd64_2_constants,%r11 and $4095,%r11 add $416,%r11 sub %r11,%rsp # qhasm: input out # qhasm: input in # qhasm: input c # qhasm: out_stack = out # asm 1: movq out_stack=stack64#1 # asm 2: movq out_stack=0(%rsp) movq %rdi,0(%rsp) # qhasm: r11_stack = r11_caller # asm 1: movq r11_stack=stack64#2 # asm 2: movq r11_stack=8(%rsp) movq %r11,8(%rsp) # qhasm: r12_stack = r12_caller # asm 1: movq r12_stack=stack64#3 # asm 2: movq r12_stack=16(%rsp) movq %r12,16(%rsp) # qhasm: r13_stack = r13_caller # asm 1: movq r13_stack=stack64#4 # asm 2: movq r13_stack=24(%rsp) movq %r13,24(%rsp) # qhasm: r14_stack = r14_caller # asm 1: movq r14_stack=stack64#5 # asm 2: movq r14_stack=32(%rsp) movq %r14,32(%rsp) # qhasm: r15_stack = r15_caller # asm 1: movq r15_stack=stack64#6 # asm 2: movq r15_stack=40(%rsp) movq %r15,40(%rsp) # qhasm: rbx_stack = rbx_caller # asm 1: movq rbx_stack=stack64#7 # asm 2: movq rbx_stack=48(%rsp) movq %rbx,48(%rsp) # qhasm: rbp_stack = rbp_caller # asm 1: movq rbp_stack=stack64#8 # asm 2: movq rbp_stack=56(%rsp) movq %rbp,56(%rsp) # qhasm: x0 = *(uint32 *) (c + 0) # asm 1: movl 0(x0=int64#1d # asm 2: movl 0(x0=%edi movl 0(%rdx),%edi # qhasm: x1 = *(uint32 *) (c + 4) # asm 1: movl 4(x1=int64#4d # asm 2: movl 4(x1=%ecx movl 4(%rdx),%ecx # qhasm: x2 = *(uint32 *) (c + 8) # asm 1: movl 8(x2=int64#5d # asm 2: movl 8(x2=%r8d movl 8(%rdx),%r8d # qhasm: x3 = *(uint32 *) (c + 12) # asm 1: movl 12(x3=int64#6d # asm 2: movl 12(x3=%r9d movl 12(%rdx),%r9d # qhasm: r0 = x0 # asm 1: movq r0=stack64#9 # asm 2: movq r0=64(%rsp) movq %rdi,64(%rsp) # qhasm: r1 = x1 # asm 1: movq r1=stack64#10 # asm 2: movq r1=72(%rsp) movq %rcx,72(%rsp) # qhasm: r2 = x2 # asm 1: movq r2=stack64#11 # asm 2: movq r2=80(%rsp) movq %r8,80(%rsp) # qhasm: r3 = x3 # asm 1: movq r3=stack64#12 # asm 2: movq r3=88(%rsp) movq %r9,88(%rsp) # qhasm: x0 = *(uint32 *) (c + 16) # asm 1: movl 16(x0=int64#1d # asm 2: movl 16(x0=%edi movl 16(%rdx),%edi # qhasm: x1 ^= x0 # asm 1: xor r4=stack64#13 # asm 2: movq r4=96(%rsp) movq %rdi,96(%rsp) # qhasm: r5 = x1 # asm 1: movq r5=stack64#14 # asm 2: movq r5=104(%rsp) movq %rcx,104(%rsp) # qhasm: r6 = x2 # asm 1: movq r6=stack64#15 # asm 2: movq r6=112(%rsp) movq %r8,112(%rsp) # qhasm: r7 = x3 # asm 1: movq r7=stack64#16 # asm 2: movq r7=120(%rsp) movq %r9,120(%rsp) # qhasm: x0 = *(uint32 *) (c + 20) # asm 1: movl 20(x0=int64#1d # asm 2: movl 20(x0=%edi movl 20(%rdx),%edi # qhasm: x1 ^= x0 # asm 1: xor r8=stack64#17 # asm 2: movq r8=128(%rsp) movq %rdi,128(%rsp) # qhasm: r9 = x1 # asm 1: movq r9=stack64#18 # asm 2: movq r9=136(%rsp) movq %rcx,136(%rsp) # qhasm: r10 = x2 # asm 1: movq r10=stack64#19 # asm 2: movq r10=144(%rsp) movq %r8,144(%rsp) # qhasm: r11 = x3 # asm 1: movq r11=stack64#20 # asm 2: movq r11=152(%rsp) movq %r9,152(%rsp) # qhasm: x0 = *(uint32 *) (c + 24) # asm 1: movl 24(x0=int64#1d # asm 2: movl 24(x0=%edi movl 24(%rdx),%edi # qhasm: x1 ^= x0 # asm 1: xor r12=stack64#21 # asm 2: movq r12=160(%rsp) movq %rdi,160(%rsp) # qhasm: r13 = x1 # asm 1: movq r13=stack64#22 # asm 2: movq r13=168(%rsp) movq %rcx,168(%rsp) # qhasm: r14 = x2 # asm 1: movq r14=stack64#23 # asm 2: movq r14=176(%rsp) movq %r8,176(%rsp) # qhasm: r15 = x3 # asm 1: movq r15=stack64#24 # asm 2: movq r15=184(%rsp) movq %r9,184(%rsp) # qhasm: x0 = *(uint32 *) (c + 28) # asm 1: movl 28(x0=int64#1d # asm 2: movl 28(x0=%edi movl 28(%rdx),%edi # qhasm: x1 ^= x0 # asm 1: xor r16=stack64#25 # asm 2: movq r16=192(%rsp) movq %rdi,192(%rsp) # qhasm: r17 = x1 # asm 1: movq r17=stack64#26 # asm 2: movq r17=200(%rsp) movq %rcx,200(%rsp) # qhasm: r18 = x2 # asm 1: movq r18=stack64#27 # asm 2: movq r18=208(%rsp) movq %r8,208(%rsp) # qhasm: r19 = x3 # asm 1: movq r19=stack64#28 # asm 2: movq r19=216(%rsp) movq %r9,216(%rsp) # qhasm: x0 = *(uint32 *) (c + 32) # asm 1: movl 32(x0=int64#1d # asm 2: movl 32(x0=%edi movl 32(%rdx),%edi # qhasm: x1 ^= x0 # asm 1: xor r20=stack64#29 # asm 2: movq r20=224(%rsp) movq %rdi,224(%rsp) # qhasm: r21 = x1 # asm 1: movq r21=stack64#30 # asm 2: movq r21=232(%rsp) movq %rcx,232(%rsp) # qhasm: r22 = x2 # asm 1: movq r22=stack64#31 # asm 2: movq r22=240(%rsp) movq %r8,240(%rsp) # qhasm: r23 = x3 # asm 1: movq r23=stack64#32 # asm 2: movq r23=248(%rsp) movq %r9,248(%rsp) # qhasm: x0 = *(uint32 *) (c + 36) # asm 1: movl 36(x0=int64#1d # asm 2: movl 36(x0=%edi movl 36(%rdx),%edi # qhasm: x1 ^= x0 # asm 1: xor r24=stack64#33 # asm 2: movq r24=256(%rsp) movq %rdi,256(%rsp) # qhasm: r25 = x1 # asm 1: movq r25=stack64#34 # asm 2: movq r25=264(%rsp) movq %rcx,264(%rsp) # qhasm: r26 = x2 # asm 1: movq r26=stack64#35 # asm 2: movq r26=272(%rsp) movq %r8,272(%rsp) # qhasm: r27 = x3 # asm 1: movq r27=stack64#36 # asm 2: movq r27=280(%rsp) movq %r9,280(%rsp) # qhasm: x0 = *(uint32 *) (c + 40) # asm 1: movl 40(x0=int64#1d # asm 2: movl 40(x0=%edi movl 40(%rdx),%edi # qhasm: x1 ^= x0 # asm 1: xor r28=stack64#37 # asm 2: movq r28=288(%rsp) movq %rdi,288(%rsp) # qhasm: r29 = x1 # asm 1: movq r29=stack64#38 # asm 2: movq r29=296(%rsp) movq %rcx,296(%rsp) # qhasm: r30 = x2 # asm 1: movq r30=stack64#39 # asm 2: movq r30=304(%rsp) movq %r8,304(%rsp) # qhasm: r31 = x3 # asm 1: movq r31=stack64#40 # asm 2: movq r31=312(%rsp) movq %r9,312(%rsp) # qhasm: x0 = *(uint32 *) (c + 44) # asm 1: movl 44(x0=int64#1d # asm 2: movl 44(x0=%edi movl 44(%rdx),%edi # qhasm: x1 ^= x0 # asm 1: xor r32=stack64#41 # asm 2: movq r32=320(%rsp) movq %rdi,320(%rsp) # qhasm: r33 = x1 # asm 1: movq r33=stack64#42 # asm 2: movq r33=328(%rsp) movq %rcx,328(%rsp) # qhasm: r34 = x2 # asm 1: movq r34=stack64#43 # asm 2: movq r34=336(%rsp) movq %r8,336(%rsp) # qhasm: r35 = x3 # asm 1: movq r35=stack64#44 # asm 2: movq r35=344(%rsp) movq %r9,344(%rsp) # qhasm: x0 = *(uint32 *) (c + 48) # asm 1: movl 48(x0=int64#1d # asm 2: movl 48(x0=%edi movl 48(%rdx),%edi # qhasm: x1 ^= x0 # asm 1: xor r36=stack64#45 # asm 2: movq r36=352(%rsp) movq %rdi,352(%rsp) # qhasm: r37 = x1 # asm 1: movq r37=stack64#46 # asm 2: movq r37=360(%rsp) movq %rcx,360(%rsp) # qhasm: r38 = x2 # asm 1: movq r38=stack64#47 # asm 2: movq r38=368(%rsp) movq %r8,368(%rsp) # qhasm: r39 = x3 # asm 1: movq r39=stack64#48 # asm 2: movq r39=376(%rsp) movq %r9,376(%rsp) # qhasm: x0 = *(uint32 *) (c + 52) # asm 1: movl 52(x0=int64#1d # asm 2: movl 52(x0=%edi movl 52(%rdx),%edi # qhasm: x1 ^= x0 # asm 1: xor r40=stack64#49 # asm 2: movq r40=384(%rsp) movq %rdi,384(%rsp) # qhasm: r41 = x1 # asm 1: movq r41=stack64#50 # asm 2: movq r41=392(%rsp) movq %rcx,392(%rsp) # qhasm: r42 = x2 # asm 1: movq r42=stack64#51 # asm 2: movq r42=400(%rsp) movq %r8,400(%rsp) # qhasm: r43 = x3 # asm 1: movq r43=stack64#52 # asm 2: movq r43=408(%rsp) movq %r9,408(%rsp) # qhasm: y0 = *(uint32 *) (in + 0) # asm 1: movl 0(y0=int64#3d # asm 2: movl 0(y0=%edx movl 0(%rsi),%edx # qhasm: y1 = *(uint32 *) (in + 4) # asm 1: movl 4(y1=int64#4d # asm 2: movl 4(y1=%ecx movl 4(%rsi),%ecx # qhasm: y2 = *(uint32 *) (in + 8) # asm 1: movl 8(y2=int64#7d # asm 2: movl 8(y2=%eax movl 8(%rsi),%eax # qhasm: y3 = *(uint32 *) (in + 12) # asm 1: movl 12(y3=int64#14d # asm 2: movl 12(y3=%ebx movl 12(%rsi),%ebx # qhasm: assign 3 to y0 # qhasm: assign 4 to y1 # qhasm: assign 7 to y2 # qhasm: assign 14 to y3 # qhasm: table = &aes128_amd64_2_tablex # asm 1: lea aes128_amd64_2_tablex(%rip),>table=int64#5 # asm 2: lea aes128_amd64_2_tablex(%rip),>table=%r8 lea aes128_amd64_2_tablex(%rip),%r8 # qhasm: y0 ^= r0 # asm 1: xorq p00=int64#2d # asm 2: movzbl p00=%esi movzbl %dl,%esi # qhasm: p01 = (y0 >> 8) & 255 # asm 1: movzbl p01=int64#1d # asm 2: movzbl p01=%edi movzbl %dh,%edi # qhasm: (uint32) y0 >>= 16 # asm 1: shr $16,p02=int64#6d # asm 2: movzbl p02=%r9d movzbl %dl,%r9d # qhasm: p03 = (y0 >> 8) & 255 # asm 1: movzbl p03=int64#15d # asm 2: movzbl p03=%ebp movzbl %dh,%ebp # qhasm: z0 = *(uint32 *) (table + 3 + p00 * 8) # asm 1: movl 3(z0=int64#3d # asm 2: movl 3(z0=%edx movl 3(%r8,%rsi,8),%edx # qhasm: assign 3 to z0 # qhasm: p10 = y1 & 255 # asm 1: movzbl p10=int64#8d # asm 2: movzbl p10=%r10d movzbl %cl,%r10d # qhasm: p11 = (y1 >> 8) & 255 # asm 1: movzbl p11=int64#2d # asm 2: movzbl p11=%esi movzbl %ch,%esi # qhasm: (uint32) z0 ^= *(uint32 *) (table + 2 + p11 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,p12=int64#9d # asm 2: movzbl p12=%r11d movzbl %cl,%r11d # qhasm: p13 = (y1 >> 8) & 255 # asm 1: movzbl p13=int64#2d # asm 2: movzbl p13=%esi movzbl %ch,%esi # qhasm: z1 = *(uint32 *) (table + 4 + p03 * 8) # asm 1: movl 4(z1=int64#4d # asm 2: movl 4(z1=%ecx movl 4(%r8,%rbp,8),%ecx # qhasm: (uint32) z1 ^= *(uint32 *) (table + 3 + p10 * 8) # asm 1: xorl 3(p20=int64#8d # asm 2: movzbl p20=%r10d movzbl %al,%r10d # qhasm: p21 = (y2 >> 8) & 255 # asm 1: movzbl p21=int64#15d # asm 2: movzbl p21=%ebp movzbl %ah,%ebp # qhasm: (uint32) z1 ^= *(uint32 *) (table + 2 + p21 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,p22=int64#10d # asm 2: movzbl p22=%r12d movzbl %al,%r12d # qhasm: (uint32) z0 ^= *(uint32 *) (table + 1 + p22 * 8) # asm 1: xorl 1(> 8) & 255 # asm 1: movzbl p23=int64#15d # asm 2: movzbl p23=%ebp movzbl %ah,%ebp # qhasm: z2 = *(uint32 *) (table + 1 + p02 * 8) # asm 1: movl 1(z2=int64#7d # asm 2: movl 1(z2=%eax movl 1(%r8,%r9,8),%eax # qhasm: (uint32) z2 ^= *(uint32 *) (table + 4 + p13 * 8) # asm 1: xorl 4(p30=int64#6d # asm 2: movzbl p30=%r9d movzbl %bl,%r9d # qhasm: p31 = (y3 >> 8) & 255 # asm 1: movzbl p31=int64#2d # asm 2: movzbl p31=%esi movzbl %bh,%esi # qhasm: (uint32) z2 ^= *(uint32 *) (table + 2 + p31 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,p32=int64#2d # asm 2: movzbl p32=%esi movzbl %bl,%esi # qhasm: (uint32) z1 ^= *(uint32 *) (table + 1 + p32 * 8) # asm 1: xorl 1(> 8) & 255 # asm 1: movzbl p33=int64#14d # asm 2: movzbl p33=%ebx movzbl %bh,%ebx # qhasm: (uint32) z0 ^= *(uint32 *) (table + 4 + p33 * 8) # asm 1: xorl 4(z3=int64#14d # asm 2: movl 2(z3=%ebx movl 2(%r8,%rdi,8),%ebx # qhasm: z2 ^= r6 # asm 1: xorq q00=int64#2d # asm 2: movzbl q00=%esi movzbl %dl,%esi # qhasm: q03 = (z0 >> 8) & 255 # asm 1: movzbl q03=int64#1d # asm 2: movzbl q03=%edi movzbl %dh,%edi # qhasm: (uint32) z0 >>= 16 # asm 1: shr $16,q02=int64#6d # asm 2: movzbl q02=%r9d movzbl %dl,%r9d # qhasm: q01 = (z0 >> 8) & 255 # asm 1: movzbl q01=int64#15d # asm 2: movzbl q01=%ebp movzbl %dh,%ebp # qhasm: y0 = *(uint32 *) (table + 3 + q00 * 8) # asm 1: movl 3(y0=int64#3d # asm 2: movl 3(y0=%edx movl 3(%r8,%rsi,8),%edx # qhasm: assign 3 to y0 # qhasm: q10 = z1 & 255 # asm 1: movzbl q10=int64#8d # asm 2: movzbl q10=%r10d movzbl %cl,%r10d # qhasm: q11 = (z1 >> 8) & 255 # asm 1: movzbl q11=int64#2d # asm 2: movzbl q11=%esi movzbl %ch,%esi # qhasm: (uint32) y0 ^= *(uint32 *) (table + 2 + q11 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,q12=int64#9d # asm 2: movzbl q12=%r11d movzbl %cl,%r11d # qhasm: q13 = (z1 >> 8) & 255 # asm 1: movzbl q13=int64#2d # asm 2: movzbl q13=%esi movzbl %ch,%esi # qhasm: y1 = *(uint32 *) (table + 4 + q01 * 8) # asm 1: movl 4(y1=int64#4d # asm 2: movl 4(y1=%ecx movl 4(%r8,%rbp,8),%ecx # qhasm: (uint32) y1 ^= *(uint32 *) (table + 3 + q10 * 8) # asm 1: xorl 3(q20=int64#8d # asm 2: movzbl q20=%r10d movzbl %al,%r10d # qhasm: q21 = (z2 >> 8) & 255 # asm 1: movzbl q21=int64#15d # asm 2: movzbl q21=%ebp movzbl %ah,%ebp # qhasm: (uint32) y1 ^= *(uint32 *) (table + 2 + q21 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,q22=int64#10d # asm 2: movzbl q22=%r12d movzbl %al,%r12d # qhasm: (uint32) y0 ^= *(uint32 *) (table + 1 + q22 * 8) # asm 1: xorl 1(> 8) & 255 # asm 1: movzbl q23=int64#15d # asm 2: movzbl q23=%ebp movzbl %ah,%ebp # qhasm: y2 = *(uint32 *) (table + 1 + q02 * 8) # asm 1: movl 1(y2=int64#7d # asm 2: movl 1(y2=%eax movl 1(%r8,%r9,8),%eax # qhasm: (uint32) y2 ^= *(uint32 *) (table + 4 + q13 * 8) # asm 1: xorl 4(q30=int64#6d # asm 2: movzbl q30=%r9d movzbl %bl,%r9d # qhasm: q31 = (z3 >> 8) & 255 # asm 1: movzbl q31=int64#2d # asm 2: movzbl q31=%esi movzbl %bh,%esi # qhasm: (uint32) y2 ^= *(uint32 *) (table + 2 + q31 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,q32=int64#2d # asm 2: movzbl q32=%esi movzbl %bl,%esi # qhasm: (uint32) y1 ^= *(uint32 *) (table + 1 + q32 * 8) # asm 1: xorl 1(> 8) & 255 # asm 1: movzbl q33=int64#14d # asm 2: movzbl q33=%ebx movzbl %bh,%ebx # qhasm: (uint32) y0 ^= *(uint32 *) (table + 4 + q33 * 8) # asm 1: xorl 4(y3=int64#14d # asm 2: movl 2(y3=%ebx movl 2(%r8,%rdi,8),%ebx # qhasm: y2 ^= r10 # asm 1: xorq p00=int64#2d # asm 2: movzbl p00=%esi movzbl %dl,%esi # qhasm: p01 = (y0 >> 8) & 255 # asm 1: movzbl p01=int64#1d # asm 2: movzbl p01=%edi movzbl %dh,%edi # qhasm: (uint32) y0 >>= 16 # asm 1: shr $16,p02=int64#6d # asm 2: movzbl p02=%r9d movzbl %dl,%r9d # qhasm: p03 = (y0 >> 8) & 255 # asm 1: movzbl p03=int64#15d # asm 2: movzbl p03=%ebp movzbl %dh,%ebp # qhasm: z0 = *(uint32 *) (table + 3 + p00 * 8) # asm 1: movl 3(z0=int64#3d # asm 2: movl 3(z0=%edx movl 3(%r8,%rsi,8),%edx # qhasm: assign 3 to z0 # qhasm: p10 = y1 & 255 # asm 1: movzbl p10=int64#8d # asm 2: movzbl p10=%r10d movzbl %cl,%r10d # qhasm: p11 = (y1 >> 8) & 255 # asm 1: movzbl p11=int64#2d # asm 2: movzbl p11=%esi movzbl %ch,%esi # qhasm: (uint32) z0 ^= *(uint32 *) (table + 2 + p11 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,p12=int64#9d # asm 2: movzbl p12=%r11d movzbl %cl,%r11d # qhasm: p13 = (y1 >> 8) & 255 # asm 1: movzbl p13=int64#2d # asm 2: movzbl p13=%esi movzbl %ch,%esi # qhasm: z1 = *(uint32 *) (table + 4 + p03 * 8) # asm 1: movl 4(z1=int64#4d # asm 2: movl 4(z1=%ecx movl 4(%r8,%rbp,8),%ecx # qhasm: (uint32) z1 ^= *(uint32 *) (table + 3 + p10 * 8) # asm 1: xorl 3(p20=int64#8d # asm 2: movzbl p20=%r10d movzbl %al,%r10d # qhasm: p21 = (y2 >> 8) & 255 # asm 1: movzbl p21=int64#15d # asm 2: movzbl p21=%ebp movzbl %ah,%ebp # qhasm: (uint32) z1 ^= *(uint32 *) (table + 2 + p21 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,p22=int64#10d # asm 2: movzbl p22=%r12d movzbl %al,%r12d # qhasm: (uint32) z0 ^= *(uint32 *) (table + 1 + p22 * 8) # asm 1: xorl 1(> 8) & 255 # asm 1: movzbl p23=int64#15d # asm 2: movzbl p23=%ebp movzbl %ah,%ebp # qhasm: z2 = *(uint32 *) (table + 1 + p02 * 8) # asm 1: movl 1(z2=int64#7d # asm 2: movl 1(z2=%eax movl 1(%r8,%r9,8),%eax # qhasm: (uint32) z2 ^= *(uint32 *) (table + 4 + p13 * 8) # asm 1: xorl 4(p30=int64#6d # asm 2: movzbl p30=%r9d movzbl %bl,%r9d # qhasm: p31 = (y3 >> 8) & 255 # asm 1: movzbl p31=int64#2d # asm 2: movzbl p31=%esi movzbl %bh,%esi # qhasm: (uint32) z2 ^= *(uint32 *) (table + 2 + p31 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,p32=int64#2d # asm 2: movzbl p32=%esi movzbl %bl,%esi # qhasm: (uint32) z1 ^= *(uint32 *) (table + 1 + p32 * 8) # asm 1: xorl 1(> 8) & 255 # asm 1: movzbl p33=int64#14d # asm 2: movzbl p33=%ebx movzbl %bh,%ebx # qhasm: (uint32) z0 ^= *(uint32 *) (table + 4 + p33 * 8) # asm 1: xorl 4(z3=int64#14d # asm 2: movl 2(z3=%ebx movl 2(%r8,%rdi,8),%ebx # qhasm: (uint32) z3 ^= *(uint32 *) (table + 1 + p12 * 8) # asm 1: xorl 1(q00=int64#2d # asm 2: movzbl q00=%esi movzbl %dl,%esi # qhasm: q03 = (z0 >> 8) & 255 # asm 1: movzbl q03=int64#1d # asm 2: movzbl q03=%edi movzbl %dh,%edi # qhasm: (uint32) z0 >>= 16 # asm 1: shr $16,q02=int64#6d # asm 2: movzbl q02=%r9d movzbl %dl,%r9d # qhasm: q01 = (z0 >> 8) & 255 # asm 1: movzbl q01=int64#15d # asm 2: movzbl q01=%ebp movzbl %dh,%ebp # qhasm: y0 = *(uint32 *) (table + 3 + q00 * 8) # asm 1: movl 3(y0=int64#3d # asm 2: movl 3(y0=%edx movl 3(%r8,%rsi,8),%edx # qhasm: assign 3 to y0 # qhasm: q10 = z1 & 255 # asm 1: movzbl q10=int64#8d # asm 2: movzbl q10=%r10d movzbl %cl,%r10d # qhasm: q11 = (z1 >> 8) & 255 # asm 1: movzbl q11=int64#2d # asm 2: movzbl q11=%esi movzbl %ch,%esi # qhasm: (uint32) y0 ^= *(uint32 *) (table + 2 + q11 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,q12=int64#9d # asm 2: movzbl q12=%r11d movzbl %cl,%r11d # qhasm: q13 = (z1 >> 8) & 255 # asm 1: movzbl q13=int64#2d # asm 2: movzbl q13=%esi movzbl %ch,%esi # qhasm: y1 = *(uint32 *) (table + 4 + q01 * 8) # asm 1: movl 4(y1=int64#4d # asm 2: movl 4(y1=%ecx movl 4(%r8,%rbp,8),%ecx # qhasm: (uint32) y1 ^= *(uint32 *) (table + 3 + q10 * 8) # asm 1: xorl 3(q20=int64#8d # asm 2: movzbl q20=%r10d movzbl %al,%r10d # qhasm: q21 = (z2 >> 8) & 255 # asm 1: movzbl q21=int64#15d # asm 2: movzbl q21=%ebp movzbl %ah,%ebp # qhasm: (uint32) y1 ^= *(uint32 *) (table + 2 + q21 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,q22=int64#10d # asm 2: movzbl q22=%r12d movzbl %al,%r12d # qhasm: (uint32) y0 ^= *(uint32 *) (table + 1 + q22 * 8) # asm 1: xorl 1(> 8) & 255 # asm 1: movzbl q23=int64#15d # asm 2: movzbl q23=%ebp movzbl %ah,%ebp # qhasm: y2 = *(uint32 *) (table + 1 + q02 * 8) # asm 1: movl 1(y2=int64#7d # asm 2: movl 1(y2=%eax movl 1(%r8,%r9,8),%eax # qhasm: (uint32) y2 ^= *(uint32 *) (table + 4 + q13 * 8) # asm 1: xorl 4(q30=int64#6d # asm 2: movzbl q30=%r9d movzbl %bl,%r9d # qhasm: q31 = (z3 >> 8) & 255 # asm 1: movzbl q31=int64#2d # asm 2: movzbl q31=%esi movzbl %bh,%esi # qhasm: (uint32) y2 ^= *(uint32 *) (table + 2 + q31 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,q32=int64#2d # asm 2: movzbl q32=%esi movzbl %bl,%esi # qhasm: (uint32) y1 ^= *(uint32 *) (table + 1 + q32 * 8) # asm 1: xorl 1(> 8) & 255 # asm 1: movzbl q33=int64#14d # asm 2: movzbl q33=%ebx movzbl %bh,%ebx # qhasm: (uint32) y0 ^= *(uint32 *) (table + 4 + q33 * 8) # asm 1: xorl 4(y3=int64#14d # asm 2: movl 2(y3=%ebx movl 2(%r8,%rdi,8),%ebx # qhasm: (uint32) y3 ^= *(uint32 *) (table + 1 + q12 * 8) # asm 1: xorl 1(p00=int64#2d # asm 2: movzbl p00=%esi movzbl %dl,%esi # qhasm: p01 = (y0 >> 8) & 255 # asm 1: movzbl p01=int64#1d # asm 2: movzbl p01=%edi movzbl %dh,%edi # qhasm: (uint32) y0 >>= 16 # asm 1: shr $16,p02=int64#6d # asm 2: movzbl p02=%r9d movzbl %dl,%r9d # qhasm: p03 = (y0 >> 8) & 255 # asm 1: movzbl p03=int64#15d # asm 2: movzbl p03=%ebp movzbl %dh,%ebp # qhasm: z0 = *(uint32 *) (table + 3 + p00 * 8) # asm 1: movl 3(z0=int64#3d # asm 2: movl 3(z0=%edx movl 3(%r8,%rsi,8),%edx # qhasm: assign 3 to z0 # qhasm: p10 = y1 & 255 # asm 1: movzbl p10=int64#8d # asm 2: movzbl p10=%r10d movzbl %cl,%r10d # qhasm: p11 = (y1 >> 8) & 255 # asm 1: movzbl p11=int64#2d # asm 2: movzbl p11=%esi movzbl %ch,%esi # qhasm: (uint32) z0 ^= *(uint32 *) (table + 2 + p11 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,p12=int64#9d # asm 2: movzbl p12=%r11d movzbl %cl,%r11d # qhasm: p13 = (y1 >> 8) & 255 # asm 1: movzbl p13=int64#2d # asm 2: movzbl p13=%esi movzbl %ch,%esi # qhasm: z1 = *(uint32 *) (table + 4 + p03 * 8) # asm 1: movl 4(z1=int64#4d # asm 2: movl 4(z1=%ecx movl 4(%r8,%rbp,8),%ecx # qhasm: (uint32) z1 ^= *(uint32 *) (table + 3 + p10 * 8) # asm 1: xorl 3(p20=int64#8d # asm 2: movzbl p20=%r10d movzbl %al,%r10d # qhasm: p21 = (y2 >> 8) & 255 # asm 1: movzbl p21=int64#15d # asm 2: movzbl p21=%ebp movzbl %ah,%ebp # qhasm: (uint32) z1 ^= *(uint32 *) (table + 2 + p21 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,p22=int64#10d # asm 2: movzbl p22=%r12d movzbl %al,%r12d # qhasm: (uint32) z0 ^= *(uint32 *) (table + 1 + p22 * 8) # asm 1: xorl 1(> 8) & 255 # asm 1: movzbl p23=int64#15d # asm 2: movzbl p23=%ebp movzbl %ah,%ebp # qhasm: z2 = *(uint32 *) (table + 1 + p02 * 8) # asm 1: movl 1(z2=int64#7d # asm 2: movl 1(z2=%eax movl 1(%r8,%r9,8),%eax # qhasm: (uint32) z2 ^= *(uint32 *) (table + 4 + p13 * 8) # asm 1: xorl 4(p30=int64#6d # asm 2: movzbl p30=%r9d movzbl %bl,%r9d # qhasm: p31 = (y3 >> 8) & 255 # asm 1: movzbl p31=int64#2d # asm 2: movzbl p31=%esi movzbl %bh,%esi # qhasm: (uint32) z2 ^= *(uint32 *) (table + 2 + p31 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,p32=int64#2d # asm 2: movzbl p32=%esi movzbl %bl,%esi # qhasm: (uint32) z1 ^= *(uint32 *) (table + 1 + p32 * 8) # asm 1: xorl 1(> 8) & 255 # asm 1: movzbl p33=int64#14d # asm 2: movzbl p33=%ebx movzbl %bh,%ebx # qhasm: (uint32) z0 ^= *(uint32 *) (table + 4 + p33 * 8) # asm 1: xorl 4(z3=int64#14d # asm 2: movl 2(z3=%ebx movl 2(%r8,%rdi,8),%ebx # qhasm: (uint32) z3 ^= *(uint32 *) (table + 1 + p12 * 8) # asm 1: xorl 1(q00=int64#2d # asm 2: movzbl q00=%esi movzbl %dl,%esi # qhasm: q03 = (z0 >> 8) & 255 # asm 1: movzbl q03=int64#1d # asm 2: movzbl q03=%edi movzbl %dh,%edi # qhasm: (uint32) z0 >>= 16 # asm 1: shr $16,q02=int64#6d # asm 2: movzbl q02=%r9d movzbl %dl,%r9d # qhasm: q01 = (z0 >> 8) & 255 # asm 1: movzbl q01=int64#15d # asm 2: movzbl q01=%ebp movzbl %dh,%ebp # qhasm: y0 = *(uint32 *) (table + 3 + q00 * 8) # asm 1: movl 3(y0=int64#3d # asm 2: movl 3(y0=%edx movl 3(%r8,%rsi,8),%edx # qhasm: assign 3 to y0 # qhasm: q10 = z1 & 255 # asm 1: movzbl q10=int64#8d # asm 2: movzbl q10=%r10d movzbl %cl,%r10d # qhasm: q11 = (z1 >> 8) & 255 # asm 1: movzbl q11=int64#2d # asm 2: movzbl q11=%esi movzbl %ch,%esi # qhasm: (uint32) y0 ^= *(uint32 *) (table + 2 + q11 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,q12=int64#9d # asm 2: movzbl q12=%r11d movzbl %cl,%r11d # qhasm: q13 = (z1 >> 8) & 255 # asm 1: movzbl q13=int64#2d # asm 2: movzbl q13=%esi movzbl %ch,%esi # qhasm: y1 = *(uint32 *) (table + 4 + q01 * 8) # asm 1: movl 4(y1=int64#4d # asm 2: movl 4(y1=%ecx movl 4(%r8,%rbp,8),%ecx # qhasm: (uint32) y1 ^= *(uint32 *) (table + 3 + q10 * 8) # asm 1: xorl 3(q20=int64#8d # asm 2: movzbl q20=%r10d movzbl %al,%r10d # qhasm: q21 = (z2 >> 8) & 255 # asm 1: movzbl q21=int64#15d # asm 2: movzbl q21=%ebp movzbl %ah,%ebp # qhasm: (uint32) y1 ^= *(uint32 *) (table + 2 + q21 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,q22=int64#10d # asm 2: movzbl q22=%r12d movzbl %al,%r12d # qhasm: (uint32) y0 ^= *(uint32 *) (table + 1 + q22 * 8) # asm 1: xorl 1(> 8) & 255 # asm 1: movzbl q23=int64#15d # asm 2: movzbl q23=%ebp movzbl %ah,%ebp # qhasm: y2 = *(uint32 *) (table + 1 + q02 * 8) # asm 1: movl 1(y2=int64#7d # asm 2: movl 1(y2=%eax movl 1(%r8,%r9,8),%eax # qhasm: (uint32) y2 ^= *(uint32 *) (table + 4 + q13 * 8) # asm 1: xorl 4(q30=int64#6d # asm 2: movzbl q30=%r9d movzbl %bl,%r9d # qhasm: q31 = (z3 >> 8) & 255 # asm 1: movzbl q31=int64#2d # asm 2: movzbl q31=%esi movzbl %bh,%esi # qhasm: (uint32) y2 ^= *(uint32 *) (table + 2 + q31 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,q32=int64#2d # asm 2: movzbl q32=%esi movzbl %bl,%esi # qhasm: (uint32) y1 ^= *(uint32 *) (table + 1 + q32 * 8) # asm 1: xorl 1(> 8) & 255 # asm 1: movzbl q33=int64#14d # asm 2: movzbl q33=%ebx movzbl %bh,%ebx # qhasm: (uint32) y0 ^= *(uint32 *) (table + 4 + q33 * 8) # asm 1: xorl 4(y3=int64#14d # asm 2: movl 2(y3=%ebx movl 2(%r8,%rdi,8),%ebx # qhasm: (uint32) y3 ^= *(uint32 *) (table + 1 + q12 * 8) # asm 1: xorl 1(p00=int64#2d # asm 2: movzbl p00=%esi movzbl %dl,%esi # qhasm: p01 = (y0 >> 8) & 255 # asm 1: movzbl p01=int64#1d # asm 2: movzbl p01=%edi movzbl %dh,%edi # qhasm: (uint32) y0 >>= 16 # asm 1: shr $16,p02=int64#6d # asm 2: movzbl p02=%r9d movzbl %dl,%r9d # qhasm: p03 = (y0 >> 8) & 255 # asm 1: movzbl p03=int64#15d # asm 2: movzbl p03=%ebp movzbl %dh,%ebp # qhasm: z0 = *(uint32 *) (table + 3 + p00 * 8) # asm 1: movl 3(z0=int64#3d # asm 2: movl 3(z0=%edx movl 3(%r8,%rsi,8),%edx # qhasm: assign 3 to z0 # qhasm: p10 = y1 & 255 # asm 1: movzbl p10=int64#8d # asm 2: movzbl p10=%r10d movzbl %cl,%r10d # qhasm: p11 = (y1 >> 8) & 255 # asm 1: movzbl p11=int64#2d # asm 2: movzbl p11=%esi movzbl %ch,%esi # qhasm: (uint32) z0 ^= *(uint32 *) (table + 2 + p11 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,p12=int64#9d # asm 2: movzbl p12=%r11d movzbl %cl,%r11d # qhasm: p13 = (y1 >> 8) & 255 # asm 1: movzbl p13=int64#2d # asm 2: movzbl p13=%esi movzbl %ch,%esi # qhasm: z1 = *(uint32 *) (table + 4 + p03 * 8) # asm 1: movl 4(z1=int64#4d # asm 2: movl 4(z1=%ecx movl 4(%r8,%rbp,8),%ecx # qhasm: (uint32) z1 ^= *(uint32 *) (table + 3 + p10 * 8) # asm 1: xorl 3(p20=int64#8d # asm 2: movzbl p20=%r10d movzbl %al,%r10d # qhasm: p21 = (y2 >> 8) & 255 # asm 1: movzbl p21=int64#15d # asm 2: movzbl p21=%ebp movzbl %ah,%ebp # qhasm: (uint32) z1 ^= *(uint32 *) (table + 2 + p21 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,p22=int64#10d # asm 2: movzbl p22=%r12d movzbl %al,%r12d # qhasm: (uint32) z0 ^= *(uint32 *) (table + 1 + p22 * 8) # asm 1: xorl 1(> 8) & 255 # asm 1: movzbl p23=int64#15d # asm 2: movzbl p23=%ebp movzbl %ah,%ebp # qhasm: z2 = *(uint32 *) (table + 1 + p02 * 8) # asm 1: movl 1(z2=int64#7d # asm 2: movl 1(z2=%eax movl 1(%r8,%r9,8),%eax # qhasm: (uint32) z2 ^= *(uint32 *) (table + 4 + p13 * 8) # asm 1: xorl 4(p30=int64#6d # asm 2: movzbl p30=%r9d movzbl %bl,%r9d # qhasm: p31 = (y3 >> 8) & 255 # asm 1: movzbl p31=int64#2d # asm 2: movzbl p31=%esi movzbl %bh,%esi # qhasm: (uint32) z2 ^= *(uint32 *) (table + 2 + p31 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,p32=int64#2d # asm 2: movzbl p32=%esi movzbl %bl,%esi # qhasm: (uint32) z1 ^= *(uint32 *) (table + 1 + p32 * 8) # asm 1: xorl 1(> 8) & 255 # asm 1: movzbl p33=int64#14d # asm 2: movzbl p33=%ebx movzbl %bh,%ebx # qhasm: (uint32) z0 ^= *(uint32 *) (table + 4 + p33 * 8) # asm 1: xorl 4(z3=int64#14d # asm 2: movl 2(z3=%ebx movl 2(%r8,%rdi,8),%ebx # qhasm: (uint32) z3 ^= *(uint32 *) (table + 1 + p12 * 8) # asm 1: xorl 1(q00=int64#2d # asm 2: movzbl q00=%esi movzbl %dl,%esi # qhasm: q03 = (z0 >> 8) & 255 # asm 1: movzbl q03=int64#1d # asm 2: movzbl q03=%edi movzbl %dh,%edi # qhasm: (uint32) z0 >>= 16 # asm 1: shr $16,q02=int64#6d # asm 2: movzbl q02=%r9d movzbl %dl,%r9d # qhasm: q01 = (z0 >> 8) & 255 # asm 1: movzbl q01=int64#15d # asm 2: movzbl q01=%ebp movzbl %dh,%ebp # qhasm: y0 = *(uint32 *) (table + 3 + q00 * 8) # asm 1: movl 3(y0=int64#3d # asm 2: movl 3(y0=%edx movl 3(%r8,%rsi,8),%edx # qhasm: assign 3 to y0 # qhasm: q10 = z1 & 255 # asm 1: movzbl q10=int64#8d # asm 2: movzbl q10=%r10d movzbl %cl,%r10d # qhasm: q11 = (z1 >> 8) & 255 # asm 1: movzbl q11=int64#2d # asm 2: movzbl q11=%esi movzbl %ch,%esi # qhasm: (uint32) y0 ^= *(uint32 *) (table + 2 + q11 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,q12=int64#9d # asm 2: movzbl q12=%r11d movzbl %cl,%r11d # qhasm: q13 = (z1 >> 8) & 255 # asm 1: movzbl q13=int64#2d # asm 2: movzbl q13=%esi movzbl %ch,%esi # qhasm: y1 = *(uint32 *) (table + 4 + q01 * 8) # asm 1: movl 4(y1=int64#4d # asm 2: movl 4(y1=%ecx movl 4(%r8,%rbp,8),%ecx # qhasm: (uint32) y1 ^= *(uint32 *) (table + 3 + q10 * 8) # asm 1: xorl 3(q20=int64#8d # asm 2: movzbl q20=%r10d movzbl %al,%r10d # qhasm: q21 = (z2 >> 8) & 255 # asm 1: movzbl q21=int64#15d # asm 2: movzbl q21=%ebp movzbl %ah,%ebp # qhasm: (uint32) y1 ^= *(uint32 *) (table + 2 + q21 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,q22=int64#10d # asm 2: movzbl q22=%r12d movzbl %al,%r12d # qhasm: (uint32) y0 ^= *(uint32 *) (table + 1 + q22 * 8) # asm 1: xorl 1(> 8) & 255 # asm 1: movzbl q23=int64#15d # asm 2: movzbl q23=%ebp movzbl %ah,%ebp # qhasm: y2 = *(uint32 *) (table + 1 + q02 * 8) # asm 1: movl 1(y2=int64#7d # asm 2: movl 1(y2=%eax movl 1(%r8,%r9,8),%eax # qhasm: (uint32) y2 ^= *(uint32 *) (table + 4 + q13 * 8) # asm 1: xorl 4(q30=int64#6d # asm 2: movzbl q30=%r9d movzbl %bl,%r9d # qhasm: q31 = (z3 >> 8) & 255 # asm 1: movzbl q31=int64#2d # asm 2: movzbl q31=%esi movzbl %bh,%esi # qhasm: (uint32) y2 ^= *(uint32 *) (table + 2 + q31 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,q32=int64#2d # asm 2: movzbl q32=%esi movzbl %bl,%esi # qhasm: (uint32) y1 ^= *(uint32 *) (table + 1 + q32 * 8) # asm 1: xorl 1(> 8) & 255 # asm 1: movzbl q33=int64#14d # asm 2: movzbl q33=%ebx movzbl %bh,%ebx # qhasm: (uint32) y0 ^= *(uint32 *) (table + 4 + q33 * 8) # asm 1: xorl 4(y3=int64#14d # asm 2: movl 2(y3=%ebx movl 2(%r8,%rdi,8),%ebx # qhasm: (uint32) y3 ^= *(uint32 *) (table + 1 + q12 * 8) # asm 1: xorl 1(p00=int64#2d # asm 2: movzbl p00=%esi movzbl %dl,%esi # qhasm: p01 = (y0 >> 8) & 255 # asm 1: movzbl p01=int64#1d # asm 2: movzbl p01=%edi movzbl %dh,%edi # qhasm: (uint32) y0 >>= 16 # asm 1: shr $16,p02=int64#6d # asm 2: movzbl p02=%r9d movzbl %dl,%r9d # qhasm: p03 = (y0 >> 8) & 255 # asm 1: movzbl p03=int64#15d # asm 2: movzbl p03=%ebp movzbl %dh,%ebp # qhasm: z0 = *(uint32 *) (table + 3 + p00 * 8) # asm 1: movl 3(z0=int64#3d # asm 2: movl 3(z0=%edx movl 3(%r8,%rsi,8),%edx # qhasm: assign 3 to z0 # qhasm: p10 = y1 & 255 # asm 1: movzbl p10=int64#8d # asm 2: movzbl p10=%r10d movzbl %cl,%r10d # qhasm: p11 = (y1 >> 8) & 255 # asm 1: movzbl p11=int64#2d # asm 2: movzbl p11=%esi movzbl %ch,%esi # qhasm: (uint32) z0 ^= *(uint32 *) (table + 2 + p11 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,p12=int64#9d # asm 2: movzbl p12=%r11d movzbl %cl,%r11d # qhasm: p13 = (y1 >> 8) & 255 # asm 1: movzbl p13=int64#2d # asm 2: movzbl p13=%esi movzbl %ch,%esi # qhasm: z1 = *(uint32 *) (table + 4 + p03 * 8) # asm 1: movl 4(z1=int64#4d # asm 2: movl 4(z1=%ecx movl 4(%r8,%rbp,8),%ecx # qhasm: (uint32) z1 ^= *(uint32 *) (table + 3 + p10 * 8) # asm 1: xorl 3(p20=int64#8d # asm 2: movzbl p20=%r10d movzbl %al,%r10d # qhasm: p21 = (y2 >> 8) & 255 # asm 1: movzbl p21=int64#15d # asm 2: movzbl p21=%ebp movzbl %ah,%ebp # qhasm: (uint32) z1 ^= *(uint32 *) (table + 2 + p21 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,p22=int64#10d # asm 2: movzbl p22=%r12d movzbl %al,%r12d # qhasm: (uint32) z0 ^= *(uint32 *) (table + 1 + p22 * 8) # asm 1: xorl 1(> 8) & 255 # asm 1: movzbl p23=int64#15d # asm 2: movzbl p23=%ebp movzbl %ah,%ebp # qhasm: z2 = *(uint32 *) (table + 1 + p02 * 8) # asm 1: movl 1(z2=int64#7d # asm 2: movl 1(z2=%eax movl 1(%r8,%r9,8),%eax # qhasm: (uint32) z2 ^= *(uint32 *) (table + 4 + p13 * 8) # asm 1: xorl 4(p30=int64#6d # asm 2: movzbl p30=%r9d movzbl %bl,%r9d # qhasm: p31 = (y3 >> 8) & 255 # asm 1: movzbl p31=int64#2d # asm 2: movzbl p31=%esi movzbl %bh,%esi # qhasm: (uint32) z2 ^= *(uint32 *) (table + 2 + p31 * 8) # asm 1: xorl 2(>= 16 # asm 1: shr $16,p32=int64#2d # asm 2: movzbl p32=%esi movzbl %bl,%esi # qhasm: (uint32) z1 ^= *(uint32 *) (table + 1 + p32 * 8) # asm 1: xorl 1(> 8) & 255 # asm 1: movzbl p33=int64#14d # asm 2: movzbl p33=%ebx movzbl %bh,%ebx # qhasm: (uint32) z0 ^= *(uint32 *) (table + 4 + p33 * 8) # asm 1: xorl 4(z3=int64#14d # asm 2: movl 2(z3=%ebx movl 2(%r8,%rdi,8),%ebx # qhasm: (uint32) z3 ^= *(uint32 *) (table + 1 + p12 * 8) # asm 1: xorl 1(y0=int64#1d # asm 2: movzbl y0=%edi movzbl %dl,%edi # qhasm: y0 = *(uint8 *) (table + 1 + y0 * 8) # asm 1: movzbq 1(y0=int64#2 # asm 2: movzbq 1(y0=%rsi movzbq 1(%r8,%rdi,8),%rsi # qhasm: y3 = (z0 >> 8) & 255 # asm 1: movzbl y3=int64#1d # asm 2: movzbl y3=%edi movzbl %dh,%edi # qhasm: y3 = *(uint16 *) (table + y3 * 8) # asm 1: movzwq (y3=int64#6 # asm 2: movzwq (y3=%r9 movzwq (%r8,%rdi,8),%r9 # qhasm: (uint32) z0 >>= 16 # asm 1: shr $16,y2=int64#1d # asm 2: movzbl y2=%edi movzbl %dl,%edi # qhasm: y2 = *(uint32 *) (table + 3 + y2 * 8) # asm 1: movl 3(y2=int64#8d # asm 2: movl 3(y2=%r10d movl 3(%r8,%rdi,8),%r10d # qhasm: (uint32) y2 &= 0x00ff0000 # asm 1: and $0x00ff0000,> 8) & 255 # asm 1: movzbl y1=int64#1d # asm 2: movzbl y1=%edi movzbl %dh,%edi # qhasm: y1 = *(uint32 *) (table + 2 + y1 * 8) # asm 1: movl 2(y1=int64#3d # asm 2: movl 2(y1=%edx movl 2(%r8,%rdi,8),%edx # qhasm: (uint32) y1 &= 0xff000000 # asm 1: and $0xff000000,q10=int64#1d # asm 2: movzbl q10=%edi movzbl %cl,%edi # qhasm: q10 = *(uint8 *) (table + 1 + q10 * 8) # asm 1: movzbq 1(q10=int64#1 # asm 2: movzbq 1(q10=%rdi movzbq 1(%r8,%rdi,8),%rdi # qhasm: y1 ^= q10 # asm 1: xor > 8) & 255 # asm 1: movzbl q11=int64#1d # asm 2: movzbl q11=%edi movzbl %ch,%edi # qhasm: q11 = *(uint16 *) (table + q11 * 8) # asm 1: movzwq (q11=int64#1 # asm 2: movzwq (q11=%rdi movzwq (%r8,%rdi,8),%rdi # qhasm: y0 ^= q11 # asm 1: xor >= 16 # asm 1: shr $16,q12=int64#1d # asm 2: movzbl q12=%edi movzbl %cl,%edi # qhasm: q12 = *(uint32 *) (table + 3 + q12 * 8) # asm 1: movl 3(q12=int64#1d # asm 2: movl 3(q12=%edi movl 3(%r8,%rdi,8),%edi # qhasm: (uint32) q12 &= 0x00ff0000 # asm 1: and $0x00ff0000,> 8) & 255 # asm 1: movzbl q13=int64#1d # asm 2: movzbl q13=%edi movzbl %ch,%edi # qhasm: q13 = *(uint32 *) (table + 2 + q13 * 8) # asm 1: movl 2(q13=int64#1d # asm 2: movl 2(q13=%edi movl 2(%r8,%rdi,8),%edi # qhasm: (uint32) q13 &= 0xff000000 # asm 1: and $0xff000000,q20=int64#1d # asm 2: movzbl q20=%edi movzbl %al,%edi # qhasm: q20 = *(uint8 *) (table + 1 + q20 * 8) # asm 1: movzbq 1(q20=int64#1 # asm 2: movzbq 1(q20=%rdi movzbq 1(%r8,%rdi,8),%rdi # qhasm: y2 ^= q20 # asm 1: xor > 8) & 255 # asm 1: movzbl q21=int64#1d # asm 2: movzbl q21=%edi movzbl %ah,%edi # qhasm: q21 = *(uint16 *) (table + q21 * 8) # asm 1: movzwq (q21=int64#1 # asm 2: movzwq (q21=%rdi movzwq (%r8,%rdi,8),%rdi # qhasm: y1 ^= q21 # asm 1: xor >= 16 # asm 1: shr $16,q22=int64#1d # asm 2: movzbl q22=%edi movzbl %al,%edi # qhasm: q22 = *(uint32 *) (table + 3 + q22 * 8) # asm 1: movl 3(q22=int64#1d # asm 2: movl 3(q22=%edi movl 3(%r8,%rdi,8),%edi # qhasm: (uint32) q22 &= 0x00ff0000 # asm 1: and $0x00ff0000,> 8) & 255 # asm 1: movzbl q23=int64#1d # asm 2: movzbl q23=%edi movzbl %ah,%edi # qhasm: q23 = *(uint32 *) (table + 2 + q23 * 8) # asm 1: movl 2(q23=int64#1d # asm 2: movl 2(q23=%edi movl 2(%r8,%rdi,8),%edi # qhasm: (uint32) q23 &= 0xff000000 # asm 1: and $0xff000000,q30=int64#1d # asm 2: movzbl q30=%edi movzbl %bl,%edi # qhasm: q30 = *(uint8 *) (table + 1 + q30 * 8) # asm 1: movzbq 1(q30=int64#1 # asm 2: movzbq 1(q30=%rdi movzbq 1(%r8,%rdi,8),%rdi # qhasm: y3 ^= q30 # asm 1: xor > 8) & 255 # asm 1: movzbl q31=int64#1d # asm 2: movzbl q31=%edi movzbl %bh,%edi # qhasm: q31 = *(uint16 *) (table + q31 * 8) # asm 1: movzwq (q31=int64#1 # asm 2: movzwq (q31=%rdi movzwq (%r8,%rdi,8),%rdi # qhasm: y2 ^= q31 # asm 1: xor >= 16 # asm 1: shr $16,q32=int64#1d # asm 2: movzbl q32=%edi movzbl %bl,%edi # qhasm: q32 = *(uint32 *) (table + 3 + q32 * 8) # asm 1: movl 3(q32=int64#1d # asm 2: movl 3(q32=%edi movl 3(%r8,%rdi,8),%edi # qhasm: (uint32) q32 &= 0x00ff0000 # asm 1: and $0x00ff0000,> 8) & 255 # asm 1: movzbl q33=int64#1d # asm 2: movzbl q33=%edi movzbl %bh,%edi # qhasm: q33 = *(uint32 *) (table + 2 + q33 * 8) # asm 1: movl 2(q33=int64#1d # asm 2: movl 2(q33=%edi movl 2(%r8,%rdi,8),%edi # qhasm: (uint32) q33 &= 0xff000000 # asm 1: and $0xff000000,out=int64#1 # asm 2: movq out=%rdi movq 0(%rsp),%rdi # qhasm: *(uint32 *) (out + 0) = y0 # asm 1: movl r11_caller=int64#9 # asm 2: movq r11_caller=%r11 movq 8(%rsp),%r11 # qhasm: r12_caller = r12_stack # asm 1: movq r12_caller=int64#10 # asm 2: movq r12_caller=%r12 movq 16(%rsp),%r12 # qhasm: r13_caller = r13_stack # asm 1: movq r13_caller=int64#11 # asm 2: movq r13_caller=%r13 movq 24(%rsp),%r13 # qhasm: r14_caller = r14_stack # asm 1: movq r14_caller=int64#12 # asm 2: movq r14_caller=%r14 movq 32(%rsp),%r14 # qhasm: r15_caller = r15_stack # asm 1: movq r15_caller=int64#13 # asm 2: movq r15_caller=%r15 movq 40(%rsp),%r15 # qhasm: rbx_caller = rbx_stack # asm 1: movq rbx_caller=int64#14 # asm 2: movq rbx_caller=%rbx movq 48(%rsp),%rbx # qhasm: rbp_caller = rbp_stack # asm 1: movq rbp_caller=int64#15 # asm 2: movq rbp_caller=%rbp movq 56(%rsp),%rbp # qhasm: leave add %r11,%rsp mov %rdi,%rax mov %rsi,%rdx ret