#include "namespace.h" #define MASK0_0 CRYPTO_NAMESPACE(MASK0_0) #define _MASK0_0 _CRYPTO_NAMESPACE(MASK0_0) #define MASK0_1 CRYPTO_NAMESPACE(MASK0_1) #define _MASK0_1 _CRYPTO_NAMESPACE(MASK0_1) #define MASK1_0 CRYPTO_NAMESPACE(MASK1_0) #define _MASK1_0 _CRYPTO_NAMESPACE(MASK1_0) #define MASK1_1 CRYPTO_NAMESPACE(MASK1_1) #define _MASK1_1 _CRYPTO_NAMESPACE(MASK1_1) #define MASK2_0 CRYPTO_NAMESPACE(MASK2_0) #define _MASK2_0 _CRYPTO_NAMESPACE(MASK2_0) #define MASK2_1 CRYPTO_NAMESPACE(MASK2_1) #define _MASK2_1 _CRYPTO_NAMESPACE(MASK2_1) #define MASK3_0 CRYPTO_NAMESPACE(MASK3_0) #define _MASK3_0 _CRYPTO_NAMESPACE(MASK3_0) #define MASK3_1 CRYPTO_NAMESPACE(MASK3_1) #define _MASK3_1 _CRYPTO_NAMESPACE(MASK3_1) #define MASK4_0 CRYPTO_NAMESPACE(MASK4_0) #define _MASK4_0 _CRYPTO_NAMESPACE(MASK4_0) #define MASK4_1 CRYPTO_NAMESPACE(MASK4_1) #define _MASK4_1 _CRYPTO_NAMESPACE(MASK4_1) #define MASK5_0 CRYPTO_NAMESPACE(MASK5_0) #define _MASK5_0 _CRYPTO_NAMESPACE(MASK5_0) #define MASK5_1 CRYPTO_NAMESPACE(MASK5_1) #define _MASK5_1 _CRYPTO_NAMESPACE(MASK5_1) #define transpose_64x64_asm CRYPTO_NAMESPACE(transpose_64x64_asm) #define _transpose_64x64_asm _CRYPTO_NAMESPACE(transpose_64x64_asm) # qhasm: int64 input_0 # qhasm: int64 input_1 # qhasm: int64 input_2 # qhasm: int64 input_3 # qhasm: int64 input_4 # qhasm: int64 input_5 # qhasm: stack64 input_6 # qhasm: stack64 input_7 # qhasm: int64 caller_r11 # qhasm: int64 caller_r12 # qhasm: int64 caller_r13 # qhasm: int64 caller_r14 # qhasm: int64 caller_r15 # qhasm: int64 caller_rbx # qhasm: int64 caller_rbp # qhasm: reg128 r0 # qhasm: reg128 r1 # qhasm: reg128 r2 # qhasm: reg128 r3 # qhasm: reg128 r4 # qhasm: reg128 r5 # qhasm: reg128 r6 # qhasm: reg128 r7 # qhasm: reg128 t0 # qhasm: reg128 t1 # qhasm: reg128 v00 # qhasm: reg128 v01 # qhasm: reg128 v10 # qhasm: reg128 v11 # qhasm: int64 buf # qhasm: reg128 mask0 # qhasm: reg128 mask1 # qhasm: reg128 mask2 # qhasm: reg128 mask3 # qhasm: reg128 mask4 # qhasm: reg128 mask5 # qhasm: enter transpose_64x64_asm .p2align 5 .global _transpose_64x64_asm .global transpose_64x64_asm _transpose_64x64_asm: transpose_64x64_asm: mov % rsp, % r11 and $31, % r11 add $0, % r11 sub % r11, % rsp # qhasm: mask0 aligned= mem128[ MASK5_0 ] # asm 1: movdqa MASK5_0(%rip),>mask0=reg128#1 # asm 2: movdqa MASK5_0(%rip),>mask0=%xmm0 movdqa MASK5_0( % rip), % xmm0 # qhasm: mask1 aligned= mem128[ MASK5_1 ] # asm 1: movdqa MASK5_1(%rip),>mask1=reg128#2 # asm 2: movdqa MASK5_1(%rip),>mask1=%xmm1 movdqa MASK5_1( % rip), % xmm1 # qhasm: mask2 aligned= mem128[ MASK4_0 ] # asm 1: movdqa MASK4_0(%rip),>mask2=reg128#3 # asm 2: movdqa MASK4_0(%rip),>mask2=%xmm2 movdqa MASK4_0( % rip), % xmm2 # qhasm: mask3 aligned= mem128[ MASK4_1 ] # asm 1: movdqa MASK4_1(%rip),>mask3=reg128#4 # asm 2: movdqa MASK4_1(%rip),>mask3=%xmm3 movdqa MASK4_1( % rip), % xmm3 # qhasm: mask4 aligned= mem128[ MASK3_0 ] # asm 1: movdqa MASK3_0(%rip),>mask4=reg128#5 # asm 2: movdqa MASK3_0(%rip),>mask4=%xmm4 movdqa MASK3_0( % rip), % xmm4 # qhasm: mask5 aligned= mem128[ MASK3_1 ] # asm 1: movdqa MASK3_1(%rip),>mask5=reg128#6 # asm 2: movdqa MASK3_1(%rip),>mask5=%xmm5 movdqa MASK3_1( % rip), % xmm5 # qhasm: r0 = mem64[ input_0 + 0 ] x2 # asm 1: movddup 0(r0=reg128#7 # asm 2: movddup 0(r0=%xmm6 movddup 0( % rdi), % xmm6 # qhasm: r1 = mem64[ input_0 + 64 ] x2 # asm 1: movddup 64(r1=reg128#8 # asm 2: movddup 64(r1=%xmm7 movddup 64( % rdi), % xmm7 # qhasm: r2 = mem64[ input_0 + 128 ] x2 # asm 1: movddup 128(r2=reg128#9 # asm 2: movddup 128(r2=%xmm8 movddup 128( % rdi), % xmm8 # qhasm: r3 = mem64[ input_0 + 192 ] x2 # asm 1: movddup 192(r3=reg128#10 # asm 2: movddup 192(r3=%xmm9 movddup 192( % rdi), % xmm9 # qhasm: r4 = mem64[ input_0 + 256 ] x2 # asm 1: movddup 256(r4=reg128#11 # asm 2: movddup 256(r4=%xmm10 movddup 256( % rdi), % xmm10 # qhasm: r5 = mem64[ input_0 + 320 ] x2 # asm 1: movddup 320(r5=reg128#12 # asm 2: movddup 320(r5=%xmm11 movddup 320( % rdi), % xmm11 # qhasm: r6 = mem64[ input_0 + 384 ] x2 # asm 1: movddup 384(r6=reg128#13 # asm 2: movddup 384(r6=%xmm12 movddup 384( % rdi), % xmm12 # qhasm: r7 = mem64[ input_0 + 448 ] x2 # asm 1: movddup 448(r7=reg128#14 # asm 2: movddup 448(r7=%xmm13 movddup 448( % rdi), % xmm13 # qhasm: v00 = r0 & mask0 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm0, % xmm6, % xmm14 # qhasm: 2x v10 = r4 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm10, % xmm15 # qhasm: 2x v01 = r0 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#7 # asm 2: vpsrlq $32,v01=%xmm6 vpsrlq $32, % xmm6, % xmm6 # qhasm: v11 = r4 & mask1 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm1, % xmm10, % xmm10 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#15 # asm 2: vpor r0=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r4 = v01 | v11 # asm 1: vpor r4=reg128#7 # asm 2: vpor r4=%xmm6 vpor % xmm10, % xmm6, % xmm6 # qhasm: v00 = r1 & mask0 # asm 1: vpand v00=reg128#11 # asm 2: vpand v00=%xmm10 vpand % xmm0, % xmm7, % xmm10 # qhasm: 2x v10 = r5 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm11, % xmm15 # qhasm: 2x v01 = r1 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#8 # asm 2: vpsrlq $32,v01=%xmm7 vpsrlq $32, % xmm7, % xmm7 # qhasm: v11 = r5 & mask1 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm1, % xmm11, % xmm11 # qhasm: r1 = v00 | v10 # asm 1: vpor r1=reg128#11 # asm 2: vpor r1=%xmm10 vpor % xmm15, % xmm10, % xmm10 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#8 # asm 2: vpor r5=%xmm7 vpor % xmm11, % xmm7, % xmm7 # qhasm: v00 = r2 & mask0 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm0, % xmm8, % xmm11 # qhasm: 2x v10 = r6 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm12, % xmm15 # qhasm: 2x v01 = r2 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#9 # asm 2: vpsrlq $32,v01=%xmm8 vpsrlq $32, % xmm8, % xmm8 # qhasm: v11 = r6 & mask1 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm1, % xmm12, % xmm12 # qhasm: r2 = v00 | v10 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#9 # asm 2: vpor r6=%xmm8 vpor % xmm12, % xmm8, % xmm8 # qhasm: v00 = r3 & mask0 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm0, % xmm9, % xmm12 # qhasm: 2x v10 = r7 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm13, % xmm15 # qhasm: 2x v01 = r3 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#10 # asm 2: vpsrlq $32,v01=%xmm9 vpsrlq $32, % xmm9, % xmm9 # qhasm: v11 = r7 & mask1 # asm 1: vpand v11=reg128#14 # asm 2: vpand v11=%xmm13 vpand % xmm1, % xmm13, % xmm13 # qhasm: r3 = v00 | v10 # asm 1: vpor r3=reg128#13 # asm 2: vpor r3=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#10 # asm 2: vpor r7=%xmm9 vpor % xmm13, % xmm9, % xmm9 # qhasm: v00 = r0 & mask2 # asm 1: vpand v00=reg128#14 # asm 2: vpand v00=%xmm13 vpand % xmm2, % xmm14, % xmm13 # qhasm: 4x v10 = r2 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm11, % xmm15 # qhasm: 4x v01 = r0 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#15 # asm 2: vpsrld $16,v01=%xmm14 vpsrld $16, % xmm14, % xmm14 # qhasm: v11 = r2 & mask3 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm3, % xmm11, % xmm11 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#14 # asm 2: vpor r0=%xmm13 vpor % xmm15, % xmm13, % xmm13 # qhasm: r2 = v01 | v11 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm11, % xmm14, % xmm11 # qhasm: v00 = r1 & mask2 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm2, % xmm10, % xmm14 # qhasm: 4x v10 = r3 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm12, % xmm15 # qhasm: 4x v01 = r1 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#11 # asm 2: vpsrld $16,v01=%xmm10 vpsrld $16, % xmm10, % xmm10 # qhasm: v11 = r3 & mask3 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm3, % xmm12, % xmm12 # qhasm: r1 = v00 | v10 # asm 1: vpor r1=reg128#15 # asm 2: vpor r1=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm12, % xmm10, % xmm10 # qhasm: v00 = r4 & mask2 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm2, % xmm6, % xmm12 # qhasm: 4x v10 = r6 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm8, % xmm15 # qhasm: 4x v01 = r4 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#7 # asm 2: vpsrld $16,v01=%xmm6 vpsrld $16, % xmm6, % xmm6 # qhasm: v11 = r6 & mask3 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm3, % xmm8, % xmm8 # qhasm: r4 = v00 | v10 # asm 1: vpor r4=reg128#13 # asm 2: vpor r4=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#7 # asm 2: vpor r6=%xmm6 vpor % xmm8, % xmm6, % xmm6 # qhasm: v00 = r5 & mask2 # asm 1: vpand v00=reg128#9 # asm 2: vpand v00=%xmm8 vpand % xmm2, % xmm7, % xmm8 # qhasm: 4x v10 = r7 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm9, % xmm15 # qhasm: 4x v01 = r5 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#8 # asm 2: vpsrld $16,v01=%xmm7 vpsrld $16, % xmm7, % xmm7 # qhasm: v11 = r7 & mask3 # asm 1: vpand v11=reg128#10 # asm 2: vpand v11=%xmm9 vpand % xmm3, % xmm9, % xmm9 # qhasm: r5 = v00 | v10 # asm 1: vpor r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm15, % xmm8, % xmm8 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#8 # asm 2: vpor r7=%xmm7 vpor % xmm9, % xmm7, % xmm7 # qhasm: v00 = r0 & mask4 # asm 1: vpand v00=reg128#10 # asm 2: vpand v00=%xmm9 vpand % xmm4, % xmm13, % xmm9 # qhasm: 8x v10 = r1 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm14, % xmm15 # qhasm: 8x v01 = r0 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#14 # asm 2: vpsrlw $8,v01=%xmm13 vpsrlw $8, % xmm13, % xmm13 # qhasm: v11 = r1 & mask5 # asm 1: vpand v11=reg128#15 # asm 2: vpand v11=%xmm14 vpand % xmm5, % xmm14, % xmm14 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#10 # asm 2: vpor r0=%xmm9 vpor % xmm15, % xmm9, % xmm9 # qhasm: r1 = v01 | v11 # asm 1: vpor r1=reg128#14 # asm 2: vpor r1=%xmm13 vpor % xmm14, % xmm13, % xmm13 # qhasm: v00 = r2 & mask4 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm4, % xmm11, % xmm14 # qhasm: 8x v10 = r3 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm10, % xmm15 # qhasm: 8x v01 = r2 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#12 # asm 2: vpsrlw $8,v01=%xmm11 vpsrlw $8, % xmm11, % xmm11 # qhasm: v11 = r3 & mask5 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm5, % xmm10, % xmm10 # qhasm: r2 = v00 | v10 # asm 1: vpor r2=reg128#15 # asm 2: vpor r2=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm10, % xmm11, % xmm10 # qhasm: v00 = r4 & mask4 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm4, % xmm12, % xmm11 # qhasm: 8x v10 = r5 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm8, % xmm15 # qhasm: 8x v01 = r4 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#13 # asm 2: vpsrlw $8,v01=%xmm12 vpsrlw $8, % xmm12, % xmm12 # qhasm: v11 = r5 & mask5 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm5, % xmm8, % xmm8 # qhasm: r4 = v00 | v10 # asm 1: vpor r4=reg128#12 # asm 2: vpor r4=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm8, % xmm12, % xmm8 # qhasm: v00 = r6 & mask4 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm4, % xmm6, % xmm12 # qhasm: 8x v10 = r7 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm7, % xmm15 # qhasm: 8x v01 = r6 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#7 # asm 2: vpsrlw $8,v01=%xmm6 vpsrlw $8, % xmm6, % xmm6 # qhasm: v11 = r7 & mask5 # asm 1: vpand v11=reg128#8 # asm 2: vpand v11=%xmm7 vpand % xmm5, % xmm7, % xmm7 # qhasm: r6 = v00 | v10 # asm 1: vpor r6=reg128#13 # asm 2: vpor r6=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#7 # asm 2: vpor r7=%xmm6 vpor % xmm7, % xmm6, % xmm6 # qhasm: buf = r0[0] # asm 1: pextrq $0x0,buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm9, % rsi # qhasm: mem64[ input_0 + 0 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm13, % rsi # qhasm: mem64[ input_0 + 64 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm14, % rsi # qhasm: mem64[ input_0 + 128 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm10, % rsi # qhasm: mem64[ input_0 + 192 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm11, % rsi # qhasm: mem64[ input_0 + 256 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm8, % rsi # qhasm: mem64[ input_0 + 320 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm12, % rsi # qhasm: mem64[ input_0 + 384 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm6, % rsi # qhasm: mem64[ input_0 + 448 ] = buf # asm 1: movq r0=reg128#7 # asm 2: movddup 8(r0=%xmm6 movddup 8( % rdi), % xmm6 # qhasm: r1 = mem64[ input_0 + 72 ] x2 # asm 1: movddup 72(r1=reg128#8 # asm 2: movddup 72(r1=%xmm7 movddup 72( % rdi), % xmm7 # qhasm: r2 = mem64[ input_0 + 136 ] x2 # asm 1: movddup 136(r2=reg128#9 # asm 2: movddup 136(r2=%xmm8 movddup 136( % rdi), % xmm8 # qhasm: r3 = mem64[ input_0 + 200 ] x2 # asm 1: movddup 200(r3=reg128#10 # asm 2: movddup 200(r3=%xmm9 movddup 200( % rdi), % xmm9 # qhasm: r4 = mem64[ input_0 + 264 ] x2 # asm 1: movddup 264(r4=reg128#11 # asm 2: movddup 264(r4=%xmm10 movddup 264( % rdi), % xmm10 # qhasm: r5 = mem64[ input_0 + 328 ] x2 # asm 1: movddup 328(r5=reg128#12 # asm 2: movddup 328(r5=%xmm11 movddup 328( % rdi), % xmm11 # qhasm: r6 = mem64[ input_0 + 392 ] x2 # asm 1: movddup 392(r6=reg128#13 # asm 2: movddup 392(r6=%xmm12 movddup 392( % rdi), % xmm12 # qhasm: r7 = mem64[ input_0 + 456 ] x2 # asm 1: movddup 456(r7=reg128#14 # asm 2: movddup 456(r7=%xmm13 movddup 456( % rdi), % xmm13 # qhasm: v00 = r0 & mask0 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm0, % xmm6, % xmm14 # qhasm: 2x v10 = r4 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm10, % xmm15 # qhasm: 2x v01 = r0 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#7 # asm 2: vpsrlq $32,v01=%xmm6 vpsrlq $32, % xmm6, % xmm6 # qhasm: v11 = r4 & mask1 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm1, % xmm10, % xmm10 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#15 # asm 2: vpor r0=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r4 = v01 | v11 # asm 1: vpor r4=reg128#7 # asm 2: vpor r4=%xmm6 vpor % xmm10, % xmm6, % xmm6 # qhasm: v00 = r1 & mask0 # asm 1: vpand v00=reg128#11 # asm 2: vpand v00=%xmm10 vpand % xmm0, % xmm7, % xmm10 # qhasm: 2x v10 = r5 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm11, % xmm15 # qhasm: 2x v01 = r1 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#8 # asm 2: vpsrlq $32,v01=%xmm7 vpsrlq $32, % xmm7, % xmm7 # qhasm: v11 = r5 & mask1 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm1, % xmm11, % xmm11 # qhasm: r1 = v00 | v10 # asm 1: vpor r1=reg128#11 # asm 2: vpor r1=%xmm10 vpor % xmm15, % xmm10, % xmm10 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#8 # asm 2: vpor r5=%xmm7 vpor % xmm11, % xmm7, % xmm7 # qhasm: v00 = r2 & mask0 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm0, % xmm8, % xmm11 # qhasm: 2x v10 = r6 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm12, % xmm15 # qhasm: 2x v01 = r2 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#9 # asm 2: vpsrlq $32,v01=%xmm8 vpsrlq $32, % xmm8, % xmm8 # qhasm: v11 = r6 & mask1 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm1, % xmm12, % xmm12 # qhasm: r2 = v00 | v10 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#9 # asm 2: vpor r6=%xmm8 vpor % xmm12, % xmm8, % xmm8 # qhasm: v00 = r3 & mask0 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm0, % xmm9, % xmm12 # qhasm: 2x v10 = r7 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm13, % xmm15 # qhasm: 2x v01 = r3 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#10 # asm 2: vpsrlq $32,v01=%xmm9 vpsrlq $32, % xmm9, % xmm9 # qhasm: v11 = r7 & mask1 # asm 1: vpand v11=reg128#14 # asm 2: vpand v11=%xmm13 vpand % xmm1, % xmm13, % xmm13 # qhasm: r3 = v00 | v10 # asm 1: vpor r3=reg128#13 # asm 2: vpor r3=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#10 # asm 2: vpor r7=%xmm9 vpor % xmm13, % xmm9, % xmm9 # qhasm: v00 = r0 & mask2 # asm 1: vpand v00=reg128#14 # asm 2: vpand v00=%xmm13 vpand % xmm2, % xmm14, % xmm13 # qhasm: 4x v10 = r2 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm11, % xmm15 # qhasm: 4x v01 = r0 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#15 # asm 2: vpsrld $16,v01=%xmm14 vpsrld $16, % xmm14, % xmm14 # qhasm: v11 = r2 & mask3 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm3, % xmm11, % xmm11 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#14 # asm 2: vpor r0=%xmm13 vpor % xmm15, % xmm13, % xmm13 # qhasm: r2 = v01 | v11 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm11, % xmm14, % xmm11 # qhasm: v00 = r1 & mask2 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm2, % xmm10, % xmm14 # qhasm: 4x v10 = r3 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm12, % xmm15 # qhasm: 4x v01 = r1 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#11 # asm 2: vpsrld $16,v01=%xmm10 vpsrld $16, % xmm10, % xmm10 # qhasm: v11 = r3 & mask3 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm3, % xmm12, % xmm12 # qhasm: r1 = v00 | v10 # asm 1: vpor r1=reg128#15 # asm 2: vpor r1=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm12, % xmm10, % xmm10 # qhasm: v00 = r4 & mask2 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm2, % xmm6, % xmm12 # qhasm: 4x v10 = r6 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm8, % xmm15 # qhasm: 4x v01 = r4 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#7 # asm 2: vpsrld $16,v01=%xmm6 vpsrld $16, % xmm6, % xmm6 # qhasm: v11 = r6 & mask3 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm3, % xmm8, % xmm8 # qhasm: r4 = v00 | v10 # asm 1: vpor r4=reg128#13 # asm 2: vpor r4=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#7 # asm 2: vpor r6=%xmm6 vpor % xmm8, % xmm6, % xmm6 # qhasm: v00 = r5 & mask2 # asm 1: vpand v00=reg128#9 # asm 2: vpand v00=%xmm8 vpand % xmm2, % xmm7, % xmm8 # qhasm: 4x v10 = r7 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm9, % xmm15 # qhasm: 4x v01 = r5 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#8 # asm 2: vpsrld $16,v01=%xmm7 vpsrld $16, % xmm7, % xmm7 # qhasm: v11 = r7 & mask3 # asm 1: vpand v11=reg128#10 # asm 2: vpand v11=%xmm9 vpand % xmm3, % xmm9, % xmm9 # qhasm: r5 = v00 | v10 # asm 1: vpor r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm15, % xmm8, % xmm8 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#8 # asm 2: vpor r7=%xmm7 vpor % xmm9, % xmm7, % xmm7 # qhasm: v00 = r0 & mask4 # asm 1: vpand v00=reg128#10 # asm 2: vpand v00=%xmm9 vpand % xmm4, % xmm13, % xmm9 # qhasm: 8x v10 = r1 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm14, % xmm15 # qhasm: 8x v01 = r0 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#14 # asm 2: vpsrlw $8,v01=%xmm13 vpsrlw $8, % xmm13, % xmm13 # qhasm: v11 = r1 & mask5 # asm 1: vpand v11=reg128#15 # asm 2: vpand v11=%xmm14 vpand % xmm5, % xmm14, % xmm14 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#10 # asm 2: vpor r0=%xmm9 vpor % xmm15, % xmm9, % xmm9 # qhasm: r1 = v01 | v11 # asm 1: vpor r1=reg128#14 # asm 2: vpor r1=%xmm13 vpor % xmm14, % xmm13, % xmm13 # qhasm: v00 = r2 & mask4 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm4, % xmm11, % xmm14 # qhasm: 8x v10 = r3 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm10, % xmm15 # qhasm: 8x v01 = r2 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#12 # asm 2: vpsrlw $8,v01=%xmm11 vpsrlw $8, % xmm11, % xmm11 # qhasm: v11 = r3 & mask5 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm5, % xmm10, % xmm10 # qhasm: r2 = v00 | v10 # asm 1: vpor r2=reg128#15 # asm 2: vpor r2=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm10, % xmm11, % xmm10 # qhasm: v00 = r4 & mask4 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm4, % xmm12, % xmm11 # qhasm: 8x v10 = r5 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm8, % xmm15 # qhasm: 8x v01 = r4 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#13 # asm 2: vpsrlw $8,v01=%xmm12 vpsrlw $8, % xmm12, % xmm12 # qhasm: v11 = r5 & mask5 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm5, % xmm8, % xmm8 # qhasm: r4 = v00 | v10 # asm 1: vpor r4=reg128#12 # asm 2: vpor r4=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm8, % xmm12, % xmm8 # qhasm: v00 = r6 & mask4 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm4, % xmm6, % xmm12 # qhasm: 8x v10 = r7 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm7, % xmm15 # qhasm: 8x v01 = r6 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#7 # asm 2: vpsrlw $8,v01=%xmm6 vpsrlw $8, % xmm6, % xmm6 # qhasm: v11 = r7 & mask5 # asm 1: vpand v11=reg128#8 # asm 2: vpand v11=%xmm7 vpand % xmm5, % xmm7, % xmm7 # qhasm: r6 = v00 | v10 # asm 1: vpor r6=reg128#13 # asm 2: vpor r6=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#7 # asm 2: vpor r7=%xmm6 vpor % xmm7, % xmm6, % xmm6 # qhasm: buf = r0[0] # asm 1: pextrq $0x0,buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm9, % rsi # qhasm: mem64[ input_0 + 8 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm13, % rsi # qhasm: mem64[ input_0 + 72 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm14, % rsi # qhasm: mem64[ input_0 + 136 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm10, % rsi # qhasm: mem64[ input_0 + 200 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm11, % rsi # qhasm: mem64[ input_0 + 264 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm8, % rsi # qhasm: mem64[ input_0 + 328 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm12, % rsi # qhasm: mem64[ input_0 + 392 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm6, % rsi # qhasm: mem64[ input_0 + 456 ] = buf # asm 1: movq r0=reg128#7 # asm 2: movddup 16(r0=%xmm6 movddup 16( % rdi), % xmm6 # qhasm: r1 = mem64[ input_0 + 80 ] x2 # asm 1: movddup 80(r1=reg128#8 # asm 2: movddup 80(r1=%xmm7 movddup 80( % rdi), % xmm7 # qhasm: r2 = mem64[ input_0 + 144 ] x2 # asm 1: movddup 144(r2=reg128#9 # asm 2: movddup 144(r2=%xmm8 movddup 144( % rdi), % xmm8 # qhasm: r3 = mem64[ input_0 + 208 ] x2 # asm 1: movddup 208(r3=reg128#10 # asm 2: movddup 208(r3=%xmm9 movddup 208( % rdi), % xmm9 # qhasm: r4 = mem64[ input_0 + 272 ] x2 # asm 1: movddup 272(r4=reg128#11 # asm 2: movddup 272(r4=%xmm10 movddup 272( % rdi), % xmm10 # qhasm: r5 = mem64[ input_0 + 336 ] x2 # asm 1: movddup 336(r5=reg128#12 # asm 2: movddup 336(r5=%xmm11 movddup 336( % rdi), % xmm11 # qhasm: r6 = mem64[ input_0 + 400 ] x2 # asm 1: movddup 400(r6=reg128#13 # asm 2: movddup 400(r6=%xmm12 movddup 400( % rdi), % xmm12 # qhasm: r7 = mem64[ input_0 + 464 ] x2 # asm 1: movddup 464(r7=reg128#14 # asm 2: movddup 464(r7=%xmm13 movddup 464( % rdi), % xmm13 # qhasm: v00 = r0 & mask0 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm0, % xmm6, % xmm14 # qhasm: 2x v10 = r4 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm10, % xmm15 # qhasm: 2x v01 = r0 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#7 # asm 2: vpsrlq $32,v01=%xmm6 vpsrlq $32, % xmm6, % xmm6 # qhasm: v11 = r4 & mask1 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm1, % xmm10, % xmm10 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#15 # asm 2: vpor r0=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r4 = v01 | v11 # asm 1: vpor r4=reg128#7 # asm 2: vpor r4=%xmm6 vpor % xmm10, % xmm6, % xmm6 # qhasm: v00 = r1 & mask0 # asm 1: vpand v00=reg128#11 # asm 2: vpand v00=%xmm10 vpand % xmm0, % xmm7, % xmm10 # qhasm: 2x v10 = r5 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm11, % xmm15 # qhasm: 2x v01 = r1 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#8 # asm 2: vpsrlq $32,v01=%xmm7 vpsrlq $32, % xmm7, % xmm7 # qhasm: v11 = r5 & mask1 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm1, % xmm11, % xmm11 # qhasm: r1 = v00 | v10 # asm 1: vpor r1=reg128#11 # asm 2: vpor r1=%xmm10 vpor % xmm15, % xmm10, % xmm10 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#8 # asm 2: vpor r5=%xmm7 vpor % xmm11, % xmm7, % xmm7 # qhasm: v00 = r2 & mask0 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm0, % xmm8, % xmm11 # qhasm: 2x v10 = r6 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm12, % xmm15 # qhasm: 2x v01 = r2 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#9 # asm 2: vpsrlq $32,v01=%xmm8 vpsrlq $32, % xmm8, % xmm8 # qhasm: v11 = r6 & mask1 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm1, % xmm12, % xmm12 # qhasm: r2 = v00 | v10 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#9 # asm 2: vpor r6=%xmm8 vpor % xmm12, % xmm8, % xmm8 # qhasm: v00 = r3 & mask0 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm0, % xmm9, % xmm12 # qhasm: 2x v10 = r7 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm13, % xmm15 # qhasm: 2x v01 = r3 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#10 # asm 2: vpsrlq $32,v01=%xmm9 vpsrlq $32, % xmm9, % xmm9 # qhasm: v11 = r7 & mask1 # asm 1: vpand v11=reg128#14 # asm 2: vpand v11=%xmm13 vpand % xmm1, % xmm13, % xmm13 # qhasm: r3 = v00 | v10 # asm 1: vpor r3=reg128#13 # asm 2: vpor r3=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#10 # asm 2: vpor r7=%xmm9 vpor % xmm13, % xmm9, % xmm9 # qhasm: v00 = r0 & mask2 # asm 1: vpand v00=reg128#14 # asm 2: vpand v00=%xmm13 vpand % xmm2, % xmm14, % xmm13 # qhasm: 4x v10 = r2 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm11, % xmm15 # qhasm: 4x v01 = r0 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#15 # asm 2: vpsrld $16,v01=%xmm14 vpsrld $16, % xmm14, % xmm14 # qhasm: v11 = r2 & mask3 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm3, % xmm11, % xmm11 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#14 # asm 2: vpor r0=%xmm13 vpor % xmm15, % xmm13, % xmm13 # qhasm: r2 = v01 | v11 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm11, % xmm14, % xmm11 # qhasm: v00 = r1 & mask2 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm2, % xmm10, % xmm14 # qhasm: 4x v10 = r3 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm12, % xmm15 # qhasm: 4x v01 = r1 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#11 # asm 2: vpsrld $16,v01=%xmm10 vpsrld $16, % xmm10, % xmm10 # qhasm: v11 = r3 & mask3 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm3, % xmm12, % xmm12 # qhasm: r1 = v00 | v10 # asm 1: vpor r1=reg128#15 # asm 2: vpor r1=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm12, % xmm10, % xmm10 # qhasm: v00 = r4 & mask2 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm2, % xmm6, % xmm12 # qhasm: 4x v10 = r6 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm8, % xmm15 # qhasm: 4x v01 = r4 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#7 # asm 2: vpsrld $16,v01=%xmm6 vpsrld $16, % xmm6, % xmm6 # qhasm: v11 = r6 & mask3 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm3, % xmm8, % xmm8 # qhasm: r4 = v00 | v10 # asm 1: vpor r4=reg128#13 # asm 2: vpor r4=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#7 # asm 2: vpor r6=%xmm6 vpor % xmm8, % xmm6, % xmm6 # qhasm: v00 = r5 & mask2 # asm 1: vpand v00=reg128#9 # asm 2: vpand v00=%xmm8 vpand % xmm2, % xmm7, % xmm8 # qhasm: 4x v10 = r7 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm9, % xmm15 # qhasm: 4x v01 = r5 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#8 # asm 2: vpsrld $16,v01=%xmm7 vpsrld $16, % xmm7, % xmm7 # qhasm: v11 = r7 & mask3 # asm 1: vpand v11=reg128#10 # asm 2: vpand v11=%xmm9 vpand % xmm3, % xmm9, % xmm9 # qhasm: r5 = v00 | v10 # asm 1: vpor r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm15, % xmm8, % xmm8 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#8 # asm 2: vpor r7=%xmm7 vpor % xmm9, % xmm7, % xmm7 # qhasm: v00 = r0 & mask4 # asm 1: vpand v00=reg128#10 # asm 2: vpand v00=%xmm9 vpand % xmm4, % xmm13, % xmm9 # qhasm: 8x v10 = r1 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm14, % xmm15 # qhasm: 8x v01 = r0 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#14 # asm 2: vpsrlw $8,v01=%xmm13 vpsrlw $8, % xmm13, % xmm13 # qhasm: v11 = r1 & mask5 # asm 1: vpand v11=reg128#15 # asm 2: vpand v11=%xmm14 vpand % xmm5, % xmm14, % xmm14 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#10 # asm 2: vpor r0=%xmm9 vpor % xmm15, % xmm9, % xmm9 # qhasm: r1 = v01 | v11 # asm 1: vpor r1=reg128#14 # asm 2: vpor r1=%xmm13 vpor % xmm14, % xmm13, % xmm13 # qhasm: v00 = r2 & mask4 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm4, % xmm11, % xmm14 # qhasm: 8x v10 = r3 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm10, % xmm15 # qhasm: 8x v01 = r2 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#12 # asm 2: vpsrlw $8,v01=%xmm11 vpsrlw $8, % xmm11, % xmm11 # qhasm: v11 = r3 & mask5 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm5, % xmm10, % xmm10 # qhasm: r2 = v00 | v10 # asm 1: vpor r2=reg128#15 # asm 2: vpor r2=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm10, % xmm11, % xmm10 # qhasm: v00 = r4 & mask4 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm4, % xmm12, % xmm11 # qhasm: 8x v10 = r5 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm8, % xmm15 # qhasm: 8x v01 = r4 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#13 # asm 2: vpsrlw $8,v01=%xmm12 vpsrlw $8, % xmm12, % xmm12 # qhasm: v11 = r5 & mask5 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm5, % xmm8, % xmm8 # qhasm: r4 = v00 | v10 # asm 1: vpor r4=reg128#12 # asm 2: vpor r4=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm8, % xmm12, % xmm8 # qhasm: v00 = r6 & mask4 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm4, % xmm6, % xmm12 # qhasm: 8x v10 = r7 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm7, % xmm15 # qhasm: 8x v01 = r6 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#7 # asm 2: vpsrlw $8,v01=%xmm6 vpsrlw $8, % xmm6, % xmm6 # qhasm: v11 = r7 & mask5 # asm 1: vpand v11=reg128#8 # asm 2: vpand v11=%xmm7 vpand % xmm5, % xmm7, % xmm7 # qhasm: r6 = v00 | v10 # asm 1: vpor r6=reg128#13 # asm 2: vpor r6=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#7 # asm 2: vpor r7=%xmm6 vpor % xmm7, % xmm6, % xmm6 # qhasm: buf = r0[0] # asm 1: pextrq $0x0,buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm9, % rsi # qhasm: mem64[ input_0 + 16 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm13, % rsi # qhasm: mem64[ input_0 + 80 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm14, % rsi # qhasm: mem64[ input_0 + 144 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm10, % rsi # qhasm: mem64[ input_0 + 208 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm11, % rsi # qhasm: mem64[ input_0 + 272 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm8, % rsi # qhasm: mem64[ input_0 + 336 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm12, % rsi # qhasm: mem64[ input_0 + 400 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm6, % rsi # qhasm: mem64[ input_0 + 464 ] = buf # asm 1: movq r0=reg128#7 # asm 2: movddup 24(r0=%xmm6 movddup 24( % rdi), % xmm6 # qhasm: r1 = mem64[ input_0 + 88 ] x2 # asm 1: movddup 88(r1=reg128#8 # asm 2: movddup 88(r1=%xmm7 movddup 88( % rdi), % xmm7 # qhasm: r2 = mem64[ input_0 + 152 ] x2 # asm 1: movddup 152(r2=reg128#9 # asm 2: movddup 152(r2=%xmm8 movddup 152( % rdi), % xmm8 # qhasm: r3 = mem64[ input_0 + 216 ] x2 # asm 1: movddup 216(r3=reg128#10 # asm 2: movddup 216(r3=%xmm9 movddup 216( % rdi), % xmm9 # qhasm: r4 = mem64[ input_0 + 280 ] x2 # asm 1: movddup 280(r4=reg128#11 # asm 2: movddup 280(r4=%xmm10 movddup 280( % rdi), % xmm10 # qhasm: r5 = mem64[ input_0 + 344 ] x2 # asm 1: movddup 344(r5=reg128#12 # asm 2: movddup 344(r5=%xmm11 movddup 344( % rdi), % xmm11 # qhasm: r6 = mem64[ input_0 + 408 ] x2 # asm 1: movddup 408(r6=reg128#13 # asm 2: movddup 408(r6=%xmm12 movddup 408( % rdi), % xmm12 # qhasm: r7 = mem64[ input_0 + 472 ] x2 # asm 1: movddup 472(r7=reg128#14 # asm 2: movddup 472(r7=%xmm13 movddup 472( % rdi), % xmm13 # qhasm: v00 = r0 & mask0 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm0, % xmm6, % xmm14 # qhasm: 2x v10 = r4 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm10, % xmm15 # qhasm: 2x v01 = r0 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#7 # asm 2: vpsrlq $32,v01=%xmm6 vpsrlq $32, % xmm6, % xmm6 # qhasm: v11 = r4 & mask1 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm1, % xmm10, % xmm10 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#15 # asm 2: vpor r0=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r4 = v01 | v11 # asm 1: vpor r4=reg128#7 # asm 2: vpor r4=%xmm6 vpor % xmm10, % xmm6, % xmm6 # qhasm: v00 = r1 & mask0 # asm 1: vpand v00=reg128#11 # asm 2: vpand v00=%xmm10 vpand % xmm0, % xmm7, % xmm10 # qhasm: 2x v10 = r5 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm11, % xmm15 # qhasm: 2x v01 = r1 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#8 # asm 2: vpsrlq $32,v01=%xmm7 vpsrlq $32, % xmm7, % xmm7 # qhasm: v11 = r5 & mask1 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm1, % xmm11, % xmm11 # qhasm: r1 = v00 | v10 # asm 1: vpor r1=reg128#11 # asm 2: vpor r1=%xmm10 vpor % xmm15, % xmm10, % xmm10 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#8 # asm 2: vpor r5=%xmm7 vpor % xmm11, % xmm7, % xmm7 # qhasm: v00 = r2 & mask0 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm0, % xmm8, % xmm11 # qhasm: 2x v10 = r6 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm12, % xmm15 # qhasm: 2x v01 = r2 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#9 # asm 2: vpsrlq $32,v01=%xmm8 vpsrlq $32, % xmm8, % xmm8 # qhasm: v11 = r6 & mask1 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm1, % xmm12, % xmm12 # qhasm: r2 = v00 | v10 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#9 # asm 2: vpor r6=%xmm8 vpor % xmm12, % xmm8, % xmm8 # qhasm: v00 = r3 & mask0 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm0, % xmm9, % xmm12 # qhasm: 2x v10 = r7 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm13, % xmm15 # qhasm: 2x v01 = r3 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#10 # asm 2: vpsrlq $32,v01=%xmm9 vpsrlq $32, % xmm9, % xmm9 # qhasm: v11 = r7 & mask1 # asm 1: vpand v11=reg128#14 # asm 2: vpand v11=%xmm13 vpand % xmm1, % xmm13, % xmm13 # qhasm: r3 = v00 | v10 # asm 1: vpor r3=reg128#13 # asm 2: vpor r3=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#10 # asm 2: vpor r7=%xmm9 vpor % xmm13, % xmm9, % xmm9 # qhasm: v00 = r0 & mask2 # asm 1: vpand v00=reg128#14 # asm 2: vpand v00=%xmm13 vpand % xmm2, % xmm14, % xmm13 # qhasm: 4x v10 = r2 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm11, % xmm15 # qhasm: 4x v01 = r0 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#15 # asm 2: vpsrld $16,v01=%xmm14 vpsrld $16, % xmm14, % xmm14 # qhasm: v11 = r2 & mask3 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm3, % xmm11, % xmm11 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#14 # asm 2: vpor r0=%xmm13 vpor % xmm15, % xmm13, % xmm13 # qhasm: r2 = v01 | v11 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm11, % xmm14, % xmm11 # qhasm: v00 = r1 & mask2 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm2, % xmm10, % xmm14 # qhasm: 4x v10 = r3 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm12, % xmm15 # qhasm: 4x v01 = r1 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#11 # asm 2: vpsrld $16,v01=%xmm10 vpsrld $16, % xmm10, % xmm10 # qhasm: v11 = r3 & mask3 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm3, % xmm12, % xmm12 # qhasm: r1 = v00 | v10 # asm 1: vpor r1=reg128#15 # asm 2: vpor r1=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm12, % xmm10, % xmm10 # qhasm: v00 = r4 & mask2 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm2, % xmm6, % xmm12 # qhasm: 4x v10 = r6 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm8, % xmm15 # qhasm: 4x v01 = r4 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#7 # asm 2: vpsrld $16,v01=%xmm6 vpsrld $16, % xmm6, % xmm6 # qhasm: v11 = r6 & mask3 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm3, % xmm8, % xmm8 # qhasm: r4 = v00 | v10 # asm 1: vpor r4=reg128#13 # asm 2: vpor r4=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#7 # asm 2: vpor r6=%xmm6 vpor % xmm8, % xmm6, % xmm6 # qhasm: v00 = r5 & mask2 # asm 1: vpand v00=reg128#9 # asm 2: vpand v00=%xmm8 vpand % xmm2, % xmm7, % xmm8 # qhasm: 4x v10 = r7 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm9, % xmm15 # qhasm: 4x v01 = r5 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#8 # asm 2: vpsrld $16,v01=%xmm7 vpsrld $16, % xmm7, % xmm7 # qhasm: v11 = r7 & mask3 # asm 1: vpand v11=reg128#10 # asm 2: vpand v11=%xmm9 vpand % xmm3, % xmm9, % xmm9 # qhasm: r5 = v00 | v10 # asm 1: vpor r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm15, % xmm8, % xmm8 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#8 # asm 2: vpor r7=%xmm7 vpor % xmm9, % xmm7, % xmm7 # qhasm: v00 = r0 & mask4 # asm 1: vpand v00=reg128#10 # asm 2: vpand v00=%xmm9 vpand % xmm4, % xmm13, % xmm9 # qhasm: 8x v10 = r1 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm14, % xmm15 # qhasm: 8x v01 = r0 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#14 # asm 2: vpsrlw $8,v01=%xmm13 vpsrlw $8, % xmm13, % xmm13 # qhasm: v11 = r1 & mask5 # asm 1: vpand v11=reg128#15 # asm 2: vpand v11=%xmm14 vpand % xmm5, % xmm14, % xmm14 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#10 # asm 2: vpor r0=%xmm9 vpor % xmm15, % xmm9, % xmm9 # qhasm: r1 = v01 | v11 # asm 1: vpor r1=reg128#14 # asm 2: vpor r1=%xmm13 vpor % xmm14, % xmm13, % xmm13 # qhasm: v00 = r2 & mask4 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm4, % xmm11, % xmm14 # qhasm: 8x v10 = r3 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm10, % xmm15 # qhasm: 8x v01 = r2 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#12 # asm 2: vpsrlw $8,v01=%xmm11 vpsrlw $8, % xmm11, % xmm11 # qhasm: v11 = r3 & mask5 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm5, % xmm10, % xmm10 # qhasm: r2 = v00 | v10 # asm 1: vpor r2=reg128#15 # asm 2: vpor r2=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm10, % xmm11, % xmm10 # qhasm: v00 = r4 & mask4 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm4, % xmm12, % xmm11 # qhasm: 8x v10 = r5 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm8, % xmm15 # qhasm: 8x v01 = r4 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#13 # asm 2: vpsrlw $8,v01=%xmm12 vpsrlw $8, % xmm12, % xmm12 # qhasm: v11 = r5 & mask5 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm5, % xmm8, % xmm8 # qhasm: r4 = v00 | v10 # asm 1: vpor r4=reg128#12 # asm 2: vpor r4=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm8, % xmm12, % xmm8 # qhasm: v00 = r6 & mask4 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm4, % xmm6, % xmm12 # qhasm: 8x v10 = r7 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm7, % xmm15 # qhasm: 8x v01 = r6 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#7 # asm 2: vpsrlw $8,v01=%xmm6 vpsrlw $8, % xmm6, % xmm6 # qhasm: v11 = r7 & mask5 # asm 1: vpand v11=reg128#8 # asm 2: vpand v11=%xmm7 vpand % xmm5, % xmm7, % xmm7 # qhasm: r6 = v00 | v10 # asm 1: vpor r6=reg128#13 # asm 2: vpor r6=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#7 # asm 2: vpor r7=%xmm6 vpor % xmm7, % xmm6, % xmm6 # qhasm: buf = r0[0] # asm 1: pextrq $0x0,buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm9, % rsi # qhasm: mem64[ input_0 + 24 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm13, % rsi # qhasm: mem64[ input_0 + 88 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm14, % rsi # qhasm: mem64[ input_0 + 152 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm10, % rsi # qhasm: mem64[ input_0 + 216 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm11, % rsi # qhasm: mem64[ input_0 + 280 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm8, % rsi # qhasm: mem64[ input_0 + 344 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm12, % rsi # qhasm: mem64[ input_0 + 408 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm6, % rsi # qhasm: mem64[ input_0 + 472 ] = buf # asm 1: movq r0=reg128#7 # asm 2: movddup 32(r0=%xmm6 movddup 32( % rdi), % xmm6 # qhasm: r1 = mem64[ input_0 + 96 ] x2 # asm 1: movddup 96(r1=reg128#8 # asm 2: movddup 96(r1=%xmm7 movddup 96( % rdi), % xmm7 # qhasm: r2 = mem64[ input_0 + 160 ] x2 # asm 1: movddup 160(r2=reg128#9 # asm 2: movddup 160(r2=%xmm8 movddup 160( % rdi), % xmm8 # qhasm: r3 = mem64[ input_0 + 224 ] x2 # asm 1: movddup 224(r3=reg128#10 # asm 2: movddup 224(r3=%xmm9 movddup 224( % rdi), % xmm9 # qhasm: r4 = mem64[ input_0 + 288 ] x2 # asm 1: movddup 288(r4=reg128#11 # asm 2: movddup 288(r4=%xmm10 movddup 288( % rdi), % xmm10 # qhasm: r5 = mem64[ input_0 + 352 ] x2 # asm 1: movddup 352(r5=reg128#12 # asm 2: movddup 352(r5=%xmm11 movddup 352( % rdi), % xmm11 # qhasm: r6 = mem64[ input_0 + 416 ] x2 # asm 1: movddup 416(r6=reg128#13 # asm 2: movddup 416(r6=%xmm12 movddup 416( % rdi), % xmm12 # qhasm: r7 = mem64[ input_0 + 480 ] x2 # asm 1: movddup 480(r7=reg128#14 # asm 2: movddup 480(r7=%xmm13 movddup 480( % rdi), % xmm13 # qhasm: v00 = r0 & mask0 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm0, % xmm6, % xmm14 # qhasm: 2x v10 = r4 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm10, % xmm15 # qhasm: 2x v01 = r0 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#7 # asm 2: vpsrlq $32,v01=%xmm6 vpsrlq $32, % xmm6, % xmm6 # qhasm: v11 = r4 & mask1 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm1, % xmm10, % xmm10 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#15 # asm 2: vpor r0=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r4 = v01 | v11 # asm 1: vpor r4=reg128#7 # asm 2: vpor r4=%xmm6 vpor % xmm10, % xmm6, % xmm6 # qhasm: v00 = r1 & mask0 # asm 1: vpand v00=reg128#11 # asm 2: vpand v00=%xmm10 vpand % xmm0, % xmm7, % xmm10 # qhasm: 2x v10 = r5 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm11, % xmm15 # qhasm: 2x v01 = r1 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#8 # asm 2: vpsrlq $32,v01=%xmm7 vpsrlq $32, % xmm7, % xmm7 # qhasm: v11 = r5 & mask1 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm1, % xmm11, % xmm11 # qhasm: r1 = v00 | v10 # asm 1: vpor r1=reg128#11 # asm 2: vpor r1=%xmm10 vpor % xmm15, % xmm10, % xmm10 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#8 # asm 2: vpor r5=%xmm7 vpor % xmm11, % xmm7, % xmm7 # qhasm: v00 = r2 & mask0 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm0, % xmm8, % xmm11 # qhasm: 2x v10 = r6 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm12, % xmm15 # qhasm: 2x v01 = r2 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#9 # asm 2: vpsrlq $32,v01=%xmm8 vpsrlq $32, % xmm8, % xmm8 # qhasm: v11 = r6 & mask1 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm1, % xmm12, % xmm12 # qhasm: r2 = v00 | v10 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#9 # asm 2: vpor r6=%xmm8 vpor % xmm12, % xmm8, % xmm8 # qhasm: v00 = r3 & mask0 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm0, % xmm9, % xmm12 # qhasm: 2x v10 = r7 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm13, % xmm15 # qhasm: 2x v01 = r3 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#10 # asm 2: vpsrlq $32,v01=%xmm9 vpsrlq $32, % xmm9, % xmm9 # qhasm: v11 = r7 & mask1 # asm 1: vpand v11=reg128#14 # asm 2: vpand v11=%xmm13 vpand % xmm1, % xmm13, % xmm13 # qhasm: r3 = v00 | v10 # asm 1: vpor r3=reg128#13 # asm 2: vpor r3=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#10 # asm 2: vpor r7=%xmm9 vpor % xmm13, % xmm9, % xmm9 # qhasm: v00 = r0 & mask2 # asm 1: vpand v00=reg128#14 # asm 2: vpand v00=%xmm13 vpand % xmm2, % xmm14, % xmm13 # qhasm: 4x v10 = r2 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm11, % xmm15 # qhasm: 4x v01 = r0 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#15 # asm 2: vpsrld $16,v01=%xmm14 vpsrld $16, % xmm14, % xmm14 # qhasm: v11 = r2 & mask3 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm3, % xmm11, % xmm11 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#14 # asm 2: vpor r0=%xmm13 vpor % xmm15, % xmm13, % xmm13 # qhasm: r2 = v01 | v11 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm11, % xmm14, % xmm11 # qhasm: v00 = r1 & mask2 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm2, % xmm10, % xmm14 # qhasm: 4x v10 = r3 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm12, % xmm15 # qhasm: 4x v01 = r1 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#11 # asm 2: vpsrld $16,v01=%xmm10 vpsrld $16, % xmm10, % xmm10 # qhasm: v11 = r3 & mask3 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm3, % xmm12, % xmm12 # qhasm: r1 = v00 | v10 # asm 1: vpor r1=reg128#15 # asm 2: vpor r1=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm12, % xmm10, % xmm10 # qhasm: v00 = r4 & mask2 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm2, % xmm6, % xmm12 # qhasm: 4x v10 = r6 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm8, % xmm15 # qhasm: 4x v01 = r4 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#7 # asm 2: vpsrld $16,v01=%xmm6 vpsrld $16, % xmm6, % xmm6 # qhasm: v11 = r6 & mask3 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm3, % xmm8, % xmm8 # qhasm: r4 = v00 | v10 # asm 1: vpor r4=reg128#13 # asm 2: vpor r4=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#7 # asm 2: vpor r6=%xmm6 vpor % xmm8, % xmm6, % xmm6 # qhasm: v00 = r5 & mask2 # asm 1: vpand v00=reg128#9 # asm 2: vpand v00=%xmm8 vpand % xmm2, % xmm7, % xmm8 # qhasm: 4x v10 = r7 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm9, % xmm15 # qhasm: 4x v01 = r5 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#8 # asm 2: vpsrld $16,v01=%xmm7 vpsrld $16, % xmm7, % xmm7 # qhasm: v11 = r7 & mask3 # asm 1: vpand v11=reg128#10 # asm 2: vpand v11=%xmm9 vpand % xmm3, % xmm9, % xmm9 # qhasm: r5 = v00 | v10 # asm 1: vpor r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm15, % xmm8, % xmm8 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#8 # asm 2: vpor r7=%xmm7 vpor % xmm9, % xmm7, % xmm7 # qhasm: v00 = r0 & mask4 # asm 1: vpand v00=reg128#10 # asm 2: vpand v00=%xmm9 vpand % xmm4, % xmm13, % xmm9 # qhasm: 8x v10 = r1 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm14, % xmm15 # qhasm: 8x v01 = r0 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#14 # asm 2: vpsrlw $8,v01=%xmm13 vpsrlw $8, % xmm13, % xmm13 # qhasm: v11 = r1 & mask5 # asm 1: vpand v11=reg128#15 # asm 2: vpand v11=%xmm14 vpand % xmm5, % xmm14, % xmm14 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#10 # asm 2: vpor r0=%xmm9 vpor % xmm15, % xmm9, % xmm9 # qhasm: r1 = v01 | v11 # asm 1: vpor r1=reg128#14 # asm 2: vpor r1=%xmm13 vpor % xmm14, % xmm13, % xmm13 # qhasm: v00 = r2 & mask4 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm4, % xmm11, % xmm14 # qhasm: 8x v10 = r3 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm10, % xmm15 # qhasm: 8x v01 = r2 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#12 # asm 2: vpsrlw $8,v01=%xmm11 vpsrlw $8, % xmm11, % xmm11 # qhasm: v11 = r3 & mask5 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm5, % xmm10, % xmm10 # qhasm: r2 = v00 | v10 # asm 1: vpor r2=reg128#15 # asm 2: vpor r2=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm10, % xmm11, % xmm10 # qhasm: v00 = r4 & mask4 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm4, % xmm12, % xmm11 # qhasm: 8x v10 = r5 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm8, % xmm15 # qhasm: 8x v01 = r4 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#13 # asm 2: vpsrlw $8,v01=%xmm12 vpsrlw $8, % xmm12, % xmm12 # qhasm: v11 = r5 & mask5 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm5, % xmm8, % xmm8 # qhasm: r4 = v00 | v10 # asm 1: vpor r4=reg128#12 # asm 2: vpor r4=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm8, % xmm12, % xmm8 # qhasm: v00 = r6 & mask4 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm4, % xmm6, % xmm12 # qhasm: 8x v10 = r7 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm7, % xmm15 # qhasm: 8x v01 = r6 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#7 # asm 2: vpsrlw $8,v01=%xmm6 vpsrlw $8, % xmm6, % xmm6 # qhasm: v11 = r7 & mask5 # asm 1: vpand v11=reg128#8 # asm 2: vpand v11=%xmm7 vpand % xmm5, % xmm7, % xmm7 # qhasm: r6 = v00 | v10 # asm 1: vpor r6=reg128#13 # asm 2: vpor r6=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#7 # asm 2: vpor r7=%xmm6 vpor % xmm7, % xmm6, % xmm6 # qhasm: buf = r0[0] # asm 1: pextrq $0x0,buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm9, % rsi # qhasm: mem64[ input_0 + 32 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm13, % rsi # qhasm: mem64[ input_0 + 96 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm14, % rsi # qhasm: mem64[ input_0 + 160 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm10, % rsi # qhasm: mem64[ input_0 + 224 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm11, % rsi # qhasm: mem64[ input_0 + 288 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm8, % rsi # qhasm: mem64[ input_0 + 352 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm12, % rsi # qhasm: mem64[ input_0 + 416 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm6, % rsi # qhasm: mem64[ input_0 + 480 ] = buf # asm 1: movq r0=reg128#7 # asm 2: movddup 40(r0=%xmm6 movddup 40( % rdi), % xmm6 # qhasm: r1 = mem64[ input_0 + 104 ] x2 # asm 1: movddup 104(r1=reg128#8 # asm 2: movddup 104(r1=%xmm7 movddup 104( % rdi), % xmm7 # qhasm: r2 = mem64[ input_0 + 168 ] x2 # asm 1: movddup 168(r2=reg128#9 # asm 2: movddup 168(r2=%xmm8 movddup 168( % rdi), % xmm8 # qhasm: r3 = mem64[ input_0 + 232 ] x2 # asm 1: movddup 232(r3=reg128#10 # asm 2: movddup 232(r3=%xmm9 movddup 232( % rdi), % xmm9 # qhasm: r4 = mem64[ input_0 + 296 ] x2 # asm 1: movddup 296(r4=reg128#11 # asm 2: movddup 296(r4=%xmm10 movddup 296( % rdi), % xmm10 # qhasm: r5 = mem64[ input_0 + 360 ] x2 # asm 1: movddup 360(r5=reg128#12 # asm 2: movddup 360(r5=%xmm11 movddup 360( % rdi), % xmm11 # qhasm: r6 = mem64[ input_0 + 424 ] x2 # asm 1: movddup 424(r6=reg128#13 # asm 2: movddup 424(r6=%xmm12 movddup 424( % rdi), % xmm12 # qhasm: r7 = mem64[ input_0 + 488 ] x2 # asm 1: movddup 488(r7=reg128#14 # asm 2: movddup 488(r7=%xmm13 movddup 488( % rdi), % xmm13 # qhasm: v00 = r0 & mask0 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm0, % xmm6, % xmm14 # qhasm: 2x v10 = r4 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm10, % xmm15 # qhasm: 2x v01 = r0 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#7 # asm 2: vpsrlq $32,v01=%xmm6 vpsrlq $32, % xmm6, % xmm6 # qhasm: v11 = r4 & mask1 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm1, % xmm10, % xmm10 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#15 # asm 2: vpor r0=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r4 = v01 | v11 # asm 1: vpor r4=reg128#7 # asm 2: vpor r4=%xmm6 vpor % xmm10, % xmm6, % xmm6 # qhasm: v00 = r1 & mask0 # asm 1: vpand v00=reg128#11 # asm 2: vpand v00=%xmm10 vpand % xmm0, % xmm7, % xmm10 # qhasm: 2x v10 = r5 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm11, % xmm15 # qhasm: 2x v01 = r1 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#8 # asm 2: vpsrlq $32,v01=%xmm7 vpsrlq $32, % xmm7, % xmm7 # qhasm: v11 = r5 & mask1 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm1, % xmm11, % xmm11 # qhasm: r1 = v00 | v10 # asm 1: vpor r1=reg128#11 # asm 2: vpor r1=%xmm10 vpor % xmm15, % xmm10, % xmm10 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#8 # asm 2: vpor r5=%xmm7 vpor % xmm11, % xmm7, % xmm7 # qhasm: v00 = r2 & mask0 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm0, % xmm8, % xmm11 # qhasm: 2x v10 = r6 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm12, % xmm15 # qhasm: 2x v01 = r2 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#9 # asm 2: vpsrlq $32,v01=%xmm8 vpsrlq $32, % xmm8, % xmm8 # qhasm: v11 = r6 & mask1 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm1, % xmm12, % xmm12 # qhasm: r2 = v00 | v10 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#9 # asm 2: vpor r6=%xmm8 vpor % xmm12, % xmm8, % xmm8 # qhasm: v00 = r3 & mask0 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm0, % xmm9, % xmm12 # qhasm: 2x v10 = r7 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm13, % xmm15 # qhasm: 2x v01 = r3 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#10 # asm 2: vpsrlq $32,v01=%xmm9 vpsrlq $32, % xmm9, % xmm9 # qhasm: v11 = r7 & mask1 # asm 1: vpand v11=reg128#14 # asm 2: vpand v11=%xmm13 vpand % xmm1, % xmm13, % xmm13 # qhasm: r3 = v00 | v10 # asm 1: vpor r3=reg128#13 # asm 2: vpor r3=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#10 # asm 2: vpor r7=%xmm9 vpor % xmm13, % xmm9, % xmm9 # qhasm: v00 = r0 & mask2 # asm 1: vpand v00=reg128#14 # asm 2: vpand v00=%xmm13 vpand % xmm2, % xmm14, % xmm13 # qhasm: 4x v10 = r2 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm11, % xmm15 # qhasm: 4x v01 = r0 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#15 # asm 2: vpsrld $16,v01=%xmm14 vpsrld $16, % xmm14, % xmm14 # qhasm: v11 = r2 & mask3 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm3, % xmm11, % xmm11 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#14 # asm 2: vpor r0=%xmm13 vpor % xmm15, % xmm13, % xmm13 # qhasm: r2 = v01 | v11 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm11, % xmm14, % xmm11 # qhasm: v00 = r1 & mask2 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm2, % xmm10, % xmm14 # qhasm: 4x v10 = r3 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm12, % xmm15 # qhasm: 4x v01 = r1 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#11 # asm 2: vpsrld $16,v01=%xmm10 vpsrld $16, % xmm10, % xmm10 # qhasm: v11 = r3 & mask3 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm3, % xmm12, % xmm12 # qhasm: r1 = v00 | v10 # asm 1: vpor r1=reg128#15 # asm 2: vpor r1=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm12, % xmm10, % xmm10 # qhasm: v00 = r4 & mask2 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm2, % xmm6, % xmm12 # qhasm: 4x v10 = r6 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm8, % xmm15 # qhasm: 4x v01 = r4 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#7 # asm 2: vpsrld $16,v01=%xmm6 vpsrld $16, % xmm6, % xmm6 # qhasm: v11 = r6 & mask3 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm3, % xmm8, % xmm8 # qhasm: r4 = v00 | v10 # asm 1: vpor r4=reg128#13 # asm 2: vpor r4=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#7 # asm 2: vpor r6=%xmm6 vpor % xmm8, % xmm6, % xmm6 # qhasm: v00 = r5 & mask2 # asm 1: vpand v00=reg128#9 # asm 2: vpand v00=%xmm8 vpand % xmm2, % xmm7, % xmm8 # qhasm: 4x v10 = r7 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm9, % xmm15 # qhasm: 4x v01 = r5 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#8 # asm 2: vpsrld $16,v01=%xmm7 vpsrld $16, % xmm7, % xmm7 # qhasm: v11 = r7 & mask3 # asm 1: vpand v11=reg128#10 # asm 2: vpand v11=%xmm9 vpand % xmm3, % xmm9, % xmm9 # qhasm: r5 = v00 | v10 # asm 1: vpor r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm15, % xmm8, % xmm8 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#8 # asm 2: vpor r7=%xmm7 vpor % xmm9, % xmm7, % xmm7 # qhasm: v00 = r0 & mask4 # asm 1: vpand v00=reg128#10 # asm 2: vpand v00=%xmm9 vpand % xmm4, % xmm13, % xmm9 # qhasm: 8x v10 = r1 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm14, % xmm15 # qhasm: 8x v01 = r0 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#14 # asm 2: vpsrlw $8,v01=%xmm13 vpsrlw $8, % xmm13, % xmm13 # qhasm: v11 = r1 & mask5 # asm 1: vpand v11=reg128#15 # asm 2: vpand v11=%xmm14 vpand % xmm5, % xmm14, % xmm14 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#10 # asm 2: vpor r0=%xmm9 vpor % xmm15, % xmm9, % xmm9 # qhasm: r1 = v01 | v11 # asm 1: vpor r1=reg128#14 # asm 2: vpor r1=%xmm13 vpor % xmm14, % xmm13, % xmm13 # qhasm: v00 = r2 & mask4 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm4, % xmm11, % xmm14 # qhasm: 8x v10 = r3 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm10, % xmm15 # qhasm: 8x v01 = r2 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#12 # asm 2: vpsrlw $8,v01=%xmm11 vpsrlw $8, % xmm11, % xmm11 # qhasm: v11 = r3 & mask5 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm5, % xmm10, % xmm10 # qhasm: r2 = v00 | v10 # asm 1: vpor r2=reg128#15 # asm 2: vpor r2=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm10, % xmm11, % xmm10 # qhasm: v00 = r4 & mask4 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm4, % xmm12, % xmm11 # qhasm: 8x v10 = r5 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm8, % xmm15 # qhasm: 8x v01 = r4 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#13 # asm 2: vpsrlw $8,v01=%xmm12 vpsrlw $8, % xmm12, % xmm12 # qhasm: v11 = r5 & mask5 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm5, % xmm8, % xmm8 # qhasm: r4 = v00 | v10 # asm 1: vpor r4=reg128#12 # asm 2: vpor r4=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm8, % xmm12, % xmm8 # qhasm: v00 = r6 & mask4 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm4, % xmm6, % xmm12 # qhasm: 8x v10 = r7 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm7, % xmm15 # qhasm: 8x v01 = r6 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#7 # asm 2: vpsrlw $8,v01=%xmm6 vpsrlw $8, % xmm6, % xmm6 # qhasm: v11 = r7 & mask5 # asm 1: vpand v11=reg128#8 # asm 2: vpand v11=%xmm7 vpand % xmm5, % xmm7, % xmm7 # qhasm: r6 = v00 | v10 # asm 1: vpor r6=reg128#13 # asm 2: vpor r6=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#7 # asm 2: vpor r7=%xmm6 vpor % xmm7, % xmm6, % xmm6 # qhasm: buf = r0[0] # asm 1: pextrq $0x0,buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm9, % rsi # qhasm: mem64[ input_0 + 40 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm13, % rsi # qhasm: mem64[ input_0 + 104 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm14, % rsi # qhasm: mem64[ input_0 + 168 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm10, % rsi # qhasm: mem64[ input_0 + 232 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm11, % rsi # qhasm: mem64[ input_0 + 296 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm8, % rsi # qhasm: mem64[ input_0 + 360 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm12, % rsi # qhasm: mem64[ input_0 + 424 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm6, % rsi # qhasm: mem64[ input_0 + 488 ] = buf # asm 1: movq r0=reg128#7 # asm 2: movddup 48(r0=%xmm6 movddup 48( % rdi), % xmm6 # qhasm: r1 = mem64[ input_0 + 112 ] x2 # asm 1: movddup 112(r1=reg128#8 # asm 2: movddup 112(r1=%xmm7 movddup 112( % rdi), % xmm7 # qhasm: r2 = mem64[ input_0 + 176 ] x2 # asm 1: movddup 176(r2=reg128#9 # asm 2: movddup 176(r2=%xmm8 movddup 176( % rdi), % xmm8 # qhasm: r3 = mem64[ input_0 + 240 ] x2 # asm 1: movddup 240(r3=reg128#10 # asm 2: movddup 240(r3=%xmm9 movddup 240( % rdi), % xmm9 # qhasm: r4 = mem64[ input_0 + 304 ] x2 # asm 1: movddup 304(r4=reg128#11 # asm 2: movddup 304(r4=%xmm10 movddup 304( % rdi), % xmm10 # qhasm: r5 = mem64[ input_0 + 368 ] x2 # asm 1: movddup 368(r5=reg128#12 # asm 2: movddup 368(r5=%xmm11 movddup 368( % rdi), % xmm11 # qhasm: r6 = mem64[ input_0 + 432 ] x2 # asm 1: movddup 432(r6=reg128#13 # asm 2: movddup 432(r6=%xmm12 movddup 432( % rdi), % xmm12 # qhasm: r7 = mem64[ input_0 + 496 ] x2 # asm 1: movddup 496(r7=reg128#14 # asm 2: movddup 496(r7=%xmm13 movddup 496( % rdi), % xmm13 # qhasm: v00 = r0 & mask0 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm0, % xmm6, % xmm14 # qhasm: 2x v10 = r4 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm10, % xmm15 # qhasm: 2x v01 = r0 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#7 # asm 2: vpsrlq $32,v01=%xmm6 vpsrlq $32, % xmm6, % xmm6 # qhasm: v11 = r4 & mask1 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm1, % xmm10, % xmm10 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#15 # asm 2: vpor r0=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r4 = v01 | v11 # asm 1: vpor r4=reg128#7 # asm 2: vpor r4=%xmm6 vpor % xmm10, % xmm6, % xmm6 # qhasm: v00 = r1 & mask0 # asm 1: vpand v00=reg128#11 # asm 2: vpand v00=%xmm10 vpand % xmm0, % xmm7, % xmm10 # qhasm: 2x v10 = r5 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm11, % xmm15 # qhasm: 2x v01 = r1 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#8 # asm 2: vpsrlq $32,v01=%xmm7 vpsrlq $32, % xmm7, % xmm7 # qhasm: v11 = r5 & mask1 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm1, % xmm11, % xmm11 # qhasm: r1 = v00 | v10 # asm 1: vpor r1=reg128#11 # asm 2: vpor r1=%xmm10 vpor % xmm15, % xmm10, % xmm10 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#8 # asm 2: vpor r5=%xmm7 vpor % xmm11, % xmm7, % xmm7 # qhasm: v00 = r2 & mask0 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm0, % xmm8, % xmm11 # qhasm: 2x v10 = r6 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm12, % xmm15 # qhasm: 2x v01 = r2 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#9 # asm 2: vpsrlq $32,v01=%xmm8 vpsrlq $32, % xmm8, % xmm8 # qhasm: v11 = r6 & mask1 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm1, % xmm12, % xmm12 # qhasm: r2 = v00 | v10 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#9 # asm 2: vpor r6=%xmm8 vpor % xmm12, % xmm8, % xmm8 # qhasm: v00 = r3 & mask0 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm0, % xmm9, % xmm12 # qhasm: 2x v10 = r7 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm13, % xmm15 # qhasm: 2x v01 = r3 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#10 # asm 2: vpsrlq $32,v01=%xmm9 vpsrlq $32, % xmm9, % xmm9 # qhasm: v11 = r7 & mask1 # asm 1: vpand v11=reg128#14 # asm 2: vpand v11=%xmm13 vpand % xmm1, % xmm13, % xmm13 # qhasm: r3 = v00 | v10 # asm 1: vpor r3=reg128#13 # asm 2: vpor r3=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#10 # asm 2: vpor r7=%xmm9 vpor % xmm13, % xmm9, % xmm9 # qhasm: v00 = r0 & mask2 # asm 1: vpand v00=reg128#14 # asm 2: vpand v00=%xmm13 vpand % xmm2, % xmm14, % xmm13 # qhasm: 4x v10 = r2 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm11, % xmm15 # qhasm: 4x v01 = r0 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#15 # asm 2: vpsrld $16,v01=%xmm14 vpsrld $16, % xmm14, % xmm14 # qhasm: v11 = r2 & mask3 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm3, % xmm11, % xmm11 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#14 # asm 2: vpor r0=%xmm13 vpor % xmm15, % xmm13, % xmm13 # qhasm: r2 = v01 | v11 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm11, % xmm14, % xmm11 # qhasm: v00 = r1 & mask2 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm2, % xmm10, % xmm14 # qhasm: 4x v10 = r3 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm12, % xmm15 # qhasm: 4x v01 = r1 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#11 # asm 2: vpsrld $16,v01=%xmm10 vpsrld $16, % xmm10, % xmm10 # qhasm: v11 = r3 & mask3 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm3, % xmm12, % xmm12 # qhasm: r1 = v00 | v10 # asm 1: vpor r1=reg128#15 # asm 2: vpor r1=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm12, % xmm10, % xmm10 # qhasm: v00 = r4 & mask2 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm2, % xmm6, % xmm12 # qhasm: 4x v10 = r6 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm8, % xmm15 # qhasm: 4x v01 = r4 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#7 # asm 2: vpsrld $16,v01=%xmm6 vpsrld $16, % xmm6, % xmm6 # qhasm: v11 = r6 & mask3 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm3, % xmm8, % xmm8 # qhasm: r4 = v00 | v10 # asm 1: vpor r4=reg128#13 # asm 2: vpor r4=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#7 # asm 2: vpor r6=%xmm6 vpor % xmm8, % xmm6, % xmm6 # qhasm: v00 = r5 & mask2 # asm 1: vpand v00=reg128#9 # asm 2: vpand v00=%xmm8 vpand % xmm2, % xmm7, % xmm8 # qhasm: 4x v10 = r7 << 16 # asm 1: vpslld $16,v10=reg128#16 # asm 2: vpslld $16,v10=%xmm15 vpslld $16, % xmm9, % xmm15 # qhasm: 4x v01 = r5 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#8 # asm 2: vpsrld $16,v01=%xmm7 vpsrld $16, % xmm7, % xmm7 # qhasm: v11 = r7 & mask3 # asm 1: vpand v11=reg128#10 # asm 2: vpand v11=%xmm9 vpand % xmm3, % xmm9, % xmm9 # qhasm: r5 = v00 | v10 # asm 1: vpor r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm15, % xmm8, % xmm8 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#8 # asm 2: vpor r7=%xmm7 vpor % xmm9, % xmm7, % xmm7 # qhasm: v00 = r0 & mask4 # asm 1: vpand v00=reg128#10 # asm 2: vpand v00=%xmm9 vpand % xmm4, % xmm13, % xmm9 # qhasm: 8x v10 = r1 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm14, % xmm15 # qhasm: 8x v01 = r0 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#14 # asm 2: vpsrlw $8,v01=%xmm13 vpsrlw $8, % xmm13, % xmm13 # qhasm: v11 = r1 & mask5 # asm 1: vpand v11=reg128#15 # asm 2: vpand v11=%xmm14 vpand % xmm5, % xmm14, % xmm14 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#10 # asm 2: vpor r0=%xmm9 vpor % xmm15, % xmm9, % xmm9 # qhasm: r1 = v01 | v11 # asm 1: vpor r1=reg128#14 # asm 2: vpor r1=%xmm13 vpor % xmm14, % xmm13, % xmm13 # qhasm: v00 = r2 & mask4 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm4, % xmm11, % xmm14 # qhasm: 8x v10 = r3 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm10, % xmm15 # qhasm: 8x v01 = r2 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#12 # asm 2: vpsrlw $8,v01=%xmm11 vpsrlw $8, % xmm11, % xmm11 # qhasm: v11 = r3 & mask5 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm5, % xmm10, % xmm10 # qhasm: r2 = v00 | v10 # asm 1: vpor r2=reg128#15 # asm 2: vpor r2=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm10, % xmm11, % xmm10 # qhasm: v00 = r4 & mask4 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm4, % xmm12, % xmm11 # qhasm: 8x v10 = r5 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm8, % xmm15 # qhasm: 8x v01 = r4 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#13 # asm 2: vpsrlw $8,v01=%xmm12 vpsrlw $8, % xmm12, % xmm12 # qhasm: v11 = r5 & mask5 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm5, % xmm8, % xmm8 # qhasm: r4 = v00 | v10 # asm 1: vpor r4=reg128#12 # asm 2: vpor r4=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm8, % xmm12, % xmm8 # qhasm: v00 = r6 & mask4 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm4, % xmm6, % xmm12 # qhasm: 8x v10 = r7 << 8 # asm 1: vpsllw $8,v10=reg128#16 # asm 2: vpsllw $8,v10=%xmm15 vpsllw $8, % xmm7, % xmm15 # qhasm: 8x v01 = r6 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#7 # asm 2: vpsrlw $8,v01=%xmm6 vpsrlw $8, % xmm6, % xmm6 # qhasm: v11 = r7 & mask5 # asm 1: vpand v11=reg128#8 # asm 2: vpand v11=%xmm7 vpand % xmm5, % xmm7, % xmm7 # qhasm: r6 = v00 | v10 # asm 1: vpor r6=reg128#13 # asm 2: vpor r6=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#7 # asm 2: vpor r7=%xmm6 vpor % xmm7, % xmm6, % xmm6 # qhasm: buf = r0[0] # asm 1: pextrq $0x0,buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm9, % rsi # qhasm: mem64[ input_0 + 48 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm13, % rsi # qhasm: mem64[ input_0 + 112 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm14, % rsi # qhasm: mem64[ input_0 + 176 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm10, % rsi # qhasm: mem64[ input_0 + 240 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm11, % rsi # qhasm: mem64[ input_0 + 304 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm8, % rsi # qhasm: mem64[ input_0 + 368 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm12, % rsi # qhasm: mem64[ input_0 + 432 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm6, % rsi # qhasm: mem64[ input_0 + 496 ] = buf # asm 1: movq r0=reg128#7 # asm 2: movddup 56(r0=%xmm6 movddup 56( % rdi), % xmm6 # qhasm: r1 = mem64[ input_0 + 120 ] x2 # asm 1: movddup 120(r1=reg128#8 # asm 2: movddup 120(r1=%xmm7 movddup 120( % rdi), % xmm7 # qhasm: r2 = mem64[ input_0 + 184 ] x2 # asm 1: movddup 184(r2=reg128#9 # asm 2: movddup 184(r2=%xmm8 movddup 184( % rdi), % xmm8 # qhasm: r3 = mem64[ input_0 + 248 ] x2 # asm 1: movddup 248(r3=reg128#10 # asm 2: movddup 248(r3=%xmm9 movddup 248( % rdi), % xmm9 # qhasm: r4 = mem64[ input_0 + 312 ] x2 # asm 1: movddup 312(r4=reg128#11 # asm 2: movddup 312(r4=%xmm10 movddup 312( % rdi), % xmm10 # qhasm: r5 = mem64[ input_0 + 376 ] x2 # asm 1: movddup 376(r5=reg128#12 # asm 2: movddup 376(r5=%xmm11 movddup 376( % rdi), % xmm11 # qhasm: r6 = mem64[ input_0 + 440 ] x2 # asm 1: movddup 440(r6=reg128#13 # asm 2: movddup 440(r6=%xmm12 movddup 440( % rdi), % xmm12 # qhasm: r7 = mem64[ input_0 + 504 ] x2 # asm 1: movddup 504(r7=reg128#14 # asm 2: movddup 504(r7=%xmm13 movddup 504( % rdi), % xmm13 # qhasm: v00 = r0 & mask0 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm0, % xmm6, % xmm14 # qhasm: 2x v10 = r4 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm10, % xmm15 # qhasm: 2x v01 = r0 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#7 # asm 2: vpsrlq $32,v01=%xmm6 vpsrlq $32, % xmm6, % xmm6 # qhasm: v11 = r4 & mask1 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm1, % xmm10, % xmm10 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#15 # asm 2: vpor r0=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r4 = v01 | v11 # asm 1: vpor r4=reg128#7 # asm 2: vpor r4=%xmm6 vpor % xmm10, % xmm6, % xmm6 # qhasm: v00 = r1 & mask0 # asm 1: vpand v00=reg128#11 # asm 2: vpand v00=%xmm10 vpand % xmm0, % xmm7, % xmm10 # qhasm: 2x v10 = r5 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm11, % xmm15 # qhasm: 2x v01 = r1 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#8 # asm 2: vpsrlq $32,v01=%xmm7 vpsrlq $32, % xmm7, % xmm7 # qhasm: v11 = r5 & mask1 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm1, % xmm11, % xmm11 # qhasm: r1 = v00 | v10 # asm 1: vpor r1=reg128#11 # asm 2: vpor r1=%xmm10 vpor % xmm15, % xmm10, % xmm10 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#8 # asm 2: vpor r5=%xmm7 vpor % xmm11, % xmm7, % xmm7 # qhasm: v00 = r2 & mask0 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm0, % xmm8, % xmm11 # qhasm: 2x v10 = r6 << 32 # asm 1: vpsllq $32,v10=reg128#16 # asm 2: vpsllq $32,v10=%xmm15 vpsllq $32, % xmm12, % xmm15 # qhasm: 2x v01 = r2 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#9 # asm 2: vpsrlq $32,v01=%xmm8 vpsrlq $32, % xmm8, % xmm8 # qhasm: v11 = r6 & mask1 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm1, % xmm12, % xmm12 # qhasm: r2 = v00 | v10 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#9 # asm 2: vpor r6=%xmm8 vpor % xmm12, % xmm8, % xmm8 # qhasm: v00 = r3 & mask0 # asm 1: vpand v00=reg128#1 # asm 2: vpand v00=%xmm0 vpand % xmm0, % xmm9, % xmm0 # qhasm: 2x v10 = r7 << 32 # asm 1: vpsllq $32,v10=reg128#13 # asm 2: vpsllq $32,v10=%xmm12 vpsllq $32, % xmm13, % xmm12 # qhasm: 2x v01 = r3 unsigned>> 32 # asm 1: vpsrlq $32,v01=reg128#10 # asm 2: vpsrlq $32,v01=%xmm9 vpsrlq $32, % xmm9, % xmm9 # qhasm: v11 = r7 & mask1 # asm 1: vpand v11=reg128#2 # asm 2: vpand v11=%xmm1 vpand % xmm1, % xmm13, % xmm1 # qhasm: r3 = v00 | v10 # asm 1: vpor r3=reg128#1 # asm 2: vpor r3=%xmm0 vpor % xmm12, % xmm0, % xmm0 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#2 # asm 2: vpor r7=%xmm1 vpor % xmm1, % xmm9, % xmm1 # qhasm: v00 = r0 & mask2 # asm 1: vpand v00=reg128#10 # asm 2: vpand v00=%xmm9 vpand % xmm2, % xmm14, % xmm9 # qhasm: 4x v10 = r2 << 16 # asm 1: vpslld $16,v10=reg128#13 # asm 2: vpslld $16,v10=%xmm12 vpslld $16, % xmm11, % xmm12 # qhasm: 4x v01 = r0 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#14 # asm 2: vpsrld $16,v01=%xmm13 vpsrld $16, % xmm14, % xmm13 # qhasm: v11 = r2 & mask3 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm3, % xmm11, % xmm11 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#10 # asm 2: vpor r0=%xmm9 vpor % xmm12, % xmm9, % xmm9 # qhasm: r2 = v01 | v11 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm11, % xmm13, % xmm11 # qhasm: v00 = r1 & mask2 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm2, % xmm10, % xmm12 # qhasm: 4x v10 = r3 << 16 # asm 1: vpslld $16,v10=reg128#14 # asm 2: vpslld $16,v10=%xmm13 vpslld $16, % xmm0, % xmm13 # qhasm: 4x v01 = r1 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#11 # asm 2: vpsrld $16,v01=%xmm10 vpsrld $16, % xmm10, % xmm10 # qhasm: v11 = r3 & mask3 # asm 1: vpand v11=reg128#1 # asm 2: vpand v11=%xmm0 vpand % xmm3, % xmm0, % xmm0 # qhasm: r1 = v00 | v10 # asm 1: vpor r1=reg128#13 # asm 2: vpor r1=%xmm12 vpor % xmm13, % xmm12, % xmm12 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#1 # asm 2: vpor r3=%xmm0 vpor % xmm0, % xmm10, % xmm0 # qhasm: v00 = r4 & mask2 # asm 1: vpand v00=reg128#11 # asm 2: vpand v00=%xmm10 vpand % xmm2, % xmm6, % xmm10 # qhasm: 4x v10 = r6 << 16 # asm 1: vpslld $16,v10=reg128#14 # asm 2: vpslld $16,v10=%xmm13 vpslld $16, % xmm8, % xmm13 # qhasm: 4x v01 = r4 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#7 # asm 2: vpsrld $16,v01=%xmm6 vpsrld $16, % xmm6, % xmm6 # qhasm: v11 = r6 & mask3 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm3, % xmm8, % xmm8 # qhasm: r4 = v00 | v10 # asm 1: vpor r4=reg128#11 # asm 2: vpor r4=%xmm10 vpor % xmm13, % xmm10, % xmm10 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#7 # asm 2: vpor r6=%xmm6 vpor % xmm8, % xmm6, % xmm6 # qhasm: v00 = r5 & mask2 # asm 1: vpand v00=reg128#3 # asm 2: vpand v00=%xmm2 vpand % xmm2, % xmm7, % xmm2 # qhasm: 4x v10 = r7 << 16 # asm 1: vpslld $16,v10=reg128#9 # asm 2: vpslld $16,v10=%xmm8 vpslld $16, % xmm1, % xmm8 # qhasm: 4x v01 = r5 unsigned>> 16 # asm 1: vpsrld $16,v01=reg128#8 # asm 2: vpsrld $16,v01=%xmm7 vpsrld $16, % xmm7, % xmm7 # qhasm: v11 = r7 & mask3 # asm 1: vpand v11=reg128#2 # asm 2: vpand v11=%xmm1 vpand % xmm3, % xmm1, % xmm1 # qhasm: r5 = v00 | v10 # asm 1: vpor r5=reg128#3 # asm 2: vpor r5=%xmm2 vpor % xmm8, % xmm2, % xmm2 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#2 # asm 2: vpor r7=%xmm1 vpor % xmm1, % xmm7, % xmm1 # qhasm: v00 = r0 & mask4 # asm 1: vpand v00=reg128#4 # asm 2: vpand v00=%xmm3 vpand % xmm4, % xmm9, % xmm3 # qhasm: 8x v10 = r1 << 8 # asm 1: vpsllw $8,v10=reg128#8 # asm 2: vpsllw $8,v10=%xmm7 vpsllw $8, % xmm12, % xmm7 # qhasm: 8x v01 = r0 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#9 # asm 2: vpsrlw $8,v01=%xmm8 vpsrlw $8, % xmm9, % xmm8 # qhasm: v11 = r1 & mask5 # asm 1: vpand v11=reg128#10 # asm 2: vpand v11=%xmm9 vpand % xmm5, % xmm12, % xmm9 # qhasm: r0 = v00 | v10 # asm 1: vpor r0=reg128#4 # asm 2: vpor r0=%xmm3 vpor % xmm7, % xmm3, % xmm3 # qhasm: r1 = v01 | v11 # asm 1: vpor r1=reg128#8 # asm 2: vpor r1=%xmm7 vpor % xmm9, % xmm8, % xmm7 # qhasm: v00 = r2 & mask4 # asm 1: vpand v00=reg128#9 # asm 2: vpand v00=%xmm8 vpand % xmm4, % xmm11, % xmm8 # qhasm: 8x v10 = r3 << 8 # asm 1: vpsllw $8,v10=reg128#10 # asm 2: vpsllw $8,v10=%xmm9 vpsllw $8, % xmm0, % xmm9 # qhasm: 8x v01 = r2 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#12 # asm 2: vpsrlw $8,v01=%xmm11 vpsrlw $8, % xmm11, % xmm11 # qhasm: v11 = r3 & mask5 # asm 1: vpand v11=reg128#1 # asm 2: vpand v11=%xmm0 vpand % xmm5, % xmm0, % xmm0 # qhasm: r2 = v00 | v10 # asm 1: vpor r2=reg128#9 # asm 2: vpor r2=%xmm8 vpor % xmm9, % xmm8, % xmm8 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#1 # asm 2: vpor r3=%xmm0 vpor % xmm0, % xmm11, % xmm0 # qhasm: v00 = r4 & mask4 # asm 1: vpand v00=reg128#10 # asm 2: vpand v00=%xmm9 vpand % xmm4, % xmm10, % xmm9 # qhasm: 8x v10 = r5 << 8 # asm 1: vpsllw $8,v10=reg128#12 # asm 2: vpsllw $8,v10=%xmm11 vpsllw $8, % xmm2, % xmm11 # qhasm: 8x v01 = r4 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#11 # asm 2: vpsrlw $8,v01=%xmm10 vpsrlw $8, % xmm10, % xmm10 # qhasm: v11 = r5 & mask5 # asm 1: vpand v11=reg128#3 # asm 2: vpand v11=%xmm2 vpand % xmm5, % xmm2, % xmm2 # qhasm: r4 = v00 | v10 # asm 1: vpor r4=reg128#10 # asm 2: vpor r4=%xmm9 vpor % xmm11, % xmm9, % xmm9 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#3 # asm 2: vpor r5=%xmm2 vpor % xmm2, % xmm10, % xmm2 # qhasm: v00 = r6 & mask4 # asm 1: vpand v00=reg128#5 # asm 2: vpand v00=%xmm4 vpand % xmm4, % xmm6, % xmm4 # qhasm: 8x v10 = r7 << 8 # asm 1: vpsllw $8,v10=reg128#11 # asm 2: vpsllw $8,v10=%xmm10 vpsllw $8, % xmm1, % xmm10 # qhasm: 8x v01 = r6 unsigned>> 8 # asm 1: vpsrlw $8,v01=reg128#7 # asm 2: vpsrlw $8,v01=%xmm6 vpsrlw $8, % xmm6, % xmm6 # qhasm: v11 = r7 & mask5 # asm 1: vpand v11=reg128#2 # asm 2: vpand v11=%xmm1 vpand % xmm5, % xmm1, % xmm1 # qhasm: r6 = v00 | v10 # asm 1: vpor r6=reg128#5 # asm 2: vpor r6=%xmm4 vpor % xmm10, % xmm4, % xmm4 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#2 # asm 2: vpor r7=%xmm1 vpor % xmm1, % xmm6, % xmm1 # qhasm: buf = r0[0] # asm 1: pextrq $0x0,buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm3, % rsi # qhasm: mem64[ input_0 + 56 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm7, % rsi # qhasm: mem64[ input_0 + 120 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm8, % rsi # qhasm: mem64[ input_0 + 184 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm0, % rsi # qhasm: mem64[ input_0 + 248 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm9, % rsi # qhasm: mem64[ input_0 + 312 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm2, % rsi # qhasm: mem64[ input_0 + 376 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm4, % rsi # qhasm: mem64[ input_0 + 440 ] = buf # asm 1: movq buf=int64#2 # asm 2: pextrq $0x0,buf=%rsi pextrq $0x0, % xmm1, % rsi # qhasm: mem64[ input_0 + 504 ] = buf # asm 1: movq mask0=reg128#1 # asm 2: movdqa MASK2_0(%rip),>mask0=%xmm0 movdqa MASK2_0( % rip), % xmm0 # qhasm: mask1 aligned= mem128[ MASK2_1 ] # asm 1: movdqa MASK2_1(%rip),>mask1=reg128#2 # asm 2: movdqa MASK2_1(%rip),>mask1=%xmm1 movdqa MASK2_1( % rip), % xmm1 # qhasm: mask2 aligned= mem128[ MASK1_0 ] # asm 1: movdqa MASK1_0(%rip),>mask2=reg128#3 # asm 2: movdqa MASK1_0(%rip),>mask2=%xmm2 movdqa MASK1_0( % rip), % xmm2 # qhasm: mask3 aligned= mem128[ MASK1_1 ] # asm 1: movdqa MASK1_1(%rip),>mask3=reg128#4 # asm 2: movdqa MASK1_1(%rip),>mask3=%xmm3 movdqa MASK1_1( % rip), % xmm3 # qhasm: mask4 aligned= mem128[ MASK0_0 ] # asm 1: movdqa MASK0_0(%rip),>mask4=reg128#5 # asm 2: movdqa MASK0_0(%rip),>mask4=%xmm4 movdqa MASK0_0( % rip), % xmm4 # qhasm: mask5 aligned= mem128[ MASK0_1 ] # asm 1: movdqa MASK0_1(%rip),>mask5=reg128#6 # asm 2: movdqa MASK0_1(%rip),>mask5=%xmm5 movdqa MASK0_1( % rip), % xmm5 # qhasm: r0 = mem64[ input_0 + 0 ] x2 # asm 1: movddup 0(r0=reg128#7 # asm 2: movddup 0(r0=%xmm6 movddup 0( % rdi), % xmm6 # qhasm: r1 = mem64[ input_0 + 8 ] x2 # asm 1: movddup 8(r1=reg128#8 # asm 2: movddup 8(r1=%xmm7 movddup 8( % rdi), % xmm7 # qhasm: r2 = mem64[ input_0 + 16 ] x2 # asm 1: movddup 16(r2=reg128#9 # asm 2: movddup 16(r2=%xmm8 movddup 16( % rdi), % xmm8 # qhasm: r3 = mem64[ input_0 + 24 ] x2 # asm 1: movddup 24(r3=reg128#10 # asm 2: movddup 24(r3=%xmm9 movddup 24( % rdi), % xmm9 # qhasm: r4 = mem64[ input_0 + 32 ] x2 # asm 1: movddup 32(r4=reg128#11 # asm 2: movddup 32(r4=%xmm10 movddup 32( % rdi), % xmm10 # qhasm: r5 = mem64[ input_0 + 40 ] x2 # asm 1: movddup 40(r5=reg128#12 # asm 2: movddup 40(r5=%xmm11 movddup 40( % rdi), % xmm11 # qhasm: r6 = mem64[ input_0 + 48 ] x2 # asm 1: movddup 48(r6=reg128#13 # asm 2: movddup 48(r6=%xmm12 movddup 48( % rdi), % xmm12 # qhasm: r7 = mem64[ input_0 + 56 ] x2 # asm 1: movddup 56(r7=reg128#14 # asm 2: movddup 56(r7=%xmm13 movddup 56( % rdi), % xmm13 # qhasm: v00 = r0 & mask0 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm0, % xmm6, % xmm14 # qhasm: v10 = r4 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm10, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm1, % xmm6, % xmm6 # qhasm: v11 = r4 & mask1 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm1, % xmm10, % xmm10 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r0=reg128#15 # asm 2: vpor r0=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r4 = v01 | v11 # asm 1: vpor r4=reg128#7 # asm 2: vpor r4=%xmm6 vpor % xmm10, % xmm6, % xmm6 # qhasm: v00 = r1 & mask0 # asm 1: vpand v00=reg128#11 # asm 2: vpand v00=%xmm10 vpand % xmm0, % xmm7, % xmm10 # qhasm: v10 = r5 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm11, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#8 # asm 2: vpand v01=%xmm7 vpand % xmm1, % xmm7, % xmm7 # qhasm: v11 = r5 & mask1 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm1, % xmm11, % xmm11 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r1=reg128#11 # asm 2: vpor r1=%xmm10 vpor % xmm15, % xmm10, % xmm10 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#8 # asm 2: vpor r5=%xmm7 vpor % xmm11, % xmm7, % xmm7 # qhasm: v00 = r2 & mask0 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm0, % xmm8, % xmm11 # qhasm: v10 = r6 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm12, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#9 # asm 2: vpand v01=%xmm8 vpand % xmm1, % xmm8, % xmm8 # qhasm: v11 = r6 & mask1 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm1, % xmm12, % xmm12 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#9 # asm 2: vpor r6=%xmm8 vpor % xmm12, % xmm8, % xmm8 # qhasm: v00 = r3 & mask0 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm0, % xmm9, % xmm12 # qhasm: v10 = r7 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm13, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#10 # asm 2: vpand v01=%xmm9 vpand % xmm1, % xmm9, % xmm9 # qhasm: v11 = r7 & mask1 # asm 1: vpand v11=reg128#14 # asm 2: vpand v11=%xmm13 vpand % xmm1, % xmm13, % xmm13 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r3=reg128#13 # asm 2: vpor r3=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#10 # asm 2: vpor r7=%xmm9 vpor % xmm13, % xmm9, % xmm9 # qhasm: v00 = r0 & mask2 # asm 1: vpand v00=reg128#14 # asm 2: vpand v00=%xmm13 vpand % xmm2, % xmm14, % xmm13 # qhasm: v10 = r2 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm11, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#15 # asm 2: vpand v01=%xmm14 vpand % xmm3, % xmm14, % xmm14 # qhasm: v11 = r2 & mask3 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm3, % xmm11, % xmm11 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r0=reg128#14 # asm 2: vpor r0=%xmm13 vpor % xmm15, % xmm13, % xmm13 # qhasm: r2 = v01 | v11 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm11, % xmm14, % xmm11 # qhasm: v00 = r1 & mask2 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm2, % xmm10, % xmm14 # qhasm: v10 = r3 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm12, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#11 # asm 2: vpand v01=%xmm10 vpand % xmm3, % xmm10, % xmm10 # qhasm: v11 = r3 & mask3 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm3, % xmm12, % xmm12 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r1=reg128#15 # asm 2: vpor r1=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm12, % xmm10, % xmm10 # qhasm: v00 = r4 & mask2 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm2, % xmm6, % xmm12 # qhasm: v10 = r6 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm8, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm3, % xmm6, % xmm6 # qhasm: v11 = r6 & mask3 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm3, % xmm8, % xmm8 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r4=reg128#13 # asm 2: vpor r4=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#7 # asm 2: vpor r6=%xmm6 vpor % xmm8, % xmm6, % xmm6 # qhasm: v00 = r5 & mask2 # asm 1: vpand v00=reg128#9 # asm 2: vpand v00=%xmm8 vpand % xmm2, % xmm7, % xmm8 # qhasm: v10 = r7 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm9, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#8 # asm 2: vpand v01=%xmm7 vpand % xmm3, % xmm7, % xmm7 # qhasm: v11 = r7 & mask3 # asm 1: vpand v11=reg128#10 # asm 2: vpand v11=%xmm9 vpand % xmm3, % xmm9, % xmm9 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm15, % xmm8, % xmm8 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#8 # asm 2: vpor r7=%xmm7 vpor % xmm9, % xmm7, % xmm7 # qhasm: v00 = r0 & mask4 # asm 1: vpand v00=reg128#10 # asm 2: vpand v00=%xmm9 vpand % xmm4, % xmm13, % xmm9 # qhasm: v10 = r1 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm14, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#14 # asm 2: vpand v01=%xmm13 vpand % xmm5, % xmm13, % xmm13 # qhasm: v11 = r1 & mask5 # asm 1: vpand v11=reg128#15 # asm 2: vpand v11=%xmm14 vpand % xmm5, % xmm14, % xmm14 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r0=reg128#10 # asm 2: vpor r0=%xmm9 vpor % xmm15, % xmm9, % xmm9 # qhasm: r1 = v01 | v11 # asm 1: vpor r1=reg128#14 # asm 2: vpor r1=%xmm13 vpor % xmm14, % xmm13, % xmm13 # qhasm: v00 = r2 & mask4 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm4, % xmm11, % xmm14 # qhasm: v10 = r3 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm10, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#12 # asm 2: vpand v01=%xmm11 vpand % xmm5, % xmm11, % xmm11 # qhasm: v11 = r3 & mask5 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm5, % xmm10, % xmm10 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r2=reg128#15 # asm 2: vpor r2=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm10, % xmm11, % xmm10 # qhasm: v00 = r4 & mask4 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm4, % xmm12, % xmm11 # qhasm: v10 = r5 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm8, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#13 # asm 2: vpand v01=%xmm12 vpand % xmm5, % xmm12, % xmm12 # qhasm: v11 = r5 & mask5 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm5, % xmm8, % xmm8 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r4=reg128#12 # asm 2: vpor r4=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm8, % xmm12, % xmm8 # qhasm: v00 = r6 & mask4 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm4, % xmm6, % xmm12 # qhasm: v10 = r7 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm7, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm5, % xmm6, % xmm6 # qhasm: v11 = r7 & mask5 # asm 1: vpand v11=reg128#8 # asm 2: vpand v11=%xmm7 vpand % xmm5, % xmm7, % xmm7 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r6=reg128#13 # asm 2: vpor r6=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#7 # asm 2: vpor r7=%xmm6 vpor % xmm7, % xmm6, % xmm6 # qhasm: t0 = r0[0]r1[0] # asm 1: vpunpcklqdq t0=reg128#8 # asm 2: vpunpcklqdq t0=%xmm7 vpunpcklqdq % xmm13, % xmm9, % xmm7 # qhasm: mem128[ input_0 + 0 ] = t0 # asm 1: movdqu t0=reg128#8 # asm 2: vpunpcklqdq t0=%xmm7 vpunpcklqdq % xmm10, % xmm14, % xmm7 # qhasm: mem128[ input_0 + 16 ] = t0 # asm 1: movdqu t0=reg128#8 # asm 2: vpunpcklqdq t0=%xmm7 vpunpcklqdq % xmm8, % xmm11, % xmm7 # qhasm: mem128[ input_0 + 32 ] = t0 # asm 1: movdqu t0=reg128#7 # asm 2: vpunpcklqdq t0=%xmm6 vpunpcklqdq % xmm6, % xmm12, % xmm6 # qhasm: mem128[ input_0 + 48 ] = t0 # asm 1: movdqu r0=reg128#7 # asm 2: movddup 64(r0=%xmm6 movddup 64( % rdi), % xmm6 # qhasm: r1 = mem64[ input_0 + 72 ] x2 # asm 1: movddup 72(r1=reg128#8 # asm 2: movddup 72(r1=%xmm7 movddup 72( % rdi), % xmm7 # qhasm: r2 = mem64[ input_0 + 80 ] x2 # asm 1: movddup 80(r2=reg128#9 # asm 2: movddup 80(r2=%xmm8 movddup 80( % rdi), % xmm8 # qhasm: r3 = mem64[ input_0 + 88 ] x2 # asm 1: movddup 88(r3=reg128#10 # asm 2: movddup 88(r3=%xmm9 movddup 88( % rdi), % xmm9 # qhasm: r4 = mem64[ input_0 + 96 ] x2 # asm 1: movddup 96(r4=reg128#11 # asm 2: movddup 96(r4=%xmm10 movddup 96( % rdi), % xmm10 # qhasm: r5 = mem64[ input_0 + 104 ] x2 # asm 1: movddup 104(r5=reg128#12 # asm 2: movddup 104(r5=%xmm11 movddup 104( % rdi), % xmm11 # qhasm: r6 = mem64[ input_0 + 112 ] x2 # asm 1: movddup 112(r6=reg128#13 # asm 2: movddup 112(r6=%xmm12 movddup 112( % rdi), % xmm12 # qhasm: r7 = mem64[ input_0 + 120 ] x2 # asm 1: movddup 120(r7=reg128#14 # asm 2: movddup 120(r7=%xmm13 movddup 120( % rdi), % xmm13 # qhasm: v00 = r0 & mask0 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm0, % xmm6, % xmm14 # qhasm: v10 = r4 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm10, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm1, % xmm6, % xmm6 # qhasm: v11 = r4 & mask1 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm1, % xmm10, % xmm10 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r0=reg128#15 # asm 2: vpor r0=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r4 = v01 | v11 # asm 1: vpor r4=reg128#7 # asm 2: vpor r4=%xmm6 vpor % xmm10, % xmm6, % xmm6 # qhasm: v00 = r1 & mask0 # asm 1: vpand v00=reg128#11 # asm 2: vpand v00=%xmm10 vpand % xmm0, % xmm7, % xmm10 # qhasm: v10 = r5 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm11, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#8 # asm 2: vpand v01=%xmm7 vpand % xmm1, % xmm7, % xmm7 # qhasm: v11 = r5 & mask1 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm1, % xmm11, % xmm11 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r1=reg128#11 # asm 2: vpor r1=%xmm10 vpor % xmm15, % xmm10, % xmm10 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#8 # asm 2: vpor r5=%xmm7 vpor % xmm11, % xmm7, % xmm7 # qhasm: v00 = r2 & mask0 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm0, % xmm8, % xmm11 # qhasm: v10 = r6 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm12, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#9 # asm 2: vpand v01=%xmm8 vpand % xmm1, % xmm8, % xmm8 # qhasm: v11 = r6 & mask1 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm1, % xmm12, % xmm12 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#9 # asm 2: vpor r6=%xmm8 vpor % xmm12, % xmm8, % xmm8 # qhasm: v00 = r3 & mask0 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm0, % xmm9, % xmm12 # qhasm: v10 = r7 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm13, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#10 # asm 2: vpand v01=%xmm9 vpand % xmm1, % xmm9, % xmm9 # qhasm: v11 = r7 & mask1 # asm 1: vpand v11=reg128#14 # asm 2: vpand v11=%xmm13 vpand % xmm1, % xmm13, % xmm13 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r3=reg128#13 # asm 2: vpor r3=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#10 # asm 2: vpor r7=%xmm9 vpor % xmm13, % xmm9, % xmm9 # qhasm: v00 = r0 & mask2 # asm 1: vpand v00=reg128#14 # asm 2: vpand v00=%xmm13 vpand % xmm2, % xmm14, % xmm13 # qhasm: v10 = r2 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm11, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#15 # asm 2: vpand v01=%xmm14 vpand % xmm3, % xmm14, % xmm14 # qhasm: v11 = r2 & mask3 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm3, % xmm11, % xmm11 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r0=reg128#14 # asm 2: vpor r0=%xmm13 vpor % xmm15, % xmm13, % xmm13 # qhasm: r2 = v01 | v11 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm11, % xmm14, % xmm11 # qhasm: v00 = r1 & mask2 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm2, % xmm10, % xmm14 # qhasm: v10 = r3 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm12, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#11 # asm 2: vpand v01=%xmm10 vpand % xmm3, % xmm10, % xmm10 # qhasm: v11 = r3 & mask3 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm3, % xmm12, % xmm12 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r1=reg128#15 # asm 2: vpor r1=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm12, % xmm10, % xmm10 # qhasm: v00 = r4 & mask2 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm2, % xmm6, % xmm12 # qhasm: v10 = r6 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm8, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm3, % xmm6, % xmm6 # qhasm: v11 = r6 & mask3 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm3, % xmm8, % xmm8 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r4=reg128#13 # asm 2: vpor r4=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#7 # asm 2: vpor r6=%xmm6 vpor % xmm8, % xmm6, % xmm6 # qhasm: v00 = r5 & mask2 # asm 1: vpand v00=reg128#9 # asm 2: vpand v00=%xmm8 vpand % xmm2, % xmm7, % xmm8 # qhasm: v10 = r7 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm9, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#8 # asm 2: vpand v01=%xmm7 vpand % xmm3, % xmm7, % xmm7 # qhasm: v11 = r7 & mask3 # asm 1: vpand v11=reg128#10 # asm 2: vpand v11=%xmm9 vpand % xmm3, % xmm9, % xmm9 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm15, % xmm8, % xmm8 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#8 # asm 2: vpor r7=%xmm7 vpor % xmm9, % xmm7, % xmm7 # qhasm: v00 = r0 & mask4 # asm 1: vpand v00=reg128#10 # asm 2: vpand v00=%xmm9 vpand % xmm4, % xmm13, % xmm9 # qhasm: v10 = r1 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm14, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#14 # asm 2: vpand v01=%xmm13 vpand % xmm5, % xmm13, % xmm13 # qhasm: v11 = r1 & mask5 # asm 1: vpand v11=reg128#15 # asm 2: vpand v11=%xmm14 vpand % xmm5, % xmm14, % xmm14 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r0=reg128#10 # asm 2: vpor r0=%xmm9 vpor % xmm15, % xmm9, % xmm9 # qhasm: r1 = v01 | v11 # asm 1: vpor r1=reg128#14 # asm 2: vpor r1=%xmm13 vpor % xmm14, % xmm13, % xmm13 # qhasm: v00 = r2 & mask4 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm4, % xmm11, % xmm14 # qhasm: v10 = r3 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm10, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#12 # asm 2: vpand v01=%xmm11 vpand % xmm5, % xmm11, % xmm11 # qhasm: v11 = r3 & mask5 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm5, % xmm10, % xmm10 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r2=reg128#15 # asm 2: vpor r2=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm10, % xmm11, % xmm10 # qhasm: v00 = r4 & mask4 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm4, % xmm12, % xmm11 # qhasm: v10 = r5 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm8, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#13 # asm 2: vpand v01=%xmm12 vpand % xmm5, % xmm12, % xmm12 # qhasm: v11 = r5 & mask5 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm5, % xmm8, % xmm8 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r4=reg128#12 # asm 2: vpor r4=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm8, % xmm12, % xmm8 # qhasm: v00 = r6 & mask4 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm4, % xmm6, % xmm12 # qhasm: v10 = r7 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm7, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm5, % xmm6, % xmm6 # qhasm: v11 = r7 & mask5 # asm 1: vpand v11=reg128#8 # asm 2: vpand v11=%xmm7 vpand % xmm5, % xmm7, % xmm7 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r6=reg128#13 # asm 2: vpor r6=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#7 # asm 2: vpor r7=%xmm6 vpor % xmm7, % xmm6, % xmm6 # qhasm: t0 = r0[0]r1[0] # asm 1: vpunpcklqdq t0=reg128#8 # asm 2: vpunpcklqdq t0=%xmm7 vpunpcklqdq % xmm13, % xmm9, % xmm7 # qhasm: mem128[ input_0 + 64 ] = t0 # asm 1: movdqu t0=reg128#8 # asm 2: vpunpcklqdq t0=%xmm7 vpunpcklqdq % xmm10, % xmm14, % xmm7 # qhasm: mem128[ input_0 + 80 ] = t0 # asm 1: movdqu t0=reg128#8 # asm 2: vpunpcklqdq t0=%xmm7 vpunpcklqdq % xmm8, % xmm11, % xmm7 # qhasm: mem128[ input_0 + 96 ] = t0 # asm 1: movdqu t0=reg128#7 # asm 2: vpunpcklqdq t0=%xmm6 vpunpcklqdq % xmm6, % xmm12, % xmm6 # qhasm: mem128[ input_0 + 112 ] = t0 # asm 1: movdqu r0=reg128#7 # asm 2: movddup 128(r0=%xmm6 movddup 128( % rdi), % xmm6 # qhasm: r1 = mem64[ input_0 + 136 ] x2 # asm 1: movddup 136(r1=reg128#8 # asm 2: movddup 136(r1=%xmm7 movddup 136( % rdi), % xmm7 # qhasm: r2 = mem64[ input_0 + 144 ] x2 # asm 1: movddup 144(r2=reg128#9 # asm 2: movddup 144(r2=%xmm8 movddup 144( % rdi), % xmm8 # qhasm: r3 = mem64[ input_0 + 152 ] x2 # asm 1: movddup 152(r3=reg128#10 # asm 2: movddup 152(r3=%xmm9 movddup 152( % rdi), % xmm9 # qhasm: r4 = mem64[ input_0 + 160 ] x2 # asm 1: movddup 160(r4=reg128#11 # asm 2: movddup 160(r4=%xmm10 movddup 160( % rdi), % xmm10 # qhasm: r5 = mem64[ input_0 + 168 ] x2 # asm 1: movddup 168(r5=reg128#12 # asm 2: movddup 168(r5=%xmm11 movddup 168( % rdi), % xmm11 # qhasm: r6 = mem64[ input_0 + 176 ] x2 # asm 1: movddup 176(r6=reg128#13 # asm 2: movddup 176(r6=%xmm12 movddup 176( % rdi), % xmm12 # qhasm: r7 = mem64[ input_0 + 184 ] x2 # asm 1: movddup 184(r7=reg128#14 # asm 2: movddup 184(r7=%xmm13 movddup 184( % rdi), % xmm13 # qhasm: v00 = r0 & mask0 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm0, % xmm6, % xmm14 # qhasm: v10 = r4 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm10, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm1, % xmm6, % xmm6 # qhasm: v11 = r4 & mask1 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm1, % xmm10, % xmm10 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r0=reg128#15 # asm 2: vpor r0=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r4 = v01 | v11 # asm 1: vpor r4=reg128#7 # asm 2: vpor r4=%xmm6 vpor % xmm10, % xmm6, % xmm6 # qhasm: v00 = r1 & mask0 # asm 1: vpand v00=reg128#11 # asm 2: vpand v00=%xmm10 vpand % xmm0, % xmm7, % xmm10 # qhasm: v10 = r5 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm11, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#8 # asm 2: vpand v01=%xmm7 vpand % xmm1, % xmm7, % xmm7 # qhasm: v11 = r5 & mask1 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm1, % xmm11, % xmm11 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r1=reg128#11 # asm 2: vpor r1=%xmm10 vpor % xmm15, % xmm10, % xmm10 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#8 # asm 2: vpor r5=%xmm7 vpor % xmm11, % xmm7, % xmm7 # qhasm: v00 = r2 & mask0 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm0, % xmm8, % xmm11 # qhasm: v10 = r6 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm12, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#9 # asm 2: vpand v01=%xmm8 vpand % xmm1, % xmm8, % xmm8 # qhasm: v11 = r6 & mask1 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm1, % xmm12, % xmm12 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#9 # asm 2: vpor r6=%xmm8 vpor % xmm12, % xmm8, % xmm8 # qhasm: v00 = r3 & mask0 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm0, % xmm9, % xmm12 # qhasm: v10 = r7 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm13, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#10 # asm 2: vpand v01=%xmm9 vpand % xmm1, % xmm9, % xmm9 # qhasm: v11 = r7 & mask1 # asm 1: vpand v11=reg128#14 # asm 2: vpand v11=%xmm13 vpand % xmm1, % xmm13, % xmm13 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r3=reg128#13 # asm 2: vpor r3=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#10 # asm 2: vpor r7=%xmm9 vpor % xmm13, % xmm9, % xmm9 # qhasm: v00 = r0 & mask2 # asm 1: vpand v00=reg128#14 # asm 2: vpand v00=%xmm13 vpand % xmm2, % xmm14, % xmm13 # qhasm: v10 = r2 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm11, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#15 # asm 2: vpand v01=%xmm14 vpand % xmm3, % xmm14, % xmm14 # qhasm: v11 = r2 & mask3 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm3, % xmm11, % xmm11 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r0=reg128#14 # asm 2: vpor r0=%xmm13 vpor % xmm15, % xmm13, % xmm13 # qhasm: r2 = v01 | v11 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm11, % xmm14, % xmm11 # qhasm: v00 = r1 & mask2 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm2, % xmm10, % xmm14 # qhasm: v10 = r3 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm12, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#11 # asm 2: vpand v01=%xmm10 vpand % xmm3, % xmm10, % xmm10 # qhasm: v11 = r3 & mask3 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm3, % xmm12, % xmm12 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r1=reg128#15 # asm 2: vpor r1=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm12, % xmm10, % xmm10 # qhasm: v00 = r4 & mask2 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm2, % xmm6, % xmm12 # qhasm: v10 = r6 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm8, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm3, % xmm6, % xmm6 # qhasm: v11 = r6 & mask3 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm3, % xmm8, % xmm8 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r4=reg128#13 # asm 2: vpor r4=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#7 # asm 2: vpor r6=%xmm6 vpor % xmm8, % xmm6, % xmm6 # qhasm: v00 = r5 & mask2 # asm 1: vpand v00=reg128#9 # asm 2: vpand v00=%xmm8 vpand % xmm2, % xmm7, % xmm8 # qhasm: v10 = r7 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm9, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#8 # asm 2: vpand v01=%xmm7 vpand % xmm3, % xmm7, % xmm7 # qhasm: v11 = r7 & mask3 # asm 1: vpand v11=reg128#10 # asm 2: vpand v11=%xmm9 vpand % xmm3, % xmm9, % xmm9 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm15, % xmm8, % xmm8 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#8 # asm 2: vpor r7=%xmm7 vpor % xmm9, % xmm7, % xmm7 # qhasm: v00 = r0 & mask4 # asm 1: vpand v00=reg128#10 # asm 2: vpand v00=%xmm9 vpand % xmm4, % xmm13, % xmm9 # qhasm: v10 = r1 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm14, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#14 # asm 2: vpand v01=%xmm13 vpand % xmm5, % xmm13, % xmm13 # qhasm: v11 = r1 & mask5 # asm 1: vpand v11=reg128#15 # asm 2: vpand v11=%xmm14 vpand % xmm5, % xmm14, % xmm14 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r0=reg128#10 # asm 2: vpor r0=%xmm9 vpor % xmm15, % xmm9, % xmm9 # qhasm: r1 = v01 | v11 # asm 1: vpor r1=reg128#14 # asm 2: vpor r1=%xmm13 vpor % xmm14, % xmm13, % xmm13 # qhasm: v00 = r2 & mask4 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm4, % xmm11, % xmm14 # qhasm: v10 = r3 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm10, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#12 # asm 2: vpand v01=%xmm11 vpand % xmm5, % xmm11, % xmm11 # qhasm: v11 = r3 & mask5 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm5, % xmm10, % xmm10 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r2=reg128#15 # asm 2: vpor r2=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm10, % xmm11, % xmm10 # qhasm: v00 = r4 & mask4 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm4, % xmm12, % xmm11 # qhasm: v10 = r5 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm8, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#13 # asm 2: vpand v01=%xmm12 vpand % xmm5, % xmm12, % xmm12 # qhasm: v11 = r5 & mask5 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm5, % xmm8, % xmm8 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r4=reg128#12 # asm 2: vpor r4=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm8, % xmm12, % xmm8 # qhasm: v00 = r6 & mask4 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm4, % xmm6, % xmm12 # qhasm: v10 = r7 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm7, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm5, % xmm6, % xmm6 # qhasm: v11 = r7 & mask5 # asm 1: vpand v11=reg128#8 # asm 2: vpand v11=%xmm7 vpand % xmm5, % xmm7, % xmm7 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r6=reg128#13 # asm 2: vpor r6=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#7 # asm 2: vpor r7=%xmm6 vpor % xmm7, % xmm6, % xmm6 # qhasm: t0 = r0[0]r1[0] # asm 1: vpunpcklqdq t0=reg128#8 # asm 2: vpunpcklqdq t0=%xmm7 vpunpcklqdq % xmm13, % xmm9, % xmm7 # qhasm: mem128[ input_0 + 128 ] = t0 # asm 1: movdqu t0=reg128#8 # asm 2: vpunpcklqdq t0=%xmm7 vpunpcklqdq % xmm10, % xmm14, % xmm7 # qhasm: mem128[ input_0 + 144 ] = t0 # asm 1: movdqu t0=reg128#8 # asm 2: vpunpcklqdq t0=%xmm7 vpunpcklqdq % xmm8, % xmm11, % xmm7 # qhasm: mem128[ input_0 + 160 ] = t0 # asm 1: movdqu t0=reg128#7 # asm 2: vpunpcklqdq t0=%xmm6 vpunpcklqdq % xmm6, % xmm12, % xmm6 # qhasm: mem128[ input_0 + 176 ] = t0 # asm 1: movdqu r0=reg128#7 # asm 2: movddup 192(r0=%xmm6 movddup 192( % rdi), % xmm6 # qhasm: r1 = mem64[ input_0 + 200 ] x2 # asm 1: movddup 200(r1=reg128#8 # asm 2: movddup 200(r1=%xmm7 movddup 200( % rdi), % xmm7 # qhasm: r2 = mem64[ input_0 + 208 ] x2 # asm 1: movddup 208(r2=reg128#9 # asm 2: movddup 208(r2=%xmm8 movddup 208( % rdi), % xmm8 # qhasm: r3 = mem64[ input_0 + 216 ] x2 # asm 1: movddup 216(r3=reg128#10 # asm 2: movddup 216(r3=%xmm9 movddup 216( % rdi), % xmm9 # qhasm: r4 = mem64[ input_0 + 224 ] x2 # asm 1: movddup 224(r4=reg128#11 # asm 2: movddup 224(r4=%xmm10 movddup 224( % rdi), % xmm10 # qhasm: r5 = mem64[ input_0 + 232 ] x2 # asm 1: movddup 232(r5=reg128#12 # asm 2: movddup 232(r5=%xmm11 movddup 232( % rdi), % xmm11 # qhasm: r6 = mem64[ input_0 + 240 ] x2 # asm 1: movddup 240(r6=reg128#13 # asm 2: movddup 240(r6=%xmm12 movddup 240( % rdi), % xmm12 # qhasm: r7 = mem64[ input_0 + 248 ] x2 # asm 1: movddup 248(r7=reg128#14 # asm 2: movddup 248(r7=%xmm13 movddup 248( % rdi), % xmm13 # qhasm: v00 = r0 & mask0 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm0, % xmm6, % xmm14 # qhasm: v10 = r4 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm10, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm1, % xmm6, % xmm6 # qhasm: v11 = r4 & mask1 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm1, % xmm10, % xmm10 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r0=reg128#15 # asm 2: vpor r0=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r4 = v01 | v11 # asm 1: vpor r4=reg128#7 # asm 2: vpor r4=%xmm6 vpor % xmm10, % xmm6, % xmm6 # qhasm: v00 = r1 & mask0 # asm 1: vpand v00=reg128#11 # asm 2: vpand v00=%xmm10 vpand % xmm0, % xmm7, % xmm10 # qhasm: v10 = r5 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm11, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#8 # asm 2: vpand v01=%xmm7 vpand % xmm1, % xmm7, % xmm7 # qhasm: v11 = r5 & mask1 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm1, % xmm11, % xmm11 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r1=reg128#11 # asm 2: vpor r1=%xmm10 vpor % xmm15, % xmm10, % xmm10 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#8 # asm 2: vpor r5=%xmm7 vpor % xmm11, % xmm7, % xmm7 # qhasm: v00 = r2 & mask0 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm0, % xmm8, % xmm11 # qhasm: v10 = r6 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm12, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#9 # asm 2: vpand v01=%xmm8 vpand % xmm1, % xmm8, % xmm8 # qhasm: v11 = r6 & mask1 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm1, % xmm12, % xmm12 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#9 # asm 2: vpor r6=%xmm8 vpor % xmm12, % xmm8, % xmm8 # qhasm: v00 = r3 & mask0 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm0, % xmm9, % xmm12 # qhasm: v10 = r7 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm13, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#10 # asm 2: vpand v01=%xmm9 vpand % xmm1, % xmm9, % xmm9 # qhasm: v11 = r7 & mask1 # asm 1: vpand v11=reg128#14 # asm 2: vpand v11=%xmm13 vpand % xmm1, % xmm13, % xmm13 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r3=reg128#13 # asm 2: vpor r3=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#10 # asm 2: vpor r7=%xmm9 vpor % xmm13, % xmm9, % xmm9 # qhasm: v00 = r0 & mask2 # asm 1: vpand v00=reg128#14 # asm 2: vpand v00=%xmm13 vpand % xmm2, % xmm14, % xmm13 # qhasm: v10 = r2 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm11, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#15 # asm 2: vpand v01=%xmm14 vpand % xmm3, % xmm14, % xmm14 # qhasm: v11 = r2 & mask3 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm3, % xmm11, % xmm11 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r0=reg128#14 # asm 2: vpor r0=%xmm13 vpor % xmm15, % xmm13, % xmm13 # qhasm: r2 = v01 | v11 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm11, % xmm14, % xmm11 # qhasm: v00 = r1 & mask2 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm2, % xmm10, % xmm14 # qhasm: v10 = r3 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm12, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#11 # asm 2: vpand v01=%xmm10 vpand % xmm3, % xmm10, % xmm10 # qhasm: v11 = r3 & mask3 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm3, % xmm12, % xmm12 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r1=reg128#15 # asm 2: vpor r1=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm12, % xmm10, % xmm10 # qhasm: v00 = r4 & mask2 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm2, % xmm6, % xmm12 # qhasm: v10 = r6 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm8, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm3, % xmm6, % xmm6 # qhasm: v11 = r6 & mask3 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm3, % xmm8, % xmm8 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r4=reg128#13 # asm 2: vpor r4=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#7 # asm 2: vpor r6=%xmm6 vpor % xmm8, % xmm6, % xmm6 # qhasm: v00 = r5 & mask2 # asm 1: vpand v00=reg128#9 # asm 2: vpand v00=%xmm8 vpand % xmm2, % xmm7, % xmm8 # qhasm: v10 = r7 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm9, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#8 # asm 2: vpand v01=%xmm7 vpand % xmm3, % xmm7, % xmm7 # qhasm: v11 = r7 & mask3 # asm 1: vpand v11=reg128#10 # asm 2: vpand v11=%xmm9 vpand % xmm3, % xmm9, % xmm9 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm15, % xmm8, % xmm8 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#8 # asm 2: vpor r7=%xmm7 vpor % xmm9, % xmm7, % xmm7 # qhasm: v00 = r0 & mask4 # asm 1: vpand v00=reg128#10 # asm 2: vpand v00=%xmm9 vpand % xmm4, % xmm13, % xmm9 # qhasm: v10 = r1 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm14, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#14 # asm 2: vpand v01=%xmm13 vpand % xmm5, % xmm13, % xmm13 # qhasm: v11 = r1 & mask5 # asm 1: vpand v11=reg128#15 # asm 2: vpand v11=%xmm14 vpand % xmm5, % xmm14, % xmm14 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r0=reg128#10 # asm 2: vpor r0=%xmm9 vpor % xmm15, % xmm9, % xmm9 # qhasm: r1 = v01 | v11 # asm 1: vpor r1=reg128#14 # asm 2: vpor r1=%xmm13 vpor % xmm14, % xmm13, % xmm13 # qhasm: v00 = r2 & mask4 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm4, % xmm11, % xmm14 # qhasm: v10 = r3 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm10, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#12 # asm 2: vpand v01=%xmm11 vpand % xmm5, % xmm11, % xmm11 # qhasm: v11 = r3 & mask5 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm5, % xmm10, % xmm10 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r2=reg128#15 # asm 2: vpor r2=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm10, % xmm11, % xmm10 # qhasm: v00 = r4 & mask4 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm4, % xmm12, % xmm11 # qhasm: v10 = r5 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm8, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#13 # asm 2: vpand v01=%xmm12 vpand % xmm5, % xmm12, % xmm12 # qhasm: v11 = r5 & mask5 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm5, % xmm8, % xmm8 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r4=reg128#12 # asm 2: vpor r4=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm8, % xmm12, % xmm8 # qhasm: v00 = r6 & mask4 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm4, % xmm6, % xmm12 # qhasm: v10 = r7 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm7, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm5, % xmm6, % xmm6 # qhasm: v11 = r7 & mask5 # asm 1: vpand v11=reg128#8 # asm 2: vpand v11=%xmm7 vpand % xmm5, % xmm7, % xmm7 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r6=reg128#13 # asm 2: vpor r6=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#7 # asm 2: vpor r7=%xmm6 vpor % xmm7, % xmm6, % xmm6 # qhasm: t0 = r0[0]r1[0] # asm 1: vpunpcklqdq t0=reg128#8 # asm 2: vpunpcklqdq t0=%xmm7 vpunpcklqdq % xmm13, % xmm9, % xmm7 # qhasm: mem128[ input_0 + 192 ] = t0 # asm 1: movdqu t0=reg128#8 # asm 2: vpunpcklqdq t0=%xmm7 vpunpcklqdq % xmm10, % xmm14, % xmm7 # qhasm: mem128[ input_0 + 208 ] = t0 # asm 1: movdqu t0=reg128#8 # asm 2: vpunpcklqdq t0=%xmm7 vpunpcklqdq % xmm8, % xmm11, % xmm7 # qhasm: mem128[ input_0 + 224 ] = t0 # asm 1: movdqu t0=reg128#7 # asm 2: vpunpcklqdq t0=%xmm6 vpunpcklqdq % xmm6, % xmm12, % xmm6 # qhasm: mem128[ input_0 + 240 ] = t0 # asm 1: movdqu r0=reg128#7 # asm 2: movddup 256(r0=%xmm6 movddup 256( % rdi), % xmm6 # qhasm: r1 = mem64[ input_0 + 264 ] x2 # asm 1: movddup 264(r1=reg128#8 # asm 2: movddup 264(r1=%xmm7 movddup 264( % rdi), % xmm7 # qhasm: r2 = mem64[ input_0 + 272 ] x2 # asm 1: movddup 272(r2=reg128#9 # asm 2: movddup 272(r2=%xmm8 movddup 272( % rdi), % xmm8 # qhasm: r3 = mem64[ input_0 + 280 ] x2 # asm 1: movddup 280(r3=reg128#10 # asm 2: movddup 280(r3=%xmm9 movddup 280( % rdi), % xmm9 # qhasm: r4 = mem64[ input_0 + 288 ] x2 # asm 1: movddup 288(r4=reg128#11 # asm 2: movddup 288(r4=%xmm10 movddup 288( % rdi), % xmm10 # qhasm: r5 = mem64[ input_0 + 296 ] x2 # asm 1: movddup 296(r5=reg128#12 # asm 2: movddup 296(r5=%xmm11 movddup 296( % rdi), % xmm11 # qhasm: r6 = mem64[ input_0 + 304 ] x2 # asm 1: movddup 304(r6=reg128#13 # asm 2: movddup 304(r6=%xmm12 movddup 304( % rdi), % xmm12 # qhasm: r7 = mem64[ input_0 + 312 ] x2 # asm 1: movddup 312(r7=reg128#14 # asm 2: movddup 312(r7=%xmm13 movddup 312( % rdi), % xmm13 # qhasm: v00 = r0 & mask0 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm0, % xmm6, % xmm14 # qhasm: v10 = r4 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm10, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm1, % xmm6, % xmm6 # qhasm: v11 = r4 & mask1 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm1, % xmm10, % xmm10 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r0=reg128#15 # asm 2: vpor r0=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r4 = v01 | v11 # asm 1: vpor r4=reg128#7 # asm 2: vpor r4=%xmm6 vpor % xmm10, % xmm6, % xmm6 # qhasm: v00 = r1 & mask0 # asm 1: vpand v00=reg128#11 # asm 2: vpand v00=%xmm10 vpand % xmm0, % xmm7, % xmm10 # qhasm: v10 = r5 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm11, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#8 # asm 2: vpand v01=%xmm7 vpand % xmm1, % xmm7, % xmm7 # qhasm: v11 = r5 & mask1 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm1, % xmm11, % xmm11 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r1=reg128#11 # asm 2: vpor r1=%xmm10 vpor % xmm15, % xmm10, % xmm10 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#8 # asm 2: vpor r5=%xmm7 vpor % xmm11, % xmm7, % xmm7 # qhasm: v00 = r2 & mask0 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm0, % xmm8, % xmm11 # qhasm: v10 = r6 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm12, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#9 # asm 2: vpand v01=%xmm8 vpand % xmm1, % xmm8, % xmm8 # qhasm: v11 = r6 & mask1 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm1, % xmm12, % xmm12 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#9 # asm 2: vpor r6=%xmm8 vpor % xmm12, % xmm8, % xmm8 # qhasm: v00 = r3 & mask0 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm0, % xmm9, % xmm12 # qhasm: v10 = r7 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm13, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#10 # asm 2: vpand v01=%xmm9 vpand % xmm1, % xmm9, % xmm9 # qhasm: v11 = r7 & mask1 # asm 1: vpand v11=reg128#14 # asm 2: vpand v11=%xmm13 vpand % xmm1, % xmm13, % xmm13 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r3=reg128#13 # asm 2: vpor r3=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#10 # asm 2: vpor r7=%xmm9 vpor % xmm13, % xmm9, % xmm9 # qhasm: v00 = r0 & mask2 # asm 1: vpand v00=reg128#14 # asm 2: vpand v00=%xmm13 vpand % xmm2, % xmm14, % xmm13 # qhasm: v10 = r2 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm11, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#15 # asm 2: vpand v01=%xmm14 vpand % xmm3, % xmm14, % xmm14 # qhasm: v11 = r2 & mask3 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm3, % xmm11, % xmm11 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r0=reg128#14 # asm 2: vpor r0=%xmm13 vpor % xmm15, % xmm13, % xmm13 # qhasm: r2 = v01 | v11 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm11, % xmm14, % xmm11 # qhasm: v00 = r1 & mask2 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm2, % xmm10, % xmm14 # qhasm: v10 = r3 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm12, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#11 # asm 2: vpand v01=%xmm10 vpand % xmm3, % xmm10, % xmm10 # qhasm: v11 = r3 & mask3 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm3, % xmm12, % xmm12 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r1=reg128#15 # asm 2: vpor r1=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm12, % xmm10, % xmm10 # qhasm: v00 = r4 & mask2 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm2, % xmm6, % xmm12 # qhasm: v10 = r6 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm8, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm3, % xmm6, % xmm6 # qhasm: v11 = r6 & mask3 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm3, % xmm8, % xmm8 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r4=reg128#13 # asm 2: vpor r4=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#7 # asm 2: vpor r6=%xmm6 vpor % xmm8, % xmm6, % xmm6 # qhasm: v00 = r5 & mask2 # asm 1: vpand v00=reg128#9 # asm 2: vpand v00=%xmm8 vpand % xmm2, % xmm7, % xmm8 # qhasm: v10 = r7 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm9, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#8 # asm 2: vpand v01=%xmm7 vpand % xmm3, % xmm7, % xmm7 # qhasm: v11 = r7 & mask3 # asm 1: vpand v11=reg128#10 # asm 2: vpand v11=%xmm9 vpand % xmm3, % xmm9, % xmm9 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm15, % xmm8, % xmm8 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#8 # asm 2: vpor r7=%xmm7 vpor % xmm9, % xmm7, % xmm7 # qhasm: v00 = r0 & mask4 # asm 1: vpand v00=reg128#10 # asm 2: vpand v00=%xmm9 vpand % xmm4, % xmm13, % xmm9 # qhasm: v10 = r1 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm14, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#14 # asm 2: vpand v01=%xmm13 vpand % xmm5, % xmm13, % xmm13 # qhasm: v11 = r1 & mask5 # asm 1: vpand v11=reg128#15 # asm 2: vpand v11=%xmm14 vpand % xmm5, % xmm14, % xmm14 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r0=reg128#10 # asm 2: vpor r0=%xmm9 vpor % xmm15, % xmm9, % xmm9 # qhasm: r1 = v01 | v11 # asm 1: vpor r1=reg128#14 # asm 2: vpor r1=%xmm13 vpor % xmm14, % xmm13, % xmm13 # qhasm: v00 = r2 & mask4 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm4, % xmm11, % xmm14 # qhasm: v10 = r3 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm10, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#12 # asm 2: vpand v01=%xmm11 vpand % xmm5, % xmm11, % xmm11 # qhasm: v11 = r3 & mask5 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm5, % xmm10, % xmm10 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r2=reg128#15 # asm 2: vpor r2=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm10, % xmm11, % xmm10 # qhasm: v00 = r4 & mask4 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm4, % xmm12, % xmm11 # qhasm: v10 = r5 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm8, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#13 # asm 2: vpand v01=%xmm12 vpand % xmm5, % xmm12, % xmm12 # qhasm: v11 = r5 & mask5 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm5, % xmm8, % xmm8 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r4=reg128#12 # asm 2: vpor r4=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm8, % xmm12, % xmm8 # qhasm: v00 = r6 & mask4 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm4, % xmm6, % xmm12 # qhasm: v10 = r7 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm7, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm5, % xmm6, % xmm6 # qhasm: v11 = r7 & mask5 # asm 1: vpand v11=reg128#8 # asm 2: vpand v11=%xmm7 vpand % xmm5, % xmm7, % xmm7 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r6=reg128#13 # asm 2: vpor r6=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#7 # asm 2: vpor r7=%xmm6 vpor % xmm7, % xmm6, % xmm6 # qhasm: t0 = r0[0]r1[0] # asm 1: vpunpcklqdq t0=reg128#8 # asm 2: vpunpcklqdq t0=%xmm7 vpunpcklqdq % xmm13, % xmm9, % xmm7 # qhasm: mem128[ input_0 + 256 ] = t0 # asm 1: movdqu t0=reg128#8 # asm 2: vpunpcklqdq t0=%xmm7 vpunpcklqdq % xmm10, % xmm14, % xmm7 # qhasm: mem128[ input_0 + 272 ] = t0 # asm 1: movdqu t0=reg128#8 # asm 2: vpunpcklqdq t0=%xmm7 vpunpcklqdq % xmm8, % xmm11, % xmm7 # qhasm: mem128[ input_0 + 288 ] = t0 # asm 1: movdqu t0=reg128#7 # asm 2: vpunpcklqdq t0=%xmm6 vpunpcklqdq % xmm6, % xmm12, % xmm6 # qhasm: mem128[ input_0 + 304 ] = t0 # asm 1: movdqu r0=reg128#7 # asm 2: movddup 320(r0=%xmm6 movddup 320( % rdi), % xmm6 # qhasm: r1 = mem64[ input_0 + 328 ] x2 # asm 1: movddup 328(r1=reg128#8 # asm 2: movddup 328(r1=%xmm7 movddup 328( % rdi), % xmm7 # qhasm: r2 = mem64[ input_0 + 336 ] x2 # asm 1: movddup 336(r2=reg128#9 # asm 2: movddup 336(r2=%xmm8 movddup 336( % rdi), % xmm8 # qhasm: r3 = mem64[ input_0 + 344 ] x2 # asm 1: movddup 344(r3=reg128#10 # asm 2: movddup 344(r3=%xmm9 movddup 344( % rdi), % xmm9 # qhasm: r4 = mem64[ input_0 + 352 ] x2 # asm 1: movddup 352(r4=reg128#11 # asm 2: movddup 352(r4=%xmm10 movddup 352( % rdi), % xmm10 # qhasm: r5 = mem64[ input_0 + 360 ] x2 # asm 1: movddup 360(r5=reg128#12 # asm 2: movddup 360(r5=%xmm11 movddup 360( % rdi), % xmm11 # qhasm: r6 = mem64[ input_0 + 368 ] x2 # asm 1: movddup 368(r6=reg128#13 # asm 2: movddup 368(r6=%xmm12 movddup 368( % rdi), % xmm12 # qhasm: r7 = mem64[ input_0 + 376 ] x2 # asm 1: movddup 376(r7=reg128#14 # asm 2: movddup 376(r7=%xmm13 movddup 376( % rdi), % xmm13 # qhasm: v00 = r0 & mask0 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm0, % xmm6, % xmm14 # qhasm: v10 = r4 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm10, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm1, % xmm6, % xmm6 # qhasm: v11 = r4 & mask1 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm1, % xmm10, % xmm10 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r0=reg128#15 # asm 2: vpor r0=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r4 = v01 | v11 # asm 1: vpor r4=reg128#7 # asm 2: vpor r4=%xmm6 vpor % xmm10, % xmm6, % xmm6 # qhasm: v00 = r1 & mask0 # asm 1: vpand v00=reg128#11 # asm 2: vpand v00=%xmm10 vpand % xmm0, % xmm7, % xmm10 # qhasm: v10 = r5 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm11, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#8 # asm 2: vpand v01=%xmm7 vpand % xmm1, % xmm7, % xmm7 # qhasm: v11 = r5 & mask1 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm1, % xmm11, % xmm11 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r1=reg128#11 # asm 2: vpor r1=%xmm10 vpor % xmm15, % xmm10, % xmm10 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#8 # asm 2: vpor r5=%xmm7 vpor % xmm11, % xmm7, % xmm7 # qhasm: v00 = r2 & mask0 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm0, % xmm8, % xmm11 # qhasm: v10 = r6 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm12, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#9 # asm 2: vpand v01=%xmm8 vpand % xmm1, % xmm8, % xmm8 # qhasm: v11 = r6 & mask1 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm1, % xmm12, % xmm12 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#9 # asm 2: vpor r6=%xmm8 vpor % xmm12, % xmm8, % xmm8 # qhasm: v00 = r3 & mask0 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm0, % xmm9, % xmm12 # qhasm: v10 = r7 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm13, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#10 # asm 2: vpand v01=%xmm9 vpand % xmm1, % xmm9, % xmm9 # qhasm: v11 = r7 & mask1 # asm 1: vpand v11=reg128#14 # asm 2: vpand v11=%xmm13 vpand % xmm1, % xmm13, % xmm13 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r3=reg128#13 # asm 2: vpor r3=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#10 # asm 2: vpor r7=%xmm9 vpor % xmm13, % xmm9, % xmm9 # qhasm: v00 = r0 & mask2 # asm 1: vpand v00=reg128#14 # asm 2: vpand v00=%xmm13 vpand % xmm2, % xmm14, % xmm13 # qhasm: v10 = r2 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm11, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#15 # asm 2: vpand v01=%xmm14 vpand % xmm3, % xmm14, % xmm14 # qhasm: v11 = r2 & mask3 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm3, % xmm11, % xmm11 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r0=reg128#14 # asm 2: vpor r0=%xmm13 vpor % xmm15, % xmm13, % xmm13 # qhasm: r2 = v01 | v11 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm11, % xmm14, % xmm11 # qhasm: v00 = r1 & mask2 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm2, % xmm10, % xmm14 # qhasm: v10 = r3 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm12, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#11 # asm 2: vpand v01=%xmm10 vpand % xmm3, % xmm10, % xmm10 # qhasm: v11 = r3 & mask3 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm3, % xmm12, % xmm12 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r1=reg128#15 # asm 2: vpor r1=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm12, % xmm10, % xmm10 # qhasm: v00 = r4 & mask2 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm2, % xmm6, % xmm12 # qhasm: v10 = r6 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm8, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm3, % xmm6, % xmm6 # qhasm: v11 = r6 & mask3 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm3, % xmm8, % xmm8 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r4=reg128#13 # asm 2: vpor r4=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#7 # asm 2: vpor r6=%xmm6 vpor % xmm8, % xmm6, % xmm6 # qhasm: v00 = r5 & mask2 # asm 1: vpand v00=reg128#9 # asm 2: vpand v00=%xmm8 vpand % xmm2, % xmm7, % xmm8 # qhasm: v10 = r7 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm9, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#8 # asm 2: vpand v01=%xmm7 vpand % xmm3, % xmm7, % xmm7 # qhasm: v11 = r7 & mask3 # asm 1: vpand v11=reg128#10 # asm 2: vpand v11=%xmm9 vpand % xmm3, % xmm9, % xmm9 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm15, % xmm8, % xmm8 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#8 # asm 2: vpor r7=%xmm7 vpor % xmm9, % xmm7, % xmm7 # qhasm: v00 = r0 & mask4 # asm 1: vpand v00=reg128#10 # asm 2: vpand v00=%xmm9 vpand % xmm4, % xmm13, % xmm9 # qhasm: v10 = r1 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm14, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#14 # asm 2: vpand v01=%xmm13 vpand % xmm5, % xmm13, % xmm13 # qhasm: v11 = r1 & mask5 # asm 1: vpand v11=reg128#15 # asm 2: vpand v11=%xmm14 vpand % xmm5, % xmm14, % xmm14 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r0=reg128#10 # asm 2: vpor r0=%xmm9 vpor % xmm15, % xmm9, % xmm9 # qhasm: r1 = v01 | v11 # asm 1: vpor r1=reg128#14 # asm 2: vpor r1=%xmm13 vpor % xmm14, % xmm13, % xmm13 # qhasm: v00 = r2 & mask4 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm4, % xmm11, % xmm14 # qhasm: v10 = r3 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm10, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#12 # asm 2: vpand v01=%xmm11 vpand % xmm5, % xmm11, % xmm11 # qhasm: v11 = r3 & mask5 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm5, % xmm10, % xmm10 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r2=reg128#15 # asm 2: vpor r2=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm10, % xmm11, % xmm10 # qhasm: v00 = r4 & mask4 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm4, % xmm12, % xmm11 # qhasm: v10 = r5 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm8, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#13 # asm 2: vpand v01=%xmm12 vpand % xmm5, % xmm12, % xmm12 # qhasm: v11 = r5 & mask5 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm5, % xmm8, % xmm8 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r4=reg128#12 # asm 2: vpor r4=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm8, % xmm12, % xmm8 # qhasm: v00 = r6 & mask4 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm4, % xmm6, % xmm12 # qhasm: v10 = r7 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm7, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm5, % xmm6, % xmm6 # qhasm: v11 = r7 & mask5 # asm 1: vpand v11=reg128#8 # asm 2: vpand v11=%xmm7 vpand % xmm5, % xmm7, % xmm7 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r6=reg128#13 # asm 2: vpor r6=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#7 # asm 2: vpor r7=%xmm6 vpor % xmm7, % xmm6, % xmm6 # qhasm: t0 = r0[0]r1[0] # asm 1: vpunpcklqdq t0=reg128#8 # asm 2: vpunpcklqdq t0=%xmm7 vpunpcklqdq % xmm13, % xmm9, % xmm7 # qhasm: mem128[ input_0 + 320 ] = t0 # asm 1: movdqu t0=reg128#8 # asm 2: vpunpcklqdq t0=%xmm7 vpunpcklqdq % xmm10, % xmm14, % xmm7 # qhasm: mem128[ input_0 + 336 ] = t0 # asm 1: movdqu t0=reg128#8 # asm 2: vpunpcklqdq t0=%xmm7 vpunpcklqdq % xmm8, % xmm11, % xmm7 # qhasm: mem128[ input_0 + 352 ] = t0 # asm 1: movdqu t0=reg128#7 # asm 2: vpunpcklqdq t0=%xmm6 vpunpcklqdq % xmm6, % xmm12, % xmm6 # qhasm: mem128[ input_0 + 368 ] = t0 # asm 1: movdqu r0=reg128#7 # asm 2: movddup 384(r0=%xmm6 movddup 384( % rdi), % xmm6 # qhasm: r1 = mem64[ input_0 + 392 ] x2 # asm 1: movddup 392(r1=reg128#8 # asm 2: movddup 392(r1=%xmm7 movddup 392( % rdi), % xmm7 # qhasm: r2 = mem64[ input_0 + 400 ] x2 # asm 1: movddup 400(r2=reg128#9 # asm 2: movddup 400(r2=%xmm8 movddup 400( % rdi), % xmm8 # qhasm: r3 = mem64[ input_0 + 408 ] x2 # asm 1: movddup 408(r3=reg128#10 # asm 2: movddup 408(r3=%xmm9 movddup 408( % rdi), % xmm9 # qhasm: r4 = mem64[ input_0 + 416 ] x2 # asm 1: movddup 416(r4=reg128#11 # asm 2: movddup 416(r4=%xmm10 movddup 416( % rdi), % xmm10 # qhasm: r5 = mem64[ input_0 + 424 ] x2 # asm 1: movddup 424(r5=reg128#12 # asm 2: movddup 424(r5=%xmm11 movddup 424( % rdi), % xmm11 # qhasm: r6 = mem64[ input_0 + 432 ] x2 # asm 1: movddup 432(r6=reg128#13 # asm 2: movddup 432(r6=%xmm12 movddup 432( % rdi), % xmm12 # qhasm: r7 = mem64[ input_0 + 440 ] x2 # asm 1: movddup 440(r7=reg128#14 # asm 2: movddup 440(r7=%xmm13 movddup 440( % rdi), % xmm13 # qhasm: v00 = r0 & mask0 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm0, % xmm6, % xmm14 # qhasm: v10 = r4 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm10, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm1, % xmm6, % xmm6 # qhasm: v11 = r4 & mask1 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm1, % xmm10, % xmm10 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r0=reg128#15 # asm 2: vpor r0=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r4 = v01 | v11 # asm 1: vpor r4=reg128#7 # asm 2: vpor r4=%xmm6 vpor % xmm10, % xmm6, % xmm6 # qhasm: v00 = r1 & mask0 # asm 1: vpand v00=reg128#11 # asm 2: vpand v00=%xmm10 vpand % xmm0, % xmm7, % xmm10 # qhasm: v10 = r5 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm11, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#8 # asm 2: vpand v01=%xmm7 vpand % xmm1, % xmm7, % xmm7 # qhasm: v11 = r5 & mask1 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm1, % xmm11, % xmm11 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r1=reg128#11 # asm 2: vpor r1=%xmm10 vpor % xmm15, % xmm10, % xmm10 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#8 # asm 2: vpor r5=%xmm7 vpor % xmm11, % xmm7, % xmm7 # qhasm: v00 = r2 & mask0 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm0, % xmm8, % xmm11 # qhasm: v10 = r6 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm12, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#9 # asm 2: vpand v01=%xmm8 vpand % xmm1, % xmm8, % xmm8 # qhasm: v11 = r6 & mask1 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm1, % xmm12, % xmm12 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#9 # asm 2: vpor r6=%xmm8 vpor % xmm12, % xmm8, % xmm8 # qhasm: v00 = r3 & mask0 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm0, % xmm9, % xmm12 # qhasm: v10 = r7 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm13, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#10 # asm 2: vpand v01=%xmm9 vpand % xmm1, % xmm9, % xmm9 # qhasm: v11 = r7 & mask1 # asm 1: vpand v11=reg128#14 # asm 2: vpand v11=%xmm13 vpand % xmm1, % xmm13, % xmm13 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r3=reg128#13 # asm 2: vpor r3=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#10 # asm 2: vpor r7=%xmm9 vpor % xmm13, % xmm9, % xmm9 # qhasm: v00 = r0 & mask2 # asm 1: vpand v00=reg128#14 # asm 2: vpand v00=%xmm13 vpand % xmm2, % xmm14, % xmm13 # qhasm: v10 = r2 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm11, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#15 # asm 2: vpand v01=%xmm14 vpand % xmm3, % xmm14, % xmm14 # qhasm: v11 = r2 & mask3 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm3, % xmm11, % xmm11 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r0=reg128#14 # asm 2: vpor r0=%xmm13 vpor % xmm15, % xmm13, % xmm13 # qhasm: r2 = v01 | v11 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm11, % xmm14, % xmm11 # qhasm: v00 = r1 & mask2 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm2, % xmm10, % xmm14 # qhasm: v10 = r3 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm12, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#11 # asm 2: vpand v01=%xmm10 vpand % xmm3, % xmm10, % xmm10 # qhasm: v11 = r3 & mask3 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm3, % xmm12, % xmm12 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r1=reg128#15 # asm 2: vpor r1=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm12, % xmm10, % xmm10 # qhasm: v00 = r4 & mask2 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm2, % xmm6, % xmm12 # qhasm: v10 = r6 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm8, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm3, % xmm6, % xmm6 # qhasm: v11 = r6 & mask3 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm3, % xmm8, % xmm8 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r4=reg128#13 # asm 2: vpor r4=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#7 # asm 2: vpor r6=%xmm6 vpor % xmm8, % xmm6, % xmm6 # qhasm: v00 = r5 & mask2 # asm 1: vpand v00=reg128#9 # asm 2: vpand v00=%xmm8 vpand % xmm2, % xmm7, % xmm8 # qhasm: v10 = r7 & mask2 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm2, % xmm9, % xmm15 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#8 # asm 2: vpand v01=%xmm7 vpand % xmm3, % xmm7, % xmm7 # qhasm: v11 = r7 & mask3 # asm 1: vpand v11=reg128#10 # asm 2: vpand v11=%xmm9 vpand % xmm3, % xmm9, % xmm9 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm15, % xmm8, % xmm8 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#8 # asm 2: vpor r7=%xmm7 vpor % xmm9, % xmm7, % xmm7 # qhasm: v00 = r0 & mask4 # asm 1: vpand v00=reg128#10 # asm 2: vpand v00=%xmm9 vpand % xmm4, % xmm13, % xmm9 # qhasm: v10 = r1 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm14, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#14 # asm 2: vpand v01=%xmm13 vpand % xmm5, % xmm13, % xmm13 # qhasm: v11 = r1 & mask5 # asm 1: vpand v11=reg128#15 # asm 2: vpand v11=%xmm14 vpand % xmm5, % xmm14, % xmm14 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r0=reg128#10 # asm 2: vpor r0=%xmm9 vpor % xmm15, % xmm9, % xmm9 # qhasm: r1 = v01 | v11 # asm 1: vpor r1=reg128#14 # asm 2: vpor r1=%xmm13 vpor % xmm14, % xmm13, % xmm13 # qhasm: v00 = r2 & mask4 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm4, % xmm11, % xmm14 # qhasm: v10 = r3 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm10, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#12 # asm 2: vpand v01=%xmm11 vpand % xmm5, % xmm11, % xmm11 # qhasm: v11 = r3 & mask5 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm5, % xmm10, % xmm10 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r2=reg128#15 # asm 2: vpor r2=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#11 # asm 2: vpor r3=%xmm10 vpor % xmm10, % xmm11, % xmm10 # qhasm: v00 = r4 & mask4 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm4, % xmm12, % xmm11 # qhasm: v10 = r5 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm8, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#13 # asm 2: vpand v01=%xmm12 vpand % xmm5, % xmm12, % xmm12 # qhasm: v11 = r5 & mask5 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm5, % xmm8, % xmm8 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r4=reg128#12 # asm 2: vpor r4=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#9 # asm 2: vpor r5=%xmm8 vpor % xmm8, % xmm12, % xmm8 # qhasm: v00 = r6 & mask4 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm4, % xmm6, % xmm12 # qhasm: v10 = r7 & mask4 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm4, % xmm7, % xmm15 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm5, % xmm6, % xmm6 # qhasm: v11 = r7 & mask5 # asm 1: vpand v11=reg128#8 # asm 2: vpand v11=%xmm7 vpand % xmm5, % xmm7, % xmm7 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r6=reg128#13 # asm 2: vpor r6=%xmm12 vpor % xmm15, % xmm12, % xmm12 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#7 # asm 2: vpor r7=%xmm6 vpor % xmm7, % xmm6, % xmm6 # qhasm: t0 = r0[0]r1[0] # asm 1: vpunpcklqdq t0=reg128#8 # asm 2: vpunpcklqdq t0=%xmm7 vpunpcklqdq % xmm13, % xmm9, % xmm7 # qhasm: mem128[ input_0 + 384 ] = t0 # asm 1: movdqu t0=reg128#8 # asm 2: vpunpcklqdq t0=%xmm7 vpunpcklqdq % xmm10, % xmm14, % xmm7 # qhasm: mem128[ input_0 + 400 ] = t0 # asm 1: movdqu t0=reg128#8 # asm 2: vpunpcklqdq t0=%xmm7 vpunpcklqdq % xmm8, % xmm11, % xmm7 # qhasm: mem128[ input_0 + 416 ] = t0 # asm 1: movdqu t0=reg128#7 # asm 2: vpunpcklqdq t0=%xmm6 vpunpcklqdq % xmm6, % xmm12, % xmm6 # qhasm: mem128[ input_0 + 432 ] = t0 # asm 1: movdqu r0=reg128#7 # asm 2: movddup 448(r0=%xmm6 movddup 448( % rdi), % xmm6 # qhasm: r1 = mem64[ input_0 + 456 ] x2 # asm 1: movddup 456(r1=reg128#8 # asm 2: movddup 456(r1=%xmm7 movddup 456( % rdi), % xmm7 # qhasm: r2 = mem64[ input_0 + 464 ] x2 # asm 1: movddup 464(r2=reg128#9 # asm 2: movddup 464(r2=%xmm8 movddup 464( % rdi), % xmm8 # qhasm: r3 = mem64[ input_0 + 472 ] x2 # asm 1: movddup 472(r3=reg128#10 # asm 2: movddup 472(r3=%xmm9 movddup 472( % rdi), % xmm9 # qhasm: r4 = mem64[ input_0 + 480 ] x2 # asm 1: movddup 480(r4=reg128#11 # asm 2: movddup 480(r4=%xmm10 movddup 480( % rdi), % xmm10 # qhasm: r5 = mem64[ input_0 + 488 ] x2 # asm 1: movddup 488(r5=reg128#12 # asm 2: movddup 488(r5=%xmm11 movddup 488( % rdi), % xmm11 # qhasm: r6 = mem64[ input_0 + 496 ] x2 # asm 1: movddup 496(r6=reg128#13 # asm 2: movddup 496(r6=%xmm12 movddup 496( % rdi), % xmm12 # qhasm: r7 = mem64[ input_0 + 504 ] x2 # asm 1: movddup 504(r7=reg128#14 # asm 2: movddup 504(r7=%xmm13 movddup 504( % rdi), % xmm13 # qhasm: v00 = r0 & mask0 # asm 1: vpand v00=reg128#15 # asm 2: vpand v00=%xmm14 vpand % xmm0, % xmm6, % xmm14 # qhasm: v10 = r4 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm10, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm1, % xmm6, % xmm6 # qhasm: v11 = r4 & mask1 # asm 1: vpand v11=reg128#11 # asm 2: vpand v11=%xmm10 vpand % xmm1, % xmm10, % xmm10 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r0=reg128#15 # asm 2: vpor r0=%xmm14 vpor % xmm15, % xmm14, % xmm14 # qhasm: r4 = v01 | v11 # asm 1: vpor r4=reg128#7 # asm 2: vpor r4=%xmm6 vpor % xmm10, % xmm6, % xmm6 # qhasm: v00 = r1 & mask0 # asm 1: vpand v00=reg128#11 # asm 2: vpand v00=%xmm10 vpand % xmm0, % xmm7, % xmm10 # qhasm: v10 = r5 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm11, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#8 # asm 2: vpand v01=%xmm7 vpand % xmm1, % xmm7, % xmm7 # qhasm: v11 = r5 & mask1 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm1, % xmm11, % xmm11 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r1=reg128#11 # asm 2: vpor r1=%xmm10 vpor % xmm15, % xmm10, % xmm10 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#8 # asm 2: vpor r5=%xmm7 vpor % xmm11, % xmm7, % xmm7 # qhasm: v00 = r2 & mask0 # asm 1: vpand v00=reg128#12 # asm 2: vpand v00=%xmm11 vpand % xmm0, % xmm8, % xmm11 # qhasm: v10 = r6 & mask0 # asm 1: vpand v10=reg128#16 # asm 2: vpand v10=%xmm15 vpand % xmm0, % xmm12, % xmm15 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#9 # asm 2: vpand v01=%xmm8 vpand % xmm1, % xmm8, % xmm8 # qhasm: v11 = r6 & mask1 # asm 1: vpand v11=reg128#13 # asm 2: vpand v11=%xmm12 vpand % xmm1, % xmm12, % xmm12 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm15, % xmm11, % xmm11 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#9 # asm 2: vpor r6=%xmm8 vpor % xmm12, % xmm8, % xmm8 # qhasm: v00 = r3 & mask0 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm0, % xmm9, % xmm12 # qhasm: v10 = r7 & mask0 # asm 1: vpand v10=reg128#1 # asm 2: vpand v10=%xmm0 vpand % xmm0, % xmm13, % xmm0 # qhasm: 2x v10 <<= 4 # asm 1: psllq $4,v01=reg128#10 # asm 2: vpand v01=%xmm9 vpand % xmm1, % xmm9, % xmm9 # qhasm: v11 = r7 & mask1 # asm 1: vpand v11=reg128#2 # asm 2: vpand v11=%xmm1 vpand % xmm1, % xmm13, % xmm1 # qhasm: 2x v01 unsigned>>= 4 # asm 1: psrlq $4,r3=reg128#1 # asm 2: vpor r3=%xmm0 vpor % xmm0, % xmm12, % xmm0 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#2 # asm 2: vpor r7=%xmm1 vpor % xmm1, % xmm9, % xmm1 # qhasm: v00 = r0 & mask2 # asm 1: vpand v00=reg128#10 # asm 2: vpand v00=%xmm9 vpand % xmm2, % xmm14, % xmm9 # qhasm: v10 = r2 & mask2 # asm 1: vpand v10=reg128#13 # asm 2: vpand v10=%xmm12 vpand % xmm2, % xmm11, % xmm12 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#14 # asm 2: vpand v01=%xmm13 vpand % xmm3, % xmm14, % xmm13 # qhasm: v11 = r2 & mask3 # asm 1: vpand v11=reg128#12 # asm 2: vpand v11=%xmm11 vpand % xmm3, % xmm11, % xmm11 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r0=reg128#10 # asm 2: vpor r0=%xmm9 vpor % xmm12, % xmm9, % xmm9 # qhasm: r2 = v01 | v11 # asm 1: vpor r2=reg128#12 # asm 2: vpor r2=%xmm11 vpor % xmm11, % xmm13, % xmm11 # qhasm: v00 = r1 & mask2 # asm 1: vpand v00=reg128#13 # asm 2: vpand v00=%xmm12 vpand % xmm2, % xmm10, % xmm12 # qhasm: v10 = r3 & mask2 # asm 1: vpand v10=reg128#14 # asm 2: vpand v10=%xmm13 vpand % xmm2, % xmm0, % xmm13 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#11 # asm 2: vpand v01=%xmm10 vpand % xmm3, % xmm10, % xmm10 # qhasm: v11 = r3 & mask3 # asm 1: vpand v11=reg128#1 # asm 2: vpand v11=%xmm0 vpand % xmm3, % xmm0, % xmm0 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r1=reg128#13 # asm 2: vpor r1=%xmm12 vpor % xmm13, % xmm12, % xmm12 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#1 # asm 2: vpor r3=%xmm0 vpor % xmm0, % xmm10, % xmm0 # qhasm: v00 = r4 & mask2 # asm 1: vpand v00=reg128#11 # asm 2: vpand v00=%xmm10 vpand % xmm2, % xmm6, % xmm10 # qhasm: v10 = r6 & mask2 # asm 1: vpand v10=reg128#14 # asm 2: vpand v10=%xmm13 vpand % xmm2, % xmm8, % xmm13 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm3, % xmm6, % xmm6 # qhasm: v11 = r6 & mask3 # asm 1: vpand v11=reg128#9 # asm 2: vpand v11=%xmm8 vpand % xmm3, % xmm8, % xmm8 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r4=reg128#11 # asm 2: vpor r4=%xmm10 vpor % xmm13, % xmm10, % xmm10 # qhasm: r6 = v01 | v11 # asm 1: vpor r6=reg128#7 # asm 2: vpor r6=%xmm6 vpor % xmm8, % xmm6, % xmm6 # qhasm: v00 = r5 & mask2 # asm 1: vpand v00=reg128#9 # asm 2: vpand v00=%xmm8 vpand % xmm2, % xmm7, % xmm8 # qhasm: v10 = r7 & mask2 # asm 1: vpand v10=reg128#3 # asm 2: vpand v10=%xmm2 vpand % xmm2, % xmm1, % xmm2 # qhasm: 2x v10 <<= 2 # asm 1: psllq $2,v01=reg128#8 # asm 2: vpand v01=%xmm7 vpand % xmm3, % xmm7, % xmm7 # qhasm: v11 = r7 & mask3 # asm 1: vpand v11=reg128#2 # asm 2: vpand v11=%xmm1 vpand % xmm3, % xmm1, % xmm1 # qhasm: 2x v01 unsigned>>= 2 # asm 1: psrlq $2,r5=reg128#3 # asm 2: vpor r5=%xmm2 vpor % xmm2, % xmm8, % xmm2 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#2 # asm 2: vpor r7=%xmm1 vpor % xmm1, % xmm7, % xmm1 # qhasm: v00 = r0 & mask4 # asm 1: vpand v00=reg128#4 # asm 2: vpand v00=%xmm3 vpand % xmm4, % xmm9, % xmm3 # qhasm: v10 = r1 & mask4 # asm 1: vpand v10=reg128#8 # asm 2: vpand v10=%xmm7 vpand % xmm4, % xmm12, % xmm7 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#9 # asm 2: vpand v01=%xmm8 vpand % xmm5, % xmm9, % xmm8 # qhasm: v11 = r1 & mask5 # asm 1: vpand v11=reg128#10 # asm 2: vpand v11=%xmm9 vpand % xmm5, % xmm12, % xmm9 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r0=reg128#4 # asm 2: vpor r0=%xmm3 vpor % xmm7, % xmm3, % xmm3 # qhasm: r1 = v01 | v11 # asm 1: vpor r1=reg128#8 # asm 2: vpor r1=%xmm7 vpor % xmm9, % xmm8, % xmm7 # qhasm: v00 = r2 & mask4 # asm 1: vpand v00=reg128#9 # asm 2: vpand v00=%xmm8 vpand % xmm4, % xmm11, % xmm8 # qhasm: v10 = r3 & mask4 # asm 1: vpand v10=reg128#10 # asm 2: vpand v10=%xmm9 vpand % xmm4, % xmm0, % xmm9 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#12 # asm 2: vpand v01=%xmm11 vpand % xmm5, % xmm11, % xmm11 # qhasm: v11 = r3 & mask5 # asm 1: vpand v11=reg128#1 # asm 2: vpand v11=%xmm0 vpand % xmm5, % xmm0, % xmm0 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r2=reg128#9 # asm 2: vpor r2=%xmm8 vpor % xmm9, % xmm8, % xmm8 # qhasm: r3 = v01 | v11 # asm 1: vpor r3=reg128#1 # asm 2: vpor r3=%xmm0 vpor % xmm0, % xmm11, % xmm0 # qhasm: v00 = r4 & mask4 # asm 1: vpand v00=reg128#10 # asm 2: vpand v00=%xmm9 vpand % xmm4, % xmm10, % xmm9 # qhasm: v10 = r5 & mask4 # asm 1: vpand v10=reg128#12 # asm 2: vpand v10=%xmm11 vpand % xmm4, % xmm2, % xmm11 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#11 # asm 2: vpand v01=%xmm10 vpand % xmm5, % xmm10, % xmm10 # qhasm: v11 = r5 & mask5 # asm 1: vpand v11=reg128#3 # asm 2: vpand v11=%xmm2 vpand % xmm5, % xmm2, % xmm2 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r4=reg128#10 # asm 2: vpor r4=%xmm9 vpor % xmm11, % xmm9, % xmm9 # qhasm: r5 = v01 | v11 # asm 1: vpor r5=reg128#3 # asm 2: vpor r5=%xmm2 vpor % xmm2, % xmm10, % xmm2 # qhasm: v00 = r6 & mask4 # asm 1: vpand v00=reg128#11 # asm 2: vpand v00=%xmm10 vpand % xmm4, % xmm6, % xmm10 # qhasm: v10 = r7 & mask4 # asm 1: vpand v10=reg128#5 # asm 2: vpand v10=%xmm4 vpand % xmm4, % xmm1, % xmm4 # qhasm: 2x v10 <<= 1 # asm 1: psllq $1,v01=reg128#7 # asm 2: vpand v01=%xmm6 vpand % xmm5, % xmm6, % xmm6 # qhasm: v11 = r7 & mask5 # asm 1: vpand v11=reg128#2 # asm 2: vpand v11=%xmm1 vpand % xmm5, % xmm1, % xmm1 # qhasm: 2x v01 unsigned>>= 1 # asm 1: psrlq $1,r6=reg128#5 # asm 2: vpor r6=%xmm4 vpor % xmm4, % xmm10, % xmm4 # qhasm: r7 = v01 | v11 # asm 1: vpor r7=reg128#2 # asm 2: vpor r7=%xmm1 vpor % xmm1, % xmm6, % xmm1 # qhasm: t0 = r0[0]r1[0] # asm 1: vpunpcklqdq t0=reg128#4 # asm 2: vpunpcklqdq t0=%xmm3 vpunpcklqdq % xmm7, % xmm3, % xmm3 # qhasm: mem128[ input_0 + 448 ] = t0 # asm 1: movdqu t0=reg128#1 # asm 2: vpunpcklqdq t0=%xmm0 vpunpcklqdq % xmm0, % xmm8, % xmm0 # qhasm: mem128[ input_0 + 464 ] = t0 # asm 1: movdqu t0=reg128#1 # asm 2: vpunpcklqdq t0=%xmm0 vpunpcklqdq % xmm2, % xmm9, % xmm0 # qhasm: mem128[ input_0 + 480 ] = t0 # asm 1: movdqu t0=reg128#1 # asm 2: vpunpcklqdq t0=%xmm0 vpunpcklqdq % xmm1, % xmm4, % xmm0 # qhasm: mem128[ input_0 + 496 ] = t0 # asm 1: movdqu