# Check 64bit AVX512{VBMI2,VL} instructions
.allow_index_reg
.text
_start:
vpcompressb %xmm30, (%rcx){%k7} # AVX512{VBMI2,VL}
vpcompressb %xmm30, 0x123(%rax,%r14,8) # AVX512{VBMI2,VL}
vpcompressb %xmm30, 127(%rdx) # AVX512{VBMI2,VL} Disp8
vpcompressb %ymm30, (%rcx){%k7} # AVX512{VBMI2,VL}
vpcompressb %ymm30, 0x123(%rax,%r14,8) # AVX512{VBMI2,VL}
vpcompressb %ymm30, 127(%rdx) # AVX512{VBMI2,VL} Disp8
vpcompressb %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpcompressb %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL}
vpcompressb %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL}
vpcompressb %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpcompressb %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL}
vpcompressb %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL}
vpcompressw %xmm30, (%rcx){%k7} # AVX512{VBMI2,VL}
vpcompressw %xmm30, 0x123(%rax,%r14,8) # AVX512{VBMI2,VL}
vpcompressw %xmm30, 254(%rdx) # AVX512{VBMI2,VL} Disp8
vpcompressw %ymm30, (%rcx){%k7} # AVX512{VBMI2,VL}
vpcompressw %ymm30, 0x123(%rax,%r14,8) # AVX512{VBMI2,VL}
vpcompressw %ymm30, 254(%rdx) # AVX512{VBMI2,VL} Disp8
vpcompressw %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpcompressw %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL}
vpcompressw %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL}
vpcompressw %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpcompressw %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL}
vpcompressw %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL}
vpexpandb (%rcx), %xmm30{%k7} # AVX512{VBMI2,VL}
vpexpandb (%rcx), %xmm30{%k7}{z} # AVX512{VBMI2,VL}
vpexpandb 0x123(%rax,%r14,8), %xmm30 # AVX512{VBMI2,VL}
vpexpandb 127(%rdx), %xmm30 # AVX512{VBMI2,VL} Disp8
vpexpandb (%rcx), %ymm30{%k7} # AVX512{VBMI2,VL}
vpexpandb (%rcx), %ymm30{%k7}{z} # AVX512{VBMI2,VL}
vpexpandb 0x123(%rax,%r14,8), %ymm30 # AVX512{VBMI2,VL}
vpexpandb 127(%rdx), %ymm30 # AVX512{VBMI2,VL} Disp8
vpexpandb %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpexpandb %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL}
vpexpandb %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL}
vpexpandb %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpexpandb %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL}
vpexpandb %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL}
vpexpandw (%rcx), %xmm30{%k7} # AVX512{VBMI2,VL}
vpexpandw (%rcx), %xmm30{%k7}{z} # AVX512{VBMI2,VL}
vpexpandw 0x123(%rax,%r14,8), %xmm30 # AVX512{VBMI2,VL}
vpexpandw 254(%rdx), %xmm30 # AVX512{VBMI2,VL} Disp8
vpexpandw (%rcx), %ymm30{%k7} # AVX512{VBMI2,VL}
vpexpandw (%rcx), %ymm30{%k7}{z} # AVX512{VBMI2,VL}
vpexpandw 0x123(%rax,%r14,8), %ymm30 # AVX512{VBMI2,VL}
vpexpandw 254(%rdx), %ymm30 # AVX512{VBMI2,VL} Disp8
vpexpandw %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpexpandw %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL}
vpexpandw %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL}
vpexpandw %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpexpandw %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL}
vpexpandw %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL}
vpshldvw %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshldvw %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL}
vpshldvw %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL}
vpshldvw 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshldvw 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8
vpshldvw %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshldvw %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL}
vpshldvw %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL}
vpshldvw 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshldvw 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8
vpshldvd %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshldvd %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL}
vpshldvd %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL}
vpshldvd 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshldvd 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8
vpshldvd 508(%rdx){1to4}, %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8
vpshldvd %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshldvd %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL}
vpshldvd %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL}
vpshldvd 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshldvd 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8
vpshldvd 508(%rdx){1to8}, %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8
vpshldvq %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshldvq %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL}
vpshldvq %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL}
vpshldvq 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshldvq 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8
vpshldvq 1016(%rdx){1to2}, %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8
vpshldvq %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshldvq %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL}
vpshldvq %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL}
vpshldvq 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshldvq 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8
vpshldvq 1016(%rdx){1to4}, %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8
vpshrdvw %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshrdvw %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL}
vpshrdvw %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL}
vpshrdvw 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshrdvw 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8
vpshrdvw %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshrdvw %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL}
vpshrdvw %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL}
vpshrdvw 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshrdvw 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8
vpshrdvd %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshrdvd %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL}
vpshrdvd %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL}
vpshrdvd 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshrdvd 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8
vpshrdvd 508(%rdx){1to4}, %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8
vpshrdvd %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshrdvd %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL}
vpshrdvd %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL}
vpshrdvd 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshrdvd 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8
vpshrdvd 508(%rdx){1to8}, %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8
vpshrdvq %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshrdvq %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL}
vpshrdvq %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL}
vpshrdvq 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshrdvq 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8
vpshrdvq 1016(%rdx){1to2}, %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8
vpshrdvq %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshrdvq %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL}
vpshrdvq %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL}
vpshrdvq 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshrdvq 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8
vpshrdvq 1016(%rdx){1to4}, %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8
vpshldw $0xab, %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshldw $0xab, %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL}
vpshldw $0xab, %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL}
vpshldw $123, 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshldw $123, 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8
vpshldw $0xab, %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshldw $0xab, %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL}
vpshldw $0xab, %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL}
vpshldw $123, 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshldw $123, 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8
vpshldd $0xab, %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshldd $0xab, %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL}
vpshldd $0xab, %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL}
vpshldd $123, 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshldd $123, 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8
vpshldd $123, 508(%rdx){1to4}, %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8
vpshldd $0xab, %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshldd $0xab, %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL}
vpshldd $0xab, %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL}
vpshldd $123, 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshldd $123, 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8
vpshldd $123, 508(%rdx){1to8}, %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8
vpshldq $0xab, %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshldq $0xab, %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL}
vpshldq $0xab, %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL}
vpshldq $123, 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshldq $123, 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8
vpshldq $123, 1016(%rdx){1to2}, %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8
vpshldq $0xab, %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshldq $0xab, %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL}
vpshldq $0xab, %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL}
vpshldq $123, 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshldq $123, 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8
vpshldq $123, 1016(%rdx){1to4}, %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8
vpshrdw $0xab, %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshrdw $0xab, %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL}
vpshrdw $0xab, %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL}
vpshrdw $123, 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshrdw $123, 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8
vpshrdw $0xab, %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshrdw $0xab, %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL}
vpshrdw $0xab, %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL}
vpshrdw $123, 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshrdw $123, 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8
vpshrdd $0xab, %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshrdd $0xab, %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL}
vpshrdd $0xab, %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL}
vpshrdd $123, 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshrdd $123, 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8
vpshrdd $123, 508(%rdx){1to4}, %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8
vpshrdd $0xab, %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshrdd $0xab, %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL}
vpshrdd $0xab, %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL}
vpshrdd $123, 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshrdd $123, 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8
vpshrdd $123, 508(%rdx){1to8}, %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8
vpshrdq $0xab, %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshrdq $0xab, %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL}
vpshrdq $0xab, %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL}
vpshrdq $123, 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL}
vpshrdq $123, 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8
vpshrdq $123, 1016(%rdx){1to2}, %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8
vpshrdq $0xab, %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshrdq $0xab, %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL}
vpshrdq $0xab, %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL}
vpshrdq $123, 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL}
vpshrdq $123, 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8
vpshrdq $123, 1016(%rdx){1to4}, %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8
.intel_syntax noprefix
vpcompressb XMMWORD PTR [rcx]{k7}, xmm30 # AVX512{VBMI2,VL}
vpcompressb XMMWORD PTR [rax+r14*8+0x1234], xmm30 # AVX512{VBMI2,VL}
vpcompressb XMMWORD PTR [rdx+127], xmm30 # AVX512{VBMI2,VL} Disp8
vpcompressb YMMWORD PTR [rcx]{k7}, ymm30 # AVX512{VBMI2,VL}
vpcompressb YMMWORD PTR [rax+r14*8+0x1234], ymm30 # AVX512{VBMI2,VL}
vpcompressb YMMWORD PTR [rdx+127], ymm30 # AVX512{VBMI2,VL} Disp8
vpcompressb xmm30, xmm29 # AVX512{VBMI2,VL}
vpcompressb xmm30{k7}, xmm29 # AVX512{VBMI2,VL}
vpcompressb xmm30{k7}{z}, xmm29 # AVX512{VBMI2,VL}
vpcompressb ymm30, ymm29 # AVX512{VBMI2,VL}
vpcompressb ymm30{k7}, ymm29 # AVX512{VBMI2,VL}
vpcompressb ymm30{k7}{z}, ymm29 # AVX512{VBMI2,VL}
vpcompressw XMMWORD PTR [rcx]{k7}, xmm30 # AVX512{VBMI2,VL}
vpcompressw XMMWORD PTR [rax+r14*8+0x1234], xmm30 # AVX512{VBMI2,VL}
vpcompressw XMMWORD PTR [rdx+254], xmm30 # AVX512{VBMI2,VL} Disp8
vpcompressw YMMWORD PTR [rcx]{k7}, ymm30 # AVX512{VBMI2,VL}
vpcompressw YMMWORD PTR [rax+r14*8+0x1234], ymm30 # AVX512{VBMI2,VL}
vpcompressw YMMWORD PTR [rdx+254], ymm30 # AVX512{VBMI2,VL} Disp8
vpcompressw xmm30, xmm29 # AVX512{VBMI2,VL}
vpcompressw xmm30{k7}, xmm29 # AVX512{VBMI2,VL}
vpcompressw xmm30{k7}{z}, xmm29 # AVX512{VBMI2,VL}
vpcompressw ymm30, ymm29 # AVX512{VBMI2,VL}
vpcompressw ymm30{k7}, ymm29 # AVX512{VBMI2,VL}
vpcompressw ymm30{k7}{z}, ymm29 # AVX512{VBMI2,VL}
vpexpandb xmm30{k7}, XMMWORD PTR [rcx] # AVX512{VBMI2,VL}
vpexpandb xmm30{k7}{z}, XMMWORD PTR [rcx] # AVX512{VBMI2,VL}
vpexpandb xmm30, XMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL}
vpexpandb xmm30, XMMWORD PTR [rdx+127] # AVX512{VBMI2,VL} Disp8
vpexpandb ymm30{k7}, YMMWORD PTR [rcx] # AVX512{VBMI2,VL}
vpexpandb ymm30{k7}{z}, YMMWORD PTR [rcx] # AVX512{VBMI2,VL}
vpexpandb ymm30, YMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL}
vpexpandb ymm30, YMMWORD PTR [rdx+127] # AVX512{VBMI2,VL} Disp8
vpexpandb xmm30, xmm29 # AVX512{VBMI2,VL}
vpexpandb xmm30{k7}, xmm29 # AVX512{VBMI2,VL}
vpexpandb xmm30{k7}{z}, xmm29 # AVX512{VBMI2,VL}
vpexpandb ymm30, ymm29 # AVX512{VBMI2,VL}
vpexpandb ymm30{k7}, ymm29 # AVX512{VBMI2,VL}
vpexpandb ymm30{k7}{z}, ymm29 # AVX512{VBMI2,VL}
vpexpandw xmm30{k7}, XMMWORD PTR [rcx] # AVX512{VBMI2,VL}
vpexpandw xmm30{k7}{z}, XMMWORD PTR [rcx] # AVX512{VBMI2,VL}
vpexpandw xmm30, XMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL}
vpexpandw xmm30, XMMWORD PTR [rdx+254] # AVX512{VBMI2,VL} Disp8
vpexpandw ymm30{k7}, YMMWORD PTR [rcx] # AVX512{VBMI2,VL}
vpexpandw ymm30{k7}{z}, YMMWORD PTR [rcx] # AVX512{VBMI2,VL}
vpexpandw ymm30, YMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL}
vpexpandw ymm30, YMMWORD PTR [rdx+254] # AVX512{VBMI2,VL} Disp8
vpexpandw xmm30, xmm29 # AVX512{VBMI2,VL}
vpexpandw xmm30{k7}, xmm29 # AVX512{VBMI2,VL}
vpexpandw xmm30{k7}{z}, xmm29 # AVX512{VBMI2,VL}
vpexpandw ymm30, ymm29 # AVX512{VBMI2,VL}
vpexpandw ymm30{k7}, ymm29 # AVX512{VBMI2,VL}
vpexpandw ymm30{k7}{z}, ymm29 # AVX512{VBMI2,VL}
vpshldvw xmm30, xmm29, xmm28 # AVX512{VBMI2,VL}
vpshldvw xmm30{k7}, xmm29, xmm28 # AVX512{VBMI2,VL}
vpshldvw xmm30{k7}{z}, xmm29, xmm28 # AVX512{VBMI2,VL}
vpshldvw xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL}
vpshldvw xmm30, xmm29, XMMWORD PTR [rdx+2032] # AVX512{VBMI2,VL} Disp8
vpshldvw ymm30, ymm29, ymm28 # AVX512{VBMI2,VL}
vpshldvw ymm30{k7}, ymm29, ymm28 # AVX512{VBMI2,VL}
vpshldvw ymm30{k7}{z}, ymm29, ymm28 # AVX512{VBMI2,VL}
vpshldvw ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL}
vpshldvw ymm30, ymm29, YMMWORD PTR [rdx+4064] # AVX512{VBMI2,VL} Disp8
vpshldvd xmm30, xmm29, xmm28 # AVX512{VBMI2,VL}
vpshldvd xmm30{k7}, xmm29, xmm28 # AVX512{VBMI2,VL}
vpshldvd xmm30{k7}{z}, xmm29, xmm28 # AVX512{VBMI2,VL}
vpshldvd xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL}
vpshldvd xmm30, xmm29, XMMWORD PTR [rdx+2032] # AVX512{VBMI2,VL} Disp8
vpshldvd xmm30, xmm29, [rdx+508]{1to4} # AVX512{VBMI2,VL} Disp8
vpshldvd ymm30, ymm29, ymm28 # AVX512{VBMI2,VL}
vpshldvd ymm30{k7}, ymm29, ymm28 # AVX512{VBMI2,VL}
vpshldvd ymm30{k7}{z}, ymm29, ymm28 # AVX512{VBMI2,VL}
vpshldvd ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL}
vpshldvd ymm30, ymm29, YMMWORD PTR [rdx+4064] # AVX512{VBMI2,VL} Disp8
vpshldvd ymm30, ymm29, [rdx+508]{1to8} # AVX512{VBMI2,VL} Disp8
vpshldvq xmm30, xmm29, xmm28 # AVX512{VBMI2,VL}
vpshldvq xmm30{k7}, xmm29, xmm28 # AVX512{VBMI2,VL}
vpshldvq xmm30{k7}{z}, xmm29, xmm28 # AVX512{VBMI2,VL}
vpshldvq xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL}
vpshldvq xmm30, xmm29, XMMWORD PTR [rdx+2032] # AVX512{VBMI2,VL} Disp8
vpshldvq xmm30, xmm29, [rdx+1016]{1to2} # AVX512{VBMI2,VL} Disp8
vpshldvq ymm30, ymm29, ymm28 # AVX512{VBMI2,VL}
vpshldvq ymm30{k7}, ymm29, ymm28 # AVX512{VBMI2,VL}
vpshldvq ymm30{k7}{z}, ymm29, ymm28 # AVX512{VBMI2,VL}
vpshldvq ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL}
vpshldvq ymm30, ymm29, YMMWORD PTR [rdx+4064] # AVX512{VBMI2,VL} Disp8
vpshldvq ymm30, ymm29, [rdx+1016]{1to4} # AVX512{VBMI2,VL} Disp8
vpshrdvw xmm30, xmm29, xmm28 # AVX512{VBMI2,VL}
vpshrdvw xmm30{k7}, xmm29, xmm28 # AVX512{VBMI2,VL}
vpshrdvw xmm30{k7}{z}, xmm29, xmm28 # AVX512{VBMI2,VL}
vpshrdvw xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL}
vpshrdvw xmm30, xmm29, XMMWORD PTR [rdx+2032] # AVX512{VBMI2,VL} Disp8
vpshrdvw ymm30, ymm29, ymm28 # AVX512{VBMI2,VL}
vpshrdvw ymm30{k7}, ymm29, ymm28 # AVX512{VBMI2,VL}
vpshrdvw ymm30{k7}{z}, ymm29, ymm28 # AVX512{VBMI2,VL}
vpshrdvw ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL}
vpshrdvw ymm30, ymm29, YMMWORD PTR [rdx+4064] # AVX512{VBMI2,VL} Disp8
vpshrdvd xmm30, xmm29, xmm28 # AVX512{VBMI2,VL}
vpshrdvd xmm30{k7}, xmm29, xmm28 # AVX512{VBMI2,VL}
vpshrdvd xmm30{k7}{z}, xmm29, xmm28 # AVX512{VBMI2,VL}
vpshrdvd xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL}
vpshrdvd xmm30, xmm29, XMMWORD PTR [rdx+2032] # AVX512{VBMI2,VL} Disp8
vpshrdvd xmm30, xmm29, [rdx+508]{1to4} # AVX512{VBMI2,VL} Disp8
vpshrdvd ymm30, ymm29, ymm28 # AVX512{VBMI2,VL}
vpshrdvd ymm30{k7}, ymm29, ymm28 # AVX512{VBMI2,VL}
vpshrdvd ymm30{k7}{z}, ymm29, ymm28 # AVX512{VBMI2,VL}
vpshrdvd ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL}
vpshrdvd ymm30, ymm29, YMMWORD PTR [rdx+4064] # AVX512{VBMI2,VL} Disp8
vpshrdvd ymm30, ymm29, [rdx+508]{1to8} # AVX512{VBMI2,VL} Disp8
vpshrdvq xmm30, xmm29, xmm28 # AVX512{VBMI2,VL}
vpshrdvq xmm30{k7}, xmm29, xmm28 # AVX512{VBMI2,VL}
vpshrdvq xmm30{k7}{z}, xmm29, xmm28 # AVX512{VBMI2,VL}
vpshrdvq xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL}
vpshrdvq xmm30, xmm29, XMMWORD PTR [rdx+2032] # AVX512{VBMI2,VL} Disp8
vpshrdvq xmm30, xmm29, [rdx+1016]{1to2} # AVX512{VBMI2,VL} Disp8
vpshrdvq ymm30, ymm29, ymm28 # AVX512{VBMI2,VL}
vpshrdvq ymm30{k7}, ymm29, ymm28 # AVX512{VBMI2,VL}
vpshrdvq ymm30{k7}{z}, ymm29, ymm28 # AVX512{VBMI2,VL}
vpshrdvq ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL}
vpshrdvq ymm30, ymm29, YMMWORD PTR [rdx+4064] # AVX512{VBMI2,VL} Disp8
vpshrdvq ymm30, ymm29, [rdx+1016]{1to4} # AVX512{VBMI2,VL} Disp8
vpshldw xmm30, xmm29, xmm28, 0xab # AVX512{VBMI2,VL}
vpshldw xmm30{k7}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL}
vpshldw xmm30{k7}{z}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL}
vpshldw xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL}
vpshldw xmm30, xmm29, XMMWORD PTR [rdx+2032], 123 # AVX512{VBMI2,VL} Disp8
vpshldw ymm30, ymm29, ymm28, 0xab # AVX512{VBMI2,VL}
vpshldw ymm30{k7}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL}
vpshldw ymm30{k7}{z}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL}
vpshldw ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL}
vpshldw ymm30, ymm29, YMMWORD PTR [rdx+4064], 123 # AVX512{VBMI2,VL} Disp8
vpshldd xmm30, xmm29, xmm28, 0xab # AVX512{VBMI2,VL}
vpshldd xmm30{k7}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL}
vpshldd xmm30{k7}{z}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL}
vpshldd xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL}
vpshldd xmm30, xmm29, XMMWORD PTR [rdx+2032], 123 # AVX512{VBMI2,VL} Disp8
vpshldd xmm30, xmm29, [rdx+508]{1to4}, 123 # AVX512{VBMI2,VL} Disp8
vpshldd ymm30, ymm29, ymm28, 0xab # AVX512{VBMI2,VL}
vpshldd ymm30{k7}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL}
vpshldd ymm30{k7}{z}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL}
vpshldd ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL}
vpshldd ymm30, ymm29, YMMWORD PTR [rdx+4064], 123 # AVX512{VBMI2,VL} Disp8
vpshldd ymm30, ymm29, [rdx+508]{1to8}, 123 # AVX512{VBMI2,VL} Disp8
vpshldq xmm30, xmm29, xmm28, 0xab # AVX512{VBMI2,VL}
vpshldq xmm30{k7}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL}
vpshldq xmm30{k7}{z}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL}
vpshldq xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL}
vpshldq xmm30, xmm29, XMMWORD PTR [rdx+2032], 123 # AVX512{VBMI2,VL} Disp8
vpshldq xmm30, xmm29, [rdx+1016]{1to2}, 123 # AVX512{VBMI2,VL} Disp8
vpshldq ymm30, ymm29, ymm28, 0xab # AVX512{VBMI2,VL}
vpshldq ymm30{k7}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL}
vpshldq ymm30{k7}{z}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL}
vpshldq ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL}
vpshldq ymm30, ymm29, YMMWORD PTR [rdx+4064], 123 # AVX512{VBMI2,VL} Disp8
vpshldq ymm30, ymm29, [rdx+1016]{1to4}, 123 # AVX512{VBMI2,VL} Disp8
vpshrdw xmm30, xmm29, xmm28, 0xab # AVX512{VBMI2,VL}
vpshrdw xmm30{k7}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL}
vpshrdw xmm30{k7}{z}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL}
vpshrdw xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL}
vpshrdw xmm30, xmm29, XMMWORD PTR [rdx+2032], 123 # AVX512{VBMI2,VL} Disp8
vpshrdw ymm30, ymm29, ymm28, 0xab # AVX512{VBMI2,VL}
vpshrdw ymm30{k7}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL}
vpshrdw ymm30{k7}{z}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL}
vpshrdw ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL}
vpshrdw ymm30, ymm29, YMMWORD PTR [rdx+4064], 123 # AVX512{VBMI2,VL} Disp8
vpshrdd xmm30, xmm29, xmm28, 0xab # AVX512{VBMI2,VL}
vpshrdd xmm30{k7}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL}
vpshrdd xmm30{k7}{z}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL}
vpshrdd xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL}
vpshrdd xmm30, xmm29, XMMWORD PTR [rdx+2032], 123 # AVX512{VBMI2,VL} Disp8
vpshrdd xmm30, xmm29, [rdx+508]{1to4}, 123 # AVX512{VBMI2,VL} Disp8
vpshrdd ymm30, ymm29, ymm28, 0xab # AVX512{VBMI2,VL}
vpshrdd ymm30{k7}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL}
vpshrdd ymm30{k7}{z}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL}
vpshrdd ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL}
vpshrdd ymm30, ymm29, YMMWORD PTR [rdx+4064], 123 # AVX512{VBMI2,VL} Disp8
vpshrdd ymm30, ymm29, [rdx+508]{1to8}, 123 # AVX512{VBMI2,VL} Disp8
vpshrdq xmm30, xmm29, xmm28, 0xab # AVX512{VBMI2,VL}
vpshrdq xmm30{k7}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL}
vpshrdq xmm30{k7}{z}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL}
vpshrdq xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL}
vpshrdq xmm30, xmm29, XMMWORD PTR [rdx+2032], 123 # AVX512{VBMI2,VL} Disp8
vpshrdq xmm30, xmm29, [rdx+1016]{1to2}, 123 # AVX512{VBMI2,VL} Disp8
vpshrdq ymm30, ymm29, ymm28, 0xab # AVX512{VBMI2,VL}
vpshrdq ymm30{k7}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL}
vpshrdq ymm30{k7}{z}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL}
vpshrdq ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL}
vpshrdq ymm30, ymm29, YMMWORD PTR [rdx+4064], 123 # AVX512{VBMI2,VL} Disp8
vpshrdq ymm30, ymm29, [rdx+1016]{1to4}, 123 # AVX512{VBMI2,VL} Disp8