+[a-f0-9]+: 62 f1 ff 89 6f d1 vmovdqu16 %xmm1,%xmm2\{%k1\}\{z\}
+[a-f0-9]+: 62 f1 7e 89 6f d1 vmovdqu32 %xmm1,%xmm2\{%k1\}\{z\}
+[a-f0-9]+: 62 f1 fe 89 6f d1 vmovdqu64 %xmm1,%xmm2\{%k1\}\{z\}
+ +[a-f0-9]+: c5 .* vpand %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpand %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpandn %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpxor %xmm2,%xmm3,%xmm4
+ +[a-f0-9]+: c5 .* vpand %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpand %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpandn %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpxor %ymm2,%ymm3,%ymm4
+ +[a-f0-9]+: c5 .* vpand 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpandn 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpxor 0x70\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%eax\),%xmm2,%xmm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpand 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpandn 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: c5 .* vpxor 0x60\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandd 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandq 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnd 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandnq 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpord 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vporq 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxord 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpxorq 0x80\(%eax\),%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandd %xmm2,%xmm3,%xmm4\{%k5\}
+ +[a-f0-9]+: 62 .* vpandq %ymm2,%ymm3,%ymm4\{%k5\}
+ +[a-f0-9]+: 62 .* vpandnd %ymm2,%ymm3,%ymm4\{%k5\}
+ +[a-f0-9]+: 62 .* vpandnq %xmm2,%xmm3,%xmm4\{%k5\}
+ +[a-f0-9]+: 62 .* vpord %xmm2,%xmm3,%xmm4\{%k5\}
+ +[a-f0-9]+: 62 .* vporq %ymm2,%ymm3,%ymm4\{%k5\}
+ +[a-f0-9]+: 62 .* vpxord %ymm2,%ymm3,%ymm4\{%k5\}
+ +[a-f0-9]+: 62 .* vpxorq %xmm2,%xmm3,%xmm4\{%k5\}
+ +[a-f0-9]+: 62 .* vpandd \(%eax\)\{1to8\},%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpandq \(%eax\)\{1to2\},%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnd \(%eax\)\{1to4\},%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpandnq \(%eax\)\{1to4\},%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vpord \(%eax\)\{1to8\},%ymm2,%ymm3
+ +[a-f0-9]+: 62 .* vporq \(%eax\)\{1to2\},%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxord \(%eax\)\{1to4\},%xmm2,%xmm3
+ +[a-f0-9]+: 62 .* vpxorq \(%eax\)\{1to4\},%ymm2,%ymm3
#pass