mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-23 20:53:53 +08:00
b24413180f
Many source files in the tree are missing licensing information, which makes it harder for compliance tools to determine the correct license. By default all files without license information are under the default license of the kernel, which is GPL version 2. Update the files which contain no license information with the 'GPL-2.0' SPDX license identifier. The SPDX identifier is a legally binding shorthand, which can be used instead of the full boiler plate text. This patch is based on work done by Thomas Gleixner and Kate Stewart and Philippe Ombredanne. How this work was done: Patches were generated and checked against linux-4.14-rc6 for a subset of the use cases: - file had no licensing information it it. - file was a */uapi/* one with no licensing information in it, - file was a */uapi/* one with existing licensing information, Further patches will be generated in subsequent months to fix up cases where non-standard license headers were used, and references to license had to be inferred by heuristics based on keywords. The analysis to determine which SPDX License Identifier to be applied to a file was done in a spreadsheet of side by side results from of the output of two independent scanners (ScanCode & Windriver) producing SPDX tag:value files created by Philippe Ombredanne. Philippe prepared the base worksheet, and did an initial spot review of a few 1000 files. The 4.13 kernel was the starting point of the analysis with 60,537 files assessed. Kate Stewart did a file by file comparison of the scanner results in the spreadsheet to determine which SPDX license identifier(s) to be applied to the file. She confirmed any determination that was not immediately clear with lawyers working with the Linux Foundation. Criteria used to select files for SPDX license identifier tagging was: - Files considered eligible had to be source code files. - Make and config files were included as candidates if they contained >5 lines of source - File already had some variant of a license header in it (even if <5 lines). All documentation files were explicitly excluded. The following heuristics were used to determine which SPDX license identifiers to apply. - when both scanners couldn't find any license traces, file was considered to have no license information in it, and the top level COPYING file license applied. For non */uapi/* files that summary was: SPDX license identifier # files ---------------------------------------------------|------- GPL-2.0 11139 and resulted in the first patch in this series. If that file was a */uapi/* path one, it was "GPL-2.0 WITH Linux-syscall-note" otherwise it was "GPL-2.0". Results of that was: SPDX license identifier # files ---------------------------------------------------|------- GPL-2.0 WITH Linux-syscall-note 930 and resulted in the second patch in this series. - if a file had some form of licensing information in it, and was one of the */uapi/* ones, it was denoted with the Linux-syscall-note if any GPL family license was found in the file or had no licensing in it (per prior point). Results summary: SPDX license identifier # files ---------------------------------------------------|------ GPL-2.0 WITH Linux-syscall-note 270 GPL-2.0+ WITH Linux-syscall-note 169 ((GPL-2.0 WITH Linux-syscall-note) OR BSD-2-Clause) 21 ((GPL-2.0 WITH Linux-syscall-note) OR BSD-3-Clause) 17 LGPL-2.1+ WITH Linux-syscall-note 15 GPL-1.0+ WITH Linux-syscall-note 14 ((GPL-2.0+ WITH Linux-syscall-note) OR BSD-3-Clause) 5 LGPL-2.0+ WITH Linux-syscall-note 4 LGPL-2.1 WITH Linux-syscall-note 3 ((GPL-2.0 WITH Linux-syscall-note) OR MIT) 3 ((GPL-2.0 WITH Linux-syscall-note) AND MIT) 1 and that resulted in the third patch in this series. - when the two scanners agreed on the detected license(s), that became the concluded license(s). - when there was disagreement between the two scanners (one detected a license but the other didn't, or they both detected different licenses) a manual inspection of the file occurred. - In most cases a manual inspection of the information in the file resulted in a clear resolution of the license that should apply (and which scanner probably needed to revisit its heuristics). - When it was not immediately clear, the license identifier was confirmed with lawyers working with the Linux Foundation. - If there was any question as to the appropriate license identifier, the file was flagged for further research and to be revisited later in time. In total, over 70 hours of logged manual review was done on the spreadsheet to determine the SPDX license identifiers to apply to the source files by Kate, Philippe, Thomas and, in some cases, confirmation by lawyers working with the Linux Foundation. Kate also obtained a third independent scan of the 4.13 code base from FOSSology, and compared selected files where the other two scanners disagreed against that SPDX file, to see if there was new insights. The Windriver scanner is based on an older version of FOSSology in part, so they are related. Thomas did random spot checks in about 500 files from the spreadsheets for the uapi headers and agreed with SPDX license identifier in the files he inspected. For the non-uapi files Thomas did random spot checks in about 15000 files. In initial set of patches against 4.14-rc6, 3 files were found to have copy/paste license identifier errors, and have been fixed to reflect the correct identifier. Additionally Philippe spent 10 hours this week doing a detailed manual inspection and review of the 12,461 patched files from the initial patch version early this week with: - a full scancode scan run, collecting the matched texts, detected license ids and scores - reviewing anything where there was a license detected (about 500+ files) to ensure that the applied SPDX license was correct - reviewing anything where there was no detection but the patch license was not GPL-2.0 WITH Linux-syscall-note to ensure that the applied SPDX license was correct This produced a worksheet with 20 files needing minor correction. This worksheet was then exported into 3 different .csv files for the different types of files to be modified. These .csv files were then reviewed by Greg. Thomas wrote a script to parse the csv files and add the proper SPDX tag to the file, in the format that the file expected. This script was further refined by Greg based on the output to detect more types of files automatically and to distinguish between header and source .c files (which need different comment types.) Finally Greg ran the script using the .csv files to generate the patches. Reviewed-by: Kate Stewart <kstewart@linuxfoundation.org> Reviewed-by: Philippe Ombredanne <pombredanne@nexb.com> Reviewed-by: Thomas Gleixner <tglx@linutronix.de> Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
2694 lines
89 KiB
C
2694 lines
89 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* This file contains instructions for testing by the test titled:
|
|
*
|
|
* "Test x86 instruction decoder - new instructions"
|
|
*
|
|
* Note that the 'Expecting' comment lines are consumed by the
|
|
* gen-insn-x86-dat.awk script and have the format:
|
|
*
|
|
* Expecting: <op> <branch> <rel>
|
|
*
|
|
* If this file is changed, remember to run the gen-insn-x86-dat.sh
|
|
* script and commit the result.
|
|
*
|
|
* Refer to insn-x86.c for more details.
|
|
*/
|
|
|
|
int main(void)
|
|
{
|
|
/* Following line is a marker for the awk script - do not change */
|
|
asm volatile("rdtsc"); /* Start here */
|
|
|
|
/* Test fix for vcvtph2ps in x86-opcode-map.txt */
|
|
|
|
asm volatile("vcvtph2ps %xmm3,%ymm5");
|
|
|
|
#ifdef __x86_64__
|
|
|
|
/* AVX-512: Instructions with the same op codes as Mask Instructions */
|
|
|
|
asm volatile("cmovno %rax,%rbx");
|
|
asm volatile("cmovno 0x12345678(%rax),%rcx");
|
|
asm volatile("cmovno 0x12345678(%rax),%cx");
|
|
|
|
asm volatile("cmove %rax,%rbx");
|
|
asm volatile("cmove 0x12345678(%rax),%rcx");
|
|
asm volatile("cmove 0x12345678(%rax),%cx");
|
|
|
|
asm volatile("seto 0x12345678(%rax)");
|
|
asm volatile("setno 0x12345678(%rax)");
|
|
asm volatile("setb 0x12345678(%rax)");
|
|
asm volatile("setc 0x12345678(%rax)");
|
|
asm volatile("setnae 0x12345678(%rax)");
|
|
asm volatile("setae 0x12345678(%rax)");
|
|
asm volatile("setnb 0x12345678(%rax)");
|
|
asm volatile("setnc 0x12345678(%rax)");
|
|
asm volatile("sets 0x12345678(%rax)");
|
|
asm volatile("setns 0x12345678(%rax)");
|
|
|
|
/* AVX-512: Mask Instructions */
|
|
|
|
asm volatile("kandw %k7,%k6,%k5");
|
|
asm volatile("kandq %k7,%k6,%k5");
|
|
asm volatile("kandb %k7,%k6,%k5");
|
|
asm volatile("kandd %k7,%k6,%k5");
|
|
|
|
asm volatile("kandnw %k7,%k6,%k5");
|
|
asm volatile("kandnq %k7,%k6,%k5");
|
|
asm volatile("kandnb %k7,%k6,%k5");
|
|
asm volatile("kandnd %k7,%k6,%k5");
|
|
|
|
asm volatile("knotw %k7,%k6");
|
|
asm volatile("knotq %k7,%k6");
|
|
asm volatile("knotb %k7,%k6");
|
|
asm volatile("knotd %k7,%k6");
|
|
|
|
asm volatile("korw %k7,%k6,%k5");
|
|
asm volatile("korq %k7,%k6,%k5");
|
|
asm volatile("korb %k7,%k6,%k5");
|
|
asm volatile("kord %k7,%k6,%k5");
|
|
|
|
asm volatile("kxnorw %k7,%k6,%k5");
|
|
asm volatile("kxnorq %k7,%k6,%k5");
|
|
asm volatile("kxnorb %k7,%k6,%k5");
|
|
asm volatile("kxnord %k7,%k6,%k5");
|
|
|
|
asm volatile("kxorw %k7,%k6,%k5");
|
|
asm volatile("kxorq %k7,%k6,%k5");
|
|
asm volatile("kxorb %k7,%k6,%k5");
|
|
asm volatile("kxord %k7,%k6,%k5");
|
|
|
|
asm volatile("kaddw %k7,%k6,%k5");
|
|
asm volatile("kaddq %k7,%k6,%k5");
|
|
asm volatile("kaddb %k7,%k6,%k5");
|
|
asm volatile("kaddd %k7,%k6,%k5");
|
|
|
|
asm volatile("kunpckbw %k7,%k6,%k5");
|
|
asm volatile("kunpckwd %k7,%k6,%k5");
|
|
asm volatile("kunpckdq %k7,%k6,%k5");
|
|
|
|
asm volatile("kmovw %k6,%k5");
|
|
asm volatile("kmovw (%rcx),%k5");
|
|
asm volatile("kmovw 0x123(%rax,%r14,8),%k5");
|
|
asm volatile("kmovw %k5,(%rcx)");
|
|
asm volatile("kmovw %k5,0x123(%rax,%r14,8)");
|
|
asm volatile("kmovw %eax,%k5");
|
|
asm volatile("kmovw %ebp,%k5");
|
|
asm volatile("kmovw %r13d,%k5");
|
|
asm volatile("kmovw %k5,%eax");
|
|
asm volatile("kmovw %k5,%ebp");
|
|
asm volatile("kmovw %k5,%r13d");
|
|
|
|
asm volatile("kmovq %k6,%k5");
|
|
asm volatile("kmovq (%rcx),%k5");
|
|
asm volatile("kmovq 0x123(%rax,%r14,8),%k5");
|
|
asm volatile("kmovq %k5,(%rcx)");
|
|
asm volatile("kmovq %k5,0x123(%rax,%r14,8)");
|
|
asm volatile("kmovq %rax,%k5");
|
|
asm volatile("kmovq %rbp,%k5");
|
|
asm volatile("kmovq %r13,%k5");
|
|
asm volatile("kmovq %k5,%rax");
|
|
asm volatile("kmovq %k5,%rbp");
|
|
asm volatile("kmovq %k5,%r13");
|
|
|
|
asm volatile("kmovb %k6,%k5");
|
|
asm volatile("kmovb (%rcx),%k5");
|
|
asm volatile("kmovb 0x123(%rax,%r14,8),%k5");
|
|
asm volatile("kmovb %k5,(%rcx)");
|
|
asm volatile("kmovb %k5,0x123(%rax,%r14,8)");
|
|
asm volatile("kmovb %eax,%k5");
|
|
asm volatile("kmovb %ebp,%k5");
|
|
asm volatile("kmovb %r13d,%k5");
|
|
asm volatile("kmovb %k5,%eax");
|
|
asm volatile("kmovb %k5,%ebp");
|
|
asm volatile("kmovb %k5,%r13d");
|
|
|
|
asm volatile("kmovd %k6,%k5");
|
|
asm volatile("kmovd (%rcx),%k5");
|
|
asm volatile("kmovd 0x123(%rax,%r14,8),%k5");
|
|
asm volatile("kmovd %k5,(%rcx)");
|
|
asm volatile("kmovd %k5,0x123(%rax,%r14,8)");
|
|
asm volatile("kmovd %eax,%k5");
|
|
asm volatile("kmovd %ebp,%k5");
|
|
asm volatile("kmovd %r13d,%k5");
|
|
asm volatile("kmovd %k5,%eax");
|
|
asm volatile("kmovd %k5,%ebp");
|
|
asm volatile("kmovd %k5,%r13d");
|
|
|
|
asm volatile("kortestw %k6,%k5");
|
|
asm volatile("kortestq %k6,%k5");
|
|
asm volatile("kortestb %k6,%k5");
|
|
asm volatile("kortestd %k6,%k5");
|
|
|
|
asm volatile("ktestw %k6,%k5");
|
|
asm volatile("ktestq %k6,%k5");
|
|
asm volatile("ktestb %k6,%k5");
|
|
asm volatile("ktestd %k6,%k5");
|
|
|
|
asm volatile("kshiftrw $0x12,%k6,%k5");
|
|
asm volatile("kshiftrq $0x5b,%k6,%k5");
|
|
asm volatile("kshiftlw $0x12,%k6,%k5");
|
|
asm volatile("kshiftlq $0x5b,%k6,%k5");
|
|
|
|
/* AVX-512: Op code 0f 5b */
|
|
asm volatile("vcvtdq2ps %xmm5,%xmm6");
|
|
asm volatile("vcvtqq2ps %zmm29,%ymm6{%k7}");
|
|
asm volatile("vcvtps2dq %xmm5,%xmm6");
|
|
asm volatile("vcvttps2dq %xmm5,%xmm6");
|
|
|
|
/* AVX-512: Op code 0f 6f */
|
|
|
|
asm volatile("movq %mm0,%mm4");
|
|
asm volatile("vmovdqa %ymm4,%ymm6");
|
|
asm volatile("vmovdqa32 %zmm25,%zmm26");
|
|
asm volatile("vmovdqa64 %zmm25,%zmm26");
|
|
asm volatile("vmovdqu %ymm4,%ymm6");
|
|
asm volatile("vmovdqu32 %zmm29,%zmm30");
|
|
asm volatile("vmovdqu64 %zmm25,%zmm26");
|
|
asm volatile("vmovdqu8 %zmm29,%zmm30");
|
|
asm volatile("vmovdqu16 %zmm25,%zmm26");
|
|
|
|
/* AVX-512: Op code 0f 78 */
|
|
|
|
asm volatile("vmread %rax,%rbx");
|
|
asm volatile("vcvttps2udq %zmm25,%zmm26");
|
|
asm volatile("vcvttpd2udq %zmm29,%ymm6{%k7}");
|
|
asm volatile("vcvttsd2usi %xmm6,%rax");
|
|
asm volatile("vcvttss2usi %xmm6,%rax");
|
|
asm volatile("vcvttps2uqq %ymm5,%zmm26{%k7}");
|
|
asm volatile("vcvttpd2uqq %zmm29,%zmm30");
|
|
|
|
/* AVX-512: Op code 0f 79 */
|
|
|
|
asm volatile("vmwrite %rax,%rbx");
|
|
asm volatile("vcvtps2udq %zmm25,%zmm26");
|
|
asm volatile("vcvtpd2udq %zmm29,%ymm6{%k7}");
|
|
asm volatile("vcvtsd2usi %xmm6,%rax");
|
|
asm volatile("vcvtss2usi %xmm6,%rax");
|
|
asm volatile("vcvtps2uqq %ymm5,%zmm26{%k7}");
|
|
asm volatile("vcvtpd2uqq %zmm29,%zmm30");
|
|
|
|
/* AVX-512: Op code 0f 7a */
|
|
|
|
asm volatile("vcvtudq2pd %ymm5,%zmm29{%k7}");
|
|
asm volatile("vcvtuqq2pd %zmm25,%zmm26");
|
|
asm volatile("vcvtudq2ps %zmm29,%zmm30");
|
|
asm volatile("vcvtuqq2ps %zmm25,%ymm26{%k7}");
|
|
asm volatile("vcvttps2qq %ymm25,%zmm26{%k7}");
|
|
asm volatile("vcvttpd2qq %zmm29,%zmm30");
|
|
|
|
/* AVX-512: Op code 0f 7b */
|
|
|
|
asm volatile("vcvtusi2sd %eax,%xmm5,%xmm6");
|
|
asm volatile("vcvtusi2ss %eax,%xmm5,%xmm6");
|
|
asm volatile("vcvtps2qq %ymm5,%zmm26{%k7}");
|
|
asm volatile("vcvtpd2qq %zmm29,%zmm30");
|
|
|
|
/* AVX-512: Op code 0f 7f */
|
|
|
|
asm volatile("movq.s %mm0,%mm4");
|
|
asm volatile("vmovdqa %ymm8,%ymm6");
|
|
asm volatile("vmovdqa32.s %zmm25,%zmm26");
|
|
asm volatile("vmovdqa64.s %zmm25,%zmm26");
|
|
asm volatile("vmovdqu %ymm8,%ymm6");
|
|
asm volatile("vmovdqu32.s %zmm25,%zmm26");
|
|
asm volatile("vmovdqu64.s %zmm25,%zmm26");
|
|
asm volatile("vmovdqu8.s %zmm30,(%rcx)");
|
|
asm volatile("vmovdqu16.s %zmm25,%zmm26");
|
|
|
|
/* AVX-512: Op code 0f db */
|
|
|
|
asm volatile("pand %mm1,%mm2");
|
|
asm volatile("pand %xmm1,%xmm2");
|
|
asm volatile("vpand %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpandd %zmm24,%zmm25,%zmm26");
|
|
asm volatile("vpandq %zmm24,%zmm25,%zmm26");
|
|
|
|
/* AVX-512: Op code 0f df */
|
|
|
|
asm volatile("pandn %mm1,%mm2");
|
|
asm volatile("pandn %xmm1,%xmm2");
|
|
asm volatile("vpandn %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpandnd %zmm24,%zmm25,%zmm26");
|
|
asm volatile("vpandnq %zmm24,%zmm25,%zmm26");
|
|
|
|
/* AVX-512: Op code 0f e6 */
|
|
|
|
asm volatile("vcvttpd2dq %xmm1,%xmm2");
|
|
asm volatile("vcvtdq2pd %xmm5,%xmm6");
|
|
asm volatile("vcvtdq2pd %ymm5,%zmm26{%k7}");
|
|
asm volatile("vcvtqq2pd %zmm25,%zmm26");
|
|
asm volatile("vcvtpd2dq %xmm1,%xmm2");
|
|
|
|
/* AVX-512: Op code 0f eb */
|
|
|
|
asm volatile("por %mm4,%mm6");
|
|
asm volatile("vpor %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpord %zmm24,%zmm25,%zmm26");
|
|
asm volatile("vporq %zmm24,%zmm25,%zmm26");
|
|
|
|
/* AVX-512: Op code 0f ef */
|
|
|
|
asm volatile("pxor %mm4,%mm6");
|
|
asm volatile("vpxor %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpxord %zmm24,%zmm25,%zmm26");
|
|
asm volatile("vpxorq %zmm24,%zmm25,%zmm26");
|
|
|
|
/* AVX-512: Op code 0f 38 10 */
|
|
|
|
asm volatile("pblendvb %xmm1,%xmm0");
|
|
asm volatile("vpsrlvw %zmm27,%zmm28,%zmm29");
|
|
asm volatile("vpmovuswb %zmm28,%ymm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 11 */
|
|
|
|
asm volatile("vpmovusdb %zmm28,%xmm6{%k7}");
|
|
asm volatile("vpsravw %zmm27,%zmm28,%zmm29");
|
|
|
|
/* AVX-512: Op code 0f 38 12 */
|
|
|
|
asm volatile("vpmovusqb %zmm27,%xmm6{%k7}");
|
|
asm volatile("vpsllvw %zmm27,%zmm28,%zmm29");
|
|
|
|
/* AVX-512: Op code 0f 38 13 */
|
|
|
|
asm volatile("vcvtph2ps %xmm3,%ymm5");
|
|
asm volatile("vcvtph2ps %ymm5,%zmm27{%k7}");
|
|
asm volatile("vpmovusdw %zmm27,%ymm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 14 */
|
|
|
|
asm volatile("blendvps %xmm1,%xmm0");
|
|
asm volatile("vpmovusqw %zmm27,%xmm6{%k7}");
|
|
asm volatile("vprorvd %zmm27,%zmm28,%zmm29");
|
|
asm volatile("vprorvq %zmm27,%zmm28,%zmm29");
|
|
|
|
/* AVX-512: Op code 0f 38 15 */
|
|
|
|
asm volatile("blendvpd %xmm1,%xmm0");
|
|
asm volatile("vpmovusqd %zmm27,%ymm6{%k7}");
|
|
asm volatile("vprolvd %zmm27,%zmm28,%zmm29");
|
|
asm volatile("vprolvq %zmm27,%zmm28,%zmm29");
|
|
|
|
/* AVX-512: Op code 0f 38 16 */
|
|
|
|
asm volatile("vpermps %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpermps %ymm24,%ymm26,%ymm22{%k7}");
|
|
asm volatile("vpermpd %ymm24,%ymm26,%ymm22{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 19 */
|
|
|
|
asm volatile("vbroadcastsd %xmm4,%ymm6");
|
|
asm volatile("vbroadcastf32x2 %xmm27,%zmm26");
|
|
|
|
/* AVX-512: Op code 0f 38 1a */
|
|
|
|
asm volatile("vbroadcastf128 (%rcx),%ymm4");
|
|
asm volatile("vbroadcastf32x4 (%rcx),%zmm26");
|
|
asm volatile("vbroadcastf64x2 (%rcx),%zmm26");
|
|
|
|
/* AVX-512: Op code 0f 38 1b */
|
|
|
|
asm volatile("vbroadcastf32x8 (%rcx),%zmm27");
|
|
asm volatile("vbroadcastf64x4 (%rcx),%zmm26");
|
|
|
|
/* AVX-512: Op code 0f 38 1f */
|
|
|
|
asm volatile("vpabsq %zmm27,%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 20 */
|
|
|
|
asm volatile("vpmovsxbw %xmm4,%xmm5");
|
|
asm volatile("vpmovswb %zmm27,%ymm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 21 */
|
|
|
|
asm volatile("vpmovsxbd %xmm4,%ymm6");
|
|
asm volatile("vpmovsdb %zmm27,%xmm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 22 */
|
|
|
|
asm volatile("vpmovsxbq %xmm4,%ymm4");
|
|
asm volatile("vpmovsqb %zmm27,%xmm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 23 */
|
|
|
|
asm volatile("vpmovsxwd %xmm4,%ymm4");
|
|
asm volatile("vpmovsdw %zmm27,%ymm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 24 */
|
|
|
|
asm volatile("vpmovsxwq %xmm4,%ymm6");
|
|
asm volatile("vpmovsqw %zmm27,%xmm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 25 */
|
|
|
|
asm volatile("vpmovsxdq %xmm4,%ymm4");
|
|
asm volatile("vpmovsqd %zmm27,%ymm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 26 */
|
|
|
|
asm volatile("vptestmb %zmm27,%zmm28,%k5");
|
|
asm volatile("vptestmw %zmm27,%zmm28,%k5");
|
|
asm volatile("vptestnmb %zmm26,%zmm27,%k5");
|
|
asm volatile("vptestnmw %zmm26,%zmm27,%k5");
|
|
|
|
/* AVX-512: Op code 0f 38 27 */
|
|
|
|
asm volatile("vptestmd %zmm27,%zmm28,%k5");
|
|
asm volatile("vptestmq %zmm27,%zmm28,%k5");
|
|
asm volatile("vptestnmd %zmm26,%zmm27,%k5");
|
|
asm volatile("vptestnmq %zmm26,%zmm27,%k5");
|
|
|
|
/* AVX-512: Op code 0f 38 28 */
|
|
|
|
asm volatile("vpmuldq %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpmovm2b %k5,%zmm28");
|
|
asm volatile("vpmovm2w %k5,%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 29 */
|
|
|
|
asm volatile("vpcmpeqq %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpmovb2m %zmm28,%k5");
|
|
asm volatile("vpmovw2m %zmm28,%k5");
|
|
|
|
/* AVX-512: Op code 0f 38 2a */
|
|
|
|
asm volatile("vmovntdqa (%rcx),%ymm4");
|
|
asm volatile("vpbroadcastmb2q %k6,%zmm30");
|
|
|
|
/* AVX-512: Op code 0f 38 2c */
|
|
|
|
asm volatile("vmaskmovps (%rcx),%ymm4,%ymm6");
|
|
asm volatile("vscalefps %zmm24,%zmm25,%zmm26");
|
|
asm volatile("vscalefpd %zmm24,%zmm25,%zmm26");
|
|
|
|
/* AVX-512: Op code 0f 38 2d */
|
|
|
|
asm volatile("vmaskmovpd (%rcx),%ymm4,%ymm6");
|
|
asm volatile("vscalefss %xmm24,%xmm25,%xmm26{%k7}");
|
|
asm volatile("vscalefsd %xmm24,%xmm25,%xmm26{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 30 */
|
|
|
|
asm volatile("vpmovzxbw %xmm4,%ymm4");
|
|
asm volatile("vpmovwb %zmm27,%ymm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 31 */
|
|
|
|
asm volatile("vpmovzxbd %xmm4,%ymm6");
|
|
asm volatile("vpmovdb %zmm27,%xmm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 32 */
|
|
|
|
asm volatile("vpmovzxbq %xmm4,%ymm4");
|
|
asm volatile("vpmovqb %zmm27,%xmm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 33 */
|
|
|
|
asm volatile("vpmovzxwd %xmm4,%ymm4");
|
|
asm volatile("vpmovdw %zmm27,%ymm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 34 */
|
|
|
|
asm volatile("vpmovzxwq %xmm4,%ymm6");
|
|
asm volatile("vpmovqw %zmm27,%xmm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 35 */
|
|
|
|
asm volatile("vpmovzxdq %xmm4,%ymm4");
|
|
asm volatile("vpmovqd %zmm27,%ymm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 38 */
|
|
|
|
asm volatile("vpermd %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpermd %ymm24,%ymm26,%ymm22{%k7}");
|
|
asm volatile("vpermq %ymm24,%ymm26,%ymm22{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 38 */
|
|
|
|
asm volatile("vpminsb %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpmovm2d %k5,%zmm28");
|
|
asm volatile("vpmovm2q %k5,%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 39 */
|
|
|
|
asm volatile("vpminsd %xmm1,%xmm2,%xmm3");
|
|
asm volatile("vpminsd %zmm24,%zmm25,%zmm26");
|
|
asm volatile("vpminsq %zmm24,%zmm25,%zmm26");
|
|
asm volatile("vpmovd2m %zmm28,%k5");
|
|
asm volatile("vpmovq2m %zmm28,%k5");
|
|
|
|
/* AVX-512: Op code 0f 38 3a */
|
|
|
|
asm volatile("vpminuw %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpbroadcastmw2d %k6,%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 3b */
|
|
|
|
asm volatile("vpminud %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpminud %zmm24,%zmm25,%zmm26");
|
|
asm volatile("vpminuq %zmm24,%zmm25,%zmm26");
|
|
|
|
/* AVX-512: Op code 0f 38 3d */
|
|
|
|
asm volatile("vpmaxsd %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpmaxsd %zmm24,%zmm25,%zmm26");
|
|
asm volatile("vpmaxsq %zmm24,%zmm25,%zmm26");
|
|
|
|
/* AVX-512: Op code 0f 38 3f */
|
|
|
|
asm volatile("vpmaxud %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpmaxud %zmm24,%zmm25,%zmm26");
|
|
asm volatile("vpmaxuq %zmm24,%zmm25,%zmm26");
|
|
|
|
/* AVX-512: Op code 0f 38 42 */
|
|
|
|
asm volatile("vpmulld %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpmulld %zmm24,%zmm25,%zmm26");
|
|
asm volatile("vpmullq %zmm24,%zmm25,%zmm26");
|
|
|
|
/* AVX-512: Op code 0f 38 42 */
|
|
|
|
asm volatile("vgetexpps %zmm25,%zmm26");
|
|
asm volatile("vgetexppd %zmm27,%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 43 */
|
|
|
|
asm volatile("vgetexpss %xmm24,%xmm25,%xmm26{%k7}");
|
|
asm volatile("vgetexpsd %xmm28,%xmm29,%xmm30{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 44 */
|
|
|
|
asm volatile("vplzcntd %zmm27,%zmm28");
|
|
asm volatile("vplzcntq %zmm27,%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 46 */
|
|
|
|
asm volatile("vpsravd %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpsravd %zmm24,%zmm25,%zmm26");
|
|
asm volatile("vpsravq %zmm24,%zmm25,%zmm26");
|
|
|
|
/* AVX-512: Op code 0f 38 4c */
|
|
|
|
asm volatile("vrcp14ps %zmm25,%zmm26");
|
|
asm volatile("vrcp14pd %zmm27,%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 4d */
|
|
|
|
asm volatile("vrcp14ss %xmm24,%xmm25,%xmm26{%k7}");
|
|
asm volatile("vrcp14sd %xmm24,%xmm25,%xmm26{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 4e */
|
|
|
|
asm volatile("vrsqrt14ps %zmm25,%zmm26");
|
|
asm volatile("vrsqrt14pd %zmm27,%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 4f */
|
|
|
|
asm volatile("vrsqrt14ss %xmm24,%xmm25,%xmm26{%k7}");
|
|
asm volatile("vrsqrt14sd %xmm24,%xmm25,%xmm26{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 59 */
|
|
|
|
asm volatile("vpbroadcastq %xmm4,%xmm6");
|
|
asm volatile("vbroadcasti32x2 %xmm27,%zmm26");
|
|
|
|
/* AVX-512: Op code 0f 38 5a */
|
|
|
|
asm volatile("vbroadcasti128 (%rcx),%ymm4");
|
|
asm volatile("vbroadcasti32x4 (%rcx),%zmm26");
|
|
asm volatile("vbroadcasti64x2 (%rcx),%zmm26");
|
|
|
|
/* AVX-512: Op code 0f 38 5b */
|
|
|
|
asm volatile("vbroadcasti32x8 (%rcx),%zmm28");
|
|
asm volatile("vbroadcasti64x4 (%rcx),%zmm26");
|
|
|
|
/* AVX-512: Op code 0f 38 64 */
|
|
|
|
asm volatile("vpblendmd %zmm26,%zmm27,%zmm28");
|
|
asm volatile("vpblendmq %zmm26,%zmm27,%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 65 */
|
|
|
|
asm volatile("vblendmps %zmm24,%zmm25,%zmm26");
|
|
asm volatile("vblendmpd %zmm26,%zmm27,%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 66 */
|
|
|
|
asm volatile("vpblendmb %zmm26,%zmm27,%zmm28");
|
|
asm volatile("vpblendmw %zmm26,%zmm27,%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 75 */
|
|
|
|
asm volatile("vpermi2b %zmm24,%zmm25,%zmm26");
|
|
asm volatile("vpermi2w %zmm26,%zmm27,%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 76 */
|
|
|
|
asm volatile("vpermi2d %zmm26,%zmm27,%zmm28");
|
|
asm volatile("vpermi2q %zmm26,%zmm27,%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 77 */
|
|
|
|
asm volatile("vpermi2ps %zmm26,%zmm27,%zmm28");
|
|
asm volatile("vpermi2pd %zmm26,%zmm27,%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 7a */
|
|
|
|
asm volatile("vpbroadcastb %eax,%xmm30");
|
|
|
|
/* AVX-512: Op code 0f 38 7b */
|
|
|
|
asm volatile("vpbroadcastw %eax,%xmm30");
|
|
|
|
/* AVX-512: Op code 0f 38 7c */
|
|
|
|
asm volatile("vpbroadcastd %eax,%xmm30");
|
|
asm volatile("vpbroadcastq %rax,%zmm30");
|
|
|
|
/* AVX-512: Op code 0f 38 7d */
|
|
|
|
asm volatile("vpermt2b %zmm26,%zmm27,%zmm28");
|
|
asm volatile("vpermt2w %zmm26,%zmm27,%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 7e */
|
|
|
|
asm volatile("vpermt2d %zmm26,%zmm27,%zmm28");
|
|
asm volatile("vpermt2q %zmm26,%zmm27,%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 7f */
|
|
|
|
asm volatile("vpermt2ps %zmm26,%zmm27,%zmm28");
|
|
asm volatile("vpermt2pd %zmm26,%zmm27,%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 83 */
|
|
|
|
asm volatile("vpmultishiftqb %zmm26,%zmm27,%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 88 */
|
|
|
|
asm volatile("vexpandps (%rcx),%zmm26");
|
|
asm volatile("vexpandpd (%rcx),%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 89 */
|
|
|
|
asm volatile("vpexpandd (%rcx),%zmm28");
|
|
asm volatile("vpexpandq (%rcx),%zmm26");
|
|
|
|
/* AVX-512: Op code 0f 38 8a */
|
|
|
|
asm volatile("vcompressps %zmm28,(%rcx)");
|
|
asm volatile("vcompresspd %zmm28,(%rcx)");
|
|
|
|
/* AVX-512: Op code 0f 38 8b */
|
|
|
|
asm volatile("vpcompressd %zmm28,(%rcx)");
|
|
asm volatile("vpcompressq %zmm26,(%rcx)");
|
|
|
|
/* AVX-512: Op code 0f 38 8d */
|
|
|
|
asm volatile("vpermb %zmm26,%zmm27,%zmm28");
|
|
asm volatile("vpermw %zmm26,%zmm27,%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 90 */
|
|
|
|
asm volatile("vpgatherdd %xmm2,0x02(%rbp,%xmm7,2),%xmm1");
|
|
asm volatile("vpgatherdq %xmm2,0x04(%rbp,%xmm7,2),%xmm1");
|
|
asm volatile("vpgatherdd 0x7b(%rbp,%zmm27,8),%zmm26{%k1}");
|
|
asm volatile("vpgatherdq 0x7b(%rbp,%ymm27,8),%zmm26{%k1}");
|
|
|
|
/* AVX-512: Op code 0f 38 91 */
|
|
|
|
asm volatile("vpgatherqd %xmm2,0x02(%rbp,%xmm7,2),%xmm1");
|
|
asm volatile("vpgatherqq %xmm2,0x02(%rbp,%xmm7,2),%xmm1");
|
|
asm volatile("vpgatherqd 0x7b(%rbp,%zmm27,8),%ymm26{%k1}");
|
|
asm volatile("vpgatherqq 0x7b(%rbp,%zmm27,8),%zmm26{%k1}");
|
|
|
|
/* AVX-512: Op code 0f 38 a0 */
|
|
|
|
asm volatile("vpscatterdd %zmm28,0x7b(%rbp,%zmm29,8){%k1}");
|
|
asm volatile("vpscatterdq %zmm26,0x7b(%rbp,%ymm27,8){%k1}");
|
|
|
|
/* AVX-512: Op code 0f 38 a1 */
|
|
|
|
asm volatile("vpscatterqd %ymm6,0x7b(%rbp,%zmm29,8){%k1}");
|
|
asm volatile("vpscatterqq %ymm6,0x7b(%rbp,%ymm27,8){%k1}");
|
|
|
|
/* AVX-512: Op code 0f 38 a2 */
|
|
|
|
asm volatile("vscatterdps %zmm28,0x7b(%rbp,%zmm29,8){%k1}");
|
|
asm volatile("vscatterdpd %zmm28,0x7b(%rbp,%ymm27,8){%k1}");
|
|
|
|
/* AVX-512: Op code 0f 38 a3 */
|
|
|
|
asm volatile("vscatterqps %ymm6,0x7b(%rbp,%zmm29,8){%k1}");
|
|
asm volatile("vscatterqpd %zmm28,0x7b(%rbp,%zmm29,8){%k1}");
|
|
|
|
/* AVX-512: Op code 0f 38 b4 */
|
|
|
|
asm volatile("vpmadd52luq %zmm26,%zmm27,%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 b5 */
|
|
|
|
asm volatile("vpmadd52huq %zmm26,%zmm27,%zmm28");
|
|
|
|
/* AVX-512: Op code 0f 38 c4 */
|
|
|
|
asm volatile("vpconflictd %zmm26,%zmm27");
|
|
asm volatile("vpconflictq %zmm26,%zmm27");
|
|
|
|
/* AVX-512: Op code 0f 38 c8 */
|
|
|
|
asm volatile("vexp2ps %zmm29,%zmm30");
|
|
asm volatile("vexp2pd %zmm26,%zmm27");
|
|
|
|
/* AVX-512: Op code 0f 38 ca */
|
|
|
|
asm volatile("vrcp28ps %zmm29,%zmm30");
|
|
asm volatile("vrcp28pd %zmm26,%zmm27");
|
|
|
|
/* AVX-512: Op code 0f 38 cb */
|
|
|
|
asm volatile("vrcp28ss %xmm28,%xmm29,%xmm30{%k7}");
|
|
asm volatile("vrcp28sd %xmm25,%xmm26,%xmm27{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 cc */
|
|
|
|
asm volatile("vrsqrt28ps %zmm29,%zmm30");
|
|
asm volatile("vrsqrt28pd %zmm26,%zmm27");
|
|
|
|
/* AVX-512: Op code 0f 38 cd */
|
|
|
|
asm volatile("vrsqrt28ss %xmm28,%xmm29,%xmm30{%k7}");
|
|
asm volatile("vrsqrt28sd %xmm25,%xmm26,%xmm27{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 03 */
|
|
|
|
asm volatile("valignd $0x12,%zmm28,%zmm29,%zmm30");
|
|
asm volatile("valignq $0x12,%zmm25,%zmm26,%zmm27");
|
|
|
|
/* AVX-512: Op code 0f 3a 08 */
|
|
|
|
asm volatile("vroundps $0x5,%ymm6,%ymm2");
|
|
asm volatile("vrndscaleps $0x12,%zmm25,%zmm26");
|
|
|
|
/* AVX-512: Op code 0f 3a 09 */
|
|
|
|
asm volatile("vroundpd $0x5,%ymm6,%ymm2");
|
|
asm volatile("vrndscalepd $0x12,%zmm25,%zmm26");
|
|
|
|
/* AVX-512: Op code 0f 3a 1a */
|
|
|
|
asm volatile("vroundss $0x5,%xmm4,%xmm6,%xmm2");
|
|
asm volatile("vrndscaless $0x12,%xmm24,%xmm25,%xmm26{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 0b */
|
|
|
|
asm volatile("vroundsd $0x5,%xmm4,%xmm6,%xmm2");
|
|
asm volatile("vrndscalesd $0x12,%xmm24,%xmm25,%xmm26{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 18 */
|
|
|
|
asm volatile("vinsertf128 $0x5,%xmm4,%ymm4,%ymm6");
|
|
asm volatile("vinsertf32x4 $0x12,%xmm24,%zmm25,%zmm26{%k7}");
|
|
asm volatile("vinsertf64x2 $0x12,%xmm24,%zmm25,%zmm26{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 19 */
|
|
|
|
asm volatile("vextractf128 $0x5,%ymm4,%xmm4");
|
|
asm volatile("vextractf32x4 $0x12,%zmm25,%xmm26{%k7}");
|
|
asm volatile("vextractf64x2 $0x12,%zmm25,%xmm26{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 1a */
|
|
|
|
asm volatile("vinsertf32x8 $0x12,%ymm25,%zmm26,%zmm27{%k7}");
|
|
asm volatile("vinsertf64x4 $0x12,%ymm28,%zmm29,%zmm30{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 1b */
|
|
|
|
asm volatile("vextractf32x8 $0x12,%zmm29,%ymm30{%k7}");
|
|
asm volatile("vextractf64x4 $0x12,%zmm26,%ymm27{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 1e */
|
|
|
|
asm volatile("vpcmpud $0x12,%zmm29,%zmm30,%k5");
|
|
asm volatile("vpcmpuq $0x12,%zmm26,%zmm27,%k5");
|
|
|
|
/* AVX-512: Op code 0f 3a 1f */
|
|
|
|
asm volatile("vpcmpd $0x12,%zmm29,%zmm30,%k5");
|
|
asm volatile("vpcmpq $0x12,%zmm26,%zmm27,%k5");
|
|
|
|
/* AVX-512: Op code 0f 3a 23 */
|
|
|
|
asm volatile("vshuff32x4 $0x12,%zmm28,%zmm29,%zmm30");
|
|
asm volatile("vshuff64x2 $0x12,%zmm25,%zmm26,%zmm27");
|
|
|
|
/* AVX-512: Op code 0f 3a 25 */
|
|
|
|
asm volatile("vpternlogd $0x12,%zmm28,%zmm29,%zmm30");
|
|
asm volatile("vpternlogq $0x12,%zmm28,%zmm29,%zmm30");
|
|
|
|
/* AVX-512: Op code 0f 3a 26 */
|
|
|
|
asm volatile("vgetmantps $0x12,%zmm26,%zmm27");
|
|
asm volatile("vgetmantpd $0x12,%zmm29,%zmm30");
|
|
|
|
/* AVX-512: Op code 0f 3a 27 */
|
|
|
|
asm volatile("vgetmantss $0x12,%xmm25,%xmm26,%xmm27{%k7}");
|
|
asm volatile("vgetmantsd $0x12,%xmm28,%xmm29,%xmm30{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 38 */
|
|
|
|
asm volatile("vinserti128 $0x5,%xmm4,%ymm4,%ymm6");
|
|
asm volatile("vinserti32x4 $0x12,%xmm24,%zmm25,%zmm26{%k7}");
|
|
asm volatile("vinserti64x2 $0x12,%xmm24,%zmm25,%zmm26{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 39 */
|
|
|
|
asm volatile("vextracti128 $0x5,%ymm4,%xmm6");
|
|
asm volatile("vextracti32x4 $0x12,%zmm25,%xmm26{%k7}");
|
|
asm volatile("vextracti64x2 $0x12,%zmm25,%xmm26{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 3a */
|
|
|
|
asm volatile("vinserti32x8 $0x12,%ymm28,%zmm29,%zmm30{%k7}");
|
|
asm volatile("vinserti64x4 $0x12,%ymm25,%zmm26,%zmm27{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 3b */
|
|
|
|
asm volatile("vextracti32x8 $0x12,%zmm29,%ymm30{%k7}");
|
|
asm volatile("vextracti64x4 $0x12,%zmm26,%ymm27{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 3e */
|
|
|
|
asm volatile("vpcmpub $0x12,%zmm29,%zmm30,%k5");
|
|
asm volatile("vpcmpuw $0x12,%zmm26,%zmm27,%k5");
|
|
|
|
/* AVX-512: Op code 0f 3a 3f */
|
|
|
|
asm volatile("vpcmpb $0x12,%zmm29,%zmm30,%k5");
|
|
asm volatile("vpcmpw $0x12,%zmm26,%zmm27,%k5");
|
|
|
|
/* AVX-512: Op code 0f 3a 43 */
|
|
|
|
asm volatile("vmpsadbw $0x5,%ymm4,%ymm6,%ymm2");
|
|
asm volatile("vdbpsadbw $0x12,%zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 3a 43 */
|
|
|
|
asm volatile("vshufi32x4 $0x12,%zmm25,%zmm26,%zmm27");
|
|
asm volatile("vshufi64x2 $0x12,%zmm28,%zmm29,%zmm30");
|
|
|
|
/* AVX-512: Op code 0f 3a 50 */
|
|
|
|
asm volatile("vrangeps $0x12,%zmm25,%zmm26,%zmm27");
|
|
asm volatile("vrangepd $0x12,%zmm28,%zmm29,%zmm30");
|
|
|
|
/* AVX-512: Op code 0f 3a 51 */
|
|
|
|
asm volatile("vrangess $0x12,%xmm25,%xmm26,%xmm27");
|
|
asm volatile("vrangesd $0x12,%xmm28,%xmm29,%xmm30");
|
|
|
|
/* AVX-512: Op code 0f 3a 54 */
|
|
|
|
asm volatile("vfixupimmps $0x12,%zmm28,%zmm29,%zmm30");
|
|
asm volatile("vfixupimmpd $0x12,%zmm25,%zmm26,%zmm27");
|
|
|
|
/* AVX-512: Op code 0f 3a 55 */
|
|
|
|
asm volatile("vfixupimmss $0x12,%xmm28,%xmm29,%xmm30{%k7}");
|
|
asm volatile("vfixupimmsd $0x12,%xmm25,%xmm26,%xmm27{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 56 */
|
|
|
|
asm volatile("vreduceps $0x12,%zmm26,%zmm27");
|
|
asm volatile("vreducepd $0x12,%zmm29,%zmm30");
|
|
|
|
/* AVX-512: Op code 0f 3a 57 */
|
|
|
|
asm volatile("vreducess $0x12,%xmm25,%xmm26,%xmm27");
|
|
asm volatile("vreducesd $0x12,%xmm28,%xmm29,%xmm30");
|
|
|
|
/* AVX-512: Op code 0f 3a 66 */
|
|
|
|
asm volatile("vfpclassps $0x12,%zmm27,%k5");
|
|
asm volatile("vfpclasspd $0x12,%zmm30,%k5");
|
|
|
|
/* AVX-512: Op code 0f 3a 67 */
|
|
|
|
asm volatile("vfpclassss $0x12,%xmm27,%k5");
|
|
asm volatile("vfpclasssd $0x12,%xmm30,%k5");
|
|
|
|
/* AVX-512: Op code 0f 72 (Grp13) */
|
|
|
|
asm volatile("vprord $0x12,%zmm25,%zmm26");
|
|
asm volatile("vprorq $0x12,%zmm25,%zmm26");
|
|
asm volatile("vprold $0x12,%zmm29,%zmm30");
|
|
asm volatile("vprolq $0x12,%zmm29,%zmm30");
|
|
asm volatile("psrad $0x2,%mm6");
|
|
asm volatile("vpsrad $0x5,%ymm6,%ymm2");
|
|
asm volatile("vpsrad $0x5,%zmm26,%zmm22");
|
|
asm volatile("vpsraq $0x5,%zmm26,%zmm22");
|
|
|
|
/* AVX-512: Op code 0f 38 c6 (Grp18) */
|
|
|
|
asm volatile("vgatherpf0dps 0x7b(%r14,%zmm31,8){%k1}");
|
|
asm volatile("vgatherpf0dpd 0x7b(%r14,%ymm31,8){%k1}");
|
|
asm volatile("vgatherpf1dps 0x7b(%r14,%zmm31,8){%k1}");
|
|
asm volatile("vgatherpf1dpd 0x7b(%r14,%ymm31,8){%k1}");
|
|
asm volatile("vscatterpf0dps 0x7b(%r14,%zmm31,8){%k1}");
|
|
asm volatile("vscatterpf0dpd 0x7b(%r14,%ymm31,8){%k1}");
|
|
asm volatile("vscatterpf1dps 0x7b(%r14,%zmm31,8){%k1}");
|
|
asm volatile("vscatterpf1dpd 0x7b(%r14,%ymm31,8){%k1}");
|
|
|
|
/* AVX-512: Op code 0f 38 c7 (Grp19) */
|
|
|
|
asm volatile("vgatherpf0qps 0x7b(%r14,%zmm31,8){%k1}");
|
|
asm volatile("vgatherpf0qpd 0x7b(%r14,%zmm31,8){%k1}");
|
|
asm volatile("vgatherpf1qps 0x7b(%r14,%zmm31,8){%k1}");
|
|
asm volatile("vgatherpf1qpd 0x7b(%r14,%zmm31,8){%k1}");
|
|
asm volatile("vscatterpf0qps 0x7b(%r14,%zmm31,8){%k1}");
|
|
asm volatile("vscatterpf0qpd 0x7b(%r14,%zmm31,8){%k1}");
|
|
asm volatile("vscatterpf1qps 0x7b(%r14,%zmm31,8){%k1}");
|
|
asm volatile("vscatterpf1qpd 0x7b(%r14,%zmm31,8){%k1}");
|
|
|
|
/* AVX-512: Examples */
|
|
|
|
asm volatile("vaddpd %zmm28,%zmm29,%zmm30");
|
|
asm volatile("vaddpd %zmm28,%zmm29,%zmm30{%k7}");
|
|
asm volatile("vaddpd %zmm28,%zmm29,%zmm30{%k7}{z}");
|
|
asm volatile("vaddpd {rn-sae},%zmm28,%zmm29,%zmm30");
|
|
asm volatile("vaddpd {ru-sae},%zmm28,%zmm29,%zmm30");
|
|
asm volatile("vaddpd {rd-sae},%zmm28,%zmm29,%zmm30");
|
|
asm volatile("vaddpd {rz-sae},%zmm28,%zmm29,%zmm30");
|
|
asm volatile("vaddpd (%rcx),%zmm29,%zmm30");
|
|
asm volatile("vaddpd 0x123(%rax,%r14,8),%zmm29,%zmm30");
|
|
asm volatile("vaddpd (%rcx){1to8},%zmm29,%zmm30");
|
|
asm volatile("vaddpd 0x1fc0(%rdx),%zmm29,%zmm30");
|
|
asm volatile("vaddpd 0x3f8(%rdx){1to8},%zmm29,%zmm30");
|
|
asm volatile("vcmpeq_uqps 0x1fc(%rdx){1to16},%zmm30,%k5");
|
|
asm volatile("vcmpltsd 0x123(%rax,%r14,8),%xmm29,%k5{%k7}");
|
|
asm volatile("vcmplesd {sae},%xmm28,%xmm29,%k5{%k7}");
|
|
asm volatile("vgetmantss $0x5b,0x123(%rax,%r14,8),%xmm29,%xmm30{%k7}");
|
|
|
|
/* bndmk m64, bnd */
|
|
|
|
asm volatile("bndmk (%rax), %bnd0");
|
|
asm volatile("bndmk (%r8), %bnd0");
|
|
asm volatile("bndmk (0x12345678), %bnd0");
|
|
asm volatile("bndmk (%rax), %bnd3");
|
|
asm volatile("bndmk (%rcx,%rax,1), %bnd0");
|
|
asm volatile("bndmk 0x12345678(,%rax,1), %bnd0");
|
|
asm volatile("bndmk (%rax,%rcx,1), %bnd0");
|
|
asm volatile("bndmk (%rax,%rcx,8), %bnd0");
|
|
asm volatile("bndmk 0x12(%rax), %bnd0");
|
|
asm volatile("bndmk 0x12(%rbp), %bnd0");
|
|
asm volatile("bndmk 0x12(%rcx,%rax,1), %bnd0");
|
|
asm volatile("bndmk 0x12(%rbp,%rax,1), %bnd0");
|
|
asm volatile("bndmk 0x12(%rax,%rcx,1), %bnd0");
|
|
asm volatile("bndmk 0x12(%rax,%rcx,8), %bnd0");
|
|
asm volatile("bndmk 0x12345678(%rax), %bnd0");
|
|
asm volatile("bndmk 0x12345678(%rbp), %bnd0");
|
|
asm volatile("bndmk 0x12345678(%rcx,%rax,1), %bnd0");
|
|
asm volatile("bndmk 0x12345678(%rbp,%rax,1), %bnd0");
|
|
asm volatile("bndmk 0x12345678(%rax,%rcx,1), %bnd0");
|
|
asm volatile("bndmk 0x12345678(%rax,%rcx,8), %bnd0");
|
|
|
|
/* bndcl r/m64, bnd */
|
|
|
|
asm volatile("bndcl (%rax), %bnd0");
|
|
asm volatile("bndcl (%r8), %bnd0");
|
|
asm volatile("bndcl (0x12345678), %bnd0");
|
|
asm volatile("bndcl (%rax), %bnd3");
|
|
asm volatile("bndcl (%rcx,%rax,1), %bnd0");
|
|
asm volatile("bndcl 0x12345678(,%rax,1), %bnd0");
|
|
asm volatile("bndcl (%rax,%rcx,1), %bnd0");
|
|
asm volatile("bndcl (%rax,%rcx,8), %bnd0");
|
|
asm volatile("bndcl 0x12(%rax), %bnd0");
|
|
asm volatile("bndcl 0x12(%rbp), %bnd0");
|
|
asm volatile("bndcl 0x12(%rcx,%rax,1), %bnd0");
|
|
asm volatile("bndcl 0x12(%rbp,%rax,1), %bnd0");
|
|
asm volatile("bndcl 0x12(%rax,%rcx,1), %bnd0");
|
|
asm volatile("bndcl 0x12(%rax,%rcx,8), %bnd0");
|
|
asm volatile("bndcl 0x12345678(%rax), %bnd0");
|
|
asm volatile("bndcl 0x12345678(%rbp), %bnd0");
|
|
asm volatile("bndcl 0x12345678(%rcx,%rax,1), %bnd0");
|
|
asm volatile("bndcl 0x12345678(%rbp,%rax,1), %bnd0");
|
|
asm volatile("bndcl 0x12345678(%rax,%rcx,1), %bnd0");
|
|
asm volatile("bndcl 0x12345678(%rax,%rcx,8), %bnd0");
|
|
asm volatile("bndcl %rax, %bnd0");
|
|
|
|
/* bndcu r/m64, bnd */
|
|
|
|
asm volatile("bndcu (%rax), %bnd0");
|
|
asm volatile("bndcu (%r8), %bnd0");
|
|
asm volatile("bndcu (0x12345678), %bnd0");
|
|
asm volatile("bndcu (%rax), %bnd3");
|
|
asm volatile("bndcu (%rcx,%rax,1), %bnd0");
|
|
asm volatile("bndcu 0x12345678(,%rax,1), %bnd0");
|
|
asm volatile("bndcu (%rax,%rcx,1), %bnd0");
|
|
asm volatile("bndcu (%rax,%rcx,8), %bnd0");
|
|
asm volatile("bndcu 0x12(%rax), %bnd0");
|
|
asm volatile("bndcu 0x12(%rbp), %bnd0");
|
|
asm volatile("bndcu 0x12(%rcx,%rax,1), %bnd0");
|
|
asm volatile("bndcu 0x12(%rbp,%rax,1), %bnd0");
|
|
asm volatile("bndcu 0x12(%rax,%rcx,1), %bnd0");
|
|
asm volatile("bndcu 0x12(%rax,%rcx,8), %bnd0");
|
|
asm volatile("bndcu 0x12345678(%rax), %bnd0");
|
|
asm volatile("bndcu 0x12345678(%rbp), %bnd0");
|
|
asm volatile("bndcu 0x12345678(%rcx,%rax,1), %bnd0");
|
|
asm volatile("bndcu 0x12345678(%rbp,%rax,1), %bnd0");
|
|
asm volatile("bndcu 0x12345678(%rax,%rcx,1), %bnd0");
|
|
asm volatile("bndcu 0x12345678(%rax,%rcx,8), %bnd0");
|
|
asm volatile("bndcu %rax, %bnd0");
|
|
|
|
/* bndcn r/m64, bnd */
|
|
|
|
asm volatile("bndcn (%rax), %bnd0");
|
|
asm volatile("bndcn (%r8), %bnd0");
|
|
asm volatile("bndcn (0x12345678), %bnd0");
|
|
asm volatile("bndcn (%rax), %bnd3");
|
|
asm volatile("bndcn (%rcx,%rax,1), %bnd0");
|
|
asm volatile("bndcn 0x12345678(,%rax,1), %bnd0");
|
|
asm volatile("bndcn (%rax,%rcx,1), %bnd0");
|
|
asm volatile("bndcn (%rax,%rcx,8), %bnd0");
|
|
asm volatile("bndcn 0x12(%rax), %bnd0");
|
|
asm volatile("bndcn 0x12(%rbp), %bnd0");
|
|
asm volatile("bndcn 0x12(%rcx,%rax,1), %bnd0");
|
|
asm volatile("bndcn 0x12(%rbp,%rax,1), %bnd0");
|
|
asm volatile("bndcn 0x12(%rax,%rcx,1), %bnd0");
|
|
asm volatile("bndcn 0x12(%rax,%rcx,8), %bnd0");
|
|
asm volatile("bndcn 0x12345678(%rax), %bnd0");
|
|
asm volatile("bndcn 0x12345678(%rbp), %bnd0");
|
|
asm volatile("bndcn 0x12345678(%rcx,%rax,1), %bnd0");
|
|
asm volatile("bndcn 0x12345678(%rbp,%rax,1), %bnd0");
|
|
asm volatile("bndcn 0x12345678(%rax,%rcx,1), %bnd0");
|
|
asm volatile("bndcn 0x12345678(%rax,%rcx,8), %bnd0");
|
|
asm volatile("bndcn %rax, %bnd0");
|
|
|
|
/* bndmov m128, bnd */
|
|
|
|
asm volatile("bndmov (%rax), %bnd0");
|
|
asm volatile("bndmov (%r8), %bnd0");
|
|
asm volatile("bndmov (0x12345678), %bnd0");
|
|
asm volatile("bndmov (%rax), %bnd3");
|
|
asm volatile("bndmov (%rcx,%rax,1), %bnd0");
|
|
asm volatile("bndmov 0x12345678(,%rax,1), %bnd0");
|
|
asm volatile("bndmov (%rax,%rcx,1), %bnd0");
|
|
asm volatile("bndmov (%rax,%rcx,8), %bnd0");
|
|
asm volatile("bndmov 0x12(%rax), %bnd0");
|
|
asm volatile("bndmov 0x12(%rbp), %bnd0");
|
|
asm volatile("bndmov 0x12(%rcx,%rax,1), %bnd0");
|
|
asm volatile("bndmov 0x12(%rbp,%rax,1), %bnd0");
|
|
asm volatile("bndmov 0x12(%rax,%rcx,1), %bnd0");
|
|
asm volatile("bndmov 0x12(%rax,%rcx,8), %bnd0");
|
|
asm volatile("bndmov 0x12345678(%rax), %bnd0");
|
|
asm volatile("bndmov 0x12345678(%rbp), %bnd0");
|
|
asm volatile("bndmov 0x12345678(%rcx,%rax,1), %bnd0");
|
|
asm volatile("bndmov 0x12345678(%rbp,%rax,1), %bnd0");
|
|
asm volatile("bndmov 0x12345678(%rax,%rcx,1), %bnd0");
|
|
asm volatile("bndmov 0x12345678(%rax,%rcx,8), %bnd0");
|
|
|
|
/* bndmov bnd, m128 */
|
|
|
|
asm volatile("bndmov %bnd0, (%rax)");
|
|
asm volatile("bndmov %bnd0, (%r8)");
|
|
asm volatile("bndmov %bnd0, (0x12345678)");
|
|
asm volatile("bndmov %bnd3, (%rax)");
|
|
asm volatile("bndmov %bnd0, (%rcx,%rax,1)");
|
|
asm volatile("bndmov %bnd0, 0x12345678(,%rax,1)");
|
|
asm volatile("bndmov %bnd0, (%rax,%rcx,1)");
|
|
asm volatile("bndmov %bnd0, (%rax,%rcx,8)");
|
|
asm volatile("bndmov %bnd0, 0x12(%rax)");
|
|
asm volatile("bndmov %bnd0, 0x12(%rbp)");
|
|
asm volatile("bndmov %bnd0, 0x12(%rcx,%rax,1)");
|
|
asm volatile("bndmov %bnd0, 0x12(%rbp,%rax,1)");
|
|
asm volatile("bndmov %bnd0, 0x12(%rax,%rcx,1)");
|
|
asm volatile("bndmov %bnd0, 0x12(%rax,%rcx,8)");
|
|
asm volatile("bndmov %bnd0, 0x12345678(%rax)");
|
|
asm volatile("bndmov %bnd0, 0x12345678(%rbp)");
|
|
asm volatile("bndmov %bnd0, 0x12345678(%rcx,%rax,1)");
|
|
asm volatile("bndmov %bnd0, 0x12345678(%rbp,%rax,1)");
|
|
asm volatile("bndmov %bnd0, 0x12345678(%rax,%rcx,1)");
|
|
asm volatile("bndmov %bnd0, 0x12345678(%rax,%rcx,8)");
|
|
|
|
/* bndmov bnd2, bnd1 */
|
|
|
|
asm volatile("bndmov %bnd0, %bnd1");
|
|
asm volatile("bndmov %bnd1, %bnd0");
|
|
|
|
/* bndldx mib, bnd */
|
|
|
|
asm volatile("bndldx (%rax), %bnd0");
|
|
asm volatile("bndldx (%r8), %bnd0");
|
|
asm volatile("bndldx (0x12345678), %bnd0");
|
|
asm volatile("bndldx (%rax), %bnd3");
|
|
asm volatile("bndldx (%rcx,%rax,1), %bnd0");
|
|
asm volatile("bndldx 0x12345678(,%rax,1), %bnd0");
|
|
asm volatile("bndldx (%rax,%rcx,1), %bnd0");
|
|
asm volatile("bndldx 0x12(%rax), %bnd0");
|
|
asm volatile("bndldx 0x12(%rbp), %bnd0");
|
|
asm volatile("bndldx 0x12(%rcx,%rax,1), %bnd0");
|
|
asm volatile("bndldx 0x12(%rbp,%rax,1), %bnd0");
|
|
asm volatile("bndldx 0x12(%rax,%rcx,1), %bnd0");
|
|
asm volatile("bndldx 0x12345678(%rax), %bnd0");
|
|
asm volatile("bndldx 0x12345678(%rbp), %bnd0");
|
|
asm volatile("bndldx 0x12345678(%rcx,%rax,1), %bnd0");
|
|
asm volatile("bndldx 0x12345678(%rbp,%rax,1), %bnd0");
|
|
asm volatile("bndldx 0x12345678(%rax,%rcx,1), %bnd0");
|
|
|
|
/* bndstx bnd, mib */
|
|
|
|
asm volatile("bndstx %bnd0, (%rax)");
|
|
asm volatile("bndstx %bnd0, (%r8)");
|
|
asm volatile("bndstx %bnd0, (0x12345678)");
|
|
asm volatile("bndstx %bnd3, (%rax)");
|
|
asm volatile("bndstx %bnd0, (%rcx,%rax,1)");
|
|
asm volatile("bndstx %bnd0, 0x12345678(,%rax,1)");
|
|
asm volatile("bndstx %bnd0, (%rax,%rcx,1)");
|
|
asm volatile("bndstx %bnd0, 0x12(%rax)");
|
|
asm volatile("bndstx %bnd0, 0x12(%rbp)");
|
|
asm volatile("bndstx %bnd0, 0x12(%rcx,%rax,1)");
|
|
asm volatile("bndstx %bnd0, 0x12(%rbp,%rax,1)");
|
|
asm volatile("bndstx %bnd0, 0x12(%rax,%rcx,1)");
|
|
asm volatile("bndstx %bnd0, 0x12345678(%rax)");
|
|
asm volatile("bndstx %bnd0, 0x12345678(%rbp)");
|
|
asm volatile("bndstx %bnd0, 0x12345678(%rcx,%rax,1)");
|
|
asm volatile("bndstx %bnd0, 0x12345678(%rbp,%rax,1)");
|
|
asm volatile("bndstx %bnd0, 0x12345678(%rax,%rcx,1)");
|
|
|
|
/* bnd prefix on call, ret, jmp and all jcc */
|
|
|
|
asm volatile("bnd call label1"); /* Expecting: call unconditional 0 */
|
|
asm volatile("bnd call *(%eax)"); /* Expecting: call indirect 0 */
|
|
asm volatile("bnd ret"); /* Expecting: ret indirect 0 */
|
|
asm volatile("bnd jmp label1"); /* Expecting: jmp unconditional 0 */
|
|
asm volatile("bnd jmp label1"); /* Expecting: jmp unconditional 0 */
|
|
asm volatile("bnd jmp *(%ecx)"); /* Expecting: jmp indirect 0 */
|
|
asm volatile("bnd jne label1"); /* Expecting: jcc conditional 0 */
|
|
|
|
/* sha1rnds4 imm8, xmm2/m128, xmm1 */
|
|
|
|
asm volatile("sha1rnds4 $0x0, %xmm1, %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, %xmm7, %xmm2");
|
|
asm volatile("sha1rnds4 $0x91, %xmm8, %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, %xmm7, %xmm8");
|
|
asm volatile("sha1rnds4 $0x91, %xmm15, %xmm8");
|
|
asm volatile("sha1rnds4 $0x91, (%rax), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, (%r8), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, (0x12345678), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, (%rax), %xmm3");
|
|
asm volatile("sha1rnds4 $0x91, (%rcx,%rax,1), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12345678(,%rax,1), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, (%rax,%rcx,1), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, (%rax,%rcx,8), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12(%rax), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12(%rbp), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12(%rcx,%rax,1), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12(%rbp,%rax,1), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12(%rax,%rcx,1), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12(%rax,%rcx,8), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12345678(%rax), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12345678(%rbp), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12345678(%rcx,%rax,1), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12345678(%rbp,%rax,1), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12345678(%rax,%rcx,1), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12345678(%rax,%rcx,8), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12345678(%rax,%rcx,8), %xmm15");
|
|
|
|
/* sha1nexte xmm2/m128, xmm1 */
|
|
|
|
asm volatile("sha1nexte %xmm1, %xmm0");
|
|
asm volatile("sha1nexte %xmm7, %xmm2");
|
|
asm volatile("sha1nexte %xmm8, %xmm0");
|
|
asm volatile("sha1nexte %xmm7, %xmm8");
|
|
asm volatile("sha1nexte %xmm15, %xmm8");
|
|
asm volatile("sha1nexte (%rax), %xmm0");
|
|
asm volatile("sha1nexte (%r8), %xmm0");
|
|
asm volatile("sha1nexte (0x12345678), %xmm0");
|
|
asm volatile("sha1nexte (%rax), %xmm3");
|
|
asm volatile("sha1nexte (%rcx,%rax,1), %xmm0");
|
|
asm volatile("sha1nexte 0x12345678(,%rax,1), %xmm0");
|
|
asm volatile("sha1nexte (%rax,%rcx,1), %xmm0");
|
|
asm volatile("sha1nexte (%rax,%rcx,8), %xmm0");
|
|
asm volatile("sha1nexte 0x12(%rax), %xmm0");
|
|
asm volatile("sha1nexte 0x12(%rbp), %xmm0");
|
|
asm volatile("sha1nexte 0x12(%rcx,%rax,1), %xmm0");
|
|
asm volatile("sha1nexte 0x12(%rbp,%rax,1), %xmm0");
|
|
asm volatile("sha1nexte 0x12(%rax,%rcx,1), %xmm0");
|
|
asm volatile("sha1nexte 0x12(%rax,%rcx,8), %xmm0");
|
|
asm volatile("sha1nexte 0x12345678(%rax), %xmm0");
|
|
asm volatile("sha1nexte 0x12345678(%rbp), %xmm0");
|
|
asm volatile("sha1nexte 0x12345678(%rcx,%rax,1), %xmm0");
|
|
asm volatile("sha1nexte 0x12345678(%rbp,%rax,1), %xmm0");
|
|
asm volatile("sha1nexte 0x12345678(%rax,%rcx,1), %xmm0");
|
|
asm volatile("sha1nexte 0x12345678(%rax,%rcx,8), %xmm0");
|
|
asm volatile("sha1nexte 0x12345678(%rax,%rcx,8), %xmm15");
|
|
|
|
/* sha1msg1 xmm2/m128, xmm1 */
|
|
|
|
asm volatile("sha1msg1 %xmm1, %xmm0");
|
|
asm volatile("sha1msg1 %xmm7, %xmm2");
|
|
asm volatile("sha1msg1 %xmm8, %xmm0");
|
|
asm volatile("sha1msg1 %xmm7, %xmm8");
|
|
asm volatile("sha1msg1 %xmm15, %xmm8");
|
|
asm volatile("sha1msg1 (%rax), %xmm0");
|
|
asm volatile("sha1msg1 (%r8), %xmm0");
|
|
asm volatile("sha1msg1 (0x12345678), %xmm0");
|
|
asm volatile("sha1msg1 (%rax), %xmm3");
|
|
asm volatile("sha1msg1 (%rcx,%rax,1), %xmm0");
|
|
asm volatile("sha1msg1 0x12345678(,%rax,1), %xmm0");
|
|
asm volatile("sha1msg1 (%rax,%rcx,1), %xmm0");
|
|
asm volatile("sha1msg1 (%rax,%rcx,8), %xmm0");
|
|
asm volatile("sha1msg1 0x12(%rax), %xmm0");
|
|
asm volatile("sha1msg1 0x12(%rbp), %xmm0");
|
|
asm volatile("sha1msg1 0x12(%rcx,%rax,1), %xmm0");
|
|
asm volatile("sha1msg1 0x12(%rbp,%rax,1), %xmm0");
|
|
asm volatile("sha1msg1 0x12(%rax,%rcx,1), %xmm0");
|
|
asm volatile("sha1msg1 0x12(%rax,%rcx,8), %xmm0");
|
|
asm volatile("sha1msg1 0x12345678(%rax), %xmm0");
|
|
asm volatile("sha1msg1 0x12345678(%rbp), %xmm0");
|
|
asm volatile("sha1msg1 0x12345678(%rcx,%rax,1), %xmm0");
|
|
asm volatile("sha1msg1 0x12345678(%rbp,%rax,1), %xmm0");
|
|
asm volatile("sha1msg1 0x12345678(%rax,%rcx,1), %xmm0");
|
|
asm volatile("sha1msg1 0x12345678(%rax,%rcx,8), %xmm0");
|
|
asm volatile("sha1msg1 0x12345678(%rax,%rcx,8), %xmm15");
|
|
|
|
/* sha1msg2 xmm2/m128, xmm1 */
|
|
|
|
asm volatile("sha1msg2 %xmm1, %xmm0");
|
|
asm volatile("sha1msg2 %xmm7, %xmm2");
|
|
asm volatile("sha1msg2 %xmm8, %xmm0");
|
|
asm volatile("sha1msg2 %xmm7, %xmm8");
|
|
asm volatile("sha1msg2 %xmm15, %xmm8");
|
|
asm volatile("sha1msg2 (%rax), %xmm0");
|
|
asm volatile("sha1msg2 (%r8), %xmm0");
|
|
asm volatile("sha1msg2 (0x12345678), %xmm0");
|
|
asm volatile("sha1msg2 (%rax), %xmm3");
|
|
asm volatile("sha1msg2 (%rcx,%rax,1), %xmm0");
|
|
asm volatile("sha1msg2 0x12345678(,%rax,1), %xmm0");
|
|
asm volatile("sha1msg2 (%rax,%rcx,1), %xmm0");
|
|
asm volatile("sha1msg2 (%rax,%rcx,8), %xmm0");
|
|
asm volatile("sha1msg2 0x12(%rax), %xmm0");
|
|
asm volatile("sha1msg2 0x12(%rbp), %xmm0");
|
|
asm volatile("sha1msg2 0x12(%rcx,%rax,1), %xmm0");
|
|
asm volatile("sha1msg2 0x12(%rbp,%rax,1), %xmm0");
|
|
asm volatile("sha1msg2 0x12(%rax,%rcx,1), %xmm0");
|
|
asm volatile("sha1msg2 0x12(%rax,%rcx,8), %xmm0");
|
|
asm volatile("sha1msg2 0x12345678(%rax), %xmm0");
|
|
asm volatile("sha1msg2 0x12345678(%rbp), %xmm0");
|
|
asm volatile("sha1msg2 0x12345678(%rcx,%rax,1), %xmm0");
|
|
asm volatile("sha1msg2 0x12345678(%rbp,%rax,1), %xmm0");
|
|
asm volatile("sha1msg2 0x12345678(%rax,%rcx,1), %xmm0");
|
|
asm volatile("sha1msg2 0x12345678(%rax,%rcx,8), %xmm0");
|
|
asm volatile("sha1msg2 0x12345678(%rax,%rcx,8), %xmm15");
|
|
|
|
/* sha256rnds2 <XMM0>, xmm2/m128, xmm1 */
|
|
/* Note sha256rnds2 has an implicit operand 'xmm0' */
|
|
|
|
asm volatile("sha256rnds2 %xmm4, %xmm1");
|
|
asm volatile("sha256rnds2 %xmm7, %xmm2");
|
|
asm volatile("sha256rnds2 %xmm8, %xmm1");
|
|
asm volatile("sha256rnds2 %xmm7, %xmm8");
|
|
asm volatile("sha256rnds2 %xmm15, %xmm8");
|
|
asm volatile("sha256rnds2 (%rax), %xmm1");
|
|
asm volatile("sha256rnds2 (%r8), %xmm1");
|
|
asm volatile("sha256rnds2 (0x12345678), %xmm1");
|
|
asm volatile("sha256rnds2 (%rax), %xmm3");
|
|
asm volatile("sha256rnds2 (%rcx,%rax,1), %xmm1");
|
|
asm volatile("sha256rnds2 0x12345678(,%rax,1), %xmm1");
|
|
asm volatile("sha256rnds2 (%rax,%rcx,1), %xmm1");
|
|
asm volatile("sha256rnds2 (%rax,%rcx,8), %xmm1");
|
|
asm volatile("sha256rnds2 0x12(%rax), %xmm1");
|
|
asm volatile("sha256rnds2 0x12(%rbp), %xmm1");
|
|
asm volatile("sha256rnds2 0x12(%rcx,%rax,1), %xmm1");
|
|
asm volatile("sha256rnds2 0x12(%rbp,%rax,1), %xmm1");
|
|
asm volatile("sha256rnds2 0x12(%rax,%rcx,1), %xmm1");
|
|
asm volatile("sha256rnds2 0x12(%rax,%rcx,8), %xmm1");
|
|
asm volatile("sha256rnds2 0x12345678(%rax), %xmm1");
|
|
asm volatile("sha256rnds2 0x12345678(%rbp), %xmm1");
|
|
asm volatile("sha256rnds2 0x12345678(%rcx,%rax,1), %xmm1");
|
|
asm volatile("sha256rnds2 0x12345678(%rbp,%rax,1), %xmm1");
|
|
asm volatile("sha256rnds2 0x12345678(%rax,%rcx,1), %xmm1");
|
|
asm volatile("sha256rnds2 0x12345678(%rax,%rcx,8), %xmm1");
|
|
asm volatile("sha256rnds2 0x12345678(%rax,%rcx,8), %xmm15");
|
|
|
|
/* sha256msg1 xmm2/m128, xmm1 */
|
|
|
|
asm volatile("sha256msg1 %xmm1, %xmm0");
|
|
asm volatile("sha256msg1 %xmm7, %xmm2");
|
|
asm volatile("sha256msg1 %xmm8, %xmm0");
|
|
asm volatile("sha256msg1 %xmm7, %xmm8");
|
|
asm volatile("sha256msg1 %xmm15, %xmm8");
|
|
asm volatile("sha256msg1 (%rax), %xmm0");
|
|
asm volatile("sha256msg1 (%r8), %xmm0");
|
|
asm volatile("sha256msg1 (0x12345678), %xmm0");
|
|
asm volatile("sha256msg1 (%rax), %xmm3");
|
|
asm volatile("sha256msg1 (%rcx,%rax,1), %xmm0");
|
|
asm volatile("sha256msg1 0x12345678(,%rax,1), %xmm0");
|
|
asm volatile("sha256msg1 (%rax,%rcx,1), %xmm0");
|
|
asm volatile("sha256msg1 (%rax,%rcx,8), %xmm0");
|
|
asm volatile("sha256msg1 0x12(%rax), %xmm0");
|
|
asm volatile("sha256msg1 0x12(%rbp), %xmm0");
|
|
asm volatile("sha256msg1 0x12(%rcx,%rax,1), %xmm0");
|
|
asm volatile("sha256msg1 0x12(%rbp,%rax,1), %xmm0");
|
|
asm volatile("sha256msg1 0x12(%rax,%rcx,1), %xmm0");
|
|
asm volatile("sha256msg1 0x12(%rax,%rcx,8), %xmm0");
|
|
asm volatile("sha256msg1 0x12345678(%rax), %xmm0");
|
|
asm volatile("sha256msg1 0x12345678(%rbp), %xmm0");
|
|
asm volatile("sha256msg1 0x12345678(%rcx,%rax,1), %xmm0");
|
|
asm volatile("sha256msg1 0x12345678(%rbp,%rax,1), %xmm0");
|
|
asm volatile("sha256msg1 0x12345678(%rax,%rcx,1), %xmm0");
|
|
asm volatile("sha256msg1 0x12345678(%rax,%rcx,8), %xmm0");
|
|
asm volatile("sha256msg1 0x12345678(%rax,%rcx,8), %xmm15");
|
|
|
|
/* sha256msg2 xmm2/m128, xmm1 */
|
|
|
|
asm volatile("sha256msg2 %xmm1, %xmm0");
|
|
asm volatile("sha256msg2 %xmm7, %xmm2");
|
|
asm volatile("sha256msg2 %xmm8, %xmm0");
|
|
asm volatile("sha256msg2 %xmm7, %xmm8");
|
|
asm volatile("sha256msg2 %xmm15, %xmm8");
|
|
asm volatile("sha256msg2 (%rax), %xmm0");
|
|
asm volatile("sha256msg2 (%r8), %xmm0");
|
|
asm volatile("sha256msg2 (0x12345678), %xmm0");
|
|
asm volatile("sha256msg2 (%rax), %xmm3");
|
|
asm volatile("sha256msg2 (%rcx,%rax,1), %xmm0");
|
|
asm volatile("sha256msg2 0x12345678(,%rax,1), %xmm0");
|
|
asm volatile("sha256msg2 (%rax,%rcx,1), %xmm0");
|
|
asm volatile("sha256msg2 (%rax,%rcx,8), %xmm0");
|
|
asm volatile("sha256msg2 0x12(%rax), %xmm0");
|
|
asm volatile("sha256msg2 0x12(%rbp), %xmm0");
|
|
asm volatile("sha256msg2 0x12(%rcx,%rax,1), %xmm0");
|
|
asm volatile("sha256msg2 0x12(%rbp,%rax,1), %xmm0");
|
|
asm volatile("sha256msg2 0x12(%rax,%rcx,1), %xmm0");
|
|
asm volatile("sha256msg2 0x12(%rax,%rcx,8), %xmm0");
|
|
asm volatile("sha256msg2 0x12345678(%rax), %xmm0");
|
|
asm volatile("sha256msg2 0x12345678(%rbp), %xmm0");
|
|
asm volatile("sha256msg2 0x12345678(%rcx,%rax,1), %xmm0");
|
|
asm volatile("sha256msg2 0x12345678(%rbp,%rax,1), %xmm0");
|
|
asm volatile("sha256msg2 0x12345678(%rax,%rcx,1), %xmm0");
|
|
asm volatile("sha256msg2 0x12345678(%rax,%rcx,8), %xmm0");
|
|
asm volatile("sha256msg2 0x12345678(%rax,%rcx,8), %xmm15");
|
|
|
|
/* clflushopt m8 */
|
|
|
|
asm volatile("clflushopt (%rax)");
|
|
asm volatile("clflushopt (%r8)");
|
|
asm volatile("clflushopt (0x12345678)");
|
|
asm volatile("clflushopt 0x12345678(%rax,%rcx,8)");
|
|
asm volatile("clflushopt 0x12345678(%r8,%rcx,8)");
|
|
/* Also check instructions in the same group encoding as clflushopt */
|
|
asm volatile("clflush (%rax)");
|
|
asm volatile("clflush (%r8)");
|
|
asm volatile("sfence");
|
|
|
|
/* clwb m8 */
|
|
|
|
asm volatile("clwb (%rax)");
|
|
asm volatile("clwb (%r8)");
|
|
asm volatile("clwb (0x12345678)");
|
|
asm volatile("clwb 0x12345678(%rax,%rcx,8)");
|
|
asm volatile("clwb 0x12345678(%r8,%rcx,8)");
|
|
/* Also check instructions in the same group encoding as clwb */
|
|
asm volatile("xsaveopt (%rax)");
|
|
asm volatile("xsaveopt (%r8)");
|
|
asm volatile("mfence");
|
|
|
|
/* xsavec mem */
|
|
|
|
asm volatile("xsavec (%rax)");
|
|
asm volatile("xsavec (%r8)");
|
|
asm volatile("xsavec (0x12345678)");
|
|
asm volatile("xsavec 0x12345678(%rax,%rcx,8)");
|
|
asm volatile("xsavec 0x12345678(%r8,%rcx,8)");
|
|
|
|
/* xsaves mem */
|
|
|
|
asm volatile("xsaves (%rax)");
|
|
asm volatile("xsaves (%r8)");
|
|
asm volatile("xsaves (0x12345678)");
|
|
asm volatile("xsaves 0x12345678(%rax,%rcx,8)");
|
|
asm volatile("xsaves 0x12345678(%r8,%rcx,8)");
|
|
|
|
/* xrstors mem */
|
|
|
|
asm volatile("xrstors (%rax)");
|
|
asm volatile("xrstors (%r8)");
|
|
asm volatile("xrstors (0x12345678)");
|
|
asm volatile("xrstors 0x12345678(%rax,%rcx,8)");
|
|
asm volatile("xrstors 0x12345678(%r8,%rcx,8)");
|
|
|
|
/* ptwrite */
|
|
|
|
asm volatile("ptwrite (%rax)");
|
|
asm volatile("ptwrite (%r8)");
|
|
asm volatile("ptwrite (0x12345678)");
|
|
asm volatile("ptwrite 0x12345678(%rax,%rcx,8)");
|
|
asm volatile("ptwrite 0x12345678(%r8,%rcx,8)");
|
|
|
|
asm volatile("ptwritel (%rax)");
|
|
asm volatile("ptwritel (%r8)");
|
|
asm volatile("ptwritel (0x12345678)");
|
|
asm volatile("ptwritel 0x12345678(%rax,%rcx,8)");
|
|
asm volatile("ptwritel 0x12345678(%r8,%rcx,8)");
|
|
|
|
asm volatile("ptwriteq (%rax)");
|
|
asm volatile("ptwriteq (%r8)");
|
|
asm volatile("ptwriteq (0x12345678)");
|
|
asm volatile("ptwriteq 0x12345678(%rax,%rcx,8)");
|
|
asm volatile("ptwriteq 0x12345678(%r8,%rcx,8)");
|
|
|
|
#else /* #ifdef __x86_64__ */
|
|
|
|
/* bound r32, mem (same op code as EVEX prefix) */
|
|
|
|
asm volatile("bound %eax, 0x12345678(%ecx)");
|
|
asm volatile("bound %ecx, 0x12345678(%eax)");
|
|
asm volatile("bound %edx, 0x12345678(%eax)");
|
|
asm volatile("bound %ebx, 0x12345678(%eax)");
|
|
asm volatile("bound %esp, 0x12345678(%eax)");
|
|
asm volatile("bound %ebp, 0x12345678(%eax)");
|
|
asm volatile("bound %esi, 0x12345678(%eax)");
|
|
asm volatile("bound %edi, 0x12345678(%eax)");
|
|
asm volatile("bound %ecx, (%eax)");
|
|
asm volatile("bound %eax, (0x12345678)");
|
|
asm volatile("bound %edx, (%ecx,%eax,1)");
|
|
asm volatile("bound %edx, 0x12345678(,%eax,1)");
|
|
asm volatile("bound %edx, (%eax,%ecx,1)");
|
|
asm volatile("bound %edx, (%eax,%ecx,8)");
|
|
asm volatile("bound %edx, 0x12(%eax)");
|
|
asm volatile("bound %edx, 0x12(%ebp)");
|
|
asm volatile("bound %edx, 0x12(%ecx,%eax,1)");
|
|
asm volatile("bound %edx, 0x12(%ebp,%eax,1)");
|
|
asm volatile("bound %edx, 0x12(%eax,%ecx,1)");
|
|
asm volatile("bound %edx, 0x12(%eax,%ecx,8)");
|
|
asm volatile("bound %edx, 0x12345678(%eax)");
|
|
asm volatile("bound %edx, 0x12345678(%ebp)");
|
|
asm volatile("bound %edx, 0x12345678(%ecx,%eax,1)");
|
|
asm volatile("bound %edx, 0x12345678(%ebp,%eax,1)");
|
|
asm volatile("bound %edx, 0x12345678(%eax,%ecx,1)");
|
|
asm volatile("bound %edx, 0x12345678(%eax,%ecx,8)");
|
|
|
|
/* bound r16, mem (same op code as EVEX prefix) */
|
|
|
|
asm volatile("bound %ax, 0x12345678(%ecx)");
|
|
asm volatile("bound %cx, 0x12345678(%eax)");
|
|
asm volatile("bound %dx, 0x12345678(%eax)");
|
|
asm volatile("bound %bx, 0x12345678(%eax)");
|
|
asm volatile("bound %sp, 0x12345678(%eax)");
|
|
asm volatile("bound %bp, 0x12345678(%eax)");
|
|
asm volatile("bound %si, 0x12345678(%eax)");
|
|
asm volatile("bound %di, 0x12345678(%eax)");
|
|
asm volatile("bound %cx, (%eax)");
|
|
asm volatile("bound %ax, (0x12345678)");
|
|
asm volatile("bound %dx, (%ecx,%eax,1)");
|
|
asm volatile("bound %dx, 0x12345678(,%eax,1)");
|
|
asm volatile("bound %dx, (%eax,%ecx,1)");
|
|
asm volatile("bound %dx, (%eax,%ecx,8)");
|
|
asm volatile("bound %dx, 0x12(%eax)");
|
|
asm volatile("bound %dx, 0x12(%ebp)");
|
|
asm volatile("bound %dx, 0x12(%ecx,%eax,1)");
|
|
asm volatile("bound %dx, 0x12(%ebp,%eax,1)");
|
|
asm volatile("bound %dx, 0x12(%eax,%ecx,1)");
|
|
asm volatile("bound %dx, 0x12(%eax,%ecx,8)");
|
|
asm volatile("bound %dx, 0x12345678(%eax)");
|
|
asm volatile("bound %dx, 0x12345678(%ebp)");
|
|
asm volatile("bound %dx, 0x12345678(%ecx,%eax,1)");
|
|
asm volatile("bound %dx, 0x12345678(%ebp,%eax,1)");
|
|
asm volatile("bound %dx, 0x12345678(%eax,%ecx,1)");
|
|
asm volatile("bound %dx, 0x12345678(%eax,%ecx,8)");
|
|
|
|
/* AVX-512: Instructions with the same op codes as Mask Instructions */
|
|
|
|
asm volatile("cmovno %eax,%ebx");
|
|
asm volatile("cmovno 0x12345678(%eax),%ecx");
|
|
asm volatile("cmovno 0x12345678(%eax),%cx");
|
|
|
|
asm volatile("cmove %eax,%ebx");
|
|
asm volatile("cmove 0x12345678(%eax),%ecx");
|
|
asm volatile("cmove 0x12345678(%eax),%cx");
|
|
|
|
asm volatile("seto 0x12345678(%eax)");
|
|
asm volatile("setno 0x12345678(%eax)");
|
|
asm volatile("setb 0x12345678(%eax)");
|
|
asm volatile("setc 0x12345678(%eax)");
|
|
asm volatile("setnae 0x12345678(%eax)");
|
|
asm volatile("setae 0x12345678(%eax)");
|
|
asm volatile("setnb 0x12345678(%eax)");
|
|
asm volatile("setnc 0x12345678(%eax)");
|
|
asm volatile("sets 0x12345678(%eax)");
|
|
asm volatile("setns 0x12345678(%eax)");
|
|
|
|
/* AVX-512: Mask Instructions */
|
|
|
|
asm volatile("kandw %k7,%k6,%k5");
|
|
asm volatile("kandq %k7,%k6,%k5");
|
|
asm volatile("kandb %k7,%k6,%k5");
|
|
asm volatile("kandd %k7,%k6,%k5");
|
|
|
|
asm volatile("kandnw %k7,%k6,%k5");
|
|
asm volatile("kandnq %k7,%k6,%k5");
|
|
asm volatile("kandnb %k7,%k6,%k5");
|
|
asm volatile("kandnd %k7,%k6,%k5");
|
|
|
|
asm volatile("knotw %k7,%k6");
|
|
asm volatile("knotq %k7,%k6");
|
|
asm volatile("knotb %k7,%k6");
|
|
asm volatile("knotd %k7,%k6");
|
|
|
|
asm volatile("korw %k7,%k6,%k5");
|
|
asm volatile("korq %k7,%k6,%k5");
|
|
asm volatile("korb %k7,%k6,%k5");
|
|
asm volatile("kord %k7,%k6,%k5");
|
|
|
|
asm volatile("kxnorw %k7,%k6,%k5");
|
|
asm volatile("kxnorq %k7,%k6,%k5");
|
|
asm volatile("kxnorb %k7,%k6,%k5");
|
|
asm volatile("kxnord %k7,%k6,%k5");
|
|
|
|
asm volatile("kxorw %k7,%k6,%k5");
|
|
asm volatile("kxorq %k7,%k6,%k5");
|
|
asm volatile("kxorb %k7,%k6,%k5");
|
|
asm volatile("kxord %k7,%k6,%k5");
|
|
|
|
asm volatile("kaddw %k7,%k6,%k5");
|
|
asm volatile("kaddq %k7,%k6,%k5");
|
|
asm volatile("kaddb %k7,%k6,%k5");
|
|
asm volatile("kaddd %k7,%k6,%k5");
|
|
|
|
asm volatile("kunpckbw %k7,%k6,%k5");
|
|
asm volatile("kunpckwd %k7,%k6,%k5");
|
|
asm volatile("kunpckdq %k7,%k6,%k5");
|
|
|
|
asm volatile("kmovw %k6,%k5");
|
|
asm volatile("kmovw (%ecx),%k5");
|
|
asm volatile("kmovw 0x123(%eax,%ecx,8),%k5");
|
|
asm volatile("kmovw %k5,(%ecx)");
|
|
asm volatile("kmovw %k5,0x123(%eax,%ecx,8)");
|
|
asm volatile("kmovw %eax,%k5");
|
|
asm volatile("kmovw %ebp,%k5");
|
|
asm volatile("kmovw %k5,%eax");
|
|
asm volatile("kmovw %k5,%ebp");
|
|
|
|
asm volatile("kmovq %k6,%k5");
|
|
asm volatile("kmovq (%ecx),%k5");
|
|
asm volatile("kmovq 0x123(%eax,%ecx,8),%k5");
|
|
asm volatile("kmovq %k5,(%ecx)");
|
|
asm volatile("kmovq %k5,0x123(%eax,%ecx,8)");
|
|
|
|
asm volatile("kmovb %k6,%k5");
|
|
asm volatile("kmovb (%ecx),%k5");
|
|
asm volatile("kmovb 0x123(%eax,%ecx,8),%k5");
|
|
asm volatile("kmovb %k5,(%ecx)");
|
|
asm volatile("kmovb %k5,0x123(%eax,%ecx,8)");
|
|
asm volatile("kmovb %eax,%k5");
|
|
asm volatile("kmovb %ebp,%k5");
|
|
asm volatile("kmovb %k5,%eax");
|
|
asm volatile("kmovb %k5,%ebp");
|
|
|
|
asm volatile("kmovd %k6,%k5");
|
|
asm volatile("kmovd (%ecx),%k5");
|
|
asm volatile("kmovd 0x123(%eax,%ecx,8),%k5");
|
|
asm volatile("kmovd %k5,(%ecx)");
|
|
asm volatile("kmovd %k5,0x123(%eax,%ecx,8)");
|
|
asm volatile("kmovd %eax,%k5");
|
|
asm volatile("kmovd %ebp,%k5");
|
|
asm volatile("kmovd %k5,%eax");
|
|
asm volatile("kmovd %k5,%ebp");
|
|
|
|
asm volatile("kortestw %k6,%k5");
|
|
asm volatile("kortestq %k6,%k5");
|
|
asm volatile("kortestb %k6,%k5");
|
|
asm volatile("kortestd %k6,%k5");
|
|
|
|
asm volatile("ktestw %k6,%k5");
|
|
asm volatile("ktestq %k6,%k5");
|
|
asm volatile("ktestb %k6,%k5");
|
|
asm volatile("ktestd %k6,%k5");
|
|
|
|
asm volatile("kshiftrw $0x12,%k6,%k5");
|
|
asm volatile("kshiftrq $0x5b,%k6,%k5");
|
|
asm volatile("kshiftlw $0x12,%k6,%k5");
|
|
asm volatile("kshiftlq $0x5b,%k6,%k5");
|
|
|
|
/* AVX-512: Op code 0f 5b */
|
|
asm volatile("vcvtdq2ps %xmm5,%xmm6");
|
|
asm volatile("vcvtqq2ps %zmm5,%ymm6{%k7}");
|
|
asm volatile("vcvtps2dq %xmm5,%xmm6");
|
|
asm volatile("vcvttps2dq %xmm5,%xmm6");
|
|
|
|
/* AVX-512: Op code 0f 6f */
|
|
|
|
asm volatile("movq %mm0,%mm4");
|
|
asm volatile("vmovdqa %ymm4,%ymm6");
|
|
asm volatile("vmovdqa32 %zmm5,%zmm6");
|
|
asm volatile("vmovdqa64 %zmm5,%zmm6");
|
|
asm volatile("vmovdqu %ymm4,%ymm6");
|
|
asm volatile("vmovdqu32 %zmm5,%zmm6");
|
|
asm volatile("vmovdqu64 %zmm5,%zmm6");
|
|
asm volatile("vmovdqu8 %zmm5,%zmm6");
|
|
asm volatile("vmovdqu16 %zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 78 */
|
|
|
|
asm volatile("vmread %eax,%ebx");
|
|
asm volatile("vcvttps2udq %zmm5,%zmm6");
|
|
asm volatile("vcvttpd2udq %zmm5,%ymm6{%k7}");
|
|
asm volatile("vcvttsd2usi %xmm6,%eax");
|
|
asm volatile("vcvttss2usi %xmm6,%eax");
|
|
asm volatile("vcvttps2uqq %ymm5,%zmm6{%k7}");
|
|
asm volatile("vcvttpd2uqq %zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 79 */
|
|
|
|
asm volatile("vmwrite %eax,%ebx");
|
|
asm volatile("vcvtps2udq %zmm5,%zmm6");
|
|
asm volatile("vcvtpd2udq %zmm5,%ymm6{%k7}");
|
|
asm volatile("vcvtsd2usi %xmm6,%eax");
|
|
asm volatile("vcvtss2usi %xmm6,%eax");
|
|
asm volatile("vcvtps2uqq %ymm5,%zmm6{%k7}");
|
|
asm volatile("vcvtpd2uqq %zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 7a */
|
|
|
|
asm volatile("vcvtudq2pd %ymm5,%zmm6{%k7}");
|
|
asm volatile("vcvtuqq2pd %zmm5,%zmm6");
|
|
asm volatile("vcvtudq2ps %zmm5,%zmm6");
|
|
asm volatile("vcvtuqq2ps %zmm5,%ymm6{%k7}");
|
|
asm volatile("vcvttps2qq %ymm5,%zmm6{%k7}");
|
|
asm volatile("vcvttpd2qq %zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 7b */
|
|
|
|
asm volatile("vcvtusi2sd %eax,%xmm5,%xmm6");
|
|
asm volatile("vcvtusi2ss %eax,%xmm5,%xmm6");
|
|
asm volatile("vcvtps2qq %ymm5,%zmm6{%k7}");
|
|
asm volatile("vcvtpd2qq %zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 7f */
|
|
|
|
asm volatile("movq.s %mm0,%mm4");
|
|
asm volatile("vmovdqa.s %ymm5,%ymm6");
|
|
asm volatile("vmovdqa32.s %zmm5,%zmm6");
|
|
asm volatile("vmovdqa64.s %zmm5,%zmm6");
|
|
asm volatile("vmovdqu.s %ymm5,%ymm6");
|
|
asm volatile("vmovdqu32.s %zmm5,%zmm6");
|
|
asm volatile("vmovdqu64.s %zmm5,%zmm6");
|
|
asm volatile("vmovdqu8.s %zmm5,%zmm6");
|
|
asm volatile("vmovdqu16.s %zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f db */
|
|
|
|
asm volatile("pand %mm1,%mm2");
|
|
asm volatile("pand %xmm1,%xmm2");
|
|
asm volatile("vpand %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpandd %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vpandq %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f df */
|
|
|
|
asm volatile("pandn %mm1,%mm2");
|
|
asm volatile("pandn %xmm1,%xmm2");
|
|
asm volatile("vpandn %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpandnd %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vpandnq %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f e6 */
|
|
|
|
asm volatile("vcvttpd2dq %xmm1,%xmm2");
|
|
asm volatile("vcvtdq2pd %xmm5,%xmm6");
|
|
asm volatile("vcvtdq2pd %ymm5,%zmm6{%k7}");
|
|
asm volatile("vcvtqq2pd %zmm5,%zmm6");
|
|
asm volatile("vcvtpd2dq %xmm1,%xmm2");
|
|
|
|
/* AVX-512: Op code 0f eb */
|
|
|
|
asm volatile("por %mm4,%mm6");
|
|
asm volatile("vpor %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpord %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vporq %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f ef */
|
|
|
|
asm volatile("pxor %mm4,%mm6");
|
|
asm volatile("vpxor %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpxord %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vpxorq %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 10 */
|
|
|
|
asm volatile("pblendvb %xmm1,%xmm0");
|
|
asm volatile("vpsrlvw %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vpmovuswb %zmm5,%ymm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 11 */
|
|
|
|
asm volatile("vpmovusdb %zmm5,%xmm6{%k7}");
|
|
asm volatile("vpsravw %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 12 */
|
|
|
|
asm volatile("vpmovusqb %zmm5,%xmm6{%k7}");
|
|
asm volatile("vpsllvw %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 13 */
|
|
|
|
asm volatile("vcvtph2ps %xmm3,%ymm5");
|
|
asm volatile("vcvtph2ps %ymm5,%zmm6{%k7}");
|
|
asm volatile("vpmovusdw %zmm5,%ymm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 14 */
|
|
|
|
asm volatile("blendvps %xmm1,%xmm0");
|
|
asm volatile("vpmovusqw %zmm5,%xmm6{%k7}");
|
|
asm volatile("vprorvd %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vprorvq %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 15 */
|
|
|
|
asm volatile("blendvpd %xmm1,%xmm0");
|
|
asm volatile("vpmovusqd %zmm5,%ymm6{%k7}");
|
|
asm volatile("vprolvd %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vprolvq %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 16 */
|
|
|
|
asm volatile("vpermps %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpermps %ymm4,%ymm6,%ymm2{%k7}");
|
|
asm volatile("vpermpd %ymm4,%ymm6,%ymm2{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 19 */
|
|
|
|
asm volatile("vbroadcastsd %xmm4,%ymm6");
|
|
asm volatile("vbroadcastf32x2 %xmm7,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 1a */
|
|
|
|
asm volatile("vbroadcastf128 (%ecx),%ymm4");
|
|
asm volatile("vbroadcastf32x4 (%ecx),%zmm6");
|
|
asm volatile("vbroadcastf64x2 (%ecx),%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 1b */
|
|
|
|
asm volatile("vbroadcastf32x8 (%ecx),%zmm6");
|
|
asm volatile("vbroadcastf64x4 (%ecx),%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 1f */
|
|
|
|
asm volatile("vpabsq %zmm4,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 20 */
|
|
|
|
asm volatile("vpmovsxbw %xmm4,%xmm5");
|
|
asm volatile("vpmovswb %zmm5,%ymm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 21 */
|
|
|
|
asm volatile("vpmovsxbd %xmm4,%ymm6");
|
|
asm volatile("vpmovsdb %zmm5,%xmm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 22 */
|
|
|
|
asm volatile("vpmovsxbq %xmm4,%ymm4");
|
|
asm volatile("vpmovsqb %zmm5,%xmm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 23 */
|
|
|
|
asm volatile("vpmovsxwd %xmm4,%ymm4");
|
|
asm volatile("vpmovsdw %zmm5,%ymm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 24 */
|
|
|
|
asm volatile("vpmovsxwq %xmm4,%ymm6");
|
|
asm volatile("vpmovsqw %zmm5,%xmm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 25 */
|
|
|
|
asm volatile("vpmovsxdq %xmm4,%ymm4");
|
|
asm volatile("vpmovsqd %zmm5,%ymm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 26 */
|
|
|
|
asm volatile("vptestmb %zmm5,%zmm6,%k5");
|
|
asm volatile("vptestmw %zmm5,%zmm6,%k5");
|
|
asm volatile("vptestnmb %zmm4,%zmm5,%k5");
|
|
asm volatile("vptestnmw %zmm4,%zmm5,%k5");
|
|
|
|
/* AVX-512: Op code 0f 38 27 */
|
|
|
|
asm volatile("vptestmd %zmm5,%zmm6,%k5");
|
|
asm volatile("vptestmq %zmm5,%zmm6,%k5");
|
|
asm volatile("vptestnmd %zmm4,%zmm5,%k5");
|
|
asm volatile("vptestnmq %zmm4,%zmm5,%k5");
|
|
|
|
/* AVX-512: Op code 0f 38 28 */
|
|
|
|
asm volatile("vpmuldq %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpmovm2b %k5,%zmm6");
|
|
asm volatile("vpmovm2w %k5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 29 */
|
|
|
|
asm volatile("vpcmpeqq %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpmovb2m %zmm6,%k5");
|
|
asm volatile("vpmovw2m %zmm6,%k5");
|
|
|
|
/* AVX-512: Op code 0f 38 2a */
|
|
|
|
asm volatile("vmovntdqa (%ecx),%ymm4");
|
|
asm volatile("vpbroadcastmb2q %k6,%zmm1");
|
|
|
|
/* AVX-512: Op code 0f 38 2c */
|
|
|
|
asm volatile("vmaskmovps (%ecx),%ymm4,%ymm6");
|
|
asm volatile("vscalefps %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vscalefpd %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 2d */
|
|
|
|
asm volatile("vmaskmovpd (%ecx),%ymm4,%ymm6");
|
|
asm volatile("vscalefss %xmm4,%xmm5,%xmm6{%k7}");
|
|
asm volatile("vscalefsd %xmm4,%xmm5,%xmm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 30 */
|
|
|
|
asm volatile("vpmovzxbw %xmm4,%ymm4");
|
|
asm volatile("vpmovwb %zmm5,%ymm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 31 */
|
|
|
|
asm volatile("vpmovzxbd %xmm4,%ymm6");
|
|
asm volatile("vpmovdb %zmm5,%xmm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 32 */
|
|
|
|
asm volatile("vpmovzxbq %xmm4,%ymm4");
|
|
asm volatile("vpmovqb %zmm5,%xmm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 33 */
|
|
|
|
asm volatile("vpmovzxwd %xmm4,%ymm4");
|
|
asm volatile("vpmovdw %zmm5,%ymm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 34 */
|
|
|
|
asm volatile("vpmovzxwq %xmm4,%ymm6");
|
|
asm volatile("vpmovqw %zmm5,%xmm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 35 */
|
|
|
|
asm volatile("vpmovzxdq %xmm4,%ymm4");
|
|
asm volatile("vpmovqd %zmm5,%ymm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 36 */
|
|
|
|
asm volatile("vpermd %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpermd %ymm4,%ymm6,%ymm2{%k7}");
|
|
asm volatile("vpermq %ymm4,%ymm6,%ymm2{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 38 */
|
|
|
|
asm volatile("vpminsb %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpmovm2d %k5,%zmm6");
|
|
asm volatile("vpmovm2q %k5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 39 */
|
|
|
|
asm volatile("vpminsd %xmm1,%xmm2,%xmm3");
|
|
asm volatile("vpminsd %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vpminsq %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vpmovd2m %zmm6,%k5");
|
|
asm volatile("vpmovq2m %zmm6,%k5");
|
|
|
|
/* AVX-512: Op code 0f 38 3a */
|
|
|
|
asm volatile("vpminuw %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpbroadcastmw2d %k6,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 3b */
|
|
|
|
asm volatile("vpminud %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpminud %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vpminuq %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 3d */
|
|
|
|
asm volatile("vpmaxsd %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpmaxsd %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vpmaxsq %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 3f */
|
|
|
|
asm volatile("vpmaxud %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpmaxud %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vpmaxuq %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 40 */
|
|
|
|
asm volatile("vpmulld %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpmulld %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vpmullq %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 42 */
|
|
|
|
asm volatile("vgetexpps %zmm5,%zmm6");
|
|
asm volatile("vgetexppd %zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 43 */
|
|
|
|
asm volatile("vgetexpss %xmm4,%xmm5,%xmm6{%k7}");
|
|
asm volatile("vgetexpsd %xmm2,%xmm3,%xmm4{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 44 */
|
|
|
|
asm volatile("vplzcntd %zmm5,%zmm6");
|
|
asm volatile("vplzcntq %zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 46 */
|
|
|
|
asm volatile("vpsravd %ymm4,%ymm6,%ymm2");
|
|
asm volatile("vpsravd %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vpsravq %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 4c */
|
|
|
|
asm volatile("vrcp14ps %zmm5,%zmm6");
|
|
asm volatile("vrcp14pd %zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 4d */
|
|
|
|
asm volatile("vrcp14ss %xmm4,%xmm5,%xmm6{%k7}");
|
|
asm volatile("vrcp14sd %xmm4,%xmm5,%xmm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 4e */
|
|
|
|
asm volatile("vrsqrt14ps %zmm5,%zmm6");
|
|
asm volatile("vrsqrt14pd %zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 4f */
|
|
|
|
asm volatile("vrsqrt14ss %xmm4,%xmm5,%xmm6{%k7}");
|
|
asm volatile("vrsqrt14sd %xmm4,%xmm5,%xmm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 59 */
|
|
|
|
asm volatile("vpbroadcastq %xmm4,%xmm6");
|
|
asm volatile("vbroadcasti32x2 %xmm7,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 5a */
|
|
|
|
asm volatile("vbroadcasti128 (%ecx),%ymm4");
|
|
asm volatile("vbroadcasti32x4 (%ecx),%zmm6");
|
|
asm volatile("vbroadcasti64x2 (%ecx),%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 5b */
|
|
|
|
asm volatile("vbroadcasti32x8 (%ecx),%zmm6");
|
|
asm volatile("vbroadcasti64x4 (%ecx),%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 64 */
|
|
|
|
asm volatile("vpblendmd %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vpblendmq %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 65 */
|
|
|
|
asm volatile("vblendmps %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vblendmpd %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 66 */
|
|
|
|
asm volatile("vpblendmb %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vpblendmw %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 75 */
|
|
|
|
asm volatile("vpermi2b %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vpermi2w %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 76 */
|
|
|
|
asm volatile("vpermi2d %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vpermi2q %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 77 */
|
|
|
|
asm volatile("vpermi2ps %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vpermi2pd %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 7a */
|
|
|
|
asm volatile("vpbroadcastb %eax,%xmm3");
|
|
|
|
/* AVX-512: Op code 0f 38 7b */
|
|
|
|
asm volatile("vpbroadcastw %eax,%xmm3");
|
|
|
|
/* AVX-512: Op code 0f 38 7c */
|
|
|
|
asm volatile("vpbroadcastd %eax,%xmm3");
|
|
|
|
/* AVX-512: Op code 0f 38 7d */
|
|
|
|
asm volatile("vpermt2b %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vpermt2w %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 7e */
|
|
|
|
asm volatile("vpermt2d %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vpermt2q %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 7f */
|
|
|
|
asm volatile("vpermt2ps %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vpermt2pd %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 83 */
|
|
|
|
asm volatile("vpmultishiftqb %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 88 */
|
|
|
|
asm volatile("vexpandps (%ecx),%zmm6");
|
|
asm volatile("vexpandpd (%ecx),%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 89 */
|
|
|
|
asm volatile("vpexpandd (%ecx),%zmm6");
|
|
asm volatile("vpexpandq (%ecx),%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 8a */
|
|
|
|
asm volatile("vcompressps %zmm6,(%ecx)");
|
|
asm volatile("vcompresspd %zmm6,(%ecx)");
|
|
|
|
/* AVX-512: Op code 0f 38 8b */
|
|
|
|
asm volatile("vpcompressd %zmm6,(%ecx)");
|
|
asm volatile("vpcompressq %zmm6,(%ecx)");
|
|
|
|
/* AVX-512: Op code 0f 38 8d */
|
|
|
|
asm volatile("vpermb %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vpermw %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 90 */
|
|
|
|
asm volatile("vpgatherdd %xmm2,0x02(%ebp,%xmm7,2),%xmm1");
|
|
asm volatile("vpgatherdq %xmm2,0x04(%ebp,%xmm7,2),%xmm1");
|
|
asm volatile("vpgatherdd 0x7b(%ebp,%zmm7,8),%zmm6{%k1}");
|
|
asm volatile("vpgatherdq 0x7b(%ebp,%ymm7,8),%zmm6{%k1}");
|
|
|
|
/* AVX-512: Op code 0f 38 91 */
|
|
|
|
asm volatile("vpgatherqd %xmm2,0x02(%ebp,%xmm7,2),%xmm1");
|
|
asm volatile("vpgatherqq %xmm2,0x02(%ebp,%xmm7,2),%xmm1");
|
|
asm volatile("vpgatherqd 0x7b(%ebp,%zmm7,8),%ymm6{%k1}");
|
|
asm volatile("vpgatherqq 0x7b(%ebp,%zmm7,8),%zmm6{%k1}");
|
|
|
|
/* AVX-512: Op code 0f 38 a0 */
|
|
|
|
asm volatile("vpscatterdd %zmm6,0x7b(%ebp,%zmm7,8){%k1}");
|
|
asm volatile("vpscatterdq %zmm6,0x7b(%ebp,%ymm7,8){%k1}");
|
|
|
|
/* AVX-512: Op code 0f 38 a1 */
|
|
|
|
asm volatile("vpscatterqd %ymm6,0x7b(%ebp,%zmm7,8){%k1}");
|
|
asm volatile("vpscatterqq %ymm6,0x7b(%ebp,%ymm7,8){%k1}");
|
|
|
|
/* AVX-512: Op code 0f 38 a2 */
|
|
|
|
asm volatile("vscatterdps %zmm6,0x7b(%ebp,%zmm7,8){%k1}");
|
|
asm volatile("vscatterdpd %zmm6,0x7b(%ebp,%ymm7,8){%k1}");
|
|
|
|
/* AVX-512: Op code 0f 38 a3 */
|
|
|
|
asm volatile("vscatterqps %ymm6,0x7b(%ebp,%zmm7,8){%k1}");
|
|
asm volatile("vscatterqpd %zmm6,0x7b(%ebp,%zmm7,8){%k1}");
|
|
|
|
/* AVX-512: Op code 0f 38 b4 */
|
|
|
|
asm volatile("vpmadd52luq %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 b5 */
|
|
|
|
asm volatile("vpmadd52huq %zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 c4 */
|
|
|
|
asm volatile("vpconflictd %zmm5,%zmm6");
|
|
asm volatile("vpconflictq %zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 38 c8 */
|
|
|
|
asm volatile("vexp2ps %zmm6,%zmm7");
|
|
asm volatile("vexp2pd %zmm6,%zmm7");
|
|
|
|
/* AVX-512: Op code 0f 38 ca */
|
|
|
|
asm volatile("vrcp28ps %zmm6,%zmm7");
|
|
asm volatile("vrcp28pd %zmm6,%zmm7");
|
|
|
|
/* AVX-512: Op code 0f 38 cb */
|
|
|
|
asm volatile("vrcp28ss %xmm5,%xmm6,%xmm7{%k7}");
|
|
asm volatile("vrcp28sd %xmm5,%xmm6,%xmm7{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 38 cc */
|
|
|
|
asm volatile("vrsqrt28ps %zmm6,%zmm7");
|
|
asm volatile("vrsqrt28pd %zmm6,%zmm7");
|
|
|
|
/* AVX-512: Op code 0f 38 cd */
|
|
|
|
asm volatile("vrsqrt28ss %xmm5,%xmm6,%xmm7{%k7}");
|
|
asm volatile("vrsqrt28sd %xmm5,%xmm6,%xmm7{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 03 */
|
|
|
|
asm volatile("valignd $0x12,%zmm5,%zmm6,%zmm7");
|
|
asm volatile("valignq $0x12,%zmm5,%zmm6,%zmm7");
|
|
|
|
/* AVX-512: Op code 0f 3a 08 */
|
|
|
|
asm volatile("vroundps $0x5,%ymm6,%ymm2");
|
|
asm volatile("vrndscaleps $0x12,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 3a 09 */
|
|
|
|
asm volatile("vroundpd $0x5,%ymm6,%ymm2");
|
|
asm volatile("vrndscalepd $0x12,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 3a 0a */
|
|
|
|
asm volatile("vroundss $0x5,%xmm4,%xmm6,%xmm2");
|
|
asm volatile("vrndscaless $0x12,%xmm4,%xmm5,%xmm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 0b */
|
|
|
|
asm volatile("vroundsd $0x5,%xmm4,%xmm6,%xmm2");
|
|
asm volatile("vrndscalesd $0x12,%xmm4,%xmm5,%xmm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 18 */
|
|
|
|
asm volatile("vinsertf128 $0x5,%xmm4,%ymm4,%ymm6");
|
|
asm volatile("vinsertf32x4 $0x12,%xmm4,%zmm5,%zmm6{%k7}");
|
|
asm volatile("vinsertf64x2 $0x12,%xmm4,%zmm5,%zmm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 19 */
|
|
|
|
asm volatile("vextractf128 $0x5,%ymm4,%xmm4");
|
|
asm volatile("vextractf32x4 $0x12,%zmm5,%xmm6{%k7}");
|
|
asm volatile("vextractf64x2 $0x12,%zmm5,%xmm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 1a */
|
|
|
|
asm volatile("vinsertf32x8 $0x12,%ymm5,%zmm6,%zmm7{%k7}");
|
|
asm volatile("vinsertf64x4 $0x12,%ymm5,%zmm6,%zmm7{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 1b */
|
|
|
|
asm volatile("vextractf32x8 $0x12,%zmm6,%ymm7{%k7}");
|
|
asm volatile("vextractf64x4 $0x12,%zmm6,%ymm7{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 1e */
|
|
|
|
asm volatile("vpcmpud $0x12,%zmm6,%zmm7,%k5");
|
|
asm volatile("vpcmpuq $0x12,%zmm6,%zmm7,%k5");
|
|
|
|
/* AVX-512: Op code 0f 3a 1f */
|
|
|
|
asm volatile("vpcmpd $0x12,%zmm6,%zmm7,%k5");
|
|
asm volatile("vpcmpq $0x12,%zmm6,%zmm7,%k5");
|
|
|
|
/* AVX-512: Op code 0f 3a 23 */
|
|
|
|
asm volatile("vshuff32x4 $0x12,%zmm5,%zmm6,%zmm7");
|
|
asm volatile("vshuff64x2 $0x12,%zmm5,%zmm6,%zmm7");
|
|
|
|
/* AVX-512: Op code 0f 3a 25 */
|
|
|
|
asm volatile("vpternlogd $0x12,%zmm5,%zmm6,%zmm7");
|
|
asm volatile("vpternlogq $0x12,%zmm5,%zmm6,%zmm7");
|
|
|
|
/* AVX-512: Op code 0f 3a 26 */
|
|
|
|
asm volatile("vgetmantps $0x12,%zmm6,%zmm7");
|
|
asm volatile("vgetmantpd $0x12,%zmm6,%zmm7");
|
|
|
|
/* AVX-512: Op code 0f 3a 27 */
|
|
|
|
asm volatile("vgetmantss $0x12,%xmm5,%xmm6,%xmm7{%k7}");
|
|
asm volatile("vgetmantsd $0x12,%xmm5,%xmm6,%xmm7{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 38 */
|
|
|
|
asm volatile("vinserti128 $0x5,%xmm4,%ymm4,%ymm6");
|
|
asm volatile("vinserti32x4 $0x12,%xmm4,%zmm5,%zmm6{%k7}");
|
|
asm volatile("vinserti64x2 $0x12,%xmm4,%zmm5,%zmm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 39 */
|
|
|
|
asm volatile("vextracti128 $0x5,%ymm4,%xmm6");
|
|
asm volatile("vextracti32x4 $0x12,%zmm5,%xmm6{%k7}");
|
|
asm volatile("vextracti64x2 $0x12,%zmm5,%xmm6{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 3a */
|
|
|
|
asm volatile("vinserti32x8 $0x12,%ymm5,%zmm6,%zmm7{%k7}");
|
|
asm volatile("vinserti64x4 $0x12,%ymm5,%zmm6,%zmm7{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 3b */
|
|
|
|
asm volatile("vextracti32x8 $0x12,%zmm6,%ymm7{%k7}");
|
|
asm volatile("vextracti64x4 $0x12,%zmm6,%ymm7{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 3e */
|
|
|
|
asm volatile("vpcmpub $0x12,%zmm6,%zmm7,%k5");
|
|
asm volatile("vpcmpuw $0x12,%zmm6,%zmm7,%k5");
|
|
|
|
/* AVX-512: Op code 0f 3a 3f */
|
|
|
|
asm volatile("vpcmpb $0x12,%zmm6,%zmm7,%k5");
|
|
asm volatile("vpcmpw $0x12,%zmm6,%zmm7,%k5");
|
|
|
|
/* AVX-512: Op code 0f 3a 42 */
|
|
|
|
asm volatile("vmpsadbw $0x5,%ymm4,%ymm6,%ymm2");
|
|
asm volatile("vdbpsadbw $0x12,%zmm4,%zmm5,%zmm6");
|
|
|
|
/* AVX-512: Op code 0f 3a 43 */
|
|
|
|
asm volatile("vshufi32x4 $0x12,%zmm5,%zmm6,%zmm7");
|
|
asm volatile("vshufi64x2 $0x12,%zmm5,%zmm6,%zmm7");
|
|
|
|
/* AVX-512: Op code 0f 3a 50 */
|
|
|
|
asm volatile("vrangeps $0x12,%zmm5,%zmm6,%zmm7");
|
|
asm volatile("vrangepd $0x12,%zmm5,%zmm6,%zmm7");
|
|
|
|
/* AVX-512: Op code 0f 3a 51 */
|
|
|
|
asm volatile("vrangess $0x12,%xmm5,%xmm6,%xmm7");
|
|
asm volatile("vrangesd $0x12,%xmm5,%xmm6,%xmm7");
|
|
|
|
/* AVX-512: Op code 0f 3a 54 */
|
|
|
|
asm volatile("vfixupimmps $0x12,%zmm5,%zmm6,%zmm7");
|
|
asm volatile("vfixupimmpd $0x12,%zmm5,%zmm6,%zmm7");
|
|
|
|
/* AVX-512: Op code 0f 3a 55 */
|
|
|
|
asm volatile("vfixupimmss $0x12,%xmm5,%xmm6,%xmm7{%k7}");
|
|
asm volatile("vfixupimmsd $0x12,%xmm5,%xmm6,%xmm7{%k7}");
|
|
|
|
/* AVX-512: Op code 0f 3a 56 */
|
|
|
|
asm volatile("vreduceps $0x12,%zmm6,%zmm7");
|
|
asm volatile("vreducepd $0x12,%zmm6,%zmm7");
|
|
|
|
/* AVX-512: Op code 0f 3a 57 */
|
|
|
|
asm volatile("vreducess $0x12,%xmm5,%xmm6,%xmm7");
|
|
asm volatile("vreducesd $0x12,%xmm5,%xmm6,%xmm7");
|
|
|
|
/* AVX-512: Op code 0f 3a 66 */
|
|
|
|
asm volatile("vfpclassps $0x12,%zmm7,%k5");
|
|
asm volatile("vfpclasspd $0x12,%zmm7,%k5");
|
|
|
|
/* AVX-512: Op code 0f 3a 67 */
|
|
|
|
asm volatile("vfpclassss $0x12,%xmm7,%k5");
|
|
asm volatile("vfpclasssd $0x12,%xmm7,%k5");
|
|
|
|
/* AVX-512: Op code 0f 72 (Grp13) */
|
|
|
|
asm volatile("vprord $0x12,%zmm5,%zmm6");
|
|
asm volatile("vprorq $0x12,%zmm5,%zmm6");
|
|
asm volatile("vprold $0x12,%zmm5,%zmm6");
|
|
asm volatile("vprolq $0x12,%zmm5,%zmm6");
|
|
asm volatile("psrad $0x2,%mm6");
|
|
asm volatile("vpsrad $0x5,%ymm6,%ymm2");
|
|
asm volatile("vpsrad $0x5,%zmm6,%zmm2");
|
|
asm volatile("vpsraq $0x5,%zmm6,%zmm2");
|
|
|
|
/* AVX-512: Op code 0f 38 c6 (Grp18) */
|
|
|
|
asm volatile("vgatherpf0dps 0x7b(%ebp,%zmm7,8){%k1}");
|
|
asm volatile("vgatherpf0dpd 0x7b(%ebp,%ymm7,8){%k1}");
|
|
asm volatile("vgatherpf1dps 0x7b(%ebp,%zmm7,8){%k1}");
|
|
asm volatile("vgatherpf1dpd 0x7b(%ebp,%ymm7,8){%k1}");
|
|
asm volatile("vscatterpf0dps 0x7b(%ebp,%zmm7,8){%k1}");
|
|
asm volatile("vscatterpf0dpd 0x7b(%ebp,%ymm7,8){%k1}");
|
|
asm volatile("vscatterpf1dps 0x7b(%ebp,%zmm7,8){%k1}");
|
|
asm volatile("vscatterpf1dpd 0x7b(%ebp,%ymm7,8){%k1}");
|
|
|
|
/* AVX-512: Op code 0f 38 c7 (Grp19) */
|
|
|
|
asm volatile("vgatherpf0qps 0x7b(%ebp,%zmm7,8){%k1}");
|
|
asm volatile("vgatherpf0qpd 0x7b(%ebp,%zmm7,8){%k1}");
|
|
asm volatile("vgatherpf1qps 0x7b(%ebp,%zmm7,8){%k1}");
|
|
asm volatile("vgatherpf1qpd 0x7b(%ebp,%zmm7,8){%k1}");
|
|
asm volatile("vscatterpf0qps 0x7b(%ebp,%zmm7,8){%k1}");
|
|
asm volatile("vscatterpf0qpd 0x7b(%ebp,%zmm7,8){%k1}");
|
|
asm volatile("vscatterpf1qps 0x7b(%ebp,%zmm7,8){%k1}");
|
|
asm volatile("vscatterpf1qpd 0x7b(%ebp,%zmm7,8){%k1}");
|
|
|
|
/* AVX-512: Examples */
|
|
|
|
asm volatile("vaddpd %zmm4,%zmm5,%zmm6");
|
|
asm volatile("vaddpd %zmm4,%zmm5,%zmm6{%k7}");
|
|
asm volatile("vaddpd %zmm4,%zmm5,%zmm6{%k7}{z}");
|
|
asm volatile("vaddpd {rn-sae},%zmm4,%zmm5,%zmm6");
|
|
asm volatile("vaddpd {ru-sae},%zmm4,%zmm5,%zmm6");
|
|
asm volatile("vaddpd {rd-sae},%zmm4,%zmm5,%zmm6");
|
|
asm volatile("vaddpd {rz-sae},%zmm4,%zmm5,%zmm6");
|
|
asm volatile("vaddpd (%ecx),%zmm5,%zmm6");
|
|
asm volatile("vaddpd 0x123(%eax,%ecx,8),%zmm5,%zmm6");
|
|
asm volatile("vaddpd (%ecx){1to8},%zmm5,%zmm6");
|
|
asm volatile("vaddpd 0x1fc0(%edx),%zmm5,%zmm6");
|
|
asm volatile("vaddpd 0x3f8(%edx){1to8},%zmm5,%zmm6");
|
|
asm volatile("vcmpeq_uqps 0x1fc(%edx){1to16},%zmm6,%k5");
|
|
asm volatile("vcmpltsd 0x123(%eax,%ecx,8),%xmm3,%k5{%k7}");
|
|
asm volatile("vcmplesd {sae},%xmm4,%xmm5,%k5{%k7}");
|
|
asm volatile("vgetmantss $0x5b,0x123(%eax,%ecx,8),%xmm4,%xmm5{%k7}");
|
|
|
|
/* bndmk m32, bnd */
|
|
|
|
asm volatile("bndmk (%eax), %bnd0");
|
|
asm volatile("bndmk (0x12345678), %bnd0");
|
|
asm volatile("bndmk (%eax), %bnd3");
|
|
asm volatile("bndmk (%ecx,%eax,1), %bnd0");
|
|
asm volatile("bndmk 0x12345678(,%eax,1), %bnd0");
|
|
asm volatile("bndmk (%eax,%ecx,1), %bnd0");
|
|
asm volatile("bndmk (%eax,%ecx,8), %bnd0");
|
|
asm volatile("bndmk 0x12(%eax), %bnd0");
|
|
asm volatile("bndmk 0x12(%ebp), %bnd0");
|
|
asm volatile("bndmk 0x12(%ecx,%eax,1), %bnd0");
|
|
asm volatile("bndmk 0x12(%ebp,%eax,1), %bnd0");
|
|
asm volatile("bndmk 0x12(%eax,%ecx,1), %bnd0");
|
|
asm volatile("bndmk 0x12(%eax,%ecx,8), %bnd0");
|
|
asm volatile("bndmk 0x12345678(%eax), %bnd0");
|
|
asm volatile("bndmk 0x12345678(%ebp), %bnd0");
|
|
asm volatile("bndmk 0x12345678(%ecx,%eax,1), %bnd0");
|
|
asm volatile("bndmk 0x12345678(%ebp,%eax,1), %bnd0");
|
|
asm volatile("bndmk 0x12345678(%eax,%ecx,1), %bnd0");
|
|
asm volatile("bndmk 0x12345678(%eax,%ecx,8), %bnd0");
|
|
|
|
/* bndcl r/m32, bnd */
|
|
|
|
asm volatile("bndcl (%eax), %bnd0");
|
|
asm volatile("bndcl (0x12345678), %bnd0");
|
|
asm volatile("bndcl (%eax), %bnd3");
|
|
asm volatile("bndcl (%ecx,%eax,1), %bnd0");
|
|
asm volatile("bndcl 0x12345678(,%eax,1), %bnd0");
|
|
asm volatile("bndcl (%eax,%ecx,1), %bnd0");
|
|
asm volatile("bndcl (%eax,%ecx,8), %bnd0");
|
|
asm volatile("bndcl 0x12(%eax), %bnd0");
|
|
asm volatile("bndcl 0x12(%ebp), %bnd0");
|
|
asm volatile("bndcl 0x12(%ecx,%eax,1), %bnd0");
|
|
asm volatile("bndcl 0x12(%ebp,%eax,1), %bnd0");
|
|
asm volatile("bndcl 0x12(%eax,%ecx,1), %bnd0");
|
|
asm volatile("bndcl 0x12(%eax,%ecx,8), %bnd0");
|
|
asm volatile("bndcl 0x12345678(%eax), %bnd0");
|
|
asm volatile("bndcl 0x12345678(%ebp), %bnd0");
|
|
asm volatile("bndcl 0x12345678(%ecx,%eax,1), %bnd0");
|
|
asm volatile("bndcl 0x12345678(%ebp,%eax,1), %bnd0");
|
|
asm volatile("bndcl 0x12345678(%eax,%ecx,1), %bnd0");
|
|
asm volatile("bndcl 0x12345678(%eax,%ecx,8), %bnd0");
|
|
asm volatile("bndcl %eax, %bnd0");
|
|
|
|
/* bndcu r/m32, bnd */
|
|
|
|
asm volatile("bndcu (%eax), %bnd0");
|
|
asm volatile("bndcu (0x12345678), %bnd0");
|
|
asm volatile("bndcu (%eax), %bnd3");
|
|
asm volatile("bndcu (%ecx,%eax,1), %bnd0");
|
|
asm volatile("bndcu 0x12345678(,%eax,1), %bnd0");
|
|
asm volatile("bndcu (%eax,%ecx,1), %bnd0");
|
|
asm volatile("bndcu (%eax,%ecx,8), %bnd0");
|
|
asm volatile("bndcu 0x12(%eax), %bnd0");
|
|
asm volatile("bndcu 0x12(%ebp), %bnd0");
|
|
asm volatile("bndcu 0x12(%ecx,%eax,1), %bnd0");
|
|
asm volatile("bndcu 0x12(%ebp,%eax,1), %bnd0");
|
|
asm volatile("bndcu 0x12(%eax,%ecx,1), %bnd0");
|
|
asm volatile("bndcu 0x12(%eax,%ecx,8), %bnd0");
|
|
asm volatile("bndcu 0x12345678(%eax), %bnd0");
|
|
asm volatile("bndcu 0x12345678(%ebp), %bnd0");
|
|
asm volatile("bndcu 0x12345678(%ecx,%eax,1), %bnd0");
|
|
asm volatile("bndcu 0x12345678(%ebp,%eax,1), %bnd0");
|
|
asm volatile("bndcu 0x12345678(%eax,%ecx,1), %bnd0");
|
|
asm volatile("bndcu 0x12345678(%eax,%ecx,8), %bnd0");
|
|
asm volatile("bndcu %eax, %bnd0");
|
|
|
|
/* bndcn r/m32, bnd */
|
|
|
|
asm volatile("bndcn (%eax), %bnd0");
|
|
asm volatile("bndcn (0x12345678), %bnd0");
|
|
asm volatile("bndcn (%eax), %bnd3");
|
|
asm volatile("bndcn (%ecx,%eax,1), %bnd0");
|
|
asm volatile("bndcn 0x12345678(,%eax,1), %bnd0");
|
|
asm volatile("bndcn (%eax,%ecx,1), %bnd0");
|
|
asm volatile("bndcn (%eax,%ecx,8), %bnd0");
|
|
asm volatile("bndcn 0x12(%eax), %bnd0");
|
|
asm volatile("bndcn 0x12(%ebp), %bnd0");
|
|
asm volatile("bndcn 0x12(%ecx,%eax,1), %bnd0");
|
|
asm volatile("bndcn 0x12(%ebp,%eax,1), %bnd0");
|
|
asm volatile("bndcn 0x12(%eax,%ecx,1), %bnd0");
|
|
asm volatile("bndcn 0x12(%eax,%ecx,8), %bnd0");
|
|
asm volatile("bndcn 0x12345678(%eax), %bnd0");
|
|
asm volatile("bndcn 0x12345678(%ebp), %bnd0");
|
|
asm volatile("bndcn 0x12345678(%ecx,%eax,1), %bnd0");
|
|
asm volatile("bndcn 0x12345678(%ebp,%eax,1), %bnd0");
|
|
asm volatile("bndcn 0x12345678(%eax,%ecx,1), %bnd0");
|
|
asm volatile("bndcn 0x12345678(%eax,%ecx,8), %bnd0");
|
|
asm volatile("bndcn %eax, %bnd0");
|
|
|
|
/* bndmov m64, bnd */
|
|
|
|
asm volatile("bndmov (%eax), %bnd0");
|
|
asm volatile("bndmov (0x12345678), %bnd0");
|
|
asm volatile("bndmov (%eax), %bnd3");
|
|
asm volatile("bndmov (%ecx,%eax,1), %bnd0");
|
|
asm volatile("bndmov 0x12345678(,%eax,1), %bnd0");
|
|
asm volatile("bndmov (%eax,%ecx,1), %bnd0");
|
|
asm volatile("bndmov (%eax,%ecx,8), %bnd0");
|
|
asm volatile("bndmov 0x12(%eax), %bnd0");
|
|
asm volatile("bndmov 0x12(%ebp), %bnd0");
|
|
asm volatile("bndmov 0x12(%ecx,%eax,1), %bnd0");
|
|
asm volatile("bndmov 0x12(%ebp,%eax,1), %bnd0");
|
|
asm volatile("bndmov 0x12(%eax,%ecx,1), %bnd0");
|
|
asm volatile("bndmov 0x12(%eax,%ecx,8), %bnd0");
|
|
asm volatile("bndmov 0x12345678(%eax), %bnd0");
|
|
asm volatile("bndmov 0x12345678(%ebp), %bnd0");
|
|
asm volatile("bndmov 0x12345678(%ecx,%eax,1), %bnd0");
|
|
asm volatile("bndmov 0x12345678(%ebp,%eax,1), %bnd0");
|
|
asm volatile("bndmov 0x12345678(%eax,%ecx,1), %bnd0");
|
|
asm volatile("bndmov 0x12345678(%eax,%ecx,8), %bnd0");
|
|
|
|
/* bndmov bnd, m64 */
|
|
|
|
asm volatile("bndmov %bnd0, (%eax)");
|
|
asm volatile("bndmov %bnd0, (0x12345678)");
|
|
asm volatile("bndmov %bnd3, (%eax)");
|
|
asm volatile("bndmov %bnd0, (%ecx,%eax,1)");
|
|
asm volatile("bndmov %bnd0, 0x12345678(,%eax,1)");
|
|
asm volatile("bndmov %bnd0, (%eax,%ecx,1)");
|
|
asm volatile("bndmov %bnd0, (%eax,%ecx,8)");
|
|
asm volatile("bndmov %bnd0, 0x12(%eax)");
|
|
asm volatile("bndmov %bnd0, 0x12(%ebp)");
|
|
asm volatile("bndmov %bnd0, 0x12(%ecx,%eax,1)");
|
|
asm volatile("bndmov %bnd0, 0x12(%ebp,%eax,1)");
|
|
asm volatile("bndmov %bnd0, 0x12(%eax,%ecx,1)");
|
|
asm volatile("bndmov %bnd0, 0x12(%eax,%ecx,8)");
|
|
asm volatile("bndmov %bnd0, 0x12345678(%eax)");
|
|
asm volatile("bndmov %bnd0, 0x12345678(%ebp)");
|
|
asm volatile("bndmov %bnd0, 0x12345678(%ecx,%eax,1)");
|
|
asm volatile("bndmov %bnd0, 0x12345678(%ebp,%eax,1)");
|
|
asm volatile("bndmov %bnd0, 0x12345678(%eax,%ecx,1)");
|
|
asm volatile("bndmov %bnd0, 0x12345678(%eax,%ecx,8)");
|
|
|
|
/* bndmov bnd2, bnd1 */
|
|
|
|
asm volatile("bndmov %bnd0, %bnd1");
|
|
asm volatile("bndmov %bnd1, %bnd0");
|
|
|
|
/* bndldx mib, bnd */
|
|
|
|
asm volatile("bndldx (%eax), %bnd0");
|
|
asm volatile("bndldx (0x12345678), %bnd0");
|
|
asm volatile("bndldx (%eax), %bnd3");
|
|
asm volatile("bndldx (%ecx,%eax,1), %bnd0");
|
|
asm volatile("bndldx 0x12345678(,%eax,1), %bnd0");
|
|
asm volatile("bndldx (%eax,%ecx,1), %bnd0");
|
|
asm volatile("bndldx 0x12(%eax), %bnd0");
|
|
asm volatile("bndldx 0x12(%ebp), %bnd0");
|
|
asm volatile("bndldx 0x12(%ecx,%eax,1), %bnd0");
|
|
asm volatile("bndldx 0x12(%ebp,%eax,1), %bnd0");
|
|
asm volatile("bndldx 0x12(%eax,%ecx,1), %bnd0");
|
|
asm volatile("bndldx 0x12345678(%eax), %bnd0");
|
|
asm volatile("bndldx 0x12345678(%ebp), %bnd0");
|
|
asm volatile("bndldx 0x12345678(%ecx,%eax,1), %bnd0");
|
|
asm volatile("bndldx 0x12345678(%ebp,%eax,1), %bnd0");
|
|
asm volatile("bndldx 0x12345678(%eax,%ecx,1), %bnd0");
|
|
|
|
/* bndstx bnd, mib */
|
|
|
|
asm volatile("bndstx %bnd0, (%eax)");
|
|
asm volatile("bndstx %bnd0, (0x12345678)");
|
|
asm volatile("bndstx %bnd3, (%eax)");
|
|
asm volatile("bndstx %bnd0, (%ecx,%eax,1)");
|
|
asm volatile("bndstx %bnd0, 0x12345678(,%eax,1)");
|
|
asm volatile("bndstx %bnd0, (%eax,%ecx,1)");
|
|
asm volatile("bndstx %bnd0, 0x12(%eax)");
|
|
asm volatile("bndstx %bnd0, 0x12(%ebp)");
|
|
asm volatile("bndstx %bnd0, 0x12(%ecx,%eax,1)");
|
|
asm volatile("bndstx %bnd0, 0x12(%ebp,%eax,1)");
|
|
asm volatile("bndstx %bnd0, 0x12(%eax,%ecx,1)");
|
|
asm volatile("bndstx %bnd0, 0x12345678(%eax)");
|
|
asm volatile("bndstx %bnd0, 0x12345678(%ebp)");
|
|
asm volatile("bndstx %bnd0, 0x12345678(%ecx,%eax,1)");
|
|
asm volatile("bndstx %bnd0, 0x12345678(%ebp,%eax,1)");
|
|
asm volatile("bndstx %bnd0, 0x12345678(%eax,%ecx,1)");
|
|
|
|
/* bnd prefix on call, ret, jmp and all jcc */
|
|
|
|
asm volatile("bnd call label1"); /* Expecting: call unconditional 0xfffffffc */
|
|
asm volatile("bnd call *(%eax)"); /* Expecting: call indirect 0 */
|
|
asm volatile("bnd ret"); /* Expecting: ret indirect 0 */
|
|
asm volatile("bnd jmp label1"); /* Expecting: jmp unconditional 0xfffffffc */
|
|
asm volatile("bnd jmp label1"); /* Expecting: jmp unconditional 0xfffffffc */
|
|
asm volatile("bnd jmp *(%ecx)"); /* Expecting: jmp indirect 0 */
|
|
asm volatile("bnd jne label1"); /* Expecting: jcc conditional 0xfffffffc */
|
|
|
|
/* sha1rnds4 imm8, xmm2/m128, xmm1 */
|
|
|
|
asm volatile("sha1rnds4 $0x0, %xmm1, %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, %xmm7, %xmm2");
|
|
asm volatile("sha1rnds4 $0x91, (%eax), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, (0x12345678), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, (%eax), %xmm3");
|
|
asm volatile("sha1rnds4 $0x91, (%ecx,%eax,1), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12345678(,%eax,1), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, (%eax,%ecx,1), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, (%eax,%ecx,8), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12(%eax), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12(%ebp), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12(%ecx,%eax,1), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12(%ebp,%eax,1), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12(%eax,%ecx,1), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12(%eax,%ecx,8), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12345678(%eax), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12345678(%ebp), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12345678(%ecx,%eax,1), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12345678(%ebp,%eax,1), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12345678(%eax,%ecx,1), %xmm0");
|
|
asm volatile("sha1rnds4 $0x91, 0x12345678(%eax,%ecx,8), %xmm0");
|
|
|
|
/* sha1nexte xmm2/m128, xmm1 */
|
|
|
|
asm volatile("sha1nexte %xmm1, %xmm0");
|
|
asm volatile("sha1nexte %xmm7, %xmm2");
|
|
asm volatile("sha1nexte (%eax), %xmm0");
|
|
asm volatile("sha1nexte (0x12345678), %xmm0");
|
|
asm volatile("sha1nexte (%eax), %xmm3");
|
|
asm volatile("sha1nexte (%ecx,%eax,1), %xmm0");
|
|
asm volatile("sha1nexte 0x12345678(,%eax,1), %xmm0");
|
|
asm volatile("sha1nexte (%eax,%ecx,1), %xmm0");
|
|
asm volatile("sha1nexte (%eax,%ecx,8), %xmm0");
|
|
asm volatile("sha1nexte 0x12(%eax), %xmm0");
|
|
asm volatile("sha1nexte 0x12(%ebp), %xmm0");
|
|
asm volatile("sha1nexte 0x12(%ecx,%eax,1), %xmm0");
|
|
asm volatile("sha1nexte 0x12(%ebp,%eax,1), %xmm0");
|
|
asm volatile("sha1nexte 0x12(%eax,%ecx,1), %xmm0");
|
|
asm volatile("sha1nexte 0x12(%eax,%ecx,8), %xmm0");
|
|
asm volatile("sha1nexte 0x12345678(%eax), %xmm0");
|
|
asm volatile("sha1nexte 0x12345678(%ebp), %xmm0");
|
|
asm volatile("sha1nexte 0x12345678(%ecx,%eax,1), %xmm0");
|
|
asm volatile("sha1nexte 0x12345678(%ebp,%eax,1), %xmm0");
|
|
asm volatile("sha1nexte 0x12345678(%eax,%ecx,1), %xmm0");
|
|
asm volatile("sha1nexte 0x12345678(%eax,%ecx,8), %xmm0");
|
|
|
|
/* sha1msg1 xmm2/m128, xmm1 */
|
|
|
|
asm volatile("sha1msg1 %xmm1, %xmm0");
|
|
asm volatile("sha1msg1 %xmm7, %xmm2");
|
|
asm volatile("sha1msg1 (%eax), %xmm0");
|
|
asm volatile("sha1msg1 (0x12345678), %xmm0");
|
|
asm volatile("sha1msg1 (%eax), %xmm3");
|
|
asm volatile("sha1msg1 (%ecx,%eax,1), %xmm0");
|
|
asm volatile("sha1msg1 0x12345678(,%eax,1), %xmm0");
|
|
asm volatile("sha1msg1 (%eax,%ecx,1), %xmm0");
|
|
asm volatile("sha1msg1 (%eax,%ecx,8), %xmm0");
|
|
asm volatile("sha1msg1 0x12(%eax), %xmm0");
|
|
asm volatile("sha1msg1 0x12(%ebp), %xmm0");
|
|
asm volatile("sha1msg1 0x12(%ecx,%eax,1), %xmm0");
|
|
asm volatile("sha1msg1 0x12(%ebp,%eax,1), %xmm0");
|
|
asm volatile("sha1msg1 0x12(%eax,%ecx,1), %xmm0");
|
|
asm volatile("sha1msg1 0x12(%eax,%ecx,8), %xmm0");
|
|
asm volatile("sha1msg1 0x12345678(%eax), %xmm0");
|
|
asm volatile("sha1msg1 0x12345678(%ebp), %xmm0");
|
|
asm volatile("sha1msg1 0x12345678(%ecx,%eax,1), %xmm0");
|
|
asm volatile("sha1msg1 0x12345678(%ebp,%eax,1), %xmm0");
|
|
asm volatile("sha1msg1 0x12345678(%eax,%ecx,1), %xmm0");
|
|
asm volatile("sha1msg1 0x12345678(%eax,%ecx,8), %xmm0");
|
|
|
|
/* sha1msg2 xmm2/m128, xmm1 */
|
|
|
|
asm volatile("sha1msg2 %xmm1, %xmm0");
|
|
asm volatile("sha1msg2 %xmm7, %xmm2");
|
|
asm volatile("sha1msg2 (%eax), %xmm0");
|
|
asm volatile("sha1msg2 (0x12345678), %xmm0");
|
|
asm volatile("sha1msg2 (%eax), %xmm3");
|
|
asm volatile("sha1msg2 (%ecx,%eax,1), %xmm0");
|
|
asm volatile("sha1msg2 0x12345678(,%eax,1), %xmm0");
|
|
asm volatile("sha1msg2 (%eax,%ecx,1), %xmm0");
|
|
asm volatile("sha1msg2 (%eax,%ecx,8), %xmm0");
|
|
asm volatile("sha1msg2 0x12(%eax), %xmm0");
|
|
asm volatile("sha1msg2 0x12(%ebp), %xmm0");
|
|
asm volatile("sha1msg2 0x12(%ecx,%eax,1), %xmm0");
|
|
asm volatile("sha1msg2 0x12(%ebp,%eax,1), %xmm0");
|
|
asm volatile("sha1msg2 0x12(%eax,%ecx,1), %xmm0");
|
|
asm volatile("sha1msg2 0x12(%eax,%ecx,8), %xmm0");
|
|
asm volatile("sha1msg2 0x12345678(%eax), %xmm0");
|
|
asm volatile("sha1msg2 0x12345678(%ebp), %xmm0");
|
|
asm volatile("sha1msg2 0x12345678(%ecx,%eax,1), %xmm0");
|
|
asm volatile("sha1msg2 0x12345678(%ebp,%eax,1), %xmm0");
|
|
asm volatile("sha1msg2 0x12345678(%eax,%ecx,1), %xmm0");
|
|
asm volatile("sha1msg2 0x12345678(%eax,%ecx,8), %xmm0");
|
|
|
|
/* sha256rnds2 <XMM0>, xmm2/m128, xmm1 */
|
|
/* Note sha256rnds2 has an implicit operand 'xmm0' */
|
|
|
|
asm volatile("sha256rnds2 %xmm4, %xmm1");
|
|
asm volatile("sha256rnds2 %xmm7, %xmm2");
|
|
asm volatile("sha256rnds2 (%eax), %xmm1");
|
|
asm volatile("sha256rnds2 (0x12345678), %xmm1");
|
|
asm volatile("sha256rnds2 (%eax), %xmm3");
|
|
asm volatile("sha256rnds2 (%ecx,%eax,1), %xmm1");
|
|
asm volatile("sha256rnds2 0x12345678(,%eax,1), %xmm1");
|
|
asm volatile("sha256rnds2 (%eax,%ecx,1), %xmm1");
|
|
asm volatile("sha256rnds2 (%eax,%ecx,8), %xmm1");
|
|
asm volatile("sha256rnds2 0x12(%eax), %xmm1");
|
|
asm volatile("sha256rnds2 0x12(%ebp), %xmm1");
|
|
asm volatile("sha256rnds2 0x12(%ecx,%eax,1), %xmm1");
|
|
asm volatile("sha256rnds2 0x12(%ebp,%eax,1), %xmm1");
|
|
asm volatile("sha256rnds2 0x12(%eax,%ecx,1), %xmm1");
|
|
asm volatile("sha256rnds2 0x12(%eax,%ecx,8), %xmm1");
|
|
asm volatile("sha256rnds2 0x12345678(%eax), %xmm1");
|
|
asm volatile("sha256rnds2 0x12345678(%ebp), %xmm1");
|
|
asm volatile("sha256rnds2 0x12345678(%ecx,%eax,1), %xmm1");
|
|
asm volatile("sha256rnds2 0x12345678(%ebp,%eax,1), %xmm1");
|
|
asm volatile("sha256rnds2 0x12345678(%eax,%ecx,1), %xmm1");
|
|
asm volatile("sha256rnds2 0x12345678(%eax,%ecx,8), %xmm1");
|
|
|
|
/* sha256msg1 xmm2/m128, xmm1 */
|
|
|
|
asm volatile("sha256msg1 %xmm1, %xmm0");
|
|
asm volatile("sha256msg1 %xmm7, %xmm2");
|
|
asm volatile("sha256msg1 (%eax), %xmm0");
|
|
asm volatile("sha256msg1 (0x12345678), %xmm0");
|
|
asm volatile("sha256msg1 (%eax), %xmm3");
|
|
asm volatile("sha256msg1 (%ecx,%eax,1), %xmm0");
|
|
asm volatile("sha256msg1 0x12345678(,%eax,1), %xmm0");
|
|
asm volatile("sha256msg1 (%eax,%ecx,1), %xmm0");
|
|
asm volatile("sha256msg1 (%eax,%ecx,8), %xmm0");
|
|
asm volatile("sha256msg1 0x12(%eax), %xmm0");
|
|
asm volatile("sha256msg1 0x12(%ebp), %xmm0");
|
|
asm volatile("sha256msg1 0x12(%ecx,%eax,1), %xmm0");
|
|
asm volatile("sha256msg1 0x12(%ebp,%eax,1), %xmm0");
|
|
asm volatile("sha256msg1 0x12(%eax,%ecx,1), %xmm0");
|
|
asm volatile("sha256msg1 0x12(%eax,%ecx,8), %xmm0");
|
|
asm volatile("sha256msg1 0x12345678(%eax), %xmm0");
|
|
asm volatile("sha256msg1 0x12345678(%ebp), %xmm0");
|
|
asm volatile("sha256msg1 0x12345678(%ecx,%eax,1), %xmm0");
|
|
asm volatile("sha256msg1 0x12345678(%ebp,%eax,1), %xmm0");
|
|
asm volatile("sha256msg1 0x12345678(%eax,%ecx,1), %xmm0");
|
|
asm volatile("sha256msg1 0x12345678(%eax,%ecx,8), %xmm0");
|
|
|
|
/* sha256msg2 xmm2/m128, xmm1 */
|
|
|
|
asm volatile("sha256msg2 %xmm1, %xmm0");
|
|
asm volatile("sha256msg2 %xmm7, %xmm2");
|
|
asm volatile("sha256msg2 (%eax), %xmm0");
|
|
asm volatile("sha256msg2 (0x12345678), %xmm0");
|
|
asm volatile("sha256msg2 (%eax), %xmm3");
|
|
asm volatile("sha256msg2 (%ecx,%eax,1), %xmm0");
|
|
asm volatile("sha256msg2 0x12345678(,%eax,1), %xmm0");
|
|
asm volatile("sha256msg2 (%eax,%ecx,1), %xmm0");
|
|
asm volatile("sha256msg2 (%eax,%ecx,8), %xmm0");
|
|
asm volatile("sha256msg2 0x12(%eax), %xmm0");
|
|
asm volatile("sha256msg2 0x12(%ebp), %xmm0");
|
|
asm volatile("sha256msg2 0x12(%ecx,%eax,1), %xmm0");
|
|
asm volatile("sha256msg2 0x12(%ebp,%eax,1), %xmm0");
|
|
asm volatile("sha256msg2 0x12(%eax,%ecx,1), %xmm0");
|
|
asm volatile("sha256msg2 0x12(%eax,%ecx,8), %xmm0");
|
|
asm volatile("sha256msg2 0x12345678(%eax), %xmm0");
|
|
asm volatile("sha256msg2 0x12345678(%ebp), %xmm0");
|
|
asm volatile("sha256msg2 0x12345678(%ecx,%eax,1), %xmm0");
|
|
asm volatile("sha256msg2 0x12345678(%ebp,%eax,1), %xmm0");
|
|
asm volatile("sha256msg2 0x12345678(%eax,%ecx,1), %xmm0");
|
|
asm volatile("sha256msg2 0x12345678(%eax,%ecx,8), %xmm0");
|
|
|
|
/* clflushopt m8 */
|
|
|
|
asm volatile("clflushopt (%eax)");
|
|
asm volatile("clflushopt (0x12345678)");
|
|
asm volatile("clflushopt 0x12345678(%eax,%ecx,8)");
|
|
/* Also check instructions in the same group encoding as clflushopt */
|
|
asm volatile("clflush (%eax)");
|
|
asm volatile("sfence");
|
|
|
|
/* clwb m8 */
|
|
|
|
asm volatile("clwb (%eax)");
|
|
asm volatile("clwb (0x12345678)");
|
|
asm volatile("clwb 0x12345678(%eax,%ecx,8)");
|
|
/* Also check instructions in the same group encoding as clwb */
|
|
asm volatile("xsaveopt (%eax)");
|
|
asm volatile("mfence");
|
|
|
|
/* xsavec mem */
|
|
|
|
asm volatile("xsavec (%eax)");
|
|
asm volatile("xsavec (0x12345678)");
|
|
asm volatile("xsavec 0x12345678(%eax,%ecx,8)");
|
|
|
|
/* xsaves mem */
|
|
|
|
asm volatile("xsaves (%eax)");
|
|
asm volatile("xsaves (0x12345678)");
|
|
asm volatile("xsaves 0x12345678(%eax,%ecx,8)");
|
|
|
|
/* xrstors mem */
|
|
|
|
asm volatile("xrstors (%eax)");
|
|
asm volatile("xrstors (0x12345678)");
|
|
asm volatile("xrstors 0x12345678(%eax,%ecx,8)");
|
|
|
|
/* ptwrite */
|
|
|
|
asm volatile("ptwrite (%eax)");
|
|
asm volatile("ptwrite (0x12345678)");
|
|
asm volatile("ptwrite 0x12345678(%eax,%ecx,8)");
|
|
|
|
asm volatile("ptwritel (%eax)");
|
|
asm volatile("ptwritel (0x12345678)");
|
|
asm volatile("ptwritel 0x12345678(%eax,%ecx,8)");
|
|
|
|
#endif /* #ifndef __x86_64__ */
|
|
|
|
/* Following line is a marker for the awk script - do not change */
|
|
asm volatile("rdtsc"); /* Stop here */
|
|
|
|
return 0;
|
|
}
|