diff --git a/clang/lib/CIR/CodeGen/CIRGenBuiltinX86.cpp b/clang/lib/CIR/CodeGen/CIRGenBuiltinX86.cpp index 5bd9f83c6b78..807c0612cef9 100644 --- a/clang/lib/CIR/CodeGen/CIRGenBuiltinX86.cpp +++ b/clang/lib/CIR/CodeGen/CIRGenBuiltinX86.cpp @@ -25,6 +25,7 @@ #include "clang/Basic/TargetBuiltins.h" #include "clang/CIR/Dialect/IR/CIRDialect.h" #include "clang/CIR/Dialect/IR/CIRTypes.h" +#include "llvm/IR/IntrinsicsX86.h" #include "llvm/Support/ErrorHandling.h" using namespace clang; @@ -186,5 +187,130 @@ mlir::Value CIRGenFunction::emitX86BuiltinExpr(unsigned BuiltinID, Ops[0].getType(), mlir::ValueRange{Ops[0], V}) .getResult(); } + case X86::BI__builtin_ia32_undef128: + case X86::BI__builtin_ia32_undef256: + case X86::BI__builtin_ia32_undef512: + // The x86 definition of "undef" is not the same as the LLVM definition + // (PR32176). We leave optimizing away an unnecessary zero constant to the + // IR optimizer and backend. + // TODO: If we had a "freeze" IR instruction to generate a fixed undef + // value, we should use that here instead of a zero. + llvm_unreachable("__builtin_ia32_undefXX NYI"); + case X86::BI__builtin_ia32_vec_ext_v4hi: + case X86::BI__builtin_ia32_vec_ext_v16qi: + case X86::BI__builtin_ia32_vec_ext_v8hi: + case X86::BI__builtin_ia32_vec_ext_v4si: + case X86::BI__builtin_ia32_vec_ext_v4sf: + case X86::BI__builtin_ia32_vec_ext_v2di: + case X86::BI__builtin_ia32_vec_ext_v32qi: + case X86::BI__builtin_ia32_vec_ext_v16hi: + case X86::BI__builtin_ia32_vec_ext_v8si: + case X86::BI__builtin_ia32_vec_ext_v4di: { + llvm_unreachable("__builtin_ia32_vec_ext_vXX NYI"); + } + case X86::BI__builtin_ia32_vec_set_v4hi: + case X86::BI__builtin_ia32_vec_set_v16qi: + case X86::BI__builtin_ia32_vec_set_v8hi: + case X86::BI__builtin_ia32_vec_set_v4si: + case X86::BI__builtin_ia32_vec_set_v2di: + case X86::BI__builtin_ia32_vec_set_v32qi: + case X86::BI__builtin_ia32_vec_set_v16hi: + case X86::BI__builtin_ia32_vec_set_v8si: + case X86::BI__builtin_ia32_vec_set_v4di: { + llvm_unreachable("__builtin_ia32_vec_set_vXX NYI"); + } + case X86::BI_mm_setcsr: + case X86::BI__builtin_ia32_ldmxcsr: { + llvm_unreachable("mm_setcsr NYI"); + } + case X86::BI_mm_getcsr: + case X86::BI__builtin_ia32_stmxcsr: { + llvm_unreachable("mm_getcsr NYI"); + } + + case X86::BI__builtin_ia32_xsave: + case X86::BI__builtin_ia32_xsave64: + case X86::BI__builtin_ia32_xrstor: + case X86::BI__builtin_ia32_xrstor64: + case X86::BI__builtin_ia32_xsaveopt: + case X86::BI__builtin_ia32_xsaveopt64: + case X86::BI__builtin_ia32_xrstors: + case X86::BI__builtin_ia32_xrstors64: + case X86::BI__builtin_ia32_xsavec: + case X86::BI__builtin_ia32_xsavec64: + case X86::BI__builtin_ia32_xsaves: + case X86::BI__builtin_ia32_xsaves64: + case X86::BI__builtin_ia32_xsetbv: + case X86::BI_xsetbv: { + std::string intrinsicName; + + // TODO(cir): Refactor this once we have the proper + // infrastructure that handles `getIntrinsic` similar to OG CodeGen. + switch (BuiltinID) { + default: + llvm_unreachable("Unsupported intrinsic!"); + case X86::BI__builtin_ia32_xsave: + intrinsicName = "x86.xsave"; + break; + case X86::BI__builtin_ia32_xsave64: + intrinsicName = "x86.xsave64"; + break; + case X86::BI__builtin_ia32_xrstor: + intrinsicName = "x86.xrstor"; + break; + case X86::BI__builtin_ia32_xrstor64: + intrinsicName = "x86.xrstor64"; + break; + case X86::BI__builtin_ia32_xsaveopt: + intrinsicName = "x86.xsaveopt"; + break; + case X86::BI__builtin_ia32_xsaveopt64: + intrinsicName = "x86.xsaveopt64"; + break; + case X86::BI__builtin_ia32_xrstors: + intrinsicName = "x86.xrstors"; + break; + case X86::BI__builtin_ia32_xrstors64: + intrinsicName = "x86.xrstors64"; + break; + case X86::BI__builtin_ia32_xsavec: + intrinsicName = "x86.xsavec"; + break; + case X86::BI__builtin_ia32_xsavec64: + intrinsicName = "x86.xsavec64"; + break; + case X86::BI__builtin_ia32_xsaves: + intrinsicName = "x86.xsaves"; + break; + case X86::BI__builtin_ia32_xsaves64: + intrinsicName = "x86.xsaves64"; + break; + case X86::BI__builtin_ia32_xsetbv: + case X86::BI_xsetbv: + intrinsicName = "x86.xsetbv"; + break; + } + auto loc = getLoc(E->getExprLoc()); + + mlir::Value mhi = builder.createShift(Ops[1], 32, false); + mhi = builder.createIntCast(mhi, builder.getSInt32Ty()); + + mlir::Value mlo = builder.createIntCast(Ops[1], builder.getSInt32Ty()); + + Ops[1] = mhi; + Ops.push_back(mlo); + + return builder + .create( + loc, builder.getStringAttr(intrinsicName), builder.getVoidTy(), Ops) + .getResult(); + } + case X86::BI__builtin_ia32_xgetbv: + case X86::BI_xgetbv: + return builder + .create(getLoc(E->getExprLoc()), + builder.getStringAttr("x86.xgetbv"), + builder.getUInt64Ty(), Ops) + .getResult(); } } diff --git a/clang/test/CIR/CodeGen/X86/x86_64-xsave.c b/clang/test/CIR/CodeGen/X86/x86_64-xsave.c new file mode 100644 index 000000000000..bfa8377df521 --- /dev/null +++ b/clang/test/CIR/CodeGen/X86/x86_64-xsave.c @@ -0,0 +1,336 @@ +// RUN: %clang_cc1 %s -DTEST_XSAVE -O0 -triple=x86_64-unknown-linux -target-feature +xsave -fclangir -emit-cir -o %t.cir -Wall -Wno-unused-but-set-variable -Werror +// RUN: FileCheck --check-prefix=CIR-XSAVE --input-file=%t.cir %s +// RUN: %clang_cc1 %s -DTEST_XSAVE -O0 -triple=x86_64-unknown-linux -target-feature +xsave -fclangir -emit-llvm -o %t.ll -Wall -Wno-unused-but-set-variable -Werror +// RUN: FileCheck --check-prefix=LLVM-XSAVE --input-file=%t.ll %s +// RUN: %clang_cc1 %s -DTEST_XSAVE -O0 -triple=x86_64-unknown-linux -target-feature +xsave -fno-signed-char -fclangir -emit-cir -o %t.cir -Wall -Wno-unused-but-set-variable -Werror +// RUN: FileCheck --check-prefix=CIR-XSAVE --input-file=%t.cir %s +// RUN: %clang_cc1 %s -DTEST_XSAVE -O0 -triple=x86_64-unknown-linux -target-feature +xsave -fno-signed-char -fclangir -emit-llvm -o %t.ll -Wall -Wno-unused-but-set-variable -Werror +// RUN: FileCheck --check-prefix=LLVM-XSAVE --input-file=%t.ll %s + +// RUN: %clang_cc1 %s -DTEST_XGETBV -O0 -triple=x86_64-unknown-linux -target-feature +xsave -fno-signed-char -fclangir -emit-cir -o %t.cir -Wall -Wno-unused-but-set-variable -Werror +// RUN: FileCheck --check-prefix=CIR-XGETBV --input-file=%t.cir %s +// RUN: %clang_cc1 %s -DTEST_XGETBV -O0 -triple=x86_64-unknown-linux -target-feature +xsave -fno-signed-char -fclangir -emit-llvm -o %t.ll -Wall -Wno-unused-but-set-variable -Werror +// RUN: FileCheck --check-prefix=LLVM-XGETBV --input-file=%t.ll %s +// RUN: %clang_cc1 %s -DTEST_XSETBV -O0 -triple=x86_64-unknown-linux -target-feature +xsave -fno-signed-char -fclangir -emit-cir -o %t.cir -Wall -Wno-unused-but-set-variable -Werror +// RUN: FileCheck --check-prefix=CIR-XSETBV --input-file=%t.cir %s +// RUN: %clang_cc1 %s -DTEST_XSETBV -O0 -triple=x86_64-unknown-linux -target-feature +xsave -fno-signed-char -fclangir -emit-llvm -o %t.ll -Wall -Wno-unused-but-set-variable -Werror +// RUN: FileCheck --check-prefix=LLVM-XSETBV --input-file=%t.ll %s + +// RUN: %clang_cc1 %s -DTEST_XSAVEOPT -O0 -triple=x86_64-unknown-linux -target-feature +xsave -target-feature +xsaveopt -fclangir -emit-cir -o %t.cir -Wall -Wno-unused-but-set-variable -Werror +// RUN: FileCheck --check-prefix=CIR-XSAVEOPT --input-file=%t.cir %s +// RUN: %clang_cc1 %s -DTEST_XSAVEOPT -O0 -triple=x86_64-unknown-linux -target-feature +xsave -target-feature +xsaveopt -fclangir -emit-llvm -o %t.ll -Wall -Wno-unused-but-set-variable -Werror +// RUN: FileCheck --check-prefix=LLVM-XSAVEOPT --input-file=%t.ll %s +// RUN: %clang_cc1 %s -DTEST_XSAVEOPT -O0 -triple=x86_64-unknown-linux -target-feature +xsave -target-feature +xsaveopt -fno-signed-char -fclangir -emit-cir -o %t.cir -Wall -Wno-unused-but-set-variable -Werror +// RUN: FileCheck --check-prefix=CIR-XSAVEOPT --input-file=%t.cir %s +// RUN: %clang_cc1 %s -DTEST_XSAVEOPT -O0 -triple=x86_64-unknown-linux -target-feature +xsave -target-feature +xsaveopt -fno-signed-char -fclangir -emit-llvm -o %t.ll -Wall -Wno-unused-but-set-variable -Werror +// RUN: FileCheck --check-prefix=LLVM-XSAVEOPT --input-file=%t.ll %s + +// RUN: %clang_cc1 %s -DTEST_XSAVEC -O0 -triple=x86_64-unknown-linux -target-feature +xsave -target-feature +xsavec -fclangir -emit-cir -o %t.cir -Wall -Wno-unused-but-set-variable -Werror +// RUN: FileCheck --check-prefix=CIR-XSAVEC --input-file=%t.cir %s +// RUN: %clang_cc1 %s -DTEST_XSAVEC -O0 -triple=x86_64-unknown-linux -target-feature +xsave -target-feature +xsavec -fclangir -emit-llvm -o %t.ll -Wall -Wno-unused-but-set-variable -Werror +// RUN: FileCheck --check-prefix=LLVM-XSAVEC --input-file=%t.ll %s +// RUN: %clang_cc1 %s -DTEST_XSAVEC -O0 -triple=x86_64-unknown-linux -target-feature +xsave -target-feature +xsavec -fno-signed-char -fclangir -emit-cir -o %t.cir -Wall -Wno-unused-but-set-variable -Werror +// RUN: FileCheck --check-prefix=CIR-XSAVEC --input-file=%t.cir %s +// RUN: %clang_cc1 %s -DTEST_XSAVEC -O0 -triple=x86_64-unknown-linux -target-feature +xsave -target-feature +xsavec -fno-signed-char -fclangir -emit-llvm -o %t.ll -Wall -Wno-unused-but-set-variable -Werror +// RUN: FileCheck --check-prefix=LLVM-XSAVEC --input-file=%t.ll %s + +// RUN: %clang_cc1 %s -DTEST_XSAVES -O0 -triple=x86_64-unknown-linux -target-feature +xsave -target-feature +xsaves -fclangir -emit-cir -o %t.cir -Wall -Wno-unused-but-set-variable -Werror +// RUN: FileCheck --check-prefix=CIR-XSAVES --input-file=%t.cir %s +// RUN: %clang_cc1 %s -DTEST_XSAVES -O0 -triple=x86_64-unknown-linux -target-feature +xsave -target-feature +xsaves -fclangir -emit-llvm -o %t.ll -Wall -Wno-unused-but-set-variable -Werror +// RUN: FileCheck --check-prefix=LLVM-XSAVES --input-file=%t.ll %s +// RUN: %clang_cc1 %s -DTEST_XSAVES -O0 -triple=x86_64-unknown-linux -target-feature +xsave -target-feature +xsaves -fno-signed-char -fclangir -emit-cir -o %t.cir -Wall -Wno-unused-but-set-variable -Werror +// RUN: FileCheck --check-prefix=CIR-XSAVES --input-file=%t.cir %s +// RUN: %clang_cc1 %s -DTEST_XSAVES -O0 -triple=x86_64-unknown-linux -target-feature +xsave -target-feature +xsaves -fno-signed-char -fclangir -emit-llvm -o %t.ll -Wall -Wno-unused-but-set-variable -Werror +// RUN: FileCheck --check-prefix=LLVM-XSAVES --input-file=%t.ll %s + +// Don't include mm_malloc.h, it's system specific. +#define __MM_MALLOC_H +#include + + +void test(void) { + unsigned long long tmp_ULLi; + unsigned int tmp_Ui; + void* tmp_vp; + tmp_ULLi = 0; tmp_Ui = 0; tmp_vp = 0; + +#ifdef TEST_XSAVE +// CIR-XSAVE: [[tmp_vp_1:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr>, !cir.ptr +// CIR-XSAVE: [[tmp_ULLi_1:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr, !u64i +// CIR-XSAVE: [[high64_1:%.*]] = cir.shift(right, [[tmp_ULLi_1]] : !u64i, %{{.*}} : !u64i) -> !u64i +// CIR-XSAVE: [[high32_1:%.*]] = cir.cast(integral, [[high64_1]] : !u64i), !s32i +// CIR-XSAVE: [[low32_1:%.*]] = cir.cast(integral, [[tmp_ULLi_1]] : !u64i), !s32i +// CIR-XSAVE: %{{.*}} = cir.llvm.intrinsic "x86.xsave" [[tmp_vp_1]], [[high32_1]], [[low32_1]] : (!cir.ptr, !s32i, !s32i) -> !void + +// LLVM-XSAVE: [[tmp_vp_1:%.*]] = load ptr, ptr %{{.*}}, align 8 +// LLVM-XSAVE: [[tmp_ULLi_1:%.*]] = load i64, ptr %{{.*}}, align 8 +// LLVM-XSAVE: [[high64_1:%.*]] = lshr i64 [[tmp_ULLi_1]], 32 +// LLVM-XSAVE: [[high32_1:%.*]] = trunc i64 [[high64_1]] to i32 +// LLVM-XSAVE: [[low32_1:%.*]] = trunc i64 [[tmp_ULLi_1]] to i32 +// LLVM-XSAVE: call void @llvm.x86.xsave(ptr [[tmp_vp_1]], i32 [[high32_1]], i32 [[low32_1]]) + (void)__builtin_ia32_xsave(tmp_vp, tmp_ULLi); + + +// CIR-XSAVE: [[tmp_vp_2:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr>, !cir.ptr +// CIR-XSAVE: [[tmp_ULLi_2:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr, !u64i +// CIR-XSAVE: [[high64_2:%.*]] = cir.shift(right, [[tmp_ULLi_2]] : !u64i, %{{.*}} : !u64i) -> !u64i +// CIR-XSAVE: [[high32_2:%.*]] = cir.cast(integral, [[high64_2]] : !u64i), !s32i +// CIR-XSAVE: [[low32_2:%.*]] = cir.cast(integral, [[tmp_ULLi_2]] : !u64i), !s32i +// CIR-XSAVE: %{{.*}} = cir.llvm.intrinsic "x86.xsave64" [[tmp_vp_2]], [[high32_2]], [[low32_2]] : (!cir.ptr, !s32i, !s32i) -> !void + +// LLVM-XSAVE: [[tmp_vp_2:%.*]] = load ptr, ptr %{{.*}}, align 8 +// LLVM-XSAVE: [[tmp_ULLi_2:%.*]] = load i64, ptr %{{.*}}, align 8 +// LLVM-XSAVE: [[high64_2:%.*]] = lshr i64 [[tmp_ULLi_2]], 32 +// LLVM-XSAVE: [[high32_2:%.*]] = trunc i64 [[high64_2]] to i32 +// LLVM-XSAVE: [[low32_2:%.*]] = trunc i64 [[tmp_ULLi_2]] to i32 +// LLVM-XSAVE: call void @llvm.x86.xsave64(ptr [[tmp_vp_2]], i32 [[high32_2]], i32 [[low32_2]]) + (void)__builtin_ia32_xsave64(tmp_vp, tmp_ULLi); + + +// CIR-XSAVE: [[tmp_vp_3:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr>, !cir.ptr +// CIR-XSAVE: [[tmp_ULLi_3:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr, !u64i +// CIR-XSAVE: [[high64_3:%.*]] = cir.shift(right, [[tmp_ULLi_3]] : !u64i, %{{.*}} : !u64i) -> !u64i +// CIR-XSAVE: [[high32_3:%.*]] = cir.cast(integral, [[high64_3]] : !u64i), !s32i +// CIR-XSAVE: [[low32_3:%.*]] = cir.cast(integral, [[tmp_ULLi_3]] : !u64i), !s32i +// CIR-XSAVE: %{{.*}} = cir.llvm.intrinsic "x86.xrstor" [[tmp_vp_3]], [[high32_3]], [[low32_3]] : (!cir.ptr, !s32i, !s32i) -> !void + +// LLVM-XSAVE: [[tmp_vp_3:%.*]] = load ptr, ptr %{{.*}}, align 8 +// LLVM-XSAVE: [[tmp_ULLi_3:%.*]] = load i64, ptr %{{.*}}, align 8 +// LLVM-XSAVE: [[high64_3:%.*]] = lshr i64 [[tmp_ULLi_3]], 32 +// LLVM-XSAVE: [[high32_3:%.*]] = trunc i64 [[high64_3]] to i32 +// LLVM-XSAVE: [[low32_3:%.*]] = trunc i64 [[tmp_ULLi_3]] to i32 +// LLVM-XSAVE: call void @llvm.x86.xrstor(ptr [[tmp_vp_3]], i32 [[high32_3]], i32 [[low32_3]]) + (void)__builtin_ia32_xrstor(tmp_vp, tmp_ULLi); + + +// CIR-XSAVE: [[tmp_vp_4:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr>, !cir.ptr +// CIR-XSAVE: [[tmp_ULLi_4:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr, !u64i +// CIR-XSAVE: [[high64_4:%.*]] = cir.shift(right, [[tmp_ULLi_4]] : !u64i, %{{.*}} : !u64i) -> !u64i +// CIR-XSAVE: [[high32_4:%.*]] = cir.cast(integral, [[high64_4]] : !u64i), !s32i +// CIR-XSAVE: [[low32_4:%.*]] = cir.cast(integral, [[tmp_ULLi_4]] : !u64i), !s32i +// CIR-XSAVE: %{{.*}} = cir.llvm.intrinsic "x86.xrstor64" [[tmp_vp_4]], [[high32_4]], [[low32_4]] : (!cir.ptr, !s32i, !s32i) -> !void + +// LLVM-XSAVE: [[tmp_vp_4:%.*]] = load ptr, ptr %{{.*}}, align 8 +// LLVM-XSAVE: [[tmp_ULLi_4:%.*]] = load i64, ptr %{{.*}}, align 8 +// LLVM-XSAVE: [[high64_4:%.*]] = lshr i64 [[tmp_ULLi_4]], 32 +// LLVM-XSAVE: [[high32_4:%.*]] = trunc i64 [[high64_4]] to i32 +// LLVM-XSAVE: [[low32_4:%.*]] = trunc i64 [[tmp_ULLi_4]] to i32 +// LLVM-XSAVE: call void @llvm.x86.xrstor64(ptr [[tmp_vp_4]], i32 [[high32_4]], i32 [[low32_4]]) + (void)__builtin_ia32_xrstor64(tmp_vp, tmp_ULLi); + + +// CIR-XSAVE: {{%.*}} = cir.llvm.intrinsic "x86.xsave" {{%.*}} : (!cir.ptr, !s32i, !s32i) -> !void +// LLVM-XSAVE: call void @llvm.x86.xsave + (void)_xsave(tmp_vp, tmp_ULLi); + +// CIR-XSAVE: {{%.*}} = cir.llvm.intrinsic "x86.xsave64" {{%.*}} : (!cir.ptr, !s32i, !s32i) -> !void +// LLVM-XSAVE: call void @llvm.x86.xsave64 + (void)_xsave64(tmp_vp, tmp_ULLi); + +// CIR-XSAVE: {{%.*}} = cir.llvm.intrinsic "x86.xrstor" {{%.*}} : (!cir.ptr, !s32i, !s32i) -> !void +// LLVM-XSAVE: call void @llvm.x86.xrstor + (void)_xrstor(tmp_vp, tmp_ULLi); + +// CIR-XSAVE: {{%.*}} = cir.llvm.intrinsic "x86.xrstor64" {{%.*}} : (!cir.ptr, !s32i, !s32i) -> !void +// LLVM-XSAVE: call void @llvm.x86.xrstor64 + (void)_xrstor64(tmp_vp, tmp_ULLi); +#endif + +#ifdef TEST_XSAVEOPT +// CIR-XSAVEOPT: [[tmp_vp_1:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr>, !cir.ptr +// CIR-XSAVEOPT: [[tmp_ULLi_1:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr, !u64i +// CIR-XSAVEOPT: [[high64_1:%.*]] = cir.shift(right, [[tmp_ULLi_1]] : !u64i, %{{.*}} : !u64i) -> !u64i +// CIR-XSAVEOPT: [[high32_1:%.*]] = cir.cast(integral, [[high64_1]] : !u64i), !s32i +// CIR-XSAVEOPT: [[low32_1:%.*]] = cir.cast(integral, [[tmp_ULLi_1]] : !u64i), !s32i +// CIR-XSAVEOPT: %{{.*}} = cir.llvm.intrinsic "x86.xsaveopt" [[tmp_vp_1]], [[high32_1]], [[low32_1]] : (!cir.ptr, !s32i, !s32i) -> !void + +// LLVM-XSAVEOPT: [[tmp_vp_1:%.*]] = load ptr, ptr %{{.*}}, align 8 +// LLVM-XSAVEOPT: [[tmp_ULLi_1:%.*]] = load i64, ptr %{{.*}}, align 8 +// LLVM-XSAVEOPT: [[high64_1:%.*]] = lshr i64 [[tmp_ULLi_1]], 32 +// LLVM-XSAVEOPT: [[high32_1:%.*]] = trunc i64 [[high64_1]] to i32 +// LLVM-XSAVEOPT: [[low32_1:%.*]] = trunc i64 [[tmp_ULLi_1]] to i32 +// LLVM-XSAVEOPT: call void @llvm.x86.xsaveopt(ptr [[tmp_vp_1]], i32 [[high32_1]], i32 [[low32_1]]) + (void)__builtin_ia32_xsaveopt(tmp_vp, tmp_ULLi); + +// CIR-XSAVEOPT: [[tmp_vp_2:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr>, !cir.ptr +// CIR-XSAVEOPT: [[tmp_ULLi_2:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr, !u64i +// CIR-XSAVEOPT: [[high64_2:%.*]] = cir.shift(right, [[tmp_ULLi_2]] : !u64i, %{{.*}} : !u64i) -> !u64i +// CIR-XSAVEOPT: [[high32_2:%.*]] = cir.cast(integral, [[high64_2]] : !u64i), !s32i +// CIR-XSAVEOPT: [[low32_2:%.*]] = cir.cast(integral, [[tmp_ULLi_2]] : !u64i), !s32i +// CIR-XSAVEOPT: %{{.*}} = cir.llvm.intrinsic "x86.xsaveopt64" [[tmp_vp_2]], [[high32_2]], [[low32_2]] : (!cir.ptr, !s32i, !s32i) -> !void + +// LLVM-XSAVEOPT: [[tmp_vp_2:%.*]] = load ptr, ptr %{{.*}}, align 8 +// LLVM-XSAVEOPT: [[tmp_ULLi_2:%.*]] = load i64, ptr %{{.*}}, align 8 +// LLVM-XSAVEOPT: [[high64_2:%.*]] = lshr i64 [[tmp_ULLi_2]], 32 +// LLVM-XSAVEOPT: [[high32_2:%.*]] = trunc i64 [[high64_2]] to i32 +// LLVM-XSAVEOPT: [[low32_2:%.*]] = trunc i64 [[tmp_ULLi_2]] to i32 +// LLVM-XSAVEOPT: call void @llvm.x86.xsaveopt64(ptr [[tmp_vp_2]], i32 [[high32_2]], i32 [[low32_2]]) + (void)__builtin_ia32_xsaveopt64(tmp_vp, tmp_ULLi); + +// CIR-XSAVEOPT: {{%.*}} = cir.llvm.intrinsic "x86.xsaveopt" {{%.*}} : (!cir.ptr, !s32i, !s32i) -> !void +// LLVM-XSAVEOPT: call void @llvm.x86.xsaveopt + (void)_xsaveopt(tmp_vp, tmp_ULLi); + +// CIR-XSAVEOPT: {{%.*}} = cir.llvm.intrinsic "x86.xsaveopt64" {{%.*}} : (!cir.ptr, !s32i, !s32i) -> !void +// LLVM-XSAVEOPT: call void @llvm.x86.xsaveopt64 + (void)_xsaveopt64(tmp_vp, tmp_ULLi); +#endif + +#ifdef TEST_XSAVEC +// CIR-XSAVEC: [[tmp_vp_1:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr>, !cir.ptr +// CIR-XSAVEC: [[tmp_ULLi_1:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr, !u64i +// CIR-XSAVEC: [[high64_1:%.*]] = cir.shift(right, [[tmp_ULLi_1]] : !u64i, %{{.*}} : !u64i) -> !u64i +// CIR-XSAVEC: [[high32_1:%.*]] = cir.cast(integral, [[high64_1]] : !u64i), !s32i +// CIR-XSAVEC: [[low32_1:%.*]] = cir.cast(integral, [[tmp_ULLi_1]] : !u64i), !s32i +// CIR-XSAVEC: %{{.*}} = cir.llvm.intrinsic "x86.xsavec" [[tmp_vp_1]], [[high32_1]], [[low32_1]] : (!cir.ptr, !s32i, !s32i) -> !void + +// LLVM-XSAVEC: [[tmp_vp_1:%.*]] = load ptr, ptr %{{.*}}, align 8 +// LLVM-XSAVEC: [[tmp_ULLi_1:%.*]] = load i64, ptr %{{.*}}, align 8 +// LLVM-XSAVEC: [[high64_1:%.*]] = lshr i64 [[tmp_ULLi_1]], 32 +// LLVM-XSAVEC: [[high32_1:%.*]] = trunc i64 [[high64_1]] to i32 +// LLVM-XSAVEC: [[low32_1:%.*]] = trunc i64 [[tmp_ULLi_1]] to i32 +// LLVM-XSAVEC: call void @llvm.x86.xsavec(ptr [[tmp_vp_1]], i32 [[high32_1]], i32 [[low32_1]]) + (void)__builtin_ia32_xsavec(tmp_vp, tmp_ULLi); + + +// CIR-XSAVEC: [[tmp_vp_2:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr>, !cir.ptr +// CIR-XSAVEC: [[tmp_ULLi_2:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr, !u64i +// CIR-XSAVEC: [[high64_2:%.*]] = cir.shift(right, [[tmp_ULLi_2]] : !u64i, %{{.*}} : !u64i) -> !u64i +// CIR-XSAVEC: [[high32_2:%.*]] = cir.cast(integral, [[high64_2]] : !u64i), !s32i +// CIR-XSAVEC: [[low32_2:%.*]] = cir.cast(integral, [[tmp_ULLi_2]] : !u64i), !s32i +// CIR-XSAVEC: %{{.*}} = cir.llvm.intrinsic "x86.xsavec64" [[tmp_vp_2]], [[high32_2]], [[low32_2]] : (!cir.ptr, !s32i, !s32i) -> !void + +// LLVM-XSAVEC: [[tmp_vp_2:%.*]] = load ptr, ptr %{{.*}}, align 8 +// LLVM-XSAVEC: [[tmp_ULLi_2:%.*]] = load i64, ptr %{{.*}}, align 8 +// LLVM-XSAVEC: [[high64_2:%.*]] = lshr i64 [[tmp_ULLi_2]], 32 +// LLVM-XSAVEC: [[high32_2:%.*]] = trunc i64 [[high64_2]] to i32 +// LLVM-XSAVEC: [[low32_2:%.*]] = trunc i64 [[tmp_ULLi_2]] to i32 +// LLVM-XSAVEC: call void @llvm.x86.xsavec64(ptr [[tmp_vp_2]], i32 [[high32_2]], i32 [[low32_2]]) + (void)__builtin_ia32_xsavec64(tmp_vp, tmp_ULLi); + +// CIR-XSAVEC: {{%.*}} = cir.llvm.intrinsic "x86.xsavec" {{%.*}} : (!cir.ptr, !s32i, !s32i) -> !void +// LLVM-XSAVEC: call void @llvm.x86.xsavec + (void)_xsavec(tmp_vp, tmp_ULLi); + +// CIR-XSAVEC: {{%.*}} = cir.llvm.intrinsic "x86.xsavec64" {{%.*}} : (!cir.ptr, !s32i, !s32i) -> !void +// LLVM-XSAVEC: call void @llvm.x86.xsavec64 + (void)_xsavec64(tmp_vp, tmp_ULLi); +#endif + +#ifdef TEST_XSAVES +// CIR-XSAVES: [[tmp_vp_1:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr>, !cir.ptr +// CIR-XSAVES: [[tmp_ULLi_1:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr, !u64i +// CIR-XSAVES: [[high64_1:%.*]] = cir.shift(right, [[tmp_ULLi_1]] : !u64i, %{{.*}} : !u64i) -> !u64i +// CIR-XSAVES: [[high32_1:%.*]] = cir.cast(integral, [[high64_1]] : !u64i), !s32i +// CIR-XSAVES: [[low32_1:%.*]] = cir.cast(integral, [[tmp_ULLi_1]] : !u64i), !s32i +// CIR-XSAVES: %{{.*}} = cir.llvm.intrinsic "x86.xsaves" [[tmp_vp_1]], [[high32_1]], [[low32_1]] : (!cir.ptr, !s32i, !s32i) -> !void + +// LLVM-XSAVES: [[tmp_vp_1:%.*]] = load ptr, ptr %{{.*}}, align 8 +// LLVM-XSAVES: [[tmp_ULLi_1:%.*]] = load i64, ptr %{{.*}}, align 8 +// LLVM-XSAVES: [[high64_1:%.*]] = lshr i64 [[tmp_ULLi_1]], 32 +// LLVM-XSAVES: [[high32_1:%.*]] = trunc i64 [[high64_1]] to i32 +// LLVM-XSAVES: [[low32_1:%.*]] = trunc i64 [[tmp_ULLi_1]] to i32 +// LLVM-XSAVES: call void @llvm.x86.xsaves(ptr [[tmp_vp_1]], i32 [[high32_1]], i32 [[low32_1]]) + (void)__builtin_ia32_xsaves(tmp_vp, tmp_ULLi); + + +// CIR-XSAVES: [[tmp_vp_2:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr>, !cir.ptr +// CIR-XSAVES: [[tmp_ULLi_2:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr, !u64i +// CIR-XSAVES: [[high64_2:%.*]] = cir.shift(right, [[tmp_ULLi_2]] : !u64i, %{{.*}} : !u64i) -> !u64i +// CIR-XSAVES: [[high32_2:%.*]] = cir.cast(integral, [[high64_2]] : !u64i), !s32i +// CIR-XSAVES: [[low32_2:%.*]] = cir.cast(integral, [[tmp_ULLi_2]] : !u64i), !s32i +// CIR-XSAVES: %{{.*}} = cir.llvm.intrinsic "x86.xsaves64" [[tmp_vp_2]], [[high32_2]], [[low32_2]] : (!cir.ptr, !s32i, !s32i) -> !void + +// LLVM-XSAVES: [[tmp_vp_2:%.*]] = load ptr, ptr %{{.*}}, align 8 +// LLVM-XSAVES: [[tmp_ULLi_2:%.*]] = load i64, ptr %{{.*}}, align 8 +// LLVM-XSAVES: [[high64_2:%.*]] = lshr i64 [[tmp_ULLi_2]], 32 +// LLVM-XSAVES: [[high32_2:%.*]] = trunc i64 [[high64_2]] to i32 +// LLVM-XSAVES: [[low32_2:%.*]] = trunc i64 [[tmp_ULLi_2]] to i32 +// LLVM-XSAVES: call void @llvm.x86.xsaves64(ptr [[tmp_vp_2]], i32 [[high32_2]], i32 [[low32_2]]) + (void)__builtin_ia32_xsaves64(tmp_vp, tmp_ULLi); + + +// CIR-XSAVES: [[tmp_vp_3:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr>, !cir.ptr +// CIR-XSAVES: [[tmp_ULLi_3:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr, !u64i +// CIR-XSAVES: [[high64_3:%.*]] = cir.shift(right, [[tmp_ULLi_3]] : !u64i, %{{.*}} : !u64i) -> !u64i +// CIR-XSAVES: [[high32_3:%.*]] = cir.cast(integral, [[high64_3]] : !u64i), !s32i +// CIR-XSAVES: [[low32_3:%.*]] = cir.cast(integral, [[tmp_ULLi_3]] : !u64i), !s32i +// CIR-XSAVES: %{{.*}} = cir.llvm.intrinsic "x86.xrstors" [[tmp_vp_3]], [[high32_3]], [[low32_3]] : (!cir.ptr, !s32i, !s32i) -> !void + +// LLVM-XSAVES: [[tmp_vp_3:%.*]] = load ptr, ptr %{{.*}}, align 8 +// LLVM-XSAVES: [[tmp_ULLi_3:%.*]] = load i64, ptr %{{.*}}, align 8 +// LLVM-XSAVES: [[high64_3:%.*]] = lshr i64 [[tmp_ULLi_3]], 32 +// LLVM-XSAVES: [[high32_3:%.*]] = trunc i64 [[high64_3]] to i32 +// LLVM-XSAVES: [[low32_3:%.*]] = trunc i64 [[tmp_ULLi_3]] to i32 +// LLVM-XSAVES: call void @llvm.x86.xrstors(ptr [[tmp_vp_3]], i32 [[high32_3]], i32 [[low32_3]]) + (void)__builtin_ia32_xrstors(tmp_vp, tmp_ULLi); + + +// CIR-XSAVES: [[tmp_vp_4:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr>, !cir.ptr +// CIR-XSAVES: [[tmp_ULLi_4:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr, !u64i +// CIR-XSAVES: [[high64_4:%.*]] = cir.shift(right, [[tmp_ULLi_4]] : !u64i, %{{.*}} : !u64i) -> !u64i +// CIR-XSAVES: [[high32_4:%.*]] = cir.cast(integral, [[high64_4]] : !u64i), !s32i +// CIR-XSAVES: [[low32_4:%.*]] = cir.cast(integral, [[tmp_ULLi_4]] : !u64i), !s32i +// CIR-XSAVES: %{{.*}} = cir.llvm.intrinsic "x86.xrstors64" [[tmp_vp_4]], [[high32_4]], [[low32_4]] : (!cir.ptr, !s32i, !s32i) -> !void + +// LLVM-XSAVES: [[tmp_vp_4:%.*]] = load ptr, ptr %{{.*}}, align 8 +// LLVM-XSAVES: [[tmp_ULLi_4:%.*]] = load i64, ptr %{{.*}}, align 8 +// LLVM-XSAVES: [[high64_4:%.*]] = lshr i64 [[tmp_ULLi_4]], 32 +// LLVM-XSAVES: [[high32_4:%.*]] = trunc i64 [[high64_4]] to i32 +// LLVM-XSAVES: [[low32_4:%.*]] = trunc i64 [[tmp_ULLi_4]] to i32 +// LLVM-XSAVES: call void @llvm.x86.xrstors64(ptr [[tmp_vp_4]], i32 [[high32_4]], i32 [[low32_4]]) + (void)__builtin_ia32_xrstors64(tmp_vp, tmp_ULLi); + + +// CIR-XSAVES: {{%.*}} = cir.llvm.intrinsic "x86.xsaves" {{%.*}} : (!cir.ptr, !s32i, !s32i) -> !void +// LLVM-XSAVES: call void @llvm.x86.xsaves + (void)_xsaves(tmp_vp, tmp_ULLi); + +// CIR-XSAVES: {{%.*}} = cir.llvm.intrinsic "x86.xsaves64" {{%.*}} : (!cir.ptr, !s32i, !s32i) -> !void +// LLVM-XSAVES: call void @llvm.x86.xsaves64 + (void)_xsaves64(tmp_vp, tmp_ULLi); + +// CIR-XSAVES: {{%.*}} = cir.llvm.intrinsic "x86.xrstors" {{%.*}} : (!cir.ptr, !s32i, !s32i) -> !void +// LLVM-XSAVES: call void @llvm.x86.xrstors + (void)_xrstors(tmp_vp, tmp_ULLi); + +// CIR-XSAVES: {{%.*}} = cir.llvm.intrinsic "x86.xrstors64" {{%.*}} : (!cir.ptr, !s32i, !s32i) -> !void +// LLVM-XSAVES: call void @llvm.x86.xrstors64 + (void)_xrstors64(tmp_vp, tmp_ULLi); +#endif + +#ifdef TEST_XGETBV + +// CIR-XGETBV: [[tmp_Ui:%.*]] = cir.load align(4) %{{.*}} : !cir.ptr, !u32i +// CIR-XGETBV: {{%.*}} = cir.llvm.intrinsic "x86.xgetbv" [[tmp_Ui]] : (!u32i) -> !u64i + +// LLVM-XGETBV: [[tmp_Ui:%.*]] = load i32, ptr %{{.*}}, align 4 +// LLVM-XGETBV: call i64 @llvm.x86.xgetbv(i32 [[tmp_Ui]]) + tmp_ULLi = __builtin_ia32_xgetbv(tmp_Ui); + +// CIR-XGETBV: {{%.*}} = cir.llvm.intrinsic "x86.xgetbv" {{%.*}} : (!u32i) -> !u64i +// LLVM-XGETBV: call i64 @llvm.x86.xgetbv + tmp_ULLi = _xgetbv(tmp_Ui); +#endif + +#ifdef TEST_XSETBV +// CIR-XSETBV: [[tmp_Ui_1:%.*]] = cir.load align(4) %{{.*}} : !cir.ptr, !u32i +// CIR-XSETBV: [[tmp_ULLi_1:%.*]] = cir.load align(8) %{{.*}} : !cir.ptr, !u64i +// CIR-XSETBV: [[high64_1:%.*]] = cir.shift(right, [[tmp_ULLi_1]] : !u64i, %{{.*}} : !u64i) -> !u64i +// CIR-XSETBV: [[high32_1:%.*]] = cir.cast(integral, [[high64_1]] : !u64i), !s32i +// CIR-XSETBV: [[low32_1:%.*]] = cir.cast(integral, [[tmp_ULLi_1]] : !u64i), !s32i +// CIR-XSETBV: %{{.*}} = cir.llvm.intrinsic "x86.xsetbv" [[tmp_Ui_1]], [[high32_1]], [[low32_1]] : (!u32i, !s32i, !s32i) -> !void + +// LLVM-XSETBV: [[tmp_Ui_1:%.*]] = load i32, ptr %{{.*}}, align 4 +// LLVM-XSETBV: [[tmp_ULLi_1:%.*]] = load i64, ptr %{{.*}}, align 8 +// LLVM-XSETBV: [[high64_1:%.*]] = lshr i64 [[tmp_ULLi_1]], 32 +// LLVM-XSETBV: [[high32_1:%.*]] = trunc i64 [[high64_1]] to i32 +// LLVM-XSETBV: [[low32_1:%.*]] = trunc i64 [[tmp_ULLi_1]] to i32 +// LLVM-XSETBV: call void @llvm.x86.xsetbv(i32 [[tmp_Ui_1]], i32 [[high32_1]], i32 [[low32_1]]) + (void)__builtin_ia32_xsetbv(tmp_Ui, tmp_ULLi); + +// CIR-XSETBV: {{%.*}} = cir.llvm.intrinsic "x86.xsetbv" {{%.*}} : (!u32i, !s32i, !s32i) -> !void +// LLVM-XSETBV: call void @llvm.x86.xsetbv + (void)_xsetbv(tmp_Ui, tmp_ULLi); +#endif +} \ No newline at end of file