diff options
author | Tom Stellard <tstellar@redhat.com> | 2018-10-19 19:11:17 +0000 |
---|---|---|
committer | Tom Stellard <tstellar@redhat.com> | 2018-10-19 19:11:17 +0000 |
commit | 9a0352f0c5ea78372340a101bb557aba3ecc3f7e (patch) | |
tree | 60da8d39ddcc465dae424532eefe6d2f18a06f5d /test/CodeGen | |
parent | 6c2e5a132cdbcc9ccea6eb3269fdcfc3ea67572c (diff) |
Merging r343443:
------------------------------------------------------------------------
r343443 | ctopper | 2018-10-01 00:08:41 -0700 (Mon, 01 Oct 2018) | 9 lines
[X86] Stop X86DomainReassignment from creating copies between GR8/GR16 physical registers and k-registers.
We can only copy between a k-register and a GR32/GR64 register.
This patch detects that the copy will be illegal and prevents the domain reassignment from happening for that closure.
This probably isn't the best fix, and we should probably figure out how to handle this correctly.
Fixes PR38803.
------------------------------------------------------------------------
git-svn-id: https://llvm.org/svn/llvm-project/llvm/branches/release_70@344804 91177308-0d34-0410-b5e6-96231b3b80d8
Diffstat (limited to 'test/CodeGen')
-rw-r--r-- | test/CodeGen/X86/pr38803.ll | 48 |
1 files changed, 48 insertions, 0 deletions
diff --git a/test/CodeGen/X86/pr38803.ll b/test/CodeGen/X86/pr38803.ll new file mode 100644 index 00000000000..a1ca27d89b0 --- /dev/null +++ b/test/CodeGen/X86/pr38803.ll @@ -0,0 +1,48 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc < %s -mcpu=skylake-avx512 -mtriple=x86_64-unknown-unknown | FileCheck %s + +@b = local_unnamed_addr global i32 0, align 4 +@c = local_unnamed_addr global i32 0, align 4 +@d = local_unnamed_addr global float 0.000000e+00, align 4 + +define float @_Z3fn2v() { +; CHECK-LABEL: _Z3fn2v: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: pushq %rax +; CHECK-NEXT: .cfi_def_cfa_offset 16 +; CHECK-NEXT: callq _Z1av +; CHECK-NEXT: # kill: def $al killed $al def $eax +; CHECK-NEXT: kmovd %eax, %k1 +; CHECK-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero +; CHECK-NEXT: vmovss %xmm0, %xmm0, %xmm0 {%k1} {z} +; CHECK-NEXT: cmpl $0, {{.*}}(%rip) +; CHECK-NEXT: je .LBB0_2 +; CHECK-NEXT: # %bb.1: # %if.then +; CHECK-NEXT: vcvtsi2ssl {{.*}}(%rip), %xmm1, %xmm1 +; CHECK-NEXT: kmovd %eax, %k1 +; CHECK-NEXT: vxorps %xmm2, %xmm2, %xmm2 +; CHECK-NEXT: vmovss %xmm2, %xmm0, %xmm1 {%k1} +; CHECK-NEXT: vmovss %xmm1, {{.*}}(%rip) +; CHECK-NEXT: .LBB0_2: # %if.end +; CHECK-NEXT: popq %rax +; CHECK-NEXT: .cfi_def_cfa_offset 8 +; CHECK-NEXT: retq +entry: + %call = tail call zeroext i1 @_Z1av() + %cond = select i1 %call, float 7.500000e-01, float 0.000000e+00 + %0 = load i32, i32* @c, align 4 + %tobool2 = icmp eq i32 %0, 0 + br i1 %tobool2, label %if.end, label %if.then + +if.then: ; preds = %entry + %1 = load i32, i32* @b, align 4 + %2 = sitofp i32 %1 to float + %conv5 = select i1 %call, float 0.000000e+00, float %2 + store float %conv5, float* @d, align 4 + br label %if.end + +if.end: ; preds = %entry, %if.then + ret float %cond +} + +declare zeroext i1 @_Z1av() |