summaryrefslogtreecommitdiff
path: root/test
diff options
context:
space:
mode:
Diffstat (limited to 'test')
-rw-r--r--test/CodeGen/X86/O0-pipeline.ll3
-rw-r--r--test/CodeGen/X86/retpoline-external.ll166
-rw-r--r--test/CodeGen/X86/retpoline.ll367
-rw-r--r--test/Transforms/IndirectBrExpand/basic.ll63
4 files changed, 599 insertions, 0 deletions
diff --git a/test/CodeGen/X86/O0-pipeline.ll b/test/CodeGen/X86/O0-pipeline.ll
index cb7dabefe45..efc2eaa89b7 100644
--- a/test/CodeGen/X86/O0-pipeline.ll
+++ b/test/CodeGen/X86/O0-pipeline.ll
@@ -25,6 +25,7 @@
; CHECK-NEXT: Instrument function entry/exit with calls to e.g. mcount() (post inlining)
; CHECK-NEXT: Scalarize Masked Memory Intrinsics
; CHECK-NEXT: Expand reduction intrinsics
+; CHECK-NEXT: Expand indirectbr instructions
; CHECK-NEXT: Rewrite Symbols
; CHECK-NEXT: FunctionPass Manager
; CHECK-NEXT: Dominator Tree Construction
@@ -57,6 +58,8 @@
; CHECK-NEXT: Machine Natural Loop Construction
; CHECK-NEXT: Insert XRay ops
; CHECK-NEXT: Implement the 'patchable-function' attribute
+; CHECK-NEXT: X86 Retpoline Thunks
+; CHECK-NEXT: FunctionPass Manager
; CHECK-NEXT: Lazy Machine Block Frequency Analysis
; CHECK-NEXT: Machine Optimization Remark Emitter
; CHECK-NEXT: MachineDominator Tree Construction
diff --git a/test/CodeGen/X86/retpoline-external.ll b/test/CodeGen/X86/retpoline-external.ll
new file mode 100644
index 00000000000..66d32ba5d73
--- /dev/null
+++ b/test/CodeGen/X86/retpoline-external.ll
@@ -0,0 +1,166 @@
+; RUN: llc -mtriple=x86_64-unknown < %s | FileCheck %s --implicit-check-not="jmp.*\*" --implicit-check-not="call.*\*" --check-prefix=X64
+; RUN: llc -mtriple=x86_64-unknown -O0 < %s | FileCheck %s --implicit-check-not="jmp.*\*" --implicit-check-not="call.*\*" --check-prefix=X64FAST
+
+; RUN: llc -mtriple=i686-unknown < %s | FileCheck %s --implicit-check-not="jmp.*\*" --implicit-check-not="call.*\*" --check-prefix=X86
+; RUN: llc -mtriple=i686-unknown -O0 < %s | FileCheck %s --implicit-check-not="jmp.*\*" --implicit-check-not="call.*\*" --check-prefix=X86FAST
+
+declare void @bar(i32)
+
+; Test a simple indirect call and tail call.
+define void @icall_reg(void (i32)* %fp, i32 %x) #0 {
+entry:
+ tail call void @bar(i32 %x)
+ tail call void %fp(i32 %x)
+ tail call void @bar(i32 %x)
+ tail call void %fp(i32 %x)
+ ret void
+}
+
+; X64-LABEL: icall_reg:
+; X64-DAG: movq %rdi, %[[fp:[^ ]*]]
+; X64-DAG: movl %esi, %[[x:[^ ]*]]
+; X64: movl %[[x]], %edi
+; X64: callq bar
+; X64-DAG: movl %[[x]], %edi
+; X64-DAG: movq %[[fp]], %r11
+; X64: callq __llvm_external_retpoline_r11
+; X64: movl %[[x]], %edi
+; X64: callq bar
+; X64-DAG: movl %[[x]], %edi
+; X64-DAG: movq %[[fp]], %r11
+; X64: jmp __llvm_external_retpoline_r11 # TAILCALL
+
+; X64FAST-LABEL: icall_reg:
+; X64FAST: callq bar
+; X64FAST: callq __llvm_external_retpoline_r11
+; X64FAST: callq bar
+; X64FAST: jmp __llvm_external_retpoline_r11 # TAILCALL
+
+; X86-LABEL: icall_reg:
+; X86-DAG: movl 12(%esp), %[[fp:[^ ]*]]
+; X86-DAG: movl 16(%esp), %[[x:[^ ]*]]
+; X86: pushl %[[x]]
+; X86: calll bar
+; X86: movl %[[fp]], %eax
+; X86: pushl %[[x]]
+; X86: calll __llvm_external_retpoline_eax
+; X86: pushl %[[x]]
+; X86: calll bar
+; X86: movl %[[fp]], %eax
+; X86: pushl %[[x]]
+; X86: calll __llvm_external_retpoline_eax
+; X86-NOT: # TAILCALL
+
+; X86FAST-LABEL: icall_reg:
+; X86FAST: calll bar
+; X86FAST: calll __llvm_external_retpoline_eax
+; X86FAST: calll bar
+; X86FAST: calll __llvm_external_retpoline_eax
+
+
+@global_fp = external global void (i32)*
+
+; Test an indirect call through a global variable.
+define void @icall_global_fp(i32 %x, void (i32)** %fpp) #0 {
+ %fp1 = load void (i32)*, void (i32)** @global_fp
+ call void %fp1(i32 %x)
+ %fp2 = load void (i32)*, void (i32)** @global_fp
+ tail call void %fp2(i32 %x)
+ ret void
+}
+
+; X64-LABEL: icall_global_fp:
+; X64-DAG: movl %edi, %[[x:[^ ]*]]
+; X64-DAG: movq global_fp(%rip), %r11
+; X64: callq __llvm_external_retpoline_r11
+; X64-DAG: movl %[[x]], %edi
+; X64-DAG: movq global_fp(%rip), %r11
+; X64: jmp __llvm_external_retpoline_r11 # TAILCALL
+
+; X64FAST-LABEL: icall_global_fp:
+; X64FAST: movq global_fp(%rip), %r11
+; X64FAST: callq __llvm_external_retpoline_r11
+; X64FAST: movq global_fp(%rip), %r11
+; X64FAST: jmp __llvm_external_retpoline_r11 # TAILCALL
+
+; X86-LABEL: icall_global_fp:
+; X86: movl global_fp, %eax
+; X86: pushl 4(%esp)
+; X86: calll __llvm_external_retpoline_eax
+; X86: addl $4, %esp
+; X86: movl global_fp, %eax
+; X86: jmp __llvm_external_retpoline_eax # TAILCALL
+
+; X86FAST-LABEL: icall_global_fp:
+; X86FAST: calll __llvm_external_retpoline_eax
+; X86FAST: jmp __llvm_external_retpoline_eax # TAILCALL
+
+
+%struct.Foo = type { void (%struct.Foo*)** }
+
+; Test an indirect call through a vtable.
+define void @vcall(%struct.Foo* %obj) #0 {
+ %vptr_field = getelementptr %struct.Foo, %struct.Foo* %obj, i32 0, i32 0
+ %vptr = load void (%struct.Foo*)**, void (%struct.Foo*)*** %vptr_field
+ %vslot = getelementptr void(%struct.Foo*)*, void(%struct.Foo*)** %vptr, i32 1
+ %fp = load void(%struct.Foo*)*, void(%struct.Foo*)** %vslot
+ tail call void %fp(%struct.Foo* %obj)
+ tail call void %fp(%struct.Foo* %obj)
+ ret void
+}
+
+; X64-LABEL: vcall:
+; X64: movq %rdi, %[[obj:[^ ]*]]
+; X64: movq (%[[obj]]), %[[vptr:[^ ]*]]
+; X64: movq 8(%[[vptr]]), %[[fp:[^ ]*]]
+; X64: movq %[[fp]], %r11
+; X64: callq __llvm_external_retpoline_r11
+; X64-DAG: movq %[[obj]], %rdi
+; X64-DAG: movq %[[fp]], %r11
+; X64: jmp __llvm_external_retpoline_r11 # TAILCALL
+
+; X64FAST-LABEL: vcall:
+; X64FAST: callq __llvm_external_retpoline_r11
+; X64FAST: jmp __llvm_external_retpoline_r11 # TAILCALL
+
+; X86-LABEL: vcall:
+; X86: movl 8(%esp), %[[obj:[^ ]*]]
+; X86: movl (%[[obj]]), %[[vptr:[^ ]*]]
+; X86: movl 4(%[[vptr]]), %[[fp:[^ ]*]]
+; X86: movl %[[fp]], %eax
+; X86: pushl %[[obj]]
+; X86: calll __llvm_external_retpoline_eax
+; X86: addl $4, %esp
+; X86: movl %[[fp]], %eax
+; X86: jmp __llvm_external_retpoline_eax # TAILCALL
+
+; X86FAST-LABEL: vcall:
+; X86FAST: calll __llvm_external_retpoline_eax
+; X86FAST: jmp __llvm_external_retpoline_eax # TAILCALL
+
+
+declare void @direct_callee()
+
+define void @direct_tail() #0 {
+ tail call void @direct_callee()
+ ret void
+}
+
+; X64-LABEL: direct_tail:
+; X64: jmp direct_callee # TAILCALL
+; X64FAST-LABEL: direct_tail:
+; X64FAST: jmp direct_callee # TAILCALL
+; X86-LABEL: direct_tail:
+; X86: jmp direct_callee # TAILCALL
+; X86FAST-LABEL: direct_tail:
+; X86FAST: jmp direct_callee # TAILCALL
+
+
+; Lastly check that no thunks were emitted.
+; X64-NOT: __{{.*}}_retpoline_{{.*}}:
+; X64FAST-NOT: __{{.*}}_retpoline_{{.*}}:
+; X86-NOT: __{{.*}}_retpoline_{{.*}}:
+; X86FAST-NOT: __{{.*}}_retpoline_{{.*}}:
+
+
+attributes #0 = { "target-features"="+retpoline-external-thunk" }
diff --git a/test/CodeGen/X86/retpoline.ll b/test/CodeGen/X86/retpoline.ll
new file mode 100644
index 00000000000..57d3388b812
--- /dev/null
+++ b/test/CodeGen/X86/retpoline.ll
@@ -0,0 +1,367 @@
+; RUN: llc -mtriple=x86_64-unknown < %s | FileCheck %s --implicit-check-not="jmp.*\*" --implicit-check-not="call.*\*" --check-prefix=X64
+; RUN: llc -mtriple=x86_64-unknown -O0 < %s | FileCheck %s --implicit-check-not="jmp.*\*" --implicit-check-not="call.*\*" --check-prefix=X64FAST
+
+; RUN: llc -mtriple=i686-unknown < %s | FileCheck %s --implicit-check-not="jmp.*\*" --implicit-check-not="call.*\*" --check-prefix=X86
+; RUN: llc -mtriple=i686-unknown -O0 < %s | FileCheck %s --implicit-check-not="jmp.*\*" --implicit-check-not="call.*\*" --check-prefix=X86FAST
+
+declare void @bar(i32)
+
+; Test a simple indirect call and tail call.
+define void @icall_reg(void (i32)* %fp, i32 %x) #0 {
+entry:
+ tail call void @bar(i32 %x)
+ tail call void %fp(i32 %x)
+ tail call void @bar(i32 %x)
+ tail call void %fp(i32 %x)
+ ret void
+}
+
+; X64-LABEL: icall_reg:
+; X64-DAG: movq %rdi, %[[fp:[^ ]*]]
+; X64-DAG: movl %esi, %[[x:[^ ]*]]
+; X64: movl %[[x]], %edi
+; X64: callq bar
+; X64-DAG: movl %[[x]], %edi
+; X64-DAG: movq %[[fp]], %r11
+; X64: callq __llvm_retpoline_r11
+; X64: movl %[[x]], %edi
+; X64: callq bar
+; X64-DAG: movl %[[x]], %edi
+; X64-DAG: movq %[[fp]], %r11
+; X64: jmp __llvm_retpoline_r11 # TAILCALL
+
+; X64FAST-LABEL: icall_reg:
+; X64FAST: callq bar
+; X64FAST: callq __llvm_retpoline_r11
+; X64FAST: callq bar
+; X64FAST: jmp __llvm_retpoline_r11 # TAILCALL
+
+; X86-LABEL: icall_reg:
+; X86-DAG: movl 12(%esp), %[[fp:[^ ]*]]
+; X86-DAG: movl 16(%esp), %[[x:[^ ]*]]
+; X86: pushl %[[x]]
+; X86: calll bar
+; X86: movl %[[fp]], %eax
+; X86: pushl %[[x]]
+; X86: calll __llvm_retpoline_eax
+; X86: pushl %[[x]]
+; X86: calll bar
+; X86: movl %[[fp]], %eax
+; X86: pushl %[[x]]
+; X86: calll __llvm_retpoline_eax
+; X86-NOT: # TAILCALL
+
+; X86FAST-LABEL: icall_reg:
+; X86FAST: calll bar
+; X86FAST: calll __llvm_retpoline_eax
+; X86FAST: calll bar
+; X86FAST: calll __llvm_retpoline_eax
+
+
+@global_fp = external global void (i32)*
+
+; Test an indirect call through a global variable.
+define void @icall_global_fp(i32 %x, void (i32)** %fpp) #0 {
+ %fp1 = load void (i32)*, void (i32)** @global_fp
+ call void %fp1(i32 %x)
+ %fp2 = load void (i32)*, void (i32)** @global_fp
+ tail call void %fp2(i32 %x)
+ ret void
+}
+
+; X64-LABEL: icall_global_fp:
+; X64-DAG: movl %edi, %[[x:[^ ]*]]
+; X64-DAG: movq global_fp(%rip), %r11
+; X64: callq __llvm_retpoline_r11
+; X64-DAG: movl %[[x]], %edi
+; X64-DAG: movq global_fp(%rip), %r11
+; X64: jmp __llvm_retpoline_r11 # TAILCALL
+
+; X64FAST-LABEL: icall_global_fp:
+; X64FAST: movq global_fp(%rip), %r11
+; X64FAST: callq __llvm_retpoline_r11
+; X64FAST: movq global_fp(%rip), %r11
+; X64FAST: jmp __llvm_retpoline_r11 # TAILCALL
+
+; X86-LABEL: icall_global_fp:
+; X86: movl global_fp, %eax
+; X86: pushl 4(%esp)
+; X86: calll __llvm_retpoline_eax
+; X86: addl $4, %esp
+; X86: movl global_fp, %eax
+; X86: jmp __llvm_retpoline_eax # TAILCALL
+
+; X86FAST-LABEL: icall_global_fp:
+; X86FAST: calll __llvm_retpoline_eax
+; X86FAST: jmp __llvm_retpoline_eax # TAILCALL
+
+
+%struct.Foo = type { void (%struct.Foo*)** }
+
+; Test an indirect call through a vtable.
+define void @vcall(%struct.Foo* %obj) #0 {
+ %vptr_field = getelementptr %struct.Foo, %struct.Foo* %obj, i32 0, i32 0
+ %vptr = load void (%struct.Foo*)**, void (%struct.Foo*)*** %vptr_field
+ %vslot = getelementptr void(%struct.Foo*)*, void(%struct.Foo*)** %vptr, i32 1
+ %fp = load void(%struct.Foo*)*, void(%struct.Foo*)** %vslot
+ tail call void %fp(%struct.Foo* %obj)
+ tail call void %fp(%struct.Foo* %obj)
+ ret void
+}
+
+; X64-LABEL: vcall:
+; X64: movq %rdi, %[[obj:[^ ]*]]
+; X64: movq (%[[obj]]), %[[vptr:[^ ]*]]
+; X64: movq 8(%[[vptr]]), %[[fp:[^ ]*]]
+; X64: movq %[[fp]], %r11
+; X64: callq __llvm_retpoline_r11
+; X64-DAG: movq %[[obj]], %rdi
+; X64-DAG: movq %[[fp]], %r11
+; X64: jmp __llvm_retpoline_r11 # TAILCALL
+
+; X64FAST-LABEL: vcall:
+; X64FAST: callq __llvm_retpoline_r11
+; X64FAST: jmp __llvm_retpoline_r11 # TAILCALL
+
+; X86-LABEL: vcall:
+; X86: movl 8(%esp), %[[obj:[^ ]*]]
+; X86: movl (%[[obj]]), %[[vptr:[^ ]*]]
+; X86: movl 4(%[[vptr]]), %[[fp:[^ ]*]]
+; X86: movl %[[fp]], %eax
+; X86: pushl %[[obj]]
+; X86: calll __llvm_retpoline_eax
+; X86: addl $4, %esp
+; X86: movl %[[fp]], %eax
+; X86: jmp __llvm_retpoline_eax # TAILCALL
+
+; X86FAST-LABEL: vcall:
+; X86FAST: calll __llvm_retpoline_eax
+; X86FAST: jmp __llvm_retpoline_eax # TAILCALL
+
+
+declare void @direct_callee()
+
+define void @direct_tail() #0 {
+ tail call void @direct_callee()
+ ret void
+}
+
+; X64-LABEL: direct_tail:
+; X64: jmp direct_callee # TAILCALL
+; X64FAST-LABEL: direct_tail:
+; X64FAST: jmp direct_callee # TAILCALL
+; X86-LABEL: direct_tail:
+; X86: jmp direct_callee # TAILCALL
+; X86FAST-LABEL: direct_tail:
+; X86FAST: jmp direct_callee # TAILCALL
+
+
+declare void @nonlazybind_callee() #1
+
+define void @nonlazybind_caller() #0 {
+ call void @nonlazybind_callee()
+ tail call void @nonlazybind_callee()
+ ret void
+}
+
+; X64-LABEL: nonlazybind_caller:
+; X64: movq nonlazybind_callee@GOTPCREL(%rip), %[[REG:.*]]
+; X64: movq %[[REG]], %r11
+; X64: callq __llvm_retpoline_r11
+; X64: movq %[[REG]], %r11
+; X64: jmp __llvm_retpoline_r11 # TAILCALL
+; X64FAST-LABEL: nonlazybind_caller:
+; X64FAST: movq nonlazybind_callee@GOTPCREL(%rip), %r11
+; X64FAST: callq __llvm_retpoline_r11
+; X64FAST: movq nonlazybind_callee@GOTPCREL(%rip), %r11
+; X64FAST: jmp __llvm_retpoline_r11 # TAILCALL
+; X86-LABEL: nonlazybind_caller:
+; X86: calll nonlazybind_callee@PLT
+; X86: jmp nonlazybind_callee@PLT # TAILCALL
+; X86FAST-LABEL: nonlazybind_caller:
+; X86FAST: calll nonlazybind_callee@PLT
+; X86FAST: jmp nonlazybind_callee@PLT # TAILCALL
+
+
+@indirectbr_rewrite.targets = constant [10 x i8*] [i8* blockaddress(@indirectbr_rewrite, %bb0),
+ i8* blockaddress(@indirectbr_rewrite, %bb1),
+ i8* blockaddress(@indirectbr_rewrite, %bb2),
+ i8* blockaddress(@indirectbr_rewrite, %bb3),
+ i8* blockaddress(@indirectbr_rewrite, %bb4),
+ i8* blockaddress(@indirectbr_rewrite, %bb5),
+ i8* blockaddress(@indirectbr_rewrite, %bb6),
+ i8* blockaddress(@indirectbr_rewrite, %bb7),
+ i8* blockaddress(@indirectbr_rewrite, %bb8),
+ i8* blockaddress(@indirectbr_rewrite, %bb9)]
+
+; Check that when retpolines are enabled a function with indirectbr gets
+; rewritten to use switch, and that in turn doesn't get lowered as a jump
+; table.
+define void @indirectbr_rewrite(i64* readonly %p, i64* %sink) #0 {
+; X64-LABEL: indirectbr_rewrite:
+; X64-NOT: jmpq
+; X86-LABEL: indirectbr_rewrite:
+; X86-NOT: jmpl
+entry:
+ %i0 = load i64, i64* %p
+ %target.i0 = getelementptr [10 x i8*], [10 x i8*]* @indirectbr_rewrite.targets, i64 0, i64 %i0
+ %target0 = load i8*, i8** %target.i0
+ indirectbr i8* %target0, [label %bb1, label %bb3]
+
+bb0:
+ store volatile i64 0, i64* %sink
+ br label %latch
+
+bb1:
+ store volatile i64 1, i64* %sink
+ br label %latch
+
+bb2:
+ store volatile i64 2, i64* %sink
+ br label %latch
+
+bb3:
+ store volatile i64 3, i64* %sink
+ br label %latch
+
+bb4:
+ store volatile i64 4, i64* %sink
+ br label %latch
+
+bb5:
+ store volatile i64 5, i64* %sink
+ br label %latch
+
+bb6:
+ store volatile i64 6, i64* %sink
+ br label %latch
+
+bb7:
+ store volatile i64 7, i64* %sink
+ br label %latch
+
+bb8:
+ store volatile i64 8, i64* %sink
+ br label %latch
+
+bb9:
+ store volatile i64 9, i64* %sink
+ br label %latch
+
+latch:
+ %i.next = load i64, i64* %p
+ %target.i.next = getelementptr [10 x i8*], [10 x i8*]* @indirectbr_rewrite.targets, i64 0, i64 %i.next
+ %target.next = load i8*, i8** %target.i.next
+ ; Potentially hit a full 10 successors here so that even if we rewrite as
+ ; a switch it will try to be lowered with a jump table.
+ indirectbr i8* %target.next, [label %bb0,
+ label %bb1,
+ label %bb2,
+ label %bb3,
+ label %bb4,
+ label %bb5,
+ label %bb6,
+ label %bb7,
+ label %bb8,
+ label %bb9]
+}
+
+; Lastly check that the necessary thunks were emitted.
+;
+; X64-LABEL: .section .text.__llvm_retpoline_r11,{{.*}},__llvm_retpoline_r11,comdat
+; X64-NEXT: .hidden __llvm_retpoline_r11
+; X64-NEXT: .weak __llvm_retpoline_r11
+; X64: __llvm_retpoline_r11:
+; X64-NEXT: # {{.*}} # %entry
+; X64-NEXT: callq [[CALL_TARGET:.*]]
+; X64-NEXT: [[CAPTURE_SPEC:.*]]: # Block address taken
+; X64-NEXT: # %entry
+; X64-NEXT: # =>This Inner Loop Header: Depth=1
+; X64-NEXT: pause
+; X64-NEXT: lfence
+; X64-NEXT: jmp [[CAPTURE_SPEC]]
+; X64-NEXT: .p2align 4, 0x90
+; X64-NEXT: [[CALL_TARGET]]: # Block address taken
+; X64-NEXT: # %entry
+; X64-NEXT: movq %r11, (%rsp)
+; X64-NEXT: retq
+;
+; X86-LABEL: .section .text.__llvm_retpoline_eax,{{.*}},__llvm_retpoline_eax,comdat
+; X86-NEXT: .hidden __llvm_retpoline_eax
+; X86-NEXT: .weak __llvm_retpoline_eax
+; X86: __llvm_retpoline_eax:
+; X86-NEXT: # {{.*}} # %entry
+; X86-NEXT: calll [[CALL_TARGET:.*]]
+; X86-NEXT: [[CAPTURE_SPEC:.*]]: # Block address taken
+; X86-NEXT: # %entry
+; X86-NEXT: # =>This Inner Loop Header: Depth=1
+; X86-NEXT: pause
+; X86-NEXT: lfence
+; X86-NEXT: jmp [[CAPTURE_SPEC]]
+; X86-NEXT: .p2align 4, 0x90
+; X86-NEXT: [[CALL_TARGET]]: # Block address taken
+; X86-NEXT: # %entry
+; X86-NEXT: movl %eax, (%esp)
+; X86-NEXT: retl
+;
+; X86-LABEL: .section .text.__llvm_retpoline_ecx,{{.*}},__llvm_retpoline_ecx,comdat
+; X86-NEXT: .hidden __llvm_retpoline_ecx
+; X86-NEXT: .weak __llvm_retpoline_ecx
+; X86: __llvm_retpoline_ecx:
+; X86-NEXT: # {{.*}} # %entry
+; X86-NEXT: calll [[CALL_TARGET:.*]]
+; X86-NEXT: [[CAPTURE_SPEC:.*]]: # Block address taken
+; X86-NEXT: # %entry
+; X86-NEXT: # =>This Inner Loop Header: Depth=1
+; X86-NEXT: pause
+; X86-NEXT: lfence
+; X86-NEXT: jmp [[CAPTURE_SPEC]]
+; X86-NEXT: .p2align 4, 0x90
+; X86-NEXT: [[CALL_TARGET]]: # Block address taken
+; X86-NEXT: # %entry
+; X86-NEXT: movl %ecx, (%esp)
+; X86-NEXT: retl
+;
+; X86-LABEL: .section .text.__llvm_retpoline_edx,{{.*}},__llvm_retpoline_edx,comdat
+; X86-NEXT: .hidden __llvm_retpoline_edx
+; X86-NEXT: .weak __llvm_retpoline_edx
+; X86: __llvm_retpoline_edx:
+; X86-NEXT: # {{.*}} # %entry
+; X86-NEXT: calll [[CALL_TARGET:.*]]
+; X86-NEXT: [[CAPTURE_SPEC:.*]]: # Block address taken
+; X86-NEXT: # %entry
+; X86-NEXT: # =>This Inner Loop Header: Depth=1
+; X86-NEXT: pause
+; X86-NEXT: lfence
+; X86-NEXT: jmp [[CAPTURE_SPEC]]
+; X86-NEXT: .p2align 4, 0x90
+; X86-NEXT: [[CALL_TARGET]]: # Block address taken
+; X86-NEXT: # %entry
+; X86-NEXT: movl %edx, (%esp)
+; X86-NEXT: retl
+;
+; X86-LABEL: .section .text.__llvm_retpoline_push,{{.*}},__llvm_retpoline_push,comdat
+; X86-NEXT: .hidden __llvm_retpoline_push
+; X86-NEXT: .weak __llvm_retpoline_push
+; X86: __llvm_retpoline_push:
+; X86-NEXT: # {{.*}} # %entry
+; X86-NEXT: calll [[CALL_TARGET:.*]]
+; X86-NEXT: [[CAPTURE_SPEC:.*]]: # Block address taken
+; X86-NEXT: # %entry
+; X86-NEXT: # =>This Inner Loop Header: Depth=1
+; X86-NEXT: pause
+; X86-NEXT: lfence
+; X86-NEXT: jmp [[CAPTURE_SPEC]]
+; X86-NEXT: .p2align 4, 0x90
+; X86-NEXT: [[CALL_TARGET]]: # Block address taken
+; X86-NEXT: # %entry
+; X86-NEXT: addl $4, %esp
+; X86-NEXT: pushl 4(%esp)
+; X86-NEXT: pushl 4(%esp)
+; X86-NEXT: popl 8(%esp)
+; X86-NEXT: popl (%esp)
+; X86-NEXT: retl
+
+
+attributes #0 = { "target-features"="+retpoline" }
+attributes #1 = { nonlazybind }
diff --git a/test/Transforms/IndirectBrExpand/basic.ll b/test/Transforms/IndirectBrExpand/basic.ll
new file mode 100644
index 00000000000..d0319c6b994
--- /dev/null
+++ b/test/Transforms/IndirectBrExpand/basic.ll
@@ -0,0 +1,63 @@
+; RUN: opt < %s -indirectbr-expand -S | FileCheck %s
+;
+; REQUIRES: x86-registered-target
+
+target triple = "x86_64-unknown-linux-gnu"
+
+@test1.targets = constant [4 x i8*] [i8* blockaddress(@test1, %bb0),
+ i8* blockaddress(@test1, %bb1),
+ i8* blockaddress(@test1, %bb2),
+ i8* blockaddress(@test1, %bb3)]
+; CHECK-LABEL: @test1.targets = constant [4 x i8*]
+; CHECK: [i8* inttoptr (i64 1 to i8*),
+; CHECK: i8* inttoptr (i64 2 to i8*),
+; CHECK: i8* inttoptr (i64 3 to i8*),
+; CHECK: i8* blockaddress(@test1, %bb3)]
+
+define void @test1(i64* readonly %p, i64* %sink) #0 {
+; CHECK-LABEL: define void @test1(
+entry:
+ %i0 = load i64, i64* %p
+ %target.i0 = getelementptr [4 x i8*], [4 x i8*]* @test1.targets, i64 0, i64 %i0
+ %target0 = load i8*, i8** %target.i0
+ ; Only a subset of blocks are viable successors here.
+ indirectbr i8* %target0, [label %bb0, label %bb1]
+; CHECK-NOT: indirectbr
+; CHECK: %[[ENTRY_V:.*]] = ptrtoint i8* %{{.*}} to i64
+; CHECK-NEXT: br label %[[SWITCH_BB:.*]]
+
+bb0:
+ store volatile i64 0, i64* %sink
+ br label %latch
+
+bb1:
+ store volatile i64 1, i64* %sink
+ br label %latch
+
+bb2:
+ store volatile i64 2, i64* %sink
+ br label %latch
+
+bb3:
+ store volatile i64 3, i64* %sink
+ br label %latch
+
+latch:
+ %i.next = load i64, i64* %p
+ %target.i.next = getelementptr [4 x i8*], [4 x i8*]* @test1.targets, i64 0, i64 %i.next
+ %target.next = load i8*, i8** %target.i.next
+ ; A different subset of blocks are viable successors here.
+ indirectbr i8* %target.next, [label %bb1, label %bb2]
+; CHECK-NOT: indirectbr
+; CHECK: %[[LATCH_V:.*]] = ptrtoint i8* %{{.*}} to i64
+; CHECK-NEXT: br label %[[SWITCH_BB]]
+;
+; CHECK: [[SWITCH_BB]]:
+; CHECK-NEXT: %[[V:.*]] = phi i64 [ %[[ENTRY_V]], %entry ], [ %[[LATCH_V]], %latch ]
+; CHECK-NEXT: switch i64 %[[V]], label %bb0 [
+; CHECK-NEXT: i64 2, label %bb1
+; CHECK-NEXT: i64 3, label %bb2
+; CHECK-NEXT: ]
+}
+
+attributes #0 = { "target-features"="+retpoline" }