
This avoids the need to have special handling at every use site. Unfortunately this means we unnecessarily emit AssertZext in the DAG (where we already directly understand the range of the intrinsic), andt we regress in undefined cases as we don't fold out asserts on undef.
180 lines
8.1 KiB
LLVM
180 lines
8.1 KiB
LLVM
; RUN: opt -mtriple=amdgcn--amdhsa -data-layout=A5 -O3 -S -inline-threshold=1 < %s | FileCheck -check-prefixes=GCN,GCN-INL1,GCN-MAXBBDEF %s
|
|
; RUN: opt -mtriple=amdgcn--amdhsa -data-layout=A5 -O3 -S < %s | FileCheck -check-prefixes=GCN,GCN-INLDEF,GCN-MAXBBDEF %s
|
|
; RUN: opt -mtriple=amdgcn--amdhsa -data-layout=A5 -passes='default<O3>' -S -inline-threshold=1 < %s | FileCheck -check-prefixes=GCN,GCN-INL1,GCN-MAXBBDEF %s
|
|
; RUN: opt -mtriple=amdgcn--amdhsa -data-layout=A5 -passes='default<O3>' -S < %s | FileCheck -check-prefixes=GCN,GCN-INLDEF,GCN-MAXBBDEF %s
|
|
; RUN: opt -mtriple=amdgcn--amdhsa -data-layout=A5 -passes='default<O3>' -S -amdgpu-inline-max-bb=1 < %s | FileCheck -check-prefixes=GCN,GCN-MAXBB1 %s
|
|
|
|
define coldcc float @foo(float %x, float %y) {
|
|
entry:
|
|
%cmp = fcmp ogt float %x, 0.000000e+00
|
|
%div = fdiv float %y, %x
|
|
%mul = fmul float %x, %y
|
|
%cond = select i1 %cmp, float %div, float %mul
|
|
ret float %cond
|
|
}
|
|
|
|
define coldcc void @foo_private_ptr(ptr addrspace(5) nocapture %p) {
|
|
entry:
|
|
%tmp1 = load float, ptr addrspace(5) %p, align 4
|
|
%cmp = fcmp ogt float %tmp1, 1.000000e+00
|
|
br i1 %cmp, label %if.then, label %if.end
|
|
|
|
if.then: ; preds = %entry
|
|
%div = fdiv float 1.000000e+00, %tmp1
|
|
store float %div, ptr addrspace(5) %p, align 4
|
|
br label %if.end
|
|
|
|
if.end: ; preds = %if.then, %entry
|
|
ret void
|
|
}
|
|
|
|
define coldcc void @foo_private_ptr2(ptr addrspace(5) nocapture %p1, ptr addrspace(5) nocapture %p2) {
|
|
entry:
|
|
call void @forbid_sroa(ptr addrspace(5) %p1)
|
|
call void @forbid_sroa(ptr addrspace(5) %p2)
|
|
%tmp1 = load float, ptr addrspace(5) %p1, align 4
|
|
%cmp = fcmp ogt float %tmp1, 1.000000e+00
|
|
br i1 %cmp, label %if.then, label %if.end
|
|
|
|
if.then:
|
|
%div = fdiv float 2.000000e+00, %tmp1
|
|
store float %div, ptr addrspace(5) %p2, align 4
|
|
br label %if.end
|
|
|
|
if.end:
|
|
ret void
|
|
}
|
|
|
|
define float @sin_wrapper(float %x) {
|
|
bb:
|
|
%call = tail call float @_Z3sinf(float %x)
|
|
ret float %call
|
|
}
|
|
|
|
define void @foo_noinline(ptr addrspace(5) nocapture %p) #0 {
|
|
entry:
|
|
%tmp1 = load float, ptr addrspace(5) %p, align 4
|
|
%mul = fmul float %tmp1, 2.000000e+00
|
|
store float %mul, ptr addrspace(5) %p, align 4
|
|
ret void
|
|
}
|
|
|
|
; GCN: define amdgpu_kernel void @test_inliner(
|
|
; GCN-INL1: %c1 = tail call coldcc float @foo(
|
|
; GCN-INLDEF: %cmp.i = fcmp ogt float %{{.+}}, 0.000000e+00
|
|
; GCN-MAXBBDEF: %div.i{{[0-9]*}} = fdiv float 1.000000e+00, %c
|
|
; GCN-MAXBBDEF: %div.i{{[0-9]*}} = fdiv float 2.000000e+00, %tmp1.i
|
|
; GCN-MAXBB1: call coldcc void @foo_private_ptr
|
|
; GCN-MAXBB1: call coldcc void @foo_private_ptr2
|
|
; GCN: call void @foo_noinline(
|
|
; GCN: tail call float @_Z3sinf(
|
|
define amdgpu_kernel void @test_inliner(ptr addrspace(1) nocapture %a, i32 %n) {
|
|
entry:
|
|
%pvt_arr = alloca [64 x float], align 4, addrspace(5)
|
|
%tid = tail call i32 @llvm.amdgcn.workitem.id.x()
|
|
%arrayidx = getelementptr inbounds float, ptr addrspace(1) %a, i32 %tid
|
|
%tmp2 = load float, ptr addrspace(1) %arrayidx, align 4
|
|
%add = add i32 %tid, 1
|
|
%arrayidx2 = getelementptr inbounds float, ptr addrspace(1) %a, i32 %add
|
|
%tmp5 = load float, ptr addrspace(1) %arrayidx2, align 4
|
|
%c1 = tail call coldcc float @foo(float %tmp2, float %tmp5)
|
|
%or = or i32 %tid, %n
|
|
%arrayidx5 = getelementptr inbounds [64 x float], ptr addrspace(5) %pvt_arr, i32 0, i32 %or
|
|
store float %c1, ptr addrspace(5) %arrayidx5, align 4
|
|
%arrayidx7 = getelementptr inbounds [64 x float], ptr addrspace(5) %pvt_arr, i32 0, i32 %or
|
|
call coldcc void @foo_private_ptr(ptr addrspace(5) %arrayidx7)
|
|
%arrayidx8 = getelementptr inbounds [64 x float], ptr addrspace(5) %pvt_arr, i32 0, i32 1
|
|
%arrayidx9 = getelementptr inbounds [64 x float], ptr addrspace(5) %pvt_arr, i32 0, i32 2
|
|
call coldcc void @foo_private_ptr2(ptr addrspace(5) %arrayidx8, ptr addrspace(5) %arrayidx9)
|
|
call void @foo_noinline(ptr addrspace(5) %arrayidx7)
|
|
%and = and i32 %tid, %n
|
|
%arrayidx11 = getelementptr inbounds [64 x float], ptr addrspace(5) %pvt_arr, i32 0, i32 %and
|
|
%tmp12 = load float, ptr addrspace(5) %arrayidx11, align 4
|
|
%c2 = call float @sin_wrapper(float %tmp12)
|
|
store float %c2, ptr addrspace(5) %arrayidx7, align 4
|
|
%xor = xor i32 %tid, %n
|
|
%arrayidx16 = getelementptr inbounds [64 x float], ptr addrspace(5) %pvt_arr, i32 0, i32 %xor
|
|
%tmp16 = load float, ptr addrspace(5) %arrayidx16, align 4
|
|
store float %tmp16, ptr addrspace(1) %arrayidx, align 4
|
|
ret void
|
|
}
|
|
|
|
; GCN: define amdgpu_kernel void @test_inliner_multi_pvt_ptr(
|
|
; GCN-MAXBBDEF: %div.i{{[0-9]*}} = fdiv float 2.000000e+00, %tmp1.i
|
|
; GCN-MAXBB1: call coldcc void @foo_private_ptr2
|
|
define amdgpu_kernel void @test_inliner_multi_pvt_ptr(ptr addrspace(1) nocapture %a, i32 %n, float %v) {
|
|
entry:
|
|
%pvt_arr1 = alloca [32 x float], align 4, addrspace(5)
|
|
%pvt_arr2 = alloca [32 x float], align 4, addrspace(5)
|
|
%tid = tail call i32 @llvm.amdgcn.workitem.id.x()
|
|
%arrayidx = getelementptr inbounds float, ptr addrspace(1) %a, i32 %tid
|
|
%or = or i32 %tid, %n
|
|
%arrayidx4 = getelementptr inbounds [32 x float], ptr addrspace(5) %pvt_arr1, i32 0, i32 %or
|
|
%arrayidx5 = getelementptr inbounds [32 x float], ptr addrspace(5) %pvt_arr2, i32 0, i32 %or
|
|
store float %v, ptr addrspace(5) %arrayidx4, align 4
|
|
store float %v, ptr addrspace(5) %arrayidx5, align 4
|
|
%arrayidx8 = getelementptr inbounds [32 x float], ptr addrspace(5) %pvt_arr1, i32 0, i32 1
|
|
%arrayidx9 = getelementptr inbounds [32 x float], ptr addrspace(5) %pvt_arr2, i32 0, i32 2
|
|
call coldcc void @foo_private_ptr2(ptr addrspace(5) %arrayidx8, ptr addrspace(5) %arrayidx9)
|
|
%xor = xor i32 %tid, %n
|
|
%arrayidx15 = getelementptr inbounds [32 x float], ptr addrspace(5) %pvt_arr1, i32 0, i32 %xor
|
|
%arrayidx16 = getelementptr inbounds [32 x float], ptr addrspace(5) %pvt_arr2, i32 0, i32 %xor
|
|
%tmp15 = load float, ptr addrspace(5) %arrayidx15, align 4
|
|
%tmp16 = load float, ptr addrspace(5) %arrayidx16, align 4
|
|
%tmp17 = fadd float %tmp15, %tmp16
|
|
store float %tmp17, ptr addrspace(1) %arrayidx, align 4
|
|
ret void
|
|
}
|
|
|
|
; GCN: define amdgpu_kernel void @test_inliner_multi_pvt_ptr_cutoff(
|
|
; GCN-INL1: call coldcc void @foo_private_ptr2
|
|
; GCN-INLDEF: %div.i{{[0-9]*}} = fdiv float 2.000000e+00, %tmp1.i
|
|
define amdgpu_kernel void @test_inliner_multi_pvt_ptr_cutoff(ptr addrspace(1) nocapture %a, i32 %n, float %v) {
|
|
entry:
|
|
%pvt_arr1 = alloca [32 x float], align 4, addrspace(5)
|
|
%pvt_arr2 = alloca [33 x float], align 4, addrspace(5)
|
|
%tid = tail call i32 @llvm.amdgcn.workitem.id.x()
|
|
%arrayidx = getelementptr inbounds float, ptr addrspace(1) %a, i32 %tid
|
|
%or = or i32 %tid, %n
|
|
%arrayidx4 = getelementptr inbounds [32 x float], ptr addrspace(5) %pvt_arr1, i32 0, i32 %or
|
|
%arrayidx5 = getelementptr inbounds [33 x float], ptr addrspace(5) %pvt_arr2, i32 0, i32 %or
|
|
store float %v, ptr addrspace(5) %arrayidx4, align 4
|
|
store float %v, ptr addrspace(5) %arrayidx5, align 4
|
|
%arrayidx8 = getelementptr inbounds [32 x float], ptr addrspace(5) %pvt_arr1, i32 0, i32 1
|
|
%arrayidx9 = getelementptr inbounds [33 x float], ptr addrspace(5) %pvt_arr2, i32 0, i32 2
|
|
call coldcc void @foo_private_ptr2(ptr addrspace(5) %arrayidx8, ptr addrspace(5) %arrayidx9)
|
|
%xor = xor i32 %tid, %n
|
|
%arrayidx15 = getelementptr inbounds [32 x float], ptr addrspace(5) %pvt_arr1, i32 0, i32 %xor
|
|
%arrayidx16 = getelementptr inbounds [33 x float], ptr addrspace(5) %pvt_arr2, i32 0, i32 %xor
|
|
%tmp15 = load float, ptr addrspace(5) %arrayidx15, align 4
|
|
%tmp16 = load float, ptr addrspace(5) %arrayidx16, align 4
|
|
%tmp17 = fadd float %tmp15, %tmp16
|
|
store float %tmp17, ptr addrspace(1) %arrayidx, align 4
|
|
ret void
|
|
}
|
|
|
|
; GCN: define amdgpu_kernel void @test_inliner_maxbb_singlebb(
|
|
; GCN: tail call float @_Z3sinf
|
|
define amdgpu_kernel void @test_inliner_maxbb_singlebb(ptr addrspace(1) nocapture %a, i32 %n) {
|
|
entry:
|
|
%cmp = icmp eq i32 %n, 1
|
|
br i1 %cmp, label %bb.1, label %bb.2
|
|
br label %bb.1
|
|
|
|
bb.1:
|
|
store float 1.0, ptr poison
|
|
br label %bb.2
|
|
|
|
bb.2:
|
|
%c = call float @sin_wrapper(float 1.0)
|
|
store float %c, ptr addrspace(1) %a
|
|
ret void
|
|
}
|
|
|
|
declare i32 @llvm.amdgcn.workitem.id.x() #1
|
|
declare float @_Z3sinf(float) #1
|
|
declare void @forbid_sroa(ptr addrspace(5) nocapture %p)
|
|
|
|
attributes #0 = { noinline }
|
|
attributes #1 = { nounwind readnone }
|