
This changes the RC priorities such that AVRegClass is the least prioritized. These registers are less constrained than the VRegClass and ARegClass as they can be either agpr or vgpr. Thus, assigning them last removes unnecessary constraints from VRegClass and ARegClass assignments, and allows the RA to make smarter decisions about whether to use vgpr / agpr for AVRegClass. We only have 5 bits for RC priorities, and we still want to prioritize larger RCs over smaller ones. Since this new prioritization uses the 5th bit for AVRegClass vs ARegClass / VRegClass, we only have 4 bits to encode the size priorities. Previously, each RC with a distinct size, had a distinct priority. However, this PR groups together multiple sizes to the same priority. Currently, this will have no effect on prioritization in practice because we only have one actually defined RC per group per vector register type. For example, a register class with 15 or 16 32bit registers will have the same size priority (14). However, we only have VReg_512 (VReg_480 doesn't exist), so only one actual RC in VRegClass has this priority. Similarly, we give register class with 17-32+ 32 bit registers a size priority of 15, but we only have VReg_1024. The effect of this PR is to prioritize first the vector register type (VReg & Areg have top priority, then AVReg), with the size of the register class having second priority. Passes PSDB. --------- Co-authored-by: Matt Arsenault <Matthew.Arsenault@amd.com>
189 lines
7.8 KiB
LLVM
189 lines
7.8 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 5
|
|
; RUN: llc -mtriple=amdgcn-amd-amdhsa -mcpu=gfx90a %s -o - | FileCheck %s
|
|
|
|
declare ptr @G()
|
|
|
|
define amdgpu_kernel void @foo(ptr addrspace(5) %ptr5, ptr %p0, double %v0, <4 x i32> %vec) {
|
|
; CHECK-LABEL: foo:
|
|
; CHECK: ; %bb.0: ; %entry
|
|
; CHECK-NEXT: s_add_u32 flat_scratch_lo, s12, s17
|
|
; CHECK-NEXT: s_addc_u32 flat_scratch_hi, s13, 0
|
|
; CHECK-NEXT: v_pk_mov_b32 v[44:45], 0, 0
|
|
; CHECK-NEXT: flat_load_dword v42, v[44:45]
|
|
; CHECK-NEXT: s_mov_b64 s[38:39], s[6:7]
|
|
; CHECK-NEXT: s_mov_b64 s[48:49], s[4:5]
|
|
; CHECK-NEXT: s_load_dwordx4 s[4:7], s[8:9], 0x8
|
|
; CHECK-NEXT: s_load_dword s64, s[8:9], 0x0
|
|
; CHECK-NEXT: s_add_u32 s0, s0, s17
|
|
; CHECK-NEXT: s_addc_u32 s1, s1, 0
|
|
; CHECK-NEXT: s_mov_b64 s[34:35], s[8:9]
|
|
; CHECK-NEXT: s_waitcnt lgkmcnt(0)
|
|
; CHECK-NEXT: v_accvgpr_write_b32 a32, s6
|
|
; CHECK-NEXT: v_accvgpr_write_b32 a33, s7
|
|
; CHECK-NEXT: s_mov_b64 s[6:7], src_private_base
|
|
; CHECK-NEXT: s_cmp_lg_u32 s64, -1
|
|
; CHECK-NEXT: s_cselect_b32 s7, s7, 0
|
|
; CHECK-NEXT: s_cselect_b32 s8, s64, 0
|
|
; CHECK-NEXT: s_add_u32 s50, s34, 48
|
|
; CHECK-NEXT: s_addc_u32 s51, s35, 0
|
|
; CHECK-NEXT: v_pk_mov_b32 v[56:57], s[4:5], s[4:5] op_sel:[0,1]
|
|
; CHECK-NEXT: s_getpc_b64 s[4:5]
|
|
; CHECK-NEXT: s_add_u32 s4, s4, G@gotpcrel32@lo+4
|
|
; CHECK-NEXT: s_addc_u32 s5, s5, G@gotpcrel32@hi+12
|
|
; CHECK-NEXT: s_load_dwordx2 s[54:55], s[4:5], 0x0
|
|
; CHECK-NEXT: s_mov_b32 s6, 0
|
|
; CHECK-NEXT: v_mov_b32_e32 v47, s7
|
|
; CHECK-NEXT: s_mov_b32 s7, s6
|
|
; CHECK-NEXT: s_mov_b32 s53, s14
|
|
; CHECK-NEXT: v_mov_b32_e32 v46, s8
|
|
; CHECK-NEXT: v_pk_mov_b32 v[58:59], s[6:7], s[6:7] op_sel:[0,1]
|
|
; CHECK-NEXT: s_mov_b64 s[4:5], s[48:49]
|
|
; CHECK-NEXT: s_mov_b64 s[6:7], s[38:39]
|
|
; CHECK-NEXT: s_mov_b64 s[8:9], s[50:51]
|
|
; CHECK-NEXT: s_mov_b32 s12, s14
|
|
; CHECK-NEXT: s_mov_b32 s13, s15
|
|
; CHECK-NEXT: s_mov_b32 s14, s16
|
|
; CHECK-NEXT: v_mov_b32_e32 v31, v0
|
|
; CHECK-NEXT: s_mov_b32 s32, 0
|
|
; CHECK-NEXT: s_mov_b32 s33, s16
|
|
; CHECK-NEXT: s_mov_b32 s52, s15
|
|
; CHECK-NEXT: s_mov_b64 s[36:37], s[10:11]
|
|
; CHECK-NEXT: v_mov_b32_e32 v40, v0
|
|
; CHECK-NEXT: flat_store_dwordx2 v[56:57], v[58:59]
|
|
; CHECK-NEXT: ; kill: def $sgpr15 killed $sgpr15
|
|
; CHECK-NEXT: s_waitcnt lgkmcnt(0)
|
|
; CHECK-NEXT: s_swappc_b64 s[30:31], s[54:55]
|
|
; CHECK-NEXT: flat_load_dwordx2 v[60:61], v[56:57]
|
|
; CHECK-NEXT: v_mov_b32_e32 v62, 0
|
|
; CHECK-NEXT: v_mov_b32_e32 v63, 0x3ff00000
|
|
; CHECK-NEXT: s_mov_b64 s[4:5], s[48:49]
|
|
; CHECK-NEXT: s_mov_b64 s[6:7], s[38:39]
|
|
; CHECK-NEXT: s_mov_b64 s[8:9], s[50:51]
|
|
; CHECK-NEXT: s_mov_b64 s[10:11], s[36:37]
|
|
; CHECK-NEXT: s_mov_b32 s12, s53
|
|
; CHECK-NEXT: s_mov_b32 s13, s52
|
|
; CHECK-NEXT: s_mov_b32 s14, s33
|
|
; CHECK-NEXT: v_mov_b32_e32 v31, v40
|
|
; CHECK-NEXT: flat_store_dwordx2 v[44:45], v[62:63]
|
|
; CHECK-NEXT: flat_store_dwordx2 v[56:57], v[58:59]
|
|
; CHECK-NEXT: s_waitcnt vmcnt(0)
|
|
; CHECK-NEXT: ; kill: def $sgpr15 killed $sgpr15
|
|
; CHECK-NEXT: s_swappc_b64 s[30:31], s[54:55]
|
|
; CHECK-NEXT: flat_load_dwordx2 v[0:1], v[46:47] glc
|
|
; CHECK-NEXT: s_waitcnt vmcnt(0) lgkmcnt(0)
|
|
; CHECK-NEXT: v_mov_b32_e32 v0, s64
|
|
; CHECK-NEXT: v_cmp_lt_i32_e32 vcc, 0, v42
|
|
; CHECK-NEXT: flat_store_dwordx2 v[56:57], v[60:61]
|
|
; CHECK-NEXT: s_waitcnt vmcnt(0)
|
|
; CHECK-NEXT: flat_store_dwordx2 v[56:57], a[32:33]
|
|
; CHECK-NEXT: buffer_store_dword a33, v0, s[0:3], 0 offen offset:4
|
|
; CHECK-NEXT: buffer_store_dword v62, v0, s[0:3], 0 offen
|
|
; CHECK-NEXT: ; implicit-def: $vgpr4
|
|
; CHECK-NEXT: s_and_saveexec_b64 s[4:5], vcc
|
|
; CHECK-NEXT: s_xor_b64 s[4:5], exec, s[4:5]
|
|
; CHECK-NEXT: s_cbranch_execz .LBB0_4
|
|
; CHECK-NEXT: ; %bb.1: ; %LeafBlock5
|
|
; CHECK-NEXT: v_cmp_eq_u32_e32 vcc, 1, v42
|
|
; CHECK-NEXT: v_mov_b32_e32 v4, 0
|
|
; CHECK-NEXT: s_and_saveexec_b64 s[6:7], vcc
|
|
; CHECK-NEXT: ; %bb.2: ; %sw.bb17.i.i.i.i
|
|
; CHECK-NEXT: v_mov_b32_e32 v4, 1
|
|
; CHECK-NEXT: ; %bb.3: ; %Flow
|
|
; CHECK-NEXT: s_or_b64 exec, exec, s[6:7]
|
|
; CHECK-NEXT: .LBB0_4: ; %Flow8
|
|
; CHECK-NEXT: s_or_saveexec_b64 s[4:5], s[4:5]
|
|
; CHECK-NEXT: v_pk_mov_b32 v[0:1], v[42:43], v[42:43] op_sel:[0,1]
|
|
; CHECK-NEXT: v_pk_mov_b32 v[2:3], v[44:45], v[44:45] op_sel:[0,1]
|
|
; CHECK-NEXT: s_xor_b64 exec, exec, s[4:5]
|
|
; CHECK-NEXT: s_cbranch_execz .LBB0_8
|
|
; CHECK-NEXT: ; %bb.5: ; %LeafBlock
|
|
; CHECK-NEXT: v_cmp_eq_u32_e32 vcc, 0, v42
|
|
; CHECK-NEXT: v_pk_mov_b32 v[0:1], v[42:43], v[42:43] op_sel:[0,1]
|
|
; CHECK-NEXT: v_pk_mov_b32 v[2:3], v[44:45], v[44:45] op_sel:[0,1]
|
|
; CHECK-NEXT: s_and_saveexec_b64 s[6:7], vcc
|
|
; CHECK-NEXT: ; %bb.6: ; %sw.bb.i.i.i.i
|
|
; CHECK-NEXT: v_mov_b32_e32 v0, 0
|
|
; CHECK-NEXT: ; %bb.7: ; %Flow7
|
|
; CHECK-NEXT: s_or_b64 exec, exec, s[6:7]
|
|
; CHECK-NEXT: v_mov_b32_e32 v4, 0
|
|
; CHECK-NEXT: .LBB0_8: ; %bb.1
|
|
; CHECK-NEXT: s_or_b64 exec, exec, s[4:5]
|
|
; CHECK-NEXT: v_cmp_eq_u32_e32 vcc, 0, v4
|
|
; CHECK-NEXT: s_and_saveexec_b64 s[4:5], vcc
|
|
; CHECK-NEXT: s_cbranch_execz .LBB0_10
|
|
; CHECK-NEXT: ; %bb.9: ; %sw.bb.i.i.i.i.i
|
|
; CHECK-NEXT: s_load_dwordx4 s[8:11], s[34:35], 0x20
|
|
; CHECK-NEXT: s_waitcnt lgkmcnt(0)
|
|
; CHECK-NEXT: v_pk_mov_b32 v[0:1], s[8:9], s[8:9] op_sel:[0,1]
|
|
; CHECK-NEXT: v_pk_mov_b32 v[2:3], s[10:11], s[10:11] op_sel:[0,1]
|
|
; CHECK-NEXT: .LBB0_10: ; %bb.2
|
|
; CHECK-NEXT: s_or_b64 exec, exec, s[4:5]
|
|
; CHECK-NEXT: v_cmp_eq_u32_e32 vcc, 0, v0
|
|
; CHECK-NEXT: s_and_saveexec_b64 s[4:5], vcc
|
|
; CHECK-NEXT: s_or_b64 exec, exec, s[4:5]
|
|
; CHECK-NEXT: buffer_store_dword v42, off, s[0:3], 0
|
|
; CHECK-NEXT: s_endpgm
|
|
entry:
|
|
%load.null = load i32, ptr null, align 8
|
|
%insert = insertelement <4 x i32> zeroinitializer, i32 %load.null, i64 0
|
|
%cast = addrspacecast ptr addrspace(5) %ptr5 to ptr
|
|
store double 0.000000e+00, ptr %p0, align 8
|
|
%call = tail call ptr @G()
|
|
store double 1.000000e+00, ptr null, align 8
|
|
%load.0 = load double, ptr %p0, align 8
|
|
store volatile double 0.000000e+00, ptr %p0, align 8
|
|
%call.1 = tail call ptr @G()
|
|
%load.1 = load volatile double, ptr %cast, align 8
|
|
store volatile double %load.0, ptr %p0, align 8
|
|
store double %v0, ptr %p0, align 8
|
|
%load.2 = load double, ptr %p0, align 8
|
|
store double %load.2, ptr addrspace(5) %ptr5, align 8
|
|
store i32 0, ptr addrspace(5) %ptr5, align 4
|
|
switch i32 %load.null, label %bb.1 [
|
|
i32 0, label %sw.bb.i.i.i.i
|
|
i32 1, label %sw.bb17.i.i.i.i
|
|
]
|
|
|
|
sw.bb.i.i.i.i: ; preds = %entry
|
|
br label %bb.1
|
|
|
|
sw.bb17.i.i.i.i: ; preds = %entry
|
|
br label %bb.1
|
|
|
|
bb.1: ; preds = %sw.bb17.i.i.i.i, %sw.bb.i.i.i.i, %entry
|
|
%phi.0 = phi i32 [ 0, %entry ], [ 0, %sw.bb.i.i.i.i ], [ 1, %sw.bb17.i.i.i.i ]
|
|
%phi.1 = phi <4 x i32> [ %insert, %entry ], [ zeroinitializer, %sw.bb.i.i.i.i ], [ %insert, %sw.bb17.i.i.i.i ]
|
|
switch i32 %phi.0, label %bb.2 [
|
|
i32 0, label %sw.bb.i.i.i.i.i
|
|
]
|
|
|
|
sw.bb.i.i.i.i.i: ; preds = %bb.1
|
|
br label %bb.2
|
|
|
|
bb.2: ; preds = %sw.bb.i.i.i.i.i, %bb.1
|
|
%phi.2 = phi <4 x i32> [ %phi.1, %bb.1 ], [ %vec, %sw.bb.i.i.i.i.i ]
|
|
%extract.1 = extractelement <4 x i32> %phi.2, i64 0
|
|
switch i32 1, label %bb.3 [
|
|
i32 0, label %sw.bb.i.i5.i.i
|
|
]
|
|
|
|
sw.bb.i.i5.i.i: ; preds = %bb.2
|
|
br label %bb.3
|
|
|
|
bb.3: ; preds = %sw.bb.i.i5.i.i, %bb.2
|
|
%phi.3 = phi <4 x i32> [ zeroinitializer, %sw.bb.i.i5.i.i ], [ %insert, %bb.2 ]
|
|
switch i32 %extract.1, label %bb.4 [
|
|
i32 0, label %sw.bb7.i.i.i3.i.i
|
|
]
|
|
|
|
sw.bb7.i.i.i3.i.i: ; preds = %bb.3
|
|
%insert.0 = insertelement <4 x i32> %insert, i32 0, i64 1
|
|
br label %bb.4
|
|
|
|
bb.4: ; preds = %sw.bb7.i.i.i3.i.i, %bb.3
|
|
%phi.4 = phi <4 x i32> [ %phi.3, %bb.3 ], [ %insert.0, %sw.bb7.i.i.i3.i.i ]
|
|
%extract = extractelement <4 x i32> %phi.4, i64 0
|
|
store i32 %extract, ptr addrspace(5) null, align 4
|
|
ret void
|
|
}
|