
The i386 psABI specifies that `__float128` has 16 byte alignment and must be passed on the stack; however, LLVM currently stores it in a stack slot that has an offset of 4. Add a custom lowering to correct this alignment to 16-byte. i386 does not specify an `__int128`, but it seems reasonable to keep the same behavior as `__float128` so this is changed as well. There also isn't a good way to distinguish whether a set of four registers came from an integer or a float. The main test demonstrating this change is `store_perturbed` in `llvm/test/CodeGen/X86/i128-fp128-abi.ll`. Referenced ABI: https://gitlab.com/x86-psABIs/i386-ABI/-/wikis/uploads/14c05f1b1e156e0e46b61bfa7c1df1e2/intel386-psABI-2020-08-07.pdf Fixes: https://github.com/llvm/llvm-project/issues/77401
151 lines
4.1 KiB
LLVM
151 lines
4.1 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
|
; RUN: llc -mtriple=i686-linux-gnu %s -o - | FileCheck %s
|
|
|
|
define i32 @branch_eq(i64 %a, i64 %b) {
|
|
; CHECK-LABEL: branch_eq:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; CHECK-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; CHECK-NEXT: xorl {{[0-9]+}}(%esp), %ecx
|
|
; CHECK-NEXT: xorl {{[0-9]+}}(%esp), %eax
|
|
; CHECK-NEXT: orl %ecx, %eax
|
|
; CHECK-NEXT: jne .LBB0_2
|
|
; CHECK-NEXT: # %bb.1: # %bb1
|
|
; CHECK-NEXT: movl $1, %eax
|
|
; CHECK-NEXT: retl
|
|
; CHECK-NEXT: .LBB0_2: # %bb2
|
|
; CHECK-NEXT: movl $2, %eax
|
|
; CHECK-NEXT: retl
|
|
entry:
|
|
%cmp = icmp eq i64 %a, %b
|
|
br i1 %cmp, label %bb1, label %bb2
|
|
bb1:
|
|
ret i32 1
|
|
bb2:
|
|
ret i32 2
|
|
}
|
|
|
|
define i32 @branch_slt(i64 %a, i64 %b) {
|
|
; CHECK-LABEL: branch_slt:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; CHECK-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; CHECK-NEXT: cmpl {{[0-9]+}}(%esp), %eax
|
|
; CHECK-NEXT: sbbl {{[0-9]+}}(%esp), %ecx
|
|
; CHECK-NEXT: jge .LBB1_2
|
|
; CHECK-NEXT: # %bb.1: # %bb1
|
|
; CHECK-NEXT: movl $1, %eax
|
|
; CHECK-NEXT: retl
|
|
; CHECK-NEXT: .LBB1_2: # %bb2
|
|
; CHECK-NEXT: movl $2, %eax
|
|
; CHECK-NEXT: retl
|
|
entry:
|
|
%cmp = icmp slt i64 %a, %b
|
|
br i1 %cmp, label %bb1, label %bb2
|
|
bb1:
|
|
ret i32 1
|
|
bb2:
|
|
ret i32 2
|
|
}
|
|
|
|
define i32 @branch_ule(i64 %a, i64 %b) {
|
|
; CHECK-LABEL: branch_ule:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; CHECK-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; CHECK-NEXT: cmpl {{[0-9]+}}(%esp), %eax
|
|
; CHECK-NEXT: sbbl {{[0-9]+}}(%esp), %ecx
|
|
; CHECK-NEXT: jb .LBB2_2
|
|
; CHECK-NEXT: # %bb.1: # %bb1
|
|
; CHECK-NEXT: movl $1, %eax
|
|
; CHECK-NEXT: retl
|
|
; CHECK-NEXT: .LBB2_2: # %bb2
|
|
; CHECK-NEXT: movl $2, %eax
|
|
; CHECK-NEXT: retl
|
|
entry:
|
|
%cmp = icmp ule i64 %a, %b
|
|
br i1 %cmp, label %bb1, label %bb2
|
|
bb1:
|
|
ret i32 1
|
|
bb2:
|
|
ret i32 2
|
|
}
|
|
|
|
define i32 @set_gt(i64 %a, i64 %b) {
|
|
; CHECK-LABEL: set_gt:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; CHECK-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; CHECK-NEXT: cmpl {{[0-9]+}}(%esp), %eax
|
|
; CHECK-NEXT: sbbl {{[0-9]+}}(%esp), %ecx
|
|
; CHECK-NEXT: setl %al
|
|
; CHECK-NEXT: movzbl %al, %eax
|
|
; CHECK-NEXT: retl
|
|
entry:
|
|
%cmp = icmp sgt i64 %a, %b
|
|
%res = select i1 %cmp, i32 1, i32 0
|
|
ret i32 %res
|
|
}
|
|
|
|
define i32 @test_wide(i128 %a, i128 %b) {
|
|
; CHECK-LABEL: test_wide:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: pushl %esi
|
|
; CHECK-NEXT: .cfi_def_cfa_offset 8
|
|
; CHECK-NEXT: subl $8, %esp
|
|
; CHECK-NEXT: .cfi_def_cfa_offset 16
|
|
; CHECK-NEXT: .cfi_offset %esi, -8
|
|
; CHECK-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; CHECK-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; CHECK-NEXT: movl {{[0-9]+}}(%esp), %edx
|
|
; CHECK-NEXT: movl {{[0-9]+}}(%esp), %esi
|
|
; CHECK-NEXT: cmpl {{[0-9]+}}(%esp), %edx
|
|
; CHECK-NEXT: sbbl {{[0-9]+}}(%esp), %esi
|
|
; CHECK-NEXT: sbbl {{[0-9]+}}(%esp), %ecx
|
|
; CHECK-NEXT: sbbl {{[0-9]+}}(%esp), %eax
|
|
; CHECK-NEXT: jge .LBB4_3
|
|
; CHECK-NEXT: # %bb.1: # %bb1
|
|
; CHECK-NEXT: movl $1, %eax
|
|
; CHECK-NEXT: jmp .LBB4_2
|
|
; CHECK-NEXT: .LBB4_3: # %bb2
|
|
; CHECK-NEXT: movl $2, %eax
|
|
; CHECK-NEXT: .LBB4_2: # %bb1
|
|
; CHECK-NEXT: addl $8, %esp
|
|
; CHECK-NEXT: .cfi_def_cfa_offset 8
|
|
; CHECK-NEXT: popl %esi
|
|
; CHECK-NEXT: .cfi_def_cfa_offset 4
|
|
; CHECK-NEXT: retl
|
|
entry:
|
|
%cmp = icmp slt i128 %a, %b
|
|
br i1 %cmp, label %bb1, label %bb2
|
|
bb1:
|
|
ret i32 1
|
|
bb2:
|
|
ret i32 2
|
|
}
|
|
|
|
; The comparison of the low bits will be folded to a CARRY_FALSE node. Make
|
|
; sure the code can handle that.
|
|
define i32 @test_carry_false(i64 %a, i64 %b) {
|
|
; CHECK-LABEL: test_carry_false:
|
|
; CHECK: # %bb.0: # %entry
|
|
; CHECK-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; CHECK-NEXT: cmpl {{[0-9]+}}(%esp), %eax
|
|
; CHECK-NEXT: jge .LBB5_2
|
|
; CHECK-NEXT: # %bb.1: # %bb1
|
|
; CHECK-NEXT: movl $1, %eax
|
|
; CHECK-NEXT: retl
|
|
; CHECK-NEXT: .LBB5_2: # %bb2
|
|
; CHECK-NEXT: movl $2, %eax
|
|
; CHECK-NEXT: retl
|
|
entry:
|
|
%x = and i64 %a, -4294967296 ;0xffffffff00000000
|
|
%y = and i64 %b, -4294967296
|
|
%cmp = icmp slt i64 %x, %y
|
|
br i1 %cmp, label %bb1, label %bb2
|
|
bb1:
|
|
ret i32 1
|
|
bb2:
|
|
ret i32 2
|
|
}
|