
On ANDNOT capable targets we can always do this profitably, without ANDNOT we only attempt this if we don't introduce an additional NOT Followup to #112547
1084 lines
37 KiB
LLVM
1084 lines
37 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
|
; RUN: llc < %s -mtriple=i686-- -mattr=-bmi | FileCheck %s --check-prefixes=X86,X86-NOBMI
|
|
; RUN: llc < %s -mtriple=i686-- -mattr=+bmi | FileCheck %s --check-prefixes=X86,X86-BMI
|
|
; RUN: llc < %s -mtriple=x86_64-- -mattr=-bmi | FileCheck %s --check-prefixes=X64,X64-NOBMI
|
|
; RUN: llc < %s -mtriple=x86_64-- -mattr=+bmi | FileCheck %s --check-prefixes=X64,X64-BMI
|
|
|
|
; TODO - PR112425 - attempt to reconstruct andnot patterns through bitwise-agnostic operations
|
|
|
|
declare void @use_i64(i64)
|
|
declare void @use_i32(i32)
|
|
|
|
;
|
|
; Fold (and X, (rotl (not Y), Z))) -> (and X, (not (rotl Y, Z)))
|
|
;
|
|
|
|
define i64 @andnot_rotl_i64(i64 %a0, i64 %a1, i64 %a2) nounwind {
|
|
; X86-NOBMI-LABEL: andnot_rotl_i64:
|
|
; X86-NOBMI: # %bb.0:
|
|
; X86-NOBMI-NEXT: pushl %esi
|
|
; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %esi
|
|
; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NOBMI-NEXT: testb $32, %cl
|
|
; X86-NOBMI-NEXT: jne .LBB0_1
|
|
; X86-NOBMI-NEXT: # %bb.2:
|
|
; X86-NOBMI-NEXT: movl %eax, %edx
|
|
; X86-NOBMI-NEXT: jmp .LBB0_3
|
|
; X86-NOBMI-NEXT: .LBB0_1:
|
|
; X86-NOBMI-NEXT: movl %esi, %edx
|
|
; X86-NOBMI-NEXT: movl %eax, %esi
|
|
; X86-NOBMI-NEXT: .LBB0_3:
|
|
; X86-NOBMI-NEXT: movl %esi, %eax
|
|
; X86-NOBMI-NEXT: shldl %cl, %edx, %eax
|
|
; X86-NOBMI-NEXT: notl %eax
|
|
; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %eax
|
|
; X86-NOBMI-NEXT: # kill: def $cl killed $cl killed $ecx
|
|
; X86-NOBMI-NEXT: shldl %cl, %esi, %edx
|
|
; X86-NOBMI-NEXT: notl %edx
|
|
; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %edx
|
|
; X86-NOBMI-NEXT: popl %esi
|
|
; X86-NOBMI-NEXT: retl
|
|
;
|
|
; X86-BMI-LABEL: andnot_rotl_i64:
|
|
; X86-BMI: # %bb.0:
|
|
; X86-BMI-NEXT: pushl %esi
|
|
; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %edx
|
|
; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-BMI-NEXT: testb $32, %cl
|
|
; X86-BMI-NEXT: jne .LBB0_1
|
|
; X86-BMI-NEXT: # %bb.2:
|
|
; X86-BMI-NEXT: movl %eax, %esi
|
|
; X86-BMI-NEXT: jmp .LBB0_3
|
|
; X86-BMI-NEXT: .LBB0_1:
|
|
; X86-BMI-NEXT: movl %edx, %esi
|
|
; X86-BMI-NEXT: movl %eax, %edx
|
|
; X86-BMI-NEXT: .LBB0_3:
|
|
; X86-BMI-NEXT: movl %edx, %eax
|
|
; X86-BMI-NEXT: shldl %cl, %esi, %eax
|
|
; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax
|
|
; X86-BMI-NEXT: # kill: def $cl killed $cl killed $ecx
|
|
; X86-BMI-NEXT: shldl %cl, %edx, %esi
|
|
; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %esi, %edx
|
|
; X86-BMI-NEXT: popl %esi
|
|
; X86-BMI-NEXT: retl
|
|
;
|
|
; X64-NOBMI-LABEL: andnot_rotl_i64:
|
|
; X64-NOBMI: # %bb.0:
|
|
; X64-NOBMI-NEXT: movq %rdx, %rcx
|
|
; X64-NOBMI-NEXT: movq %rsi, %rax
|
|
; X64-NOBMI-NEXT: # kill: def $cl killed $cl killed $rcx
|
|
; X64-NOBMI-NEXT: rolq %cl, %rax
|
|
; X64-NOBMI-NEXT: notq %rax
|
|
; X64-NOBMI-NEXT: andq %rdi, %rax
|
|
; X64-NOBMI-NEXT: retq
|
|
;
|
|
; X64-BMI-LABEL: andnot_rotl_i64:
|
|
; X64-BMI: # %bb.0:
|
|
; X64-BMI-NEXT: movq %rdx, %rcx
|
|
; X64-BMI-NEXT: # kill: def $cl killed $cl killed $rcx
|
|
; X64-BMI-NEXT: rolq %cl, %rsi
|
|
; X64-BMI-NEXT: andnq %rdi, %rsi, %rax
|
|
; X64-BMI-NEXT: retq
|
|
%not = xor i64 %a1, -1
|
|
%rot = tail call i64 @llvm.fshl.i64(i64 %not, i64 %not, i64 %a2)
|
|
%and = and i64 %rot, %a0
|
|
ret i64 %and
|
|
}
|
|
|
|
define i32 @andnot_rotl_i32(i32 %a0, i32 %a1, i32 %a2) nounwind {
|
|
; X86-NOBMI-LABEL: andnot_rotl_i32:
|
|
; X86-NOBMI: # %bb.0:
|
|
; X86-NOBMI-NEXT: movzbl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NOBMI-NEXT: roll %cl, %eax
|
|
; X86-NOBMI-NEXT: notl %eax
|
|
; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %eax
|
|
; X86-NOBMI-NEXT: retl
|
|
;
|
|
; X86-BMI-LABEL: andnot_rotl_i32:
|
|
; X86-BMI: # %bb.0:
|
|
; X86-BMI-NEXT: movzbl {{[0-9]+}}(%esp), %ecx
|
|
; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-BMI-NEXT: roll %cl, %eax
|
|
; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax
|
|
; X86-BMI-NEXT: retl
|
|
;
|
|
; X64-NOBMI-LABEL: andnot_rotl_i32:
|
|
; X64-NOBMI: # %bb.0:
|
|
; X64-NOBMI-NEXT: movl %edx, %ecx
|
|
; X64-NOBMI-NEXT: movl %esi, %eax
|
|
; X64-NOBMI-NEXT: # kill: def $cl killed $cl killed $ecx
|
|
; X64-NOBMI-NEXT: roll %cl, %eax
|
|
; X64-NOBMI-NEXT: notl %eax
|
|
; X64-NOBMI-NEXT: andl %edi, %eax
|
|
; X64-NOBMI-NEXT: retq
|
|
;
|
|
; X64-BMI-LABEL: andnot_rotl_i32:
|
|
; X64-BMI: # %bb.0:
|
|
; X64-BMI-NEXT: movl %edx, %ecx
|
|
; X64-BMI-NEXT: # kill: def $cl killed $cl killed $ecx
|
|
; X64-BMI-NEXT: roll %cl, %esi
|
|
; X64-BMI-NEXT: andnl %edi, %esi, %eax
|
|
; X64-BMI-NEXT: retq
|
|
%not = xor i32 %a1, -1
|
|
%rot = tail call i32 @llvm.fshl.i32(i32 %not, i32 %not, i32 %a2)
|
|
%and = and i32 %rot, %a0
|
|
ret i32 %and
|
|
}
|
|
|
|
define i16 @andnot_rotl_i16(i16 %a0, i16 %a1, i16 %a2) nounwind {
|
|
; X86-LABEL: andnot_rotl_i16:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movzbl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: movzwl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: rolw %cl, %ax
|
|
; X86-NEXT: notl %eax
|
|
; X86-NEXT: andw {{[0-9]+}}(%esp), %ax
|
|
; X86-NEXT: # kill: def $ax killed $ax killed $eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-NOBMI-LABEL: andnot_rotl_i16:
|
|
; X64-NOBMI: # %bb.0:
|
|
; X64-NOBMI-NEXT: movl %edx, %ecx
|
|
; X64-NOBMI-NEXT: movl %esi, %eax
|
|
; X64-NOBMI-NEXT: # kill: def $cl killed $cl killed $ecx
|
|
; X64-NOBMI-NEXT: rolw %cl, %ax
|
|
; X64-NOBMI-NEXT: notl %eax
|
|
; X64-NOBMI-NEXT: andl %edi, %eax
|
|
; X64-NOBMI-NEXT: # kill: def $ax killed $ax killed $eax
|
|
; X64-NOBMI-NEXT: retq
|
|
;
|
|
; X64-BMI-LABEL: andnot_rotl_i16:
|
|
; X64-BMI: # %bb.0:
|
|
; X64-BMI-NEXT: movl %edx, %ecx
|
|
; X64-BMI-NEXT: # kill: def $cl killed $cl killed $ecx
|
|
; X64-BMI-NEXT: rolw %cl, %si
|
|
; X64-BMI-NEXT: andnl %edi, %esi, %eax
|
|
; X64-BMI-NEXT: # kill: def $ax killed $ax killed $eax
|
|
; X64-BMI-NEXT: retq
|
|
%not = xor i16 %a1, -1
|
|
%rot = tail call i16 @llvm.fshl.i16(i16 %not, i16 %not, i16 %a2)
|
|
%and = and i16 %rot, %a0
|
|
ret i16 %and
|
|
}
|
|
|
|
define i8 @andnot_rotl_i8(i8 %a0, i8 %a1, i8 %a2) nounwind {
|
|
; X86-LABEL: andnot_rotl_i8:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movzbl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: movzbl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: rolb %cl, %al
|
|
; X86-NEXT: notb %al
|
|
; X86-NEXT: andb {{[0-9]+}}(%esp), %al
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: andnot_rotl_i8:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: movl %edx, %ecx
|
|
; X64-NEXT: movl %esi, %eax
|
|
; X64-NEXT: # kill: def $cl killed $cl killed $ecx
|
|
; X64-NEXT: rolb %cl, %al
|
|
; X64-NEXT: notb %al
|
|
; X64-NEXT: andb %dil, %al
|
|
; X64-NEXT: # kill: def $al killed $al killed $eax
|
|
; X64-NEXT: retq
|
|
%not = xor i8 %a1, -1
|
|
%rot = tail call i8 @llvm.fshl.i8(i8 %not, i8 %not, i8 %a2)
|
|
%and = and i8 %rot, %a0
|
|
ret i8 %and
|
|
}
|
|
|
|
define i64 @andnot_rotl_i64_multiuse_rot(i64 %a0, i64 %a1, i64 %a2) nounwind {
|
|
; X86-LABEL: andnot_rotl_i64_multiuse_rot:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: pushl %ebx
|
|
; X86-NEXT: pushl %edi
|
|
; X86-NEXT: pushl %esi
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %esi
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
|
|
; X86-NEXT: notl %edx
|
|
; X86-NEXT: notl %esi
|
|
; X86-NEXT: testb $32, %cl
|
|
; X86-NEXT: jne .LBB4_1
|
|
; X86-NEXT: # %bb.2:
|
|
; X86-NEXT: movl %esi, %eax
|
|
; X86-NEXT: jmp .LBB4_3
|
|
; X86-NEXT: .LBB4_1:
|
|
; X86-NEXT: movl %edx, %eax
|
|
; X86-NEXT: movl %esi, %edx
|
|
; X86-NEXT: .LBB4_3:
|
|
; X86-NEXT: movl %edx, %ebx
|
|
; X86-NEXT: shldl %cl, %eax, %ebx
|
|
; X86-NEXT: # kill: def $cl killed $cl killed $ecx
|
|
; X86-NEXT: shldl %cl, %edx, %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %esi
|
|
; X86-NEXT: andl %eax, %esi
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %edi
|
|
; X86-NEXT: andl %ebx, %edi
|
|
; X86-NEXT: pushl %ebx
|
|
; X86-NEXT: pushl %eax
|
|
; X86-NEXT: calll use_i64@PLT
|
|
; X86-NEXT: addl $8, %esp
|
|
; X86-NEXT: movl %esi, %eax
|
|
; X86-NEXT: movl %edi, %edx
|
|
; X86-NEXT: popl %esi
|
|
; X86-NEXT: popl %edi
|
|
; X86-NEXT: popl %ebx
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: andnot_rotl_i64_multiuse_rot:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: pushq %rbx
|
|
; X64-NEXT: movq %rdx, %rcx
|
|
; X64-NEXT: movq %rdi, %rbx
|
|
; X64-NEXT: notq %rsi
|
|
; X64-NEXT: # kill: def $cl killed $cl killed $rcx
|
|
; X64-NEXT: rolq %cl, %rsi
|
|
; X64-NEXT: andq %rsi, %rbx
|
|
; X64-NEXT: movq %rsi, %rdi
|
|
; X64-NEXT: callq use_i64@PLT
|
|
; X64-NEXT: movq %rbx, %rax
|
|
; X64-NEXT: popq %rbx
|
|
; X64-NEXT: retq
|
|
%not = xor i64 %a1, -1
|
|
%rot = tail call i64 @llvm.fshl.i64(i64 %not, i64 %not, i64 %a2)
|
|
%and = and i64 %rot, %a0
|
|
call void @use_i64(i64 %rot)
|
|
ret i64 %and
|
|
}
|
|
|
|
;
|
|
; Fold (and X, (rotr (not Y), Z))) -> (and X, (not (rotr Y, Z)))
|
|
;
|
|
|
|
define i64 @andnot_rotr_i64(i64 %a0, i64 %a1, i64 %a2) nounwind {
|
|
; X86-NOBMI-LABEL: andnot_rotr_i64:
|
|
; X86-NOBMI: # %bb.0:
|
|
; X86-NOBMI-NEXT: pushl %esi
|
|
; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %esi
|
|
; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NOBMI-NEXT: testb $32, %cl
|
|
; X86-NOBMI-NEXT: je .LBB5_1
|
|
; X86-NOBMI-NEXT: # %bb.2:
|
|
; X86-NOBMI-NEXT: movl %eax, %edx
|
|
; X86-NOBMI-NEXT: jmp .LBB5_3
|
|
; X86-NOBMI-NEXT: .LBB5_1:
|
|
; X86-NOBMI-NEXT: movl %esi, %edx
|
|
; X86-NOBMI-NEXT: movl %eax, %esi
|
|
; X86-NOBMI-NEXT: .LBB5_3:
|
|
; X86-NOBMI-NEXT: movl %esi, %eax
|
|
; X86-NOBMI-NEXT: shrdl %cl, %edx, %eax
|
|
; X86-NOBMI-NEXT: notl %eax
|
|
; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %eax
|
|
; X86-NOBMI-NEXT: # kill: def $cl killed $cl killed $ecx
|
|
; X86-NOBMI-NEXT: shrdl %cl, %esi, %edx
|
|
; X86-NOBMI-NEXT: notl %edx
|
|
; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %edx
|
|
; X86-NOBMI-NEXT: popl %esi
|
|
; X86-NOBMI-NEXT: retl
|
|
;
|
|
; X86-BMI-LABEL: andnot_rotr_i64:
|
|
; X86-BMI: # %bb.0:
|
|
; X86-BMI-NEXT: pushl %esi
|
|
; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %edx
|
|
; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-BMI-NEXT: testb $32, %cl
|
|
; X86-BMI-NEXT: je .LBB5_1
|
|
; X86-BMI-NEXT: # %bb.2:
|
|
; X86-BMI-NEXT: movl %eax, %esi
|
|
; X86-BMI-NEXT: jmp .LBB5_3
|
|
; X86-BMI-NEXT: .LBB5_1:
|
|
; X86-BMI-NEXT: movl %edx, %esi
|
|
; X86-BMI-NEXT: movl %eax, %edx
|
|
; X86-BMI-NEXT: .LBB5_3:
|
|
; X86-BMI-NEXT: movl %edx, %eax
|
|
; X86-BMI-NEXT: shrdl %cl, %esi, %eax
|
|
; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax
|
|
; X86-BMI-NEXT: # kill: def $cl killed $cl killed $ecx
|
|
; X86-BMI-NEXT: shrdl %cl, %edx, %esi
|
|
; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %esi, %edx
|
|
; X86-BMI-NEXT: popl %esi
|
|
; X86-BMI-NEXT: retl
|
|
;
|
|
; X64-NOBMI-LABEL: andnot_rotr_i64:
|
|
; X64-NOBMI: # %bb.0:
|
|
; X64-NOBMI-NEXT: movq %rdx, %rcx
|
|
; X64-NOBMI-NEXT: movq %rsi, %rax
|
|
; X64-NOBMI-NEXT: # kill: def $cl killed $cl killed $rcx
|
|
; X64-NOBMI-NEXT: rorq %cl, %rax
|
|
; X64-NOBMI-NEXT: notq %rax
|
|
; X64-NOBMI-NEXT: andq %rdi, %rax
|
|
; X64-NOBMI-NEXT: retq
|
|
;
|
|
; X64-BMI-LABEL: andnot_rotr_i64:
|
|
; X64-BMI: # %bb.0:
|
|
; X64-BMI-NEXT: movq %rdx, %rcx
|
|
; X64-BMI-NEXT: # kill: def $cl killed $cl killed $rcx
|
|
; X64-BMI-NEXT: rorq %cl, %rsi
|
|
; X64-BMI-NEXT: andnq %rdi, %rsi, %rax
|
|
; X64-BMI-NEXT: retq
|
|
%not = xor i64 %a1, -1
|
|
%rot = tail call i64 @llvm.fshr.i64(i64 %not, i64 %not, i64 %a2)
|
|
%and = and i64 %rot, %a0
|
|
ret i64 %and
|
|
}
|
|
|
|
define i32 @andnot_rotr_i32(i32 %a0, i32 %a1, i32 %a2) nounwind {
|
|
; X86-NOBMI-LABEL: andnot_rotr_i32:
|
|
; X86-NOBMI: # %bb.0:
|
|
; X86-NOBMI-NEXT: movzbl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NOBMI-NEXT: rorl %cl, %eax
|
|
; X86-NOBMI-NEXT: notl %eax
|
|
; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %eax
|
|
; X86-NOBMI-NEXT: retl
|
|
;
|
|
; X86-BMI-LABEL: andnot_rotr_i32:
|
|
; X86-BMI: # %bb.0:
|
|
; X86-BMI-NEXT: movzbl {{[0-9]+}}(%esp), %ecx
|
|
; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-BMI-NEXT: rorl %cl, %eax
|
|
; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax
|
|
; X86-BMI-NEXT: retl
|
|
;
|
|
; X64-NOBMI-LABEL: andnot_rotr_i32:
|
|
; X64-NOBMI: # %bb.0:
|
|
; X64-NOBMI-NEXT: movl %edx, %ecx
|
|
; X64-NOBMI-NEXT: movl %esi, %eax
|
|
; X64-NOBMI-NEXT: # kill: def $cl killed $cl killed $ecx
|
|
; X64-NOBMI-NEXT: rorl %cl, %eax
|
|
; X64-NOBMI-NEXT: notl %eax
|
|
; X64-NOBMI-NEXT: andl %edi, %eax
|
|
; X64-NOBMI-NEXT: retq
|
|
;
|
|
; X64-BMI-LABEL: andnot_rotr_i32:
|
|
; X64-BMI: # %bb.0:
|
|
; X64-BMI-NEXT: movl %edx, %ecx
|
|
; X64-BMI-NEXT: # kill: def $cl killed $cl killed $ecx
|
|
; X64-BMI-NEXT: rorl %cl, %esi
|
|
; X64-BMI-NEXT: andnl %edi, %esi, %eax
|
|
; X64-BMI-NEXT: retq
|
|
%not = xor i32 %a1, -1
|
|
%rot = tail call i32 @llvm.fshr.i32(i32 %not, i32 %not, i32 %a2)
|
|
%and = and i32 %rot, %a0
|
|
ret i32 %and
|
|
}
|
|
|
|
define i16 @andnot_rotr_i16(i16 %a0, i16 %a1, i16 %a2) nounwind {
|
|
; X86-LABEL: andnot_rotr_i16:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movzbl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: movzwl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: rorw %cl, %ax
|
|
; X86-NEXT: notl %eax
|
|
; X86-NEXT: andw {{[0-9]+}}(%esp), %ax
|
|
; X86-NEXT: # kill: def $ax killed $ax killed $eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-NOBMI-LABEL: andnot_rotr_i16:
|
|
; X64-NOBMI: # %bb.0:
|
|
; X64-NOBMI-NEXT: movl %edx, %ecx
|
|
; X64-NOBMI-NEXT: movl %esi, %eax
|
|
; X64-NOBMI-NEXT: # kill: def $cl killed $cl killed $ecx
|
|
; X64-NOBMI-NEXT: rorw %cl, %ax
|
|
; X64-NOBMI-NEXT: notl %eax
|
|
; X64-NOBMI-NEXT: andl %edi, %eax
|
|
; X64-NOBMI-NEXT: # kill: def $ax killed $ax killed $eax
|
|
; X64-NOBMI-NEXT: retq
|
|
;
|
|
; X64-BMI-LABEL: andnot_rotr_i16:
|
|
; X64-BMI: # %bb.0:
|
|
; X64-BMI-NEXT: movl %edx, %ecx
|
|
; X64-BMI-NEXT: # kill: def $cl killed $cl killed $ecx
|
|
; X64-BMI-NEXT: rorw %cl, %si
|
|
; X64-BMI-NEXT: andnl %edi, %esi, %eax
|
|
; X64-BMI-NEXT: # kill: def $ax killed $ax killed $eax
|
|
; X64-BMI-NEXT: retq
|
|
%not = xor i16 %a1, -1
|
|
%rot = tail call i16 @llvm.fshr.i16(i16 %not, i16 %not, i16 %a2)
|
|
%and = and i16 %rot, %a0
|
|
ret i16 %and
|
|
}
|
|
|
|
define i8 @andnot_rotr_i8(i8 %a0, i8 %a1, i8 %a2) nounwind {
|
|
; X86-LABEL: andnot_rotr_i8:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movzbl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: movzbl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: rorb %cl, %al
|
|
; X86-NEXT: notb %al
|
|
; X86-NEXT: andb {{[0-9]+}}(%esp), %al
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: andnot_rotr_i8:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: movl %edx, %ecx
|
|
; X64-NEXT: movl %esi, %eax
|
|
; X64-NEXT: # kill: def $cl killed $cl killed $ecx
|
|
; X64-NEXT: rorb %cl, %al
|
|
; X64-NEXT: notb %al
|
|
; X64-NEXT: andb %dil, %al
|
|
; X64-NEXT: # kill: def $al killed $al killed $eax
|
|
; X64-NEXT: retq
|
|
%not = xor i8 %a1, -1
|
|
%rot = tail call i8 @llvm.fshr.i8(i8 %not, i8 %not, i8 %a2)
|
|
%and = and i8 %rot, %a0
|
|
ret i8 %and
|
|
}
|
|
|
|
define i32 @andnot_rotr_i32_multiuse_not(i32 %a0, i32 %a1, i32 %a2) nounwind {
|
|
; X86-NOBMI-LABEL: andnot_rotr_i32_multiuse_not:
|
|
; X86-NOBMI: # %bb.0:
|
|
; X86-NOBMI-NEXT: pushl %esi
|
|
; X86-NOBMI-NEXT: movzbl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NOBMI-NEXT: notl %eax
|
|
; X86-NOBMI-NEXT: movl %eax, %esi
|
|
; X86-NOBMI-NEXT: rorl %cl, %esi
|
|
; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %esi
|
|
; X86-NOBMI-NEXT: pushl %eax
|
|
; X86-NOBMI-NEXT: calll use_i32@PLT
|
|
; X86-NOBMI-NEXT: addl $4, %esp
|
|
; X86-NOBMI-NEXT: movl %esi, %eax
|
|
; X86-NOBMI-NEXT: popl %esi
|
|
; X86-NOBMI-NEXT: retl
|
|
;
|
|
; X86-BMI-LABEL: andnot_rotr_i32_multiuse_not:
|
|
; X86-BMI: # %bb.0:
|
|
; X86-BMI-NEXT: pushl %esi
|
|
; X86-BMI-NEXT: movzbl {{[0-9]+}}(%esp), %ecx
|
|
; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-BMI-NEXT: movl %eax, %edx
|
|
; X86-BMI-NEXT: notl %edx
|
|
; X86-BMI-NEXT: rorl %cl, %eax
|
|
; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %eax, %esi
|
|
; X86-BMI-NEXT: pushl %edx
|
|
; X86-BMI-NEXT: calll use_i32@PLT
|
|
; X86-BMI-NEXT: addl $4, %esp
|
|
; X86-BMI-NEXT: movl %esi, %eax
|
|
; X86-BMI-NEXT: popl %esi
|
|
; X86-BMI-NEXT: retl
|
|
;
|
|
; X64-NOBMI-LABEL: andnot_rotr_i32_multiuse_not:
|
|
; X64-NOBMI: # %bb.0:
|
|
; X64-NOBMI-NEXT: pushq %rbx
|
|
; X64-NOBMI-NEXT: movl %edx, %ecx
|
|
; X64-NOBMI-NEXT: notl %esi
|
|
; X64-NOBMI-NEXT: movl %esi, %ebx
|
|
; X64-NOBMI-NEXT: # kill: def $cl killed $cl killed $ecx
|
|
; X64-NOBMI-NEXT: rorl %cl, %ebx
|
|
; X64-NOBMI-NEXT: andl %edi, %ebx
|
|
; X64-NOBMI-NEXT: movl %esi, %edi
|
|
; X64-NOBMI-NEXT: callq use_i32@PLT
|
|
; X64-NOBMI-NEXT: movl %ebx, %eax
|
|
; X64-NOBMI-NEXT: popq %rbx
|
|
; X64-NOBMI-NEXT: retq
|
|
;
|
|
; X64-BMI-LABEL: andnot_rotr_i32_multiuse_not:
|
|
; X64-BMI: # %bb.0:
|
|
; X64-BMI-NEXT: pushq %rbx
|
|
; X64-BMI-NEXT: movl %edx, %ecx
|
|
; X64-BMI-NEXT: movl %esi, %eax
|
|
; X64-BMI-NEXT: notl %eax
|
|
; X64-BMI-NEXT: # kill: def $cl killed $cl killed $ecx
|
|
; X64-BMI-NEXT: rorl %cl, %esi
|
|
; X64-BMI-NEXT: andnl %edi, %esi, %ebx
|
|
; X64-BMI-NEXT: movl %eax, %edi
|
|
; X64-BMI-NEXT: callq use_i32@PLT
|
|
; X64-BMI-NEXT: movl %ebx, %eax
|
|
; X64-BMI-NEXT: popq %rbx
|
|
; X64-BMI-NEXT: retq
|
|
%not = xor i32 %a1, -1
|
|
%rot = tail call i32 @llvm.fshr.i32(i32 %not, i32 %not, i32 %a2)
|
|
%and = and i32 %rot, %a0
|
|
call void @use_i32(i32 %not)
|
|
ret i32 %and
|
|
}
|
|
|
|
;
|
|
; Fold (and X, (bswap (not Y)))) -> (and X, (not (bswap Y)))
|
|
;
|
|
|
|
define i64 @andnot_bswap_i64(i64 %a0, i64 %a1) nounwind {
|
|
; X86-NOBMI-LABEL: andnot_bswap_i64:
|
|
; X86-NOBMI: # %bb.0:
|
|
; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %edx
|
|
; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NOBMI-NEXT: bswapl %eax
|
|
; X86-NOBMI-NEXT: notl %eax
|
|
; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %eax
|
|
; X86-NOBMI-NEXT: bswapl %edx
|
|
; X86-NOBMI-NEXT: notl %edx
|
|
; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %edx
|
|
; X86-NOBMI-NEXT: retl
|
|
;
|
|
; X86-BMI-LABEL: andnot_bswap_i64:
|
|
; X86-BMI: # %bb.0:
|
|
; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-BMI-NEXT: bswapl %eax
|
|
; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax
|
|
; X86-BMI-NEXT: bswapl %ecx
|
|
; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %ecx, %edx
|
|
; X86-BMI-NEXT: retl
|
|
;
|
|
; X64-NOBMI-LABEL: andnot_bswap_i64:
|
|
; X64-NOBMI: # %bb.0:
|
|
; X64-NOBMI-NEXT: movq %rsi, %rax
|
|
; X64-NOBMI-NEXT: bswapq %rax
|
|
; X64-NOBMI-NEXT: notq %rax
|
|
; X64-NOBMI-NEXT: andq %rdi, %rax
|
|
; X64-NOBMI-NEXT: retq
|
|
;
|
|
; X64-BMI-LABEL: andnot_bswap_i64:
|
|
; X64-BMI: # %bb.0:
|
|
; X64-BMI-NEXT: bswapq %rsi
|
|
; X64-BMI-NEXT: andnq %rdi, %rsi, %rax
|
|
; X64-BMI-NEXT: retq
|
|
%not = xor i64 %a1, -1
|
|
%bswap = tail call i64 @llvm.bswap.i64(i64 %not)
|
|
%and = and i64 %bswap, %a0
|
|
ret i64 %and
|
|
}
|
|
|
|
define i32 @andnot_bswap_i32(i32 %a0, i32 %a1) nounwind {
|
|
; X86-NOBMI-LABEL: andnot_bswap_i32:
|
|
; X86-NOBMI: # %bb.0:
|
|
; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NOBMI-NEXT: bswapl %eax
|
|
; X86-NOBMI-NEXT: notl %eax
|
|
; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %eax
|
|
; X86-NOBMI-NEXT: retl
|
|
;
|
|
; X86-BMI-LABEL: andnot_bswap_i32:
|
|
; X86-BMI: # %bb.0:
|
|
; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-BMI-NEXT: bswapl %eax
|
|
; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax
|
|
; X86-BMI-NEXT: retl
|
|
;
|
|
; X64-NOBMI-LABEL: andnot_bswap_i32:
|
|
; X64-NOBMI: # %bb.0:
|
|
; X64-NOBMI-NEXT: movl %esi, %eax
|
|
; X64-NOBMI-NEXT: bswapl %eax
|
|
; X64-NOBMI-NEXT: notl %eax
|
|
; X64-NOBMI-NEXT: andl %edi, %eax
|
|
; X64-NOBMI-NEXT: retq
|
|
;
|
|
; X64-BMI-LABEL: andnot_bswap_i32:
|
|
; X64-BMI: # %bb.0:
|
|
; X64-BMI-NEXT: bswapl %esi
|
|
; X64-BMI-NEXT: andnl %edi, %esi, %eax
|
|
; X64-BMI-NEXT: retq
|
|
%not = xor i32 %a1, -1
|
|
%bswap = tail call i32 @llvm.bswap.i32(i32 %not)
|
|
%and = and i32 %bswap, %a0
|
|
ret i32 %and
|
|
}
|
|
|
|
define i16 @andnot_bswap_i16(i16 %a0, i16 %a1) nounwind {
|
|
; X86-LABEL: andnot_bswap_i16:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movzwl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: rolw $8, %ax
|
|
; X86-NEXT: notl %eax
|
|
; X86-NEXT: andw {{[0-9]+}}(%esp), %ax
|
|
; X86-NEXT: # kill: def $ax killed $ax killed $eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-NOBMI-LABEL: andnot_bswap_i16:
|
|
; X64-NOBMI: # %bb.0:
|
|
; X64-NOBMI-NEXT: movl %esi, %eax
|
|
; X64-NOBMI-NEXT: rolw $8, %ax
|
|
; X64-NOBMI-NEXT: notl %eax
|
|
; X64-NOBMI-NEXT: andl %edi, %eax
|
|
; X64-NOBMI-NEXT: # kill: def $ax killed $ax killed $eax
|
|
; X64-NOBMI-NEXT: retq
|
|
;
|
|
; X64-BMI-LABEL: andnot_bswap_i16:
|
|
; X64-BMI: # %bb.0:
|
|
; X64-BMI-NEXT: rolw $8, %si
|
|
; X64-BMI-NEXT: andnl %edi, %esi, %eax
|
|
; X64-BMI-NEXT: # kill: def $ax killed $ax killed $eax
|
|
; X64-BMI-NEXT: retq
|
|
%not = xor i16 %a1, -1
|
|
%bswap = tail call i16 @llvm.bswap.i16(i16 %not)
|
|
%and = and i16 %bswap, %a0
|
|
ret i16 %and
|
|
}
|
|
|
|
define i32 @andnot_bswap_i32_multiuse_bswap(i32 %a0, i32 %a1) nounwind {
|
|
; X86-LABEL: andnot_bswap_i32_multiuse_bswap:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: pushl %esi
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: notl %eax
|
|
; X86-NEXT: bswapl %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %esi
|
|
; X86-NEXT: andl %eax, %esi
|
|
; X86-NEXT: pushl %eax
|
|
; X86-NEXT: calll use_i32@PLT
|
|
; X86-NEXT: addl $4, %esp
|
|
; X86-NEXT: movl %esi, %eax
|
|
; X86-NEXT: popl %esi
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: andnot_bswap_i32_multiuse_bswap:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: pushq %rbx
|
|
; X64-NEXT: movl %edi, %ebx
|
|
; X64-NEXT: notl %esi
|
|
; X64-NEXT: bswapl %esi
|
|
; X64-NEXT: andl %esi, %ebx
|
|
; X64-NEXT: movl %esi, %edi
|
|
; X64-NEXT: callq use_i32@PLT
|
|
; X64-NEXT: movl %ebx, %eax
|
|
; X64-NEXT: popq %rbx
|
|
; X64-NEXT: retq
|
|
%not = xor i32 %a1, -1
|
|
%bswap = tail call i32 @llvm.bswap.i32(i32 %not)
|
|
%and = and i32 %bswap, %a0
|
|
call void @use_i32(i32 %bswap)
|
|
ret i32 %and
|
|
}
|
|
|
|
define i32 @andnot_bswap_i32_multiuse_not(i32 %a0, i32 %a1) nounwind {
|
|
; X86-NOBMI-LABEL: andnot_bswap_i32_multiuse_not:
|
|
; X86-NOBMI: # %bb.0:
|
|
; X86-NOBMI-NEXT: pushl %esi
|
|
; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NOBMI-NEXT: notl %eax
|
|
; X86-NOBMI-NEXT: movl %eax, %esi
|
|
; X86-NOBMI-NEXT: bswapl %esi
|
|
; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %esi
|
|
; X86-NOBMI-NEXT: pushl %eax
|
|
; X86-NOBMI-NEXT: calll use_i32@PLT
|
|
; X86-NOBMI-NEXT: addl $4, %esp
|
|
; X86-NOBMI-NEXT: movl %esi, %eax
|
|
; X86-NOBMI-NEXT: popl %esi
|
|
; X86-NOBMI-NEXT: retl
|
|
;
|
|
; X86-BMI-LABEL: andnot_bswap_i32_multiuse_not:
|
|
; X86-BMI: # %bb.0:
|
|
; X86-BMI-NEXT: pushl %esi
|
|
; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-BMI-NEXT: movl %eax, %ecx
|
|
; X86-BMI-NEXT: notl %ecx
|
|
; X86-BMI-NEXT: bswapl %eax
|
|
; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %eax, %esi
|
|
; X86-BMI-NEXT: pushl %ecx
|
|
; X86-BMI-NEXT: calll use_i32@PLT
|
|
; X86-BMI-NEXT: addl $4, %esp
|
|
; X86-BMI-NEXT: movl %esi, %eax
|
|
; X86-BMI-NEXT: popl %esi
|
|
; X86-BMI-NEXT: retl
|
|
;
|
|
; X64-NOBMI-LABEL: andnot_bswap_i32_multiuse_not:
|
|
; X64-NOBMI: # %bb.0:
|
|
; X64-NOBMI-NEXT: pushq %rbx
|
|
; X64-NOBMI-NEXT: notl %esi
|
|
; X64-NOBMI-NEXT: movl %esi, %ebx
|
|
; X64-NOBMI-NEXT: bswapl %ebx
|
|
; X64-NOBMI-NEXT: andl %edi, %ebx
|
|
; X64-NOBMI-NEXT: movl %esi, %edi
|
|
; X64-NOBMI-NEXT: callq use_i32@PLT
|
|
; X64-NOBMI-NEXT: movl %ebx, %eax
|
|
; X64-NOBMI-NEXT: popq %rbx
|
|
; X64-NOBMI-NEXT: retq
|
|
;
|
|
; X64-BMI-LABEL: andnot_bswap_i32_multiuse_not:
|
|
; X64-BMI: # %bb.0:
|
|
; X64-BMI-NEXT: pushq %rbx
|
|
; X64-BMI-NEXT: movl %esi, %eax
|
|
; X64-BMI-NEXT: notl %eax
|
|
; X64-BMI-NEXT: bswapl %esi
|
|
; X64-BMI-NEXT: andnl %edi, %esi, %ebx
|
|
; X64-BMI-NEXT: movl %eax, %edi
|
|
; X64-BMI-NEXT: callq use_i32@PLT
|
|
; X64-BMI-NEXT: movl %ebx, %eax
|
|
; X64-BMI-NEXT: popq %rbx
|
|
; X64-BMI-NEXT: retq
|
|
%not = xor i32 %a1, -1
|
|
%bswap = tail call i32 @llvm.bswap.i32(i32 %not)
|
|
%and = and i32 %bswap, %a0
|
|
call void @use_i32(i32 %not)
|
|
ret i32 %and
|
|
}
|
|
|
|
;
|
|
; Fold (and X, (bitreverse (not Y)))) -> (and X, (not (bitreverse Y)))
|
|
;
|
|
|
|
define i64 @andnot_bitreverse_i64(i64 %a0, i64 %a1) nounwind {
|
|
; X86-NOBMI-LABEL: andnot_bitreverse_i64:
|
|
; X86-NOBMI: # %bb.0:
|
|
; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NOBMI-NEXT: bswapl %eax
|
|
; X86-NOBMI-NEXT: movl %eax, %edx
|
|
; X86-NOBMI-NEXT: andl $252645135, %edx # imm = 0xF0F0F0F
|
|
; X86-NOBMI-NEXT: shll $4, %edx
|
|
; X86-NOBMI-NEXT: shrl $4, %eax
|
|
; X86-NOBMI-NEXT: andl $252645135, %eax # imm = 0xF0F0F0F
|
|
; X86-NOBMI-NEXT: orl %edx, %eax
|
|
; X86-NOBMI-NEXT: movl %eax, %edx
|
|
; X86-NOBMI-NEXT: andl $858993459, %edx # imm = 0x33333333
|
|
; X86-NOBMI-NEXT: shrl $2, %eax
|
|
; X86-NOBMI-NEXT: andl $858993459, %eax # imm = 0x33333333
|
|
; X86-NOBMI-NEXT: leal (%eax,%edx,4), %eax
|
|
; X86-NOBMI-NEXT: movl %eax, %edx
|
|
; X86-NOBMI-NEXT: andl $1431655765, %edx # imm = 0x55555555
|
|
; X86-NOBMI-NEXT: shrl %eax
|
|
; X86-NOBMI-NEXT: andl $1431655765, %eax # imm = 0x55555555
|
|
; X86-NOBMI-NEXT: leal (%eax,%edx,2), %eax
|
|
; X86-NOBMI-NEXT: notl %eax
|
|
; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %eax
|
|
; X86-NOBMI-NEXT: bswapl %ecx
|
|
; X86-NOBMI-NEXT: movl %ecx, %edx
|
|
; X86-NOBMI-NEXT: andl $252645135, %edx # imm = 0xF0F0F0F
|
|
; X86-NOBMI-NEXT: shll $4, %edx
|
|
; X86-NOBMI-NEXT: shrl $4, %ecx
|
|
; X86-NOBMI-NEXT: andl $252645135, %ecx # imm = 0xF0F0F0F
|
|
; X86-NOBMI-NEXT: orl %edx, %ecx
|
|
; X86-NOBMI-NEXT: movl %ecx, %edx
|
|
; X86-NOBMI-NEXT: andl $858993459, %edx # imm = 0x33333333
|
|
; X86-NOBMI-NEXT: shrl $2, %ecx
|
|
; X86-NOBMI-NEXT: andl $858993459, %ecx # imm = 0x33333333
|
|
; X86-NOBMI-NEXT: leal (%ecx,%edx,4), %ecx
|
|
; X86-NOBMI-NEXT: movl %ecx, %edx
|
|
; X86-NOBMI-NEXT: andl $1431655765, %edx # imm = 0x55555555
|
|
; X86-NOBMI-NEXT: shrl %ecx
|
|
; X86-NOBMI-NEXT: andl $1431655765, %ecx # imm = 0x55555555
|
|
; X86-NOBMI-NEXT: leal (%ecx,%edx,2), %edx
|
|
; X86-NOBMI-NEXT: notl %edx
|
|
; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %edx
|
|
; X86-NOBMI-NEXT: retl
|
|
;
|
|
; X86-BMI-LABEL: andnot_bitreverse_i64:
|
|
; X86-BMI: # %bb.0:
|
|
; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-BMI-NEXT: bswapl %eax
|
|
; X86-BMI-NEXT: movl %eax, %edx
|
|
; X86-BMI-NEXT: andl $252645135, %edx # imm = 0xF0F0F0F
|
|
; X86-BMI-NEXT: shll $4, %edx
|
|
; X86-BMI-NEXT: shrl $4, %eax
|
|
; X86-BMI-NEXT: andl $252645135, %eax # imm = 0xF0F0F0F
|
|
; X86-BMI-NEXT: orl %edx, %eax
|
|
; X86-BMI-NEXT: movl %eax, %edx
|
|
; X86-BMI-NEXT: andl $858993459, %edx # imm = 0x33333333
|
|
; X86-BMI-NEXT: shrl $2, %eax
|
|
; X86-BMI-NEXT: andl $858993459, %eax # imm = 0x33333333
|
|
; X86-BMI-NEXT: leal (%eax,%edx,4), %eax
|
|
; X86-BMI-NEXT: movl %eax, %edx
|
|
; X86-BMI-NEXT: andl $1431655765, %edx # imm = 0x55555555
|
|
; X86-BMI-NEXT: shrl %eax
|
|
; X86-BMI-NEXT: andl $1431655765, %eax # imm = 0x55555555
|
|
; X86-BMI-NEXT: leal (%eax,%edx,2), %eax
|
|
; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax
|
|
; X86-BMI-NEXT: bswapl %ecx
|
|
; X86-BMI-NEXT: movl %ecx, %edx
|
|
; X86-BMI-NEXT: andl $252645135, %edx # imm = 0xF0F0F0F
|
|
; X86-BMI-NEXT: shll $4, %edx
|
|
; X86-BMI-NEXT: shrl $4, %ecx
|
|
; X86-BMI-NEXT: andl $252645135, %ecx # imm = 0xF0F0F0F
|
|
; X86-BMI-NEXT: orl %edx, %ecx
|
|
; X86-BMI-NEXT: movl %ecx, %edx
|
|
; X86-BMI-NEXT: andl $858993459, %edx # imm = 0x33333333
|
|
; X86-BMI-NEXT: shrl $2, %ecx
|
|
; X86-BMI-NEXT: andl $858993459, %ecx # imm = 0x33333333
|
|
; X86-BMI-NEXT: leal (%ecx,%edx,4), %ecx
|
|
; X86-BMI-NEXT: movl %ecx, %edx
|
|
; X86-BMI-NEXT: andl $1431655765, %edx # imm = 0x55555555
|
|
; X86-BMI-NEXT: shrl %ecx
|
|
; X86-BMI-NEXT: andl $1431655765, %ecx # imm = 0x55555555
|
|
; X86-BMI-NEXT: leal (%ecx,%edx,2), %ecx
|
|
; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %ecx, %edx
|
|
; X86-BMI-NEXT: retl
|
|
;
|
|
; X64-NOBMI-LABEL: andnot_bitreverse_i64:
|
|
; X64-NOBMI: # %bb.0:
|
|
; X64-NOBMI-NEXT: bswapq %rsi
|
|
; X64-NOBMI-NEXT: movq %rsi, %rax
|
|
; X64-NOBMI-NEXT: shrq $4, %rax
|
|
; X64-NOBMI-NEXT: movabsq $1085102592571150095, %rcx # imm = 0xF0F0F0F0F0F0F0F
|
|
; X64-NOBMI-NEXT: andq %rcx, %rax
|
|
; X64-NOBMI-NEXT: andq %rcx, %rsi
|
|
; X64-NOBMI-NEXT: shlq $4, %rsi
|
|
; X64-NOBMI-NEXT: orq %rax, %rsi
|
|
; X64-NOBMI-NEXT: movabsq $3689348814741910323, %rax # imm = 0x3333333333333333
|
|
; X64-NOBMI-NEXT: movq %rsi, %rcx
|
|
; X64-NOBMI-NEXT: andq %rax, %rcx
|
|
; X64-NOBMI-NEXT: shrq $2, %rsi
|
|
; X64-NOBMI-NEXT: andq %rax, %rsi
|
|
; X64-NOBMI-NEXT: leaq (%rsi,%rcx,4), %rax
|
|
; X64-NOBMI-NEXT: movabsq $6148914691236517205, %rcx # imm = 0x5555555555555555
|
|
; X64-NOBMI-NEXT: movq %rax, %rdx
|
|
; X64-NOBMI-NEXT: andq %rcx, %rdx
|
|
; X64-NOBMI-NEXT: shrq %rax
|
|
; X64-NOBMI-NEXT: andq %rcx, %rax
|
|
; X64-NOBMI-NEXT: leaq (%rax,%rdx,2), %rax
|
|
; X64-NOBMI-NEXT: notq %rax
|
|
; X64-NOBMI-NEXT: andq %rdi, %rax
|
|
; X64-NOBMI-NEXT: retq
|
|
;
|
|
; X64-BMI-LABEL: andnot_bitreverse_i64:
|
|
; X64-BMI: # %bb.0:
|
|
; X64-BMI-NEXT: bswapq %rsi
|
|
; X64-BMI-NEXT: movq %rsi, %rax
|
|
; X64-BMI-NEXT: shrq $4, %rax
|
|
; X64-BMI-NEXT: movabsq $1085102592571150095, %rcx # imm = 0xF0F0F0F0F0F0F0F
|
|
; X64-BMI-NEXT: andq %rcx, %rax
|
|
; X64-BMI-NEXT: andq %rcx, %rsi
|
|
; X64-BMI-NEXT: shlq $4, %rsi
|
|
; X64-BMI-NEXT: orq %rax, %rsi
|
|
; X64-BMI-NEXT: movabsq $3689348814741910323, %rax # imm = 0x3333333333333333
|
|
; X64-BMI-NEXT: movq %rsi, %rcx
|
|
; X64-BMI-NEXT: andq %rax, %rcx
|
|
; X64-BMI-NEXT: shrq $2, %rsi
|
|
; X64-BMI-NEXT: andq %rax, %rsi
|
|
; X64-BMI-NEXT: leaq (%rsi,%rcx,4), %rax
|
|
; X64-BMI-NEXT: movabsq $6148914691236517205, %rcx # imm = 0x5555555555555555
|
|
; X64-BMI-NEXT: movq %rax, %rdx
|
|
; X64-BMI-NEXT: andq %rcx, %rdx
|
|
; X64-BMI-NEXT: shrq %rax
|
|
; X64-BMI-NEXT: andq %rcx, %rax
|
|
; X64-BMI-NEXT: leaq (%rax,%rdx,2), %rax
|
|
; X64-BMI-NEXT: andnq %rdi, %rax, %rax
|
|
; X64-BMI-NEXT: retq
|
|
%not = xor i64 %a1, -1
|
|
%bitrev = tail call i64 @llvm.bitreverse.i64(i64 %not)
|
|
%and = and i64 %bitrev, %a0
|
|
ret i64 %and
|
|
}
|
|
|
|
define i32 @andnot_bitreverse_i32(i32 %a0, i32 %a1) nounwind {
|
|
; X86-NOBMI-LABEL: andnot_bitreverse_i32:
|
|
; X86-NOBMI: # %bb.0:
|
|
; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NOBMI-NEXT: bswapl %eax
|
|
; X86-NOBMI-NEXT: movl %eax, %ecx
|
|
; X86-NOBMI-NEXT: andl $252645135, %ecx # imm = 0xF0F0F0F
|
|
; X86-NOBMI-NEXT: shll $4, %ecx
|
|
; X86-NOBMI-NEXT: shrl $4, %eax
|
|
; X86-NOBMI-NEXT: andl $252645135, %eax # imm = 0xF0F0F0F
|
|
; X86-NOBMI-NEXT: orl %ecx, %eax
|
|
; X86-NOBMI-NEXT: movl %eax, %ecx
|
|
; X86-NOBMI-NEXT: andl $858993459, %ecx # imm = 0x33333333
|
|
; X86-NOBMI-NEXT: shrl $2, %eax
|
|
; X86-NOBMI-NEXT: andl $858993459, %eax # imm = 0x33333333
|
|
; X86-NOBMI-NEXT: leal (%eax,%ecx,4), %eax
|
|
; X86-NOBMI-NEXT: movl %eax, %ecx
|
|
; X86-NOBMI-NEXT: andl $1431655765, %ecx # imm = 0x55555555
|
|
; X86-NOBMI-NEXT: shrl %eax
|
|
; X86-NOBMI-NEXT: andl $1431655765, %eax # imm = 0x55555555
|
|
; X86-NOBMI-NEXT: leal (%eax,%ecx,2), %eax
|
|
; X86-NOBMI-NEXT: notl %eax
|
|
; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %eax
|
|
; X86-NOBMI-NEXT: retl
|
|
;
|
|
; X86-BMI-LABEL: andnot_bitreverse_i32:
|
|
; X86-BMI: # %bb.0:
|
|
; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-BMI-NEXT: bswapl %eax
|
|
; X86-BMI-NEXT: movl %eax, %ecx
|
|
; X86-BMI-NEXT: andl $252645135, %ecx # imm = 0xF0F0F0F
|
|
; X86-BMI-NEXT: shll $4, %ecx
|
|
; X86-BMI-NEXT: shrl $4, %eax
|
|
; X86-BMI-NEXT: andl $252645135, %eax # imm = 0xF0F0F0F
|
|
; X86-BMI-NEXT: orl %ecx, %eax
|
|
; X86-BMI-NEXT: movl %eax, %ecx
|
|
; X86-BMI-NEXT: andl $858993459, %ecx # imm = 0x33333333
|
|
; X86-BMI-NEXT: shrl $2, %eax
|
|
; X86-BMI-NEXT: andl $858993459, %eax # imm = 0x33333333
|
|
; X86-BMI-NEXT: leal (%eax,%ecx,4), %eax
|
|
; X86-BMI-NEXT: movl %eax, %ecx
|
|
; X86-BMI-NEXT: andl $1431655765, %ecx # imm = 0x55555555
|
|
; X86-BMI-NEXT: shrl %eax
|
|
; X86-BMI-NEXT: andl $1431655765, %eax # imm = 0x55555555
|
|
; X86-BMI-NEXT: leal (%eax,%ecx,2), %eax
|
|
; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax
|
|
; X86-BMI-NEXT: retl
|
|
;
|
|
; X64-NOBMI-LABEL: andnot_bitreverse_i32:
|
|
; X64-NOBMI: # %bb.0:
|
|
; X64-NOBMI-NEXT: # kill: def $esi killed $esi def $rsi
|
|
; X64-NOBMI-NEXT: bswapl %esi
|
|
; X64-NOBMI-NEXT: movl %esi, %eax
|
|
; X64-NOBMI-NEXT: andl $252645135, %eax # imm = 0xF0F0F0F
|
|
; X64-NOBMI-NEXT: shll $4, %eax
|
|
; X64-NOBMI-NEXT: shrl $4, %esi
|
|
; X64-NOBMI-NEXT: andl $252645135, %esi # imm = 0xF0F0F0F
|
|
; X64-NOBMI-NEXT: orl %eax, %esi
|
|
; X64-NOBMI-NEXT: movl %esi, %eax
|
|
; X64-NOBMI-NEXT: andl $858993459, %eax # imm = 0x33333333
|
|
; X64-NOBMI-NEXT: shrl $2, %esi
|
|
; X64-NOBMI-NEXT: andl $858993459, %esi # imm = 0x33333333
|
|
; X64-NOBMI-NEXT: leal (%rsi,%rax,4), %eax
|
|
; X64-NOBMI-NEXT: movl %eax, %ecx
|
|
; X64-NOBMI-NEXT: andl $1431655765, %ecx # imm = 0x55555555
|
|
; X64-NOBMI-NEXT: shrl %eax
|
|
; X64-NOBMI-NEXT: andl $1431655765, %eax # imm = 0x55555555
|
|
; X64-NOBMI-NEXT: leal (%rax,%rcx,2), %eax
|
|
; X64-NOBMI-NEXT: notl %eax
|
|
; X64-NOBMI-NEXT: andl %edi, %eax
|
|
; X64-NOBMI-NEXT: retq
|
|
;
|
|
; X64-BMI-LABEL: andnot_bitreverse_i32:
|
|
; X64-BMI: # %bb.0:
|
|
; X64-BMI-NEXT: # kill: def $esi killed $esi def $rsi
|
|
; X64-BMI-NEXT: bswapl %esi
|
|
; X64-BMI-NEXT: movl %esi, %eax
|
|
; X64-BMI-NEXT: andl $252645135, %eax # imm = 0xF0F0F0F
|
|
; X64-BMI-NEXT: shll $4, %eax
|
|
; X64-BMI-NEXT: shrl $4, %esi
|
|
; X64-BMI-NEXT: andl $252645135, %esi # imm = 0xF0F0F0F
|
|
; X64-BMI-NEXT: orl %eax, %esi
|
|
; X64-BMI-NEXT: movl %esi, %eax
|
|
; X64-BMI-NEXT: andl $858993459, %eax # imm = 0x33333333
|
|
; X64-BMI-NEXT: shrl $2, %esi
|
|
; X64-BMI-NEXT: andl $858993459, %esi # imm = 0x33333333
|
|
; X64-BMI-NEXT: leal (%rsi,%rax,4), %eax
|
|
; X64-BMI-NEXT: movl %eax, %ecx
|
|
; X64-BMI-NEXT: andl $1431655765, %ecx # imm = 0x55555555
|
|
; X64-BMI-NEXT: shrl %eax
|
|
; X64-BMI-NEXT: andl $1431655765, %eax # imm = 0x55555555
|
|
; X64-BMI-NEXT: leal (%rax,%rcx,2), %eax
|
|
; X64-BMI-NEXT: andnl %edi, %eax, %eax
|
|
; X64-BMI-NEXT: retq
|
|
%not = xor i32 %a1, -1
|
|
%bitrev = tail call i32 @llvm.bitreverse.i32(i32 %not)
|
|
%and = and i32 %bitrev, %a0
|
|
ret i32 %and
|
|
}
|
|
|
|
define i16 @andnot_bitreverse_i16(i16 %a0, i16 %a1) nounwind {
|
|
; X86-LABEL: andnot_bitreverse_i16:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movzwl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: rolw $8, %ax
|
|
; X86-NEXT: movl %eax, %ecx
|
|
; X86-NEXT: andl $3855, %ecx # imm = 0xF0F
|
|
; X86-NEXT: shll $4, %ecx
|
|
; X86-NEXT: shrl $4, %eax
|
|
; X86-NEXT: andl $3855, %eax # imm = 0xF0F
|
|
; X86-NEXT: orl %ecx, %eax
|
|
; X86-NEXT: movl %eax, %ecx
|
|
; X86-NEXT: andl $13107, %ecx # imm = 0x3333
|
|
; X86-NEXT: shrl $2, %eax
|
|
; X86-NEXT: andl $13107, %eax # imm = 0x3333
|
|
; X86-NEXT: leal (%eax,%ecx,4), %eax
|
|
; X86-NEXT: movl %eax, %ecx
|
|
; X86-NEXT: andl $21845, %ecx # imm = 0x5555
|
|
; X86-NEXT: shrl %eax
|
|
; X86-NEXT: andl $21845, %eax # imm = 0x5555
|
|
; X86-NEXT: leal (%eax,%ecx,2), %eax
|
|
; X86-NEXT: notl %eax
|
|
; X86-NEXT: andw {{[0-9]+}}(%esp), %ax
|
|
; X86-NEXT: # kill: def $ax killed $ax killed $eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-NOBMI-LABEL: andnot_bitreverse_i16:
|
|
; X64-NOBMI: # %bb.0:
|
|
; X64-NOBMI-NEXT: # kill: def $esi killed $esi def $rsi
|
|
; X64-NOBMI-NEXT: rolw $8, %si
|
|
; X64-NOBMI-NEXT: movl %esi, %eax
|
|
; X64-NOBMI-NEXT: andl $3855, %eax # imm = 0xF0F
|
|
; X64-NOBMI-NEXT: shll $4, %eax
|
|
; X64-NOBMI-NEXT: shrl $4, %esi
|
|
; X64-NOBMI-NEXT: andl $3855, %esi # imm = 0xF0F
|
|
; X64-NOBMI-NEXT: orl %eax, %esi
|
|
; X64-NOBMI-NEXT: movl %esi, %eax
|
|
; X64-NOBMI-NEXT: andl $13107, %eax # imm = 0x3333
|
|
; X64-NOBMI-NEXT: shrl $2, %esi
|
|
; X64-NOBMI-NEXT: andl $13107, %esi # imm = 0x3333
|
|
; X64-NOBMI-NEXT: leal (%rsi,%rax,4), %eax
|
|
; X64-NOBMI-NEXT: movl %eax, %ecx
|
|
; X64-NOBMI-NEXT: andl $21845, %ecx # imm = 0x5555
|
|
; X64-NOBMI-NEXT: shrl %eax
|
|
; X64-NOBMI-NEXT: andl $21845, %eax # imm = 0x5555
|
|
; X64-NOBMI-NEXT: leal (%rax,%rcx,2), %eax
|
|
; X64-NOBMI-NEXT: notl %eax
|
|
; X64-NOBMI-NEXT: andl %edi, %eax
|
|
; X64-NOBMI-NEXT: # kill: def $ax killed $ax killed $eax
|
|
; X64-NOBMI-NEXT: retq
|
|
;
|
|
; X64-BMI-LABEL: andnot_bitreverse_i16:
|
|
; X64-BMI: # %bb.0:
|
|
; X64-BMI-NEXT: # kill: def $esi killed $esi def $rsi
|
|
; X64-BMI-NEXT: rolw $8, %si
|
|
; X64-BMI-NEXT: movl %esi, %eax
|
|
; X64-BMI-NEXT: andl $3855, %eax # imm = 0xF0F
|
|
; X64-BMI-NEXT: shll $4, %eax
|
|
; X64-BMI-NEXT: shrl $4, %esi
|
|
; X64-BMI-NEXT: andl $3855, %esi # imm = 0xF0F
|
|
; X64-BMI-NEXT: orl %eax, %esi
|
|
; X64-BMI-NEXT: movl %esi, %eax
|
|
; X64-BMI-NEXT: andl $13107, %eax # imm = 0x3333
|
|
; X64-BMI-NEXT: shrl $2, %esi
|
|
; X64-BMI-NEXT: andl $13107, %esi # imm = 0x3333
|
|
; X64-BMI-NEXT: leal (%rsi,%rax,4), %eax
|
|
; X64-BMI-NEXT: movl %eax, %ecx
|
|
; X64-BMI-NEXT: andl $21845, %ecx # imm = 0x5555
|
|
; X64-BMI-NEXT: shrl %eax
|
|
; X64-BMI-NEXT: andl $21845, %eax # imm = 0x5555
|
|
; X64-BMI-NEXT: leal (%rax,%rcx,2), %eax
|
|
; X64-BMI-NEXT: andnl %edi, %eax, %eax
|
|
; X64-BMI-NEXT: # kill: def $ax killed $ax killed $eax
|
|
; X64-BMI-NEXT: retq
|
|
%not = xor i16 %a1, -1
|
|
%bitrev = tail call i16 @llvm.bitreverse.i16(i16 %not)
|
|
%and = and i16 %bitrev, %a0
|
|
ret i16 %and
|
|
}
|
|
|
|
define i8 @andnot_bitreverse_i8(i8 %a0, i8 %a1) nounwind {
|
|
; X86-LABEL: andnot_bitreverse_i8:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movzbl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: rolb $4, %al
|
|
; X86-NEXT: movl %eax, %ecx
|
|
; X86-NEXT: andb $51, %cl
|
|
; X86-NEXT: shlb $2, %cl
|
|
; X86-NEXT: shrb $2, %al
|
|
; X86-NEXT: andb $51, %al
|
|
; X86-NEXT: orb %cl, %al
|
|
; X86-NEXT: movl %eax, %ecx
|
|
; X86-NEXT: andb $85, %cl
|
|
; X86-NEXT: addb %cl, %cl
|
|
; X86-NEXT: shrb %al
|
|
; X86-NEXT: andb $85, %al
|
|
; X86-NEXT: orb %cl, %al
|
|
; X86-NEXT: notb %al
|
|
; X86-NEXT: andb {{[0-9]+}}(%esp), %al
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: andnot_bitreverse_i8:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: rolb $4, %sil
|
|
; X64-NEXT: movl %esi, %eax
|
|
; X64-NEXT: andb $51, %al
|
|
; X64-NEXT: shlb $2, %al
|
|
; X64-NEXT: shrb $2, %sil
|
|
; X64-NEXT: andb $51, %sil
|
|
; X64-NEXT: orb %sil, %al
|
|
; X64-NEXT: movl %eax, %ecx
|
|
; X64-NEXT: andb $85, %cl
|
|
; X64-NEXT: addb %cl, %cl
|
|
; X64-NEXT: shrb %al
|
|
; X64-NEXT: andb $85, %al
|
|
; X64-NEXT: orb %cl, %al
|
|
; X64-NEXT: notb %al
|
|
; X64-NEXT: andb %dil, %al
|
|
; X64-NEXT: retq
|
|
%not = xor i8 %a1, -1
|
|
%bitrev = tail call i8 @llvm.bitreverse.i8(i8 %not)
|
|
%and = and i8 %bitrev, %a0
|
|
ret i8 %and
|
|
}
|