llvm-project/llvm/lib/Target/X86/X86PartialReduction.cpp
Haohai Wen 18a1085e02 [X86] Fix collectLeaves for adds used by phi that forms loop
When add has additional users, we should indentify whether add's
user is phi that forms loop rather than root's.

Reviewed By: LuoYuanke

Differential Revision: https://reviews.llvm.org/D129169
2022-07-08 10:39:02 +08:00

546 lines
17 KiB
C++

//===-- X86PartialReduction.cpp -------------------------------------------===//
//
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
// See https://llvm.org/LICENSE.txt for license information.
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
//
//===----------------------------------------------------------------------===//
//
// This pass looks for add instructions used by a horizontal reduction to see
// if we might be able to use pmaddwd or psadbw. Some cases of this require
// cross basic block knowledge and can't be done in SelectionDAG.
//
//===----------------------------------------------------------------------===//
#include "X86.h"
#include "X86TargetMachine.h"
#include "llvm/Analysis/ValueTracking.h"
#include "llvm/CodeGen/TargetPassConfig.h"
#include "llvm/IR/Constants.h"
#include "llvm/IR/IRBuilder.h"
#include "llvm/IR/Instructions.h"
#include "llvm/IR/IntrinsicInst.h"
#include "llvm/IR/IntrinsicsX86.h"
#include "llvm/IR/Operator.h"
#include "llvm/IR/PatternMatch.h"
#include "llvm/Pass.h"
#include "llvm/Support/KnownBits.h"
using namespace llvm;
#define DEBUG_TYPE "x86-partial-reduction"
namespace {
class X86PartialReduction : public FunctionPass {
const DataLayout *DL;
const X86Subtarget *ST;
public:
static char ID; // Pass identification, replacement for typeid.
X86PartialReduction() : FunctionPass(ID) { }
bool runOnFunction(Function &Fn) override;
void getAnalysisUsage(AnalysisUsage &AU) const override {
AU.setPreservesCFG();
}
StringRef getPassName() const override {
return "X86 Partial Reduction";
}
private:
bool tryMAddReplacement(Instruction *Op, bool ReduceInOneBB);
bool trySADReplacement(Instruction *Op);
};
}
FunctionPass *llvm::createX86PartialReductionPass() {
return new X86PartialReduction();
}
char X86PartialReduction::ID = 0;
INITIALIZE_PASS(X86PartialReduction, DEBUG_TYPE,
"X86 Partial Reduction", false, false)
// This function should be aligned with detectExtMul() in X86ISelLowering.cpp.
static bool matchVPDPBUSDPattern(const X86Subtarget *ST, BinaryOperator *Mul,
const DataLayout *DL) {
if (!ST->hasVNNI() && !ST->hasAVXVNNI())
return false;
Value *LHS = Mul->getOperand(0);
Value *RHS = Mul->getOperand(1);
if (isa<SExtInst>(LHS))
std::swap(LHS, RHS);
auto IsFreeTruncation = [&](Value *Op) {
if (auto *Cast = dyn_cast<CastInst>(Op)) {
if (Cast->getParent() == Mul->getParent() &&
(Cast->getOpcode() == Instruction::SExt ||
Cast->getOpcode() == Instruction::ZExt) &&
Cast->getOperand(0)->getType()->getScalarSizeInBits() <= 8)
return true;
}
return isa<Constant>(Op);
};
// (dpbusd (zext a), (sext, b)). Since the first operand should be unsigned
// value, we need to check LHS is zero extended value. RHS should be signed
// value, so we just check the signed bits.
if ((IsFreeTruncation(LHS) &&
computeKnownBits(LHS, *DL).countMaxActiveBits() <= 8) &&
(IsFreeTruncation(RHS) && ComputeMaxSignificantBits(RHS, *DL) <= 8))
return true;
return false;
}
bool X86PartialReduction::tryMAddReplacement(Instruction *Op,
bool ReduceInOneBB) {
if (!ST->hasSSE2())
return false;
// Need at least 8 elements.
if (cast<FixedVectorType>(Op->getType())->getNumElements() < 8)
return false;
// Element type should be i32.
if (!cast<VectorType>(Op->getType())->getElementType()->isIntegerTy(32))
return false;
auto *Mul = dyn_cast<BinaryOperator>(Op);
if (!Mul || Mul->getOpcode() != Instruction::Mul)
return false;
Value *LHS = Mul->getOperand(0);
Value *RHS = Mul->getOperand(1);
// If the target support VNNI, leave it to ISel to combine reduce operation
// to VNNI instruction.
// TODO: we can support transforming reduce to VNNI intrinsic for across block
// in this pass.
if (ReduceInOneBB && matchVPDPBUSDPattern(ST, Mul, DL))
return false;
// LHS and RHS should be only used once or if they are the same then only
// used twice. Only check this when SSE4.1 is enabled and we have zext/sext
// instructions, otherwise we use punpck to emulate zero extend in stages. The
// trunc/ we need to do likely won't introduce new instructions in that case.
if (ST->hasSSE41()) {
if (LHS == RHS) {
if (!isa<Constant>(LHS) && !LHS->hasNUses(2))
return false;
} else {
if (!isa<Constant>(LHS) && !LHS->hasOneUse())
return false;
if (!isa<Constant>(RHS) && !RHS->hasOneUse())
return false;
}
}
auto CanShrinkOp = [&](Value *Op) {
auto IsFreeTruncation = [&](Value *Op) {
if (auto *Cast = dyn_cast<CastInst>(Op)) {
if (Cast->getParent() == Mul->getParent() &&
(Cast->getOpcode() == Instruction::SExt ||
Cast->getOpcode() == Instruction::ZExt) &&
Cast->getOperand(0)->getType()->getScalarSizeInBits() <= 16)
return true;
}
return isa<Constant>(Op);
};
// If the operation can be freely truncated and has enough sign bits we
// can shrink.
if (IsFreeTruncation(Op) &&
ComputeNumSignBits(Op, *DL, 0, nullptr, Mul) > 16)
return true;
// SelectionDAG has limited support for truncating through an add or sub if
// the inputs are freely truncatable.
if (auto *BO = dyn_cast<BinaryOperator>(Op)) {
if (BO->getParent() == Mul->getParent() &&
IsFreeTruncation(BO->getOperand(0)) &&
IsFreeTruncation(BO->getOperand(1)) &&
ComputeNumSignBits(Op, *DL, 0, nullptr, Mul) > 16)
return true;
}
return false;
};
// Both Ops need to be shrinkable.
if (!CanShrinkOp(LHS) && !CanShrinkOp(RHS))
return false;
IRBuilder<> Builder(Mul);
auto *MulTy = cast<FixedVectorType>(Op->getType());
unsigned NumElts = MulTy->getNumElements();
// Extract even elements and odd elements and add them together. This will
// be pattern matched by SelectionDAG to pmaddwd. This instruction will be
// half the original width.
SmallVector<int, 16> EvenMask(NumElts / 2);
SmallVector<int, 16> OddMask(NumElts / 2);
for (int i = 0, e = NumElts / 2; i != e; ++i) {
EvenMask[i] = i * 2;
OddMask[i] = i * 2 + 1;
}
// Creating a new mul so the replaceAllUsesWith below doesn't replace the
// uses in the shuffles we're creating.
Value *NewMul = Builder.CreateMul(Mul->getOperand(0), Mul->getOperand(1));
Value *EvenElts = Builder.CreateShuffleVector(NewMul, NewMul, EvenMask);
Value *OddElts = Builder.CreateShuffleVector(NewMul, NewMul, OddMask);
Value *MAdd = Builder.CreateAdd(EvenElts, OddElts);
// Concatenate zeroes to extend back to the original type.
SmallVector<int, 32> ConcatMask(NumElts);
std::iota(ConcatMask.begin(), ConcatMask.end(), 0);
Value *Zero = Constant::getNullValue(MAdd->getType());
Value *Concat = Builder.CreateShuffleVector(MAdd, Zero, ConcatMask);
Mul->replaceAllUsesWith(Concat);
Mul->eraseFromParent();
return true;
}
bool X86PartialReduction::trySADReplacement(Instruction *Op) {
if (!ST->hasSSE2())
return false;
// TODO: There's nothing special about i32, any integer type above i16 should
// work just as well.
if (!cast<VectorType>(Op->getType())->getElementType()->isIntegerTy(32))
return false;
Value *LHS;
if (match(Op, PatternMatch::m_Intrinsic<Intrinsic::abs>())) {
LHS = Op->getOperand(0);
} else {
// Operand should be a select.
auto *SI = dyn_cast<SelectInst>(Op);
if (!SI)
return false;
Value *RHS;
// Select needs to implement absolute value.
auto SPR = matchSelectPattern(SI, LHS, RHS);
if (SPR.Flavor != SPF_ABS)
return false;
}
// Need a subtract of two values.
auto *Sub = dyn_cast<BinaryOperator>(LHS);
if (!Sub || Sub->getOpcode() != Instruction::Sub)
return false;
// Look for zero extend from i8.
auto getZeroExtendedVal = [](Value *Op) -> Value * {
if (auto *ZExt = dyn_cast<ZExtInst>(Op))
if (cast<VectorType>(ZExt->getOperand(0)->getType())
->getElementType()
->isIntegerTy(8))
return ZExt->getOperand(0);
return nullptr;
};
// Both operands of the subtract should be extends from vXi8.
Value *Op0 = getZeroExtendedVal(Sub->getOperand(0));
Value *Op1 = getZeroExtendedVal(Sub->getOperand(1));
if (!Op0 || !Op1)
return false;
IRBuilder<> Builder(Op);
auto *OpTy = cast<FixedVectorType>(Op->getType());
unsigned NumElts = OpTy->getNumElements();
unsigned IntrinsicNumElts;
Intrinsic::ID IID;
if (ST->hasBWI() && NumElts >= 64) {
IID = Intrinsic::x86_avx512_psad_bw_512;
IntrinsicNumElts = 64;
} else if (ST->hasAVX2() && NumElts >= 32) {
IID = Intrinsic::x86_avx2_psad_bw;
IntrinsicNumElts = 32;
} else {
IID = Intrinsic::x86_sse2_psad_bw;
IntrinsicNumElts = 16;
}
Function *PSADBWFn = Intrinsic::getDeclaration(Op->getModule(), IID);
if (NumElts < 16) {
// Pad input with zeroes.
SmallVector<int, 32> ConcatMask(16);
for (unsigned i = 0; i != NumElts; ++i)
ConcatMask[i] = i;
for (unsigned i = NumElts; i != 16; ++i)
ConcatMask[i] = (i % NumElts) + NumElts;
Value *Zero = Constant::getNullValue(Op0->getType());
Op0 = Builder.CreateShuffleVector(Op0, Zero, ConcatMask);
Op1 = Builder.CreateShuffleVector(Op1, Zero, ConcatMask);
NumElts = 16;
}
// Intrinsics produce vXi64 and need to be casted to vXi32.
auto *I32Ty =
FixedVectorType::get(Builder.getInt32Ty(), IntrinsicNumElts / 4);
assert(NumElts % IntrinsicNumElts == 0 && "Unexpected number of elements!");
unsigned NumSplits = NumElts / IntrinsicNumElts;
// First collect the pieces we need.
SmallVector<Value *, 4> Ops(NumSplits);
for (unsigned i = 0; i != NumSplits; ++i) {
SmallVector<int, 64> ExtractMask(IntrinsicNumElts);
std::iota(ExtractMask.begin(), ExtractMask.end(), i * IntrinsicNumElts);
Value *ExtractOp0 = Builder.CreateShuffleVector(Op0, Op0, ExtractMask);
Value *ExtractOp1 = Builder.CreateShuffleVector(Op1, Op0, ExtractMask);
Ops[i] = Builder.CreateCall(PSADBWFn, {ExtractOp0, ExtractOp1});
Ops[i] = Builder.CreateBitCast(Ops[i], I32Ty);
}
assert(isPowerOf2_32(NumSplits) && "Expected power of 2 splits");
unsigned Stages = Log2_32(NumSplits);
for (unsigned s = Stages; s > 0; --s) {
unsigned NumConcatElts =
cast<FixedVectorType>(Ops[0]->getType())->getNumElements() * 2;
for (unsigned i = 0; i != 1U << (s - 1); ++i) {
SmallVector<int, 64> ConcatMask(NumConcatElts);
std::iota(ConcatMask.begin(), ConcatMask.end(), 0);
Ops[i] = Builder.CreateShuffleVector(Ops[i*2], Ops[i*2+1], ConcatMask);
}
}
// At this point the final value should be in Ops[0]. Now we need to adjust
// it to the final original type.
NumElts = cast<FixedVectorType>(OpTy)->getNumElements();
if (NumElts == 2) {
// Extract down to 2 elements.
Ops[0] = Builder.CreateShuffleVector(Ops[0], Ops[0], ArrayRef<int>{0, 1});
} else if (NumElts >= 8) {
SmallVector<int, 32> ConcatMask(NumElts);
unsigned SubElts =
cast<FixedVectorType>(Ops[0]->getType())->getNumElements();
for (unsigned i = 0; i != SubElts; ++i)
ConcatMask[i] = i;
for (unsigned i = SubElts; i != NumElts; ++i)
ConcatMask[i] = (i % SubElts) + SubElts;
Value *Zero = Constant::getNullValue(Ops[0]->getType());
Ops[0] = Builder.CreateShuffleVector(Ops[0], Zero, ConcatMask);
}
Op->replaceAllUsesWith(Ops[0]);
Op->eraseFromParent();
return true;
}
// Walk backwards from the ExtractElementInst and determine if it is the end of
// a horizontal reduction. Return the input to the reduction if we find one.
static Value *matchAddReduction(const ExtractElementInst &EE,
bool &ReduceInOneBB) {
ReduceInOneBB = true;
// Make sure we're extracting index 0.
auto *Index = dyn_cast<ConstantInt>(EE.getIndexOperand());
if (!Index || !Index->isNullValue())
return nullptr;
const auto *BO = dyn_cast<BinaryOperator>(EE.getVectorOperand());
if (!BO || BO->getOpcode() != Instruction::Add || !BO->hasOneUse())
return nullptr;
if (EE.getParent() != BO->getParent())
ReduceInOneBB = false;
unsigned NumElems = cast<FixedVectorType>(BO->getType())->getNumElements();
// Ensure the reduction size is a power of 2.
if (!isPowerOf2_32(NumElems))
return nullptr;
const Value *Op = BO;
unsigned Stages = Log2_32(NumElems);
for (unsigned i = 0; i != Stages; ++i) {
const auto *BO = dyn_cast<BinaryOperator>(Op);
if (!BO || BO->getOpcode() != Instruction::Add)
return nullptr;
if (EE.getParent() != BO->getParent())
ReduceInOneBB = false;
// If this isn't the first add, then it should only have 2 users, the
// shuffle and another add which we checked in the previous iteration.
if (i != 0 && !BO->hasNUses(2))
return nullptr;
Value *LHS = BO->getOperand(0);
Value *RHS = BO->getOperand(1);
auto *Shuffle = dyn_cast<ShuffleVectorInst>(LHS);
if (Shuffle) {
Op = RHS;
} else {
Shuffle = dyn_cast<ShuffleVectorInst>(RHS);
Op = LHS;
}
// The first operand of the shuffle should be the same as the other operand
// of the bin op.
if (!Shuffle || Shuffle->getOperand(0) != Op)
return nullptr;
// Verify the shuffle has the expected (at this stage of the pyramid) mask.
unsigned MaskEnd = 1 << i;
for (unsigned Index = 0; Index < MaskEnd; ++Index)
if (Shuffle->getMaskValue(Index) != (int)(MaskEnd + Index))
return nullptr;
}
return const_cast<Value *>(Op);
}
// See if this BO is reachable from this Phi by walking forward through single
// use BinaryOperators with the same opcode. If we get back then we know we've
// found a loop and it is safe to step through this Add to find more leaves.
static bool isReachableFromPHI(PHINode *Phi, BinaryOperator *BO) {
// The PHI itself should only have one use.
if (!Phi->hasOneUse())
return false;
Instruction *U = cast<Instruction>(*Phi->user_begin());
if (U == BO)
return true;
while (U->hasOneUse() && U->getOpcode() == BO->getOpcode())
U = cast<Instruction>(*U->user_begin());
return U == BO;
}
// Collect all the leaves of the tree of adds that feeds into the horizontal
// reduction. Root is the Value that is used by the horizontal reduction.
// We look through single use phis, single use adds, or adds that are used by
// a phi that forms a loop with the add.
static void collectLeaves(Value *Root, SmallVectorImpl<Instruction *> &Leaves) {
SmallPtrSet<Value *, 8> Visited;
SmallVector<Value *, 8> Worklist;
Worklist.push_back(Root);
while (!Worklist.empty()) {
Value *V = Worklist.pop_back_val();
if (!Visited.insert(V).second)
continue;
if (auto *PN = dyn_cast<PHINode>(V)) {
// PHI node should have single use unless it is the root node, then it
// has 2 uses.
if (!PN->hasNUses(PN == Root ? 2 : 1))
break;
// Push incoming values to the worklist.
append_range(Worklist, PN->incoming_values());
continue;
}
if (auto *BO = dyn_cast<BinaryOperator>(V)) {
if (BO->getOpcode() == Instruction::Add) {
// Simple case. Single use, just push its operands to the worklist.
if (BO->hasNUses(BO == Root ? 2 : 1)) {
append_range(Worklist, BO->operands());
continue;
}
// If there is additional use, make sure it is an unvisited phi that
// gets us back to this node.
if (BO->hasNUses(BO == Root ? 3 : 2)) {
PHINode *PN = nullptr;
for (auto *U : BO->users())
if (auto *P = dyn_cast<PHINode>(U))
if (!Visited.count(P))
PN = P;
// If we didn't find a 2-input PHI then this isn't a case we can
// handle.
if (!PN || PN->getNumIncomingValues() != 2)
continue;
// Walk forward from this phi to see if it reaches back to this add.
if (!isReachableFromPHI(PN, BO))
continue;
// The phi forms a loop with this Add, push its operands.
append_range(Worklist, BO->operands());
}
}
}
// Not an add or phi, make it a leaf.
if (auto *I = dyn_cast<Instruction>(V)) {
if (!V->hasNUses(I == Root ? 2 : 1))
continue;
// Add this as a leaf.
Leaves.push_back(I);
}
}
}
bool X86PartialReduction::runOnFunction(Function &F) {
if (skipFunction(F))
return false;
auto *TPC = getAnalysisIfAvailable<TargetPassConfig>();
if (!TPC)
return false;
auto &TM = TPC->getTM<X86TargetMachine>();
ST = TM.getSubtargetImpl(F);
DL = &F.getParent()->getDataLayout();
bool MadeChange = false;
for (auto &BB : F) {
for (auto &I : BB) {
auto *EE = dyn_cast<ExtractElementInst>(&I);
if (!EE)
continue;
bool ReduceInOneBB;
// First find a reduction tree.
// FIXME: Do we need to handle other opcodes than Add?
Value *Root = matchAddReduction(*EE, ReduceInOneBB);
if (!Root)
continue;
SmallVector<Instruction *, 8> Leaves;
collectLeaves(Root, Leaves);
for (Instruction *I : Leaves) {
if (tryMAddReplacement(I, ReduceInOneBB)) {
MadeChange = true;
continue;
}
// Don't do SAD matching on the root node. SelectionDAG already
// has support for that and currently generates better code.
if (I != Root && trySADReplacement(I))
MadeChange = true;
}
}
}
return MadeChange;
}