Compare commits
8 Commits
main
...
users/zhao
Author | SHA1 | Date | |
---|---|---|---|
![]() |
636914e0f3 | ||
![]() |
49294ba29e | ||
![]() |
6e525512d0 | ||
![]() |
b99249e3f7 | ||
![]() |
66aae1ebd9 | ||
![]() |
6acd00f096 | ||
![]() |
3674bad63b | ||
![]() |
8375c79afe |
@ -2434,6 +2434,7 @@ static SDValue lowerBUILD_VECTORAsBroadCastLoad(BuildVectorSDNode *BVOp,
|
|||||||
SDValue LoongArchTargetLowering::lowerBUILD_VECTOR(SDValue Op,
|
SDValue LoongArchTargetLowering::lowerBUILD_VECTOR(SDValue Op,
|
||||||
SelectionDAG &DAG) const {
|
SelectionDAG &DAG) const {
|
||||||
BuildVectorSDNode *Node = cast<BuildVectorSDNode>(Op);
|
BuildVectorSDNode *Node = cast<BuildVectorSDNode>(Op);
|
||||||
|
MVT VT = Node->getSimpleValueType(0);
|
||||||
EVT ResTy = Op->getValueType(0);
|
EVT ResTy = Op->getValueType(0);
|
||||||
unsigned NumElts = ResTy.getVectorNumElements();
|
unsigned NumElts = ResTy.getVectorNumElements();
|
||||||
SDLoc DL(Op);
|
SDLoc DL(Op);
|
||||||
@ -2517,24 +2518,107 @@ SDValue LoongArchTargetLowering::lowerBUILD_VECTOR(SDValue Op,
|
|||||||
}
|
}
|
||||||
|
|
||||||
if (!IsConstant) {
|
if (!IsConstant) {
|
||||||
// Use INSERT_VECTOR_ELT operations rather than expand to stores.
|
// If the BUILD_VECTOR has a repeated pattern, use INSERT_VECTOR_ELT to fill
|
||||||
// The resulting code is the same length as the expansion, but it doesn't
|
// the sub-sequence of the vector and then broadcast the sub-sequence.
|
||||||
// use memory operations.
|
//
|
||||||
assert(ResTy.isVector());
|
// TODO: If the BUILD_VECTOR contains undef elements, consider falling
|
||||||
|
// back to use INSERT_VECTOR_ELT to materialize the vector, because it
|
||||||
SDValue Op0 = Node->getOperand(0);
|
// generates worse code in some cases. This could be further optimized
|
||||||
|
// with more consideration.
|
||||||
|
SmallVector<SDValue> Sequence;
|
||||||
|
BitVector UndefElements;
|
||||||
|
if (Node->getRepeatedSequence(Sequence, &UndefElements) &&
|
||||||
|
UndefElements.count() == 0) {
|
||||||
SDValue Vector = DAG.getUNDEF(ResTy);
|
SDValue Vector = DAG.getUNDEF(ResTy);
|
||||||
|
SDValue FillVec = Vector;
|
||||||
|
EVT FillTy = ResTy;
|
||||||
|
|
||||||
|
// Using LSX instructions to fill the sub-sequence of 256-bits vector,
|
||||||
|
// because the high part can be simply treated as undef.
|
||||||
|
if (Is256Vec) {
|
||||||
|
FillTy = ResTy.getHalfNumVectorElementsVT(*DAG.getContext());
|
||||||
|
FillVec = DAG.getExtractSubvector(DL, FillTy, Vector, 0);
|
||||||
|
}
|
||||||
|
|
||||||
|
SDValue Op0 = Sequence[0];
|
||||||
|
unsigned SeqLen = Sequence.size();
|
||||||
if (!Op0.isUndef())
|
if (!Op0.isUndef())
|
||||||
Vector = DAG.getNode(ISD::SCALAR_TO_VECTOR, DL, ResTy, Op0);
|
FillVec = DAG.getNode(ISD::SCALAR_TO_VECTOR, DL, FillTy, Op0);
|
||||||
for (unsigned i = 1; i < NumElts; ++i) {
|
for (unsigned i = 1; i < SeqLen; ++i) {
|
||||||
|
SDValue Opi = Sequence[i];
|
||||||
|
if (Opi.isUndef())
|
||||||
|
continue;
|
||||||
|
FillVec = DAG.getNode(ISD::INSERT_VECTOR_ELT, DL, FillTy, FillVec, Opi,
|
||||||
|
DAG.getConstant(i, DL, Subtarget.getGRLenVT()));
|
||||||
|
}
|
||||||
|
|
||||||
|
unsigned SplatLen = NumElts / SeqLen;
|
||||||
|
MVT SplatEltTy = MVT::getIntegerVT(VT.getScalarSizeInBits() * SeqLen);
|
||||||
|
MVT SplatTy = MVT::getVectorVT(SplatEltTy, SplatLen);
|
||||||
|
|
||||||
|
// If size of the sub-sequence is half of a 256-bits vector, bitcast the
|
||||||
|
// vector to v4i64 type in order to match the pattern of XVREPLVE0Q.
|
||||||
|
if (SplatEltTy == MVT::i128)
|
||||||
|
SplatTy = MVT::v4i64;
|
||||||
|
|
||||||
|
SDValue SplatVec;
|
||||||
|
SDValue SrcVec = DAG.getBitcast(
|
||||||
|
SplatTy,
|
||||||
|
Is256Vec ? DAG.getInsertSubvector(DL, Vector, FillVec, 0) : FillVec);
|
||||||
|
if (Is256Vec) {
|
||||||
|
SplatVec =
|
||||||
|
DAG.getNode((SplatEltTy == MVT::i128) ? LoongArchISD::XVREPLVE0Q
|
||||||
|
: LoongArchISD::XVREPLVE0,
|
||||||
|
DL, SplatTy, SrcVec);
|
||||||
|
} else {
|
||||||
|
SplatVec = DAG.getNode(LoongArchISD::VREPLVEI, DL, SplatTy, SrcVec,
|
||||||
|
DAG.getConstant(0, DL, Subtarget.getGRLenVT()));
|
||||||
|
}
|
||||||
|
|
||||||
|
return DAG.getBitcast(ResTy, SplatVec);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Use INSERT_VECTOR_ELT operations rather than expand to stores, because
|
||||||
|
// using memory operations is much lower.
|
||||||
|
EVT VecTy = ResTy;
|
||||||
|
unsigned VecNumElts = NumElts;
|
||||||
|
|
||||||
|
// Split the 256-bits vector and fill them separately, concat the two parts
|
||||||
|
// to get the result vector.
|
||||||
|
if (Is256Vec) {
|
||||||
|
VecTy = ResTy.getHalfNumVectorElementsVT(*DAG.getContext());
|
||||||
|
VecNumElts = NumElts / 2;
|
||||||
|
}
|
||||||
|
|
||||||
|
SDValue Vector = DAG.getUNDEF(VecTy);
|
||||||
|
SDValue Op0 = Node->getOperand(0);
|
||||||
|
if (!Op0.isUndef())
|
||||||
|
Vector = DAG.getNode(ISD::SCALAR_TO_VECTOR, DL, VecTy, Op0);
|
||||||
|
for (unsigned i = 1; i < VecNumElts; ++i) {
|
||||||
SDValue Opi = Node->getOperand(i);
|
SDValue Opi = Node->getOperand(i);
|
||||||
if (Opi.isUndef())
|
if (Opi.isUndef())
|
||||||
continue;
|
continue;
|
||||||
Vector = DAG.getNode(ISD::INSERT_VECTOR_ELT, DL, ResTy, Vector, Opi,
|
Vector = DAG.getNode(ISD::INSERT_VECTOR_ELT, DL, VecTy, Vector, Opi,
|
||||||
DAG.getConstant(i, DL, Subtarget.getGRLenVT()));
|
DAG.getConstant(i, DL, Subtarget.getGRLenVT()));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (Is128Vec)
|
||||||
return Vector;
|
return Vector;
|
||||||
|
|
||||||
|
SDValue VectorHi = DAG.getUNDEF(VecTy);
|
||||||
|
SDValue OpHi0 = Node->getOperand(VecNumElts);
|
||||||
|
if (!OpHi0.isUndef())
|
||||||
|
VectorHi = DAG.getNode(ISD::SCALAR_TO_VECTOR, DL, VecTy, OpHi0);
|
||||||
|
for (unsigned i = VecNumElts + 1; i < NumElts; ++i) {
|
||||||
|
SDValue Opi = Node->getOperand(i);
|
||||||
|
if (Opi.isUndef())
|
||||||
|
continue;
|
||||||
|
VectorHi = DAG.getNode(
|
||||||
|
ISD::INSERT_VECTOR_ELT, DL, VecTy, VectorHi, Opi,
|
||||||
|
DAG.getConstant(i - VecNumElts, DL, Subtarget.getGRLenVT()));
|
||||||
|
}
|
||||||
|
|
||||||
|
return DAG.getNode(ISD::CONCAT_VECTORS, DL, ResTy, Vector, VectorHi);
|
||||||
}
|
}
|
||||||
|
|
||||||
return SDValue();
|
return SDValue();
|
||||||
@ -6637,6 +6721,8 @@ const char *LoongArchTargetLowering::getTargetNodeName(unsigned Opcode) const {
|
|||||||
NODE_NAME_CASE(VREPLVEI)
|
NODE_NAME_CASE(VREPLVEI)
|
||||||
NODE_NAME_CASE(VREPLGR2VR)
|
NODE_NAME_CASE(VREPLGR2VR)
|
||||||
NODE_NAME_CASE(XVPERMI)
|
NODE_NAME_CASE(XVPERMI)
|
||||||
|
NODE_NAME_CASE(XVREPLVE0)
|
||||||
|
NODE_NAME_CASE(XVREPLVE0Q)
|
||||||
NODE_NAME_CASE(VPICK_SEXT_ELT)
|
NODE_NAME_CASE(VPICK_SEXT_ELT)
|
||||||
NODE_NAME_CASE(VPICK_ZEXT_ELT)
|
NODE_NAME_CASE(VPICK_ZEXT_ELT)
|
||||||
NODE_NAME_CASE(VREPLVE)
|
NODE_NAME_CASE(VREPLVE)
|
||||||
|
@ -141,6 +141,8 @@ enum NodeType : unsigned {
|
|||||||
VREPLVEI,
|
VREPLVEI,
|
||||||
VREPLGR2VR,
|
VREPLGR2VR,
|
||||||
XVPERMI,
|
XVPERMI,
|
||||||
|
XVREPLVE0,
|
||||||
|
XVREPLVE0Q,
|
||||||
|
|
||||||
// Extended vector element extraction
|
// Extended vector element extraction
|
||||||
VPICK_SEXT_ELT,
|
VPICK_SEXT_ELT,
|
||||||
|
@ -10,8 +10,13 @@
|
|||||||
//
|
//
|
||||||
//===----------------------------------------------------------------------===//
|
//===----------------------------------------------------------------------===//
|
||||||
|
|
||||||
|
def SDT_LoongArchXVREPLVE0 : SDTypeProfile<1, 1, [SDTCisVec<0>,
|
||||||
|
SDTCisSameAs<0, 1>]>;
|
||||||
|
|
||||||
// Target nodes.
|
// Target nodes.
|
||||||
def loongarch_xvpermi: SDNode<"LoongArchISD::XVPERMI", SDT_LoongArchV1RUimm>;
|
def loongarch_xvpermi: SDNode<"LoongArchISD::XVPERMI", SDT_LoongArchV1RUimm>;
|
||||||
|
def loongarch_xvreplve0: SDNode<"LoongArchISD::XVREPLVE0", SDT_LoongArchXVREPLVE0>;
|
||||||
|
def loongarch_xvreplve0q: SDNode<"LoongArchISD::XVREPLVE0Q", SDT_LoongArchXVREPLVE0>;
|
||||||
def loongarch_xvmskltz: SDNode<"LoongArchISD::XVMSKLTZ", SDT_LoongArchVMSKCOND>;
|
def loongarch_xvmskltz: SDNode<"LoongArchISD::XVMSKLTZ", SDT_LoongArchVMSKCOND>;
|
||||||
def loongarch_xvmskgez: SDNode<"LoongArchISD::XVMSKGEZ", SDT_LoongArchVMSKCOND>;
|
def loongarch_xvmskgez: SDNode<"LoongArchISD::XVMSKGEZ", SDT_LoongArchVMSKCOND>;
|
||||||
def loongarch_xvmskeqz: SDNode<"LoongArchISD::XVMSKEQZ", SDT_LoongArchVMSKCOND>;
|
def loongarch_xvmskeqz: SDNode<"LoongArchISD::XVMSKEQZ", SDT_LoongArchVMSKCOND>;
|
||||||
@ -1852,11 +1857,26 @@ def : Pat<(loongarch_xvpermi v4i64:$xj, immZExt8: $ui8),
|
|||||||
def : Pat<(loongarch_xvpermi v4f64:$xj, immZExt8: $ui8),
|
def : Pat<(loongarch_xvpermi v4f64:$xj, immZExt8: $ui8),
|
||||||
(XVPERMI_D v4f64:$xj, immZExt8: $ui8)>;
|
(XVPERMI_D v4f64:$xj, immZExt8: $ui8)>;
|
||||||
|
|
||||||
// XVREPLVE0_{W/D}
|
// XVREPLVE0_{B/H/W/D/Q}
|
||||||
|
def : Pat<(loongarch_xvreplve0 v32i8:$xj),
|
||||||
|
(XVREPLVE0_B v32i8:$xj)>;
|
||||||
|
def : Pat<(loongarch_xvreplve0 v16i16:$xj),
|
||||||
|
(XVREPLVE0_H v16i16:$xj)>;
|
||||||
|
def : Pat<(loongarch_xvreplve0 v8i32:$xj),
|
||||||
|
(XVREPLVE0_W v8i32:$xj)>;
|
||||||
|
def : Pat<(loongarch_xvreplve0 v4i64:$xj),
|
||||||
|
(XVREPLVE0_D v4i64:$xj)>;
|
||||||
|
def : Pat<(loongarch_xvreplve0 v8f32:$xj),
|
||||||
|
(XVREPLVE0_W v8f32:$xj)>;
|
||||||
|
def : Pat<(loongarch_xvreplve0 v4f64:$xj),
|
||||||
|
(XVREPLVE0_D v4f64:$xj)>;
|
||||||
def : Pat<(lasxsplatf32 FPR32:$fj),
|
def : Pat<(lasxsplatf32 FPR32:$fj),
|
||||||
(XVREPLVE0_W (SUBREG_TO_REG (i64 0), FPR32:$fj, sub_32))>;
|
(XVREPLVE0_W (SUBREG_TO_REG (i64 0), FPR32:$fj, sub_32))>;
|
||||||
def : Pat<(lasxsplatf64 FPR64:$fj),
|
def : Pat<(lasxsplatf64 FPR64:$fj),
|
||||||
(XVREPLVE0_D (SUBREG_TO_REG (i64 0), FPR64:$fj, sub_64))>;
|
(XVREPLVE0_D (SUBREG_TO_REG (i64 0), FPR64:$fj, sub_64))>;
|
||||||
|
foreach vt = [v32i8, v16i16, v8i32, v4i64, v8f32, v4f64] in
|
||||||
|
def : Pat<(vt (loongarch_xvreplve0q LASX256:$xj)),
|
||||||
|
(XVREPLVE0_Q LASX256:$xj)>;
|
||||||
|
|
||||||
// VSTELM
|
// VSTELM
|
||||||
defm : VstelmPat<truncstorei8, v32i8, XVSTELM_B, simm8, uimm5>;
|
defm : VstelmPat<truncstorei8, v32i8, XVSTELM_B, simm8, uimm5>;
|
||||||
|
@ -7,18 +7,19 @@ declare <32 x i8> @llvm.bitreverse.v32i8(<32 x i8>)
|
|||||||
define <32 x i8> @test_bitreverse_v32i8(<32 x i8> %a) nounwind {
|
define <32 x i8> @test_bitreverse_v32i8(<32 x i8> %a) nounwind {
|
||||||
; CHECK-LABEL: test_bitreverse_v32i8:
|
; CHECK-LABEL: test_bitreverse_v32i8:
|
||||||
; CHECK: # %bb.0:
|
; CHECK: # %bb.0:
|
||||||
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 0
|
|
||||||
; CHECK-NEXT: bitrev.8b $a0, $a0
|
|
||||||
; CHECK-NEXT: xvinsgr2vr.d $xr1, $a0, 0
|
|
||||||
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 1
|
|
||||||
; CHECK-NEXT: bitrev.8b $a0, $a0
|
|
||||||
; CHECK-NEXT: xvinsgr2vr.d $xr1, $a0, 1
|
|
||||||
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 2
|
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 2
|
||||||
; CHECK-NEXT: bitrev.8b $a0, $a0
|
; CHECK-NEXT: bitrev.8b $a0, $a0
|
||||||
; CHECK-NEXT: xvinsgr2vr.d $xr1, $a0, 2
|
; CHECK-NEXT: vinsgr2vr.d $vr2, $a0, 0
|
||||||
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 3
|
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 3
|
||||||
; CHECK-NEXT: bitrev.8b $a0, $a0
|
; CHECK-NEXT: bitrev.8b $a0, $a0
|
||||||
; CHECK-NEXT: xvinsgr2vr.d $xr1, $a0, 3
|
; CHECK-NEXT: vinsgr2vr.d $vr2, $a0, 1
|
||||||
|
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 0
|
||||||
|
; CHECK-NEXT: bitrev.8b $a0, $a0
|
||||||
|
; CHECK-NEXT: vinsgr2vr.d $vr1, $a0, 0
|
||||||
|
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 1
|
||||||
|
; CHECK-NEXT: bitrev.8b $a0, $a0
|
||||||
|
; CHECK-NEXT: vinsgr2vr.d $vr1, $a0, 1
|
||||||
|
; CHECK-NEXT: xvpermi.q $xr1, $xr2, 2
|
||||||
; CHECK-NEXT: xvori.b $xr0, $xr1, 0
|
; CHECK-NEXT: xvori.b $xr0, $xr1, 0
|
||||||
; CHECK-NEXT: ret
|
; CHECK-NEXT: ret
|
||||||
%b = call <32 x i8> @llvm.bitreverse.v32i8(<32 x i8> %a)
|
%b = call <32 x i8> @llvm.bitreverse.v32i8(<32 x i8> %a)
|
||||||
@ -30,19 +31,20 @@ declare <16 x i16> @llvm.bitreverse.v16i16(<16 x i16>)
|
|||||||
define <16 x i16> @test_bitreverse_v16i16(<16 x i16> %a) nounwind {
|
define <16 x i16> @test_bitreverse_v16i16(<16 x i16> %a) nounwind {
|
||||||
; CHECK-LABEL: test_bitreverse_v16i16:
|
; CHECK-LABEL: test_bitreverse_v16i16:
|
||||||
; CHECK: # %bb.0:
|
; CHECK: # %bb.0:
|
||||||
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 0
|
|
||||||
; CHECK-NEXT: bitrev.d $a0, $a0
|
|
||||||
; CHECK-NEXT: xvinsgr2vr.d $xr1, $a0, 0
|
|
||||||
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 1
|
|
||||||
; CHECK-NEXT: bitrev.d $a0, $a0
|
|
||||||
; CHECK-NEXT: xvinsgr2vr.d $xr1, $a0, 1
|
|
||||||
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 2
|
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 2
|
||||||
; CHECK-NEXT: bitrev.d $a0, $a0
|
; CHECK-NEXT: bitrev.d $a0, $a0
|
||||||
; CHECK-NEXT: xvinsgr2vr.d $xr1, $a0, 2
|
; CHECK-NEXT: vinsgr2vr.d $vr1, $a0, 0
|
||||||
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 3
|
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 3
|
||||||
; CHECK-NEXT: bitrev.d $a0, $a0
|
; CHECK-NEXT: bitrev.d $a0, $a0
|
||||||
; CHECK-NEXT: xvinsgr2vr.d $xr1, $a0, 3
|
; CHECK-NEXT: vinsgr2vr.d $vr1, $a0, 1
|
||||||
; CHECK-NEXT: xvshuf4i.h $xr0, $xr1, 27
|
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 0
|
||||||
|
; CHECK-NEXT: bitrev.d $a0, $a0
|
||||||
|
; CHECK-NEXT: vinsgr2vr.d $vr2, $a0, 0
|
||||||
|
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 1
|
||||||
|
; CHECK-NEXT: bitrev.d $a0, $a0
|
||||||
|
; CHECK-NEXT: vinsgr2vr.d $vr2, $a0, 1
|
||||||
|
; CHECK-NEXT: xvpermi.q $xr2, $xr1, 2
|
||||||
|
; CHECK-NEXT: xvshuf4i.h $xr0, $xr2, 27
|
||||||
; CHECK-NEXT: ret
|
; CHECK-NEXT: ret
|
||||||
%b = call <16 x i16> @llvm.bitreverse.v16i16(<16 x i16> %a)
|
%b = call <16 x i16> @llvm.bitreverse.v16i16(<16 x i16> %a)
|
||||||
ret <16 x i16> %b
|
ret <16 x i16> %b
|
||||||
@ -53,19 +55,20 @@ declare <8 x i32> @llvm.bitreverse.v8i32(<8 x i32>)
|
|||||||
define <8 x i32> @test_bitreverse_v8i32(<8 x i32> %a) nounwind {
|
define <8 x i32> @test_bitreverse_v8i32(<8 x i32> %a) nounwind {
|
||||||
; CHECK-LABEL: test_bitreverse_v8i32:
|
; CHECK-LABEL: test_bitreverse_v8i32:
|
||||||
; CHECK: # %bb.0:
|
; CHECK: # %bb.0:
|
||||||
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 0
|
|
||||||
; CHECK-NEXT: bitrev.d $a0, $a0
|
|
||||||
; CHECK-NEXT: xvinsgr2vr.d $xr1, $a0, 0
|
|
||||||
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 1
|
|
||||||
; CHECK-NEXT: bitrev.d $a0, $a0
|
|
||||||
; CHECK-NEXT: xvinsgr2vr.d $xr1, $a0, 1
|
|
||||||
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 2
|
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 2
|
||||||
; CHECK-NEXT: bitrev.d $a0, $a0
|
; CHECK-NEXT: bitrev.d $a0, $a0
|
||||||
; CHECK-NEXT: xvinsgr2vr.d $xr1, $a0, 2
|
; CHECK-NEXT: vinsgr2vr.d $vr1, $a0, 0
|
||||||
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 3
|
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 3
|
||||||
; CHECK-NEXT: bitrev.d $a0, $a0
|
; CHECK-NEXT: bitrev.d $a0, $a0
|
||||||
; CHECK-NEXT: xvinsgr2vr.d $xr1, $a0, 3
|
; CHECK-NEXT: vinsgr2vr.d $vr1, $a0, 1
|
||||||
; CHECK-NEXT: xvshuf4i.w $xr0, $xr1, 177
|
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 0
|
||||||
|
; CHECK-NEXT: bitrev.d $a0, $a0
|
||||||
|
; CHECK-NEXT: vinsgr2vr.d $vr2, $a0, 0
|
||||||
|
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 1
|
||||||
|
; CHECK-NEXT: bitrev.d $a0, $a0
|
||||||
|
; CHECK-NEXT: vinsgr2vr.d $vr2, $a0, 1
|
||||||
|
; CHECK-NEXT: xvpermi.q $xr2, $xr1, 2
|
||||||
|
; CHECK-NEXT: xvshuf4i.w $xr0, $xr2, 177
|
||||||
; CHECK-NEXT: ret
|
; CHECK-NEXT: ret
|
||||||
%b = call <8 x i32> @llvm.bitreverse.v8i32(<8 x i32> %a)
|
%b = call <8 x i32> @llvm.bitreverse.v8i32(<8 x i32> %a)
|
||||||
ret <8 x i32> %b
|
ret <8 x i32> %b
|
||||||
@ -76,18 +79,19 @@ declare <4 x i64> @llvm.bitreverse.v4i64(<4 x i64>)
|
|||||||
define <4 x i64> @test_bitreverse_v4i64(<4 x i64> %a) nounwind {
|
define <4 x i64> @test_bitreverse_v4i64(<4 x i64> %a) nounwind {
|
||||||
; CHECK-LABEL: test_bitreverse_v4i64:
|
; CHECK-LABEL: test_bitreverse_v4i64:
|
||||||
; CHECK: # %bb.0:
|
; CHECK: # %bb.0:
|
||||||
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 0
|
|
||||||
; CHECK-NEXT: bitrev.d $a0, $a0
|
|
||||||
; CHECK-NEXT: xvinsgr2vr.d $xr1, $a0, 0
|
|
||||||
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 1
|
|
||||||
; CHECK-NEXT: bitrev.d $a0, $a0
|
|
||||||
; CHECK-NEXT: xvinsgr2vr.d $xr1, $a0, 1
|
|
||||||
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 2
|
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 2
|
||||||
; CHECK-NEXT: bitrev.d $a0, $a0
|
; CHECK-NEXT: bitrev.d $a0, $a0
|
||||||
; CHECK-NEXT: xvinsgr2vr.d $xr1, $a0, 2
|
; CHECK-NEXT: vinsgr2vr.d $vr2, $a0, 0
|
||||||
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 3
|
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 3
|
||||||
; CHECK-NEXT: bitrev.d $a0, $a0
|
; CHECK-NEXT: bitrev.d $a0, $a0
|
||||||
; CHECK-NEXT: xvinsgr2vr.d $xr1, $a0, 3
|
; CHECK-NEXT: vinsgr2vr.d $vr2, $a0, 1
|
||||||
|
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 0
|
||||||
|
; CHECK-NEXT: bitrev.d $a0, $a0
|
||||||
|
; CHECK-NEXT: vinsgr2vr.d $vr1, $a0, 0
|
||||||
|
; CHECK-NEXT: xvpickve2gr.d $a0, $xr0, 1
|
||||||
|
; CHECK-NEXT: bitrev.d $a0, $a0
|
||||||
|
; CHECK-NEXT: vinsgr2vr.d $vr1, $a0, 1
|
||||||
|
; CHECK-NEXT: xvpermi.q $xr1, $xr2, 2
|
||||||
; CHECK-NEXT: xvori.b $xr0, $xr1, 0
|
; CHECK-NEXT: xvori.b $xr0, $xr1, 0
|
||||||
; CHECK-NEXT: ret
|
; CHECK-NEXT: ret
|
||||||
%b = call <4 x i64> @llvm.bitreverse.v4i64(<4 x i64> %a)
|
%b = call <4 x i64> @llvm.bitreverse.v4i64(<4 x i64> %a)
|
||||||
|
File diff suppressed because it is too large
Load Diff
@ -6,91 +6,91 @@ declare <8 x float> @llvm.powi.v8f32.i32(<8 x float>, i32)
|
|||||||
define <8 x float> @powi_v8f32(<8 x float> %va, i32 %b) nounwind {
|
define <8 x float> @powi_v8f32(<8 x float> %va, i32 %b) nounwind {
|
||||||
; CHECK-LABEL: powi_v8f32:
|
; CHECK-LABEL: powi_v8f32:
|
||||||
; CHECK: # %bb.0: # %entry
|
; CHECK: # %bb.0: # %entry
|
||||||
; CHECK-NEXT: addi.d $sp, $sp, -96
|
; CHECK-NEXT: addi.d $sp, $sp, -128
|
||||||
; CHECK-NEXT: st.d $ra, $sp, 88 # 8-byte Folded Spill
|
; CHECK-NEXT: st.d $ra, $sp, 120 # 8-byte Folded Spill
|
||||||
; CHECK-NEXT: st.d $fp, $sp, 80 # 8-byte Folded Spill
|
; CHECK-NEXT: st.d $fp, $sp, 112 # 8-byte Folded Spill
|
||||||
; CHECK-NEXT: xvst $xr0, $sp, 16 # 32-byte Folded Spill
|
; CHECK-NEXT: xvst $xr0, $sp, 80 # 32-byte Folded Spill
|
||||||
; CHECK-NEXT: addi.w $fp, $a0, 0
|
; CHECK-NEXT: addi.w $fp, $a0, 0
|
||||||
; CHECK-NEXT: xvpickve.w $xr0, $xr0, 1
|
; CHECK-NEXT: xvpickve.w $xr0, $xr0, 5
|
||||||
; CHECK-NEXT: # kill: def $f0 killed $f0 killed $xr0
|
; CHECK-NEXT: # kill: def $f0 killed $f0 killed $xr0
|
||||||
; CHECK-NEXT: move $a0, $fp
|
; CHECK-NEXT: move $a0, $fp
|
||||||
; CHECK-NEXT: pcaddu18i $ra, %call36(__powisf2)
|
; CHECK-NEXT: pcaddu18i $ra, %call36(__powisf2)
|
||||||
; CHECK-NEXT: jirl $ra, $ra, 0
|
; CHECK-NEXT: jirl $ra, $ra, 0
|
||||||
; CHECK-NEXT: # kill: def $f0 killed $f0 def $xr0
|
; CHECK-NEXT: # kill: def $f0 killed $f0 def $vr0
|
||||||
; CHECK-NEXT: xvst $xr0, $sp, 48 # 32-byte Folded Spill
|
; CHECK-NEXT: vst $vr0, $sp, 48 # 16-byte Folded Spill
|
||||||
; CHECK-NEXT: xvld $xr0, $sp, 16 # 32-byte Folded Reload
|
; CHECK-NEXT: xvld $xr0, $sp, 80 # 32-byte Folded Reload
|
||||||
; CHECK-NEXT: xvpickve.w $xr0, $xr0, 0
|
|
||||||
; CHECK-NEXT: # kill: def $f0 killed $f0 killed $xr0
|
|
||||||
; CHECK-NEXT: move $a0, $fp
|
|
||||||
; CHECK-NEXT: pcaddu18i $ra, %call36(__powisf2)
|
|
||||||
; CHECK-NEXT: jirl $ra, $ra, 0
|
|
||||||
; CHECK-NEXT: # kill: def $f0 killed $f0 def $xr0
|
|
||||||
; CHECK-NEXT: xvld $xr1, $sp, 48 # 32-byte Folded Reload
|
|
||||||
; CHECK-NEXT: xvinsve0.w $xr0, $xr1, 1
|
|
||||||
; CHECK-NEXT: xvst $xr0, $sp, 48 # 32-byte Folded Spill
|
|
||||||
; CHECK-NEXT: xvld $xr0, $sp, 16 # 32-byte Folded Reload
|
|
||||||
; CHECK-NEXT: xvpickve.w $xr0, $xr0, 2
|
|
||||||
; CHECK-NEXT: # kill: def $f0 killed $f0 killed $xr0
|
|
||||||
; CHECK-NEXT: move $a0, $fp
|
|
||||||
; CHECK-NEXT: pcaddu18i $ra, %call36(__powisf2)
|
|
||||||
; CHECK-NEXT: jirl $ra, $ra, 0
|
|
||||||
; CHECK-NEXT: # kill: def $f0 killed $f0 def $xr0
|
|
||||||
; CHECK-NEXT: xvld $xr1, $sp, 48 # 32-byte Folded Reload
|
|
||||||
; CHECK-NEXT: xvinsve0.w $xr1, $xr0, 2
|
|
||||||
; CHECK-NEXT: xvst $xr1, $sp, 48 # 32-byte Folded Spill
|
|
||||||
; CHECK-NEXT: xvld $xr0, $sp, 16 # 32-byte Folded Reload
|
|
||||||
; CHECK-NEXT: xvpickve.w $xr0, $xr0, 3
|
|
||||||
; CHECK-NEXT: # kill: def $f0 killed $f0 killed $xr0
|
|
||||||
; CHECK-NEXT: move $a0, $fp
|
|
||||||
; CHECK-NEXT: pcaddu18i $ra, %call36(__powisf2)
|
|
||||||
; CHECK-NEXT: jirl $ra, $ra, 0
|
|
||||||
; CHECK-NEXT: # kill: def $f0 killed $f0 def $xr0
|
|
||||||
; CHECK-NEXT: xvld $xr1, $sp, 48 # 32-byte Folded Reload
|
|
||||||
; CHECK-NEXT: xvinsve0.w $xr1, $xr0, 3
|
|
||||||
; CHECK-NEXT: xvst $xr1, $sp, 48 # 32-byte Folded Spill
|
|
||||||
; CHECK-NEXT: xvld $xr0, $sp, 16 # 32-byte Folded Reload
|
|
||||||
; CHECK-NEXT: xvpickve.w $xr0, $xr0, 4
|
; CHECK-NEXT: xvpickve.w $xr0, $xr0, 4
|
||||||
; CHECK-NEXT: # kill: def $f0 killed $f0 killed $xr0
|
; CHECK-NEXT: # kill: def $f0 killed $f0 killed $xr0
|
||||||
; CHECK-NEXT: move $a0, $fp
|
; CHECK-NEXT: move $a0, $fp
|
||||||
; CHECK-NEXT: pcaddu18i $ra, %call36(__powisf2)
|
; CHECK-NEXT: pcaddu18i $ra, %call36(__powisf2)
|
||||||
; CHECK-NEXT: jirl $ra, $ra, 0
|
; CHECK-NEXT: jirl $ra, $ra, 0
|
||||||
; CHECK-NEXT: # kill: def $f0 killed $f0 def $xr0
|
; CHECK-NEXT: # kill: def $f0 killed $f0 def $xr0
|
||||||
; CHECK-NEXT: xvld $xr1, $sp, 48 # 32-byte Folded Reload
|
; CHECK-NEXT: vld $vr1, $sp, 48 # 16-byte Folded Reload
|
||||||
; CHECK-NEXT: xvinsve0.w $xr1, $xr0, 4
|
; CHECK-NEXT: vextrins.w $vr0, $vr1, 16
|
||||||
; CHECK-NEXT: xvst $xr1, $sp, 48 # 32-byte Folded Spill
|
; CHECK-NEXT: xvst $xr0, $sp, 48 # 32-byte Folded Spill
|
||||||
; CHECK-NEXT: xvld $xr0, $sp, 16 # 32-byte Folded Reload
|
; CHECK-NEXT: xvld $xr0, $sp, 80 # 32-byte Folded Reload
|
||||||
; CHECK-NEXT: xvpickve.w $xr0, $xr0, 5
|
|
||||||
; CHECK-NEXT: # kill: def $f0 killed $f0 killed $xr0
|
|
||||||
; CHECK-NEXT: move $a0, $fp
|
|
||||||
; CHECK-NEXT: pcaddu18i $ra, %call36(__powisf2)
|
|
||||||
; CHECK-NEXT: jirl $ra, $ra, 0
|
|
||||||
; CHECK-NEXT: # kill: def $f0 killed $f0 def $xr0
|
|
||||||
; CHECK-NEXT: xvld $xr1, $sp, 48 # 32-byte Folded Reload
|
|
||||||
; CHECK-NEXT: xvinsve0.w $xr1, $xr0, 5
|
|
||||||
; CHECK-NEXT: xvst $xr1, $sp, 48 # 32-byte Folded Spill
|
|
||||||
; CHECK-NEXT: xvld $xr0, $sp, 16 # 32-byte Folded Reload
|
|
||||||
; CHECK-NEXT: xvpickve.w $xr0, $xr0, 6
|
; CHECK-NEXT: xvpickve.w $xr0, $xr0, 6
|
||||||
; CHECK-NEXT: # kill: def $f0 killed $f0 killed $xr0
|
; CHECK-NEXT: # kill: def $f0 killed $f0 killed $xr0
|
||||||
; CHECK-NEXT: move $a0, $fp
|
; CHECK-NEXT: move $a0, $fp
|
||||||
; CHECK-NEXT: pcaddu18i $ra, %call36(__powisf2)
|
; CHECK-NEXT: pcaddu18i $ra, %call36(__powisf2)
|
||||||
; CHECK-NEXT: jirl $ra, $ra, 0
|
; CHECK-NEXT: jirl $ra, $ra, 0
|
||||||
; CHECK-NEXT: # kill: def $f0 killed $f0 def $xr0
|
; CHECK-NEXT: # kill: def $f0 killed $f0 def $vr0
|
||||||
; CHECK-NEXT: xvld $xr1, $sp, 48 # 32-byte Folded Reload
|
; CHECK-NEXT: xvld $xr1, $sp, 48 # 32-byte Folded Reload
|
||||||
; CHECK-NEXT: xvinsve0.w $xr1, $xr0, 6
|
; CHECK-NEXT: vextrins.w $vr1, $vr0, 32
|
||||||
; CHECK-NEXT: xvst $xr1, $sp, 48 # 32-byte Folded Spill
|
; CHECK-NEXT: xvst $xr1, $sp, 48 # 32-byte Folded Spill
|
||||||
; CHECK-NEXT: xvld $xr0, $sp, 16 # 32-byte Folded Reload
|
; CHECK-NEXT: xvld $xr0, $sp, 80 # 32-byte Folded Reload
|
||||||
; CHECK-NEXT: xvpickve.w $xr0, $xr0, 7
|
; CHECK-NEXT: xvpickve.w $xr0, $xr0, 7
|
||||||
; CHECK-NEXT: # kill: def $f0 killed $f0 killed $xr0
|
; CHECK-NEXT: # kill: def $f0 killed $f0 killed $xr0
|
||||||
; CHECK-NEXT: move $a0, $fp
|
; CHECK-NEXT: move $a0, $fp
|
||||||
; CHECK-NEXT: pcaddu18i $ra, %call36(__powisf2)
|
; CHECK-NEXT: pcaddu18i $ra, %call36(__powisf2)
|
||||||
; CHECK-NEXT: jirl $ra, $ra, 0
|
; CHECK-NEXT: jirl $ra, $ra, 0
|
||||||
; CHECK-NEXT: # kill: def $f0 killed $f0 def $xr0
|
; CHECK-NEXT: # kill: def $f0 killed $f0 def $vr0
|
||||||
; CHECK-NEXT: xvld $xr1, $sp, 48 # 32-byte Folded Reload
|
; CHECK-NEXT: xvld $xr1, $sp, 48 # 32-byte Folded Reload
|
||||||
; CHECK-NEXT: xvinsve0.w $xr1, $xr0, 7
|
; CHECK-NEXT: vextrins.w $vr1, $vr0, 48
|
||||||
|
; CHECK-NEXT: xvst $xr1, $sp, 48 # 32-byte Folded Spill
|
||||||
|
; CHECK-NEXT: xvld $xr0, $sp, 80 # 32-byte Folded Reload
|
||||||
|
; CHECK-NEXT: xvpickve.w $xr0, $xr0, 1
|
||||||
|
; CHECK-NEXT: # kill: def $f0 killed $f0 killed $xr0
|
||||||
|
; CHECK-NEXT: move $a0, $fp
|
||||||
|
; CHECK-NEXT: pcaddu18i $ra, %call36(__powisf2)
|
||||||
|
; CHECK-NEXT: jirl $ra, $ra, 0
|
||||||
|
; CHECK-NEXT: # kill: def $f0 killed $f0 def $vr0
|
||||||
|
; CHECK-NEXT: vst $vr0, $sp, 16 # 16-byte Folded Spill
|
||||||
|
; CHECK-NEXT: xvld $xr0, $sp, 80 # 32-byte Folded Reload
|
||||||
|
; CHECK-NEXT: xvpickve.w $xr0, $xr0, 0
|
||||||
|
; CHECK-NEXT: # kill: def $f0 killed $f0 killed $xr0
|
||||||
|
; CHECK-NEXT: move $a0, $fp
|
||||||
|
; CHECK-NEXT: pcaddu18i $ra, %call36(__powisf2)
|
||||||
|
; CHECK-NEXT: jirl $ra, $ra, 0
|
||||||
|
; CHECK-NEXT: # kill: def $f0 killed $f0 def $xr0
|
||||||
|
; CHECK-NEXT: vld $vr1, $sp, 16 # 16-byte Folded Reload
|
||||||
|
; CHECK-NEXT: vextrins.w $vr0, $vr1, 16
|
||||||
|
; CHECK-NEXT: xvst $xr0, $sp, 16 # 32-byte Folded Spill
|
||||||
|
; CHECK-NEXT: xvld $xr0, $sp, 80 # 32-byte Folded Reload
|
||||||
|
; CHECK-NEXT: xvpickve.w $xr0, $xr0, 2
|
||||||
|
; CHECK-NEXT: # kill: def $f0 killed $f0 killed $xr0
|
||||||
|
; CHECK-NEXT: move $a0, $fp
|
||||||
|
; CHECK-NEXT: pcaddu18i $ra, %call36(__powisf2)
|
||||||
|
; CHECK-NEXT: jirl $ra, $ra, 0
|
||||||
|
; CHECK-NEXT: # kill: def $f0 killed $f0 def $vr0
|
||||||
|
; CHECK-NEXT: xvld $xr1, $sp, 16 # 32-byte Folded Reload
|
||||||
|
; CHECK-NEXT: vextrins.w $vr1, $vr0, 32
|
||||||
|
; CHECK-NEXT: xvst $xr1, $sp, 16 # 32-byte Folded Spill
|
||||||
|
; CHECK-NEXT: xvld $xr0, $sp, 80 # 32-byte Folded Reload
|
||||||
|
; CHECK-NEXT: xvpickve.w $xr0, $xr0, 3
|
||||||
|
; CHECK-NEXT: # kill: def $f0 killed $f0 killed $xr0
|
||||||
|
; CHECK-NEXT: move $a0, $fp
|
||||||
|
; CHECK-NEXT: pcaddu18i $ra, %call36(__powisf2)
|
||||||
|
; CHECK-NEXT: jirl $ra, $ra, 0
|
||||||
|
; CHECK-NEXT: # kill: def $f0 killed $f0 def $vr0
|
||||||
|
; CHECK-NEXT: xvld $xr1, $sp, 16 # 32-byte Folded Reload
|
||||||
|
; CHECK-NEXT: vextrins.w $vr1, $vr0, 48
|
||||||
|
; CHECK-NEXT: xvld $xr0, $sp, 48 # 32-byte Folded Reload
|
||||||
|
; CHECK-NEXT: xvpermi.q $xr1, $xr0, 2
|
||||||
; CHECK-NEXT: xvori.b $xr0, $xr1, 0
|
; CHECK-NEXT: xvori.b $xr0, $xr1, 0
|
||||||
; CHECK-NEXT: ld.d $fp, $sp, 80 # 8-byte Folded Reload
|
; CHECK-NEXT: ld.d $fp, $sp, 112 # 8-byte Folded Reload
|
||||||
; CHECK-NEXT: ld.d $ra, $sp, 88 # 8-byte Folded Reload
|
; CHECK-NEXT: ld.d $ra, $sp, 120 # 8-byte Folded Reload
|
||||||
; CHECK-NEXT: addi.d $sp, $sp, 96
|
; CHECK-NEXT: addi.d $sp, $sp, 128
|
||||||
; CHECK-NEXT: ret
|
; CHECK-NEXT: ret
|
||||||
entry:
|
entry:
|
||||||
%res = call <8 x float> @llvm.powi.v8f32.i32(<8 x float> %va, i32 %b)
|
%res = call <8 x float> @llvm.powi.v8f32.i32(<8 x float> %va, i32 %b)
|
||||||
@ -102,51 +102,50 @@ declare <4 x double> @llvm.powi.v4f64.i32(<4 x double>, i32)
|
|||||||
define <4 x double> @powi_v4f64(<4 x double> %va, i32 %b) nounwind {
|
define <4 x double> @powi_v4f64(<4 x double> %va, i32 %b) nounwind {
|
||||||
; CHECK-LABEL: powi_v4f64:
|
; CHECK-LABEL: powi_v4f64:
|
||||||
; CHECK: # %bb.0: # %entry
|
; CHECK: # %bb.0: # %entry
|
||||||
; CHECK-NEXT: addi.d $sp, $sp, -96
|
; CHECK-NEXT: addi.d $sp, $sp, -112
|
||||||
; CHECK-NEXT: st.d $ra, $sp, 88 # 8-byte Folded Spill
|
; CHECK-NEXT: st.d $ra, $sp, 104 # 8-byte Folded Spill
|
||||||
; CHECK-NEXT: st.d $fp, $sp, 80 # 8-byte Folded Spill
|
; CHECK-NEXT: st.d $fp, $sp, 96 # 8-byte Folded Spill
|
||||||
; CHECK-NEXT: xvst $xr0, $sp, 48 # 32-byte Folded Spill
|
; CHECK-NEXT: xvst $xr0, $sp, 64 # 32-byte Folded Spill
|
||||||
; CHECK-NEXT: addi.w $fp, $a0, 0
|
; CHECK-NEXT: addi.w $fp, $a0, 0
|
||||||
; CHECK-NEXT: xvpickve.d $xr0, $xr0, 1
|
; CHECK-NEXT: xvpickve.d $xr0, $xr0, 3
|
||||||
; CHECK-NEXT: # kill: def $f0_64 killed $f0_64 killed $xr0
|
; CHECK-NEXT: # kill: def $f0_64 killed $f0_64 killed $xr0
|
||||||
; CHECK-NEXT: move $a0, $fp
|
; CHECK-NEXT: move $a0, $fp
|
||||||
; CHECK-NEXT: pcaddu18i $ra, %call36(__powidf2)
|
; CHECK-NEXT: pcaddu18i $ra, %call36(__powidf2)
|
||||||
; CHECK-NEXT: jirl $ra, $ra, 0
|
; CHECK-NEXT: jirl $ra, $ra, 0
|
||||||
; CHECK-NEXT: # kill: def $f0_64 killed $f0_64 def $xr0
|
; CHECK-NEXT: # kill: def $f0_64 killed $f0_64 def $vr0
|
||||||
; CHECK-NEXT: xvst $xr0, $sp, 16 # 32-byte Folded Spill
|
; CHECK-NEXT: vst $vr0, $sp, 32 # 16-byte Folded Spill
|
||||||
; CHECK-NEXT: xvld $xr0, $sp, 48 # 32-byte Folded Reload
|
; CHECK-NEXT: xvld $xr0, $sp, 64 # 32-byte Folded Reload
|
||||||
; CHECK-NEXT: xvpickve.d $xr0, $xr0, 0
|
|
||||||
; CHECK-NEXT: # kill: def $f0_64 killed $f0_64 killed $xr0
|
|
||||||
; CHECK-NEXT: move $a0, $fp
|
|
||||||
; CHECK-NEXT: pcaddu18i $ra, %call36(__powidf2)
|
|
||||||
; CHECK-NEXT: jirl $ra, $ra, 0
|
|
||||||
; CHECK-NEXT: # kill: def $f0_64 killed $f0_64 def $xr0
|
|
||||||
; CHECK-NEXT: xvld $xr1, $sp, 16 # 32-byte Folded Reload
|
|
||||||
; CHECK-NEXT: xvinsve0.d $xr0, $xr1, 1
|
|
||||||
; CHECK-NEXT: xvst $xr0, $sp, 16 # 32-byte Folded Spill
|
|
||||||
; CHECK-NEXT: xvld $xr0, $sp, 48 # 32-byte Folded Reload
|
|
||||||
; CHECK-NEXT: xvpickve.d $xr0, $xr0, 2
|
; CHECK-NEXT: xvpickve.d $xr0, $xr0, 2
|
||||||
; CHECK-NEXT: # kill: def $f0_64 killed $f0_64 killed $xr0
|
; CHECK-NEXT: # kill: def $f0_64 killed $f0_64 killed $xr0
|
||||||
; CHECK-NEXT: move $a0, $fp
|
; CHECK-NEXT: move $a0, $fp
|
||||||
; CHECK-NEXT: pcaddu18i $ra, %call36(__powidf2)
|
; CHECK-NEXT: pcaddu18i $ra, %call36(__powidf2)
|
||||||
; CHECK-NEXT: jirl $ra, $ra, 0
|
; CHECK-NEXT: jirl $ra, $ra, 0
|
||||||
; CHECK-NEXT: # kill: def $f0_64 killed $f0_64 def $xr0
|
; CHECK-NEXT: # kill: def $f0_64 killed $f0_64 def $xr0
|
||||||
; CHECK-NEXT: xvld $xr1, $sp, 16 # 32-byte Folded Reload
|
; CHECK-NEXT: vld $vr1, $sp, 32 # 16-byte Folded Reload
|
||||||
; CHECK-NEXT: xvinsve0.d $xr1, $xr0, 2
|
; CHECK-NEXT: vextrins.d $vr0, $vr1, 16
|
||||||
; CHECK-NEXT: xvst $xr1, $sp, 16 # 32-byte Folded Spill
|
; CHECK-NEXT: xvst $xr0, $sp, 32 # 32-byte Folded Spill
|
||||||
; CHECK-NEXT: xvld $xr0, $sp, 48 # 32-byte Folded Reload
|
; CHECK-NEXT: xvld $xr0, $sp, 64 # 32-byte Folded Reload
|
||||||
; CHECK-NEXT: xvpickve.d $xr0, $xr0, 3
|
; CHECK-NEXT: xvpickve.d $xr0, $xr0, 1
|
||||||
|
; CHECK-NEXT: # kill: def $f0_64 killed $f0_64 killed $xr0
|
||||||
|
; CHECK-NEXT: move $a0, $fp
|
||||||
|
; CHECK-NEXT: pcaddu18i $ra, %call36(__powidf2)
|
||||||
|
; CHECK-NEXT: jirl $ra, $ra, 0
|
||||||
|
; CHECK-NEXT: # kill: def $f0_64 killed $f0_64 def $vr0
|
||||||
|
; CHECK-NEXT: vst $vr0, $sp, 16 # 16-byte Folded Spill
|
||||||
|
; CHECK-NEXT: xvld $xr0, $sp, 64 # 32-byte Folded Reload
|
||||||
|
; CHECK-NEXT: xvpickve.d $xr0, $xr0, 0
|
||||||
; CHECK-NEXT: # kill: def $f0_64 killed $f0_64 killed $xr0
|
; CHECK-NEXT: # kill: def $f0_64 killed $f0_64 killed $xr0
|
||||||
; CHECK-NEXT: move $a0, $fp
|
; CHECK-NEXT: move $a0, $fp
|
||||||
; CHECK-NEXT: pcaddu18i $ra, %call36(__powidf2)
|
; CHECK-NEXT: pcaddu18i $ra, %call36(__powidf2)
|
||||||
; CHECK-NEXT: jirl $ra, $ra, 0
|
; CHECK-NEXT: jirl $ra, $ra, 0
|
||||||
; CHECK-NEXT: # kill: def $f0_64 killed $f0_64 def $xr0
|
; CHECK-NEXT: # kill: def $f0_64 killed $f0_64 def $xr0
|
||||||
; CHECK-NEXT: xvld $xr1, $sp, 16 # 32-byte Folded Reload
|
; CHECK-NEXT: vld $vr1, $sp, 16 # 16-byte Folded Reload
|
||||||
; CHECK-NEXT: xvinsve0.d $xr1, $xr0, 3
|
; CHECK-NEXT: vextrins.d $vr0, $vr1, 16
|
||||||
; CHECK-NEXT: xvori.b $xr0, $xr1, 0
|
; CHECK-NEXT: xvld $xr1, $sp, 32 # 32-byte Folded Reload
|
||||||
; CHECK-NEXT: ld.d $fp, $sp, 80 # 8-byte Folded Reload
|
; CHECK-NEXT: xvpermi.q $xr0, $xr1, 2
|
||||||
; CHECK-NEXT: ld.d $ra, $sp, 88 # 8-byte Folded Reload
|
; CHECK-NEXT: ld.d $fp, $sp, 96 # 8-byte Folded Reload
|
||||||
; CHECK-NEXT: addi.d $sp, $sp, 96
|
; CHECK-NEXT: ld.d $ra, $sp, 104 # 8-byte Folded Reload
|
||||||
|
; CHECK-NEXT: addi.d $sp, $sp, 112
|
||||||
; CHECK-NEXT: ret
|
; CHECK-NEXT: ret
|
||||||
entry:
|
entry:
|
||||||
%res = call <4 x double> @llvm.powi.v4f64.i32(<4 x double> %va, i32 %b)
|
%res = call <4 x double> @llvm.powi.v4f64.i32(<4 x double> %va, i32 %b)
|
||||||
|
@ -6,12 +6,13 @@
|
|||||||
define <4 x double> @shufflevector_v4f64(<4 x double> %a, <4 x double> %b) {
|
define <4 x double> @shufflevector_v4f64(<4 x double> %a, <4 x double> %b) {
|
||||||
; CHECK-LABEL: shufflevector_v4f64:
|
; CHECK-LABEL: shufflevector_v4f64:
|
||||||
; CHECK: # %bb.0: # %entry
|
; CHECK: # %bb.0: # %entry
|
||||||
; CHECK-NEXT: xvpickve.d $xr2, $xr1, 2
|
; CHECK-NEXT: xvpickve.d $xr2, $xr1, 3
|
||||||
; CHECK-NEXT: xvpickve.d $xr3, $xr0, 3
|
; CHECK-NEXT: xvpermi.d $xr3, $xr0, 78
|
||||||
; CHECK-NEXT: xvinsve0.d $xr0, $xr2, 1
|
; CHECK-NEXT: xvrepl128vei.d $xr3, $xr3, 1
|
||||||
; CHECK-NEXT: xvinsve0.d $xr0, $xr3, 2
|
; CHECK-NEXT: vextrins.d $vr3, $vr2, 16
|
||||||
; CHECK-NEXT: xvpickve.d $xr1, $xr1, 3
|
; CHECK-NEXT: xvpickve.d $xr1, $xr1, 2
|
||||||
; CHECK-NEXT: xvinsve0.d $xr0, $xr1, 3
|
; CHECK-NEXT: vextrins.d $vr0, $vr1, 16
|
||||||
|
; CHECK-NEXT: xvpermi.q $xr0, $xr3, 2
|
||||||
; CHECK-NEXT: ret
|
; CHECK-NEXT: ret
|
||||||
entry:
|
entry:
|
||||||
%c = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 0, i32 6, i32 3, i32 7>
|
%c = shufflevector <4 x double> %a, <4 x double> %b, <4 x i32> <i32 0, i32 6, i32 3, i32 7>
|
||||||
|
@ -45,9 +45,8 @@ define void @insert_32xi8_undef(ptr %dst, i8 %in) nounwind {
|
|||||||
define void @insert_32xi8_undef_upper(ptr %dst, i8 %in) nounwind {
|
define void @insert_32xi8_undef_upper(ptr %dst, i8 %in) nounwind {
|
||||||
; CHECK-LABEL: insert_32xi8_undef_upper:
|
; CHECK-LABEL: insert_32xi8_undef_upper:
|
||||||
; CHECK: # %bb.0:
|
; CHECK: # %bb.0:
|
||||||
; CHECK-NEXT: xvreplgr2vr.b $xr0, $a1
|
; CHECK-NEXT: vinsgr2vr.b $vr0, $a1, 6
|
||||||
; CHECK-NEXT: xvpermi.q $xr0, $xr0, 48
|
; CHECK-NEXT: xvpermi.q $xr0, $xr0, 2
|
||||||
; CHECK-NEXT: xvextrins.b $xr0, $xr0, 102
|
|
||||||
; CHECK-NEXT: xvst $xr0, $a0, 0
|
; CHECK-NEXT: xvst $xr0, $a0, 0
|
||||||
; CHECK-NEXT: ret
|
; CHECK-NEXT: ret
|
||||||
%v = insertelement <32 x i8> poison, i8 %in, i32 22
|
%v = insertelement <32 x i8> poison, i8 %in, i32 22
|
||||||
@ -99,9 +98,8 @@ define void @insert_16xi16_undef(ptr %dst, i16 %in) nounwind {
|
|||||||
define void @insert_16xi16_undef_upper(ptr %dst, i16 %in) nounwind {
|
define void @insert_16xi16_undef_upper(ptr %dst, i16 %in) nounwind {
|
||||||
; CHECK-LABEL: insert_16xi16_undef_upper:
|
; CHECK-LABEL: insert_16xi16_undef_upper:
|
||||||
; CHECK: # %bb.0:
|
; CHECK: # %bb.0:
|
||||||
; CHECK-NEXT: xvreplgr2vr.h $xr0, $a1
|
; CHECK-NEXT: vinsgr2vr.h $vr0, $a1, 2
|
||||||
; CHECK-NEXT: xvpermi.q $xr0, $xr0, 48
|
; CHECK-NEXT: xvpermi.q $xr0, $xr0, 2
|
||||||
; CHECK-NEXT: xvextrins.h $xr0, $xr0, 34
|
|
||||||
; CHECK-NEXT: xvst $xr0, $a0, 0
|
; CHECK-NEXT: xvst $xr0, $a0, 0
|
||||||
; CHECK-NEXT: ret
|
; CHECK-NEXT: ret
|
||||||
%v = insertelement <16 x i16> poison, i16 %in, i32 10
|
%v = insertelement <16 x i16> poison, i16 %in, i32 10
|
||||||
|
@ -24,7 +24,7 @@ define <16 x i16> @scalar_to_16xi16(i16 %val) {
|
|||||||
define <8 x i32> @scalar_to_8xi32(i32 %val) {
|
define <8 x i32> @scalar_to_8xi32(i32 %val) {
|
||||||
; CHECK-LABEL: scalar_to_8xi32:
|
; CHECK-LABEL: scalar_to_8xi32:
|
||||||
; CHECK: # %bb.0:
|
; CHECK: # %bb.0:
|
||||||
; CHECK-NEXT: xvinsgr2vr.w $xr0, $a0, 0
|
; CHECK-NEXT: vinsgr2vr.w $vr0, $a0, 0
|
||||||
; CHECK-NEXT: ret
|
; CHECK-NEXT: ret
|
||||||
%ret = insertelement <8 x i32> poison, i32 %val, i32 0
|
%ret = insertelement <8 x i32> poison, i32 %val, i32 0
|
||||||
ret <8 x i32> %ret
|
ret <8 x i32> %ret
|
||||||
@ -33,7 +33,7 @@ define <8 x i32> @scalar_to_8xi32(i32 %val) {
|
|||||||
define <4 x i64> @scalar_to_4xi64(i64 %val) {
|
define <4 x i64> @scalar_to_4xi64(i64 %val) {
|
||||||
; CHECK-LABEL: scalar_to_4xi64:
|
; CHECK-LABEL: scalar_to_4xi64:
|
||||||
; CHECK: # %bb.0:
|
; CHECK: # %bb.0:
|
||||||
; CHECK-NEXT: xvinsgr2vr.d $xr0, $a0, 0
|
; CHECK-NEXT: vinsgr2vr.d $vr0, $a0, 0
|
||||||
; CHECK-NEXT: ret
|
; CHECK-NEXT: ret
|
||||||
%ret = insertelement <4 x i64> poison, i64 %val, i32 0
|
%ret = insertelement <4 x i64> poison, i64 %val, i32 0
|
||||||
ret <4 x i64> %ret
|
ret <4 x i64> %ret
|
||||||
|
@ -338,6 +338,102 @@ entry:
|
|||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
|
define void @buildvector_v16i8_subseq_2(ptr %dst, i8 %a0, i8 %a1, i8 %a2, i8 %a3, i8 %a4, i8 %a5, i8 %a6, i8 %a7) nounwind {
|
||||||
|
; CHECK-LABEL: buildvector_v16i8_subseq_2:
|
||||||
|
; CHECK: # %bb.0: # %entry
|
||||||
|
; CHECK-NEXT: ld.b $t0, $sp, 0
|
||||||
|
; CHECK-NEXT: vinsgr2vr.b $vr0, $a1, 0
|
||||||
|
; CHECK-NEXT: vinsgr2vr.b $vr0, $a2, 1
|
||||||
|
; CHECK-NEXT: vinsgr2vr.b $vr0, $a3, 2
|
||||||
|
; CHECK-NEXT: vinsgr2vr.b $vr0, $a4, 3
|
||||||
|
; CHECK-NEXT: vinsgr2vr.b $vr0, $a5, 4
|
||||||
|
; CHECK-NEXT: vinsgr2vr.b $vr0, $a6, 5
|
||||||
|
; CHECK-NEXT: vinsgr2vr.b $vr0, $a7, 6
|
||||||
|
; CHECK-NEXT: vinsgr2vr.b $vr0, $t0, 7
|
||||||
|
; CHECK-NEXT: vreplvei.d $vr0, $vr0, 0
|
||||||
|
; CHECK-NEXT: vst $vr0, $a0, 0
|
||||||
|
; CHECK-NEXT: ret
|
||||||
|
entry:
|
||||||
|
%ins0 = insertelement <16 x i8> undef, i8 %a0, i32 0
|
||||||
|
%ins1 = insertelement <16 x i8> %ins0, i8 %a1, i32 1
|
||||||
|
%ins2 = insertelement <16 x i8> %ins1, i8 %a2, i32 2
|
||||||
|
%ins3 = insertelement <16 x i8> %ins2, i8 %a3, i32 3
|
||||||
|
%ins4 = insertelement <16 x i8> %ins3, i8 %a4, i32 4
|
||||||
|
%ins5 = insertelement <16 x i8> %ins4, i8 %a5, i32 5
|
||||||
|
%ins6 = insertelement <16 x i8> %ins5, i8 %a6, i32 6
|
||||||
|
%ins7 = insertelement <16 x i8> %ins6, i8 %a7, i32 7
|
||||||
|
%ins8 = insertelement <16 x i8> %ins7, i8 %a0, i32 8
|
||||||
|
%ins9 = insertelement <16 x i8> %ins8, i8 %a1, i32 9
|
||||||
|
%ins10 = insertelement <16 x i8> %ins9, i8 %a2, i32 10
|
||||||
|
%ins11 = insertelement <16 x i8> %ins10, i8 %a3, i32 11
|
||||||
|
%ins12 = insertelement <16 x i8> %ins11, i8 %a4, i32 12
|
||||||
|
%ins13 = insertelement <16 x i8> %ins12, i8 %a5, i32 13
|
||||||
|
%ins14 = insertelement <16 x i8> %ins13, i8 %a6, i32 14
|
||||||
|
%ins15 = insertelement <16 x i8> %ins14, i8 %a7, i32 15
|
||||||
|
store <16 x i8> %ins15, ptr %dst
|
||||||
|
ret void
|
||||||
|
}
|
||||||
|
|
||||||
|
define void @buildvector_v16i8_subseq_4(ptr %dst, i8 %a0, i8 %a1, i8 %a2, i8 %a3) nounwind {
|
||||||
|
; CHECK-LABEL: buildvector_v16i8_subseq_4:
|
||||||
|
; CHECK: # %bb.0: # %entry
|
||||||
|
; CHECK-NEXT: vinsgr2vr.b $vr0, $a1, 0
|
||||||
|
; CHECK-NEXT: vinsgr2vr.b $vr0, $a2, 1
|
||||||
|
; CHECK-NEXT: vinsgr2vr.b $vr0, $a3, 2
|
||||||
|
; CHECK-NEXT: vinsgr2vr.b $vr0, $a4, 3
|
||||||
|
; CHECK-NEXT: vreplvei.w $vr0, $vr0, 0
|
||||||
|
; CHECK-NEXT: vst $vr0, $a0, 0
|
||||||
|
; CHECK-NEXT: ret
|
||||||
|
entry:
|
||||||
|
%ins0 = insertelement <16 x i8> undef, i8 %a0, i32 0
|
||||||
|
%ins1 = insertelement <16 x i8> %ins0, i8 %a1, i32 1
|
||||||
|
%ins2 = insertelement <16 x i8> %ins1, i8 %a2, i32 2
|
||||||
|
%ins3 = insertelement <16 x i8> %ins2, i8 %a3, i32 3
|
||||||
|
%ins4 = insertelement <16 x i8> %ins3, i8 %a0, i32 4
|
||||||
|
%ins5 = insertelement <16 x i8> %ins4, i8 %a1, i32 5
|
||||||
|
%ins6 = insertelement <16 x i8> %ins5, i8 %a2, i32 6
|
||||||
|
%ins7 = insertelement <16 x i8> %ins6, i8 %a3, i32 7
|
||||||
|
%ins8 = insertelement <16 x i8> %ins7, i8 %a0, i32 8
|
||||||
|
%ins9 = insertelement <16 x i8> %ins8, i8 %a1, i32 9
|
||||||
|
%ins10 = insertelement <16 x i8> %ins9, i8 %a2, i32 10
|
||||||
|
%ins11 = insertelement <16 x i8> %ins10, i8 %a3, i32 11
|
||||||
|
%ins12 = insertelement <16 x i8> %ins11, i8 %a0, i32 12
|
||||||
|
%ins13 = insertelement <16 x i8> %ins12, i8 %a1, i32 13
|
||||||
|
%ins14 = insertelement <16 x i8> %ins13, i8 %a2, i32 14
|
||||||
|
%ins15 = insertelement <16 x i8> %ins14, i8 %a3, i32 15
|
||||||
|
store <16 x i8> %ins15, ptr %dst
|
||||||
|
ret void
|
||||||
|
}
|
||||||
|
|
||||||
|
define void @buildvector_v16i8_subseq_8(ptr %dst, i8 %a0, i8 %a1) nounwind {
|
||||||
|
; CHECK-LABEL: buildvector_v16i8_subseq_8:
|
||||||
|
; CHECK: # %bb.0: # %entry
|
||||||
|
; CHECK-NEXT: vinsgr2vr.b $vr0, $a1, 0
|
||||||
|
; CHECK-NEXT: vinsgr2vr.b $vr0, $a2, 1
|
||||||
|
; CHECK-NEXT: vreplvei.h $vr0, $vr0, 0
|
||||||
|
; CHECK-NEXT: vst $vr0, $a0, 0
|
||||||
|
; CHECK-NEXT: ret
|
||||||
|
entry:
|
||||||
|
%ins0 = insertelement <16 x i8> undef, i8 %a0, i32 0
|
||||||
|
%ins1 = insertelement <16 x i8> %ins0, i8 %a1, i32 1
|
||||||
|
%ins2 = insertelement <16 x i8> %ins1, i8 %a0, i32 2
|
||||||
|
%ins3 = insertelement <16 x i8> %ins2, i8 %a1, i32 3
|
||||||
|
%ins4 = insertelement <16 x i8> %ins3, i8 %a0, i32 4
|
||||||
|
%ins5 = insertelement <16 x i8> %ins4, i8 %a1, i32 5
|
||||||
|
%ins6 = insertelement <16 x i8> %ins5, i8 %a0, i32 6
|
||||||
|
%ins7 = insertelement <16 x i8> %ins6, i8 %a1, i32 7
|
||||||
|
%ins8 = insertelement <16 x i8> %ins7, i8 %a0, i32 8
|
||||||
|
%ins9 = insertelement <16 x i8> %ins8, i8 %a1, i32 9
|
||||||
|
%ins10 = insertelement <16 x i8> %ins9, i8 %a0, i32 10
|
||||||
|
%ins11 = insertelement <16 x i8> %ins10, i8 %a1, i32 11
|
||||||
|
%ins12 = insertelement <16 x i8> %ins11, i8 %a0, i32 12
|
||||||
|
%ins13 = insertelement <16 x i8> %ins12, i8 %a1, i32 13
|
||||||
|
%ins14 = insertelement <16 x i8> %ins13, i8 %a0, i32 14
|
||||||
|
%ins15 = insertelement <16 x i8> %ins14, i8 %a1, i32 15
|
||||||
|
store <16 x i8> %ins15, ptr %dst
|
||||||
|
ret void
|
||||||
|
}
|
||||||
|
|
||||||
define void @buildvector_v8i16(ptr %dst, i16 %a0, i16 %a1, i16 %a2, i16 %a3, i16 %a4, i16 %a5, i16 %a6, i16 %a7) nounwind {
|
define void @buildvector_v8i16(ptr %dst, i16 %a0, i16 %a1, i16 %a2, i16 %a3, i16 %a4, i16 %a5, i16 %a6, i16 %a7) nounwind {
|
||||||
; CHECK-LABEL: buildvector_v8i16:
|
; CHECK-LABEL: buildvector_v8i16:
|
||||||
; CHECK: # %bb.0: # %entry
|
; CHECK: # %bb.0: # %entry
|
||||||
@ -410,6 +506,50 @@ entry:
|
|||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
|
define void @buildvector_v8i16_subseq_2(ptr %dst, i16 %a0, i16 %a1, i16 %a2, i16 %a3) nounwind {
|
||||||
|
; CHECK-LABEL: buildvector_v8i16_subseq_2:
|
||||||
|
; CHECK: # %bb.0: # %entry
|
||||||
|
; CHECK-NEXT: vinsgr2vr.h $vr0, $a1, 0
|
||||||
|
; CHECK-NEXT: vinsgr2vr.h $vr0, $a2, 1
|
||||||
|
; CHECK-NEXT: vinsgr2vr.h $vr0, $a3, 2
|
||||||
|
; CHECK-NEXT: vinsgr2vr.h $vr0, $a4, 3
|
||||||
|
; CHECK-NEXT: vreplvei.d $vr0, $vr0, 0
|
||||||
|
; CHECK-NEXT: vst $vr0, $a0, 0
|
||||||
|
; CHECK-NEXT: ret
|
||||||
|
entry:
|
||||||
|
%ins0 = insertelement <8 x i16> undef, i16 %a0, i32 0
|
||||||
|
%ins1 = insertelement <8 x i16> %ins0, i16 %a1, i32 1
|
||||||
|
%ins2 = insertelement <8 x i16> %ins1, i16 %a2, i32 2
|
||||||
|
%ins3 = insertelement <8 x i16> %ins2, i16 %a3, i32 3
|
||||||
|
%ins4 = insertelement <8 x i16> %ins3, i16 %a0, i32 4
|
||||||
|
%ins5 = insertelement <8 x i16> %ins4, i16 %a1, i32 5
|
||||||
|
%ins6 = insertelement <8 x i16> %ins5, i16 %a2, i32 6
|
||||||
|
%ins7 = insertelement <8 x i16> %ins6, i16 %a3, i32 7
|
||||||
|
store <8 x i16> %ins7, ptr %dst
|
||||||
|
ret void
|
||||||
|
}
|
||||||
|
|
||||||
|
define void @buildvector_v8i16_subseq_4(ptr %dst, i16 %a0, i16 %a1) nounwind {
|
||||||
|
; CHECK-LABEL: buildvector_v8i16_subseq_4:
|
||||||
|
; CHECK: # %bb.0: # %entry
|
||||||
|
; CHECK-NEXT: vinsgr2vr.h $vr0, $a1, 0
|
||||||
|
; CHECK-NEXT: vinsgr2vr.h $vr0, $a2, 1
|
||||||
|
; CHECK-NEXT: vreplvei.w $vr0, $vr0, 0
|
||||||
|
; CHECK-NEXT: vst $vr0, $a0, 0
|
||||||
|
; CHECK-NEXT: ret
|
||||||
|
entry:
|
||||||
|
%ins0 = insertelement <8 x i16> undef, i16 %a0, i32 0
|
||||||
|
%ins1 = insertelement <8 x i16> %ins0, i16 %a1, i32 1
|
||||||
|
%ins2 = insertelement <8 x i16> %ins1, i16 %a0, i32 2
|
||||||
|
%ins3 = insertelement <8 x i16> %ins2, i16 %a1, i32 3
|
||||||
|
%ins4 = insertelement <8 x i16> %ins3, i16 %a0, i32 4
|
||||||
|
%ins5 = insertelement <8 x i16> %ins4, i16 %a1, i32 5
|
||||||
|
%ins6 = insertelement <8 x i16> %ins5, i16 %a0, i32 6
|
||||||
|
%ins7 = insertelement <8 x i16> %ins6, i16 %a1, i32 7
|
||||||
|
store <8 x i16> %ins7, ptr %dst
|
||||||
|
ret void
|
||||||
|
}
|
||||||
|
|
||||||
define void @buildvector_v4i32(ptr %dst, i32 %a0, i32 %a1, i32 %a2, i32 %a3) nounwind {
|
define void @buildvector_v4i32(ptr %dst, i32 %a0, i32 %a1, i32 %a2, i32 %a3) nounwind {
|
||||||
; CHECK-LABEL: buildvector_v4i32:
|
; CHECK-LABEL: buildvector_v4i32:
|
||||||
; CHECK: # %bb.0: # %entry
|
; CHECK: # %bb.0: # %entry
|
||||||
@ -462,6 +602,23 @@ entry:
|
|||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
|
define void @buildvector_v4i32_subseq_2(ptr %dst, i32 %a0, i32 %a1) nounwind {
|
||||||
|
; CHECK-LABEL: buildvector_v4i32_subseq_2:
|
||||||
|
; CHECK: # %bb.0: # %entry
|
||||||
|
; CHECK-NEXT: vinsgr2vr.w $vr0, $a1, 0
|
||||||
|
; CHECK-NEXT: vinsgr2vr.w $vr0, $a2, 1
|
||||||
|
; CHECK-NEXT: vreplvei.d $vr0, $vr0, 0
|
||||||
|
; CHECK-NEXT: vst $vr0, $a0, 0
|
||||||
|
; CHECK-NEXT: ret
|
||||||
|
entry:
|
||||||
|
%ins0 = insertelement <4 x i32> undef, i32 %a0, i32 0
|
||||||
|
%ins1 = insertelement <4 x i32> %ins0, i32 %a1, i32 1
|
||||||
|
%ins2 = insertelement <4 x i32> %ins1, i32 %a0, i32 2
|
||||||
|
%ins3 = insertelement <4 x i32> %ins2, i32 %a1, i32 3
|
||||||
|
store <4 x i32> %ins3, ptr %dst
|
||||||
|
ret void
|
||||||
|
}
|
||||||
|
|
||||||
define void @buildvector_v2i64(ptr %dst, i64 %a0, i64 %a1) nounwind {
|
define void @buildvector_v2i64(ptr %dst, i64 %a0, i64 %a1) nounwind {
|
||||||
; CHECK-LABEL: buildvector_v2i64:
|
; CHECK-LABEL: buildvector_v2i64:
|
||||||
; CHECK: # %bb.0: # %entry
|
; CHECK: # %bb.0: # %entry
|
||||||
@ -562,6 +719,24 @@ entry:
|
|||||||
ret void
|
ret void
|
||||||
}
|
}
|
||||||
|
|
||||||
|
define void @buildvector_v4f32_subseq_2(ptr %dst, float %a0, float %a1) nounwind {
|
||||||
|
; CHECK-LABEL: buildvector_v4f32_subseq_2:
|
||||||
|
; CHECK: # %bb.0: # %entry
|
||||||
|
; CHECK-NEXT: # kill: def $f1 killed $f1 def $vr1
|
||||||
|
; CHECK-NEXT: # kill: def $f0 killed $f0 def $vr0
|
||||||
|
; CHECK-NEXT: vextrins.w $vr0, $vr1, 16
|
||||||
|
; CHECK-NEXT: vreplvei.d $vr0, $vr0, 0
|
||||||
|
; CHECK-NEXT: vst $vr0, $a0, 0
|
||||||
|
; CHECK-NEXT: ret
|
||||||
|
entry:
|
||||||
|
%ins0 = insertelement <4 x float> undef, float %a0, i32 0
|
||||||
|
%ins1 = insertelement <4 x float> %ins0, float %a1, i32 1
|
||||||
|
%ins2 = insertelement <4 x float> %ins1, float %a0, i32 2
|
||||||
|
%ins3 = insertelement <4 x float> %ins2, float %a1, i32 3
|
||||||
|
store <4 x float> %ins3, ptr %dst
|
||||||
|
ret void
|
||||||
|
}
|
||||||
|
|
||||||
define void @buildvector_v2f64(ptr %dst, double %a0, double %a1) nounwind {
|
define void @buildvector_v2f64(ptr %dst, double %a0, double %a1) nounwind {
|
||||||
; CHECK-LABEL: buildvector_v2f64:
|
; CHECK-LABEL: buildvector_v2f64:
|
||||||
; CHECK: # %bb.0: # %entry
|
; CHECK: # %bb.0: # %entry
|
||||||
|
Loading…
x
Reference in New Issue
Block a user