; NOTE: Assertions have been autogenerated by utils/update_test_checks.py UTC_ARGS: --version 5 ; RUN: opt < %s -passes=instcombine -S | FileCheck %s ; PR125228 define <16 x i8> @knownbits_bitcast_masked_shift(<16 x i8> %arg1, <16 x i8> %arg2) { ; CHECK-LABEL: define <16 x i8> @knownbits_bitcast_masked_shift( ; CHECK-SAME: <16 x i8> [[ARG1:%.*]], <16 x i8> [[ARG2:%.*]]) { ; CHECK-NEXT: [[AND:%.*]] = and <16 x i8> [[ARG1]], splat (i8 3) ; CHECK-NEXT: [[AND3:%.*]] = and <16 x i8> [[ARG2]], splat (i8 48) ; CHECK-NEXT: [[OR:%.*]] = or disjoint <16 x i8> [[AND3]], [[AND]] ; CHECK-NEXT: [[BITCAST4:%.*]] = bitcast <16 x i8> [[OR]] to <8 x i16> ; CHECK-NEXT: [[SHL5:%.*]] = shl nuw <8 x i16> [[BITCAST4]], splat (i16 2) ; CHECK-NEXT: [[BITCAST6:%.*]] = bitcast <8 x i16> [[SHL5]] to <16 x i8> ; CHECK-NEXT: ret <16 x i8> [[BITCAST6]] ; %and = and <16 x i8> %arg1, splat (i8 3) %and3 = and <16 x i8> %arg2, splat (i8 48) %or = or disjoint <16 x i8> %and3, %and %bitcast4 = bitcast <16 x i8> %or to <8 x i16> %shl5 = shl nuw <8 x i16> %bitcast4, splat (i16 2) %bitcast6 = bitcast <8 x i16> %shl5 to <16 x i8> %and7 = and <16 x i8> %bitcast6, splat (i8 -52) ret <16 x i8> %and7 } define <16 x i8> @knownbits_shuffle_masked_nibble_shift(<16 x i8> %arg) { ; CHECK-LABEL: define <16 x i8> @knownbits_shuffle_masked_nibble_shift( ; CHECK-SAME: <16 x i8> [[ARG:%.*]]) { ; CHECK-NEXT: [[AND:%.*]] = and <16 x i8> [[ARG]], splat (i8 15) ; CHECK-NEXT: [[SHUFFLEVECTOR:%.*]] = shufflevector <16 x i8> [[AND]], <16 x i8> poison, <16 x i32> ; CHECK-NEXT: [[BITCAST1:%.*]] = bitcast <16 x i8> [[SHUFFLEVECTOR]] to <8 x i16> ; CHECK-NEXT: [[SHL:%.*]] = shl nuw <8 x i16> [[BITCAST1]], splat (i16 4) ; CHECK-NEXT: [[BITCAST2:%.*]] = bitcast <8 x i16> [[SHL]] to <16 x i8> ; CHECK-NEXT: ret <16 x i8> [[BITCAST2]] ; %and = and <16 x i8> %arg, splat (i8 15) %shufflevector = shufflevector <16 x i8> %and, <16 x i8> poison, <16 x i32> %bitcast1 = bitcast <16 x i8> %shufflevector to <8 x i16> %shl = shl nuw <8 x i16> %bitcast1, splat (i16 4) %bitcast2 = bitcast <8 x i16> %shl to <16 x i8> %and3 = and <16 x i8> %bitcast2, splat (i8 -16) ret <16 x i8> %and3 } define <16 x i8> @knownbits_reverse_shuffle_masked_shift(<16 x i8> %arg) { ; CHECK-LABEL: define <16 x i8> @knownbits_reverse_shuffle_masked_shift( ; CHECK-SAME: <16 x i8> [[ARG:%.*]]) { ; CHECK-NEXT: [[AND:%.*]] = and <16 x i8> [[ARG]], splat (i8 15) ; CHECK-NEXT: [[SHUFFLEVECTOR:%.*]] = shufflevector <16 x i8> [[AND]], <16 x i8> poison, <16 x i32> ; CHECK-NEXT: [[BITCAST1:%.*]] = bitcast <16 x i8> [[SHUFFLEVECTOR]] to <8 x i16> ; CHECK-NEXT: [[SHL:%.*]] = shl nuw <8 x i16> [[BITCAST1]], splat (i16 4) ; CHECK-NEXT: [[BITCAST2:%.*]] = bitcast <8 x i16> [[SHL]] to <16 x i8> ; CHECK-NEXT: ret <16 x i8> [[BITCAST2]] ; %and = and <16 x i8> %arg, splat (i8 15) %shufflevector = shufflevector <16 x i8> %and, <16 x i8> poison, <16 x i32> %bitcast1 = bitcast <16 x i8> %shufflevector to <8 x i16> %shl = shl nuw <8 x i16> %bitcast1, splat (i16 4) %bitcast2 = bitcast <8 x i16> %shl to <16 x i8> %and3 = and <16 x i8> %bitcast2, splat (i8 -16) ret <16 x i8> %and3 } define <16 x i8> @knownbits_extract_bit(<8 x i16> %arg) { ; CHECK-LABEL: define <16 x i8> @knownbits_extract_bit( ; CHECK-SAME: <8 x i16> [[ARG:%.*]]) { ; CHECK-NEXT: [[LSHR:%.*]] = lshr <8 x i16> [[ARG]], splat (i16 15) ; CHECK-NEXT: [[BITCAST1:%.*]] = bitcast <8 x i16> [[LSHR]] to <16 x i8> ; CHECK-NEXT: ret <16 x i8> [[BITCAST1]] ; %lshr = lshr <8 x i16> %arg, splat (i16 15) %bitcast1 = bitcast <8 x i16> %lshr to <16 x i8> %and = and <16 x i8> %bitcast1, splat (i8 1) ret <16 x i8> %and } define { i32, i1 } @knownbits_popcount_add_with_overflow(<2 x i64> %arg1, <2 x i64> %arg2) { ; CHECK-LABEL: define { i32, i1 } @knownbits_popcount_add_with_overflow( ; CHECK-SAME: <2 x i64> [[ARG1:%.*]], <2 x i64> [[ARG2:%.*]]) { ; CHECK-NEXT: [[CALL:%.*]] = tail call range(i64 0, 65) <2 x i64> @llvm.ctpop.v2i64(<2 x i64> [[ARG1]]) ; CHECK-NEXT: [[BITCAST5:%.*]] = bitcast <2 x i64> [[CALL]] to <4 x i32> ; CHECK-NEXT: [[EXTRACTELEMENT:%.*]] = extractelement <4 x i32> [[BITCAST5]], i64 0 ; CHECK-NEXT: [[CALL9:%.*]] = tail call range(i64 0, 65) <2 x i64> @llvm.ctpop.v2i64(<2 x i64> [[ARG2]]) ; CHECK-NEXT: [[BITCAST10:%.*]] = bitcast <2 x i64> [[CALL9]] to <4 x i32> ; CHECK-NEXT: [[EXTRACTELEMENT11:%.*]] = extractelement <4 x i32> [[BITCAST10]], i64 0 ; CHECK-NEXT: [[CALL12:%.*]] = add nuw nsw i32 [[EXTRACTELEMENT]], [[EXTRACTELEMENT11]] ; CHECK-NEXT: [[TMP1:%.*]] = insertvalue { i32, i1 } { i32 poison, i1 false }, i32 [[CALL12]], 0 ; CHECK-NEXT: ret { i32, i1 } [[TMP1]] ; %call = tail call <2 x i64> @llvm.ctpop.v2i64(<2 x i64> %arg1) %bitcast5 = bitcast <2 x i64> %call to <4 x i32> %extractelement = extractelement <4 x i32> %bitcast5, i64 0 %call9 = tail call <2 x i64> @llvm.ctpop.v2i64(<2 x i64> %arg2) %bitcast10 = bitcast <2 x i64> %call9 to <4 x i32> %extractelement11 = extractelement <4 x i32> %bitcast10, i64 0 %call12 = tail call { i32, i1 } @llvm.uadd.with.overflow.i32(i32 %extractelement, i32 %extractelement11) ret { i32, i1 } %call12 } define <16 x i8> @knownbits_shuffle_add_shift_v32i8(<16 x i8> %arg1, <8 x i16> %arg2, <8 x i16> %arg3) { ; CHECK-LABEL: define <16 x i8> @knownbits_shuffle_add_shift_v32i8( ; CHECK-SAME: <16 x i8> [[ARG1:%.*]], <8 x i16> [[ARG2:%.*]], <8 x i16> [[ARG3:%.*]]) { ; CHECK-NEXT: [[SHL6:%.*]] = shl <8 x i16> [[ARG2]], splat (i16 8) ; CHECK-NEXT: [[BITCAST7:%.*]] = bitcast <8 x i16> [[SHL6]] to <16 x i8> ; CHECK-NEXT: [[SHL10:%.*]] = shl <8 x i16> [[ARG3]], splat (i16 8) ; CHECK-NEXT: [[BITCAST11:%.*]] = bitcast <8 x i16> [[SHL10]] to <16 x i8> ; CHECK-NEXT: [[ADD12:%.*]] = add <16 x i8> [[BITCAST11]], [[BITCAST7]] ; CHECK-NEXT: [[ADD14:%.*]] = add <16 x i8> [[ADD12]], [[ARG1]] ; CHECK-NEXT: ret <16 x i8> [[ADD14]] ; %shl6 = shl <8 x i16> %arg2, splat (i16 8) %bitcast7 = bitcast <8 x i16> %shl6 to <16 x i8> %shl10 = shl <8 x i16> %arg3, splat (i16 8) %bitcast11 = bitcast <8 x i16> %shl10 to <16 x i8> %add12 = add <16 x i8> %bitcast11, %bitcast7 %add13 = add <16 x i8> %add12, %arg1 %bitcast14 = bitcast <16 x i8> %add12 to <8 x i16> %shl15 = shl <8 x i16> %bitcast14, splat (i16 8) %bitcast16 = bitcast <8 x i16> %shl15 to <16 x i8> %add17 = add <16 x i8> %add13, %bitcast16 ret <16 x i8> %add17 } declare <2 x i64> @llvm.ctpop.v2i64(<2 x i64>) declare { i32, i1 } @llvm.uadd.with.overflow.i32(i32, i32)