Skip to content

Conversation

@AZero13
Copy link
Contributor

@AZero13 AZero13 commented Oct 17, 2025

@AZero13 AZero13 requested a review from nikic as a code owner October 17, 2025 20:26
@llvmbot llvmbot added llvm:instcombine Covers the InstCombine, InstSimplify and AggressiveInstCombine passes llvm:transforms labels Oct 17, 2025
@llvmbot
Copy link
Member

llvmbot commented Oct 17, 2025

@llvm/pr-subscribers-llvm-transforms

Author: AZero13 (AZero13)

Changes

Alive2: https://alive2.llvm.org/ce/z/QiMA_i


Full diff: https://github.com/llvm/llvm-project/pull/164009.diff

2 Files Affected:

  • (modified) llvm/lib/Transforms/InstCombine/InstCombineAddSub.cpp (+16-7)
  • (modified) llvm/test/Transforms/InstCombine/sext-a-lt-b-plus-zext-a-gt-b-to-uscmp.ll (+174)
diff --git a/llvm/lib/Transforms/InstCombine/InstCombineAddSub.cpp b/llvm/lib/Transforms/InstCombine/InstCombineAddSub.cpp index 73ec4514f8414..d882921a7d28b 100644 --- a/llvm/lib/Transforms/InstCombine/InstCombineAddSub.cpp +++ b/llvm/lib/Transforms/InstCombine/InstCombineAddSub.cpp @@ -1660,6 +1660,7 @@ Instruction *InstCombinerImpl::visitAdd(BinaryOperator &I) { return replaceInstUsesWith(I, Constant::getNullValue(I.getType())); // sext(A < B) + zext(A > B) => ucmp/scmp(A, B) + // sext(A <= B) + zext(A >= B) => ucmp/scmp(A, B) CmpPredicate LTPred, GTPred; if (match(&I, m_c_Add(m_SExt(m_c_ICmp(LTPred, m_Value(A), m_Value(B))), @@ -1670,13 +1671,21 @@ Instruction *InstCombinerImpl::visitAdd(BinaryOperator &I) { std::swap(A, B); } - if (ICmpInst::isLT(LTPred) && ICmpInst::isGT(GTPred) && - ICmpInst::isSigned(LTPred) == ICmpInst::isSigned(GTPred)) - return replaceInstUsesWith( - I, Builder.CreateIntrinsic( - Ty, - ICmpInst::isSigned(LTPred) ? Intrinsic::scmp : Intrinsic::ucmp, - {A, B})); + if (ICmpInst::isSigned(LTPred) == ICmpInst::isSigned(GTPred)) { + Intrinsic::ID IID = ICmpInst::isSigned(LTPred) ? Intrinsic::scmp : Intrinsic::ucmp; +  + // Handle strict inequalities: sext(A < B) + zext(A > B) => scmp/ucmp(A, B) + if (ICmpInst::isLT(LTPred) && ICmpInst::isGT(GTPred)) { + return replaceInstUsesWith( + I, Builder.CreateIntrinsic(Ty, IID, {A, B})); + } +  + // Handle non-strict inequalities: sext(A <= B) + zext(A >= B) => scmp/ucmp(A, B) + if (ICmpInst::isLE(LTPred) && ICmpInst::isGE(GTPred)) { + return replaceInstUsesWith( + I, Builder.CreateIntrinsic(Ty, IID, {A, B})); + } + } } // A+B --> A|B iff A and B have no bits set in common. diff --git a/llvm/test/Transforms/InstCombine/sext-a-lt-b-plus-zext-a-gt-b-to-uscmp.ll b/llvm/test/Transforms/InstCombine/sext-a-lt-b-plus-zext-a-gt-b-to-uscmp.ll index 02ae7ce82f13c..720219bb49c5f 100644 --- a/llvm/test/Transforms/InstCombine/sext-a-lt-b-plus-zext-a-gt-b-to-uscmp.ll +++ b/llvm/test/Transforms/InstCombine/sext-a-lt-b-plus-zext-a-gt-b-to-uscmp.ll @@ -182,3 +182,177 @@ define i8 @signed_add_neg5(i32 %a, i32 %b) { %r = add i8 %lt8, %gt8 ret i8 %r } + +; sext(A s<= B) + zext(A s>= B) => scmp(A, B) +define i8 @signed_add_ge_le(i32 %a, i32 %b) { +; CHECK-LABEL: define i8 @signed_add_ge_le( +; CHECK-SAME: i32 [[A:%.*]], i32 [[B:%.*]]) { +; CHECK-NEXT: [[R:%.*]] = call i8 @llvm.scmp.i8.i32(i32 [[A]], i32 [[B]]) +; CHECK-NEXT: ret i8 [[R]] +; + %le = icmp sle i32 %a, %b + %le8 = sext i1 %le to i8 + %ge = icmp sge i32 %a, %b + %ge8 = zext i1 %ge to i8 + %r = add i8 %le8, %ge8 + ret i8 %r +} + +; Unsigned version of >= and <= +define i8 @unsigned_add_ge_le(i32 %a, i32 %b) { +; CHECK-LABEL: define i8 @unsigned_add_ge_le( +; CHECK-SAME: i32 [[A:%.*]], i32 [[B:%.*]]) { +; CHECK-NEXT: [[R:%.*]] = call i8 @llvm.ucmp.i8.i32(i32 [[A]], i32 [[B]]) +; CHECK-NEXT: ret i8 [[R]] +; + %le = icmp ule i32 %a, %b + %le8 = sext i1 %le to i8 + %ge = icmp uge i32 %a, %b + %ge8 = zext i1 %ge to i8 + %r = add i8 %le8, %ge8 + ret i8 %r +} + +; zext(A s>= B) - zext(A s<= B) => scmp(A, B) +define i8 @signed_sub_ge_le(i32 %a, i32 %b) { +; CHECK-LABEL: define i8 @signed_sub_ge_le( +; CHECK-SAME: i32 [[A:%.*]], i32 [[B:%.*]]) { +; CHECK-NEXT: [[R:%.*]] = call i8 @llvm.scmp.i8.i32(i32 [[A]], i32 [[B]]) +; CHECK-NEXT: ret i8 [[R]] +; + %le = icmp sle i32 %a, %b + %le8 = zext i1 %le to i8 + %ge = icmp sge i32 %a, %b + %ge8 = zext i1 %ge to i8 + %r = sub i8 %ge8, %le8 + ret i8 %r +} + +; Unsigned version of >= and <= subtraction +define i8 @unsigned_sub_ge_le(i32 %a, i32 %b) { +; CHECK-LABEL: define i8 @unsigned_sub_ge_le( +; CHECK-SAME: i32 [[A:%.*]], i32 [[B:%.*]]) { +; CHECK-NEXT: [[R:%.*]] = call i8 @llvm.ucmp.i8.i32(i32 [[A]], i32 [[B]]) +; CHECK-NEXT: ret i8 [[R]] +; + %le = icmp ule i32 %a, %b + %le8 = zext i1 %le to i8 + %ge = icmp uge i32 %a, %b + %ge8 = zext i1 %ge to i8 + %r = sub i8 %ge8, %le8 + ret i8 %r +} + +; Constant canonicalization: (a > 4) - (a < 6) => scmp(a, 5) +define i8 @signed_sub_const_canonicalization(i32 %a) { +; CHECK-LABEL: define i8 @signed_sub_const_canonicalization( +; CHECK-SAME: i32 [[A:%.*]]) { +; CHECK-NEXT: [[LT:%.*]] = icmp slt i32 [[A]], 6 +; CHECK-NEXT: [[LT8_NEG:%.*]] = sext i1 [[LT]] to i8 +; CHECK-NEXT: [[GT:%.*]] = icmp sgt i32 [[A]], 4 +; CHECK-NEXT: [[GT8:%.*]] = zext i1 [[GT]] to i8 +; CHECK-NEXT: [[R:%.*]] = add nsw i8 [[LT8_NEG]], [[GT8]] +; CHECK-NEXT: ret i8 [[R]] +; + %lt = icmp slt i32 %a, 6 + %lt8 = zext i1 %lt to i8 + %gt = icmp sgt i32 %a, 4 + %gt8 = zext i1 %gt to i8 + %r = sub i8 %gt8, %lt8 + ret i8 %r +} + +; Constant canonicalization: (a >= 5) - (a <= 5) => scmp(a, 5) +define i8 @signed_sub_const_canonicalization2(i32 %a) { +; CHECK-LABEL: define i8 @signed_sub_const_canonicalization2( +; CHECK-SAME: i32 [[A:%.*]]) { +; CHECK-NEXT: [[LE:%.*]] = icmp slt i32 [[A]], 6 +; CHECK-NEXT: [[LE8_NEG:%.*]] = sext i1 [[LE]] to i8 +; CHECK-NEXT: [[GE:%.*]] = icmp sgt i32 [[A]], 4 +; CHECK-NEXT: [[GE8:%.*]] = zext i1 [[GE]] to i8 +; CHECK-NEXT: [[R:%.*]] = add nsw i8 [[LE8_NEG]], [[GE8]] +; CHECK-NEXT: ret i8 [[R]] +; + %le = icmp sle i32 %a, 5 + %le8 = zext i1 %le to i8 + %ge = icmp sge i32 %a, 5 + %ge8 = zext i1 %ge to i8 + %r = sub i8 %ge8, %le8 + ret i8 %r +} + +; Unsigned constant canonicalization: (a > 4) - (a < 6) => ucmp(a, 5) +define i8 @unsigned_sub_const_canonicalization(i32 %a) { +; CHECK-LABEL: define i8 @unsigned_sub_const_canonicalization( +; CHECK-SAME: i32 [[A:%.*]]) { +; CHECK-NEXT: [[LT:%.*]] = icmp ult i32 [[A]], 6 +; CHECK-NEXT: [[LT8_NEG:%.*]] = sext i1 [[LT]] to i8 +; CHECK-NEXT: [[GT:%.*]] = icmp ugt i32 [[A]], 4 +; CHECK-NEXT: [[GT8:%.*]] = zext i1 [[GT]] to i8 +; CHECK-NEXT: [[R:%.*]] = add nsw i8 [[LT8_NEG]], [[GT8]] +; CHECK-NEXT: ret i8 [[R]] +; + %lt = icmp ult i32 %a, 6 + %lt8 = zext i1 %lt to i8 + %gt = icmp ugt i32 %a, 4 + %gt8 = zext i1 %gt to i8 + %r = sub i8 %gt8, %lt8 + ret i8 %r +} + +; Constant canonicalization with >= and <=: (a >= 5) - (a <= 5) => scmp(a, 5) +define i8 @signed_sub_const_canonicalization_ge_le(i32 %a) { +; CHECK-LABEL: define i8 @signed_sub_const_canonicalization_ge_le( +; CHECK-SAME: i32 [[A:%.*]]) { +; CHECK-NEXT: [[LE:%.*]] = icmp slt i32 [[A]], 6 +; CHECK-NEXT: [[LE8_NEG:%.*]] = sext i1 [[LE]] to i8 +; CHECK-NEXT: [[GE:%.*]] = icmp sgt i32 [[A]], 4 +; CHECK-NEXT: [[GE8:%.*]] = zext i1 [[GE]] to i8 +; CHECK-NEXT: [[R:%.*]] = add nsw i8 [[LE8_NEG]], [[GE8]] +; CHECK-NEXT: ret i8 [[R]] +; + %le = icmp sle i32 %a, 5 + %le8 = zext i1 %le to i8 + %ge = icmp sge i32 %a, 5 + %ge8 = zext i1 %ge to i8 + %r = sub i8 %ge8, %le8 + ret i8 %r +} + +; More constant canonicalization: (a > 2) - (a < 4) => scmp(a, 3) +define i8 @signed_sub_const_canonicalization3(i32 %a) { +; CHECK-LABEL: define i8 @signed_sub_const_canonicalization3( +; CHECK-SAME: i32 [[A:%.*]]) { +; CHECK-NEXT: [[LT:%.*]] = icmp slt i32 [[A]], 4 +; CHECK-NEXT: [[LT8_NEG:%.*]] = sext i1 [[LT]] to i8 +; CHECK-NEXT: [[GT:%.*]] = icmp sgt i32 [[A]], 2 +; CHECK-NEXT: [[GT8:%.*]] = zext i1 [[GT]] to i8 +; CHECK-NEXT: [[R:%.*]] = add nsw i8 [[LT8_NEG]], [[GT8]] +; CHECK-NEXT: ret i8 [[R]] +; + %lt = icmp slt i32 %a, 4 + %lt8 = zext i1 %lt to i8 + %gt = icmp sgt i32 %a, 2 + %gt8 = zext i1 %gt to i8 + %r = sub i8 %gt8, %lt8 + ret i8 %r +} + +; Negative test: constants that are more than one apart - should NOT canonicalize +define i8 @signed_sub_const_no_canonicalization(i32 %a) { +; CHECK-LABEL: define i8 @signed_sub_const_no_canonicalization( +; CHECK-SAME: i32 [[A:%.*]]) { +; CHECK-NEXT: [[LT:%.*]] = icmp slt i32 [[A]], 10 +; CHECK-NEXT: [[LT8_NEG:%.*]] = sext i1 [[LT]] to i8 +; CHECK-NEXT: [[GT:%.*]] = icmp sgt i32 [[A]], 4 +; CHECK-NEXT: [[GT8:%.*]] = zext i1 [[GT]] to i8 +; CHECK-NEXT: [[R:%.*]] = add nsw i8 [[LT8_NEG]], [[GT8]] +; CHECK-NEXT: ret i8 [[R]] +; + %lt = icmp slt i32 %a, 10 + %lt8 = zext i1 %lt to i8 + %gt = icmp sgt i32 %a, 4 + %gt8 = zext i1 %gt to i8 + %r = sub i8 %gt8, %lt8 + ret i8 %r +} 
@github-actions
Copy link

github-actions bot commented Oct 17, 2025

✅ With the latest revision this PR passed the C/C++ code formatter.

@AZero13
Copy link
Contributor Author

AZero13 commented Oct 23, 2025

Okay I am going to add constant support and then it will have a few hits.

@AZero13 AZero13 force-pushed the a-b branch 3 times, most recently from 7f79822 to 5246701 Compare October 23, 2025 14:23
@AZero13
Copy link
Contributor Author

AZero13 commented Oct 24, 2025

@dtcxzyw can we try one more time?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

llvm:instcombine Covers the InstCombine, InstSimplify and AggressiveInstCombine passes llvm:transforms

2 participants