| ; RUN: llc -verify-machineinstrs < %s -mtriple=aarch64-none-linux-gnu | FileCheck %s |
| |
| @var32 = global i32 0 |
| @var64 = global i64 0 |
| |
| define void @test_lsl_arith(i32 %lhs32, i32 %rhs32, i64 %lhs64, i64 %rhs64) { |
| ; CHECK: test_lsl_arith: |
| |
| %rhs1 = load volatile i32* @var32 |
| %shift1 = shl i32 %rhs1, 18 |
| %val1 = add i32 %lhs32, %shift1 |
| store volatile i32 %val1, i32* @var32 |
| ; CHECK: add {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}}, lsl #18 |
| |
| %rhs2 = load volatile i32* @var32 |
| %shift2 = shl i32 %rhs2, 31 |
| %val2 = add i32 %shift2, %lhs32 |
| store volatile i32 %val2, i32* @var32 |
| ; CHECK: add {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}}, lsl #31 |
| |
| %rhs3 = load volatile i32* @var32 |
| %shift3 = shl i32 %rhs3, 5 |
| %val3 = sub i32 %lhs32, %shift3 |
| store volatile i32 %val3, i32* @var32 |
| ; CHECK: sub {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}}, lsl #5 |
| |
| ; Subtraction is not commutative! |
| %rhs4 = load volatile i32* @var32 |
| %shift4 = shl i32 %rhs4, 19 |
| %val4 = sub i32 %shift4, %lhs32 |
| store volatile i32 %val4, i32* @var32 |
| ; CHECK-NOT: sub {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}}, lsl #19 |
| |
| %lhs4a = load volatile i32* @var32 |
| %shift4a = shl i32 %lhs4a, 15 |
| %val4a = sub i32 0, %shift4a |
| store volatile i32 %val4a, i32* @var32 |
| ; CHECK: sub {{w[0-9]+}}, wzr, {{w[0-9]+}}, lsl #15 |
| |
| %rhs5 = load volatile i64* @var64 |
| %shift5 = shl i64 %rhs5, 18 |
| %val5 = add i64 %lhs64, %shift5 |
| store volatile i64 %val5, i64* @var64 |
| ; CHECK: add {{x[0-9]+}}, {{x[0-9]+}}, {{x[0-9]+}}, lsl #18 |
| |
| %rhs6 = load volatile i64* @var64 |
| %shift6 = shl i64 %rhs6, 31 |
| %val6 = add i64 %shift6, %lhs64 |
| store volatile i64 %val6, i64* @var64 |
| ; CHECK: add {{x[0-9]+}}, {{x[0-9]+}}, {{x[0-9]+}}, lsl #31 |
| |
| %rhs7 = load volatile i64* @var64 |
| %shift7 = shl i64 %rhs7, 5 |
| %val7 = sub i64 %lhs64, %shift7 |
| store volatile i64 %val7, i64* @var64 |
| ; CHECK: sub {{x[0-9]+}}, {{x[0-9]+}}, {{x[0-9]+}}, lsl #5 |
| |
| ; Subtraction is not commutative! |
| %rhs8 = load volatile i64* @var64 |
| %shift8 = shl i64 %rhs8, 19 |
| %val8 = sub i64 %shift8, %lhs64 |
| store volatile i64 %val8, i64* @var64 |
| ; CHECK-NOT: sub {{x[0-9]+}}, {{x[0-9]+}}, {{x[0-9]+}}, lsl #19 |
| |
| %lhs8a = load volatile i64* @var64 |
| %shift8a = shl i64 %lhs8a, 60 |
| %val8a = sub i64 0, %shift8a |
| store volatile i64 %val8a, i64* @var64 |
| ; CHECK: sub {{x[0-9]+}}, xzr, {{x[0-9]+}}, lsl #60 |
| |
| ret void |
| ; CHECK: ret |
| } |
| |
| define void @test_lsr_arith(i32 %lhs32, i32 %rhs32, i64 %lhs64, i64 %rhs64) { |
| ; CHECK: test_lsr_arith: |
| |
| %shift1 = lshr i32 %rhs32, 18 |
| %val1 = add i32 %lhs32, %shift1 |
| store volatile i32 %val1, i32* @var32 |
| ; CHECK: add {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}}, lsr #18 |
| |
| %shift2 = lshr i32 %rhs32, 31 |
| %val2 = add i32 %shift2, %lhs32 |
| store volatile i32 %val2, i32* @var32 |
| ; CHECK: add {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}}, lsr #31 |
| |
| %shift3 = lshr i32 %rhs32, 5 |
| %val3 = sub i32 %lhs32, %shift3 |
| store volatile i32 %val3, i32* @var32 |
| ; CHECK: sub {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}}, lsr #5 |
| |
| ; Subtraction is not commutative! |
| %shift4 = lshr i32 %rhs32, 19 |
| %val4 = sub i32 %shift4, %lhs32 |
| store volatile i32 %val4, i32* @var32 |
| ; CHECK-NOT: sub {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}}, lsr #19 |
| |
| %shift4a = lshr i32 %lhs32, 15 |
| %val4a = sub i32 0, %shift4a |
| store volatile i32 %val4a, i32* @var32 |
| ; CHECK: sub {{w[0-9]+}}, wzr, {{w[0-9]+}}, lsr #15 |
| |
| %shift5 = lshr i64 %rhs64, 18 |
| %val5 = add i64 %lhs64, %shift5 |
| store volatile i64 %val5, i64* @var64 |
| ; CHECK: add {{x[0-9]+}}, {{x[0-9]+}}, {{x[0-9]+}}, lsr #18 |
| |
| %shift6 = lshr i64 %rhs64, 31 |
| %val6 = add i64 %shift6, %lhs64 |
| store volatile i64 %val6, i64* @var64 |
| ; CHECK: add {{x[0-9]+}}, {{x[0-9]+}}, {{x[0-9]+}}, lsr #31 |
| |
| %shift7 = lshr i64 %rhs64, 5 |
| %val7 = sub i64 %lhs64, %shift7 |
| store volatile i64 %val7, i64* @var64 |
| ; CHECK: sub {{x[0-9]+}}, {{x[0-9]+}}, {{x[0-9]+}}, lsr #5 |
| |
| ; Subtraction is not commutative! |
| %shift8 = lshr i64 %rhs64, 19 |
| %val8 = sub i64 %shift8, %lhs64 |
| store volatile i64 %val8, i64* @var64 |
| ; CHECK-NOT: sub {{x[0-9]+}}, {{x[0-9]+}}, {{x[0-9]+}}, lsr #19 |
| |
| %shift8a = lshr i64 %lhs64, 45 |
| %val8a = sub i64 0, %shift8a |
| store volatile i64 %val8a, i64* @var64 |
| ; CHECK: sub {{x[0-9]+}}, xzr, {{x[0-9]+}}, lsr #45 |
| |
| ret void |
| ; CHECK: ret |
| } |
| |
| define void @test_asr_arith(i32 %lhs32, i32 %rhs32, i64 %lhs64, i64 %rhs64) { |
| ; CHECK: test_asr_arith: |
| |
| %shift1 = ashr i32 %rhs32, 18 |
| %val1 = add i32 %lhs32, %shift1 |
| store volatile i32 %val1, i32* @var32 |
| ; CHECK: add {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}}, asr #18 |
| |
| %shift2 = ashr i32 %rhs32, 31 |
| %val2 = add i32 %shift2, %lhs32 |
| store volatile i32 %val2, i32* @var32 |
| ; CHECK: add {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}}, asr #31 |
| |
| %shift3 = ashr i32 %rhs32, 5 |
| %val3 = sub i32 %lhs32, %shift3 |
| store volatile i32 %val3, i32* @var32 |
| ; CHECK: sub {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}}, asr #5 |
| |
| ; Subtraction is not commutative! |
| %shift4 = ashr i32 %rhs32, 19 |
| %val4 = sub i32 %shift4, %lhs32 |
| store volatile i32 %val4, i32* @var32 |
| ; CHECK-NOT: sub {{w[0-9]+}}, {{w[0-9]+}}, {{w[0-9]+}}, asr #19 |
| |
| %shift4a = ashr i32 %lhs32, 15 |
| %val4a = sub i32 0, %shift4a |
| store volatile i32 %val4a, i32* @var32 |
| ; CHECK: sub {{w[0-9]+}}, wzr, {{w[0-9]+}}, asr #15 |
| |
| %shift5 = ashr i64 %rhs64, 18 |
| %val5 = add i64 %lhs64, %shift5 |
| store volatile i64 %val5, i64* @var64 |
| ; CHECK: add {{x[0-9]+}}, {{x[0-9]+}}, {{x[0-9]+}}, asr #18 |
| |
| %shift6 = ashr i64 %rhs64, 31 |
| %val6 = add i64 %shift6, %lhs64 |
| store volatile i64 %val6, i64* @var64 |
| ; CHECK: add {{x[0-9]+}}, {{x[0-9]+}}, {{x[0-9]+}}, asr #31 |
| |
| %shift7 = ashr i64 %rhs64, 5 |
| %val7 = sub i64 %lhs64, %shift7 |
| store volatile i64 %val7, i64* @var64 |
| ; CHECK: sub {{x[0-9]+}}, {{x[0-9]+}}, {{x[0-9]+}}, asr #5 |
| |
| ; Subtraction is not commutative! |
| %shift8 = ashr i64 %rhs64, 19 |
| %val8 = sub i64 %shift8, %lhs64 |
| store volatile i64 %val8, i64* @var64 |
| ; CHECK-NOT: sub {{x[0-9]+}}, {{x[0-9]+}}, {{x[0-9]+}}, asr #19 |
| |
| %shift8a = ashr i64 %lhs64, 45 |
| %val8a = sub i64 0, %shift8a |
| store volatile i64 %val8a, i64* @var64 |
| ; CHECK: sub {{x[0-9]+}}, xzr, {{x[0-9]+}}, asr #45 |
| |
| ret void |
| ; CHECK: ret |
| } |
| |
| define i32 @test_cmp(i32 %lhs32, i32 %rhs32, i64 %lhs64, i64 %rhs64) { |
| ; CHECK: test_cmp: |
| |
| %shift1 = shl i32 %rhs32, 13 |
| %tst1 = icmp uge i32 %lhs32, %shift1 |
| br i1 %tst1, label %t2, label %end |
| ; CHECK: cmp {{w[0-9]+}}, {{w[0-9]+}}, lsl #13 |
| |
| t2: |
| %shift2 = lshr i32 %rhs32, 20 |
| %tst2 = icmp ne i32 %lhs32, %shift2 |
| br i1 %tst2, label %t3, label %end |
| ; CHECK: cmp {{w[0-9]+}}, {{w[0-9]+}}, lsr #20 |
| |
| t3: |
| %shift3 = ashr i32 %rhs32, 9 |
| %tst3 = icmp ne i32 %lhs32, %shift3 |
| br i1 %tst3, label %t4, label %end |
| ; CHECK: cmp {{w[0-9]+}}, {{w[0-9]+}}, asr #9 |
| |
| t4: |
| %shift4 = shl i64 %rhs64, 43 |
| %tst4 = icmp uge i64 %lhs64, %shift4 |
| br i1 %tst4, label %t5, label %end |
| ; CHECK: cmp {{x[0-9]+}}, {{x[0-9]+}}, lsl #43 |
| |
| t5: |
| %shift5 = lshr i64 %rhs64, 20 |
| %tst5 = icmp ne i64 %lhs64, %shift5 |
| br i1 %tst5, label %t6, label %end |
| ; CHECK: cmp {{x[0-9]+}}, {{x[0-9]+}}, lsr #20 |
| |
| t6: |
| %shift6 = ashr i64 %rhs64, 59 |
| %tst6 = icmp ne i64 %lhs64, %shift6 |
| br i1 %tst6, label %t7, label %end |
| ; CHECK: cmp {{x[0-9]+}}, {{x[0-9]+}}, asr #59 |
| |
| t7: |
| ret i32 1 |
| end: |
| |
| ret i32 0 |
| ; CHECK: ret |
| } |
| |
| define i32 @test_cmn(i32 %lhs32, i32 %rhs32, i64 %lhs64, i64 %rhs64) { |
| ; CHECK: test_cmn: |
| |
| %shift1 = shl i32 %rhs32, 13 |
| %val1 = sub i32 0, %shift1 |
| %tst1 = icmp uge i32 %lhs32, %val1 |
| br i1 %tst1, label %t2, label %end |
| ; Important that this isn't lowered to a cmn instruction because if %rhs32 == |
| ; 0 then the results will differ. |
| ; CHECK: sub [[RHS:w[0-9]+]], wzr, {{w[0-9]+}}, lsl #13 |
| ; CHECK: cmp {{w[0-9]+}}, [[RHS]] |
| |
| t2: |
| %shift2 = lshr i32 %rhs32, 20 |
| %val2 = sub i32 0, %shift2 |
| %tst2 = icmp ne i32 %lhs32, %val2 |
| br i1 %tst2, label %t3, label %end |
| ; CHECK: cmn {{w[0-9]+}}, {{w[0-9]+}}, lsr #20 |
| |
| t3: |
| %shift3 = ashr i32 %rhs32, 9 |
| %val3 = sub i32 0, %shift3 |
| %tst3 = icmp eq i32 %lhs32, %val3 |
| br i1 %tst3, label %t4, label %end |
| ; CHECK: cmn {{w[0-9]+}}, {{w[0-9]+}}, asr #9 |
| |
| t4: |
| %shift4 = shl i64 %rhs64, 43 |
| %val4 = sub i64 0, %shift4 |
| %tst4 = icmp slt i64 %lhs64, %val4 |
| br i1 %tst4, label %t5, label %end |
| ; Again, it's important that cmn isn't used here in case %rhs64 == 0. |
| ; CHECK: sub [[RHS:x[0-9]+]], xzr, {{x[0-9]+}}, lsl #43 |
| ; CHECK: cmp {{x[0-9]+}}, [[RHS]] |
| |
| t5: |
| %shift5 = lshr i64 %rhs64, 20 |
| %val5 = sub i64 0, %shift5 |
| %tst5 = icmp ne i64 %lhs64, %val5 |
| br i1 %tst5, label %t6, label %end |
| ; CHECK: cmn {{x[0-9]+}}, {{x[0-9]+}}, lsr #20 |
| |
| t6: |
| %shift6 = ashr i64 %rhs64, 59 |
| %val6 = sub i64 0, %shift6 |
| %tst6 = icmp ne i64 %lhs64, %val6 |
| br i1 %tst6, label %t7, label %end |
| ; CHECK: cmn {{x[0-9]+}}, {{x[0-9]+}}, asr #59 |
| |
| t7: |
| ret i32 1 |
| end: |
| |
| ret i32 0 |
| ; CHECK: ret |
| } |
| |