aboutsummaryrefslogtreecommitdiff
path: root/test/CodeGen/AArch64/fast-isel-folded-shift.ll
diff options
context:
space:
mode:
Diffstat (limited to 'test/CodeGen/AArch64/fast-isel-folded-shift.ll')
-rw-r--r--test/CodeGen/AArch64/fast-isel-folded-shift.ll125
1 files changed, 125 insertions, 0 deletions
diff --git a/test/CodeGen/AArch64/fast-isel-folded-shift.ll b/test/CodeGen/AArch64/fast-isel-folded-shift.ll
new file mode 100644
index 000000000000..b881ef5c6d52
--- /dev/null
+++ b/test/CodeGen/AArch64/fast-isel-folded-shift.ll
@@ -0,0 +1,125 @@
+; RUN: llc -mtriple=aarch64-apple-darwin -fast-isel=1 -verify-machineinstrs < %s | FileCheck %s
+
+; Test invalid shift values. This will fall-back to SDAG.
+; AND
+define zeroext i8 @and_rs_i8(i8 signext %a, i8 signext %b) {
+; CHECK-LABEL: and_rs_i8
+; CHECK: and [[REG:w[0-9]+]], w0, w8
+; CHECK-NEXT: and {{w[0-9]+}}, [[REG]], #0xff
+ %1 = shl i8 %b, 8
+ %2 = and i8 %a, %1
+ ret i8 %2
+}
+
+define zeroext i16 @and_rs_i16(i16 signext %a, i16 signext %b) {
+; CHECK-LABEL: and_rs_i16
+; CHECK: and [[REG:w[0-9]+]], w0, w8
+; CHECK-NEXT: and {{w[0-9]+}}, [[REG]], #0xffff
+ %1 = shl i16 %b, 16
+ %2 = and i16 %a, %1
+ ret i16 %2
+}
+
+define i32 @and_rs_i32(i32 %a, i32 %b) {
+; CHECK-LABEL: and_rs_i32
+; CHECK: and w0, w0, w8
+ %1 = shl i32 %b, 32
+ %2 = and i32 %a, %1
+ ret i32 %2
+}
+
+define i64 @and_rs_i64(i64 %a, i64 %b) {
+; CHECK-LABEL: and_rs_i64
+; CHECK: and x0, x0, x8
+ %1 = shl i64 %b, 64
+ %2 = and i64 %a, %1
+ ret i64 %2
+}
+
+; OR
+define zeroext i8 @or_rs_i8(i8 signext %a, i8 signext %b) {
+; CHECK-LABEL: or_rs_i8
+; CHECK: orr [[REG:w[0-9]+]], w0, w8
+; CHECK-NEXT: and {{w[0-9]+}}, [[REG]], #0xff
+ %1 = shl i8 %b, 8
+ %2 = or i8 %a, %1
+ ret i8 %2
+}
+
+define zeroext i16 @or_rs_i16(i16 signext %a, i16 signext %b) {
+; CHECK-LABEL: or_rs_i16
+; CHECK: orr [[REG:w[0-9]+]], w0, w8
+; CHECK-NEXT: and {{w[0-9]+}}, [[REG]], #0xffff
+ %1 = shl i16 %b, 16
+ %2 = or i16 %a, %1
+ ret i16 %2
+}
+
+define i32 @or_rs_i32(i32 %a, i32 %b) {
+; CHECK-LABEL: or_rs_i32
+; CHECK: orr w0, w0, w8
+ %1 = shl i32 %b, 32
+ %2 = or i32 %a, %1
+ ret i32 %2
+}
+
+define i64 @or_rs_i64(i64 %a, i64 %b) {
+; CHECK-LABEL: or_rs_i64
+; CHECK: orr x0, x0, x8
+ %1 = shl i64 %b, 64
+ %2 = or i64 %a, %1
+ ret i64 %2
+}
+
+; XOR
+define zeroext i8 @xor_rs_i8(i8 %a, i8 %b) {
+; CHECK-LABEL: xor_rs_i8
+; CHECK: eor [[REG:w[0-9]+]], w0, w8
+; CHECK-NEXT: and {{w[0-9]+}}, [[REG]], #0xff
+ %1 = shl i8 %b, 8
+ %2 = xor i8 %a, %1
+ ret i8 %2
+}
+
+define zeroext i16 @xor_rs_i16(i16 %a, i16 %b) {
+; CHECK-LABEL: xor_rs_i16
+; CHECK: eor [[REG:w[0-9]+]], w0, w8
+; CHECK-NEXT: and {{w[0-9]+}}, [[REG]], #0xffff
+ %1 = shl i16 %b, 16
+ %2 = xor i16 %a, %1
+ ret i16 %2
+}
+
+define i32 @xor_rs_i32(i32 %a, i32 %b) {
+; CHECK-LABEL: xor_rs_i32
+; CHECK: eor w0, w0, w8
+ %1 = shl i32 %b, 32
+ %2 = xor i32 %a, %1
+ ret i32 %2
+}
+
+define i64 @xor_rs_i64(i64 %a, i64 %b) {
+; CHECK-LABEL: xor_rs_i64
+; CHECK: eor x0, x0, x8
+ %1 = shl i64 %b, 64
+ %2 = xor i64 %a, %1
+ ret i64 %2
+}
+
+;ADD
+define i32 @add_rs_i32(i32 %a, i32 %b) {
+; CHECK-LABEL: add_rs_i32
+; CHECK: add w0, w0, w8
+ %1 = shl i32 %b, 32
+ %2 = add i32 %a, %1
+ ret i32 %2
+}
+
+define i64 @add_rs_i64(i64 %a, i64 %b) {
+; CHECK-LABEL: add_rs_i64
+; CHECK: add x0, x0, x8
+ %1 = shl i64 %b, 64
+ %2 = add i64 %a, %1
+ ret i64 %2
+}
+