diff options
Diffstat (limited to 'test/CodeGen/SystemZ/loop-01.ll')
| -rw-r--r-- | test/CodeGen/SystemZ/loop-01.ll | 241 |
1 files changed, 0 insertions, 241 deletions
diff --git a/test/CodeGen/SystemZ/loop-01.ll b/test/CodeGen/SystemZ/loop-01.ll deleted file mode 100644 index 321be4b8e62b..000000000000 --- a/test/CodeGen/SystemZ/loop-01.ll +++ /dev/null @@ -1,241 +0,0 @@ -; Test loop tuning. -; -; RUN: llc < %s -mtriple=s390x-linux-gnu -mcpu=z10 | FileCheck %s -; RUN: llc < %s -mtriple=s390x-linux-gnu -mcpu=z13 \ -; RUN: | FileCheck %s -check-prefix=CHECK -check-prefix=CHECK-Z13 - -; Test that strength reduction is applied to addresses with a scale factor, -; but that indexed addressing can still be used. -define void @f1(i32 *%dest, i32 %a) { -; CHECK-LABEL: f1: -; CHECK-NOT: sllg -; CHECK: st %r3, 0({{%r[1-5],%r[1-5]}}) -; CHECK: br %r14 -entry: - br label %loop - -loop: - %index = phi i64 [ 0, %entry ], [ %next, %loop ] - %ptr = getelementptr i32, i32 *%dest, i64 %index - store i32 %a, i32 *%ptr - %next = add i64 %index, 1 - %cmp = icmp ne i64 %next, 100 - br i1 %cmp, label %loop, label %exit - -exit: - ret void -} - -; Test a loop that should be converted into dbr form and then use BRCT. -define void @f2(i32 *%src, i32 *%dest) { -; CHECK-LABEL: f2: -; CHECK: lhi [[REG:%r[0-5]]], 100 -; CHECK: [[LABEL:\.[^:]*]]:{{.*}} %loop -; CHECK: brct [[REG]], [[LABEL]] -; CHECK: br %r14 -entry: - br label %loop - -loop: - %count = phi i32 [ 0, %entry ], [ %next, %loop.next ] - %next = add i32 %count, 1 - %val = load volatile i32 , i32 *%src - %cmp = icmp eq i32 %val, 0 - br i1 %cmp, label %loop.next, label %loop.store - -loop.store: - %add = add i32 %val, 1 - store volatile i32 %add, i32 *%dest - br label %loop.next - -loop.next: - %cont = icmp ne i32 %next, 100 - br i1 %cont, label %loop, label %exit - -exit: - ret void -} - -; Like f2, but for BRCTG. -define void @f3(i64 *%src, i64 *%dest) { -; CHECK-LABEL: f3: -; CHECK: lghi [[REG:%r[0-5]]], 100 -; CHECK: [[LABEL:\.[^:]*]]:{{.*}} %loop -; CHECK: brctg [[REG]], [[LABEL]] -; CHECK: br %r14 -entry: - br label %loop - -loop: - %count = phi i64 [ 0, %entry ], [ %next, %loop.next ] - %next = add i64 %count, 1 - %val = load volatile i64 , i64 *%src - %cmp = icmp eq i64 %val, 0 - br i1 %cmp, label %loop.next, label %loop.store - -loop.store: - %add = add i64 %val, 1 - store volatile i64 %add, i64 *%dest - br label %loop.next - -loop.next: - %cont = icmp ne i64 %next, 100 - br i1 %cont, label %loop, label %exit - -exit: - ret void -} - -; Test a loop with a 64-bit decremented counter in which the 32-bit -; low part of the counter is used after the decrement. This is an example -; of a subregister use being the only thing that blocks a conversion to BRCTG. -define void @f4(i32 *%src, i32 *%dest, i64 *%dest2, i64 %count) { -; CHECK-LABEL: f4: -; CHECK: aghi [[REG:%r[0-5]]], -1 -; CHECK: lr [[REG2:%r[0-5]]], [[REG]] -; CHECK: stg [[REG2]], -; CHECK: jne {{\..*}} -; CHECK: br %r14 -entry: - br label %loop - -loop: - %left = phi i64 [ %count, %entry ], [ %next, %loop.next ] - store volatile i64 %left, i64 *%dest2 - %val = load volatile i32 , i32 *%src - %cmp = icmp eq i32 %val, 0 - br i1 %cmp, label %loop.next, label %loop.store - -loop.store: - %add = add i32 %val, 1 - store volatile i32 %add, i32 *%dest - br label %loop.next - -loop.next: - %next = add i64 %left, -1 - %ext = zext i32 %val to i64 - %shl = shl i64 %ext, 32 - %and = and i64 %next, 4294967295 - %or = or i64 %shl, %and - store volatile i64 %or, i64 *%dest2 - %cont = icmp ne i64 %next, 0 - br i1 %cont, label %loop, label %exit - -exit: - ret void -} - -; Test that negative offsets are avoided for loads of floating point. -%s.float = type { float, float, float } -define void @f5(%s.float* nocapture %a, - %s.float* nocapture readonly %b, - i32 zeroext %S) { -; CHECK-Z13-LABEL: f5: -; CHECK-Z13-NOT: -{{[0-9]+}}(%r - -entry: - %cmp9 = icmp eq i32 %S, 0 - br i1 %cmp9, label %for.cond.cleanup, label %for.body.preheader - -for.body.preheader: ; preds = %entry - br label %for.body - -for.cond.cleanup.loopexit: ; preds = %for.body - br label %for.cond.cleanup - -for.cond.cleanup: ; preds = %for.cond.cleanup.loopexit, %entry - ret void - -for.body: ; preds = %for.body.preheader, %for.body - %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %for.body.preheader ] - %a1 = getelementptr inbounds %s.float, %s.float* %b, i64 %indvars.iv, i32 0 - %tmp = load float, float* %a1, align 4 - %b4 = getelementptr inbounds %s.float, %s.float* %b, i64 %indvars.iv, i32 1 - %tmp1 = load float, float* %b4, align 4 - %add = fadd float %tmp, %tmp1 - %c = getelementptr inbounds %s.float, %s.float* %b, i64 %indvars.iv, i32 2 - %tmp2 = load float, float* %c, align 4 - %add7 = fadd float %add, %tmp2 - %a10 = getelementptr inbounds %s.float, %s.float* %a, i64 %indvars.iv, i32 0 - store float %add7, float* %a10, align 4 - %indvars.iv.next = add nuw nsw i64 %indvars.iv, 1 - %lftr.wideiv = trunc i64 %indvars.iv.next to i32 - %exitcond = icmp eq i32 %lftr.wideiv, %S - br i1 %exitcond, label %for.cond.cleanup.loopexit, label %for.body -} - -; Test that negative offsets are avoided for loads of double. -%s.double = type { double, double, double } -define void @f6(%s.double* nocapture %a, - %s.double* nocapture readonly %b, - i32 zeroext %S) { -; CHECK-Z13-LABEL: f6: -; CHECK-Z13-NOT: -{{[0-9]+}}(%r -entry: - %cmp9 = icmp eq i32 %S, 0 - br i1 %cmp9, label %for.cond.cleanup, label %for.body.preheader - -for.body.preheader: ; preds = %entry - br label %for.body - -for.cond.cleanup.loopexit: ; preds = %for.body - br label %for.cond.cleanup - -for.cond.cleanup: ; preds = %for.cond.cleanup.loopexit, %entry - ret void - -for.body: ; preds = %for.body.preheader, %for.body - %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %for.body.preheader ] - %a1 = getelementptr inbounds %s.double, %s.double* %b, i64 %indvars.iv, i32 0 - %tmp = load double, double* %a1, align 4 - %b4 = getelementptr inbounds %s.double, %s.double* %b, i64 %indvars.iv, i32 1 - %tmp1 = load double, double* %b4, align 4 - %add = fadd double %tmp, %tmp1 - %c = getelementptr inbounds %s.double, %s.double* %b, i64 %indvars.iv, i32 2 - %tmp2 = load double, double* %c, align 4 - %add7 = fadd double %add, %tmp2 - %a10 = getelementptr inbounds %s.double, %s.double* %a, i64 %indvars.iv, i32 0 - store double %add7, double* %a10, align 4 - %indvars.iv.next = add nuw nsw i64 %indvars.iv, 1 - %lftr.wideiv = trunc i64 %indvars.iv.next to i32 - %exitcond = icmp eq i32 %lftr.wideiv, %S - br i1 %exitcond, label %for.cond.cleanup.loopexit, label %for.body -} - -; Test that negative offsets are avoided for memory accesses of vector type. -%s.vec = type { <4 x i32>, <4 x i32>, <4 x i32> } -define void @f7(%s.vec* nocapture %a, - %s.vec* nocapture readonly %b, - i32 zeroext %S) { -; CHECK-Z13-LABEL: f7: -; CHECK-Z13-NOT: -{{[0-9]+}}(%r -entry: - %cmp9 = icmp eq i32 %S, 0 - br i1 %cmp9, label %for.cond.cleanup, label %for.body.preheader - -for.body.preheader: ; preds = %entry - br label %for.body - -for.cond.cleanup.loopexit: ; preds = %for.body - br label %for.cond.cleanup - -for.cond.cleanup: ; preds = %for.cond.cleanup.loopexit, %entry - ret void - -for.body: ; preds = %for.body.preheader, %for.body - %indvars.iv = phi i64 [ %indvars.iv.next, %for.body ], [ 0, %for.body.preheader ] - %a1 = getelementptr inbounds %s.vec, %s.vec* %b, i64 %indvars.iv, i32 0 - %tmp = load <4 x i32>, <4 x i32>* %a1, align 4 - %b4 = getelementptr inbounds %s.vec, %s.vec* %b, i64 %indvars.iv, i32 1 - %tmp1 = load <4 x i32>, <4 x i32>* %b4, align 4 - %add = add <4 x i32> %tmp1, %tmp - %c = getelementptr inbounds %s.vec, %s.vec* %b, i64 %indvars.iv, i32 2 - %tmp2 = load <4 x i32>, <4 x i32>* %c, align 4 - %add7 = add <4 x i32> %add, %tmp2 - %a10 = getelementptr inbounds %s.vec, %s.vec* %a, i64 %indvars.iv, i32 0 - store <4 x i32> %add7, <4 x i32>* %a10, align 4 - %indvars.iv.next = add nuw nsw i64 %indvars.iv, 1 - %lftr.wideiv = trunc i64 %indvars.iv.next to i32 - %exitcond = icmp eq i32 %lftr.wideiv, %S - br i1 %exitcond, label %for.cond.cleanup.loopexit, label %for.body -} |
