diff options
Diffstat (limited to 'test/CodeGen/X86/bmi-intrinsics-fast-isel-x86_64.ll')
| -rw-r--r-- | test/CodeGen/X86/bmi-intrinsics-fast-isel-x86_64.ll | 165 |
1 files changed, 0 insertions, 165 deletions
diff --git a/test/CodeGen/X86/bmi-intrinsics-fast-isel-x86_64.ll b/test/CodeGen/X86/bmi-intrinsics-fast-isel-x86_64.ll deleted file mode 100644 index 8b15a1591b67..000000000000 --- a/test/CodeGen/X86/bmi-intrinsics-fast-isel-x86_64.ll +++ /dev/null @@ -1,165 +0,0 @@ -; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py -; RUN: llc < %s -fast-isel -mtriple=x86_64-unknown-unknown -mattr=+bmi | FileCheck %s --check-prefix=X64 - -; NOTE: This should use IR equivalent to what is generated by clang/test/CodeGen/bmi-builtins.c - -; -; AMD Intrinsics -; - -define i64 @test__andn_u64(i64 %a0, i64 %a1) { -; X64-LABEL: test__andn_u64: -; X64: # BB#0: -; X64-NEXT: xorq $-1, %rdi -; X64-NEXT: andq %rsi, %rdi -; X64-NEXT: movq %rdi, %rax -; X64-NEXT: retq - %xor = xor i64 %a0, -1 - %res = and i64 %xor, %a1 - ret i64 %res -} - -define i64 @test__bextr_u64(i64 %a0, i64 %a1) { -; X64-LABEL: test__bextr_u64: -; X64: # BB#0: -; X64-NEXT: bextrq %rsi, %rdi, %rax -; X64-NEXT: retq - %res = call i64 @llvm.x86.bmi.bextr.64(i64 %a0, i64 %a1) - ret i64 %res -} - -define i64 @test__blsi_u64(i64 %a0) { -; X64-LABEL: test__blsi_u64: -; X64: # BB#0: -; X64-NEXT: xorl %eax, %eax -; X64-NEXT: subq %rdi, %rax -; X64-NEXT: andq %rdi, %rax -; X64-NEXT: retq - %neg = sub i64 0, %a0 - %res = and i64 %a0, %neg - ret i64 %res -} - -define i64 @test__blsmsk_u64(i64 %a0) { -; X64-LABEL: test__blsmsk_u64: -; X64: # BB#0: -; X64-NEXT: movq %rdi, %rax -; X64-NEXT: subq $1, %rax -; X64-NEXT: xorq %rdi, %rax -; X64-NEXT: retq - %dec = sub i64 %a0, 1 - %res = xor i64 %a0, %dec - ret i64 %res -} - -define i64 @test__blsr_u64(i64 %a0) { -; X64-LABEL: test__blsr_u64: -; X64: # BB#0: -; X64-NEXT: movq %rdi, %rax -; X64-NEXT: subq $1, %rax -; X64-NEXT: andq %rdi, %rax -; X64-NEXT: retq - %dec = sub i64 %a0, 1 - %res = and i64 %a0, %dec - ret i64 %res -} - -define i64 @test__tzcnt_u64(i64 %a0) { -; X64-LABEL: test__tzcnt_u64: -; X64: # BB#0: -; X64-NEXT: movl $64, %ecx -; X64-NEXT: tzcntq %rdi, %rax -; X64-NEXT: cmovbq %rcx, %rax -; X64-NEXT: retq - %cmp = icmp ne i64 %a0, 0 - %cttz = call i64 @llvm.cttz.i64(i64 %a0, i1 true) - %res = select i1 %cmp, i64 %cttz, i64 64 - ret i64 %res -} - -; -; Intel intrinsics -; - -define i64 @test_andn_u64(i64 %a0, i64 %a1) { -; X64-LABEL: test_andn_u64: -; X64: # BB#0: -; X64-NEXT: xorq $-1, %rdi -; X64-NEXT: andq %rsi, %rdi -; X64-NEXT: movq %rdi, %rax -; X64-NEXT: retq - %xor = xor i64 %a0, -1 - %res = and i64 %xor, %a1 - ret i64 %res -} - -define i64 @test_bextr_u64(i64 %a0, i32 %a1, i32 %a2) { -; X64-LABEL: test_bextr_u64: -; X64: # BB#0: -; X64-NEXT: andl $255, %esi -; X64-NEXT: andl $255, %edx -; X64-NEXT: shll $8, %edx -; X64-NEXT: orl %esi, %edx -; X64-NEXT: movl %edx, %eax -; X64-NEXT: bextrq %rax, %rdi, %rax -; X64-NEXT: retq - %and1 = and i32 %a1, 255 - %and2 = and i32 %a2, 255 - %shl = shl i32 %and2, 8 - %or = or i32 %and1, %shl - %zext = zext i32 %or to i64 - %res = call i64 @llvm.x86.bmi.bextr.64(i64 %a0, i64 %zext) - ret i64 %res -} - -define i64 @test_blsi_u64(i64 %a0) { -; X64-LABEL: test_blsi_u64: -; X64: # BB#0: -; X64-NEXT: xorl %eax, %eax -; X64-NEXT: subq %rdi, %rax -; X64-NEXT: andq %rdi, %rax -; X64-NEXT: retq - %neg = sub i64 0, %a0 - %res = and i64 %a0, %neg - ret i64 %res -} - -define i64 @test_blsmsk_u64(i64 %a0) { -; X64-LABEL: test_blsmsk_u64: -; X64: # BB#0: -; X64-NEXT: movq %rdi, %rax -; X64-NEXT: subq $1, %rax -; X64-NEXT: xorq %rdi, %rax -; X64-NEXT: retq - %dec = sub i64 %a0, 1 - %res = xor i64 %a0, %dec - ret i64 %res -} - -define i64 @test_blsr_u64(i64 %a0) { -; X64-LABEL: test_blsr_u64: -; X64: # BB#0: -; X64-NEXT: movq %rdi, %rax -; X64-NEXT: subq $1, %rax -; X64-NEXT: andq %rdi, %rax -; X64-NEXT: retq - %dec = sub i64 %a0, 1 - %res = and i64 %a0, %dec - ret i64 %res -} - -define i64 @test_tzcnt_u64(i64 %a0) { -; X64-LABEL: test_tzcnt_u64: -; X64: # BB#0: -; X64-NEXT: movl $64, %ecx -; X64-NEXT: tzcntq %rdi, %rax -; X64-NEXT: cmovbq %rcx, %rax -; X64-NEXT: retq - %cmp = icmp ne i64 %a0, 0 - %cttz = call i64 @llvm.cttz.i64(i64 %a0, i1 true) - %res = select i1 %cmp, i64 %cttz, i64 64 - ret i64 %res -} - -declare i64 @llvm.cttz.i64(i64, i1) -declare i64 @llvm.x86.bmi.bextr.64(i64, i64) |
