aboutsummaryrefslogtreecommitdiff
path: root/test/CodeGen/AMDGPU/fdiv.f16.ll
diff options
context:
space:
mode:
Diffstat (limited to 'test/CodeGen/AMDGPU/fdiv.f16.ll')
-rw-r--r--test/CodeGen/AMDGPU/fdiv.f16.ll249
1 files changed, 0 insertions, 249 deletions
diff --git a/test/CodeGen/AMDGPU/fdiv.f16.ll b/test/CodeGen/AMDGPU/fdiv.f16.ll
deleted file mode 100644
index 7f84e973c958..000000000000
--- a/test/CodeGen/AMDGPU/fdiv.f16.ll
+++ /dev/null
@@ -1,249 +0,0 @@
-; RUN: llc -march=amdgcn -verify-machineinstrs < %s | FileCheck -check-prefix=GCN -check-prefix=SI %s
-; RUN: llc -march=amdgcn -mcpu=fiji -mattr=-flat-for-global -mattr=+fp16-denormals -verify-machineinstrs < %s | FileCheck -check-prefix=GCN -check-prefix=VI %s
-; RUN: llc -march=amdgcn -mcpu=fiji -mattr=-flat-for-global -mattr=-fp16-denormals -verify-machineinstrs < %s | FileCheck -check-prefix=GCN -check-prefix=VI %s
-
-; Make sure fdiv is promoted to f32.
-
-; GCN-LABEL: {{^}}v_fdiv_f16
-; SI: v_cvt_f32_f16
-; SI: v_cvt_f32_f16
-; SI: v_div_scale_f32
-; SI-DAG: v_div_scale_f32
-; SI-DAG: v_rcp_f32
-; SI: v_fma_f32
-; SI: v_fma_f32
-; SI: v_mul_f32
-; SI: v_fma_f32
-; SI: v_fma_f32
-; SI: v_fma_f32
-; SI: v_div_fmas_f32
-; SI: v_div_fixup_f32
-; SI: v_cvt_f16_f32
-
-; VI: flat_load_ushort [[LHS:v[0-9]+]]
-; VI: flat_load_ushort [[RHS:v[0-9]+]]
-
-; VI-DAG: v_cvt_f32_f16_e32 [[CVT_LHS:v[0-9]+]], [[LHS]]
-; VI-DAG: v_cvt_f32_f16_e32 [[CVT_RHS:v[0-9]+]], [[RHS]]
-
-; VI-DAG: v_rcp_f32_e32 [[RCP_RHS:v[0-9]+]], [[CVT_RHS]]
-; VI: v_mul_f32_e32 [[MUL:v[0-9]+]], [[RCP_RHS]], [[CVT_LHS]]
-; VI: v_cvt_f16_f32_e32 [[CVT_BACK:v[0-9]+]], [[MUL]]
-; VI: v_div_fixup_f16 [[RESULT:v[0-9]+]], [[CVT_BACK]], [[RHS]], [[LHS]]
-; VI: flat_store_short v{{\[[0-9]+:[0-9]+\]}}, [[RESULT]]
-define amdgpu_kernel void @v_fdiv_f16(
- half addrspace(1)* %r,
- half addrspace(1)* %a,
- half addrspace(1)* %b) #0 {
-entry:
- %tid = call i32 @llvm.amdgcn.workitem.id.x()
- %tid.ext = sext i32 %tid to i64
- %gep.a = getelementptr inbounds half, half addrspace(1)* %a, i64 %tid.ext
- %gep.b = getelementptr inbounds half, half addrspace(1)* %b, i64 %tid.ext
- %gep.r = getelementptr inbounds half, half addrspace(1)* %r, i64 %tid.ext
- %a.val = load volatile half, half addrspace(1)* %gep.a
- %b.val = load volatile half, half addrspace(1)* %gep.b
- %r.val = fdiv half %a.val, %b.val
- store half %r.val, half addrspace(1)* %gep.r
- ret void
-}
-
-; GCN-LABEL: {{^}}v_rcp_f16:
-; VI: flat_load_ushort [[VAL:v[0-9]+]]
-; VI-NOT: [[VAL]]
-; VI: v_rcp_f16_e32 [[RESULT:v[0-9]+]], [[VAL]]
-; VI-NOT: [[RESULT]]
-; VI: flat_store_short v{{\[[0-9]+:[0-9]+\]}}, [[RESULT]]
-define amdgpu_kernel void @v_rcp_f16(half addrspace(1)* %r, half addrspace(1)* %b) #0 {
-entry:
- %tid = call i32 @llvm.amdgcn.workitem.id.x()
- %tid.ext = sext i32 %tid to i64
- %gep.b = getelementptr inbounds half, half addrspace(1)* %b, i64 %tid.ext
- %gep.r = getelementptr inbounds half, half addrspace(1)* %r, i64 %tid.ext
- %b.val = load volatile half, half addrspace(1)* %gep.b
- %r.val = fdiv half 1.0, %b.val
- store half %r.val, half addrspace(1)* %gep.r
- ret void
-}
-
-; GCN-LABEL: {{^}}v_rcp_f16_abs:
-; VI: flat_load_ushort [[VAL:v[0-9]+]]
-; VI-NOT: [[VAL]]
-; VI: v_rcp_f16_e64 [[RESULT:v[0-9]+]], |[[VAL]]|
-; VI-NOT: [RESULT]]
-; VI: flat_store_short v{{\[[0-9]+:[0-9]+\]}}, [[RESULT]]
-define amdgpu_kernel void @v_rcp_f16_abs(half addrspace(1)* %r, half addrspace(1)* %b) #0 {
-entry:
- %tid = call i32 @llvm.amdgcn.workitem.id.x()
- %tid.ext = sext i32 %tid to i64
- %gep.b = getelementptr inbounds half, half addrspace(1)* %b, i64 %tid.ext
- %gep.r = getelementptr inbounds half, half addrspace(1)* %r, i64 %tid.ext
- %b.val = load volatile half, half addrspace(1)* %gep.b
- %b.abs = call half @llvm.fabs.f16(half %b.val)
- %r.val = fdiv half 1.0, %b.abs
- store half %r.val, half addrspace(1)* %gep.r
- ret void
-}
-
-; GCN-LABEL: {{^}}v_rcp_f16_arcp:
-; VI: flat_load_ushort [[VAL:v[0-9]+]]
-; VI-NOT: [[VAL]]
-; VI: v_rcp_f16_e32 [[RESULT:v[0-9]+]], [[VAL]]
-; VI-NOT: [[RESULT]]
-; VI: flat_store_short v{{\[[0-9]+:[0-9]+\]}}, [[RESULT]]
-define amdgpu_kernel void @v_rcp_f16_arcp(half addrspace(1)* %r, half addrspace(1)* %b) #0 {
-entry:
- %tid = call i32 @llvm.amdgcn.workitem.id.x()
- %tid.ext = sext i32 %tid to i64
- %gep.b = getelementptr inbounds half, half addrspace(1)* %b, i64 %tid.ext
- %gep.r = getelementptr inbounds half, half addrspace(1)* %r, i64 %tid.ext
- %b.val = load volatile half, half addrspace(1)* %gep.b
- %r.val = fdiv arcp half 1.0, %b.val
- store half %r.val, half addrspace(1)* %gep.r
- ret void
-}
-
-; GCN-LABEL: {{^}}v_rcp_f16_neg:
-; VI: flat_load_ushort [[VAL:v[0-9]+]]
-; VI-NOT: [[VAL]]
-; VI: v_rcp_f16_e64 [[RESULT:v[0-9]+]], -[[VAL]]
-; VI-NOT: [RESULT]]
-; VI: flat_store_short v{{\[[0-9]+:[0-9]+\]}}, [[RESULT]]
-define amdgpu_kernel void @v_rcp_f16_neg(half addrspace(1)* %r, half addrspace(1)* %b) #0 {
-entry:
- %tid = call i32 @llvm.amdgcn.workitem.id.x()
- %tid.ext = sext i32 %tid to i64
- %gep.b = getelementptr inbounds half, half addrspace(1)* %b, i64 %tid.ext
- %gep.r = getelementptr inbounds half, half addrspace(1)* %r, i64 %tid.ext
- %b.val = load volatile half, half addrspace(1)* %gep.b
- %r.val = fdiv half -1.0, %b.val
- store half %r.val, half addrspace(1)* %gep.r
- ret void
-}
-
-; GCN-LABEL: {{^}}v_rsq_f16:
-; VI: flat_load_ushort [[VAL:v[0-9]+]]
-; VI-NOT: [[VAL]]
-; VI: v_rsq_f16_e32 [[RESULT:v[0-9]+]], [[VAL]]
-; VI-NOT: [RESULT]]
-; VI: flat_store_short v{{\[[0-9]+:[0-9]+\]}}, [[RESULT]]
-define amdgpu_kernel void @v_rsq_f16(half addrspace(1)* %r, half addrspace(1)* %b) #0 {
-entry:
- %tid = call i32 @llvm.amdgcn.workitem.id.x()
- %tid.ext = sext i32 %tid to i64
- %gep.b = getelementptr inbounds half, half addrspace(1)* %b, i64 %tid.ext
- %gep.r = getelementptr inbounds half, half addrspace(1)* %r, i64 %tid.ext
- %b.val = load volatile half, half addrspace(1)* %gep.b
- %b.sqrt = call half @llvm.sqrt.f16(half %b.val)
- %r.val = fdiv half 1.0, %b.sqrt
- store half %r.val, half addrspace(1)* %gep.r
- ret void
-}
-
-; GCN-LABEL: {{^}}v_rsq_f16_neg:
-; VI: flat_load_ushort [[VAL:v[0-9]+]]
-; VI-NOT: [[VAL]]
-; VI: v_sqrt_f16_e32 [[SQRT:v[0-9]+]], [[VAL]]
-; VI-NEXT: v_rcp_f16_e64 [[RESULT:v[0-9]+]], -[[SQRT]]
-; VI-NOT: [RESULT]]
-; VI: flat_store_short v{{\[[0-9]+:[0-9]+\]}}, [[RESULT]]
-define amdgpu_kernel void @v_rsq_f16_neg(half addrspace(1)* %r, half addrspace(1)* %b) #0 {
-entry:
- %tid = call i32 @llvm.amdgcn.workitem.id.x()
- %tid.ext = sext i32 %tid to i64
- %gep.b = getelementptr inbounds half, half addrspace(1)* %b, i64 %tid.ext
- %gep.r = getelementptr inbounds half, half addrspace(1)* %r, i64 %tid.ext
- %b.val = load volatile half, half addrspace(1)* %gep.b
- %b.sqrt = call half @llvm.sqrt.f16(half %b.val)
- %r.val = fdiv half -1.0, %b.sqrt
- store half %r.val, half addrspace(1)* %gep.r
- ret void
-}
-
-; GCN-LABEL: {{^}}v_fdiv_f16_arcp:
-; VI: flat_load_ushort [[LHS:v[0-9]+]]
-; VI: flat_load_ushort [[RHS:v[0-9]+]]
-
-; VI: v_rcp_f16_e32 [[RCP:v[0-9]+]], [[RHS]]
-; VI: v_mul_f16_e32 [[RESULT:v[0-9]+]], [[RCP]], [[LHS]]
-
-; VI: flat_store_short v{{\[[0-9]+:[0-9]+\]}}, [[RESULT]]
-define amdgpu_kernel void @v_fdiv_f16_arcp(half addrspace(1)* %r, half addrspace(1)* %a, half addrspace(1)* %b) #0 {
-entry:
- %tid = call i32 @llvm.amdgcn.workitem.id.x()
- %tid.ext = sext i32 %tid to i64
- %gep.a = getelementptr inbounds half, half addrspace(1)* %a, i64 %tid.ext
- %gep.b = getelementptr inbounds half, half addrspace(1)* %b, i64 %tid.ext
- %gep.r = getelementptr inbounds half, half addrspace(1)* %r, i64 %tid.ext
- %a.val = load volatile half, half addrspace(1)* %gep.a
- %b.val = load volatile half, half addrspace(1)* %gep.b
- %r.val = fdiv arcp half %a.val, %b.val
- store half %r.val, half addrspace(1)* %gep.r
- ret void
-}
-
-; GCN-LABEL: {{^}}v_fdiv_f16_unsafe:
-; VI: flat_load_ushort [[LHS:v[0-9]+]]
-; VI: flat_load_ushort [[RHS:v[0-9]+]]
-
-; VI: v_rcp_f16_e32 [[RCP:v[0-9]+]], [[RHS]]
-; VI: v_mul_f16_e32 [[RESULT:v[0-9]+]], [[RCP]], [[LHS]]
-
-; VI: flat_store_short v{{\[[0-9]+:[0-9]+\]}}, [[RESULT]]
-define amdgpu_kernel void @v_fdiv_f16_unsafe(half addrspace(1)* %r, half addrspace(1)* %a, half addrspace(1)* %b) #2 {
-entry:
- %tid = call i32 @llvm.amdgcn.workitem.id.x()
- %tid.ext = sext i32 %tid to i64
- %gep.a = getelementptr inbounds half, half addrspace(1)* %a, i64 %tid.ext
- %gep.b = getelementptr inbounds half, half addrspace(1)* %b, i64 %tid.ext
- %gep.r = getelementptr inbounds half, half addrspace(1)* %r, i64 %tid.ext
- %a.val = load volatile half, half addrspace(1)* %gep.a
- %b.val = load volatile half, half addrspace(1)* %gep.b
- %r.val = fdiv half %a.val, %b.val
- store half %r.val, half addrspace(1)* %gep.r
- ret void
-}
-
-; FUNC-LABEL: {{^}}div_arcp_2_x_pat_f16:
-; SI: v_mul_f32_e32 v{{[0-9]+}}, 0.5, v{{[0-9]+}}
-
-; VI: v_mul_f16_e32 [[MUL:v[0-9]+]], 0.5, v{{[0-9]+}}
-; VI: buffer_store_short [[MUL]]
-define amdgpu_kernel void @div_arcp_2_x_pat_f16(half addrspace(1)* %out) #0 {
- %x = load half, half addrspace(1)* undef
- %rcp = fdiv arcp half %x, 2.0
- store half %rcp, half addrspace(1)* %out, align 4
- ret void
-}
-
-; FUNC-LABEL: {{^}}div_arcp_k_x_pat_f16:
-; SI: v_mul_f32_e32 v{{[0-9]+}}, 0x3dcccccd, v{{[0-9]+}}
-
-; VI: v_mul_f16_e32 [[MUL:v[0-9]+]], 0x2e66, v{{[0-9]+}}
-; VI: buffer_store_short [[MUL]]
-define amdgpu_kernel void @div_arcp_k_x_pat_f16(half addrspace(1)* %out) #0 {
- %x = load half, half addrspace(1)* undef
- %rcp = fdiv arcp half %x, 10.0
- store half %rcp, half addrspace(1)* %out, align 4
- ret void
-}
-
-; FUNC-LABEL: {{^}}div_arcp_neg_k_x_pat_f16:
-; SI: v_mul_f32_e32 v{{[0-9]+}}, 0xbdcccccd, v{{[0-9]+}}
-
-; VI: v_mul_f16_e32 [[MUL:v[0-9]+]], 0xae66, v{{[0-9]+}}
-; VI: buffer_store_short [[MUL]]
-define amdgpu_kernel void @div_arcp_neg_k_x_pat_f16(half addrspace(1)* %out) #0 {
- %x = load half, half addrspace(1)* undef
- %rcp = fdiv arcp half %x, -10.0
- store half %rcp, half addrspace(1)* %out, align 4
- ret void
-}
-
-declare i32 @llvm.amdgcn.workitem.id.x() #1
-declare half @llvm.sqrt.f16(half) #1
-declare half @llvm.fabs.f16(half) #1
-
-attributes #0 = { nounwind }
-attributes #1 = { nounwind readnone }
-attributes #2 = { nounwind "unsafe-fp-math"="true" }