; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mcpu=verde -verify-machineinstrs < %s | FileCheck -allow-deprecated-dag-overlap -check-prefixes=GCN,SI,FUNC %s ; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mcpu=tonga -mattr=-flat-for-global -verify-machineinstrs < %s | FileCheck -allow-deprecated-dag-overlap -check-prefixes=GCN,VI,FUNC %s ; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mcpu=gfx900 -mattr=-flat-for-global -verify-machineinstrs < %s | FileCheck -allow-deprecated-dag-overlap -check-prefixes=FUNC,GFX9PLUS %s ; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mcpu=gfx1010 -mattr=-flat-for-global -verify-machineinstrs < %s | FileCheck -allow-deprecated-dag-overlap -check-prefixes=FUNC,GFX9PLUS %s ; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mcpu=gfx1100 -mattr=-flat-for-global -verify-machineinstrs < %s | FileCheck -allow-deprecated-dag-overlap -check-prefixes=FUNC,GFX9PLUS %s ; RUN: llc -amdgpu-scalarize-global-loads=false -march=r600 -mcpu=redwood < %s | FileCheck -allow-deprecated-dag-overlap -check-prefixes=EG,FUNC %s ; mul24 and mad24 are affected ; FUNC-LABEL: {{^}}test_mul_v2i32: ; EG: MULLO_INT {{\*? *}}T{{[0-9]+\.[XYZW], T[0-9]+\.[XYZW], T[0-9]+\.[XYZW]}} ; EG: MULLO_INT {{\*? *}}T{{[0-9]+\.[XYZW], T[0-9]+\.[XYZW], T[0-9]+\.[XYZW]}} ; GCN: v_mul_lo_u32 v{{[0-9]+, v[0-9]+, v[0-9]+}} ; GCN: v_mul_lo_u32 v{{[0-9]+, v[0-9]+, v[0-9]+}} define amdgpu_kernel void @test_mul_v2i32(ptr addrspace(1) %out, ptr addrspace(1) %in) { %b_ptr = getelementptr <2 x i32>, ptr addrspace(1) %in, i32 1 %a = load <2 x i32>, ptr addrspace(1) %in %b = load <2 x i32>, ptr addrspace(1) %b_ptr %result = mul <2 x i32> %a, %b store <2 x i32> %result, ptr addrspace(1) %out ret void } ; FUNC-LABEL: {{^}}v_mul_v4i32: ; EG: MULLO_INT {{\*? *}}T{{[0-9]+\.[XYZW], T[0-9]+\.[XYZW], T[0-9]+\.[XYZW]}} ; EG: MULLO_INT {{\*? *}}T{{[0-9]+\.[XYZW], T[0-9]+\.[XYZW], T[0-9]+\.[XYZW]}} ; EG: MULLO_INT {{\*? *}}T{{[0-9]+\.[XYZW], T[0-9]+\.[XYZW], T[0-9]+\.[XYZW]}} ; EG: MULLO_INT {{\*? *}}T{{[0-9]+\.[XYZW], T[0-9]+\.[XYZW], T[0-9]+\.[XYZW]}} ; GCN: v_mul_lo_u32 v{{[0-9]+, v[0-9]+, v[0-9]+}} ; GCN: v_mul_lo_u32 v{{[0-9]+, v[0-9]+, v[0-9]+}} ; GCN: v_mul_lo_u32 v{{[0-9]+, v[0-9]+, v[0-9]+}} ; GCN: v_mul_lo_u32 v{{[0-9]+, v[0-9]+, v[0-9]+}} define amdgpu_kernel void @v_mul_v4i32(ptr addrspace(1) %out, ptr addrspace(1) %in) { %b_ptr = getelementptr <4 x i32>, ptr addrspace(1) %in, i32 1 %a = load <4 x i32>, ptr addrspace(1) %in %b = load <4 x i32>, ptr addrspace(1) %b_ptr %result = mul <4 x i32> %a, %b store <4 x i32> %result, ptr addrspace(1) %out ret void } ; FUNC-LABEL: {{^}}s_trunc_i64_mul_to_i32: ; GCN: s_load_dword ; GCN: s_load_dword ; GCN: s_mul_i32 ; GCN: buffer_store_dword define amdgpu_kernel void @s_trunc_i64_mul_to_i32(ptr addrspace(1) %out, i64 %a, i64 %b) { %mul = mul i64 %b, %a %trunc = trunc i64 %mul to i32 store i32 %trunc, ptr addrspace(1) %out, align 8 ret void } ; FUNC-LABEL: {{^}}v_trunc_i64_mul_to_i32: ; GCN: s_load_dword ; GCN: s_load_dword ; GCN: v_mul_lo_u32 ; GCN: buffer_store_dword define amdgpu_kernel void @v_trunc_i64_mul_to_i32(ptr addrspace(1) %out, ptr addrspace(1) %aptr, ptr addrspace(1) %bptr) nounwind { %a = load i64, ptr addrspace(1) %aptr, align 8 %b = load i64, ptr addrspace(1) %bptr, align 8 %mul = mul i64 %b, %a %trunc = trunc i64 %mul to i32 store i32 %trunc, ptr addrspace(1) %out, align 8 ret void } ; This 64-bit multiply should just use MUL_HI and MUL_LO, since the top ; 32-bits of both arguments are sign bits. ; FUNC-LABEL: {{^}}mul64_sext_c: ; EG-DAG: MULLO_INT ; EG-DAG: MULHI_INT ; SI-DAG: s_mulk_i32 ; SI-DAG: v_mul_hi_i32 ; VI: v_mad_i64_i32 define amdgpu_kernel void @mul64_sext_c(ptr addrspace(1) %out, i32 %in) { entry: %0 = sext i32 %in to i64 %1 = mul i64 %0, 80 store i64 %1, ptr addrspace(1) %out ret void } ; FUNC-LABEL: {{^}}v_mul64_sext_c: ; EG-DAG: MULLO_INT ; EG-DAG: MULHI_INT ; SI-DAG: v_mul_lo_u32 ; SI-DAG: v_mul_hi_i32 ; VI: v_mad_i64_i32 ; GCN: s_endpgm define amdgpu_kernel void @v_mul64_sext_c(ptr addrspace(1) %out, ptr addrspace(1) %in) { %val = load i32, ptr addrspace(1) %in, align 4 %ext = sext i32 %val to i64 %mul = mul i64 %ext, 80 store i64 %mul, ptr addrspace(1) %out, align 8 ret void } ; FUNC-LABEL: {{^}}v_mul64_sext_inline_imm: ; SI-DAG: v_mul_lo_u32 v{{[0-9]+}}, v{{[0-9]+}}, 9 ; SI-DAG: v_mul_hi_i32 v{{[0-9]+}}, v{{[0-9]+}}, 9 ; VI: v_mad_i64_i32 v[{{[0-9]+}}:{{[0-9]+}}], s[{{[0-9]+}}:{{[0-9]+}}], v{{[0-9]+}}, 9, 0 ; GCN: s_endpgm define amdgpu_kernel void @v_mul64_sext_inline_imm(ptr addrspace(1) %out, ptr addrspace(1) %in) { %val = load i32, ptr addrspace(1) %in, align 4 %ext = sext i32 %val to i64 %mul = mul i64 %ext, 9 store i64 %mul, ptr addrspace(1) %out, align 8 ret void } ; FUNC-LABEL: {{^}}s_mul_i32: ; GCN: s_load_dword [[SRC0:s[0-9]+]], ; GCN: s_load_dword [[SRC1:s[0-9]+]], ; GCN: s_mul_i32 [[SRESULT:s[0-9]+]], [[SRC0]], [[SRC1]] ; GCN: v_mov_b32_e32 [[VRESULT:v[0-9]+]], [[SRESULT]] ; GCN: buffer_store_dword [[VRESULT]], ; GCN: s_endpgm define amdgpu_kernel void @s_mul_i32(ptr addrspace(1) %out, [8 x i32], i32 %a, [8 x i32], i32 %b) nounwind { %mul = mul i32 %a, %b store i32 %mul, ptr addrspace(1) %out, align 4 ret void } ; FUNC-LABEL: {{^}}v_mul_i32: ; GCN: v_mul_lo_u32 v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]+}} define amdgpu_kernel void @v_mul_i32(ptr addrspace(1) %out, ptr addrspace(1) %in) { %b_ptr = getelementptr i32, ptr addrspace(1) %in, i32 1 %a = load i32, ptr addrspace(1) %in %b = load i32, ptr addrspace(1) %b_ptr %result = mul i32 %a, %b store i32 %result, ptr addrspace(1) %out ret void } ; A standard 64-bit multiply. The expansion should be around 6 instructions. ; It would be difficult to match the expansion correctly without writing ; a really complicated list of FileCheck expressions. I don't want ; to confuse people who may 'break' this test with a correct optimization, ; so this test just uses FUNC-LABEL to make sure the compiler does not ; crash with a 'failed to select' error. ; FUNC-LABEL: {{^}}s_mul_i64: ; GFX9PLUS-DAG: s_mul_i32 ; GFX9PLUS-DAG: s_mul_hi_u32 ; GFX9PLUS-DAG: s_mul_i32 ; GFX9PLUS-DAG: s_mul_i32 ; GFX9PLUS: s_endpgm define amdgpu_kernel void @s_mul_i64(ptr addrspace(1) %out, i64 %a, i64 %b) nounwind { %mul = mul i64 %a, %b store i64 %mul, ptr addrspace(1) %out, align 8 ret void } ; FUNC-LABEL: {{^}}v_mul_i64: ; GCN: v_mul_lo_u32 define amdgpu_kernel void @v_mul_i64(ptr addrspace(1) %out, ptr addrspace(1) %aptr, ptr addrspace(1) %bptr) { %a = load i64, ptr addrspace(1) %aptr, align 8 %b = load i64, ptr addrspace(1) %bptr, align 8 %mul = mul i64 %a, %b store i64 %mul, ptr addrspace(1) %out, align 8 ret void } ; FUNC-LABEL: {{^}}mul32_in_branch: ; GCN: s_mul_i32 define amdgpu_kernel void @mul32_in_branch(ptr addrspace(1) %out, ptr addrspace(1) %in, i32 %a, i32 %b, i32 %c) { entry: %0 = icmp eq i32 %a, 0 br i1 %0, label %if, label %else if: %1 = load i32, ptr addrspace(1) %in br label %endif else: %2 = mul i32 %a, %b br label %endif endif: %3 = phi i32 [%1, %if], [%2, %else] store i32 %3, ptr addrspace(1) %out ret void } ; FUNC-LABEL: {{^}}mul64_in_branch: ; SI-DAG: s_mul_i32 ; SI-DAG: v_mul_hi_u32 ; VI: v_mad_u64_u32 ; GCN: s_endpgm define amdgpu_kernel void @mul64_in_branch(ptr addrspace(1) %out, ptr addrspace(1) %in, i64 %a, i64 %b, i64 %c) { entry: %0 = icmp eq i64 %a, 0 br i1 %0, label %if, label %else if: %1 = load i64, ptr addrspace(1) %in br label %endif else: %2 = mul i64 %a, %b br label %endif endif: %3 = phi i64 [%1, %if], [%2, %else] store i64 %3, ptr addrspace(1) %out ret void } ; FIXME: Load dwordx4 ; FUNC-LABEL: {{^}}s_mul_i128: ; GCN: s_load_dwordx4 ; GCN: s_load_dwordx4 ; SI: v_mul_hi_u32 ; SI: v_mul_hi_u32 ; SI: s_mul_i32 ; SI: v_mul_hi_u32 ; SI: s_mul_i32 ; SI: s_mul_i32 ; SI-DAG: s_mul_i32 ; SI-DAG: v_mul_hi_u32 ; SI-DAG: v_mul_hi_u32 ; SI-DAG: s_mul_i32 ; SI-DAG: s_mul_i32 ; SI-DAG: v_mul_hi_u32 ; VI-DAG: v_mad_u64_u32 ; VI-DAG: v_mad_u64_u32 ; VI-DAG: v_mad_u64_u32 ; VI-DAG: v_mad_u64_u32 ; VI-DAG: v_mad_u64_u32 ; VI-DAG: v_mad_u64_u32 ; VI-DAG: s_mul_i32 ; VI-DAG: s_mul_i32 ; VI-DAG: s_mul_i32 ; VI-DAG: s_mul_i32 ; GCN: buffer_store_dwordx4 define amdgpu_kernel void @s_mul_i128(ptr addrspace(1) %out, [8 x i32], i128 %a, [8 x i32], i128 %b) nounwind #0 { %mul = mul i128 %a, %b store i128 %mul, ptr addrspace(1) %out ret void } ; FUNC-LABEL: {{^}}v_mul_i128: ; GCN: {{buffer|flat}}_load_dwordx4 ; GCN: {{buffer|flat}}_load_dwordx4 ; SI-DAG: v_mul_lo_u32 ; SI-DAG: v_mul_hi_u32 ; SI-DAG: v_mul_hi_u32 ; SI-DAG: v_mul_lo_u32 ; SI-DAG: v_mul_hi_u32 ; SI-DAG: v_mul_hi_u32 ; SI-DAG: v_mul_lo_u32 ; SI-DAG: v_mul_lo_u32 ; SI-DAG: v_add_i32_e32 ; SI-DAG: v_mul_hi_u32 ; SI-DAG: v_mul_lo_u32 ; SI-DAG: v_mul_hi_u32 ; SI-DAG: v_mul_lo_u32 ; SI-DAG: v_mul_lo_u32 ; SI-DAG: v_mul_lo_u32 ; SI-DAG: v_mul_lo_u32 ; SI-DAG: v_mul_lo_u32 ; VI-DAG: v_mad_u64_u32 ; VI-DAG: v_mad_u64_u32 ; VI-DAG: v_mad_u64_u32 ; VI-DAG: v_mad_u64_u32 ; VI-DAG: v_mad_u64_u32 ; VI-DAG: v_mad_u64_u32 ; VI-DAG: v_mul_lo_u32 ; VI-DAG: v_mul_lo_u32 ; VI-DAG: v_mul_lo_u32 ; GCN: {{buffer|flat}}_store_dwordx4 define amdgpu_kernel void @v_mul_i128(ptr addrspace(1) %out, ptr addrspace(1) %aptr, ptr addrspace(1) %bptr) #0 { %tid = call i32 @llvm.amdgcn.workitem.id.x() %gep.a = getelementptr inbounds i128, ptr addrspace(1) %aptr, i32 %tid %gep.b = getelementptr inbounds i128, ptr addrspace(1) %bptr, i32 %tid %gep.out = getelementptr inbounds i128, ptr addrspace(1) %bptr, i32 %tid %a = load i128, ptr addrspace(1) %gep.a %b = load i128, ptr addrspace(1) %gep.b %mul = mul i128 %a, %b store i128 %mul, ptr addrspace(1) %gep.out ret void } declare i32 @llvm.amdgcn.workitem.id.x() #1 attributes #0 = { nounwind } attributes #1 = { nounwind readnone}