1; NOTE: Assertions have been autogenerated by utils/update_mir_test_checks.py 2; RUN: llc -global-isel -amdgpu-fixed-function-abi -mtriple=amdgcn-amd-amdhsa -mcpu=gfx900 -stop-after=irtranslator -verify-machineinstrs < %s | FileCheck -check-prefix=GCN %s 3 4declare hidden void @external_void_func_sret_struct_i8_i32_byval_struct_i8_i32({ i8, i32 } addrspace(5)* sret({ i8, i32 }), { i8, i32 } addrspace(5)* byval({ i8, i32 })) #0 5 6define amdgpu_kernel void @test_call_external_void_func_sret_struct_i8_i32_byval_struct_i8_i32(i32) #0 { 7 ; GCN-LABEL: name: test_call_external_void_func_sret_struct_i8_i32_byval_struct_i8_i32 8 ; GCN: bb.1 (%ir-block.1): 9 ; GCN: liveins: $sgpr14, $sgpr15, $sgpr16, $vgpr0, $vgpr1, $vgpr2, $sgpr4_sgpr5, $sgpr6_sgpr7, $sgpr8_sgpr9, $sgpr10_sgpr11 10 ; GCN: [[COPY:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr2 11 ; GCN: [[COPY1:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr1 12 ; GCN: [[COPY2:%[0-9]+]]:vgpr_32(s32) = COPY $vgpr0 13 ; GCN: [[COPY3:%[0-9]+]]:sgpr_32 = COPY $sgpr16 14 ; GCN: [[COPY4:%[0-9]+]]:sgpr_32 = COPY $sgpr15 15 ; GCN: [[COPY5:%[0-9]+]]:sgpr_32 = COPY $sgpr14 16 ; GCN: [[COPY6:%[0-9]+]]:sgpr_64 = COPY $sgpr10_sgpr11 17 ; GCN: [[COPY7:%[0-9]+]]:sgpr_64 = COPY $sgpr6_sgpr7 18 ; GCN: [[COPY8:%[0-9]+]]:sgpr_64 = COPY $sgpr4_sgpr5 19 ; GCN: [[COPY9:%[0-9]+]]:_(p4) = COPY $sgpr8_sgpr9 20 ; GCN: [[C:%[0-9]+]]:_(s8) = G_CONSTANT i8 3 21 ; GCN: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 8 22 ; GCN: [[DEF:%[0-9]+]]:_(p1) = G_IMPLICIT_DEF 23 ; GCN: [[COPY10:%[0-9]+]]:_(p1) = COPY [[DEF]](p1) 24 ; GCN: [[FRAME_INDEX:%[0-9]+]]:_(p5) = G_FRAME_INDEX %stack.0.in.val 25 ; GCN: [[FRAME_INDEX1:%[0-9]+]]:_(p5) = G_FRAME_INDEX %stack.1.out.val 26 ; GCN: [[INT:%[0-9]+]]:_(p4) = G_INTRINSIC intrinsic(@llvm.amdgcn.kernarg.segment.ptr) 27 ; GCN: [[C2:%[0-9]+]]:_(s32) = G_CONSTANT i32 4 28 ; GCN: [[PTR_ADD:%[0-9]+]]:_(p5) = G_PTR_ADD [[FRAME_INDEX]], [[C2]](s32) 29 ; GCN: G_STORE [[C]](s8), [[FRAME_INDEX]](p5) :: (store 1 into %ir.in.gep01, addrspace 5) 30 ; GCN: G_STORE [[C1]](s32), [[PTR_ADD]](p5) :: (store 4 into %ir.in.gep1, addrspace 5) 31 ; GCN: ADJCALLSTACKUP 0, 0, implicit-def $scc 32 ; GCN: [[GV:%[0-9]+]]:sreg_64(p0) = G_GLOBAL_VALUE @external_void_func_sret_struct_i8_i32_byval_struct_i8_i32 33 ; GCN: [[COPY11:%[0-9]+]]:_(p4) = COPY [[COPY8]] 34 ; GCN: [[COPY12:%[0-9]+]]:_(p4) = COPY [[COPY7]] 35 ; GCN: [[COPY13:%[0-9]+]]:_(p4) = COPY [[COPY9]](p4) 36 ; GCN: [[C3:%[0-9]+]]:_(s64) = G_CONSTANT i64 8 37 ; GCN: [[PTR_ADD1:%[0-9]+]]:_(p4) = G_PTR_ADD [[COPY13]], [[C3]](s64) 38 ; GCN: [[COPY14:%[0-9]+]]:_(s64) = COPY [[COPY6]] 39 ; GCN: [[COPY15:%[0-9]+]]:_(s32) = COPY [[COPY5]] 40 ; GCN: [[COPY16:%[0-9]+]]:_(s32) = COPY [[COPY4]] 41 ; GCN: [[COPY17:%[0-9]+]]:_(s32) = COPY [[COPY3]] 42 ; GCN: [[COPY18:%[0-9]+]]:_(s32) = COPY [[COPY2]](s32) 43 ; GCN: [[COPY19:%[0-9]+]]:_(s32) = COPY [[COPY1]](s32) 44 ; GCN: [[C4:%[0-9]+]]:_(s32) = G_CONSTANT i32 10 45 ; GCN: [[SHL:%[0-9]+]]:_(s32) = G_SHL [[COPY19]], [[C4]](s32) 46 ; GCN: [[OR:%[0-9]+]]:_(s32) = G_OR [[COPY18]], [[SHL]] 47 ; GCN: [[COPY20:%[0-9]+]]:_(s32) = COPY [[COPY]](s32) 48 ; GCN: [[C5:%[0-9]+]]:_(s32) = G_CONSTANT i32 20 49 ; GCN: [[SHL1:%[0-9]+]]:_(s32) = G_SHL [[COPY20]], [[C5]](s32) 50 ; GCN: [[OR1:%[0-9]+]]:_(s32) = G_OR [[OR]], [[SHL1]] 51 ; GCN: $vgpr0 = COPY [[FRAME_INDEX1]](p5) 52 ; GCN: [[COPY21:%[0-9]+]]:_(p5) = COPY $sp_reg 53 ; GCN: [[C6:%[0-9]+]]:_(s32) = G_CONSTANT i32 0 54 ; GCN: [[PTR_ADD2:%[0-9]+]]:_(p5) = G_PTR_ADD [[COPY21]], [[C6]](s32) 55 ; GCN: G_STORE [[FRAME_INDEX]](p5), [[PTR_ADD2]](p5) :: (store 4 into stack, align 16, addrspace 5) 56 ; GCN: [[COPY22:%[0-9]+]]:_(<4 x s32>) = COPY $private_rsrc_reg 57 ; GCN: $sgpr0_sgpr1_sgpr2_sgpr3 = COPY [[COPY22]](<4 x s32>) 58 ; GCN: $sgpr4_sgpr5 = COPY [[COPY11]](p4) 59 ; GCN: $sgpr6_sgpr7 = COPY [[COPY12]](p4) 60 ; GCN: $sgpr8_sgpr9 = COPY [[PTR_ADD1]](p4) 61 ; GCN: $sgpr10_sgpr11 = COPY [[COPY14]](s64) 62 ; GCN: $sgpr12 = COPY [[COPY15]](s32) 63 ; GCN: $sgpr13 = COPY [[COPY16]](s32) 64 ; GCN: $sgpr14 = COPY [[COPY17]](s32) 65 ; GCN: $vgpr31 = COPY [[OR1]](s32) 66 ; GCN: $sgpr30_sgpr31 = SI_CALL [[GV]](p0), @external_void_func_sret_struct_i8_i32_byval_struct_i8_i32, csr_amdgpu_highregs, implicit $vgpr0, implicit $sgpr0_sgpr1_sgpr2_sgpr3, implicit $sgpr4_sgpr5, implicit $sgpr6_sgpr7, implicit $sgpr8_sgpr9, implicit $sgpr10_sgpr11, implicit $sgpr12, implicit $sgpr13, implicit $sgpr14, implicit $vgpr31 67 ; GCN: ADJCALLSTACKDOWN 0, 8, implicit-def $scc 68 ; GCN: [[PTR_ADD3:%[0-9]+]]:_(p5) = G_PTR_ADD [[FRAME_INDEX1]], [[C2]](s32) 69 ; GCN: [[LOAD:%[0-9]+]]:_(s8) = G_LOAD [[FRAME_INDEX1]](p5) :: (dereferenceable load 1 from %ir.out.gep02, addrspace 5) 70 ; GCN: [[LOAD1:%[0-9]+]]:_(s32) = G_LOAD [[PTR_ADD3]](p5) :: (dereferenceable load 4 from %ir.out.gep1, addrspace 5) 71 ; GCN: G_STORE [[LOAD]](s8), [[DEF]](p1) :: (volatile store 1 into `i8 addrspace(1)* undef`, addrspace 1) 72 ; GCN: G_STORE [[LOAD1]](s32), [[COPY10]](p1) :: (volatile store 4 into `i32 addrspace(1)* undef`, addrspace 1) 73 ; GCN: S_ENDPGM 0 74 %in.val = alloca { i8, i32 }, align 4, addrspace(5) 75 %out.val = alloca { i8, i32 }, align 4, addrspace(5) 76 %in.gep0 = getelementptr inbounds { i8, i32 }, { i8, i32 } addrspace(5)* %in.val, i32 0, i32 0 77 %in.gep1 = getelementptr inbounds { i8, i32 }, { i8, i32 } addrspace(5)* %in.val, i32 0, i32 1 78 store i8 3, i8 addrspace(5)* %in.gep0 79 store i32 8, i32 addrspace(5)* %in.gep1 80 call void @external_void_func_sret_struct_i8_i32_byval_struct_i8_i32({ i8, i32 } addrspace(5)* %out.val, { i8, i32 } addrspace(5)* %in.val) 81 %out.gep0 = getelementptr inbounds { i8, i32 }, { i8, i32 } addrspace(5)* %out.val, i32 0, i32 0 82 %out.gep1 = getelementptr inbounds { i8, i32 }, { i8, i32 } addrspace(5)* %out.val, i32 0, i32 1 83 %out.val0 = load i8, i8 addrspace(5)* %out.gep0 84 %out.val1 = load i32, i32 addrspace(5)* %out.gep1 85 store volatile i8 %out.val0, i8 addrspace(1)* undef 86 store volatile i32 %out.val1, i32 addrspace(1)* undef 87 ret void 88} 89