/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/MSP430/ |
D | BranchSelector.ll | 12 store volatile i16 11, i16* @reg, align 2 13 store volatile i16 13, i16* @reg, align 2 14 store volatile i16 17, i16* @reg, align 2 15 store volatile i16 11, i16* @reg, align 2 16 store volatile i16 13, i16* @reg, align 2 17 store volatile i16 17, i16* @reg, align 2 18 store volatile i16 11, i16* @reg, align 2 19 store volatile i16 13, i16* @reg, align 2 20 store volatile i16 17, i16* @reg, align 2 21 store volatile i16 11, i16* @reg, align 2 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/SystemZ/ |
D | frame-19.ll | 30 %v0 = load volatile <16 x i8>, <16 x i8> *%ptr 31 %v1 = load volatile <16 x i8>, <16 x i8> *%ptr 32 %v2 = load volatile <16 x i8>, <16 x i8> *%ptr 33 %v3 = load volatile <16 x i8>, <16 x i8> *%ptr 34 %v4 = load volatile <16 x i8>, <16 x i8> *%ptr 35 %v5 = load volatile <16 x i8>, <16 x i8> *%ptr 36 %v6 = load volatile <16 x i8>, <16 x i8> *%ptr 37 %v7 = load volatile <16 x i8>, <16 x i8> *%ptr 38 %v8 = load volatile <16 x i8>, <16 x i8> *%ptr 39 %v9 = load volatile <16 x i8>, <16 x i8> *%ptr [all …]
|
D | int-conv-11.ll | 11 %val0 = load volatile i32, i32 *%ptr 12 %val1 = load volatile i32, i32 *%ptr 13 %val2 = load volatile i32, i32 *%ptr 14 %val3 = load volatile i32, i32 *%ptr 15 %val4 = load volatile i32, i32 *%ptr 16 %val5 = load volatile i32, i32 *%ptr 17 %val6 = load volatile i32, i32 *%ptr 18 %val7 = load volatile i32, i32 *%ptr 19 %val8 = load volatile i32, i32 *%ptr 20 %val9 = load volatile i32, i32 *%ptr [all …]
|
D | frame-20.ll | 48 %l0 = load volatile double, double *%ptr 49 %l1 = load volatile double, double *%ptr 50 %l2 = load volatile double, double *%ptr 51 %l3 = load volatile double, double *%ptr 52 %l4 = load volatile double, double *%ptr 53 %l5 = load volatile double, double *%ptr 54 %l6 = load volatile double, double *%ptr 55 %l7 = load volatile double, double *%ptr 56 %l8 = load volatile double, double *%ptr 57 %l9 = load volatile double, double *%ptr [all …]
|
D | fp-conv-17.mir | 5 %val0 = load volatile float, float* %ptr2 6 %val1 = load volatile float, float* %ptr2 7 %val2 = load volatile float, float* %ptr2 8 %val3 = load volatile float, float* %ptr2 9 %val4 = load volatile float, float* %ptr2 10 %val5 = load volatile float, float* %ptr2 11 %val6 = load volatile float, float* %ptr2 12 %val7 = load volatile float, float* %ptr2 13 %val8 = load volatile float, float* %ptr2 14 %val9 = load volatile float, float* %ptr2 [all …]
|
D | frame-18.ll | 19 %l0 = load volatile i32, i32 *%ptr 20 %l1 = load volatile i32, i32 *%ptr 21 %l3 = load volatile i32, i32 *%ptr 22 %l4 = load volatile i32, i32 *%ptr 23 %l5 = load volatile i32, i32 *%ptr 24 %l6 = load volatile i32, i32 *%ptr 25 %l7 = load volatile i32, i32 *%ptr 26 %l8 = load volatile i32, i32 *%ptr 27 %l9 = load volatile i32, i32 *%ptr 28 %l10 = load volatile i32, i32 *%ptr [all …]
|
D | frame-17.ll | 34 %l0 = load volatile float, float *%ptr 35 %l1 = load volatile float, float *%ptr 36 %l2 = load volatile float, float *%ptr 37 %l3 = load volatile float, float *%ptr 38 %l4 = load volatile float, float *%ptr 39 %l5 = load volatile float, float *%ptr 40 %l6 = load volatile float, float *%ptr 41 %l7 = load volatile float, float *%ptr 42 %l8 = load volatile float, float *%ptr 43 %l9 = load volatile float, float *%ptr [all …]
|
D | frame-03.ll | 42 %l0 = load volatile double, double *%ptr 43 %l1 = load volatile double, double *%ptr 44 %l2 = load volatile double, double *%ptr 45 %l3 = load volatile double, double *%ptr 46 %l4 = load volatile double, double *%ptr 47 %l5 = load volatile double, double *%ptr 48 %l6 = load volatile double, double *%ptr 49 %l7 = load volatile double, double *%ptr 50 %l8 = load volatile double, double *%ptr 51 %l9 = load volatile double, double *%ptr [all …]
|
D | frame-02.ll | 40 %l0 = load volatile float, float *%ptr 41 %l1 = load volatile float, float *%ptr 42 %l2 = load volatile float, float *%ptr 43 %l3 = load volatile float, float *%ptr 44 %l4 = load volatile float, float *%ptr 45 %l5 = load volatile float, float *%ptr 46 %l6 = load volatile float, float *%ptr 47 %l7 = load volatile float, float *%ptr 48 %l8 = load volatile float, float *%ptr 49 %l9 = load volatile float, float *%ptr [all …]
|
D | frame-08.ll | 28 %l0 = load volatile i32, i32 *%ptr 29 %l1 = load volatile i32, i32 *%ptr 30 %l4 = load volatile i32, i32 *%ptr 31 %l5 = load volatile i32, i32 *%ptr 32 %l6 = load volatile i32, i32 *%ptr 33 %l7 = load volatile i32, i32 *%ptr 34 %l8 = load volatile i32, i32 *%ptr 35 %l9 = load volatile i32, i32 *%ptr 36 %l10 = load volatile i32, i32 *%ptr 37 %l11 = load volatile i32, i32 *%ptr [all …]
|
/external/llvm/test/CodeGen/SystemZ/ |
D | frame-19.ll | 30 %v0 = load volatile <16 x i8>, <16 x i8> *%ptr 31 %v1 = load volatile <16 x i8>, <16 x i8> *%ptr 32 %v2 = load volatile <16 x i8>, <16 x i8> *%ptr 33 %v3 = load volatile <16 x i8>, <16 x i8> *%ptr 34 %v4 = load volatile <16 x i8>, <16 x i8> *%ptr 35 %v5 = load volatile <16 x i8>, <16 x i8> *%ptr 36 %v6 = load volatile <16 x i8>, <16 x i8> *%ptr 37 %v7 = load volatile <16 x i8>, <16 x i8> *%ptr 38 %v8 = load volatile <16 x i8>, <16 x i8> *%ptr 39 %v9 = load volatile <16 x i8>, <16 x i8> *%ptr [all …]
|
D | int-conv-11.ll | 11 %val0 = load volatile i32 , i32 *%ptr 12 %val1 = load volatile i32 , i32 *%ptr 13 %val2 = load volatile i32 , i32 *%ptr 14 %val3 = load volatile i32 , i32 *%ptr 15 %val4 = load volatile i32 , i32 *%ptr 16 %val5 = load volatile i32 , i32 *%ptr 17 %val6 = load volatile i32 , i32 *%ptr 18 %val7 = load volatile i32 , i32 *%ptr 19 %val8 = load volatile i32 , i32 *%ptr 20 %val9 = load volatile i32 , i32 *%ptr [all …]
|
D | frame-20.ll | 48 %l0 = load volatile double, double *%ptr 49 %l1 = load volatile double, double *%ptr 50 %l2 = load volatile double, double *%ptr 51 %l3 = load volatile double, double *%ptr 52 %l4 = load volatile double, double *%ptr 53 %l5 = load volatile double, double *%ptr 54 %l6 = load volatile double, double *%ptr 55 %l7 = load volatile double, double *%ptr 56 %l8 = load volatile double, double *%ptr 57 %l9 = load volatile double, double *%ptr [all …]
|
D | frame-18.ll | 19 %l0 = load volatile i32 , i32 *%ptr 20 %l1 = load volatile i32 , i32 *%ptr 21 %l3 = load volatile i32 , i32 *%ptr 22 %l4 = load volatile i32 , i32 *%ptr 23 %l5 = load volatile i32 , i32 *%ptr 24 %l6 = load volatile i32 , i32 *%ptr 25 %l7 = load volatile i32 , i32 *%ptr 26 %l8 = load volatile i32 , i32 *%ptr 27 %l9 = load volatile i32 , i32 *%ptr 28 %l10 = load volatile i32 , i32 *%ptr [all …]
|
D | frame-17.ll | 34 %l0 = load volatile float , float *%ptr 35 %l1 = load volatile float , float *%ptr 36 %l2 = load volatile float , float *%ptr 37 %l3 = load volatile float , float *%ptr 38 %l4 = load volatile float , float *%ptr 39 %l5 = load volatile float , float *%ptr 40 %l6 = load volatile float , float *%ptr 41 %l7 = load volatile float , float *%ptr 42 %l8 = load volatile float , float *%ptr 43 %l9 = load volatile float , float *%ptr [all …]
|
D | frame-03.ll | 42 %l0 = load volatile double , double *%ptr 43 %l1 = load volatile double , double *%ptr 44 %l2 = load volatile double , double *%ptr 45 %l3 = load volatile double , double *%ptr 46 %l4 = load volatile double , double *%ptr 47 %l5 = load volatile double , double *%ptr 48 %l6 = load volatile double , double *%ptr 49 %l7 = load volatile double , double *%ptr 50 %l8 = load volatile double , double *%ptr 51 %l9 = load volatile double , double *%ptr [all …]
|
D | frame-02.ll | 40 %l0 = load volatile float , float *%ptr 41 %l1 = load volatile float , float *%ptr 42 %l2 = load volatile float , float *%ptr 43 %l3 = load volatile float , float *%ptr 44 %l4 = load volatile float , float *%ptr 45 %l5 = load volatile float , float *%ptr 46 %l6 = load volatile float , float *%ptr 47 %l7 = load volatile float , float *%ptr 48 %l8 = load volatile float , float *%ptr 49 %l9 = load volatile float , float *%ptr [all …]
|
D | frame-08.ll | 28 %l0 = load volatile i32 , i32 *%ptr 29 %l1 = load volatile i32 , i32 *%ptr 30 %l4 = load volatile i32 , i32 *%ptr 31 %l5 = load volatile i32 , i32 *%ptr 32 %l6 = load volatile i32 , i32 *%ptr 33 %l7 = load volatile i32 , i32 *%ptr 34 %l8 = load volatile i32 , i32 *%ptr 35 %l9 = load volatile i32 , i32 *%ptr 36 %l10 = load volatile i32 , i32 *%ptr 37 %l11 = load volatile i32 , i32 *%ptr [all …]
|
/external/llvm/test/CodeGen/AArch64/ |
D | callee-save.ll | 15 %val1 = load volatile float, float* @var 16 %val2 = load volatile float, float* @var 17 %val3 = load volatile float, float* @var 18 %val4 = load volatile float, float* @var 19 %val5 = load volatile float, float* @var 20 %val6 = load volatile float, float* @var 21 %val7 = load volatile float, float* @var 22 %val8 = load volatile float, float* @var 23 %val9 = load volatile float, float* @var 24 %val10 = load volatile float, float* @var [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/AArch64/ |
D | callee-save.ll | 15 %val1 = load volatile float, float* @var 16 %val2 = load volatile float, float* @var 17 %val3 = load volatile float, float* @var 18 %val4 = load volatile float, float* @var 19 %val5 = load volatile float, float* @var 20 %val6 = load volatile float, float* @var 21 %val7 = load volatile float, float* @var 22 %val8 = load volatile float, float* @var 23 %val9 = load volatile float, float* @var 24 %val10 = load volatile float, float* @var [all …]
|
D | arm64-zeroreg.ll | 18 %v0 = load volatile i64, i64* %addr 19 %v1 = load volatile i64, i64* %addr 20 %v2 = load volatile i64, i64* %addr 21 %v3 = load volatile i64, i64* %addr 22 %v4 = load volatile i64, i64* %addr 23 %v5 = load volatile i64, i64* %addr 24 %v6 = load volatile i64, i64* %addr 25 %v7 = load volatile i64, i64* %addr 26 %v8 = load volatile i64, i64* %addr 27 %v9 = load volatile i64, i64* %addr [all …]
|
D | swiftself-scavenger.ll | 19 %v0 = load volatile i64, i64* @ptr64, align 8 20 %v1 = load volatile i64, i64* @ptr64, align 8 21 %v2 = load volatile i64, i64* @ptr64, align 8 22 %v3 = load volatile i64, i64* @ptr64, align 8 23 %v4 = load volatile i64, i64* @ptr64, align 8 24 %v5 = load volatile i64, i64* @ptr64, align 8 25 %v6 = load volatile i64, i64* @ptr64, align 8 26 %v7 = load volatile i64, i64* @ptr64, align 8 27 %v8 = load volatile i64, i64* @ptr64, align 8 28 %v9 = load volatile i64, i64* @ptr64, align 8 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/AMDGPU/ |
D | attr-amdgpu-flat-work-group-size.ll | 44 %val0 = load volatile float, float addrspace(1)* @var 45 %val1 = load volatile float, float addrspace(1)* @var 46 %val2 = load volatile float, float addrspace(1)* @var 47 %val3 = load volatile float, float addrspace(1)* @var 48 %val4 = load volatile float, float addrspace(1)* @var 49 %val5 = load volatile float, float addrspace(1)* @var 50 %val6 = load volatile float, float addrspace(1)* @var 51 %val7 = load volatile float, float addrspace(1)* @var 52 %val8 = load volatile float, float addrspace(1)* @var 53 %val9 = load volatile float, float addrspace(1)* @var [all …]
|
D | attr-amdgpu-waves-per-eu.ll | 124 %val0 = load volatile float, float addrspace(1)* @var 125 %val1 = load volatile float, float addrspace(1)* @var 126 %val2 = load volatile float, float addrspace(1)* @var 127 %val3 = load volatile float, float addrspace(1)* @var 128 %val4 = load volatile float, float addrspace(1)* @var 129 %val5 = load volatile float, float addrspace(1)* @var 130 %val6 = load volatile float, float addrspace(1)* @var 131 %val7 = load volatile float, float addrspace(1)* @var 132 %val8 = load volatile float, float addrspace(1)* @var 133 %val9 = load volatile float, float addrspace(1)* @var [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/Mips/ |
D | emergency-spill-slot-near-fp.ll | 15 store volatile i8* %stackspace_casted, i8** @ptrvar 18 %v0 = load volatile i32, i32* @var 19 %v1 = load volatile i32, i32* @var 20 %v2 = load volatile i32, i32* @var 21 %v3 = load volatile i32, i32* @var 22 %v4 = load volatile i32, i32* @var 23 %v5 = load volatile i32, i32* @var 24 %v6 = load volatile i32, i32* @var 25 %v7 = load volatile i32, i32* @var 26 %v8 = load volatile i32, i32* @var [all …]
|