/external/libhevc/common/arm64/ |
D | ihevc_intra_pred_chroma_mode2.s | 123 ld2 {v0.8b, v1.8b},[x0],x8 129 ld2 {v2.8b, v3.8b},[x10],x8 132 ld2 {v4.8b, v5.8b},[x0],x8 133 ld2 {v6.8b, v7.8b},[x10],x8 136 ld2 {v8.8b, v9.8b},[x0],x8 137 ld2 {v10.8b, v11.8b},[x10],x8 138 ld2 {v12.8b, v13.8b},[x0],x8 141 ld2 {v14.8b, v15.8b},[x10],x8 188 ld2 {v0.8b, v1.8b},[x0],x8 191 ld2 {v2.8b, v3.8b},[x10],x8 [all …]
|
D | ihevc_intra_pred_chroma_dc.s | 131 ld2 {v30.8b, v31.8b}, [x6], #16 //load from src[nt] 137 ld2 {v26.8b, v27.8b}, [x8],#16 //load from src[2nt+1] 159 ld2 {v30.8b, v31.8b}, [x6],#16 //load from src[nt] 163 ld2 {v26.8b, v27.8b}, [x8],#16 //load from src[2nt+1] 255 ld2 {v30.8b, v31.8b},[x6] //load from src[nt] 258 ld2 {v26.8b, v27.8b},[x8] //load from src[2nt+1]
|
/external/llvm/test/Transforms/SROA/ |
D | ppcf128-no-fold.ll | 5 %struct.ld2 = type { [2 x ppc_fp128] } 11 %z = alloca %struct.ld2, align 16 13 %dat = getelementptr inbounds %struct.ld2, %struct.ld2* %z, i32 0, i32 0 16 %dat1 = getelementptr inbounds %struct.ld2, %struct.ld2* %z, i32 0, i32 0 20 %coerce.dive = getelementptr %struct.ld2, %struct.ld2* %z, i32 0, i32 0
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/SROA/ |
D | ppcf128-no-fold.ll | 5 %struct.ld2 = type { [2 x ppc_fp128] } 11 %z = alloca %struct.ld2, align 16 13 %dat = getelementptr inbounds %struct.ld2, %struct.ld2* %z, i32 0, i32 0 16 %dat1 = getelementptr inbounds %struct.ld2, %struct.ld2* %z, i32 0, i32 0 20 %coerce.dive = getelementptr %struct.ld2, %struct.ld2* %z, i32 0, i32 0
|
/external/libhevc/decoder/arm64/ |
D | ihevcd_itrans_recon_dc_chroma.s | 117 ld2 {v2.8b, v3.8b},[x7],x2 118 ld2 {v4.8b, v5.8b},[x7],x2 119 ld2 {v6.8b, v7.8b},[x7],x2 120 ld2 {v8.8b, v9.8b},[x7],x2 122 ld2 {v10.8b, v11.8b},[x7],x2 123 ld2 {v12.8b, v13.8b},[x7],x2 124 ld2 {v14.8b, v15.8b},[x7],x2 125 ld2 {v16.8b, v17.8b},[x7] 184 ld2 {v2.8b, v3.8b},[x0],x2 185 ld2 {v4.8b, v5.8b},[x0],x2 [all …]
|
/external/llvm/test/CodeGen/PowerPC/ |
D | ppcf128sf.ll | 4 @ld2 = common global ppc_fp128 0xM00000000000000000000000000000000, align 16 15 %1 = load ppc_fp128, ppc_fp128* @ld2, align 16 19 %3 = load ppc_fp128, ppc_fp128* @ld2, align 16 23 %5 = load ppc_fp128, ppc_fp128* @ld2, align 16 27 %7 = load ppc_fp128, ppc_fp128* @ld2, align 16 91 %1 = load ppc_fp128, ppc_fp128* @ld2, align 16 104 %1 = load ppc_fp128, ppc_fp128* @ld2, align 16 117 %1 = load ppc_fp128, ppc_fp128* @ld2, align 16 130 %1 = load ppc_fp128, ppc_fp128* @ld2, align 16 143 %1 = load ppc_fp128, ppc_fp128* @ld2, align 16 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/PowerPC/ |
D | ppcf128sf.ll | 4 @ld2 = common global ppc_fp128 0xM00000000000000000000000000000000, align 16 15 %1 = load ppc_fp128, ppc_fp128* @ld2, align 16 19 %3 = load ppc_fp128, ppc_fp128* @ld2, align 16 23 %5 = load ppc_fp128, ppc_fp128* @ld2, align 16 27 %7 = load ppc_fp128, ppc_fp128* @ld2, align 16 91 %1 = load ppc_fp128, ppc_fp128* @ld2, align 16 104 %1 = load ppc_fp128, ppc_fp128* @ld2, align 16 117 %1 = load ppc_fp128, ppc_fp128* @ld2, align 16 130 %1 = load ppc_fp128, ppc_fp128* @ld2, align 16 143 %1 = load ppc_fp128, ppc_fp128* @ld2, align 16 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/MC/AArch64/ |
D | neon-simd-ldst-multi-elem.s | 369 ld2 { v0.16b, v1.16b }, [x0] 370 ld2 { v15.8h, v16.8h }, [x15] 371 ld2 { v31.4s, v0.4s }, [sp] 372 ld2 { v0.2d, v1.2d }, [x0] 373 ld2 { v0.8b, v1.8b }, [x0] 374 ld2 { v15.4h, v16.4h }, [x15] 375 ld2 { v31.2s, v0.2s }, [sp] 384 ld2 { v0.16b-v1.16b }, [x0] 385 ld2 { v15.8h-v16.8h }, [x15] 386 ld2 { v31.4s-v0.4s }, [sp] [all …]
|
D | arm64-simd-ldst.s | 187 ld2.8b {v4, v5}, [x19] 188 ld2.16b {v4, v5}, [x19] 189 ld2.4h {v4, v5}, [x19] 190 ld2.8h {v4, v5}, [x19] 191 ld2.2s {v4, v5}, [x19] 192 ld2.4s {v4, v5}, [x19] 193 ld2.2d {v4, v5}, [x19] 205 ; CHECK: ld2.8b { v4, v5 }, [x19] ; encoding: [0x64,0x82,0x40,0x0c] 206 ; CHECK: ld2.16b { v4, v5 }, [x19] ; encoding: [0x64,0x82,0x40,0x4c] 207 ; CHECK: ld2.4h { v4, v5 }, [x19] ; encoding: [0x64,0x86,0x40,0x0c] [all …]
|
D | neon-simd-ldst-one-elem.s | 96 ld2 { v0.b, v1.b }[9], [x0] 97 ld2 { v15.h, v16.h }[7], [x15] 98 ld2 { v31.s, v0.s }[3], [sp] 99 ld2 { v0.d, v1.d }[1], [x0] 257 ld2 { v0.b, v1.b }[9], [x0], x3 258 ld2 { v15.h, v16.h }[7], [x15], #4 259 ld2 { v31.s, v0.s }[3], [sp], #8 260 ld2 { v0.d, v1.d }[1], [x0], x0
|
D | neon-simd-post-ldst-multi-elem.s | 124 ld2 { v0.16b, v1.16b }, [x0], x1 125 ld2 { v15.8h, v16.8h }, [x15], x2 126 ld2 { v31.4s, v0.4s }, [sp], #32 127 ld2 { v0.2d, v1.2d }, [x0], #32 128 ld2 { v0.8b, v1.8b }, [x0], x2 129 ld2 { v15.4h, v16.4h }, [x15], x3 130 ld2 { v31.2s, v0.2s }, [sp], #16
|
/external/llvm/test/MC/AArch64/ |
D | neon-simd-ldst-multi-elem.s | 369 ld2 { v0.16b, v1.16b }, [x0] 370 ld2 { v15.8h, v16.8h }, [x15] 371 ld2 { v31.4s, v0.4s }, [sp] 372 ld2 { v0.2d, v1.2d }, [x0] 373 ld2 { v0.8b, v1.8b }, [x0] 374 ld2 { v15.4h, v16.4h }, [x15] 375 ld2 { v31.2s, v0.2s }, [sp] 384 ld2 { v0.16b-v1.16b }, [x0] 385 ld2 { v15.8h-v16.8h }, [x15] 386 ld2 { v31.4s-v0.4s }, [sp] [all …]
|
D | arm64-simd-ldst.s | 187 ld2.8b {v4, v5}, [x19] 188 ld2.16b {v4, v5}, [x19] 189 ld2.4h {v4, v5}, [x19] 190 ld2.8h {v4, v5}, [x19] 191 ld2.2s {v4, v5}, [x19] 192 ld2.4s {v4, v5}, [x19] 193 ld2.2d {v4, v5}, [x19] 205 ; CHECK: ld2.8b { v4, v5 }, [x19] ; encoding: [0x64,0x82,0x40,0x0c] 206 ; CHECK: ld2.16b { v4, v5 }, [x19] ; encoding: [0x64,0x82,0x40,0x4c] 207 ; CHECK: ld2.4h { v4, v5 }, [x19] ; encoding: [0x64,0x86,0x40,0x0c] [all …]
|
D | neon-simd-ldst-one-elem.s | 96 ld2 { v0.b, v1.b }[9], [x0] 97 ld2 { v15.h, v16.h }[7], [x15] 98 ld2 { v31.s, v0.s }[3], [sp] 99 ld2 { v0.d, v1.d }[1], [x0] 257 ld2 { v0.b, v1.b }[9], [x0], x3 258 ld2 { v15.h, v16.h }[7], [x15], #4 259 ld2 { v31.s, v0.s }[3], [sp], #8 260 ld2 { v0.d, v1.d }[1], [x0], x0
|
D | neon-simd-post-ldst-multi-elem.s | 124 ld2 { v0.16b, v1.16b }, [x0], x1 125 ld2 { v15.8h, v16.8h }, [x15], x2 126 ld2 { v31.4s, v0.4s }, [sp], #32 127 ld2 { v0.2d, v1.2d }, [x0], #32 128 ld2 { v0.8b, v1.8b }, [x0], x2 129 ld2 { v15.4h, v16.4h }, [x15], x3 130 ld2 { v31.2s, v0.2s }, [sp], #16
|
/external/capstone/suite/MC/AArch64/ |
D | neon-simd-ldst-multi-elem.s.cs | 156 0x00,0x80,0x40,0x4c = ld2 {v0.16b, v1.16b}, [x0] 157 0xef,0x85,0x40,0x4c = ld2 {v15.8h, v16.8h}, [x15] 158 0xff,0x8b,0x40,0x4c = ld2 {v31.4s, v0.4s}, [sp] 159 0x00,0x8c,0x40,0x4c = ld2 {v0.2d, v1.2d}, [x0] 160 0x00,0x80,0x40,0x0c = ld2 {v0.8b, v1.8b}, [x0] 161 0xef,0x85,0x40,0x0c = ld2 {v15.4h, v16.4h}, [x15] 162 0xff,0x8b,0x40,0x0c = ld2 {v31.2s, v0.2s}, [sp] 163 0x00,0x80,0x40,0x4c = ld2 {v0.16b, v1.16b}, [x0] 164 0xef,0x85,0x40,0x4c = ld2 {v15.8h, v16.8h}, [x15] 165 0xff,0x8b,0x40,0x4c = ld2 {v31.4s, v0.4s}, [sp] [all …]
|
D | neon-simd-ldst-one-elem.s.cs | 38 0x00,0x04,0x60,0x4d = ld2 {v0.b, v1.b}[9], [x0] 39 0xef,0x59,0x60,0x4d = ld2 {v15.h, v16.h}[7], [x15] 40 0xff,0x93,0x60,0x4d = ld2 {v31.s, v0.s}[3], [sp] 41 0x00,0x84,0x60,0x4d = ld2 {v0.d, v1.d}[1], [x0] 102 0x00,0x04,0xe3,0x4d = ld2 {v0.b, v1.b}[9], [x0], x3 103 0xef,0x59,0xff,0x4d = ld2 {v15.h, v16.h}[7], [x15], #4 104 0xff,0x93,0xff,0x4d = ld2 {v31.s, v0.s}[3], [sp], #8 105 0x00,0x84,0xe0,0x4d = ld2 {v0.d, v1.d}[1], [x0], x0
|
D | neon-simd-post-ldst-multi-elem.s.cs | 34 0x00,0x80,0xc1,0x4c = ld2 {v0.16b, v1.16b}, [x0], x1 35 0xef,0x85,0xc2,0x4c = ld2 {v15.8h, v16.8h}, [x15], x2 36 0xff,0x8b,0xdf,0x4c = ld2 {v31.4s, v0.4s}, [sp], #32 37 0x00,0x8c,0xdf,0x4c = ld2 {v0.2d, v1.2d}, [x0], #32 38 0x00,0x80,0xc2,0x0c = ld2 {v0.8b, v1.8b}, [x0], x2 39 0xef,0x85,0xc3,0x0c = ld2 {v15.4h, v16.4h}, [x15], x3 40 0xff,0x8b,0xdf,0x0c = ld2 {v31.2s, v0.2s}, [sp], #16
|
/external/llvm/test/CodeGen/AArch64/ |
D | aarch64-interleaved-accesses-extract-user.ll | 4 ; CHECK: %ldN = call { <4 x i32>, <4 x i32> } @llvm.aarch64.neon.ld2.v4i32.p0v4i32 22 ; CHECK: %ldN = call { <4 x i32>, <4 x i32> } @llvm.aarch64.neon.ld2.v4i32.p0v4i32 42 ; CHECK-NOT: %ldN = call { <4 x i32>, <4 x i32> } @llvm.aarch64.neon.ld2.v4i32.p0v4i32 59 ; CHECK-NOT: %ldN = call { <4 x i32>, <4 x i32> } @llvm.aarch64.neon.ld2.v4i32.p0v4i32 69 ; CHECK-NOT: %ldN = call { <4 x i32>, <4 x i32> } @llvm.aarch64.neon.ld2.v4i32.p0v4i32 79 ; CHECK-NOT: %ldN = call { <4 x i32>, <4 x i32> } @llvm.aarch64.neon.ld2.v4i32.p0v4i32
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | promote-alloca-array-allocation.ll | 19 %ld2 = load i32, i32* %arrayidx10, align 4 20 store i32 %ld2, i32 addrspace(1)* %out, align 4 41 %ld2 = load i32, i32* %arrayidx10, align 4 42 store i32 %ld2, i32 addrspace(1)* %out, align 4
|
/external/llvm/test/CodeGen/X86/ |
D | atom-cmpb.ll | 17 %ld2 = load i8, i8* %incdec.ptr, align 1 19 %x5 = xor i8 %ld2, -1 20 %cmp34 = icmp ult i8 %ld2, %ld1
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/X86/ |
D | atom-cmpb.ll | 17 %ld2 = load i8, i8* %incdec.ptr, align 1 19 %x5 = xor i8 %ld2, -1 20 %cmp34 = icmp ult i8 %ld2, %ld1
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/AMDGPU/ |
D | promote-alloca-array-allocation.ll | 19 %ld2 = load i32, i32* %arrayidx10, align 4 20 store i32 %ld2, i32 addrspace(1)* %out, align 4 41 %ld2 = load i32, i32* %arrayidx10, align 4 42 store i32 %ld2, i32 addrspace(1)* %out, align 4
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/EarlyCSE/AArch64/ |
D | intrinsics.ll | 8 ; Check that @llvm.aarch64.neon.ld2 is optimized away by Early CSE. 10 ; CHECK-NOT: call { <4 x i32>, <4 x i32> } @llvm.aarch64.neon.ld2.v4i32.p0i8 29 %vld2 = call { <4 x i32>, <4 x i32> } @llvm.aarch64.neon.ld2.v4i32.p0i8(i8* %5) 65 %vld2 = call { <4 x i32>, <4 x i32> } @llvm.aarch64.neon.ld2.v4i32.p0i8(i8* %5) 78 ; Check that the first @llvm.aarch64.neon.ld2 is optimized away by Early CSE. 80 ; CHECK: call { <4 x i32>, <4 x i32> } @llvm.aarch64.neon.ld2.v4i32.p0i8 81 ; CHECK-NOT: call { <4 x i32>, <4 x i32> } @llvm.aarch64.neon.ld2.v4i32.p0i8 94 %vld2 = call { <4 x i32>, <4 x i32> } @llvm.aarch64.neon.ld2.v4i32.p0i8(i8* %0) 98 %vld22 = call { <4 x i32>, <4 x i32> } @llvm.aarch64.neon.ld2.v4i32.p0i8(i8* %1) 112 ; Check that the store prevents @llvm.aarch64.neon.ld2 from being optimized [all …]
|
/external/llvm/test/Transforms/EarlyCSE/AArch64/ |
D | intrinsics.ll | 6 ; Check that @llvm.aarch64.neon.ld2 is optimized away by Early CSE. 8 ; CHECK-NOT: call { <4 x i32>, <4 x i32> } @llvm.aarch64.neon.ld2.v4i32.p0i8 27 %vld2 = call { <4 x i32>, <4 x i32> } @llvm.aarch64.neon.ld2.v4i32.p0i8(i8* %5) 63 %vld2 = call { <4 x i32>, <4 x i32> } @llvm.aarch64.neon.ld2.v4i32.p0i8(i8* %5) 76 ; Check that the first @llvm.aarch64.neon.ld2 is optimized away by Early CSE. 78 ; CHECK: call { <4 x i32>, <4 x i32> } @llvm.aarch64.neon.ld2.v4i32.p0i8 79 ; CHECK-NOT: call { <4 x i32>, <4 x i32> } @llvm.aarch64.neon.ld2.v4i32.p0i8 92 %vld2 = call { <4 x i32>, <4 x i32> } @llvm.aarch64.neon.ld2.v4i32.p0i8(i8* %0) 96 %vld22 = call { <4 x i32>, <4 x i32> } @llvm.aarch64.neon.ld2.v4i32.p0i8(i8* %1) 110 ; Check that the store prevents @llvm.aarch64.neon.ld2 from being optimized [all …]
|