Searched refs:tensor_to_memref (Results 1 – 13 of 13) sorted by relevance
/external/llvm-project/mlir/test/Dialect/Standard/ |
D | canonicalize.mlir | 3 // Test case: Basic folding of tensor_load(tensor_to_memref(t)) -> t 8 %0 = tensor_to_memref %arg0 : memref<?xf32> 13 // Test case: Basic folding of tensor_to_memref(tensor_load(m)) -> m 19 %1 = tensor_to_memref %0 : memref<?xf32> 27 // CHECK: %[[MEMREF_ADDRSPACE7:.*]] = tensor_to_memref %[[TENSOR]] : memref<?xf32, 7> 31 %1 = tensor_to_memref %0 : memref<?xf32, 7> 48 // Test case: Folding of load(tensor_to_memref(%v, %idxs)) 57 %0 = tensor_to_memref %arg2 : memref<?x?xf32>
|
D | func-bufferize.mlir | 6 // CHECK: %[[MEMREF:.*]] = tensor_to_memref %[[TENSOR]] : memref<f32> 15 // CHECK: %[[M1:.*]] = tensor_to_memref %[[T1]] : memref<f32> 19 // CHECK: %[[M2:.*]] = tensor_to_memref %[[T2]] : memref<f32> 39 // CHECK: %[[MEMREF:.*]] = tensor_to_memref %[[TENSOR]] : memref<f32> 50 // CHECK: %[[MEMREF:.*]] = tensor_to_memref %[[TENSOR]] : memref<f32>
|
D | bufferize.mlir | 6 // CHECK: %[[MEMREF:.*]] = tensor_to_memref %[[TENSOR]] : memref<f32> 21 // CHECK: %[[ARG_MEMREF:.*]] = tensor_to_memref %[[ARG]] : memref<*xf32> 67 // CHECK: %[[MEMREF:.*]] = tensor_to_memref %[[TENSOR]] : memref<?xf32> 80 // CHECK: %[[TRUE_VAL_MEMREF:.*]] = tensor_to_memref %[[TRUE_VAL]] : memref<f32> 81 // CHECK: %[[FALSE_VAL_MEMREF:.*]] = tensor_to_memref %[[FALSE_VAL]] : memref<f32> 92 // CHECK: %[[MEMREF:.*]] = tensor_to_memref %[[TENSOR]] 103 // CHECK: %[[MEMREF:.*]] = tensor_to_memref %[[TENSOR]] : memref<*xf32> 114 // CHECK: %[[MEMREF:.*]] = tensor_to_memref %[[TENSOR]] : memref<2xf32>
|
D | ops.mlir | 24 %0 = tensor_to_memref %arg0 : memref<?xi64, affine_map<(d0) -> (d0 + 7)>> 25 %1 = tensor_to_memref %arg1 : memref<*xi64, 1>
|
/external/llvm-project/mlir/test/Dialect/SCF/ |
D | bufferize.mlir | 8 // CHECK: %[[TRUE_MEMREF:.*]] = tensor_to_memref %[[TRUE_TENSOR]] : memref<?xf32> 11 // CHECK: %[[FALSE_MEMREF:.*]] = tensor_to_memref %[[FALSE_TENSOR]] : memref<?xf32> 30 // CHECK: %[[MEMREF:.*]] = tensor_to_memref %[[TENSOR]] : memref<f32> 33 // CHECK: %[[MEMREF_YIELDED:.*]] = tensor_to_memref %[[TENSOR_ITER]] : memref<f32> 65 // CHECK: %[[MEMREF:.*]] = tensor_to_memref %[[TENSOR]] : memref<f32> 69 // CHECK: %[[MEMREF_MUNGED:.*]] = tensor_to_memref %[[TENSOR_MUNGED]] : memref<f32>
|
/external/llvm-project/mlir/test/Dialect/Linalg/ |
D | bufferize.mlir | 6 // - tensor_to_memref / tensor_load materializations are properly inserted 14 // CHECK: %[[MEMREF:.*]] = tensor_to_memref %[[TENSOR]] : memref<4xf32> 98 // CHECK: %[[MEMREF_ARG:.*]] = tensor_to_memref %[[ARG]] : memref<?x?xf32> 138 // CHECK: %[[ARG0_MEMREF:.*]] = tensor_to_memref %[[ARG0_TENSOR]] : memref<2x3x4xvector<3… 139 // CHECK: %[[ARG1_MEMREF:.*]] = tensor_to_memref %[[ARG1_TENSOR]] : memref<3x2xf32> 172 // CHECK: %[[M0:.*]] = tensor_to_memref %[[T]] : memref<?x?xf32> 180 // CHECK: %[[M1:.*]] = tensor_to_memref %[[T]] : memref<?x?xf32> 211 // CHECK-DAG: %[[M0:.*]] = tensor_to_memref %[[T]] : memref<?x?xf32> 212 // CHECK-DAG: %[[SM0:.*]] = tensor_to_memref %[[ST0]] : memref<2x3xf32> 225 // CHECK-DAG: %[[M1:.*]] = tensor_to_memref %[[T]] : memref<?x?xf32> [all …]
|
/external/llvm-project/mlir/test/Transforms/ |
D | finalizing-bufferize.mlir | 8 %1 = tensor_to_memref %0 : memref<f32> 17 %1 = tensor_to_memref %0 : memref<f32>
|
/external/tensorflow/tensorflow/compiler/mlir/hlo/lib/Dialect/mhlo/transforms/ |
D | lhlo_fuse_linalg.cc | 106 if (auto tensor_to_memref = dyn_cast<TensorToMemrefOp>(definingOp)) { in runOnFunction() local 107 auto alias = tensor_to_memref.tensor(); in runOnFunction()
|
/external/llvm-project/mlir/test/Dialect/Shape/ |
D | bufferize.mlir | 9 // CHECK: %[[YIELDED_MEMREF:.*]] = tensor_to_memref %[[TENSOR_VAL]] : memref<2xf16>
|
/external/tensorflow/tensorflow/compiler/mlir/hlo/tests/ |
D | lhlo-fuse-linalg.mlir | 394 %4 = tensor_to_memref %3 : memref<?xf32> 407 // CHECK: tensor_to_memref 418 // TILED: tensor_to_memref 429 // PLOOP: tensor_to_memref
|
/external/llvm-project/mlir/include/mlir/Transforms/ |
D | Passes.td | 301 remaining `tensor_load` and `tensor_to_memref` operations. 305 `tensor_to_memref` operations.
|
/external/tensorflow/tensorflow/compiler/mlir/tools/kernel_gen/tests/ |
D | buffer_reuse.mlir | 442 %10 = tensor_to_memref %0 : memref<?xindex>
|
/external/llvm-project/mlir/include/mlir/Dialect/StandardOps/IR/ |
D | Ops.td | 3516 The opposite of this op is tensor_to_memref. Together, these two ops are 3591 def TensorToMemrefOp : Std_Op<"tensor_to_memref", 3606 `tensor_to_memref(tensor_load(%memref)) -> %memref`, which is the property 3618 %12 = tensor_to_memref %10 : memref<4x?xf32, #map0, 42>
|