/external/clang/test/SemaCUDA/ |
D | cuda-builtin-vars.cu | 28 out[i++] = gridDim.x; in kernel() 29 gridDim.x = 0; // expected-error {{no setter defined for property 'x'}} in kernel() 30 out[i++] = gridDim.y; in kernel() 31 gridDim.y = 0; // expected-error {{no setter defined for property 'y'}} in kernel() 32 out[i++] = gridDim.z; in kernel() 33 gridDim.z = 0; // expected-error {{no setter defined for property 'z'}} in kernel()
|
/external/llvm-project/clang/test/SemaCUDA/ |
D | cuda-builtin-vars.cu | 28 out[i++] = gridDim.x; in kernel() 29 gridDim.x = 0; // expected-error {{no setter defined for property 'x'}} in kernel() 30 out[i++] = gridDim.y; in kernel() 31 gridDim.y = 0; // expected-error {{no setter defined for property 'y'}} in kernel() 32 out[i++] = gridDim.z; in kernel() 33 gridDim.z = 0; // expected-error {{no setter defined for property 'z'}} in kernel()
|
/external/tensorflow/tensorflow/stream_executor/cuda/ |
D | cudart_stub.cc | 114 extern cudaError_t CUDARTAPI __cudaPopCallConfiguration(dim3 *gridDim, in __cudaPopCallConfiguration() argument 118 using FuncPtr = cudaError_t(CUDARTAPI *)(dim3 * gridDim, dim3 * blockDim, in __cudaPopCallConfiguration() 122 return func_ptr(gridDim, blockDim, sharedMem, stream); in __cudaPopCallConfiguration() 126 dim3 gridDim, dim3 blockDim, size_t sharedMem = 0, void *stream = 0) { in __cudaPushCallConfiguration() argument 127 using FuncPtr = unsigned(CUDARTAPI *)(dim3 gridDim, dim3 blockDim, in __cudaPushCallConfiguration() 131 return func_ptr(gridDim, blockDim, sharedMem, stream); in __cudaPushCallConfiguration()
|
/external/llvm-project/clang/test/CodeGenCUDA/ |
D | cuda-builtin-vars.cu | 21 out[i++] = gridDim.x; // CHECK: call i32 @llvm.nvvm.read.ptx.sreg.nctaid.x() in kernel() 22 out[i++] = gridDim.y; // CHECK: call i32 @llvm.nvvm.read.ptx.sreg.nctaid.y() in kernel() 23 out[i++] = gridDim.z; // CHECK: call i32 @llvm.nvvm.read.ptx.sreg.nctaid.z() in kernel()
|
/external/clang/test/CodeGenCUDA/ |
D | cuda-builtin-vars.cu | 21 out[i++] = gridDim.x; // CHECK: call i32 @llvm.nvvm.read.ptx.sreg.nctaid.x() in kernel() 22 out[i++] = gridDim.y; // CHECK: call i32 @llvm.nvvm.read.ptx.sreg.nctaid.y() in kernel() 23 out[i++] = gridDim.z; // CHECK: call i32 @llvm.nvvm.read.ptx.sreg.nctaid.z() in kernel()
|
/external/eigen/unsupported/Eigen/CXX11/src/Tensor/ |
D | TensorReductionCuda.h | 116 const Index num_threads = blockDim.x * gridDim.x; in ReductionInitKernel() 130 if (gridDim.x == 1) { in FullReductionKernel() 158 eigen_assert(gridDim.x == 1 || *semaphore >= 2u); in FullReductionKernel() 178 if (gridDim.x > 1 && threadIdx.x == 0) { in FullReductionKernel() 180 atomicInc(semaphore, gridDim.x + 1); in FullReductionKernel() 193 eigen_assert(gridDim.x == 1); in ReductionInitFullReduxKernelHalfFloat() 206 const Index num_threads = blockDim.x * gridDim.x; in ReductionInitKernelHalfFloat() 226 if (gridDim.x == 1 && first_index == 0) { in FullReductionKernelHalfFloat() 256 if (gridDim.x == 1 && first_index == 0) { in FullReductionKernelHalfFloat() 380 eigen_assert(gridDim.y == 1); [all …]
|
/external/tensorflow/tensorflow/core/kernels/ |
D | debug_ops_gpu.cu.cc | 42 const int32 total_thread_count = gridDim.x * blockDim.x; in CurtHealthKernel() 60 const int32 total_thread_count = gridDim.x * blockDim.x; in ConciseHealthKernel() 91 const int32 total_thread_count = gridDim.x * blockDim.x; in FullHealthKernel() 134 const int32 total_thread_count = gridDim.x * blockDim.x; in ReduceInfNanThreeSlotsKernel()
|
D | concat_lib_gpu_impl.cu.cc | 43 for (; gidx < total_cols; gidx += blockDim.x * gridDim.x) { in concat_fixed_kernel() 50 for (; gidy < total_rows; gidy += blockDim.y * gridDim.y) { in concat_fixed_kernel() 97 for (; gidx < total_cols; gidx += blockDim.x * gridDim.x) { in concat_variable_kernel() 109 for (; gidy < total_rows; gidy += blockDim.y * gridDim.y) in concat_variable_kernel()
|
D | check_numerics_op_gpu.cu.cc | 42 const int32 total_thread_count = gridDim.x * blockDim.x; in CheckNumericsKernel() 65 const int32 total_thread_count = gridDim.x * blockDim.x; in CheckNumericsKernelV2()
|
D | random_op_gpu.h | 147 const int32 total_thread_count = gridDim.x * blockDim.x; 189 const int32 total_thread_count = gridDim.x * blockDim.x;
|
D | split_lib_gpu.cu.cc | 148 for (; gidx < total_cols; gidx += blockDim.x * gridDim.x) { in split_v_kernel() 160 for (; gidy < total_rows; gidy += blockDim.y * gridDim.y) in split_v_kernel()
|
D | softmax_op_gpu.cu.cc | 90 tid += gridDim.x * blockDim.x; in GenerateNormalizedProb() 105 tid += gridDim.x * blockDim.x; in GenerateNormalizedProb()
|
D | stateful_random_ops_gpu.cu.cc | 56 auto total_thread_count = gridDim.x * blockDim.x; in FillKernel()
|
D | reduction_gpu_kernels.cu.h | 171 const int stride = blockDim.x * gridDim.x; 303 row += rows_per_warp * gridDim.y * blockDim.y; 304 for (; row < num_rows; row += rows_per_warp * gridDim.y * blockDim.y) { 333 out[col * gridDim.y + blockIdx.y] = s; 364 row += gridDim.y * blockDim.y; 367 for (; row < num_rows; row += gridDim.y * blockDim.y) { 397 out[col * gridDim.y + blockIdx.y] = s;
|
/external/llvm-project/clang/test/CodeGenCUDA/Inputs/ |
D | cuda.h | 25 extern "C" hipError_t hipLaunchKernel(const void *func, dim3 gridDim, 38 extern "C" cudaError_t cudaLaunchKernel(const void *func, dim3 gridDim,
|
/external/llvm-project/clang/test/SemaCUDA/Inputs/ |
D | cuda.h | 28 extern "C" hipError_t hipLaunchKernel(const void *func, dim3 gridDim, 42 extern "C" cudaError_t cudaLaunchKernel(const void *func, dim3 gridDim,
|
/external/llvm-project/mlir/test/Transforms/ |
D | parametric-mapping.mlir | 38 // stepXgdimx = step * gridDim.x 41 // new_step = step * gridDim.x * blockDim.x
|
/external/tensorflow/tensorflow/tools/ci_build/builds/user_ops/ |
D | cuda_op_kernel.cu.cc | 23 i += blockDim.x * gridDim.x) { in AddOneKernel()
|
/external/tensorflow/tensorflow/examples/adding_an_op/ |
D | cuda_op_kernel.cu.cc | 24 i += blockDim.x * gridDim.x) { in AddOneKernel()
|
/external/tensorflow/tensorflow/core/util/ |
D | gpu_device_functions.h | 168 /*delta=*/gridDim.x * blockDim.x, /*end=*/count); in GpuGridRangeX() 178 /*delta=*/gridDim.y * blockDim.y, /*end=*/count); in GpuGridRangeY() 188 /*delta=*/gridDim.z * blockDim.z, /*end=*/count); in GpuGridRangeZ() 564 assert(blockDim.x * gridDim.x / blockDim.x == gridDim.x); in SetZero() 576 assert(blockDim.x * gridDim.x / blockDim.x == gridDim.x); in SetToValue()
|
/external/clang/lib/Headers/ |
D | cuda_builtin_vars.h | 115 __CUDA_BUILTIN_VAR __cuda_builtin_gridDim_t gridDim; variable
|
/external/llvm-project/clang/lib/Headers/ |
D | __clang_cuda_builtin_vars.h | 109 __CUDA_BUILTIN_VAR __cuda_builtin_gridDim_t gridDim; variable
|
D | __clang_cuda_runtime_wrapper.h | 437 extern "C" unsigned __cudaPushCallConfiguration(dim3 gridDim, dim3 blockDim,
|
/external/llvm-project/openmp/libomptarget/deviceRTLs/nvptx/src/ |
D | target_impl.h | 208 INLINE int GetNumberOfBlocksInKernel() { return gridDim.x; } in GetNumberOfBlocksInKernel()
|
/external/tensorflow/tensorflow/core/kernels/image/ |
D | adjust_hsv_gpu.cu.h | 103 idx < number_elements; idx += blockDim.x * gridDim.x * 3) { in adjust_hsv_nhwc()
|