/external/clang/test/SemaCUDA/ |
D | cuda-builtin-vars.cu | 21 out[i++] = blockDim.x; in kernel() 22 blockDim.x = 0; // expected-error {{no setter defined for property 'x'}} in kernel() 23 out[i++] = blockDim.y; in kernel() 24 blockDim.y = 0; // expected-error {{no setter defined for property 'y'}} in kernel() 25 out[i++] = blockDim.z; in kernel() 26 blockDim.z = 0; // expected-error {{no setter defined for property 'z'}} in kernel()
|
/external/opencv3/modules/cudaimgproc/src/cuda/ |
D | canny.cu | 108 const int x = blockIdx.x * blockDim.x + threadIdx.x; in calcMagnitudeKernel() 109 const int y = blockIdx.y * blockDim.y + threadIdx.y; in calcMagnitudeKernel() 174 const int x = blockIdx.x * blockDim.x + threadIdx.x; in calcMapKernel() 175 const int y = blockIdx.y * blockDim.y + threadIdx.y; in calcMapKernel() 251 const int x = blockIdx.x * blockDim.x + threadIdx.x; in edgesHysteresisLocalKernel() 252 const int y = blockIdx.y * blockDim.y + threadIdx.y; in edgesHysteresisLocalKernel() 257 if (threadIdx.y == blockDim.y - 1) in edgesHysteresisLocalKernel() 258 …smem[blockDim.y + 1][threadIdx.x + 1] = checkIdx(y + 1, x, map.rows, map.cols) ? map(y + 1, x) : 0; in edgesHysteresisLocalKernel() 261 if (threadIdx.x == blockDim.x - 1) in edgesHysteresisLocalKernel() 262 …smem[threadIdx.y + 1][blockDim.x + 1] = checkIdx(y, x + 1, map.rows, map.cols) ? map(y, x + 1) : 0; in edgesHysteresisLocalKernel() [all …]
|
D | build_point_list.cu | 61 const int x = blockIdx.x * blockDim.x * PIXELS_PER_THREAD + threadIdx.x; in buildPointList() 62 const int y = blockIdx.y * blockDim.y + threadIdx.y; in buildPointList() 72 … for (int i = 0, xx = x; i < PIXELS_PER_THREAD && xx < src.cols; ++i, xx += blockDim.x) in buildPointList() 90 for (int i = 0; i < blockDim.y; ++i) in buildPointList() 98 for (int i = 0; i < blockDim.y; ++i) in buildPointList() 107 for(int i = threadIdx.x; i < qsize; i += blockDim.x, gidx += blockDim.x) in buildPointList()
|
D | mean_shift.cu | 116 …int base = (blockIdx.y * blockDim.y + threadIdx.y) * out_step + (blockIdx.x * blockDim.x + threadI… in do_mean_shift() 124 int x0 = blockIdx.x * blockDim.x + threadIdx.x; in meanshift_kernel() 125 int y0 = blockIdx.y * blockDim.y + threadIdx.y; in meanshift_kernel() 153 int x0 = blockIdx.x * blockDim.x + threadIdx.x; in meanshiftproc_kernel() 154 int y0 = blockIdx.y * blockDim.y + threadIdx.y; in meanshiftproc_kernel() 158 …int basesp = (blockIdx.y * blockDim.y + threadIdx.y) * outspstep + (blockIdx.x * blockDim.x + thre… in meanshiftproc_kernel()
|
D | generalized_hough.cu | 72 const int x = blockIdx.x * blockDim.x * PIXELS_PER_THREAD + threadIdx.x; in buildEdgePointList() 73 const int y = blockIdx.y * blockDim.y + threadIdx.y; in buildEdgePointList() 86 … for (int i = 0, xx = x; i < PIXELS_PER_THREAD && xx < edges.cols; ++i, xx += blockDim.x) in buildEdgePointList() 114 for (int i = 0; i < blockDim.y; ++i) in buildEdgePointList() 122 for (int i = 0; i < blockDim.y; ++i) in buildEdgePointList() 131 for(int i = threadIdx.x; i < qsize; i += blockDim.x, gidx += blockDim.x) in buildEdgePointList() 172 const int tid = blockIdx.x * blockDim.x + threadIdx.x; in buildRTable() 213 const int tid = blockIdx.x * blockDim.x + threadIdx.x; in Ballard_Pos_calcHist() 261 const int x = blockIdx.x * blockDim.x + threadIdx.x; in Ballard_Pos_findPosInHist() 262 const int y = blockIdx.y * blockDim.y + threadIdx.y; in Ballard_Pos_findPosInHist() [all …]
|
D | corners.cu | 66 const int x = blockIdx.x * blockDim.x + threadIdx.x; in cornerHarris_kernel() 67 const int y = blockIdx.y * blockDim.y + threadIdx.y; in cornerHarris_kernel() 100 const int x = blockIdx.x * blockDim.x + threadIdx.x; in cornerHarris_kernel() 101 const int y = blockIdx.y * blockDim.y + threadIdx.y; in cornerHarris_kernel() 171 const int x = blockIdx.x * blockDim.x + threadIdx.x; in cornerMinEigenVal_kernel() 172 const int y = blockIdx.y * blockDim.y + threadIdx.y; in cornerMinEigenVal_kernel() 209 const int x = blockIdx.x * blockDim.x + threadIdx.x; in cornerMinEigenVal_kernel() 210 const int y = blockIdx.y * blockDim.y + threadIdx.y; in cornerMinEigenVal_kernel()
|
D | hough_lines.cu | 75 for (int i = threadIdx.x; i < count; i += blockDim.x) in linesAccumGlobal() 93 for (int i = threadIdx.x; i < numrho + 1; i += blockDim.x) in linesAccumShared() 109 for (int i = threadIdx.x; i < count; i += blockDim.x) in linesAccumShared() 125 for (int i = threadIdx.x; i < numrho + 1; i += blockDim.x) in linesAccumShared() 151 const int r = blockIdx.x * blockDim.x + threadIdx.x; in linesGetResult() 152 const int n = blockIdx.y * blockDim.y + threadIdx.y; in linesGetResult()
|
D | hist.cu | 59 const int y = blockIdx.x * blockDim.y + threadIdx.y; in histogram256Kernel() 60 const int tid = threadIdx.y * blockDim.x + threadIdx.x; in histogram256Kernel() 70 for (int x = threadIdx.x; x < cols_4; x += blockDim.x) in histogram256Kernel() 128 const int y = blockIdx.x * blockDim.y + threadIdx.y; in histEven8u() 129 const int tid = threadIdx.y * blockDim.x + threadIdx.x; in histEven8u() 142 for (int x = threadIdx.x; x < cols_4; x += blockDim.x) in histEven8u()
|
D | match_template.cu | 91 int x = blockDim.x * blockIdx.x + threadIdx.x; in matchTemplateNaiveKernel_CCORR() 92 int y = blockDim.y * blockIdx.y + threadIdx.y; in matchTemplateNaiveKernel_CCORR() 157 int x = blockDim.x * blockIdx.x + threadIdx.x; in matchTemplateNaiveKernel_SQDIFF() 158 int y = blockDim.y * blockIdx.y + threadIdx.y; in matchTemplateNaiveKernel_SQDIFF() 223 const int x = blockIdx.x * blockDim.x + threadIdx.x; in matchTemplatePreparedKernel_SQDIFF_8U() 224 const int y = blockIdx.y * blockDim.y + threadIdx.y; in matchTemplatePreparedKernel_SQDIFF_8U() 293 const int x = blockIdx.x * blockDim.x + threadIdx.x; in matchTemplatePreparedKernel_SQDIFF_NORMED_8U() 294 const int y = blockIdx.y * blockDim.y + threadIdx.y; in matchTemplatePreparedKernel_SQDIFF_NORMED_8U() 339 const int x = blockIdx.x * blockDim.x + threadIdx.x; in matchTemplatePreparedKernel_CCOFF_8U() 340 const int y = blockIdx.y * blockDim.y + threadIdx.y; in matchTemplatePreparedKernel_CCOFF_8U() [all …]
|
D | clahe.cu | 65 const unsigned int tid = threadIdx.y * blockDim.x + threadIdx.x; in calcLutKernel() 70 for (int i = threadIdx.y; i < tileSize.y; i += blockDim.y) in calcLutKernel() 73 for (int j = threadIdx.x; j < tileSize.x; j += blockDim.x) in calcLutKernel() 139 const int x = blockIdx.x * blockDim.x + threadIdx.x; in tranformKernel() 140 const int y = blockIdx.y * blockDim.y + threadIdx.y; in tranformKernel()
|
D | hough_circles.cu | 68 const int tid = blockIdx.x * blockDim.x + threadIdx.x; in circlesAccumCenters() 132 const int x = blockIdx.x * blockDim.x + threadIdx.x; in buildCentersList() 133 const int y = blockIdx.y * blockDim.y + threadIdx.y; in buildCentersList() 186 for (int i = threadIdx.x; i < histSize + 2; i += blockDim.x) in circlesAccumRadius() 198 for (int i = threadIdx.x; i < count; i += blockDim.x) in circlesAccumRadius() 216 for (int i = threadIdx.x; i < histSize; i += blockDim.x) in circlesAccumRadius()
|
D | blend.cu | 55 int x = blockIdx.x * blockDim.x + threadIdx.x; in blendLinearKernel() 56 int y = blockIdx.y * blockDim.y + threadIdx.y; in blendLinearKernel() 89 int x = blockIdx.x * blockDim.x + threadIdx.x; in blendLinearKernel8UC4() 90 int y = blockIdx.y * blockDim.y + threadIdx.y; in blendLinearKernel8UC4()
|
/external/opencv3/modules/cudaobjdetect/src/cuda/ |
D | hog.cu | 126 if (blockIdx.x * blockDim.z + block_x >= img_block_width) in compute_hists_kernel_many_blocks() 133 const int offset_x = (blockIdx.x * blockDim.z + block_x) * cblock_stride_x + in compute_hists_kernel_many_blocks() 143 float* hist = hists + 12 * (cell_y * blockDim.z * CELLS_PER_BLOCK_Y + in compute_hists_kernel_many_blocks() 186 blockIdx.x * blockDim.z + block_x) * in compute_hists_kernel_many_blocks() 267 if (blockIdx.x * blockDim.z + threadIdx.z >= img_block_width) in normalize_hists_kernel_many_blocks() 271 blockIdx.x * blockDim.z + threadIdx.z) * in normalize_hists_kernel_many_blocks() 340 if (blockIdx.x * blockDim.z + win_x >= img_win_width) in compute_confidence_hists_kernel_many_blocks() 344 … blockIdx.x * win_block_stride_x * blockDim.z + win_x) * in compute_confidence_hists_kernel_many_blocks() 362 … confidences[blockIdx.y * img_win_width + blockIdx.x * blockDim.z + win_x] = product + free_coef; in compute_confidence_hists_kernel_many_blocks() 402 if (blockIdx.x * blockDim.z + win_x >= img_win_width) in classify_hists_kernel_many_blocks() [all …]
|
/external/opencv3/modules/cudev/include/opencv2/cudev/grid/detail/ |
D | split_merge.hpp | 67 const int x = blockIdx.x * blockDim.x + threadIdx.x; in mergeC2() 68 const int y = blockIdx.y * blockDim.y + threadIdx.y; in mergeC2() 97 const int x = blockIdx.x * blockDim.x + threadIdx.x; in mergeC3() 98 const int y = blockIdx.y * blockDim.y + threadIdx.y; in mergeC3() 128 const int x = blockIdx.x * blockDim.x + threadIdx.x; in mergeC4() 129 const int y = blockIdx.y * blockDim.y + threadIdx.y; in mergeC4() 191 const int x = blockIdx.x * blockDim.x + threadIdx.x; in split() 192 const int y = blockIdx.y * blockDim.y + threadIdx.y; in split() 221 const int x = blockIdx.x * blockDim.x + threadIdx.x; in split() 222 const int y = blockIdx.y * blockDim.y + threadIdx.y; in split() [all …]
|
D | minmaxloc.hpp | 67 const int x0 = blockIdx.x * blockDim.x * patch_x + threadIdx.x; in minMaxLoc_pass_1() 68 const int y0 = blockIdx.y * blockDim.y * patch_y + threadIdx.y; in minMaxLoc_pass_1() 75 for (int i = 0, y = y0; i < patch_y && y < rows; ++i, y += blockDim.y) in minMaxLoc_pass_1() 77 for (int j = 0, x = x0; j < patch_x && x < cols; ++j, x += blockDim.x) in minMaxLoc_pass_1() 98 const int tid = threadIdx.y * blockDim.x + threadIdx.x; in minMaxLoc_pass_1()
|
D | copy.hpp | 62 const int x = blockIdx.x * blockDim.x + threadIdx.x; in copy() 63 const int y = blockIdx.y * blockDim.y + threadIdx.y; in copy() 107 const int x = blockIdx.x * blockDim.x + threadIdx.x; in copy_tuple() 108 const int y = blockIdx.y * blockDim.y + threadIdx.y; in copy_tuple()
|
D | histogram.hpp | 62 const int y = blockIdx.x * blockDim.y + threadIdx.y; in histogram() 63 const int tid = threadIdx.y * blockDim.x + threadIdx.x; in histogram() 72 for (int x = threadIdx.x; x < cols; x += blockDim.x) in histogram()
|
/external/opencv3/modules/cudafeatures2d/src/cuda/ |
D | orb.cu | 78 const int ptidx = blockIdx.x * blockDim.y + threadIdx.y; in HarrisResponses() 90 for (int ind = threadIdx.x; ind < blockSize * blockSize; ind += blockDim.x) in HarrisResponses() 108 int* srow0 = smem0 + threadIdx.y * blockDim.x; in HarrisResponses() 109 int* srow1 = smem1 + threadIdx.y * blockDim.x; in HarrisResponses() 110 int* srow2 = smem2 + threadIdx.y * blockDim.x; in HarrisResponses() 156 int* srow0 = smem0 + threadIdx.y * blockDim.x; in IC_Angle() 157 int* srow1 = smem1 + threadIdx.y * blockDim.x; in IC_Angle() 161 const int ptidx = blockIdx.x * blockDim.y + threadIdx.y; in IC_Angle() 170 for (int u = threadIdx.x - half_k; u <= half_k; u += blockDim.x) in IC_Angle() 182 for (int u = threadIdx.x - d; u <= d; u += blockDim.x) in IC_Angle() [all …]
|
/external/opencv3/modules/cudev/include/opencv2/cudev/block/ |
D | block.hpp | 65 return blockDim.x * blockDim.y * blockDim.z; in blockSize() 70 return (threadIdx.z * blockDim.y + threadIdx.y) * blockDim.x + threadIdx.x; in threadLineId()
|
/external/clang/test/CodeGenCUDA/ |
D | cuda-builtin-vars.cu | 17 out[i++] = blockDim.x; // CHECK: call i32 @llvm.ptx.read.ntid.x() in kernel() 18 out[i++] = blockDim.y; // CHECK: call i32 @llvm.ptx.read.ntid.y() in kernel() 19 out[i++] = blockDim.z; // CHECK: call i32 @llvm.ptx.read.ntid.z() in kernel()
|
/external/opencv3/modules/core/include/opencv2/core/cuda/ |
D | block.hpp | 63 return blockDim.x * blockDim.y * blockDim.z; in stride() 73 return threadIdx.z * blockDim.x * blockDim.y + threadIdx.y * blockDim.x + threadIdx.x; in flattenedThreadId()
|
/external/opencv3/modules/cudaoptflow/src/cuda/ |
D | pyrlk.cu | 133 const unsigned int tid = threadIdx.y * blockDim.x + threadIdx.x; in sparseKernel() 162 for (int yBase = threadIdx.y, i = 0; yBase < c_winSize_y; yBase += blockDim.y, ++i) in sparseKernel() 164 for (int xBase = threadIdx.x, j = 0; xBase < c_winSize_x; xBase += blockDim.x, ++j) in sparseKernel() 241 for (int y = threadIdx.y, i = 0; y < c_winSize_y; y += blockDim.y, ++i) in sparseKernel() 243 for (int x = threadIdx.x, j = 0; x < c_winSize_x; x += blockDim.x, ++j) in sparseKernel() 284 for (int y = threadIdx.y, i = 0; y < c_winSize_y; y += blockDim.y, ++i) in sparseKernel() 286 for (int x = threadIdx.x, j = 0; x < c_winSize_x; x += blockDim.x, ++j) in sparseKernel() 334 const int patchWidth = blockDim.x + 2 * c_halfWin_x; in denseKernel() 335 const int patchHeight = blockDim.y + 2 * c_halfWin_y; in denseKernel() 341 const int xBase = blockIdx.x * blockDim.x; in denseKernel() [all …]
|
D | tvl1flow.cu | 59 const int x = blockIdx.x * blockDim.x + threadIdx.x; in centeredGradientKernel() 60 const int y = blockIdx.y * blockDim.y + threadIdx.y; in centeredGradientKernel() 110 const int x = blockIdx.x * blockDim.x + threadIdx.x; in warpBackwardKernel() 111 const int y = blockIdx.y * blockDim.y + threadIdx.y; in warpBackwardKernel() 223 const int x = blockIdx.x * blockDim.x + threadIdx.x; in estimateUKernel() 224 const int y = blockIdx.y * blockDim.y + threadIdx.y; in estimateUKernel() 322 const int x = blockIdx.x * blockDim.x + threadIdx.x; in estimateDualVariablesKernel() 323 const int y = blockIdx.y * blockDim.y + threadIdx.y; in estimateDualVariablesKernel()
|
/external/opencv3/modules/cudastereo/src/cuda/ |
D | util.cu | 59 const int x = blockIdx.x * blockDim.x + threadIdx.x; in reprojectImageTo3D() 60 const int y = blockIdx.y * blockDim.y + threadIdx.y; in reprojectImageTo3D() 168 const int x = (blockIdx.x * blockDim.x + threadIdx.x) << 2; in drawColorDisp() 169 const int y = blockIdx.y * blockDim.y + threadIdx.y; in drawColorDisp() 188 const int x = (blockIdx.x * blockDim.x + threadIdx.x) << 1; in drawColorDisp() 189 const int y = blockIdx.y * blockDim.y + threadIdx.y; in drawColorDisp()
|
/external/opencv3/modules/superres/src/cuda/ |
D | btv_l1_gpu.cu | 78 const int x = blockIdx.x * blockDim.x + threadIdx.x; in buildMotionMapsKernel() 79 const int y = blockIdx.y * blockDim.y + threadIdx.y; in buildMotionMapsKernel() 117 const int x = blockIdx.x * blockDim.x + threadIdx.x; in upscaleKernel() 118 const int y = blockIdx.y * blockDim.y + threadIdx.y; in upscaleKernel() 197 const int x = blockIdx.x * blockDim.x + threadIdx.x + ksize; in calcBtvRegularizationKernel() 198 const int y = blockIdx.y * blockDim.y + threadIdx.y + ksize; in calcBtvRegularizationKernel()
|