Lines Matching refs:H
128 #define VAR(W, H) \ argument
129 uint32_t vpx_variance##W##x##H##_c(const uint8_t *src_ptr, int src_stride, \
133 variance(src_ptr, src_stride, ref_ptr, ref_stride, W, H, sse, &sum); \
134 return *sse - (uint32_t)(((int64_t)sum * sum) / (W * H)); \
137 #define SUBPIX_VAR(W, H) \ argument
138 uint32_t vpx_sub_pixel_variance##W##x##H##_c( \
141 uint16_t fdata3[(H + 1) * W]; \
142 uint8_t temp2[H * W]; \
144 var_filter_block2d_bil_first_pass(src_ptr, fdata3, src_stride, 1, H + 1, \
146 var_filter_block2d_bil_second_pass(fdata3, temp2, W, W, H, W, \
149 return vpx_variance##W##x##H##_c(temp2, W, ref_ptr, ref_stride, sse); \
152 #define SUBPIX_AVG_VAR(W, H) \ argument
153 uint32_t vpx_sub_pixel_avg_variance##W##x##H##_c( \
157 uint16_t fdata3[(H + 1) * W]; \
158 uint8_t temp2[H * W]; \
159 DECLARE_ALIGNED(16, uint8_t, temp3[H * W]); \
161 var_filter_block2d_bil_first_pass(src_ptr, fdata3, src_stride, 1, H + 1, \
163 var_filter_block2d_bil_second_pass(fdata3, temp2, W, W, H, W, \
166 vpx_comp_avg_pred_c(temp3, second_pred, W, H, temp2, W); \
168 return vpx_variance##W##x##H##_c(temp3, W, ref_ptr, ref_stride, sse); \
175 #define GET_VAR(W, H) \ argument
176 void vpx_get##W##x##H##var_c(const uint8_t *src_ptr, int src_stride, \
179 variance(src_ptr, src_stride, ref_ptr, ref_stride, W, H, sse, sum); \
186 #define MSE(W, H) \ argument
187 uint32_t vpx_mse##W##x##H##_c(const uint8_t *src_ptr, int src_stride, \
191 variance(src_ptr, src_stride, ref_ptr, ref_stride, W, H, sse, &sum); \
196 #define VARIANCES(W, H) \ argument
197 VAR(W, H) \
198 SUBPIX_VAR(W, H) \
199 SUBPIX_AVG_VAR(W, H)
293 #define HIGHBD_VAR(W, H) \ argument
294 uint32_t vpx_highbd_8_variance##W##x##H##_c( \
298 highbd_8_variance(src_ptr, src_stride, ref_ptr, ref_stride, W, H, sse, \
300 return *sse - (uint32_t)(((int64_t)sum * sum) / (W * H)); \
303 uint32_t vpx_highbd_10_variance##W##x##H##_c( \
308 highbd_10_variance(src_ptr, src_stride, ref_ptr, ref_stride, W, H, sse, \
310 var = (int64_t)(*sse) - (((int64_t)sum * sum) / (W * H)); \
314 uint32_t vpx_highbd_12_variance##W##x##H##_c( \
319 highbd_12_variance(src_ptr, src_stride, ref_ptr, ref_stride, W, H, sse, \
321 var = (int64_t)(*sse) - (((int64_t)sum * sum) / (W * H)); \
347 #define HIGHBD_MSE(W, H) \ argument
348 uint32_t vpx_highbd_8_mse##W##x##H##_c( \
352 highbd_8_variance(src_ptr, src_stride, ref_ptr, ref_stride, W, H, sse, \
357 uint32_t vpx_highbd_10_mse##W##x##H##_c( \
361 highbd_10_variance(src_ptr, src_stride, ref_ptr, ref_stride, W, H, sse, \
366 uint32_t vpx_highbd_12_mse##W##x##H##_c( \
370 highbd_12_variance(src_ptr, src_stride, ref_ptr, ref_stride, W, H, sse, \
417 #define HIGHBD_SUBPIX_VAR(W, H) \ argument
418 uint32_t vpx_highbd_8_sub_pixel_variance##W##x##H##_c( \
421 uint16_t fdata3[(H + 1) * W]; \
422 uint16_t temp2[H * W]; \
425 src_ptr, fdata3, src_stride, 1, H + 1, W, bilinear_filters[x_offset]); \
426 highbd_var_filter_block2d_bil_second_pass(fdata3, temp2, W, W, H, W, \
429 return vpx_highbd_8_variance##W##x##H##_c(CONVERT_TO_BYTEPTR(temp2), W, \
433 uint32_t vpx_highbd_10_sub_pixel_variance##W##x##H##_c( \
436 uint16_t fdata3[(H + 1) * W]; \
437 uint16_t temp2[H * W]; \
440 src_ptr, fdata3, src_stride, 1, H + 1, W, bilinear_filters[x_offset]); \
441 highbd_var_filter_block2d_bil_second_pass(fdata3, temp2, W, W, H, W, \
444 return vpx_highbd_10_variance##W##x##H##_c(CONVERT_TO_BYTEPTR(temp2), W, \
448 uint32_t vpx_highbd_12_sub_pixel_variance##W##x##H##_c( \
451 uint16_t fdata3[(H + 1) * W]; \
452 uint16_t temp2[H * W]; \
455 src_ptr, fdata3, src_stride, 1, H + 1, W, bilinear_filters[x_offset]); \
456 highbd_var_filter_block2d_bil_second_pass(fdata3, temp2, W, W, H, W, \
459 return vpx_highbd_12_variance##W##x##H##_c(CONVERT_TO_BYTEPTR(temp2), W, \
463 #define HIGHBD_SUBPIX_AVG_VAR(W, H) \ argument
464 uint32_t vpx_highbd_8_sub_pixel_avg_variance##W##x##H##_c( \
468 uint16_t fdata3[(H + 1) * W]; \
469 uint16_t temp2[H * W]; \
470 DECLARE_ALIGNED(16, uint16_t, temp3[H * W]); \
473 src_ptr, fdata3, src_stride, 1, H + 1, W, bilinear_filters[x_offset]); \
474 highbd_var_filter_block2d_bil_second_pass(fdata3, temp2, W, W, H, W, \
477 vpx_highbd_comp_avg_pred_c(temp3, CONVERT_TO_SHORTPTR(second_pred), W, H, \
480 return vpx_highbd_8_variance##W##x##H##_c(CONVERT_TO_BYTEPTR(temp3), W, \
484 uint32_t vpx_highbd_10_sub_pixel_avg_variance##W##x##H##_c( \
488 uint16_t fdata3[(H + 1) * W]; \
489 uint16_t temp2[H * W]; \
490 DECLARE_ALIGNED(16, uint16_t, temp3[H * W]); \
493 src_ptr, fdata3, src_stride, 1, H + 1, W, bilinear_filters[x_offset]); \
494 highbd_var_filter_block2d_bil_second_pass(fdata3, temp2, W, W, H, W, \
497 vpx_highbd_comp_avg_pred_c(temp3, CONVERT_TO_SHORTPTR(second_pred), W, H, \
500 return vpx_highbd_10_variance##W##x##H##_c(CONVERT_TO_BYTEPTR(temp3), W, \
504 uint32_t vpx_highbd_12_sub_pixel_avg_variance##W##x##H##_c( \
508 uint16_t fdata3[(H + 1) * W]; \
509 uint16_t temp2[H * W]; \
510 DECLARE_ALIGNED(16, uint16_t, temp3[H * W]); \
513 src_ptr, fdata3, src_stride, 1, H + 1, W, bilinear_filters[x_offset]); \
514 highbd_var_filter_block2d_bil_second_pass(fdata3, temp2, W, W, H, W, \
517 vpx_highbd_comp_avg_pred_c(temp3, CONVERT_TO_SHORTPTR(second_pred), W, H, \
520 return vpx_highbd_12_variance##W##x##H##_c(CONVERT_TO_BYTEPTR(temp3), W, \
525 #define HIGHBD_VARIANCES(W, H) \ argument
526 HIGHBD_VAR(W, H) \
527 HIGHBD_SUBPIX_VAR(W, H) \
528 HIGHBD_SUBPIX_AVG_VAR(W, H)