/*M/////////////////////////////////////////////////////////////////////////////////////// // // IMPORTANT: READ BEFORE DOWNLOADING, COPYING, INSTALLING OR USING. // // By downloading, copying, installing or using the software you agree to this license. // If you do not agree to this license, do not download, install, // copy or use the software. // // // License Agreement // For Open Source Computer Vision Library // // Copyright (C) 2000-2008, Intel Corporation, all rights reserved. // Copyright (C) 2009-2010, Willow Garage Inc., all rights reserved. // Copyright (C) 2014-2015, Itseez Inc., all rights reserved. // Third party copyrights are property of their respective owners. // // Redistribution and use in source and binary forms, with or without modification, // are permitted provided that the following conditions are met: // // * Redistribution's of source code must retain the above copyright notice, // this list of conditions and the following disclaimer. // // * Redistribution's in binary form must reproduce the above copyright notice, // this list of conditions and the following disclaimer in the documentation // and/or other materials provided with the distribution. // // * The name of the copyright holders may not be used to endorse or promote products // derived from this software without specific prior written permission. // // This software is provided by the copyright holders and contributors "as is" and // any express or implied warranties, including, but not limited to, the implied // warranties of merchantability and fitness for a particular purpose are disclaimed. // In no event shall the Intel Corporation or contributors be liable for any direct, // indirect, incidental, special, exemplary, or consequential damages // (including, but not limited to, procurement of substitute goods or services; // loss of use, data, or profits; or business interruption) however caused // and on any theory of liability, whether in contract, strict liability, // or tort (including negligence or otherwise) arising in any way out of // the use of this software, even if advised of the possibility of such damage. // //M*/ /********************************* COPYRIGHT NOTICE *******************************\ The function for RGB to Lab conversion is based on the MATLAB script RGB2Lab.m translated by Mark Ruzon from C code by Yossi Rubner, 23 September 1997. See the page [http://vision.stanford.edu/~ruzon/software/rgblab.html] \**********************************************************************************/ /********************************* COPYRIGHT NOTICE *******************************\ Original code for Bayer->BGR/RGB conversion is provided by Dirk Schaefer from MD-Mathematische Dienste GmbH. Below is the copyright notice: IMPORTANT: READ BEFORE DOWNLOADING, COPYING, INSTALLING OR USING. By downloading, copying, installing or using the software you agree to this license. If you do not agree to this license, do not download, install, copy or use the software. Contributors License Agreement: Copyright (c) 2002, MD-Mathematische Dienste GmbH Im Defdahl 5-10 44141 Dortmund Germany www.md-it.de Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. The name of Contributor may not be used to endorse or promote products derived from this software without specific prior written permission. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. \**********************************************************************************/ #include "precomp.hpp" #include "opencl_kernels_imgproc.hpp" #include #define CV_DESCALE(x,n) (((x) + (1 << ((n)-1))) >> (n)) #if defined (HAVE_IPP) && (IPP_VERSION_MAJOR >= 7) #define MAX_IPP8u 255 #define MAX_IPP16u 65535 #define MAX_IPP32f 1.0 static IppStatus sts = ippInit(); #endif namespace cv { // computes cubic spline coefficients for a function: (xi=i, yi=f[i]), i=0..n template static void splineBuild(const _Tp* f, int n, _Tp* tab) { _Tp cn = 0; int i; tab[0] = tab[1] = (_Tp)0; for(i = 1; i < n-1; i++) { _Tp t = 3*(f[i+1] - 2*f[i] + f[i-1]); _Tp l = 1/(4 - tab[(i-1)*4]); tab[i*4] = l; tab[i*4+1] = (t - tab[(i-1)*4+1])*l; } for(i = n-1; i >= 0; i--) { _Tp c = tab[i*4+1] - tab[i*4]*cn; _Tp b = f[i+1] - f[i] - (cn + c*2)*(_Tp)0.3333333333333333; _Tp d = (cn - c)*(_Tp)0.3333333333333333; tab[i*4] = f[i]; tab[i*4+1] = b; tab[i*4+2] = c; tab[i*4+3] = d; cn = c; } } // interpolates value of a function at x, 0 <= x <= n using a cubic spline. template static inline _Tp splineInterpolate(_Tp x, const _Tp* tab, int n) { // don't touch this function without urgent need - some versions of gcc fail to inline it correctly int ix = std::min(std::max(int(x), 0), n-1); x -= ix; tab += ix*4; return ((tab[3]*x + tab[2])*x + tab[1])*x + tab[0]; } template struct ColorChannel { typedef float worktype_f; static _Tp max() { return std::numeric_limits<_Tp>::max(); } static _Tp half() { return (_Tp)(max()/2 + 1); } }; template<> struct ColorChannel { typedef float worktype_f; static float max() { return 1.f; } static float half() { return 0.5f; } }; /*template<> struct ColorChannel { typedef double worktype_f; static double max() { return 1.; } static double half() { return 0.5; } };*/ ///////////////////////////// Top-level template function //////////////////////////////// template class CvtColorLoop_Invoker : public ParallelLoopBody { typedef typename Cvt::channel_type _Tp; public: CvtColorLoop_Invoker(const Mat& _src, Mat& _dst, const Cvt& _cvt) : ParallelLoopBody(), src(_src), dst(_dst), cvt(_cvt) { } virtual void operator()(const Range& range) const { const uchar* yS = src.ptr(range.start); uchar* yD = dst.ptr(range.start); for( int i = range.start; i < range.end; ++i, yS += src.step, yD += dst.step ) cvt((const _Tp*)yS, (_Tp*)yD, src.cols); } private: const Mat& src; Mat& dst; const Cvt& cvt; const CvtColorLoop_Invoker& operator= (const CvtColorLoop_Invoker&); }; template void CvtColorLoop(const Mat& src, Mat& dst, const Cvt& cvt) { parallel_for_(Range(0, src.rows), CvtColorLoop_Invoker(src, dst, cvt), src.total()/(double)(1<<16) ); } #if defined (HAVE_IPP) && (IPP_VERSION_MAJOR >= 7) typedef IppStatus (CV_STDCALL* ippiReorderFunc)(const void *, int, void *, int, IppiSize, const int *); typedef IppStatus (CV_STDCALL* ippiGeneralFunc)(const void *, int, void *, int, IppiSize); typedef IppStatus (CV_STDCALL* ippiColor2GrayFunc)(const void *, int, void *, int, IppiSize, const Ipp32f *); template class CvtColorIPPLoop_Invoker : public ParallelLoopBody { public: CvtColorIPPLoop_Invoker(const Mat& _src, Mat& _dst, const Cvt& _cvt, bool *_ok) : ParallelLoopBody(), src(_src), dst(_dst), cvt(_cvt), ok(_ok) { *ok = true; } virtual void operator()(const Range& range) const { const void *yS = src.ptr(range.start); void *yD = dst.ptr(range.start); if( !cvt(yS, (int)src.step[0], yD, (int)dst.step[0], src.cols, range.end - range.start) ) *ok = false; else { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); } } private: const Mat& src; Mat& dst; const Cvt& cvt; bool *ok; const CvtColorIPPLoop_Invoker& operator= (const CvtColorIPPLoop_Invoker&); }; template bool CvtColorIPPLoop(const Mat& src, Mat& dst, const Cvt& cvt) { bool ok; parallel_for_(Range(0, src.rows), CvtColorIPPLoop_Invoker(src, dst, cvt, &ok), src.total()/(double)(1<<16) ); return ok; } template bool CvtColorIPPLoopCopy(Mat& src, Mat& dst, const Cvt& cvt) { Mat temp; Mat &source = src; if( src.data == dst.data ) { src.copyTo(temp); source = temp; } bool ok; parallel_for_(Range(0, source.rows), CvtColorIPPLoop_Invoker(source, dst, cvt, &ok), source.total()/(double)(1<<16) ); return ok; } static IppStatus CV_STDCALL ippiSwapChannels_8u_C3C4Rf(const Ipp8u* pSrc, int srcStep, Ipp8u* pDst, int dstStep, IppiSize roiSize, const int *dstOrder) { return ippiSwapChannels_8u_C3C4R(pSrc, srcStep, pDst, dstStep, roiSize, dstOrder, MAX_IPP8u); } static IppStatus CV_STDCALL ippiSwapChannels_16u_C3C4Rf(const Ipp16u* pSrc, int srcStep, Ipp16u* pDst, int dstStep, IppiSize roiSize, const int *dstOrder) { return ippiSwapChannels_16u_C3C4R(pSrc, srcStep, pDst, dstStep, roiSize, dstOrder, MAX_IPP16u); } static IppStatus CV_STDCALL ippiSwapChannels_32f_C3C4Rf(const Ipp32f* pSrc, int srcStep, Ipp32f* pDst, int dstStep, IppiSize roiSize, const int *dstOrder) { return ippiSwapChannels_32f_C3C4R(pSrc, srcStep, pDst, dstStep, roiSize, dstOrder, MAX_IPP32f); } static ippiReorderFunc ippiSwapChannelsC3C4RTab[] = { (ippiReorderFunc)ippiSwapChannels_8u_C3C4Rf, 0, (ippiReorderFunc)ippiSwapChannels_16u_C3C4Rf, 0, 0, (ippiReorderFunc)ippiSwapChannels_32f_C3C4Rf, 0, 0 }; static ippiGeneralFunc ippiCopyAC4C3RTab[] = { (ippiGeneralFunc)ippiCopy_8u_AC4C3R, 0, (ippiGeneralFunc)ippiCopy_16u_AC4C3R, 0, 0, (ippiGeneralFunc)ippiCopy_32f_AC4C3R, 0, 0 }; static ippiReorderFunc ippiSwapChannelsC4C3RTab[] = { (ippiReorderFunc)ippiSwapChannels_8u_C4C3R, 0, (ippiReorderFunc)ippiSwapChannels_16u_C4C3R, 0, 0, (ippiReorderFunc)ippiSwapChannels_32f_C4C3R, 0, 0 }; static ippiReorderFunc ippiSwapChannelsC3RTab[] = { (ippiReorderFunc)ippiSwapChannels_8u_C3R, 0, (ippiReorderFunc)ippiSwapChannels_16u_C3R, 0, 0, (ippiReorderFunc)ippiSwapChannels_32f_C3R, 0, 0 }; #if IPP_VERSION_X100 >= 801 static ippiReorderFunc ippiSwapChannelsC4RTab[] = { (ippiReorderFunc)ippiSwapChannels_8u_C4R, 0, (ippiReorderFunc)ippiSwapChannels_16u_C4R, 0, 0, (ippiReorderFunc)ippiSwapChannels_32f_C4R, 0, 0 }; #endif static ippiColor2GrayFunc ippiColor2GrayC3Tab[] = { (ippiColor2GrayFunc)ippiColorToGray_8u_C3C1R, 0, (ippiColor2GrayFunc)ippiColorToGray_16u_C3C1R, 0, 0, (ippiColor2GrayFunc)ippiColorToGray_32f_C3C1R, 0, 0 }; static ippiColor2GrayFunc ippiColor2GrayC4Tab[] = { (ippiColor2GrayFunc)ippiColorToGray_8u_AC4C1R, 0, (ippiColor2GrayFunc)ippiColorToGray_16u_AC4C1R, 0, 0, (ippiColor2GrayFunc)ippiColorToGray_32f_AC4C1R, 0, 0 }; static ippiGeneralFunc ippiRGB2GrayC3Tab[] = { (ippiGeneralFunc)ippiRGBToGray_8u_C3C1R, 0, (ippiGeneralFunc)ippiRGBToGray_16u_C3C1R, 0, 0, (ippiGeneralFunc)ippiRGBToGray_32f_C3C1R, 0, 0 }; static ippiGeneralFunc ippiRGB2GrayC4Tab[] = { (ippiGeneralFunc)ippiRGBToGray_8u_AC4C1R, 0, (ippiGeneralFunc)ippiRGBToGray_16u_AC4C1R, 0, 0, (ippiGeneralFunc)ippiRGBToGray_32f_AC4C1R, 0, 0 }; static ippiGeneralFunc ippiCopyP3C3RTab[] = { (ippiGeneralFunc)ippiCopy_8u_P3C3R, 0, (ippiGeneralFunc)ippiCopy_16u_P3C3R, 0, 0, (ippiGeneralFunc)ippiCopy_32f_P3C3R, 0, 0 }; static ippiGeneralFunc ippiRGB2XYZTab[] = { (ippiGeneralFunc)ippiRGBToXYZ_8u_C3R, 0, (ippiGeneralFunc)ippiRGBToXYZ_16u_C3R, 0, 0, (ippiGeneralFunc)ippiRGBToXYZ_32f_C3R, 0, 0 }; static ippiGeneralFunc ippiXYZ2RGBTab[] = { (ippiGeneralFunc)ippiXYZToRGB_8u_C3R, 0, (ippiGeneralFunc)ippiXYZToRGB_16u_C3R, 0, 0, (ippiGeneralFunc)ippiXYZToRGB_32f_C3R, 0, 0 }; static ippiGeneralFunc ippiRGB2HSVTab[] = { (ippiGeneralFunc)ippiRGBToHSV_8u_C3R, 0, (ippiGeneralFunc)ippiRGBToHSV_16u_C3R, 0, 0, 0, 0, 0 }; static ippiGeneralFunc ippiHSV2RGBTab[] = { (ippiGeneralFunc)ippiHSVToRGB_8u_C3R, 0, (ippiGeneralFunc)ippiHSVToRGB_16u_C3R, 0, 0, 0, 0, 0 }; static ippiGeneralFunc ippiRGB2HLSTab[] = { (ippiGeneralFunc)ippiRGBToHLS_8u_C3R, 0, (ippiGeneralFunc)ippiRGBToHLS_16u_C3R, 0, 0, (ippiGeneralFunc)ippiRGBToHLS_32f_C3R, 0, 0 }; static ippiGeneralFunc ippiHLS2RGBTab[] = { (ippiGeneralFunc)ippiHLSToRGB_8u_C3R, 0, (ippiGeneralFunc)ippiHLSToRGB_16u_C3R, 0, 0, (ippiGeneralFunc)ippiHLSToRGB_32f_C3R, 0, 0 }; #if !defined(HAVE_IPP_ICV_ONLY) && 0 static ippiGeneralFunc ippiRGBToLUVTab[] = { (ippiGeneralFunc)ippiRGBToLUV_8u_C3R, 0, (ippiGeneralFunc)ippiRGBToLUV_16u_C3R, 0, 0, (ippiGeneralFunc)ippiRGBToLUV_32f_C3R, 0, 0 }; static ippiGeneralFunc ippiLUVToRGBTab[] = { (ippiGeneralFunc)ippiLUVToRGB_8u_C3R, 0, (ippiGeneralFunc)ippiLUVToRGB_16u_C3R, 0, 0, (ippiGeneralFunc)ippiLUVToRGB_32f_C3R, 0, 0 }; #endif struct IPPGeneralFunctor { IPPGeneralFunctor(ippiGeneralFunc _func) : func(_func){} bool operator()(const void *src, int srcStep, void *dst, int dstStep, int cols, int rows) const { return func ? func(src, srcStep, dst, dstStep, ippiSize(cols, rows)) >= 0 : false; } private: ippiGeneralFunc func; }; struct IPPReorderFunctor { IPPReorderFunctor(ippiReorderFunc _func, int _order0, int _order1, int _order2) : func(_func) { order[0] = _order0; order[1] = _order1; order[2] = _order2; order[3] = 3; } bool operator()(const void *src, int srcStep, void *dst, int dstStep, int cols, int rows) const { return func ? func(src, srcStep, dst, dstStep, ippiSize(cols, rows), order) >= 0 : false; } private: ippiReorderFunc func; int order[4]; }; struct IPPColor2GrayFunctor { IPPColor2GrayFunctor(ippiColor2GrayFunc _func) : func(_func) { coeffs[0] = 0.114f; coeffs[1] = 0.587f; coeffs[2] = 0.299f; } bool operator()(const void *src, int srcStep, void *dst, int dstStep, int cols, int rows) const { return func ? func(src, srcStep, dst, dstStep, ippiSize(cols, rows), coeffs) >= 0 : false; } private: ippiColor2GrayFunc func; Ipp32f coeffs[3]; }; struct IPPGray2BGRFunctor { IPPGray2BGRFunctor(ippiGeneralFunc _func) : func(_func) { } bool operator()(const void *src, int srcStep, void *dst, int dstStep, int cols, int rows) const { if (func == 0) return false; const void* srcarray[3] = { src, src, src }; return func(srcarray, srcStep, dst, dstStep, ippiSize(cols, rows)) >= 0; } private: ippiGeneralFunc func; }; struct IPPGray2BGRAFunctor { IPPGray2BGRAFunctor(ippiGeneralFunc _func1, ippiReorderFunc _func2, int _depth) : func1(_func1), func2(_func2), depth(_depth) { } bool operator()(const void *src, int srcStep, void *dst, int dstStep, int cols, int rows) const { if (func1 == 0 || func2 == 0) return false; const void* srcarray[3] = { src, src, src }; Mat temp(rows, cols, CV_MAKETYPE(depth, 3)); if(func1(srcarray, srcStep, temp.ptr(), (int)temp.step[0], ippiSize(cols, rows)) < 0) return false; int order[4] = {0, 1, 2, 3}; return func2(temp.ptr(), (int)temp.step[0], dst, dstStep, ippiSize(cols, rows), order) >= 0; } private: ippiGeneralFunc func1; ippiReorderFunc func2; int depth; }; struct IPPReorderGeneralFunctor { IPPReorderGeneralFunctor(ippiReorderFunc _func1, ippiGeneralFunc _func2, int _order0, int _order1, int _order2, int _depth) : func1(_func1), func2(_func2), depth(_depth) { order[0] = _order0; order[1] = _order1; order[2] = _order2; order[3] = 3; } bool operator()(const void *src, int srcStep, void *dst, int dstStep, int cols, int rows) const { if (func1 == 0 || func2 == 0) return false; Mat temp; temp.create(rows, cols, CV_MAKETYPE(depth, 3)); if(func1(src, srcStep, temp.ptr(), (int)temp.step[0], ippiSize(cols, rows), order) < 0) return false; return func2(temp.ptr(), (int)temp.step[0], dst, dstStep, ippiSize(cols, rows)) >= 0; } private: ippiReorderFunc func1; ippiGeneralFunc func2; int order[4]; int depth; }; struct IPPGeneralReorderFunctor { IPPGeneralReorderFunctor(ippiGeneralFunc _func1, ippiReorderFunc _func2, int _order0, int _order1, int _order2, int _depth) : func1(_func1), func2(_func2), depth(_depth) { order[0] = _order0; order[1] = _order1; order[2] = _order2; order[3] = 3; } bool operator()(const void *src, int srcStep, void *dst, int dstStep, int cols, int rows) const { if (func1 == 0 || func2 == 0) return false; Mat temp; temp.create(rows, cols, CV_MAKETYPE(depth, 3)); if(func1(src, srcStep, temp.ptr(), (int)temp.step[0], ippiSize(cols, rows)) < 0) return false; return func2(temp.ptr(), (int)temp.step[0], dst, dstStep, ippiSize(cols, rows), order) >= 0; } private: ippiGeneralFunc func1; ippiReorderFunc func2; int order[4]; int depth; }; #endif ////////////////// Various 3/4-channel to 3/4-channel RGB transformations ///////////////// template struct RGB2RGB { typedef _Tp channel_type; RGB2RGB(int _srccn, int _dstcn, int _blueIdx) : srccn(_srccn), dstcn(_dstcn), blueIdx(_blueIdx) {} void operator()(const _Tp* src, _Tp* dst, int n) const { int scn = srccn, dcn = dstcn, bidx = blueIdx; if( dcn == 3 ) { n *= 3; for( int i = 0; i < n; i += 3, src += scn ) { _Tp t0 = src[bidx], t1 = src[1], t2 = src[bidx ^ 2]; dst[i] = t0; dst[i+1] = t1; dst[i+2] = t2; } } else if( scn == 3 ) { n *= 3; _Tp alpha = ColorChannel<_Tp>::max(); for( int i = 0; i < n; i += 3, dst += 4 ) { _Tp t0 = src[i], t1 = src[i+1], t2 = src[i+2]; dst[bidx] = t0; dst[1] = t1; dst[bidx^2] = t2; dst[3] = alpha; } } else { n *= 4; for( int i = 0; i < n; i += 4 ) { _Tp t0 = src[i], t1 = src[i+1], t2 = src[i+2], t3 = src[i+3]; dst[i] = t2; dst[i+1] = t1; dst[i+2] = t0; dst[i+3] = t3; } } } int srccn, dstcn, blueIdx; }; #if CV_NEON template<> struct RGB2RGB { typedef uchar channel_type; RGB2RGB(int _srccn, int _dstcn, int _blueIdx) : srccn(_srccn), dstcn(_dstcn), blueIdx(_blueIdx) { v_alpha = vdupq_n_u8(ColorChannel::max()); v_alpha2 = vget_low_u8(v_alpha); } void operator()(const uchar * src, uchar * dst, int n) const { int scn = srccn, dcn = dstcn, bidx = blueIdx, i = 0; if (dcn == 3) { n *= 3; if (scn == 3) { for ( ; i <= n - 48; i += 48, src += 48 ) { uint8x16x3_t v_src = vld3q_u8(src), v_dst; v_dst.val[0] = v_src.val[bidx]; v_dst.val[1] = v_src.val[1]; v_dst.val[2] = v_src.val[bidx ^ 2]; vst3q_u8(dst + i, v_dst); } for ( ; i <= n - 24; i += 24, src += 24 ) { uint8x8x3_t v_src = vld3_u8(src), v_dst; v_dst.val[0] = v_src.val[bidx]; v_dst.val[1] = v_src.val[1]; v_dst.val[2] = v_src.val[bidx ^ 2]; vst3_u8(dst + i, v_dst); } for ( ; i < n; i += 3, src += 3 ) { uchar t0 = src[bidx], t1 = src[1], t2 = src[bidx ^ 2]; dst[i] = t0; dst[i+1] = t1; dst[i+2] = t2; } } else { for ( ; i <= n - 48; i += 48, src += 64 ) { uint8x16x4_t v_src = vld4q_u8(src); uint8x16x3_t v_dst; v_dst.val[0] = v_src.val[bidx]; v_dst.val[1] = v_src.val[1]; v_dst.val[2] = v_src.val[bidx ^ 2]; vst3q_u8(dst + i, v_dst); } for ( ; i <= n - 24; i += 24, src += 32 ) { uint8x8x4_t v_src = vld4_u8(src); uint8x8x3_t v_dst; v_dst.val[0] = v_src.val[bidx]; v_dst.val[1] = v_src.val[1]; v_dst.val[2] = v_src.val[bidx ^ 2]; vst3_u8(dst + i, v_dst); } for ( ; i < n; i += 3, src += 4 ) { uchar t0 = src[bidx], t1 = src[1], t2 = src[bidx ^ 2]; dst[i] = t0; dst[i+1] = t1; dst[i+2] = t2; } } } else if (scn == 3) { n *= 3; for ( ; i <= n - 48; i += 48, dst += 64 ) { uint8x16x3_t v_src = vld3q_u8(src + i); uint8x16x4_t v_dst; v_dst.val[bidx] = v_src.val[0]; v_dst.val[1] = v_src.val[1]; v_dst.val[bidx ^ 2] = v_src.val[2]; v_dst.val[3] = v_alpha; vst4q_u8(dst, v_dst); } for ( ; i <= n - 24; i += 24, dst += 32 ) { uint8x8x3_t v_src = vld3_u8(src + i); uint8x8x4_t v_dst; v_dst.val[bidx] = v_src.val[0]; v_dst.val[1] = v_src.val[1]; v_dst.val[bidx ^ 2] = v_src.val[2]; v_dst.val[3] = v_alpha2; vst4_u8(dst, v_dst); } uchar alpha = ColorChannel::max(); for (; i < n; i += 3, dst += 4 ) { uchar t0 = src[i], t1 = src[i+1], t2 = src[i+2]; dst[bidx] = t0; dst[1] = t1; dst[bidx^2] = t2; dst[3] = alpha; } } else { n *= 4; for ( ; i <= n - 64; i += 64 ) { uint8x16x4_t v_src = vld4q_u8(src + i), v_dst; v_dst.val[0] = v_src.val[2]; v_dst.val[1] = v_src.val[1]; v_dst.val[2] = v_src.val[0]; v_dst.val[3] = v_src.val[3]; vst4q_u8(dst + i, v_dst); } for ( ; i <= n - 32; i += 32 ) { uint8x8x4_t v_src = vld4_u8(src + i), v_dst; v_dst.val[0] = v_src.val[2]; v_dst.val[1] = v_src.val[1]; v_dst.val[2] = v_src.val[0]; v_dst.val[3] = v_src.val[3]; vst4_u8(dst + i, v_dst); } for ( ; i < n; i += 4) { uchar t0 = src[i], t1 = src[i+1], t2 = src[i+2], t3 = src[i+3]; dst[i] = t2; dst[i+1] = t1; dst[i+2] = t0; dst[i+3] = t3; } } } int srccn, dstcn, blueIdx; uint8x16_t v_alpha; uint8x8_t v_alpha2; }; #endif /////////// Transforming 16-bit (565 or 555) RGB to/from 24/32-bit (888[8]) RGB ////////// struct RGB5x52RGB { typedef uchar channel_type; RGB5x52RGB(int _dstcn, int _blueIdx, int _greenBits) : dstcn(_dstcn), blueIdx(_blueIdx), greenBits(_greenBits) { #if CV_NEON v_n3 = vdupq_n_u16(~3); v_n7 = vdupq_n_u16(~7); v_255 = vdupq_n_u8(255); v_0 = vdupq_n_u8(0); v_mask = vdupq_n_u16(0x8000); #endif } void operator()(const uchar* src, uchar* dst, int n) const { int dcn = dstcn, bidx = blueIdx, i = 0; if( greenBits == 6 ) { #if CV_NEON for ( ; i <= n - 16; i += 16, dst += dcn * 16) { uint16x8_t v_src0 = vld1q_u16((const ushort *)src + i), v_src1 = vld1q_u16((const ushort *)src + i + 8); uint8x16_t v_b = vcombine_u8(vmovn_u16(vshlq_n_u16(v_src0, 3)), vmovn_u16(vshlq_n_u16(v_src1, 3))); uint8x16_t v_g = vcombine_u8(vmovn_u16(vandq_u16(vshrq_n_u16(v_src0, 3), v_n3)), vmovn_u16(vandq_u16(vshrq_n_u16(v_src1, 3), v_n3))); uint8x16_t v_r = vcombine_u8(vmovn_u16(vandq_u16(vshrq_n_u16(v_src0, 8), v_n7)), vmovn_u16(vandq_u16(vshrq_n_u16(v_src1, 8), v_n7))); if (dcn == 3) { uint8x16x3_t v_dst; v_dst.val[bidx] = v_b; v_dst.val[1] = v_g; v_dst.val[bidx^2] = v_r; vst3q_u8(dst, v_dst); } else { uint8x16x4_t v_dst; v_dst.val[bidx] = v_b; v_dst.val[1] = v_g; v_dst.val[bidx^2] = v_r; v_dst.val[3] = v_255; vst4q_u8(dst, v_dst); } } #endif for( ; i < n; i++, dst += dcn ) { unsigned t = ((const ushort*)src)[i]; dst[bidx] = (uchar)(t << 3); dst[1] = (uchar)((t >> 3) & ~3); dst[bidx ^ 2] = (uchar)((t >> 8) & ~7); if( dcn == 4 ) dst[3] = 255; } } else { #if CV_NEON for ( ; i <= n - 16; i += 16, dst += dcn * 16) { uint16x8_t v_src0 = vld1q_u16((const ushort *)src + i), v_src1 = vld1q_u16((const ushort *)src + i + 8); uint8x16_t v_b = vcombine_u8(vmovn_u16(vshlq_n_u16(v_src0, 3)), vmovn_u16(vshlq_n_u16(v_src1, 3))); uint8x16_t v_g = vcombine_u8(vmovn_u16(vandq_u16(vshrq_n_u16(v_src0, 2), v_n7)), vmovn_u16(vandq_u16(vshrq_n_u16(v_src1, 2), v_n7))); uint8x16_t v_r = vcombine_u8(vmovn_u16(vandq_u16(vshrq_n_u16(v_src0, 7), v_n7)), vmovn_u16(vandq_u16(vshrq_n_u16(v_src1, 7), v_n7))); if (dcn == 3) { uint8x16x3_t v_dst; v_dst.val[bidx] = v_b; v_dst.val[1] = v_g; v_dst.val[bidx^2] = v_r; vst3q_u8(dst, v_dst); } else { uint8x16x4_t v_dst; v_dst.val[bidx] = v_b; v_dst.val[1] = v_g; v_dst.val[bidx^2] = v_r; v_dst.val[3] = vbslq_u8(vcombine_u8(vqmovn_u16(vandq_u16(v_src0, v_mask)), vqmovn_u16(vandq_u16(v_src1, v_mask))), v_255, v_0); vst4q_u8(dst, v_dst); } } #endif for( ; i < n; i++, dst += dcn ) { unsigned t = ((const ushort*)src)[i]; dst[bidx] = (uchar)(t << 3); dst[1] = (uchar)((t >> 2) & ~7); dst[bidx ^ 2] = (uchar)((t >> 7) & ~7); if( dcn == 4 ) dst[3] = t & 0x8000 ? 255 : 0; } } } int dstcn, blueIdx, greenBits; #if CV_NEON uint16x8_t v_n3, v_n7, v_mask; uint8x16_t v_255, v_0; #endif }; struct RGB2RGB5x5 { typedef uchar channel_type; RGB2RGB5x5(int _srccn, int _blueIdx, int _greenBits) : srccn(_srccn), blueIdx(_blueIdx), greenBits(_greenBits) { #if CV_NEON v_n3 = vdup_n_u8(~3); v_n7 = vdup_n_u8(~7); v_mask = vdupq_n_u16(0x8000); v_0 = vdupq_n_u16(0); v_full = vdupq_n_u16(0xffff); #endif } void operator()(const uchar* src, uchar* dst, int n) const { int scn = srccn, bidx = blueIdx, i = 0; if (greenBits == 6) { if (scn == 3) { #if CV_NEON for ( ; i <= n - 8; i += 8, src += 24 ) { uint8x8x3_t v_src = vld3_u8(src); uint16x8_t v_dst = vmovl_u8(vshr_n_u8(v_src.val[bidx], 3)); v_dst = vorrq_u16(v_dst, vshlq_n_u16(vmovl_u8(vand_u8(v_src.val[1], v_n3)), 3)); v_dst = vorrq_u16(v_dst, vshlq_n_u16(vmovl_u8(vand_u8(v_src.val[bidx^2], v_n7)), 8)); vst1q_u16((ushort *)dst + i, v_dst); } #endif for ( ; i < n; i++, src += 3 ) ((ushort*)dst)[i] = (ushort)((src[bidx] >> 3)|((src[1]&~3) << 3)|((src[bidx^2]&~7) << 8)); } else { #if CV_NEON for ( ; i <= n - 8; i += 8, src += 32 ) { uint8x8x4_t v_src = vld4_u8(src); uint16x8_t v_dst = vmovl_u8(vshr_n_u8(v_src.val[bidx], 3)); v_dst = vorrq_u16(v_dst, vshlq_n_u16(vmovl_u8(vand_u8(v_src.val[1], v_n3)), 3)); v_dst = vorrq_u16(v_dst, vshlq_n_u16(vmovl_u8(vand_u8(v_src.val[bidx^2], v_n7)), 8)); vst1q_u16((ushort *)dst + i, v_dst); } #endif for ( ; i < n; i++, src += 4 ) ((ushort*)dst)[i] = (ushort)((src[bidx] >> 3)|((src[1]&~3) << 3)|((src[bidx^2]&~7) << 8)); } } else if (scn == 3) { #if CV_NEON for ( ; i <= n - 8; i += 8, src += 24 ) { uint8x8x3_t v_src = vld3_u8(src); uint16x8_t v_dst = vmovl_u8(vshr_n_u8(v_src.val[bidx], 3)); v_dst = vorrq_u16(v_dst, vshlq_n_u16(vmovl_u8(vand_u8(v_src.val[1], v_n7)), 2)); v_dst = vorrq_u16(v_dst, vshlq_n_u16(vmovl_u8(vand_u8(v_src.val[bidx^2], v_n7)), 7)); vst1q_u16((ushort *)dst + i, v_dst); } #endif for ( ; i < n; i++, src += 3 ) ((ushort*)dst)[i] = (ushort)((src[bidx] >> 3)|((src[1]&~7) << 2)|((src[bidx^2]&~7) << 7)); } else { #if CV_NEON for ( ; i <= n - 8; i += 8, src += 32 ) { uint8x8x4_t v_src = vld4_u8(src); uint16x8_t v_dst = vmovl_u8(vshr_n_u8(v_src.val[bidx], 3)); v_dst = vorrq_u16(v_dst, vshlq_n_u16(vmovl_u8(vand_u8(v_src.val[1], v_n7)), 2)); v_dst = vorrq_u16(v_dst, vorrq_u16(vshlq_n_u16(vmovl_u8(vand_u8(v_src.val[bidx^2], v_n7)), 7), vbslq_u16(veorq_u16(vceqq_u16(vmovl_u8(v_src.val[3]), v_0), v_full), v_mask, v_0))); vst1q_u16((ushort *)dst + i, v_dst); } #endif for ( ; i < n; i++, src += 4 ) ((ushort*)dst)[i] = (ushort)((src[bidx] >> 3)|((src[1]&~7) << 2)| ((src[bidx^2]&~7) << 7)|(src[3] ? 0x8000 : 0)); } } int srccn, blueIdx, greenBits; #if CV_NEON uint8x8_t v_n3, v_n7; uint16x8_t v_mask, v_0, v_full; #endif }; ///////////////////////////////// Color to/from Grayscale //////////////////////////////// template struct Gray2RGB { typedef _Tp channel_type; Gray2RGB(int _dstcn) : dstcn(_dstcn) {} void operator()(const _Tp* src, _Tp* dst, int n) const { if( dstcn == 3 ) for( int i = 0; i < n; i++, dst += 3 ) { dst[0] = dst[1] = dst[2] = src[i]; } else { _Tp alpha = ColorChannel<_Tp>::max(); for( int i = 0; i < n; i++, dst += 4 ) { dst[0] = dst[1] = dst[2] = src[i]; dst[3] = alpha; } } } int dstcn; }; struct Gray2RGB5x5 { typedef uchar channel_type; Gray2RGB5x5(int _greenBits) : greenBits(_greenBits) { #if CV_NEON v_n7 = vdup_n_u8(~7); v_n3 = vdup_n_u8(~3); #elif CV_SSE2 haveSIMD = checkHardwareSupport(CV_CPU_SSE2); v_n7 = _mm_set1_epi16(~7); v_n3 = _mm_set1_epi16(~3); v_zero = _mm_setzero_si128(); #endif } void operator()(const uchar* src, uchar* dst, int n) const { int i = 0; if( greenBits == 6 ) { #if CV_NEON for ( ; i <= n - 8; i += 8 ) { uint8x8_t v_src = vld1_u8(src + i); uint16x8_t v_dst = vmovl_u8(vshr_n_u8(v_src, 3)); v_dst = vorrq_u16(v_dst, vshlq_n_u16(vmovl_u8(vand_u8(v_src, v_n3)), 3)); v_dst = vorrq_u16(v_dst, vshlq_n_u16(vmovl_u8(vand_u8(v_src, v_n7)), 8)); vst1q_u16((ushort *)dst + i, v_dst); } #elif CV_SSE2 if (haveSIMD) { for ( ; i <= n - 16; i += 16 ) { __m128i v_src = _mm_loadu_si128((__m128i const *)(src + i)); __m128i v_src_p = _mm_unpacklo_epi8(v_src, v_zero); __m128i v_dst = _mm_or_si128(_mm_srli_epi16(v_src_p, 3), _mm_or_si128(_mm_slli_epi16(_mm_and_si128(v_src_p, v_n3), 3), _mm_slli_epi16(_mm_and_si128(v_src_p, v_n7), 8))); _mm_storeu_si128((__m128i *)((ushort *)dst + i), v_dst); v_src_p = _mm_unpackhi_epi8(v_src, v_zero); v_dst = _mm_or_si128(_mm_srli_epi16(v_src_p, 3), _mm_or_si128(_mm_slli_epi16(_mm_and_si128(v_src_p, v_n3), 3), _mm_slli_epi16(_mm_and_si128(v_src_p, v_n7), 8))); _mm_storeu_si128((__m128i *)((ushort *)dst + i + 8), v_dst); } } #endif for ( ; i < n; i++ ) { int t = src[i]; ((ushort*)dst)[i] = (ushort)((t >> 3)|((t & ~3) << 3)|((t & ~7) << 8)); } } else { #if CV_NEON for ( ; i <= n - 8; i += 8 ) { uint16x8_t v_src = vmovl_u8(vshr_n_u8(vld1_u8(src + i), 3)); uint16x8_t v_dst = vorrq_u16(vorrq_u16(v_src, vshlq_n_u16(v_src, 5)), vshlq_n_u16(v_src, 10)); vst1q_u16((ushort *)dst + i, v_dst); } #elif CV_SSE2 if (haveSIMD) { for ( ; i <= n - 16; i += 8 ) { __m128i v_src = _mm_loadu_si128((__m128i const *)(src + i)); __m128i v_src_p = _mm_srli_epi16(_mm_unpacklo_epi8(v_src, v_zero), 3); __m128i v_dst = _mm_or_si128(v_src_p, _mm_or_si128(_mm_slli_epi32(v_src_p, 5), _mm_slli_epi16(v_src_p, 10))); _mm_storeu_si128((__m128i *)((ushort *)dst + i), v_dst); v_src_p = _mm_srli_epi16(_mm_unpackhi_epi8(v_src, v_zero), 3); v_dst = _mm_or_si128(v_src_p, _mm_or_si128(_mm_slli_epi16(v_src_p, 5), _mm_slli_epi16(v_src_p, 10))); _mm_storeu_si128((__m128i *)((ushort *)dst + i + 8), v_dst); } } #endif for( ; i < n; i++ ) { int t = src[i] >> 3; ((ushort*)dst)[i] = (ushort)(t|(t << 5)|(t << 10)); } } } int greenBits; #if CV_NEON uint8x8_t v_n7, v_n3; #elif CV_SSE2 __m128i v_n7, v_n3, v_zero; bool haveSIMD; #endif }; #undef R2Y #undef G2Y #undef B2Y enum { yuv_shift = 14, xyz_shift = 12, R2Y = 4899, G2Y = 9617, B2Y = 1868, BLOCK_SIZE = 256 }; struct RGB5x52Gray { typedef uchar channel_type; RGB5x52Gray(int _greenBits) : greenBits(_greenBits) { #if CV_NEON v_b2y = vdup_n_u16(B2Y); v_g2y = vdup_n_u16(G2Y); v_r2y = vdup_n_u16(R2Y); v_delta = vdupq_n_u32(1 << (yuv_shift - 1)); v_f8 = vdupq_n_u16(0xf8); v_fc = vdupq_n_u16(0xfc); #elif CV_SSE2 haveSIMD = checkHardwareSupport(CV_CPU_SSE2); v_b2y = _mm_set1_epi16(B2Y); v_g2y = _mm_set1_epi16(G2Y); v_r2y = _mm_set1_epi16(R2Y); v_delta = _mm_set1_epi32(1 << (yuv_shift - 1)); v_f8 = _mm_set1_epi16(0xf8); v_fc = _mm_set1_epi16(0xfc); #endif } void operator()(const uchar* src, uchar* dst, int n) const { int i = 0; if( greenBits == 6 ) { #if CV_NEON for ( ; i <= n - 8; i += 8) { uint16x8_t v_src = vld1q_u16((ushort *)src + i); uint16x8_t v_t0 = vandq_u16(vshlq_n_u16(v_src, 3), v_f8), v_t1 = vandq_u16(vshrq_n_u16(v_src, 3), v_fc), v_t2 = vandq_u16(vshrq_n_u16(v_src, 8), v_f8); uint32x4_t v_dst0 = vmlal_u16(vmlal_u16(vmull_u16(vget_low_u16(v_t0), v_b2y), vget_low_u16(v_t1), v_g2y), vget_low_u16(v_t2), v_r2y); uint32x4_t v_dst1 = vmlal_u16(vmlal_u16(vmull_u16(vget_high_u16(v_t0), v_b2y), vget_high_u16(v_t1), v_g2y), vget_high_u16(v_t2), v_r2y); v_dst0 = vshrq_n_u32(vaddq_u32(v_dst0, v_delta), yuv_shift); v_dst1 = vshrq_n_u32(vaddq_u32(v_dst1, v_delta), yuv_shift); vst1_u8(dst + i, vmovn_u16(vcombine_u16(vmovn_u32(v_dst0), vmovn_u32(v_dst1)))); } #elif CV_SSE2 if (haveSIMD) { __m128i v_zero = _mm_setzero_si128(); for ( ; i <= n - 8; i += 8) { __m128i v_src = _mm_loadu_si128((__m128i const *)((ushort *)src + i)); __m128i v_t0 = _mm_and_si128(_mm_slli_epi16(v_src, 3), v_f8), v_t1 = _mm_and_si128(_mm_srli_epi16(v_src, 3), v_fc), v_t2 = _mm_and_si128(_mm_srli_epi16(v_src, 8), v_f8); __m128i v_mullo_b = _mm_mullo_epi16(v_t0, v_b2y); __m128i v_mullo_g = _mm_mullo_epi16(v_t1, v_g2y); __m128i v_mullo_r = _mm_mullo_epi16(v_t2, v_r2y); __m128i v_mulhi_b = _mm_mulhi_epi16(v_t0, v_b2y); __m128i v_mulhi_g = _mm_mulhi_epi16(v_t1, v_g2y); __m128i v_mulhi_r = _mm_mulhi_epi16(v_t2, v_r2y); __m128i v_dst0 = _mm_add_epi32(_mm_unpacklo_epi16(v_mullo_b, v_mulhi_b), _mm_unpacklo_epi16(v_mullo_g, v_mulhi_g)); v_dst0 = _mm_add_epi32(_mm_add_epi32(v_dst0, v_delta), _mm_unpacklo_epi16(v_mullo_r, v_mulhi_r)); __m128i v_dst1 = _mm_add_epi32(_mm_unpackhi_epi16(v_mullo_b, v_mulhi_b), _mm_unpackhi_epi16(v_mullo_g, v_mulhi_g)); v_dst1 = _mm_add_epi32(_mm_add_epi32(v_dst1, v_delta), _mm_unpackhi_epi16(v_mullo_r, v_mulhi_r)); v_dst0 = _mm_srli_epi32(v_dst0, yuv_shift); v_dst1 = _mm_srli_epi32(v_dst1, yuv_shift); __m128i v_dst = _mm_packs_epi32(v_dst0, v_dst1); _mm_storel_epi64((__m128i *)(dst + i), _mm_packus_epi16(v_dst, v_zero)); } } #endif for ( ; i < n; i++) { int t = ((ushort*)src)[i]; dst[i] = (uchar)CV_DESCALE(((t << 3) & 0xf8)*B2Y + ((t >> 3) & 0xfc)*G2Y + ((t >> 8) & 0xf8)*R2Y, yuv_shift); } } else { #if CV_NEON for ( ; i <= n - 8; i += 8) { uint16x8_t v_src = vld1q_u16((ushort *)src + i); uint16x8_t v_t0 = vandq_u16(vshlq_n_u16(v_src, 3), v_f8), v_t1 = vandq_u16(vshrq_n_u16(v_src, 2), v_f8), v_t2 = vandq_u16(vshrq_n_u16(v_src, 7), v_f8); uint32x4_t v_dst0 = vmlal_u16(vmlal_u16(vmull_u16(vget_low_u16(v_t0), v_b2y), vget_low_u16(v_t1), v_g2y), vget_low_u16(v_t2), v_r2y); uint32x4_t v_dst1 = vmlal_u16(vmlal_u16(vmull_u16(vget_high_u16(v_t0), v_b2y), vget_high_u16(v_t1), v_g2y), vget_high_u16(v_t2), v_r2y); v_dst0 = vshrq_n_u32(vaddq_u32(v_dst0, v_delta), yuv_shift); v_dst1 = vshrq_n_u32(vaddq_u32(v_dst1, v_delta), yuv_shift); vst1_u8(dst + i, vmovn_u16(vcombine_u16(vmovn_u32(v_dst0), vmovn_u32(v_dst1)))); } #elif CV_SSE2 if (haveSIMD) { __m128i v_zero = _mm_setzero_si128(); for ( ; i <= n - 8; i += 8) { __m128i v_src = _mm_loadu_si128((__m128i const *)((ushort *)src + i)); __m128i v_t0 = _mm_and_si128(_mm_slli_epi16(v_src, 3), v_f8), v_t1 = _mm_and_si128(_mm_srli_epi16(v_src, 2), v_f8), v_t2 = _mm_and_si128(_mm_srli_epi16(v_src, 7), v_f8); __m128i v_mullo_b = _mm_mullo_epi16(v_t0, v_b2y); __m128i v_mullo_g = _mm_mullo_epi16(v_t1, v_g2y); __m128i v_mullo_r = _mm_mullo_epi16(v_t2, v_r2y); __m128i v_mulhi_b = _mm_mulhi_epi16(v_t0, v_b2y); __m128i v_mulhi_g = _mm_mulhi_epi16(v_t1, v_g2y); __m128i v_mulhi_r = _mm_mulhi_epi16(v_t2, v_r2y); __m128i v_dst0 = _mm_add_epi32(_mm_unpacklo_epi16(v_mullo_b, v_mulhi_b), _mm_unpacklo_epi16(v_mullo_g, v_mulhi_g)); v_dst0 = _mm_add_epi32(_mm_add_epi32(v_dst0, v_delta), _mm_unpacklo_epi16(v_mullo_r, v_mulhi_r)); __m128i v_dst1 = _mm_add_epi32(_mm_unpackhi_epi16(v_mullo_b, v_mulhi_b), _mm_unpackhi_epi16(v_mullo_g, v_mulhi_g)); v_dst1 = _mm_add_epi32(_mm_add_epi32(v_dst1, v_delta), _mm_unpackhi_epi16(v_mullo_r, v_mulhi_r)); v_dst0 = _mm_srli_epi32(v_dst0, yuv_shift); v_dst1 = _mm_srli_epi32(v_dst1, yuv_shift); __m128i v_dst = _mm_packs_epi32(v_dst0, v_dst1); _mm_storel_epi64((__m128i *)(dst + i), _mm_packus_epi16(v_dst, v_zero)); } } #endif for ( ; i < n; i++) { int t = ((ushort*)src)[i]; dst[i] = (uchar)CV_DESCALE(((t << 3) & 0xf8)*B2Y + ((t >> 2) & 0xf8)*G2Y + ((t >> 7) & 0xf8)*R2Y, yuv_shift); } } } int greenBits; #if CV_NEON uint16x4_t v_b2y, v_g2y, v_r2y; uint32x4_t v_delta; uint16x8_t v_f8, v_fc; #elif CV_SSE2 bool haveSIMD; __m128i v_b2y, v_g2y, v_r2y; __m128i v_delta; __m128i v_f8, v_fc; #endif }; template struct RGB2Gray { typedef _Tp channel_type; RGB2Gray(int _srccn, int blueIdx, const float* _coeffs) : srccn(_srccn) { static const float coeffs0[] = { 0.299f, 0.587f, 0.114f }; memcpy( coeffs, _coeffs ? _coeffs : coeffs0, 3*sizeof(coeffs[0]) ); if(blueIdx == 0) std::swap(coeffs[0], coeffs[2]); } void operator()(const _Tp* src, _Tp* dst, int n) const { int scn = srccn; float cb = coeffs[0], cg = coeffs[1], cr = coeffs[2]; for(int i = 0; i < n; i++, src += scn) dst[i] = saturate_cast<_Tp>(src[0]*cb + src[1]*cg + src[2]*cr); } int srccn; float coeffs[3]; }; template<> struct RGB2Gray { typedef uchar channel_type; RGB2Gray(int _srccn, int blueIdx, const int* coeffs) : srccn(_srccn) { const int coeffs0[] = { R2Y, G2Y, B2Y }; if(!coeffs) coeffs = coeffs0; int b = 0, g = 0, r = (1 << (yuv_shift-1)); int db = coeffs[blueIdx^2], dg = coeffs[1], dr = coeffs[blueIdx]; for( int i = 0; i < 256; i++, b += db, g += dg, r += dr ) { tab[i] = b; tab[i+256] = g; tab[i+512] = r; } } void operator()(const uchar* src, uchar* dst, int n) const { int scn = srccn; const int* _tab = tab; for(int i = 0; i < n; i++, src += scn) dst[i] = (uchar)((_tab[src[0]] + _tab[src[1]+256] + _tab[src[2]+512]) >> yuv_shift); } int srccn; int tab[256*3]; }; #if CV_NEON template <> struct RGB2Gray { typedef ushort channel_type; RGB2Gray(int _srccn, int blueIdx, const int* _coeffs) : srccn(_srccn) { static const int coeffs0[] = { R2Y, G2Y, B2Y }; memcpy(coeffs, _coeffs ? _coeffs : coeffs0, 3*sizeof(coeffs[0])); if( blueIdx == 0 ) std::swap(coeffs[0], coeffs[2]); v_cb = vdup_n_u16(coeffs[0]); v_cg = vdup_n_u16(coeffs[1]); v_cr = vdup_n_u16(coeffs[2]); v_delta = vdupq_n_u32(1 << (yuv_shift - 1)); } void operator()(const ushort* src, ushort* dst, int n) const { int scn = srccn, cb = coeffs[0], cg = coeffs[1], cr = coeffs[2], i = 0; for ( ; i <= n - 8; i += 8, src += scn * 8) { uint16x8_t v_b, v_r, v_g; if (scn == 3) { uint16x8x3_t v_src = vld3q_u16(src); v_b = v_src.val[0]; v_g = v_src.val[1]; v_r = v_src.val[2]; } else { uint16x8x4_t v_src = vld4q_u16(src); v_b = v_src.val[0]; v_g = v_src.val[1]; v_r = v_src.val[2]; } uint32x4_t v_dst0_ = vmlal_u16(vmlal_u16( vmull_u16(vget_low_u16(v_b), v_cb), vget_low_u16(v_g), v_cg), vget_low_u16(v_r), v_cr); uint32x4_t v_dst1_ = vmlal_u16(vmlal_u16( vmull_u16(vget_high_u16(v_b), v_cb), vget_high_u16(v_g), v_cg), vget_high_u16(v_r), v_cr); uint16x4_t v_dst0 = vmovn_u32(vshrq_n_u32(vaddq_u32(v_dst0_, v_delta), yuv_shift)); uint16x4_t v_dst1 = vmovn_u32(vshrq_n_u32(vaddq_u32(v_dst1_, v_delta), yuv_shift)); vst1q_u16(dst + i, vcombine_u16(v_dst0, v_dst1)); } for ( ; i <= n - 4; i += 4, src += scn * 4) { uint16x4_t v_b, v_r, v_g; if (scn == 3) { uint16x4x3_t v_src = vld3_u16(src); v_b = v_src.val[0]; v_g = v_src.val[1]; v_r = v_src.val[2]; } else { uint16x4x4_t v_src = vld4_u16(src); v_b = v_src.val[0]; v_g = v_src.val[1]; v_r = v_src.val[2]; } uint32x4_t v_dst = vmlal_u16(vmlal_u16( vmull_u16(v_b, v_cb), v_g, v_cg), v_r, v_cr); vst1_u16(dst + i, vmovn_u32(vshrq_n_u32(vaddq_u32(v_dst, v_delta), yuv_shift))); } for( ; i < n; i++, src += scn) dst[i] = (ushort)CV_DESCALE((unsigned)(src[0]*cb + src[1]*cg + src[2]*cr), yuv_shift); } int srccn, coeffs[3]; uint16x4_t v_cb, v_cg, v_cr; uint32x4_t v_delta; }; template <> struct RGB2Gray { typedef float channel_type; RGB2Gray(int _srccn, int blueIdx, const float* _coeffs) : srccn(_srccn) { static const float coeffs0[] = { 0.299f, 0.587f, 0.114f }; memcpy( coeffs, _coeffs ? _coeffs : coeffs0, 3*sizeof(coeffs[0]) ); if(blueIdx == 0) std::swap(coeffs[0], coeffs[2]); v_cb = vdupq_n_f32(coeffs[0]); v_cg = vdupq_n_f32(coeffs[1]); v_cr = vdupq_n_f32(coeffs[2]); } void operator()(const float * src, float * dst, int n) const { int scn = srccn, i = 0; float cb = coeffs[0], cg = coeffs[1], cr = coeffs[2]; if (scn == 3) { for ( ; i <= n - 8; i += 8, src += scn * 8) { float32x4x3_t v_src = vld3q_f32(src); vst1q_f32(dst + i, vmlaq_f32(vmlaq_f32(vmulq_f32(v_src.val[0], v_cb), v_src.val[1], v_cg), v_src.val[2], v_cr)); v_src = vld3q_f32(src + scn * 4); vst1q_f32(dst + i + 4, vmlaq_f32(vmlaq_f32(vmulq_f32(v_src.val[0], v_cb), v_src.val[1], v_cg), v_src.val[2], v_cr)); } for ( ; i <= n - 4; i += 4, src += scn * 4) { float32x4x3_t v_src = vld3q_f32(src); vst1q_f32(dst + i, vmlaq_f32(vmlaq_f32(vmulq_f32(v_src.val[0], v_cb), v_src.val[1], v_cg), v_src.val[2], v_cr)); } } else { for ( ; i <= n - 8; i += 8, src += scn * 8) { float32x4x4_t v_src = vld4q_f32(src); vst1q_f32(dst + i, vmlaq_f32(vmlaq_f32(vmulq_f32(v_src.val[0], v_cb), v_src.val[1], v_cg), v_src.val[2], v_cr)); v_src = vld4q_f32(src + scn * 4); vst1q_f32(dst + i + 4, vmlaq_f32(vmlaq_f32(vmulq_f32(v_src.val[0], v_cb), v_src.val[1], v_cg), v_src.val[2], v_cr)); } for ( ; i <= n - 4; i += 4, src += scn * 4) { float32x4x4_t v_src = vld4q_f32(src); vst1q_f32(dst + i, vmlaq_f32(vmlaq_f32(vmulq_f32(v_src.val[0], v_cb), v_src.val[1], v_cg), v_src.val[2], v_cr)); } } for ( ; i < n; i++, src += scn) dst[i] = src[0]*cb + src[1]*cg + src[2]*cr; } int srccn; float coeffs[3]; float32x4_t v_cb, v_cg, v_cr; }; #elif CV_SSE2 #if CV_SSE4_1 template <> struct RGB2Gray { typedef ushort channel_type; RGB2Gray(int _srccn, int blueIdx, const int* _coeffs) : srccn(_srccn) { static const int coeffs0[] = { R2Y, G2Y, B2Y }; memcpy(coeffs, _coeffs ? _coeffs : coeffs0, 3*sizeof(coeffs[0])); if( blueIdx == 0 ) std::swap(coeffs[0], coeffs[2]); v_cb = _mm_set1_epi16((short)coeffs[0]); v_cg = _mm_set1_epi16((short)coeffs[1]); v_cr = _mm_set1_epi16((short)coeffs[2]); v_delta = _mm_set1_epi32(1 << (yuv_shift - 1)); haveSIMD = checkHardwareSupport(CV_CPU_SSE4_1); } // 16s x 8 void process(__m128i v_b, __m128i v_g, __m128i v_r, __m128i & v_gray) const { __m128i v_mullo_r = _mm_mullo_epi16(v_r, v_cr); __m128i v_mullo_g = _mm_mullo_epi16(v_g, v_cg); __m128i v_mullo_b = _mm_mullo_epi16(v_b, v_cb); __m128i v_mulhi_r = _mm_mulhi_epu16(v_r, v_cr); __m128i v_mulhi_g = _mm_mulhi_epu16(v_g, v_cg); __m128i v_mulhi_b = _mm_mulhi_epu16(v_b, v_cb); __m128i v_gray0 = _mm_add_epi32(_mm_unpacklo_epi16(v_mullo_r, v_mulhi_r), _mm_unpacklo_epi16(v_mullo_g, v_mulhi_g)); v_gray0 = _mm_add_epi32(_mm_unpacklo_epi16(v_mullo_b, v_mulhi_b), v_gray0); v_gray0 = _mm_srli_epi32(_mm_add_epi32(v_gray0, v_delta), yuv_shift); __m128i v_gray1 = _mm_add_epi32(_mm_unpackhi_epi16(v_mullo_r, v_mulhi_r), _mm_unpackhi_epi16(v_mullo_g, v_mulhi_g)); v_gray1 = _mm_add_epi32(_mm_unpackhi_epi16(v_mullo_b, v_mulhi_b), v_gray1); v_gray1 = _mm_srli_epi32(_mm_add_epi32(v_gray1, v_delta), yuv_shift); v_gray = _mm_packus_epi32(v_gray0, v_gray1); } void operator()(const ushort* src, ushort* dst, int n) const { int scn = srccn, cb = coeffs[0], cg = coeffs[1], cr = coeffs[2], i = 0; if (scn == 3 && haveSIMD) { for ( ; i <= n - 16; i += 16, src += scn * 16) { __m128i v_r0 = _mm_loadu_si128((__m128i const *)(src)); __m128i v_r1 = _mm_loadu_si128((__m128i const *)(src + 8)); __m128i v_g0 = _mm_loadu_si128((__m128i const *)(src + 16)); __m128i v_g1 = _mm_loadu_si128((__m128i const *)(src + 24)); __m128i v_b0 = _mm_loadu_si128((__m128i const *)(src + 32)); __m128i v_b1 = _mm_loadu_si128((__m128i const *)(src + 40)); _mm_deinterleave_epi16(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1); __m128i v_gray0; process(v_r0, v_g0, v_b0, v_gray0); __m128i v_gray1; process(v_r1, v_g1, v_b1, v_gray1); _mm_storeu_si128((__m128i *)(dst + i), v_gray0); _mm_storeu_si128((__m128i *)(dst + i + 8), v_gray1); } } else if (scn == 4 && haveSIMD) { for ( ; i <= n - 16; i += 16, src += scn * 16) { __m128i v_r0 = _mm_loadu_si128((__m128i const *)(src)); __m128i v_r1 = _mm_loadu_si128((__m128i const *)(src + 8)); __m128i v_g0 = _mm_loadu_si128((__m128i const *)(src + 16)); __m128i v_g1 = _mm_loadu_si128((__m128i const *)(src + 24)); __m128i v_b0 = _mm_loadu_si128((__m128i const *)(src + 32)); __m128i v_b1 = _mm_loadu_si128((__m128i const *)(src + 40)); __m128i v_a0 = _mm_loadu_si128((__m128i const *)(src + 48)); __m128i v_a1 = _mm_loadu_si128((__m128i const *)(src + 56)); _mm_deinterleave_epi16(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1, v_a0, v_a1); __m128i v_gray0; process(v_r0, v_g0, v_b0, v_gray0); __m128i v_gray1; process(v_r1, v_g1, v_b1, v_gray1); _mm_storeu_si128((__m128i *)(dst + i), v_gray0); _mm_storeu_si128((__m128i *)(dst + i + 8), v_gray1); } } for( ; i < n; i++, src += scn) dst[i] = (ushort)CV_DESCALE((unsigned)(src[0]*cb + src[1]*cg + src[2]*cr), yuv_shift); } int srccn, coeffs[3]; __m128i v_cb, v_cg, v_cr; __m128i v_delta; bool haveSIMD; }; #endif // CV_SSE4_1 template <> struct RGB2Gray { typedef float channel_type; RGB2Gray(int _srccn, int blueIdx, const float* _coeffs) : srccn(_srccn) { static const float coeffs0[] = { 0.299f, 0.587f, 0.114f }; memcpy( coeffs, _coeffs ? _coeffs : coeffs0, 3*sizeof(coeffs[0]) ); if(blueIdx == 0) std::swap(coeffs[0], coeffs[2]); v_cb = _mm_set1_ps(coeffs[0]); v_cg = _mm_set1_ps(coeffs[1]); v_cr = _mm_set1_ps(coeffs[2]); haveSIMD = checkHardwareSupport(CV_CPU_SSE2); } void process(__m128 v_b, __m128 v_g, __m128 v_r, __m128 & v_gray) const { v_gray = _mm_mul_ps(v_r, v_cr); v_gray = _mm_add_ps(v_gray, _mm_mul_ps(v_g, v_cg)); v_gray = _mm_add_ps(v_gray, _mm_mul_ps(v_b, v_cb)); } void operator()(const float * src, float * dst, int n) const { int scn = srccn, i = 0; float cb = coeffs[0], cg = coeffs[1], cr = coeffs[2]; if (scn == 3 && haveSIMD) { for ( ; i <= n - 8; i += 8, src += scn * 8) { __m128 v_r0 = _mm_loadu_ps(src); __m128 v_r1 = _mm_loadu_ps(src + 4); __m128 v_g0 = _mm_loadu_ps(src + 8); __m128 v_g1 = _mm_loadu_ps(src + 12); __m128 v_b0 = _mm_loadu_ps(src + 16); __m128 v_b1 = _mm_loadu_ps(src + 20); _mm_deinterleave_ps(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1); __m128 v_gray0; process(v_r0, v_g0, v_b0, v_gray0); __m128 v_gray1; process(v_r1, v_g1, v_b1, v_gray1); _mm_storeu_ps(dst + i, v_gray0); _mm_storeu_ps(dst + i + 4, v_gray1); } } else if (scn == 4 && haveSIMD) { for ( ; i <= n - 8; i += 8, src += scn * 8) { __m128 v_r0 = _mm_loadu_ps(src); __m128 v_r1 = _mm_loadu_ps(src + 4); __m128 v_g0 = _mm_loadu_ps(src + 8); __m128 v_g1 = _mm_loadu_ps(src + 12); __m128 v_b0 = _mm_loadu_ps(src + 16); __m128 v_b1 = _mm_loadu_ps(src + 20); __m128 v_a0 = _mm_loadu_ps(src + 24); __m128 v_a1 = _mm_loadu_ps(src + 28); _mm_deinterleave_ps(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1, v_a0, v_a1); __m128 v_gray0; process(v_r0, v_g0, v_b0, v_gray0); __m128 v_gray1; process(v_r1, v_g1, v_b1, v_gray1); _mm_storeu_ps(dst + i, v_gray0); _mm_storeu_ps(dst + i + 4, v_gray1); } } for ( ; i < n; i++, src += scn) dst[i] = src[0]*cb + src[1]*cg + src[2]*cr; } int srccn; float coeffs[3]; __m128 v_cb, v_cg, v_cr; bool haveSIMD; }; #else template<> struct RGB2Gray { typedef ushort channel_type; RGB2Gray(int _srccn, int blueIdx, const int* _coeffs) : srccn(_srccn) { static const int coeffs0[] = { R2Y, G2Y, B2Y }; memcpy(coeffs, _coeffs ? _coeffs : coeffs0, 3*sizeof(coeffs[0])); if( blueIdx == 0 ) std::swap(coeffs[0], coeffs[2]); } void operator()(const ushort* src, ushort* dst, int n) const { int scn = srccn, cb = coeffs[0], cg = coeffs[1], cr = coeffs[2]; for(int i = 0; i < n; i++, src += scn) dst[i] = (ushort)CV_DESCALE((unsigned)(src[0]*cb + src[1]*cg + src[2]*cr), yuv_shift); } int srccn; int coeffs[3]; }; #endif ///////////////////////////////////// RGB <-> YCrCb ////////////////////////////////////// template struct RGB2YCrCb_f { typedef _Tp channel_type; RGB2YCrCb_f(int _srccn, int _blueIdx, const float* _coeffs) : srccn(_srccn), blueIdx(_blueIdx) { static const float coeffs0[] = {0.299f, 0.587f, 0.114f, 0.713f, 0.564f}; memcpy(coeffs, _coeffs ? _coeffs : coeffs0, 5*sizeof(coeffs[0])); if(blueIdx==0) std::swap(coeffs[0], coeffs[2]); } void operator()(const _Tp* src, _Tp* dst, int n) const { int scn = srccn, bidx = blueIdx; const _Tp delta = ColorChannel<_Tp>::half(); float C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4]; n *= 3; for(int i = 0; i < n; i += 3, src += scn) { _Tp Y = saturate_cast<_Tp>(src[0]*C0 + src[1]*C1 + src[2]*C2); _Tp Cr = saturate_cast<_Tp>((src[bidx^2] - Y)*C3 + delta); _Tp Cb = saturate_cast<_Tp>((src[bidx] - Y)*C4 + delta); dst[i] = Y; dst[i+1] = Cr; dst[i+2] = Cb; } } int srccn, blueIdx; float coeffs[5]; }; #if CV_NEON template <> struct RGB2YCrCb_f { typedef float channel_type; RGB2YCrCb_f(int _srccn, int _blueIdx, const float* _coeffs) : srccn(_srccn), blueIdx(_blueIdx) { static const float coeffs0[] = {0.299f, 0.587f, 0.114f, 0.713f, 0.564f}; memcpy(coeffs, _coeffs ? _coeffs : coeffs0, 5*sizeof(coeffs[0])); if(blueIdx==0) std::swap(coeffs[0], coeffs[2]); v_c0 = vdupq_n_f32(coeffs[0]); v_c1 = vdupq_n_f32(coeffs[1]); v_c2 = vdupq_n_f32(coeffs[2]); v_c3 = vdupq_n_f32(coeffs[3]); v_c4 = vdupq_n_f32(coeffs[4]); v_delta = vdupq_n_f32(ColorChannel::half()); } void operator()(const float * src, float * dst, int n) const { int scn = srccn, bidx = blueIdx, i = 0; const float delta = ColorChannel::half(); float C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4]; n *= 3; if (scn == 3) for ( ; i <= n - 12; i += 12, src += 12) { float32x4x3_t v_src = vld3q_f32(src), v_dst; v_dst.val[0] = vmlaq_f32(vmlaq_f32(vmulq_f32(v_src.val[0], v_c0), v_src.val[1], v_c1), v_src.val[2], v_c2); v_dst.val[1] = vmlaq_f32(v_delta, vsubq_f32(v_src.val[bidx^2], v_dst.val[0]), v_c3); v_dst.val[2] = vmlaq_f32(v_delta, vsubq_f32(v_src.val[bidx], v_dst.val[0]), v_c4); vst3q_f32(dst + i, v_dst); } else for ( ; i <= n - 12; i += 12, src += 16) { float32x4x4_t v_src = vld4q_f32(src); float32x4x3_t v_dst; v_dst.val[0] = vmlaq_f32(vmlaq_f32(vmulq_f32(v_src.val[0], v_c0), v_src.val[1], v_c1), v_src.val[2], v_c2); v_dst.val[1] = vmlaq_f32(v_delta, vsubq_f32(v_src.val[bidx^2], v_dst.val[0]), v_c3); v_dst.val[2] = vmlaq_f32(v_delta, vsubq_f32(v_src.val[bidx], v_dst.val[0]), v_c4); vst3q_f32(dst + i, v_dst); } for ( ; i < n; i += 3, src += scn) { float Y = src[0]*C0 + src[1]*C1 + src[2]*C2; float Cr = (src[bidx^2] - Y)*C3 + delta; float Cb = (src[bidx] - Y)*C4 + delta; dst[i] = Y; dst[i+1] = Cr; dst[i+2] = Cb; } } int srccn, blueIdx; float coeffs[5]; float32x4_t v_c0, v_c1, v_c2, v_c3, v_c4, v_delta; }; #elif CV_SSE2 template <> struct RGB2YCrCb_f { typedef float channel_type; RGB2YCrCb_f(int _srccn, int _blueIdx, const float* _coeffs) : srccn(_srccn), blueIdx(_blueIdx) { static const float coeffs0[] = {0.299f, 0.587f, 0.114f, 0.713f, 0.564f}; memcpy(coeffs, _coeffs ? _coeffs : coeffs0, 5*sizeof(coeffs[0])); if (blueIdx==0) std::swap(coeffs[0], coeffs[2]); v_c0 = _mm_set1_ps(coeffs[0]); v_c1 = _mm_set1_ps(coeffs[1]); v_c2 = _mm_set1_ps(coeffs[2]); v_c3 = _mm_set1_ps(coeffs[3]); v_c4 = _mm_set1_ps(coeffs[4]); v_delta = _mm_set1_ps(ColorChannel::half()); haveSIMD = checkHardwareSupport(CV_CPU_SSE2); } void process(__m128 v_r, __m128 v_g, __m128 v_b, __m128 & v_y, __m128 & v_cr, __m128 & v_cb) const { v_y = _mm_mul_ps(v_r, v_c0); v_y = _mm_add_ps(v_y, _mm_mul_ps(v_g, v_c1)); v_y = _mm_add_ps(v_y, _mm_mul_ps(v_b, v_c2)); v_cr = _mm_add_ps(_mm_mul_ps(_mm_sub_ps(blueIdx == 0 ? v_b : v_r, v_y), v_c3), v_delta); v_cb = _mm_add_ps(_mm_mul_ps(_mm_sub_ps(blueIdx == 2 ? v_b : v_r, v_y), v_c4), v_delta); } void operator()(const float * src, float * dst, int n) const { int scn = srccn, bidx = blueIdx, i = 0; const float delta = ColorChannel::half(); float C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4]; n *= 3; if (haveSIMD) { for ( ; i <= n - 24; i += 24, src += 8 * scn) { __m128 v_r0 = _mm_loadu_ps(src); __m128 v_r1 = _mm_loadu_ps(src + 4); __m128 v_g0 = _mm_loadu_ps(src + 8); __m128 v_g1 = _mm_loadu_ps(src + 12); __m128 v_b0 = _mm_loadu_ps(src + 16); __m128 v_b1 = _mm_loadu_ps(src + 20); if (scn == 4) { __m128 v_a0 = _mm_loadu_ps(src + 24); __m128 v_a1 = _mm_loadu_ps(src + 28); _mm_deinterleave_ps(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1, v_a0, v_a1); } else _mm_deinterleave_ps(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1); __m128 v_y0, v_cr0, v_cb0; process(v_r0, v_g0, v_b0, v_y0, v_cr0, v_cb0); __m128 v_y1, v_cr1, v_cb1; process(v_r1, v_g1, v_b1, v_y1, v_cr1, v_cb1); _mm_interleave_ps(v_y0, v_y1, v_cr0, v_cr1, v_cb0, v_cb1); _mm_storeu_ps(dst + i, v_y0); _mm_storeu_ps(dst + i + 4, v_y1); _mm_storeu_ps(dst + i + 8, v_cr0); _mm_storeu_ps(dst + i + 12, v_cr1); _mm_storeu_ps(dst + i + 16, v_cb0); _mm_storeu_ps(dst + i + 20, v_cb1); } } for ( ; i < n; i += 3, src += scn) { float Y = src[0]*C0 + src[1]*C1 + src[2]*C2; float Cr = (src[bidx^2] - Y)*C3 + delta; float Cb = (src[bidx] - Y)*C4 + delta; dst[i] = Y; dst[i+1] = Cr; dst[i+2] = Cb; } } int srccn, blueIdx; float coeffs[5]; __m128 v_c0, v_c1, v_c2, v_c3, v_c4, v_delta; bool haveSIMD; }; #endif template struct RGB2YCrCb_i { typedef _Tp channel_type; RGB2YCrCb_i(int _srccn, int _blueIdx, const int* _coeffs) : srccn(_srccn), blueIdx(_blueIdx) { static const int coeffs0[] = {R2Y, G2Y, B2Y, 11682, 9241}; memcpy(coeffs, _coeffs ? _coeffs : coeffs0, 5*sizeof(coeffs[0])); if(blueIdx==0) std::swap(coeffs[0], coeffs[2]); } void operator()(const _Tp* src, _Tp* dst, int n) const { int scn = srccn, bidx = blueIdx; int C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4]; int delta = ColorChannel<_Tp>::half()*(1 << yuv_shift); n *= 3; for(int i = 0; i < n; i += 3, src += scn) { int Y = CV_DESCALE(src[0]*C0 + src[1]*C1 + src[2]*C2, yuv_shift); int Cr = CV_DESCALE((src[bidx^2] - Y)*C3 + delta, yuv_shift); int Cb = CV_DESCALE((src[bidx] - Y)*C4 + delta, yuv_shift); dst[i] = saturate_cast<_Tp>(Y); dst[i+1] = saturate_cast<_Tp>(Cr); dst[i+2] = saturate_cast<_Tp>(Cb); } } int srccn, blueIdx; int coeffs[5]; }; #if CV_NEON template <> struct RGB2YCrCb_i { typedef uchar channel_type; RGB2YCrCb_i(int _srccn, int _blueIdx, const int* _coeffs) : srccn(_srccn), blueIdx(_blueIdx) { static const int coeffs0[] = {R2Y, G2Y, B2Y, 11682, 9241}; memcpy(coeffs, _coeffs ? _coeffs : coeffs0, 5*sizeof(coeffs[0])); if (blueIdx==0) std::swap(coeffs[0], coeffs[2]); v_c0 = vdup_n_s16(coeffs[0]); v_c1 = vdup_n_s16(coeffs[1]); v_c2 = vdup_n_s16(coeffs[2]); v_c3 = vdupq_n_s32(coeffs[3]); v_c4 = vdupq_n_s32(coeffs[4]); v_delta = vdupq_n_s32(ColorChannel::half()*(1 << yuv_shift)); v_delta2 = vdupq_n_s32(1 << (yuv_shift - 1)); } void operator()(const uchar * src, uchar * dst, int n) const { int scn = srccn, bidx = blueIdx, i = 0; int C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4]; int delta = ColorChannel::half()*(1 << yuv_shift); n *= 3; for ( ; i <= n - 24; i += 24, src += scn * 8) { uint8x8x3_t v_dst; int16x8x3_t v_src16; if (scn == 3) { uint8x8x3_t v_src = vld3_u8(src); v_src16.val[0] = vreinterpretq_s16_u16(vmovl_u8(v_src.val[0])); v_src16.val[1] = vreinterpretq_s16_u16(vmovl_u8(v_src.val[1])); v_src16.val[2] = vreinterpretq_s16_u16(vmovl_u8(v_src.val[2])); } else { uint8x8x4_t v_src = vld4_u8(src); v_src16.val[0] = vreinterpretq_s16_u16(vmovl_u8(v_src.val[0])); v_src16.val[1] = vreinterpretq_s16_u16(vmovl_u8(v_src.val[1])); v_src16.val[2] = vreinterpretq_s16_u16(vmovl_u8(v_src.val[2])); } int16x4x3_t v_src0; v_src0.val[0] = vget_low_s16(v_src16.val[0]); v_src0.val[1] = vget_low_s16(v_src16.val[1]); v_src0.val[2] = vget_low_s16(v_src16.val[2]); int32x4_t v_Y0 = vmlal_s16(vmlal_s16(vmull_s16(v_src0.val[0], v_c0), v_src0.val[1], v_c1), v_src0.val[2], v_c2); v_Y0 = vshrq_n_s32(vaddq_s32(v_Y0, v_delta2), yuv_shift); int32x4_t v_Cr0 = vmlaq_s32(v_delta, vsubq_s32(vmovl_s16(v_src0.val[bidx^2]), v_Y0), v_c3); v_Cr0 = vshrq_n_s32(vaddq_s32(v_Cr0, v_delta2), yuv_shift); int32x4_t v_Cb0 = vmlaq_s32(v_delta, vsubq_s32(vmovl_s16(v_src0.val[bidx]), v_Y0), v_c4); v_Cb0 = vshrq_n_s32(vaddq_s32(v_Cb0, v_delta2), yuv_shift); v_src0.val[0] = vget_high_s16(v_src16.val[0]); v_src0.val[1] = vget_high_s16(v_src16.val[1]); v_src0.val[2] = vget_high_s16(v_src16.val[2]); int32x4_t v_Y1 = vmlal_s16(vmlal_s16(vmull_s16(v_src0.val[0], v_c0), v_src0.val[1], v_c1), v_src0.val[2], v_c2); v_Y1 = vshrq_n_s32(vaddq_s32(v_Y1, v_delta2), yuv_shift); int32x4_t v_Cr1 = vmlaq_s32(v_delta, vsubq_s32(vmovl_s16(v_src0.val[bidx^2]), v_Y1), v_c3); v_Cr1 = vshrq_n_s32(vaddq_s32(v_Cr1, v_delta2), yuv_shift); int32x4_t v_Cb1 = vmlaq_s32(v_delta, vsubq_s32(vmovl_s16(v_src0.val[bidx]), v_Y1), v_c4); v_Cb1 = vshrq_n_s32(vaddq_s32(v_Cb1, v_delta2), yuv_shift); v_dst.val[0] = vqmovun_s16(vcombine_s16(vqmovn_s32(v_Y0), vqmovn_s32(v_Y1))); v_dst.val[1] = vqmovun_s16(vcombine_s16(vqmovn_s32(v_Cr0), vqmovn_s32(v_Cr1))); v_dst.val[2] = vqmovun_s16(vcombine_s16(vqmovn_s32(v_Cb0), vqmovn_s32(v_Cb1))); vst3_u8(dst + i, v_dst); } for ( ; i < n; i += 3, src += scn) { int Y = CV_DESCALE(src[0]*C0 + src[1]*C1 + src[2]*C2, yuv_shift); int Cr = CV_DESCALE((src[bidx^2] - Y)*C3 + delta, yuv_shift); int Cb = CV_DESCALE((src[bidx] - Y)*C4 + delta, yuv_shift); dst[i] = saturate_cast(Y); dst[i+1] = saturate_cast(Cr); dst[i+2] = saturate_cast(Cb); } } int srccn, blueIdx, coeffs[5]; int16x4_t v_c0, v_c1, v_c2; int32x4_t v_c3, v_c4, v_delta, v_delta2; }; template <> struct RGB2YCrCb_i { typedef ushort channel_type; RGB2YCrCb_i(int _srccn, int _blueIdx, const int* _coeffs) : srccn(_srccn), blueIdx(_blueIdx) { static const int coeffs0[] = {R2Y, G2Y, B2Y, 11682, 9241}; memcpy(coeffs, _coeffs ? _coeffs : coeffs0, 5*sizeof(coeffs[0])); if (blueIdx==0) std::swap(coeffs[0], coeffs[2]); v_c0 = vdupq_n_s32(coeffs[0]); v_c1 = vdupq_n_s32(coeffs[1]); v_c2 = vdupq_n_s32(coeffs[2]); v_c3 = vdupq_n_s32(coeffs[3]); v_c4 = vdupq_n_s32(coeffs[4]); v_delta = vdupq_n_s32(ColorChannel::half()*(1 << yuv_shift)); v_delta2 = vdupq_n_s32(1 << (yuv_shift - 1)); } void operator()(const ushort * src, ushort * dst, int n) const { int scn = srccn, bidx = blueIdx, i = 0; int C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4]; int delta = ColorChannel::half()*(1 << yuv_shift); n *= 3; for ( ; i <= n - 24; i += 24, src += scn * 8) { uint16x8x3_t v_src, v_dst; int32x4x3_t v_src0; if (scn == 3) v_src = vld3q_u16(src); else { uint16x8x4_t v_src_ = vld4q_u16(src); v_src.val[0] = v_src_.val[0]; v_src.val[1] = v_src_.val[1]; v_src.val[2] = v_src_.val[2]; } v_src0.val[0] = vreinterpretq_s32_u32(vmovl_u16(vget_low_u16(v_src.val[0]))); v_src0.val[1] = vreinterpretq_s32_u32(vmovl_u16(vget_low_u16(v_src.val[1]))); v_src0.val[2] = vreinterpretq_s32_u32(vmovl_u16(vget_low_u16(v_src.val[2]))); int32x4_t v_Y0 = vmlaq_s32(vmlaq_s32(vmulq_s32(v_src0.val[0], v_c0), v_src0.val[1], v_c1), v_src0.val[2], v_c2); v_Y0 = vshrq_n_s32(vaddq_s32(v_Y0, v_delta2), yuv_shift); int32x4_t v_Cr0 = vmlaq_s32(v_delta, vsubq_s32(v_src0.val[bidx^2], v_Y0), v_c3); v_Cr0 = vshrq_n_s32(vaddq_s32(v_Cr0, v_delta2), yuv_shift); int32x4_t v_Cb0 = vmlaq_s32(v_delta, vsubq_s32(v_src0.val[bidx], v_Y0), v_c4); v_Cb0 = vshrq_n_s32(vaddq_s32(v_Cb0, v_delta2), yuv_shift); v_src0.val[0] = vreinterpretq_s32_u32(vmovl_u16(vget_high_u16(v_src.val[0]))); v_src0.val[1] = vreinterpretq_s32_u32(vmovl_u16(vget_high_u16(v_src.val[1]))); v_src0.val[2] = vreinterpretq_s32_u32(vmovl_u16(vget_high_u16(v_src.val[2]))); int32x4_t v_Y1 = vmlaq_s32(vmlaq_s32(vmulq_s32(v_src0.val[0], v_c0), v_src0.val[1], v_c1), v_src0.val[2], v_c2); v_Y1 = vshrq_n_s32(vaddq_s32(v_Y1, v_delta2), yuv_shift); int32x4_t v_Cr1 = vmlaq_s32(v_delta, vsubq_s32(v_src0.val[bidx^2], v_Y1), v_c3); v_Cr1 = vshrq_n_s32(vaddq_s32(v_Cr1, v_delta2), yuv_shift); int32x4_t v_Cb1 = vmlaq_s32(v_delta, vsubq_s32(v_src0.val[bidx], v_Y1), v_c4); v_Cb1 = vshrq_n_s32(vaddq_s32(v_Cb1, v_delta2), yuv_shift); v_dst.val[0] = vcombine_u16(vqmovun_s32(v_Y0), vqmovun_s32(v_Y1)); v_dst.val[1] = vcombine_u16(vqmovun_s32(v_Cr0), vqmovun_s32(v_Cr1)); v_dst.val[2] = vcombine_u16(vqmovun_s32(v_Cb0), vqmovun_s32(v_Cb1)); vst3q_u16(dst + i, v_dst); } for ( ; i <= n - 12; i += 12, src += scn * 4) { uint16x4x3_t v_dst; int32x4x3_t v_src0; if (scn == 3) { uint16x4x3_t v_src = vld3_u16(src); v_src0.val[0] = vreinterpretq_s32_u32(vmovl_u16(v_src.val[0])); v_src0.val[1] = vreinterpretq_s32_u32(vmovl_u16(v_src.val[1])); v_src0.val[2] = vreinterpretq_s32_u32(vmovl_u16(v_src.val[2])); } else { uint16x4x4_t v_src = vld4_u16(src); v_src0.val[0] = vreinterpretq_s32_u32(vmovl_u16(v_src.val[0])); v_src0.val[1] = vreinterpretq_s32_u32(vmovl_u16(v_src.val[1])); v_src0.val[2] = vreinterpretq_s32_u32(vmovl_u16(v_src.val[2])); } int32x4_t v_Y = vmlaq_s32(vmlaq_s32(vmulq_s32(v_src0.val[0], v_c0), v_src0.val[1], v_c1), v_src0.val[2], v_c2); v_Y = vshrq_n_s32(vaddq_s32(v_Y, v_delta2), yuv_shift); int32x4_t v_Cr = vmlaq_s32(v_delta, vsubq_s32(v_src0.val[bidx^2], v_Y), v_c3); v_Cr = vshrq_n_s32(vaddq_s32(v_Cr, v_delta2), yuv_shift); int32x4_t v_Cb = vmlaq_s32(v_delta, vsubq_s32(v_src0.val[bidx], v_Y), v_c4); v_Cb = vshrq_n_s32(vaddq_s32(v_Cb, v_delta2), yuv_shift); v_dst.val[0] = vqmovun_s32(v_Y); v_dst.val[1] = vqmovun_s32(v_Cr); v_dst.val[2] = vqmovun_s32(v_Cb); vst3_u16(dst + i, v_dst); } for ( ; i < n; i += 3, src += scn) { int Y = CV_DESCALE(src[0]*C0 + src[1]*C1 + src[2]*C2, yuv_shift); int Cr = CV_DESCALE((src[bidx^2] - Y)*C3 + delta, yuv_shift); int Cb = CV_DESCALE((src[bidx] - Y)*C4 + delta, yuv_shift); dst[i] = saturate_cast(Y); dst[i+1] = saturate_cast(Cr); dst[i+2] = saturate_cast(Cb); } } int srccn, blueIdx, coeffs[5]; int32x4_t v_c0, v_c1, v_c2, v_c3, v_c4, v_delta, v_delta2; }; #elif CV_SSE4_1 template <> struct RGB2YCrCb_i { typedef uchar channel_type; RGB2YCrCb_i(int _srccn, int _blueIdx, const int* _coeffs) : srccn(_srccn), blueIdx(_blueIdx) { static const int coeffs0[] = {R2Y, G2Y, B2Y, 11682, 9241}; memcpy(coeffs, _coeffs ? _coeffs : coeffs0, 5*sizeof(coeffs[0])); if (blueIdx==0) std::swap(coeffs[0], coeffs[2]); v_c0 = _mm_set1_epi32(coeffs[0]); v_c1 = _mm_set1_epi32(coeffs[1]); v_c2 = _mm_set1_epi32(coeffs[2]); v_c3 = _mm_set1_epi32(coeffs[3]); v_c4 = _mm_set1_epi32(coeffs[4]); v_delta2 = _mm_set1_epi32(1 << (yuv_shift - 1)); v_delta = _mm_set1_epi32(ColorChannel::half()*(1 << yuv_shift)); v_delta = _mm_add_epi32(v_delta, v_delta2); v_zero = _mm_setzero_si128(); haveSIMD = checkHardwareSupport(CV_CPU_SSE4_1); } // 16u x 8 void process(__m128i v_r, __m128i v_g, __m128i v_b, __m128i & v_y, __m128i & v_cr, __m128i & v_cb) const { __m128i v_r_p = _mm_unpacklo_epi16(v_r, v_zero); __m128i v_g_p = _mm_unpacklo_epi16(v_g, v_zero); __m128i v_b_p = _mm_unpacklo_epi16(v_b, v_zero); __m128i v_y0 = _mm_add_epi32(_mm_mullo_epi32(v_r_p, v_c0), _mm_add_epi32(_mm_mullo_epi32(v_g_p, v_c1), _mm_mullo_epi32(v_b_p, v_c2))); v_y0 = _mm_srli_epi32(_mm_add_epi32(v_delta2, v_y0), yuv_shift); __m128i v_cr0 = _mm_mullo_epi32(_mm_sub_epi32(blueIdx == 2 ? v_r_p : v_b_p, v_y0), v_c3); __m128i v_cb0 = _mm_mullo_epi32(_mm_sub_epi32(blueIdx == 0 ? v_r_p : v_b_p, v_y0), v_c4); v_cr0 = _mm_srai_epi32(_mm_add_epi32(v_delta, v_cr0), yuv_shift); v_cb0 = _mm_srai_epi32(_mm_add_epi32(v_delta, v_cb0), yuv_shift); v_r_p = _mm_unpackhi_epi16(v_r, v_zero); v_g_p = _mm_unpackhi_epi16(v_g, v_zero); v_b_p = _mm_unpackhi_epi16(v_b, v_zero); __m128i v_y1 = _mm_add_epi32(_mm_mullo_epi32(v_r_p, v_c0), _mm_add_epi32(_mm_mullo_epi32(v_g_p, v_c1), _mm_mullo_epi32(v_b_p, v_c2))); v_y1 = _mm_srli_epi32(_mm_add_epi32(v_delta2, v_y1), yuv_shift); __m128i v_cr1 = _mm_mullo_epi32(_mm_sub_epi32(blueIdx == 2 ? v_r_p : v_b_p, v_y1), v_c3); __m128i v_cb1 = _mm_mullo_epi32(_mm_sub_epi32(blueIdx == 0 ? v_r_p : v_b_p, v_y1), v_c4); v_cr1 = _mm_srai_epi32(_mm_add_epi32(v_delta, v_cr1), yuv_shift); v_cb1 = _mm_srai_epi32(_mm_add_epi32(v_delta, v_cb1), yuv_shift); v_y = _mm_packs_epi32(v_y0, v_y1); v_cr = _mm_packs_epi32(v_cr0, v_cr1); v_cb = _mm_packs_epi32(v_cb0, v_cb1); } void operator()(const uchar * src, uchar * dst, int n) const { int scn = srccn, bidx = blueIdx, i = 0; int C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4]; int delta = ColorChannel::half()*(1 << yuv_shift); n *= 3; if (haveSIMD) { for ( ; i <= n - 96; i += 96, src += scn * 32) { __m128i v_r0 = _mm_loadu_si128((__m128i const *)(src)); __m128i v_r1 = _mm_loadu_si128((__m128i const *)(src + 16)); __m128i v_g0 = _mm_loadu_si128((__m128i const *)(src + 32)); __m128i v_g1 = _mm_loadu_si128((__m128i const *)(src + 48)); __m128i v_b0 = _mm_loadu_si128((__m128i const *)(src + 64)); __m128i v_b1 = _mm_loadu_si128((__m128i const *)(src + 80)); if (scn == 4) { __m128i v_a0 = _mm_loadu_si128((__m128i const *)(src + 96)); __m128i v_a1 = _mm_loadu_si128((__m128i const *)(src + 112)); _mm_deinterleave_epi8(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1, v_a0, v_a1); } else _mm_deinterleave_epi8(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1); __m128i v_y0 = v_zero, v_cr0 = v_zero, v_cb0 = v_zero; process(_mm_unpacklo_epi8(v_r0, v_zero), _mm_unpacklo_epi8(v_g0, v_zero), _mm_unpacklo_epi8(v_b0, v_zero), v_y0, v_cr0, v_cb0); __m128i v_y1 = v_zero, v_cr1 = v_zero, v_cb1 = v_zero; process(_mm_unpackhi_epi8(v_r0, v_zero), _mm_unpackhi_epi8(v_g0, v_zero), _mm_unpackhi_epi8(v_b0, v_zero), v_y1, v_cr1, v_cb1); __m128i v_y_0 = _mm_packus_epi16(v_y0, v_y1); __m128i v_cr_0 = _mm_packus_epi16(v_cr0, v_cr1); __m128i v_cb_0 = _mm_packus_epi16(v_cb0, v_cb1); process(_mm_unpacklo_epi8(v_r1, v_zero), _mm_unpacklo_epi8(v_g1, v_zero), _mm_unpacklo_epi8(v_b1, v_zero), v_y0, v_cr0, v_cb0); process(_mm_unpackhi_epi8(v_r1, v_zero), _mm_unpackhi_epi8(v_g1, v_zero), _mm_unpackhi_epi8(v_b1, v_zero), v_y1, v_cr1, v_cb1); __m128i v_y_1 = _mm_packus_epi16(v_y0, v_y1); __m128i v_cr_1 = _mm_packus_epi16(v_cr0, v_cr1); __m128i v_cb_1 = _mm_packus_epi16(v_cb0, v_cb1); _mm_interleave_epi8(v_y_0, v_y_1, v_cr_0, v_cr_1, v_cb_0, v_cb_1); _mm_storeu_si128((__m128i *)(dst + i), v_y_0); _mm_storeu_si128((__m128i *)(dst + i + 16), v_y_1); _mm_storeu_si128((__m128i *)(dst + i + 32), v_cr_0); _mm_storeu_si128((__m128i *)(dst + i + 48), v_cr_1); _mm_storeu_si128((__m128i *)(dst + i + 64), v_cb_0); _mm_storeu_si128((__m128i *)(dst + i + 80), v_cb_1); } } for ( ; i < n; i += 3, src += scn) { int Y = CV_DESCALE(src[0]*C0 + src[1]*C1 + src[2]*C2, yuv_shift); int Cr = CV_DESCALE((src[bidx^2] - Y)*C3 + delta, yuv_shift); int Cb = CV_DESCALE((src[bidx] - Y)*C4 + delta, yuv_shift); dst[i] = saturate_cast(Y); dst[i+1] = saturate_cast(Cr); dst[i+2] = saturate_cast(Cb); } } int srccn, blueIdx, coeffs[5]; __m128i v_c0, v_c1, v_c2; __m128i v_c3, v_c4, v_delta, v_delta2; __m128i v_zero; bool haveSIMD; }; template <> struct RGB2YCrCb_i { typedef ushort channel_type; RGB2YCrCb_i(int _srccn, int _blueIdx, const int* _coeffs) : srccn(_srccn), blueIdx(_blueIdx) { static const int coeffs0[] = {R2Y, G2Y, B2Y, 11682, 9241}; memcpy(coeffs, _coeffs ? _coeffs : coeffs0, 5*sizeof(coeffs[0])); if (blueIdx==0) std::swap(coeffs[0], coeffs[2]); v_c0 = _mm_set1_epi32(coeffs[0]); v_c1 = _mm_set1_epi32(coeffs[1]); v_c2 = _mm_set1_epi32(coeffs[2]); v_c3 = _mm_set1_epi32(coeffs[3]); v_c4 = _mm_set1_epi32(coeffs[4]); v_delta2 = _mm_set1_epi32(1 << (yuv_shift - 1)); v_delta = _mm_set1_epi32(ColorChannel::half()*(1 << yuv_shift)); v_delta = _mm_add_epi32(v_delta, v_delta2); v_zero = _mm_setzero_si128(); haveSIMD = checkHardwareSupport(CV_CPU_SSE4_1); } // 16u x 8 void process(__m128i v_r, __m128i v_g, __m128i v_b, __m128i & v_y, __m128i & v_cr, __m128i & v_cb) const { __m128i v_r_p = _mm_unpacklo_epi16(v_r, v_zero); __m128i v_g_p = _mm_unpacklo_epi16(v_g, v_zero); __m128i v_b_p = _mm_unpacklo_epi16(v_b, v_zero); __m128i v_y0 = _mm_add_epi32(_mm_mullo_epi32(v_r_p, v_c0), _mm_add_epi32(_mm_mullo_epi32(v_g_p, v_c1), _mm_mullo_epi32(v_b_p, v_c2))); v_y0 = _mm_srli_epi32(_mm_add_epi32(v_delta2, v_y0), yuv_shift); __m128i v_cr0 = _mm_mullo_epi32(_mm_sub_epi32(blueIdx == 2 ? v_r_p : v_b_p, v_y0), v_c3); __m128i v_cb0 = _mm_mullo_epi32(_mm_sub_epi32(blueIdx == 0 ? v_r_p : v_b_p, v_y0), v_c4); v_cr0 = _mm_srai_epi32(_mm_add_epi32(v_delta, v_cr0), yuv_shift); v_cb0 = _mm_srai_epi32(_mm_add_epi32(v_delta, v_cb0), yuv_shift); v_r_p = _mm_unpackhi_epi16(v_r, v_zero); v_g_p = _mm_unpackhi_epi16(v_g, v_zero); v_b_p = _mm_unpackhi_epi16(v_b, v_zero); __m128i v_y1 = _mm_add_epi32(_mm_mullo_epi32(v_r_p, v_c0), _mm_add_epi32(_mm_mullo_epi32(v_g_p, v_c1), _mm_mullo_epi32(v_b_p, v_c2))); v_y1 = _mm_srli_epi32(_mm_add_epi32(v_delta2, v_y1), yuv_shift); __m128i v_cr1 = _mm_mullo_epi32(_mm_sub_epi32(blueIdx == 2 ? v_r_p : v_b_p, v_y1), v_c3); __m128i v_cb1 = _mm_mullo_epi32(_mm_sub_epi32(blueIdx == 0 ? v_r_p : v_b_p, v_y1), v_c4); v_cr1 = _mm_srai_epi32(_mm_add_epi32(v_delta, v_cr1), yuv_shift); v_cb1 = _mm_srai_epi32(_mm_add_epi32(v_delta, v_cb1), yuv_shift); v_y = _mm_packus_epi32(v_y0, v_y1); v_cr = _mm_packus_epi32(v_cr0, v_cr1); v_cb = _mm_packus_epi32(v_cb0, v_cb1); } void operator()(const ushort * src, ushort * dst, int n) const { int scn = srccn, bidx = blueIdx, i = 0; int C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4]; int delta = ColorChannel::half()*(1 << yuv_shift); n *= 3; if (haveSIMD) { for ( ; i <= n - 48; i += 48, src += scn * 16) { __m128i v_r0 = _mm_loadu_si128((__m128i const *)(src)); __m128i v_r1 = _mm_loadu_si128((__m128i const *)(src + 8)); __m128i v_g0 = _mm_loadu_si128((__m128i const *)(src + 16)); __m128i v_g1 = _mm_loadu_si128((__m128i const *)(src + 24)); __m128i v_b0 = _mm_loadu_si128((__m128i const *)(src + 32)); __m128i v_b1 = _mm_loadu_si128((__m128i const *)(src + 40)); if (scn == 4) { __m128i v_a0 = _mm_loadu_si128((__m128i const *)(src + 48)); __m128i v_a1 = _mm_loadu_si128((__m128i const *)(src + 56)); _mm_deinterleave_epi16(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1, v_a0, v_a1); } else _mm_deinterleave_epi16(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1); __m128i v_y0 = v_zero, v_cr0 = v_zero, v_cb0 = v_zero; process(v_r0, v_g0, v_b0, v_y0, v_cr0, v_cb0); __m128i v_y1 = v_zero, v_cr1 = v_zero, v_cb1 = v_zero; process(v_r1, v_g1, v_b1, v_y1, v_cr1, v_cb1); _mm_interleave_epi16(v_y0, v_y1, v_cr0, v_cr1, v_cb0, v_cb1); _mm_storeu_si128((__m128i *)(dst + i), v_y0); _mm_storeu_si128((__m128i *)(dst + i + 8), v_y1); _mm_storeu_si128((__m128i *)(dst + i + 16), v_cr0); _mm_storeu_si128((__m128i *)(dst + i + 24), v_cr1); _mm_storeu_si128((__m128i *)(dst + i + 32), v_cb0); _mm_storeu_si128((__m128i *)(dst + i + 40), v_cb1); } } for ( ; i < n; i += 3, src += scn) { int Y = CV_DESCALE(src[0]*C0 + src[1]*C1 + src[2]*C2, yuv_shift); int Cr = CV_DESCALE((src[bidx^2] - Y)*C3 + delta, yuv_shift); int Cb = CV_DESCALE((src[bidx] - Y)*C4 + delta, yuv_shift); dst[i] = saturate_cast(Y); dst[i+1] = saturate_cast(Cr); dst[i+2] = saturate_cast(Cb); } } int srccn, blueIdx, coeffs[5]; __m128i v_c0, v_c1, v_c2; __m128i v_c3, v_c4, v_delta, v_delta2; __m128i v_zero; bool haveSIMD; }; #endif // CV_SSE4_1 template struct YCrCb2RGB_f { typedef _Tp channel_type; YCrCb2RGB_f(int _dstcn, int _blueIdx, const float* _coeffs) : dstcn(_dstcn), blueIdx(_blueIdx) { static const float coeffs0[] = {1.403f, -0.714f, -0.344f, 1.773f}; memcpy(coeffs, _coeffs ? _coeffs : coeffs0, 4*sizeof(coeffs[0])); } void operator()(const _Tp* src, _Tp* dst, int n) const { int dcn = dstcn, bidx = blueIdx; const _Tp delta = ColorChannel<_Tp>::half(), alpha = ColorChannel<_Tp>::max(); float C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3]; n *= 3; for(int i = 0; i < n; i += 3, dst += dcn) { _Tp Y = src[i]; _Tp Cr = src[i+1]; _Tp Cb = src[i+2]; _Tp b = saturate_cast<_Tp>(Y + (Cb - delta)*C3); _Tp g = saturate_cast<_Tp>(Y + (Cb - delta)*C2 + (Cr - delta)*C1); _Tp r = saturate_cast<_Tp>(Y + (Cr - delta)*C0); dst[bidx] = b; dst[1] = g; dst[bidx^2] = r; if( dcn == 4 ) dst[3] = alpha; } } int dstcn, blueIdx; float coeffs[4]; }; #if CV_NEON template <> struct YCrCb2RGB_f { typedef float channel_type; YCrCb2RGB_f(int _dstcn, int _blueIdx, const float* _coeffs) : dstcn(_dstcn), blueIdx(_blueIdx) { static const float coeffs0[] = {1.403f, -0.714f, -0.344f, 1.773f}; memcpy(coeffs, _coeffs ? _coeffs : coeffs0, 4*sizeof(coeffs[0])); v_c0 = vdupq_n_f32(coeffs[0]); v_c1 = vdupq_n_f32(coeffs[1]); v_c2 = vdupq_n_f32(coeffs[2]); v_c3 = vdupq_n_f32(coeffs[3]); v_delta = vdupq_n_f32(ColorChannel::half()); v_alpha = vdupq_n_f32(ColorChannel::max()); } void operator()(const float* src, float* dst, int n) const { int dcn = dstcn, bidx = blueIdx, i = 0; const float delta = ColorChannel::half(), alpha = ColorChannel::max(); float C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3]; n *= 3; if (dcn == 3) for ( ; i <= n - 12; i += 12, dst += 12) { float32x4x3_t v_src = vld3q_f32(src + i), v_dst; float32x4_t v_Y = v_src.val[0], v_Cr = v_src.val[1], v_Cb = v_src.val[2]; v_dst.val[bidx] = vmlaq_f32(v_Y, vsubq_f32(v_Cb, v_delta), v_c3); v_dst.val[1] = vaddq_f32(vmlaq_f32(vmulq_f32(vsubq_f32(v_Cb, v_delta), v_c2), vsubq_f32(v_Cr, v_delta), v_c1), v_Y); v_dst.val[bidx^2] = vmlaq_f32(v_Y, vsubq_f32(v_Cr, v_delta), v_c0); vst3q_f32(dst, v_dst); } else for ( ; i <= n - 12; i += 12, dst += 16) { float32x4x3_t v_src = vld3q_f32(src + i); float32x4x4_t v_dst; float32x4_t v_Y = v_src.val[0], v_Cr = v_src.val[1], v_Cb = v_src.val[2]; v_dst.val[bidx] = vmlaq_f32(v_Y, vsubq_f32(v_Cb, v_delta), v_c3); v_dst.val[1] = vaddq_f32(vmlaq_f32(vmulq_f32(vsubq_f32(v_Cb, v_delta), v_c2), vsubq_f32(v_Cr, v_delta), v_c1), v_Y); v_dst.val[bidx^2] = vmlaq_f32(v_Y, vsubq_f32(v_Cr, v_delta), v_c0); v_dst.val[3] = v_alpha; vst4q_f32(dst, v_dst); } for ( ; i < n; i += 3, dst += dcn) { float Y = src[i], Cr = src[i+1], Cb = src[i+2]; float b = Y + (Cb - delta)*C3; float g = Y + (Cb - delta)*C2 + (Cr - delta)*C1; float r = Y + (Cr - delta)*C0; dst[bidx] = b; dst[1] = g; dst[bidx^2] = r; if( dcn == 4 ) dst[3] = alpha; } } int dstcn, blueIdx; float coeffs[4]; float32x4_t v_c0, v_c1, v_c2, v_c3, v_alpha, v_delta; }; #elif CV_SSE2 template <> struct YCrCb2RGB_f { typedef float channel_type; YCrCb2RGB_f(int _dstcn, int _blueIdx, const float* _coeffs) : dstcn(_dstcn), blueIdx(_blueIdx) { static const float coeffs0[] = {1.403f, -0.714f, -0.344f, 1.773f}; memcpy(coeffs, _coeffs ? _coeffs : coeffs0, 4*sizeof(coeffs[0])); v_c0 = _mm_set1_ps(coeffs[0]); v_c1 = _mm_set1_ps(coeffs[1]); v_c2 = _mm_set1_ps(coeffs[2]); v_c3 = _mm_set1_ps(coeffs[3]); v_delta = _mm_set1_ps(ColorChannel::half()); v_alpha = _mm_set1_ps(ColorChannel::max()); haveSIMD = checkHardwareSupport(CV_CPU_SSE2); } void process(__m128 v_y, __m128 v_cr, __m128 v_cb, __m128 & v_r, __m128 & v_g, __m128 & v_b) const { v_cb = _mm_sub_ps(v_cb, v_delta); v_cr = _mm_sub_ps(v_cr, v_delta); v_b = _mm_mul_ps(v_cb, v_c3); v_g = _mm_add_ps(_mm_mul_ps(v_cb, v_c2), _mm_mul_ps(v_cr, v_c1)); v_r = _mm_mul_ps(v_cr, v_c0); v_b = _mm_add_ps(v_b, v_y); v_g = _mm_add_ps(v_g, v_y); v_r = _mm_add_ps(v_r, v_y); if (blueIdx == 0) std::swap(v_b, v_r); } void operator()(const float* src, float* dst, int n) const { int dcn = dstcn, bidx = blueIdx, i = 0; const float delta = ColorChannel::half(), alpha = ColorChannel::max(); float C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3]; n *= 3; if (haveSIMD) { for ( ; i <= n - 24; i += 24, dst += 8 * dcn) { __m128 v_y0 = _mm_loadu_ps(src + i); __m128 v_y1 = _mm_loadu_ps(src + i + 4); __m128 v_cr0 = _mm_loadu_ps(src + i + 8); __m128 v_cr1 = _mm_loadu_ps(src + i + 12); __m128 v_cb0 = _mm_loadu_ps(src + i + 16); __m128 v_cb1 = _mm_loadu_ps(src + i + 20); _mm_deinterleave_ps(v_y0, v_y1, v_cr0, v_cr1, v_cb0, v_cb1); __m128 v_r0, v_g0, v_b0; process(v_y0, v_cr0, v_cb0, v_r0, v_g0, v_b0); __m128 v_r1, v_g1, v_b1; process(v_y1, v_cr1, v_cb1, v_r1, v_g1, v_b1); __m128 v_a0 = v_alpha, v_a1 = v_alpha; if (dcn == 3) _mm_interleave_ps(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1); else _mm_interleave_ps(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1, v_a0, v_a1); _mm_storeu_ps(dst, v_r0); _mm_storeu_ps(dst + 4, v_r1); _mm_storeu_ps(dst + 8, v_g0); _mm_storeu_ps(dst + 12, v_g1); _mm_storeu_ps(dst + 16, v_b0); _mm_storeu_ps(dst + 20, v_b1); if (dcn == 4) { _mm_storeu_ps(dst + 24, v_a0); _mm_storeu_ps(dst + 28, v_a1); } } } for ( ; i < n; i += 3, dst += dcn) { float Y = src[i], Cr = src[i+1], Cb = src[i+2]; float b = Y + (Cb - delta)*C3; float g = Y + (Cb - delta)*C2 + (Cr - delta)*C1; float r = Y + (Cr - delta)*C0; dst[bidx] = b; dst[1] = g; dst[bidx^2] = r; if( dcn == 4 ) dst[3] = alpha; } } int dstcn, blueIdx; float coeffs[4]; __m128 v_c0, v_c1, v_c2, v_c3, v_alpha, v_delta; bool haveSIMD; }; #endif template struct YCrCb2RGB_i { typedef _Tp channel_type; YCrCb2RGB_i(int _dstcn, int _blueIdx, const int* _coeffs) : dstcn(_dstcn), blueIdx(_blueIdx) { static const int coeffs0[] = {22987, -11698, -5636, 29049}; memcpy(coeffs, _coeffs ? _coeffs : coeffs0, 4*sizeof(coeffs[0])); } void operator()(const _Tp* src, _Tp* dst, int n) const { int dcn = dstcn, bidx = blueIdx; const _Tp delta = ColorChannel<_Tp>::half(), alpha = ColorChannel<_Tp>::max(); int C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3]; n *= 3; for(int i = 0; i < n; i += 3, dst += dcn) { _Tp Y = src[i]; _Tp Cr = src[i+1]; _Tp Cb = src[i+2]; int b = Y + CV_DESCALE((Cb - delta)*C3, yuv_shift); int g = Y + CV_DESCALE((Cb - delta)*C2 + (Cr - delta)*C1, yuv_shift); int r = Y + CV_DESCALE((Cr - delta)*C0, yuv_shift); dst[bidx] = saturate_cast<_Tp>(b); dst[1] = saturate_cast<_Tp>(g); dst[bidx^2] = saturate_cast<_Tp>(r); if( dcn == 4 ) dst[3] = alpha; } } int dstcn, blueIdx; int coeffs[4]; }; #if CV_NEON template <> struct YCrCb2RGB_i { typedef uchar channel_type; YCrCb2RGB_i(int _dstcn, int _blueIdx, const int* _coeffs) : dstcn(_dstcn), blueIdx(_blueIdx) { static const int coeffs0[] = {22987, -11698, -5636, 29049}; memcpy(coeffs, _coeffs ? _coeffs : coeffs0, 4*sizeof(coeffs[0])); v_c0 = vdupq_n_s32(coeffs[0]); v_c1 = vdupq_n_s32(coeffs[1]); v_c2 = vdupq_n_s32(coeffs[2]); v_c3 = vdupq_n_s32(coeffs[3]); v_delta = vdup_n_s16(ColorChannel::half()); v_delta2 = vdupq_n_s32(1 << (yuv_shift - 1)); v_alpha = vdup_n_u8(ColorChannel::max()); } void operator()(const uchar* src, uchar* dst, int n) const { int dcn = dstcn, bidx = blueIdx, i = 0; const uchar delta = ColorChannel::half(), alpha = ColorChannel::max(); int C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3]; n *= 3; for ( ; i <= n - 24; i += 24, dst += dcn * 8) { uint8x8x3_t v_src = vld3_u8(src + i); int16x8x3_t v_src16; v_src16.val[0] = vreinterpretq_s16_u16(vmovl_u8(v_src.val[0])); v_src16.val[1] = vreinterpretq_s16_u16(vmovl_u8(v_src.val[1])); v_src16.val[2] = vreinterpretq_s16_u16(vmovl_u8(v_src.val[2])); int16x4_t v_Y = vget_low_s16(v_src16.val[0]), v_Cr = vget_low_s16(v_src16.val[1]), v_Cb = vget_low_s16(v_src16.val[2]); int32x4_t v_b0 = vmulq_s32(v_c3, vsubl_s16(v_Cb, v_delta)); v_b0 = vaddw_s16(vshrq_n_s32(vaddq_s32(v_b0, v_delta2), yuv_shift), v_Y); int32x4_t v_g0 = vmlaq_s32(vmulq_s32(vsubl_s16(v_Cr, v_delta), v_c1), vsubl_s16(v_Cb, v_delta), v_c2); v_g0 = vaddw_s16(vshrq_n_s32(vaddq_s32(v_g0, v_delta2), yuv_shift), v_Y); int32x4_t v_r0 = vmulq_s32(v_c0, vsubl_s16(v_Cr, v_delta)); v_r0 = vaddw_s16(vshrq_n_s32(vaddq_s32(v_r0, v_delta2), yuv_shift), v_Y); v_Y = vget_high_s16(v_src16.val[0]); v_Cr = vget_high_s16(v_src16.val[1]); v_Cb = vget_high_s16(v_src16.val[2]); int32x4_t v_b1 = vmulq_s32(v_c3, vsubl_s16(v_Cb, v_delta)); v_b1 = vaddw_s16(vshrq_n_s32(vaddq_s32(v_b1, v_delta2), yuv_shift), v_Y); int32x4_t v_g1 = vmlaq_s32(vmulq_s32(vsubl_s16(v_Cr, v_delta), v_c1), vsubl_s16(v_Cb, v_delta), v_c2); v_g1 = vaddw_s16(vshrq_n_s32(vaddq_s32(v_g1, v_delta2), yuv_shift), v_Y); int32x4_t v_r1 = vmulq_s32(v_c0, vsubl_s16(v_Cr, v_delta)); v_r1 = vaddw_s16(vshrq_n_s32(vaddq_s32(v_r1, v_delta2), yuv_shift), v_Y); uint8x8_t v_b = vqmovun_s16(vcombine_s16(vmovn_s32(v_b0), vmovn_s32(v_b1))); uint8x8_t v_g = vqmovun_s16(vcombine_s16(vmovn_s32(v_g0), vmovn_s32(v_g1))); uint8x8_t v_r = vqmovun_s16(vcombine_s16(vmovn_s32(v_r0), vmovn_s32(v_r1))); if (dcn == 3) { uint8x8x3_t v_dst; v_dst.val[bidx] = v_b; v_dst.val[1] = v_g; v_dst.val[bidx^2] = v_r; vst3_u8(dst, v_dst); } else { uint8x8x4_t v_dst; v_dst.val[bidx] = v_b; v_dst.val[1] = v_g; v_dst.val[bidx^2] = v_r; v_dst.val[3] = v_alpha; vst4_u8(dst, v_dst); } } for ( ; i < n; i += 3, dst += dcn) { uchar Y = src[i]; uchar Cr = src[i+1]; uchar Cb = src[i+2]; int b = Y + CV_DESCALE((Cb - delta)*C3, yuv_shift); int g = Y + CV_DESCALE((Cb - delta)*C2 + (Cr - delta)*C1, yuv_shift); int r = Y + CV_DESCALE((Cr - delta)*C0, yuv_shift); dst[bidx] = saturate_cast(b); dst[1] = saturate_cast(g); dst[bidx^2] = saturate_cast(r); if( dcn == 4 ) dst[3] = alpha; } } int dstcn, blueIdx; int coeffs[4]; int32x4_t v_c0, v_c1, v_c2, v_c3, v_delta2; int16x4_t v_delta; uint8x8_t v_alpha; }; template <> struct YCrCb2RGB_i { typedef ushort channel_type; YCrCb2RGB_i(int _dstcn, int _blueIdx, const int* _coeffs) : dstcn(_dstcn), blueIdx(_blueIdx) { static const int coeffs0[] = {22987, -11698, -5636, 29049}; memcpy(coeffs, _coeffs ? _coeffs : coeffs0, 4*sizeof(coeffs[0])); v_c0 = vdupq_n_s32(coeffs[0]); v_c1 = vdupq_n_s32(coeffs[1]); v_c2 = vdupq_n_s32(coeffs[2]); v_c3 = vdupq_n_s32(coeffs[3]); v_delta = vdupq_n_s32(ColorChannel::half()); v_delta2 = vdupq_n_s32(1 << (yuv_shift - 1)); v_alpha = vdupq_n_u16(ColorChannel::max()); v_alpha2 = vget_low_u16(v_alpha); } void operator()(const ushort* src, ushort* dst, int n) const { int dcn = dstcn, bidx = blueIdx, i = 0; const ushort delta = ColorChannel::half(), alpha = ColorChannel::max(); int C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3]; n *= 3; for ( ; i <= n - 24; i += 24, dst += dcn * 8) { uint16x8x3_t v_src = vld3q_u16(src + i); int32x4_t v_Y = vreinterpretq_s32_u32(vmovl_u16(vget_low_u16(v_src.val[0]))), v_Cr = vreinterpretq_s32_u32(vmovl_u16(vget_low_u16(v_src.val[1]))), v_Cb = vreinterpretq_s32_u32(vmovl_u16(vget_low_u16(v_src.val[2]))); int32x4_t v_b0 = vmulq_s32(v_c3, vsubq_s32(v_Cb, v_delta)); v_b0 = vaddq_s32(vshrq_n_s32(vaddq_s32(v_b0, v_delta2), yuv_shift), v_Y); int32x4_t v_g0 = vmlaq_s32(vmulq_s32(vsubq_s32(v_Cr, v_delta), v_c1), vsubq_s32(v_Cb, v_delta), v_c2); v_g0 = vaddq_s32(vshrq_n_s32(vaddq_s32(v_g0, v_delta2), yuv_shift), v_Y); int32x4_t v_r0 = vmulq_s32(v_c0, vsubq_s32(v_Cr, v_delta)); v_r0 = vaddq_s32(vshrq_n_s32(vaddq_s32(v_r0, v_delta2), yuv_shift), v_Y); v_Y = vreinterpretq_s32_u32(vmovl_u16(vget_high_u16(v_src.val[0]))), v_Cr = vreinterpretq_s32_u32(vmovl_u16(vget_high_u16(v_src.val[1]))), v_Cb = vreinterpretq_s32_u32(vmovl_u16(vget_high_u16(v_src.val[2]))); int32x4_t v_b1 = vmulq_s32(v_c3, vsubq_s32(v_Cb, v_delta)); v_b1 = vaddq_s32(vshrq_n_s32(vaddq_s32(v_b1, v_delta2), yuv_shift), v_Y); int32x4_t v_g1 = vmlaq_s32(vmulq_s32(vsubq_s32(v_Cr, v_delta), v_c1), vsubq_s32(v_Cb, v_delta), v_c2); v_g1 = vaddq_s32(vshrq_n_s32(vaddq_s32(v_g1, v_delta2), yuv_shift), v_Y); int32x4_t v_r1 = vmulq_s32(v_c0, vsubq_s32(v_Cr, v_delta)); v_r1 = vaddq_s32(vshrq_n_s32(vaddq_s32(v_r1, v_delta2), yuv_shift), v_Y); uint16x8_t v_b = vcombine_u16(vqmovun_s32(v_b0), vqmovun_s32(v_b1)); uint16x8_t v_g = vcombine_u16(vqmovun_s32(v_g0), vqmovun_s32(v_g1)); uint16x8_t v_r = vcombine_u16(vqmovun_s32(v_r0), vqmovun_s32(v_r1)); if (dcn == 3) { uint16x8x3_t v_dst; v_dst.val[bidx] = v_b; v_dst.val[1] = v_g; v_dst.val[bidx^2] = v_r; vst3q_u16(dst, v_dst); } else { uint16x8x4_t v_dst; v_dst.val[bidx] = v_b; v_dst.val[1] = v_g; v_dst.val[bidx^2] = v_r; v_dst.val[3] = v_alpha; vst4q_u16(dst, v_dst); } } for ( ; i <= n - 12; i += 12, dst += dcn * 4) { uint16x4x3_t v_src = vld3_u16(src + i); int32x4_t v_Y = vreinterpretq_s32_u32(vmovl_u16(v_src.val[0])), v_Cr = vreinterpretq_s32_u32(vmovl_u16(v_src.val[1])), v_Cb = vreinterpretq_s32_u32(vmovl_u16(v_src.val[2])); int32x4_t v_b = vmulq_s32(v_c3, vsubq_s32(v_Cb, v_delta)); v_b = vaddq_s32(vshrq_n_s32(vaddq_s32(v_b, v_delta2), yuv_shift), v_Y); int32x4_t v_g = vmlaq_s32(vmulq_s32(vsubq_s32(v_Cr, v_delta), v_c1), vsubq_s32(v_Cb, v_delta), v_c2); v_g = vaddq_s32(vshrq_n_s32(vaddq_s32(v_g, v_delta2), yuv_shift), v_Y); int32x4_t v_r = vmulq_s32(vsubq_s32(v_Cr, v_delta), v_c0); v_r = vaddq_s32(vshrq_n_s32(vaddq_s32(v_r, v_delta2), yuv_shift), v_Y); uint16x4_t v_bd = vqmovun_s32(v_b); uint16x4_t v_gd = vqmovun_s32(v_g); uint16x4_t v_rd = vqmovun_s32(v_r); if (dcn == 3) { uint16x4x3_t v_dst; v_dst.val[bidx] = v_bd; v_dst.val[1] = v_gd; v_dst.val[bidx^2] = v_rd; vst3_u16(dst, v_dst); } else { uint16x4x4_t v_dst; v_dst.val[bidx] = v_bd; v_dst.val[1] = v_gd; v_dst.val[bidx^2] = v_rd; v_dst.val[3] = v_alpha2; vst4_u16(dst, v_dst); } } for ( ; i < n; i += 3, dst += dcn) { ushort Y = src[i]; ushort Cr = src[i+1]; ushort Cb = src[i+2]; int b = Y + CV_DESCALE((Cb - delta)*C3, yuv_shift); int g = Y + CV_DESCALE((Cb - delta)*C2 + (Cr - delta)*C1, yuv_shift); int r = Y + CV_DESCALE((Cr - delta)*C0, yuv_shift); dst[bidx] = saturate_cast(b); dst[1] = saturate_cast(g); dst[bidx^2] = saturate_cast(r); if( dcn == 4 ) dst[3] = alpha; } } int dstcn, blueIdx; int coeffs[4]; int32x4_t v_c0, v_c1, v_c2, v_c3, v_delta2, v_delta; uint16x8_t v_alpha; uint16x4_t v_alpha2; }; #elif CV_SSE2 template <> struct YCrCb2RGB_i { typedef uchar channel_type; YCrCb2RGB_i(int _dstcn, int _blueIdx, const int* _coeffs) : dstcn(_dstcn), blueIdx(_blueIdx) { static const int coeffs0[] = {22987, -11698, -5636, 29049}; memcpy(coeffs, _coeffs ? _coeffs : coeffs0, 4*sizeof(coeffs[0])); v_c0 = _mm_set1_epi16((short)coeffs[0]); v_c1 = _mm_set1_epi16((short)coeffs[1]); v_c2 = _mm_set1_epi16((short)coeffs[2]); v_c3 = _mm_set1_epi16((short)coeffs[3]); v_delta = _mm_set1_epi16(ColorChannel::half()); v_delta2 = _mm_set1_epi32(1 << (yuv_shift - 1)); v_zero = _mm_setzero_si128(); uchar alpha = ColorChannel::max(); v_alpha = _mm_set1_epi8(*(char *)&alpha); useSSE = coeffs[0] <= std::numeric_limits::max(); haveSIMD = checkHardwareSupport(CV_CPU_SSE2); } // 16s x 8 void process(__m128i v_y, __m128i v_cr, __m128i v_cb, __m128i & v_r, __m128i & v_g, __m128i & v_b) const { v_cr = _mm_sub_epi16(v_cr, v_delta); v_cb = _mm_sub_epi16(v_cb, v_delta); __m128i v_y_p = _mm_unpacklo_epi16(v_y, v_zero); __m128i v_mullo_3 = _mm_mullo_epi16(v_cb, v_c3); __m128i v_mullo_2 = _mm_mullo_epi16(v_cb, v_c2); __m128i v_mullo_1 = _mm_mullo_epi16(v_cr, v_c1); __m128i v_mullo_0 = _mm_mullo_epi16(v_cr, v_c0); __m128i v_mulhi_3 = _mm_mulhi_epi16(v_cb, v_c3); __m128i v_mulhi_2 = _mm_mulhi_epi16(v_cb, v_c2); __m128i v_mulhi_1 = _mm_mulhi_epi16(v_cr, v_c1); __m128i v_mulhi_0 = _mm_mulhi_epi16(v_cr, v_c0); __m128i v_b0 = _mm_srai_epi32(_mm_add_epi32(_mm_unpacklo_epi16(v_mullo_3, v_mulhi_3), v_delta2), yuv_shift); __m128i v_g0 = _mm_srai_epi32(_mm_add_epi32(_mm_add_epi32(_mm_unpacklo_epi16(v_mullo_2, v_mulhi_2), _mm_unpacklo_epi16(v_mullo_1, v_mulhi_1)), v_delta2), yuv_shift); __m128i v_r0 = _mm_srai_epi32(_mm_add_epi32(_mm_unpacklo_epi16(v_mullo_0, v_mulhi_0), v_delta2), yuv_shift); v_r0 = _mm_add_epi32(v_r0, v_y_p); v_g0 = _mm_add_epi32(v_g0, v_y_p); v_b0 = _mm_add_epi32(v_b0, v_y_p); v_y_p = _mm_unpackhi_epi16(v_y, v_zero); __m128i v_b1 = _mm_srai_epi32(_mm_add_epi32(_mm_unpackhi_epi16(v_mullo_3, v_mulhi_3), v_delta2), yuv_shift); __m128i v_g1 = _mm_srai_epi32(_mm_add_epi32(_mm_add_epi32(_mm_unpackhi_epi16(v_mullo_2, v_mulhi_2), _mm_unpackhi_epi16(v_mullo_1, v_mulhi_1)), v_delta2), yuv_shift); __m128i v_r1 = _mm_srai_epi32(_mm_add_epi32(_mm_unpackhi_epi16(v_mullo_0, v_mulhi_0), v_delta2), yuv_shift); v_r1 = _mm_add_epi32(v_r1, v_y_p); v_g1 = _mm_add_epi32(v_g1, v_y_p); v_b1 = _mm_add_epi32(v_b1, v_y_p); v_r = _mm_packs_epi32(v_r0, v_r1); v_g = _mm_packs_epi32(v_g0, v_g1); v_b = _mm_packs_epi32(v_b0, v_b1); } void operator()(const uchar* src, uchar* dst, int n) const { int dcn = dstcn, bidx = blueIdx, i = 0; const uchar delta = ColorChannel::half(), alpha = ColorChannel::max(); int C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3]; n *= 3; if (haveSIMD && useSSE) { for ( ; i <= n - 96; i += 96, dst += dcn * 32) { __m128i v_y0 = _mm_loadu_si128((__m128i const *)(src + i)); __m128i v_y1 = _mm_loadu_si128((__m128i const *)(src + i + 16)); __m128i v_cr0 = _mm_loadu_si128((__m128i const *)(src + i + 32)); __m128i v_cr1 = _mm_loadu_si128((__m128i const *)(src + i + 48)); __m128i v_cb0 = _mm_loadu_si128((__m128i const *)(src + i + 64)); __m128i v_cb1 = _mm_loadu_si128((__m128i const *)(src + i + 80)); _mm_deinterleave_epi8(v_y0, v_y1, v_cr0, v_cr1, v_cb0, v_cb1); __m128i v_r_0 = v_zero, v_g_0 = v_zero, v_b_0 = v_zero; process(_mm_unpacklo_epi8(v_y0, v_zero), _mm_unpacklo_epi8(v_cr0, v_zero), _mm_unpacklo_epi8(v_cb0, v_zero), v_r_0, v_g_0, v_b_0); __m128i v_r_1 = v_zero, v_g_1 = v_zero, v_b_1 = v_zero; process(_mm_unpackhi_epi8(v_y0, v_zero), _mm_unpackhi_epi8(v_cr0, v_zero), _mm_unpackhi_epi8(v_cb0, v_zero), v_r_1, v_g_1, v_b_1); __m128i v_r0 = _mm_packus_epi16(v_r_0, v_r_1); __m128i v_g0 = _mm_packus_epi16(v_g_0, v_g_1); __m128i v_b0 = _mm_packus_epi16(v_b_0, v_b_1); process(_mm_unpacklo_epi8(v_y1, v_zero), _mm_unpacklo_epi8(v_cr1, v_zero), _mm_unpacklo_epi8(v_cb1, v_zero), v_r_0, v_g_0, v_b_0); process(_mm_unpackhi_epi8(v_y1, v_zero), _mm_unpackhi_epi8(v_cr1, v_zero), _mm_unpackhi_epi8(v_cb1, v_zero), v_r_1, v_g_1, v_b_1); __m128i v_r1 = _mm_packus_epi16(v_r_0, v_r_1); __m128i v_g1 = _mm_packus_epi16(v_g_0, v_g_1); __m128i v_b1 = _mm_packus_epi16(v_b_0, v_b_1); if (bidx == 0) { std::swap(v_r0, v_b0); std::swap(v_r1, v_b1); } __m128i v_a0 = v_alpha, v_a1 = v_alpha; if (dcn == 3) _mm_interleave_epi8(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1); else _mm_interleave_epi8(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1, v_a0, v_a1); _mm_storeu_si128((__m128i *)(dst), v_r0); _mm_storeu_si128((__m128i *)(dst + 16), v_r1); _mm_storeu_si128((__m128i *)(dst + 32), v_g0); _mm_storeu_si128((__m128i *)(dst + 48), v_g1); _mm_storeu_si128((__m128i *)(dst + 64), v_b0); _mm_storeu_si128((__m128i *)(dst + 80), v_b1); if (dcn == 4) { _mm_storeu_si128((__m128i *)(dst + 96), v_a0); _mm_storeu_si128((__m128i *)(dst + 112), v_a1); } } } for ( ; i < n; i += 3, dst += dcn) { uchar Y = src[i]; uchar Cr = src[i+1]; uchar Cb = src[i+2]; int b = Y + CV_DESCALE((Cb - delta)*C3, yuv_shift); int g = Y + CV_DESCALE((Cb - delta)*C2 + (Cr - delta)*C1, yuv_shift); int r = Y + CV_DESCALE((Cr - delta)*C0, yuv_shift); dst[bidx] = saturate_cast(b); dst[1] = saturate_cast(g); dst[bidx^2] = saturate_cast(r); if( dcn == 4 ) dst[3] = alpha; } } int dstcn, blueIdx; int coeffs[4]; bool useSSE, haveSIMD; __m128i v_c0, v_c1, v_c2, v_c3, v_delta2; __m128i v_delta, v_alpha, v_zero; }; #endif // CV_SSE2 ////////////////////////////////////// RGB <-> XYZ /////////////////////////////////////// static const float sRGB2XYZ_D65[] = { 0.412453f, 0.357580f, 0.180423f, 0.212671f, 0.715160f, 0.072169f, 0.019334f, 0.119193f, 0.950227f }; static const float XYZ2sRGB_D65[] = { 3.240479f, -1.53715f, -0.498535f, -0.969256f, 1.875991f, 0.041556f, 0.055648f, -0.204043f, 1.057311f }; template struct RGB2XYZ_f { typedef _Tp channel_type; RGB2XYZ_f(int _srccn, int blueIdx, const float* _coeffs) : srccn(_srccn) { memcpy(coeffs, _coeffs ? _coeffs : sRGB2XYZ_D65, 9*sizeof(coeffs[0])); if(blueIdx == 0) { std::swap(coeffs[0], coeffs[2]); std::swap(coeffs[3], coeffs[5]); std::swap(coeffs[6], coeffs[8]); } } void operator()(const _Tp* src, _Tp* dst, int n) const { int scn = srccn; float C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4], C5 = coeffs[5], C6 = coeffs[6], C7 = coeffs[7], C8 = coeffs[8]; n *= 3; for(int i = 0; i < n; i += 3, src += scn) { _Tp X = saturate_cast<_Tp>(src[0]*C0 + src[1]*C1 + src[2]*C2); _Tp Y = saturate_cast<_Tp>(src[0]*C3 + src[1]*C4 + src[2]*C5); _Tp Z = saturate_cast<_Tp>(src[0]*C6 + src[1]*C7 + src[2]*C8); dst[i] = X; dst[i+1] = Y; dst[i+2] = Z; } } int srccn; float coeffs[9]; }; #if CV_NEON template <> struct RGB2XYZ_f { typedef float channel_type; RGB2XYZ_f(int _srccn, int blueIdx, const float* _coeffs) : srccn(_srccn) { memcpy(coeffs, _coeffs ? _coeffs : sRGB2XYZ_D65, 9*sizeof(coeffs[0])); if(blueIdx == 0) { std::swap(coeffs[0], coeffs[2]); std::swap(coeffs[3], coeffs[5]); std::swap(coeffs[6], coeffs[8]); } v_c0 = vdupq_n_f32(coeffs[0]); v_c1 = vdupq_n_f32(coeffs[1]); v_c2 = vdupq_n_f32(coeffs[2]); v_c3 = vdupq_n_f32(coeffs[3]); v_c4 = vdupq_n_f32(coeffs[4]); v_c5 = vdupq_n_f32(coeffs[5]); v_c6 = vdupq_n_f32(coeffs[6]); v_c7 = vdupq_n_f32(coeffs[7]); v_c8 = vdupq_n_f32(coeffs[8]); } void operator()(const float* src, float* dst, int n) const { int scn = srccn, i = 0; float C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4], C5 = coeffs[5], C6 = coeffs[6], C7 = coeffs[7], C8 = coeffs[8]; n *= 3; if (scn == 3) for ( ; i <= n - 12; i += 12, src += 12) { float32x4x3_t v_src = vld3q_f32(src), v_dst; v_dst.val[0] = vmlaq_f32(vmlaq_f32(vmulq_f32(v_src.val[0], v_c0), v_src.val[1], v_c1), v_src.val[2], v_c2); v_dst.val[1] = vmlaq_f32(vmlaq_f32(vmulq_f32(v_src.val[0], v_c3), v_src.val[1], v_c4), v_src.val[2], v_c5); v_dst.val[2] = vmlaq_f32(vmlaq_f32(vmulq_f32(v_src.val[0], v_c6), v_src.val[1], v_c7), v_src.val[2], v_c8); vst3q_f32(dst + i, v_dst); } else for ( ; i <= n - 12; i += 12, src += 16) { float32x4x4_t v_src = vld4q_f32(src); float32x4x3_t v_dst; v_dst.val[0] = vmlaq_f32(vmlaq_f32(vmulq_f32(v_src.val[0], v_c0), v_src.val[1], v_c1), v_src.val[2], v_c2); v_dst.val[1] = vmlaq_f32(vmlaq_f32(vmulq_f32(v_src.val[0], v_c3), v_src.val[1], v_c4), v_src.val[2], v_c5); v_dst.val[2] = vmlaq_f32(vmlaq_f32(vmulq_f32(v_src.val[0], v_c6), v_src.val[1], v_c7), v_src.val[2], v_c8); vst3q_f32(dst + i, v_dst); } for ( ; i < n; i += 3, src += scn) { float X = saturate_cast(src[0]*C0 + src[1]*C1 + src[2]*C2); float Y = saturate_cast(src[0]*C3 + src[1]*C4 + src[2]*C5); float Z = saturate_cast(src[0]*C6 + src[1]*C7 + src[2]*C8); dst[i] = X; dst[i+1] = Y; dst[i+2] = Z; } } int srccn; float coeffs[9]; float32x4_t v_c0, v_c1, v_c2, v_c3, v_c4, v_c5, v_c6, v_c7, v_c8; }; #elif CV_SSE2 template <> struct RGB2XYZ_f { typedef float channel_type; RGB2XYZ_f(int _srccn, int blueIdx, const float* _coeffs) : srccn(_srccn) { memcpy(coeffs, _coeffs ? _coeffs : sRGB2XYZ_D65, 9*sizeof(coeffs[0])); if(blueIdx == 0) { std::swap(coeffs[0], coeffs[2]); std::swap(coeffs[3], coeffs[5]); std::swap(coeffs[6], coeffs[8]); } v_c0 = _mm_set1_ps(coeffs[0]); v_c1 = _mm_set1_ps(coeffs[1]); v_c2 = _mm_set1_ps(coeffs[2]); v_c3 = _mm_set1_ps(coeffs[3]); v_c4 = _mm_set1_ps(coeffs[4]); v_c5 = _mm_set1_ps(coeffs[5]); v_c6 = _mm_set1_ps(coeffs[6]); v_c7 = _mm_set1_ps(coeffs[7]); v_c8 = _mm_set1_ps(coeffs[8]); haveSIMD = checkHardwareSupport(CV_CPU_SSE2); } void process(__m128 v_r, __m128 v_g, __m128 v_b, __m128 & v_x, __m128 & v_y, __m128 & v_z) const { v_x = _mm_mul_ps(v_r, v_c0); v_x = _mm_add_ps(v_x, _mm_mul_ps(v_g, v_c1)); v_x = _mm_add_ps(v_x, _mm_mul_ps(v_b, v_c2)); v_y = _mm_mul_ps(v_r, v_c3); v_y = _mm_add_ps(v_y, _mm_mul_ps(v_g, v_c4)); v_y = _mm_add_ps(v_y, _mm_mul_ps(v_b, v_c5)); v_z = _mm_mul_ps(v_r, v_c6); v_z = _mm_add_ps(v_z, _mm_mul_ps(v_g, v_c7)); v_z = _mm_add_ps(v_z, _mm_mul_ps(v_b, v_c8)); } void operator()(const float* src, float* dst, int n) const { int scn = srccn, i = 0; float C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4], C5 = coeffs[5], C6 = coeffs[6], C7 = coeffs[7], C8 = coeffs[8]; n *= 3; if (haveSIMD) { for ( ; i <= n - 24; i += 24, src += 8 * scn) { __m128 v_r0 = _mm_loadu_ps(src); __m128 v_r1 = _mm_loadu_ps(src + 4); __m128 v_g0 = _mm_loadu_ps(src + 8); __m128 v_g1 = _mm_loadu_ps(src + 12); __m128 v_b0 = _mm_loadu_ps(src + 16); __m128 v_b1 = _mm_loadu_ps(src + 20); if (scn == 4) { __m128 v_a0 = _mm_loadu_ps(src + 24); __m128 v_a1 = _mm_loadu_ps(src + 28); _mm_deinterleave_ps(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1, v_a0, v_a1); } else _mm_deinterleave_ps(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1); __m128 v_x0, v_y0, v_z0; process(v_r0, v_g0, v_b0, v_x0, v_y0, v_z0); __m128 v_x1, v_y1, v_z1; process(v_r1, v_g1, v_b1, v_x1, v_y1, v_z1); _mm_interleave_ps(v_x0, v_x1, v_y0, v_y1, v_z0, v_z1); _mm_storeu_ps(dst + i, v_x0); _mm_storeu_ps(dst + i + 4, v_x1); _mm_storeu_ps(dst + i + 8, v_y0); _mm_storeu_ps(dst + i + 12, v_y1); _mm_storeu_ps(dst + i + 16, v_z0); _mm_storeu_ps(dst + i + 20, v_z1); } } for ( ; i < n; i += 3, src += scn) { float X = saturate_cast(src[0]*C0 + src[1]*C1 + src[2]*C2); float Y = saturate_cast(src[0]*C3 + src[1]*C4 + src[2]*C5); float Z = saturate_cast(src[0]*C6 + src[1]*C7 + src[2]*C8); dst[i] = X; dst[i+1] = Y; dst[i+2] = Z; } } int srccn; float coeffs[9]; __m128 v_c0, v_c1, v_c2, v_c3, v_c4, v_c5, v_c6, v_c7, v_c8; bool haveSIMD; }; #endif template struct RGB2XYZ_i { typedef _Tp channel_type; RGB2XYZ_i(int _srccn, int blueIdx, const float* _coeffs) : srccn(_srccn) { static const int coeffs0[] = { 1689, 1465, 739, 871, 2929, 296, 79, 488, 3892 }; for( int i = 0; i < 9; i++ ) coeffs[i] = _coeffs ? cvRound(_coeffs[i]*(1 << xyz_shift)) : coeffs0[i]; if(blueIdx == 0) { std::swap(coeffs[0], coeffs[2]); std::swap(coeffs[3], coeffs[5]); std::swap(coeffs[6], coeffs[8]); } } void operator()(const _Tp* src, _Tp* dst, int n) const { int scn = srccn; int C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4], C5 = coeffs[5], C6 = coeffs[6], C7 = coeffs[7], C8 = coeffs[8]; n *= 3; for(int i = 0; i < n; i += 3, src += scn) { int X = CV_DESCALE(src[0]*C0 + src[1]*C1 + src[2]*C2, xyz_shift); int Y = CV_DESCALE(src[0]*C3 + src[1]*C4 + src[2]*C5, xyz_shift); int Z = CV_DESCALE(src[0]*C6 + src[1]*C7 + src[2]*C8, xyz_shift); dst[i] = saturate_cast<_Tp>(X); dst[i+1] = saturate_cast<_Tp>(Y); dst[i+2] = saturate_cast<_Tp>(Z); } } int srccn; int coeffs[9]; }; #if CV_NEON template <> struct RGB2XYZ_i { typedef uchar channel_type; RGB2XYZ_i(int _srccn, int blueIdx, const float* _coeffs) : srccn(_srccn) { static const int coeffs0[] = { 1689, 1465, 739, 871, 2929, 296, 79, 488, 3892 }; for( int i = 0; i < 9; i++ ) coeffs[i] = _coeffs ? cvRound(_coeffs[i]*(1 << xyz_shift)) : coeffs0[i]; if(blueIdx == 0) { std::swap(coeffs[0], coeffs[2]); std::swap(coeffs[3], coeffs[5]); std::swap(coeffs[6], coeffs[8]); } v_c0 = vdup_n_u16(coeffs[0]); v_c1 = vdup_n_u16(coeffs[1]); v_c2 = vdup_n_u16(coeffs[2]); v_c3 = vdup_n_u16(coeffs[3]); v_c4 = vdup_n_u16(coeffs[4]); v_c5 = vdup_n_u16(coeffs[5]); v_c6 = vdup_n_u16(coeffs[6]); v_c7 = vdup_n_u16(coeffs[7]); v_c8 = vdup_n_u16(coeffs[8]); v_delta = vdupq_n_u32(1 << (xyz_shift - 1)); } void operator()(const uchar * src, uchar * dst, int n) const { int scn = srccn, i = 0; int C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4], C5 = coeffs[5], C6 = coeffs[6], C7 = coeffs[7], C8 = coeffs[8]; n *= 3; for ( ; i <= n - 24; i += 24, src += scn * 8) { uint8x8x3_t v_dst; uint16x8x3_t v_src16; if (scn == 3) { uint8x8x3_t v_src = vld3_u8(src); v_src16.val[0] = vmovl_u8(v_src.val[0]); v_src16.val[1] = vmovl_u8(v_src.val[1]); v_src16.val[2] = vmovl_u8(v_src.val[2]); } else { uint8x8x4_t v_src = vld4_u8(src); v_src16.val[0] = vmovl_u8(v_src.val[0]); v_src16.val[1] = vmovl_u8(v_src.val[1]); v_src16.val[2] = vmovl_u8(v_src.val[2]); } uint16x4_t v_s0 = vget_low_u16(v_src16.val[0]), v_s1 = vget_low_u16(v_src16.val[1]), v_s2 = vget_low_u16(v_src16.val[2]); uint32x4_t v_X0 = vmlal_u16(vmlal_u16(vmull_u16(v_s0, v_c0), v_s1, v_c1), v_s2, v_c2); uint32x4_t v_Y0 = vmlal_u16(vmlal_u16(vmull_u16(v_s0, v_c3), v_s1, v_c4), v_s2, v_c5); uint32x4_t v_Z0 = vmlal_u16(vmlal_u16(vmull_u16(v_s0, v_c6), v_s1, v_c7), v_s2, v_c8); v_X0 = vshrq_n_u32(vaddq_u32(v_X0, v_delta), xyz_shift); v_Y0 = vshrq_n_u32(vaddq_u32(v_Y0, v_delta), xyz_shift); v_Z0 = vshrq_n_u32(vaddq_u32(v_Z0, v_delta), xyz_shift); v_s0 = vget_high_u16(v_src16.val[0]), v_s1 = vget_high_u16(v_src16.val[1]), v_s2 = vget_high_u16(v_src16.val[2]); uint32x4_t v_X1 = vmlal_u16(vmlal_u16(vmull_u16(v_s0, v_c0), v_s1, v_c1), v_s2, v_c2); uint32x4_t v_Y1 = vmlal_u16(vmlal_u16(vmull_u16(v_s0, v_c3), v_s1, v_c4), v_s2, v_c5); uint32x4_t v_Z1 = vmlal_u16(vmlal_u16(vmull_u16(v_s0, v_c6), v_s1, v_c7), v_s2, v_c8); v_X1 = vshrq_n_u32(vaddq_u32(v_X1, v_delta), xyz_shift); v_Y1 = vshrq_n_u32(vaddq_u32(v_Y1, v_delta), xyz_shift); v_Z1 = vshrq_n_u32(vaddq_u32(v_Z1, v_delta), xyz_shift); v_dst.val[0] = vqmovn_u16(vcombine_u16(vmovn_u32(v_X0), vmovn_u32(v_X1))); v_dst.val[1] = vqmovn_u16(vcombine_u16(vmovn_u32(v_Y0), vmovn_u32(v_Y1))); v_dst.val[2] = vqmovn_u16(vcombine_u16(vmovn_u32(v_Z0), vmovn_u32(v_Z1))); vst3_u8(dst + i, v_dst); } for ( ; i < n; i += 3, src += scn) { int X = CV_DESCALE(src[0]*C0 + src[1]*C1 + src[2]*C2, xyz_shift); int Y = CV_DESCALE(src[0]*C3 + src[1]*C4 + src[2]*C5, xyz_shift); int Z = CV_DESCALE(src[0]*C6 + src[1]*C7 + src[2]*C8, xyz_shift); dst[i] = saturate_cast(X); dst[i+1] = saturate_cast(Y); dst[i+2] = saturate_cast(Z); } } int srccn, coeffs[9]; uint16x4_t v_c0, v_c1, v_c2, v_c3, v_c4, v_c5, v_c6, v_c7, v_c8; uint32x4_t v_delta; }; template <> struct RGB2XYZ_i { typedef ushort channel_type; RGB2XYZ_i(int _srccn, int blueIdx, const float* _coeffs) : srccn(_srccn) { static const int coeffs0[] = { 1689, 1465, 739, 871, 2929, 296, 79, 488, 3892 }; for( int i = 0; i < 9; i++ ) coeffs[i] = _coeffs ? cvRound(_coeffs[i]*(1 << xyz_shift)) : coeffs0[i]; if(blueIdx == 0) { std::swap(coeffs[0], coeffs[2]); std::swap(coeffs[3], coeffs[5]); std::swap(coeffs[6], coeffs[8]); } v_c0 = vdup_n_u16(coeffs[0]); v_c1 = vdup_n_u16(coeffs[1]); v_c2 = vdup_n_u16(coeffs[2]); v_c3 = vdup_n_u16(coeffs[3]); v_c4 = vdup_n_u16(coeffs[4]); v_c5 = vdup_n_u16(coeffs[5]); v_c6 = vdup_n_u16(coeffs[6]); v_c7 = vdup_n_u16(coeffs[7]); v_c8 = vdup_n_u16(coeffs[8]); v_delta = vdupq_n_u32(1 << (xyz_shift - 1)); } void operator()(const ushort * src, ushort * dst, int n) const { int scn = srccn, i = 0; int C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4], C5 = coeffs[5], C6 = coeffs[6], C7 = coeffs[7], C8 = coeffs[8]; n *= 3; for ( ; i <= n - 24; i += 24, src += scn * 8) { uint16x8x3_t v_src, v_dst; if (scn == 3) v_src = vld3q_u16(src); else { uint16x8x4_t v_src4 = vld4q_u16(src); v_src.val[0] = v_src4.val[0]; v_src.val[1] = v_src4.val[1]; v_src.val[2] = v_src4.val[2]; } uint16x4_t v_s0 = vget_low_u16(v_src.val[0]), v_s1 = vget_low_u16(v_src.val[1]), v_s2 = vget_low_u16(v_src.val[2]); uint32x4_t v_X0 = vmlal_u16(vmlal_u16(vmull_u16(v_s0, v_c0), v_s1, v_c1), v_s2, v_c2); uint32x4_t v_Y0 = vmlal_u16(vmlal_u16(vmull_u16(v_s0, v_c3), v_s1, v_c4), v_s2, v_c5); uint32x4_t v_Z0 = vmlal_u16(vmlal_u16(vmull_u16(v_s0, v_c6), v_s1, v_c7), v_s2, v_c8); v_X0 = vshrq_n_u32(vaddq_u32(v_X0, v_delta), xyz_shift); v_Y0 = vshrq_n_u32(vaddq_u32(v_Y0, v_delta), xyz_shift); v_Z0 = vshrq_n_u32(vaddq_u32(v_Z0, v_delta), xyz_shift); v_s0 = vget_high_u16(v_src.val[0]), v_s1 = vget_high_u16(v_src.val[1]), v_s2 = vget_high_u16(v_src.val[2]); uint32x4_t v_X1 = vmlal_u16(vmlal_u16(vmull_u16(v_s0, v_c0), v_s1, v_c1), v_s2, v_c2); uint32x4_t v_Y1 = vmlal_u16(vmlal_u16(vmull_u16(v_s0, v_c3), v_s1, v_c4), v_s2, v_c5); uint32x4_t v_Z1 = vmlal_u16(vmlal_u16(vmull_u16(v_s0, v_c6), v_s1, v_c7), v_s2, v_c8); v_X1 = vshrq_n_u32(vaddq_u32(v_X1, v_delta), xyz_shift); v_Y1 = vshrq_n_u32(vaddq_u32(v_Y1, v_delta), xyz_shift); v_Z1 = vshrq_n_u32(vaddq_u32(v_Z1, v_delta), xyz_shift); v_dst.val[0] = vcombine_u16(vqmovn_u32(v_X0), vqmovn_u32(v_X1)); v_dst.val[1] = vcombine_u16(vqmovn_u32(v_Y0), vqmovn_u32(v_Y1)); v_dst.val[2] = vcombine_u16(vqmovn_u32(v_Z0), vqmovn_u32(v_Z1)); vst3q_u16(dst + i, v_dst); } for ( ; i <= n - 12; i += 12, src += scn * 4) { uint16x4x3_t v_dst; uint16x4_t v_s0, v_s1, v_s2; if (scn == 3) { uint16x4x3_t v_src = vld3_u16(src); v_s0 = v_src.val[0]; v_s1 = v_src.val[1]; v_s2 = v_src.val[2]; } else { uint16x4x4_t v_src = vld4_u16(src); v_s0 = v_src.val[0]; v_s1 = v_src.val[1]; v_s2 = v_src.val[2]; } uint32x4_t v_X = vmlal_u16(vmlal_u16(vmull_u16(v_s0, v_c0), v_s1, v_c1), v_s2, v_c2); uint32x4_t v_Y = vmlal_u16(vmlal_u16(vmull_u16(v_s0, v_c3), v_s1, v_c4), v_s2, v_c5); uint32x4_t v_Z = vmlal_u16(vmlal_u16(vmull_u16(v_s0, v_c6), v_s1, v_c7), v_s2, v_c8); v_dst.val[0] = vqmovn_u32(vshrq_n_u32(vaddq_u32(v_X, v_delta), xyz_shift)); v_dst.val[1] = vqmovn_u32(vshrq_n_u32(vaddq_u32(v_Y, v_delta), xyz_shift)); v_dst.val[2] = vqmovn_u32(vshrq_n_u32(vaddq_u32(v_Z, v_delta), xyz_shift)); vst3_u16(dst + i, v_dst); } for ( ; i < n; i += 3, src += scn) { int X = CV_DESCALE(src[0]*C0 + src[1]*C1 + src[2]*C2, xyz_shift); int Y = CV_DESCALE(src[0]*C3 + src[1]*C4 + src[2]*C5, xyz_shift); int Z = CV_DESCALE(src[0]*C6 + src[1]*C7 + src[2]*C8, xyz_shift); dst[i] = saturate_cast(X); dst[i+1] = saturate_cast(Y); dst[i+2] = saturate_cast(Z); } } int srccn, coeffs[9]; uint16x4_t v_c0, v_c1, v_c2, v_c3, v_c4, v_c5, v_c6, v_c7, v_c8; uint32x4_t v_delta; }; #endif template struct XYZ2RGB_f { typedef _Tp channel_type; XYZ2RGB_f(int _dstcn, int _blueIdx, const float* _coeffs) : dstcn(_dstcn), blueIdx(_blueIdx) { memcpy(coeffs, _coeffs ? _coeffs : XYZ2sRGB_D65, 9*sizeof(coeffs[0])); if(blueIdx == 0) { std::swap(coeffs[0], coeffs[6]); std::swap(coeffs[1], coeffs[7]); std::swap(coeffs[2], coeffs[8]); } } void operator()(const _Tp* src, _Tp* dst, int n) const { int dcn = dstcn; _Tp alpha = ColorChannel<_Tp>::max(); float C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4], C5 = coeffs[5], C6 = coeffs[6], C7 = coeffs[7], C8 = coeffs[8]; n *= 3; for(int i = 0; i < n; i += 3, dst += dcn) { _Tp B = saturate_cast<_Tp>(src[i]*C0 + src[i+1]*C1 + src[i+2]*C2); _Tp G = saturate_cast<_Tp>(src[i]*C3 + src[i+1]*C4 + src[i+2]*C5); _Tp R = saturate_cast<_Tp>(src[i]*C6 + src[i+1]*C7 + src[i+2]*C8); dst[0] = B; dst[1] = G; dst[2] = R; if( dcn == 4 ) dst[3] = alpha; } } int dstcn, blueIdx; float coeffs[9]; }; #if CV_SSE2 template <> struct XYZ2RGB_f { typedef float channel_type; XYZ2RGB_f(int _dstcn, int _blueIdx, const float* _coeffs) : dstcn(_dstcn), blueIdx(_blueIdx) { memcpy(coeffs, _coeffs ? _coeffs : XYZ2sRGB_D65, 9*sizeof(coeffs[0])); if(blueIdx == 0) { std::swap(coeffs[0], coeffs[6]); std::swap(coeffs[1], coeffs[7]); std::swap(coeffs[2], coeffs[8]); } v_c0 = _mm_set1_ps(coeffs[0]); v_c1 = _mm_set1_ps(coeffs[1]); v_c2 = _mm_set1_ps(coeffs[2]); v_c3 = _mm_set1_ps(coeffs[3]); v_c4 = _mm_set1_ps(coeffs[4]); v_c5 = _mm_set1_ps(coeffs[5]); v_c6 = _mm_set1_ps(coeffs[6]); v_c7 = _mm_set1_ps(coeffs[7]); v_c8 = _mm_set1_ps(coeffs[8]); v_alpha = _mm_set1_ps(ColorChannel::max()); haveSIMD = checkHardwareSupport(CV_CPU_SSE2); } void process(__m128 v_x, __m128 v_y, __m128 v_z, __m128 & v_r, __m128 & v_g, __m128 & v_b) const { v_b = _mm_mul_ps(v_x, v_c0); v_b = _mm_add_ps(v_b, _mm_mul_ps(v_y, v_c1)); v_b = _mm_add_ps(v_b, _mm_mul_ps(v_z, v_c2)); v_g = _mm_mul_ps(v_x, v_c3); v_g = _mm_add_ps(v_g, _mm_mul_ps(v_y, v_c4)); v_g = _mm_add_ps(v_g, _mm_mul_ps(v_z, v_c5)); v_r = _mm_mul_ps(v_x, v_c6); v_r = _mm_add_ps(v_r, _mm_mul_ps(v_y, v_c7)); v_r = _mm_add_ps(v_r, _mm_mul_ps(v_z, v_c8)); } void operator()(const float* src, float* dst, int n) const { int dcn = dstcn; float alpha = ColorChannel::max(); float C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4], C5 = coeffs[5], C6 = coeffs[6], C7 = coeffs[7], C8 = coeffs[8]; n *= 3; int i = 0; if (haveSIMD) { for ( ; i <= n - 24; i += 24, dst += 8 * dcn) { __m128 v_x0 = _mm_loadu_ps(src + i); __m128 v_x1 = _mm_loadu_ps(src + i + 4); __m128 v_y0 = _mm_loadu_ps(src + i + 8); __m128 v_y1 = _mm_loadu_ps(src + i + 12); __m128 v_z0 = _mm_loadu_ps(src + i + 16); __m128 v_z1 = _mm_loadu_ps(src + i + 20); _mm_deinterleave_ps(v_x0, v_x1, v_y0, v_y1, v_z0, v_z1); __m128 v_r0, v_g0, v_b0; process(v_x0, v_y0, v_z0, v_r0, v_g0, v_b0); __m128 v_r1, v_g1, v_b1; process(v_x1, v_y1, v_z1, v_r1, v_g1, v_b1); __m128 v_a0 = v_alpha, v_a1 = v_alpha; if (dcn == 4) _mm_interleave_ps(v_b0, v_b1, v_g0, v_g1, v_r0, v_r1, v_a0, v_a1); else _mm_interleave_ps(v_b0, v_b1, v_g0, v_g1, v_r0, v_r1); _mm_storeu_ps(dst, v_b0); _mm_storeu_ps(dst + 4, v_b1); _mm_storeu_ps(dst + 8, v_g0); _mm_storeu_ps(dst + 12, v_g1); _mm_storeu_ps(dst + 16, v_r0); _mm_storeu_ps(dst + 20, v_r1); if (dcn == 4) { _mm_storeu_ps(dst + 24, v_a0); _mm_storeu_ps(dst + 28, v_a1); } } } for( ; i < n; i += 3, dst += dcn) { float B = src[i]*C0 + src[i+1]*C1 + src[i+2]*C2; float G = src[i]*C3 + src[i+1]*C4 + src[i+2]*C5; float R = src[i]*C6 + src[i+1]*C7 + src[i+2]*C8; dst[0] = B; dst[1] = G; dst[2] = R; if( dcn == 4 ) dst[3] = alpha; } } int dstcn, blueIdx; float coeffs[9]; __m128 v_c0, v_c1, v_c2, v_c3, v_c4, v_c5, v_c6, v_c7, v_c8; __m128 v_alpha; bool haveSIMD; }; #endif // CV_SSE2 template struct XYZ2RGB_i { typedef _Tp channel_type; XYZ2RGB_i(int _dstcn, int _blueIdx, const int* _coeffs) : dstcn(_dstcn), blueIdx(_blueIdx) { static const int coeffs0[] = { 13273, -6296, -2042, -3970, 7684, 170, 228, -836, 4331 }; for(int i = 0; i < 9; i++) coeffs[i] = _coeffs ? cvRound(_coeffs[i]*(1 << xyz_shift)) : coeffs0[i]; if(blueIdx == 0) { std::swap(coeffs[0], coeffs[6]); std::swap(coeffs[1], coeffs[7]); std::swap(coeffs[2], coeffs[8]); } } void operator()(const _Tp* src, _Tp* dst, int n) const { int dcn = dstcn; _Tp alpha = ColorChannel<_Tp>::max(); int C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4], C5 = coeffs[5], C6 = coeffs[6], C7 = coeffs[7], C8 = coeffs[8]; n *= 3; for(int i = 0; i < n; i += 3, dst += dcn) { int B = CV_DESCALE(src[i]*C0 + src[i+1]*C1 + src[i+2]*C2, xyz_shift); int G = CV_DESCALE(src[i]*C3 + src[i+1]*C4 + src[i+2]*C5, xyz_shift); int R = CV_DESCALE(src[i]*C6 + src[i+1]*C7 + src[i+2]*C8, xyz_shift); dst[0] = saturate_cast<_Tp>(B); dst[1] = saturate_cast<_Tp>(G); dst[2] = saturate_cast<_Tp>(R); if( dcn == 4 ) dst[3] = alpha; } } int dstcn, blueIdx; int coeffs[9]; }; #if CV_NEON template <> struct XYZ2RGB_i { typedef uchar channel_type; XYZ2RGB_i(int _dstcn, int _blueIdx, const int* _coeffs) : dstcn(_dstcn), blueIdx(_blueIdx) { static const int coeffs0[] = { 13273, -6296, -2042, -3970, 7684, 170, 228, -836, 4331 }; for(int i = 0; i < 9; i++) coeffs[i] = _coeffs ? cvRound(_coeffs[i]*(1 << xyz_shift)) : coeffs0[i]; if(blueIdx == 0) { std::swap(coeffs[0], coeffs[6]); std::swap(coeffs[1], coeffs[7]); std::swap(coeffs[2], coeffs[8]); } v_c0 = vdup_n_s16(coeffs[0]); v_c1 = vdup_n_s16(coeffs[1]); v_c2 = vdup_n_s16(coeffs[2]); v_c3 = vdup_n_s16(coeffs[3]); v_c4 = vdup_n_s16(coeffs[4]); v_c5 = vdup_n_s16(coeffs[5]); v_c6 = vdup_n_s16(coeffs[6]); v_c7 = vdup_n_s16(coeffs[7]); v_c8 = vdup_n_s16(coeffs[8]); v_delta = vdupq_n_s32(1 << (xyz_shift - 1)); v_alpha = vmovn_u16(vdupq_n_u16(ColorChannel::max())); } void operator()(const uchar* src, uchar* dst, int n) const { int dcn = dstcn, i = 0; uchar alpha = ColorChannel::max(); int C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4], C5 = coeffs[5], C6 = coeffs[6], C7 = coeffs[7], C8 = coeffs[8]; n *= 3; for ( ; i <= n - 24; i += 24, dst += dcn * 8) { uint8x8x3_t v_src = vld3_u8(src + i); int16x8x3_t v_src16; v_src16.val[0] = vreinterpretq_s16_u16(vmovl_u8(v_src.val[0])); v_src16.val[1] = vreinterpretq_s16_u16(vmovl_u8(v_src.val[1])); v_src16.val[2] = vreinterpretq_s16_u16(vmovl_u8(v_src.val[2])); int16x4_t v_s0 = vget_low_s16(v_src16.val[0]), v_s1 = vget_low_s16(v_src16.val[1]), v_s2 = vget_low_s16(v_src16.val[2]); int32x4_t v_X0 = vmlal_s16(vmlal_s16(vmull_s16(v_s0, v_c0), v_s1, v_c1), v_s2, v_c2); int32x4_t v_Y0 = vmlal_s16(vmlal_s16(vmull_s16(v_s0, v_c3), v_s1, v_c4), v_s2, v_c5); int32x4_t v_Z0 = vmlal_s16(vmlal_s16(vmull_s16(v_s0, v_c6), v_s1, v_c7), v_s2, v_c8); v_X0 = vshrq_n_s32(vaddq_s32(v_X0, v_delta), xyz_shift); v_Y0 = vshrq_n_s32(vaddq_s32(v_Y0, v_delta), xyz_shift); v_Z0 = vshrq_n_s32(vaddq_s32(v_Z0, v_delta), xyz_shift); v_s0 = vget_high_s16(v_src16.val[0]), v_s1 = vget_high_s16(v_src16.val[1]), v_s2 = vget_high_s16(v_src16.val[2]); int32x4_t v_X1 = vmlal_s16(vmlal_s16(vmull_s16(v_s0, v_c0), v_s1, v_c1), v_s2, v_c2); int32x4_t v_Y1 = vmlal_s16(vmlal_s16(vmull_s16(v_s0, v_c3), v_s1, v_c4), v_s2, v_c5); int32x4_t v_Z1 = vmlal_s16(vmlal_s16(vmull_s16(v_s0, v_c6), v_s1, v_c7), v_s2, v_c8); v_X1 = vshrq_n_s32(vaddq_s32(v_X1, v_delta), xyz_shift); v_Y1 = vshrq_n_s32(vaddq_s32(v_Y1, v_delta), xyz_shift); v_Z1 = vshrq_n_s32(vaddq_s32(v_Z1, v_delta), xyz_shift); uint8x8_t v_b = vqmovun_s16(vcombine_s16(vqmovn_s32(v_X0), vqmovn_s32(v_X1))); uint8x8_t v_g = vqmovun_s16(vcombine_s16(vqmovn_s32(v_Y0), vqmovn_s32(v_Y1))); uint8x8_t v_r = vqmovun_s16(vcombine_s16(vqmovn_s32(v_Z0), vqmovn_s32(v_Z1))); if (dcn == 3) { uint8x8x3_t v_dst; v_dst.val[0] = v_b; v_dst.val[1] = v_g; v_dst.val[2] = v_r; vst3_u8(dst, v_dst); } else { uint8x8x4_t v_dst; v_dst.val[0] = v_b; v_dst.val[1] = v_g; v_dst.val[2] = v_r; v_dst.val[3] = v_alpha; vst4_u8(dst, v_dst); } } for ( ; i < n; i += 3, dst += dcn) { int B = CV_DESCALE(src[i]*C0 + src[i+1]*C1 + src[i+2]*C2, xyz_shift); int G = CV_DESCALE(src[i]*C3 + src[i+1]*C4 + src[i+2]*C5, xyz_shift); int R = CV_DESCALE(src[i]*C6 + src[i+1]*C7 + src[i+2]*C8, xyz_shift); dst[0] = saturate_cast(B); dst[1] = saturate_cast(G); dst[2] = saturate_cast(R); if( dcn == 4 ) dst[3] = alpha; } } int dstcn, blueIdx; int coeffs[9]; int16x4_t v_c0, v_c1, v_c2, v_c3, v_c4, v_c5, v_c6, v_c7, v_c8; uint8x8_t v_alpha; int32x4_t v_delta; }; template <> struct XYZ2RGB_i { typedef ushort channel_type; XYZ2RGB_i(int _dstcn, int _blueIdx, const int* _coeffs) : dstcn(_dstcn), blueIdx(_blueIdx) { static const int coeffs0[] = { 13273, -6296, -2042, -3970, 7684, 170, 228, -836, 4331 }; for(int i = 0; i < 9; i++) coeffs[i] = _coeffs ? cvRound(_coeffs[i]*(1 << xyz_shift)) : coeffs0[i]; if(blueIdx == 0) { std::swap(coeffs[0], coeffs[6]); std::swap(coeffs[1], coeffs[7]); std::swap(coeffs[2], coeffs[8]); } v_c0 = vdupq_n_s32(coeffs[0]); v_c1 = vdupq_n_s32(coeffs[1]); v_c2 = vdupq_n_s32(coeffs[2]); v_c3 = vdupq_n_s32(coeffs[3]); v_c4 = vdupq_n_s32(coeffs[4]); v_c5 = vdupq_n_s32(coeffs[5]); v_c6 = vdupq_n_s32(coeffs[6]); v_c7 = vdupq_n_s32(coeffs[7]); v_c8 = vdupq_n_s32(coeffs[8]); v_delta = vdupq_n_s32(1 << (xyz_shift - 1)); v_alpha = vdupq_n_u16(ColorChannel::max()); v_alpha2 = vget_low_u16(v_alpha); } void operator()(const ushort* src, ushort* dst, int n) const { int dcn = dstcn, i = 0; ushort alpha = ColorChannel::max(); int C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4], C5 = coeffs[5], C6 = coeffs[6], C7 = coeffs[7], C8 = coeffs[8]; n *= 3; for ( ; i <= n - 24; i += 24, dst += dcn * 8) { uint16x8x3_t v_src = vld3q_u16(src + i); int32x4_t v_s0 = vreinterpretq_s32_u32(vmovl_u16(vget_low_u16(v_src.val[0]))), v_s1 = vreinterpretq_s32_u32(vmovl_u16(vget_low_u16(v_src.val[1]))), v_s2 = vreinterpretq_s32_u32(vmovl_u16(vget_low_u16(v_src.val[2]))); int32x4_t v_X0 = vmlaq_s32(vmlaq_s32(vmulq_s32(v_s0, v_c0), v_s1, v_c1), v_s2, v_c2); int32x4_t v_Y0 = vmlaq_s32(vmlaq_s32(vmulq_s32(v_s0, v_c3), v_s1, v_c4), v_s2, v_c5); int32x4_t v_Z0 = vmlaq_s32(vmlaq_s32(vmulq_s32(v_s0, v_c6), v_s1, v_c7), v_s2, v_c8); v_X0 = vshrq_n_s32(vaddq_s32(v_X0, v_delta), xyz_shift); v_Y0 = vshrq_n_s32(vaddq_s32(v_Y0, v_delta), xyz_shift); v_Z0 = vshrq_n_s32(vaddq_s32(v_Z0, v_delta), xyz_shift); v_s0 = vreinterpretq_s32_u32(vmovl_u16(vget_high_u16(v_src.val[0]))); v_s1 = vreinterpretq_s32_u32(vmovl_u16(vget_high_u16(v_src.val[1]))); v_s2 = vreinterpretq_s32_u32(vmovl_u16(vget_high_u16(v_src.val[2]))); int32x4_t v_X1 = vmlaq_s32(vmlaq_s32(vmulq_s32(v_s0, v_c0), v_s1, v_c1), v_s2, v_c2); int32x4_t v_Y1 = vmlaq_s32(vmlaq_s32(vmulq_s32(v_s0, v_c3), v_s1, v_c4), v_s2, v_c5); int32x4_t v_Z1 = vmlaq_s32(vmlaq_s32(vmulq_s32(v_s0, v_c6), v_s1, v_c7), v_s2, v_c8); v_X1 = vshrq_n_s32(vaddq_s32(v_X1, v_delta), xyz_shift); v_Y1 = vshrq_n_s32(vaddq_s32(v_Y1, v_delta), xyz_shift); v_Z1 = vshrq_n_s32(vaddq_s32(v_Z1, v_delta), xyz_shift); uint16x8_t v_b = vcombine_u16(vqmovun_s32(v_X0), vqmovun_s32(v_X1)); uint16x8_t v_g = vcombine_u16(vqmovun_s32(v_Y0), vqmovun_s32(v_Y1)); uint16x8_t v_r = vcombine_u16(vqmovun_s32(v_Z0), vqmovun_s32(v_Z1)); if (dcn == 3) { uint16x8x3_t v_dst; v_dst.val[0] = v_b; v_dst.val[1] = v_g; v_dst.val[2] = v_r; vst3q_u16(dst, v_dst); } else { uint16x8x4_t v_dst; v_dst.val[0] = v_b; v_dst.val[1] = v_g; v_dst.val[2] = v_r; v_dst.val[3] = v_alpha; vst4q_u16(dst, v_dst); } } for ( ; i <= n - 12; i += 12, dst += dcn * 4) { uint16x4x3_t v_src = vld3_u16(src + i); int32x4_t v_s0 = vreinterpretq_s32_u32(vmovl_u16(v_src.val[0])), v_s1 = vreinterpretq_s32_u32(vmovl_u16(v_src.val[1])), v_s2 = vreinterpretq_s32_u32(vmovl_u16(v_src.val[2])); int32x4_t v_X = vmlaq_s32(vmlaq_s32(vmulq_s32(v_s0, v_c0), v_s1, v_c1), v_s2, v_c2); int32x4_t v_Y = vmlaq_s32(vmlaq_s32(vmulq_s32(v_s0, v_c3), v_s1, v_c4), v_s2, v_c5); int32x4_t v_Z = vmlaq_s32(vmlaq_s32(vmulq_s32(v_s0, v_c6), v_s1, v_c7), v_s2, v_c8); v_X = vshrq_n_s32(vaddq_s32(v_X, v_delta), xyz_shift); v_Y = vshrq_n_s32(vaddq_s32(v_Y, v_delta), xyz_shift); v_Z = vshrq_n_s32(vaddq_s32(v_Z, v_delta), xyz_shift); uint16x4_t v_b = vqmovun_s32(v_X); uint16x4_t v_g = vqmovun_s32(v_Y); uint16x4_t v_r = vqmovun_s32(v_Z); if (dcn == 3) { uint16x4x3_t v_dst; v_dst.val[0] = v_b; v_dst.val[1] = v_g; v_dst.val[2] = v_r; vst3_u16(dst, v_dst); } else { uint16x4x4_t v_dst; v_dst.val[0] = v_b; v_dst.val[1] = v_g; v_dst.val[2] = v_r; v_dst.val[3] = v_alpha2; vst4_u16(dst, v_dst); } } for ( ; i < n; i += 3, dst += dcn) { int B = CV_DESCALE(src[i]*C0 + src[i+1]*C1 + src[i+2]*C2, xyz_shift); int G = CV_DESCALE(src[i]*C3 + src[i+1]*C4 + src[i+2]*C5, xyz_shift); int R = CV_DESCALE(src[i]*C6 + src[i+1]*C7 + src[i+2]*C8, xyz_shift); dst[0] = saturate_cast(B); dst[1] = saturate_cast(G); dst[2] = saturate_cast(R); if( dcn == 4 ) dst[3] = alpha; } } int dstcn, blueIdx; int coeffs[9]; int32x4_t v_c0, v_c1, v_c2, v_c3, v_c4, v_c5, v_c6, v_c7, v_c8, v_delta; uint16x4_t v_alpha2; uint16x8_t v_alpha; }; #endif ////////////////////////////////////// RGB <-> HSV /////////////////////////////////////// struct RGB2HSV_b { typedef uchar channel_type; RGB2HSV_b(int _srccn, int _blueIdx, int _hrange) : srccn(_srccn), blueIdx(_blueIdx), hrange(_hrange) { CV_Assert( hrange == 180 || hrange == 256 ); } void operator()(const uchar* src, uchar* dst, int n) const { int i, bidx = blueIdx, scn = srccn; const int hsv_shift = 12; static int sdiv_table[256]; static int hdiv_table180[256]; static int hdiv_table256[256]; static volatile bool initialized = false; int hr = hrange; const int* hdiv_table = hr == 180 ? hdiv_table180 : hdiv_table256; n *= 3; if( !initialized ) { sdiv_table[0] = hdiv_table180[0] = hdiv_table256[0] = 0; for( i = 1; i < 256; i++ ) { sdiv_table[i] = saturate_cast((255 << hsv_shift)/(1.*i)); hdiv_table180[i] = saturate_cast((180 << hsv_shift)/(6.*i)); hdiv_table256[i] = saturate_cast((256 << hsv_shift)/(6.*i)); } initialized = true; } for( i = 0; i < n; i += 3, src += scn ) { int b = src[bidx], g = src[1], r = src[bidx^2]; int h, s, v = b; int vmin = b, diff; int vr, vg; CV_CALC_MAX_8U( v, g ); CV_CALC_MAX_8U( v, r ); CV_CALC_MIN_8U( vmin, g ); CV_CALC_MIN_8U( vmin, r ); diff = v - vmin; vr = v == r ? -1 : 0; vg = v == g ? -1 : 0; s = (diff * sdiv_table[v] + (1 << (hsv_shift-1))) >> hsv_shift; h = (vr & (g - b)) + (~vr & ((vg & (b - r + 2 * diff)) + ((~vg) & (r - g + 4 * diff)))); h = (h * hdiv_table[diff] + (1 << (hsv_shift-1))) >> hsv_shift; h += h < 0 ? hr : 0; dst[i] = saturate_cast(h); dst[i+1] = (uchar)s; dst[i+2] = (uchar)v; } } int srccn, blueIdx, hrange; }; struct RGB2HSV_f { typedef float channel_type; RGB2HSV_f(int _srccn, int _blueIdx, float _hrange) : srccn(_srccn), blueIdx(_blueIdx), hrange(_hrange) {} void operator()(const float* src, float* dst, int n) const { int i, bidx = blueIdx, scn = srccn; float hscale = hrange*(1.f/360.f); n *= 3; for( i = 0; i < n; i += 3, src += scn ) { float b = src[bidx], g = src[1], r = src[bidx^2]; float h, s, v; float vmin, diff; v = vmin = r; if( v < g ) v = g; if( v < b ) v = b; if( vmin > g ) vmin = g; if( vmin > b ) vmin = b; diff = v - vmin; s = diff/(float)(fabs(v) + FLT_EPSILON); diff = (float)(60./(diff + FLT_EPSILON)); if( v == r ) h = (g - b)*diff; else if( v == g ) h = (b - r)*diff + 120.f; else h = (r - g)*diff + 240.f; if( h < 0 ) h += 360.f; dst[i] = h*hscale; dst[i+1] = s; dst[i+2] = v; } } int srccn, blueIdx; float hrange; }; struct HSV2RGB_f { typedef float channel_type; HSV2RGB_f(int _dstcn, int _blueIdx, float _hrange) : dstcn(_dstcn), blueIdx(_blueIdx), hscale(6.f/_hrange) {} void operator()(const float* src, float* dst, int n) const { int i, bidx = blueIdx, dcn = dstcn; float _hscale = hscale; float alpha = ColorChannel::max(); n *= 3; for( i = 0; i < n; i += 3, dst += dcn ) { float h = src[i], s = src[i+1], v = src[i+2]; float b, g, r; if( s == 0 ) b = g = r = v; else { static const int sector_data[][3]= {{1,3,0}, {1,0,2}, {3,0,1}, {0,2,1}, {0,1,3}, {2,1,0}}; float tab[4]; int sector; h *= _hscale; if( h < 0 ) do h += 6; while( h < 0 ); else if( h >= 6 ) do h -= 6; while( h >= 6 ); sector = cvFloor(h); h -= sector; if( (unsigned)sector >= 6u ) { sector = 0; h = 0.f; } tab[0] = v; tab[1] = v*(1.f - s); tab[2] = v*(1.f - s*h); tab[3] = v*(1.f - s*(1.f - h)); b = tab[sector_data[sector][0]]; g = tab[sector_data[sector][1]]; r = tab[sector_data[sector][2]]; } dst[bidx] = b; dst[1] = g; dst[bidx^2] = r; if( dcn == 4 ) dst[3] = alpha; } } int dstcn, blueIdx; float hscale; }; struct HSV2RGB_b { typedef uchar channel_type; HSV2RGB_b(int _dstcn, int _blueIdx, int _hrange) : dstcn(_dstcn), cvt(3, _blueIdx, (float)_hrange) { #if CV_NEON v_scale_inv = vdupq_n_f32(1.f/255.f); v_scale = vdupq_n_f32(255.f); v_alpha = vdup_n_u8(ColorChannel::max()); #elif CV_SSE2 v_scale_inv = _mm_set1_ps(1.f/255.f); v_scale = _mm_set1_ps(255.0f); v_zero = _mm_setzero_si128(); haveSIMD = checkHardwareSupport(CV_CPU_SSE2); #endif } #if CV_SSE2 // 16s x 8 void process(__m128i v_r, __m128i v_g, __m128i v_b, float * buf) const { __m128 v_r0 = _mm_cvtepi32_ps(_mm_unpacklo_epi16(v_r, v_zero)); __m128 v_g0 = _mm_cvtepi32_ps(_mm_unpacklo_epi16(v_g, v_zero)); __m128 v_b0 = _mm_cvtepi32_ps(_mm_unpacklo_epi16(v_b, v_zero)); __m128 v_r1 = _mm_cvtepi32_ps(_mm_unpackhi_epi16(v_r, v_zero)); __m128 v_g1 = _mm_cvtepi32_ps(_mm_unpackhi_epi16(v_g, v_zero)); __m128 v_b1 = _mm_cvtepi32_ps(_mm_unpackhi_epi16(v_b, v_zero)); v_g0 = _mm_mul_ps(v_g0, v_scale_inv); v_b0 = _mm_mul_ps(v_b0, v_scale_inv); v_g1 = _mm_mul_ps(v_g1, v_scale_inv); v_b1 = _mm_mul_ps(v_b1, v_scale_inv); _mm_interleave_ps(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1); _mm_store_ps(buf, v_r0); _mm_store_ps(buf + 4, v_r1); _mm_store_ps(buf + 8, v_g0); _mm_store_ps(buf + 12, v_g1); _mm_store_ps(buf + 16, v_b0); _mm_store_ps(buf + 20, v_b1); } #endif void operator()(const uchar* src, uchar* dst, int n) const { int i, j, dcn = dstcn; uchar alpha = ColorChannel::max(); float CV_DECL_ALIGNED(16) buf[3*BLOCK_SIZE]; for( i = 0; i < n; i += BLOCK_SIZE, src += BLOCK_SIZE*3 ) { int dn = std::min(n - i, (int)BLOCK_SIZE); j = 0; #if CV_NEON for ( ; j <= (dn - 8) * 3; j += 24) { uint8x8x3_t v_src = vld3_u8(src + j); uint16x8_t v_t0 = vmovl_u8(v_src.val[0]), v_t1 = vmovl_u8(v_src.val[1]), v_t2 = vmovl_u8(v_src.val[2]); float32x4x3_t v_dst; v_dst.val[0] = vcvtq_f32_u32(vmovl_u16(vget_low_u16(v_t0))); v_dst.val[1] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_low_u16(v_t1))), v_scale_inv); v_dst.val[2] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_low_u16(v_t2))), v_scale_inv); vst3q_f32(buf + j, v_dst); v_dst.val[0] = vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_t0))); v_dst.val[1] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_t1))), v_scale_inv); v_dst.val[2] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_t2))), v_scale_inv); vst3q_f32(buf + j + 12, v_dst); } #elif CV_SSE2 if (haveSIMD) { for ( ; j <= (dn - 32) * 3; j += 96) { __m128i v_r0 = _mm_loadu_si128((__m128i const *)(src + j)); __m128i v_r1 = _mm_loadu_si128((__m128i const *)(src + j + 16)); __m128i v_g0 = _mm_loadu_si128((__m128i const *)(src + j + 32)); __m128i v_g1 = _mm_loadu_si128((__m128i const *)(src + j + 48)); __m128i v_b0 = _mm_loadu_si128((__m128i const *)(src + j + 64)); __m128i v_b1 = _mm_loadu_si128((__m128i const *)(src + j + 80)); _mm_deinterleave_epi8(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1); process(_mm_unpacklo_epi8(v_r0, v_zero), _mm_unpacklo_epi8(v_g0, v_zero), _mm_unpacklo_epi8(v_b0, v_zero), buf + j); process(_mm_unpackhi_epi8(v_r0, v_zero), _mm_unpackhi_epi8(v_g0, v_zero), _mm_unpackhi_epi8(v_b0, v_zero), buf + j + 24); process(_mm_unpacklo_epi8(v_r1, v_zero), _mm_unpacklo_epi8(v_g1, v_zero), _mm_unpacklo_epi8(v_b1, v_zero), buf + j + 48); process(_mm_unpackhi_epi8(v_r1, v_zero), _mm_unpackhi_epi8(v_g1, v_zero), _mm_unpackhi_epi8(v_b1, v_zero), buf + j + 72); } } #endif for( ; j < dn*3; j += 3 ) { buf[j] = src[j]; buf[j+1] = src[j+1]*(1.f/255.f); buf[j+2] = src[j+2]*(1.f/255.f); } cvt(buf, buf, dn); j = 0; #if CV_NEON for ( ; j <= (dn - 8) * 3; j += 24, dst += dcn * 8) { float32x4x3_t v_src0 = vld3q_f32(buf + j), v_src1 = vld3q_f32(buf + j + 12); uint8x8_t v_dst0 = vqmovn_u16(vcombine_u16(vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src0.val[0], v_scale))), vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src1.val[0], v_scale))))); uint8x8_t v_dst1 = vqmovn_u16(vcombine_u16(vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src0.val[1], v_scale))), vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src1.val[1], v_scale))))); uint8x8_t v_dst2 = vqmovn_u16(vcombine_u16(vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src0.val[2], v_scale))), vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src1.val[2], v_scale))))); if (dcn == 4) { uint8x8x4_t v_dst; v_dst.val[0] = v_dst0; v_dst.val[1] = v_dst1; v_dst.val[2] = v_dst2; v_dst.val[3] = v_alpha; vst4_u8(dst, v_dst); } else { uint8x8x3_t v_dst; v_dst.val[0] = v_dst0; v_dst.val[1] = v_dst1; v_dst.val[2] = v_dst2; vst3_u8(dst, v_dst); } } #elif CV_SSE2 if (dcn == 3 && haveSIMD) { for ( ; j <= (dn * 3 - 16); j += 16, dst += 16) { __m128 v_src0 = _mm_mul_ps(_mm_load_ps(buf + j), v_scale); __m128 v_src1 = _mm_mul_ps(_mm_load_ps(buf + j + 4), v_scale); __m128 v_src2 = _mm_mul_ps(_mm_load_ps(buf + j + 8), v_scale); __m128 v_src3 = _mm_mul_ps(_mm_load_ps(buf + j + 12), v_scale); __m128i v_dst0 = _mm_packs_epi32(_mm_cvtps_epi32(v_src0), _mm_cvtps_epi32(v_src1)); __m128i v_dst1 = _mm_packs_epi32(_mm_cvtps_epi32(v_src2), _mm_cvtps_epi32(v_src3)); _mm_storeu_si128((__m128i *)dst, _mm_packus_epi16(v_dst0, v_dst1)); } int jr = j % 3; if (jr) dst -= jr, j -= jr; } #endif for( ; j < dn*3; j += 3, dst += dcn ) { dst[0] = saturate_cast(buf[j]*255.f); dst[1] = saturate_cast(buf[j+1]*255.f); dst[2] = saturate_cast(buf[j+2]*255.f); if( dcn == 4 ) dst[3] = alpha; } } } int dstcn; HSV2RGB_f cvt; #if CV_NEON float32x4_t v_scale, v_scale_inv; uint8x8_t v_alpha; #elif CV_SSE2 __m128 v_scale_inv, v_scale; __m128i v_zero; bool haveSIMD; #endif }; ///////////////////////////////////// RGB <-> HLS //////////////////////////////////////// struct RGB2HLS_f { typedef float channel_type; RGB2HLS_f(int _srccn, int _blueIdx, float _hrange) : srccn(_srccn), blueIdx(_blueIdx), hrange(_hrange) {} void operator()(const float* src, float* dst, int n) const { int i, bidx = blueIdx, scn = srccn; float hscale = hrange*(1.f/360.f); n *= 3; for( i = 0; i < n; i += 3, src += scn ) { float b = src[bidx], g = src[1], r = src[bidx^2]; float h = 0.f, s = 0.f, l; float vmin, vmax, diff; vmax = vmin = r; if( vmax < g ) vmax = g; if( vmax < b ) vmax = b; if( vmin > g ) vmin = g; if( vmin > b ) vmin = b; diff = vmax - vmin; l = (vmax + vmin)*0.5f; if( diff > FLT_EPSILON ) { s = l < 0.5f ? diff/(vmax + vmin) : diff/(2 - vmax - vmin); diff = 60.f/diff; if( vmax == r ) h = (g - b)*diff; else if( vmax == g ) h = (b - r)*diff + 120.f; else h = (r - g)*diff + 240.f; if( h < 0.f ) h += 360.f; } dst[i] = h*hscale; dst[i+1] = l; dst[i+2] = s; } } int srccn, blueIdx; float hrange; }; struct RGB2HLS_b { typedef uchar channel_type; RGB2HLS_b(int _srccn, int _blueIdx, int _hrange) : srccn(_srccn), cvt(3, _blueIdx, (float)_hrange) { #if CV_NEON v_scale_inv = vdupq_n_f32(1.f/255.f); v_scale = vdupq_n_f32(255.f); v_alpha = vdup_n_u8(ColorChannel::max()); #elif CV_SSE2 v_scale_inv = _mm_set1_ps(1.f/255.f); v_scale = _mm_set1_ps(255.f); v_zero = _mm_setzero_si128(); haveSIMD = checkHardwareSupport(CV_CPU_SSE2); #endif } #if CV_SSE2 void process(const float * buf, __m128i & v_h, __m128i & v_l, __m128i & v_s) const { __m128 v_h0f = _mm_load_ps(buf); __m128 v_h1f = _mm_load_ps(buf + 4); __m128 v_l0f = _mm_load_ps(buf + 8); __m128 v_l1f = _mm_load_ps(buf + 12); __m128 v_s0f = _mm_load_ps(buf + 16); __m128 v_s1f = _mm_load_ps(buf + 20); _mm_deinterleave_ps(v_h0f, v_h1f, v_l0f, v_l1f, v_s0f, v_s1f); v_l0f = _mm_mul_ps(v_l0f, v_scale); v_l1f = _mm_mul_ps(v_l1f, v_scale); v_s0f = _mm_mul_ps(v_s0f, v_scale); v_s1f = _mm_mul_ps(v_s1f, v_scale); v_h = _mm_packs_epi32(_mm_cvtps_epi32(v_h0f), _mm_cvtps_epi32(v_h1f)); v_l = _mm_packs_epi32(_mm_cvtps_epi32(v_l0f), _mm_cvtps_epi32(v_l1f)); v_s = _mm_packs_epi32(_mm_cvtps_epi32(v_s0f), _mm_cvtps_epi32(v_s1f)); } #endif void operator()(const uchar* src, uchar* dst, int n) const { int i, j, scn = srccn; float CV_DECL_ALIGNED(16) buf[3*BLOCK_SIZE]; for( i = 0; i < n; i += BLOCK_SIZE, dst += BLOCK_SIZE*3 ) { int dn = std::min(n - i, (int)BLOCK_SIZE); j = 0; #if CV_NEON for ( ; j <= (dn - 8) * 3; j += 24, src += 8 * scn) { uint16x8_t v_t0, v_t1, v_t2; if (scn == 3) { uint8x8x3_t v_src = vld3_u8(src); v_t0 = vmovl_u8(v_src.val[0]); v_t1 = vmovl_u8(v_src.val[1]); v_t2 = vmovl_u8(v_src.val[2]); } else { uint8x8x4_t v_src = vld4_u8(src); v_t0 = vmovl_u8(v_src.val[0]); v_t1 = vmovl_u8(v_src.val[1]); v_t2 = vmovl_u8(v_src.val[2]); } float32x4x3_t v_dst; v_dst.val[0] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_low_u16(v_t0))), v_scale_inv); v_dst.val[1] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_low_u16(v_t1))), v_scale_inv); v_dst.val[2] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_low_u16(v_t2))), v_scale_inv); vst3q_f32(buf + j, v_dst); v_dst.val[0] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_t0))), v_scale_inv); v_dst.val[1] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_t1))), v_scale_inv); v_dst.val[2] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_t2))), v_scale_inv); vst3q_f32(buf + j + 12, v_dst); } #elif CV_SSE2 if (scn == 3 && haveSIMD) { for ( ; j <= (dn * 3 - 16); j += 16, src += 16) { __m128i v_src = _mm_loadu_si128((__m128i const *)src); __m128i v_src_p = _mm_unpacklo_epi8(v_src, v_zero); _mm_store_ps(buf + j, _mm_mul_ps(_mm_cvtepi32_ps(_mm_unpacklo_epi16(v_src_p, v_zero)), v_scale_inv)); _mm_store_ps(buf + j + 4, _mm_mul_ps(_mm_cvtepi32_ps(_mm_unpackhi_epi16(v_src_p, v_zero)), v_scale_inv)); v_src_p = _mm_unpackhi_epi8(v_src, v_zero); _mm_store_ps(buf + j + 8, _mm_mul_ps(_mm_cvtepi32_ps(_mm_unpacklo_epi16(v_src_p, v_zero)), v_scale_inv)); _mm_store_ps(buf + j + 12, _mm_mul_ps(_mm_cvtepi32_ps(_mm_unpackhi_epi16(v_src_p, v_zero)), v_scale_inv)); } int jr = j % 3; if (jr) src -= jr, j -= jr; } #endif for( ; j < dn*3; j += 3, src += scn ) { buf[j] = src[0]*(1.f/255.f); buf[j+1] = src[1]*(1.f/255.f); buf[j+2] = src[2]*(1.f/255.f); } cvt(buf, buf, dn); j = 0; #if CV_NEON for ( ; j <= (dn - 8) * 3; j += 24) { float32x4x3_t v_src0 = vld3q_f32(buf + j), v_src1 = vld3q_f32(buf + j + 12); uint8x8x3_t v_dst; v_dst.val[0] = vqmovn_u16(vcombine_u16(vqmovn_u32(cv_vrndq_u32_f32(v_src0.val[0])), vqmovn_u32(cv_vrndq_u32_f32(v_src1.val[0])))); v_dst.val[1] = vqmovn_u16(vcombine_u16(vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src0.val[1], v_scale))), vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src1.val[1], v_scale))))); v_dst.val[2] = vqmovn_u16(vcombine_u16(vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src0.val[2], v_scale))), vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src1.val[2], v_scale))))); vst3_u8(dst + j, v_dst); } #elif CV_SSE2 if (haveSIMD) { for ( ; j <= (dn - 32) * 3; j += 96) { __m128i v_h_0, v_l_0, v_s_0; process(buf + j, v_h_0, v_l_0, v_s_0); __m128i v_h_1, v_l_1, v_s_1; process(buf + j + 24, v_h_1, v_l_1, v_s_1); __m128i v_h0 = _mm_packus_epi16(v_h_0, v_h_1); __m128i v_l0 = _mm_packus_epi16(v_l_0, v_l_1); __m128i v_s0 = _mm_packus_epi16(v_s_0, v_s_1); process(buf + j + 48, v_h_0, v_l_0, v_s_0); process(buf + j + 72, v_h_1, v_l_1, v_s_1); __m128i v_h1 = _mm_packus_epi16(v_h_0, v_h_1); __m128i v_l1 = _mm_packus_epi16(v_l_0, v_l_1); __m128i v_s1 = _mm_packus_epi16(v_s_0, v_s_1); _mm_interleave_epi8(v_h0, v_h1, v_l0, v_l1, v_s0, v_s1); _mm_storeu_si128((__m128i *)(dst + j), v_h0); _mm_storeu_si128((__m128i *)(dst + j + 16), v_h1); _mm_storeu_si128((__m128i *)(dst + j + 32), v_l0); _mm_storeu_si128((__m128i *)(dst + j + 48), v_l1); _mm_storeu_si128((__m128i *)(dst + j + 64), v_s0); _mm_storeu_si128((__m128i *)(dst + j + 80), v_s1); } } #endif for( ; j < dn*3; j += 3 ) { dst[j] = saturate_cast(buf[j]); dst[j+1] = saturate_cast(buf[j+1]*255.f); dst[j+2] = saturate_cast(buf[j+2]*255.f); } } } int srccn; RGB2HLS_f cvt; #if CV_NEON float32x4_t v_scale, v_scale_inv; uint8x8_t v_alpha; #elif CV_SSE2 __m128 v_scale, v_scale_inv; __m128i v_zero; bool haveSIMD; #endif }; struct HLS2RGB_f { typedef float channel_type; HLS2RGB_f(int _dstcn, int _blueIdx, float _hrange) : dstcn(_dstcn), blueIdx(_blueIdx), hscale(6.f/_hrange) {} void operator()(const float* src, float* dst, int n) const { int i, bidx = blueIdx, dcn = dstcn; float _hscale = hscale; float alpha = ColorChannel::max(); n *= 3; for( i = 0; i < n; i += 3, dst += dcn ) { float h = src[i], l = src[i+1], s = src[i+2]; float b, g, r; if( s == 0 ) b = g = r = l; else { static const int sector_data[][3]= {{1,3,0}, {1,0,2}, {3,0,1}, {0,2,1}, {0,1,3}, {2,1,0}}; float tab[4]; int sector; float p2 = l <= 0.5f ? l*(1 + s) : l + s - l*s; float p1 = 2*l - p2; h *= _hscale; if( h < 0 ) do h += 6; while( h < 0 ); else if( h >= 6 ) do h -= 6; while( h >= 6 ); assert( 0 <= h && h < 6 ); sector = cvFloor(h); h -= sector; tab[0] = p2; tab[1] = p1; tab[2] = p1 + (p2 - p1)*(1-h); tab[3] = p1 + (p2 - p1)*h; b = tab[sector_data[sector][0]]; g = tab[sector_data[sector][1]]; r = tab[sector_data[sector][2]]; } dst[bidx] = b; dst[1] = g; dst[bidx^2] = r; if( dcn == 4 ) dst[3] = alpha; } } int dstcn, blueIdx; float hscale; }; struct HLS2RGB_b { typedef uchar channel_type; HLS2RGB_b(int _dstcn, int _blueIdx, int _hrange) : dstcn(_dstcn), cvt(3, _blueIdx, (float)_hrange) { #if CV_NEON v_scale_inv = vdupq_n_f32(1.f/255.f); v_scale = vdupq_n_f32(255.f); v_alpha = vdup_n_u8(ColorChannel::max()); #elif CV_SSE2 v_scale_inv = _mm_set1_ps(1.f/255.f); v_scale = _mm_set1_ps(255.f); v_zero = _mm_setzero_si128(); haveSIMD = checkHardwareSupport(CV_CPU_SSE2); #endif } #if CV_SSE2 // 16s x 8 void process(__m128i v_r, __m128i v_g, __m128i v_b, float * buf) const { __m128 v_r0 = _mm_cvtepi32_ps(_mm_unpacklo_epi16(v_r, v_zero)); __m128 v_g0 = _mm_cvtepi32_ps(_mm_unpacklo_epi16(v_g, v_zero)); __m128 v_b0 = _mm_cvtepi32_ps(_mm_unpacklo_epi16(v_b, v_zero)); __m128 v_r1 = _mm_cvtepi32_ps(_mm_unpackhi_epi16(v_r, v_zero)); __m128 v_g1 = _mm_cvtepi32_ps(_mm_unpackhi_epi16(v_g, v_zero)); __m128 v_b1 = _mm_cvtepi32_ps(_mm_unpackhi_epi16(v_b, v_zero)); v_g0 = _mm_mul_ps(v_g0, v_scale_inv); v_b0 = _mm_mul_ps(v_b0, v_scale_inv); v_g1 = _mm_mul_ps(v_g1, v_scale_inv); v_b1 = _mm_mul_ps(v_b1, v_scale_inv); _mm_interleave_ps(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1); _mm_store_ps(buf, v_r0); _mm_store_ps(buf + 4, v_r1); _mm_store_ps(buf + 8, v_g0); _mm_store_ps(buf + 12, v_g1); _mm_store_ps(buf + 16, v_b0); _mm_store_ps(buf + 20, v_b1); } #endif void operator()(const uchar* src, uchar* dst, int n) const { int i, j, dcn = dstcn; uchar alpha = ColorChannel::max(); float CV_DECL_ALIGNED(16) buf[3*BLOCK_SIZE]; for( i = 0; i < n; i += BLOCK_SIZE, src += BLOCK_SIZE*3 ) { int dn = std::min(n - i, (int)BLOCK_SIZE); j = 0; #if CV_NEON for ( ; j <= (dn - 8) * 3; j += 24) { uint8x8x3_t v_src = vld3_u8(src + j); uint16x8_t v_t0 = vmovl_u8(v_src.val[0]), v_t1 = vmovl_u8(v_src.val[1]), v_t2 = vmovl_u8(v_src.val[2]); float32x4x3_t v_dst; v_dst.val[0] = vcvtq_f32_u32(vmovl_u16(vget_low_u16(v_t0))); v_dst.val[1] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_low_u16(v_t1))), v_scale_inv); v_dst.val[2] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_low_u16(v_t2))), v_scale_inv); vst3q_f32(buf + j, v_dst); v_dst.val[0] = vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_t0))); v_dst.val[1] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_t1))), v_scale_inv); v_dst.val[2] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_t2))), v_scale_inv); vst3q_f32(buf + j + 12, v_dst); } #elif CV_SSE2 if (haveSIMD) { for ( ; j <= (dn - 32) * 3; j += 96) { __m128i v_r0 = _mm_loadu_si128((__m128i const *)(src + j)); __m128i v_r1 = _mm_loadu_si128((__m128i const *)(src + j + 16)); __m128i v_g0 = _mm_loadu_si128((__m128i const *)(src + j + 32)); __m128i v_g1 = _mm_loadu_si128((__m128i const *)(src + j + 48)); __m128i v_b0 = _mm_loadu_si128((__m128i const *)(src + j + 64)); __m128i v_b1 = _mm_loadu_si128((__m128i const *)(src + j + 80)); _mm_deinterleave_epi8(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1); process(_mm_unpacklo_epi8(v_r0, v_zero), _mm_unpacklo_epi8(v_g0, v_zero), _mm_unpacklo_epi8(v_b0, v_zero), buf + j); process(_mm_unpackhi_epi8(v_r0, v_zero), _mm_unpackhi_epi8(v_g0, v_zero), _mm_unpackhi_epi8(v_b0, v_zero), buf + j + 24); process(_mm_unpacklo_epi8(v_r1, v_zero), _mm_unpacklo_epi8(v_g1, v_zero), _mm_unpacklo_epi8(v_b1, v_zero), buf + j + 48); process(_mm_unpackhi_epi8(v_r1, v_zero), _mm_unpackhi_epi8(v_g1, v_zero), _mm_unpackhi_epi8(v_b1, v_zero), buf + j + 72); } } #endif for( ; j < dn*3; j += 3 ) { buf[j] = src[j]; buf[j+1] = src[j+1]*(1.f/255.f); buf[j+2] = src[j+2]*(1.f/255.f); } cvt(buf, buf, dn); j = 0; #if CV_NEON for ( ; j <= (dn - 8) * 3; j += 24, dst += dcn * 8) { float32x4x3_t v_src0 = vld3q_f32(buf + j), v_src1 = vld3q_f32(buf + j + 12); uint8x8_t v_dst0 = vqmovn_u16(vcombine_u16(vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src0.val[0], v_scale))), vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src1.val[0], v_scale))))); uint8x8_t v_dst1 = vqmovn_u16(vcombine_u16(vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src0.val[1], v_scale))), vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src1.val[1], v_scale))))); uint8x8_t v_dst2 = vqmovn_u16(vcombine_u16(vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src0.val[2], v_scale))), vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src1.val[2], v_scale))))); if (dcn == 4) { uint8x8x4_t v_dst; v_dst.val[0] = v_dst0; v_dst.val[1] = v_dst1; v_dst.val[2] = v_dst2; v_dst.val[3] = v_alpha; vst4_u8(dst, v_dst); } else { uint8x8x3_t v_dst; v_dst.val[0] = v_dst0; v_dst.val[1] = v_dst1; v_dst.val[2] = v_dst2; vst3_u8(dst, v_dst); } } #elif CV_SSE2 if (dcn == 3 && haveSIMD) { for ( ; j <= (dn * 3 - 16); j += 16, dst += 16) { __m128 v_src0 = _mm_mul_ps(_mm_load_ps(buf + j), v_scale); __m128 v_src1 = _mm_mul_ps(_mm_load_ps(buf + j + 4), v_scale); __m128 v_src2 = _mm_mul_ps(_mm_load_ps(buf + j + 8), v_scale); __m128 v_src3 = _mm_mul_ps(_mm_load_ps(buf + j + 12), v_scale); __m128i v_dst0 = _mm_packs_epi32(_mm_cvtps_epi32(v_src0), _mm_cvtps_epi32(v_src1)); __m128i v_dst1 = _mm_packs_epi32(_mm_cvtps_epi32(v_src2), _mm_cvtps_epi32(v_src3)); _mm_storeu_si128((__m128i *)dst, _mm_packus_epi16(v_dst0, v_dst1)); } int jr = j % 3; if (jr) dst -= jr, j -= jr; } #endif for( ; j < dn*3; j += 3, dst += dcn ) { dst[0] = saturate_cast(buf[j]*255.f); dst[1] = saturate_cast(buf[j+1]*255.f); dst[2] = saturate_cast(buf[j+2]*255.f); if( dcn == 4 ) dst[3] = alpha; } } } int dstcn; HLS2RGB_f cvt; #if CV_NEON float32x4_t v_scale, v_scale_inv; uint8x8_t v_alpha; #elif CV_SSE2 __m128 v_scale, v_scale_inv; __m128i v_zero; bool haveSIMD; #endif }; ///////////////////////////////////// RGB <-> L*a*b* ///////////////////////////////////// static const float D65[] = { 0.950456f, 1.f, 1.088754f }; enum { LAB_CBRT_TAB_SIZE = 1024, GAMMA_TAB_SIZE = 1024 }; static float LabCbrtTab[LAB_CBRT_TAB_SIZE*4]; static const float LabCbrtTabScale = LAB_CBRT_TAB_SIZE/1.5f; static float sRGBGammaTab[GAMMA_TAB_SIZE*4], sRGBInvGammaTab[GAMMA_TAB_SIZE*4]; static const float GammaTabScale = (float)GAMMA_TAB_SIZE; static ushort sRGBGammaTab_b[256], linearGammaTab_b[256]; #undef lab_shift #define lab_shift xyz_shift #define gamma_shift 3 #define lab_shift2 (lab_shift + gamma_shift) #define LAB_CBRT_TAB_SIZE_B (256*3/2*(1<(255.f*(1 << gamma_shift)*(x <= 0.04045f ? x*(1.f/12.92f) : (float)std::pow((double)(x + 0.055)*(1./1.055), 2.4))); linearGammaTab_b[i] = (ushort)(i*(1 << gamma_shift)); } for(i = 0; i < LAB_CBRT_TAB_SIZE_B; i++) { float x = i*(1.f/(255.f*(1 << gamma_shift))); LabCbrtTab_b[i] = saturate_cast((1 << lab_shift2)*(x < 0.008856f ? x*7.787f + 0.13793103448275862f : cvCbrt(x))); } initialized = true; } } struct RGB2Lab_b { typedef uchar channel_type; RGB2Lab_b(int _srccn, int blueIdx, const float* _coeffs, const float* _whitept, bool _srgb) : srccn(_srccn), srgb(_srgb) { static volatile int _3 = 3; initLabTabs(); if (!_coeffs) _coeffs = sRGB2XYZ_D65; if (!_whitept) _whitept = D65; float scale[] = { (1 << lab_shift)/_whitept[0], (float)(1 << lab_shift), (1 << lab_shift)/_whitept[2] }; for( int i = 0; i < _3; i++ ) { coeffs[i*3+(blueIdx^2)] = cvRound(_coeffs[i*3]*scale[i]); coeffs[i*3+1] = cvRound(_coeffs[i*3+1]*scale[i]); coeffs[i*3+blueIdx] = cvRound(_coeffs[i*3+2]*scale[i]); CV_Assert( coeffs[i] >= 0 && coeffs[i*3+1] >= 0 && coeffs[i*3+2] >= 0 && coeffs[i*3] + coeffs[i*3+1] + coeffs[i*3+2] < 2*(1 << lab_shift) ); } } void operator()(const uchar* src, uchar* dst, int n) const { const int Lscale = (116*255+50)/100; const int Lshift = -((16*255*(1 << lab_shift2) + 50)/100); const ushort* tab = srgb ? sRGBGammaTab_b : linearGammaTab_b; int i, scn = srccn; int C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4], C5 = coeffs[5], C6 = coeffs[6], C7 = coeffs[7], C8 = coeffs[8]; n *= 3; for( i = 0; i < n; i += 3, src += scn ) { int R = tab[src[0]], G = tab[src[1]], B = tab[src[2]]; int fX = LabCbrtTab_b[CV_DESCALE(R*C0 + G*C1 + B*C2, lab_shift)]; int fY = LabCbrtTab_b[CV_DESCALE(R*C3 + G*C4 + B*C5, lab_shift)]; int fZ = LabCbrtTab_b[CV_DESCALE(R*C6 + G*C7 + B*C8, lab_shift)]; int L = CV_DESCALE( Lscale*fY + Lshift, lab_shift2 ); int a = CV_DESCALE( 500*(fX - fY) + 128*(1 << lab_shift2), lab_shift2 ); int b = CV_DESCALE( 200*(fY - fZ) + 128*(1 << lab_shift2), lab_shift2 ); dst[i] = saturate_cast(L); dst[i+1] = saturate_cast(a); dst[i+2] = saturate_cast(b); } } int srccn; int coeffs[9]; bool srgb; }; #define clip(value) \ value < 0.0f ? 0.0f : value > 1.0f ? 1.0f : value; struct RGB2Lab_f { typedef float channel_type; RGB2Lab_f(int _srccn, int blueIdx, const float* _coeffs, const float* _whitept, bool _srgb) : srccn(_srccn), srgb(_srgb) { volatile int _3 = 3; initLabTabs(); if (!_coeffs) _coeffs = sRGB2XYZ_D65; if (!_whitept) _whitept = D65; float scale[] = { 1.0f / _whitept[0], 1.0f, 1.0f / _whitept[2] }; for( int i = 0; i < _3; i++ ) { int j = i * 3; coeffs[j + (blueIdx ^ 2)] = _coeffs[j] * scale[i]; coeffs[j + 1] = _coeffs[j + 1] * scale[i]; coeffs[j + blueIdx] = _coeffs[j + 2] * scale[i]; CV_Assert( coeffs[j] >= 0 && coeffs[j + 1] >= 0 && coeffs[j + 2] >= 0 && coeffs[j] + coeffs[j + 1] + coeffs[j + 2] < 1.5f*LabCbrtTabScale ); } } void operator()(const float* src, float* dst, int n) const { int i, scn = srccn; float gscale = GammaTabScale; const float* gammaTab = srgb ? sRGBGammaTab : 0; float C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4], C5 = coeffs[5], C6 = coeffs[6], C7 = coeffs[7], C8 = coeffs[8]; n *= 3; static const float _1_3 = 1.0f / 3.0f; static const float _a = 16.0f / 116.0f; for (i = 0; i < n; i += 3, src += scn ) { float R = clip(src[0]); float G = clip(src[1]); float B = clip(src[2]); if (gammaTab) { R = splineInterpolate(R * gscale, gammaTab, GAMMA_TAB_SIZE); G = splineInterpolate(G * gscale, gammaTab, GAMMA_TAB_SIZE); B = splineInterpolate(B * gscale, gammaTab, GAMMA_TAB_SIZE); } float X = R*C0 + G*C1 + B*C2; float Y = R*C3 + G*C4 + B*C5; float Z = R*C6 + G*C7 + B*C8; float FX = X > 0.008856f ? std::pow(X, _1_3) : (7.787f * X + _a); float FY = Y > 0.008856f ? std::pow(Y, _1_3) : (7.787f * Y + _a); float FZ = Z > 0.008856f ? std::pow(Z, _1_3) : (7.787f * Z + _a); float L = Y > 0.008856f ? (116.f * FY - 16.f) : (903.3f * Y); float a = 500.f * (FX - FY); float b = 200.f * (FY - FZ); dst[i] = L; dst[i + 1] = a; dst[i + 2] = b; } } int srccn; float coeffs[9]; bool srgb; }; struct Lab2RGB_f { typedef float channel_type; Lab2RGB_f( int _dstcn, int blueIdx, const float* _coeffs, const float* _whitept, bool _srgb ) : dstcn(_dstcn), srgb(_srgb) { initLabTabs(); if(!_coeffs) _coeffs = XYZ2sRGB_D65; if(!_whitept) _whitept = D65; for( int i = 0; i < 3; i++ ) { coeffs[i+(blueIdx^2)*3] = _coeffs[i]*_whitept[i]; coeffs[i+3] = _coeffs[i+3]*_whitept[i]; coeffs[i+blueIdx*3] = _coeffs[i+6]*_whitept[i]; } } void operator()(const float* src, float* dst, int n) const { int i, dcn = dstcn; const float* gammaTab = srgb ? sRGBInvGammaTab : 0; float gscale = GammaTabScale; float C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4], C5 = coeffs[5], C6 = coeffs[6], C7 = coeffs[7], C8 = coeffs[8]; float alpha = ColorChannel::max(); n *= 3; static const float lThresh = 0.008856f * 903.3f; static const float fThresh = 7.787f * 0.008856f + 16.0f / 116.0f; for (i = 0; i < n; i += 3, dst += dcn) { float li = src[i]; float ai = src[i + 1]; float bi = src[i + 2]; float y, fy; if (li <= lThresh) { y = li / 903.3f; fy = 7.787f * y + 16.0f / 116.0f; } else { fy = (li + 16.0f) / 116.0f; y = fy * fy * fy; } float fxz[] = { ai / 500.0f + fy, fy - bi / 200.0f }; for (int j = 0; j < 2; j++) if (fxz[j] <= fThresh) fxz[j] = (fxz[j] - 16.0f / 116.0f) / 7.787f; else fxz[j] = fxz[j] * fxz[j] * fxz[j]; float x = fxz[0], z = fxz[1]; float ro = C0 * x + C1 * y + C2 * z; float go = C3 * x + C4 * y + C5 * z; float bo = C6 * x + C7 * y + C8 * z; ro = clip(ro); go = clip(go); bo = clip(bo); if (gammaTab) { ro = splineInterpolate(ro * gscale, gammaTab, GAMMA_TAB_SIZE); go = splineInterpolate(go * gscale, gammaTab, GAMMA_TAB_SIZE); bo = splineInterpolate(bo * gscale, gammaTab, GAMMA_TAB_SIZE); } dst[0] = ro, dst[1] = go, dst[2] = bo; if( dcn == 4 ) dst[3] = alpha; } } int dstcn; float coeffs[9]; bool srgb; }; #undef clip struct Lab2RGB_b { typedef uchar channel_type; Lab2RGB_b( int _dstcn, int blueIdx, const float* _coeffs, const float* _whitept, bool _srgb ) : dstcn(_dstcn), cvt(3, blueIdx, _coeffs, _whitept, _srgb ) { #if CV_NEON v_scale_inv = vdupq_n_f32(100.f/255.f); v_scale = vdupq_n_f32(255.f); v_alpha = vdup_n_u8(ColorChannel::max()); v_128 = vdupq_n_f32(128.0f); #elif CV_SSE2 v_scale_inv = _mm_set1_ps(100.f/255.f); v_scale = _mm_set1_ps(255.f); v_128 = _mm_set1_ps(128.0f); v_zero = _mm_setzero_si128(); haveSIMD = checkHardwareSupport(CV_CPU_SSE2); #endif } #if CV_SSE2 // 16s x 8 void process(__m128i v_r, __m128i v_g, __m128i v_b, float * buf) const { __m128 v_r0 = _mm_cvtepi32_ps(_mm_unpacklo_epi16(v_r, v_zero)); __m128 v_g0 = _mm_cvtepi32_ps(_mm_unpacklo_epi16(v_g, v_zero)); __m128 v_b0 = _mm_cvtepi32_ps(_mm_unpacklo_epi16(v_b, v_zero)); __m128 v_r1 = _mm_cvtepi32_ps(_mm_unpackhi_epi16(v_r, v_zero)); __m128 v_g1 = _mm_cvtepi32_ps(_mm_unpackhi_epi16(v_g, v_zero)); __m128 v_b1 = _mm_cvtepi32_ps(_mm_unpackhi_epi16(v_b, v_zero)); v_r0 = _mm_mul_ps(v_r0, v_scale_inv); v_r1 = _mm_mul_ps(v_r1, v_scale_inv); v_g0 = _mm_sub_ps(v_g0, v_128); v_g1 = _mm_sub_ps(v_g1, v_128); v_b0 = _mm_sub_ps(v_b0, v_128); v_b1 = _mm_sub_ps(v_b1, v_128); _mm_interleave_ps(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1); _mm_store_ps(buf, v_r0); _mm_store_ps(buf + 4, v_r1); _mm_store_ps(buf + 8, v_g0); _mm_store_ps(buf + 12, v_g1); _mm_store_ps(buf + 16, v_b0); _mm_store_ps(buf + 20, v_b1); } #endif void operator()(const uchar* src, uchar* dst, int n) const { int i, j, dcn = dstcn; uchar alpha = ColorChannel::max(); float CV_DECL_ALIGNED(16) buf[3*BLOCK_SIZE]; for( i = 0; i < n; i += BLOCK_SIZE, src += BLOCK_SIZE*3 ) { int dn = std::min(n - i, (int)BLOCK_SIZE); j = 0; #if CV_NEON for ( ; j <= (dn - 8) * 3; j += 24) { uint8x8x3_t v_src = vld3_u8(src + j); uint16x8_t v_t0 = vmovl_u8(v_src.val[0]), v_t1 = vmovl_u8(v_src.val[1]), v_t2 = vmovl_u8(v_src.val[2]); float32x4x3_t v_dst; v_dst.val[0] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_low_u16(v_t0))), v_scale_inv); v_dst.val[1] = vsubq_f32(vcvtq_f32_u32(vmovl_u16(vget_low_u16(v_t1))), v_128); v_dst.val[2] = vsubq_f32(vcvtq_f32_u32(vmovl_u16(vget_low_u16(v_t2))), v_128); vst3q_f32(buf + j, v_dst); v_dst.val[0] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_t0))), v_scale_inv); v_dst.val[1] = vsubq_f32(vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_t1))), v_128); v_dst.val[2] = vsubq_f32(vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_t2))), v_128); vst3q_f32(buf + j + 12, v_dst); } #elif CV_SSE2 if (haveSIMD) { for ( ; j <= (dn - 32) * 3; j += 96) { __m128i v_r0 = _mm_loadu_si128((__m128i const *)(src + j)); __m128i v_r1 = _mm_loadu_si128((__m128i const *)(src + j + 16)); __m128i v_g0 = _mm_loadu_si128((__m128i const *)(src + j + 32)); __m128i v_g1 = _mm_loadu_si128((__m128i const *)(src + j + 48)); __m128i v_b0 = _mm_loadu_si128((__m128i const *)(src + j + 64)); __m128i v_b1 = _mm_loadu_si128((__m128i const *)(src + j + 80)); _mm_deinterleave_epi8(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1); process(_mm_unpacklo_epi8(v_r0, v_zero), _mm_unpacklo_epi8(v_g0, v_zero), _mm_unpacklo_epi8(v_b0, v_zero), buf + j); process(_mm_unpackhi_epi8(v_r0, v_zero), _mm_unpackhi_epi8(v_g0, v_zero), _mm_unpackhi_epi8(v_b0, v_zero), buf + j + 24); process(_mm_unpacklo_epi8(v_r1, v_zero), _mm_unpacklo_epi8(v_g1, v_zero), _mm_unpacklo_epi8(v_b1, v_zero), buf + j + 48); process(_mm_unpackhi_epi8(v_r1, v_zero), _mm_unpackhi_epi8(v_g1, v_zero), _mm_unpackhi_epi8(v_b1, v_zero), buf + j + 72); } } #endif for( ; j < dn*3; j += 3 ) { buf[j] = src[j]*(100.f/255.f); buf[j+1] = (float)(src[j+1] - 128); buf[j+2] = (float)(src[j+2] - 128); } cvt(buf, buf, dn); j = 0; #if CV_NEON for ( ; j <= (dn - 8) * 3; j += 24, dst += dcn * 8) { float32x4x3_t v_src0 = vld3q_f32(buf + j), v_src1 = vld3q_f32(buf + j + 12); uint8x8_t v_dst0 = vqmovn_u16(vcombine_u16(vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src0.val[0], v_scale))), vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src1.val[0], v_scale))))); uint8x8_t v_dst1 = vqmovn_u16(vcombine_u16(vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src0.val[1], v_scale))), vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src1.val[1], v_scale))))); uint8x8_t v_dst2 = vqmovn_u16(vcombine_u16(vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src0.val[2], v_scale))), vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src1.val[2], v_scale))))); if (dcn == 4) { uint8x8x4_t v_dst; v_dst.val[0] = v_dst0; v_dst.val[1] = v_dst1; v_dst.val[2] = v_dst2; v_dst.val[3] = v_alpha; vst4_u8(dst, v_dst); } else { uint8x8x3_t v_dst; v_dst.val[0] = v_dst0; v_dst.val[1] = v_dst1; v_dst.val[2] = v_dst2; vst3_u8(dst, v_dst); } } #elif CV_SSE2 if (dcn == 3 && haveSIMD) { for ( ; j <= (dn * 3 - 16); j += 16, dst += 16) { __m128 v_src0 = _mm_mul_ps(_mm_load_ps(buf + j), v_scale); __m128 v_src1 = _mm_mul_ps(_mm_load_ps(buf + j + 4), v_scale); __m128 v_src2 = _mm_mul_ps(_mm_load_ps(buf + j + 8), v_scale); __m128 v_src3 = _mm_mul_ps(_mm_load_ps(buf + j + 12), v_scale); __m128i v_dst0 = _mm_packs_epi32(_mm_cvtps_epi32(v_src0), _mm_cvtps_epi32(v_src1)); __m128i v_dst1 = _mm_packs_epi32(_mm_cvtps_epi32(v_src2), _mm_cvtps_epi32(v_src3)); _mm_storeu_si128((__m128i *)dst, _mm_packus_epi16(v_dst0, v_dst1)); } int jr = j % 3; if (jr) dst -= jr, j -= jr; } #endif for( ; j < dn*3; j += 3, dst += dcn ) { dst[0] = saturate_cast(buf[j]*255.f); dst[1] = saturate_cast(buf[j+1]*255.f); dst[2] = saturate_cast(buf[j+2]*255.f); if( dcn == 4 ) dst[3] = alpha; } } } int dstcn; Lab2RGB_f cvt; #if CV_NEON float32x4_t v_scale, v_scale_inv, v_128; uint8x8_t v_alpha; #elif CV_SSE2 __m128 v_scale, v_scale_inv, v_128; __m128i v_zero; bool haveSIMD; #endif }; ///////////////////////////////////// RGB <-> L*u*v* ///////////////////////////////////// struct RGB2Luv_f { typedef float channel_type; RGB2Luv_f( int _srccn, int blueIdx, const float* _coeffs, const float* whitept, bool _srgb ) : srccn(_srccn), srgb(_srgb) { volatile int i; initLabTabs(); if(!_coeffs) _coeffs = sRGB2XYZ_D65; if(!whitept) whitept = D65; for( i = 0; i < 3; i++ ) { coeffs[i*3] = _coeffs[i*3]; coeffs[i*3+1] = _coeffs[i*3+1]; coeffs[i*3+2] = _coeffs[i*3+2]; if( blueIdx == 0 ) std::swap(coeffs[i*3], coeffs[i*3+2]); CV_Assert( coeffs[i*3] >= 0 && coeffs[i*3+1] >= 0 && coeffs[i*3+2] >= 0 && coeffs[i*3] + coeffs[i*3+1] + coeffs[i*3+2] < 1.5f ); } float d = 1.f/(whitept[0] + whitept[1]*15 + whitept[2]*3); un = 4*whitept[0]*d; vn = 9*whitept[1]*d; CV_Assert(whitept[1] == 1.f); } void operator()(const float* src, float* dst, int n) const { int i, scn = srccn; float gscale = GammaTabScale; const float* gammaTab = srgb ? sRGBGammaTab : 0; float C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4], C5 = coeffs[5], C6 = coeffs[6], C7 = coeffs[7], C8 = coeffs[8]; float _un = 13*un, _vn = 13*vn; n *= 3; for( i = 0; i < n; i += 3, src += scn ) { float R = src[0], G = src[1], B = src[2]; if( gammaTab ) { R = splineInterpolate(R*gscale, gammaTab, GAMMA_TAB_SIZE); G = splineInterpolate(G*gscale, gammaTab, GAMMA_TAB_SIZE); B = splineInterpolate(B*gscale, gammaTab, GAMMA_TAB_SIZE); } float X = R*C0 + G*C1 + B*C2; float Y = R*C3 + G*C4 + B*C5; float Z = R*C6 + G*C7 + B*C8; float L = splineInterpolate(Y*LabCbrtTabScale, LabCbrtTab, LAB_CBRT_TAB_SIZE); L = 116.f*L - 16.f; float d = (4*13) / std::max(X + 15 * Y + 3 * Z, FLT_EPSILON); float u = L*(X*d - _un); float v = L*((9*0.25f)*Y*d - _vn); dst[i] = L; dst[i+1] = u; dst[i+2] = v; } } int srccn; float coeffs[9], un, vn; bool srgb; }; struct Luv2RGB_f { typedef float channel_type; Luv2RGB_f( int _dstcn, int blueIdx, const float* _coeffs, const float* whitept, bool _srgb ) : dstcn(_dstcn), srgb(_srgb) { initLabTabs(); if(!_coeffs) _coeffs = XYZ2sRGB_D65; if(!whitept) whitept = D65; for( int i = 0; i < 3; i++ ) { coeffs[i+(blueIdx^2)*3] = _coeffs[i]; coeffs[i+3] = _coeffs[i+3]; coeffs[i+blueIdx*3] = _coeffs[i+6]; } float d = 1.f/(whitept[0] + whitept[1]*15 + whitept[2]*3); un = 4*whitept[0]*d; vn = 9*whitept[1]*d; CV_Assert(whitept[1] == 1.f); } void operator()(const float* src, float* dst, int n) const { int i, dcn = dstcn; const float* gammaTab = srgb ? sRGBInvGammaTab : 0; float gscale = GammaTabScale; float C0 = coeffs[0], C1 = coeffs[1], C2 = coeffs[2], C3 = coeffs[3], C4 = coeffs[4], C5 = coeffs[5], C6 = coeffs[6], C7 = coeffs[7], C8 = coeffs[8]; float alpha = ColorChannel::max(); float _un = un, _vn = vn; n *= 3; for( i = 0; i < n; i += 3, dst += dcn ) { float L = src[i], u = src[i+1], v = src[i+2], d, X, Y, Z; Y = (L + 16.f) * (1.f/116.f); Y = Y*Y*Y; d = (1.f/13.f)/L; u = u*d + _un; v = v*d + _vn; float iv = 1.f/v; X = 2.25f * u * Y * iv ; Z = (12 - 3 * u - 20 * v) * Y * 0.25f * iv; float R = X*C0 + Y*C1 + Z*C2; float G = X*C3 + Y*C4 + Z*C5; float B = X*C6 + Y*C7 + Z*C8; R = std::min(std::max(R, 0.f), 1.f); G = std::min(std::max(G, 0.f), 1.f); B = std::min(std::max(B, 0.f), 1.f); if( gammaTab ) { R = splineInterpolate(R*gscale, gammaTab, GAMMA_TAB_SIZE); G = splineInterpolate(G*gscale, gammaTab, GAMMA_TAB_SIZE); B = splineInterpolate(B*gscale, gammaTab, GAMMA_TAB_SIZE); } dst[0] = R; dst[1] = G; dst[2] = B; if( dcn == 4 ) dst[3] = alpha; } } int dstcn; float coeffs[9], un, vn; bool srgb; }; struct RGB2Luv_b { typedef uchar channel_type; RGB2Luv_b( int _srccn, int blueIdx, const float* _coeffs, const float* _whitept, bool _srgb ) : srccn(_srccn), cvt(3, blueIdx, _coeffs, _whitept, _srgb) { #if CV_NEON v_scale_inv = vdupq_n_f32(1.f/255.f); v_scale = vdupq_n_f32(2.55f); v_coeff1 = vdupq_n_f32(0.72033898305084743f); v_coeff2 = vdupq_n_f32(96.525423728813564f); v_coeff3 = vdupq_n_f32(0.9732824427480916f); v_coeff4 = vdupq_n_f32(136.259541984732824f); v_alpha = vdup_n_u8(ColorChannel::max()); #elif CV_SSE2 v_zero = _mm_setzero_si128(); v_scale_inv = _mm_set1_ps(1.f/255.f); v_scale = _mm_set1_ps(2.55f); v_coeff1 = _mm_set1_ps(0.72033898305084743f); v_coeff2 = _mm_set1_ps(96.525423728813564f); v_coeff3 = _mm_set1_ps(0.9732824427480916f); v_coeff4 = _mm_set1_ps(136.259541984732824f); haveSIMD = checkHardwareSupport(CV_CPU_SSE2); #endif } #if CV_SSE2 void process(const float * buf, __m128i & v_l, __m128i & v_u, __m128i & v_v) const { __m128 v_l0f = _mm_load_ps(buf); __m128 v_l1f = _mm_load_ps(buf + 4); __m128 v_u0f = _mm_load_ps(buf + 8); __m128 v_u1f = _mm_load_ps(buf + 12); __m128 v_v0f = _mm_load_ps(buf + 16); __m128 v_v1f = _mm_load_ps(buf + 20); _mm_deinterleave_ps(v_l0f, v_l1f, v_u0f, v_u1f, v_v0f, v_v1f); v_l0f = _mm_mul_ps(v_l0f, v_scale); v_l1f = _mm_mul_ps(v_l1f, v_scale); v_u0f = _mm_add_ps(_mm_mul_ps(v_u0f, v_coeff1), v_coeff2); v_u1f = _mm_add_ps(_mm_mul_ps(v_u1f, v_coeff1), v_coeff2); v_v0f = _mm_add_ps(_mm_mul_ps(v_v0f, v_coeff3), v_coeff4); v_v1f = _mm_add_ps(_mm_mul_ps(v_v1f, v_coeff3), v_coeff4); v_l = _mm_packs_epi32(_mm_cvtps_epi32(v_l0f), _mm_cvtps_epi32(v_l1f)); v_u = _mm_packs_epi32(_mm_cvtps_epi32(v_u0f), _mm_cvtps_epi32(v_u1f)); v_v = _mm_packs_epi32(_mm_cvtps_epi32(v_v0f), _mm_cvtps_epi32(v_v1f)); } #endif void operator()(const uchar* src, uchar* dst, int n) const { int i, j, scn = srccn; float CV_DECL_ALIGNED(16) buf[3*BLOCK_SIZE]; for( i = 0; i < n; i += BLOCK_SIZE, dst += BLOCK_SIZE*3 ) { int dn = std::min(n - i, (int)BLOCK_SIZE); j = 0; #if CV_NEON for ( ; j <= (dn - 8) * 3; j += 24, src += 8 * scn) { uint16x8_t v_t0, v_t1, v_t2; if (scn == 3) { uint8x8x3_t v_src = vld3_u8(src); v_t0 = vmovl_u8(v_src.val[0]); v_t1 = vmovl_u8(v_src.val[1]); v_t2 = vmovl_u8(v_src.val[2]); } else { uint8x8x4_t v_src = vld4_u8(src); v_t0 = vmovl_u8(v_src.val[0]); v_t1 = vmovl_u8(v_src.val[1]); v_t2 = vmovl_u8(v_src.val[2]); } float32x4x3_t v_dst; v_dst.val[0] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_low_u16(v_t0))), v_scale_inv); v_dst.val[1] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_low_u16(v_t1))), v_scale_inv); v_dst.val[2] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_low_u16(v_t2))), v_scale_inv); vst3q_f32(buf + j, v_dst); v_dst.val[0] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_t0))), v_scale_inv); v_dst.val[1] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_t1))), v_scale_inv); v_dst.val[2] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_t2))), v_scale_inv); vst3q_f32(buf + j + 12, v_dst); } #elif CV_SSE2 if (scn == 3 && haveSIMD) { for ( ; j <= (dn * 3 - 16); j += 16, src += 16) { __m128i v_src = _mm_loadu_si128((__m128i const *)src); __m128i v_src_p = _mm_unpacklo_epi8(v_src, v_zero); _mm_store_ps(buf + j, _mm_mul_ps(_mm_cvtepi32_ps(_mm_unpacklo_epi16(v_src_p, v_zero)), v_scale_inv)); _mm_store_ps(buf + j + 4, _mm_mul_ps(_mm_cvtepi32_ps(_mm_unpackhi_epi16(v_src_p, v_zero)), v_scale_inv)); v_src_p = _mm_unpackhi_epi8(v_src, v_zero); _mm_store_ps(buf + j + 8, _mm_mul_ps(_mm_cvtepi32_ps(_mm_unpacklo_epi16(v_src_p, v_zero)), v_scale_inv)); _mm_store_ps(buf + j + 12, _mm_mul_ps(_mm_cvtepi32_ps(_mm_unpackhi_epi16(v_src_p, v_zero)), v_scale_inv)); } int jr = j % 3; if (jr) src -= jr, j -= jr; } #endif for( ; j < dn*3; j += 3, src += scn ) { buf[j] = src[0]*(1.f/255.f); buf[j+1] = (float)(src[1]*(1.f/255.f)); buf[j+2] = (float)(src[2]*(1.f/255.f)); } cvt(buf, buf, dn); j = 0; #if CV_NEON for ( ; j <= (dn - 8) * 3; j += 24) { float32x4x3_t v_src0 = vld3q_f32(buf + j), v_src1 = vld3q_f32(buf + j + 12); uint8x8x3_t v_dst; v_dst.val[0] = vqmovn_u16(vcombine_u16(vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src0.val[0], v_scale))), vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src1.val[0], v_scale))))); v_dst.val[1] = vqmovn_u16(vcombine_u16(vqmovn_u32(cv_vrndq_u32_f32(vaddq_f32(vmulq_f32(v_src0.val[1], v_coeff1), v_coeff2))), vqmovn_u32(cv_vrndq_u32_f32(vaddq_f32(vmulq_f32(v_src1.val[1], v_coeff1), v_coeff2))))); v_dst.val[2] = vqmovn_u16(vcombine_u16(vqmovn_u32(cv_vrndq_u32_f32(vaddq_f32(vmulq_f32(v_src0.val[2], v_coeff3), v_coeff4))), vqmovn_u32(cv_vrndq_u32_f32(vaddq_f32(vmulq_f32(v_src1.val[2], v_coeff3), v_coeff4))))); vst3_u8(dst + j, v_dst); } #elif CV_SSE2 if (haveSIMD) { for ( ; j <= (dn - 32) * 3; j += 96) { __m128i v_l_0, v_u_0, v_v_0; process(buf + j, v_l_0, v_u_0, v_v_0); __m128i v_l_1, v_u_1, v_v_1; process(buf + j + 24, v_l_1, v_u_1, v_v_1); __m128i v_l0 = _mm_packus_epi16(v_l_0, v_l_1); __m128i v_u0 = _mm_packus_epi16(v_u_0, v_u_1); __m128i v_v0 = _mm_packus_epi16(v_v_0, v_v_1); process(buf + j + 48, v_l_0, v_u_0, v_v_0); process(buf + j + 72, v_l_1, v_u_1, v_v_1); __m128i v_l1 = _mm_packus_epi16(v_l_0, v_l_1); __m128i v_u1 = _mm_packus_epi16(v_u_0, v_u_1); __m128i v_v1 = _mm_packus_epi16(v_v_0, v_v_1); _mm_interleave_epi8(v_l0, v_l1, v_u0, v_u1, v_v0, v_v1); _mm_storeu_si128((__m128i *)(dst + j), v_l0); _mm_storeu_si128((__m128i *)(dst + j + 16), v_l1); _mm_storeu_si128((__m128i *)(dst + j + 32), v_u0); _mm_storeu_si128((__m128i *)(dst + j + 48), v_u1); _mm_storeu_si128((__m128i *)(dst + j + 64), v_v0); _mm_storeu_si128((__m128i *)(dst + j + 80), v_v1); } } #endif for( ; j < dn*3; j += 3 ) { dst[j] = saturate_cast(buf[j]*2.55f); dst[j+1] = saturate_cast(buf[j+1]*0.72033898305084743f + 96.525423728813564f); dst[j+2] = saturate_cast(buf[j+2]*0.9732824427480916f + 136.259541984732824f); } } } int srccn; RGB2Luv_f cvt; #if CV_NEON float32x4_t v_scale, v_scale_inv, v_coeff1, v_coeff2, v_coeff3, v_coeff4; uint8x8_t v_alpha; #elif CV_SSE2 __m128 v_scale, v_scale_inv, v_coeff1, v_coeff2, v_coeff3, v_coeff4; __m128i v_zero; bool haveSIMD; #endif }; struct Luv2RGB_b { typedef uchar channel_type; Luv2RGB_b( int _dstcn, int blueIdx, const float* _coeffs, const float* _whitept, bool _srgb ) : dstcn(_dstcn), cvt(3, blueIdx, _coeffs, _whitept, _srgb ) { #if CV_NEON v_scale_inv = vdupq_n_f32(100.f/255.f); v_coeff1 = vdupq_n_f32(1.388235294117647f); v_coeff2 = vdupq_n_f32(1.027450980392157f); v_134 = vdupq_n_f32(134.f); v_140 = vdupq_n_f32(140.f); v_scale = vdupq_n_f32(255.f); v_alpha = vdup_n_u8(ColorChannel::max()); #elif CV_SSE2 v_scale_inv = _mm_set1_ps(100.f/255.f); v_coeff1 = _mm_set1_ps(1.388235294117647f); v_coeff2 = _mm_set1_ps(1.027450980392157f); v_134 = _mm_set1_ps(134.f); v_140 = _mm_set1_ps(140.f); v_scale = _mm_set1_ps(255.f); v_zero = _mm_setzero_si128(); haveSIMD = checkHardwareSupport(CV_CPU_SSE2); #endif } #if CV_SSE2 // 16s x 8 void process(__m128i v_l, __m128i v_u, __m128i v_v, float * buf) const { __m128 v_l0 = _mm_cvtepi32_ps(_mm_unpacklo_epi16(v_l, v_zero)); __m128 v_u0 = _mm_cvtepi32_ps(_mm_unpacklo_epi16(v_u, v_zero)); __m128 v_v0 = _mm_cvtepi32_ps(_mm_unpacklo_epi16(v_v, v_zero)); __m128 v_l1 = _mm_cvtepi32_ps(_mm_unpackhi_epi16(v_l, v_zero)); __m128 v_u1 = _mm_cvtepi32_ps(_mm_unpackhi_epi16(v_u, v_zero)); __m128 v_v1 = _mm_cvtepi32_ps(_mm_unpackhi_epi16(v_v, v_zero)); v_l0 = _mm_mul_ps(v_l0, v_scale_inv); v_l1 = _mm_mul_ps(v_l1, v_scale_inv); v_u0 = _mm_sub_ps(_mm_mul_ps(v_u0, v_coeff1), v_134); v_u1 = _mm_sub_ps(_mm_mul_ps(v_u1, v_coeff1), v_134); v_v0 = _mm_sub_ps(_mm_mul_ps(v_v0, v_coeff2), v_140); v_v1 = _mm_sub_ps(_mm_mul_ps(v_v1, v_coeff2), v_140); _mm_interleave_ps(v_l0, v_l1, v_u0, v_u1, v_v0, v_v1); _mm_store_ps(buf, v_l0); _mm_store_ps(buf + 4, v_l1); _mm_store_ps(buf + 8, v_u0); _mm_store_ps(buf + 12, v_u1); _mm_store_ps(buf + 16, v_v0); _mm_store_ps(buf + 20, v_v1); } #endif void operator()(const uchar* src, uchar* dst, int n) const { int i, j, dcn = dstcn; uchar alpha = ColorChannel::max(); float CV_DECL_ALIGNED(16) buf[3*BLOCK_SIZE]; for( i = 0; i < n; i += BLOCK_SIZE, src += BLOCK_SIZE*3 ) { int dn = std::min(n - i, (int)BLOCK_SIZE); j = 0; #if CV_NEON for ( ; j <= (dn - 8) * 3; j += 24) { uint8x8x3_t v_src = vld3_u8(src + j); uint16x8_t v_t0 = vmovl_u8(v_src.val[0]), v_t1 = vmovl_u8(v_src.val[1]), v_t2 = vmovl_u8(v_src.val[2]); float32x4x3_t v_dst; v_dst.val[0] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_low_u16(v_t0))), v_scale_inv); v_dst.val[1] = vsubq_f32(vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_low_u16(v_t1))), v_coeff1), v_134); v_dst.val[2] = vsubq_f32(vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_low_u16(v_t2))), v_coeff2), v_140); vst3q_f32(buf + j, v_dst); v_dst.val[0] = vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_t0))), v_scale_inv); v_dst.val[1] = vsubq_f32(vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_t1))), v_coeff1), v_134); v_dst.val[2] = vsubq_f32(vmulq_f32(vcvtq_f32_u32(vmovl_u16(vget_high_u16(v_t2))), v_coeff2), v_140); vst3q_f32(buf + j + 12, v_dst); } #elif CV_SSE2 if (haveSIMD) { for ( ; j <= (dn - 32) * 3; j += 96) { __m128i v_r0 = _mm_loadu_si128((__m128i const *)(src + j)); __m128i v_r1 = _mm_loadu_si128((__m128i const *)(src + j + 16)); __m128i v_g0 = _mm_loadu_si128((__m128i const *)(src + j + 32)); __m128i v_g1 = _mm_loadu_si128((__m128i const *)(src + j + 48)); __m128i v_b0 = _mm_loadu_si128((__m128i const *)(src + j + 64)); __m128i v_b1 = _mm_loadu_si128((__m128i const *)(src + j + 80)); _mm_deinterleave_epi8(v_r0, v_r1, v_g0, v_g1, v_b0, v_b1); process(_mm_unpacklo_epi8(v_r0, v_zero), _mm_unpacklo_epi8(v_g0, v_zero), _mm_unpacklo_epi8(v_b0, v_zero), buf + j); process(_mm_unpackhi_epi8(v_r0, v_zero), _mm_unpackhi_epi8(v_g0, v_zero), _mm_unpackhi_epi8(v_b0, v_zero), buf + j + 24); process(_mm_unpacklo_epi8(v_r1, v_zero), _mm_unpacklo_epi8(v_g1, v_zero), _mm_unpacklo_epi8(v_b1, v_zero), buf + j + 48); process(_mm_unpackhi_epi8(v_r1, v_zero), _mm_unpackhi_epi8(v_g1, v_zero), _mm_unpackhi_epi8(v_b1, v_zero), buf + j + 72); } } #endif for( ; j < dn*3; j += 3 ) { buf[j] = src[j]*(100.f/255.f); buf[j+1] = (float)(src[j+1]*1.388235294117647f - 134.f); buf[j+2] = (float)(src[j+2]*1.027450980392157f - 140.f); } cvt(buf, buf, dn); j = 0; #if CV_NEON for ( ; j <= (dn - 8) * 3; j += 24, dst += dcn * 8) { float32x4x3_t v_src0 = vld3q_f32(buf + j), v_src1 = vld3q_f32(buf + j + 12); uint8x8_t v_dst0 = vqmovn_u16(vcombine_u16(vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src0.val[0], v_scale))), vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src1.val[0], v_scale))))); uint8x8_t v_dst1 = vqmovn_u16(vcombine_u16(vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src0.val[1], v_scale))), vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src1.val[1], v_scale))))); uint8x8_t v_dst2 = vqmovn_u16(vcombine_u16(vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src0.val[2], v_scale))), vqmovn_u32(cv_vrndq_u32_f32(vmulq_f32(v_src1.val[2], v_scale))))); if (dcn == 4) { uint8x8x4_t v_dst; v_dst.val[0] = v_dst0; v_dst.val[1] = v_dst1; v_dst.val[2] = v_dst2; v_dst.val[3] = v_alpha; vst4_u8(dst, v_dst); } else { uint8x8x3_t v_dst; v_dst.val[0] = v_dst0; v_dst.val[1] = v_dst1; v_dst.val[2] = v_dst2; vst3_u8(dst, v_dst); } } #elif CV_SSE2 if (dcn == 3 && haveSIMD) { for ( ; j <= (dn * 3 - 16); j += 16, dst += 16) { __m128 v_src0 = _mm_mul_ps(_mm_load_ps(buf + j), v_scale); __m128 v_src1 = _mm_mul_ps(_mm_load_ps(buf + j + 4), v_scale); __m128 v_src2 = _mm_mul_ps(_mm_load_ps(buf + j + 8), v_scale); __m128 v_src3 = _mm_mul_ps(_mm_load_ps(buf + j + 12), v_scale); __m128i v_dst0 = _mm_packs_epi32(_mm_cvtps_epi32(v_src0), _mm_cvtps_epi32(v_src1)); __m128i v_dst1 = _mm_packs_epi32(_mm_cvtps_epi32(v_src2), _mm_cvtps_epi32(v_src3)); _mm_storeu_si128((__m128i *)dst, _mm_packus_epi16(v_dst0, v_dst1)); } int jr = j % 3; if (jr) dst -= jr, j -= jr; } #endif for( ; j < dn*3; j += 3, dst += dcn ) { dst[0] = saturate_cast(buf[j]*255.f); dst[1] = saturate_cast(buf[j+1]*255.f); dst[2] = saturate_cast(buf[j+2]*255.f); if( dcn == 4 ) dst[3] = alpha; } } } int dstcn; Luv2RGB_f cvt; #if CV_NEON float32x4_t v_scale, v_scale_inv, v_coeff1, v_coeff2, v_134, v_140; uint8x8_t v_alpha; #elif CV_SSE2 __m128 v_scale, v_scale_inv, v_coeff1, v_coeff2, v_134, v_140; __m128i v_zero; bool haveSIMD; #endif }; ///////////////////////////////////// YUV420 -> RGB ///////////////////////////////////// const int ITUR_BT_601_CY = 1220542; const int ITUR_BT_601_CUB = 2116026; const int ITUR_BT_601_CUG = -409993; const int ITUR_BT_601_CVG = -852492; const int ITUR_BT_601_CVR = 1673527; const int ITUR_BT_601_SHIFT = 20; // Coefficients for RGB to YUV420p conversion const int ITUR_BT_601_CRY = 269484; const int ITUR_BT_601_CGY = 528482; const int ITUR_BT_601_CBY = 102760; const int ITUR_BT_601_CRU = -155188; const int ITUR_BT_601_CGU = -305135; const int ITUR_BT_601_CBU = 460324; const int ITUR_BT_601_CGV = -385875; const int ITUR_BT_601_CBV = -74448; template struct YUV420sp2RGB888Invoker : ParallelLoopBody { Mat* dst; const uchar* my1, *muv; int width, stride; YUV420sp2RGB888Invoker(Mat* _dst, int _stride, const uchar* _y1, const uchar* _uv) : dst(_dst), my1(_y1), muv(_uv), width(_dst->cols), stride(_stride) {} void operator()(const Range& range) const { int rangeBegin = range.start * 2; int rangeEnd = range.end * 2; //R = 1.164(Y - 16) + 1.596(V - 128) //G = 1.164(Y - 16) - 0.813(V - 128) - 0.391(U - 128) //B = 1.164(Y - 16) + 2.018(U - 128) //R = (1220542(Y - 16) + 1673527(V - 128) + (1 << 19)) >> 20 //G = (1220542(Y - 16) - 852492(V - 128) - 409993(U - 128) + (1 << 19)) >> 20 //B = (1220542(Y - 16) + 2116026(U - 128) + (1 << 19)) >> 20 const uchar* y1 = my1 + rangeBegin * stride, *uv = muv + rangeBegin * stride / 2; #ifdef HAVE_TEGRA_OPTIMIZATION if(tegra::useTegra() && tegra::cvtYUV4202RGB(bIdx, uIdx, 3, y1, uv, stride, dst->ptr(rangeBegin), dst->step, rangeEnd - rangeBegin, dst->cols)) return; #endif for (int j = rangeBegin; j < rangeEnd; j += 2, y1 += stride * 2, uv += stride) { uchar* row1 = dst->ptr(j); uchar* row2 = dst->ptr(j + 1); const uchar* y2 = y1 + stride; for (int i = 0; i < width; i += 2, row1 += 6, row2 += 6) { int u = int(uv[i + 0 + uIdx]) - 128; int v = int(uv[i + 1 - uIdx]) - 128; int ruv = (1 << (ITUR_BT_601_SHIFT - 1)) + ITUR_BT_601_CVR * v; int guv = (1 << (ITUR_BT_601_SHIFT - 1)) + ITUR_BT_601_CVG * v + ITUR_BT_601_CUG * u; int buv = (1 << (ITUR_BT_601_SHIFT - 1)) + ITUR_BT_601_CUB * u; int y00 = std::max(0, int(y1[i]) - 16) * ITUR_BT_601_CY; row1[2-bIdx] = saturate_cast((y00 + ruv) >> ITUR_BT_601_SHIFT); row1[1] = saturate_cast((y00 + guv) >> ITUR_BT_601_SHIFT); row1[bIdx] = saturate_cast((y00 + buv) >> ITUR_BT_601_SHIFT); int y01 = std::max(0, int(y1[i + 1]) - 16) * ITUR_BT_601_CY; row1[5-bIdx] = saturate_cast((y01 + ruv) >> ITUR_BT_601_SHIFT); row1[4] = saturate_cast((y01 + guv) >> ITUR_BT_601_SHIFT); row1[3+bIdx] = saturate_cast((y01 + buv) >> ITUR_BT_601_SHIFT); int y10 = std::max(0, int(y2[i]) - 16) * ITUR_BT_601_CY; row2[2-bIdx] = saturate_cast((y10 + ruv) >> ITUR_BT_601_SHIFT); row2[1] = saturate_cast((y10 + guv) >> ITUR_BT_601_SHIFT); row2[bIdx] = saturate_cast((y10 + buv) >> ITUR_BT_601_SHIFT); int y11 = std::max(0, int(y2[i + 1]) - 16) * ITUR_BT_601_CY; row2[5-bIdx] = saturate_cast((y11 + ruv) >> ITUR_BT_601_SHIFT); row2[4] = saturate_cast((y11 + guv) >> ITUR_BT_601_SHIFT); row2[3+bIdx] = saturate_cast((y11 + buv) >> ITUR_BT_601_SHIFT); } } } }; template struct YUV420sp2RGBA8888Invoker : ParallelLoopBody { Mat* dst; const uchar* my1, *muv; int width, stride; YUV420sp2RGBA8888Invoker(Mat* _dst, int _stride, const uchar* _y1, const uchar* _uv) : dst(_dst), my1(_y1), muv(_uv), width(_dst->cols), stride(_stride) {} void operator()(const Range& range) const { int rangeBegin = range.start * 2; int rangeEnd = range.end * 2; //R = 1.164(Y - 16) + 1.596(V - 128) //G = 1.164(Y - 16) - 0.813(V - 128) - 0.391(U - 128) //B = 1.164(Y - 16) + 2.018(U - 128) //R = (1220542(Y - 16) + 1673527(V - 128) + (1 << 19)) >> 20 //G = (1220542(Y - 16) - 852492(V - 128) - 409993(U - 128) + (1 << 19)) >> 20 //B = (1220542(Y - 16) + 2116026(U - 128) + (1 << 19)) >> 20 const uchar* y1 = my1 + rangeBegin * stride, *uv = muv + rangeBegin * stride / 2; #ifdef HAVE_TEGRA_OPTIMIZATION if(tegra::useTegra() && tegra::cvtYUV4202RGB(bIdx, uIdx, 4, y1, uv, stride, dst->ptr(rangeBegin), dst->step, rangeEnd - rangeBegin, dst->cols)) return; #endif for (int j = rangeBegin; j < rangeEnd; j += 2, y1 += stride * 2, uv += stride) { uchar* row1 = dst->ptr(j); uchar* row2 = dst->ptr(j + 1); const uchar* y2 = y1 + stride; for (int i = 0; i < width; i += 2, row1 += 8, row2 += 8) { int u = int(uv[i + 0 + uIdx]) - 128; int v = int(uv[i + 1 - uIdx]) - 128; int ruv = (1 << (ITUR_BT_601_SHIFT - 1)) + ITUR_BT_601_CVR * v; int guv = (1 << (ITUR_BT_601_SHIFT - 1)) + ITUR_BT_601_CVG * v + ITUR_BT_601_CUG * u; int buv = (1 << (ITUR_BT_601_SHIFT - 1)) + ITUR_BT_601_CUB * u; int y00 = std::max(0, int(y1[i]) - 16) * ITUR_BT_601_CY; row1[2-bIdx] = saturate_cast((y00 + ruv) >> ITUR_BT_601_SHIFT); row1[1] = saturate_cast((y00 + guv) >> ITUR_BT_601_SHIFT); row1[bIdx] = saturate_cast((y00 + buv) >> ITUR_BT_601_SHIFT); row1[3] = uchar(0xff); int y01 = std::max(0, int(y1[i + 1]) - 16) * ITUR_BT_601_CY; row1[6-bIdx] = saturate_cast((y01 + ruv) >> ITUR_BT_601_SHIFT); row1[5] = saturate_cast((y01 + guv) >> ITUR_BT_601_SHIFT); row1[4+bIdx] = saturate_cast((y01 + buv) >> ITUR_BT_601_SHIFT); row1[7] = uchar(0xff); int y10 = std::max(0, int(y2[i]) - 16) * ITUR_BT_601_CY; row2[2-bIdx] = saturate_cast((y10 + ruv) >> ITUR_BT_601_SHIFT); row2[1] = saturate_cast((y10 + guv) >> ITUR_BT_601_SHIFT); row2[bIdx] = saturate_cast((y10 + buv) >> ITUR_BT_601_SHIFT); row2[3] = uchar(0xff); int y11 = std::max(0, int(y2[i + 1]) - 16) * ITUR_BT_601_CY; row2[6-bIdx] = saturate_cast((y11 + ruv) >> ITUR_BT_601_SHIFT); row2[5] = saturate_cast((y11 + guv) >> ITUR_BT_601_SHIFT); row2[4+bIdx] = saturate_cast((y11 + buv) >> ITUR_BT_601_SHIFT); row2[7] = uchar(0xff); } } } }; template struct YUV420p2RGB888Invoker : ParallelLoopBody { Mat* dst; const uchar* my1, *mu, *mv; int width, stride; int ustepIdx, vstepIdx; YUV420p2RGB888Invoker(Mat* _dst, int _stride, const uchar* _y1, const uchar* _u, const uchar* _v, int _ustepIdx, int _vstepIdx) : dst(_dst), my1(_y1), mu(_u), mv(_v), width(_dst->cols), stride(_stride), ustepIdx(_ustepIdx), vstepIdx(_vstepIdx) {} void operator()(const Range& range) const { const int rangeBegin = range.start * 2; const int rangeEnd = range.end * 2; int uvsteps[2] = {width/2, stride - width/2}; int usIdx = ustepIdx, vsIdx = vstepIdx; const uchar* y1 = my1 + rangeBegin * stride; const uchar* u1 = mu + (range.start / 2) * stride; const uchar* v1 = mv + (range.start / 2) * stride; if(range.start % 2 == 1) { u1 += uvsteps[(usIdx++) & 1]; v1 += uvsteps[(vsIdx++) & 1]; } for (int j = rangeBegin; j < rangeEnd; j += 2, y1 += stride * 2, u1 += uvsteps[(usIdx++) & 1], v1 += uvsteps[(vsIdx++) & 1]) { uchar* row1 = dst->ptr(j); uchar* row2 = dst->ptr(j + 1); const uchar* y2 = y1 + stride; for (int i = 0; i < width / 2; i += 1, row1 += 6, row2 += 6) { int u = int(u1[i]) - 128; int v = int(v1[i]) - 128; int ruv = (1 << (ITUR_BT_601_SHIFT - 1)) + ITUR_BT_601_CVR * v; int guv = (1 << (ITUR_BT_601_SHIFT - 1)) + ITUR_BT_601_CVG * v + ITUR_BT_601_CUG * u; int buv = (1 << (ITUR_BT_601_SHIFT - 1)) + ITUR_BT_601_CUB * u; int y00 = std::max(0, int(y1[2 * i]) - 16) * ITUR_BT_601_CY; row1[2-bIdx] = saturate_cast((y00 + ruv) >> ITUR_BT_601_SHIFT); row1[1] = saturate_cast((y00 + guv) >> ITUR_BT_601_SHIFT); row1[bIdx] = saturate_cast((y00 + buv) >> ITUR_BT_601_SHIFT); int y01 = std::max(0, int(y1[2 * i + 1]) - 16) * ITUR_BT_601_CY; row1[5-bIdx] = saturate_cast((y01 + ruv) >> ITUR_BT_601_SHIFT); row1[4] = saturate_cast((y01 + guv) >> ITUR_BT_601_SHIFT); row1[3+bIdx] = saturate_cast((y01 + buv) >> ITUR_BT_601_SHIFT); int y10 = std::max(0, int(y2[2 * i]) - 16) * ITUR_BT_601_CY; row2[2-bIdx] = saturate_cast((y10 + ruv) >> ITUR_BT_601_SHIFT); row2[1] = saturate_cast((y10 + guv) >> ITUR_BT_601_SHIFT); row2[bIdx] = saturate_cast((y10 + buv) >> ITUR_BT_601_SHIFT); int y11 = std::max(0, int(y2[2 * i + 1]) - 16) * ITUR_BT_601_CY; row2[5-bIdx] = saturate_cast((y11 + ruv) >> ITUR_BT_601_SHIFT); row2[4] = saturate_cast((y11 + guv) >> ITUR_BT_601_SHIFT); row2[3+bIdx] = saturate_cast((y11 + buv) >> ITUR_BT_601_SHIFT); } } } }; template struct YUV420p2RGBA8888Invoker : ParallelLoopBody { Mat* dst; const uchar* my1, *mu, *mv; int width, stride; int ustepIdx, vstepIdx; YUV420p2RGBA8888Invoker(Mat* _dst, int _stride, const uchar* _y1, const uchar* _u, const uchar* _v, int _ustepIdx, int _vstepIdx) : dst(_dst), my1(_y1), mu(_u), mv(_v), width(_dst->cols), stride(_stride), ustepIdx(_ustepIdx), vstepIdx(_vstepIdx) {} void operator()(const Range& range) const { int rangeBegin = range.start * 2; int rangeEnd = range.end * 2; int uvsteps[2] = {width/2, stride - width/2}; int usIdx = ustepIdx, vsIdx = vstepIdx; const uchar* y1 = my1 + rangeBegin * stride; const uchar* u1 = mu + (range.start / 2) * stride; const uchar* v1 = mv + (range.start / 2) * stride; if(range.start % 2 == 1) { u1 += uvsteps[(usIdx++) & 1]; v1 += uvsteps[(vsIdx++) & 1]; } for (int j = rangeBegin; j < rangeEnd; j += 2, y1 += stride * 2, u1 += uvsteps[(usIdx++) & 1], v1 += uvsteps[(vsIdx++) & 1]) { uchar* row1 = dst->ptr(j); uchar* row2 = dst->ptr(j + 1); const uchar* y2 = y1 + stride; for (int i = 0; i < width / 2; i += 1, row1 += 8, row2 += 8) { int u = int(u1[i]) - 128; int v = int(v1[i]) - 128; int ruv = (1 << (ITUR_BT_601_SHIFT - 1)) + ITUR_BT_601_CVR * v; int guv = (1 << (ITUR_BT_601_SHIFT - 1)) + ITUR_BT_601_CVG * v + ITUR_BT_601_CUG * u; int buv = (1 << (ITUR_BT_601_SHIFT - 1)) + ITUR_BT_601_CUB * u; int y00 = std::max(0, int(y1[2 * i]) - 16) * ITUR_BT_601_CY; row1[2-bIdx] = saturate_cast((y00 + ruv) >> ITUR_BT_601_SHIFT); row1[1] = saturate_cast((y00 + guv) >> ITUR_BT_601_SHIFT); row1[bIdx] = saturate_cast((y00 + buv) >> ITUR_BT_601_SHIFT); row1[3] = uchar(0xff); int y01 = std::max(0, int(y1[2 * i + 1]) - 16) * ITUR_BT_601_CY; row1[6-bIdx] = saturate_cast((y01 + ruv) >> ITUR_BT_601_SHIFT); row1[5] = saturate_cast((y01 + guv) >> ITUR_BT_601_SHIFT); row1[4+bIdx] = saturate_cast((y01 + buv) >> ITUR_BT_601_SHIFT); row1[7] = uchar(0xff); int y10 = std::max(0, int(y2[2 * i]) - 16) * ITUR_BT_601_CY; row2[2-bIdx] = saturate_cast((y10 + ruv) >> ITUR_BT_601_SHIFT); row2[1] = saturate_cast((y10 + guv) >> ITUR_BT_601_SHIFT); row2[bIdx] = saturate_cast((y10 + buv) >> ITUR_BT_601_SHIFT); row2[3] = uchar(0xff); int y11 = std::max(0, int(y2[2 * i + 1]) - 16) * ITUR_BT_601_CY; row2[6-bIdx] = saturate_cast((y11 + ruv) >> ITUR_BT_601_SHIFT); row2[5] = saturate_cast((y11 + guv) >> ITUR_BT_601_SHIFT); row2[4+bIdx] = saturate_cast((y11 + buv) >> ITUR_BT_601_SHIFT); row2[7] = uchar(0xff); } } } }; #define MIN_SIZE_FOR_PARALLEL_YUV420_CONVERSION (320*240) template inline void cvtYUV420sp2RGB(Mat& _dst, int _stride, const uchar* _y1, const uchar* _uv) { YUV420sp2RGB888Invoker converter(&_dst, _stride, _y1, _uv); if (_dst.total() >= MIN_SIZE_FOR_PARALLEL_YUV420_CONVERSION) parallel_for_(Range(0, _dst.rows/2), converter); else converter(Range(0, _dst.rows/2)); } template inline void cvtYUV420sp2RGBA(Mat& _dst, int _stride, const uchar* _y1, const uchar* _uv) { YUV420sp2RGBA8888Invoker converter(&_dst, _stride, _y1, _uv); if (_dst.total() >= MIN_SIZE_FOR_PARALLEL_YUV420_CONVERSION) parallel_for_(Range(0, _dst.rows/2), converter); else converter(Range(0, _dst.rows/2)); } template inline void cvtYUV420p2RGB(Mat& _dst, int _stride, const uchar* _y1, const uchar* _u, const uchar* _v, int ustepIdx, int vstepIdx) { YUV420p2RGB888Invoker converter(&_dst, _stride, _y1, _u, _v, ustepIdx, vstepIdx); if (_dst.total() >= MIN_SIZE_FOR_PARALLEL_YUV420_CONVERSION) parallel_for_(Range(0, _dst.rows/2), converter); else converter(Range(0, _dst.rows/2)); } template inline void cvtYUV420p2RGBA(Mat& _dst, int _stride, const uchar* _y1, const uchar* _u, const uchar* _v, int ustepIdx, int vstepIdx) { YUV420p2RGBA8888Invoker converter(&_dst, _stride, _y1, _u, _v, ustepIdx, vstepIdx); if (_dst.total() >= MIN_SIZE_FOR_PARALLEL_YUV420_CONVERSION) parallel_for_(Range(0, _dst.rows/2), converter); else converter(Range(0, _dst.rows/2)); } ///////////////////////////////////// RGB -> YUV420p ///////////////////////////////////// template struct RGB888toYUV420pInvoker: public ParallelLoopBody { RGB888toYUV420pInvoker( const Mat& src, Mat* dst, const int uIdx ) : src_(src), dst_(dst), uIdx_(uIdx) { } void operator()(const Range& rowRange) const { const int w = src_.cols; const int h = src_.rows; const int cn = src_.channels(); for( int i = rowRange.start; i < rowRange.end; i++ ) { const uchar* row0 = src_.ptr(2 * i); const uchar* row1 = src_.ptr(2 * i + 1); uchar* y = dst_->ptr(2*i); uchar* u = dst_->ptr(h + i/2) + (i % 2) * (w/2); uchar* v = dst_->ptr(h + (i + h/2)/2) + ((i + h/2) % 2) * (w/2); if( uIdx_ == 2 ) std::swap(u, v); for( int j = 0, k = 0; j < w * cn; j += 2 * cn, k++ ) { int r00 = row0[2-bIdx + j]; int g00 = row0[1 + j]; int b00 = row0[bIdx + j]; int r01 = row0[2-bIdx + cn + j]; int g01 = row0[1 + cn + j]; int b01 = row0[bIdx + cn + j]; int r10 = row1[2-bIdx + j]; int g10 = row1[1 + j]; int b10 = row1[bIdx + j]; int r11 = row1[2-bIdx + cn + j]; int g11 = row1[1 + cn + j]; int b11 = row1[bIdx + cn + j]; const int shifted16 = (16 << ITUR_BT_601_SHIFT); const int halfShift = (1 << (ITUR_BT_601_SHIFT - 1)); int y00 = ITUR_BT_601_CRY * r00 + ITUR_BT_601_CGY * g00 + ITUR_BT_601_CBY * b00 + halfShift + shifted16; int y01 = ITUR_BT_601_CRY * r01 + ITUR_BT_601_CGY * g01 + ITUR_BT_601_CBY * b01 + halfShift + shifted16; int y10 = ITUR_BT_601_CRY * r10 + ITUR_BT_601_CGY * g10 + ITUR_BT_601_CBY * b10 + halfShift + shifted16; int y11 = ITUR_BT_601_CRY * r11 + ITUR_BT_601_CGY * g11 + ITUR_BT_601_CBY * b11 + halfShift + shifted16; y[2*k + 0] = saturate_cast(y00 >> ITUR_BT_601_SHIFT); y[2*k + 1] = saturate_cast(y01 >> ITUR_BT_601_SHIFT); y[2*k + dst_->step + 0] = saturate_cast(y10 >> ITUR_BT_601_SHIFT); y[2*k + dst_->step + 1] = saturate_cast(y11 >> ITUR_BT_601_SHIFT); const int shifted128 = (128 << ITUR_BT_601_SHIFT); int u00 = ITUR_BT_601_CRU * r00 + ITUR_BT_601_CGU * g00 + ITUR_BT_601_CBU * b00 + halfShift + shifted128; int v00 = ITUR_BT_601_CBU * r00 + ITUR_BT_601_CGV * g00 + ITUR_BT_601_CBV * b00 + halfShift + shifted128; u[k] = saturate_cast(u00 >> ITUR_BT_601_SHIFT); v[k] = saturate_cast(v00 >> ITUR_BT_601_SHIFT); } } } static bool isFit( const Mat& src ) { return (src.total() >= 320*240); } private: RGB888toYUV420pInvoker& operator=(const RGB888toYUV420pInvoker&); const Mat& src_; Mat* const dst_; const int uIdx_; }; template static void cvtRGBtoYUV420p(const Mat& src, Mat& dst) { RGB888toYUV420pInvoker colorConverter(src, &dst, uIdx); if( RGB888toYUV420pInvoker::isFit(src) ) parallel_for_(Range(0, src.rows/2), colorConverter); else colorConverter(Range(0, src.rows/2)); } ///////////////////////////////////// YUV422 -> RGB ///////////////////////////////////// template struct YUV422toRGB888Invoker : ParallelLoopBody { Mat* dst; const uchar* src; int width, stride; YUV422toRGB888Invoker(Mat* _dst, int _stride, const uchar* _yuv) : dst(_dst), src(_yuv), width(_dst->cols), stride(_stride) {} void operator()(const Range& range) const { int rangeBegin = range.start; int rangeEnd = range.end; const int uidx = 1 - yIdx + uIdx * 2; const int vidx = (2 + uidx) % 4; const uchar* yuv_src = src + rangeBegin * stride; for (int j = rangeBegin; j < rangeEnd; j++, yuv_src += stride) { uchar* row = dst->ptr(j); for (int i = 0; i < 2 * width; i += 4, row += 6) { int u = int(yuv_src[i + uidx]) - 128; int v = int(yuv_src[i + vidx]) - 128; int ruv = (1 << (ITUR_BT_601_SHIFT - 1)) + ITUR_BT_601_CVR * v; int guv = (1 << (ITUR_BT_601_SHIFT - 1)) + ITUR_BT_601_CVG * v + ITUR_BT_601_CUG * u; int buv = (1 << (ITUR_BT_601_SHIFT - 1)) + ITUR_BT_601_CUB * u; int y00 = std::max(0, int(yuv_src[i + yIdx]) - 16) * ITUR_BT_601_CY; row[2-bIdx] = saturate_cast((y00 + ruv) >> ITUR_BT_601_SHIFT); row[1] = saturate_cast((y00 + guv) >> ITUR_BT_601_SHIFT); row[bIdx] = saturate_cast((y00 + buv) >> ITUR_BT_601_SHIFT); int y01 = std::max(0, int(yuv_src[i + yIdx + 2]) - 16) * ITUR_BT_601_CY; row[5-bIdx] = saturate_cast((y01 + ruv) >> ITUR_BT_601_SHIFT); row[4] = saturate_cast((y01 + guv) >> ITUR_BT_601_SHIFT); row[3+bIdx] = saturate_cast((y01 + buv) >> ITUR_BT_601_SHIFT); } } } }; template struct YUV422toRGBA8888Invoker : ParallelLoopBody { Mat* dst; const uchar* src; int width, stride; YUV422toRGBA8888Invoker(Mat* _dst, int _stride, const uchar* _yuv) : dst(_dst), src(_yuv), width(_dst->cols), stride(_stride) {} void operator()(const Range& range) const { int rangeBegin = range.start; int rangeEnd = range.end; const int uidx = 1 - yIdx + uIdx * 2; const int vidx = (2 + uidx) % 4; const uchar* yuv_src = src + rangeBegin * stride; for (int j = rangeBegin; j < rangeEnd; j++, yuv_src += stride) { uchar* row = dst->ptr(j); for (int i = 0; i < 2 * width; i += 4, row += 8) { int u = int(yuv_src[i + uidx]) - 128; int v = int(yuv_src[i + vidx]) - 128; int ruv = (1 << (ITUR_BT_601_SHIFT - 1)) + ITUR_BT_601_CVR * v; int guv = (1 << (ITUR_BT_601_SHIFT - 1)) + ITUR_BT_601_CVG * v + ITUR_BT_601_CUG * u; int buv = (1 << (ITUR_BT_601_SHIFT - 1)) + ITUR_BT_601_CUB * u; int y00 = std::max(0, int(yuv_src[i + yIdx]) - 16) * ITUR_BT_601_CY; row[2-bIdx] = saturate_cast((y00 + ruv) >> ITUR_BT_601_SHIFT); row[1] = saturate_cast((y00 + guv) >> ITUR_BT_601_SHIFT); row[bIdx] = saturate_cast((y00 + buv) >> ITUR_BT_601_SHIFT); row[3] = uchar(0xff); int y01 = std::max(0, int(yuv_src[i + yIdx + 2]) - 16) * ITUR_BT_601_CY; row[6-bIdx] = saturate_cast((y01 + ruv) >> ITUR_BT_601_SHIFT); row[5] = saturate_cast((y01 + guv) >> ITUR_BT_601_SHIFT); row[4+bIdx] = saturate_cast((y01 + buv) >> ITUR_BT_601_SHIFT); row[7] = uchar(0xff); } } } }; #define MIN_SIZE_FOR_PARALLEL_YUV422_CONVERSION (320*240) template inline void cvtYUV422toRGB(Mat& _dst, int _stride, const uchar* _yuv) { YUV422toRGB888Invoker converter(&_dst, _stride, _yuv); if (_dst.total() >= MIN_SIZE_FOR_PARALLEL_YUV422_CONVERSION) parallel_for_(Range(0, _dst.rows), converter); else converter(Range(0, _dst.rows)); } template inline void cvtYUV422toRGBA(Mat& _dst, int _stride, const uchar* _yuv) { YUV422toRGBA8888Invoker converter(&_dst, _stride, _yuv); if (_dst.total() >= MIN_SIZE_FOR_PARALLEL_YUV422_CONVERSION) parallel_for_(Range(0, _dst.rows), converter); else converter(Range(0, _dst.rows)); } /////////////////////////// RGBA <-> mRGBA (alpha premultiplied) ////////////// template struct RGBA2mRGBA { typedef _Tp channel_type; void operator()(const _Tp* src, _Tp* dst, int n) const { _Tp max_val = ColorChannel<_Tp>::max(); _Tp half_val = ColorChannel<_Tp>::half(); for( int i = 0; i < n; i++ ) { _Tp v0 = *src++; _Tp v1 = *src++; _Tp v2 = *src++; _Tp v3 = *src++; *dst++ = (v0 * v3 + half_val) / max_val; *dst++ = (v1 * v3 + half_val) / max_val; *dst++ = (v2 * v3 + half_val) / max_val; *dst++ = v3; } } }; template struct mRGBA2RGBA { typedef _Tp channel_type; void operator()(const _Tp* src, _Tp* dst, int n) const { _Tp max_val = ColorChannel<_Tp>::max(); for( int i = 0; i < n; i++ ) { _Tp v0 = *src++; _Tp v1 = *src++; _Tp v2 = *src++; _Tp v3 = *src++; _Tp v3_half = v3 / 2; *dst++ = (v3==0)? 0 : (v0 * max_val + v3_half) / v3; *dst++ = (v3==0)? 0 : (v1 * max_val + v3_half) / v3; *dst++ = (v3==0)? 0 : (v2 * max_val + v3_half) / v3; *dst++ = v3; } } }; #ifdef HAVE_OPENCL static bool ocl_cvtColor( InputArray _src, OutputArray _dst, int code, int dcn ) { bool ok = false; UMat src = _src.getUMat(), dst; Size sz = src.size(), dstSz = sz; int scn = src.channels(), depth = src.depth(), bidx, uidx, yidx; int dims = 2, stripeSize = 1; ocl::Kernel k; if (depth != CV_8U && depth != CV_16U && depth != CV_32F) return false; ocl::Device dev = ocl::Device::getDefault(); int pxPerWIy = dev.isIntel() && (dev.type() & ocl::Device::TYPE_GPU) ? 4 : 1; int pxPerWIx = 1; size_t globalsize[] = { src.cols, (src.rows + pxPerWIy - 1) / pxPerWIy }; cv::String opts = format("-D depth=%d -D scn=%d -D PIX_PER_WI_Y=%d ", depth, scn, pxPerWIy); switch (code) { case COLOR_BGR2BGRA: case COLOR_RGB2BGRA: case COLOR_BGRA2BGR: case COLOR_RGBA2BGR: case COLOR_RGB2BGR: case COLOR_BGRA2RGBA: { CV_Assert(scn == 3 || scn == 4); dcn = code == COLOR_BGR2BGRA || code == COLOR_RGB2BGRA || code == COLOR_BGRA2RGBA ? 4 : 3; bool reverse = !(code == COLOR_BGR2BGRA || code == COLOR_BGRA2BGR); k.create("RGB", ocl::imgproc::cvtcolor_oclsrc, opts + format("-D dcn=%d -D bidx=0 -D %s", dcn, reverse ? "REVERSE" : "ORDER")); break; } case COLOR_BGR5652BGR: case COLOR_BGR5552BGR: case COLOR_BGR5652RGB: case COLOR_BGR5552RGB: case COLOR_BGR5652BGRA: case COLOR_BGR5552BGRA: case COLOR_BGR5652RGBA: case COLOR_BGR5552RGBA: { dcn = code == COLOR_BGR5652BGRA || code == COLOR_BGR5552BGRA || code == COLOR_BGR5652RGBA || code == COLOR_BGR5552RGBA ? 4 : 3; CV_Assert((dcn == 3 || dcn == 4) && scn == 2 && depth == CV_8U); bidx = code == COLOR_BGR5652BGR || code == COLOR_BGR5552BGR || code == COLOR_BGR5652BGRA || code == COLOR_BGR5552BGRA ? 0 : 2; int greenbits = code == COLOR_BGR5652BGR || code == COLOR_BGR5652RGB || code == COLOR_BGR5652BGRA || code == COLOR_BGR5652RGBA ? 6 : 5; k.create("RGB5x52RGB", ocl::imgproc::cvtcolor_oclsrc, opts + format("-D dcn=%d -D bidx=%d -D greenbits=%d", dcn, bidx, greenbits)); break; } case COLOR_BGR2BGR565: case COLOR_BGR2BGR555: case COLOR_RGB2BGR565: case COLOR_RGB2BGR555: case COLOR_BGRA2BGR565: case COLOR_BGRA2BGR555: case COLOR_RGBA2BGR565: case COLOR_RGBA2BGR555: { CV_Assert((scn == 3 || scn == 4) && depth == CV_8U ); bidx = code == COLOR_BGR2BGR565 || code == COLOR_BGR2BGR555 || code == COLOR_BGRA2BGR565 || code == COLOR_BGRA2BGR555 ? 0 : 2; int greenbits = code == COLOR_BGR2BGR565 || code == COLOR_RGB2BGR565 || code == COLOR_BGRA2BGR565 || code == COLOR_RGBA2BGR565 ? 6 : 5; dcn = 2; k.create("RGB2RGB5x5", ocl::imgproc::cvtcolor_oclsrc, opts + format("-D dcn=2 -D bidx=%d -D greenbits=%d", bidx, greenbits)); break; } case COLOR_BGR5652GRAY: case COLOR_BGR5552GRAY: { CV_Assert(scn == 2 && depth == CV_8U); dcn = 1; int greenbits = code == COLOR_BGR5652GRAY ? 6 : 5; k.create("BGR5x52Gray", ocl::imgproc::cvtcolor_oclsrc, opts + format("-D dcn=1 -D bidx=0 -D greenbits=%d", greenbits)); break; } case COLOR_GRAY2BGR565: case COLOR_GRAY2BGR555: { CV_Assert(scn == 1 && depth == CV_8U); dcn = 2; int greenbits = code == COLOR_GRAY2BGR565 ? 6 : 5; k.create("Gray2BGR5x5", ocl::imgproc::cvtcolor_oclsrc, opts + format("-D dcn=2 -D bidx=0 -D greenbits=%d", greenbits)); break; } case COLOR_BGR2GRAY: case COLOR_BGRA2GRAY: case COLOR_RGB2GRAY: case COLOR_RGBA2GRAY: { CV_Assert(scn == 3 || scn == 4); bidx = code == COLOR_BGR2GRAY || code == COLOR_BGRA2GRAY ? 0 : 2; dcn = 1; k.create("RGB2Gray", ocl::imgproc::cvtcolor_oclsrc, opts + format("-D dcn=1 -D bidx=%d -D STRIPE_SIZE=%d", bidx, stripeSize)); globalsize[0] = (src.cols + stripeSize-1)/stripeSize; break; } case COLOR_GRAY2BGR: case COLOR_GRAY2BGRA: { CV_Assert(scn == 1); dcn = code == COLOR_GRAY2BGRA ? 4 : 3; k.create("Gray2RGB", ocl::imgproc::cvtcolor_oclsrc, opts + format("-D bidx=0 -D dcn=%d", dcn)); break; } case COLOR_BGR2YUV: case COLOR_RGB2YUV: { CV_Assert(scn == 3 || scn == 4); bidx = code == COLOR_RGB2YUV ? 0 : 2; dcn = 3; k.create("RGB2YUV", ocl::imgproc::cvtcolor_oclsrc, opts + format("-D dcn=3 -D bidx=%d", bidx)); break; } case COLOR_YUV2BGR: case COLOR_YUV2RGB: { if(dcn < 0) dcn = 3; CV_Assert(dcn == 3 || dcn == 4); bidx = code == COLOR_YUV2RGB ? 0 : 2; k.create("YUV2RGB", ocl::imgproc::cvtcolor_oclsrc, opts + format("-D dcn=%d -D bidx=%d", dcn, bidx)); break; } case COLOR_YUV2RGB_NV12: case COLOR_YUV2BGR_NV12: case COLOR_YUV2RGB_NV21: case COLOR_YUV2BGR_NV21: case COLOR_YUV2RGBA_NV12: case COLOR_YUV2BGRA_NV12: case COLOR_YUV2RGBA_NV21: case COLOR_YUV2BGRA_NV21: { CV_Assert( scn == 1 ); CV_Assert( sz.width % 2 == 0 && sz.height % 3 == 0 && depth == CV_8U ); dcn = code == COLOR_YUV2BGRA_NV12 || code == COLOR_YUV2RGBA_NV12 || code == COLOR_YUV2BGRA_NV21 || code == COLOR_YUV2RGBA_NV21 ? 4 : 3; bidx = code == COLOR_YUV2BGRA_NV12 || code == COLOR_YUV2BGR_NV12 || code == COLOR_YUV2BGRA_NV21 || code == COLOR_YUV2BGR_NV21 ? 0 : 2; uidx = code == COLOR_YUV2RGBA_NV21 || code == COLOR_YUV2RGB_NV21 || code == COLOR_YUV2BGRA_NV21 || code == COLOR_YUV2BGR_NV21 ? 1 : 0; dstSz = Size(sz.width, sz.height * 2 / 3); globalsize[0] = dstSz.width / 2; globalsize[1] = (dstSz.height/2 + pxPerWIy - 1) / pxPerWIy; k.create("YUV2RGB_NVx", ocl::imgproc::cvtcolor_oclsrc, opts + format("-D dcn=%d -D bidx=%d -D uidx=%d", dcn, bidx, uidx)); break; } case COLOR_YUV2BGR_YV12: case COLOR_YUV2RGB_YV12: case COLOR_YUV2BGRA_YV12: case COLOR_YUV2RGBA_YV12: case COLOR_YUV2BGR_IYUV: case COLOR_YUV2RGB_IYUV: case COLOR_YUV2BGRA_IYUV: case COLOR_YUV2RGBA_IYUV: { CV_Assert( scn == 1 ); CV_Assert( sz.width % 2 == 0 && sz.height % 3 == 0 && depth == CV_8U ); dcn = code == COLOR_YUV2BGRA_YV12 || code == COLOR_YUV2RGBA_YV12 || code == COLOR_YUV2BGRA_IYUV || code == COLOR_YUV2RGBA_IYUV ? 4 : 3; bidx = code == COLOR_YUV2BGRA_YV12 || code == COLOR_YUV2BGR_YV12 || code == COLOR_YUV2BGRA_IYUV || code == COLOR_YUV2BGR_IYUV ? 0 : 2; uidx = code == COLOR_YUV2BGRA_YV12 || code == COLOR_YUV2BGR_YV12 || code == COLOR_YUV2RGBA_YV12 || code == COLOR_YUV2RGB_YV12 ? 1 : 0; dstSz = Size(sz.width, sz.height * 2 / 3); globalsize[0] = dstSz.width / 2; globalsize[1] = (dstSz.height/2 + pxPerWIy - 1) / pxPerWIy; k.create("YUV2RGB_YV12_IYUV", ocl::imgproc::cvtcolor_oclsrc, opts + format("-D dcn=%d -D bidx=%d -D uidx=%d%s", dcn, bidx, uidx, src.isContinuous() ? " -D SRC_CONT" : "")); break; } case COLOR_YUV2GRAY_420: { if (dcn <= 0) dcn = 1; CV_Assert( dcn == 1 ); CV_Assert( sz.width % 2 == 0 && sz.height % 3 == 0 && depth == CV_8U ); dstSz = Size(sz.width, sz.height * 2 / 3); _dst.create(dstSz, CV_MAKETYPE(depth, dcn)); dst = _dst.getUMat(); src.rowRange(0, dstSz.height).copyTo(dst); return true; } case COLOR_RGB2YUV_YV12: case COLOR_BGR2YUV_YV12: case COLOR_RGBA2YUV_YV12: case COLOR_BGRA2YUV_YV12: case COLOR_RGB2YUV_IYUV: case COLOR_BGR2YUV_IYUV: case COLOR_RGBA2YUV_IYUV: case COLOR_BGRA2YUV_IYUV: { if (dcn <= 0) dcn = 1; bidx = code == COLOR_BGRA2YUV_YV12 || code == COLOR_BGR2YUV_YV12 || code == COLOR_BGRA2YUV_IYUV || code == COLOR_BGR2YUV_IYUV ? 0 : 2; uidx = code == COLOR_RGBA2YUV_YV12 || code == COLOR_RGB2YUV_YV12 || code == COLOR_BGRA2YUV_YV12 || code == COLOR_BGR2YUV_YV12 ? 1 : 0; CV_Assert( (scn == 3 || scn == 4) && depth == CV_8U ); CV_Assert( dcn == 1 ); CV_Assert( sz.width % 2 == 0 && sz.height % 2 == 0 ); dstSz = Size(sz.width, sz.height / 2 * 3); _dst.create(dstSz, CV_MAKETYPE(depth, dcn)); dst = _dst.getUMat(); if (dev.isIntel() && src.cols % 4 == 0 && src.step % 4 == 0 && src.offset % 4 == 0 && dst.step % 4 == 0 && dst.offset % 4 == 0) { pxPerWIx = 2; } globalsize[0] = dstSz.width / (2 * pxPerWIx); globalsize[1] = (dstSz.height/3 + pxPerWIy - 1) / pxPerWIy; k.create("RGB2YUV_YV12_IYUV", ocl::imgproc::cvtcolor_oclsrc, opts + format("-D dcn=%d -D bidx=%d -D uidx=%d -D PIX_PER_WI_X=%d", dcn, bidx, uidx, pxPerWIx)); k.args(ocl::KernelArg::ReadOnlyNoSize(src), ocl::KernelArg::WriteOnly(dst)); return k.run(2, globalsize, NULL, false); } case COLOR_YUV2RGB_UYVY: case COLOR_YUV2BGR_UYVY: case COLOR_YUV2RGBA_UYVY: case COLOR_YUV2BGRA_UYVY: case COLOR_YUV2RGB_YUY2: case COLOR_YUV2BGR_YUY2: case COLOR_YUV2RGB_YVYU: case COLOR_YUV2BGR_YVYU: case COLOR_YUV2RGBA_YUY2: case COLOR_YUV2BGRA_YUY2: case COLOR_YUV2RGBA_YVYU: case COLOR_YUV2BGRA_YVYU: { if (dcn <= 0) dcn = (code==COLOR_YUV2RGBA_UYVY || code==COLOR_YUV2BGRA_UYVY || code==COLOR_YUV2RGBA_YUY2 || code==COLOR_YUV2BGRA_YUY2 || code==COLOR_YUV2RGBA_YVYU || code==COLOR_YUV2BGRA_YVYU) ? 4 : 3; bidx = (code==COLOR_YUV2BGR_UYVY || code==COLOR_YUV2BGRA_UYVY || code==COLOR_YUV2BGRA_YUY2 || code==COLOR_YUV2BGR_YUY2 || code==COLOR_YUV2BGRA_YVYU || code==COLOR_YUV2BGR_YVYU) ? 0 : 2; yidx = (code==COLOR_YUV2RGB_UYVY || code==COLOR_YUV2RGBA_UYVY || code==COLOR_YUV2BGR_UYVY || code==COLOR_YUV2BGRA_UYVY) ? 1 : 0; uidx = (code==COLOR_YUV2RGB_YVYU || code==COLOR_YUV2RGBA_YVYU || code==COLOR_YUV2BGR_YVYU || code==COLOR_YUV2BGRA_YVYU) ? 2 : 0; uidx = 1 - yidx + uidx; CV_Assert( dcn == 3 || dcn == 4 ); CV_Assert( scn == 2 && depth == CV_8U ); k.create("YUV2RGB_422", ocl::imgproc::cvtcolor_oclsrc, opts + format("-D dcn=%d -D bidx=%d -D uidx=%d -D yidx=%d%s", dcn, bidx, uidx, yidx, src.offset % 4 == 0 && src.step % 4 == 0 ? " -D USE_OPTIMIZED_LOAD" : "")); break; } case COLOR_BGR2YCrCb: case COLOR_RGB2YCrCb: { CV_Assert(scn == 3 || scn == 4); bidx = code == COLOR_BGR2YCrCb ? 0 : 2; dcn = 3; k.create("RGB2YCrCb", ocl::imgproc::cvtcolor_oclsrc, opts + format("-D dcn=3 -D bidx=%d", bidx)); break; } case COLOR_YCrCb2BGR: case COLOR_YCrCb2RGB: { if( dcn <= 0 ) dcn = 3; CV_Assert(scn == 3 && (dcn == 3 || dcn == 4)); bidx = code == COLOR_YCrCb2BGR ? 0 : 2; k.create("YCrCb2RGB", ocl::imgproc::cvtcolor_oclsrc, opts + format("-D dcn=%d -D bidx=%d", dcn, bidx)); break; } case COLOR_BGR2XYZ: case COLOR_RGB2XYZ: { CV_Assert(scn == 3 || scn == 4); bidx = code == COLOR_BGR2XYZ ? 0 : 2; UMat c; if (depth == CV_32F) { float coeffs[] = { 0.412453f, 0.357580f, 0.180423f, 0.212671f, 0.715160f, 0.072169f, 0.019334f, 0.119193f, 0.950227f }; if (bidx == 0) { std::swap(coeffs[0], coeffs[2]); std::swap(coeffs[3], coeffs[5]); std::swap(coeffs[6], coeffs[8]); } Mat(1, 9, CV_32FC1, &coeffs[0]).copyTo(c); } else { int coeffs[] = { 1689, 1465, 739, 871, 2929, 296, 79, 488, 3892 }; if (bidx == 0) { std::swap(coeffs[0], coeffs[2]); std::swap(coeffs[3], coeffs[5]); std::swap(coeffs[6], coeffs[8]); } Mat(1, 9, CV_32SC1, &coeffs[0]).copyTo(c); } _dst.create(dstSz, CV_MAKETYPE(depth, 3)); dst = _dst.getUMat(); k.create("RGB2XYZ", ocl::imgproc::cvtcolor_oclsrc, opts + format("-D dcn=3 -D bidx=%d", bidx)); if (k.empty()) return false; k.args(ocl::KernelArg::ReadOnlyNoSize(src), ocl::KernelArg::WriteOnly(dst), ocl::KernelArg::PtrReadOnly(c)); return k.run(2, globalsize, 0, false); } case COLOR_XYZ2BGR: case COLOR_XYZ2RGB: { if (dcn <= 0) dcn = 3; CV_Assert(scn == 3 && (dcn == 3 || dcn == 4)); bidx = code == COLOR_XYZ2BGR ? 0 : 2; UMat c; if (depth == CV_32F) { float coeffs[] = { 3.240479f, -1.53715f, -0.498535f, -0.969256f, 1.875991f, 0.041556f, 0.055648f, -0.204043f, 1.057311f }; if (bidx == 0) { std::swap(coeffs[0], coeffs[6]); std::swap(coeffs[1], coeffs[7]); std::swap(coeffs[2], coeffs[8]); } Mat(1, 9, CV_32FC1, &coeffs[0]).copyTo(c); } else { int coeffs[] = { 13273, -6296, -2042, -3970, 7684, 170, 228, -836, 4331 }; if (bidx == 0) { std::swap(coeffs[0], coeffs[6]); std::swap(coeffs[1], coeffs[7]); std::swap(coeffs[2], coeffs[8]); } Mat(1, 9, CV_32SC1, &coeffs[0]).copyTo(c); } _dst.create(dstSz, CV_MAKETYPE(depth, dcn)); dst = _dst.getUMat(); k.create("XYZ2RGB", ocl::imgproc::cvtcolor_oclsrc, opts + format("-D dcn=%d -D bidx=%d", dcn, bidx)); if (k.empty()) return false; k.args(ocl::KernelArg::ReadOnlyNoSize(src), ocl::KernelArg::WriteOnly(dst), ocl::KernelArg::PtrReadOnly(c)); return k.run(2, globalsize, 0, false); } case COLOR_BGR2HSV: case COLOR_RGB2HSV: case COLOR_BGR2HSV_FULL: case COLOR_RGB2HSV_FULL: case COLOR_BGR2HLS: case COLOR_RGB2HLS: case COLOR_BGR2HLS_FULL: case COLOR_RGB2HLS_FULL: { CV_Assert((scn == 3 || scn == 4) && (depth == CV_8U || depth == CV_32F)); bidx = code == COLOR_BGR2HSV || code == COLOR_BGR2HLS || code == COLOR_BGR2HSV_FULL || code == COLOR_BGR2HLS_FULL ? 0 : 2; int hrange = depth == CV_32F ? 360 : code == COLOR_BGR2HSV || code == COLOR_RGB2HSV || code == COLOR_BGR2HLS || code == COLOR_RGB2HLS ? 180 : 256; bool is_hsv = code == COLOR_BGR2HSV || code == COLOR_RGB2HSV || code == COLOR_BGR2HSV_FULL || code == COLOR_RGB2HSV_FULL; String kernelName = String("RGB2") + (is_hsv ? "HSV" : "HLS"); dcn = 3; if (is_hsv && depth == CV_8U) { static UMat sdiv_data; static UMat hdiv_data180; static UMat hdiv_data256; static int sdiv_table[256]; static int hdiv_table180[256]; static int hdiv_table256[256]; static volatile bool initialized180 = false, initialized256 = false; volatile bool & initialized = hrange == 180 ? initialized180 : initialized256; if (!initialized) { int * const hdiv_table = hrange == 180 ? hdiv_table180 : hdiv_table256, hsv_shift = 12; UMat & hdiv_data = hrange == 180 ? hdiv_data180 : hdiv_data256; sdiv_table[0] = hdiv_table180[0] = hdiv_table256[0] = 0; int v = 255 << hsv_shift; if (!initialized180 && !initialized256) { for(int i = 1; i < 256; i++ ) sdiv_table[i] = saturate_cast(v/(1.*i)); Mat(1, 256, CV_32SC1, sdiv_table).copyTo(sdiv_data); } v = hrange << hsv_shift; for (int i = 1; i < 256; i++ ) hdiv_table[i] = saturate_cast(v/(6.*i)); Mat(1, 256, CV_32SC1, hdiv_table).copyTo(hdiv_data); initialized = true; } _dst.create(dstSz, CV_8UC3); dst = _dst.getUMat(); k.create("RGB2HSV", ocl::imgproc::cvtcolor_oclsrc, opts + format("-D hrange=%d -D bidx=%d -D dcn=3", hrange, bidx)); if (k.empty()) return false; k.args(ocl::KernelArg::ReadOnlyNoSize(src), ocl::KernelArg::WriteOnly(dst), ocl::KernelArg::PtrReadOnly(sdiv_data), hrange == 256 ? ocl::KernelArg::PtrReadOnly(hdiv_data256) : ocl::KernelArg::PtrReadOnly(hdiv_data180)); return k.run(2, globalsize, NULL, false); } else k.create(kernelName.c_str(), ocl::imgproc::cvtcolor_oclsrc, opts + format("-D hscale=%ff -D bidx=%d -D dcn=3", hrange*(1.f/360.f), bidx)); break; } case COLOR_HSV2BGR: case COLOR_HSV2RGB: case COLOR_HSV2BGR_FULL: case COLOR_HSV2RGB_FULL: case COLOR_HLS2BGR: case COLOR_HLS2RGB: case COLOR_HLS2BGR_FULL: case COLOR_HLS2RGB_FULL: { if (dcn <= 0) dcn = 3; CV_Assert(scn == 3 && (dcn == 3 || dcn == 4) && (depth == CV_8U || depth == CV_32F)); bidx = code == COLOR_HSV2BGR || code == COLOR_HLS2BGR || code == COLOR_HSV2BGR_FULL || code == COLOR_HLS2BGR_FULL ? 0 : 2; int hrange = depth == CV_32F ? 360 : code == COLOR_HSV2BGR || code == COLOR_HSV2RGB || code == COLOR_HLS2BGR || code == COLOR_HLS2RGB ? 180 : 255; bool is_hsv = code == COLOR_HSV2BGR || code == COLOR_HSV2RGB || code == COLOR_HSV2BGR_FULL || code == COLOR_HSV2RGB_FULL; String kernelName = String(is_hsv ? "HSV" : "HLS") + "2RGB"; k.create(kernelName.c_str(), ocl::imgproc::cvtcolor_oclsrc, opts + format("-D dcn=%d -D bidx=%d -D hrange=%d -D hscale=%ff", dcn, bidx, hrange, 6.f/hrange)); break; } case COLOR_RGBA2mRGBA: case COLOR_mRGBA2RGBA: { CV_Assert(scn == 4 && depth == CV_8U); dcn = 4; k.create(code == COLOR_RGBA2mRGBA ? "RGBA2mRGBA" : "mRGBA2RGBA", ocl::imgproc::cvtcolor_oclsrc, opts + "-D dcn=4 -D bidx=3"); break; } case CV_BGR2Lab: case CV_RGB2Lab: case CV_LBGR2Lab: case CV_LRGB2Lab: case CV_BGR2Luv: case CV_RGB2Luv: case CV_LBGR2Luv: case CV_LRGB2Luv: { CV_Assert( (scn == 3 || scn == 4) && (depth == CV_8U || depth == CV_32F) ); bidx = code == CV_BGR2Lab || code == CV_LBGR2Lab || code == CV_BGR2Luv || code == CV_LBGR2Luv ? 0 : 2; bool srgb = code == CV_BGR2Lab || code == CV_RGB2Lab || code == CV_RGB2Luv || code == CV_BGR2Luv; bool lab = code == CV_BGR2Lab || code == CV_RGB2Lab || code == CV_LBGR2Lab || code == CV_LRGB2Lab; float un, vn; dcn = 3; k.create(format("BGR2%s", lab ? "Lab" : "Luv").c_str(), ocl::imgproc::cvtcolor_oclsrc, opts + format("-D dcn=%d -D bidx=%d%s", dcn, bidx, srgb ? " -D SRGB" : "")); if (k.empty()) return false; initLabTabs(); _dst.create(dstSz, CV_MAKETYPE(depth, dcn)); dst = _dst.getUMat(); ocl::KernelArg srcarg = ocl::KernelArg::ReadOnlyNoSize(src), dstarg = ocl::KernelArg::WriteOnly(dst); if (depth == CV_8U && lab) { static UMat usRGBGammaTab, ulinearGammaTab, uLabCbrtTab, ucoeffs; if (srgb && usRGBGammaTab.empty()) Mat(1, 256, CV_16UC1, sRGBGammaTab_b).copyTo(usRGBGammaTab); else if (ulinearGammaTab.empty()) Mat(1, 256, CV_16UC1, linearGammaTab_b).copyTo(ulinearGammaTab); if (uLabCbrtTab.empty()) Mat(1, LAB_CBRT_TAB_SIZE_B, CV_16UC1, LabCbrtTab_b).copyTo(uLabCbrtTab); { int coeffs[9]; const float * const _coeffs = sRGB2XYZ_D65, * const _whitept = D65; const float scale[] = { (1 << lab_shift)/_whitept[0], (float)(1 << lab_shift), (1 << lab_shift)/_whitept[2] }; for (int i = 0; i < 3; i++ ) { coeffs[i*3+(bidx^2)] = cvRound(_coeffs[i*3]*scale[i]); coeffs[i*3+1] = cvRound(_coeffs[i*3+1]*scale[i]); coeffs[i*3+bidx] = cvRound(_coeffs[i*3+2]*scale[i]); CV_Assert( coeffs[i] >= 0 && coeffs[i*3+1] >= 0 && coeffs[i*3+2] >= 0 && coeffs[i*3] + coeffs[i*3+1] + coeffs[i*3+2] < 2*(1 << lab_shift) ); } Mat(1, 9, CV_32SC1, coeffs).copyTo(ucoeffs); } const int Lscale = (116*255+50)/100; const int Lshift = -((16*255*(1 << lab_shift2) + 50)/100); k.args(srcarg, dstarg, ocl::KernelArg::PtrReadOnly(srgb ? usRGBGammaTab : ulinearGammaTab), ocl::KernelArg::PtrReadOnly(uLabCbrtTab), ocl::KernelArg::PtrReadOnly(ucoeffs), Lscale, Lshift); } else { static UMat usRGBGammaTab, ucoeffs, uLabCbrtTab; if (srgb && usRGBGammaTab.empty()) Mat(1, GAMMA_TAB_SIZE * 4, CV_32FC1, sRGBGammaTab).copyTo(usRGBGammaTab); if (!lab && uLabCbrtTab.empty()) Mat(1, LAB_CBRT_TAB_SIZE * 4, CV_32FC1, LabCbrtTab).copyTo(uLabCbrtTab); { float coeffs[9]; const float * const _coeffs = sRGB2XYZ_D65, * const _whitept = D65; float scale[] = { 1.0f / _whitept[0], 1.0f, 1.0f / _whitept[2] }; for (int i = 0; i < 3; i++) { int j = i * 3; coeffs[j + (bidx ^ 2)] = _coeffs[j] * (lab ? scale[i] : 1); coeffs[j + 1] = _coeffs[j + 1] * (lab ? scale[i] : 1); coeffs[j + bidx] = _coeffs[j + 2] * (lab ? scale[i] : 1); CV_Assert( coeffs[j] >= 0 && coeffs[j + 1] >= 0 && coeffs[j + 2] >= 0 && coeffs[j] + coeffs[j + 1] + coeffs[j + 2] < 1.5f*(lab ? LabCbrtTabScale : 1) ); } float d = 1.f/(_whitept[0] + _whitept[1]*15 + _whitept[2]*3); un = 13*4*_whitept[0]*d; vn = 13*9*_whitept[1]*d; Mat(1, 9, CV_32FC1, coeffs).copyTo(ucoeffs); } float _1_3 = 1.0f / 3.0f, _a = 16.0f / 116.0f; ocl::KernelArg ucoeffsarg = ocl::KernelArg::PtrReadOnly(ucoeffs); if (lab) { if (srgb) k.args(srcarg, dstarg, ocl::KernelArg::PtrReadOnly(usRGBGammaTab), ucoeffsarg, _1_3, _a); else k.args(srcarg, dstarg, ucoeffsarg, _1_3, _a); } else { ocl::KernelArg LabCbrtTabarg = ocl::KernelArg::PtrReadOnly(uLabCbrtTab); if (srgb) k.args(srcarg, dstarg, ocl::KernelArg::PtrReadOnly(usRGBGammaTab), LabCbrtTabarg, ucoeffsarg, un, vn); else k.args(srcarg, dstarg, LabCbrtTabarg, ucoeffsarg, un, vn); } } return k.run(dims, globalsize, NULL, false); } case CV_Lab2BGR: case CV_Lab2RGB: case CV_Lab2LBGR: case CV_Lab2LRGB: case CV_Luv2BGR: case CV_Luv2RGB: case CV_Luv2LBGR: case CV_Luv2LRGB: { if( dcn <= 0 ) dcn = 3; CV_Assert( scn == 3 && (dcn == 3 || dcn == 4) && (depth == CV_8U || depth == CV_32F) ); bidx = code == CV_Lab2BGR || code == CV_Lab2LBGR || code == CV_Luv2BGR || code == CV_Luv2LBGR ? 0 : 2; bool srgb = code == CV_Lab2BGR || code == CV_Lab2RGB || code == CV_Luv2BGR || code == CV_Luv2RGB; bool lab = code == CV_Lab2BGR || code == CV_Lab2RGB || code == CV_Lab2LBGR || code == CV_Lab2LRGB; float un, vn; k.create(format("%s2BGR", lab ? "Lab" : "Luv").c_str(), ocl::imgproc::cvtcolor_oclsrc, opts + format("-D dcn=%d -D bidx=%d%s", dcn, bidx, srgb ? " -D SRGB" : "")); if (k.empty()) return false; initLabTabs(); static UMat ucoeffs, usRGBInvGammaTab; if (srgb && usRGBInvGammaTab.empty()) Mat(1, GAMMA_TAB_SIZE*4, CV_32FC1, sRGBInvGammaTab).copyTo(usRGBInvGammaTab); { float coeffs[9]; const float * const _coeffs = XYZ2sRGB_D65, * const _whitept = D65; for( int i = 0; i < 3; i++ ) { coeffs[i+(bidx^2)*3] = _coeffs[i] * (lab ? _whitept[i] : 1); coeffs[i+3] = _coeffs[i+3] * (lab ? _whitept[i] : 1); coeffs[i+bidx*3] = _coeffs[i+6] * (lab ? _whitept[i] : 1); } float d = 1.f/(_whitept[0] + _whitept[1]*15 + _whitept[2]*3); un = 4*_whitept[0]*d; vn = 9*_whitept[1]*d; Mat(1, 9, CV_32FC1, coeffs).copyTo(ucoeffs); } _dst.create(sz, CV_MAKETYPE(depth, dcn)); dst = _dst.getUMat(); float lThresh = 0.008856f * 903.3f; float fThresh = 7.787f * 0.008856f + 16.0f / 116.0f; ocl::KernelArg srcarg = ocl::KernelArg::ReadOnlyNoSize(src), dstarg = ocl::KernelArg::WriteOnly(dst), coeffsarg = ocl::KernelArg::PtrReadOnly(ucoeffs); if (lab) { if (srgb) k.args(srcarg, dstarg, ocl::KernelArg::PtrReadOnly(usRGBInvGammaTab), coeffsarg, lThresh, fThresh); else k.args(srcarg, dstarg, coeffsarg, lThresh, fThresh); } else { if (srgb) k.args(srcarg, dstarg, ocl::KernelArg::PtrReadOnly(usRGBInvGammaTab), coeffsarg, un, vn); else k.args(srcarg, dstarg, coeffsarg, un, vn); } return k.run(dims, globalsize, NULL, false); } default: break; } if( !k.empty() ) { _dst.create(dstSz, CV_MAKETYPE(depth, dcn)); dst = _dst.getUMat(); k.args(ocl::KernelArg::ReadOnlyNoSize(src), ocl::KernelArg::WriteOnly(dst)); ok = k.run(dims, globalsize, NULL, false); } return ok; } #endif }//namespace cv ////////////////////////////////////////////////////////////////////////////////////////// // The main function // ////////////////////////////////////////////////////////////////////////////////////////// void cv::cvtColor( InputArray _src, OutputArray _dst, int code, int dcn ) { int stype = _src.type(); int scn = CV_MAT_CN(stype), depth = CV_MAT_DEPTH(stype), bidx; CV_OCL_RUN( _src.dims() <= 2 && _dst.isUMat() && !(depth == CV_8U && (code == CV_Luv2BGR || code == CV_Luv2RGB)), ocl_cvtColor(_src, _dst, code, dcn) ) Mat src = _src.getMat(), dst; Size sz = src.size(); CV_Assert( depth == CV_8U || depth == CV_16U || depth == CV_32F ); switch( code ) { case CV_BGR2BGRA: case CV_RGB2BGRA: case CV_BGRA2BGR: case CV_RGBA2BGR: case CV_RGB2BGR: case CV_BGRA2RGBA: CV_Assert( scn == 3 || scn == 4 ); dcn = code == CV_BGR2BGRA || code == CV_RGB2BGRA || code == CV_BGRA2RGBA ? 4 : 3; bidx = code == CV_BGR2BGRA || code == CV_BGRA2BGR ? 0 : 2; _dst.create( sz, CV_MAKETYPE(depth, dcn)); dst = _dst.getMat(); #if defined (HAVE_IPP) && (IPP_VERSION_MAJOR >= 7) CV_IPP_CHECK() { if( code == CV_BGR2BGRA) { if ( CvtColorIPPLoop(src, dst, IPPReorderFunctor(ippiSwapChannelsC3C4RTab[depth], 0, 1, 2)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_BGRA2BGR ) { if ( CvtColorIPPLoop(src, dst, IPPGeneralFunctor(ippiCopyAC4C3RTab[depth])) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_BGR2RGBA ) { if( CvtColorIPPLoop(src, dst, IPPReorderFunctor(ippiSwapChannelsC3C4RTab[depth], 2, 1, 0)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_RGBA2BGR ) { if( CvtColorIPPLoop(src, dst, IPPReorderFunctor(ippiSwapChannelsC4C3RTab[depth], 2, 1, 0)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_RGB2BGR ) { if( CvtColorIPPLoopCopy(src, dst, IPPReorderFunctor(ippiSwapChannelsC3RTab[depth], 2, 1, 0)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } #if IPP_VERSION_X100 >= 801 else if( code == CV_RGBA2BGRA ) { if( CvtColorIPPLoopCopy(src, dst, IPPReorderFunctor(ippiSwapChannelsC4RTab[depth], 2, 1, 0)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } #endif } #endif if( depth == CV_8U ) { #ifdef HAVE_TEGRA_OPTIMIZATION if(tegra::useTegra() && tegra::cvtBGR2RGB(src, dst, bidx)) break; #endif CvtColorLoop(src, dst, RGB2RGB(scn, dcn, bidx)); } else if( depth == CV_16U ) CvtColorLoop(src, dst, RGB2RGB(scn, dcn, bidx)); else CvtColorLoop(src, dst, RGB2RGB(scn, dcn, bidx)); break; case CV_BGR2BGR565: case CV_BGR2BGR555: case CV_RGB2BGR565: case CV_RGB2BGR555: case CV_BGRA2BGR565: case CV_BGRA2BGR555: case CV_RGBA2BGR565: case CV_RGBA2BGR555: CV_Assert( (scn == 3 || scn == 4) && depth == CV_8U ); _dst.create(sz, CV_8UC2); dst = _dst.getMat(); #if defined(HAVE_IPP) && 0 // breaks OCL accuracy tests CV_IPP_CHECK() { CV_SUPPRESS_DEPRECATED_START if (code == CV_BGR2BGR565 && scn == 3) { if (CvtColorIPPLoop(src, dst, IPPGeneralFunctor((ippiGeneralFunc)ippiBGRToBGR565_8u16u_C3R))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if (code == CV_BGRA2BGR565 && scn == 4) { if (CvtColorIPPLoopCopy(src, dst, IPPReorderGeneralFunctor(ippiSwapChannelsC4C3RTab[depth], (ippiGeneralFunc)ippiBGRToBGR565_8u16u_C3R, 0, 1, 2, depth))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if (code == CV_RGB2BGR565 && scn == 3) { if( CvtColorIPPLoopCopy(src, dst, IPPReorderGeneralFunctor(ippiSwapChannelsC3RTab[depth], (ippiGeneralFunc)ippiBGRToBGR565_8u16u_C3R, 2, 1, 0, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if (code == CV_RGBA2BGR565 && scn == 4) { if( CvtColorIPPLoopCopy(src, dst, IPPReorderGeneralFunctor(ippiSwapChannelsC4C3RTab[depth], (ippiGeneralFunc)ippiBGRToBGR565_8u16u_C3R, 2, 1, 0, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } CV_SUPPRESS_DEPRECATED_END } #endif #ifdef HAVE_TEGRA_OPTIMIZATION if(code == CV_BGR2BGR565 || code == CV_BGRA2BGR565 || code == CV_RGB2BGR565 || code == CV_RGBA2BGR565) if(tegra::useTegra() && tegra::cvtRGB2RGB565(src, dst, code == CV_RGB2BGR565 || code == CV_RGBA2BGR565 ? 0 : 2)) break; #endif CvtColorLoop(src, dst, RGB2RGB5x5(scn, code == CV_BGR2BGR565 || code == CV_BGR2BGR555 || code == CV_BGRA2BGR565 || code == CV_BGRA2BGR555 ? 0 : 2, code == CV_BGR2BGR565 || code == CV_RGB2BGR565 || code == CV_BGRA2BGR565 || code == CV_RGBA2BGR565 ? 6 : 5 // green bits )); break; case CV_BGR5652BGR: case CV_BGR5552BGR: case CV_BGR5652RGB: case CV_BGR5552RGB: case CV_BGR5652BGRA: case CV_BGR5552BGRA: case CV_BGR5652RGBA: case CV_BGR5552RGBA: if(dcn <= 0) dcn = (code==CV_BGR5652BGRA || code==CV_BGR5552BGRA || code==CV_BGR5652RGBA || code==CV_BGR5552RGBA) ? 4 : 3; CV_Assert( (dcn == 3 || dcn == 4) && scn == 2 && depth == CV_8U ); _dst.create(sz, CV_MAKETYPE(depth, dcn)); dst = _dst.getMat(); #ifdef HAVE_IPP CV_IPP_CHECK() { CV_SUPPRESS_DEPRECATED_START if (code == CV_BGR5652BGR && dcn == 3) { if (CvtColorIPPLoop(src, dst, IPPGeneralFunctor((ippiGeneralFunc)ippiBGR565ToBGR_16u8u_C3R))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if (code == CV_BGR5652RGB && dcn == 3) { if (CvtColorIPPLoop(src, dst, IPPGeneralReorderFunctor((ippiGeneralFunc)ippiBGR565ToBGR_16u8u_C3R, ippiSwapChannelsC3RTab[depth], 2, 1, 0, depth))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if (code == CV_BGR5652BGRA && dcn == 4) { if (CvtColorIPPLoop(src, dst, IPPGeneralReorderFunctor((ippiGeneralFunc)ippiBGR565ToBGR_16u8u_C3R, ippiSwapChannelsC3C4RTab[depth], 0, 1, 2, depth))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if (code == CV_BGR5652RGBA && dcn == 4) { if (CvtColorIPPLoop(src, dst, IPPGeneralReorderFunctor((ippiGeneralFunc)ippiBGR565ToBGR_16u8u_C3R, ippiSwapChannelsC3C4RTab[depth], 2, 1, 0, depth))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } CV_SUPPRESS_DEPRECATED_END } #endif CvtColorLoop(src, dst, RGB5x52RGB(dcn, code == CV_BGR5652BGR || code == CV_BGR5552BGR || code == CV_BGR5652BGRA || code == CV_BGR5552BGRA ? 0 : 2, // blue idx code == CV_BGR5652BGR || code == CV_BGR5652RGB || code == CV_BGR5652BGRA || code == CV_BGR5652RGBA ? 6 : 5 // green bits )); break; case CV_BGR2GRAY: case CV_BGRA2GRAY: case CV_RGB2GRAY: case CV_RGBA2GRAY: CV_Assert( scn == 3 || scn == 4 ); _dst.create(sz, CV_MAKETYPE(depth, 1)); dst = _dst.getMat(); #if defined (HAVE_IPP) && (IPP_VERSION_MAJOR >= 7) CV_IPP_CHECK() { if( code == CV_BGR2GRAY && depth == CV_32F ) { if( CvtColorIPPLoop(src, dst, IPPColor2GrayFunctor(ippiColor2GrayC3Tab[depth])) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_RGB2GRAY && depth == CV_32F ) { if( CvtColorIPPLoop(src, dst, IPPGeneralFunctor(ippiRGB2GrayC3Tab[depth])) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_BGRA2GRAY && depth == CV_32F ) { if( CvtColorIPPLoop(src, dst, IPPColor2GrayFunctor(ippiColor2GrayC4Tab[depth])) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_RGBA2GRAY && depth == CV_32F ) { if( CvtColorIPPLoop(src, dst, IPPGeneralFunctor(ippiRGB2GrayC4Tab[depth])) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } } #endif bidx = code == CV_BGR2GRAY || code == CV_BGRA2GRAY ? 0 : 2; if( depth == CV_8U ) { #ifdef HAVE_TEGRA_OPTIMIZATION if(tegra::useTegra() && tegra::cvtRGB2Gray(src, dst, bidx)) break; #endif CvtColorLoop(src, dst, RGB2Gray(scn, bidx, 0)); } else if( depth == CV_16U ) CvtColorLoop(src, dst, RGB2Gray(scn, bidx, 0)); else CvtColorLoop(src, dst, RGB2Gray(scn, bidx, 0)); break; case CV_BGR5652GRAY: case CV_BGR5552GRAY: CV_Assert( scn == 2 && depth == CV_8U ); _dst.create(sz, CV_8UC1); dst = _dst.getMat(); CvtColorLoop(src, dst, RGB5x52Gray(code == CV_BGR5652GRAY ? 6 : 5)); break; case CV_GRAY2BGR: case CV_GRAY2BGRA: if( dcn <= 0 ) dcn = (code==CV_GRAY2BGRA) ? 4 : 3; CV_Assert( scn == 1 && (dcn == 3 || dcn == 4)); _dst.create(sz, CV_MAKETYPE(depth, dcn)); dst = _dst.getMat(); #if defined (HAVE_IPP) && (IPP_VERSION_MAJOR >= 7) CV_IPP_CHECK() { if( code == CV_GRAY2BGR ) { if( CvtColorIPPLoop(src, dst, IPPGray2BGRFunctor(ippiCopyP3C3RTab[depth])) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_GRAY2BGRA ) { if( CvtColorIPPLoop(src, dst, IPPGray2BGRAFunctor(ippiCopyP3C3RTab[depth], ippiSwapChannelsC3C4RTab[depth], depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } } #endif if( depth == CV_8U ) { #ifdef HAVE_TEGRA_OPTIMIZATION if(tegra::useTegra() && tegra::cvtGray2RGB(src, dst)) break; #endif CvtColorLoop(src, dst, Gray2RGB(dcn)); } else if( depth == CV_16U ) CvtColorLoop(src, dst, Gray2RGB(dcn)); else CvtColorLoop(src, dst, Gray2RGB(dcn)); break; case CV_GRAY2BGR565: case CV_GRAY2BGR555: CV_Assert( scn == 1 && depth == CV_8U ); _dst.create(sz, CV_8UC2); dst = _dst.getMat(); CvtColorLoop(src, dst, Gray2RGB5x5(code == CV_GRAY2BGR565 ? 6 : 5)); break; case CV_BGR2YCrCb: case CV_RGB2YCrCb: case CV_BGR2YUV: case CV_RGB2YUV: { CV_Assert( scn == 3 || scn == 4 ); bidx = code == CV_BGR2YCrCb || code == CV_BGR2YUV ? 0 : 2; static const float yuv_f[] = { 0.114f, 0.587f, 0.299f, 0.492f, 0.877f }; static const int yuv_i[] = { B2Y, G2Y, R2Y, 8061, 14369 }; const float* coeffs_f = code == CV_BGR2YCrCb || code == CV_RGB2YCrCb ? 0 : yuv_f; const int* coeffs_i = code == CV_BGR2YCrCb || code == CV_RGB2YCrCb ? 0 : yuv_i; _dst.create(sz, CV_MAKETYPE(depth, 3)); dst = _dst.getMat(); #if defined HAVE_IPP && 0 CV_IPP_CHECK() { if (code == CV_RGB2YUV && scn == 3 && depth == CV_8U) { if (CvtColorIPPLoop(src, dst, IPPGeneralFunctor((ippiGeneralFunc)ippiRGBToYUV_8u_C3R))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if (code == CV_BGR2YUV && scn == 3 && depth == CV_8U) { if (CvtColorIPPLoop(src, dst, IPPReorderGeneralFunctor(ippiSwapChannelsC3RTab[depth], (ippiGeneralFunc)ippiRGBToYUV_8u_C3R, 2, 1, 0, depth))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if (code == CV_RGB2YUV && scn == 4 && depth == CV_8U) { if (CvtColorIPPLoop(src, dst, IPPReorderGeneralFunctor(ippiSwapChannelsC4C3RTab[depth], (ippiGeneralFunc)ippiRGBToYUV_8u_C3R, 0, 1, 2, depth))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if (code == CV_BGR2YUV && scn == 4 && depth == CV_8U) { if (CvtColorIPPLoop(src, dst, IPPReorderGeneralFunctor(ippiSwapChannelsC4C3RTab[depth], (ippiGeneralFunc)ippiRGBToYUV_8u_C3R, 2, 1, 0, depth))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } } #endif if( depth == CV_8U ) { #ifdef HAVE_TEGRA_OPTIMIZATION if((code == CV_RGB2YCrCb || code == CV_BGR2YCrCb) && tegra::useTegra() && tegra::cvtRGB2YCrCb(src, dst, bidx)) break; #endif CvtColorLoop(src, dst, RGB2YCrCb_i(scn, bidx, coeffs_i)); } else if( depth == CV_16U ) CvtColorLoop(src, dst, RGB2YCrCb_i(scn, bidx, coeffs_i)); else CvtColorLoop(src, dst, RGB2YCrCb_f(scn, bidx, coeffs_f)); } break; case CV_YCrCb2BGR: case CV_YCrCb2RGB: case CV_YUV2BGR: case CV_YUV2RGB: { if( dcn <= 0 ) dcn = 3; CV_Assert( scn == 3 && (dcn == 3 || dcn == 4) ); bidx = code == CV_YCrCb2BGR || code == CV_YUV2BGR ? 0 : 2; static const float yuv_f[] = { 2.032f, -0.395f, -0.581f, 1.140f }; static const int yuv_i[] = { 33292, -6472, -9519, 18678 }; const float* coeffs_f = code == CV_YCrCb2BGR || code == CV_YCrCb2RGB ? 0 : yuv_f; const int* coeffs_i = code == CV_YCrCb2BGR || code == CV_YCrCb2RGB ? 0 : yuv_i; _dst.create(sz, CV_MAKETYPE(depth, dcn)); dst = _dst.getMat(); #if defined HAVE_IPP && 0 CV_IPP_CHECK() { if (code == CV_YUV2RGB && dcn == 3 && depth == CV_8U) { if (CvtColorIPPLoop(src, dst, IPPGeneralFunctor((ippiGeneralFunc)ippiYUVToRGB_8u_C3R))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if (code == CV_YUV2BGR && dcn == 3 && depth == CV_8U) { if (CvtColorIPPLoop(src, dst, IPPGeneralReorderFunctor((ippiGeneralFunc)ippiYUVToRGB_8u_C3R, ippiSwapChannelsC3RTab[depth], 2, 1, 0, depth))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if (code == CV_YUV2RGB && dcn == 4 && depth == CV_8U) { if (CvtColorIPPLoop(src, dst, IPPGeneralReorderFunctor((ippiGeneralFunc)ippiYUVToRGB_8u_C3R, ippiSwapChannelsC3C4RTab[depth], 0, 1, 2, depth))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if (code == CV_YUV2BGR && dcn == 4 && depth == CV_8U) { if (CvtColorIPPLoop(src, dst, IPPGeneralReorderFunctor((ippiGeneralFunc)ippiYUVToRGB_8u_C3R, ippiSwapChannelsC3C4RTab[depth], 2, 1, 0, depth))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } } #endif if( depth == CV_8U ) CvtColorLoop(src, dst, YCrCb2RGB_i(dcn, bidx, coeffs_i)); else if( depth == CV_16U ) CvtColorLoop(src, dst, YCrCb2RGB_i(dcn, bidx, coeffs_i)); else CvtColorLoop(src, dst, YCrCb2RGB_f(dcn, bidx, coeffs_f)); } break; case CV_BGR2XYZ: case CV_RGB2XYZ: CV_Assert( scn == 3 || scn == 4 ); bidx = code == CV_BGR2XYZ ? 0 : 2; _dst.create(sz, CV_MAKETYPE(depth, 3)); dst = _dst.getMat(); #if defined (HAVE_IPP) && (IPP_VERSION_MAJOR >= 7) CV_IPP_CHECK() { if( code == CV_BGR2XYZ && scn == 3 && depth != CV_32F ) { if( CvtColorIPPLoopCopy(src, dst, IPPReorderGeneralFunctor(ippiSwapChannelsC3RTab[depth], ippiRGB2XYZTab[depth], 2, 1, 0, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_BGR2XYZ && scn == 4 && depth != CV_32F ) { if( CvtColorIPPLoop(src, dst, IPPReorderGeneralFunctor(ippiSwapChannelsC4C3RTab[depth], ippiRGB2XYZTab[depth], 2, 1, 0, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_RGB2XYZ && scn == 3 && depth != CV_32F ) { if( CvtColorIPPLoopCopy(src, dst, IPPGeneralFunctor(ippiRGB2XYZTab[depth])) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_RGB2XYZ && scn == 4 && depth != CV_32F ) { if( CvtColorIPPLoop(src, dst, IPPReorderGeneralFunctor(ippiSwapChannelsC4C3RTab[depth], ippiRGB2XYZTab[depth], 0, 1, 2, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } } #endif if( depth == CV_8U ) CvtColorLoop(src, dst, RGB2XYZ_i(scn, bidx, 0)); else if( depth == CV_16U ) CvtColorLoop(src, dst, RGB2XYZ_i(scn, bidx, 0)); else CvtColorLoop(src, dst, RGB2XYZ_f(scn, bidx, 0)); break; case CV_XYZ2BGR: case CV_XYZ2RGB: if( dcn <= 0 ) dcn = 3; CV_Assert( scn == 3 && (dcn == 3 || dcn == 4) ); bidx = code == CV_XYZ2BGR ? 0 : 2; _dst.create(sz, CV_MAKETYPE(depth, dcn)); dst = _dst.getMat(); #if defined (HAVE_IPP) && (IPP_VERSION_MAJOR >= 7) CV_IPP_CHECK() { if( code == CV_XYZ2BGR && dcn == 3 && depth != CV_32F ) { if( CvtColorIPPLoopCopy(src, dst, IPPGeneralReorderFunctor(ippiXYZ2RGBTab[depth], ippiSwapChannelsC3RTab[depth], 2, 1, 0, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_XYZ2BGR && dcn == 4 && depth != CV_32F ) { if( CvtColorIPPLoop(src, dst, IPPGeneralReorderFunctor(ippiXYZ2RGBTab[depth], ippiSwapChannelsC3C4RTab[depth], 2, 1, 0, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } if( code == CV_XYZ2RGB && dcn == 3 && depth != CV_32F ) { if( CvtColorIPPLoopCopy(src, dst, IPPGeneralFunctor(ippiXYZ2RGBTab[depth])) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_XYZ2RGB && dcn == 4 && depth != CV_32F ) { if( CvtColorIPPLoop(src, dst, IPPGeneralReorderFunctor(ippiXYZ2RGBTab[depth], ippiSwapChannelsC3C4RTab[depth], 0, 1, 2, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } } #endif if( depth == CV_8U ) CvtColorLoop(src, dst, XYZ2RGB_i(dcn, bidx, 0)); else if( depth == CV_16U ) CvtColorLoop(src, dst, XYZ2RGB_i(dcn, bidx, 0)); else CvtColorLoop(src, dst, XYZ2RGB_f(dcn, bidx, 0)); break; case CV_BGR2HSV: case CV_RGB2HSV: case CV_BGR2HSV_FULL: case CV_RGB2HSV_FULL: case CV_BGR2HLS: case CV_RGB2HLS: case CV_BGR2HLS_FULL: case CV_RGB2HLS_FULL: { CV_Assert( (scn == 3 || scn == 4) && (depth == CV_8U || depth == CV_32F) ); bidx = code == CV_BGR2HSV || code == CV_BGR2HLS || code == CV_BGR2HSV_FULL || code == CV_BGR2HLS_FULL ? 0 : 2; int hrange = depth == CV_32F ? 360 : code == CV_BGR2HSV || code == CV_RGB2HSV || code == CV_BGR2HLS || code == CV_RGB2HLS ? 180 : 256; _dst.create(sz, CV_MAKETYPE(depth, 3)); dst = _dst.getMat(); #if defined (HAVE_IPP) && (IPP_VERSION_MAJOR >= 7) CV_IPP_CHECK() { if( depth == CV_8U || depth == CV_16U ) { #if 0 // breaks OCL accuracy tests if( code == CV_BGR2HSV_FULL && scn == 3 ) { if( CvtColorIPPLoopCopy(src, dst, IPPReorderGeneralFunctor(ippiSwapChannelsC3RTab[depth], ippiRGB2HSVTab[depth], 2, 1, 0, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_BGR2HSV_FULL && scn == 4 ) { if( CvtColorIPPLoop(src, dst, IPPReorderGeneralFunctor(ippiSwapChannelsC4C3RTab[depth], ippiRGB2HSVTab[depth], 2, 1, 0, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_RGB2HSV_FULL && scn == 4 ) { if( CvtColorIPPLoop(src, dst, IPPReorderGeneralFunctor(ippiSwapChannelsC4C3RTab[depth], ippiRGB2HSVTab[depth], 0, 1, 2, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else #endif if( code == CV_RGB2HSV_FULL && scn == 3 && depth == CV_16U ) { if( CvtColorIPPLoopCopy(src, dst, IPPGeneralFunctor(ippiRGB2HSVTab[depth])) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_BGR2HLS_FULL && scn == 3 ) { if( CvtColorIPPLoopCopy(src, dst, IPPReorderGeneralFunctor(ippiSwapChannelsC3RTab[depth], ippiRGB2HLSTab[depth], 2, 1, 0, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_BGR2HLS_FULL && scn == 4 ) { if( CvtColorIPPLoop(src, dst, IPPReorderGeneralFunctor(ippiSwapChannelsC4C3RTab[depth], ippiRGB2HLSTab[depth], 2, 1, 0, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_RGB2HLS_FULL && scn == 3 ) { if( CvtColorIPPLoopCopy(src, dst, IPPGeneralFunctor(ippiRGB2HLSTab[depth])) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_RGB2HLS_FULL && scn == 4 ) { if( CvtColorIPPLoop(src, dst, IPPReorderGeneralFunctor(ippiSwapChannelsC4C3RTab[depth], ippiRGB2HLSTab[depth], 0, 1, 2, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } } } #endif if( code == CV_BGR2HSV || code == CV_RGB2HSV || code == CV_BGR2HSV_FULL || code == CV_RGB2HSV_FULL ) { #ifdef HAVE_TEGRA_OPTIMIZATION if(tegra::useTegra() && tegra::cvtRGB2HSV(src, dst, bidx, hrange)) break; #endif if( depth == CV_8U ) CvtColorLoop(src, dst, RGB2HSV_b(scn, bidx, hrange)); else CvtColorLoop(src, dst, RGB2HSV_f(scn, bidx, (float)hrange)); } else { if( depth == CV_8U ) CvtColorLoop(src, dst, RGB2HLS_b(scn, bidx, hrange)); else CvtColorLoop(src, dst, RGB2HLS_f(scn, bidx, (float)hrange)); } } break; case CV_HSV2BGR: case CV_HSV2RGB: case CV_HSV2BGR_FULL: case CV_HSV2RGB_FULL: case CV_HLS2BGR: case CV_HLS2RGB: case CV_HLS2BGR_FULL: case CV_HLS2RGB_FULL: { if( dcn <= 0 ) dcn = 3; CV_Assert( scn == 3 && (dcn == 3 || dcn == 4) && (depth == CV_8U || depth == CV_32F) ); bidx = code == CV_HSV2BGR || code == CV_HLS2BGR || code == CV_HSV2BGR_FULL || code == CV_HLS2BGR_FULL ? 0 : 2; int hrange = depth == CV_32F ? 360 : code == CV_HSV2BGR || code == CV_HSV2RGB || code == CV_HLS2BGR || code == CV_HLS2RGB ? 180 : 255; _dst.create(sz, CV_MAKETYPE(depth, dcn)); dst = _dst.getMat(); #if defined (HAVE_IPP) && (IPP_VERSION_MAJOR >= 7) CV_IPP_CHECK() { if( depth == CV_8U || depth == CV_16U ) { if( code == CV_HSV2BGR_FULL && dcn == 3 ) { if( CvtColorIPPLoopCopy(src, dst, IPPGeneralReorderFunctor(ippiHSV2RGBTab[depth], ippiSwapChannelsC3RTab[depth], 2, 1, 0, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_HSV2BGR_FULL && dcn == 4 ) { if( CvtColorIPPLoop(src, dst, IPPGeneralReorderFunctor(ippiHSV2RGBTab[depth], ippiSwapChannelsC3C4RTab[depth], 2, 1, 0, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_HSV2RGB_FULL && dcn == 3 ) { if( CvtColorIPPLoopCopy(src, dst, IPPGeneralFunctor(ippiHSV2RGBTab[depth])) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_HSV2RGB_FULL && dcn == 4 ) { if( CvtColorIPPLoop(src, dst, IPPGeneralReorderFunctor(ippiHSV2RGBTab[depth], ippiSwapChannelsC3C4RTab[depth], 0, 1, 2, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_HLS2BGR_FULL && dcn == 3 ) { if( CvtColorIPPLoopCopy(src, dst, IPPGeneralReorderFunctor(ippiHLS2RGBTab[depth], ippiSwapChannelsC3RTab[depth], 2, 1, 0, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_HLS2BGR_FULL && dcn == 4 ) { if( CvtColorIPPLoop(src, dst, IPPGeneralReorderFunctor(ippiHLS2RGBTab[depth], ippiSwapChannelsC3C4RTab[depth], 2, 1, 0, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_HLS2RGB_FULL && dcn == 3 ) { if( CvtColorIPPLoopCopy(src, dst, IPPGeneralFunctor(ippiHLS2RGBTab[depth])) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_HLS2RGB_FULL && dcn == 4 ) { if( CvtColorIPPLoop(src, dst, IPPGeneralReorderFunctor(ippiHLS2RGBTab[depth], ippiSwapChannelsC3C4RTab[depth], 0, 1, 2, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } } } #endif if( code == CV_HSV2BGR || code == CV_HSV2RGB || code == CV_HSV2BGR_FULL || code == CV_HSV2RGB_FULL ) { if( depth == CV_8U ) CvtColorLoop(src, dst, HSV2RGB_b(dcn, bidx, hrange)); else CvtColorLoop(src, dst, HSV2RGB_f(dcn, bidx, (float)hrange)); } else { if( depth == CV_8U ) CvtColorLoop(src, dst, HLS2RGB_b(dcn, bidx, hrange)); else CvtColorLoop(src, dst, HLS2RGB_f(dcn, bidx, (float)hrange)); } } break; case CV_BGR2Lab: case CV_RGB2Lab: case CV_LBGR2Lab: case CV_LRGB2Lab: case CV_BGR2Luv: case CV_RGB2Luv: case CV_LBGR2Luv: case CV_LRGB2Luv: { CV_Assert( (scn == 3 || scn == 4) && (depth == CV_8U || depth == CV_32F) ); bidx = code == CV_BGR2Lab || code == CV_BGR2Luv || code == CV_LBGR2Lab || code == CV_LBGR2Luv ? 0 : 2; bool srgb = code == CV_BGR2Lab || code == CV_RGB2Lab || code == CV_BGR2Luv || code == CV_RGB2Luv; _dst.create(sz, CV_MAKETYPE(depth, 3)); dst = _dst.getMat(); #if defined HAVE_IPP && 0 CV_IPP_CHECK() { if (code == CV_LBGR2Lab && scn == 3 && depth == CV_8U) { if (CvtColorIPPLoop(src, dst, IPPGeneralFunctor((ippiGeneralFunc)ippiBGRToLab_8u_C3R))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if (code == CV_LBGR2Lab && scn == 4 && depth == CV_8U) { if (CvtColorIPPLoop(src, dst, IPPReorderGeneralFunctor(ippiSwapChannelsC4C3RTab[depth], (ippiGeneralFunc)ippiBGRToLab_8u_C3R, 0, 1, 2, depth))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if (code == CV_LRGB2Lab && scn == 3 && depth == CV_8U) // slower than OpenCV { if (CvtColorIPPLoop(src, dst, IPPReorderGeneralFunctor(ippiSwapChannelsC3RTab[depth], (ippiGeneralFunc)ippiBGRToLab_8u_C3R, 2, 1, 0, depth))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if (code == CV_LRGB2Lab && scn == 4 && depth == CV_8U) // slower than OpenCV { if (CvtColorIPPLoop(src, dst, IPPReorderGeneralFunctor(ippiSwapChannelsC4C3RTab[depth], (ippiGeneralFunc)ippiBGRToLab_8u_C3R, 2, 1, 0, depth))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if (code == CV_LRGB2Luv && scn == 3) { if (CvtColorIPPLoop(src, dst, IPPGeneralFunctor(ippiRGBToLUVTab[depth]))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if (code == CV_LRGB2Luv && scn == 4) { if (CvtColorIPPLoop(src, dst, IPPReorderGeneralFunctor(ippiSwapChannelsC4C3RTab[depth], ippiRGBToLUVTab[depth], 0, 1, 2, depth))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if (code == CV_LBGR2Luv && scn == 3) { if (CvtColorIPPLoop(src, dst, IPPReorderGeneralFunctor(ippiSwapChannelsC3RTab[depth], ippiRGBToLUVTab[depth], 2, 1, 0, depth))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if (code == CV_LBGR2Luv && scn == 4) { if (CvtColorIPPLoop(src, dst, IPPReorderGeneralFunctor(ippiSwapChannelsC4C3RTab[depth], ippiRGBToLUVTab[depth], 2, 1, 0, depth))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } } #endif if( code == CV_BGR2Lab || code == CV_RGB2Lab || code == CV_LBGR2Lab || code == CV_LRGB2Lab ) { if( depth == CV_8U ) CvtColorLoop(src, dst, RGB2Lab_b(scn, bidx, 0, 0, srgb)); else CvtColorLoop(src, dst, RGB2Lab_f(scn, bidx, 0, 0, srgb)); } else { if( depth == CV_8U ) CvtColorLoop(src, dst, RGB2Luv_b(scn, bidx, 0, 0, srgb)); else CvtColorLoop(src, dst, RGB2Luv_f(scn, bidx, 0, 0, srgb)); } } break; case CV_Lab2BGR: case CV_Lab2RGB: case CV_Lab2LBGR: case CV_Lab2LRGB: case CV_Luv2BGR: case CV_Luv2RGB: case CV_Luv2LBGR: case CV_Luv2LRGB: { if( dcn <= 0 ) dcn = 3; CV_Assert( scn == 3 && (dcn == 3 || dcn == 4) && (depth == CV_8U || depth == CV_32F) ); bidx = code == CV_Lab2BGR || code == CV_Luv2BGR || code == CV_Lab2LBGR || code == CV_Luv2LBGR ? 0 : 2; bool srgb = code == CV_Lab2BGR || code == CV_Lab2RGB || code == CV_Luv2BGR || code == CV_Luv2RGB; _dst.create(sz, CV_MAKETYPE(depth, dcn)); dst = _dst.getMat(); #if defined HAVE_IPP && 0 CV_IPP_CHECK() { if( code == CV_Lab2LBGR && dcn == 3 && depth == CV_8U) { if( CvtColorIPPLoop(src, dst, IPPGeneralFunctor((ippiGeneralFunc)ippiLabToBGR_8u_C3R)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_Lab2LBGR && dcn == 4 && depth == CV_8U ) { if( CvtColorIPPLoop(src, dst, IPPGeneralReorderFunctor((ippiGeneralFunc)ippiLabToBGR_8u_C3R, ippiSwapChannelsC3C4RTab[depth], 0, 1, 2, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } if( code == CV_Lab2LRGB && dcn == 3 && depth == CV_8U ) { if( CvtColorIPPLoop(src, dst, IPPGeneralReorderFunctor((ippiGeneralFunc)ippiLabToBGR_8u_C3R, ippiSwapChannelsC3RTab[depth], 2, 1, 0, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } else if( code == CV_Lab2LRGB && dcn == 4 && depth == CV_8U ) { if( CvtColorIPPLoop(src, dst, IPPGeneralReorderFunctor((ippiGeneralFunc)ippiLabToBGR_8u_C3R, ippiSwapChannelsC3C4RTab[depth], 2, 1, 0, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } if( code == CV_Luv2LRGB && dcn == 3 ) { if( CvtColorIPPLoop(src, dst, IPPGeneralFunctor(ippiLUVToRGBTab[depth])) ) return; } else if( code == CV_Luv2LRGB && dcn == 4 ) { if( CvtColorIPPLoop(src, dst, IPPGeneralReorderFunctor(ippiLUVToRGBTab[depth], ippiSwapChannelsC3C4RTab[depth], 0, 1, 2, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } } if( code == CV_Luv2LBGR && dcn == 3 ) { if( CvtColorIPPLoop(src, dst, IPPGeneralReorderFunctor(ippiLUVToRGBTab[depth], ippiSwapChannelsC3RTab[depth], 2, 1, 0, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } } else if( code == CV_Luv2LBGR && dcn == 4 ) { if( CvtColorIPPLoop(src, dst, IPPGeneralReorderFunctor(ippiLUVToRGBTab[depth], ippiSwapChannelsC3C4RTab[depth], 2, 1, 0, depth)) ) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } } } #endif if( code == CV_Lab2BGR || code == CV_Lab2RGB || code == CV_Lab2LBGR || code == CV_Lab2LRGB ) { if( depth == CV_8U ) CvtColorLoop(src, dst, Lab2RGB_b(dcn, bidx, 0, 0, srgb)); else CvtColorLoop(src, dst, Lab2RGB_f(dcn, bidx, 0, 0, srgb)); } else { if( depth == CV_8U ) CvtColorLoop(src, dst, Luv2RGB_b(dcn, bidx, 0, 0, srgb)); else CvtColorLoop(src, dst, Luv2RGB_f(dcn, bidx, 0, 0, srgb)); } } break; case CV_BayerBG2GRAY: case CV_BayerGB2GRAY: case CV_BayerRG2GRAY: case CV_BayerGR2GRAY: case CV_BayerBG2BGR: case CV_BayerGB2BGR: case CV_BayerRG2BGR: case CV_BayerGR2BGR: case CV_BayerBG2BGR_VNG: case CV_BayerGB2BGR_VNG: case CV_BayerRG2BGR_VNG: case CV_BayerGR2BGR_VNG: case CV_BayerBG2BGR_EA: case CV_BayerGB2BGR_EA: case CV_BayerRG2BGR_EA: case CV_BayerGR2BGR_EA: demosaicing(src, _dst, code, dcn); break; case CV_YUV2BGR_NV21: case CV_YUV2RGB_NV21: case CV_YUV2BGR_NV12: case CV_YUV2RGB_NV12: case CV_YUV2BGRA_NV21: case CV_YUV2RGBA_NV21: case CV_YUV2BGRA_NV12: case CV_YUV2RGBA_NV12: { // http://www.fourcc.org/yuv.php#NV21 == yuv420sp -> a plane of 8 bit Y samples followed by an interleaved V/U plane containing 8 bit 2x2 subsampled chroma samples // http://www.fourcc.org/yuv.php#NV12 -> a plane of 8 bit Y samples followed by an interleaved U/V plane containing 8 bit 2x2 subsampled colour difference samples if (dcn <= 0) dcn = (code==CV_YUV420sp2BGRA || code==CV_YUV420sp2RGBA || code==CV_YUV2BGRA_NV12 || code==CV_YUV2RGBA_NV12) ? 4 : 3; const int bIdx = (code==CV_YUV2BGR_NV21 || code==CV_YUV2BGRA_NV21 || code==CV_YUV2BGR_NV12 || code==CV_YUV2BGRA_NV12) ? 0 : 2; const int uIdx = (code==CV_YUV2BGR_NV21 || code==CV_YUV2BGRA_NV21 || code==CV_YUV2RGB_NV21 || code==CV_YUV2RGBA_NV21) ? 1 : 0; CV_Assert( dcn == 3 || dcn == 4 ); CV_Assert( sz.width % 2 == 0 && sz.height % 3 == 0 && depth == CV_8U ); Size dstSz(sz.width, sz.height * 2 / 3); _dst.create(dstSz, CV_MAKETYPE(depth, dcn)); dst = _dst.getMat(); int srcstep = (int)src.step; const uchar* y = src.ptr(); const uchar* uv = y + srcstep * dstSz.height; switch(dcn*100 + bIdx * 10 + uIdx) { case 300: cvtYUV420sp2RGB<0, 0> (dst, srcstep, y, uv); break; case 301: cvtYUV420sp2RGB<0, 1> (dst, srcstep, y, uv); break; case 320: cvtYUV420sp2RGB<2, 0> (dst, srcstep, y, uv); break; case 321: cvtYUV420sp2RGB<2, 1> (dst, srcstep, y, uv); break; case 400: cvtYUV420sp2RGBA<0, 0>(dst, srcstep, y, uv); break; case 401: cvtYUV420sp2RGBA<0, 1>(dst, srcstep, y, uv); break; case 420: cvtYUV420sp2RGBA<2, 0>(dst, srcstep, y, uv); break; case 421: cvtYUV420sp2RGBA<2, 1>(dst, srcstep, y, uv); break; default: CV_Error( CV_StsBadFlag, "Unknown/unsupported color conversion code" ); break; }; } break; case CV_YUV2BGR_YV12: case CV_YUV2RGB_YV12: case CV_YUV2BGRA_YV12: case CV_YUV2RGBA_YV12: case CV_YUV2BGR_IYUV: case CV_YUV2RGB_IYUV: case CV_YUV2BGRA_IYUV: case CV_YUV2RGBA_IYUV: { //http://www.fourcc.org/yuv.php#YV12 == yuv420p -> It comprises an NxM Y plane followed by (N/2)x(M/2) V and U planes. //http://www.fourcc.org/yuv.php#IYUV == I420 -> It comprises an NxN Y plane followed by (N/2)x(N/2) U and V planes if (dcn <= 0) dcn = (code==CV_YUV2BGRA_YV12 || code==CV_YUV2RGBA_YV12 || code==CV_YUV2RGBA_IYUV || code==CV_YUV2BGRA_IYUV) ? 4 : 3; const int bIdx = (code==CV_YUV2BGR_YV12 || code==CV_YUV2BGRA_YV12 || code==CV_YUV2BGR_IYUV || code==CV_YUV2BGRA_IYUV) ? 0 : 2; const int uIdx = (code==CV_YUV2BGR_YV12 || code==CV_YUV2RGB_YV12 || code==CV_YUV2BGRA_YV12 || code==CV_YUV2RGBA_YV12) ? 1 : 0; CV_Assert( dcn == 3 || dcn == 4 ); CV_Assert( sz.width % 2 == 0 && sz.height % 3 == 0 && depth == CV_8U ); Size dstSz(sz.width, sz.height * 2 / 3); _dst.create(dstSz, CV_MAKETYPE(depth, dcn)); dst = _dst.getMat(); int srcstep = (int)src.step; const uchar* y = src.ptr(); const uchar* u = y + srcstep * dstSz.height; const uchar* v = y + srcstep * (dstSz.height + dstSz.height/4) + (dstSz.width/2) * ((dstSz.height % 4)/2); int ustepIdx = 0; int vstepIdx = dstSz.height % 4 == 2 ? 1 : 0; if(uIdx == 1) { std::swap(u ,v), std::swap(ustepIdx, vstepIdx); } switch(dcn*10 + bIdx) { case 30: cvtYUV420p2RGB<0>(dst, srcstep, y, u, v, ustepIdx, vstepIdx); break; case 32: cvtYUV420p2RGB<2>(dst, srcstep, y, u, v, ustepIdx, vstepIdx); break; case 40: cvtYUV420p2RGBA<0>(dst, srcstep, y, u, v, ustepIdx, vstepIdx); break; case 42: cvtYUV420p2RGBA<2>(dst, srcstep, y, u, v, ustepIdx, vstepIdx); break; default: CV_Error( CV_StsBadFlag, "Unknown/unsupported color conversion code" ); break; }; } break; case CV_YUV2GRAY_420: { if (dcn <= 0) dcn = 1; CV_Assert( dcn == 1 ); CV_Assert( sz.width % 2 == 0 && sz.height % 3 == 0 && depth == CV_8U ); Size dstSz(sz.width, sz.height * 2 / 3); _dst.create(dstSz, CV_MAKETYPE(depth, dcn)); dst = _dst.getMat(); #if defined HAVE_IPP CV_IPP_CHECK() { if (ippStsNoErr == ippiCopy_8u_C1R(src.data, (int)src.step, dst.data, (int)dst.step, ippiSize(dstSz.width, dstSz.height))) { CV_IMPL_ADD(CV_IMPL_IPP); return; } setIppErrorStatus(); } #endif src(Range(0, dstSz.height), Range::all()).copyTo(dst); } break; case CV_RGB2YUV_YV12: case CV_BGR2YUV_YV12: case CV_RGBA2YUV_YV12: case CV_BGRA2YUV_YV12: case CV_RGB2YUV_IYUV: case CV_BGR2YUV_IYUV: case CV_RGBA2YUV_IYUV: case CV_BGRA2YUV_IYUV: { if (dcn <= 0) dcn = 1; const int bIdx = (code == CV_BGR2YUV_IYUV || code == CV_BGRA2YUV_IYUV || code == CV_BGR2YUV_YV12 || code == CV_BGRA2YUV_YV12) ? 0 : 2; const int uIdx = (code == CV_BGR2YUV_IYUV || code == CV_BGRA2YUV_IYUV || code == CV_RGB2YUV_IYUV || code == CV_RGBA2YUV_IYUV) ? 1 : 2; CV_Assert( (scn == 3 || scn == 4) && depth == CV_8U ); CV_Assert( dcn == 1 ); CV_Assert( sz.width % 2 == 0 && sz.height % 2 == 0 ); Size dstSz(sz.width, sz.height / 2 * 3); _dst.create(dstSz, CV_MAKETYPE(depth, dcn)); dst = _dst.getMat(); switch(bIdx + uIdx*10) { case 10: cvtRGBtoYUV420p<0, 1>(src, dst); break; case 12: cvtRGBtoYUV420p<2, 1>(src, dst); break; case 20: cvtRGBtoYUV420p<0, 2>(src, dst); break; case 22: cvtRGBtoYUV420p<2, 2>(src, dst); break; default: CV_Error( CV_StsBadFlag, "Unknown/unsupported color conversion code" ); break; }; } break; case CV_YUV2RGB_UYVY: case CV_YUV2BGR_UYVY: case CV_YUV2RGBA_UYVY: case CV_YUV2BGRA_UYVY: case CV_YUV2RGB_YUY2: case CV_YUV2BGR_YUY2: case CV_YUV2RGB_YVYU: case CV_YUV2BGR_YVYU: case CV_YUV2RGBA_YUY2: case CV_YUV2BGRA_YUY2: case CV_YUV2RGBA_YVYU: case CV_YUV2BGRA_YVYU: { //http://www.fourcc.org/yuv.php#UYVY //http://www.fourcc.org/yuv.php#YUY2 //http://www.fourcc.org/yuv.php#YVYU if (dcn <= 0) dcn = (code==CV_YUV2RGBA_UYVY || code==CV_YUV2BGRA_UYVY || code==CV_YUV2RGBA_YUY2 || code==CV_YUV2BGRA_YUY2 || code==CV_YUV2RGBA_YVYU || code==CV_YUV2BGRA_YVYU) ? 4 : 3; const int bIdx = (code==CV_YUV2BGR_UYVY || code==CV_YUV2BGRA_UYVY || code==CV_YUV2BGR_YUY2 || code==CV_YUV2BGRA_YUY2 || code==CV_YUV2BGR_YVYU || code==CV_YUV2BGRA_YVYU) ? 0 : 2; const int ycn = (code==CV_YUV2RGB_UYVY || code==CV_YUV2BGR_UYVY || code==CV_YUV2RGBA_UYVY || code==CV_YUV2BGRA_UYVY) ? 1 : 0; const int uIdx = (code==CV_YUV2RGB_YVYU || code==CV_YUV2BGR_YVYU || code==CV_YUV2RGBA_YVYU || code==CV_YUV2BGRA_YVYU) ? 1 : 0; CV_Assert( dcn == 3 || dcn == 4 ); CV_Assert( scn == 2 && depth == CV_8U ); _dst.create(sz, CV_8UC(dcn)); dst = _dst.getMat(); switch(dcn*1000 + bIdx*100 + uIdx*10 + ycn) { case 3000: cvtYUV422toRGB<0,0,0>(dst, (int)src.step, src.ptr()); break; case 3001: cvtYUV422toRGB<0,0,1>(dst, (int)src.step, src.ptr()); break; case 3010: cvtYUV422toRGB<0,1,0>(dst, (int)src.step, src.ptr()); break; case 3011: cvtYUV422toRGB<0,1,1>(dst, (int)src.step, src.ptr()); break; case 3200: cvtYUV422toRGB<2,0,0>(dst, (int)src.step, src.ptr()); break; case 3201: cvtYUV422toRGB<2,0,1>(dst, (int)src.step, src.ptr()); break; case 3210: cvtYUV422toRGB<2,1,0>(dst, (int)src.step, src.ptr()); break; case 3211: cvtYUV422toRGB<2,1,1>(dst, (int)src.step, src.ptr()); break; case 4000: cvtYUV422toRGBA<0,0,0>(dst, (int)src.step, src.ptr()); break; case 4001: cvtYUV422toRGBA<0,0,1>(dst, (int)src.step, src.ptr()); break; case 4010: cvtYUV422toRGBA<0,1,0>(dst, (int)src.step, src.ptr()); break; case 4011: cvtYUV422toRGBA<0,1,1>(dst, (int)src.step, src.ptr()); break; case 4200: cvtYUV422toRGBA<2,0,0>(dst, (int)src.step, src.ptr()); break; case 4201: cvtYUV422toRGBA<2,0,1>(dst, (int)src.step, src.ptr()); break; case 4210: cvtYUV422toRGBA<2,1,0>(dst, (int)src.step, src.ptr()); break; case 4211: cvtYUV422toRGBA<2,1,1>(dst, (int)src.step, src.ptr()); break; default: CV_Error( CV_StsBadFlag, "Unknown/unsupported color conversion code" ); break; }; } break; case CV_YUV2GRAY_UYVY: case CV_YUV2GRAY_YUY2: { if (dcn <= 0) dcn = 1; CV_Assert( dcn == 1 ); CV_Assert( scn == 2 && depth == CV_8U ); extractChannel(_src, _dst, code == CV_YUV2GRAY_UYVY ? 1 : 0); } break; case CV_RGBA2mRGBA: { if (dcn <= 0) dcn = 4; CV_Assert( scn == 4 && dcn == 4 ); _dst.create(sz, CV_MAKETYPE(depth, dcn)); dst = _dst.getMat(); if( depth == CV_8U ) { #if defined(HAVE_IPP) CV_IPP_CHECK() { if (CvtColorIPPLoop(src, dst, IPPGeneralFunctor((ippiGeneralFunc)ippiAlphaPremul_8u_AC4R))) { CV_IMPL_ADD(CV_IMPL_IPP|CV_IMPL_MT); return; } setIppErrorStatus(); } #endif CvtColorLoop(src, dst, RGBA2mRGBA()); } else { CV_Error( CV_StsBadArg, "Unsupported image depth" ); } } break; case CV_mRGBA2RGBA: { if (dcn <= 0) dcn = 4; CV_Assert( scn == 4 && dcn == 4 ); _dst.create(sz, CV_MAKETYPE(depth, dcn)); dst = _dst.getMat(); if( depth == CV_8U ) CvtColorLoop(src, dst, mRGBA2RGBA()); else { CV_Error( CV_StsBadArg, "Unsupported image depth" ); } } break; default: CV_Error( CV_StsBadFlag, "Unknown/unsupported color conversion code" ); } } CV_IMPL void cvCvtColor( const CvArr* srcarr, CvArr* dstarr, int code ) { cv::Mat src = cv::cvarrToMat(srcarr), dst0 = cv::cvarrToMat(dstarr), dst = dst0; CV_Assert( src.depth() == dst.depth() ); cv::cvtColor(src, dst, code, dst.channels()); CV_Assert( dst.data == dst0.data ); } /* End of file. */