[opencv] 16/251: AVX and SSE4.1 optimized conversion implementations migrated to separate files

Nobuhiro Iwamatsu iwamatsu at moszumanska.debian.org
Sun Aug 27 23:27:19 UTC 2017


This is an automated email from the git hooks/post-receive script.

iwamatsu pushed a commit to annotated tag 3.3.0
in repository opencv.

commit 5448d9186a740bbf816e4fdb453f52c806a67f43
Author: Vitaly Tuzov <terfendail at mediana.jetos.com>
Date:   Fri Jun 30 18:27:20 2017 +0300

    AVX and SSE4.1 optimized conversion implementations migrated to separate files
---
 modules/core/src/convert.avx2.cpp   |  78 +++++++++++
 modules/core/src/convert.cpp        | 263 +++++++-----------------------------
 modules/core/src/convert.hpp        |  15 ++
 modules/core/src/convert.sse4_1.cpp | 241 +++++++++++++++++++++++++++++++++
 4 files changed, 386 insertions(+), 211 deletions(-)

diff --git a/modules/core/src/convert.avx2.cpp b/modules/core/src/convert.avx2.cpp
new file mode 100644
index 0000000..ac1f94d
--- /dev/null
+++ b/modules/core/src/convert.avx2.cpp
@@ -0,0 +1,78 @@
+/*M///////////////////////////////////////////////////////////////////////////////////////
+//
+//  IMPORTANT: READ BEFORE DOWNLOADING, COPYING, INSTALLING OR USING.
+//
+//  By downloading, copying, installing or using the software you agree to this license.
+//  If you do not agree to this license, do not download, install,
+//  copy or use the software.
+//
+//
+//                           License Agreement
+//                For Open Source Computer Vision Library
+//
+// Copyright (C) 2000-2008, Intel Corporation, all rights reserved.
+// Copyright (C) 2009-2011, Willow Garage Inc., all rights reserved.
+// Copyright (C) 2014-2015, Itseez Inc., all rights reserved.
+// Third party copyrights are property of their respective owners.
+//
+// Redistribution and use in source and binary forms, with or without modification,
+// are permitted provided that the following conditions are met:
+//
+//   * Redistribution's of source code must retain the above copyright notice,
+//     this list of conditions and the following disclaimer.
+//
+//   * Redistribution's in binary form must reproduce the above copyright notice,
+//     this list of conditions and the following disclaimer in the documentation
+//     and/or other materials provided with the distribution.
+//
+//   * The name of the copyright holders may not be used to endorse or promote products
+//     derived from this software without specific prior written permission.
+//
+// This software is provided by the copyright holders and contributors "as is" and
+// any express or implied warranties, including, but not limited to, the implied
+// warranties of merchantability and fitness for a particular purpose are disclaimed.
+// In no event shall the Intel Corporation or contributors be liable for any direct,
+// indirect, incidental, special, exemplary, or consequential damages
+// (including, but not limited to, procurement of substitute goods or services;
+// loss of use, data, or profits; or business interruption) however caused
+// and on any theory of liability, whether in contract, strict liability,
+// or tort (including negligence or otherwise) arising in any way out of
+// the use of this software, even if advised of the possibility of such damage.
+//
+//M*/
+
+#include "precomp.hpp"
+#include "convert.hpp"
+
+namespace cv
+{
+namespace opt_AVX2
+{
+
+void cvtScale_s16s32f32Line_AVX2(const short* src, int* dst, float scale, float shift, int width)
+{
+    int x = 0;
+
+    __m256 scale256 = _mm256_set1_ps(scale);
+    __m256 shift256 = _mm256_set1_ps(shift);
+    const int shuffle = 0xD8;
+
+    for (; x <= width - 16; x += 16)
+    {
+        __m256i v_src = _mm256_loadu_si256((const __m256i *)(src + x));
+        v_src = _mm256_permute4x64_epi64(v_src, shuffle);
+        __m256i v_src_lo = _mm256_srai_epi32(_mm256_unpacklo_epi16(v_src, v_src), 16);
+        __m256i v_src_hi = _mm256_srai_epi32(_mm256_unpackhi_epi16(v_src, v_src), 16);
+        __m256 v_dst0 = _mm256_add_ps(_mm256_mul_ps(_mm256_cvtepi32_ps(v_src_lo), scale256), shift256);
+        __m256 v_dst1 = _mm256_add_ps(_mm256_mul_ps(_mm256_cvtepi32_ps(v_src_hi), scale256), shift256);
+        _mm256_storeu_si256((__m256i *)(dst + x), _mm256_cvtps_epi32(v_dst0));
+        _mm256_storeu_si256((__m256i *)(dst + x + 8), _mm256_cvtps_epi32(v_dst1));
+    }
+
+    for (; x < width; x++)
+        dst[x] = saturate_cast<int>(src[x] * scale + shift);
+}
+
+}
+}
+/* End of file. */
diff --git a/modules/core/src/convert.cpp b/modules/core/src/convert.cpp
index 94005d7..0955c78 100644
--- a/modules/core/src/convert.cpp
+++ b/modules/core/src/convert.cpp
@@ -1489,41 +1489,22 @@ struct cvtScale_SIMD<uchar, schar, float>
     }
 };
 
-#if CV_SSE4_1
+#if CV_TRY_SSE4_1
 
 template <>
 struct cvtScale_SIMD<uchar, ushort, float>
 {
     cvtScale_SIMD()
     {
-        haveSSE = checkHardwareSupport(CV_CPU_SSE4_1);
+        haveSSE = CV_CPU_HAS_SUPPORT_SSE4_1;
     }
 
     int operator () (const uchar * src, ushort * dst, int width, float scale, float shift) const
     {
-        int x = 0;
-
-        if (!haveSSE)
-            return x;
-
-        __m128i v_zero = _mm_setzero_si128();
-        __m128 v_scale = _mm_set1_ps(scale), v_shift = _mm_set1_ps(shift);
-
-        for ( ; x <= width - 8; x += 8)
-        {
-            __m128i v_src = _mm_unpacklo_epi8(_mm_loadl_epi64((__m128i const *)(src + x)), v_zero);
-            __m128 v_src_f = _mm_cvtepi32_ps(_mm_unpacklo_epi16(v_src, v_zero));
-            __m128 v_dst_0 = _mm_add_ps(_mm_mul_ps(v_src_f, v_scale), v_shift);
-
-            v_src_f = _mm_cvtepi32_ps(_mm_unpackhi_epi16(v_src, v_zero));
-            __m128 v_dst_1 = _mm_add_ps(_mm_mul_ps(v_src_f, v_scale), v_shift);
-
-            __m128i v_dst = _mm_packus_epi32(_mm_cvtps_epi32(v_dst_0),
-                                             _mm_cvtps_epi32(v_dst_1));
-            _mm_storeu_si128((__m128i *)(dst + x), v_dst);
-        }
-
-        return x;
+        if (haveSSE)
+            return opt_SSE4_1::cvtScale_SIMD_u8u16f32_SSE41(src, dst, width, scale, shift);
+        else
+            return 0;
     }
 
     bool haveSSE;
@@ -1720,41 +1701,22 @@ struct cvtScale_SIMD<schar, schar, float>
     }
 };
 
-#if CV_SSE4_1
+#if CV_TRY_SSE4_1
 
 template <>
 struct cvtScale_SIMD<schar, ushort, float>
 {
     cvtScale_SIMD()
     {
-        haveSSE = checkHardwareSupport(CV_CPU_SSE4_1);
+        haveSSE = CV_CPU_HAS_SUPPORT_SSE4_1;
     }
 
     int operator () (const schar * src, ushort * dst, int width, float scale, float shift) const
     {
-        int x = 0;
-
-        if (!haveSSE)
-            return x;
-
-        __m128i v_zero = _mm_setzero_si128();
-        __m128 v_scale = _mm_set1_ps(scale), v_shift = _mm_set1_ps(shift);
-
-        for ( ; x <= width - 8; x += 8)
-        {
-            __m128i v_src = _mm_srai_epi16(_mm_unpacklo_epi8(v_zero, _mm_loadl_epi64((__m128i const *)(src + x))), 8);
-            __m128 v_src_f = _mm_cvtepi32_ps(_mm_srai_epi32(_mm_unpacklo_epi16(v_zero, v_src), 16));
-            __m128 v_dst_0 = _mm_add_ps(_mm_mul_ps(v_src_f, v_scale), v_shift);
-
-            v_src_f = _mm_cvtepi32_ps(_mm_srai_epi32(_mm_unpackhi_epi16(v_zero, v_src), 16));
-            __m128 v_dst_1 = _mm_add_ps(_mm_mul_ps(v_src_f, v_scale), v_shift);
-
-            __m128i v_dst = _mm_packus_epi32(_mm_cvtps_epi32(v_dst_0),
-                                             _mm_cvtps_epi32(v_dst_1));
-            _mm_storeu_si128((__m128i *)(dst + x), v_dst);
-        }
-
-        return x;
+        if (haveSSE)
+            return opt_SSE4_1::cvtScale_SIMD_s8u16f32_SSE41(src, dst, width, scale, shift);
+        else
+            return 0;
     }
 
     bool haveSSE;
@@ -1952,41 +1914,22 @@ struct cvtScale_SIMD<ushort, schar, float>
     }
 };
 
-#if CV_SSE4_1
+#if CV_TRY_SSE4_1
 
 template <>
 struct cvtScale_SIMD<ushort, ushort, float>
 {
     cvtScale_SIMD()
     {
-        haveSSE = checkHardwareSupport(CV_CPU_SSE4_1);
+        haveSSE = CV_CPU_HAS_SUPPORT_SSE4_1;
     }
 
     int operator () (const ushort * src, ushort * dst, int width, float scale, float shift) const
     {
-        int x = 0;
-
-        if (!haveSSE)
-            return x;
-
-        __m128i v_zero = _mm_setzero_si128();
-        __m128 v_scale = _mm_set1_ps(scale), v_shift = _mm_set1_ps(shift);
-
-        for ( ; x <= width - 8; x += 8)
-        {
-            __m128i v_src = _mm_loadu_si128((__m128i const *)(src + x));
-            __m128 v_src_f = _mm_cvtepi32_ps(_mm_unpacklo_epi16(v_src, v_zero));
-            __m128 v_dst_0 = _mm_add_ps(_mm_mul_ps(v_src_f, v_scale), v_shift);
-
-            v_src_f = _mm_cvtepi32_ps(_mm_unpackhi_epi16(v_src, v_zero));
-            __m128 v_dst_1 = _mm_add_ps(_mm_mul_ps(v_src_f, v_scale), v_shift);
-
-            __m128i v_dst = _mm_packus_epi32(_mm_cvtps_epi32(v_dst_0),
-                                             _mm_cvtps_epi32(v_dst_1));
-            _mm_storeu_si128((__m128i *)(dst + x), v_dst);
-        }
-
-        return x;
+        if (haveSSE)
+            return opt_SSE4_1::cvtScale_SIMD_u16u16f32_SSE41(src, dst, width, scale, shift);
+        else
+            return 0;
     }
 
     bool haveSSE;
@@ -2183,41 +2126,22 @@ struct cvtScale_SIMD<short, schar, float>
     }
 };
 
-#if CV_SSE4_1
+#if CV_TRY_SSE4_1
 
 template <>
 struct cvtScale_SIMD<short, ushort, float>
 {
     cvtScale_SIMD()
     {
-        haveSSE = checkHardwareSupport(CV_CPU_SSE4_1);
+        haveSSE = CV_CPU_HAS_SUPPORT_SSE4_1;
     }
 
     int operator () (const short * src, ushort * dst, int width, float scale, float shift) const
     {
-        int x = 0;
-
-        if (!haveSSE)
-            return x;
-
-        __m128i v_zero = _mm_setzero_si128();
-        __m128 v_scale = _mm_set1_ps(scale), v_shift = _mm_set1_ps(shift);
-
-        for ( ; x <= width - 8; x += 8)
-        {
-            __m128i v_src = _mm_loadu_si128((__m128i const *)(src + x));
-            __m128 v_src_f = _mm_cvtepi32_ps(_mm_srai_epi32(_mm_unpacklo_epi16(v_zero, v_src), 16));
-            __m128 v_dst_0 = _mm_add_ps(_mm_mul_ps(v_src_f, v_scale), v_shift);
-
-            v_src_f = _mm_cvtepi32_ps(_mm_srai_epi32(_mm_unpackhi_epi16(v_zero, v_src), 16));
-            __m128 v_dst_1 = _mm_add_ps(_mm_mul_ps(v_src_f, v_scale), v_shift);
-
-            __m128i v_dst = _mm_packus_epi32(_mm_cvtps_epi32(v_dst_0),
-                                             _mm_cvtps_epi32(v_dst_1));
-            _mm_storeu_si128((__m128i *)(dst + x), v_dst);
-        }
-
-        return x;
+        if (haveSSE)
+            return opt_SSE4_1::cvtScale_SIMD_s16u16f32_SSE41(src, dst, width, scale, shift);
+        else
+            return 0;
     }
 
     bool haveSSE;
@@ -2412,39 +2336,22 @@ struct cvtScale_SIMD<int, schar, float>
     }
 };
 
-#if CV_SSE4_1
+#if CV_TRY_SSE4_1
 
 template <>
 struct cvtScale_SIMD<int, ushort, float>
 {
     cvtScale_SIMD()
     {
-        haveSSE = checkHardwareSupport(CV_CPU_SSE4_1);
+        haveSSE = CV_CPU_HAS_SUPPORT_SSE4_1;
     }
 
     int operator () (const int * src, ushort * dst, int width, float scale, float shift) const
     {
-        int x = 0;
-
-        if (!haveSSE)
-            return x;
-
-        __m128 v_scale = _mm_set1_ps(scale), v_shift = _mm_set1_ps(shift);
-
-        for ( ; x <= width - 8; x += 8)
-        {
-            __m128i v_src = _mm_loadu_si128((__m128i const *)(src + x));
-            __m128 v_dst_0 = _mm_add_ps(_mm_mul_ps(_mm_cvtepi32_ps(v_src), v_scale), v_shift);
-
-            v_src = _mm_loadu_si128((__m128i const *)(src + x + 4));
-            __m128 v_dst_1 = _mm_add_ps(_mm_mul_ps(_mm_cvtepi32_ps(v_src), v_scale), v_shift);
-
-            __m128i v_dst = _mm_packus_epi32(_mm_cvtps_epi32(v_dst_0),
-                                             _mm_cvtps_epi32(v_dst_1));
-            _mm_storeu_si128((__m128i *)(dst + x), v_dst);
-        }
-
-        return x;
+        if (haveSSE)
+            return opt_SSE4_1::cvtScale_SIMD_s32u16f32_SSE41(src, dst, width, scale, shift);
+        else
+            return 0;
     }
 
     bool haveSSE;
@@ -2629,39 +2536,22 @@ struct cvtScale_SIMD<float, schar, float>
     }
 };
 
-#if CV_SSE4_1
+#if CV_TRY_SSE4_1
 
 template <>
 struct cvtScale_SIMD<float, ushort, float>
 {
     cvtScale_SIMD()
     {
-        haveSSE = checkHardwareSupport(CV_CPU_SSE4_1);
+        haveSSE = CV_CPU_HAS_SUPPORT_SSE4_1;
     }
 
     int operator () (const float * src, ushort * dst, int width, float scale, float shift) const
     {
-        int x = 0;
-
-        if (!haveSSE)
-            return x;
-
-        __m128 v_scale = _mm_set1_ps(scale), v_shift = _mm_set1_ps(shift);
-
-        for ( ; x <= width - 8; x += 8)
-        {
-            __m128 v_src = _mm_loadu_ps(src + x);
-            __m128 v_dst_0 = _mm_add_ps(_mm_mul_ps(v_src, v_scale), v_shift);
-
-            v_src = _mm_loadu_ps(src + x + 4);
-            __m128 v_dst_1 = _mm_add_ps(_mm_mul_ps(v_src, v_scale), v_shift);
-
-            __m128i v_dst = _mm_packus_epi32(_mm_cvtps_epi32(v_dst_0),
-                                             _mm_cvtps_epi32(v_dst_1));
-            _mm_storeu_si128((__m128i *)(dst + x), v_dst);
-        }
-
-        return x;
+        if (haveSSE)
+            return opt_SSE4_1::cvtScale_SIMD_f32u16f32_SSE41(src, dst, width, scale, shift);
+        else
+            return 0;
     }
 
     bool haveSSE;
@@ -2842,41 +2732,22 @@ struct cvtScale_SIMD<double, schar, float>
     }
 };
 
-#if CV_SSE4_1
+#if CV_TRY_SSE4_1
 
 template <>
 struct cvtScale_SIMD<double, ushort, float>
 {
     cvtScale_SIMD()
     {
-        haveSSE = checkHardwareSupport(CV_CPU_SSE4_1);
+        haveSSE = CV_CPU_HAS_SUPPORT_SSE4_1;
     }
 
     int operator () (const double * src, ushort * dst, int width, float scale, float shift) const
     {
-        int x = 0;
-
-        if (!haveSSE)
-            return x;
-
-        __m128 v_scale = _mm_set1_ps(scale), v_shift = _mm_set1_ps(shift);
-
-        for ( ; x <= width - 8; x += 8)
-        {
-            __m128 v_src = _mm_movelh_ps(_mm_cvtpd_ps(_mm_loadu_pd(src + x)),
-                                         _mm_cvtpd_ps(_mm_loadu_pd(src + x + 2)));
-            __m128 v_dst_0 = _mm_add_ps(_mm_mul_ps(v_src, v_scale), v_shift);
-
-            v_src = _mm_movelh_ps(_mm_cvtpd_ps(_mm_loadu_pd(src + x + 4)),
-                                  _mm_cvtpd_ps(_mm_loadu_pd(src + x + 6)));
-            __m128 v_dst_1 = _mm_add_ps(_mm_mul_ps(v_src, v_scale), v_shift);
-
-            __m128i v_dst = _mm_packus_epi32(_mm_cvtps_epi32(v_dst_0),
-                                             _mm_cvtps_epi32(v_dst_1));
-            _mm_storeu_si128((__m128i *)(dst + x), v_dst);
-        }
-
-        return x;
+        if (haveSSE)
+            return opt_SSE4_1::cvtScale_SIMD_f64u16f32_SSE41(src, dst, width, scale, shift);
+        else
+            return 0;
     }
 
     bool haveSSE;
@@ -3803,24 +3674,11 @@ cvtScale_<short, int, float>( const short* src, size_t sstep,
     {
         int x = 0;
 
-        #if CV_AVX2
-        if (USE_AVX2)
+        #if CV_TRY_AVX2
+        if (CV_CPU_HAS_SUPPORT_AVX2)
         {
-            __m256 scale256 = _mm256_set1_ps(scale);
-            __m256 shift256 = _mm256_set1_ps(shift);
-            const int shuffle = 0xD8;
-
-            for ( ; x <= size.width - 16; x += 16)
-            {
-                __m256i v_src = _mm256_loadu_si256((const __m256i *)(src + x));
-                v_src = _mm256_permute4x64_epi64(v_src, shuffle);
-                __m256i v_src_lo = _mm256_srai_epi32(_mm256_unpacklo_epi16(v_src, v_src), 16);
-                __m256i v_src_hi = _mm256_srai_epi32(_mm256_unpackhi_epi16(v_src, v_src), 16);
-                __m256 v_dst0 = _mm256_add_ps(_mm256_mul_ps(_mm256_cvtepi32_ps(v_src_lo), scale256), shift256);
-                __m256 v_dst1 = _mm256_add_ps(_mm256_mul_ps(_mm256_cvtepi32_ps(v_src_hi), scale256), shift256);
-                _mm256_storeu_si256((__m256i *)(dst + x), _mm256_cvtps_epi32(v_dst0));
-                _mm256_storeu_si256((__m256i *)(dst + x + 8), _mm256_cvtps_epi32(v_dst1));
-            }
+            opt_AVX2::cvtScale_s16s32f32Line_AVX2(src, dst, scale, shift, size.width);
+            return;
         }
         #endif
         #if CV_SSE2
@@ -3933,37 +3791,20 @@ struct Cvt_SIMD<double, schar>
     }
 };
 
-#if CV_SSE4_1
+#if CV_TRY_SSE4_1
 
 template <>
 struct Cvt_SIMD<double, ushort>
 {
     bool haveSIMD;
-    Cvt_SIMD() { haveSIMD = checkHardwareSupport(CV_CPU_SSE4_1); }
+    Cvt_SIMD() { haveSIMD = CV_CPU_HAS_SUPPORT_SSE4_1; }
 
     int operator() (const double * src, ushort * dst, int width) const
     {
-        int x = 0;
-
-        if (!haveSIMD)
-            return x;
-
-        for ( ; x <= width - 8; x += 8)
-        {
-            __m128 v_src0 = _mm_cvtpd_ps(_mm_loadu_pd(src + x));
-            __m128 v_src1 = _mm_cvtpd_ps(_mm_loadu_pd(src + x + 2));
-            __m128 v_src2 = _mm_cvtpd_ps(_mm_loadu_pd(src + x + 4));
-            __m128 v_src3 = _mm_cvtpd_ps(_mm_loadu_pd(src + x + 6));
-
-            v_src0 = _mm_movelh_ps(v_src0, v_src1);
-            v_src1 = _mm_movelh_ps(v_src2, v_src3);
-
-            __m128i v_dst = _mm_packus_epi32(_mm_cvtps_epi32(v_src0),
-                                             _mm_cvtps_epi32(v_src1));
-            _mm_storeu_si128((__m128i *)(dst + x), v_dst);
-        }
-
-        return x;
+        if (haveSIMD)
+            return opt_SSE4_1::Cvt_SIMD_f64u16_SSE41(src, dst, width);
+        else
+            return 0;
     }
 };
 
diff --git a/modules/core/src/convert.hpp b/modules/core/src/convert.hpp
index c149d21..ce1d313 100644
--- a/modules/core/src/convert.hpp
+++ b/modules/core/src/convert.hpp
@@ -170,4 +170,19 @@ namespace opt_FP16
 void cvtScaleHalf_SIMD32f16f( const float* src, size_t sstep, short* dst, size_t dstep, cv::Size size );
 void cvtScaleHalf_SIMD16f32f( const short* src, size_t sstep, float* dst, size_t dstep, cv::Size size );
 }
+namespace opt_AVX2
+{
+void cvtScale_s16s32f32Line_AVX2(const short* src, int* dst, float scale, float shift, int width);
+}
+namespace opt_SSE4_1
+{
+    int cvtScale_SIMD_u8u16f32_SSE41(const uchar * src, ushort * dst, int width, float scale, float shift);
+    int cvtScale_SIMD_s8u16f32_SSE41(const schar * src, ushort * dst, int width, float scale, float shift);
+    int cvtScale_SIMD_u16u16f32_SSE41(const ushort * src, ushort * dst, int width, float scale, float shift);
+    int cvtScale_SIMD_s16u16f32_SSE41(const short * src, ushort * dst, int width, float scale, float shift);
+    int cvtScale_SIMD_s32u16f32_SSE41(const int * src, ushort * dst, int width, float scale, float shift);
+    int cvtScale_SIMD_f32u16f32_SSE41(const float * src, ushort * dst, int width, float scale, float shift);
+    int cvtScale_SIMD_f64u16f32_SSE41(const double * src, ushort * dst, int width, float scale, float shift);
+    int Cvt_SIMD_f64u16_SSE41(const double * src, ushort * dst, int width);
+}
 }
\ No newline at end of file
diff --git a/modules/core/src/convert.sse4_1.cpp b/modules/core/src/convert.sse4_1.cpp
new file mode 100644
index 0000000..9a5234a
--- /dev/null
+++ b/modules/core/src/convert.sse4_1.cpp
@@ -0,0 +1,241 @@
+/*M///////////////////////////////////////////////////////////////////////////////////////
+//
+//  IMPORTANT: READ BEFORE DOWNLOADING, COPYING, INSTALLING OR USING.
+//
+//  By downloading, copying, installing or using the software you agree to this license.
+//  If you do not agree to this license, do not download, install,
+//  copy or use the software.
+//
+//
+//                           License Agreement
+//                For Open Source Computer Vision Library
+//
+// Copyright (C) 2000-2008, Intel Corporation, all rights reserved.
+// Copyright (C) 2009-2011, Willow Garage Inc., all rights reserved.
+// Copyright (C) 2014-2015, Itseez Inc., all rights reserved.
+// Third party copyrights are property of their respective owners.
+//
+// Redistribution and use in source and binary forms, with or without modification,
+// are permitted provided that the following conditions are met:
+//
+//   * Redistribution's of source code must retain the above copyright notice,
+//     this list of conditions and the following disclaimer.
+//
+//   * Redistribution's in binary form must reproduce the above copyright notice,
+//     this list of conditions and the following disclaimer in the documentation
+//     and/or other materials provided with the distribution.
+//
+//   * The name of the copyright holders may not be used to endorse or promote products
+//     derived from this software without specific prior written permission.
+//
+// This software is provided by the copyright holders and contributors "as is" and
+// any express or implied warranties, including, but not limited to, the implied
+// warranties of merchantability and fitness for a particular purpose are disclaimed.
+// In no event shall the Intel Corporation or contributors be liable for any direct,
+// indirect, incidental, special, exemplary, or consequential damages
+// (including, but not limited to, procurement of substitute goods or services;
+// loss of use, data, or profits; or business interruption) however caused
+// and on any theory of liability, whether in contract, strict liability,
+// or tort (including negligence or otherwise) arising in any way out of
+// the use of this software, even if advised of the possibility of such damage.
+//
+//M*/
+
+#include "precomp.hpp"
+#include "convert.hpp"
+
+namespace cv
+{
+namespace opt_SSE4_1
+{
+
+int cvtScale_SIMD_u8u16f32_SSE41(const uchar * src, ushort * dst, int width, float scale, float shift)
+{
+    int x = 0;
+
+    __m128i v_zero = _mm_setzero_si128();
+    __m128 v_scale = _mm_set1_ps(scale), v_shift = _mm_set1_ps(shift);
+
+    for ( ; x <= width - 8; x += 8)
+    {
+        __m128i v_src = _mm_unpacklo_epi8(_mm_loadl_epi64((__m128i const *)(src + x)), v_zero);
+        __m128 v_src_f = _mm_cvtepi32_ps(_mm_unpacklo_epi16(v_src, v_zero));
+        __m128 v_dst_0 = _mm_add_ps(_mm_mul_ps(v_src_f, v_scale), v_shift);
+
+        v_src_f = _mm_cvtepi32_ps(_mm_unpackhi_epi16(v_src, v_zero));
+        __m128 v_dst_1 = _mm_add_ps(_mm_mul_ps(v_src_f, v_scale), v_shift);
+
+        __m128i v_dst = _mm_packus_epi32(_mm_cvtps_epi32(v_dst_0),
+                                            _mm_cvtps_epi32(v_dst_1));
+        _mm_storeu_si128((__m128i *)(dst + x), v_dst);
+    }
+
+    return x;
+}
+
+int cvtScale_SIMD_s8u16f32_SSE41(const schar * src, ushort * dst, int width, float scale, float shift)
+{
+    int x = 0;
+
+    __m128i v_zero = _mm_setzero_si128();
+    __m128 v_scale = _mm_set1_ps(scale), v_shift = _mm_set1_ps(shift);
+
+    for ( ; x <= width - 8; x += 8)
+    {
+        __m128i v_src = _mm_srai_epi16(_mm_unpacklo_epi8(v_zero, _mm_loadl_epi64((__m128i const *)(src + x))), 8);
+        __m128 v_src_f = _mm_cvtepi32_ps(_mm_srai_epi32(_mm_unpacklo_epi16(v_zero, v_src), 16));
+        __m128 v_dst_0 = _mm_add_ps(_mm_mul_ps(v_src_f, v_scale), v_shift);
+
+        v_src_f = _mm_cvtepi32_ps(_mm_srai_epi32(_mm_unpackhi_epi16(v_zero, v_src), 16));
+        __m128 v_dst_1 = _mm_add_ps(_mm_mul_ps(v_src_f, v_scale), v_shift);
+
+        __m128i v_dst = _mm_packus_epi32(_mm_cvtps_epi32(v_dst_0),
+                                            _mm_cvtps_epi32(v_dst_1));
+        _mm_storeu_si128((__m128i *)(dst + x), v_dst);
+    }
+
+    return x;
+}
+
+int cvtScale_SIMD_u16u16f32_SSE41(const ushort * src, ushort * dst, int width, float scale, float shift)
+{
+    int x = 0;
+
+    __m128i v_zero = _mm_setzero_si128();
+    __m128 v_scale = _mm_set1_ps(scale), v_shift = _mm_set1_ps(shift);
+
+    for ( ; x <= width - 8; x += 8)
+    {
+        __m128i v_src = _mm_loadu_si128((__m128i const *)(src + x));
+        __m128 v_src_f = _mm_cvtepi32_ps(_mm_unpacklo_epi16(v_src, v_zero));
+        __m128 v_dst_0 = _mm_add_ps(_mm_mul_ps(v_src_f, v_scale), v_shift);
+
+        v_src_f = _mm_cvtepi32_ps(_mm_unpackhi_epi16(v_src, v_zero));
+        __m128 v_dst_1 = _mm_add_ps(_mm_mul_ps(v_src_f, v_scale), v_shift);
+
+        __m128i v_dst = _mm_packus_epi32(_mm_cvtps_epi32(v_dst_0),
+                                            _mm_cvtps_epi32(v_dst_1));
+        _mm_storeu_si128((__m128i *)(dst + x), v_dst);
+    }
+
+    return x;
+}
+
+int cvtScale_SIMD_s16u16f32_SSE41(const short * src, ushort * dst, int width, float scale, float shift)
+{
+    int x = 0;
+
+    __m128i v_zero = _mm_setzero_si128();
+    __m128 v_scale = _mm_set1_ps(scale), v_shift = _mm_set1_ps(shift);
+
+    for ( ; x <= width - 8; x += 8)
+    {
+        __m128i v_src = _mm_loadu_si128((__m128i const *)(src + x));
+        __m128 v_src_f = _mm_cvtepi32_ps(_mm_srai_epi32(_mm_unpacklo_epi16(v_zero, v_src), 16));
+        __m128 v_dst_0 = _mm_add_ps(_mm_mul_ps(v_src_f, v_scale), v_shift);
+
+        v_src_f = _mm_cvtepi32_ps(_mm_srai_epi32(_mm_unpackhi_epi16(v_zero, v_src), 16));
+        __m128 v_dst_1 = _mm_add_ps(_mm_mul_ps(v_src_f, v_scale), v_shift);
+
+        __m128i v_dst = _mm_packus_epi32(_mm_cvtps_epi32(v_dst_0),
+                                            _mm_cvtps_epi32(v_dst_1));
+        _mm_storeu_si128((__m128i *)(dst + x), v_dst);
+    }
+
+    return x;
+}
+
+int cvtScale_SIMD_s32u16f32_SSE41(const int * src, ushort * dst, int width, float scale, float shift)
+{
+    int x = 0;
+
+    __m128 v_scale = _mm_set1_ps(scale), v_shift = _mm_set1_ps(shift);
+
+    for ( ; x <= width - 8; x += 8)
+    {
+        __m128i v_src = _mm_loadu_si128((__m128i const *)(src + x));
+        __m128 v_dst_0 = _mm_add_ps(_mm_mul_ps(_mm_cvtepi32_ps(v_src), v_scale), v_shift);
+
+        v_src = _mm_loadu_si128((__m128i const *)(src + x + 4));
+        __m128 v_dst_1 = _mm_add_ps(_mm_mul_ps(_mm_cvtepi32_ps(v_src), v_scale), v_shift);
+
+        __m128i v_dst = _mm_packus_epi32(_mm_cvtps_epi32(v_dst_0),
+                                            _mm_cvtps_epi32(v_dst_1));
+        _mm_storeu_si128((__m128i *)(dst + x), v_dst);
+    }
+
+    return x;
+}
+
+int cvtScale_SIMD_f32u16f32_SSE41(const float * src, ushort * dst, int width, float scale, float shift)
+{
+    int x = 0;
+
+    __m128 v_scale = _mm_set1_ps(scale), v_shift = _mm_set1_ps(shift);
+
+    for ( ; x <= width - 8; x += 8)
+    {
+        __m128 v_src = _mm_loadu_ps(src + x);
+        __m128 v_dst_0 = _mm_add_ps(_mm_mul_ps(v_src, v_scale), v_shift);
+
+        v_src = _mm_loadu_ps(src + x + 4);
+        __m128 v_dst_1 = _mm_add_ps(_mm_mul_ps(v_src, v_scale), v_shift);
+
+        __m128i v_dst = _mm_packus_epi32(_mm_cvtps_epi32(v_dst_0),
+                                            _mm_cvtps_epi32(v_dst_1));
+        _mm_storeu_si128((__m128i *)(dst + x), v_dst);
+    }
+
+    return x;
+}
+
+int cvtScale_SIMD_f64u16f32_SSE41(const double * src, ushort * dst, int width, float scale, float shift)
+{
+    int x = 0;
+
+    __m128 v_scale = _mm_set1_ps(scale), v_shift = _mm_set1_ps(shift);
+
+    for ( ; x <= width - 8; x += 8)
+    {
+        __m128 v_src = _mm_movelh_ps(_mm_cvtpd_ps(_mm_loadu_pd(src + x)),
+                                        _mm_cvtpd_ps(_mm_loadu_pd(src + x + 2)));
+        __m128 v_dst_0 = _mm_add_ps(_mm_mul_ps(v_src, v_scale), v_shift);
+
+        v_src = _mm_movelh_ps(_mm_cvtpd_ps(_mm_loadu_pd(src + x + 4)),
+                                _mm_cvtpd_ps(_mm_loadu_pd(src + x + 6)));
+        __m128 v_dst_1 = _mm_add_ps(_mm_mul_ps(v_src, v_scale), v_shift);
+
+        __m128i v_dst = _mm_packus_epi32(_mm_cvtps_epi32(v_dst_0),
+                                            _mm_cvtps_epi32(v_dst_1));
+        _mm_storeu_si128((__m128i *)(dst + x), v_dst);
+    }
+
+    return x;
+}
+
+int Cvt_SIMD_f64u16_SSE41(const double * src, ushort * dst, int width)
+{
+    int x = 0;
+
+    for ( ; x <= width - 8; x += 8)
+    {
+        __m128 v_src0 = _mm_cvtpd_ps(_mm_loadu_pd(src + x));
+        __m128 v_src1 = _mm_cvtpd_ps(_mm_loadu_pd(src + x + 2));
+        __m128 v_src2 = _mm_cvtpd_ps(_mm_loadu_pd(src + x + 4));
+        __m128 v_src3 = _mm_cvtpd_ps(_mm_loadu_pd(src + x + 6));
+
+        v_src0 = _mm_movelh_ps(v_src0, v_src1);
+        v_src1 = _mm_movelh_ps(v_src2, v_src3);
+
+        __m128i v_dst = _mm_packus_epi32(_mm_cvtps_epi32(v_src0),
+                                            _mm_cvtps_epi32(v_src1));
+        _mm_storeu_si128((__m128i *)(dst + x), v_dst);
+    }
+
+    return x;
+}
+
+}
+}
+
+/* End of file. */

-- 
Alioth's /usr/local/bin/git-commit-notice on /srv/git.debian.org/git/debian-science/packages/opencv.git



More information about the debian-science-commits mailing list