| 1 | // Copyright 2022 Google Inc. All Rights Reserved. |
| 2 | // |
| 3 | // Use of this source code is governed by a BSD-style license |
| 4 | // that can be found in the COPYING file in the root of the source |
| 5 | // tree. An additional intellectual property rights grant can be found |
| 6 | // in the file PATENTS. All contributing project authors may |
| 7 | // be found in the AUTHORS file in the root of the source tree. |
| 8 | // ----------------------------------------------------------------------------- |
| 9 | // |
| 10 | // Speed-critical functions for Sharp YUV. |
| 11 | // |
| 12 | // Author: Skal (pascal.massimino@gmail.com) |
| 13 | |
| 14 | #include "sharpyuv/sharpyuv_dsp.h" |
| 15 | |
| 16 | #if defined(WEBP_USE_SSE2) |
| 17 | #include <stdlib.h> |
| 18 | #include <emmintrin.h> |
| 19 | |
| 20 | static uint16_t clip_SSE2(int v, int max) { |
| 21 | return (v < 0) ? 0 : (v > max) ? max : (uint16_t)v; |
| 22 | } |
| 23 | |
| 24 | static uint64_t SharpYuvUpdateY_SSE2(const uint16_t* ref, const uint16_t* src, |
| 25 | uint16_t* dst, int len, int bit_depth) { |
| 26 | const int max_y = (1 << bit_depth) - 1; |
| 27 | uint64_t diff = 0; |
| 28 | uint32_t tmp[4]; |
| 29 | int i; |
| 30 | const __m128i zero = _mm_setzero_si128(); |
| 31 | const __m128i max = _mm_set1_epi16(max_y); |
| 32 | const __m128i one = _mm_set1_epi16(1); |
| 33 | __m128i sum = zero; |
| 34 | |
| 35 | for (i = 0; i + 8 <= len; i += 8) { |
| 36 | const __m128i A = _mm_loadu_si128((const __m128i*)(ref + i)); |
| 37 | const __m128i B = _mm_loadu_si128((const __m128i*)(src + i)); |
| 38 | const __m128i C = _mm_loadu_si128((const __m128i*)(dst + i)); |
| 39 | const __m128i D = _mm_sub_epi16(A, B); // diff_y |
| 40 | const __m128i E = _mm_cmpgt_epi16(zero, D); // sign (-1 or 0) |
| 41 | const __m128i F = _mm_add_epi16(C, D); // new_y |
| 42 | const __m128i G = _mm_or_si128(E, one); // -1 or 1 |
| 43 | const __m128i H = _mm_max_epi16(_mm_min_epi16(F, max), zero); |
| 44 | const __m128i I = _mm_madd_epi16(D, G); // sum(abs(...)) |
| 45 | _mm_storeu_si128((__m128i*)(dst + i), H); |
| 46 | sum = _mm_add_epi32(sum, I); |
| 47 | } |
| 48 | _mm_storeu_si128((__m128i*)tmp, sum); |
| 49 | diff = tmp[3] + tmp[2] + tmp[1] + tmp[0]; |
| 50 | for (; i < len; ++i) { |
| 51 | const int diff_y = ref[i] - src[i]; |
| 52 | const int new_y = (int)dst[i] + diff_y; |
| 53 | dst[i] = clip_SSE2(new_y, max_y); |
| 54 | diff += (uint64_t)abs(diff_y); |
| 55 | } |
| 56 | return diff; |
| 57 | } |
| 58 | |
| 59 | static void SharpYuvUpdateRGB_SSE2(const int16_t* ref, const int16_t* src, |
| 60 | int16_t* dst, int len) { |
| 61 | int i = 0; |
| 62 | for (i = 0; i + 8 <= len; i += 8) { |
| 63 | const __m128i A = _mm_loadu_si128((const __m128i*)(ref + i)); |
| 64 | const __m128i B = _mm_loadu_si128((const __m128i*)(src + i)); |
| 65 | const __m128i C = _mm_loadu_si128((const __m128i*)(dst + i)); |
| 66 | const __m128i D = _mm_sub_epi16(A, B); // diff_uv |
| 67 | const __m128i E = _mm_add_epi16(C, D); // new_uv |
| 68 | _mm_storeu_si128((__m128i*)(dst + i), E); |
| 69 | } |
| 70 | for (; i < len; ++i) { |
| 71 | const int diff_uv = ref[i] - src[i]; |
| 72 | dst[i] += diff_uv; |
| 73 | } |
| 74 | } |
| 75 | |
| 76 | static void SharpYuvFilterRow16_SSE2(const int16_t* A, const int16_t* B, |
| 77 | int len, const uint16_t* best_y, |
| 78 | uint16_t* out, int bit_depth) { |
| 79 | const int max_y = (1 << bit_depth) - 1; |
| 80 | int i; |
| 81 | const __m128i kCst8 = _mm_set1_epi16(8); |
| 82 | const __m128i max = _mm_set1_epi16(max_y); |
| 83 | const __m128i zero = _mm_setzero_si128(); |
| 84 | for (i = 0; i + 8 <= len; i += 8) { |
| 85 | const __m128i a0 = _mm_loadu_si128((const __m128i*)(A + i + 0)); |
| 86 | const __m128i a1 = _mm_loadu_si128((const __m128i*)(A + i + 1)); |
| 87 | const __m128i b0 = _mm_loadu_si128((const __m128i*)(B + i + 0)); |
| 88 | const __m128i b1 = _mm_loadu_si128((const __m128i*)(B + i + 1)); |
| 89 | const __m128i a0b1 = _mm_add_epi16(a0, b1); |
| 90 | const __m128i a1b0 = _mm_add_epi16(a1, b0); |
| 91 | const __m128i a0a1b0b1 = _mm_add_epi16(a0b1, a1b0); // A0+A1+B0+B1 |
| 92 | const __m128i a0a1b0b1_8 = _mm_add_epi16(a0a1b0b1, kCst8); |
| 93 | const __m128i a0b1_2 = _mm_add_epi16(a0b1, a0b1); // 2*(A0+B1) |
| 94 | const __m128i a1b0_2 = _mm_add_epi16(a1b0, a1b0); // 2*(A1+B0) |
| 95 | const __m128i c0 = _mm_srai_epi16(_mm_add_epi16(a0b1_2, a0a1b0b1_8), 3); |
| 96 | const __m128i c1 = _mm_srai_epi16(_mm_add_epi16(a1b0_2, a0a1b0b1_8), 3); |
| 97 | const __m128i d0 = _mm_add_epi16(c1, a0); |
| 98 | const __m128i d1 = _mm_add_epi16(c0, a1); |
| 99 | const __m128i e0 = _mm_srai_epi16(d0, 1); |
| 100 | const __m128i e1 = _mm_srai_epi16(d1, 1); |
| 101 | const __m128i f0 = _mm_unpacklo_epi16(e0, e1); |
| 102 | const __m128i f1 = _mm_unpackhi_epi16(e0, e1); |
| 103 | const __m128i g0 = _mm_loadu_si128((const __m128i*)(best_y + 2 * i + 0)); |
| 104 | const __m128i g1 = _mm_loadu_si128((const __m128i*)(best_y + 2 * i + 8)); |
| 105 | const __m128i h0 = _mm_add_epi16(g0, f0); |
| 106 | const __m128i h1 = _mm_add_epi16(g1, f1); |
| 107 | const __m128i i0 = _mm_max_epi16(_mm_min_epi16(h0, max), zero); |
| 108 | const __m128i i1 = _mm_max_epi16(_mm_min_epi16(h1, max), zero); |
| 109 | _mm_storeu_si128((__m128i*)(out + 2 * i + 0), i0); |
| 110 | _mm_storeu_si128((__m128i*)(out + 2 * i + 8), i1); |
| 111 | } |
| 112 | for (; i < len; ++i) { |
| 113 | // (9 * A0 + 3 * A1 + 3 * B0 + B1 + 8) >> 4 = |
| 114 | // = (8 * A0 + 2 * (A1 + B0) + (A0 + A1 + B0 + B1 + 8)) >> 4 |
| 115 | // We reuse the common sub-expressions. |
| 116 | const int a0b1 = A[i + 0] + B[i + 1]; |
| 117 | const int a1b0 = A[i + 1] + B[i + 0]; |
| 118 | const int a0a1b0b1 = a0b1 + a1b0 + 8; |
| 119 | const int v0 = (8 * A[i + 0] + 2 * a1b0 + a0a1b0b1) >> 4; |
| 120 | const int v1 = (8 * A[i + 1] + 2 * a0b1 + a0a1b0b1) >> 4; |
| 121 | out[2 * i + 0] = clip_SSE2(best_y[2 * i + 0] + v0, max_y); |
| 122 | out[2 * i + 1] = clip_SSE2(best_y[2 * i + 1] + v1, max_y); |
| 123 | } |
| 124 | } |
| 125 | |
| 126 | static WEBP_INLINE __m128i s16_to_s32(__m128i in) { |
| 127 | return _mm_srai_epi32(_mm_unpacklo_epi16(in, in), 16); |
| 128 | } |
| 129 | |
| 130 | static void SharpYuvFilterRow32_SSE2(const int16_t* A, const int16_t* B, |
| 131 | int len, const uint16_t* best_y, |
| 132 | uint16_t* out, int bit_depth) { |
| 133 | const int max_y = (1 << bit_depth) - 1; |
| 134 | int i; |
| 135 | const __m128i kCst8 = _mm_set1_epi32(8); |
| 136 | const __m128i max = _mm_set1_epi16(max_y); |
| 137 | const __m128i zero = _mm_setzero_si128(); |
| 138 | for (i = 0; i + 4 <= len; i += 4) { |
| 139 | const __m128i a0 = s16_to_s32(_mm_loadl_epi64((const __m128i*)(A + i + 0))); |
| 140 | const __m128i a1 = s16_to_s32(_mm_loadl_epi64((const __m128i*)(A + i + 1))); |
| 141 | const __m128i b0 = s16_to_s32(_mm_loadl_epi64((const __m128i*)(B + i + 0))); |
| 142 | const __m128i b1 = s16_to_s32(_mm_loadl_epi64((const __m128i*)(B + i + 1))); |
| 143 | const __m128i a0b1 = _mm_add_epi32(a0, b1); |
| 144 | const __m128i a1b0 = _mm_add_epi32(a1, b0); |
| 145 | const __m128i a0a1b0b1 = _mm_add_epi32(a0b1, a1b0); // A0+A1+B0+B1 |
| 146 | const __m128i a0a1b0b1_8 = _mm_add_epi32(a0a1b0b1, kCst8); |
| 147 | const __m128i a0b1_2 = _mm_add_epi32(a0b1, a0b1); // 2*(A0+B1) |
| 148 | const __m128i a1b0_2 = _mm_add_epi32(a1b0, a1b0); // 2*(A1+B0) |
| 149 | const __m128i c0 = _mm_srai_epi32(_mm_add_epi32(a0b1_2, a0a1b0b1_8), 3); |
| 150 | const __m128i c1 = _mm_srai_epi32(_mm_add_epi32(a1b0_2, a0a1b0b1_8), 3); |
| 151 | const __m128i d0 = _mm_add_epi32(c1, a0); |
| 152 | const __m128i d1 = _mm_add_epi32(c0, a1); |
| 153 | const __m128i e0 = _mm_srai_epi32(d0, 1); |
| 154 | const __m128i e1 = _mm_srai_epi32(d1, 1); |
| 155 | const __m128i f0 = _mm_unpacklo_epi32(e0, e1); |
| 156 | const __m128i f1 = _mm_unpackhi_epi32(e0, e1); |
| 157 | const __m128i g = _mm_loadu_si128((const __m128i*)(best_y + 2 * i + 0)); |
| 158 | const __m128i h_16 = _mm_add_epi16(g, _mm_packs_epi32(f0, f1)); |
| 159 | const __m128i final = _mm_max_epi16(_mm_min_epi16(h_16, max), zero); |
| 160 | _mm_storeu_si128((__m128i*)(out + 2 * i + 0), final); |
| 161 | } |
| 162 | for (; i < len; ++i) { |
| 163 | // (9 * A0 + 3 * A1 + 3 * B0 + B1 + 8) >> 4 = |
| 164 | // = (8 * A0 + 2 * (A1 + B0) + (A0 + A1 + B0 + B1 + 8)) >> 4 |
| 165 | // We reuse the common sub-expressions. |
| 166 | const int a0b1 = A[i + 0] + B[i + 1]; |
| 167 | const int a1b0 = A[i + 1] + B[i + 0]; |
| 168 | const int a0a1b0b1 = a0b1 + a1b0 + 8; |
| 169 | const int v0 = (8 * A[i + 0] + 2 * a1b0 + a0a1b0b1) >> 4; |
| 170 | const int v1 = (8 * A[i + 1] + 2 * a0b1 + a0a1b0b1) >> 4; |
| 171 | out[2 * i + 0] = clip_SSE2(best_y[2 * i + 0] + v0, max_y); |
| 172 | out[2 * i + 1] = clip_SSE2(best_y[2 * i + 1] + v1, max_y); |
| 173 | } |
| 174 | } |
| 175 | |
| 176 | static void SharpYuvFilterRow_SSE2(const int16_t* A, const int16_t* B, int len, |
| 177 | const uint16_t* best_y, uint16_t* out, |
| 178 | int bit_depth) { |
| 179 | if (bit_depth <= 10) { |
| 180 | SharpYuvFilterRow16_SSE2(A, B, len, best_y, out, bit_depth); |
| 181 | } else { |
| 182 | SharpYuvFilterRow32_SSE2(A, B, len, best_y, out, bit_depth); |
| 183 | } |
| 184 | } |
| 185 | |
| 186 | //------------------------------------------------------------------------------ |
| 187 | |
| 188 | extern void InitSharpYuvSSE2(void); |
| 189 | |
| 190 | WEBP_TSAN_IGNORE_FUNCTION void InitSharpYuvSSE2(void) { |
| 191 | SharpYuvUpdateY = SharpYuvUpdateY_SSE2; |
| 192 | SharpYuvUpdateRGB = SharpYuvUpdateRGB_SSE2; |
| 193 | SharpYuvFilterRow = SharpYuvFilterRow_SSE2; |
| 194 | } |
| 195 | #else // !WEBP_USE_SSE2 |
| 196 | |
| 197 | extern void InitSharpYuvSSE2(void); |
| 198 | |
| 199 | void InitSharpYuvSSE2(void) {} |
| 200 | |
| 201 | #endif // WEBP_USE_SSE2 |
| 202 | |