1 | /* |
2 | Simple DirectMedia Layer |
3 | Copyright (C) 1997-2025 Sam Lantinga <slouken@libsdl.org> |
4 | |
5 | This software is provided 'as-is', without any express or implied |
6 | warranty. In no event will the authors be held liable for any damages |
7 | arising from the use of this software. |
8 | |
9 | Permission is granted to anyone to use this software for any purpose, |
10 | including commercial applications, and to alter it and redistribute it |
11 | freely, subject to the following restrictions: |
12 | |
13 | 1. The origin of this software must not be misrepresented; you must not |
14 | claim that you wrote the original software. If you use this software |
15 | in a product, an acknowledgment in the product documentation would be |
16 | appreciated but is not required. |
17 | 2. Altered source versions must be plainly marked as such, and must not be |
18 | misrepresented as being the original software. |
19 | 3. This notice may not be removed or altered from any source distribution. |
20 | */ |
21 | #include "SDL_internal.h" |
22 | |
23 | #include "SDL_surface_c.h" |
24 | |
25 | static bool SDL_StretchSurfaceUncheckedNearest(SDL_Surface *src, const SDL_Rect *srcrect, SDL_Surface *dst, const SDL_Rect *dstrect); |
26 | static bool SDL_StretchSurfaceUncheckedLinear(SDL_Surface *src, const SDL_Rect *srcrect, SDL_Surface *dst, const SDL_Rect *dstrect); |
27 | |
28 | bool SDL_StretchSurface(SDL_Surface *src, const SDL_Rect *srcrect, SDL_Surface *dst, const SDL_Rect *dstrect, SDL_ScaleMode scaleMode) |
29 | { |
30 | bool result; |
31 | int src_locked; |
32 | int dst_locked; |
33 | SDL_Rect full_src; |
34 | SDL_Rect full_dst; |
35 | |
36 | if (!src) { |
37 | return SDL_InvalidParamError("src" ); |
38 | } |
39 | if (!dst) { |
40 | return SDL_InvalidParamError("dst" ); |
41 | } |
42 | |
43 | if (src->format != dst->format) { |
44 | // Slow! |
45 | SDL_Surface *src_tmp = SDL_ConvertSurfaceAndColorspace(src, dst->format, dst->palette, dst->colorspace, dst->props); |
46 | if (!src_tmp) { |
47 | return false; |
48 | } |
49 | result = SDL_StretchSurface(src_tmp, srcrect, dst, dstrect, scaleMode); |
50 | SDL_DestroySurface(src_tmp); |
51 | return result; |
52 | } |
53 | |
54 | if (SDL_ISPIXELFORMAT_FOURCC(src->format)) { |
55 | // Slow! |
56 | if (!dstrect) { |
57 | full_dst.x = 0; |
58 | full_dst.y = 0; |
59 | full_dst.w = dst->w; |
60 | full_dst.h = dst->h; |
61 | dstrect = &full_dst; |
62 | } |
63 | |
64 | SDL_Surface *src_tmp = SDL_ConvertSurface(src, SDL_PIXELFORMAT_XRGB8888); |
65 | SDL_Surface *dst_tmp = SDL_CreateSurface(dstrect->w, dstrect->h, SDL_PIXELFORMAT_XRGB8888); |
66 | if (src_tmp && dst_tmp) { |
67 | result = SDL_StretchSurface(src_tmp, srcrect, dst_tmp, NULL, scaleMode); |
68 | if (result) { |
69 | result = SDL_ConvertPixelsAndColorspace(dstrect->w, dstrect->h, |
70 | dst_tmp->format, SDL_COLORSPACE_SRGB, 0, |
71 | dst_tmp->pixels, dst_tmp->pitch, |
72 | dst->format, dst->colorspace, SDL_GetSurfaceProperties(dst), |
73 | (Uint8 *)dst->pixels + dstrect->y * dst->pitch + dstrect->x * SDL_BYTESPERPIXEL(dst->format), dst->pitch); |
74 | } |
75 | } else { |
76 | result = false; |
77 | } |
78 | SDL_DestroySurface(src_tmp); |
79 | SDL_DestroySurface(dst_tmp); |
80 | return result; |
81 | } |
82 | |
83 | switch (scaleMode) { |
84 | case SDL_SCALEMODE_NEAREST: |
85 | break; |
86 | case SDL_SCALEMODE_LINEAR: |
87 | break; |
88 | case SDL_SCALEMODE_PIXELART: |
89 | scaleMode = SDL_SCALEMODE_NEAREST; |
90 | break; |
91 | default: |
92 | return SDL_InvalidParamError("scaleMode" ); |
93 | } |
94 | |
95 | if (scaleMode == SDL_SCALEMODE_LINEAR) { |
96 | if (SDL_BYTESPERPIXEL(src->format) != 4 || src->format == SDL_PIXELFORMAT_ARGB2101010) { |
97 | return SDL_SetError("Wrong format" ); |
98 | } |
99 | } |
100 | |
101 | // Verify the blit rectangles |
102 | if (srcrect) { |
103 | if ((srcrect->x < 0) || (srcrect->y < 0) || |
104 | ((srcrect->x + srcrect->w) > src->w) || |
105 | ((srcrect->y + srcrect->h) > src->h)) { |
106 | return SDL_SetError("Invalid source blit rectangle" ); |
107 | } |
108 | } else { |
109 | full_src.x = 0; |
110 | full_src.y = 0; |
111 | full_src.w = src->w; |
112 | full_src.h = src->h; |
113 | srcrect = &full_src; |
114 | } |
115 | if (dstrect) { |
116 | if ((dstrect->x < 0) || (dstrect->y < 0) || |
117 | ((dstrect->x + dstrect->w) > dst->w) || |
118 | ((dstrect->y + dstrect->h) > dst->h)) { |
119 | return SDL_SetError("Invalid destination blit rectangle" ); |
120 | } |
121 | } else { |
122 | full_dst.x = 0; |
123 | full_dst.y = 0; |
124 | full_dst.w = dst->w; |
125 | full_dst.h = dst->h; |
126 | dstrect = &full_dst; |
127 | } |
128 | |
129 | if (dstrect->w <= 0 || dstrect->h <= 0) { |
130 | return true; |
131 | } |
132 | |
133 | if (srcrect->w > SDL_MAX_UINT16 || srcrect->h > SDL_MAX_UINT16 || |
134 | dstrect->w > SDL_MAX_UINT16 || dstrect->h > SDL_MAX_UINT16) { |
135 | return SDL_SetError("Size too large for scaling" ); |
136 | } |
137 | |
138 | // Lock the destination if it's in hardware |
139 | dst_locked = 0; |
140 | if (SDL_MUSTLOCK(dst)) { |
141 | if (!SDL_LockSurface(dst)) { |
142 | return SDL_SetError("Unable to lock destination surface" ); |
143 | } |
144 | dst_locked = 1; |
145 | } |
146 | // Lock the source if it's in hardware |
147 | src_locked = 0; |
148 | if (SDL_MUSTLOCK(src)) { |
149 | if (!SDL_LockSurface(src)) { |
150 | if (dst_locked) { |
151 | SDL_UnlockSurface(dst); |
152 | } |
153 | return SDL_SetError("Unable to lock source surface" ); |
154 | } |
155 | src_locked = 1; |
156 | } |
157 | |
158 | if (scaleMode == SDL_SCALEMODE_NEAREST) { |
159 | result = SDL_StretchSurfaceUncheckedNearest(src, srcrect, dst, dstrect); |
160 | } else { |
161 | result = SDL_StretchSurfaceUncheckedLinear(src, srcrect, dst, dstrect); |
162 | } |
163 | |
164 | // We need to unlock the surfaces if they're locked |
165 | if (dst_locked) { |
166 | SDL_UnlockSurface(dst); |
167 | } |
168 | if (src_locked) { |
169 | SDL_UnlockSurface(src); |
170 | } |
171 | |
172 | return result; |
173 | } |
174 | |
175 | /* bilinear interpolation precision must be < 8 |
176 | Because with SSE: add-multiply: _mm_madd_epi16 works with signed int |
177 | so pixels 0xb1...... are negatives and false the result |
178 | same in NEON probably */ |
179 | #define PRECISION 7 |
180 | |
181 | #define FIXED_POINT(i) ((Uint32)(i) << 16) |
182 | #define SRC_INDEX(fp) ((Uint32)(fp) >> 16) |
183 | #define INTEGER(fp) ((Uint32)(fp) >> PRECISION) |
184 | #define FRAC(fp) ((Uint32)((fp) >> (16 - PRECISION)) & ((1 << PRECISION) - 1)) |
185 | #define FRAC_ZERO 0 |
186 | #define FRAC_ONE (1 << PRECISION) |
187 | #define FP_ONE FIXED_POINT(1) |
188 | |
189 | #define BILINEAR___START \ |
190 | int i; \ |
191 | Sint64 fp_sum_h; \ |
192 | int fp_step_h, left_pad_h, right_pad_h; \ |
193 | Sint64 fp_sum_w; \ |
194 | int fp_step_w, left_pad_w, right_pad_w; \ |
195 | Sint64 fp_sum_w_init; \ |
196 | int left_pad_w_init, right_pad_w_init, dst_gap, middle_init; \ |
197 | get_scaler_datas(src_h, dst_h, &fp_sum_h, &fp_step_h, &left_pad_h, &right_pad_h); \ |
198 | get_scaler_datas(src_w, dst_w, &fp_sum_w, &fp_step_w, &left_pad_w, &right_pad_w); \ |
199 | fp_sum_w_init = fp_sum_w + left_pad_w * fp_step_w; \ |
200 | left_pad_w_init = left_pad_w; \ |
201 | right_pad_w_init = right_pad_w; \ |
202 | dst_gap = dst_pitch - 4 * dst_w; \ |
203 | middle_init = dst_w - left_pad_w - right_pad_w; |
204 | |
205 | #define BILINEAR___HEIGHT \ |
206 | int index_h, frac_h0, frac_h1, middle; \ |
207 | const Uint32 *src_h0, *src_h1; \ |
208 | int no_padding; \ |
209 | Uint64 incr_h0, incr_h1; \ |
210 | \ |
211 | no_padding = !(i < left_pad_h || i > dst_h - 1 - right_pad_h); \ |
212 | index_h = SRC_INDEX(fp_sum_h); \ |
213 | frac_h0 = FRAC(fp_sum_h); \ |
214 | \ |
215 | index_h = no_padding ? index_h : (i < left_pad_h ? 0 : src_h - 1); \ |
216 | frac_h0 = no_padding ? frac_h0 : 0; \ |
217 | incr_h1 = no_padding ? src_pitch : 0; \ |
218 | incr_h0 = (Uint64)index_h * src_pitch; \ |
219 | \ |
220 | src_h0 = (const Uint32 *)((const Uint8 *)src + incr_h0); \ |
221 | src_h1 = (const Uint32 *)((const Uint8 *)src_h0 + incr_h1); \ |
222 | \ |
223 | fp_sum_h += fp_step_h; \ |
224 | \ |
225 | frac_h1 = FRAC_ONE - frac_h0; \ |
226 | fp_sum_w = fp_sum_w_init; \ |
227 | right_pad_w = right_pad_w_init; \ |
228 | left_pad_w = left_pad_w_init; \ |
229 | middle = middle_init; |
230 | |
231 | #ifdef __clang__ |
232 | // Remove inlining of this function |
233 | // Compiler crash with clang 9.0.8 / android-ndk-r21d |
234 | // Compiler crash with clang 11.0.3 / Xcode |
235 | // OK with clang 11.0.5 / android-ndk-22 |
236 | // OK with clang 12.0.0 / Xcode |
237 | __attribute__((noinline)) |
238 | #endif |
239 | static void get_scaler_datas(int src_nb, int dst_nb, Sint64 *fp_start, int *fp_step, int *left_pad, int *right_pad) |
240 | { |
241 | |
242 | int step = FIXED_POINT(src_nb) / (dst_nb); // source step in fixed point |
243 | int x0 = FP_ONE / 2; // dst first pixel center at 0.5 in fixed point |
244 | Sint64 fp_sum; |
245 | int i; |
246 | #if 0 |
247 | // scale to source coordinates |
248 | x0 *= src_nb; |
249 | x0 /= dst_nb; // x0 == step / 2 |
250 | #else |
251 | // Use this code for perfect match with pixman |
252 | Sint64 tmp[2]; |
253 | tmp[0] = (Sint64)step * (x0 >> 16); |
254 | tmp[1] = (Sint64)step * (x0 & 0xFFFF); |
255 | x0 = (int)(tmp[0] + ((tmp[1] + 0x8000) >> 16)); // x0 == (step + 1) / 2 |
256 | #endif |
257 | // -= 0.5, get back the pixel origin, in source coordinates |
258 | x0 -= FP_ONE / 2; |
259 | |
260 | *fp_start = x0; |
261 | *fp_step = step; |
262 | *left_pad = 0; |
263 | *right_pad = 0; |
264 | |
265 | fp_sum = x0; |
266 | for (i = 0; i < dst_nb; i++) { |
267 | if (fp_sum < 0) { |
268 | *left_pad += 1; |
269 | } else { |
270 | int index = SRC_INDEX(fp_sum); |
271 | if (index > src_nb - 2) { |
272 | *right_pad += 1; |
273 | } |
274 | } |
275 | fp_sum += step; |
276 | } |
277 | // SDL_Log("%d -> %d x0=%d step=%d left_pad=%d right_pad=%d", src_nb, dst_nb, *fp_start, *fp_step, *left_pad, *right_pad); |
278 | } |
279 | |
280 | typedef struct color_t |
281 | { |
282 | Uint8 a; |
283 | Uint8 b; |
284 | Uint8 c; |
285 | Uint8 d; |
286 | } color_t; |
287 | |
288 | #if 0 |
289 | static void printf_64(const char *str, void *var) |
290 | { |
291 | uint8_t *val = (uint8_t*) var; |
292 | printf(" * %s: %02x %02x %02x %02x _ %02x %02x %02x %02x\n" , |
293 | str, val[0], val[1], val[2], val[3], val[4], val[5], val[6], val[7]); |
294 | } |
295 | #endif |
296 | |
297 | /* Interpolated == x0 + frac * (x1 - x0) == x0 * (1 - frac) + x1 * frac */ |
298 | |
299 | static SDL_INLINE void INTERPOL(const Uint32 *src_x0, const Uint32 *src_x1, int frac0, int frac1, Uint32 *dst) |
300 | { |
301 | const color_t *c0 = (const color_t *)src_x0; |
302 | const color_t *c1 = (const color_t *)src_x1; |
303 | color_t *cx = (color_t *)dst; |
304 | #if 0 |
305 | cx->a = c0->a + INTEGER(frac0 * (c1->a - c0->a)); |
306 | cx->b = c0->b + INTEGER(frac0 * (c1->b - c0->b)); |
307 | cx->c = c0->c + INTEGER(frac0 * (c1->c - c0->c)); |
308 | cx->d = c0->d + INTEGER(frac0 * (c1->d - c0->d)); |
309 | #else |
310 | cx->a = (Uint8)INTEGER(frac1 * c0->a + frac0 * c1->a); |
311 | cx->b = (Uint8)INTEGER(frac1 * c0->b + frac0 * c1->b); |
312 | cx->c = (Uint8)INTEGER(frac1 * c0->c + frac0 * c1->c); |
313 | cx->d = (Uint8)INTEGER(frac1 * c0->d + frac0 * c1->d); |
314 | #endif |
315 | } |
316 | |
317 | static SDL_INLINE void INTERPOL_BILINEAR(const Uint32 *s0, const Uint32 *s1, int frac_w0, int frac_h0, int frac_h1, Uint32 *dst) |
318 | { |
319 | Uint32 tmp[2]; |
320 | unsigned int frac_w1 = FRAC_ONE - frac_w0; |
321 | |
322 | // Vertical first, store to 'tmp' |
323 | INTERPOL(s0, s1, frac_h0, frac_h1, tmp); |
324 | INTERPOL(s0 + 1, s1 + 1, frac_h0, frac_h1, tmp + 1); |
325 | |
326 | // Horizontal, store to 'dst' |
327 | INTERPOL(tmp, tmp + 1, frac_w0, frac_w1, dst); |
328 | } |
329 | |
330 | static bool scale_mat(const Uint32 *src, int src_w, int src_h, int src_pitch, Uint32 *dst, int dst_w, int dst_h, int dst_pitch) |
331 | { |
332 | BILINEAR___START |
333 | |
334 | for (i = 0; i < dst_h; i++) { |
335 | |
336 | BILINEAR___HEIGHT |
337 | |
338 | while (left_pad_w--) { |
339 | INTERPOL_BILINEAR(src_h0, src_h1, FRAC_ZERO, frac_h0, frac_h1, dst); |
340 | dst += 1; |
341 | } |
342 | |
343 | while (middle--) { |
344 | const Uint32 *s_00_01; |
345 | const Uint32 *s_10_11; |
346 | int index_w = 4 * SRC_INDEX(fp_sum_w); |
347 | int frac_w = FRAC(fp_sum_w); |
348 | fp_sum_w += fp_step_w; |
349 | |
350 | /* |
351 | x00 ... x0_ ..... x01 |
352 | . . . |
353 | . x . |
354 | . . . |
355 | . . . |
356 | x10 ... x1_ ..... x11 |
357 | */ |
358 | s_00_01 = (const Uint32 *)((const Uint8 *)src_h0 + index_w); |
359 | s_10_11 = (const Uint32 *)((const Uint8 *)src_h1 + index_w); |
360 | |
361 | INTERPOL_BILINEAR(s_00_01, s_10_11, frac_w, frac_h0, frac_h1, dst); |
362 | |
363 | dst += 1; |
364 | } |
365 | |
366 | while (right_pad_w--) { |
367 | int index_w = 4 * (src_w - 2); |
368 | const Uint32 *s_00_01 = (const Uint32 *)((const Uint8 *)src_h0 + index_w); |
369 | const Uint32 *s_10_11 = (const Uint32 *)((const Uint8 *)src_h1 + index_w); |
370 | INTERPOL_BILINEAR(s_00_01, s_10_11, FRAC_ONE, frac_h0, frac_h1, dst); |
371 | dst += 1; |
372 | } |
373 | dst = (Uint32 *)((Uint8 *)dst + dst_gap); |
374 | } |
375 | return true; |
376 | } |
377 | |
378 | #ifdef SDL_NEON_INTRINSICS |
379 | #define CAST_uint8x8_t (uint8x8_t) |
380 | #define CAST_uint32x2_t (uint32x2_t) |
381 | #endif |
382 | |
383 | #if defined(_MSC_VER) |
384 | #ifdef SDL_NEON_INTRINSICS |
385 | #undef CAST_uint8x8_t |
386 | #undef CAST_uint32x2_t |
387 | #define CAST_uint8x8_t |
388 | #define CAST_uint32x2_t |
389 | #endif |
390 | #endif |
391 | |
392 | #ifdef SDL_SSE2_INTRINSICS |
393 | |
394 | #if 0 |
395 | static void SDL_TARGETING("sse2" ) printf_128(const char *str, __m128i var) |
396 | { |
397 | uint16_t *val = (uint16_t*) &var; |
398 | printf(" * %s: %04x %04x %04x %04x _ %04x %04x %04x %04x\n" , |
399 | str, val[0], val[1], val[2], val[3], val[4], val[5], val[6], val[7]); |
400 | } |
401 | #endif |
402 | |
403 | static SDL_INLINE int hasSSE2(void) |
404 | { |
405 | static int val = -1; |
406 | if (val != -1) { |
407 | return val; |
408 | } |
409 | val = SDL_HasSSE2(); |
410 | return val; |
411 | } |
412 | |
413 | static SDL_INLINE void SDL_TARGETING("sse2" ) INTERPOL_BILINEAR_SSE(const Uint32 *s0, const Uint32 *s1, int frac_w, __m128i v_frac_h0, __m128i v_frac_h1, Uint32 *dst, __m128i zero) |
414 | { |
415 | __m128i x_00_01, x_10_11; /* Pixels in 4*uint8 in row */ |
416 | __m128i v_frac_w0, k0, l0, d0, e0; |
417 | |
418 | int f, f2; |
419 | f = frac_w; |
420 | f2 = FRAC_ONE - frac_w; |
421 | v_frac_w0 = _mm_set_epi16((short)f, (short)f2, (short)f, (short)f2, (short)f, (short)f2, (short)f, (short)f2); |
422 | |
423 | x_00_01 = _mm_loadl_epi64((const __m128i *)s0); // Load x00 and x01 |
424 | x_10_11 = _mm_loadl_epi64((const __m128i *)s1); |
425 | |
426 | /* Interpolated == x0 + frac * (x1 - x0) == x0 * (1 - frac) + x1 * frac */ |
427 | |
428 | // Interpolation vertical |
429 | k0 = _mm_mullo_epi16(_mm_unpacklo_epi8(x_00_01, zero), v_frac_h1); |
430 | l0 = _mm_mullo_epi16(_mm_unpacklo_epi8(x_10_11, zero), v_frac_h0); |
431 | k0 = _mm_add_epi16(k0, l0); |
432 | |
433 | // For perfect match, clear the factionnal part eventually. |
434 | /* |
435 | k0 = _mm_srli_epi16(k0, PRECISION); |
436 | k0 = _mm_slli_epi16(k0, PRECISION); |
437 | */ |
438 | |
439 | // Interpolation horizontal |
440 | l0 = _mm_unpacklo_epi64(/* unused */ l0, k0); |
441 | k0 = _mm_madd_epi16(_mm_unpackhi_epi16(l0, k0), v_frac_w0); |
442 | |
443 | // Store 1 pixel |
444 | d0 = _mm_srli_epi32(k0, PRECISION * 2); |
445 | e0 = _mm_packs_epi32(d0, d0); |
446 | e0 = _mm_packus_epi16(e0, e0); |
447 | *dst = _mm_cvtsi128_si32(e0); |
448 | } |
449 | |
450 | static bool SDL_TARGETING("sse2" ) scale_mat_SSE(const Uint32 *src, int src_w, int src_h, int src_pitch, Uint32 *dst, int dst_w, int dst_h, int dst_pitch) |
451 | { |
452 | BILINEAR___START |
453 | |
454 | for (i = 0; i < dst_h; i++) { |
455 | int nb_block2; |
456 | __m128i v_frac_h0; |
457 | __m128i v_frac_h1; |
458 | __m128i zero; |
459 | |
460 | BILINEAR___HEIGHT |
461 | |
462 | nb_block2 = middle / 2; |
463 | |
464 | v_frac_h0 = _mm_set_epi16((short)frac_h0, (short)frac_h0, (short)frac_h0, (short)frac_h0, (short)frac_h0, (short)frac_h0, (short)frac_h0, (short)frac_h0); |
465 | v_frac_h1 = _mm_set_epi16((short)frac_h1, (short)frac_h1, (short)frac_h1, (short)frac_h1, (short)frac_h1, (short)frac_h1, (short)frac_h1, (short)frac_h1); |
466 | zero = _mm_setzero_si128(); |
467 | |
468 | while (left_pad_w--) { |
469 | INTERPOL_BILINEAR_SSE(src_h0, src_h1, FRAC_ZERO, v_frac_h0, v_frac_h1, dst, zero); |
470 | dst += 1; |
471 | } |
472 | |
473 | while (nb_block2--) { |
474 | int index_w_0, frac_w_0; |
475 | int index_w_1, frac_w_1; |
476 | |
477 | const Uint32 *s_00_01, *s_02_03, *s_10_11, *s_12_13; |
478 | |
479 | __m128i x_00_01, x_10_11, x_02_03, x_12_13; /* Pixels in 4*uint8 in row */ |
480 | __m128i v_frac_w0, k0, l0, d0, e0; |
481 | __m128i v_frac_w1, k1, l1, d1, e1; |
482 | |
483 | int f, f2; |
484 | index_w_0 = 4 * SRC_INDEX(fp_sum_w); |
485 | frac_w_0 = FRAC(fp_sum_w); |
486 | fp_sum_w += fp_step_w; |
487 | index_w_1 = 4 * SRC_INDEX(fp_sum_w); |
488 | frac_w_1 = FRAC(fp_sum_w); |
489 | fp_sum_w += fp_step_w; |
490 | /* |
491 | x00............ x01 x02...........x03 |
492 | . . . . . . |
493 | j0 f0 j1 j2 f1 j3 |
494 | . . . . . . |
495 | . . . . . . |
496 | . . . . . . |
497 | x10............ x11 x12...........x13 |
498 | */ |
499 | s_00_01 = (const Uint32 *)((const Uint8 *)src_h0 + index_w_0); |
500 | s_02_03 = (const Uint32 *)((const Uint8 *)src_h0 + index_w_1); |
501 | s_10_11 = (const Uint32 *)((const Uint8 *)src_h1 + index_w_0); |
502 | s_12_13 = (const Uint32 *)((const Uint8 *)src_h1 + index_w_1); |
503 | |
504 | f = frac_w_0; |
505 | f2 = FRAC_ONE - frac_w_0; |
506 | v_frac_w0 = _mm_set_epi16((short)f, (short)f2, (short)f, (short)f2, (short)f, (short)f2, (short)f, (short)f2); |
507 | |
508 | f = frac_w_1; |
509 | f2 = FRAC_ONE - frac_w_1; |
510 | v_frac_w1 = _mm_set_epi16((short)f, (short)f2, (short)f, (short)f2, (short)f, (short)f2, (short)f, (short)f2); |
511 | |
512 | x_00_01 = _mm_loadl_epi64((const __m128i *)s_00_01); // Load x00 and x01 |
513 | x_02_03 = _mm_loadl_epi64((const __m128i *)s_02_03); |
514 | x_10_11 = _mm_loadl_epi64((const __m128i *)s_10_11); |
515 | x_12_13 = _mm_loadl_epi64((const __m128i *)s_12_13); |
516 | |
517 | // Interpolation vertical |
518 | k0 = _mm_mullo_epi16(_mm_unpacklo_epi8(x_00_01, zero), v_frac_h1); |
519 | l0 = _mm_mullo_epi16(_mm_unpacklo_epi8(x_10_11, zero), v_frac_h0); |
520 | k0 = _mm_add_epi16(k0, l0); |
521 | k1 = _mm_mullo_epi16(_mm_unpacklo_epi8(x_02_03, zero), v_frac_h1); |
522 | l1 = _mm_mullo_epi16(_mm_unpacklo_epi8(x_12_13, zero), v_frac_h0); |
523 | k1 = _mm_add_epi16(k1, l1); |
524 | |
525 | // Interpolation horizontal |
526 | l0 = _mm_unpacklo_epi64(/* unused */ l0, k0); |
527 | k0 = _mm_madd_epi16(_mm_unpackhi_epi16(l0, k0), v_frac_w0); |
528 | l1 = _mm_unpacklo_epi64(/* unused */ l1, k1); |
529 | k1 = _mm_madd_epi16(_mm_unpackhi_epi16(l1, k1), v_frac_w1); |
530 | |
531 | // Store 1 pixel |
532 | d0 = _mm_srli_epi32(k0, PRECISION * 2); |
533 | e0 = _mm_packs_epi32(d0, d0); |
534 | e0 = _mm_packus_epi16(e0, e0); |
535 | *dst++ = _mm_cvtsi128_si32(e0); |
536 | |
537 | // Store 1 pixel |
538 | d1 = _mm_srli_epi32(k1, PRECISION * 2); |
539 | e1 = _mm_packs_epi32(d1, d1); |
540 | e1 = _mm_packus_epi16(e1, e1); |
541 | *dst++ = _mm_cvtsi128_si32(e1); |
542 | } |
543 | |
544 | // Last point |
545 | if (middle & 0x1) { |
546 | const Uint32 *s_00_01; |
547 | const Uint32 *s_10_11; |
548 | int index_w = 4 * SRC_INDEX(fp_sum_w); |
549 | int frac_w = FRAC(fp_sum_w); |
550 | fp_sum_w += fp_step_w; |
551 | s_00_01 = (const Uint32 *)((const Uint8 *)src_h0 + index_w); |
552 | s_10_11 = (const Uint32 *)((const Uint8 *)src_h1 + index_w); |
553 | INTERPOL_BILINEAR_SSE(s_00_01, s_10_11, frac_w, v_frac_h0, v_frac_h1, dst, zero); |
554 | dst += 1; |
555 | } |
556 | |
557 | while (right_pad_w--) { |
558 | int index_w = 4 * (src_w - 2); |
559 | const Uint32 *s_00_01 = (const Uint32 *)((const Uint8 *)src_h0 + index_w); |
560 | const Uint32 *s_10_11 = (const Uint32 *)((const Uint8 *)src_h1 + index_w); |
561 | INTERPOL_BILINEAR_SSE(s_00_01, s_10_11, FRAC_ONE, v_frac_h0, v_frac_h1, dst, zero); |
562 | dst += 1; |
563 | } |
564 | dst = (Uint32 *)((Uint8 *)dst + dst_gap); |
565 | } |
566 | return true; |
567 | } |
568 | #endif |
569 | |
570 | #ifdef SDL_NEON_INTRINSICS |
571 | |
572 | static SDL_INLINE int hasNEON(void) |
573 | { |
574 | static int val = -1; |
575 | if (val != -1) { |
576 | return val; |
577 | } |
578 | val = SDL_HasNEON(); |
579 | return val; |
580 | } |
581 | |
582 | static SDL_INLINE void INTERPOL_BILINEAR_NEON(const Uint32 *s0, const Uint32 *s1, int frac_w, uint8x8_t v_frac_h0, uint8x8_t v_frac_h1, Uint32 *dst) |
583 | { |
584 | uint8x8_t x_00_01, x_10_11; /* Pixels in 4*uint8 in row */ |
585 | uint16x8_t k0; |
586 | uint32x4_t l0; |
587 | uint16x8_t d0; |
588 | uint8x8_t e0; |
589 | |
590 | x_00_01 = CAST_uint8x8_t vld1_u32(s0); // Load 2 pixels |
591 | x_10_11 = CAST_uint8x8_t vld1_u32(s1); |
592 | |
593 | /* Interpolated == x0 + frac * (x1 - x0) == x0 * (1 - frac) + x1 * frac */ |
594 | k0 = vmull_u8(x_00_01, v_frac_h1); /* k0 := x0 * (1 - frac) */ |
595 | k0 = vmlal_u8(k0, x_10_11, v_frac_h0); /* k0 += x1 * frac */ |
596 | |
597 | // k0 now contains 2 interpolated pixels { j0, j1 } |
598 | l0 = vshll_n_u16(vget_low_u16(k0), PRECISION); |
599 | l0 = vmlsl_n_u16(l0, vget_low_u16(k0), frac_w); |
600 | l0 = vmlal_n_u16(l0, vget_high_u16(k0), frac_w); |
601 | |
602 | // Shift and narrow |
603 | d0 = vcombine_u16( |
604 | /* uint16x4_t */ vshrn_n_u32(l0, 2 * PRECISION), |
605 | /* uint16x4_t */ vshrn_n_u32(l0, 2 * PRECISION)); |
606 | |
607 | // Narrow again |
608 | e0 = vmovn_u16(d0); |
609 | |
610 | // Store 1 pixel |
611 | *dst = vget_lane_u32(CAST_uint32x2_t e0, 0); |
612 | } |
613 | |
614 | static bool scale_mat_NEON(const Uint32 *src, int src_w, int src_h, int src_pitch, Uint32 *dst, int dst_w, int dst_h, int dst_pitch) |
615 | { |
616 | BILINEAR___START |
617 | |
618 | for (i = 0; i < dst_h; i++) { |
619 | int nb_block4; |
620 | uint8x8_t v_frac_h0, v_frac_h1; |
621 | |
622 | BILINEAR___HEIGHT |
623 | |
624 | nb_block4 = middle / 4; |
625 | |
626 | v_frac_h0 = vmov_n_u8(frac_h0); |
627 | v_frac_h1 = vmov_n_u8(frac_h1); |
628 | |
629 | while (left_pad_w--) { |
630 | INTERPOL_BILINEAR_NEON(src_h0, src_h1, FRAC_ZERO, v_frac_h0, v_frac_h1, dst); |
631 | dst += 1; |
632 | } |
633 | |
634 | while (nb_block4--) { |
635 | int index_w_0, frac_w_0; |
636 | int index_w_1, frac_w_1; |
637 | int index_w_2, frac_w_2; |
638 | int index_w_3, frac_w_3; |
639 | |
640 | const Uint32 *s_00_01, *s_02_03, *s_04_05, *s_06_07; |
641 | const Uint32 *s_10_11, *s_12_13, *s_14_15, *s_16_17; |
642 | |
643 | uint8x8_t x_00_01, x_10_11, x_02_03, x_12_13; /* Pixels in 4*uint8 in row */ |
644 | uint8x8_t x_04_05, x_14_15, x_06_07, x_16_17; |
645 | |
646 | uint16x8_t k0, k1, k2, k3; |
647 | uint32x4_t l0, l1, l2, l3; |
648 | uint16x8_t d0, d1; |
649 | uint8x8_t e0, e1; |
650 | uint32x4_t f0; |
651 | |
652 | index_w_0 = 4 * SRC_INDEX(fp_sum_w); |
653 | frac_w_0 = FRAC(fp_sum_w); |
654 | fp_sum_w += fp_step_w; |
655 | index_w_1 = 4 * SRC_INDEX(fp_sum_w); |
656 | frac_w_1 = FRAC(fp_sum_w); |
657 | fp_sum_w += fp_step_w; |
658 | index_w_2 = 4 * SRC_INDEX(fp_sum_w); |
659 | frac_w_2 = FRAC(fp_sum_w); |
660 | fp_sum_w += fp_step_w; |
661 | index_w_3 = 4 * SRC_INDEX(fp_sum_w); |
662 | frac_w_3 = FRAC(fp_sum_w); |
663 | fp_sum_w += fp_step_w; |
664 | |
665 | s_00_01 = (const Uint32 *)((const Uint8 *)src_h0 + index_w_0); |
666 | s_02_03 = (const Uint32 *)((const Uint8 *)src_h0 + index_w_1); |
667 | s_04_05 = (const Uint32 *)((const Uint8 *)src_h0 + index_w_2); |
668 | s_06_07 = (const Uint32 *)((const Uint8 *)src_h0 + index_w_3); |
669 | s_10_11 = (const Uint32 *)((const Uint8 *)src_h1 + index_w_0); |
670 | s_12_13 = (const Uint32 *)((const Uint8 *)src_h1 + index_w_1); |
671 | s_14_15 = (const Uint32 *)((const Uint8 *)src_h1 + index_w_2); |
672 | s_16_17 = (const Uint32 *)((const Uint8 *)src_h1 + index_w_3); |
673 | |
674 | // Interpolation vertical |
675 | x_00_01 = CAST_uint8x8_t vld1_u32(s_00_01); // Load 2 pixels |
676 | x_02_03 = CAST_uint8x8_t vld1_u32(s_02_03); |
677 | x_04_05 = CAST_uint8x8_t vld1_u32(s_04_05); |
678 | x_06_07 = CAST_uint8x8_t vld1_u32(s_06_07); |
679 | x_10_11 = CAST_uint8x8_t vld1_u32(s_10_11); |
680 | x_12_13 = CAST_uint8x8_t vld1_u32(s_12_13); |
681 | x_14_15 = CAST_uint8x8_t vld1_u32(s_14_15); |
682 | x_16_17 = CAST_uint8x8_t vld1_u32(s_16_17); |
683 | |
684 | /* Interpolated == x0 + frac * (x1 - x0) == x0 * (1 - frac) + x1 * frac */ |
685 | k0 = vmull_u8(x_00_01, v_frac_h1); /* k0 := x0 * (1 - frac) */ |
686 | k0 = vmlal_u8(k0, x_10_11, v_frac_h0); /* k0 += x1 * frac */ |
687 | |
688 | k1 = vmull_u8(x_02_03, v_frac_h1); |
689 | k1 = vmlal_u8(k1, x_12_13, v_frac_h0); |
690 | |
691 | k2 = vmull_u8(x_04_05, v_frac_h1); |
692 | k2 = vmlal_u8(k2, x_14_15, v_frac_h0); |
693 | |
694 | k3 = vmull_u8(x_06_07, v_frac_h1); |
695 | k3 = vmlal_u8(k3, x_16_17, v_frac_h0); |
696 | |
697 | // k0 now contains 2 interpolated pixels { j0, j1 } |
698 | // k1 now contains 2 interpolated pixels { j2, j3 } |
699 | // k2 now contains 2 interpolated pixels { j4, j5 } |
700 | // k3 now contains 2 interpolated pixels { j6, j7 } |
701 | |
702 | l0 = vshll_n_u16(vget_low_u16(k0), PRECISION); |
703 | l0 = vmlsl_n_u16(l0, vget_low_u16(k0), frac_w_0); |
704 | l0 = vmlal_n_u16(l0, vget_high_u16(k0), frac_w_0); |
705 | |
706 | l1 = vshll_n_u16(vget_low_u16(k1), PRECISION); |
707 | l1 = vmlsl_n_u16(l1, vget_low_u16(k1), frac_w_1); |
708 | l1 = vmlal_n_u16(l1, vget_high_u16(k1), frac_w_1); |
709 | |
710 | l2 = vshll_n_u16(vget_low_u16(k2), PRECISION); |
711 | l2 = vmlsl_n_u16(l2, vget_low_u16(k2), frac_w_2); |
712 | l2 = vmlal_n_u16(l2, vget_high_u16(k2), frac_w_2); |
713 | |
714 | l3 = vshll_n_u16(vget_low_u16(k3), PRECISION); |
715 | l3 = vmlsl_n_u16(l3, vget_low_u16(k3), frac_w_3); |
716 | l3 = vmlal_n_u16(l3, vget_high_u16(k3), frac_w_3); |
717 | |
718 | // shift and narrow |
719 | d0 = vcombine_u16( |
720 | /* uint16x4_t */ vshrn_n_u32(l0, 2 * PRECISION), |
721 | /* uint16x4_t */ vshrn_n_u32(l1, 2 * PRECISION)); |
722 | // narrow again |
723 | e0 = vmovn_u16(d0); |
724 | |
725 | // Shift and narrow |
726 | d1 = vcombine_u16( |
727 | /* uint16x4_t */ vshrn_n_u32(l2, 2 * PRECISION), |
728 | /* uint16x4_t */ vshrn_n_u32(l3, 2 * PRECISION)); |
729 | // Narrow again |
730 | e1 = vmovn_u16(d1); |
731 | |
732 | f0 = vcombine_u32(CAST_uint32x2_t e0, CAST_uint32x2_t e1); |
733 | // Store 4 pixels |
734 | vst1q_u32(dst, f0); |
735 | |
736 | dst += 4; |
737 | } |
738 | |
739 | if (middle & 0x2) { |
740 | int index_w_0, frac_w_0; |
741 | int index_w_1, frac_w_1; |
742 | const Uint32 *s_00_01, *s_02_03; |
743 | const Uint32 *s_10_11, *s_12_13; |
744 | uint8x8_t x_00_01, x_10_11, x_02_03, x_12_13; /* Pixels in 4*uint8 in row */ |
745 | uint16x8_t k0, k1; |
746 | uint32x4_t l0, l1; |
747 | uint16x8_t d0; |
748 | uint8x8_t e0; |
749 | |
750 | index_w_0 = 4 * SRC_INDEX(fp_sum_w); |
751 | frac_w_0 = FRAC(fp_sum_w); |
752 | fp_sum_w += fp_step_w; |
753 | index_w_1 = 4 * SRC_INDEX(fp_sum_w); |
754 | frac_w_1 = FRAC(fp_sum_w); |
755 | fp_sum_w += fp_step_w; |
756 | /* |
757 | x00............ x01 x02...........x03 |
758 | . . . . . . |
759 | j0 dest0 j1 j2 dest1 j3 |
760 | . . . . . . |
761 | . . . . . . |
762 | . . . . . . |
763 | x10............ x11 x12...........x13 |
764 | */ |
765 | s_00_01 = (const Uint32 *)((const Uint8 *)src_h0 + index_w_0); |
766 | s_02_03 = (const Uint32 *)((const Uint8 *)src_h0 + index_w_1); |
767 | s_10_11 = (const Uint32 *)((const Uint8 *)src_h1 + index_w_0); |
768 | s_12_13 = (const Uint32 *)((const Uint8 *)src_h1 + index_w_1); |
769 | |
770 | // Interpolation vertical |
771 | x_00_01 = CAST_uint8x8_t vld1_u32(s_00_01); // Load 2 pixels |
772 | x_02_03 = CAST_uint8x8_t vld1_u32(s_02_03); |
773 | x_10_11 = CAST_uint8x8_t vld1_u32(s_10_11); |
774 | x_12_13 = CAST_uint8x8_t vld1_u32(s_12_13); |
775 | |
776 | /* Interpolated == x0 + frac * (x1 - x0) == x0 * (1 - frac) + x1 * frac */ |
777 | k0 = vmull_u8(x_00_01, v_frac_h1); /* k0 := x0 * (1 - frac) */ |
778 | k0 = vmlal_u8(k0, x_10_11, v_frac_h0); /* k0 += x1 * frac */ |
779 | |
780 | k1 = vmull_u8(x_02_03, v_frac_h1); |
781 | k1 = vmlal_u8(k1, x_12_13, v_frac_h0); |
782 | |
783 | // k0 now contains 2 interpolated pixels { j0, j1 } |
784 | // k1 now contains 2 interpolated pixels { j2, j3 } |
785 | |
786 | l0 = vshll_n_u16(vget_low_u16(k0), PRECISION); |
787 | l0 = vmlsl_n_u16(l0, vget_low_u16(k0), frac_w_0); |
788 | l0 = vmlal_n_u16(l0, vget_high_u16(k0), frac_w_0); |
789 | |
790 | l1 = vshll_n_u16(vget_low_u16(k1), PRECISION); |
791 | l1 = vmlsl_n_u16(l1, vget_low_u16(k1), frac_w_1); |
792 | l1 = vmlal_n_u16(l1, vget_high_u16(k1), frac_w_1); |
793 | |
794 | // Shift and narrow |
795 | |
796 | d0 = vcombine_u16( |
797 | /* uint16x4_t */ vshrn_n_u32(l0, 2 * PRECISION), |
798 | /* uint16x4_t */ vshrn_n_u32(l1, 2 * PRECISION)); |
799 | |
800 | // Narrow again |
801 | e0 = vmovn_u16(d0); |
802 | |
803 | // Store 2 pixels |
804 | vst1_u32(dst, CAST_uint32x2_t e0); |
805 | dst += 2; |
806 | } |
807 | |
808 | // Last point |
809 | if (middle & 0x1) { |
810 | int index_w = 4 * SRC_INDEX(fp_sum_w); |
811 | int frac_w = FRAC(fp_sum_w); |
812 | const Uint32 *s_00_01 = (const Uint32 *)((const Uint8 *)src_h0 + index_w); |
813 | const Uint32 *s_10_11 = (const Uint32 *)((const Uint8 *)src_h1 + index_w); |
814 | INTERPOL_BILINEAR_NEON(s_00_01, s_10_11, frac_w, v_frac_h0, v_frac_h1, dst); |
815 | dst += 1; |
816 | } |
817 | |
818 | while (right_pad_w--) { |
819 | int index_w = 4 * (src_w - 2); |
820 | const Uint32 *s_00_01 = (const Uint32 *)((const Uint8 *)src_h0 + index_w); |
821 | const Uint32 *s_10_11 = (const Uint32 *)((const Uint8 *)src_h1 + index_w); |
822 | INTERPOL_BILINEAR_NEON(s_00_01, s_10_11, FRAC_ONE, v_frac_h0, v_frac_h1, dst); |
823 | dst += 1; |
824 | } |
825 | |
826 | dst = (Uint32 *)((Uint8 *)dst + dst_gap); |
827 | } |
828 | return true; |
829 | } |
830 | #endif |
831 | |
832 | bool SDL_StretchSurfaceUncheckedLinear(SDL_Surface *s, const SDL_Rect *srcrect, SDL_Surface *d, const SDL_Rect *dstrect) |
833 | { |
834 | bool result = false; |
835 | int src_w = srcrect->w; |
836 | int src_h = srcrect->h; |
837 | int dst_w = dstrect->w; |
838 | int dst_h = dstrect->h; |
839 | int src_pitch = s->pitch; |
840 | int dst_pitch = d->pitch; |
841 | Uint32 *src = (Uint32 *)((Uint8 *)s->pixels + srcrect->x * 4 + srcrect->y * src_pitch); |
842 | Uint32 *dst = (Uint32 *)((Uint8 *)d->pixels + dstrect->x * 4 + dstrect->y * dst_pitch); |
843 | |
844 | #ifdef SDL_NEON_INTRINSICS |
845 | if (!result && hasNEON()) { |
846 | result = scale_mat_NEON(src, src_w, src_h, src_pitch, dst, dst_w, dst_h, dst_pitch); |
847 | } |
848 | #endif |
849 | |
850 | #ifdef SDL_SSE2_INTRINSICS |
851 | if (!result && hasSSE2()) { |
852 | result = scale_mat_SSE(src, src_w, src_h, src_pitch, dst, dst_w, dst_h, dst_pitch); |
853 | } |
854 | #endif |
855 | |
856 | if (!result) { |
857 | result = scale_mat(src, src_w, src_h, src_pitch, dst, dst_w, dst_h, dst_pitch); |
858 | } |
859 | |
860 | return result; |
861 | } |
862 | |
863 | #define SDL_SCALE_NEAREST__START \ |
864 | int i; \ |
865 | Uint64 posy, incy; \ |
866 | Uint64 posx, incx; \ |
867 | Uint64 srcy, srcx; \ |
868 | int dst_gap, n; \ |
869 | const Uint32 *src_h0; \ |
870 | incy = ((Uint64)src_h << 16) / dst_h; \ |
871 | incx = ((Uint64)src_w << 16) / dst_w; \ |
872 | dst_gap = dst_pitch - bpp * dst_w; \ |
873 | posy = incy / 2; |
874 | |
875 | #define SDL_SCALE_NEAREST__HEIGHT \ |
876 | srcy = (posy >> 16); \ |
877 | src_h0 = (const Uint32 *)((const Uint8 *)src_ptr + srcy * src_pitch); \ |
878 | posy += incy; \ |
879 | posx = incx / 2; \ |
880 | n = dst_w; |
881 | |
882 | static bool scale_mat_nearest_1(const Uint32 *src_ptr, int src_w, int src_h, int src_pitch, Uint32 *dst, int dst_w, int dst_h, int dst_pitch) |
883 | { |
884 | Uint32 bpp = 1; |
885 | SDL_SCALE_NEAREST__START |
886 | for (i = 0; i < dst_h; i++) { |
887 | SDL_SCALE_NEAREST__HEIGHT |
888 | while (n--) { |
889 | const Uint8 *src; |
890 | srcx = bpp * (posx >> 16); |
891 | posx += incx; |
892 | src = (const Uint8 *)src_h0 + srcx; |
893 | *(Uint8 *)dst = *src; |
894 | dst = (Uint32 *)((Uint8 *)dst + bpp); |
895 | } |
896 | dst = (Uint32 *)((Uint8 *)dst + dst_gap); |
897 | } |
898 | return true; |
899 | } |
900 | |
901 | static bool scale_mat_nearest_2(const Uint32 *src_ptr, int src_w, int src_h, int src_pitch, Uint32 *dst, int dst_w, int dst_h, int dst_pitch) |
902 | { |
903 | Uint32 bpp = 2; |
904 | SDL_SCALE_NEAREST__START |
905 | for (i = 0; i < dst_h; i++) { |
906 | SDL_SCALE_NEAREST__HEIGHT |
907 | while (n--) { |
908 | const Uint16 *src; |
909 | srcx = bpp * (posx >> 16); |
910 | posx += incx; |
911 | src = (const Uint16 *)((const Uint8 *)src_h0 + srcx); |
912 | *(Uint16 *)dst = *src; |
913 | dst = (Uint32 *)((Uint8 *)dst + bpp); |
914 | } |
915 | dst = (Uint32 *)((Uint8 *)dst + dst_gap); |
916 | } |
917 | return true; |
918 | } |
919 | |
920 | static bool scale_mat_nearest_3(const Uint32 *src_ptr, int src_w, int src_h, int src_pitch, Uint32 *dst, int dst_w, int dst_h, int dst_pitch) |
921 | { |
922 | Uint32 bpp = 3; |
923 | SDL_SCALE_NEAREST__START |
924 | for (i = 0; i < dst_h; i++) { |
925 | SDL_SCALE_NEAREST__HEIGHT |
926 | while (n--) { |
927 | const Uint8 *src; |
928 | srcx = bpp * (posx >> 16); |
929 | posx += incx; |
930 | src = (const Uint8 *)src_h0 + srcx; |
931 | ((Uint8 *)dst)[0] = src[0]; |
932 | ((Uint8 *)dst)[1] = src[1]; |
933 | ((Uint8 *)dst)[2] = src[2]; |
934 | dst = (Uint32 *)((Uint8 *)dst + bpp); |
935 | } |
936 | dst = (Uint32 *)((Uint8 *)dst + dst_gap); |
937 | } |
938 | return true; |
939 | } |
940 | |
941 | static bool scale_mat_nearest_4(const Uint32 *src_ptr, int src_w, int src_h, int src_pitch, Uint32 *dst, int dst_w, int dst_h, int dst_pitch) |
942 | { |
943 | Uint32 bpp = 4; |
944 | SDL_SCALE_NEAREST__START |
945 | for (i = 0; i < dst_h; i++) { |
946 | SDL_SCALE_NEAREST__HEIGHT |
947 | while (n--) { |
948 | const Uint32 *src; |
949 | srcx = bpp * (posx >> 16); |
950 | posx += incx; |
951 | src = (const Uint32 *)((const Uint8 *)src_h0 + srcx); |
952 | *dst = *src; |
953 | dst = (Uint32 *)((Uint8 *)dst + bpp); |
954 | } |
955 | dst = (Uint32 *)((Uint8 *)dst + dst_gap); |
956 | } |
957 | return true; |
958 | } |
959 | |
960 | bool SDL_StretchSurfaceUncheckedNearest(SDL_Surface *s, const SDL_Rect *srcrect, SDL_Surface *d, const SDL_Rect *dstrect) |
961 | { |
962 | int src_w = srcrect->w; |
963 | int src_h = srcrect->h; |
964 | int dst_w = dstrect->w; |
965 | int dst_h = dstrect->h; |
966 | int src_pitch = s->pitch; |
967 | int dst_pitch = d->pitch; |
968 | int bpp = SDL_BYTESPERPIXEL(d->format); |
969 | |
970 | Uint32 *src = (Uint32 *)((Uint8 *)s->pixels + srcrect->x * bpp + srcrect->y * src_pitch); |
971 | Uint32 *dst = (Uint32 *)((Uint8 *)d->pixels + dstrect->x * bpp + dstrect->y * dst_pitch); |
972 | |
973 | if (bpp == 4) { |
974 | return scale_mat_nearest_4(src, src_w, src_h, src_pitch, dst, dst_w, dst_h, dst_pitch); |
975 | } else if (bpp == 3) { |
976 | return scale_mat_nearest_3(src, src_w, src_h, src_pitch, dst, dst_w, dst_h, dst_pitch); |
977 | } else if (bpp == 2) { |
978 | return scale_mat_nearest_2(src, src_w, src_h, src_pitch, dst, dst_w, dst_h, dst_pitch); |
979 | } else { |
980 | return scale_mat_nearest_1(src, src_w, src_h, src_pitch, dst, dst_w, dst_h, dst_pitch); |
981 | } |
982 | } |
983 | |