comparison src/video/SDL_blit_N.c @ 1895:c121d94672cb

SDL 1.2 is moving to a branch, and SDL 1.3 is becoming the head.
author Sam Lantinga <slouken@libsdl.org>
date Mon, 10 Jul 2006 21:04:37 +0000
parents 398ac0f88e4d
children 8055185ae4ed
comparison
equal deleted inserted replaced
1894:c69cee13dd76 1895:c121d94672cb
36 #include <altivec.h> 36 #include <altivec.h>
37 #endif 37 #endif
38 #define assert(X) 38 #define assert(X)
39 #ifdef __MACOSX__ 39 #ifdef __MACOSX__
40 #include <sys/sysctl.h> 40 #include <sys/sysctl.h>
41 static size_t GetL3CacheSize( void ) 41 static size_t
42 GetL3CacheSize(void)
42 { 43 {
43 const char key[] = "hw.l3cachesize"; 44 const char key[] = "hw.l3cachesize";
44 u_int64_t result = 0; 45 u_int64_t result = 0;
45 size_t typeSize = sizeof( result ); 46 size_t typeSize = sizeof(result);
46 47
47 48
48 int err = sysctlbyname( key, &result, &typeSize, NULL, 0 ); 49 int err = sysctlbyname(key, &result, &typeSize, NULL, 0);
49 if( 0 != err ) return 0; 50 if (0 != err)
51 return 0;
50 52
51 return result; 53 return result;
52 } 54 }
53 #else 55 #else
54 static size_t GetL3CacheSize( void ) 56 static size_t
57 GetL3CacheSize(void)
55 { 58 {
56 /* XXX: Just guess G4 */ 59 /* XXX: Just guess G4 */
57 return 2097152; 60 return 2097152;
58 } 61 }
59 #endif /* __MACOSX__ */ 62 #endif /* __MACOSX__ */
60 63
61 #if (defined(__MACOSX__) && (__GNUC__ < 4)) 64 #if (defined(__MACOSX__) && (__GNUC__ < 4))
62 #define VECUINT8_LITERAL(a,b,c,d,e,f,g,h,i,j,k,l,m,n,o,p) \ 65 #define VECUINT8_LITERAL(a,b,c,d,e,f,g,h,i,j,k,l,m,n,o,p) \
63 (vector unsigned char) ( a,b,c,d,e,f,g,h,i,j,k,l,m,n,o,p ) 66 (vector unsigned char) ( a,b,c,d,e,f,g,h,i,j,k,l,m,n,o,p )
64 #define VECUINT16_LITERAL(a,b,c,d,e,f,g,h) \ 67 #define VECUINT16_LITERAL(a,b,c,d,e,f,g,h) \
65 (vector unsigned short) ( a,b,c,d,e,f,g,h ) 68 (vector unsigned short) ( a,b,c,d,e,f,g,h )
66 #else 69 #else
67 #define VECUINT8_LITERAL(a,b,c,d,e,f,g,h,i,j,k,l,m,n,o,p) \ 70 #define VECUINT8_LITERAL(a,b,c,d,e,f,g,h,i,j,k,l,m,n,o,p) \
68 (vector unsigned char) { a,b,c,d,e,f,g,h,i,j,k,l,m,n,o,p } 71 (vector unsigned char) { a,b,c,d,e,f,g,h,i,j,k,l,m,n,o,p }
69 #define VECUINT16_LITERAL(a,b,c,d,e,f,g,h) \ 72 #define VECUINT16_LITERAL(a,b,c,d,e,f,g,h) \
70 (vector unsigned short) { a,b,c,d,e,f,g,h } 73 (vector unsigned short) { a,b,c,d,e,f,g,h }
71 #endif 74 #endif
72 75
73 #define UNALIGNED_PTR(x) (((size_t) x) & 0x0000000F) 76 #define UNALIGNED_PTR(x) (((size_t) x) & 0x0000000F)
74 #define VSWIZZLE32(a,b,c,d) (vector unsigned char) \ 77 #define VSWIZZLE32(a,b,c,d) (vector unsigned char) \
99 #define VEC_ALIGNER(src) ((UNALIGNED_PTR(src)) \ 102 #define VEC_ALIGNER(src) ((UNALIGNED_PTR(src)) \
100 ? vec_lvsl(0, src) \ 103 ? vec_lvsl(0, src) \
101 : vec_add(vec_lvsl(8, src), vec_splat_u8(8))) 104 : vec_add(vec_lvsl(8, src), vec_splat_u8(8)))
102 105
103 /* Calculate the permute vector used for 32->32 swizzling */ 106 /* Calculate the permute vector used for 32->32 swizzling */
104 static vector unsigned char calc_swizzle32(const SDL_PixelFormat *srcfmt, 107 static vector unsigned char
105 const SDL_PixelFormat *dstfmt) 108 calc_swizzle32(const SDL_PixelFormat * srcfmt, const SDL_PixelFormat * dstfmt)
106 { 109 {
107 /* 110 /*
108 * We have to assume that the bits that aren't used by other 111 * We have to assume that the bits that aren't used by other
109 * colors is alpha, and it's one complete byte, since some formats 112 * colors is alpha, and it's one complete byte, since some formats
110 * leave alpha with a zero mask, but we should still swizzle the bits. 113 * leave alpha with a zero mask, but we should still swizzle the bits.
111 */ 114 */
112 /* ARGB */ 115 /* ARGB */
113 const static struct SDL_PixelFormat default_pixel_format = { 116 const static struct SDL_PixelFormat default_pixel_format = {
114 NULL, 0, 0, 117 NULL, 0, 0,
115 0, 0, 0, 0, 118 0, 0, 0, 0,
116 16, 8, 0, 24, 119 16, 8, 0, 24,
117 0x00FF0000, 0x0000FF00, 0x000000FF, 0xFF000000, 120 0x00FF0000, 0x0000FF00, 0x000000FF, 0xFF000000,
118 0, 0}; 121 0, 0
122 };
119 if (!srcfmt) { 123 if (!srcfmt) {
120 srcfmt = &default_pixel_format; 124 srcfmt = &default_pixel_format;
121 } 125 }
122 if (!dstfmt) { 126 if (!dstfmt) {
123 dstfmt = &default_pixel_format; 127 dstfmt = &default_pixel_format;
124 } 128 }
125 const vector unsigned char plus = VECUINT8_LITERAL( 129 const vector unsigned char plus = VECUINT8_LITERAL(0x00, 0x00, 0x00, 0x00,
126 0x00, 0x00, 0x00, 0x00, 130 0x04, 0x04, 0x04, 0x04,
127 0x04, 0x04, 0x04, 0x04, 131 0x08, 0x08, 0x08, 0x08,
128 0x08, 0x08, 0x08, 0x08, 132 0x0C, 0x0C, 0x0C,
129 0x0C, 0x0C, 0x0C, 0x0C ); 133 0x0C);
130 vector unsigned char vswiz; 134 vector unsigned char vswiz;
131 vector unsigned int srcvec; 135 vector unsigned int srcvec;
132 #define RESHIFT(X) (3 - ((X) >> 3)) 136 #define RESHIFT(X) (3 - ((X) >> 3))
133 Uint32 rmask = RESHIFT(srcfmt->Rshift) << (dstfmt->Rshift); 137 Uint32 rmask = RESHIFT(srcfmt->Rshift) << (dstfmt->Rshift);
134 Uint32 gmask = RESHIFT(srcfmt->Gshift) << (dstfmt->Gshift); 138 Uint32 gmask = RESHIFT(srcfmt->Gshift) << (dstfmt->Gshift);
135 Uint32 bmask = RESHIFT(srcfmt->Bshift) << (dstfmt->Bshift); 139 Uint32 bmask = RESHIFT(srcfmt->Bshift) << (dstfmt->Bshift);
136 Uint32 amask; 140 Uint32 amask;
137 /* Use zero for alpha if either surface doesn't have alpha */ 141 /* Use zero for alpha if either surface doesn't have alpha */
138 if (dstfmt->Amask) { 142 if (dstfmt->Amask) {
139 amask = ((srcfmt->Amask) ? RESHIFT(srcfmt->Ashift) : 0x10) << (dstfmt->Ashift); 143 amask =
140 } else { 144 ((srcfmt->Amask) ? RESHIFT(srcfmt->Ashift) : 0x10) << (dstfmt->
141 amask = 0x10101010 & ((dstfmt->Rmask | dstfmt->Gmask | dstfmt->Bmask) ^ 0xFFFFFFFF); 145 Ashift);
142 } 146 } else {
143 #undef RESHIFT 147 amask =
144 ((unsigned int *)(char*)&srcvec)[0] = (rmask | gmask | bmask | amask); 148 0x10101010 & ((dstfmt->Rmask | dstfmt->Gmask | dstfmt->Bmask) ^
145 vswiz = vec_add(plus, (vector unsigned char)vec_splat(srcvec, 0)); 149 0xFFFFFFFF);
146 return(vswiz); 150 }
147 } 151 #undef RESHIFT
148 152 ((unsigned int *) (char *) &srcvec)[0] = (rmask | gmask | bmask | amask);
149 static void Blit_RGB888_RGB565(SDL_BlitInfo *info); 153 vswiz = vec_add(plus, (vector unsigned char) vec_splat(srcvec, 0));
150 static void Blit_RGB888_RGB565Altivec(SDL_BlitInfo *info) { 154 return (vswiz);
155 }
156
157 static void Blit_RGB888_RGB565(SDL_BlitInfo * info);
158 static void
159 Blit_RGB888_RGB565Altivec(SDL_BlitInfo * info)
160 {
151 int height = info->d_height; 161 int height = info->d_height;
152 Uint8 *src = (Uint8 *) info->s_pixels; 162 Uint8 *src = (Uint8 *) info->s_pixels;
153 int srcskip = info->s_skip; 163 int srcskip = info->s_skip;
154 Uint8 *dst = (Uint8 *) info->d_pixels; 164 Uint8 *dst = (Uint8 *) info->d_pixels;
155 int dstskip = info->d_skip; 165 int dstskip = info->d_skip;
156 SDL_PixelFormat *srcfmt = info->src; 166 SDL_PixelFormat *srcfmt = info->src;
157 vector unsigned char valpha = vec_splat_u8(0); 167 vector unsigned char valpha = vec_splat_u8(0);
158 vector unsigned char vpermute = calc_swizzle32(srcfmt, NULL); 168 vector unsigned char vpermute = calc_swizzle32(srcfmt, NULL);
159 vector unsigned char vgmerge = VECUINT8_LITERAL( 169 vector unsigned char vgmerge = VECUINT8_LITERAL(0x00, 0x02, 0x00, 0x06,
160 0x00, 0x02, 0x00, 0x06, 170 0x00, 0x0a, 0x00, 0x0e,
161 0x00, 0x0a, 0x00, 0x0e, 171 0x00, 0x12, 0x00, 0x16,
162 0x00, 0x12, 0x00, 0x16, 172 0x00, 0x1a, 0x00, 0x1e);
163 0x00, 0x1a, 0x00, 0x1e);
164 vector unsigned short v1 = vec_splat_u16(1); 173 vector unsigned short v1 = vec_splat_u16(1);
165 vector unsigned short v3 = vec_splat_u16(3); 174 vector unsigned short v3 = vec_splat_u16(3);
166 vector unsigned short v3f = VECUINT16_LITERAL( 175 vector unsigned short v3f =
167 0x003f, 0x003f, 0x003f, 0x003f, 176 VECUINT16_LITERAL(0x003f, 0x003f, 0x003f, 0x003f,
168 0x003f, 0x003f, 0x003f, 0x003f); 177 0x003f, 0x003f, 0x003f, 0x003f);
169 vector unsigned short vfc = VECUINT16_LITERAL( 178 vector unsigned short vfc =
170 0x00fc, 0x00fc, 0x00fc, 0x00fc, 179 VECUINT16_LITERAL(0x00fc, 0x00fc, 0x00fc, 0x00fc,
171 0x00fc, 0x00fc, 0x00fc, 0x00fc); 180 0x00fc, 0x00fc, 0x00fc, 0x00fc);
172 vector unsigned short vf800 = (vector unsigned short)vec_splat_u8(-7); 181 vector unsigned short vf800 = (vector unsigned short) vec_splat_u8(-7);
173 vf800 = vec_sl(vf800, vec_splat_u16(8)); 182 vf800 = vec_sl(vf800, vec_splat_u16(8));
174 183
175 while (height--) { 184 while (height--) {
176 vector unsigned char valigner; 185 vector unsigned char valigner;
177 vector unsigned char voverflow; 186 vector unsigned char voverflow;
196 } 205 }
197 206
198 ONE_PIXEL_BLEND(((UNALIGNED_PTR(dst)) && (width)), width); 207 ONE_PIXEL_BLEND(((UNALIGNED_PTR(dst)) && (width)), width);
199 208
200 /* After all that work, here's the vector part! */ 209 /* After all that work, here's the vector part! */
201 extrawidth = (width % 8); /* trailing unaligned stores */ 210 extrawidth = (width % 8); /* trailing unaligned stores */
202 width -= extrawidth; 211 width -= extrawidth;
203 vsrc = vec_ld(0, src); 212 vsrc = vec_ld(0, src);
204 valigner = VEC_ALIGNER(src); 213 valigner = VEC_ALIGNER(src);
205 214
206 while (width) { 215 while (width) {
208 vector unsigned int vsrc1, vsrc2; 217 vector unsigned int vsrc1, vsrc2;
209 vector unsigned char vdst; 218 vector unsigned char vdst;
210 219
211 voverflow = vec_ld(15, src); 220 voverflow = vec_ld(15, src);
212 vsrc = vec_perm(vsrc, voverflow, valigner); 221 vsrc = vec_perm(vsrc, voverflow, valigner);
213 vsrc1 = (vector unsigned int)vec_perm(vsrc, valpha, vpermute); 222 vsrc1 = (vector unsigned int) vec_perm(vsrc, valpha, vpermute);
214 src += 16; 223 src += 16;
215 vsrc = voverflow; 224 vsrc = voverflow;
216 voverflow = vec_ld(15, src); 225 voverflow = vec_ld(15, src);
217 vsrc = vec_perm(vsrc, voverflow, valigner); 226 vsrc = vec_perm(vsrc, voverflow, valigner);
218 vsrc2 = (vector unsigned int)vec_perm(vsrc, valpha, vpermute); 227 vsrc2 = (vector unsigned int) vec_perm(vsrc, valpha, vpermute);
219 /* 1555 */ 228 /* 1555 */
220 vpixel = (vector unsigned short)vec_packpx(vsrc1, vsrc2); 229 vpixel = (vector unsigned short) vec_packpx(vsrc1, vsrc2);
221 vgpixel = (vector unsigned short)vec_perm(vsrc1, vsrc2, vgmerge); 230 vgpixel = (vector unsigned short) vec_perm(vsrc1, vsrc2, vgmerge);
222 vgpixel = vec_and(vgpixel, vfc); 231 vgpixel = vec_and(vgpixel, vfc);
223 vgpixel = vec_sl(vgpixel, v3); 232 vgpixel = vec_sl(vgpixel, v3);
224 vrpixel = vec_sl(vpixel, v1); 233 vrpixel = vec_sl(vpixel, v1);
225 vrpixel = vec_and(vrpixel, vf800); 234 vrpixel = vec_and(vrpixel, vf800);
226 vbpixel = vec_and(vpixel, v3f); 235 vbpixel = vec_and(vpixel, v3f);
227 vdst = vec_or((vector unsigned char)vrpixel, (vector unsigned char)vgpixel); 236 vdst =
237 vec_or((vector unsigned char) vrpixel,
238 (vector unsigned char) vgpixel);
228 /* 565 */ 239 /* 565 */
229 vdst = vec_or(vdst, (vector unsigned char)vbpixel); 240 vdst = vec_or(vdst, (vector unsigned char) vbpixel);
230 vec_st(vdst, 0, dst); 241 vec_st(vdst, 0, dst);
231 242
232 width -= 8; 243 width -= 8;
233 src += 16; 244 src += 16;
234 dst += 16; 245 dst += 16;
239 250
240 /* do scalar until we can align... */ 251 /* do scalar until we can align... */
241 ONE_PIXEL_BLEND((extrawidth), extrawidth); 252 ONE_PIXEL_BLEND((extrawidth), extrawidth);
242 #undef ONE_PIXEL_BLEND 253 #undef ONE_PIXEL_BLEND
243 254
244 src += srcskip; /* move to next row, accounting for pitch. */ 255 src += srcskip; /* move to next row, accounting for pitch. */
245 dst += dstskip; 256 dst += dstskip;
246 } 257 }
247 258
248 259
249 } 260 }
250 261
251 static void Blit_RGB565_32Altivec(SDL_BlitInfo *info) { 262 static void
263 Blit_RGB565_32Altivec(SDL_BlitInfo * info)
264 {
252 int height = info->d_height; 265 int height = info->d_height;
253 Uint8 *src = (Uint8 *) info->s_pixels; 266 Uint8 *src = (Uint8 *) info->s_pixels;
254 int srcskip = info->s_skip; 267 int srcskip = info->s_skip;
255 Uint8 *dst = (Uint8 *) info->d_pixels; 268 Uint8 *dst = (Uint8 *) info->d_pixels;
256 int dstskip = info->d_skip; 269 int dstskip = info->d_skip;
263 vector unsigned int v8 = vec_splat_u32(8); 276 vector unsigned int v8 = vec_splat_u32(8);
264 vector unsigned int v16 = vec_add(v8, v8); 277 vector unsigned int v16 = vec_add(v8, v8);
265 vector unsigned short v2 = vec_splat_u16(2); 278 vector unsigned short v2 = vec_splat_u16(2);
266 vector unsigned short v3 = vec_splat_u16(3); 279 vector unsigned short v3 = vec_splat_u16(3);
267 /* 280 /*
268 0x10 - 0x1f is the alpha 281 0x10 - 0x1f is the alpha
269 0x00 - 0x0e evens are the red 282 0x00 - 0x0e evens are the red
270 0x01 - 0x0f odds are zero 283 0x01 - 0x0f odds are zero
271 */ 284 */
272 vector unsigned char vredalpha1 = VECUINT8_LITERAL( 285 vector unsigned char vredalpha1 = VECUINT8_LITERAL(0x10, 0x00, 0x01, 0x01,
273 0x10, 0x00, 0x01, 0x01, 286 0x10, 0x02, 0x01, 0x01,
274 0x10, 0x02, 0x01, 0x01, 287 0x10, 0x04, 0x01, 0x01,
275 0x10, 0x04, 0x01, 0x01, 288 0x10, 0x06, 0x01,
276 0x10, 0x06, 0x01, 0x01 289 0x01);
277 ); 290 vector unsigned char vredalpha2 =
278 vector unsigned char vredalpha2 = (vector unsigned char) ( 291 (vector unsigned
279 vec_add((vector unsigned int)vredalpha1, vec_sl(v8, v16)) 292 char) (vec_add((vector unsigned int) vredalpha1, vec_sl(v8, v16))
280 ); 293 );
281 /* 294 /*
282 0x00 - 0x0f is ARxx ARxx ARxx ARxx 295 0x00 - 0x0f is ARxx ARxx ARxx ARxx
283 0x11 - 0x0f odds are blue 296 0x11 - 0x0f odds are blue
284 */ 297 */
285 vector unsigned char vblue1 = VECUINT8_LITERAL( 298 vector unsigned char vblue1 = VECUINT8_LITERAL(0x00, 0x01, 0x02, 0x11,
286 0x00, 0x01, 0x02, 0x11, 299 0x04, 0x05, 0x06, 0x13,
287 0x04, 0x05, 0x06, 0x13, 300 0x08, 0x09, 0x0a, 0x15,
288 0x08, 0x09, 0x0a, 0x15, 301 0x0c, 0x0d, 0x0e, 0x17);
289 0x0c, 0x0d, 0x0e, 0x17 302 vector unsigned char vblue2 =
290 ); 303 (vector unsigned char) (vec_add((vector unsigned int) vblue1, v8)
291 vector unsigned char vblue2 = (vector unsigned char)( 304 );
292 vec_add((vector unsigned int)vblue1, v8)
293 );
294 /* 305 /*
295 0x00 - 0x0f is ARxB ARxB ARxB ARxB 306 0x00 - 0x0f is ARxB ARxB ARxB ARxB
296 0x10 - 0x0e evens are green 307 0x10 - 0x0e evens are green
297 */ 308 */
298 vector unsigned char vgreen1 = VECUINT8_LITERAL( 309 vector unsigned char vgreen1 = VECUINT8_LITERAL(0x00, 0x01, 0x10, 0x03,
299 0x00, 0x01, 0x10, 0x03, 310 0x04, 0x05, 0x12, 0x07,
300 0x04, 0x05, 0x12, 0x07, 311 0x08, 0x09, 0x14, 0x0b,
301 0x08, 0x09, 0x14, 0x0b, 312 0x0c, 0x0d, 0x16, 0x0f);
302 0x0c, 0x0d, 0x16, 0x0f 313 vector unsigned char vgreen2 =
303 ); 314 (vector unsigned
304 vector unsigned char vgreen2 = (vector unsigned char)( 315 char) (vec_add((vector unsigned int) vgreen1, vec_sl(v8, v8))
305 vec_add((vector unsigned int)vgreen1, vec_sl(v8, v8)) 316 );
306 ); 317
307
308 318
309 assert(srcfmt->BytesPerPixel == 2); 319 assert(srcfmt->BytesPerPixel == 2);
310 assert(dstfmt->BytesPerPixel == 4); 320 assert(dstfmt->BytesPerPixel == 4);
311 321
312 vf800 = (vector unsigned short)vec_splat_u8(-7); 322 vf800 = (vector unsigned short) vec_splat_u8(-7);
313 vf800 = vec_sl(vf800, vec_splat_u16(8)); 323 vf800 = vec_sl(vf800, vec_splat_u16(8));
314 324
315 if (dstfmt->Amask && srcfmt->alpha) { 325 if (dstfmt->Amask && srcfmt->alpha) {
316 ((unsigned char *)&valpha)[0] = alpha = srcfmt->alpha; 326 ((unsigned char *) &valpha)[0] = alpha = srcfmt->alpha;
317 valpha = vec_splat(valpha, 0); 327 valpha = vec_splat(valpha, 0);
318 } else { 328 } else {
319 alpha = 0; 329 alpha = 0;
320 valpha = vec_splat_u8(0); 330 valpha = vec_splat_u8(0);
321 } 331 }
343 widthvar--; \ 353 widthvar--; \
344 } 354 }
345 ONE_PIXEL_BLEND(((UNALIGNED_PTR(dst)) && (width)), width); 355 ONE_PIXEL_BLEND(((UNALIGNED_PTR(dst)) && (width)), width);
346 356
347 /* After all that work, here's the vector part! */ 357 /* After all that work, here's the vector part! */
348 extrawidth = (width % 8); /* trailing unaligned stores */ 358 extrawidth = (width % 8); /* trailing unaligned stores */
349 width -= extrawidth; 359 width -= extrawidth;
350 vsrc = vec_ld(0, src); 360 vsrc = vec_ld(0, src);
351 valigner = VEC_ALIGNER(src); 361 valigner = VEC_ALIGNER(src);
352 362
353 while (width) { 363 while (width) {
355 vector unsigned char vdst1, vdst2; 365 vector unsigned char vdst1, vdst2;
356 366
357 voverflow = vec_ld(15, src); 367 voverflow = vec_ld(15, src);
358 vsrc = vec_perm(vsrc, voverflow, valigner); 368 vsrc = vec_perm(vsrc, voverflow, valigner);
359 369
360 vR = vec_and((vector unsigned short)vsrc, vf800); 370 vR = vec_and((vector unsigned short) vsrc, vf800);
361 vB = vec_sl((vector unsigned short)vsrc, v3); 371 vB = vec_sl((vector unsigned short) vsrc, v3);
362 vG = vec_sl(vB, v2); 372 vG = vec_sl(vB, v2);
363 373
364 vdst1 = (vector unsigned char)vec_perm((vector unsigned char)vR, valpha, vredalpha1); 374 vdst1 =
365 vdst1 = vec_perm(vdst1, (vector unsigned char)vB, vblue1); 375 (vector unsigned char) vec_perm((vector unsigned char) vR,
366 vdst1 = vec_perm(vdst1, (vector unsigned char)vG, vgreen1); 376 valpha, vredalpha1);
377 vdst1 = vec_perm(vdst1, (vector unsigned char) vB, vblue1);
378 vdst1 = vec_perm(vdst1, (vector unsigned char) vG, vgreen1);
367 vdst1 = vec_perm(vdst1, valpha, vpermute); 379 vdst1 = vec_perm(vdst1, valpha, vpermute);
368 vec_st(vdst1, 0, dst); 380 vec_st(vdst1, 0, dst);
369 381
370 vdst2 = (vector unsigned char)vec_perm((vector unsigned char)vR, valpha, vredalpha2); 382 vdst2 =
371 vdst2 = vec_perm(vdst2, (vector unsigned char)vB, vblue2); 383 (vector unsigned char) vec_perm((vector unsigned char) vR,
372 vdst2 = vec_perm(vdst2, (vector unsigned char)vG, vgreen2); 384 valpha, vredalpha2);
385 vdst2 = vec_perm(vdst2, (vector unsigned char) vB, vblue2);
386 vdst2 = vec_perm(vdst2, (vector unsigned char) vG, vgreen2);
373 vdst2 = vec_perm(vdst2, valpha, vpermute); 387 vdst2 = vec_perm(vdst2, valpha, vpermute);
374 vec_st(vdst2, 16, dst); 388 vec_st(vdst2, 16, dst);
375 389
376 width -= 8; 390 width -= 8;
377 dst += 32; 391 dst += 32;
378 src += 16; 392 src += 16;
379 vsrc = voverflow; 393 vsrc = voverflow;
380 } 394 }
384 398
385 /* do scalar until we can align... */ 399 /* do scalar until we can align... */
386 ONE_PIXEL_BLEND((extrawidth), extrawidth); 400 ONE_PIXEL_BLEND((extrawidth), extrawidth);
387 #undef ONE_PIXEL_BLEND 401 #undef ONE_PIXEL_BLEND
388 402
389 src += srcskip; /* move to next row, accounting for pitch. */ 403 src += srcskip; /* move to next row, accounting for pitch. */
390 dst += dstskip; 404 dst += dstskip;
391 } 405 }
392 406
393 } 407 }
394 408
395 409
396 static void Blit_RGB555_32Altivec(SDL_BlitInfo *info) { 410 static void
411 Blit_RGB555_32Altivec(SDL_BlitInfo * info)
412 {
397 int height = info->d_height; 413 int height = info->d_height;
398 Uint8 *src = (Uint8 *) info->s_pixels; 414 Uint8 *src = (Uint8 *) info->s_pixels;
399 int srcskip = info->s_skip; 415 int srcskip = info->s_skip;
400 Uint8 *dst = (Uint8 *) info->d_pixels; 416 Uint8 *dst = (Uint8 *) info->d_pixels;
401 int dstskip = info->d_skip; 417 int dstskip = info->d_skip;
408 vector unsigned int v8 = vec_splat_u32(8); 424 vector unsigned int v8 = vec_splat_u32(8);
409 vector unsigned int v16 = vec_add(v8, v8); 425 vector unsigned int v16 = vec_add(v8, v8);
410 vector unsigned short v1 = vec_splat_u16(1); 426 vector unsigned short v1 = vec_splat_u16(1);
411 vector unsigned short v3 = vec_splat_u16(3); 427 vector unsigned short v3 = vec_splat_u16(3);
412 /* 428 /*
413 0x10 - 0x1f is the alpha 429 0x10 - 0x1f is the alpha
414 0x00 - 0x0e evens are the red 430 0x00 - 0x0e evens are the red
415 0x01 - 0x0f odds are zero 431 0x01 - 0x0f odds are zero
416 */ 432 */
417 vector unsigned char vredalpha1 = VECUINT8_LITERAL( 433 vector unsigned char vredalpha1 = VECUINT8_LITERAL(0x10, 0x00, 0x01, 0x01,
418 0x10, 0x00, 0x01, 0x01, 434 0x10, 0x02, 0x01, 0x01,
419 0x10, 0x02, 0x01, 0x01, 435 0x10, 0x04, 0x01, 0x01,
420 0x10, 0x04, 0x01, 0x01, 436 0x10, 0x06, 0x01,
421 0x10, 0x06, 0x01, 0x01 437 0x01);
422 ); 438 vector unsigned char vredalpha2 =
423 vector unsigned char vredalpha2 = (vector unsigned char)( 439 (vector unsigned
424 vec_add((vector unsigned int)vredalpha1, vec_sl(v8, v16)) 440 char) (vec_add((vector unsigned int) vredalpha1, vec_sl(v8, v16))
425 ); 441 );
426 /* 442 /*
427 0x00 - 0x0f is ARxx ARxx ARxx ARxx 443 0x00 - 0x0f is ARxx ARxx ARxx ARxx
428 0x11 - 0x0f odds are blue 444 0x11 - 0x0f odds are blue
429 */ 445 */
430 vector unsigned char vblue1 = VECUINT8_LITERAL( 446 vector unsigned char vblue1 = VECUINT8_LITERAL(0x00, 0x01, 0x02, 0x11,
431 0x00, 0x01, 0x02, 0x11, 447 0x04, 0x05, 0x06, 0x13,
432 0x04, 0x05, 0x06, 0x13, 448 0x08, 0x09, 0x0a, 0x15,
433 0x08, 0x09, 0x0a, 0x15, 449 0x0c, 0x0d, 0x0e, 0x17);
434 0x0c, 0x0d, 0x0e, 0x17 450 vector unsigned char vblue2 =
435 ); 451 (vector unsigned char) (vec_add((vector unsigned int) vblue1, v8)
436 vector unsigned char vblue2 = (vector unsigned char)( 452 );
437 vec_add((vector unsigned int)vblue1, v8)
438 );
439 /* 453 /*
440 0x00 - 0x0f is ARxB ARxB ARxB ARxB 454 0x00 - 0x0f is ARxB ARxB ARxB ARxB
441 0x10 - 0x0e evens are green 455 0x10 - 0x0e evens are green
442 */ 456 */
443 vector unsigned char vgreen1 = VECUINT8_LITERAL( 457 vector unsigned char vgreen1 = VECUINT8_LITERAL(0x00, 0x01, 0x10, 0x03,
444 0x00, 0x01, 0x10, 0x03, 458 0x04, 0x05, 0x12, 0x07,
445 0x04, 0x05, 0x12, 0x07, 459 0x08, 0x09, 0x14, 0x0b,
446 0x08, 0x09, 0x14, 0x0b, 460 0x0c, 0x0d, 0x16, 0x0f);
447 0x0c, 0x0d, 0x16, 0x0f 461 vector unsigned char vgreen2 =
448 ); 462 (vector unsigned
449 vector unsigned char vgreen2 = (vector unsigned char)( 463 char) (vec_add((vector unsigned int) vgreen1, vec_sl(v8, v8))
450 vec_add((vector unsigned int)vgreen1, vec_sl(v8, v8)) 464 );
451 ); 465
452
453 466
454 assert(srcfmt->BytesPerPixel == 2); 467 assert(srcfmt->BytesPerPixel == 2);
455 assert(dstfmt->BytesPerPixel == 4); 468 assert(dstfmt->BytesPerPixel == 4);
456 469
457 vf800 = (vector unsigned short)vec_splat_u8(-7); 470 vf800 = (vector unsigned short) vec_splat_u8(-7);
458 vf800 = vec_sl(vf800, vec_splat_u16(8)); 471 vf800 = vec_sl(vf800, vec_splat_u16(8));
459 472
460 if (dstfmt->Amask && srcfmt->alpha) { 473 if (dstfmt->Amask && srcfmt->alpha) {
461 ((unsigned char *)&valpha)[0] = alpha = srcfmt->alpha; 474 ((unsigned char *) &valpha)[0] = alpha = srcfmt->alpha;
462 valpha = vec_splat(valpha, 0); 475 valpha = vec_splat(valpha, 0);
463 } else { 476 } else {
464 alpha = 0; 477 alpha = 0;
465 valpha = vec_splat_u8(0); 478 valpha = vec_splat_u8(0);
466 } 479 }
488 widthvar--; \ 501 widthvar--; \
489 } 502 }
490 ONE_PIXEL_BLEND(((UNALIGNED_PTR(dst)) && (width)), width); 503 ONE_PIXEL_BLEND(((UNALIGNED_PTR(dst)) && (width)), width);
491 504
492 /* After all that work, here's the vector part! */ 505 /* After all that work, here's the vector part! */
493 extrawidth = (width % 8); /* trailing unaligned stores */ 506 extrawidth = (width % 8); /* trailing unaligned stores */
494 width -= extrawidth; 507 width -= extrawidth;
495 vsrc = vec_ld(0, src); 508 vsrc = vec_ld(0, src);
496 valigner = VEC_ALIGNER(src); 509 valigner = VEC_ALIGNER(src);
497 510
498 while (width) { 511 while (width) {
500 vector unsigned char vdst1, vdst2; 513 vector unsigned char vdst1, vdst2;
501 514
502 voverflow = vec_ld(15, src); 515 voverflow = vec_ld(15, src);
503 vsrc = vec_perm(vsrc, voverflow, valigner); 516 vsrc = vec_perm(vsrc, voverflow, valigner);
504 517
505 vR = vec_and(vec_sl((vector unsigned short)vsrc,v1), vf800); 518 vR = vec_and(vec_sl((vector unsigned short) vsrc, v1), vf800);
506 vB = vec_sl((vector unsigned short)vsrc, v3); 519 vB = vec_sl((vector unsigned short) vsrc, v3);
507 vG = vec_sl(vB, v3); 520 vG = vec_sl(vB, v3);
508 521
509 vdst1 = (vector unsigned char)vec_perm((vector unsigned char)vR, valpha, vredalpha1); 522 vdst1 =
510 vdst1 = vec_perm(vdst1, (vector unsigned char)vB, vblue1); 523 (vector unsigned char) vec_perm((vector unsigned char) vR,
511 vdst1 = vec_perm(vdst1, (vector unsigned char)vG, vgreen1); 524 valpha, vredalpha1);
525 vdst1 = vec_perm(vdst1, (vector unsigned char) vB, vblue1);
526 vdst1 = vec_perm(vdst1, (vector unsigned char) vG, vgreen1);
512 vdst1 = vec_perm(vdst1, valpha, vpermute); 527 vdst1 = vec_perm(vdst1, valpha, vpermute);
513 vec_st(vdst1, 0, dst); 528 vec_st(vdst1, 0, dst);
514 529
515 vdst2 = (vector unsigned char)vec_perm((vector unsigned char)vR, valpha, vredalpha2); 530 vdst2 =
516 vdst2 = vec_perm(vdst2, (vector unsigned char)vB, vblue2); 531 (vector unsigned char) vec_perm((vector unsigned char) vR,
517 vdst2 = vec_perm(vdst2, (vector unsigned char)vG, vgreen2); 532 valpha, vredalpha2);
533 vdst2 = vec_perm(vdst2, (vector unsigned char) vB, vblue2);
534 vdst2 = vec_perm(vdst2, (vector unsigned char) vG, vgreen2);
518 vdst2 = vec_perm(vdst2, valpha, vpermute); 535 vdst2 = vec_perm(vdst2, valpha, vpermute);
519 vec_st(vdst2, 16, dst); 536 vec_st(vdst2, 16, dst);
520 537
521 width -= 8; 538 width -= 8;
522 dst += 32; 539 dst += 32;
523 src += 16; 540 src += 16;
524 vsrc = voverflow; 541 vsrc = voverflow;
525 } 542 }
529 546
530 /* do scalar until we can align... */ 547 /* do scalar until we can align... */
531 ONE_PIXEL_BLEND((extrawidth), extrawidth); 548 ONE_PIXEL_BLEND((extrawidth), extrawidth);
532 #undef ONE_PIXEL_BLEND 549 #undef ONE_PIXEL_BLEND
533 550
534 src += srcskip; /* move to next row, accounting for pitch. */ 551 src += srcskip; /* move to next row, accounting for pitch. */
535 dst += dstskip; 552 dst += dstskip;
536 } 553 }
537 554
538 } 555 }
539 556
540 static void BlitNtoNKey(SDL_BlitInfo *info); 557 static void BlitNtoNKey(SDL_BlitInfo * info);
541 static void BlitNtoNKeyCopyAlpha(SDL_BlitInfo *info); 558 static void BlitNtoNKeyCopyAlpha(SDL_BlitInfo * info);
542 static void Blit32to32KeyAltivec(SDL_BlitInfo *info) 559 static void
560 Blit32to32KeyAltivec(SDL_BlitInfo * info)
543 { 561 {
544 int height = info->d_height; 562 int height = info->d_height;
545 Uint32 *srcp = (Uint32 *) info->s_pixels; 563 Uint32 *srcp = (Uint32 *) info->s_pixels;
546 int srcskip = info->s_skip; 564 int srcskip = info->s_skip;
547 Uint32 *dstp = (Uint32 *) info->d_pixels; 565 Uint32 *dstp = (Uint32 *) info->d_pixels;
549 SDL_PixelFormat *srcfmt = info->src; 567 SDL_PixelFormat *srcfmt = info->src;
550 int srcbpp = srcfmt->BytesPerPixel; 568 int srcbpp = srcfmt->BytesPerPixel;
551 SDL_PixelFormat *dstfmt = info->dst; 569 SDL_PixelFormat *dstfmt = info->dst;
552 int dstbpp = dstfmt->BytesPerPixel; 570 int dstbpp = dstfmt->BytesPerPixel;
553 int copy_alpha = (srcfmt->Amask && dstfmt->Amask); 571 int copy_alpha = (srcfmt->Amask && dstfmt->Amask);
554 unsigned alpha = dstfmt->Amask ? srcfmt->alpha : 0; 572 unsigned alpha = dstfmt->Amask ? srcfmt->alpha : 0;
555 Uint32 rgbmask = srcfmt->Rmask | srcfmt->Gmask | srcfmt->Bmask; 573 Uint32 rgbmask = srcfmt->Rmask | srcfmt->Gmask | srcfmt->Bmask;
556 Uint32 ckey = info->src->colorkey; 574 Uint32 ckey = info->src->colorkey;
557 vector unsigned int valpha; 575 vector unsigned int valpha;
558 vector unsigned char vpermute; 576 vector unsigned char vpermute;
559 vector unsigned char vzero; 577 vector unsigned char vzero;
560 vector unsigned int vckey; 578 vector unsigned int vckey;
561 vector unsigned int vrgbmask; 579 vector unsigned int vrgbmask;
562 vpermute = calc_swizzle32(srcfmt, dstfmt); 580 vpermute = calc_swizzle32(srcfmt, dstfmt);
563 if (info->d_width < 16) { 581 if (info->d_width < 16) {
564 if(copy_alpha) { 582 if (copy_alpha) {
565 BlitNtoNKeyCopyAlpha(info); 583 BlitNtoNKeyCopyAlpha(info);
566 } else { 584 } else {
567 BlitNtoNKey(info); 585 BlitNtoNKey(info);
568 } 586 }
569 return; 587 return;
570 } 588 }
571 vzero = vec_splat_u8(0); 589 vzero = vec_splat_u8(0);
572 if (alpha) { 590 if (alpha) {
573 ((unsigned char *)&valpha)[0] = (unsigned char)alpha; 591 ((unsigned char *) &valpha)[0] = (unsigned char) alpha;
574 valpha = (vector unsigned int)vec_splat((vector unsigned char)valpha, 0); 592 valpha =
593 (vector unsigned int) vec_splat((vector unsigned char) valpha, 0);
575 } else { 594 } else {
576 valpha = (vector unsigned int)vzero; 595 valpha = (vector unsigned int) vzero;
577 } 596 }
578 ckey &= rgbmask; 597 ckey &= rgbmask;
579 ((unsigned int *)(char*)&vckey)[0] = ckey; 598 ((unsigned int *) (char *) &vckey)[0] = ckey;
580 vckey = vec_splat(vckey, 0); 599 vckey = vec_splat(vckey, 0);
581 ((unsigned int *)(char*)&vrgbmask)[0] = rgbmask; 600 ((unsigned int *) (char *) &vrgbmask)[0] = rgbmask;
582 vrgbmask = vec_splat(vrgbmask, 0); 601 vrgbmask = vec_splat(vrgbmask, 0);
583 602
584 while (height--) { 603 while (height--) {
585 #define ONE_PIXEL_BLEND(condition, widthvar) \ 604 #define ONE_PIXEL_BLEND(condition, widthvar) \
586 if (copy_alpha) { \ 605 if (copy_alpha) { \
626 vector unsigned int vd; 645 vector unsigned int vd;
627 vector unsigned int voverflow = vec_ld(15, srcp); 646 vector unsigned int voverflow = vec_ld(15, srcp);
628 /* load the source vec */ 647 /* load the source vec */
629 vs = vec_perm(vs, voverflow, valigner); 648 vs = vec_perm(vs, voverflow, valigner);
630 /* vsel is set for items that match the key */ 649 /* vsel is set for items that match the key */
631 vsel = (vector unsigned char)vec_and(vs, vrgbmask); 650 vsel = (vector unsigned char) vec_and(vs, vrgbmask);
632 vsel = (vector unsigned char)vec_cmpeq(vs, vckey); 651 vsel = (vector unsigned char) vec_cmpeq(vs, vckey);
633 /* permute the src vec to the dest format */ 652 /* permute the src vec to the dest format */
634 vs = vec_perm(vs, valpha, vpermute); 653 vs = vec_perm(vs, valpha, vpermute);
635 /* load the destination vec */ 654 /* load the destination vec */
636 vd = vec_ld(0, dstp); 655 vd = vec_ld(0, dstp);
637 /* select the source and dest into vs */ 656 /* select the source and dest into vs */
638 vd = (vector unsigned int)vec_sel((vector unsigned char)vs, (vector unsigned char)vd, vsel); 657 vd = (vector unsigned int) vec_sel((vector unsigned char) vs,
639 658 (vector unsigned char) vd,
659 vsel);
660
640 vec_st(vd, 0, dstp); 661 vec_st(vd, 0, dstp);
641 srcp += 4; 662 srcp += 4;
642 width -= 4; 663 width -= 4;
643 dstp += 4; 664 dstp += 4;
644 vs = voverflow; 665 vs = voverflow;
651 } 672 }
652 } 673 }
653 674
654 /* Altivec code to swizzle one 32-bit surface to a different 32-bit format. */ 675 /* Altivec code to swizzle one 32-bit surface to a different 32-bit format. */
655 /* Use this on a G5 */ 676 /* Use this on a G5 */
656 static void ConvertAltivec32to32_noprefetch(SDL_BlitInfo *info) 677 static void
678 ConvertAltivec32to32_noprefetch(SDL_BlitInfo * info)
657 { 679 {
658 int height = info->d_height; 680 int height = info->d_height;
659 Uint32 *src = (Uint32 *) info->s_pixels; 681 Uint32 *src = (Uint32 *) info->s_pixels;
660 int srcskip = info->s_skip; 682 int srcskip = info->s_skip;
661 Uint32 *dst = (Uint32 *) info->d_pixels; 683 Uint32 *dst = (Uint32 *) info->d_pixels;
665 vector unsigned int vzero = vec_splat_u32(0); 687 vector unsigned int vzero = vec_splat_u32(0);
666 vector unsigned char vpermute = calc_swizzle32(srcfmt, dstfmt); 688 vector unsigned char vpermute = calc_swizzle32(srcfmt, dstfmt);
667 if (dstfmt->Amask && !srcfmt->Amask) { 689 if (dstfmt->Amask && !srcfmt->Amask) {
668 if (srcfmt->alpha) { 690 if (srcfmt->alpha) {
669 vector unsigned char valpha; 691 vector unsigned char valpha;
670 ((unsigned char *)&valpha)[0] = srcfmt->alpha; 692 ((unsigned char *) &valpha)[0] = srcfmt->alpha;
671 vzero = (vector unsigned int)vec_splat(valpha, 0); 693 vzero = (vector unsigned int) vec_splat(valpha, 0);
672 } 694 }
673 } 695 }
674 696
675 assert(srcfmt->BytesPerPixel == 4); 697 assert(srcfmt->BytesPerPixel == 4);
676 assert(dstfmt->BytesPerPixel == 4); 698 assert(dstfmt->BytesPerPixel == 4);
697 extrawidth = (width % 4); 719 extrawidth = (width % 4);
698 width -= extrawidth; 720 width -= extrawidth;
699 valigner = VEC_ALIGNER(src); 721 valigner = VEC_ALIGNER(src);
700 vbits = vec_ld(0, src); 722 vbits = vec_ld(0, src);
701 723
702 while (width) { 724 while (width) {
703 voverflow = vec_ld(15, src); 725 voverflow = vec_ld(15, src);
704 src += 4; 726 src += 4;
705 width -= 4; 727 width -= 4;
706 vbits = vec_perm(vbits, voverflow, valigner); /* src is ready. */ 728 vbits = vec_perm(vbits, voverflow, valigner); /* src is ready. */
707 vbits = vec_perm(vbits, vzero, vpermute); /* swizzle it. */ 729 vbits = vec_perm(vbits, vzero, vpermute); /* swizzle it. */
708 vec_st(vbits, 0, dst); /* store it back out. */ 730 vec_st(vbits, 0, dst); /* store it back out. */
709 dst += 4; 731 dst += 4;
710 vbits = voverflow; 732 vbits = voverflow;
711 } 733 }
712 734
713 assert(width == 0); 735 assert(width == 0);
714 736
715 /* cover pixels at the end of the row that didn't fit in 16 bytes. */ 737 /* cover pixels at the end of the row that didn't fit in 16 bytes. */
716 while (extrawidth) { 738 while (extrawidth) {
717 bits = *(src++); /* max 7 pixels, don't bother with prefetch. */ 739 bits = *(src++); /* max 7 pixels, don't bother with prefetch. */
718 RGBA_FROM_8888(bits, srcfmt, r, g, b, a); 740 RGBA_FROM_8888(bits, srcfmt, r, g, b, a);
719 *(dst++) = MAKE8888(dstfmt, r, g, b, a); 741 *(dst++) = MAKE8888(dstfmt, r, g, b, a);
720 extrawidth--; 742 extrawidth--;
721 } 743 }
722 744
723 src += srcskip >> 2; /* move to next row, accounting for pitch. */ 745 src += srcskip >> 2; /* move to next row, accounting for pitch. */
724 dst += dstskip >> 2; 746 dst += dstskip >> 2;
725 } 747 }
726 748
727 } 749 }
728 750
729 /* Altivec code to swizzle one 32-bit surface to a different 32-bit format. */ 751 /* Altivec code to swizzle one 32-bit surface to a different 32-bit format. */
730 /* Use this on a G4 */ 752 /* Use this on a G4 */
731 static void ConvertAltivec32to32_prefetch(SDL_BlitInfo *info) 753 static void
732 { 754 ConvertAltivec32to32_prefetch(SDL_BlitInfo * info)
733 const int scalar_dst_lead = sizeof (Uint32) * 4; 755 {
734 const int vector_dst_lead = sizeof (Uint32) * 16; 756 const int scalar_dst_lead = sizeof(Uint32) * 4;
757 const int vector_dst_lead = sizeof(Uint32) * 16;
735 758
736 int height = info->d_height; 759 int height = info->d_height;
737 Uint32 *src = (Uint32 *) info->s_pixels; 760 Uint32 *src = (Uint32 *) info->s_pixels;
738 int srcskip = info->s_skip; 761 int srcskip = info->s_skip;
739 Uint32 *dst = (Uint32 *) info->d_pixels; 762 Uint32 *dst = (Uint32 *) info->d_pixels;
743 vector unsigned int vzero = vec_splat_u32(0); 766 vector unsigned int vzero = vec_splat_u32(0);
744 vector unsigned char vpermute = calc_swizzle32(srcfmt, dstfmt); 767 vector unsigned char vpermute = calc_swizzle32(srcfmt, dstfmt);
745 if (dstfmt->Amask && !srcfmt->Amask) { 768 if (dstfmt->Amask && !srcfmt->Amask) {
746 if (srcfmt->alpha) { 769 if (srcfmt->alpha) {
747 vector unsigned char valpha; 770 vector unsigned char valpha;
748 ((unsigned char *)&valpha)[0] = srcfmt->alpha; 771 ((unsigned char *) &valpha)[0] = srcfmt->alpha;
749 vzero = (vector unsigned int)vec_splat(valpha, 0); 772 vzero = (vector unsigned int) vec_splat(valpha, 0);
750 } 773 }
751 } 774 }
752 775
753 assert(srcfmt->BytesPerPixel == 4); 776 assert(srcfmt->BytesPerPixel == 4);
754 assert(dstfmt->BytesPerPixel == 4); 777 assert(dstfmt->BytesPerPixel == 4);
763 int width = info->d_width; 786 int width = info->d_width;
764 int extrawidth; 787 int extrawidth;
765 788
766 /* do scalar until we can align... */ 789 /* do scalar until we can align... */
767 while ((UNALIGNED_PTR(dst)) && (width)) { 790 while ((UNALIGNED_PTR(dst)) && (width)) {
768 vec_dstt(src+scalar_dst_lead, DST_CTRL(2,32,1024), DST_CHAN_SRC); 791 vec_dstt(src + scalar_dst_lead, DST_CTRL(2, 32, 1024),
769 vec_dstst(dst+scalar_dst_lead, DST_CTRL(2,32,1024), DST_CHAN_DEST); 792 DST_CHAN_SRC);
793 vec_dstst(dst + scalar_dst_lead, DST_CTRL(2, 32, 1024),
794 DST_CHAN_DEST);
770 bits = *(src++); 795 bits = *(src++);
771 RGBA_FROM_8888(bits, srcfmt, r, g, b, a); 796 RGBA_FROM_8888(bits, srcfmt, r, g, b, a);
772 *(dst++) = MAKE8888(dstfmt, r, g, b, a); 797 *(dst++) = MAKE8888(dstfmt, r, g, b, a);
773 width--; 798 width--;
774 } 799 }
778 width -= extrawidth; 803 width -= extrawidth;
779 valigner = VEC_ALIGNER(src); 804 valigner = VEC_ALIGNER(src);
780 vbits = vec_ld(0, src); 805 vbits = vec_ld(0, src);
781 806
782 while (width) { 807 while (width) {
783 vec_dstt(src+vector_dst_lead, DST_CTRL(2,32,1024), DST_CHAN_SRC); 808 vec_dstt(src + vector_dst_lead, DST_CTRL(2, 32, 1024),
784 vec_dstst(dst+vector_dst_lead, DST_CTRL(2,32,1024), DST_CHAN_DEST); 809 DST_CHAN_SRC);
810 vec_dstst(dst + vector_dst_lead, DST_CTRL(2, 32, 1024),
811 DST_CHAN_DEST);
785 voverflow = vec_ld(15, src); 812 voverflow = vec_ld(15, src);
786 src += 4; 813 src += 4;
787 width -= 4; 814 width -= 4;
788 vbits = vec_perm(vbits, voverflow, valigner); /* src is ready. */ 815 vbits = vec_perm(vbits, voverflow, valigner); /* src is ready. */
789 vbits = vec_perm(vbits, vzero, vpermute); /* swizzle it. */ 816 vbits = vec_perm(vbits, vzero, vpermute); /* swizzle it. */
790 vec_st(vbits, 0, dst); /* store it back out. */ 817 vec_st(vbits, 0, dst); /* store it back out. */
791 dst += 4; 818 dst += 4;
792 vbits = voverflow; 819 vbits = voverflow;
793 } 820 }
794 821
795 assert(width == 0); 822 assert(width == 0);
796 823
797 /* cover pixels at the end of the row that didn't fit in 16 bytes. */ 824 /* cover pixels at the end of the row that didn't fit in 16 bytes. */
798 while (extrawidth) { 825 while (extrawidth) {
799 bits = *(src++); /* max 7 pixels, don't bother with prefetch. */ 826 bits = *(src++); /* max 7 pixels, don't bother with prefetch. */
800 RGBA_FROM_8888(bits, srcfmt, r, g, b, a); 827 RGBA_FROM_8888(bits, srcfmt, r, g, b, a);
801 *(dst++) = MAKE8888(dstfmt, r, g, b, a); 828 *(dst++) = MAKE8888(dstfmt, r, g, b, a);
802 extrawidth--; 829 extrawidth--;
803 } 830 }
804 831
805 src += srcskip >> 2; /* move to next row, accounting for pitch. */ 832 src += srcskip >> 2; /* move to next row, accounting for pitch. */
806 dst += dstskip >> 2; 833 dst += dstskip >> 2;
807 } 834 }
808 835
809 vec_dss(DST_CHAN_SRC); 836 vec_dss(DST_CHAN_SRC);
810 vec_dss(DST_CHAN_DEST); 837 vec_dss(DST_CHAN_DEST);
811 } 838 }
812 839
813 static Uint32 GetBlitFeatures( void ) 840 static Uint32
841 GetBlitFeatures(void)
814 { 842 {
815 static Uint32 features = 0xffffffff; 843 static Uint32 features = 0xffffffff;
816 if (features == 0xffffffff) { 844 if (features == 0xffffffff) {
817 /* Provide an override for testing .. */ 845 /* Provide an override for testing .. */
818 char *override = SDL_getenv("SDL_ALTIVEC_BLIT_FEATURES"); 846 char *override = SDL_getenv("SDL_ALTIVEC_BLIT_FEATURES");
819 if (override) { 847 if (override) {
820 features = 0; 848 features = 0;
821 SDL_sscanf(override, "%u", &features); 849 SDL_sscanf(override, "%u", &features);
822 } else { 850 } else {
823 features = ( 0 851 features = (0
824 /* Feature 1 is has-MMX */ 852 /* Feature 1 is has-MMX */
825 | ((SDL_HasMMX()) ? 1 : 0) 853 | ((SDL_HasMMX())? 1 : 0)
826 /* Feature 2 is has-AltiVec */ 854 /* Feature 2 is has-AltiVec */
827 | ((SDL_HasAltiVec()) ? 2 : 0) 855 | ((SDL_HasAltiVec())? 2 : 0)
828 /* Feature 4 is dont-use-prefetch */ 856 /* Feature 4 is dont-use-prefetch */
829 /* !!!! FIXME: Check for G5 or later, not the cache size! Always prefetch on a G4. */ 857 /* !!!! FIXME: Check for G5 or later, not the cache size! Always prefetch on a G4. */
830 | ((GetL3CacheSize() == 0) ? 4 : 0) 858 | ((GetL3CacheSize() == 0) ? 4 : 0)
831 ); 859 );
832 } 860 }
833 } 861 }
834 return features; 862 return features;
835 } 863 }
864
836 #if __MWERKS__ 865 #if __MWERKS__
837 #pragma altivec_model off 866 #pragma altivec_model off
838 #endif 867 #endif
839 #else 868 #else
840 /* Feature 1 is has-MMX */ 869 /* Feature 1 is has-MMX */
867 #define RGB888_RGB332(dst, src) { \ 896 #define RGB888_RGB332(dst, src) { \
868 dst = (Uint8)((((src)&0x00E00000)>>16)| \ 897 dst = (Uint8)((((src)&0x00E00000)>>16)| \
869 (((src)&0x0000E000)>>11)| \ 898 (((src)&0x0000E000)>>11)| \
870 (((src)&0x000000C0)>>6)); \ 899 (((src)&0x000000C0)>>6)); \
871 } 900 }
872 static void Blit_RGB888_index8(SDL_BlitInfo *info) 901 static void
902 Blit_RGB888_index8(SDL_BlitInfo * info)
873 { 903 {
874 #ifndef USE_DUFFS_LOOP 904 #ifndef USE_DUFFS_LOOP
875 int c; 905 int c;
876 #endif 906 #endif
877 int width, height; 907 int width, height;
878 Uint32 *src; 908 Uint32 *src;
879 const Uint8 *map; 909 const Uint8 *map;
880 Uint8 *dst; 910 Uint8 *dst;
881 int srcskip, dstskip; 911 int srcskip, dstskip;
882 912
883 /* Set up some basic variables */ 913 /* Set up some basic variables */
884 width = info->d_width; 914 width = info->d_width;
885 height = info->d_height; 915 height = info->d_height;
886 src = (Uint32 *)info->s_pixels; 916 src = (Uint32 *) info->s_pixels;
887 srcskip = info->s_skip/4; 917 srcskip = info->s_skip / 4;
888 dst = info->d_pixels; 918 dst = info->d_pixels;
889 dstskip = info->d_skip; 919 dstskip = info->d_skip;
890 map = info->table; 920 map = info->table;
891 921
892 if ( map == NULL ) { 922 if (map == NULL) {
893 while ( height-- ) { 923 while (height--) {
894 #ifdef USE_DUFFS_LOOP 924 #ifdef USE_DUFFS_LOOP
925 /* *INDENT-OFF* */
895 DUFFS_LOOP( 926 DUFFS_LOOP(
896 RGB888_RGB332(*dst++, *src); 927 RGB888_RGB332(*dst++, *src);
897 , width); 928 , width);
929 /* *INDENT-ON* */
898 #else 930 #else
899 for ( c=width/4; c; --c ) { 931 for (c = width / 4; c; --c) {
900 /* Pack RGB into 8bit pixel */ 932 /* Pack RGB into 8bit pixel */
901 ++src; 933 ++src;
902 RGB888_RGB332(*dst++, *src); 934 RGB888_RGB332(*dst++, *src);
903 ++src; 935 ++src;
904 RGB888_RGB332(*dst++, *src); 936 RGB888_RGB332(*dst++, *src);
905 ++src; 937 ++src;
906 RGB888_RGB332(*dst++, *src); 938 RGB888_RGB332(*dst++, *src);
907 ++src; 939 ++src;
908 } 940 }
909 switch ( width & 3 ) { 941 switch (width & 3) {
910 case 3: 942 case 3:
911 RGB888_RGB332(*dst++, *src); 943 RGB888_RGB332(*dst++, *src);
912 ++src; 944 ++src;
913 case 2: 945 case 2:
914 RGB888_RGB332(*dst++, *src); 946 RGB888_RGB332(*dst++, *src);
915 ++src; 947 ++src;
916 case 1: 948 case 1:
917 RGB888_RGB332(*dst++, *src); 949 RGB888_RGB332(*dst++, *src);
918 ++src; 950 ++src;
919 } 951 }
920 #endif /* USE_DUFFS_LOOP */ 952 #endif /* USE_DUFFS_LOOP */
921 src += srcskip; 953 src += srcskip;
922 dst += dstskip; 954 dst += dstskip;
923 } 955 }
924 } else { 956 } else {
925 int Pixel; 957 int Pixel;
926 958
927 while ( height-- ) { 959 while (height--) {
928 #ifdef USE_DUFFS_LOOP 960 #ifdef USE_DUFFS_LOOP
961 /* *INDENT-OFF* */
929 DUFFS_LOOP( 962 DUFFS_LOOP(
930 RGB888_RGB332(Pixel, *src); 963 RGB888_RGB332(Pixel, *src);
931 *dst++ = map[Pixel]; 964 *dst++ = map[Pixel];
932 ++src; 965 ++src;
933 , width); 966 , width);
967 /* *INDENT-ON* */
934 #else 968 #else
935 for ( c=width/4; c; --c ) { 969 for (c = width / 4; c; --c) {
936 /* Pack RGB into 8bit pixel */ 970 /* Pack RGB into 8bit pixel */
937 RGB888_RGB332(Pixel, *src); 971 RGB888_RGB332(Pixel, *src);
938 *dst++ = map[Pixel]; 972 *dst++ = map[Pixel];
939 ++src; 973 ++src;
940 RGB888_RGB332(Pixel, *src); 974 RGB888_RGB332(Pixel, *src);
941 *dst++ = map[Pixel]; 975 *dst++ = map[Pixel];
942 ++src; 976 ++src;
943 RGB888_RGB332(Pixel, *src); 977 RGB888_RGB332(Pixel, *src);
944 *dst++ = map[Pixel]; 978 *dst++ = map[Pixel];
945 ++src; 979 ++src;
946 RGB888_RGB332(Pixel, *src); 980 RGB888_RGB332(Pixel, *src);
947 *dst++ = map[Pixel]; 981 *dst++ = map[Pixel];
948 ++src; 982 ++src;
949 } 983 }
950 switch ( width & 3 ) { 984 switch (width & 3) {
951 case 3: 985 case 3:
952 RGB888_RGB332(Pixel, *src); 986 RGB888_RGB332(Pixel, *src);
953 *dst++ = map[Pixel]; 987 *dst++ = map[Pixel];
954 ++src; 988 ++src;
955 case 2: 989 case 2:
956 RGB888_RGB332(Pixel, *src); 990 RGB888_RGB332(Pixel, *src);
957 *dst++ = map[Pixel]; 991 *dst++ = map[Pixel];
958 ++src; 992 ++src;
959 case 1: 993 case 1:
960 RGB888_RGB332(Pixel, *src); 994 RGB888_RGB332(Pixel, *src);
961 *dst++ = map[Pixel]; 995 *dst++ = map[Pixel];
962 ++src; 996 ++src;
963 } 997 }
964 #endif /* USE_DUFFS_LOOP */ 998 #endif /* USE_DUFFS_LOOP */
965 src += srcskip; 999 src += srcskip;
966 dst += dstskip; 1000 dst += dstskip;
967 } 1001 }
968 } 1002 }
969 } 1003 }
1004
970 /* Special optimized blit for RGB 8-8-8 --> RGB 5-5-5 */ 1005 /* Special optimized blit for RGB 8-8-8 --> RGB 5-5-5 */
971 #define RGB888_RGB555(dst, src) { \ 1006 #define RGB888_RGB555(dst, src) { \
972 *(Uint16 *)(dst) = (Uint16)((((*src)&0x00F80000)>>9)| \ 1007 *(Uint16 *)(dst) = (Uint16)((((*src)&0x00F80000)>>9)| \
973 (((*src)&0x0000F800)>>6)| \ 1008 (((*src)&0x0000F800)>>6)| \
974 (((*src)&0x000000F8)>>3)); \ 1009 (((*src)&0x000000F8)>>3)); \
979 (((src[HI])&0x000000F8)>>3))<<16)| \ 1014 (((src[HI])&0x000000F8)>>3))<<16)| \
980 (((src[LO])&0x00F80000)>>9)| \ 1015 (((src[LO])&0x00F80000)>>9)| \
981 (((src[LO])&0x0000F800)>>6)| \ 1016 (((src[LO])&0x0000F800)>>6)| \
982 (((src[LO])&0x000000F8)>>3); \ 1017 (((src[LO])&0x000000F8)>>3); \
983 } 1018 }
984 static void Blit_RGB888_RGB555(SDL_BlitInfo *info) 1019 static void
1020 Blit_RGB888_RGB555(SDL_BlitInfo * info)
985 { 1021 {
986 #ifndef USE_DUFFS_LOOP 1022 #ifndef USE_DUFFS_LOOP
987 int c; 1023 int c;
988 #endif 1024 #endif
989 int width, height; 1025 int width, height;
990 Uint32 *src; 1026 Uint32 *src;
991 Uint16 *dst; 1027 Uint16 *dst;
992 int srcskip, dstskip; 1028 int srcskip, dstskip;
993 1029
994 /* Set up some basic variables */ 1030 /* Set up some basic variables */
995 width = info->d_width; 1031 width = info->d_width;
996 height = info->d_height; 1032 height = info->d_height;
997 src = (Uint32 *)info->s_pixels; 1033 src = (Uint32 *) info->s_pixels;
998 srcskip = info->s_skip/4; 1034 srcskip = info->s_skip / 4;
999 dst = (Uint16 *)info->d_pixels; 1035 dst = (Uint16 *) info->d_pixels;
1000 dstskip = info->d_skip/2; 1036 dstskip = info->d_skip / 2;
1001 1037
1002 #ifdef USE_DUFFS_LOOP 1038 #ifdef USE_DUFFS_LOOP
1003 while ( height-- ) { 1039 while (height--) {
1040 /* *INDENT-OFF* */
1004 DUFFS_LOOP( 1041 DUFFS_LOOP(
1005 RGB888_RGB555(dst, src); 1042 RGB888_RGB555(dst, src);
1006 ++src; 1043 ++src;
1007 ++dst; 1044 ++dst;
1008 , width); 1045 , width);
1009 src += srcskip; 1046 /* *INDENT-ON* */
1010 dst += dstskip; 1047 src += srcskip;
1011 } 1048 dst += dstskip;
1049 }
1012 #else 1050 #else
1013 /* Memory align at 4-byte boundary, if necessary */ 1051 /* Memory align at 4-byte boundary, if necessary */
1014 if ( (long)dst & 0x03 ) { 1052 if ((long) dst & 0x03) {
1015 /* Don't do anything if width is 0 */ 1053 /* Don't do anything if width is 0 */
1016 if ( width == 0 ) { 1054 if (width == 0) {
1017 return; 1055 return;
1018 } 1056 }
1019 --width; 1057 --width;
1020 1058
1021 while ( height-- ) { 1059 while (height--) {
1022 /* Perform copy alignment */ 1060 /* Perform copy alignment */
1023 RGB888_RGB555(dst, src); 1061 RGB888_RGB555(dst, src);
1024 ++src; 1062 ++src;
1025 ++dst; 1063 ++dst;
1026 1064
1027 /* Copy in 4 pixel chunks */ 1065 /* Copy in 4 pixel chunks */
1028 for ( c=width/4; c; --c ) { 1066 for (c = width / 4; c; --c) {
1029 RGB888_RGB555_TWO(dst, src); 1067 RGB888_RGB555_TWO(dst, src);
1030 src += 2; 1068 src += 2;
1031 dst += 2; 1069 dst += 2;
1032 RGB888_RGB555_TWO(dst, src); 1070 RGB888_RGB555_TWO(dst, src);
1033 src += 2; 1071 src += 2;
1034 dst += 2; 1072 dst += 2;
1035 } 1073 }
1036 /* Get any leftovers */ 1074 /* Get any leftovers */
1037 switch (width & 3) { 1075 switch (width & 3) {
1038 case 3: 1076 case 3:
1039 RGB888_RGB555(dst, src); 1077 RGB888_RGB555(dst, src);
1040 ++src; 1078 ++src;
1041 ++dst; 1079 ++dst;
1042 case 2: 1080 case 2:
1043 RGB888_RGB555_TWO(dst, src); 1081 RGB888_RGB555_TWO(dst, src);
1044 src += 2; 1082 src += 2;
1045 dst += 2; 1083 dst += 2;
1046 break; 1084 break;
1047 case 1: 1085 case 1:
1048 RGB888_RGB555(dst, src); 1086 RGB888_RGB555(dst, src);
1049 ++src; 1087 ++src;
1050 ++dst; 1088 ++dst;
1051 break; 1089 break;
1052 } 1090 }
1053 src += srcskip; 1091 src += srcskip;
1054 dst += dstskip; 1092 dst += dstskip;
1055 } 1093 }
1056 } else { 1094 } else {
1057 while ( height-- ) { 1095 while (height--) {
1058 /* Copy in 4 pixel chunks */ 1096 /* Copy in 4 pixel chunks */
1059 for ( c=width/4; c; --c ) { 1097 for (c = width / 4; c; --c) {
1060 RGB888_RGB555_TWO(dst, src); 1098 RGB888_RGB555_TWO(dst, src);
1061 src += 2; 1099 src += 2;
1062 dst += 2; 1100 dst += 2;
1063 RGB888_RGB555_TWO(dst, src); 1101 RGB888_RGB555_TWO(dst, src);
1064 src += 2; 1102 src += 2;
1065 dst += 2; 1103 dst += 2;
1066 } 1104 }
1067 /* Get any leftovers */ 1105 /* Get any leftovers */
1068 switch (width & 3) { 1106 switch (width & 3) {
1069 case 3: 1107 case 3:
1070 RGB888_RGB555(dst, src); 1108 RGB888_RGB555(dst, src);
1071 ++src; 1109 ++src;
1072 ++dst; 1110 ++dst;
1073 case 2: 1111 case 2:
1074 RGB888_RGB555_TWO(dst, src); 1112 RGB888_RGB555_TWO(dst, src);
1075 src += 2; 1113 src += 2;
1076 dst += 2; 1114 dst += 2;
1077 break; 1115 break;
1078 case 1: 1116 case 1:
1079 RGB888_RGB555(dst, src); 1117 RGB888_RGB555(dst, src);
1080 ++src; 1118 ++src;
1081 ++dst; 1119 ++dst;
1082 break; 1120 break;
1083 } 1121 }
1084 src += srcskip; 1122 src += srcskip;
1085 dst += dstskip; 1123 dst += dstskip;
1086 } 1124 }
1087 } 1125 }
1088 #endif /* USE_DUFFS_LOOP */ 1126 #endif /* USE_DUFFS_LOOP */
1089 } 1127 }
1128
1090 /* Special optimized blit for RGB 8-8-8 --> RGB 5-6-5 */ 1129 /* Special optimized blit for RGB 8-8-8 --> RGB 5-6-5 */
1091 #define RGB888_RGB565(dst, src) { \ 1130 #define RGB888_RGB565(dst, src) { \
1092 *(Uint16 *)(dst) = (Uint16)((((*src)&0x00F80000)>>8)| \ 1131 *(Uint16 *)(dst) = (Uint16)((((*src)&0x00F80000)>>8)| \
1093 (((*src)&0x0000FC00)>>5)| \ 1132 (((*src)&0x0000FC00)>>5)| \
1094 (((*src)&0x000000F8)>>3)); \ 1133 (((*src)&0x000000F8)>>3)); \
1099 (((src[HI])&0x000000F8)>>3))<<16)| \ 1138 (((src[HI])&0x000000F8)>>3))<<16)| \
1100 (((src[LO])&0x00F80000)>>8)| \ 1139 (((src[LO])&0x00F80000)>>8)| \
1101 (((src[LO])&0x0000FC00)>>5)| \ 1140 (((src[LO])&0x0000FC00)>>5)| \
1102 (((src[LO])&0x000000F8)>>3); \ 1141 (((src[LO])&0x000000F8)>>3); \
1103 } 1142 }
1104 static void Blit_RGB888_RGB565(SDL_BlitInfo *info) 1143 static void
1144 Blit_RGB888_RGB565(SDL_BlitInfo * info)
1105 { 1145 {
1106 #ifndef USE_DUFFS_LOOP 1146 #ifndef USE_DUFFS_LOOP
1107 int c; 1147 int c;
1108 #endif 1148 #endif
1109 int width, height; 1149 int width, height;
1110 Uint32 *src; 1150 Uint32 *src;
1111 Uint16 *dst; 1151 Uint16 *dst;
1112 int srcskip, dstskip; 1152 int srcskip, dstskip;
1113 1153
1114 /* Set up some basic variables */ 1154 /* Set up some basic variables */
1115 width = info->d_width; 1155 width = info->d_width;
1116 height = info->d_height; 1156 height = info->d_height;
1117 src = (Uint32 *)info->s_pixels; 1157 src = (Uint32 *) info->s_pixels;
1118 srcskip = info->s_skip/4; 1158 srcskip = info->s_skip / 4;
1119 dst = (Uint16 *)info->d_pixels; 1159 dst = (Uint16 *) info->d_pixels;
1120 dstskip = info->d_skip/2; 1160 dstskip = info->d_skip / 2;
1121 1161
1122 #ifdef USE_DUFFS_LOOP 1162 #ifdef USE_DUFFS_LOOP
1123 while ( height-- ) { 1163 while (height--) {
1164 /* *INDENT-OFF* */
1124 DUFFS_LOOP( 1165 DUFFS_LOOP(
1125 RGB888_RGB565(dst, src); 1166 RGB888_RGB565(dst, src);
1126 ++src; 1167 ++src;
1127 ++dst; 1168 ++dst;
1128 , width); 1169 , width);
1129 src += srcskip; 1170 /* *INDENT-ON* */
1130 dst += dstskip; 1171 src += srcskip;
1131 } 1172 dst += dstskip;
1173 }
1132 #else 1174 #else
1133 /* Memory align at 4-byte boundary, if necessary */ 1175 /* Memory align at 4-byte boundary, if necessary */
1134 if ( (long)dst & 0x03 ) { 1176 if ((long) dst & 0x03) {
1135 /* Don't do anything if width is 0 */ 1177 /* Don't do anything if width is 0 */
1136 if ( width == 0 ) { 1178 if (width == 0) {
1137 return; 1179 return;
1138 } 1180 }
1139 --width; 1181 --width;
1140 1182
1141 while ( height-- ) { 1183 while (height--) {
1142 /* Perform copy alignment */ 1184 /* Perform copy alignment */
1143 RGB888_RGB565(dst, src); 1185 RGB888_RGB565(dst, src);
1144 ++src; 1186 ++src;
1145 ++dst; 1187 ++dst;
1146 1188
1147 /* Copy in 4 pixel chunks */ 1189 /* Copy in 4 pixel chunks */
1148 for ( c=width/4; c; --c ) { 1190 for (c = width / 4; c; --c) {
1149 RGB888_RGB565_TWO(dst, src); 1191 RGB888_RGB565_TWO(dst, src);
1150 src += 2; 1192 src += 2;
1151 dst += 2; 1193 dst += 2;
1152 RGB888_RGB565_TWO(dst, src); 1194 RGB888_RGB565_TWO(dst, src);
1153 src += 2; 1195 src += 2;
1154 dst += 2; 1196 dst += 2;
1155 } 1197 }
1156 /* Get any leftovers */ 1198 /* Get any leftovers */
1157 switch (width & 3) { 1199 switch (width & 3) {
1158 case 3: 1200 case 3:
1159 RGB888_RGB565(dst, src); 1201 RGB888_RGB565(dst, src);
1160 ++src; 1202 ++src;
1161 ++dst; 1203 ++dst;
1162 case 2: 1204 case 2:
1163 RGB888_RGB565_TWO(dst, src); 1205 RGB888_RGB565_TWO(dst, src);
1164 src += 2; 1206 src += 2;
1165 dst += 2; 1207 dst += 2;
1166 break; 1208 break;
1167 case 1: 1209 case 1:
1168 RGB888_RGB565(dst, src); 1210 RGB888_RGB565(dst, src);
1169 ++src; 1211 ++src;
1170 ++dst; 1212 ++dst;
1171 break; 1213 break;
1172 } 1214 }
1173 src += srcskip; 1215 src += srcskip;
1174 dst += dstskip; 1216 dst += dstskip;
1175 } 1217 }
1176 } else { 1218 } else {
1177 while ( height-- ) { 1219 while (height--) {
1178 /* Copy in 4 pixel chunks */ 1220 /* Copy in 4 pixel chunks */
1179 for ( c=width/4; c; --c ) { 1221 for (c = width / 4; c; --c) {
1180 RGB888_RGB565_TWO(dst, src); 1222 RGB888_RGB565_TWO(dst, src);
1181 src += 2; 1223 src += 2;
1182 dst += 2; 1224 dst += 2;
1183 RGB888_RGB565_TWO(dst, src); 1225 RGB888_RGB565_TWO(dst, src);
1184 src += 2; 1226 src += 2;
1185 dst += 2; 1227 dst += 2;
1186 } 1228 }
1187 /* Get any leftovers */ 1229 /* Get any leftovers */
1188 switch (width & 3) { 1230 switch (width & 3) {
1189 case 3: 1231 case 3:
1190 RGB888_RGB565(dst, src); 1232 RGB888_RGB565(dst, src);
1191 ++src; 1233 ++src;
1192 ++dst; 1234 ++dst;
1193 case 2: 1235 case 2:
1194 RGB888_RGB565_TWO(dst, src); 1236 RGB888_RGB565_TWO(dst, src);
1195 src += 2; 1237 src += 2;
1196 dst += 2; 1238 dst += 2;
1197 break; 1239 break;
1198 case 1: 1240 case 1:
1199 RGB888_RGB565(dst, src); 1241 RGB888_RGB565(dst, src);
1200 ++src; 1242 ++src;
1201 ++dst; 1243 ++dst;
1202 break; 1244 break;
1203 } 1245 }
1204 src += srcskip; 1246 src += srcskip;
1205 dst += dstskip; 1247 dst += dstskip;
1206 } 1248 }
1207 } 1249 }
1208 #endif /* USE_DUFFS_LOOP */ 1250 #endif /* USE_DUFFS_LOOP */
1209 } 1251 }
1210 1252
1211 #endif /* SDL_HERMES_BLITTERS */ 1253 #endif /* SDL_HERMES_BLITTERS */
1212 1254
1213 1255
1214 /* Special optimized blit for RGB 5-6-5 --> 32-bit RGB surfaces */ 1256 /* Special optimized blit for RGB 5-6-5 --> 32-bit RGB surfaces */
1215 #define RGB565_32(dst, src, map) (map[src[LO]*2] + map[src[HI]*2+1]) 1257 #define RGB565_32(dst, src, map) (map[src[LO]*2] + map[src[HI]*2+1])
1216 static void Blit_RGB565_32(SDL_BlitInfo *info, const Uint32 *map) 1258 static void
1259 Blit_RGB565_32(SDL_BlitInfo * info, const Uint32 * map)
1217 { 1260 {
1218 #ifndef USE_DUFFS_LOOP 1261 #ifndef USE_DUFFS_LOOP
1219 int c; 1262 int c;
1220 #endif 1263 #endif
1221 int width, height; 1264 int width, height;
1222 Uint8 *src; 1265 Uint8 *src;
1223 Uint32 *dst; 1266 Uint32 *dst;
1224 int srcskip, dstskip; 1267 int srcskip, dstskip;
1225 1268
1226 /* Set up some basic variables */ 1269 /* Set up some basic variables */
1227 width = info->d_width; 1270 width = info->d_width;
1228 height = info->d_height; 1271 height = info->d_height;
1229 src = (Uint8 *)info->s_pixels; 1272 src = (Uint8 *) info->s_pixels;
1230 srcskip = info->s_skip; 1273 srcskip = info->s_skip;
1231 dst = (Uint32 *)info->d_pixels; 1274 dst = (Uint32 *) info->d_pixels;
1232 dstskip = info->d_skip/4; 1275 dstskip = info->d_skip / 4;
1233 1276
1234 #ifdef USE_DUFFS_LOOP 1277 #ifdef USE_DUFFS_LOOP
1235 while ( height-- ) { 1278 while (height--) {
1279 /* *INDENT-OFF* */
1236 DUFFS_LOOP( 1280 DUFFS_LOOP(
1237 { 1281 {
1238 *dst++ = RGB565_32(dst, src, map); 1282 *dst++ = RGB565_32(dst, src, map);
1239 src += 2; 1283 src += 2;
1240 }, 1284 },
1241 width); 1285 width);
1242 src += srcskip; 1286 /* *INDENT-ON* */
1243 dst += dstskip; 1287 src += srcskip;
1244 } 1288 dst += dstskip;
1289 }
1245 #else 1290 #else
1246 while ( height-- ) { 1291 while (height--) {
1247 /* Copy in 4 pixel chunks */ 1292 /* Copy in 4 pixel chunks */
1248 for ( c=width/4; c; --c ) { 1293 for (c = width / 4; c; --c) {
1249 *dst++ = RGB565_32(dst, src, map); 1294 *dst++ = RGB565_32(dst, src, map);
1250 src += 2; 1295 src += 2;
1251 *dst++ = RGB565_32(dst, src, map); 1296 *dst++ = RGB565_32(dst, src, map);
1252 src += 2; 1297 src += 2;
1253 *dst++ = RGB565_32(dst, src, map); 1298 *dst++ = RGB565_32(dst, src, map);
1254 src += 2; 1299 src += 2;
1255 *dst++ = RGB565_32(dst, src, map); 1300 *dst++ = RGB565_32(dst, src, map);
1256 src += 2; 1301 src += 2;
1257 } 1302 }
1258 /* Get any leftovers */ 1303 /* Get any leftovers */
1259 switch (width & 3) { 1304 switch (width & 3) {
1260 case 3: 1305 case 3:
1261 *dst++ = RGB565_32(dst, src, map); 1306 *dst++ = RGB565_32(dst, src, map);
1262 src += 2; 1307 src += 2;
1263 case 2: 1308 case 2:
1264 *dst++ = RGB565_32(dst, src, map); 1309 *dst++ = RGB565_32(dst, src, map);
1265 src += 2; 1310 src += 2;
1266 case 1: 1311 case 1:
1267 *dst++ = RGB565_32(dst, src, map); 1312 *dst++ = RGB565_32(dst, src, map);
1268 src += 2; 1313 src += 2;
1269 break; 1314 break;
1270 } 1315 }
1271 src += srcskip; 1316 src += srcskip;
1272 dst += dstskip; 1317 dst += dstskip;
1273 } 1318 }
1274 #endif /* USE_DUFFS_LOOP */ 1319 #endif /* USE_DUFFS_LOOP */
1275 } 1320 }
1276 1321
1277 /* Special optimized blit for RGB 5-6-5 --> ARGB 8-8-8-8 */ 1322 /* Special optimized blit for RGB 5-6-5 --> ARGB 8-8-8-8 */
1278 static const Uint32 RGB565_ARGB8888_LUT[512] = { 1323 static const Uint32 RGB565_ARGB8888_LUT[512] = {
1279 0x00000000, 0xff000000, 0x00000008, 0xff002000, 1324 0x00000000, 0xff000000, 0x00000008, 0xff002000,
1280 0x00000010, 0xff004000, 0x00000018, 0xff006100, 1325 0x00000010, 0xff004000, 0x00000018, 0xff006100,
1281 0x00000020, 0xff008100, 0x00000029, 0xff00a100, 1326 0x00000020, 0xff008100, 0x00000029, 0xff00a100,
1282 0x00000031, 0xff00c200, 0x00000039, 0xff00e200, 1327 0x00000031, 0xff00c200, 0x00000039, 0xff00e200,
1283 0x00000041, 0xff080000, 0x0000004a, 0xff082000, 1328 0x00000041, 0xff080000, 0x0000004a, 0xff082000,
1284 0x00000052, 0xff084000, 0x0000005a, 0xff086100, 1329 0x00000052, 0xff084000, 0x0000005a, 0xff086100,
1285 0x00000062, 0xff088100, 0x0000006a, 0xff08a100, 1330 0x00000062, 0xff088100, 0x0000006a, 0xff08a100,
1286 0x00000073, 0xff08c200, 0x0000007b, 0xff08e200, 1331 0x00000073, 0xff08c200, 0x0000007b, 0xff08e200,
1287 0x00000083, 0xff100000, 0x0000008b, 0xff102000, 1332 0x00000083, 0xff100000, 0x0000008b, 0xff102000,
1288 0x00000094, 0xff104000, 0x0000009c, 0xff106100, 1333 0x00000094, 0xff104000, 0x0000009c, 0xff106100,
1289 0x000000a4, 0xff108100, 0x000000ac, 0xff10a100, 1334 0x000000a4, 0xff108100, 0x000000ac, 0xff10a100,
1290 0x000000b4, 0xff10c200, 0x000000bd, 0xff10e200, 1335 0x000000b4, 0xff10c200, 0x000000bd, 0xff10e200,
1291 0x000000c5, 0xff180000, 0x000000cd, 0xff182000, 1336 0x000000c5, 0xff180000, 0x000000cd, 0xff182000,
1292 0x000000d5, 0xff184000, 0x000000de, 0xff186100, 1337 0x000000d5, 0xff184000, 0x000000de, 0xff186100,
1293 0x000000e6, 0xff188100, 0x000000ee, 0xff18a100, 1338 0x000000e6, 0xff188100, 0x000000ee, 0xff18a100,
1294 0x000000f6, 0xff18c200, 0x000000ff, 0xff18e200, 1339 0x000000f6, 0xff18c200, 0x000000ff, 0xff18e200,
1295 0x00000400, 0xff200000, 0x00000408, 0xff202000, 1340 0x00000400, 0xff200000, 0x00000408, 0xff202000,
1296 0x00000410, 0xff204000, 0x00000418, 0xff206100, 1341 0x00000410, 0xff204000, 0x00000418, 0xff206100,
1297 0x00000420, 0xff208100, 0x00000429, 0xff20a100, 1342 0x00000420, 0xff208100, 0x00000429, 0xff20a100,
1298 0x00000431, 0xff20c200, 0x00000439, 0xff20e200, 1343 0x00000431, 0xff20c200, 0x00000439, 0xff20e200,
1299 0x00000441, 0xff290000, 0x0000044a, 0xff292000, 1344 0x00000441, 0xff290000, 0x0000044a, 0xff292000,
1300 0x00000452, 0xff294000, 0x0000045a, 0xff296100, 1345 0x00000452, 0xff294000, 0x0000045a, 0xff296100,
1301 0x00000462, 0xff298100, 0x0000046a, 0xff29a100, 1346 0x00000462, 0xff298100, 0x0000046a, 0xff29a100,
1302 0x00000473, 0xff29c200, 0x0000047b, 0xff29e200, 1347 0x00000473, 0xff29c200, 0x0000047b, 0xff29e200,
1303 0x00000483, 0xff310000, 0x0000048b, 0xff312000, 1348 0x00000483, 0xff310000, 0x0000048b, 0xff312000,
1304 0x00000494, 0xff314000, 0x0000049c, 0xff316100, 1349 0x00000494, 0xff314000, 0x0000049c, 0xff316100,
1305 0x000004a4, 0xff318100, 0x000004ac, 0xff31a100, 1350 0x000004a4, 0xff318100, 0x000004ac, 0xff31a100,
1306 0x000004b4, 0xff31c200, 0x000004bd, 0xff31e200, 1351 0x000004b4, 0xff31c200, 0x000004bd, 0xff31e200,
1307 0x000004c5, 0xff390000, 0x000004cd, 0xff392000, 1352 0x000004c5, 0xff390000, 0x000004cd, 0xff392000,
1308 0x000004d5, 0xff394000, 0x000004de, 0xff396100, 1353 0x000004d5, 0xff394000, 0x000004de, 0xff396100,
1309 0x000004e6, 0xff398100, 0x000004ee, 0xff39a100, 1354 0x000004e6, 0xff398100, 0x000004ee, 0xff39a100,
1310 0x000004f6, 0xff39c200, 0x000004ff, 0xff39e200, 1355 0x000004f6, 0xff39c200, 0x000004ff, 0xff39e200,
1311 0x00000800, 0xff410000, 0x00000808, 0xff412000, 1356 0x00000800, 0xff410000, 0x00000808, 0xff412000,
1312 0x00000810, 0xff414000, 0x00000818, 0xff416100, 1357 0x00000810, 0xff414000, 0x00000818, 0xff416100,
1313 0x00000820, 0xff418100, 0x00000829, 0xff41a100, 1358 0x00000820, 0xff418100, 0x00000829, 0xff41a100,
1314 0x00000831, 0xff41c200, 0x00000839, 0xff41e200, 1359 0x00000831, 0xff41c200, 0x00000839, 0xff41e200,
1315 0x00000841, 0xff4a0000, 0x0000084a, 0xff4a2000, 1360 0x00000841, 0xff4a0000, 0x0000084a, 0xff4a2000,
1316 0x00000852, 0xff4a4000, 0x0000085a, 0xff4a6100, 1361 0x00000852, 0xff4a4000, 0x0000085a, 0xff4a6100,
1317 0x00000862, 0xff4a8100, 0x0000086a, 0xff4aa100, 1362 0x00000862, 0xff4a8100, 0x0000086a, 0xff4aa100,
1318 0x00000873, 0xff4ac200, 0x0000087b, 0xff4ae200, 1363 0x00000873, 0xff4ac200, 0x0000087b, 0xff4ae200,
1319 0x00000883, 0xff520000, 0x0000088b, 0xff522000, 1364 0x00000883, 0xff520000, 0x0000088b, 0xff522000,
1320 0x00000894, 0xff524000, 0x0000089c, 0xff526100, 1365 0x00000894, 0xff524000, 0x0000089c, 0xff526100,
1321 0x000008a4, 0xff528100, 0x000008ac, 0xff52a100, 1366 0x000008a4, 0xff528100, 0x000008ac, 0xff52a100,
1322 0x000008b4, 0xff52c200, 0x000008bd, 0xff52e200, 1367 0x000008b4, 0xff52c200, 0x000008bd, 0xff52e200,
1323 0x000008c5, 0xff5a0000, 0x000008cd, 0xff5a2000, 1368 0x000008c5, 0xff5a0000, 0x000008cd, 0xff5a2000,
1324 0x000008d5, 0xff5a4000, 0x000008de, 0xff5a6100, 1369 0x000008d5, 0xff5a4000, 0x000008de, 0xff5a6100,
1325 0x000008e6, 0xff5a8100, 0x000008ee, 0xff5aa100, 1370 0x000008e6, 0xff5a8100, 0x000008ee, 0xff5aa100,
1326 0x000008f6, 0xff5ac200, 0x000008ff, 0xff5ae200, 1371 0x000008f6, 0xff5ac200, 0x000008ff, 0xff5ae200,
1327 0x00000c00, 0xff620000, 0x00000c08, 0xff622000, 1372 0x00000c00, 0xff620000, 0x00000c08, 0xff622000,
1328 0x00000c10, 0xff624000, 0x00000c18, 0xff626100, 1373 0x00000c10, 0xff624000, 0x00000c18, 0xff626100,
1329 0x00000c20, 0xff628100, 0x00000c29, 0xff62a100, 1374 0x00000c20, 0xff628100, 0x00000c29, 0xff62a100,
1330 0x00000c31, 0xff62c200, 0x00000c39, 0xff62e200, 1375 0x00000c31, 0xff62c200, 0x00000c39, 0xff62e200,
1331 0x00000c41, 0xff6a0000, 0x00000c4a, 0xff6a2000, 1376 0x00000c41, 0xff6a0000, 0x00000c4a, 0xff6a2000,
1332 0x00000c52, 0xff6a4000, 0x00000c5a, 0xff6a6100, 1377 0x00000c52, 0xff6a4000, 0x00000c5a, 0xff6a6100,
1333 0x00000c62, 0xff6a8100, 0x00000c6a, 0xff6aa100, 1378 0x00000c62, 0xff6a8100, 0x00000c6a, 0xff6aa100,
1334 0x00000c73, 0xff6ac200, 0x00000c7b, 0xff6ae200, 1379 0x00000c73, 0xff6ac200, 0x00000c7b, 0xff6ae200,
1335 0x00000c83, 0xff730000, 0x00000c8b, 0xff732000, 1380 0x00000c83, 0xff730000, 0x00000c8b, 0xff732000,
1336 0x00000c94, 0xff734000, 0x00000c9c, 0xff736100, 1381 0x00000c94, 0xff734000, 0x00000c9c, 0xff736100,
1337 0x00000ca4, 0xff738100, 0x00000cac, 0xff73a100, 1382 0x00000ca4, 0xff738100, 0x00000cac, 0xff73a100,
1338 0x00000cb4, 0xff73c200, 0x00000cbd, 0xff73e200, 1383 0x00000cb4, 0xff73c200, 0x00000cbd, 0xff73e200,
1339 0x00000cc5, 0xff7b0000, 0x00000ccd, 0xff7b2000, 1384 0x00000cc5, 0xff7b0000, 0x00000ccd, 0xff7b2000,
1340 0x00000cd5, 0xff7b4000, 0x00000cde, 0xff7b6100, 1385 0x00000cd5, 0xff7b4000, 0x00000cde, 0xff7b6100,
1341 0x00000ce6, 0xff7b8100, 0x00000cee, 0xff7ba100, 1386 0x00000ce6, 0xff7b8100, 0x00000cee, 0xff7ba100,
1342 0x00000cf6, 0xff7bc200, 0x00000cff, 0xff7be200, 1387 0x00000cf6, 0xff7bc200, 0x00000cff, 0xff7be200,
1343 0x00001000, 0xff830000, 0x00001008, 0xff832000, 1388 0x00001000, 0xff830000, 0x00001008, 0xff832000,
1344 0x00001010, 0xff834000, 0x00001018, 0xff836100, 1389 0x00001010, 0xff834000, 0x00001018, 0xff836100,
1345 0x00001020, 0xff838100, 0x00001029, 0xff83a100, 1390 0x00001020, 0xff838100, 0x00001029, 0xff83a100,
1346 0x00001031, 0xff83c200, 0x00001039, 0xff83e200, 1391 0x00001031, 0xff83c200, 0x00001039, 0xff83e200,
1347 0x00001041, 0xff8b0000, 0x0000104a, 0xff8b2000, 1392 0x00001041, 0xff8b0000, 0x0000104a, 0xff8b2000,
1348 0x00001052, 0xff8b4000, 0x0000105a, 0xff8b6100, 1393 0x00001052, 0xff8b4000, 0x0000105a, 0xff8b6100,
1349 0x00001062, 0xff8b8100, 0x0000106a, 0xff8ba100, 1394 0x00001062, 0xff8b8100, 0x0000106a, 0xff8ba100,
1350 0x00001073, 0xff8bc200, 0x0000107b, 0xff8be200, 1395 0x00001073, 0xff8bc200, 0x0000107b, 0xff8be200,
1351 0x00001083, 0xff940000, 0x0000108b, 0xff942000, 1396 0x00001083, 0xff940000, 0x0000108b, 0xff942000,
1352 0x00001094, 0xff944000, 0x0000109c, 0xff946100, 1397 0x00001094, 0xff944000, 0x0000109c, 0xff946100,
1353 0x000010a4, 0xff948100, 0x000010ac, 0xff94a100, 1398 0x000010a4, 0xff948100, 0x000010ac, 0xff94a100,
1354 0x000010b4, 0xff94c200, 0x000010bd, 0xff94e200, 1399 0x000010b4, 0xff94c200, 0x000010bd, 0xff94e200,
1355 0x000010c5, 0xff9c0000, 0x000010cd, 0xff9c2000, 1400 0x000010c5, 0xff9c0000, 0x000010cd, 0xff9c2000,
1356 0x000010d5, 0xff9c4000, 0x000010de, 0xff9c6100, 1401 0x000010d5, 0xff9c4000, 0x000010de, 0xff9c6100,
1357 0x000010e6, 0xff9c8100, 0x000010ee, 0xff9ca100, 1402 0x000010e6, 0xff9c8100, 0x000010ee, 0xff9ca100,
1358 0x000010f6, 0xff9cc200, 0x000010ff, 0xff9ce200, 1403 0x000010f6, 0xff9cc200, 0x000010ff, 0xff9ce200,
1359 0x00001400, 0xffa40000, 0x00001408, 0xffa42000, 1404 0x00001400, 0xffa40000, 0x00001408, 0xffa42000,
1360 0x00001410, 0xffa44000, 0x00001418, 0xffa46100, 1405 0x00001410, 0xffa44000, 0x00001418, 0xffa46100,
1361 0x00001420, 0xffa48100, 0x00001429, 0xffa4a100, 1406 0x00001420, 0xffa48100, 0x00001429, 0xffa4a100,
1362 0x00001431, 0xffa4c200, 0x00001439, 0xffa4e200, 1407 0x00001431, 0xffa4c200, 0x00001439, 0xffa4e200,
1363 0x00001441, 0xffac0000, 0x0000144a, 0xffac2000, 1408 0x00001441, 0xffac0000, 0x0000144a, 0xffac2000,
1364 0x00001452, 0xffac4000, 0x0000145a, 0xffac6100, 1409 0x00001452, 0xffac4000, 0x0000145a, 0xffac6100,
1365 0x00001462, 0xffac8100, 0x0000146a, 0xffaca100, 1410 0x00001462, 0xffac8100, 0x0000146a, 0xffaca100,
1366 0x00001473, 0xffacc200, 0x0000147b, 0xfface200, 1411 0x00001473, 0xffacc200, 0x0000147b, 0xfface200,
1367 0x00001483, 0xffb40000, 0x0000148b, 0xffb42000, 1412 0x00001483, 0xffb40000, 0x0000148b, 0xffb42000,
1368 0x00001494, 0xffb44000, 0x0000149c, 0xffb46100, 1413 0x00001494, 0xffb44000, 0x0000149c, 0xffb46100,
1369 0x000014a4, 0xffb48100, 0x000014ac, 0xffb4a100, 1414 0x000014a4, 0xffb48100, 0x000014ac, 0xffb4a100,
1370 0x000014b4, 0xffb4c200, 0x000014bd, 0xffb4e200, 1415 0x000014b4, 0xffb4c200, 0x000014bd, 0xffb4e200,
1371 0x000014c5, 0xffbd0000, 0x000014cd, 0xffbd2000, 1416 0x000014c5, 0xffbd0000, 0x000014cd, 0xffbd2000,
1372 0x000014d5, 0xffbd4000, 0x000014de, 0xffbd6100, 1417 0x000014d5, 0xffbd4000, 0x000014de, 0xffbd6100,
1373 0x000014e6, 0xffbd8100, 0x000014ee, 0xffbda100, 1418 0x000014e6, 0xffbd8100, 0x000014ee, 0xffbda100,
1374 0x000014f6, 0xffbdc200, 0x000014ff, 0xffbde200, 1419 0x000014f6, 0xffbdc200, 0x000014ff, 0xffbde200,
1375 0x00001800, 0xffc50000, 0x00001808, 0xffc52000, 1420 0x00001800, 0xffc50000, 0x00001808, 0xffc52000,
1376 0x00001810, 0xffc54000, 0x00001818, 0xffc56100, 1421 0x00001810, 0xffc54000, 0x00001818, 0xffc56100,
1377 0x00001820, 0xffc58100, 0x00001829, 0xffc5a100, 1422 0x00001820, 0xffc58100, 0x00001829, 0xffc5a100,
1378 0x00001831, 0xffc5c200, 0x00001839, 0xffc5e200, 1423 0x00001831, 0xffc5c200, 0x00001839, 0xffc5e200,
1379 0x00001841, 0xffcd0000, 0x0000184a, 0xffcd2000, 1424 0x00001841, 0xffcd0000, 0x0000184a, 0xffcd2000,
1380 0x00001852, 0xffcd4000, 0x0000185a, 0xffcd6100, 1425 0x00001852, 0xffcd4000, 0x0000185a, 0xffcd6100,
1381 0x00001862, 0xffcd8100, 0x0000186a, 0xffcda100, 1426 0x00001862, 0xffcd8100, 0x0000186a, 0xffcda100,
1382 0x00001873, 0xffcdc200, 0x0000187b, 0xffcde200, 1427 0x00001873, 0xffcdc200, 0x0000187b, 0xffcde200,
1383 0x00001883, 0xffd50000, 0x0000188b, 0xffd52000, 1428 0x00001883, 0xffd50000, 0x0000188b, 0xffd52000,
1384 0x00001894, 0xffd54000, 0x0000189c, 0xffd56100, 1429 0x00001894, 0xffd54000, 0x0000189c, 0xffd56100,
1385 0x000018a4, 0xffd58100, 0x000018ac, 0xffd5a100, 1430 0x000018a4, 0xffd58100, 0x000018ac, 0xffd5a100,
1386 0x000018b4, 0xffd5c200, 0x000018bd, 0xffd5e200, 1431 0x000018b4, 0xffd5c200, 0x000018bd, 0xffd5e200,
1387 0x000018c5, 0xffde0000, 0x000018cd, 0xffde2000, 1432 0x000018c5, 0xffde0000, 0x000018cd, 0xffde2000,
1388 0x000018d5, 0xffde4000, 0x000018de, 0xffde6100, 1433 0x000018d5, 0xffde4000, 0x000018de, 0xffde6100,
1389 0x000018e6, 0xffde8100, 0x000018ee, 0xffdea100, 1434 0x000018e6, 0xffde8100, 0x000018ee, 0xffdea100,
1390 0x000018f6, 0xffdec200, 0x000018ff, 0xffdee200, 1435 0x000018f6, 0xffdec200, 0x000018ff, 0xffdee200,
1391 0x00001c00, 0xffe60000, 0x00001c08, 0xffe62000, 1436 0x00001c00, 0xffe60000, 0x00001c08, 0xffe62000,
1392 0x00001c10, 0xffe64000, 0x00001c18, 0xffe66100, 1437 0x00001c10, 0xffe64000, 0x00001c18, 0xffe66100,
1393 0x00001c20, 0xffe68100, 0x00001c29, 0xffe6a100, 1438 0x00001c20, 0xffe68100, 0x00001c29, 0xffe6a100,
1394 0x00001c31, 0xffe6c200, 0x00001c39, 0xffe6e200, 1439 0x00001c31, 0xffe6c200, 0x00001c39, 0xffe6e200,
1395 0x00001c41, 0xffee0000, 0x00001c4a, 0xffee2000, 1440 0x00001c41, 0xffee0000, 0x00001c4a, 0xffee2000,
1396 0x00001c52, 0xffee4000, 0x00001c5a, 0xffee6100, 1441 0x00001c52, 0xffee4000, 0x00001c5a, 0xffee6100,
1397 0x00001c62, 0xffee8100, 0x00001c6a, 0xffeea100, 1442 0x00001c62, 0xffee8100, 0x00001c6a, 0xffeea100,
1398 0x00001c73, 0xffeec200, 0x00001c7b, 0xffeee200, 1443 0x00001c73, 0xffeec200, 0x00001c7b, 0xffeee200,
1399 0x00001c83, 0xfff60000, 0x00001c8b, 0xfff62000, 1444 0x00001c83, 0xfff60000, 0x00001c8b, 0xfff62000,
1400 0x00001c94, 0xfff64000, 0x00001c9c, 0xfff66100, 1445 0x00001c94, 0xfff64000, 0x00001c9c, 0xfff66100,
1401 0x00001ca4, 0xfff68100, 0x00001cac, 0xfff6a100, 1446 0x00001ca4, 0xfff68100, 0x00001cac, 0xfff6a100,
1402 0x00001cb4, 0xfff6c200, 0x00001cbd, 0xfff6e200, 1447 0x00001cb4, 0xfff6c200, 0x00001cbd, 0xfff6e200,
1403 0x00001cc5, 0xffff0000, 0x00001ccd, 0xffff2000, 1448 0x00001cc5, 0xffff0000, 0x00001ccd, 0xffff2000,
1404 0x00001cd5, 0xffff4000, 0x00001cde, 0xffff6100, 1449 0x00001cd5, 0xffff4000, 0x00001cde, 0xffff6100,
1405 0x00001ce6, 0xffff8100, 0x00001cee, 0xffffa100, 1450 0x00001ce6, 0xffff8100, 0x00001cee, 0xffffa100,
1406 0x00001cf6, 0xffffc200, 0x00001cff, 0xffffe200 1451 0x00001cf6, 0xffffc200, 0x00001cff, 0xffffe200
1407 }; 1452 };
1408 static void Blit_RGB565_ARGB8888(SDL_BlitInfo *info) 1453 static void
1454 Blit_RGB565_ARGB8888(SDL_BlitInfo * info)
1409 { 1455 {
1410 Blit_RGB565_32(info, RGB565_ARGB8888_LUT); 1456 Blit_RGB565_32(info, RGB565_ARGB8888_LUT);
1411 } 1457 }
1412 1458
1413 /* Special optimized blit for RGB 5-6-5 --> ABGR 8-8-8-8 */ 1459 /* Special optimized blit for RGB 5-6-5 --> ABGR 8-8-8-8 */
1414 static const Uint32 RGB565_ABGR8888_LUT[512] = { 1460 static const Uint32 RGB565_ABGR8888_LUT[512] = {
1415 0xff000000, 0x00000000, 0xff080000, 0x00002000, 1461 0xff000000, 0x00000000, 0xff080000, 0x00002000,
1416 0xff100000, 0x00004000, 0xff180000, 0x00006100, 1462 0xff100000, 0x00004000, 0xff180000, 0x00006100,
1417 0xff200000, 0x00008100, 0xff290000, 0x0000a100, 1463 0xff200000, 0x00008100, 0xff290000, 0x0000a100,
1418 0xff310000, 0x0000c200, 0xff390000, 0x0000e200, 1464 0xff310000, 0x0000c200, 0xff390000, 0x0000e200,
1419 0xff410000, 0x00000008, 0xff4a0000, 0x00002008, 1465 0xff410000, 0x00000008, 0xff4a0000, 0x00002008,
1420 0xff520000, 0x00004008, 0xff5a0000, 0x00006108, 1466 0xff520000, 0x00004008, 0xff5a0000, 0x00006108,
1421 0xff620000, 0x00008108, 0xff6a0000, 0x0000a108, 1467 0xff620000, 0x00008108, 0xff6a0000, 0x0000a108,
1422 0xff730000, 0x0000c208, 0xff7b0000, 0x0000e208, 1468 0xff730000, 0x0000c208, 0xff7b0000, 0x0000e208,
1423 0xff830000, 0x00000010, 0xff8b0000, 0x00002010, 1469 0xff830000, 0x00000010, 0xff8b0000, 0x00002010,
1424 0xff940000, 0x00004010, 0xff9c0000, 0x00006110, 1470 0xff940000, 0x00004010, 0xff9c0000, 0x00006110,
1425 0xffa40000, 0x00008110, 0xffac0000, 0x0000a110, 1471 0xffa40000, 0x00008110, 0xffac0000, 0x0000a110,
1426 0xffb40000, 0x0000c210, 0xffbd0000, 0x0000e210, 1472 0xffb40000, 0x0000c210, 0xffbd0000, 0x0000e210,
1427 0xffc50000, 0x00000018, 0xffcd0000, 0x00002018, 1473 0xffc50000, 0x00000018, 0xffcd0000, 0x00002018,
1428 0xffd50000, 0x00004018, 0xffde0000, 0x00006118, 1474 0xffd50000, 0x00004018, 0xffde0000, 0x00006118,
1429 0xffe60000, 0x00008118, 0xffee0000, 0x0000a118, 1475 0xffe60000, 0x00008118, 0xffee0000, 0x0000a118,
1430 0xfff60000, 0x0000c218, 0xffff0000, 0x0000e218, 1476 0xfff60000, 0x0000c218, 0xffff0000, 0x0000e218,
1431 0xff000400, 0x00000020, 0xff080400, 0x00002020, 1477 0xff000400, 0x00000020, 0xff080400, 0x00002020,
1432 0xff100400, 0x00004020, 0xff180400, 0x00006120, 1478 0xff100400, 0x00004020, 0xff180400, 0x00006120,
1433 0xff200400, 0x00008120, 0xff290400, 0x0000a120, 1479 0xff200400, 0x00008120, 0xff290400, 0x0000a120,
1434 0xff310400, 0x0000c220, 0xff390400, 0x0000e220, 1480 0xff310400, 0x0000c220, 0xff390400, 0x0000e220,
1435 0xff410400, 0x00000029, 0xff4a0400, 0x00002029, 1481 0xff410400, 0x00000029, 0xff4a0400, 0x00002029,
1436 0xff520400, 0x00004029, 0xff5a0400, 0x00006129, 1482 0xff520400, 0x00004029, 0xff5a0400, 0x00006129,
1437 0xff620400, 0x00008129, 0xff6a0400, 0x0000a129, 1483 0xff620400, 0x00008129, 0xff6a0400, 0x0000a129,
1438 0xff730400, 0x0000c229, 0xff7b0400, 0x0000e229, 1484 0xff730400, 0x0000c229, 0xff7b0400, 0x0000e229,
1439 0xff830400, 0x00000031, 0xff8b0400, 0x00002031, 1485 0xff830400, 0x00000031, 0xff8b0400, 0x00002031,
1440 0xff940400, 0x00004031, 0xff9c0400, 0x00006131, 1486 0xff940400, 0x00004031, 0xff9c0400, 0x00006131,
1441 0xffa40400, 0x00008131, 0xffac0400, 0x0000a131, 1487 0xffa40400, 0x00008131, 0xffac0400, 0x0000a131,
1442 0xffb40400, 0x0000c231, 0xffbd0400, 0x0000e231, 1488 0xffb40400, 0x0000c231, 0xffbd0400, 0x0000e231,
1443 0xffc50400, 0x00000039, 0xffcd0400, 0x00002039, 1489 0xffc50400, 0x00000039, 0xffcd0400, 0x00002039,
1444 0xffd50400, 0x00004039, 0xffde0400, 0x00006139, 1490 0xffd50400, 0x00004039, 0xffde0400, 0x00006139,
1445 0xffe60400, 0x00008139, 0xffee0400, 0x0000a139, 1491 0xffe60400, 0x00008139, 0xffee0400, 0x0000a139,
1446 0xfff60400, 0x0000c239, 0xffff0400, 0x0000e239, 1492 0xfff60400, 0x0000c239, 0xffff0400, 0x0000e239,
1447 0xff000800, 0x00000041, 0xff080800, 0x00002041, 1493 0xff000800, 0x00000041, 0xff080800, 0x00002041,
1448 0xff100800, 0x00004041, 0xff180800, 0x00006141, 1494 0xff100800, 0x00004041, 0xff180800, 0x00006141,
1449 0xff200800, 0x00008141, 0xff290800, 0x0000a141, 1495 0xff200800, 0x00008141, 0xff290800, 0x0000a141,
1450 0xff310800, 0x0000c241, 0xff390800, 0x0000e241, 1496 0xff310800, 0x0000c241, 0xff390800, 0x0000e241,
1451 0xff410800, 0x0000004a, 0xff4a0800, 0x0000204a, 1497 0xff410800, 0x0000004a, 0xff4a0800, 0x0000204a,
1452 0xff520800, 0x0000404a, 0xff5a0800, 0x0000614a, 1498 0xff520800, 0x0000404a, 0xff5a0800, 0x0000614a,
1453 0xff620800, 0x0000814a, 0xff6a0800, 0x0000a14a, 1499 0xff620800, 0x0000814a, 0xff6a0800, 0x0000a14a,
1454 0xff730800, 0x0000c24a, 0xff7b0800, 0x0000e24a, 1500 0xff730800, 0x0000c24a, 0xff7b0800, 0x0000e24a,
1455 0xff830800, 0x00000052, 0xff8b0800, 0x00002052, 1501 0xff830800, 0x00000052, 0xff8b0800, 0x00002052,
1456 0xff940800, 0x00004052, 0xff9c0800, 0x00006152, 1502 0xff940800, 0x00004052, 0xff9c0800, 0x00006152,
1457 0xffa40800, 0x00008152, 0xffac0800, 0x0000a152, 1503 0xffa40800, 0x00008152, 0xffac0800, 0x0000a152,
1458 0xffb40800, 0x0000c252, 0xffbd0800, 0x0000e252, 1504 0xffb40800, 0x0000c252, 0xffbd0800, 0x0000e252,
1459 0xffc50800, 0x0000005a, 0xffcd0800, 0x0000205a, 1505 0xffc50800, 0x0000005a, 0xffcd0800, 0x0000205a,
1460 0xffd50800, 0x0000405a, 0xffde0800, 0x0000615a, 1506 0xffd50800, 0x0000405a, 0xffde0800, 0x0000615a,
1461 0xffe60800, 0x0000815a, 0xffee0800, 0x0000a15a, 1507 0xffe60800, 0x0000815a, 0xffee0800, 0x0000a15a,
1462 0xfff60800, 0x0000c25a, 0xffff0800, 0x0000e25a, 1508 0xfff60800, 0x0000c25a, 0xffff0800, 0x0000e25a,
1463 0xff000c00, 0x00000062, 0xff080c00, 0x00002062, 1509 0xff000c00, 0x00000062, 0xff080c00, 0x00002062,
1464 0xff100c00, 0x00004062, 0xff180c00, 0x00006162, 1510 0xff100c00, 0x00004062, 0xff180c00, 0x00006162,
1465 0xff200c00, 0x00008162, 0xff290c00, 0x0000a162, 1511 0xff200c00, 0x00008162, 0xff290c00, 0x0000a162,
1466 0xff310c00, 0x0000c262, 0xff390c00, 0x0000e262, 1512 0xff310c00, 0x0000c262, 0xff390c00, 0x0000e262,
1467 0xff410c00, 0x0000006a, 0xff4a0c00, 0x0000206a, 1513 0xff410c00, 0x0000006a, 0xff4a0c00, 0x0000206a,
1468 0xff520c00, 0x0000406a, 0xff5a0c00, 0x0000616a, 1514 0xff520c00, 0x0000406a, 0xff5a0c00, 0x0000616a,
1469 0xff620c00, 0x0000816a, 0xff6a0c00, 0x0000a16a, 1515 0xff620c00, 0x0000816a, 0xff6a0c00, 0x0000a16a,
1470 0xff730c00, 0x0000c26a, 0xff7b0c00, 0x0000e26a, 1516 0xff730c00, 0x0000c26a, 0xff7b0c00, 0x0000e26a,
1471 0xff830c00, 0x00000073, 0xff8b0c00, 0x00002073, 1517 0xff830c00, 0x00000073, 0xff8b0c00, 0x00002073,
1472 0xff940c00, 0x00004073, 0xff9c0c00, 0x00006173, 1518 0xff940c00, 0x00004073, 0xff9c0c00, 0x00006173,
1473 0xffa40c00, 0x00008173, 0xffac0c00, 0x0000a173, 1519 0xffa40c00, 0x00008173, 0xffac0c00, 0x0000a173,
1474 0xffb40c00, 0x0000c273, 0xffbd0c00, 0x0000e273, 1520 0xffb40c00, 0x0000c273, 0xffbd0c00, 0x0000e273,
1475 0xffc50c00, 0x0000007b, 0xffcd0c00, 0x0000207b, 1521 0xffc50c00, 0x0000007b, 0xffcd0c00, 0x0000207b,
1476 0xffd50c00, 0x0000407b, 0xffde0c00, 0x0000617b, 1522 0xffd50c00, 0x0000407b, 0xffde0c00, 0x0000617b,
1477 0xffe60c00, 0x0000817b, 0xffee0c00, 0x0000a17b, 1523 0xffe60c00, 0x0000817b, 0xffee0c00, 0x0000a17b,
1478 0xfff60c00, 0x0000c27b, 0xffff0c00, 0x0000e27b, 1524 0xfff60c00, 0x0000c27b, 0xffff0c00, 0x0000e27b,
1479 0xff001000, 0x00000083, 0xff081000, 0x00002083, 1525 0xff001000, 0x00000083, 0xff081000, 0x00002083,
1480 0xff101000, 0x00004083, 0xff181000, 0x00006183, 1526 0xff101000, 0x00004083, 0xff181000, 0x00006183,
1481 0xff201000, 0x00008183, 0xff291000, 0x0000a183, 1527 0xff201000, 0x00008183, 0xff291000, 0x0000a183,
1482 0xff311000, 0x0000c283, 0xff391000, 0x0000e283, 1528 0xff311000, 0x0000c283, 0xff391000, 0x0000e283,
1483 0xff411000, 0x0000008b, 0xff4a1000, 0x0000208b, 1529 0xff411000, 0x0000008b, 0xff4a1000, 0x0000208b,
1484 0xff521000, 0x0000408b, 0xff5a1000, 0x0000618b, 1530 0xff521000, 0x0000408b, 0xff5a1000, 0x0000618b,
1485 0xff621000, 0x0000818b, 0xff6a1000, 0x0000a18b, 1531 0xff621000, 0x0000818b, 0xff6a1000, 0x0000a18b,
1486 0xff731000, 0x0000c28b, 0xff7b1000, 0x0000e28b, 1532 0xff731000, 0x0000c28b, 0xff7b1000, 0x0000e28b,
1487 0xff831000, 0x00000094, 0xff8b1000, 0x00002094, 1533 0xff831000, 0x00000094, 0xff8b1000, 0x00002094,
1488 0xff941000, 0x00004094, 0xff9c1000, 0x00006194, 1534 0xff941000, 0x00004094, 0xff9c1000, 0x00006194,
1489 0xffa41000, 0x00008194, 0xffac1000, 0x0000a194, 1535 0xffa41000, 0x00008194, 0xffac1000, 0x0000a194,
1490 0xffb41000, 0x0000c294, 0xffbd1000, 0x0000e294, 1536 0xffb41000, 0x0000c294, 0xffbd1000, 0x0000e294,
1491 0xffc51000, 0x0000009c, 0xffcd1000, 0x0000209c, 1537 0xffc51000, 0x0000009c, 0xffcd1000, 0x0000209c,
1492 0xffd51000, 0x0000409c, 0xffde1000, 0x0000619c, 1538 0xffd51000, 0x0000409c, 0xffde1000, 0x0000619c,
1493 0xffe61000, 0x0000819c, 0xffee1000, 0x0000a19c, 1539 0xffe61000, 0x0000819c, 0xffee1000, 0x0000a19c,
1494 0xfff61000, 0x0000c29c, 0xffff1000, 0x0000e29c, 1540 0xfff61000, 0x0000c29c, 0xffff1000, 0x0000e29c,
1495 0xff001400, 0x000000a4, 0xff081400, 0x000020a4, 1541 0xff001400, 0x000000a4, 0xff081400, 0x000020a4,
1496 0xff101400, 0x000040a4, 0xff181400, 0x000061a4, 1542 0xff101400, 0x000040a4, 0xff181400, 0x000061a4,
1497 0xff201400, 0x000081a4, 0xff291400, 0x0000a1a4, 1543 0xff201400, 0x000081a4, 0xff291400, 0x0000a1a4,
1498 0xff311400, 0x0000c2a4, 0xff391400, 0x0000e2a4, 1544 0xff311400, 0x0000c2a4, 0xff391400, 0x0000e2a4,
1499 0xff411400, 0x000000ac, 0xff4a1400, 0x000020ac, 1545 0xff411400, 0x000000ac, 0xff4a1400, 0x000020ac,
1500 0xff521400, 0x000040ac, 0xff5a1400, 0x000061ac, 1546 0xff521400, 0x000040ac, 0xff5a1400, 0x000061ac,
1501 0xff621400, 0x000081ac, 0xff6a1400, 0x0000a1ac, 1547 0xff621400, 0x000081ac, 0xff6a1400, 0x0000a1ac,
1502 0xff731400, 0x0000c2ac, 0xff7b1400, 0x0000e2ac, 1548 0xff731400, 0x0000c2ac, 0xff7b1400, 0x0000e2ac,
1503 0xff831400, 0x000000b4, 0xff8b1400, 0x000020b4, 1549 0xff831400, 0x000000b4, 0xff8b1400, 0x000020b4,
1504 0xff941400, 0x000040b4, 0xff9c1400, 0x000061b4, 1550 0xff941400, 0x000040b4, 0xff9c1400, 0x000061b4,
1505 0xffa41400, 0x000081b4, 0xffac1400, 0x0000a1b4, 1551 0xffa41400, 0x000081b4, 0xffac1400, 0x0000a1b4,
1506 0xffb41400, 0x0000c2b4, 0xffbd1400, 0x0000e2b4, 1552 0xffb41400, 0x0000c2b4, 0xffbd1400, 0x0000e2b4,
1507 0xffc51400, 0x000000bd, 0xffcd1400, 0x000020bd, 1553 0xffc51400, 0x000000bd, 0xffcd1400, 0x000020bd,
1508 0xffd51400, 0x000040bd, 0xffde1400, 0x000061bd, 1554 0xffd51400, 0x000040bd, 0xffde1400, 0x000061bd,
1509 0xffe61400, 0x000081bd, 0xffee1400, 0x0000a1bd, 1555 0xffe61400, 0x000081bd, 0xffee1400, 0x0000a1bd,
1510 0xfff61400, 0x0000c2bd, 0xffff1400, 0x0000e2bd, 1556 0xfff61400, 0x0000c2bd, 0xffff1400, 0x0000e2bd,
1511 0xff001800, 0x000000c5, 0xff081800, 0x000020c5, 1557 0xff001800, 0x000000c5, 0xff081800, 0x000020c5,
1512 0xff101800, 0x000040c5, 0xff181800, 0x000061c5, 1558 0xff101800, 0x000040c5, 0xff181800, 0x000061c5,
1513 0xff201800, 0x000081c5, 0xff291800, 0x0000a1c5, 1559 0xff201800, 0x000081c5, 0xff291800, 0x0000a1c5,
1514 0xff311800, 0x0000c2c5, 0xff391800, 0x0000e2c5, 1560 0xff311800, 0x0000c2c5, 0xff391800, 0x0000e2c5,
1515 0xff411800, 0x000000cd, 0xff4a1800, 0x000020cd, 1561 0xff411800, 0x000000cd, 0xff4a1800, 0x000020cd,
1516 0xff521800, 0x000040cd, 0xff5a1800, 0x000061cd, 1562 0xff521800, 0x000040cd, 0xff5a1800, 0x000061cd,
1517 0xff621800, 0x000081cd, 0xff6a1800, 0x0000a1cd, 1563 0xff621800, 0x000081cd, 0xff6a1800, 0x0000a1cd,
1518 0xff731800, 0x0000c2cd, 0xff7b1800, 0x0000e2cd, 1564 0xff731800, 0x0000c2cd, 0xff7b1800, 0x0000e2cd,
1519 0xff831800, 0x000000d5, 0xff8b1800, 0x000020d5, 1565 0xff831800, 0x000000d5, 0xff8b1800, 0x000020d5,
1520 0xff941800, 0x000040d5, 0xff9c1800, 0x000061d5, 1566 0xff941800, 0x000040d5, 0xff9c1800, 0x000061d5,
1521 0xffa41800, 0x000081d5, 0xffac1800, 0x0000a1d5, 1567 0xffa41800, 0x000081d5, 0xffac1800, 0x0000a1d5,
1522 0xffb41800, 0x0000c2d5, 0xffbd1800, 0x0000e2d5, 1568 0xffb41800, 0x0000c2d5, 0xffbd1800, 0x0000e2d5,
1523 0xffc51800, 0x000000de, 0xffcd1800, 0x000020de, 1569 0xffc51800, 0x000000de, 0xffcd1800, 0x000020de,
1524 0xffd51800, 0x000040de, 0xffde1800, 0x000061de, 1570 0xffd51800, 0x000040de, 0xffde1800, 0x000061de,
1525 0xffe61800, 0x000081de, 0xffee1800, 0x0000a1de, 1571 0xffe61800, 0x000081de, 0xffee1800, 0x0000a1de,
1526 0xfff61800, 0x0000c2de, 0xffff1800, 0x0000e2de, 1572 0xfff61800, 0x0000c2de, 0xffff1800, 0x0000e2de,
1527 0xff001c00, 0x000000e6, 0xff081c00, 0x000020e6, 1573 0xff001c00, 0x000000e6, 0xff081c00, 0x000020e6,
1528 0xff101c00, 0x000040e6, 0xff181c00, 0x000061e6, 1574 0xff101c00, 0x000040e6, 0xff181c00, 0x000061e6,
1529 0xff201c00, 0x000081e6, 0xff291c00, 0x0000a1e6, 1575 0xff201c00, 0x000081e6, 0xff291c00, 0x0000a1e6,
1530 0xff311c00, 0x0000c2e6, 0xff391c00, 0x0000e2e6, 1576 0xff311c00, 0x0000c2e6, 0xff391c00, 0x0000e2e6,
1531 0xff411c00, 0x000000ee, 0xff4a1c00, 0x000020ee, 1577 0xff411c00, 0x000000ee, 0xff4a1c00, 0x000020ee,
1532 0xff521c00, 0x000040ee, 0xff5a1c00, 0x000061ee, 1578 0xff521c00, 0x000040ee, 0xff5a1c00, 0x000061ee,
1533 0xff621c00, 0x000081ee, 0xff6a1c00, 0x0000a1ee, 1579 0xff621c00, 0x000081ee, 0xff6a1c00, 0x0000a1ee,
1534 0xff731c00, 0x0000c2ee, 0xff7b1c00, 0x0000e2ee, 1580 0xff731c00, 0x0000c2ee, 0xff7b1c00, 0x0000e2ee,
1535 0xff831c00, 0x000000f6, 0xff8b1c00, 0x000020f6, 1581 0xff831c00, 0x000000f6, 0xff8b1c00, 0x000020f6,
1536 0xff941c00, 0x000040f6, 0xff9c1c00, 0x000061f6, 1582 0xff941c00, 0x000040f6, 0xff9c1c00, 0x000061f6,
1537 0xffa41c00, 0x000081f6, 0xffac1c00, 0x0000a1f6, 1583 0xffa41c00, 0x000081f6, 0xffac1c00, 0x0000a1f6,
1538 0xffb41c00, 0x0000c2f6, 0xffbd1c00, 0x0000e2f6, 1584 0xffb41c00, 0x0000c2f6, 0xffbd1c00, 0x0000e2f6,
1539 0xffc51c00, 0x000000ff, 0xffcd1c00, 0x000020ff, 1585 0xffc51c00, 0x000000ff, 0xffcd1c00, 0x000020ff,
1540 0xffd51c00, 0x000040ff, 0xffde1c00, 0x000061ff, 1586 0xffd51c00, 0x000040ff, 0xffde1c00, 0x000061ff,
1541 0xffe61c00, 0x000081ff, 0xffee1c00, 0x0000a1ff, 1587 0xffe61c00, 0x000081ff, 0xffee1c00, 0x0000a1ff,
1542 0xfff61c00, 0x0000c2ff, 0xffff1c00, 0x0000e2ff 1588 0xfff61c00, 0x0000c2ff, 0xffff1c00, 0x0000e2ff
1543 }; 1589 };
1544 static void Blit_RGB565_ABGR8888(SDL_BlitInfo *info) 1590 static void
1591 Blit_RGB565_ABGR8888(SDL_BlitInfo * info)
1545 { 1592 {
1546 Blit_RGB565_32(info, RGB565_ABGR8888_LUT); 1593 Blit_RGB565_32(info, RGB565_ABGR8888_LUT);
1547 } 1594 }
1548 1595
1549 /* Special optimized blit for RGB 5-6-5 --> RGBA 8-8-8-8 */ 1596 /* Special optimized blit for RGB 5-6-5 --> RGBA 8-8-8-8 */
1550 static const Uint32 RGB565_RGBA8888_LUT[512] = { 1597 static const Uint32 RGB565_RGBA8888_LUT[512] = {
1551 0x000000ff, 0x00000000, 0x000008ff, 0x00200000, 1598 0x000000ff, 0x00000000, 0x000008ff, 0x00200000,
1552 0x000010ff, 0x00400000, 0x000018ff, 0x00610000, 1599 0x000010ff, 0x00400000, 0x000018ff, 0x00610000,
1553 0x000020ff, 0x00810000, 0x000029ff, 0x00a10000, 1600 0x000020ff, 0x00810000, 0x000029ff, 0x00a10000,
1554 0x000031ff, 0x00c20000, 0x000039ff, 0x00e20000, 1601 0x000031ff, 0x00c20000, 0x000039ff, 0x00e20000,
1555 0x000041ff, 0x08000000, 0x00004aff, 0x08200000, 1602 0x000041ff, 0x08000000, 0x00004aff, 0x08200000,
1556 0x000052ff, 0x08400000, 0x00005aff, 0x08610000, 1603 0x000052ff, 0x08400000, 0x00005aff, 0x08610000,
1557 0x000062ff, 0x08810000, 0x00006aff, 0x08a10000, 1604 0x000062ff, 0x08810000, 0x00006aff, 0x08a10000,
1558 0x000073ff, 0x08c20000, 0x00007bff, 0x08e20000, 1605 0x000073ff, 0x08c20000, 0x00007bff, 0x08e20000,
1559 0x000083ff, 0x10000000, 0x00008bff, 0x10200000, 1606 0x000083ff, 0x10000000, 0x00008bff, 0x10200000,
1560 0x000094ff, 0x10400000, 0x00009cff, 0x10610000, 1607 0x000094ff, 0x10400000, 0x00009cff, 0x10610000,
1561 0x0000a4ff, 0x10810000, 0x0000acff, 0x10a10000, 1608 0x0000a4ff, 0x10810000, 0x0000acff, 0x10a10000,
1562 0x0000b4ff, 0x10c20000, 0x0000bdff, 0x10e20000, 1609 0x0000b4ff, 0x10c20000, 0x0000bdff, 0x10e20000,
1563 0x0000c5ff, 0x18000000, 0x0000cdff, 0x18200000, 1610 0x0000c5ff, 0x18000000, 0x0000cdff, 0x18200000,
1564 0x0000d5ff, 0x18400000, 0x0000deff, 0x18610000, 1611 0x0000d5ff, 0x18400000, 0x0000deff, 0x18610000,
1565 0x0000e6ff, 0x18810000, 0x0000eeff, 0x18a10000, 1612 0x0000e6ff, 0x18810000, 0x0000eeff, 0x18a10000,
1566 0x0000f6ff, 0x18c20000, 0x0000ffff, 0x18e20000, 1613 0x0000f6ff, 0x18c20000, 0x0000ffff, 0x18e20000,
1567 0x000400ff, 0x20000000, 0x000408ff, 0x20200000, 1614 0x000400ff, 0x20000000, 0x000408ff, 0x20200000,
1568 0x000410ff, 0x20400000, 0x000418ff, 0x20610000, 1615 0x000410ff, 0x20400000, 0x000418ff, 0x20610000,
1569 0x000420ff, 0x20810000, 0x000429ff, 0x20a10000, 1616 0x000420ff, 0x20810000, 0x000429ff, 0x20a10000,
1570 0x000431ff, 0x20c20000, 0x000439ff, 0x20e20000, 1617 0x000431ff, 0x20c20000, 0x000439ff, 0x20e20000,
1571 0x000441ff, 0x29000000, 0x00044aff, 0x29200000, 1618 0x000441ff, 0x29000000, 0x00044aff, 0x29200000,
1572 0x000452ff, 0x29400000, 0x00045aff, 0x29610000, 1619 0x000452ff, 0x29400000, 0x00045aff, 0x29610000,
1573 0x000462ff, 0x29810000, 0x00046aff, 0x29a10000, 1620 0x000462ff, 0x29810000, 0x00046aff, 0x29a10000,
1574 0x000473ff, 0x29c20000, 0x00047bff, 0x29e20000, 1621 0x000473ff, 0x29c20000, 0x00047bff, 0x29e20000,
1575 0x000483ff, 0x31000000, 0x00048bff, 0x31200000, 1622 0x000483ff, 0x31000000, 0x00048bff, 0x31200000,
1576 0x000494ff, 0x31400000, 0x00049cff, 0x31610000, 1623 0x000494ff, 0x31400000, 0x00049cff, 0x31610000,
1577 0x0004a4ff, 0x31810000, 0x0004acff, 0x31a10000, 1624 0x0004a4ff, 0x31810000, 0x0004acff, 0x31a10000,
1578 0x0004b4ff, 0x31c20000, 0x0004bdff, 0x31e20000, 1625 0x0004b4ff, 0x31c20000, 0x0004bdff, 0x31e20000,
1579 0x0004c5ff, 0x39000000, 0x0004cdff, 0x39200000, 1626 0x0004c5ff, 0x39000000, 0x0004cdff, 0x39200000,
1580 0x0004d5ff, 0x39400000, 0x0004deff, 0x39610000, 1627 0x0004d5ff, 0x39400000, 0x0004deff, 0x39610000,
1581 0x0004e6ff, 0x39810000, 0x0004eeff, 0x39a10000, 1628 0x0004e6ff, 0x39810000, 0x0004eeff, 0x39a10000,
1582 0x0004f6ff, 0x39c20000, 0x0004ffff, 0x39e20000, 1629 0x0004f6ff, 0x39c20000, 0x0004ffff, 0x39e20000,
1583 0x000800ff, 0x41000000, 0x000808ff, 0x41200000, 1630 0x000800ff, 0x41000000, 0x000808ff, 0x41200000,
1584 0x000810ff, 0x41400000, 0x000818ff, 0x41610000, 1631 0x000810ff, 0x41400000, 0x000818ff, 0x41610000,
1585 0x000820ff, 0x41810000, 0x000829ff, 0x41a10000, 1632 0x000820ff, 0x41810000, 0x000829ff, 0x41a10000,
1586 0x000831ff, 0x41c20000, 0x000839ff, 0x41e20000, 1633 0x000831ff, 0x41c20000, 0x000839ff, 0x41e20000,
1587 0x000841ff, 0x4a000000, 0x00084aff, 0x4a200000, 1634 0x000841ff, 0x4a000000, 0x00084aff, 0x4a200000,
1588 0x000852ff, 0x4a400000, 0x00085aff, 0x4a610000, 1635 0x000852ff, 0x4a400000, 0x00085aff, 0x4a610000,
1589 0x000862ff, 0x4a810000, 0x00086aff, 0x4aa10000, 1636 0x000862ff, 0x4a810000, 0x00086aff, 0x4aa10000,
1590 0x000873ff, 0x4ac20000, 0x00087bff, 0x4ae20000, 1637 0x000873ff, 0x4ac20000, 0x00087bff, 0x4ae20000,
1591 0x000883ff, 0x52000000, 0x00088bff, 0x52200000, 1638 0x000883ff, 0x52000000, 0x00088bff, 0x52200000,
1592 0x000894ff, 0x52400000, 0x00089cff, 0x52610000, 1639 0x000894ff, 0x52400000, 0x00089cff, 0x52610000,
1593 0x0008a4ff, 0x52810000, 0x0008acff, 0x52a10000, 1640 0x0008a4ff, 0x52810000, 0x0008acff, 0x52a10000,
1594 0x0008b4ff, 0x52c20000, 0x0008bdff, 0x52e20000, 1641 0x0008b4ff, 0x52c20000, 0x0008bdff, 0x52e20000,
1595 0x0008c5ff, 0x5a000000, 0x0008cdff, 0x5a200000, 1642 0x0008c5ff, 0x5a000000, 0x0008cdff, 0x5a200000,
1596 0x0008d5ff, 0x5a400000, 0x0008deff, 0x5a610000, 1643 0x0008d5ff, 0x5a400000, 0x0008deff, 0x5a610000,
1597 0x0008e6ff, 0x5a810000, 0x0008eeff, 0x5aa10000, 1644 0x0008e6ff, 0x5a810000, 0x0008eeff, 0x5aa10000,
1598 0x0008f6ff, 0x5ac20000, 0x0008ffff, 0x5ae20000, 1645 0x0008f6ff, 0x5ac20000, 0x0008ffff, 0x5ae20000,
1599 0x000c00ff, 0x62000000, 0x000c08ff, 0x62200000, 1646 0x000c00ff, 0x62000000, 0x000c08ff, 0x62200000,
1600 0x000c10ff, 0x62400000, 0x000c18ff, 0x62610000, 1647 0x000c10ff, 0x62400000, 0x000c18ff, 0x62610000,
1601 0x000c20ff, 0x62810000, 0x000c29ff, 0x62a10000, 1648 0x000c20ff, 0x62810000, 0x000c29ff, 0x62a10000,
1602 0x000c31ff, 0x62c20000, 0x000c39ff, 0x62e20000, 1649 0x000c31ff, 0x62c20000, 0x000c39ff, 0x62e20000,
1603 0x000c41ff, 0x6a000000, 0x000c4aff, 0x6a200000, 1650 0x000c41ff, 0x6a000000, 0x000c4aff, 0x6a200000,
1604 0x000c52ff, 0x6a400000, 0x000c5aff, 0x6a610000, 1651 0x000c52ff, 0x6a400000, 0x000c5aff, 0x6a610000,
1605 0x000c62ff, 0x6a810000, 0x000c6aff, 0x6aa10000, 1652 0x000c62ff, 0x6a810000, 0x000c6aff, 0x6aa10000,
1606 0x000c73ff, 0x6ac20000, 0x000c7bff, 0x6ae20000, 1653 0x000c73ff, 0x6ac20000, 0x000c7bff, 0x6ae20000,
1607 0x000c83ff, 0x73000000, 0x000c8bff, 0x73200000, 1654 0x000c83ff, 0x73000000, 0x000c8bff, 0x73200000,
1608 0x000c94ff, 0x73400000, 0x000c9cff, 0x73610000, 1655 0x000c94ff, 0x73400000, 0x000c9cff, 0x73610000,
1609 0x000ca4ff, 0x73810000, 0x000cacff, 0x73a10000, 1656 0x000ca4ff, 0x73810000, 0x000cacff, 0x73a10000,
1610 0x000cb4ff, 0x73c20000, 0x000cbdff, 0x73e20000, 1657 0x000cb4ff, 0x73c20000, 0x000cbdff, 0x73e20000,
1611 0x000cc5ff, 0x7b000000, 0x000ccdff, 0x7b200000, 1658 0x000cc5ff, 0x7b000000, 0x000ccdff, 0x7b200000,
1612 0x000cd5ff, 0x7b400000, 0x000cdeff, 0x7b610000, 1659 0x000cd5ff, 0x7b400000, 0x000cdeff, 0x7b610000,
1613 0x000ce6ff, 0x7b810000, 0x000ceeff, 0x7ba10000, 1660 0x000ce6ff, 0x7b810000, 0x000ceeff, 0x7ba10000,
1614 0x000cf6ff, 0x7bc20000, 0x000cffff, 0x7be20000, 1661 0x000cf6ff, 0x7bc20000, 0x000cffff, 0x7be20000,
1615 0x001000ff, 0x83000000, 0x001008ff, 0x83200000, 1662 0x001000ff, 0x83000000, 0x001008ff, 0x83200000,
1616 0x001010ff, 0x83400000, 0x001018ff, 0x83610000, 1663 0x001010ff, 0x83400000, 0x001018ff, 0x83610000,
1617 0x001020ff, 0x83810000, 0x001029ff, 0x83a10000, 1664 0x001020ff, 0x83810000, 0x001029ff, 0x83a10000,
1618 0x001031ff, 0x83c20000, 0x001039ff, 0x83e20000, 1665 0x001031ff, 0x83c20000, 0x001039ff, 0x83e20000,
1619 0x001041ff, 0x8b000000, 0x00104aff, 0x8b200000, 1666 0x001041ff, 0x8b000000, 0x00104aff, 0x8b200000,
1620 0x001052ff, 0x8b400000, 0x00105aff, 0x8b610000, 1667 0x001052ff, 0x8b400000, 0x00105aff, 0x8b610000,
1621 0x001062ff, 0x8b810000, 0x00106aff, 0x8ba10000, 1668 0x001062ff, 0x8b810000, 0x00106aff, 0x8ba10000,
1622 0x001073ff, 0x8bc20000, 0x00107bff, 0x8be20000, 1669 0x001073ff, 0x8bc20000, 0x00107bff, 0x8be20000,
1623 0x001083ff, 0x94000000, 0x00108bff, 0x94200000, 1670 0x001083ff, 0x94000000, 0x00108bff, 0x94200000,
1624 0x001094ff, 0x94400000, 0x00109cff, 0x94610000, 1671 0x001094ff, 0x94400000, 0x00109cff, 0x94610000,
1625 0x0010a4ff, 0x94810000, 0x0010acff, 0x94a10000, 1672 0x0010a4ff, 0x94810000, 0x0010acff, 0x94a10000,
1626 0x0010b4ff, 0x94c20000, 0x0010bdff, 0x94e20000, 1673 0x0010b4ff, 0x94c20000, 0x0010bdff, 0x94e20000,
1627 0x0010c5ff, 0x9c000000, 0x0010cdff, 0x9c200000, 1674 0x0010c5ff, 0x9c000000, 0x0010cdff, 0x9c200000,
1628 0x0010d5ff, 0x9c400000, 0x0010deff, 0x9c610000, 1675 0x0010d5ff, 0x9c400000, 0x0010deff, 0x9c610000,
1629 0x0010e6ff, 0x9c810000, 0x0010eeff, 0x9ca10000, 1676 0x0010e6ff, 0x9c810000, 0x0010eeff, 0x9ca10000,
1630 0x0010f6ff, 0x9cc20000, 0x0010ffff, 0x9ce20000, 1677 0x0010f6ff, 0x9cc20000, 0x0010ffff, 0x9ce20000,
1631 0x001400ff, 0xa4000000, 0x001408ff, 0xa4200000, 1678 0x001400ff, 0xa4000000, 0x001408ff, 0xa4200000,
1632 0x001410ff, 0xa4400000, 0x001418ff, 0xa4610000, 1679 0x001410ff, 0xa4400000, 0x001418ff, 0xa4610000,
1633 0x001420ff, 0xa4810000, 0x001429ff, 0xa4a10000, 1680 0x001420ff, 0xa4810000, 0x001429ff, 0xa4a10000,
1634 0x001431ff, 0xa4c20000, 0x001439ff, 0xa4e20000, 1681 0x001431ff, 0xa4c20000, 0x001439ff, 0xa4e20000,
1635 0x001441ff, 0xac000000, 0x00144aff, 0xac200000, 1682 0x001441ff, 0xac000000, 0x00144aff, 0xac200000,
1636 0x001452ff, 0xac400000, 0x00145aff, 0xac610000, 1683 0x001452ff, 0xac400000, 0x00145aff, 0xac610000,
1637 0x001462ff, 0xac810000, 0x00146aff, 0xaca10000, 1684 0x001462ff, 0xac810000, 0x00146aff, 0xaca10000,
1638 0x001473ff, 0xacc20000, 0x00147bff, 0xace20000, 1685 0x001473ff, 0xacc20000, 0x00147bff, 0xace20000,
1639 0x001483ff, 0xb4000000, 0x00148bff, 0xb4200000, 1686 0x001483ff, 0xb4000000, 0x00148bff, 0xb4200000,
1640 0x001494ff, 0xb4400000, 0x00149cff, 0xb4610000, 1687 0x001494ff, 0xb4400000, 0x00149cff, 0xb4610000,
1641 0x0014a4ff, 0xb4810000, 0x0014acff, 0xb4a10000, 1688 0x0014a4ff, 0xb4810000, 0x0014acff, 0xb4a10000,
1642 0x0014b4ff, 0xb4c20000, 0x0014bdff, 0xb4e20000, 1689 0x0014b4ff, 0xb4c20000, 0x0014bdff, 0xb4e20000,
1643 0x0014c5ff, 0xbd000000, 0x0014cdff, 0xbd200000, 1690 0x0014c5ff, 0xbd000000, 0x0014cdff, 0xbd200000,
1644 0x0014d5ff, 0xbd400000, 0x0014deff, 0xbd610000, 1691 0x0014d5ff, 0xbd400000, 0x0014deff, 0xbd610000,
1645 0x0014e6ff, 0xbd810000, 0x0014eeff, 0xbda10000, 1692 0x0014e6ff, 0xbd810000, 0x0014eeff, 0xbda10000,
1646 0x0014f6ff, 0xbdc20000, 0x0014ffff, 0xbde20000, 1693 0x0014f6ff, 0xbdc20000, 0x0014ffff, 0xbde20000,
1647 0x001800ff, 0xc5000000, 0x001808ff, 0xc5200000, 1694 0x001800ff, 0xc5000000, 0x001808ff, 0xc5200000,
1648 0x001810ff, 0xc5400000, 0x001818ff, 0xc5610000, 1695 0x001810ff, 0xc5400000, 0x001818ff, 0xc5610000,
1649 0x001820ff, 0xc5810000, 0x001829ff, 0xc5a10000, 1696 0x001820ff, 0xc5810000, 0x001829ff, 0xc5a10000,
1650 0x001831ff, 0xc5c20000, 0x001839ff, 0xc5e20000, 1697 0x001831ff, 0xc5c20000, 0x001839ff, 0xc5e20000,
1651 0x001841ff, 0xcd000000, 0x00184aff, 0xcd200000, 1698 0x001841ff, 0xcd000000, 0x00184aff, 0xcd200000,
1652 0x001852ff, 0xcd400000, 0x00185aff, 0xcd610000, 1699 0x001852ff, 0xcd400000, 0x00185aff, 0xcd610000,
1653 0x001862ff, 0xcd810000, 0x00186aff, 0xcda10000, 1700 0x001862ff, 0xcd810000, 0x00186aff, 0xcda10000,
1654 0x001873ff, 0xcdc20000, 0x00187bff, 0xcde20000, 1701 0x001873ff, 0xcdc20000, 0x00187bff, 0xcde20000,
1655 0x001883ff, 0xd5000000, 0x00188bff, 0xd5200000, 1702 0x001883ff, 0xd5000000, 0x00188bff, 0xd5200000,
1656 0x001894ff, 0xd5400000, 0x00189cff, 0xd5610000, 1703 0x001894ff, 0xd5400000, 0x00189cff, 0xd5610000,
1657 0x0018a4ff, 0xd5810000, 0x0018acff, 0xd5a10000, 1704 0x0018a4ff, 0xd5810000, 0x0018acff, 0xd5a10000,
1658 0x0018b4ff, 0xd5c20000, 0x0018bdff, 0xd5e20000, 1705 0x0018b4ff, 0xd5c20000, 0x0018bdff, 0xd5e20000,
1659 0x0018c5ff, 0xde000000, 0x0018cdff, 0xde200000, 1706 0x0018c5ff, 0xde000000, 0x0018cdff, 0xde200000,
1660 0x0018d5ff, 0xde400000, 0x0018deff, 0xde610000, 1707 0x0018d5ff, 0xde400000, 0x0018deff, 0xde610000,
1661 0x0018e6ff, 0xde810000, 0x0018eeff, 0xdea10000, 1708 0x0018e6ff, 0xde810000, 0x0018eeff, 0xdea10000,
1662 0x0018f6ff, 0xdec20000, 0x0018ffff, 0xdee20000, 1709 0x0018f6ff, 0xdec20000, 0x0018ffff, 0xdee20000,
1663 0x001c00ff, 0xe6000000, 0x001c08ff, 0xe6200000, 1710 0x001c00ff, 0xe6000000, 0x001c08ff, 0xe6200000,
1664 0x001c10ff, 0xe6400000, 0x001c18ff, 0xe6610000, 1711 0x001c10ff, 0xe6400000, 0x001c18ff, 0xe6610000,
1665 0x001c20ff, 0xe6810000, 0x001c29ff, 0xe6a10000, 1712 0x001c20ff, 0xe6810000, 0x001c29ff, 0xe6a10000,
1666 0x001c31ff, 0xe6c20000, 0x001c39ff, 0xe6e20000, 1713 0x001c31ff, 0xe6c20000, 0x001c39ff, 0xe6e20000,
1667 0x001c41ff, 0xee000000, 0x001c4aff, 0xee200000, 1714 0x001c41ff, 0xee000000, 0x001c4aff, 0xee200000,
1668 0x001c52ff, 0xee400000, 0x001c5aff, 0xee610000, 1715 0x001c52ff, 0xee400000, 0x001c5aff, 0xee610000,
1669 0x001c62ff, 0xee810000, 0x001c6aff, 0xeea10000, 1716 0x001c62ff, 0xee810000, 0x001c6aff, 0xeea10000,
1670 0x001c73ff, 0xeec20000, 0x001c7bff, 0xeee20000, 1717 0x001c73ff, 0xeec20000, 0x001c7bff, 0xeee20000,
1671 0x001c83ff, 0xf6000000, 0x001c8bff, 0xf6200000, 1718 0x001c83ff, 0xf6000000, 0x001c8bff, 0xf6200000,
1672 0x001c94ff, 0xf6400000, 0x001c9cff, 0xf6610000, 1719 0x001c94ff, 0xf6400000, 0x001c9cff, 0xf6610000,
1673 0x001ca4ff, 0xf6810000, 0x001cacff, 0xf6a10000, 1720 0x001ca4ff, 0xf6810000, 0x001cacff, 0xf6a10000,
1674 0x001cb4ff, 0xf6c20000, 0x001cbdff, 0xf6e20000, 1721 0x001cb4ff, 0xf6c20000, 0x001cbdff, 0xf6e20000,
1675 0x001cc5ff, 0xff000000, 0x001ccdff, 0xff200000, 1722 0x001cc5ff, 0xff000000, 0x001ccdff, 0xff200000,
1676 0x001cd5ff, 0xff400000, 0x001cdeff, 0xff610000, 1723 0x001cd5ff, 0xff400000, 0x001cdeff, 0xff610000,
1677 0x001ce6ff, 0xff810000, 0x001ceeff, 0xffa10000, 1724 0x001ce6ff, 0xff810000, 0x001ceeff, 0xffa10000,
1678 0x001cf6ff, 0xffc20000, 0x001cffff, 0xffe20000, 1725 0x001cf6ff, 0xffc20000, 0x001cffff, 0xffe20000,
1679 }; 1726 };
1680 static void Blit_RGB565_RGBA8888(SDL_BlitInfo *info) 1727 static void
1728 Blit_RGB565_RGBA8888(SDL_BlitInfo * info)
1681 { 1729 {
1682 Blit_RGB565_32(info, RGB565_RGBA8888_LUT); 1730 Blit_RGB565_32(info, RGB565_RGBA8888_LUT);
1683 } 1731 }
1684 1732
1685 /* Special optimized blit for RGB 5-6-5 --> BGRA 8-8-8-8 */ 1733 /* Special optimized blit for RGB 5-6-5 --> BGRA 8-8-8-8 */
1686 static const Uint32 RGB565_BGRA8888_LUT[512] = { 1734 static const Uint32 RGB565_BGRA8888_LUT[512] = {
1687 0x00000000, 0x000000ff, 0x08000000, 0x002000ff, 1735 0x00000000, 0x000000ff, 0x08000000, 0x002000ff,
1688 0x10000000, 0x004000ff, 0x18000000, 0x006100ff, 1736 0x10000000, 0x004000ff, 0x18000000, 0x006100ff,
1689 0x20000000, 0x008100ff, 0x29000000, 0x00a100ff, 1737 0x20000000, 0x008100ff, 0x29000000, 0x00a100ff,
1690 0x31000000, 0x00c200ff, 0x39000000, 0x00e200ff, 1738 0x31000000, 0x00c200ff, 0x39000000, 0x00e200ff,
1691 0x41000000, 0x000008ff, 0x4a000000, 0x002008ff, 1739 0x41000000, 0x000008ff, 0x4a000000, 0x002008ff,
1692 0x52000000, 0x004008ff, 0x5a000000, 0x006108ff, 1740 0x52000000, 0x004008ff, 0x5a000000, 0x006108ff,
1693 0x62000000, 0x008108ff, 0x6a000000, 0x00a108ff, 1741 0x62000000, 0x008108ff, 0x6a000000, 0x00a108ff,
1694 0x73000000, 0x00c208ff, 0x7b000000, 0x00e208ff, 1742 0x73000000, 0x00c208ff, 0x7b000000, 0x00e208ff,
1695 0x83000000, 0x000010ff, 0x8b000000, 0x002010ff, 1743 0x83000000, 0x000010ff, 0x8b000000, 0x002010ff,
1696 0x94000000, 0x004010ff, 0x9c000000, 0x006110ff, 1744 0x94000000, 0x004010ff, 0x9c000000, 0x006110ff,
1697 0xa4000000, 0x008110ff, 0xac000000, 0x00a110ff, 1745 0xa4000000, 0x008110ff, 0xac000000, 0x00a110ff,
1698 0xb4000000, 0x00c210ff, 0xbd000000, 0x00e210ff, 1746 0xb4000000, 0x00c210ff, 0xbd000000, 0x00e210ff,
1699 0xc5000000, 0x000018ff, 0xcd000000, 0x002018ff, 1747 0xc5000000, 0x000018ff, 0xcd000000, 0x002018ff,
1700 0xd5000000, 0x004018ff, 0xde000000, 0x006118ff, 1748 0xd5000000, 0x004018ff, 0xde000000, 0x006118ff,
1701 0xe6000000, 0x008118ff, 0xee000000, 0x00a118ff, 1749 0xe6000000, 0x008118ff, 0xee000000, 0x00a118ff,
1702 0xf6000000, 0x00c218ff, 0xff000000, 0x00e218ff, 1750 0xf6000000, 0x00c218ff, 0xff000000, 0x00e218ff,
1703 0x00040000, 0x000020ff, 0x08040000, 0x002020ff, 1751 0x00040000, 0x000020ff, 0x08040000, 0x002020ff,
1704 0x10040000, 0x004020ff, 0x18040000, 0x006120ff, 1752 0x10040000, 0x004020ff, 0x18040000, 0x006120ff,
1705 0x20040000, 0x008120ff, 0x29040000, 0x00a120ff, 1753 0x20040000, 0x008120ff, 0x29040000, 0x00a120ff,
1706 0x31040000, 0x00c220ff, 0x39040000, 0x00e220ff, 1754 0x31040000, 0x00c220ff, 0x39040000, 0x00e220ff,
1707 0x41040000, 0x000029ff, 0x4a040000, 0x002029ff, 1755 0x41040000, 0x000029ff, 0x4a040000, 0x002029ff,
1708 0x52040000, 0x004029ff, 0x5a040000, 0x006129ff, 1756 0x52040000, 0x004029ff, 0x5a040000, 0x006129ff,
1709 0x62040000, 0x008129ff, 0x6a040000, 0x00a129ff, 1757 0x62040000, 0x008129ff, 0x6a040000, 0x00a129ff,
1710 0x73040000, 0x00c229ff, 0x7b040000, 0x00e229ff, 1758 0x73040000, 0x00c229ff, 0x7b040000, 0x00e229ff,
1711 0x83040000, 0x000031ff, 0x8b040000, 0x002031ff, 1759 0x83040000, 0x000031ff, 0x8b040000, 0x002031ff,
1712 0x94040000, 0x004031ff, 0x9c040000, 0x006131ff, 1760 0x94040000, 0x004031ff, 0x9c040000, 0x006131ff,
1713 0xa4040000, 0x008131ff, 0xac040000, 0x00a131ff, 1761 0xa4040000, 0x008131ff, 0xac040000, 0x00a131ff,
1714 0xb4040000, 0x00c231ff, 0xbd040000, 0x00e231ff, 1762 0xb4040000, 0x00c231ff, 0xbd040000, 0x00e231ff,
1715 0xc5040000, 0x000039ff, 0xcd040000, 0x002039ff, 1763 0xc5040000, 0x000039ff, 0xcd040000, 0x002039ff,
1716 0xd5040000, 0x004039ff, 0xde040000, 0x006139ff, 1764 0xd5040000, 0x004039ff, 0xde040000, 0x006139ff,
1717 0xe6040000, 0x008139ff, 0xee040000, 0x00a139ff, 1765 0xe6040000, 0x008139ff, 0xee040000, 0x00a139ff,
1718 0xf6040000, 0x00c239ff, 0xff040000, 0x00e239ff, 1766 0xf6040000, 0x00c239ff, 0xff040000, 0x00e239ff,
1719 0x00080000, 0x000041ff, 0x08080000, 0x002041ff, 1767 0x00080000, 0x000041ff, 0x08080000, 0x002041ff,
1720 0x10080000, 0x004041ff, 0x18080000, 0x006141ff, 1768 0x10080000, 0x004041ff, 0x18080000, 0x006141ff,
1721 0x20080000, 0x008141ff, 0x29080000, 0x00a141ff, 1769 0x20080000, 0x008141ff, 0x29080000, 0x00a141ff,
1722 0x31080000, 0x00c241ff, 0x39080000, 0x00e241ff, 1770 0x31080000, 0x00c241ff, 0x39080000, 0x00e241ff,
1723 0x41080000, 0x00004aff, 0x4a080000, 0x00204aff, 1771 0x41080000, 0x00004aff, 0x4a080000, 0x00204aff,
1724 0x52080000, 0x00404aff, 0x5a080000, 0x00614aff, 1772 0x52080000, 0x00404aff, 0x5a080000, 0x00614aff,
1725 0x62080000, 0x00814aff, 0x6a080000, 0x00a14aff, 1773 0x62080000, 0x00814aff, 0x6a080000, 0x00a14aff,
1726 0x73080000, 0x00c24aff, 0x7b080000, 0x00e24aff, 1774 0x73080000, 0x00c24aff, 0x7b080000, 0x00e24aff,
1727 0x83080000, 0x000052ff, 0x8b080000, 0x002052ff, 1775 0x83080000, 0x000052ff, 0x8b080000, 0x002052ff,
1728 0x94080000, 0x004052ff, 0x9c080000, 0x006152ff, 1776 0x94080000, 0x004052ff, 0x9c080000, 0x006152ff,
1729 0xa4080000, 0x008152ff, 0xac080000, 0x00a152ff, 1777 0xa4080000, 0x008152ff, 0xac080000, 0x00a152ff,
1730 0xb4080000, 0x00c252ff, 0xbd080000, 0x00e252ff, 1778 0xb4080000, 0x00c252ff, 0xbd080000, 0x00e252ff,
1731 0xc5080000, 0x00005aff, 0xcd080000, 0x00205aff, 1779 0xc5080000, 0x00005aff, 0xcd080000, 0x00205aff,
1732 0xd5080000, 0x00405aff, 0xde080000, 0x00615aff, 1780 0xd5080000, 0x00405aff, 0xde080000, 0x00615aff,
1733 0xe6080000, 0x00815aff, 0xee080000, 0x00a15aff, 1781 0xe6080000, 0x00815aff, 0xee080000, 0x00a15aff,
1734 0xf6080000, 0x00c25aff, 0xff080000, 0x00e25aff, 1782 0xf6080000, 0x00c25aff, 0xff080000, 0x00e25aff,
1735 0x000c0000, 0x000062ff, 0x080c0000, 0x002062ff, 1783 0x000c0000, 0x000062ff, 0x080c0000, 0x002062ff,
1736 0x100c0000, 0x004062ff, 0x180c0000, 0x006162ff, 1784 0x100c0000, 0x004062ff, 0x180c0000, 0x006162ff,
1737 0x200c0000, 0x008162ff, 0x290c0000, 0x00a162ff, 1785 0x200c0000, 0x008162ff, 0x290c0000, 0x00a162ff,
1738 0x310c0000, 0x00c262ff, 0x390c0000, 0x00e262ff, 1786 0x310c0000, 0x00c262ff, 0x390c0000, 0x00e262ff,
1739 0x410c0000, 0x00006aff, 0x4a0c0000, 0x00206aff, 1787 0x410c0000, 0x00006aff, 0x4a0c0000, 0x00206aff,
1740 0x520c0000, 0x00406aff, 0x5a0c0000, 0x00616aff, 1788 0x520c0000, 0x00406aff, 0x5a0c0000, 0x00616aff,
1741 0x620c0000, 0x00816aff, 0x6a0c0000, 0x00a16aff, 1789 0x620c0000, 0x00816aff, 0x6a0c0000, 0x00a16aff,
1742 0x730c0000, 0x00c26aff, 0x7b0c0000, 0x00e26aff, 1790 0x730c0000, 0x00c26aff, 0x7b0c0000, 0x00e26aff,
1743 0x830c0000, 0x000073ff, 0x8b0c0000, 0x002073ff, 1791 0x830c0000, 0x000073ff, 0x8b0c0000, 0x002073ff,
1744 0x940c0000, 0x004073ff, 0x9c0c0000, 0x006173ff, 1792 0x940c0000, 0x004073ff, 0x9c0c0000, 0x006173ff,
1745 0xa40c0000, 0x008173ff, 0xac0c0000, 0x00a173ff, 1793 0xa40c0000, 0x008173ff, 0xac0c0000, 0x00a173ff,
1746 0xb40c0000, 0x00c273ff, 0xbd0c0000, 0x00e273ff, 1794 0xb40c0000, 0x00c273ff, 0xbd0c0000, 0x00e273ff,
1747 0xc50c0000, 0x00007bff, 0xcd0c0000, 0x00207bff, 1795 0xc50c0000, 0x00007bff, 0xcd0c0000, 0x00207bff,
1748 0xd50c0000, 0x00407bff, 0xde0c0000, 0x00617bff, 1796 0xd50c0000, 0x00407bff, 0xde0c0000, 0x00617bff,
1749 0xe60c0000, 0x00817bff, 0xee0c0000, 0x00a17bff, 1797 0xe60c0000, 0x00817bff, 0xee0c0000, 0x00a17bff,
1750 0xf60c0000, 0x00c27bff, 0xff0c0000, 0x00e27bff, 1798 0xf60c0000, 0x00c27bff, 0xff0c0000, 0x00e27bff,
1751 0x00100000, 0x000083ff, 0x08100000, 0x002083ff, 1799 0x00100000, 0x000083ff, 0x08100000, 0x002083ff,
1752 0x10100000, 0x004083ff, 0x18100000, 0x006183ff, 1800 0x10100000, 0x004083ff, 0x18100000, 0x006183ff,
1753 0x20100000, 0x008183ff, 0x29100000, 0x00a183ff, 1801 0x20100000, 0x008183ff, 0x29100000, 0x00a183ff,
1754 0x31100000, 0x00c283ff, 0x39100000, 0x00e283ff, 1802 0x31100000, 0x00c283ff, 0x39100000, 0x00e283ff,
1755 0x41100000, 0x00008bff, 0x4a100000, 0x00208bff, 1803 0x41100000, 0x00008bff, 0x4a100000, 0x00208bff,
1756 0x52100000, 0x00408bff, 0x5a100000, 0x00618bff, 1804 0x52100000, 0x00408bff, 0x5a100000, 0x00618bff,
1757 0x62100000, 0x00818bff, 0x6a100000, 0x00a18bff, 1805 0x62100000, 0x00818bff, 0x6a100000, 0x00a18bff,
1758 0x73100000, 0x00c28bff, 0x7b100000, 0x00e28bff, 1806 0x73100000, 0x00c28bff, 0x7b100000, 0x00e28bff,
1759 0x83100000, 0x000094ff, 0x8b100000, 0x002094ff, 1807 0x83100000, 0x000094ff, 0x8b100000, 0x002094ff,
1760 0x94100000, 0x004094ff, 0x9c100000, 0x006194ff, 1808 0x94100000, 0x004094ff, 0x9c100000, 0x006194ff,
1761 0xa4100000, 0x008194ff, 0xac100000, 0x00a194ff, 1809 0xa4100000, 0x008194ff, 0xac100000, 0x00a194ff,
1762 0xb4100000, 0x00c294ff, 0xbd100000, 0x00e294ff, 1810 0xb4100000, 0x00c294ff, 0xbd100000, 0x00e294ff,
1763 0xc5100000, 0x00009cff, 0xcd100000, 0x00209cff, 1811 0xc5100000, 0x00009cff, 0xcd100000, 0x00209cff,
1764 0xd5100000, 0x00409cff, 0xde100000, 0x00619cff, 1812 0xd5100000, 0x00409cff, 0xde100000, 0x00619cff,
1765 0xe6100000, 0x00819cff, 0xee100000, 0x00a19cff, 1813 0xe6100000, 0x00819cff, 0xee100000, 0x00a19cff,
1766 0xf6100000, 0x00c29cff, 0xff100000, 0x00e29cff, 1814 0xf6100000, 0x00c29cff, 0xff100000, 0x00e29cff,
1767 0x00140000, 0x0000a4ff, 0x08140000, 0x0020a4ff, 1815 0x00140000, 0x0000a4ff, 0x08140000, 0x0020a4ff,
1768 0x10140000, 0x0040a4ff, 0x18140000, 0x0061a4ff, 1816 0x10140000, 0x0040a4ff, 0x18140000, 0x0061a4ff,
1769 0x20140000, 0x0081a4ff, 0x29140000, 0x00a1a4ff, 1817 0x20140000, 0x0081a4ff, 0x29140000, 0x00a1a4ff,
1770 0x31140000, 0x00c2a4ff, 0x39140000, 0x00e2a4ff, 1818 0x31140000, 0x00c2a4ff, 0x39140000, 0x00e2a4ff,
1771 0x41140000, 0x0000acff, 0x4a140000, 0x0020acff, 1819 0x41140000, 0x0000acff, 0x4a140000, 0x0020acff,
1772 0x52140000, 0x0040acff, 0x5a140000, 0x0061acff, 1820 0x52140000, 0x0040acff, 0x5a140000, 0x0061acff,
1773 0x62140000, 0x0081acff, 0x6a140000, 0x00a1acff, 1821 0x62140000, 0x0081acff, 0x6a140000, 0x00a1acff,
1774 0x73140000, 0x00c2acff, 0x7b140000, 0x00e2acff, 1822 0x73140000, 0x00c2acff, 0x7b140000, 0x00e2acff,
1775 0x83140000, 0x0000b4ff, 0x8b140000, 0x0020b4ff, 1823 0x83140000, 0x0000b4ff, 0x8b140000, 0x0020b4ff,
1776 0x94140000, 0x0040b4ff, 0x9c140000, 0x0061b4ff, 1824 0x94140000, 0x0040b4ff, 0x9c140000, 0x0061b4ff,
1777 0xa4140000, 0x0081b4ff, 0xac140000, 0x00a1b4ff, 1825 0xa4140000, 0x0081b4ff, 0xac140000, 0x00a1b4ff,
1778 0xb4140000, 0x00c2b4ff, 0xbd140000, 0x00e2b4ff, 1826 0xb4140000, 0x00c2b4ff, 0xbd140000, 0x00e2b4ff,
1779 0xc5140000, 0x0000bdff, 0xcd140000, 0x0020bdff, 1827 0xc5140000, 0x0000bdff, 0xcd140000, 0x0020bdff,
1780 0xd5140000, 0x0040bdff, 0xde140000, 0x0061bdff, 1828 0xd5140000, 0x0040bdff, 0xde140000, 0x0061bdff,
1781 0xe6140000, 0x0081bdff, 0xee140000, 0x00a1bdff, 1829 0xe6140000, 0x0081bdff, 0xee140000, 0x00a1bdff,
1782 0xf6140000, 0x00c2bdff, 0xff140000, 0x00e2bdff, 1830 0xf6140000, 0x00c2bdff, 0xff140000, 0x00e2bdff,
1783 0x00180000, 0x0000c5ff, 0x08180000, 0x0020c5ff, 1831 0x00180000, 0x0000c5ff, 0x08180000, 0x0020c5ff,
1784 0x10180000, 0x0040c5ff, 0x18180000, 0x0061c5ff, 1832 0x10180000, 0x0040c5ff, 0x18180000, 0x0061c5ff,
1785 0x20180000, 0x0081c5ff, 0x29180000, 0x00a1c5ff, 1833 0x20180000, 0x0081c5ff, 0x29180000, 0x00a1c5ff,
1786 0x31180000, 0x00c2c5ff, 0x39180000, 0x00e2c5ff, 1834 0x31180000, 0x00c2c5ff, 0x39180000, 0x00e2c5ff,
1787 0x41180000, 0x0000cdff, 0x4a180000, 0x0020cdff, 1835 0x41180000, 0x0000cdff, 0x4a180000, 0x0020cdff,
1788 0x52180000, 0x0040cdff, 0x5a180000, 0x0061cdff, 1836 0x52180000, 0x0040cdff, 0x5a180000, 0x0061cdff,
1789 0x62180000, 0x0081cdff, 0x6a180000, 0x00a1cdff, 1837 0x62180000, 0x0081cdff, 0x6a180000, 0x00a1cdff,
1790 0x73180000, 0x00c2cdff, 0x7b180000, 0x00e2cdff, 1838 0x73180000, 0x00c2cdff, 0x7b180000, 0x00e2cdff,
1791 0x83180000, 0x0000d5ff, 0x8b180000, 0x0020d5ff, 1839 0x83180000, 0x0000d5ff, 0x8b180000, 0x0020d5ff,
1792 0x94180000, 0x0040d5ff, 0x9c180000, 0x0061d5ff, 1840 0x94180000, 0x0040d5ff, 0x9c180000, 0x0061d5ff,
1793 0xa4180000, 0x0081d5ff, 0xac180000, 0x00a1d5ff, 1841 0xa4180000, 0x0081d5ff, 0xac180000, 0x00a1d5ff,
1794 0xb4180000, 0x00c2d5ff, 0xbd180000, 0x00e2d5ff, 1842 0xb4180000, 0x00c2d5ff, 0xbd180000, 0x00e2d5ff,
1795 0xc5180000, 0x0000deff, 0xcd180000, 0x0020deff, 1843 0xc5180000, 0x0000deff, 0xcd180000, 0x0020deff,
1796 0xd5180000, 0x0040deff, 0xde180000, 0x0061deff, 1844 0xd5180000, 0x0040deff, 0xde180000, 0x0061deff,
1797 0xe6180000, 0x0081deff, 0xee180000, 0x00a1deff, 1845 0xe6180000, 0x0081deff, 0xee180000, 0x00a1deff,
1798 0xf6180000, 0x00c2deff, 0xff180000, 0x00e2deff, 1846 0xf6180000, 0x00c2deff, 0xff180000, 0x00e2deff,
1799 0x001c0000, 0x0000e6ff, 0x081c0000, 0x0020e6ff, 1847 0x001c0000, 0x0000e6ff, 0x081c0000, 0x0020e6ff,
1800 0x101c0000, 0x0040e6ff, 0x181c0000, 0x0061e6ff, 1848 0x101c0000, 0x0040e6ff, 0x181c0000, 0x0061e6ff,
1801 0x201c0000, 0x0081e6ff, 0x291c0000, 0x00a1e6ff, 1849 0x201c0000, 0x0081e6ff, 0x291c0000, 0x00a1e6ff,
1802 0x311c0000, 0x00c2e6ff, 0x391c0000, 0x00e2e6ff, 1850 0x311c0000, 0x00c2e6ff, 0x391c0000, 0x00e2e6ff,
1803 0x411c0000, 0x0000eeff, 0x4a1c0000, 0x0020eeff, 1851 0x411c0000, 0x0000eeff, 0x4a1c0000, 0x0020eeff,
1804 0x521c0000, 0x0040eeff, 0x5a1c0000, 0x0061eeff, 1852 0x521c0000, 0x0040eeff, 0x5a1c0000, 0x0061eeff,
1805 0x621c0000, 0x0081eeff, 0x6a1c0000, 0x00a1eeff, 1853 0x621c0000, 0x0081eeff, 0x6a1c0000, 0x00a1eeff,
1806 0x731c0000, 0x00c2eeff, 0x7b1c0000, 0x00e2eeff, 1854 0x731c0000, 0x00c2eeff, 0x7b1c0000, 0x00e2eeff,
1807 0x831c0000, 0x0000f6ff, 0x8b1c0000, 0x0020f6ff, 1855 0x831c0000, 0x0000f6ff, 0x8b1c0000, 0x0020f6ff,
1808 0x941c0000, 0x0040f6ff, 0x9c1c0000, 0x0061f6ff, 1856 0x941c0000, 0x0040f6ff, 0x9c1c0000, 0x0061f6ff,
1809 0xa41c0000, 0x0081f6ff, 0xac1c0000, 0x00a1f6ff, 1857 0xa41c0000, 0x0081f6ff, 0xac1c0000, 0x00a1f6ff,
1810 0xb41c0000, 0x00c2f6ff, 0xbd1c0000, 0x00e2f6ff, 1858 0xb41c0000, 0x00c2f6ff, 0xbd1c0000, 0x00e2f6ff,
1811 0xc51c0000, 0x0000ffff, 0xcd1c0000, 0x0020ffff, 1859 0xc51c0000, 0x0000ffff, 0xcd1c0000, 0x0020ffff,
1812 0xd51c0000, 0x0040ffff, 0xde1c0000, 0x0061ffff, 1860 0xd51c0000, 0x0040ffff, 0xde1c0000, 0x0061ffff,
1813 0xe61c0000, 0x0081ffff, 0xee1c0000, 0x00a1ffff, 1861 0xe61c0000, 0x0081ffff, 0xee1c0000, 0x00a1ffff,
1814 0xf61c0000, 0x00c2ffff, 0xff1c0000, 0x00e2ffff 1862 0xf61c0000, 0x00c2ffff, 0xff1c0000, 0x00e2ffff
1815 }; 1863 };
1816 static void Blit_RGB565_BGRA8888(SDL_BlitInfo *info) 1864 static void
1865 Blit_RGB565_BGRA8888(SDL_BlitInfo * info)
1817 { 1866 {
1818 Blit_RGB565_32(info, RGB565_BGRA8888_LUT); 1867 Blit_RGB565_32(info, RGB565_BGRA8888_LUT);
1819 } 1868 }
1820 1869
1821 /* Special optimized blit for RGB 8-8-8 --> RGB 3-3-2 */ 1870 /* Special optimized blit for RGB 8-8-8 --> RGB 3-3-2 */
1824 dst = (((src)&0x00E00000)>>16)| \ 1873 dst = (((src)&0x00E00000)>>16)| \
1825 (((src)&0x0000E000)>>11)| \ 1874 (((src)&0x0000E000)>>11)| \
1826 (((src)&0x000000C0)>>6); \ 1875 (((src)&0x000000C0)>>6); \
1827 } 1876 }
1828 #endif 1877 #endif
1829 static void Blit_RGB888_index8_map(SDL_BlitInfo *info) 1878 static void
1879 Blit_RGB888_index8_map(SDL_BlitInfo * info)
1830 { 1880 {
1831 #ifndef USE_DUFFS_LOOP 1881 #ifndef USE_DUFFS_LOOP
1832 int c; 1882 int c;
1833 #endif 1883 #endif
1834 int Pixel; 1884 int Pixel;
1835 int width, height; 1885 int width, height;
1836 Uint32 *src; 1886 Uint32 *src;
1837 const Uint8 *map; 1887 const Uint8 *map;
1838 Uint8 *dst; 1888 Uint8 *dst;
1839 int srcskip, dstskip; 1889 int srcskip, dstskip;
1840 1890
1841 /* Set up some basic variables */ 1891 /* Set up some basic variables */
1842 width = info->d_width; 1892 width = info->d_width;
1843 height = info->d_height; 1893 height = info->d_height;
1844 src = (Uint32 *)info->s_pixels; 1894 src = (Uint32 *) info->s_pixels;
1845 srcskip = info->s_skip/4; 1895 srcskip = info->s_skip / 4;
1846 dst = info->d_pixels; 1896 dst = info->d_pixels;
1847 dstskip = info->d_skip; 1897 dstskip = info->d_skip;
1848 map = info->table; 1898 map = info->table;
1849 1899
1850 #ifdef USE_DUFFS_LOOP 1900 #ifdef USE_DUFFS_LOOP
1851 while ( height-- ) { 1901 while (height--) {
1902 /* *INDENT-OFF* */
1852 DUFFS_LOOP( 1903 DUFFS_LOOP(
1853 RGB888_RGB332(Pixel, *src); 1904 RGB888_RGB332(Pixel, *src);
1854 *dst++ = map[Pixel]; 1905 *dst++ = map[Pixel];
1855 ++src; 1906 ++src;
1856 , width); 1907 , width);
1857 src += srcskip; 1908 /* *INDENT-ON* */
1858 dst += dstskip; 1909 src += srcskip;
1859 } 1910 dst += dstskip;
1911 }
1860 #else 1912 #else
1861 while ( height-- ) { 1913 while (height--) {
1862 for ( c=width/4; c; --c ) { 1914 for (c = width / 4; c; --c) {
1863 /* Pack RGB into 8bit pixel */ 1915 /* Pack RGB into 8bit pixel */
1864 RGB888_RGB332(Pixel, *src); 1916 RGB888_RGB332(Pixel, *src);
1865 *dst++ = map[Pixel]; 1917 *dst++ = map[Pixel];
1866 ++src; 1918 ++src;
1867 RGB888_RGB332(Pixel, *src); 1919 RGB888_RGB332(Pixel, *src);
1868 *dst++ = map[Pixel]; 1920 *dst++ = map[Pixel];
1869 ++src; 1921 ++src;
1870 RGB888_RGB332(Pixel, *src); 1922 RGB888_RGB332(Pixel, *src);
1871 *dst++ = map[Pixel]; 1923 *dst++ = map[Pixel];
1872 ++src; 1924 ++src;
1873 RGB888_RGB332(Pixel, *src); 1925 RGB888_RGB332(Pixel, *src);
1874 *dst++ = map[Pixel]; 1926 *dst++ = map[Pixel];
1875 ++src; 1927 ++src;
1876 } 1928 }
1877 switch ( width & 3 ) { 1929 switch (width & 3) {
1878 case 3: 1930 case 3:
1879 RGB888_RGB332(Pixel, *src); 1931 RGB888_RGB332(Pixel, *src);
1880 *dst++ = map[Pixel]; 1932 *dst++ = map[Pixel];
1881 ++src; 1933 ++src;
1882 case 2: 1934 case 2:
1883 RGB888_RGB332(Pixel, *src); 1935 RGB888_RGB332(Pixel, *src);
1884 *dst++ = map[Pixel]; 1936 *dst++ = map[Pixel];
1885 ++src; 1937 ++src;
1886 case 1: 1938 case 1:
1887 RGB888_RGB332(Pixel, *src); 1939 RGB888_RGB332(Pixel, *src);
1888 *dst++ = map[Pixel]; 1940 *dst++ = map[Pixel];
1889 ++src; 1941 ++src;
1890 } 1942 }
1891 src += srcskip; 1943 src += srcskip;
1892 dst += dstskip; 1944 dst += dstskip;
1893 } 1945 }
1894 #endif /* USE_DUFFS_LOOP */ 1946 #endif /* USE_DUFFS_LOOP */
1895 } 1947 }
1896 static void BlitNto1(SDL_BlitInfo *info) 1948 static void
1949 BlitNto1(SDL_BlitInfo * info)
1897 { 1950 {
1898 #ifndef USE_DUFFS_LOOP 1951 #ifndef USE_DUFFS_LOOP
1899 int c; 1952 int c;
1900 #endif 1953 #endif
1901 int width, height; 1954 int width, height;
1902 Uint8 *src; 1955 Uint8 *src;
1903 const Uint8 *map; 1956 const Uint8 *map;
1904 Uint8 *dst; 1957 Uint8 *dst;
1905 int srcskip, dstskip; 1958 int srcskip, dstskip;
1906 int srcbpp; 1959 int srcbpp;
1907 Uint32 Pixel; 1960 Uint32 Pixel;
1908 int sR, sG, sB; 1961 int sR, sG, sB;
1909 SDL_PixelFormat *srcfmt; 1962 SDL_PixelFormat *srcfmt;
1910 1963
1911 /* Set up some basic variables */ 1964 /* Set up some basic variables */
1912 width = info->d_width; 1965 width = info->d_width;
1913 height = info->d_height; 1966 height = info->d_height;
1914 src = info->s_pixels; 1967 src = info->s_pixels;
1915 srcskip = info->s_skip; 1968 srcskip = info->s_skip;
1916 dst = info->d_pixels; 1969 dst = info->d_pixels;
1917 dstskip = info->d_skip; 1970 dstskip = info->d_skip;
1918 map = info->table; 1971 map = info->table;
1919 srcfmt = info->src; 1972 srcfmt = info->src;
1920 srcbpp = srcfmt->BytesPerPixel; 1973 srcbpp = srcfmt->BytesPerPixel;
1921 1974
1922 if ( map == NULL ) { 1975 if (map == NULL) {
1923 while ( height-- ) { 1976 while (height--) {
1924 #ifdef USE_DUFFS_LOOP 1977 #ifdef USE_DUFFS_LOOP
1978 /* *INDENT-OFF* */
1925 DUFFS_LOOP( 1979 DUFFS_LOOP(
1926 DISEMBLE_RGB(src, srcbpp, srcfmt, Pixel, 1980 DISEMBLE_RGB(src, srcbpp, srcfmt, Pixel,
1927 sR, sG, sB); 1981 sR, sG, sB);
1928 if ( 1 ) { 1982 if ( 1 ) {
1929 /* Pack RGB into 8bit pixel */ 1983 /* Pack RGB into 8bit pixel */
1932 ((sB>>6)<<(0)) ; 1986 ((sB>>6)<<(0)) ;
1933 } 1987 }
1934 dst++; 1988 dst++;
1935 src += srcbpp; 1989 src += srcbpp;
1936 , width); 1990 , width);
1991 /* *INDENT-ON* */
1937 #else 1992 #else
1938 for ( c=width; c; --c ) { 1993 for (c = width; c; --c) {
1939 DISEMBLE_RGB(src, srcbpp, srcfmt, Pixel, 1994 DISEMBLE_RGB(src, srcbpp, srcfmt, Pixel, sR, sG, sB);
1940 sR, sG, sB); 1995 if (1) {
1941 if ( 1 ) { 1996 /* Pack RGB into 8bit pixel */
1942 /* Pack RGB into 8bit pixel */ 1997 *dst = ((sR >> 5) << (3 + 2)) |
1943 *dst = ((sR>>5)<<(3+2))| 1998 ((sG >> 5) << (2)) | ((sB >> 6) << (0));
1944 ((sG>>5)<<(2)) | 1999 }
1945 ((sB>>6)<<(0)) ; 2000 dst++;
1946 } 2001 src += srcbpp;
1947 dst++; 2002 }
1948 src += srcbpp;
1949 }
1950 #endif 2003 #endif
1951 src += srcskip; 2004 src += srcskip;
1952 dst += dstskip; 2005 dst += dstskip;
1953 } 2006 }
1954 } else { 2007 } else {
1955 while ( height-- ) { 2008 while (height--) {
1956 #ifdef USE_DUFFS_LOOP 2009 #ifdef USE_DUFFS_LOOP
2010 /* *INDENT-OFF* */
1957 DUFFS_LOOP( 2011 DUFFS_LOOP(
1958 DISEMBLE_RGB(src, srcbpp, srcfmt, Pixel, 2012 DISEMBLE_RGB(src, srcbpp, srcfmt, Pixel,
1959 sR, sG, sB); 2013 sR, sG, sB);
1960 if ( 1 ) { 2014 if ( 1 ) {
1961 /* Pack RGB into 8bit pixel */ 2015 /* Pack RGB into 8bit pixel */
1964 ((sB>>6)<<(0)) ]; 2018 ((sB>>6)<<(0)) ];
1965 } 2019 }
1966 dst++; 2020 dst++;
1967 src += srcbpp; 2021 src += srcbpp;
1968 , width); 2022 , width);
2023 /* *INDENT-ON* */
1969 #else 2024 #else
1970 for ( c=width; c; --c ) { 2025 for (c = width; c; --c) {
1971 DISEMBLE_RGB(src, srcbpp, srcfmt, Pixel, 2026 DISEMBLE_RGB(src, srcbpp, srcfmt, Pixel, sR, sG, sB);
1972 sR, sG, sB); 2027 if (1) {
1973 if ( 1 ) { 2028 /* Pack RGB into 8bit pixel */
1974 /* Pack RGB into 8bit pixel */ 2029 *dst = map[((sR >> 5) << (3 + 2)) |
1975 *dst = map[((sR>>5)<<(3+2))| 2030 ((sG >> 5) << (2)) | ((sB >> 6) << (0))];
1976 ((sG>>5)<<(2)) | 2031 }
1977 ((sB>>6)<<(0)) ]; 2032 dst++;
1978 } 2033 src += srcbpp;
1979 dst++; 2034 }
1980 src += srcbpp;
1981 }
1982 #endif /* USE_DUFFS_LOOP */ 2035 #endif /* USE_DUFFS_LOOP */
1983 src += srcskip; 2036 src += srcskip;
1984 dst += dstskip; 2037 dst += dstskip;
1985 } 2038 }
1986 } 2039 }
1987 } 2040 }
1988 2041
1989 /* blits 32 bit RGB<->RGBA with both surfaces having the same R,G,B fields */ 2042 /* blits 32 bit RGB<->RGBA with both surfaces having the same R,G,B fields */
1990 static void Blit4to4MaskAlpha(SDL_BlitInfo *info) 2043 static void
1991 { 2044 Blit4to4MaskAlpha(SDL_BlitInfo * info)
1992 int width = info->d_width; 2045 {
1993 int height = info->d_height; 2046 int width = info->d_width;
1994 Uint32 *src = (Uint32 *)info->s_pixels; 2047 int height = info->d_height;
1995 int srcskip = info->s_skip; 2048 Uint32 *src = (Uint32 *) info->s_pixels;
1996 Uint32 *dst = (Uint32 *)info->d_pixels; 2049 int srcskip = info->s_skip;
1997 int dstskip = info->d_skip; 2050 Uint32 *dst = (Uint32 *) info->d_pixels;
1998 SDL_PixelFormat *srcfmt = info->src; 2051 int dstskip = info->d_skip;
1999 SDL_PixelFormat *dstfmt = info->dst; 2052 SDL_PixelFormat *srcfmt = info->src;
2000 2053 SDL_PixelFormat *dstfmt = info->dst;
2001 if (dstfmt->Amask) { 2054
2002 /* RGB->RGBA, SET_ALPHA */ 2055 if (dstfmt->Amask) {
2003 Uint32 mask = (srcfmt->alpha >> dstfmt->Aloss) << dstfmt->Ashift; 2056 /* RGB->RGBA, SET_ALPHA */
2004 2057 Uint32 mask = (srcfmt->alpha >> dstfmt->Aloss) << dstfmt->Ashift;
2005 while ( height-- ) { 2058
2059 while (height--) {
2060 /* *INDENT-OFF* */
2006 DUFFS_LOOP( 2061 DUFFS_LOOP(
2007 { 2062 {
2008 *dst = *src | mask; 2063 *dst = *src | mask;
2009 ++dst; 2064 ++dst;
2010 ++src; 2065 ++src;
2011 }, 2066 },
2012 width); 2067 width);
2013 src = (Uint32*)((Uint8*)src + srcskip); 2068 /* *INDENT-ON* */
2014 dst = (Uint32*)((Uint8*)dst + dstskip); 2069 src = (Uint32 *) ((Uint8 *) src + srcskip);
2015 } 2070 dst = (Uint32 *) ((Uint8 *) dst + dstskip);
2016 } else { 2071 }
2017 /* RGBA->RGB, NO_ALPHA */ 2072 } else {
2018 Uint32 mask = srcfmt->Rmask | srcfmt->Gmask | srcfmt->Bmask; 2073 /* RGBA->RGB, NO_ALPHA */
2019 2074 Uint32 mask = srcfmt->Rmask | srcfmt->Gmask | srcfmt->Bmask;
2020 while ( height-- ) { 2075
2076 while (height--) {
2077 /* *INDENT-OFF* */
2021 DUFFS_LOOP( 2078 DUFFS_LOOP(
2022 { 2079 {
2023 *dst = *src & mask; 2080 *dst = *src & mask;
2024 ++dst; 2081 ++dst;
2025 ++src; 2082 ++src;
2026 }, 2083 },
2027 width); 2084 width);
2028 src = (Uint32*)((Uint8*)src + srcskip); 2085 /* *INDENT-ON* */
2029 dst = (Uint32*)((Uint8*)dst + dstskip); 2086 src = (Uint32 *) ((Uint8 *) src + srcskip);
2030 } 2087 dst = (Uint32 *) ((Uint8 *) dst + dstskip);
2031 } 2088 }
2032 } 2089 }
2033 2090 }
2034 static void BlitNtoN(SDL_BlitInfo *info) 2091
2035 { 2092 static void
2036 int width = info->d_width; 2093 BlitNtoN(SDL_BlitInfo * info)
2037 int height = info->d_height; 2094 {
2038 Uint8 *src = info->s_pixels; 2095 int width = info->d_width;
2039 int srcskip = info->s_skip; 2096 int height = info->d_height;
2040 Uint8 *dst = info->d_pixels; 2097 Uint8 *src = info->s_pixels;
2041 int dstskip = info->d_skip; 2098 int srcskip = info->s_skip;
2042 SDL_PixelFormat *srcfmt = info->src; 2099 Uint8 *dst = info->d_pixels;
2043 int srcbpp = srcfmt->BytesPerPixel; 2100 int dstskip = info->d_skip;
2044 SDL_PixelFormat *dstfmt = info->dst; 2101 SDL_PixelFormat *srcfmt = info->src;
2045 int dstbpp = dstfmt->BytesPerPixel; 2102 int srcbpp = srcfmt->BytesPerPixel;
2046 unsigned alpha = dstfmt->Amask ? srcfmt->alpha : 0; 2103 SDL_PixelFormat *dstfmt = info->dst;
2047 2104 int dstbpp = dstfmt->BytesPerPixel;
2048 while ( height-- ) { 2105 unsigned alpha = dstfmt->Amask ? srcfmt->alpha : 0;
2106
2107 while (height--) {
2108 /* *INDENT-OFF* */
2049 DUFFS_LOOP( 2109 DUFFS_LOOP(
2050 { 2110 {
2051 Uint32 Pixel; 2111 Uint32 Pixel;
2052 unsigned sR; 2112 unsigned sR;
2053 unsigned sG; 2113 unsigned sG;
2056 ASSEMBLE_RGBA(dst, dstbpp, dstfmt, sR, sG, sB, alpha); 2116 ASSEMBLE_RGBA(dst, dstbpp, dstfmt, sR, sG, sB, alpha);
2057 dst += dstbpp; 2117 dst += dstbpp;
2058 src += srcbpp; 2118 src += srcbpp;
2059 }, 2119 },
2060 width); 2120 width);
2061 src += srcskip; 2121 /* *INDENT-ON* */
2062 dst += dstskip; 2122 src += srcskip;
2063 } 2123 dst += dstskip;
2064 } 2124 }
2065 2125 }
2066 static void BlitNtoNCopyAlpha(SDL_BlitInfo *info) 2126
2067 { 2127 static void
2068 int width = info->d_width; 2128 BlitNtoNCopyAlpha(SDL_BlitInfo * info)
2069 int height = info->d_height; 2129 {
2070 Uint8 *src = info->s_pixels; 2130 int width = info->d_width;
2071 int srcskip = info->s_skip; 2131 int height = info->d_height;
2072 Uint8 *dst = info->d_pixels; 2132 Uint8 *src = info->s_pixels;
2073 int dstskip = info->d_skip; 2133 int srcskip = info->s_skip;
2074 SDL_PixelFormat *srcfmt = info->src; 2134 Uint8 *dst = info->d_pixels;
2075 int srcbpp = srcfmt->BytesPerPixel; 2135 int dstskip = info->d_skip;
2076 SDL_PixelFormat *dstfmt = info->dst; 2136 SDL_PixelFormat *srcfmt = info->src;
2077 int dstbpp = dstfmt->BytesPerPixel; 2137 int srcbpp = srcfmt->BytesPerPixel;
2078 int c; 2138 SDL_PixelFormat *dstfmt = info->dst;
2079 2139 int dstbpp = dstfmt->BytesPerPixel;
2080 /* FIXME: should map alpha to [0..255] correctly! */ 2140 int c;
2081 while ( height-- ) { 2141
2082 for ( c=width; c; --c ) { 2142 /* FIXME: should map alpha to [0..255] correctly! */
2083 Uint32 Pixel; 2143 while (height--) {
2084 unsigned sR, sG, sB, sA; 2144 for (c = width; c; --c) {
2085 DISEMBLE_RGBA(src, srcbpp, srcfmt, Pixel, 2145 Uint32 Pixel;
2086 sR, sG, sB, sA); 2146 unsigned sR, sG, sB, sA;
2087 ASSEMBLE_RGBA(dst, dstbpp, dstfmt, 2147 DISEMBLE_RGBA(src, srcbpp, srcfmt, Pixel, sR, sG, sB, sA);
2088 sR, sG, sB, sA); 2148 ASSEMBLE_RGBA(dst, dstbpp, dstfmt, sR, sG, sB, sA);
2089 dst += dstbpp; 2149 dst += dstbpp;
2090 src += srcbpp; 2150 src += srcbpp;
2091 } 2151 }
2092 src += srcskip; 2152 src += srcskip;
2093 dst += dstskip; 2153 dst += dstskip;
2094 } 2154 }
2095 } 2155 }
2096 2156
2097 static void BlitNto1Key(SDL_BlitInfo *info) 2157 static void
2098 { 2158 BlitNto1Key(SDL_BlitInfo * info)
2099 int width = info->d_width; 2159 {
2100 int height = info->d_height; 2160 int width = info->d_width;
2101 Uint8 *src = info->s_pixels; 2161 int height = info->d_height;
2102 int srcskip = info->s_skip; 2162 Uint8 *src = info->s_pixels;
2103 Uint8 *dst = info->d_pixels; 2163 int srcskip = info->s_skip;
2104 int dstskip = info->d_skip; 2164 Uint8 *dst = info->d_pixels;
2105 SDL_PixelFormat *srcfmt = info->src; 2165 int dstskip = info->d_skip;
2106 const Uint8 *palmap = info->table; 2166 SDL_PixelFormat *srcfmt = info->src;
2107 Uint32 ckey = srcfmt->colorkey; 2167 const Uint8 *palmap = info->table;
2108 Uint32 rgbmask = ~srcfmt->Amask; 2168 Uint32 ckey = srcfmt->colorkey;
2109 int srcbpp; 2169 Uint32 rgbmask = ~srcfmt->Amask;
2110 Uint32 Pixel; 2170 int srcbpp;
2111 unsigned sR, sG, sB; 2171 Uint32 Pixel;
2112 2172 unsigned sR, sG, sB;
2113 /* Set up some basic variables */ 2173
2114 srcbpp = srcfmt->BytesPerPixel; 2174 /* Set up some basic variables */
2115 ckey &= rgbmask; 2175 srcbpp = srcfmt->BytesPerPixel;
2116 2176 ckey &= rgbmask;
2117 if ( palmap == NULL ) { 2177
2118 while ( height-- ) { 2178 if (palmap == NULL) {
2179 while (height--) {
2180 /* *INDENT-OFF* */
2119 DUFFS_LOOP( 2181 DUFFS_LOOP(
2120 { 2182 {
2121 DISEMBLE_RGB(src, srcbpp, srcfmt, Pixel, 2183 DISEMBLE_RGB(src, srcbpp, srcfmt, Pixel,
2122 sR, sG, sB); 2184 sR, sG, sB);
2123 if ( (Pixel & rgbmask) != ckey ) { 2185 if ( (Pixel & rgbmask) != ckey ) {
2128 } 2190 }
2129 dst++; 2191 dst++;
2130 src += srcbpp; 2192 src += srcbpp;
2131 }, 2193 },
2132 width); 2194 width);
2133 src += srcskip; 2195 /* *INDENT-ON* */
2134 dst += dstskip; 2196 src += srcskip;
2135 } 2197 dst += dstskip;
2136 } else { 2198 }
2137 while ( height-- ) { 2199 } else {
2200 while (height--) {
2201 /* *INDENT-OFF* */
2138 DUFFS_LOOP( 2202 DUFFS_LOOP(
2139 { 2203 {
2140 DISEMBLE_RGB(src, srcbpp, srcfmt, Pixel, 2204 DISEMBLE_RGB(src, srcbpp, srcfmt, Pixel,
2141 sR, sG, sB); 2205 sR, sG, sB);
2142 if ( (Pixel & rgbmask) != ckey ) { 2206 if ( (Pixel & rgbmask) != ckey ) {
2147 } 2211 }
2148 dst++; 2212 dst++;
2149 src += srcbpp; 2213 src += srcbpp;
2150 }, 2214 },
2151 width); 2215 width);
2152 src += srcskip; 2216 /* *INDENT-ON* */
2153 dst += dstskip; 2217 src += srcskip;
2154 } 2218 dst += dstskip;
2155 } 2219 }
2156 } 2220 }
2157 2221 }
2158 static void Blit2to2Key(SDL_BlitInfo *info) 2222
2159 { 2223 static void
2160 int width = info->d_width; 2224 Blit2to2Key(SDL_BlitInfo * info)
2161 int height = info->d_height; 2225 {
2162 Uint16 *srcp = (Uint16 *)info->s_pixels; 2226 int width = info->d_width;
2163 int srcskip = info->s_skip; 2227 int height = info->d_height;
2164 Uint16 *dstp = (Uint16 *)info->d_pixels; 2228 Uint16 *srcp = (Uint16 *) info->s_pixels;
2165 int dstskip = info->d_skip; 2229 int srcskip = info->s_skip;
2166 Uint32 ckey = info->src->colorkey; 2230 Uint16 *dstp = (Uint16 *) info->d_pixels;
2167 Uint32 rgbmask = ~info->src->Amask; 2231 int dstskip = info->d_skip;
2168 2232 Uint32 ckey = info->src->colorkey;
2169 /* Set up some basic variables */ 2233 Uint32 rgbmask = ~info->src->Amask;
2170 srcskip /= 2; 2234
2171 dstskip /= 2; 2235 /* Set up some basic variables */
2172 ckey &= rgbmask; 2236 srcskip /= 2;
2173 2237 dstskip /= 2;
2174 while ( height-- ) { 2238 ckey &= rgbmask;
2239
2240 while (height--) {
2241 /* *INDENT-OFF* */
2175 DUFFS_LOOP( 2242 DUFFS_LOOP(
2176 { 2243 {
2177 if ( (*srcp & rgbmask) != ckey ) { 2244 if ( (*srcp & rgbmask) != ckey ) {
2178 *dstp = *srcp; 2245 *dstp = *srcp;
2179 } 2246 }
2180 dstp++; 2247 dstp++;
2181 srcp++; 2248 srcp++;
2182 }, 2249 },
2183 width); 2250 width);
2184 srcp += srcskip; 2251 /* *INDENT-ON* */
2185 dstp += dstskip; 2252 srcp += srcskip;
2186 } 2253 dstp += dstskip;
2187 } 2254 }
2188 2255 }
2189 static void BlitNtoNKey(SDL_BlitInfo *info) 2256
2190 { 2257 static void
2191 int width = info->d_width; 2258 BlitNtoNKey(SDL_BlitInfo * info)
2192 int height = info->d_height; 2259 {
2193 Uint8 *src = info->s_pixels; 2260 int width = info->d_width;
2194 int srcskip = info->s_skip; 2261 int height = info->d_height;
2195 Uint8 *dst = info->d_pixels; 2262 Uint8 *src = info->s_pixels;
2196 int dstskip = info->d_skip; 2263 int srcskip = info->s_skip;
2197 Uint32 ckey = info->src->colorkey; 2264 Uint8 *dst = info->d_pixels;
2198 SDL_PixelFormat *srcfmt = info->src; 2265 int dstskip = info->d_skip;
2199 SDL_PixelFormat *dstfmt = info->dst; 2266 Uint32 ckey = info->src->colorkey;
2200 int srcbpp = srcfmt->BytesPerPixel; 2267 SDL_PixelFormat *srcfmt = info->src;
2201 int dstbpp = dstfmt->BytesPerPixel; 2268 SDL_PixelFormat *dstfmt = info->dst;
2202 unsigned alpha = dstfmt->Amask ? srcfmt->alpha : 0; 2269 int srcbpp = srcfmt->BytesPerPixel;
2203 Uint32 rgbmask = ~srcfmt->Amask; 2270 int dstbpp = dstfmt->BytesPerPixel;
2204 2271 unsigned alpha = dstfmt->Amask ? srcfmt->alpha : 0;
2205 /* Set up some basic variables */ 2272 Uint32 rgbmask = ~srcfmt->Amask;
2206 ckey &= rgbmask; 2273
2207 2274 /* Set up some basic variables */
2208 while ( height-- ) { 2275 ckey &= rgbmask;
2276
2277 while (height--) {
2278 /* *INDENT-OFF* */
2209 DUFFS_LOOP( 2279 DUFFS_LOOP(
2210 { 2280 {
2211 Uint32 Pixel; 2281 Uint32 Pixel;
2212 unsigned sR; 2282 unsigned sR;
2213 unsigned sG; 2283 unsigned sG;
2220 } 2290 }
2221 dst += dstbpp; 2291 dst += dstbpp;
2222 src += srcbpp; 2292 src += srcbpp;
2223 }, 2293 },
2224 width); 2294 width);
2225 src += srcskip; 2295 /* *INDENT-ON* */
2226 dst += dstskip; 2296 src += srcskip;
2227 } 2297 dst += dstskip;
2228 } 2298 }
2229 2299 }
2230 static void BlitNtoNKeyCopyAlpha(SDL_BlitInfo *info) 2300
2231 { 2301 static void
2232 int width = info->d_width; 2302 BlitNtoNKeyCopyAlpha(SDL_BlitInfo * info)
2233 int height = info->d_height; 2303 {
2234 Uint8 *src = info->s_pixels; 2304 int width = info->d_width;
2235 int srcskip = info->s_skip; 2305 int height = info->d_height;
2236 Uint8 *dst = info->d_pixels; 2306 Uint8 *src = info->s_pixels;
2237 int dstskip = info->d_skip; 2307 int srcskip = info->s_skip;
2238 Uint32 ckey = info->src->colorkey; 2308 Uint8 *dst = info->d_pixels;
2239 SDL_PixelFormat *srcfmt = info->src; 2309 int dstskip = info->d_skip;
2240 SDL_PixelFormat *dstfmt = info->dst; 2310 Uint32 ckey = info->src->colorkey;
2241 Uint32 rgbmask = ~srcfmt->Amask; 2311 SDL_PixelFormat *srcfmt = info->src;
2242 2312 SDL_PixelFormat *dstfmt = info->dst;
2243 Uint8 srcbpp; 2313 Uint32 rgbmask = ~srcfmt->Amask;
2244 Uint8 dstbpp; 2314
2245 Uint32 Pixel; 2315 Uint8 srcbpp;
2246 unsigned sR, sG, sB, sA; 2316 Uint8 dstbpp;
2247 2317 Uint32 Pixel;
2248 /* Set up some basic variables */ 2318 unsigned sR, sG, sB, sA;
2249 srcbpp = srcfmt->BytesPerPixel; 2319
2250 dstbpp = dstfmt->BytesPerPixel; 2320 /* Set up some basic variables */
2251 ckey &= rgbmask; 2321 srcbpp = srcfmt->BytesPerPixel;
2252 2322 dstbpp = dstfmt->BytesPerPixel;
2253 /* FIXME: should map alpha to [0..255] correctly! */ 2323 ckey &= rgbmask;
2254 while ( height-- ) { 2324
2325 /* FIXME: should map alpha to [0..255] correctly! */
2326 while (height--) {
2327 /* *INDENT-OFF* */
2255 DUFFS_LOOP( 2328 DUFFS_LOOP(
2256 { 2329 {
2257 DISEMBLE_RGBA(src, srcbpp, srcfmt, Pixel, 2330 DISEMBLE_RGBA(src, srcbpp, srcfmt, Pixel,
2258 sR, sG, sB, sA); 2331 sR, sG, sB, sA);
2259 if ( (Pixel & rgbmask) != ckey ) { 2332 if ( (Pixel & rgbmask) != ckey ) {
2262 } 2335 }
2263 dst += dstbpp; 2336 dst += dstbpp;
2264 src += srcbpp; 2337 src += srcbpp;
2265 }, 2338 },
2266 width); 2339 width);
2267 src += srcskip; 2340 /* *INDENT-ON* */
2268 dst += dstskip; 2341 src += srcskip;
2269 } 2342 dst += dstskip;
2343 }
2270 } 2344 }
2271 2345
2272 /* Normal N to N optimized blitters */ 2346 /* Normal N to N optimized blitters */
2273 struct blit_table { 2347 struct blit_table
2274 Uint32 srcR, srcG, srcB; 2348 {
2275 int dstbpp; 2349 Uint32 srcR, srcG, srcB;
2276 Uint32 dstR, dstG, dstB; 2350 int dstbpp;
2277 Uint32 blit_features; 2351 Uint32 dstR, dstG, dstB;
2278 void *aux_data; 2352 Uint32 blit_features;
2279 SDL_loblit blitfunc; 2353 void *aux_data;
2280 enum { NO_ALPHA=1, SET_ALPHA=2, COPY_ALPHA=4 } alpha; 2354 SDL_loblit blitfunc;
2355 enum
2356 { NO_ALPHA = 1, SET_ALPHA = 2, COPY_ALPHA = 4 } alpha;
2281 }; 2357 };
2282 static const struct blit_table normal_blit_1[] = { 2358 static const struct blit_table normal_blit_1[] = {
2283 /* Default for 8-bit RGB source, an invalid combination */ 2359 /* Default for 8-bit RGB source, an invalid combination */
2284 { 0,0,0, 0, 0,0,0, 0, NULL, NULL }, 2360 {0, 0, 0, 0, 0, 0, 0, 0, NULL, NULL},
2285 }; 2361 };
2286 static const struct blit_table normal_blit_2[] = { 2362 static const struct blit_table normal_blit_2[] = {
2287 #if SDL_HERMES_BLITTERS 2363 #if SDL_HERMES_BLITTERS
2288 { 0x0000F800,0x000007E0,0x0000001F, 2, 0x0000001F,0x000007E0,0x0000F800, 2364 {0x0000F800, 0x000007E0, 0x0000001F, 2, 0x0000001F, 0x000007E0,
2289 0, ConvertX86p16_16BGR565, ConvertX86, NO_ALPHA }, 2365 0x0000F800,
2290 { 0x0000F800,0x000007E0,0x0000001F, 2, 0x00007C00,0x000003E0,0x0000001F, 2366 0, ConvertX86p16_16BGR565, ConvertX86, NO_ALPHA},
2291 0, ConvertX86p16_16RGB555, ConvertX86, NO_ALPHA }, 2367 {0x0000F800, 0x000007E0, 0x0000001F, 2, 0x00007C00, 0x000003E0,
2292 { 0x0000F800,0x000007E0,0x0000001F, 2, 0x0000001F,0x000003E0,0x00007C00, 2368 0x0000001F,
2293 0, ConvertX86p16_16BGR555, ConvertX86, NO_ALPHA }, 2369 0, ConvertX86p16_16RGB555, ConvertX86, NO_ALPHA},
2370 {0x0000F800, 0x000007E0, 0x0000001F, 2, 0x0000001F, 0x000003E0,
2371 0x00007C00,
2372 0, ConvertX86p16_16BGR555, ConvertX86, NO_ALPHA},
2294 #elif SDL_ALTIVEC_BLITTERS 2373 #elif SDL_ALTIVEC_BLITTERS
2295 /* has-altivec */ 2374 /* has-altivec */
2296 { 0x0000F800,0x000007E0,0x0000001F, 4, 0x00000000,0x00000000,0x00000000, 2375 {0x0000F800, 0x000007E0, 0x0000001F, 4, 0x00000000, 0x00000000,
2297 2, NULL, Blit_RGB565_32Altivec, NO_ALPHA | COPY_ALPHA | SET_ALPHA }, 2376 0x00000000,
2298 { 0x00007C00,0x000003E0,0x0000001F, 4, 0x00000000,0x00000000,0x00000000, 2377 2, NULL, Blit_RGB565_32Altivec, NO_ALPHA | COPY_ALPHA | SET_ALPHA},
2299 2, NULL, Blit_RGB555_32Altivec, NO_ALPHA | COPY_ALPHA | SET_ALPHA }, 2378 {0x00007C00, 0x000003E0, 0x0000001F, 4, 0x00000000, 0x00000000,
2379 0x00000000,
2380 2, NULL, Blit_RGB555_32Altivec, NO_ALPHA | COPY_ALPHA | SET_ALPHA},
2300 #endif 2381 #endif
2301 { 0x0000F800,0x000007E0,0x0000001F, 4, 0x00FF0000,0x0000FF00,0x000000FF, 2382 {0x0000F800, 0x000007E0, 0x0000001F, 4, 0x00FF0000, 0x0000FF00,
2302 0, NULL, Blit_RGB565_ARGB8888, SET_ALPHA }, 2383 0x000000FF,
2303 { 0x0000F800,0x000007E0,0x0000001F, 4, 0x000000FF,0x0000FF00,0x00FF0000, 2384 0, NULL, Blit_RGB565_ARGB8888, SET_ALPHA},
2304 0, NULL, Blit_RGB565_ABGR8888, SET_ALPHA }, 2385 {0x0000F800, 0x000007E0, 0x0000001F, 4, 0x000000FF, 0x0000FF00,
2305 { 0x0000F800,0x000007E0,0x0000001F, 4, 0xFF000000,0x00FF0000,0x0000FF00, 2386 0x00FF0000,
2306 0, NULL, Blit_RGB565_RGBA8888, SET_ALPHA }, 2387 0, NULL, Blit_RGB565_ABGR8888, SET_ALPHA},
2307 { 0x0000F800,0x000007E0,0x0000001F, 4, 0x0000FF00,0x00FF0000,0xFF000000, 2388 {0x0000F800, 0x000007E0, 0x0000001F, 4, 0xFF000000, 0x00FF0000,
2308 0, NULL, Blit_RGB565_BGRA8888, SET_ALPHA }, 2389 0x0000FF00,
2390 0, NULL, Blit_RGB565_RGBA8888, SET_ALPHA},
2391 {0x0000F800, 0x000007E0, 0x0000001F, 4, 0x0000FF00, 0x00FF0000,
2392 0xFF000000,
2393 0, NULL, Blit_RGB565_BGRA8888, SET_ALPHA},
2309 2394
2310 /* Default for 16-bit RGB source, used if no other blitter matches */ 2395 /* Default for 16-bit RGB source, used if no other blitter matches */
2311 { 0,0,0, 0, 0,0,0, 0, NULL, BlitNtoN, 0 } 2396 {0, 0, 0, 0, 0, 0, 0, 0, NULL, BlitNtoN, 0}
2312 }; 2397 };
2313 static const struct blit_table normal_blit_3[] = { 2398 static const struct blit_table normal_blit_3[] = {
2314 /* Default for 24-bit RGB source, never optimized */ 2399 /* Default for 24-bit RGB source, never optimized */
2315 { 0,0,0, 0, 0,0,0, 0, NULL, BlitNtoN, 0 } 2400 {0, 0, 0, 0, 0, 0, 0, 0, NULL, BlitNtoN, 0}
2316 }; 2401 };
2317 static const struct blit_table normal_blit_4[] = { 2402 static const struct blit_table normal_blit_4[] = {
2318 #if SDL_HERMES_BLITTERS 2403 #if SDL_HERMES_BLITTERS
2319 { 0x00FF0000,0x0000FF00,0x000000FF, 2, 0x0000F800,0x000007E0,0x0000001F, 2404 {0x00FF0000, 0x0000FF00, 0x000000FF, 2, 0x0000F800, 0x000007E0,
2320 1, ConvertMMXpII32_16RGB565, ConvertMMX, NO_ALPHA }, 2405 0x0000001F,
2321 { 0x00FF0000,0x0000FF00,0x000000FF, 2, 0x0000F800,0x000007E0,0x0000001F, 2406 1, ConvertMMXpII32_16RGB565, ConvertMMX, NO_ALPHA},
2322 0, ConvertX86p32_16RGB565, ConvertX86, NO_ALPHA }, 2407 {0x00FF0000, 0x0000FF00, 0x000000FF, 2, 0x0000F800, 0x000007E0,
2323 { 0x00FF0000,0x0000FF00,0x000000FF, 2, 0x0000001F,0x000007E0,0x0000F800, 2408 0x0000001F,
2324 1, ConvertMMXpII32_16BGR565, ConvertMMX, NO_ALPHA }, 2409 0, ConvertX86p32_16RGB565, ConvertX86, NO_ALPHA},
2325 { 0x00FF0000,0x0000FF00,0x000000FF, 2, 0x0000001F,0x000007E0,0x0000F800, 2410 {0x00FF0000, 0x0000FF00, 0x000000FF, 2, 0x0000001F, 0x000007E0,
2326 0, ConvertX86p32_16BGR565, ConvertX86, NO_ALPHA }, 2411 0x0000F800,
2327 { 0x00FF0000,0x0000FF00,0x000000FF, 2, 0x00007C00,0x000003E0,0x0000001F, 2412 1, ConvertMMXpII32_16BGR565, ConvertMMX, NO_ALPHA},
2328 1, ConvertMMXpII32_16RGB555, ConvertMMX, NO_ALPHA }, 2413 {0x00FF0000, 0x0000FF00, 0x000000FF, 2, 0x0000001F, 0x000007E0,
2329 { 0x00FF0000,0x0000FF00,0x000000FF, 2, 0x00007C00,0x000003E0,0x0000001F, 2414 0x0000F800,
2330 0, ConvertX86p32_16RGB555, ConvertX86, NO_ALPHA }, 2415 0, ConvertX86p32_16BGR565, ConvertX86, NO_ALPHA},
2331 { 0x00FF0000,0x0000FF00,0x000000FF, 2, 0x0000001F,0x000003E0,0x00007C00, 2416 {0x00FF0000, 0x0000FF00, 0x000000FF, 2, 0x00007C00, 0x000003E0,
2332 1, ConvertMMXpII32_16BGR555, ConvertMMX, NO_ALPHA }, 2417 0x0000001F,
2333 { 0x00FF0000,0x0000FF00,0x000000FF, 2, 0x0000001F,0x000003E0,0x00007C00, 2418 1, ConvertMMXpII32_16RGB555, ConvertMMX, NO_ALPHA},
2334 0, ConvertX86p32_16BGR555, ConvertX86, NO_ALPHA }, 2419 {0x00FF0000, 0x0000FF00, 0x000000FF, 2, 0x00007C00, 0x000003E0,
2335 { 0x00FF0000,0x0000FF00,0x000000FF, 3, 0x00FF0000,0x0000FF00,0x000000FF, 2420 0x0000001F,
2336 0, ConvertX86p32_24RGB888, ConvertX86, NO_ALPHA }, 2421 0, ConvertX86p32_16RGB555, ConvertX86, NO_ALPHA},
2337 { 0x00FF0000,0x0000FF00,0x000000FF, 3, 0x000000FF,0x0000FF00,0x00FF0000, 2422 {0x00FF0000, 0x0000FF00, 0x000000FF, 2, 0x0000001F, 0x000003E0,
2338 0, ConvertX86p32_24BGR888, ConvertX86, NO_ALPHA }, 2423 0x00007C00,
2339 { 0x00FF0000,0x0000FF00,0x000000FF, 4, 0x000000FF,0x0000FF00,0x00FF0000, 2424 1, ConvertMMXpII32_16BGR555, ConvertMMX, NO_ALPHA},
2340 0, ConvertX86p32_32BGR888, ConvertX86, NO_ALPHA }, 2425 {0x00FF0000, 0x0000FF00, 0x000000FF, 2, 0x0000001F, 0x000003E0,
2341 { 0x00FF0000,0x0000FF00,0x000000FF, 4, 0xFF000000,0x00FF0000,0x0000FF00, 2426 0x00007C00,
2342 0, ConvertX86p32_32RGBA888, ConvertX86, NO_ALPHA }, 2427 0, ConvertX86p32_16BGR555, ConvertX86, NO_ALPHA},
2343 { 0x00FF0000,0x0000FF00,0x000000FF, 4, 0x0000FF00,0x00FF0000,0xFF000000, 2428 {0x00FF0000, 0x0000FF00, 0x000000FF, 3, 0x00FF0000, 0x0000FF00,
2344 0, ConvertX86p32_32BGRA888, ConvertX86, NO_ALPHA }, 2429 0x000000FF,
2430 0, ConvertX86p32_24RGB888, ConvertX86, NO_ALPHA},
2431 {0x00FF0000, 0x0000FF00, 0x000000FF, 3, 0x000000FF, 0x0000FF00,
2432 0x00FF0000,
2433 0, ConvertX86p32_24BGR888, ConvertX86, NO_ALPHA},
2434 {0x00FF0000, 0x0000FF00, 0x000000FF, 4, 0x000000FF, 0x0000FF00,
2435 0x00FF0000,
2436 0, ConvertX86p32_32BGR888, ConvertX86, NO_ALPHA},
2437 {0x00FF0000, 0x0000FF00, 0x000000FF, 4, 0xFF000000, 0x00FF0000,
2438 0x0000FF00,
2439 0, ConvertX86p32_32RGBA888, ConvertX86, NO_ALPHA},
2440 {0x00FF0000, 0x0000FF00, 0x000000FF, 4, 0x0000FF00, 0x00FF0000,
2441 0xFF000000,
2442 0, ConvertX86p32_32BGRA888, ConvertX86, NO_ALPHA},
2345 #else 2443 #else
2346 #if SDL_ALTIVEC_BLITTERS 2444 #if SDL_ALTIVEC_BLITTERS
2347 /* has-altivec | dont-use-prefetch */ 2445 /* has-altivec | dont-use-prefetch */
2348 { 0x00000000,0x00000000,0x00000000, 4, 0x00000000,0x00000000,0x00000000, 2446 {0x00000000, 0x00000000, 0x00000000, 4, 0x00000000, 0x00000000,
2349 6, NULL, ConvertAltivec32to32_noprefetch, NO_ALPHA | COPY_ALPHA | SET_ALPHA }, 2447 0x00000000,
2448 6, NULL, ConvertAltivec32to32_noprefetch,
2449 NO_ALPHA | COPY_ALPHA | SET_ALPHA},
2350 /* has-altivec */ 2450 /* has-altivec */
2351 { 0x00000000,0x00000000,0x00000000, 4, 0x00000000,0x00000000,0x00000000, 2451 {0x00000000, 0x00000000, 0x00000000, 4, 0x00000000, 0x00000000,
2352 2, NULL, ConvertAltivec32to32_prefetch, NO_ALPHA | COPY_ALPHA | SET_ALPHA }, 2452 0x00000000,
2453 2, NULL, ConvertAltivec32to32_prefetch,
2454 NO_ALPHA | COPY_ALPHA | SET_ALPHA},
2353 /* has-altivec */ 2455 /* has-altivec */
2354 { 0x00000000,0x00000000,0x00000000, 2, 0x0000F800,0x000007E0,0x0000001F, 2456 {0x00000000, 0x00000000, 0x00000000, 2, 0x0000F800, 0x000007E0,
2355 2, NULL, Blit_RGB888_RGB565Altivec, NO_ALPHA }, 2457 0x0000001F,
2458 2, NULL, Blit_RGB888_RGB565Altivec, NO_ALPHA},
2356 #endif 2459 #endif
2357 { 0x00FF0000,0x0000FF00,0x000000FF, 2, 0x0000F800,0x000007E0,0x0000001F, 2460 {0x00FF0000, 0x0000FF00, 0x000000FF, 2, 0x0000F800, 0x000007E0,
2358 0, NULL, Blit_RGB888_RGB565, NO_ALPHA }, 2461 0x0000001F,
2359 { 0x00FF0000,0x0000FF00,0x000000FF, 2, 0x00007C00,0x000003E0,0x0000001F, 2462 0, NULL, Blit_RGB888_RGB565, NO_ALPHA},
2360 0, NULL, Blit_RGB888_RGB555, NO_ALPHA }, 2463 {0x00FF0000, 0x0000FF00, 0x000000FF, 2, 0x00007C00, 0x000003E0,
2464 0x0000001F,
2465 0, NULL, Blit_RGB888_RGB555, NO_ALPHA},
2361 #endif 2466 #endif
2362 /* Default for 32-bit RGB source, used if no other blitter matches */ 2467 /* Default for 32-bit RGB source, used if no other blitter matches */
2363 { 0,0,0, 0, 0,0,0, 0, NULL, BlitNtoN, 0 } 2468 {0, 0, 0, 0, 0, 0, 0, 0, NULL, BlitNtoN, 0}
2364 }; 2469 };
2365 static const struct blit_table *normal_blit[] = { 2470 static const struct blit_table *normal_blit[] = {
2366 normal_blit_1, normal_blit_2, normal_blit_3, normal_blit_4 2471 normal_blit_1, normal_blit_2, normal_blit_3, normal_blit_4
2367 }; 2472 };
2368 2473
2369 /* Mask matches table, or table entry is zero */ 2474 /* Mask matches table, or table entry is zero */
2370 #define MASKOK(x, y) (((x) == (y)) || ((y) == 0x00000000)) 2475 #define MASKOK(x, y) (((x) == (y)) || ((y) == 0x00000000))
2371 2476
2372 SDL_loblit SDL_CalculateBlitN(SDL_Surface *surface, int blit_index) 2477 SDL_loblit
2373 { 2478 SDL_CalculateBlitN(SDL_Surface * surface, int blit_index)
2374 struct private_swaccel *sdata; 2479 {
2375 SDL_PixelFormat *srcfmt; 2480 struct private_swaccel *sdata;
2376 SDL_PixelFormat *dstfmt; 2481 SDL_PixelFormat *srcfmt;
2377 const struct blit_table *table; 2482 SDL_PixelFormat *dstfmt;
2378 int which; 2483 const struct blit_table *table;
2379 SDL_loblit blitfun; 2484 int which;
2380 2485 SDL_loblit blitfun;
2381 /* Set up data for choosing the blit */ 2486
2382 sdata = surface->map->sw_data; 2487 /* Set up data for choosing the blit */
2383 srcfmt = surface->format; 2488 sdata = surface->map->sw_data;
2384 dstfmt = surface->map->dst->format; 2489 srcfmt = surface->format;
2385 2490 dstfmt = surface->map->dst->format;
2386 if ( blit_index & 2 ) { 2491
2387 /* alpha or alpha+colorkey */ 2492 if (blit_index & 2) {
2388 return SDL_CalculateAlphaBlit(surface, blit_index); 2493 /* alpha or alpha+colorkey */
2389 } 2494 return SDL_CalculateAlphaBlit(surface, blit_index);
2390 2495 }
2391 /* We don't support destinations less than 8-bits */ 2496
2392 if ( dstfmt->BitsPerPixel < 8 ) { 2497 /* We don't support destinations less than 8-bits */
2393 return(NULL); 2498 if (dstfmt->BitsPerPixel < 8) {
2394 } 2499 return (NULL);
2395 2500 }
2396 if(blit_index == 1) { 2501
2397 /* colorkey blit: Here we don't have too many options, mostly 2502 if (blit_index == 1) {
2398 because RLE is the preferred fast way to deal with this. 2503 /* colorkey blit: Here we don't have too many options, mostly
2399 If a particular case turns out to be useful we'll add it. */ 2504 because RLE is the preferred fast way to deal with this.
2400 2505 If a particular case turns out to be useful we'll add it. */
2401 if(srcfmt->BytesPerPixel == 2 2506
2402 && surface->map->identity) 2507 if (srcfmt->BytesPerPixel == 2 && surface->map->identity)
2403 return Blit2to2Key; 2508 return Blit2to2Key;
2404 else if(dstfmt->BytesPerPixel == 1) 2509 else if (dstfmt->BytesPerPixel == 1)
2405 return BlitNto1Key; 2510 return BlitNto1Key;
2406 else { 2511 else {
2407 #if SDL_ALTIVEC_BLITTERS 2512 #if SDL_ALTIVEC_BLITTERS
2408 if((srcfmt->BytesPerPixel == 4) && (dstfmt->BytesPerPixel == 4) && SDL_HasAltiVec()) { 2513 if ((srcfmt->BytesPerPixel == 4) && (dstfmt->BytesPerPixel == 4)
2409 return Blit32to32KeyAltivec; 2514 && SDL_HasAltiVec()) {
2410 } else 2515 return Blit32to32KeyAltivec;
2516 } else
2411 #endif 2517 #endif
2412 2518
2413 if(srcfmt->Amask && dstfmt->Amask) 2519 if (srcfmt->Amask && dstfmt->Amask)
2414 return BlitNtoNKeyCopyAlpha; 2520 return BlitNtoNKeyCopyAlpha;
2415 else 2521 else
2416 return BlitNtoNKey; 2522 return BlitNtoNKey;
2417 } 2523 }
2418 } 2524 }
2419 2525
2420 blitfun = NULL; 2526 blitfun = NULL;
2421 if ( dstfmt->BitsPerPixel == 8 ) { 2527 if (dstfmt->BitsPerPixel == 8) {
2422 /* We assume 8-bit destinations are palettized */ 2528 /* We assume 8-bit destinations are palettized */
2423 if ( (srcfmt->BytesPerPixel == 4) && 2529 if ((srcfmt->BytesPerPixel == 4) &&
2424 (srcfmt->Rmask == 0x00FF0000) && 2530 (srcfmt->Rmask == 0x00FF0000) &&
2425 (srcfmt->Gmask == 0x0000FF00) && 2531 (srcfmt->Gmask == 0x0000FF00) && (srcfmt->Bmask == 0x000000FF)) {
2426 (srcfmt->Bmask == 0x000000FF) ) { 2532 if (surface->map->table) {
2427 if ( surface->map->table ) { 2533 blitfun = Blit_RGB888_index8_map;
2428 blitfun = Blit_RGB888_index8_map; 2534 } else {
2429 } else {
2430 #if SDL_HERMES_BLITTERS 2535 #if SDL_HERMES_BLITTERS
2431 sdata->aux_data = ConvertX86p32_8RGB332; 2536 sdata->aux_data = ConvertX86p32_8RGB332;
2432 blitfun = ConvertX86; 2537 blitfun = ConvertX86;
2433 #else 2538 #else
2434 blitfun = Blit_RGB888_index8; 2539 blitfun = Blit_RGB888_index8;
2435 #endif 2540 #endif
2436 } 2541 }
2437 } else { 2542 } else {
2438 blitfun = BlitNto1; 2543 blitfun = BlitNto1;
2439 } 2544 }
2440 } else { 2545 } else {
2441 /* Now the meat, choose the blitter we want */ 2546 /* Now the meat, choose the blitter we want */
2442 int a_need = NO_ALPHA; 2547 int a_need = NO_ALPHA;
2443 if(dstfmt->Amask) 2548 if (dstfmt->Amask)
2444 a_need = srcfmt->Amask ? COPY_ALPHA : SET_ALPHA; 2549 a_need = srcfmt->Amask ? COPY_ALPHA : SET_ALPHA;
2445 table = normal_blit[srcfmt->BytesPerPixel-1]; 2550 table = normal_blit[srcfmt->BytesPerPixel - 1];
2446 for ( which=0; table[which].dstbpp; ++which ) { 2551 for (which = 0; table[which].dstbpp; ++which) {
2447 if ( MASKOK(srcfmt->Rmask, table[which].srcR) && 2552 if (MASKOK(srcfmt->Rmask, table[which].srcR) &&
2448 MASKOK(srcfmt->Gmask, table[which].srcG) && 2553 MASKOK(srcfmt->Gmask, table[which].srcG) &&
2449 MASKOK(srcfmt->Bmask, table[which].srcB) && 2554 MASKOK(srcfmt->Bmask, table[which].srcB) &&
2450 MASKOK(dstfmt->Rmask, table[which].dstR) && 2555 MASKOK(dstfmt->Rmask, table[which].dstR) &&
2451 MASKOK(dstfmt->Gmask, table[which].dstG) && 2556 MASKOK(dstfmt->Gmask, table[which].dstG) &&
2452 MASKOK(dstfmt->Bmask, table[which].dstB) && 2557 MASKOK(dstfmt->Bmask, table[which].dstB) &&
2453 dstfmt->BytesPerPixel == table[which].dstbpp && 2558 dstfmt->BytesPerPixel == table[which].dstbpp &&
2454 (a_need & table[which].alpha) == a_need && 2559 (a_need & table[which].alpha) == a_need &&
2455 ((table[which].blit_features & GetBlitFeatures()) == table[which].blit_features) ) 2560 ((table[which].blit_features & GetBlitFeatures()) ==
2456 break; 2561 table[which].blit_features))
2457 } 2562 break;
2458 sdata->aux_data = table[which].aux_data; 2563 }
2459 blitfun = table[which].blitfunc; 2564 sdata->aux_data = table[which].aux_data;
2460 2565 blitfun = table[which].blitfunc;
2461 if(blitfun == BlitNtoN) { /* default C fallback catch-all. Slow! */ 2566
2462 /* Fastpath C fallback: 32bit RGB<->RGBA blit with matching RGB */ 2567 if (blitfun == BlitNtoN) { /* default C fallback catch-all. Slow! */
2463 if ( srcfmt->BytesPerPixel == 4 && dstfmt->BytesPerPixel == 4 && 2568 /* Fastpath C fallback: 32bit RGB<->RGBA blit with matching RGB */
2464 srcfmt->Rmask == dstfmt->Rmask && 2569 if (srcfmt->BytesPerPixel == 4 && dstfmt->BytesPerPixel == 4 &&
2465 srcfmt->Gmask == dstfmt->Gmask && 2570 srcfmt->Rmask == dstfmt->Rmask &&
2466 srcfmt->Bmask == dstfmt->Bmask ) { 2571 srcfmt->Gmask == dstfmt->Gmask &&
2467 blitfun = Blit4to4MaskAlpha; 2572 srcfmt->Bmask == dstfmt->Bmask) {
2468 } else if ( a_need == COPY_ALPHA ) { 2573 blitfun = Blit4to4MaskAlpha;
2469 blitfun = BlitNtoNCopyAlpha; 2574 } else if (a_need == COPY_ALPHA) {
2470 } 2575 blitfun = BlitNtoNCopyAlpha;
2471 } 2576 }
2472 } 2577 }
2578 }
2473 2579
2474 #ifdef DEBUG_ASM 2580 #ifdef DEBUG_ASM
2475 #if SDL_HERMES_BLITTERS 2581 #if SDL_HERMES_BLITTERS
2476 if ( blitfun == ConvertMMX ) 2582 if (blitfun == ConvertMMX)
2477 fprintf(stderr, "Using mmx blit\n"); 2583 fprintf(stderr, "Using mmx blit\n");
2478 else 2584 else if (blitfun == ConvertX86)
2479 if ( blitfun == ConvertX86 ) 2585 fprintf(stderr, "Using asm blit\n");
2480 fprintf(stderr, "Using asm blit\n"); 2586 else
2481 else
2482 #endif 2587 #endif
2483 if ( (blitfun == BlitNtoN) || (blitfun == BlitNto1) ) 2588 if ((blitfun == BlitNtoN) || (blitfun == BlitNto1))
2484 fprintf(stderr, "Using C blit\n"); 2589 fprintf(stderr, "Using C blit\n");
2485 else 2590 else
2486 fprintf(stderr, "Using optimized C blit\n"); 2591 fprintf(stderr, "Using optimized C blit\n");
2487 #endif /* DEBUG_ASM */ 2592 #endif /* DEBUG_ASM */
2488 2593
2489 return(blitfun); 2594 return (blitfun);
2490 } 2595 }
2596
2597 /* vi: set ts=4 sw=4 expandtab: */