/* * * rgb2rgb.c, Software RGB to RGB convertor * Written by Nick Kurshev. */ #include #include "../config.h" #include "rgb2rgb.h" #include "../mmx_defs.h" #ifdef HAVE_MMX static const uint64_t mask32 __attribute__((aligned(8))) = 0x00FFFFFF00FFFFFFULL; static const uint64_t mask24l __attribute__((aligned(8))) = 0x0000000000FFFFFFULL; static const uint64_t mask24h __attribute__((aligned(8))) = 0x0000FFFFFF000000ULL; static const uint64_t mask15b __attribute__((aligned(8))) = 0x001F001F001F001FULL; /* 00000000 00011111 xxB */ static const uint64_t mask15rg __attribute__((aligned(8))) = 0x7FE07FE07FE07FE0ULL; /* 01111111 11100000 RGx */ #endif void rgb24to32(uint8_t *src,uint8_t *dst,uint32_t src_size) { uint8_t *dest = dst; uint8_t *s = src; uint8_t *end; #ifdef HAVE_MMX uint8_t *mm_end; #endif end = s + src_size; #ifdef HAVE_MMX __asm __volatile(PREFETCH" %0"::"m"(*s):"memory"); mm_end = (uint8_t*)((((unsigned long)end)/(MMREG_SIZE*2))*(MMREG_SIZE*2)); __asm __volatile("movq %0, %%mm7"::"m"(mask32):"memory"); if(mm_end == end) mm_end -= MMREG_SIZE*2; while(s < mm_end) { __asm __volatile( PREFETCH" 32%1\n\t" "movd %1, %%mm0\n\t" "movd 3%1, %%mm1\n\t" "movd 6%1, %%mm2\n\t" "movd 9%1, %%mm3\n\t" "punpckldq %%mm1, %%mm0\n\t" "punpckldq %%mm3, %%mm2\n\t" "pand %%mm7, %%mm0\n\t" "pand %%mm7, %%mm2\n\t" MOVNTQ" %%mm0, %0\n\t" MOVNTQ" %%mm2, 8%0" :"=m"(*dest) :"m"(*s) :"memory"); dest += 16; s += 12; } __asm __volatile(SFENCE:::"memory"); __asm __volatile(EMMS:::"memory"); #endif while(s < end) { *dest++ = *s++; *dest++ = *s++; *dest++ = *s++; *dest++ = 0; } } void rgb32to24(uint8_t *src,uint8_t *dst,uint32_t src_size) { uint8_t *dest = dst; uint8_t *s = src; uint8_t *end; #ifdef HAVE_MMX uint8_t *mm_end; #endif end = s + src_size; #ifdef HAVE_MMX __asm __volatile(PREFETCH" %0"::"m"(*s):"memory"); mm_end = (uint8_t*)((((unsigned long)end)/(MMREG_SIZE*2))*(MMREG_SIZE*2)); __asm __volatile( "movq %0, %%mm7\n\t" "movq %1, %%mm6" ::"m"(mask24l),"m"(mask24h):"memory"); if(mm_end == end) mm_end -= MMREG_SIZE*2; while(s < mm_end) { __asm __volatile( PREFETCH" 32%1\n\t" "movq %1, %%mm0\n\t" "movq 8%1, %%mm1\n\t" "movq %%mm0, %%mm2\n\t" "movq %%mm1, %%mm3\n\t" "psrlq $8, %%mm2\n\t" "psrlq $8, %%mm3\n\t" "pand %%mm7, %%mm0\n\t" "pand %%mm7, %%mm1\n\t" "pand %%mm6, %%mm2\n\t" "pand %%mm6, %%mm3\n\t" "por %%mm2, %%mm0\n\t" "por %%mm3, %%mm1\n\t" MOVNTQ" %%mm0, %0\n\t" MOVNTQ" %%mm1, 6%0" :"=m"(*dest) :"m"(*s) :"memory"); dest += 12; s += 16; } __asm __volatile(SFENCE:::"memory"); __asm __volatile(EMMS:::"memory"); #endif while(s < end) { *dest++ = *s++; *dest++ = *s++; *dest++ = *s++; s++; } } /* Original by Strepto/Astral ported to gcc & bugfixed : A'rpi MMX2, 3DNOW optimization by Nick Kurshev */ void rgb15to16(uint8_t *src,uint8_t *dst,uint32_t src_size) { #ifdef HAVE_MMX register char* s=src+src_size; register char* d=dst+src_size; register int offs=-src_size; __asm __volatile(PREFETCH" %0"::"m"(*(s+offs)):"memory"); __asm __volatile( "movq %0, %%mm4\n\t" "movq %1, %%mm5" ::"m"(mask15b), "m"(mask15rg):"memory"); while(offs<0) { __asm __volatile( PREFETCH" 32%1\n\t" "movq %1, %%mm0\n\t" "movq 8%1, %%mm2\n\t" "movq %%mm0, %%mm1\n\t" "movq %%mm2, %%mm3\n\t" "pand %%mm4, %%mm0\n\t" "pand %%mm5, %%mm1\n\t" "pand %%mm4, %%mm2\n\t" "pand %%mm5, %%mm3\n\t" "psllq $1, %%mm1\n\t" "psllq $1, %%mm3\n\t" "por %%mm1, %%mm0\n\t" "por %%mm3, %%mm2\n\t" MOVNTQ" %%mm0, %0\n\t" MOVNTQ" %%mm2, 8%0" :"=m"(*(d+offs)) :"m"(*(s+offs)) :"memory"); offs+=16; } __asm __volatile(SFENCE:::"memory"); __asm __volatile(EMMS:::"memory"); #else uint16_t *s1=( uint16_t * )src; uint16_t *d1=( uint16_t * )dst; uint16_t *e=((uint8_t *)s1)+src_size; while( s1