[SCM] libav/experimental: uyvytoyv12 in MMX (untested)

siretart at users.alioth.debian.org siretart at users.alioth.debian.org
Sun Jun 30 15:30:36 UTC 2013


The following commit has been merged in the experimental branch:
commit ed8c06708e3d8da1fd56933e5f5e5815a884792e
Author: Michael Niedermayer <michaelni at gmx.at>
Date:   Sun Nov 11 22:26:15 2001 +0000

    uyvytoyv12 in MMX (untested)
    
    Originally committed as revision 2848 to svn://svn.mplayerhq.hu/mplayer/trunk/postproc

diff --git a/postproc/rgb2rgb.c b/postproc/rgb2rgb.c
index ab2878f..d345854 100644
--- a/postproc/rgb2rgb.c
+++ b/postproc/rgb2rgb.c
@@ -811,9 +811,9 @@ void yuy2toyv12(const uint8_t *src, uint8_t *ydst, uint8_t *udst, uint8_t *vdst,
 		src  += srcStride;
 	}
 #ifdef HAVE_MMX
-asm(    EMMS" \n\t"
-        SFENCE" \n\t"
-        :::"memory");
+asm volatile(   EMMS" \n\t"
+        	SFENCE" \n\t"
+        	:::"memory");
 #endif
 }
 
@@ -830,6 +830,89 @@ void uyvytoyv12(const uint8_t *src, uint8_t *ydst, uint8_t *udst, uint8_t *vdst,
 	const int chromWidth= width>>1;
 	for(y=0; y<height; y+=2)
 	{
+#ifdef HAVE_MMX
+		asm volatile(
+			"xorl %%eax, %%eax		\n\t"
+			"pcmpeqw %%mm7, %%mm7		\n\t"
+			"psrlw $8, %%mm7		\n\t" // FF,00,FF,00...
+			".balign 16			\n\t"
+			"1:				\n\t"
+			PREFETCH" 64(%0, %%eax, 4)	\n\t"
+			"movq (%0, %%eax, 4), %%mm0	\n\t" // UYVY UYVY(0)
+			"movq 8(%0, %%eax, 4), %%mm1	\n\t" // UYVY UYVY(4)
+			"movq %%mm0, %%mm2		\n\t" // UYVY UYVY(0)
+			"movq %%mm1, %%mm3		\n\t" // UYVY UYVY(4)
+			"pand %%mm7, %%mm0		\n\t" // U0V0 U0V0(0)
+			"pand %%mm7, %%mm1		\n\t" // U0V0 U0V0(4)
+			"psrlw $8, %%mm2		\n\t" // Y0Y0 Y0Y0(0)
+			"psrlw $8, %%mm3		\n\t" // Y0Y0 Y0Y0(4)
+			"packuswb %%mm1, %%mm0		\n\t" // UVUV UVUV(0)
+			"packuswb %%mm3, %%mm2		\n\t" // YYYY YYYY(0)
+
+			MOVNTQ" %%mm2, (%1, %%eax, 2)	\n\t"
+
+			"movq 16(%0, %%eax, 4), %%mm1	\n\t" // UYVY UYVY(8)
+			"movq 24(%0, %%eax, 4), %%mm2	\n\t" // UYVY UYVY(12)
+			"movq %%mm1, %%mm3		\n\t" // UYVY UYVY(8)
+			"movq %%mm2, %%mm4		\n\t" // UYVY UYVY(12)
+			"pand %%mm7, %%mm1		\n\t" // U0V0 U0V0(8)
+			"pand %%mm7, %%mm2		\n\t" // U0V0 U0V0(12)
+			"psrlw $8, %%mm3		\n\t" // Y0Y0 Y0Y0(8)
+			"psrlw $8, %%mm4		\n\t" // Y0Y0 Y0Y0(12)
+			"packuswb %%mm2, %%mm1		\n\t" // UVUV UVUV(8)
+			"packuswb %%mm4, %%mm3		\n\t" // YYYY YYYY(8)
+
+			MOVNTQ" %%mm3, 8(%1, %%eax, 2)	\n\t"
+
+			"movq %%mm0, %%mm2		\n\t" // UVUV UVUV(0)
+			"movq %%mm1, %%mm3		\n\t" // UVUV UVUV(8)
+			"psrlw $8, %%mm0		\n\t" // V0V0 V0V0(0)
+			"psrlw $8, %%mm1		\n\t" // V0V0 V0V0(8)
+			"pand %%mm7, %%mm2		\n\t" // U0U0 U0U0(0)
+			"pand %%mm7, %%mm3		\n\t" // U0U0 U0U0(8)
+			"packuswb %%mm1, %%mm0		\n\t" // VVVV VVVV(0)
+			"packuswb %%mm3, %%mm2		\n\t" // UUUU UUUU(0)
+
+			MOVNTQ" %%mm0, (%3, %%eax)	\n\t"
+			MOVNTQ" %%mm2, (%2, %%eax)	\n\t"
+
+			"addl $8, %%eax			\n\t"
+			"cmpl %4, %%eax			\n\t"
+			" jb 1b				\n\t"
+			::"r"(src), "r"(ydst), "r"(udst), "r"(vdst), "r" (chromWidth)
+			: "memory", "%eax"
+		);
+
+		ydst += lumStride;
+		src  += srcStride;
+
+		asm volatile(
+			"xorl %%eax, %%eax		\n\t"
+			".balign 16			\n\t"
+			"1:				\n\t"
+			PREFETCH" 64(%0, %%eax, 4)	\n\t"
+			"movq (%0, %%eax, 4), %%mm0	\n\t" // YUYV YUYV(0)
+			"movq 8(%0, %%eax, 4), %%mm1	\n\t" // YUYV YUYV(4)
+			"movq 16(%0, %%eax, 4), %%mm2	\n\t" // YUYV YUYV(8)
+			"movq 24(%0, %%eax, 4), %%mm3	\n\t" // YUYV YUYV(12)
+			"psrlw $8, %%mm0		\n\t" // Y0Y0 Y0Y0(0)
+			"psrlw $8, %%mm1		\n\t" // Y0Y0 Y0Y0(4)
+			"psrlw $8, %%mm2		\n\t" // Y0Y0 Y0Y0(8)
+			"psrlw $8, %%mm3		\n\t" // Y0Y0 Y0Y0(12)
+			"packuswb %%mm1, %%mm0		\n\t" // YYYY YYYY(0)
+			"packuswb %%mm3, %%mm2		\n\t" // YYYY YYYY(8)
+
+			MOVNTQ" %%mm0, (%1, %%eax, 2)	\n\t"
+			MOVNTQ" %%mm2, 8(%1, %%eax, 2)	\n\t"
+
+			"addl $8, %%eax			\n\t"
+			"cmpl %4, %%eax			\n\t"
+			" jb 1b				\n\t"
+
+			::"r"(src), "r"(ydst), "r"(udst), "r"(vdst), "r" (chromWidth)
+			: "memory", "%eax"
+		);
+#else
 		int i;
 		for(i=0; i<chromWidth; i++)
 		{
@@ -846,11 +929,17 @@ void uyvytoyv12(const uint8_t *src, uint8_t *ydst, uint8_t *udst, uint8_t *vdst,
 			ydst[2*i+0] 	= src[4*i+1];
 			ydst[2*i+1] 	= src[4*i+3];
 		}
+#endif
 		udst += chromStride;
 		vdst += chromStride;
 		ydst += lumStride;
 		src  += srcStride;
 	}
+#ifdef HAVE_MMX
+asm volatile(   EMMS" \n\t"
+        	SFENCE" \n\t"
+        	:::"memory");
+#endif
 }
 
 
diff --git a/postproc/rgb2rgb_template.c b/postproc/rgb2rgb_template.c
index ab2878f..d345854 100644
--- a/postproc/rgb2rgb_template.c
+++ b/postproc/rgb2rgb_template.c
@@ -811,9 +811,9 @@ void yuy2toyv12(const uint8_t *src, uint8_t *ydst, uint8_t *udst, uint8_t *vdst,
 		src  += srcStride;
 	}
 #ifdef HAVE_MMX
-asm(    EMMS" \n\t"
-        SFENCE" \n\t"
-        :::"memory");
+asm volatile(   EMMS" \n\t"
+        	SFENCE" \n\t"
+        	:::"memory");
 #endif
 }
 
@@ -830,6 +830,89 @@ void uyvytoyv12(const uint8_t *src, uint8_t *ydst, uint8_t *udst, uint8_t *vdst,
 	const int chromWidth= width>>1;
 	for(y=0; y<height; y+=2)
 	{
+#ifdef HAVE_MMX
+		asm volatile(
+			"xorl %%eax, %%eax		\n\t"
+			"pcmpeqw %%mm7, %%mm7		\n\t"
+			"psrlw $8, %%mm7		\n\t" // FF,00,FF,00...
+			".balign 16			\n\t"
+			"1:				\n\t"
+			PREFETCH" 64(%0, %%eax, 4)	\n\t"
+			"movq (%0, %%eax, 4), %%mm0	\n\t" // UYVY UYVY(0)
+			"movq 8(%0, %%eax, 4), %%mm1	\n\t" // UYVY UYVY(4)
+			"movq %%mm0, %%mm2		\n\t" // UYVY UYVY(0)
+			"movq %%mm1, %%mm3		\n\t" // UYVY UYVY(4)
+			"pand %%mm7, %%mm0		\n\t" // U0V0 U0V0(0)
+			"pand %%mm7, %%mm1		\n\t" // U0V0 U0V0(4)
+			"psrlw $8, %%mm2		\n\t" // Y0Y0 Y0Y0(0)
+			"psrlw $8, %%mm3		\n\t" // Y0Y0 Y0Y0(4)
+			"packuswb %%mm1, %%mm0		\n\t" // UVUV UVUV(0)
+			"packuswb %%mm3, %%mm2		\n\t" // YYYY YYYY(0)
+
+			MOVNTQ" %%mm2, (%1, %%eax, 2)	\n\t"
+
+			"movq 16(%0, %%eax, 4), %%mm1	\n\t" // UYVY UYVY(8)
+			"movq 24(%0, %%eax, 4), %%mm2	\n\t" // UYVY UYVY(12)
+			"movq %%mm1, %%mm3		\n\t" // UYVY UYVY(8)
+			"movq %%mm2, %%mm4		\n\t" // UYVY UYVY(12)
+			"pand %%mm7, %%mm1		\n\t" // U0V0 U0V0(8)
+			"pand %%mm7, %%mm2		\n\t" // U0V0 U0V0(12)
+			"psrlw $8, %%mm3		\n\t" // Y0Y0 Y0Y0(8)
+			"psrlw $8, %%mm4		\n\t" // Y0Y0 Y0Y0(12)
+			"packuswb %%mm2, %%mm1		\n\t" // UVUV UVUV(8)
+			"packuswb %%mm4, %%mm3		\n\t" // YYYY YYYY(8)
+
+			MOVNTQ" %%mm3, 8(%1, %%eax, 2)	\n\t"
+
+			"movq %%mm0, %%mm2		\n\t" // UVUV UVUV(0)
+			"movq %%mm1, %%mm3		\n\t" // UVUV UVUV(8)
+			"psrlw $8, %%mm0		\n\t" // V0V0 V0V0(0)
+			"psrlw $8, %%mm1		\n\t" // V0V0 V0V0(8)
+			"pand %%mm7, %%mm2		\n\t" // U0U0 U0U0(0)
+			"pand %%mm7, %%mm3		\n\t" // U0U0 U0U0(8)
+			"packuswb %%mm1, %%mm0		\n\t" // VVVV VVVV(0)
+			"packuswb %%mm3, %%mm2		\n\t" // UUUU UUUU(0)
+
+			MOVNTQ" %%mm0, (%3, %%eax)	\n\t"
+			MOVNTQ" %%mm2, (%2, %%eax)	\n\t"
+
+			"addl $8, %%eax			\n\t"
+			"cmpl %4, %%eax			\n\t"
+			" jb 1b				\n\t"
+			::"r"(src), "r"(ydst), "r"(udst), "r"(vdst), "r" (chromWidth)
+			: "memory", "%eax"
+		);
+
+		ydst += lumStride;
+		src  += srcStride;
+
+		asm volatile(
+			"xorl %%eax, %%eax		\n\t"
+			".balign 16			\n\t"
+			"1:				\n\t"
+			PREFETCH" 64(%0, %%eax, 4)	\n\t"
+			"movq (%0, %%eax, 4), %%mm0	\n\t" // YUYV YUYV(0)
+			"movq 8(%0, %%eax, 4), %%mm1	\n\t" // YUYV YUYV(4)
+			"movq 16(%0, %%eax, 4), %%mm2	\n\t" // YUYV YUYV(8)
+			"movq 24(%0, %%eax, 4), %%mm3	\n\t" // YUYV YUYV(12)
+			"psrlw $8, %%mm0		\n\t" // Y0Y0 Y0Y0(0)
+			"psrlw $8, %%mm1		\n\t" // Y0Y0 Y0Y0(4)
+			"psrlw $8, %%mm2		\n\t" // Y0Y0 Y0Y0(8)
+			"psrlw $8, %%mm3		\n\t" // Y0Y0 Y0Y0(12)
+			"packuswb %%mm1, %%mm0		\n\t" // YYYY YYYY(0)
+			"packuswb %%mm3, %%mm2		\n\t" // YYYY YYYY(8)
+
+			MOVNTQ" %%mm0, (%1, %%eax, 2)	\n\t"
+			MOVNTQ" %%mm2, 8(%1, %%eax, 2)	\n\t"
+
+			"addl $8, %%eax			\n\t"
+			"cmpl %4, %%eax			\n\t"
+			" jb 1b				\n\t"
+
+			::"r"(src), "r"(ydst), "r"(udst), "r"(vdst), "r" (chromWidth)
+			: "memory", "%eax"
+		);
+#else
 		int i;
 		for(i=0; i<chromWidth; i++)
 		{
@@ -846,11 +929,17 @@ void uyvytoyv12(const uint8_t *src, uint8_t *ydst, uint8_t *udst, uint8_t *vdst,
 			ydst[2*i+0] 	= src[4*i+1];
 			ydst[2*i+1] 	= src[4*i+3];
 		}
+#endif
 		udst += chromStride;
 		vdst += chromStride;
 		ydst += lumStride;
 		src  += srcStride;
 	}
+#ifdef HAVE_MMX
+asm volatile(   EMMS" \n\t"
+        	SFENCE" \n\t"
+        	:::"memory");
+#endif
 }
 
 

-- 
Libav/FFmpeg packaging



More information about the pkg-multimedia-commits mailing list