⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 yuv_to_yv12_mmx.asm

📁 MPEG4的VC代码
💻 ASM
字号:
;/*****************************************************************************; *; *  XVID MPEG-4 VIDEO CODEC; *	 mmx yuv planar to yv12 conversion ; *; *  Copyright (C) 2001 - Michael Militzer <isibaar@xvid.org>; *; *  This file is part of XviD, a free MPEG-4 video encoder/decoder; *; *  XviD is free software; you can redistribute it and/or modify it; *  under the terms of the GNU General Public License as published by; *  the Free Software Foundation; either version 2 of the License, or; *  (at your option) any later version.; *; *  This program is distributed in the hope that it will be useful,; *  but WITHOUT ANY WARRANTY; without even the implied warranty of; *  MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the; *  GNU General Public License for more details.; *; *  You should have received a copy of the GNU General Public License; *  along with this program; if not, write to the Free Software; *  Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307 USA; *; *  Under section 8 of the GNU General Public License, the copyright; *  holders of XVID explicitly forbid distribution in the following; *  countries:; *; *    - Japan; *    - United States of America; *; *  Linking XviD statically or dynamically with other modules is making a; *  combined work based on XviD.  Thus, the terms and conditions of the; *  GNU General Public License cover the whole combination.; *; *  As a special exception, the copyright holders of XviD give you; *  permission to link XviD with independent modules that communicate with; *  XviD solely through the VFW1.1 and DShow interfaces, regardless of the; *  license terms of these independent modules, and to copy and distribute; *  the resulting combined work under terms of your choice, provided that; *  every copy of the combined work is accompanied by a complete copy of; *  the source code of XviD (the version of XviD used to produce the; *  combined work), being distributed under the terms of the GNU General; *  Public License plus this exception.  An independent module is a module; *  which is not derived from or based on XviD.; *; *  Note that people who make modified versions of XviD are not obligated; *  to grant this special exception for their modified versions; it is; *  their choice whether to do so.  The GNU General Public License gives; *  permission to release a modified version without this exception; this; *  exception also makes it possible to release a modified version which; *  carries forward this exception.; *; * $Id: yuv_to_yv12_mmx.asm,v 1.7 2002/11/17 00:20:30 edgomez Exp $; *; ****************************************************************************/BITS 32%macro cglobal 1 %ifdef PREFIX	global _%1 		%define %1 _%1	%else		global %1	%endif%endmacroSECTION .textALIGN 64;------------------------------------------------------------------------------; ; void yuv_to_yv12_xmm(uint8_t *y_out,;                      uint8_t *u_out,;                      uint8_t *v_out,;                      uint8_t *src,;                      int width, int height, int stride);;; This function probably also runs on PentiumII class cpu's; ; Attention: This code assumes that width is a multiple of 16;;------------------------------------------------------------------------------	cglobal yuv_to_yv12_xmmyuv_to_yv12_xmm:		push ebx	push esi	push edi	push ebp	; local vars allocation%define localsize 4%define remainder esp	sub esp, localsize	; function code	mov eax, [esp + 40 + localsize]		; height -> eax	mov ebx, [esp + 44 + localsize]		; stride -> ebx	mov esi, [esp + 32 + localsize] 	; src -> esi 	mov edi, [esp + 20 + localsize] 	; y_out -> edi 	mov ecx, [esp + 36 + localsize] 	; width -> ecx	sub ebx, ecx			; stride - width -> ebx		mov edx, ecx	mov ebp, ecx	shr edx, 6					mov ecx, edx			; 64 bytes copied per iteration	shl edx, 6	sub ebp, edx			; remainder -> ebp	shr ebp, 4			; 16 bytes per iteration	add ebp, 1				mov [remainder], ebp	mov edx, ecx			.y_inner_loop:	prefetchnta [esi + 64]	; non temporal prefetch 	prefetchnta [esi + 96] 	movq mm1, [esi]			; read from src 	movq mm2, [esi + 8] 	movq mm3, [esi + 16] 	movq mm4, [esi + 24] 	movq mm5, [esi + 32] 	movq mm6, [esi + 40] 	movq mm7, [esi + 48] 	movq mm0, [esi + 56] 	movntq [edi], mm1		; write to y_out 	movntq [edi + 8], mm2 	movntq [edi + 16], mm3 	movntq [edi + 24], mm4 	movntq [edi + 32], mm5 	movntq [edi + 40], mm6 	movntq [edi + 48], mm7 	movntq [edi + 56], mm0 	add esi, 64	add edi, 64 	dec ecx	jnz .y_inner_loop    		dec ebp	jz .y_outer_loop.y_remainder_loop:	movq mm1, [esi]			; read from src 	movq mm2, [esi + 8] 	movntq [edi], mm1		; write to y_out 	movntq [edi + 8], mm2 	add esi, 16	add edi, 16 	dec ebp	jnz .y_remainder_loop	.y_outer_loop:	mov ebp, [remainder]	mov ecx, edx	add edi, ebx	    	dec eax	jnz near .y_inner_loop	mov eax, [esp + 40 + localsize]		; height -> eax	mov ebx, [esp + 44 + localsize]		; stride -> ebx	mov ecx, [esp + 36 + localsize]	 	; width -> ecx	mov edi, [esp + 24 + localsize] 	; u_out -> edi 	shr ecx, 1				; width / 2 -> ecx	shr ebx, 1				; stride / 2 -> ebx	shr eax, 1				; height / 2 -> eax	sub ebx, ecx			; stride / 2 - width / 2 -> ebx	mov edx, ecx	mov ebp, ecx	shr edx, 6					mov ecx, edx			; 64 bytes copied per iteration	shl edx, 6	sub ebp, edx			; remainder -> ebp	shr ebp, 3			; 8 bytes per iteration	add ebp, 1				mov [remainder], ebp		mov edx, ecx			.u_inner_loop:	prefetchnta [esi + 64]	; non temporal prefetch 	prefetchnta [esi + 96] 	movq mm1, [esi]			; read from src 	movq mm2, [esi + 8] 	movq mm3, [esi + 16] 	movq mm4, [esi + 24] 	movq mm5, [esi + 32] 	movq mm6, [esi + 40] 	movq mm7, [esi + 48] 	movq mm0, [esi + 56] 	movntq [edi], mm1		; write to u_out 	movntq [edi + 8], mm2 	movntq [edi + 16], mm3 	movntq [edi + 24], mm4 	movntq [edi + 32], mm5 	movntq [edi + 40], mm6 	movntq [edi + 48], mm7 	movntq [edi + 56], mm0 	add esi, 64	add edi, 64 	dec ecx	jnz .u_inner_loop    	dec ebp	jz .u_outer_loop.u_remainder_loop:	movq mm1, [esi]			; read from src 	movntq [edi], mm1		; write to y_out 	add esi, 8	add edi, 8 	dec ebp	jnz .u_remainder_loop.u_outer_loop:	mov ebp, [remainder]		mov ecx, edx	add edi, ebx	    	dec eax	jnz .u_inner_loop	mov eax, [esp + 40 + localsize]		; height -> eax	mov ecx, [esp + 36 + localsize] 	; width -> ecx	mov edi, [esp + 28 + localsize] 	; v_out -> edi 	shr ecx, 1				; width / 2 -> ecx	shr eax, 1				; height / 2 -> eax	mov edx, ecx	mov ebp, ecx	shr edx, 6					mov ecx, edx			; 64 bytes copied per iteration	shl edx, 6	sub ebp, edx			; remainder -> ebp	shr ebp, 3			; 8 bytes per iteration	add ebp, 1				mov [remainder], ebp		mov edx, ecx			.v_inner_loop:	prefetchnta [esi + 64]	; non temporal prefetch 	prefetchnta [esi + 96] 	movq mm1, [esi]			; read from src 	movq mm2, [esi + 8] 	movq mm3, [esi + 16] 	movq mm4, [esi + 24] 	movq mm5, [esi + 32] 	movq mm6, [esi + 40] 	movq mm7, [esi + 48] 	movq mm0, [esi + 56] 	movntq [edi], mm1		; write to u_out 	movntq [edi + 8], mm2 	movntq [edi + 16], mm3 	movntq [edi + 24], mm4 	movntq [edi + 32], mm5 	movntq [edi + 40], mm6 	movntq [edi + 48], mm7 	movntq [edi + 56], mm0 	add esi, 64	add edi, 64 	dec ecx	jnz .v_inner_loop    	dec ebp	jz .v_outer_loop.v_remainder_loop:	movq mm1, [esi]			; read from src 	movntq [edi], mm1		; write to y_out 	add esi, 8	add edi, 8 	dec ebp	jnz .v_remainder_loop.v_outer_loop:	mov ebp, [remainder]		mov ecx, edx	add edi, ebx	    	dec eax	jnz .v_inner_loop	; local vars deallocation	add esp, localsize%undef localsize%undef remainder	pop ebp	pop edi	pop esi	pop ebx	    	emms	ret;------------------------------------------------------------------------------ ;; void yuv_to_yv12_mmx(uint8_t *y_out,;                      uint8_t *u_out,;                      uint8_t *v_out,;                      uint8_t *src,;                      int width, int height, int stride);;; Attention: This code assumes that width is a multiple of 16; ;------------------------------------------------------------------------------cglobal yuv_to_yv12_mmxyuv_to_yv12_mmx:		push ebx	push esi	push edi	push ebp	; local vars allocation%define localsize 4%define remainder esp	sub esp, localsize	;  function code	mov eax, [esp + 40 + localsize]	; height -> eax	mov ebx, [esp + 44 + localsize]	; stride -> ebx	mov esi, [esp + 32 + localsize] ; src -> esi 	mov edi, [esp + 20 + localsize] ; y_out -> edi 	mov ecx, [esp + 36 + localsize] ; width -> ecx	sub ebx, ecx		; stride - width -> ebx	mov edx, ecx	mov ebp, ecx	shr edx, 6					mov ecx, edx		; 64 bytes copied per iteration	shl edx, 6	sub ebp, edx		; mainder -> ebp	shr ebp, 4		; 16 bytes per iteration	add ebp, 1				mov [remainder], ebp	mov edx, ecx			.y_inner_loop:	movq mm1, [esi]		; read from src 	movq mm2, [esi + 8] 	movq mm3, [esi + 16] 	movq mm4, [esi + 24] 	movq mm5, [esi + 32] 	movq mm6, [esi + 40] 	movq mm7, [esi + 48] 	movq mm0, [esi + 56] 	movq [edi], mm1		; write to y_out 	movq [edi + 8], mm2 	movq [edi + 16], mm3 	movq [edi + 24], mm4 	movq [edi + 32], mm5 	movq [edi + 40], mm6 	movq [edi + 48], mm7 	movq [edi + 56], mm0 	add esi, 64	add edi, 64 	dec ecx	jnz .y_inner_loop    		dec ebp	jz .y_outer_loop.y_remainder_loop:	movq mm1, [esi]		; read from src 	movq mm2, [esi + 8] 	movq [edi], mm1		; write to y_out 	movq [edi + 8], mm2 	add esi, 16	add edi, 16 	dec ebp	jnz .y_remainder_loop	.y_outer_loop:	mov ebp, [remainder]		mov ecx, edx	add edi, ebx	    	dec eax	jnz near .y_inner_loop	mov eax, [esp + 40 + localsize]	; height -> eax	mov ebx, [esp + 44 + localsize]	; stride -> ebx	mov ecx, [esp + 36 + localsize]	; width -> ecx	mov edi, [esp + 24 + localsize]	; u_out -> edi 	shr ecx, 1		; width / 2 -> ecx	shr ebx, 1		; stride / 2 -> ebx	shr eax, 1		; height / 2 -> eax	sub ebx, ecx		; stride / 2 - width / 2 -> ebx	mov edx, ecx	mov ebp, ecx	shr edx, 6					mov ecx, edx		; 64 bytes copied per iteration	shl edx, 6	sub ebp, edx		; remainder -> ebp	shr ebp, 3		; 8 bytes per iteration	add ebp, 1				mov [remainder], ebp		mov edx, ecx			.u_inner_loop:	movq mm1, [esi]		; read from src 	movq mm2, [esi + 8] 	movq mm3, [esi + 16] 	movq mm4, [esi + 24] 	movq mm5, [esi + 32] 	movq mm6, [esi + 40] 	movq mm7, [esi + 48] 	movq mm0, [esi + 56] 	movq [edi], mm1		; write to u_out 	movq [edi + 8], mm2 	movq [edi + 16], mm3 	movq [edi + 24], mm4 	movq [edi + 32], mm5 	movq [edi + 40], mm6 	movq [edi + 48], mm7 	movq [edi + 56], mm0 	add esi, 64	add edi, 64 	dec ecx	jnz .u_inner_loop    	dec ebp	jz .u_outer_loop.u_remainder_loop:	movq mm1, [esi]		; read from src 	movq [edi], mm1		; write to y_out 	add esi, 8	add edi, 8 	dec ebp	jnz .u_remainder_loop.u_outer_loop:	mov ebp, [remainder]		mov ecx, edx	add edi, ebx	    	dec eax	jnz .u_inner_loop	mov eax, [esp + 40 + localsize]	; height -> eax	mov ecx, [esp + 36 + localsize]	; width -> ecx	mov edi, [esp + 28 + localsize]	; v_out -> edi 	shr ecx, 1		; width / 2 -> ecx	shr eax, 1		; height / 2 -> eax	mov edx, ecx	mov ebp, ecx	shr edx, 6					mov ecx, edx		; 64 bytes copied per iteration	shl edx, 6	sub ebp, edx		; remainder -> ebp	shr ebp, 3		; 8 bytes per iteration	add ebp, 1				mov [remainder], ebp		mov edx, ecx			.v_inner_loop:	movq mm1, [esi]		; read from src 	movq mm2, [esi + 8] 	movq mm3, [esi + 16] 	movq mm4, [esi + 24] 	movq mm5, [esi + 32] 	movq mm6, [esi + 40] 	movq mm7, [esi + 48] 	movq mm0, [esi + 56] 	movq [edi], mm1		; write to u_out 	movq [edi + 8], mm2 	movq [edi + 16], mm3 	movq [edi + 24], mm4 	movq [edi + 32], mm5 	movq [edi + 40], mm6 	movq [edi + 48], mm7 	movq [edi + 56], mm0 	add esi, 64	add edi, 64 	dec ecx	jnz .v_inner_loop    	dec ebp	jz .v_outer_loop.v_remainder_loop:	movq mm1, [esi]		; read from src 	movq [edi], mm1		; write to y_out 	add esi, 8	add edi, 8 	dec ebp	jnz .v_remainder_loop.v_outer_loop:	mov ebp, [remainder]	mov ecx, edx	add edi, ebx	    	dec eax	jnz .v_inner_loop	; local vars deallocation	add esp, localsize%undef localsize%undef remainder			pop ebp	pop edi	pop esi	pop ebx	    	emms	ret

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -