⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 interpolate8x8_mmx.asm

📁 MPEG-4编解码的实现(包括MPEG4视音频编解码)
💻 ASM
📖 第 1 页 / 共 2 页
字号:
;/**************************************************************************
; *
; *	XVID MPEG-4 VIDEO CODEC
; *	mmx/xmm/3dnow 8x8 block-based halfpel interpolation
; *
; *	This program is an implementation of a part of one or more MPEG-4
; *	Video tools as specified in ISO/IEC 14496-2 standard.  Those intending
; *	to use this software module in hardware or software products are
; *	advised that its use may infringe existing patents or copyrights, and
; *	any such use would be at such party's own risk.  The original
; *	developer of this software module and his/her company, and subsequent
; *	editors and their companies, will have no liability for use of this
; *	software or modifications or derivatives thereof.
; *
; *	This program is free software; you can redistribute it and/or modify
; *	it under the terms of the GNU General Public License as published by
; *	the Free Software Foundation; either version 2 of the License, or
; *	(at your option) any later version.
; *
; *	This program is distributed in the hope that it will be useful,
; *	but WITHOUT ANY WARRANTY; without even the implied warranty of
; *	MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
; *	GNU General Public License for more details.
; *
; *	You should have received a copy of the GNU General Public License
; *	along with this program; if not, write to the Free Software
; *	Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
; *
; *************************************************************************/

;/**************************************************************************
; *
; *	History:
; *
; * 04.02.2002  added xmm and additional 3dnow optimizations (Isibaar)
; * 27.12.2001	mofified "compensate_halfpel"
; *	22.12.2001	inital version; (c)2001 peter ross <pross@cs.rmit.edu.au>
; *
; *************************************************************************/


bits 32

%macro cglobal 1 
	%ifdef PREFIX
		global _%1 
		%define %1 _%1
	%else
		global %1
	%endif
%endmacro

section .data


align 16


;===========================================================================
; (1 - r) rounding table
;===========================================================================

rounding1_mmx
times 4 dw 1
times 4 dw 0

;===========================================================================
; (2 - r) rounding table  
;===========================================================================

rounding2_mmx
times 4 dw 2
times 4 dw 1

mmx_one
times 8 db 1

section .text

%macro  CALC_AVG 6
	punpcklbw %3, %6
	punpckhbw %4, %6

	paddusw %1, %3		; mm01 += mm23
	paddusw %2, %4
	paddusw %1, %5		; mm01 += rounding
	paddusw %2, %5
		
	psrlw %1, 1			; mm01 >>= 1
	psrlw %2, 1

%endmacro



;===========================================================================
;
; void interpolate8x8_halfpel_h_xmm(uint8_t * const dst,
;						const uint8_t * const src,
;						const uint32_t stride,
;						const uint32_t rounding);
;
;===========================================================================

align 16
cglobal interpolate8x8_halfpel_h_xmm
interpolate8x8_halfpel_h_xmm
		push	esi
		push	edi

		mov	edi, [esp + 8 + 4]		; dst
		mov	esi, [esp + 8 + 8]		; src
		mov	edx, [esp + 8 + 12]		; stride

		mov ecx, edx
		shl edx, 1

		mov	eax, [esp + 8 + 16]		; rounding
		
		or al,al
		jnz near halfpel_h_xmm_rounding1	; needs extra hack, still faster than mmx

halfpel_h_xmm_rounding0
;		mov eax, 4

;.loop
		movq mm0, [esi]
		movq mm1, [esi + ecx]
		pavgb mm0, [esi + 1]			; mm0 = avg([src], [src+1])
		pavgb mm1, [esi + ecx + 1]		; mm1 = avg([src+stride], [src+stride+1])

		movq [edi], mm0
		movq [edi + ecx], mm1

		add esi, edx	; src += 2*stride
		add edi, edx	; dst += 2*stride

		movq mm0, [esi]
		movq mm1, [esi + ecx]
		pavgb mm0, [esi + 1]
		pavgb mm1, [esi + ecx + 1]

		movq [edi], mm0
		movq [edi + ecx], mm1

		add esi, edx
		add edi, edx

		movq mm0, [esi]
		movq mm1, [esi + ecx]
		pavgb mm0, [esi + 1]
		pavgb mm1, [esi + ecx + 1]

		movq [edi], mm0
		movq [edi + ecx], mm1

		add esi, edx
		add edi, edx

		movq mm0, [esi]
		movq mm1, [esi + ecx]
		pavgb mm0, [esi + 1]
		pavgb mm1, [esi + ecx + 1]

		movq [edi], mm0
		movq [edi + ecx], mm1

;		dec eax
;		jnz	.loop

		pop edi
		pop esi

		ret

halfpel_h_xmm_rounding1
		movq mm7, [mmx_one]
		mov eax, 4

.loop
		movq mm0, [esi]
		movq mm1, [esi + ecx]
		movq mm2, [esi + 1]
		movq mm3, [esi + ecx + 1]
		movq mm4, mm0				; backup mm01
		movq mm5, mm1

		pavgb mm0, mm2				; mm0 = avg([src], [src+1])
		pavgb mm1, mm3				; mm1 = avg([src+stride], [src+stride+1])

		paddb mm2, mm4				; add values, see if lsb is set
		paddb mm3, mm5				; (to patch pavgb's (a+b+1)/2 rounding)
		pand mm2, mm7				; isolate lsb's
		pand mm3, mm7
		psubusb mm0, mm2			; mm23 = 0 if lsb wasn't set
		psubusb mm1, mm3

		movq [edi], mm0
		movq [edi + ecx], mm1

		add esi, edx	; src += 2*stride
		add edi, edx	; dst += 2*stride

		dec eax
		jnz	.loop

		pop edi
		pop esi

		ret


;===========================================================================
;
; void interpolate8x8_halfpel_h_3dn(uint8_t * const dst,
;						const uint8_t * const src,
;						const uint32_t stride,
;						const uint32_t rounding);
;
;===========================================================================

align 16
cglobal interpolate8x8_halfpel_h_3dn
interpolate8x8_halfpel_h_3dn
		push	esi
		push	edi

		mov	edi, [esp + 8 + 4]		; dst
		mov	esi, [esp + 8 + 8]		; src
		mov	edx, [esp + 8 + 12]		; stride

		mov ecx, edx
		shl edx, 1

		mov	eax, [esp + 8 + 16]		; rounding
		
		or al,al					; perform hack for pavgusb rounding issue
		jnz near halfpel_h_3dn_rounding1
		
halfpel_h_3dn_rounding0
;		mov eax, 4

;.loop
		movq mm0, [esi]
		movq mm1, [esi + ecx]
		pavgusb mm0, [esi + 1]			; mm0 = avg([src], [src+1])
		pavgusb mm1, [esi + ecx + 1]	; mm0 = avg([src+stride], [src+stride+1])

		movq [edi], mm0
		movq [edi + ecx], mm1

		add esi, edx	; src += 2*stride
		add edi, edx	; dst += 2*stride

		movq mm0, [esi]
		movq mm1, [esi + ecx]
		pavgusb mm0, [esi + 1]
		pavgusb mm1, [esi + ecx + 1]

		movq [edi], mm0
		movq [edi + ecx], mm1

		add esi, edx
		add edi, edx

		movq mm0, [esi]
		movq mm1, [esi + ecx]
		pavgusb mm0, [esi + 1]
		pavgusb mm1, [esi + ecx + 1]

		movq [edi], mm0
		movq [edi + ecx], mm1

		add esi, edx
		add edi, edx

		movq mm0, [esi]
		movq mm1, [esi + ecx]
		pavgusb mm0, [esi + 1]
		pavgusb mm1, [esi + ecx + 1]

		movq [edi], mm0
		movq [edi + ecx], mm1

;		dec eax
;		jnz	.loop

		pop edi
		pop esi

		ret

halfpel_h_3dn_rounding1
		movq mm7, [mmx_one]
		mov eax, 4

.loop
		movq mm0, [esi]
		movq mm1, [esi + ecx]
		movq mm2, [esi + 1]
		movq mm3, [esi + ecx + 1]
		movq mm4, mm0				; backup mm01
		movq mm5, mm1

		pavgusb mm0, mm2			; mm0 = avg([src], [src+1])
		pavgusb mm1, mm3			; mm1 = avg([src+stride], [src+stride+1])

		paddb mm2, mm4				; add values, see if lsb is set
		paddb mm3, mm5				; (to patch pavgb's (a+b+1)/2 rounding)
		pand mm2, mm7				; isolate lsb's
		pand mm3, mm7
		psubusb mm0, mm2			; mm23 = 0 if lsb wasn't set
		psubusb mm1, mm3

		movq [edi], mm0
		movq [edi + ecx], mm1

		add esi, edx	; src += 2*stride
		add edi, edx	; dst += 2*stride

		dec eax
		jnz	.loop

		pop edi
		pop esi

		ret




;===========================================================================
;
; void interpolate8x8_halfpel_h_mmx(uint8_t * const dst,
;						const uint8_t * const src,
;						const uint32_t stride,
;						const uint32_t rounding);
;
;===========================================================================

align 16
cglobal interpolate8x8_halfpel_h_mmx
interpolate8x8_halfpel_h_mmx

		push	esi
		push	edi

		mov	eax, [esp + 8 + 16]		; rounding

interpolate8x8_halfpel_h_mmx.start
		movq mm7, [rounding1_mmx + eax * 8]

		mov	edi, [esp + 8 + 4]		; dst
		mov	esi, [esp + 8 + 8]		; src
		mov	edx, [esp + 8 + 12]	; stride

		mov eax, 8
		pxor	mm6, mm6		; zero

.loop
		movq mm0, [esi]
		movq mm2, [esi + 1]
		movq mm1, mm0
		movq mm3, mm2

		punpcklbw mm0, mm6	; mm01 = [src]
		punpckhbw mm1, mm6	; mm23 = [src + 1]

		CALC_AVG mm0, mm1, mm2, mm3, mm7, mm6

		packuswb mm0, mm1
		movq [edi], mm0			; [dst] = mm01

		add esi, edx		; src += stride
		add edi, edx		; dst += stride

		dec eax
		jnz .loop

		pop edi
		pop esi

		ret


;===========================================================================
;
; void interpolate8x8_halfpel_v_xmm(uint8_t * const dst,
;						const uint8_t * const src,
;						const uint32_t stride,
;						const uint32_t rounding);
;
;===========================================================================

align 16
cglobal interpolate8x8_halfpel_v_xmm
interpolate8x8_halfpel_v_xmm
		push	esi
		push	edi

		mov	edi, [esp + 8 + 4]		; dst
		mov	esi, [esp + 8 + 8]		; src
		mov	edx, [esp + 8 + 12]		; stride

		mov	eax, [esp + 8 + 16]		; rounding
		
		or al,al
		jnz near halfpel_v_xmm_rounding1

halfpel_v_xmm_rounding0
;		mov eax, 8

;.loop
		movq mm0, [esi]
		pavgb mm0, [esi + edx]		; mm0 = avg([src], [src+stride])
		movq [edi], mm0

		add edi, edx	; dst += stride
		add esi, edx	; src += stride

		movq mm0, [esi]
		pavgb mm0, [esi + edx]
		movq [edi], mm0

		add edi, edx
		add esi, edx

		movq mm0, [esi]
		pavgb mm0, [esi + edx]
		movq [edi], mm0

		add edi, edx
		add esi, edx

		movq mm0, [esi]
		pavgb mm0, [esi + edx]
		movq [edi], mm0

		add edi, edx
		add esi, edx

		movq mm0, [esi]
		pavgb mm0, [esi + edx]
		movq [edi], mm0

		add edi, edx
		add esi, edx

		movq mm0, [esi]
		pavgb mm0, [esi + edx]
		movq [edi], mm0

		add edi, edx
		add esi, edx

		movq mm0, [esi]
		pavgb mm0, [esi + edx]
		movq [edi], mm0

		add edi, edx
		add esi, edx

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -