📄 pixel.c
字号:
/***************************************************************************** * pixel.c: h264 encoder ***************************************************************************** * Copyright (C) 2003-2008 x264 project * * Authors: Eric Petit <eric.petit@lapsus.org> * Guillaume Poirier <gpoirier@mplayerhq.hu> * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02111, USA. *****************************************************************************/#include "common/common.h"#include "ppccommon.h"/*********************************************************************** * SAD routines **********************************************************************/#define PIXEL_SAD_ALTIVEC( name, lx, ly, a, b ) \static int name( uint8_t *pix1, int i_pix1, \ uint8_t *pix2, int i_pix2 ) \{ \ int y; \ DECLARE_ALIGNED_16( int sum ); \ \ LOAD_ZERO; \ PREP_LOAD; \ vec_u8_t pix1v, pix2v; \ vec_s32_t sumv = zero_s32v; \ for( y = 0; y < ly; y++ ) \ { \ VEC_LOAD_G( pix1, pix1v, lx, vec_u8_t ); \ VEC_LOAD_G( pix2, pix2v, lx, vec_u8_t ); \ sumv = (vec_s32_t) vec_sum4s( \ vec_sub( vec_max( pix1v, pix2v ), \ vec_min( pix1v, pix2v ) ), \ (vec_u32_t) sumv ); \ pix1 += i_pix1; \ pix2 += i_pix2; \ } \ sumv = vec_sum##a( sumv, zero_s32v ); \ sumv = vec_splat( sumv, b ); \ vec_ste( sumv, 0, &sum ); \ return sum; \}PIXEL_SAD_ALTIVEC( pixel_sad_16x16_altivec, 16, 16, s, 3 )PIXEL_SAD_ALTIVEC( pixel_sad_8x16_altivec, 8, 16, 2s, 1 )PIXEL_SAD_ALTIVEC( pixel_sad_16x8_altivec, 16, 8, s, 3 )PIXEL_SAD_ALTIVEC( pixel_sad_8x8_altivec, 8, 8, 2s, 1 )/*********************************************************************** * SATD routines **********************************************************************//*********************************************************************** * VEC_HADAMAR *********************************************************************** * b[0] = a[0] + a[1] + a[2] + a[3] * b[1] = a[0] + a[1] - a[2] - a[3] * b[2] = a[0] - a[1] - a[2] + a[3] * b[3] = a[0] - a[1] + a[2] - a[3] **********************************************************************/#define VEC_HADAMAR(a0,a1,a2,a3,b0,b1,b2,b3) \ b2 = vec_add( a0, a1 ); \ b3 = vec_add( a2, a3 ); \ a0 = vec_sub( a0, a1 ); \ a2 = vec_sub( a2, a3 ); \ b0 = vec_add( b2, b3 ); \ b1 = vec_sub( b2, b3 ); \ b2 = vec_sub( a0, a2 ); \ b3 = vec_add( a0, a2 )/*********************************************************************** * VEC_ABS *********************************************************************** * a: s16v * * a = abs(a) * * Call vec_sub()/vec_max() instead of vec_abs() because vec_abs() * actually also calls vec_splat(0), but we already have a null vector. **********************************************************************/#define VEC_ABS(a) \ a = vec_max( a, vec_sub( zero_s16v, a ) );#define VEC_ABSOLUTE(a) (vec_u16_t)vec_max( a, vec_sub( zero_s16v, a ) )/*********************************************************************** * VEC_ADD_ABS *********************************************************************** * a: s16v * b, c: s32v * * c[i] = abs(a[2*i]) + abs(a[2*i+1]) + [bi] **********************************************************************/#define VEC_ADD_ABS(a,b,c) \ VEC_ABS( a ); \ c = vec_sum4s( a, b )/*********************************************************************** * SATD 4x4 **********************************************************************/static int pixel_satd_4x4_altivec( uint8_t *pix1, int i_pix1, uint8_t *pix2, int i_pix2 ){ DECLARE_ALIGNED_16( int i_satd ); PREP_DIFF; PREP_LOAD_SRC( pix1 ); vec_s16_t diff0v, diff1v, diff2v, diff3v; vec_s16_t temp0v, temp1v, temp2v, temp3v; vec_s32_t satdv; vec_u8_t _offset1v_ = vec_lvsl(0, pix2); vec_u8_t _offset2v_ = vec_lvsl(0, pix2 + i_pix2); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 4, diff0v, offset1v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 4, diff1v, offset2v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 4, diff2v, offset1v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 4, diff3v, offset2v ); /* Hadamar H */ VEC_HADAMAR( diff0v, diff1v, diff2v, diff3v, temp0v, temp1v, temp2v, temp3v ); VEC_TRANSPOSE_4( temp0v, temp1v, temp2v, temp3v, diff0v, diff1v, diff2v, diff3v ); /* Hadamar V */ VEC_HADAMAR( diff0v, diff1v, diff2v, diff3v, temp0v, temp1v, temp2v, temp3v ); VEC_ADD_ABS( temp0v, zero_s32v, satdv ); VEC_ADD_ABS( temp1v, satdv, satdv ); VEC_ADD_ABS( temp2v, satdv, satdv ); VEC_ADD_ABS( temp3v, satdv, satdv ); satdv = vec_sum2s( satdv, zero_s32v ); satdv = vec_splat( satdv, 1 ); vec_ste( satdv, 0, &i_satd ); return i_satd / 2;}/*********************************************************************** * SATD 4x8 **********************************************************************/static int pixel_satd_4x8_altivec( uint8_t *pix1, int i_pix1, uint8_t *pix2, int i_pix2 ){ DECLARE_ALIGNED_16( int i_satd ); PREP_DIFF; vec_s16_t diff0v, diff1v, diff2v, diff3v; vec_s16_t temp0v, temp1v, temp2v, temp3v; vec_s32_t satdv; PREP_LOAD_SRC( pix1 ); vec_u8_t _offset1v_ = vec_lvsl(0, pix2); vec_u8_t _offset2v_ = vec_lvsl(0, pix2 + i_pix2); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 4, diff0v, offset1v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 4, diff1v, offset2v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 4, diff2v, offset1v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 4, diff3v, offset2v ); VEC_HADAMAR( diff0v, diff1v, diff2v, diff3v, temp0v, temp1v, temp2v, temp3v ); VEC_TRANSPOSE_4( temp0v, temp1v, temp2v, temp3v, diff0v, diff1v, diff2v, diff3v ); VEC_HADAMAR( diff0v, diff1v, diff2v, diff3v, temp0v, temp1v, temp2v, temp3v ); VEC_ADD_ABS( temp0v, zero_s32v, satdv ); VEC_ADD_ABS( temp1v, satdv, satdv ); VEC_ADD_ABS( temp2v, satdv, satdv ); VEC_ADD_ABS( temp3v, satdv, satdv ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 4, diff0v, offset1v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 4, diff1v, offset2v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 4, diff2v, offset1v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 4, diff3v, offset2v ); VEC_HADAMAR( diff0v, diff1v, diff2v, diff3v, temp0v, temp1v, temp2v, temp3v ); VEC_TRANSPOSE_4( temp0v, temp1v, temp2v, temp3v, diff0v, diff1v, diff2v, diff3v ); VEC_HADAMAR( diff0v, diff1v, diff2v, diff3v, temp0v, temp1v, temp2v, temp3v ); VEC_ADD_ABS( temp0v, satdv, satdv ); VEC_ADD_ABS( temp1v, satdv, satdv ); VEC_ADD_ABS( temp2v, satdv, satdv ); VEC_ADD_ABS( temp3v, satdv, satdv ); satdv = vec_sum2s( satdv, zero_s32v ); satdv = vec_splat( satdv, 1 ); vec_ste( satdv, 0, &i_satd ); return i_satd / 2;}/*********************************************************************** * SATD 8x4 **********************************************************************/static int pixel_satd_8x4_altivec( uint8_t *pix1, int i_pix1, uint8_t *pix2, int i_pix2 ){ DECLARE_ALIGNED_16( int i_satd ); PREP_DIFF; vec_s16_t diff0v, diff1v, diff2v, diff3v, diff4v, diff5v, diff6v, diff7v; vec_s16_t temp0v, temp1v, temp2v, temp3v, temp4v, temp5v, temp6v, temp7v; vec_s32_t satdv; PREP_LOAD_SRC( pix1 ); vec_u8_t _offset1v_ = vec_lvsl(0, pix2); vec_u8_t _offset2v_ = vec_lvsl(0, pix2 + i_pix2); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 8, diff0v, offset1v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 8, diff1v, offset2v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 8, diff2v, offset1v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 8, diff3v, offset2v ); VEC_HADAMAR( diff0v, diff1v, diff2v, diff3v, temp0v, temp1v, temp2v, temp3v ); /* This causes warnings because temp4v...temp7v haven't be set, but we don't care */ VEC_TRANSPOSE_8( temp0v, temp1v, temp2v, temp3v, temp4v, temp5v, temp6v, temp7v, diff0v, diff1v, diff2v, diff3v, diff4v, diff5v, diff6v, diff7v ); VEC_HADAMAR( diff0v, diff1v, diff2v, diff3v, temp0v, temp1v, temp2v, temp3v ); VEC_HADAMAR( diff4v, diff5v, diff6v, diff7v, temp4v, temp5v, temp6v, temp7v ); VEC_ADD_ABS( temp0v, zero_s32v, satdv ); VEC_ADD_ABS( temp1v, satdv, satdv ); VEC_ADD_ABS( temp2v, satdv, satdv ); VEC_ADD_ABS( temp3v, satdv, satdv ); VEC_ADD_ABS( temp4v, satdv, satdv ); VEC_ADD_ABS( temp5v, satdv, satdv ); VEC_ADD_ABS( temp6v, satdv, satdv ); VEC_ADD_ABS( temp7v, satdv, satdv ); satdv = vec_sum2s( satdv, zero_s32v ); satdv = vec_splat( satdv, 1 ); vec_ste( satdv, 0, &i_satd ); return i_satd / 2;}/*********************************************************************** * SATD 8x8 **********************************************************************/static int pixel_satd_8x8_altivec( uint8_t *pix1, int i_pix1, uint8_t *pix2, int i_pix2 ){ DECLARE_ALIGNED_16( int i_satd ); PREP_DIFF; vec_s16_t diff0v, diff1v, diff2v, diff3v, diff4v, diff5v, diff6v, diff7v; vec_s16_t temp0v, temp1v, temp2v, temp3v, temp4v, temp5v, temp6v, temp7v; vec_s32_t satdv; PREP_LOAD_SRC( pix1 ); vec_u8_t _offset1v_ = vec_lvsl(0, pix2); vec_u8_t _offset2v_ = vec_lvsl(0, pix2 + i_pix2); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 8, diff0v, offset1v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 8, diff1v, offset2v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 8, diff2v, offset1v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 8, diff3v, offset2v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 8, diff4v, offset1v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 8, diff5v, offset2v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 8, diff6v, offset1v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 8, diff7v, offset2v ); VEC_HADAMAR( diff0v, diff1v, diff2v, diff3v, temp0v, temp1v, temp2v, temp3v ); VEC_HADAMAR( diff4v, diff5v, diff6v, diff7v, temp4v, temp5v, temp6v, temp7v ); VEC_TRANSPOSE_8( temp0v, temp1v, temp2v, temp3v, temp4v, temp5v, temp6v, temp7v, diff0v, diff1v, diff2v, diff3v, diff4v, diff5v, diff6v, diff7v ); VEC_HADAMAR( diff0v, diff1v, diff2v, diff3v, temp0v, temp1v, temp2v, temp3v ); VEC_HADAMAR( diff4v, diff5v, diff6v, diff7v, temp4v, temp5v, temp6v, temp7v ); VEC_ADD_ABS( temp0v, zero_s32v, satdv ); VEC_ADD_ABS( temp1v, satdv, satdv ); VEC_ADD_ABS( temp2v, satdv, satdv ); VEC_ADD_ABS( temp3v, satdv, satdv ); VEC_ADD_ABS( temp4v, satdv, satdv ); VEC_ADD_ABS( temp5v, satdv, satdv ); VEC_ADD_ABS( temp6v, satdv, satdv ); VEC_ADD_ABS( temp7v, satdv, satdv ); satdv = vec_sums( satdv, zero_s32v ); satdv = vec_splat( satdv, 3 ); vec_ste( satdv, 0, &i_satd ); return i_satd / 2;}/*********************************************************************** * SATD 8x16 **********************************************************************/static int pixel_satd_8x16_altivec( uint8_t *pix1, int i_pix1, uint8_t *pix2, int i_pix2 ){ DECLARE_ALIGNED_16( int i_satd ); PREP_DIFF; vec_s16_t diff0v, diff1v, diff2v, diff3v, diff4v, diff5v, diff6v, diff7v; vec_s16_t temp0v, temp1v, temp2v, temp3v, temp4v, temp5v, temp6v, temp7v; vec_s32_t satdv; PREP_LOAD_SRC( pix1 ); vec_u8_t _offset1v_ = vec_lvsl(0, pix2); vec_u8_t _offset2v_ = vec_lvsl(0, pix2 + i_pix2); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 8, diff0v, offset1v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 8, diff1v, offset2v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 8, diff2v, offset1v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 8, diff3v, offset2v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 8, diff4v, offset1v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 8, diff5v, offset2v ); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 8, diff6v , offset1v); VEC_DIFF_H( pix1, i_pix1, pix2, i_pix2, 8, diff7v, offset2v ); VEC_HADAMAR( diff0v, diff1v, diff2v, diff3v, temp0v, temp1v, temp2v, temp3v ); VEC_HADAMAR( diff4v, diff5v, diff6v, diff7v, temp4v, temp5v, temp6v, temp7v ); VEC_TRANSPOSE_8( temp0v, temp1v, temp2v, temp3v, temp4v, temp5v, temp6v, temp7v, diff0v, diff1v, diff2v, diff3v, diff4v, diff5v, diff6v, diff7v );
⌨️ 快捷键说明
复制代码
Ctrl + C
搜索代码
Ctrl + F
全屏模式
F11
切换主题
Ctrl + Shift + D
显示快捷键
?
增大字号
Ctrl + =
减小字号
Ctrl + -