⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 dsputil_altivec.c

📁 tcpmp播放器的flv插件
💻 C
📖 第 1 页 / 共 5 页
字号:
  16x8 works with 16 elements ; it allows to avoid replicating  loads, and give the compiler more rooms for scheduling.  It's only used from inside hadamard8_diff16_altivec.    Unfortunately, it seems gcc-3.3 is a bit dumb, and  the compiled code has a LOT of spill code, it seems  gcc (unlike xlc) cannot keep everything in registers  by itself. The following code include hand-made  registers allocation. It's not clean, but on  a 7450 the resulting code is much faster (best case  fall from 700+ cycles to 550).    xlc doesn't add spill code, but it doesn't know how to  schedule for the 7450, and its code isn't much faster than  gcc-3.3 on the 7450 (but uses 25% less instructions...)    On the 970, the hand-made RA is still a win (arount 690  vs. around 780), but xlc goes to around 660 on the  regular C code...*/static int hadamard8_diff16x8_altivec(/*MpegEncContext*/ void *s, uint8_t *dst, uint8_t *src, int stride, int h) {  int sum;  register vector signed short    temp0 asm ("v0"),    temp1 asm ("v1"),    temp2 asm ("v2"),    temp3 asm ("v3"),    temp4 asm ("v4"),    temp5 asm ("v5"),    temp6 asm ("v6"),    temp7 asm ("v7");  register vector signed short    temp0S asm ("v8"),    temp1S asm ("v9"),    temp2S asm ("v10"),    temp3S asm ("v11"),    temp4S asm ("v12"),    temp5S asm ("v13"),    temp6S asm ("v14"),    temp7S asm ("v15");  register const_vector unsigned char vzero asm ("v31")= (const_vector unsigned char)vec_splat_u8(0);  {    register const_vector signed short vprod1 asm ("v16")= (const_vector signed short)AVV( 1,-1, 1,-1, 1,-1, 1,-1);    register const_vector signed short vprod2 asm ("v17")= (const_vector signed short)AVV( 1, 1,-1,-1, 1, 1,-1,-1);    register const_vector signed short vprod3 asm ("v18")= (const_vector signed short)AVV( 1, 1, 1, 1,-1,-1,-1,-1);    register const_vector unsigned char perm1 asm ("v19")= (const_vector unsigned char)      AVV(0x02, 0x03, 0x00, 0x01,       0x06, 0x07, 0x04, 0x05,       0x0A, 0x0B, 0x08, 0x09,       0x0E, 0x0F, 0x0C, 0x0D);    register const_vector unsigned char perm2 asm ("v20")= (const_vector unsigned char)      AVV(0x04, 0x05, 0x06, 0x07,       0x00, 0x01, 0x02, 0x03,       0x0C, 0x0D, 0x0E, 0x0F,       0x08, 0x09, 0x0A, 0x0B);    register const_vector unsigned char perm3 asm ("v21")= (const_vector unsigned char)      AVV(0x08, 0x09, 0x0A, 0x0B,       0x0C, 0x0D, 0x0E, 0x0F,       0x00, 0x01, 0x02, 0x03,       0x04, 0x05, 0x06, 0x07);#define ONEITERBUTTERFLY(i, res1, res2)					\    {									\      register vector unsigned char src1 asm ("v22"), src2 asm ("v23"); \      register vector unsigned char dst1 asm ("v24"), dst2 asm ("v25"); \      src1 = vec_ld(stride * i, src);					\      src2 = vec_ld((stride * i) + 16, src);				\      register vector unsigned char srcO asm ("v22") = vec_perm(src1, src2, vec_lvsl(stride * i, src)); \      dst1 = vec_ld(stride * i, dst);					\      dst2 = vec_ld((stride * i) + 16, dst);				\      register vector unsigned char dstO asm ("v23") = vec_perm(dst1, dst2, vec_lvsl(stride * i, dst)); \      /* promote the unsigned chars to signed shorts */			\      register vector signed short srcV asm ("v24") =                   \	(vector signed short)vec_mergeh((vector signed char)vzero, (vector signed char)srcO); \      register vector signed short dstV asm ("v25") =                   \	(vector signed short)vec_mergeh((vector signed char)vzero, (vector signed char)dstO); \      register vector signed short srcW asm ("v26") =                   \	(vector signed short)vec_mergel((vector signed char)vzero, (vector signed char)srcO); \      register vector signed short dstW asm ("v27") =                   \	(vector signed short)vec_mergel((vector signed char)vzero, (vector signed char)dstO); \      /* substractions inside the first butterfly */			\      register vector signed short but0 asm ("v28") = vec_sub(srcV, dstV); \      register vector signed short but0S asm ("v29") = vec_sub(srcW, dstW); \      register vector signed short op1 asm ("v30") = vec_perm(but0, but0, perm1); \      register vector signed short but1 asm ("v22") = vec_mladd(but0, vprod1, op1); \      register vector signed short op1S asm ("v23") = vec_perm(but0S, but0S, perm1); \      register vector signed short but1S asm ("v24") = vec_mladd(but0S, vprod1, op1S); \      register vector signed short op2 asm ("v25") = vec_perm(but1, but1, perm2); \      register vector signed short but2 asm ("v26") = vec_mladd(but1, vprod2, op2); \      register vector signed short op2S asm ("v27") = vec_perm(but1S, but1S, perm2); \      register vector signed short but2S asm ("v28") = vec_mladd(but1S, vprod2, op2S); \      register vector signed short op3 asm ("v29") = vec_perm(but2, but2, perm3); \      res1 = vec_mladd(but2, vprod3, op3);				\      register vector signed short op3S asm ("v30") = vec_perm(but2S, but2S, perm3); \      res2 = vec_mladd(but2S, vprod3, op3S);				\    }    ONEITERBUTTERFLY(0, temp0, temp0S);    ONEITERBUTTERFLY(1, temp1, temp1S);    ONEITERBUTTERFLY(2, temp2, temp2S);    ONEITERBUTTERFLY(3, temp3, temp3S);    ONEITERBUTTERFLY(4, temp4, temp4S);    ONEITERBUTTERFLY(5, temp5, temp5S);    ONEITERBUTTERFLY(6, temp6, temp6S);    ONEITERBUTTERFLY(7, temp7, temp7S);  }#undef ONEITERBUTTERFLY  {    register vector signed int vsum;    register vector signed short line0 = vec_add(temp0, temp1);    register vector signed short line1 = vec_sub(temp0, temp1);    register vector signed short line2 = vec_add(temp2, temp3);    register vector signed short line3 = vec_sub(temp2, temp3);    register vector signed short line4 = vec_add(temp4, temp5);    register vector signed short line5 = vec_sub(temp4, temp5);    register vector signed short line6 = vec_add(temp6, temp7);    register vector signed short line7 = vec_sub(temp6, temp7);          register vector signed short line0B = vec_add(line0, line2);    register vector signed short line2B = vec_sub(line0, line2);    register vector signed short line1B = vec_add(line1, line3);    register vector signed short line3B = vec_sub(line1, line3);    register vector signed short line4B = vec_add(line4, line6);    register vector signed short line6B = vec_sub(line4, line6);    register vector signed short line5B = vec_add(line5, line7);    register vector signed short line7B = vec_sub(line5, line7);          register vector signed short line0C = vec_add(line0B, line4B);    register vector signed short line4C = vec_sub(line0B, line4B);    register vector signed short line1C = vec_add(line1B, line5B);    register vector signed short line5C = vec_sub(line1B, line5B);    register vector signed short line2C = vec_add(line2B, line6B);    register vector signed short line6C = vec_sub(line2B, line6B);    register vector signed short line3C = vec_add(line3B, line7B);    register vector signed short line7C = vec_sub(line3B, line7B);          vsum = vec_sum4s(vec_abs(line0C), vec_splat_s32(0));    vsum = vec_sum4s(vec_abs(line1C), vsum);    vsum = vec_sum4s(vec_abs(line2C), vsum);    vsum = vec_sum4s(vec_abs(line3C), vsum);    vsum = vec_sum4s(vec_abs(line4C), vsum);    vsum = vec_sum4s(vec_abs(line5C), vsum);    vsum = vec_sum4s(vec_abs(line6C), vsum);    vsum = vec_sum4s(vec_abs(line7C), vsum);    register vector signed short line0S = vec_add(temp0S, temp1S);    register vector signed short line1S = vec_sub(temp0S, temp1S);    register vector signed short line2S = vec_add(temp2S, temp3S);    register vector signed short line3S = vec_sub(temp2S, temp3S);    register vector signed short line4S = vec_add(temp4S, temp5S);    register vector signed short line5S = vec_sub(temp4S, temp5S);    register vector signed short line6S = vec_add(temp6S, temp7S);    register vector signed short line7S = vec_sub(temp6S, temp7S);    register vector signed short line0BS = vec_add(line0S, line2S);    register vector signed short line2BS = vec_sub(line0S, line2S);    register vector signed short line1BS = vec_add(line1S, line3S);    register vector signed short line3BS = vec_sub(line1S, line3S);    register vector signed short line4BS = vec_add(line4S, line6S);    register vector signed short line6BS = vec_sub(line4S, line6S);    register vector signed short line5BS = vec_add(line5S, line7S);    register vector signed short line7BS = vec_sub(line5S, line7S);    register vector signed short line0CS = vec_add(line0BS, line4BS);    register vector signed short line4CS = vec_sub(line0BS, line4BS);    register vector signed short line1CS = vec_add(line1BS, line5BS);    register vector signed short line5CS = vec_sub(line1BS, line5BS);    register vector signed short line2CS = vec_add(line2BS, line6BS);    register vector signed short line6CS = vec_sub(line2BS, line6BS);    register vector signed short line3CS = vec_add(line3BS, line7BS);    register vector signed short line7CS = vec_sub(line3BS, line7BS);    vsum = vec_sum4s(vec_abs(line0CS), vsum);    vsum = vec_sum4s(vec_abs(line1CS), vsum);    vsum = vec_sum4s(vec_abs(line2CS), vsum);    vsum = vec_sum4s(vec_abs(line3CS), vsum);    vsum = vec_sum4s(vec_abs(line4CS), vsum);    vsum = vec_sum4s(vec_abs(line5CS), vsum);    vsum = vec_sum4s(vec_abs(line6CS), vsum);    vsum = vec_sum4s(vec_abs(line7CS), vsum);    vsum = vec_sums(vsum, (vector signed int)vzero);    vsum = vec_splat(vsum, 3);    vec_ste(vsum, 0, &sum);  }  return sum;}int hadamard8_diff16_altivec(/*MpegEncContext*/ void *s, uint8_t *dst, uint8_t *src, int stride, int h){POWERPC_PERF_DECLARE(altivec_hadamard8_diff16_num, 1);  int score;POWERPC_PERF_START_COUNT(altivec_hadamard8_diff16_num, 1);  score = hadamard8_diff16x8_altivec(s, dst, src, stride, 8);  if (h==16) {    dst += 8*stride;    src += 8*stride;    score += hadamard8_diff16x8_altivec(s, dst, src, stride, 8);  }POWERPC_PERF_STOP_COUNT(altivec_hadamard8_diff16_num, 1);  return score;}#endif //CONFIG_DARWINint has_altivec(void){#ifdef __AMIGAOS4__	ULONG result = 0;	extern struct ExecIFace *IExec;	IExec->GetCPUInfoTags(GCIT_VectorUnit, &result, TAG_DONE);	if (result == VECTORTYPE_ALTIVEC) return 1;	return 0;#else /* __AMIGAOS4__ */#ifdef CONFIG_DARWIN    int sels[2] = {CTL_HW, HW_VECTORUNIT};    int has_vu = 0;    size_t len = sizeof(has_vu);    int err;    err = sysctl(sels, 2, &has_vu, &len, NULL, 0);    if (err == 0) return (has_vu != 0);#else /* CONFIG_DARWIN *//* no Darwin, do it the brute-force way *//* this is borrowed from the libmpeg2 library */    {      signal (SIGILL, sigill_handler);      if (sigsetjmp (jmpbuf, 1)) {        signal (SIGILL, SIG_DFL);      } else {        canjump = 1;                asm volatile ("mtspr 256, %0\n\t"                      "vand %%v0, %%v0, %%v0"                      :                      : "r" (-1));                signal (SIGILL, SIG_DFL);        return 1;      }    }#endif /* CONFIG_DARWIN */    return 0;#endif /* __AMIGAOS4__ */}/* next one assumes that ((line_size % 8) == 0) */void avg_pixels8_xy2_altivec(uint8_t *block, const uint8_t *pixels, int line_size, int h){POWERPC_PERF_DECLARE(altivec_avg_pixels8_xy2_num, 1);#ifdef ALTIVEC_USE_REFERENCE_C_CODE    int j;POWERPC_PERF_START_COUNT(altivec_avg_pixels8_xy2_num, 1); for (j = 0; j < 2; j++) {   int             i;   const uint32_t  a = (((const struct unaligned_32 *) (pixels))->l);   const uint32_t  b = (((const struct unaligned_32 *) (pixels + 1))->l);   uint32_t        l0 = (a & 0x03030303UL) + (b & 0x03030303UL) + 0x02020202UL;   uint32_t        h0 = ((a & 0xFCFCFCFCUL) >> 2) + ((b & 0xFCFCFCFCUL) >> 2);   uint32_t        l1, h1;   pixels += line_size;   for (i = 0; i < h; i += 2) {     uint32_t        a = (((const struct unaligned_32 *) (pixels))->l);     uint32_t        b = (((const struct unaligned_32 *) (pixels + 1))->l);     l1 = (a & 0x03030303UL) + (b & 0x03030303UL);     h1 = ((a & 0xFCFCFCFCUL) >> 2) + ((b & 0xFCFCFCFCUL) >> 2);     *((uint32_t *) block) = rnd_avg32(*((uint32_t *) block), h0 + h1 + (((l0 + l1) >> 2) & 0x0F0F0F0FUL));     pixels += line_size;     block += line_size;     a = (((const struct unaligned_32 *) (pixels))->l);     b = (((const struct unaligned_32 *) (pixels + 1))->l);     l0 = (a & 0x03030303UL) + (b & 0x03030303UL) + 0x02020202UL;     h0 = ((a & 0xFCFCFCFCUL) >> 2) + ((b & 0xFCFCFCFCUL) >> 2);     *((uint32_t *) block) = rnd_avg32(*((uint32_t *) block), h0 + h1 + (((l0 + l1) >> 2) & 0x0F0F0F0FUL));     pixels += line_size;     block += line_size;   } pixels += 4 - line_size * (h + 1);   block += 4 - line_size * h; }POWERPC_PERF_STOP_COUNT(altivec_avg_pixels8_xy2_num, 1);#else /* ALTIVEC_USE_REFERENCE_C_CODE */   register int i;   register vector unsigned char     pixelsv1, pixelsv2,     pixelsavg;   register vector unsigned char     blockv, temp1, temp2, blocktemp;   register vector unsigned short     pixelssum1, pixelssum2, temp3;   register const_vector unsigned char vczero = (const_vector unsigned char)vec_splat_u8(0);   register const_vector unsigned short vctwo = (const_vector unsigned short)vec_splat_u16(2);      temp1 = vec_ld(0, pixels);   temp2 = vec_ld(16, pixels);   pixelsv1 = vec_perm(temp1, temp2, vec_lvsl(0, pixels));   if ((((unsigned long)pixels) & 0x0000000F) ==  0x0000000F)   {     pixelsv2 = temp2;   }   else   {     pixelsv2 = vec_perm(temp1, temp2, vec_lvsl(1, pixels));   }   pixelsv1 = vec_mergeh(vczero, pixelsv1);   pixelsv2 = vec_mergeh(vczero, pixelsv2);   pixelssum1 = vec_add((vector unsigned short)pixelsv1,                        (vector unsigned short)pixelsv2);   pixelssum1 = vec_add(pixelssum1, vctwo);   POWERPC_PERF_START_COUNT(altivec_avg_pixels8_xy2_num, 1);    for (i = 0; i < h ; i++) {     int rightside = ((unsigned long)block & 0x0000000F);     blockv = vec_ld(0, block);     temp1 = vec_ld(line_size, pixels);     temp2 = vec_ld(line_size + 16, pixels);     pixelsv1 = vec_perm(temp1, temp2, vec_lvsl(line_size, pixels));     if (((((unsigned long)pixels) + line_size) & 0x0000000F) ==  0x0000000F)     {       pixelsv2 = temp2;     }     else     {       pixelsv2 = vec_perm(temp1, temp2, vec_lvsl(line_size + 1, pixels));     }     pixelsv1 = vec_mergeh(vczero, pixelsv1);     pixelsv2 = vec_mergeh(vczero, pixelsv2);     pixelssum2 = vec_add((vector unsigned short)pixelsv1,                          (vector unsigned short)pixelsv2);     temp3 = vec_add(pixelssum1, pixelssum2);     temp3 = vec_sra(temp3, vctwo);     pixelssum1 = vec_add(pixelssum2, vctwo);     pixelsavg = vec_packsu(temp3, (vector unsigned short) vczero);          if (rightside)     {       blocktemp = vec_perm(blockv, pixelsavg, vcprm(0, 1, s0, s1));     }     else     {       blocktemp = vec_perm(blockv, pixelsavg, vcprm(s0, s1, 2, 3));     }          blockv = vec_avg(blocktemp, blockv);     vec_st(blockv, 0, block);          block += line_size;     pixels += line_size;   }   POWERPC_PERF_STOP_COUNT(altivec_avg_pixels8_xy2_

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -