SOURCES: libtheora-x86_64.patch (NEW) - missing file

arekm arekm at pld-linux.org
Thu Oct 9 10:07:29 CEST 2008


Author: arekm                        Date: Thu Oct  9 08:07:29 2008 GMT
Module: SOURCES                       Tag: HEAD
---- Log message:
- missing file

---- Files affected:
SOURCES:
   libtheora-x86_64.patch (NONE -> 1.1)  (NEW)

---- Diffs:

================================================================
Index: SOURCES/libtheora-x86_64.patch
diff -u /dev/null SOURCES/libtheora-x86_64.patch:1.1
--- /dev/null	Thu Oct  9 10:07:29 2008
+++ SOURCES/libtheora-x86_64.patch	Thu Oct  9 10:07:23 2008
@@ -0,0 +1,412 @@
+--- /dev/null	2007-02-13 17:29:53.000000000 +0100
++++ libtheora-1.0RC1/lib/enc/x86_64/dct_decode_mmx.c	2008-08-04 20:37:55.000000000 +0200
+@@ -0,0 +1,409 @@
++/********************************************************************
++ *                                                                  *
++ * THIS FILE IS PART OF THE OggTheora SOFTWARE CODEC SOURCE CODE.   *
++ * USE, DISTRIBUTION AND REPRODUCTION OF THIS LIBRARY SOURCE IS     *
++ * GOVERNED BY A BSD-STYLE SOURCE LICENSE INCLUDED WITH THIS SOURCE *
++ * IN 'COPYING'. PLEASE READ THESE TERMS BEFORE DISTRIBUTING.       *
++ *                                                                  *
++ * THE Theora SOURCE CODE IS COPYRIGHT (C) 2002-2008                *
++ * by the Xiph.Org Foundation http://www.xiph.org/                  *
++ *                                                                  *
++ ********************************************************************
++
++  function:
++  last mod: $Id$
++
++ ********************************************************************/
++
++#include <stdlib.h>
++
++#include "codec_internal.h"
++
++#if defined(USE_ASM)
++
++static const __attribute__((aligned(8),used)) ogg_int64_t OC_V3=
++ 0x0003000300030003LL;
++static const __attribute__((aligned(8),used)) ogg_int64_t OC_V4=
++ 0x0004000400040004LL;
++
++static void loop_filter_v(unsigned char *_pix,int _ystride,
++                          const ogg_int16_t *_ll){
++  long esi;
++  _pix-=_ystride*2;
++  __asm__ __volatile__(
++    /*mm0=0*/
++    "pxor %%mm0,%%mm0\n\t"
++    /*esi=_ystride*3*/
++    "lea (%[ystride],%[ystride],2),%[s]\n\t"
++    /*mm7=_pix[0...8]*/
++    "movq (%[pix]),%%mm7\n\t"
++    /*mm4=_pix[0...8+_ystride*3]*/
++    "movq (%[pix],%[s]),%%mm4\n\t"
++    /*mm6=_pix[0...8]*/
++    "movq %%mm7,%%mm6\n\t"
++    /*Expand unsigned _pix[0...3] to 16 bits.*/
++    "punpcklbw %%mm0,%%mm6\n\t"
++    "movq %%mm4,%%mm5\n\t"
++    /*Expand unsigned _pix[4...8] to 16 bits.*/
++    "punpckhbw %%mm0,%%mm7\n\t"
++    /*Expand other arrays too.*/
++    "punpcklbw %%mm0,%%mm4\n\t"
++    "punpckhbw %%mm0,%%mm5\n\t"
++    /*mm7:mm6=_p[0...8]-_p[0...8+_ystride*3]:*/
++    "psubw %%mm4,%%mm6\n\t"
++    "psubw %%mm5,%%mm7\n\t"
++    /*mm5=mm4=_pix[0...8+_ystride]*/
++    "movq (%[pix],%[ystride]),%%mm4\n\t"
++    /*mm1=mm3=mm2=_pix[0..8]+_ystride*2]*/
++    "movq (%[pix],%[ystride],2),%%mm2\n\t"
++    "movq %%mm4,%%mm5\n\t"
++    "movq %%mm2,%%mm3\n\t"
++    "movq %%mm2,%%mm1\n\t"
++    /*Expand these arrays.*/
++    "punpckhbw %%mm0,%%mm5\n\t"
++    "punpcklbw %%mm0,%%mm4\n\t"
++    "punpckhbw %%mm0,%%mm3\n\t"
++    "punpcklbw %%mm0,%%mm2\n\t"
++    /*Preload...*/
++    "movq %[OC_V3],%%mm0\n\t"
++    /*mm3:mm2=_pix[0...8+_ystride*2]-_pix[0...8+_ystride]*/
++    "psubw %%mm5,%%mm3\n\t"
++    "psubw %%mm4,%%mm2\n\t"
++    /*Scale by 3.*/
++    "pmullw %%mm0,%%mm3\n\t"
++    "pmullw %%mm0,%%mm2\n\t"
++    /*Preload...*/
++    "movq %[OC_V4],%%mm0\n\t"
++    /*f=mm3:mm2==_pix[0...8]-_pix[0...8+_ystride*3]+
++       3*(_pix[0...8+_ystride*2]-_pix[0...8+_ystride])*/
++    "paddw %%mm7,%%mm3\n\t"
++    "paddw %%mm6,%%mm2\n\t"
++    /*Add 4.*/
++    "paddw %%mm0,%%mm3\n\t"
++    "paddw %%mm0,%%mm2\n\t"
++    /*"Divide" by 8.*/
++    "psraw $3,%%mm3\n\t"
++    "psraw $3,%%mm2\n\t"
++    /*Now compute lflim of mm3:mm2 cf. Section 7.10 of the sepc.*/
++    /*Free up mm5.*/
++    "packuswb %%mm5,%%mm4\n\t"
++    /*mm0=L L L L*/
++    "movq (%[ll]),%%mm0\n\t"
++    /*if(R_i<-2L||R_i>2L)R_i=0:*/
++    "movq %%mm2,%%mm5\n\t"
++    "pxor %%mm6,%%mm6\n\t"
++    "movq %%mm0,%%mm7\n\t"
++    "psubw %%mm0,%%mm6\n\t"
++    "psllw $1,%%mm7\n\t"
++    "psllw $1,%%mm6\n\t"
++    /*mm2==R_3 R_2 R_1 R_0*/
++    /*mm5==R_3 R_2 R_1 R_0*/
++    /*mm6==-2L -2L -2L -2L*/
++    /*mm7==2L 2L 2L 2L*/
++    "pcmpgtw %%mm2,%%mm7\n\t"
++    "pcmpgtw %%mm6,%%mm5\n\t"
++    "pand %%mm7,%%mm2\n\t"
++    "movq %%mm0,%%mm7\n\t"
++    "pand %%mm5,%%mm2\n\t"
++    "psllw $1,%%mm7\n\t"
++    "movq %%mm3,%%mm5\n\t"
++    /*mm3==R_7 R_6 R_5 R_4*/
++    /*mm5==R_7 R_6 R_5 R_4*/
++    /*mm6==-2L -2L -2L -2L*/
++    /*mm7==2L 2L 2L 2L*/
++    "pcmpgtw %%mm3,%%mm7\n\t"
++    "pcmpgtw %%mm6,%%mm5\n\t"
++    "pand %%mm7,%%mm3\n\t"
++    "movq %%mm0,%%mm7\n\t"
++    "pand %%mm5,%%mm3\n\t"
++    /*if(R_i<-L)R_i'=R_i+2L;
++      if(R_i>L)R_i'=R_i-2L;
++      if(R_i<-L||R_i>L)R_i=-R_i':*/
++    "psraw $1,%%mm6\n\t"
++    "movq %%mm2,%%mm5\n\t"
++    "psllw $1,%%mm7\n\t"
++    /*mm2==R_3 R_2 R_1 R_0*/
++    /*mm5==R_3 R_2 R_1 R_0*/
++    /*mm6==-L -L -L -L*/
++    /*mm0==L L L L*/
++    /*mm5=R_i>L?FF:00*/
++    "pcmpgtw %%mm0,%%mm5\n\t"
++    /*mm6=-L>R_i?FF:00*/
++    "pcmpgtw %%mm2,%%mm6\n\t"
++    /*mm7=R_i>L?2L:0*/
++    "pand %%mm5,%%mm7\n\t"
++    /*mm2=R_i>L?R_i-2L:R_i*/
++    "psubw %%mm7,%%mm2\n\t"
++    "movq %%mm0,%%mm7\n\t"
++    /*mm5=-L>R_i||R_i>L*/
++    "por %%mm6,%%mm5\n\t"
++    "psllw $1,%%mm7\n\t"
++    /*mm7=-L>R_i?2L:0*/
++    "pand %%mm6,%%mm7\n\t"
++    "pxor %%mm6,%%mm6\n\t"
++    /*mm2=-L>R_i?R_i+2L:R_i*/
++    "paddw %%mm7,%%mm2\n\t"
++    "psubw %%mm0,%%mm6\n\t"
++    /*mm5=-L>R_i||R_i>L?-R_i':0*/
++    "pand %%mm2,%%mm5\n\t"
++    "movq %%mm0,%%mm7\n\t"
++    /*mm2=-L>R_i||R_i>L?0:R_i*/
++    "psubw %%mm5,%%mm2\n\t"
++    "psllw $1,%%mm7\n\t"
++    /*mm2=-L>R_i||R_i>L?-R_i':R_i*/
++    "psubw %%mm5,%%mm2\n\t"
++    "movq %%mm3,%%mm5\n\t"
++    /*mm3==R_7 R_6 R_5 R_4*/
++    /*mm5==R_7 R_6 R_5 R_4*/
++    /*mm6==-L -L -L -L*/
++    /*mm0==L L L L*/
++    /*mm6=-L>R_i?FF:00*/
++    "pcmpgtw %%mm3,%%mm6\n\t"
++    /*mm5=R_i>L?FF:00*/
++    "pcmpgtw %%mm0,%%mm5\n\t"
++    /*mm7=R_i>L?2L:0*/
++    "pand %%mm5,%%mm7\n\t"
++    /*mm2=R_i>L?R_i-2L:R_i*/
++    "psubw %%mm7,%%mm3\n\t"
++    "psllw $1,%%mm0\n\t"
++    /*mm5=-L>R_i||R_i>L*/
++    "por %%mm6,%%mm5\n\t"
++    /*mm0=-L>R_i?2L:0*/
++    "pand %%mm6,%%mm0\n\t"
++    /*mm3=-L>R_i?R_i+2L:R_i*/
++    "paddw %%mm0,%%mm3\n\t"
++    /*mm5=-L>R_i||R_i>L?-R_i':0*/
++    "pand %%mm3,%%mm5\n\t"
++    /*mm2=-L>R_i||R_i>L?0:R_i*/
++    "psubw %%mm5,%%mm3\n\t"
++    /*mm2=-L>R_i||R_i>L?-R_i':R_i*/
++    "psubw %%mm5,%%mm3\n\t"
++    /*Unfortunately, there's no unsigned byte+signed byte with unsigned
++       saturation op code, so we have to promote things back 16 bits.*/
++    "pxor %%mm0,%%mm0\n\t"
++    "movq %%mm4,%%mm5\n\t"
++    "punpcklbw %%mm0,%%mm4\n\t"
++    "punpckhbw %%mm0,%%mm5\n\t"
++    "movq %%mm1,%%mm6\n\t"
++    "punpcklbw %%mm0,%%mm1\n\t"
++    "punpckhbw %%mm0,%%mm6\n\t"
++    /*_pix[0...8+_ystride]+=R_i*/
++    "paddw %%mm2,%%mm4\n\t"
++    "paddw %%mm3,%%mm5\n\t"
++    /*_pix[0...8+_ystride*2]-=R_i*/
++    "psubw %%mm2,%%mm1\n\t"
++    "psubw %%mm3,%%mm6\n\t"
++    "packuswb %%mm5,%%mm4\n\t"
++    "packuswb %%mm6,%%mm1\n\t"
++    /*Write it back out.*/
++    "movq %%mm4,(%[pix],%[ystride])\n\t"
++    "movq %%mm1,(%[pix],%[ystride],2)\n\t"
++    :[s]"=&S"(esi)
++    :[pix]"r"(_pix),[ystride]"r"((long)_ystride),[ll]"r"(_ll),
++     [OC_V3]"m"(OC_V3),[OC_V4]"m"(OC_V4)
++    :"memory"
++  );
++}
++
++/*This code implements the bulk of loop_filter_h().
++  Data are striped p0 p1 p2 p3 ... p0 p1 p2 p3 ..., so in order to load all
++   four p0's to one register we must transpose the values in four mmx regs.
++  When half is done we repeat this for the rest.*/
++static void loop_filter_h4(unsigned char *_pix,long _ystride,
++                           const ogg_int16_t *_ll){
++  long esi;
++  long edi;
++  __asm__ __volatile__(
++    /*x x x x 3 2 1 0*/
++    "movd (%[pix]),%%mm0\n\t"
++    /*esi=_ystride*3*/
++    "lea (%[ystride],%[ystride],2),%[s]\n\t"
++    /*x x x x 7 6 5 4*/
++    "movd (%[pix],%[ystride]),%%mm1\n\t"
++    /*x x x x B A 9 8*/
++    "movd (%[pix],%[ystride],2),%%mm2\n\t"
++    /*x x x x F E D C*/
++    "movd (%[pix],%[s]),%%mm3\n\t"
++    /*mm0=7 3 6 2 5 1 4 0*/
++    "punpcklbw %%mm1,%%mm0\n\t"
++    /*mm2=F B E A D 9 C 8*/
++    "punpcklbw %%mm3,%%mm2\n\t"
++    /*mm1=7 3 6 2 5 1 4 0*/
++    "movq %%mm0,%%mm1\n\t"
++    /*mm0=F B 7 3 E A 6 2*/
++    "punpckhwd %%mm2,%%mm0\n\t"
++    /*mm1=D 9 5 1 C 8 4 0*/
++    "punpcklwd %%mm2,%%mm1\n\t"
++    "pxor %%mm7,%%mm7\n\t"
++    /*mm5=D 9 5 1 C 8 4 0*/
++    "movq %%mm1,%%mm5\n\t"
++    /*mm1=x C x 8 x 4 x 0==pix[0]*/
++    "punpcklbw %%mm7,%%mm1\n\t"
++    /*mm5=x D x 9 x 5 x 1==pix[1]*/
++    "punpckhbw %%mm7,%%mm5\n\t"
++    /*mm3=F B 7 3 E A 6 2*/
++    "movq %%mm0,%%mm3\n\t"
++    /*mm0=x E x A x 6 x 2==pix[2]*/
++    "punpcklbw %%mm7,%%mm0\n\t"
++    /*mm3=x F x B x 7 x 3==pix[3]*/
++    "punpckhbw %%mm7,%%mm3\n\t"
++    /*mm1=mm1-mm3==pix[0]-pix[3]*/
++    "psubw %%mm3,%%mm1\n\t"
++    /*Save a copy of pix[2] for later.*/
++    "movq %%mm0,%%mm4\n\t"
++    /*mm0=mm0-mm5==pix[2]-pix[1]*/
++    "psubw %%mm5,%%mm0\n\t"
++    /*Scale by 3.*/
++    "pmullw %[OC_V3],%%mm0\n\t"
++    /*f=mm1==_pix[0]-_pix[3]+ 3*(_pix[2]-_pix[1])*/
++    "paddw %%mm1,%%mm0\n\t"
++    /*Add 4.*/
++    "paddw %[OC_V4],%%mm0\n\t"
++    /*"Divide" by 8, producing the residuals R_i.*/
++    "psraw $3,%%mm0\n\t"
++    /*Now compute lflim of mm0 cf. Section 7.10 of the sepc.*/
++    /*mm6=L L L L*/
++    "movq (%[ll]),%%mm6\n\t"
++    /*if(R_i<-2L||R_i>2L)R_i=0:*/
++    "movq %%mm0,%%mm1\n\t"
++    "pxor %%mm2,%%mm2\n\t"
++    "movq %%mm6,%%mm3\n\t"
++    "psubw %%mm6,%%mm2\n\t"
++    "psllw $1,%%mm3\n\t"
++    "psllw $1,%%mm2\n\t"
++    /*mm0==R_3 R_2 R_1 R_0*/
++    /*mm1==R_3 R_2 R_1 R_0*/
++    /*mm2==-2L -2L -2L -2L*/
++    /*mm3==2L 2L 2L 2L*/
++    "pcmpgtw %%mm0,%%mm3\n\t"
++    "pcmpgtw %%mm2,%%mm1\n\t"
++    "pand %%mm3,%%mm0\n\t"
++    "pand %%mm1,%%mm0\n\t"
++    /*if(R_i<-L)R_i'=R_i+2L;
++      if(R_i>L)R_i'=R_i-2L;
++      if(R_i<-L||R_i>L)R_i=-R_i':*/
++    "psraw $1,%%mm2\n\t"
++    "movq %%mm0,%%mm1\n\t"
++    "movq %%mm6,%%mm3\n\t"
++    /*mm0==R_3 R_2 R_1 R_0*/
++    /*mm1==R_3 R_2 R_1 R_0*/
++    /*mm2==-L -L -L -L*/
++    /*mm6==L L L L*/
++    /*mm2=-L>R_i?FF:00*/
++    "pcmpgtw %%mm0,%%mm2\n\t"
++    /*mm1=R_i>L?FF:00*/
++    "pcmpgtw %%mm6,%%mm1\n\t"
++    /*mm3=2L 2L 2L 2L*/
++    "psllw $1,%%mm3\n\t"
++    /*mm6=2L 2L 2L 2L*/
++    "psllw $1,%%mm6\n\t"
++    /*mm3=R_i>L?2L:0*/
++    "pand %%mm1,%%mm3\n\t"
++    /*mm6=-L>R_i?2L:0*/
++    "pand %%mm2,%%mm6\n\t"
++    /*mm0=R_i>L?R_i-2L:R_i*/
++    "psubw %%mm3,%%mm0\n\t"
++    /*mm1=-L>R_i||R_i>L*/
++    "por %%mm2,%%mm1\n\t"
++    /*mm0=-L>R_i?R_i+2L:R_i*/
++    "paddw %%mm6,%%mm0\n\t"
++    /*mm1=-L>R_i||R_i>L?R_i':0*/
++    "pand %%mm0,%%mm1\n\t"
++    /*mm0=-L>R_i||R_i>L?0:R_i*/
++    "psubw %%mm1,%%mm0\n\t"
++    /*mm0=-L>R_i||R_i>L?-R_i':R_i*/
++    "psubw %%mm1,%%mm0\n\t"
++    /*_pix[1]+=R_i;*/
++    "paddw %%mm0,%%mm5\n\t"
++    /*_pix[2]-=R_i;*/
++    "psubw %%mm0,%%mm4\n\t"
++    /*mm5=x x x x D 9 5 1*/
++    "packuswb %%mm7,%%mm5\n\t"
++    /*mm4=x x x x E A 6 2*/
++    "packuswb %%mm7,%%mm4\n\t"
++    /*mm5=E D A 9 6 5 2 1*/
++    "punpcklbw %%mm4,%%mm5\n\t"
++    /*edi=6 5 2 1*/
++    "movd %%mm5,%%edi\n\t"
++    "movw %%di,1(%[pix])\n\t"
++    /*Why is there such a big stall here?*/
++    "psrlq $32,%%mm5\n\t"
++    "shrl $16,%%edi\n\t"
++    "movw %%di,1(%[pix],%[ystride])\n\t"
++    /*edi=E D A 9*/
++    "movd %%mm5,%%edi\n\t"
++    "movw %%di,1(%[pix],%[ystride],2)\n\t"
++    "shrl $16,%%edi\n\t"
++    "movw %%di,1(%[pix],%[s])\n\t"
++    :[s]"=&S"(esi),[d]"=&D"(edi),
++     [pix]"+r"(_pix),[ystride]"+r"(_ystride),[ll]"+r"(_ll)
++    :[OC_V3]"m"(OC_V3),[OC_V4]"m"(OC_V4)
++    :"memory"
++  );
++}
++
++static void loop_filter_h(unsigned char *_pix,int _ystride,
++                          const ogg_int16_t *_ll){
++  _pix-=2;
++  loop_filter_h4(_pix,_ystride,_ll);
++  loop_filter_h4(_pix+(_ystride<<2),_ystride,_ll);
++}
++
++static void loop_filter_mmx(PB_INSTANCE *pbi, int FLimit){
++  int j;
++  ogg_int16_t __attribute__((aligned(8)))  ll[4];
++  unsigned char *cp = pbi->display_fragments;
++  ogg_uint32_t *bp = pbi->recon_pixel_index_table;
++
++  if ( FLimit == 0 ) return;
++  ll[0]=ll[1]=ll[2]=ll[3]=FLimit;
++
++  for ( j = 0; j < 3 ; j++){
++    ogg_uint32_t *bp_begin = bp;
++    ogg_uint32_t *bp_end;
++    int stride;
++    int h;
++
++    switch(j) {
++    case 0: /* y */
++      bp_end = bp + pbi->YPlaneFragments;
++      h = pbi->HFragments;
++      stride = pbi->YStride;
++      break;
++    default: /* u,v, 4:20 specific */
++      bp_end = bp + pbi->UVPlaneFragments;
++      h = pbi->HFragments >> 1;
++      stride = pbi->UVStride;
++      break;
++    }
++
++    while(bp<bp_end){
++      ogg_uint32_t *bp_left = bp;
++      ogg_uint32_t *bp_right = bp + h;
++      while(bp<bp_right){
++        if(cp[0]){
++          if(bp>bp_left)
++            loop_filter_h(&pbi->LastFrameRecon[bp[0]],stride,ll);
++          if(bp_left>bp_begin)
++            loop_filter_v(&pbi->LastFrameRecon[bp[0]],stride,ll);
++          if(bp+1<bp_right && !cp[1])
++            loop_filter_h(&pbi->LastFrameRecon[bp[0]]+8,stride,ll);
++          if(bp+h<bp_end && !cp[h])
++            loop_filter_v(&pbi->LastFrameRecon[bp[h]],stride,ll);
++        }
++        bp++;
++        cp++;
++      }
++    }
++  }
++
++  __asm__ __volatile__("emms\n\t");
++}
++
++/* install our implementation in the function table */
++void dsp_mmx_dct_decode_init(DspFunctions *funcs)
++{
++  funcs->LoopFilter = loop_filter_mmx;
++}
++
++#endif /* USE_ASM */
================================================================


More information about the pld-cvs-commit mailing list