media/libtheora/lib/x86_vc/mmxloop.h

changeset 0
6474c204b198
     1.1 --- /dev/null	Thu Jan 01 00:00:00 1970 +0000
     1.2 +++ b/media/libtheora/lib/x86_vc/mmxloop.h	Wed Dec 31 06:09:35 2014 +0100
     1.3 @@ -0,0 +1,219 @@
     1.4 +#if !defined(_x86_vc_mmxloop_H)
     1.5 +# define _x86_vc_mmxloop_H (1)
     1.6 +# include <stddef.h>
     1.7 +# include "x86int.h"
     1.8 +
     1.9 +#if defined(OC_X86_ASM)
    1.10 +
    1.11 +/*On entry, mm0={a0,...,a7}, mm1={b0,...,b7}, mm2={c0,...,c7}, mm3={d0,...d7}.
    1.12 +  On exit, mm1={b0+lflim(R_0,L),...,b7+lflim(R_7,L)} and
    1.13 +   mm2={c0-lflim(R_0,L),...,c7-lflim(R_7,L)}; mm0 and mm3 are clobbered.*/
    1.14 +#define OC_LOOP_FILTER8_MMX __asm{ \
    1.15 +  /*mm7=0*/ \
    1.16 +  __asm pxor mm7,mm7 \
    1.17 +  /*mm6:mm0={a0,...,a7}*/ \
    1.18 +  __asm movq mm6,mm0 \
    1.19 +  __asm punpcklbw mm0,mm7 \
    1.20 +  __asm punpckhbw mm6,mm7 \
    1.21 +  /*mm3:mm5={d0,...,d7}*/ \
    1.22 +  __asm movq mm5,mm3 \
    1.23 +  __asm punpcklbw mm3,mm7 \
    1.24 +  __asm punpckhbw mm5,mm7 \
    1.25 +  /*mm6:mm0={a0-d0,...,a7-d7}*/ \
    1.26 +  __asm psubw mm0,mm3 \
    1.27 +  __asm psubw mm6,mm5 \
    1.28 +  /*mm3:mm1={b0,...,b7}*/ \
    1.29 +  __asm movq mm3,mm1 \
    1.30 +  __asm punpcklbw mm1,mm7 \
    1.31 +  __asm movq mm4,mm2 \
    1.32 +  __asm punpckhbw mm3,mm7 \
    1.33 +  /*mm5:mm4={c0,...,c7}*/ \
    1.34 +  __asm movq mm5,mm2 \
    1.35 +  __asm punpcklbw mm4,mm7 \
    1.36 +  __asm punpckhbw mm5,mm7 \
    1.37 +  /*mm7={3}x4 \
    1.38 +    mm5:mm4={c0-b0,...,c7-b7}*/ \
    1.39 +  __asm pcmpeqw mm7,mm7 \
    1.40 +  __asm psubw mm4,mm1 \
    1.41 +  __asm psrlw mm7,14 \
    1.42 +  __asm psubw mm5,mm3 \
    1.43 +  /*Scale by 3.*/ \
    1.44 +  __asm pmullw mm4,mm7 \
    1.45 +  __asm pmullw mm5,mm7 \
    1.46 +  /*mm7={4}x4 \
    1.47 +    mm5:mm4=f={a0-d0+3*(c0-b0),...,a7-d7+3*(c7-b7)}*/ \
    1.48 +  __asm psrlw mm7,1 \
    1.49 +  __asm paddw mm4,mm0 \
    1.50 +  __asm psllw mm7,2 \
    1.51 +  __asm movq mm0,[LL] \
    1.52 +  __asm paddw mm5,mm6 \
    1.53 +  /*R_i has the range [-127,128], so we compute -R_i instead. \
    1.54 +    mm4=-R_i=-(f+4>>3)=0xFF^(f-4>>3)*/ \
    1.55 +  __asm psubw mm4,mm7 \
    1.56 +  __asm psubw mm5,mm7 \
    1.57 +  __asm psraw mm4,3 \
    1.58 +  __asm psraw mm5,3 \
    1.59 +  __asm pcmpeqb mm7,mm7 \
    1.60 +  __asm packsswb mm4,mm5 \
    1.61 +  __asm pxor mm6,mm6 \
    1.62 +  __asm pxor mm4,mm7 \
    1.63 +  __asm packuswb mm1,mm3 \
    1.64 +  /*Now compute lflim of -mm4 cf. Section 7.10 of the sepc.*/ \
    1.65 +  /*There's no unsigned byte+signed byte with unsigned saturation op code, so \
    1.66 +     we have to split things by sign (the other option is to work in 16 bits, \
    1.67 +     but working in 8 bits gives much better parallelism). \
    1.68 +    We compute abs(R_i), but save a mask of which terms were negative in mm6. \
    1.69 +    Then we compute mm4=abs(lflim(R_i,L))=min(abs(R_i),max(2*L-abs(R_i),0)). \
    1.70 +    Finally, we split mm4 into positive and negative pieces using the mask in \
    1.71 +     mm6, and add and subtract them as appropriate.*/ \
    1.72 +  /*mm4=abs(-R_i)*/ \
    1.73 +  /*mm7=255-2*L*/ \
    1.74 +  __asm pcmpgtb mm6,mm4 \
    1.75 +  __asm psubb mm7,mm0 \
    1.76 +  __asm pxor mm4,mm6 \
    1.77 +  __asm psubb mm7,mm0 \
    1.78 +  __asm psubb mm4,mm6 \
    1.79 +  /*mm7=255-max(2*L-abs(R_i),0)*/ \
    1.80 +  __asm paddusb mm7,mm4 \
    1.81 +  /*mm4=min(abs(R_i),max(2*L-abs(R_i),0))*/ \
    1.82 +  __asm paddusb mm4,mm7 \
    1.83 +  __asm psubusb mm4,mm7 \
    1.84 +  /*Now split mm4 by the original sign of -R_i.*/ \
    1.85 +  __asm movq mm5,mm4 \
    1.86 +  __asm pand mm4,mm6 \
    1.87 +  __asm pandn mm6,mm5 \
    1.88 +  /*mm1={b0+lflim(R_0,L),...,b7+lflim(R_7,L)}*/ \
    1.89 +  /*mm2={c0-lflim(R_0,L),...,c7-lflim(R_7,L)}*/ \
    1.90 +  __asm paddusb mm1,mm4 \
    1.91 +  __asm psubusb mm2,mm4 \
    1.92 +  __asm psubusb mm1,mm6 \
    1.93 +  __asm paddusb mm2,mm6 \
    1.94 +}
    1.95 +
    1.96 +#define OC_LOOP_FILTER_V_MMX(_pix,_ystride,_ll) \
    1.97 +  do{ \
    1.98 +    /*Used local variable pix__ in order to fix compilation errors like: \
    1.99 +       "error C2425: 'SHL' : non-constant expression in 'second operand'".*/ \
   1.100 +    unsigned char *pix__; \
   1.101 +    unsigned char *ll__; \
   1.102 +    ll__=(_ll); \
   1.103 +    pix__=(_pix); \
   1.104 +    __asm mov YSTRIDE,_ystride \
   1.105 +    __asm mov LL,ll__ \
   1.106 +    __asm mov PIX,pix__ \
   1.107 +    __asm sub PIX,YSTRIDE \
   1.108 +    __asm sub PIX,YSTRIDE \
   1.109 +    /*mm0={a0,...,a7}*/ \
   1.110 +    __asm movq mm0,[PIX] \
   1.111 +    /*ystride3=_ystride*3*/ \
   1.112 +    __asm lea YSTRIDE3,[YSTRIDE+YSTRIDE*2] \
   1.113 +    /*mm3={d0,...,d7}*/ \
   1.114 +    __asm movq mm3,[PIX+YSTRIDE3] \
   1.115 +    /*mm1={b0,...,b7}*/ \
   1.116 +    __asm movq mm1,[PIX+YSTRIDE] \
   1.117 +    /*mm2={c0,...,c7}*/ \
   1.118 +    __asm movq mm2,[PIX+YSTRIDE*2] \
   1.119 +    OC_LOOP_FILTER8_MMX \
   1.120 +    /*Write it back out.*/ \
   1.121 +    __asm movq [PIX+YSTRIDE],mm1 \
   1.122 +    __asm movq [PIX+YSTRIDE*2],mm2 \
   1.123 +  } \
   1.124 +  while(0)
   1.125 +
   1.126 +#define OC_LOOP_FILTER_H_MMX(_pix,_ystride,_ll) \
   1.127 +  do{ \
   1.128 +    /*Used local variable ll__ in order to fix compilation errors like: \
   1.129 +       "error C2443: operand size conflict".*/ \
   1.130 +    unsigned char *ll__; \
   1.131 +    unsigned char *pix__; \
   1.132 +    ll__=(_ll); \
   1.133 +    pix__=(_pix)-2; \
   1.134 +    __asm mov PIX,pix__ \
   1.135 +    __asm mov YSTRIDE,_ystride \
   1.136 +    __asm mov LL,ll__ \
   1.137 +    /*x x x x d0 c0 b0 a0*/ \
   1.138 +    __asm movd mm0,[PIX] \
   1.139 +    /*x x x x d1 c1 b1 a1*/ \
   1.140 +    __asm movd mm1,[PIX+YSTRIDE] \
   1.141 +    /*ystride3=_ystride*3*/ \
   1.142 +    __asm lea YSTRIDE3,[YSTRIDE+YSTRIDE*2] \
   1.143 +    /*x x x x d2 c2 b2 a2*/ \
   1.144 +    __asm movd mm2,[PIX+YSTRIDE*2] \
   1.145 +    /*x x x x d3 c3 b3 a3*/ \
   1.146 +    __asm lea D,[PIX+YSTRIDE*4] \
   1.147 +    __asm movd mm3,[PIX+YSTRIDE3] \
   1.148 +    /*x x x x d4 c4 b4 a4*/ \
   1.149 +    __asm movd mm4,[D] \
   1.150 +    /*x x x x d5 c5 b5 a5*/ \
   1.151 +    __asm movd mm5,[D+YSTRIDE] \
   1.152 +    /*x x x x d6 c6 b6 a6*/ \
   1.153 +    __asm movd mm6,[D+YSTRIDE*2] \
   1.154 +    /*x x x x d7 c7 b7 a7*/ \
   1.155 +    __asm movd mm7,[D+YSTRIDE3] \
   1.156 +    /*mm0=d1 d0 c1 c0 b1 b0 a1 a0*/ \
   1.157 +    __asm punpcklbw mm0,mm1 \
   1.158 +    /*mm2=d3 d2 c3 c2 b3 b2 a3 a2*/ \
   1.159 +    __asm punpcklbw mm2,mm3 \
   1.160 +    /*mm3=d1 d0 c1 c0 b1 b0 a1 a0*/ \
   1.161 +    __asm movq mm3,mm0 \
   1.162 +    /*mm0=b3 b2 b1 b0 a3 a2 a1 a0*/ \
   1.163 +    __asm punpcklwd mm0,mm2 \
   1.164 +    /*mm3=d3 d2 d1 d0 c3 c2 c1 c0*/ \
   1.165 +    __asm punpckhwd mm3,mm2 \
   1.166 +    /*mm1=b3 b2 b1 b0 a3 a2 a1 a0*/ \
   1.167 +    __asm movq mm1,mm0 \
   1.168 +    /*mm4=d5 d4 c5 c4 b5 b4 a5 a4*/ \
   1.169 +    __asm punpcklbw mm4,mm5 \
   1.170 +    /*mm6=d7 d6 c7 c6 b7 b6 a7 a6*/ \
   1.171 +    __asm punpcklbw mm6,mm7 \
   1.172 +    /*mm5=d5 d4 c5 c4 b5 b4 a5 a4*/ \
   1.173 +    __asm movq mm5,mm4 \
   1.174 +    /*mm4=b7 b6 b5 b4 a7 a6 a5 a4*/ \
   1.175 +    __asm punpcklwd mm4,mm6 \
   1.176 +    /*mm5=d7 d6 d5 d4 c7 c6 c5 c4*/ \
   1.177 +    __asm punpckhwd mm5,mm6 \
   1.178 +    /*mm2=d3 d2 d1 d0 c3 c2 c1 c0*/ \
   1.179 +    __asm movq mm2,mm3 \
   1.180 +    /*mm0=a7 a6 a5 a4 a3 a2 a1 a0*/ \
   1.181 +    __asm punpckldq mm0,mm4 \
   1.182 +    /*mm1=b7 b6 b5 b4 b3 b2 b1 b0*/ \
   1.183 +    __asm punpckhdq mm1,mm4 \
   1.184 +    /*mm2=c7 c6 c5 c4 c3 c2 c1 c0*/ \
   1.185 +    __asm punpckldq mm2,mm5 \
   1.186 +    /*mm3=d7 d6 d5 d4 d3 d2 d1 d0*/ \
   1.187 +    __asm punpckhdq mm3,mm5 \
   1.188 +    OC_LOOP_FILTER8_MMX \
   1.189 +    /*mm2={b0+R_0'',...,b7+R_7''}*/ \
   1.190 +    __asm movq mm0,mm1 \
   1.191 +    /*mm1={b0+R_0'',c0-R_0'',...,b3+R_3'',c3-R_3''}*/ \
   1.192 +    __asm punpcklbw mm1,mm2 \
   1.193 +    /*mm2={b4+R_4'',c4-R_4'',...,b7+R_7'',c7-R_7''}*/ \
   1.194 +    __asm punpckhbw mm0,mm2 \
   1.195 +    /*[d]=c1 b1 c0 b0*/ \
   1.196 +    __asm movd D,mm1 \
   1.197 +    __asm mov [PIX+1],D_WORD \
   1.198 +    __asm psrlq mm1,32 \
   1.199 +    __asm shr D,16 \
   1.200 +    __asm mov [PIX+YSTRIDE+1],D_WORD \
   1.201 +    /*[d]=c3 b3 c2 b2*/ \
   1.202 +    __asm movd D,mm1 \
   1.203 +    __asm mov [PIX+YSTRIDE*2+1],D_WORD \
   1.204 +    __asm shr D,16 \
   1.205 +    __asm mov [PIX+YSTRIDE3+1],D_WORD \
   1.206 +    __asm lea PIX,[PIX+YSTRIDE*4] \
   1.207 +    /*[d]=c5 b5 c4 b4*/ \
   1.208 +    __asm movd D,mm0 \
   1.209 +    __asm mov [PIX+1],D_WORD \
   1.210 +    __asm psrlq mm0,32 \
   1.211 +    __asm shr D,16 \
   1.212 +    __asm mov [PIX+YSTRIDE+1],D_WORD \
   1.213 +    /*[d]=c7 b7 c6 b6*/ \
   1.214 +    __asm movd D,mm0 \
   1.215 +    __asm mov [PIX+YSTRIDE*2+1],D_WORD \
   1.216 +    __asm shr D,16 \
   1.217 +    __asm mov [PIX+YSTRIDE3+1],D_WORD \
   1.218 +  } \
   1.219 +  while(0)
   1.220 +
   1.221 +# endif
   1.222 +#endif

mercurial