media/libtheora/lib/x86/mmxloop.h

Wed, 31 Dec 2014 06:09:35 +0100

author
Michael Schloh von Bennewitz <michael@schloh.com>
date
Wed, 31 Dec 2014 06:09:35 +0100
changeset 0
6474c204b198
permissions
-rw-r--r--

Cloned upstream origin tor-browser at tor-browser-31.3.0esr-4.5-1-build1
revision ID fc1c9ff7c1b2defdbc039f12214767608f46423f for hacking purpose.

     1 #if !defined(_x86_mmxloop_H)
     2 # define _x86_mmxloop_H (1)
     3 # include <stddef.h>
     4 # include "x86int.h"
     6 #if defined(OC_X86_ASM)
     8 /*On entry, mm0={a0,...,a7}, mm1={b0,...,b7}, mm2={c0,...,c7}, mm3={d0,...d7}.
     9   On exit, mm1={b0+lflim(R_0,L),...,b7+lflim(R_7,L)} and
    10    mm2={c0-lflim(R_0,L),...,c7-lflim(R_7,L)}; mm0 and mm3 are clobbered.*/
    11 #define OC_LOOP_FILTER8_MMX \
    12   "#OC_LOOP_FILTER8_MMX\n\t" \
    13   /*mm7=0*/ \
    14   "pxor %%mm7,%%mm7\n\t" \
    15   /*mm6:mm0={a0,...,a7}*/ \
    16   "movq %%mm0,%%mm6\n\t" \
    17   "punpcklbw %%mm7,%%mm0\n\t" \
    18   "punpckhbw %%mm7,%%mm6\n\t" \
    19   /*mm3:mm5={d0,...,d7}*/ \
    20   "movq %%mm3,%%mm5\n\t" \
    21   "punpcklbw %%mm7,%%mm3\n\t" \
    22   "punpckhbw %%mm7,%%mm5\n\t" \
    23   /*mm6:mm0={a0-d0,...,a7-d7}*/ \
    24   "psubw %%mm3,%%mm0\n\t" \
    25   "psubw %%mm5,%%mm6\n\t" \
    26   /*mm3:mm1={b0,...,b7}*/ \
    27   "movq %%mm1,%%mm3\n\t" \
    28   "punpcklbw %%mm7,%%mm1\n\t" \
    29   "movq %%mm2,%%mm4\n\t" \
    30   "punpckhbw %%mm7,%%mm3\n\t" \
    31   /*mm5:mm4={c0,...,c7}*/ \
    32   "movq %%mm2,%%mm5\n\t" \
    33   "punpcklbw %%mm7,%%mm4\n\t" \
    34   "punpckhbw %%mm7,%%mm5\n\t" \
    35   /*mm7={3}x4 \
    36     mm5:mm4={c0-b0,...,c7-b7}*/ \
    37   "pcmpeqw %%mm7,%%mm7\n\t" \
    38   "psubw %%mm1,%%mm4\n\t" \
    39   "psrlw $14,%%mm7\n\t" \
    40   "psubw %%mm3,%%mm5\n\t" \
    41   /*Scale by 3.*/ \
    42   "pmullw %%mm7,%%mm4\n\t" \
    43   "pmullw %%mm7,%%mm5\n\t" \
    44   /*mm7={4}x4 \
    45     mm5:mm4=f={a0-d0+3*(c0-b0),...,a7-d7+3*(c7-b7)}*/ \
    46   "psrlw $1,%%mm7\n\t" \
    47   "paddw %%mm0,%%mm4\n\t" \
    48   "psllw $2,%%mm7\n\t" \
    49   "movq (%[ll]),%%mm0\n\t" \
    50   "paddw %%mm6,%%mm5\n\t" \
    51   /*R_i has the range [-127,128], so we compute -R_i instead. \
    52     mm4=-R_i=-(f+4>>3)=0xFF^(f-4>>3)*/ \
    53   "psubw %%mm7,%%mm4\n\t" \
    54   "psubw %%mm7,%%mm5\n\t" \
    55   "psraw $3,%%mm4\n\t" \
    56   "psraw $3,%%mm5\n\t" \
    57   "pcmpeqb %%mm7,%%mm7\n\t" \
    58   "packsswb %%mm5,%%mm4\n\t" \
    59   "pxor %%mm6,%%mm6\n\t" \
    60   "pxor %%mm7,%%mm4\n\t" \
    61   "packuswb %%mm3,%%mm1\n\t" \
    62   /*Now compute lflim of -mm4 cf. Section 7.10 of the sepc.*/ \
    63   /*There's no unsigned byte+signed byte with unsigned saturation op code, so \
    64      we have to split things by sign (the other option is to work in 16 bits, \
    65      but working in 8 bits gives much better parallelism). \
    66     We compute abs(R_i), but save a mask of which terms were negative in mm6. \
    67     Then we compute mm4=abs(lflim(R_i,L))=min(abs(R_i),max(2*L-abs(R_i),0)). \
    68     Finally, we split mm4 into positive and negative pieces using the mask in \
    69      mm6, and add and subtract them as appropriate.*/ \
    70   /*mm4=abs(-R_i)*/ \
    71   /*mm7=255-2*L*/ \
    72   "pcmpgtb %%mm4,%%mm6\n\t" \
    73   "psubb %%mm0,%%mm7\n\t" \
    74   "pxor %%mm6,%%mm4\n\t" \
    75   "psubb %%mm0,%%mm7\n\t" \
    76   "psubb %%mm6,%%mm4\n\t" \
    77   /*mm7=255-max(2*L-abs(R_i),0)*/ \
    78   "paddusb %%mm4,%%mm7\n\t" \
    79   /*mm4=min(abs(R_i),max(2*L-abs(R_i),0))*/ \
    80   "paddusb %%mm7,%%mm4\n\t" \
    81   "psubusb %%mm7,%%mm4\n\t" \
    82   /*Now split mm4 by the original sign of -R_i.*/ \
    83   "movq %%mm4,%%mm5\n\t" \
    84   "pand %%mm6,%%mm4\n\t" \
    85   "pandn %%mm5,%%mm6\n\t" \
    86   /*mm1={b0+lflim(R_0,L),...,b7+lflim(R_7,L)}*/ \
    87   /*mm2={c0-lflim(R_0,L),...,c7-lflim(R_7,L)}*/ \
    88   "paddusb %%mm4,%%mm1\n\t" \
    89   "psubusb %%mm4,%%mm2\n\t" \
    90   "psubusb %%mm6,%%mm1\n\t" \
    91   "paddusb %%mm6,%%mm2\n\t" \
    93 /*On entry, mm0={a0,...,a7}, mm1={b0,...,b7}, mm2={c0,...,c7}, mm3={d0,...d7}.
    94   On exit, mm1={b0+lflim(R_0,L),...,b7+lflim(R_7,L)} and
    95    mm2={c0-lflim(R_0,L),...,c7-lflim(R_7,L)}.
    96   All other MMX registers are clobbered.*/
    97 #define OC_LOOP_FILTER8_MMXEXT \
    98   "#OC_LOOP_FILTER8_MMXEXT\n\t" \
    99   /*R_i=(a_i-3*b_i+3*c_i-d_i+4>>3) has the range [-127,128], so we compute \
   100      -R_i=(-a_i+3*b_i-3*c_i+d_i+3>>3) instead.*/ \
   101   /*This first part is based on the transformation \
   102       f = -(3*(c-b)+a-d+4>>3) \
   103         = -(3*(c+255-b)+(a+255-d)+4-1020>>3) \
   104         = -(3*(c+~b)+(a+~d)-1016>>3) \
   105         = 127-(3*(c+~b)+(a+~d)>>3) \
   106         = 128+~(3*(c+~b)+(a+~d)>>3) (mod 256). \
   107     Although pavgb(a,b) = (a+b+1>>1) (biased up), we rely heavily on the \
   108      fact that ~pavgb(~a,~b) = (a+b>>1) (biased down). \
   109     Using this, the last expression above can be computed in 8 bits of working \
   110      precision via: \
   111       u = ~pavgb(~b,c); \
   112       v = pavgb(b,~c); \
   113       This mask is 0 or 0xFF, and controls whether t is biased up or down: \
   114       m = u-v; \
   115       t = m^pavgb(m^~a,m^d); \
   116       f = 128+pavgb(pavgb(t,u),v); \
   117     This required some careful analysis to ensure that carries are propagated \
   118      correctly in all cases, but has been checked exhaustively.*/ \
   119   /*input (a, b, c, d, ., ., ., .)*/ \
   120   /*ff=0xFF; \
   121     u=b; \
   122     v=c; \
   123     ll=255-2*L;*/ \
   124   "pcmpeqb %%mm7,%%mm7\n\t" \
   125   "movq %%mm1,%%mm4\n\t" \
   126   "movq %%mm2,%%mm5\n\t" \
   127   "movq (%[ll]),%%mm6\n\t" \
   128   /*allocated u, v, ll, ff: (a, b, c, d, u, v, ll, ff)*/ \
   129   /*u^=ff; \
   130     v^=ff;*/ \
   131   "pxor %%mm7,%%mm4\n\t" \
   132   "pxor %%mm7,%%mm5\n\t" \
   133   /*allocated ll: (a, b, c, d, u, v, ll, ff)*/ \
   134   /*u=pavgb(u,c); \
   135     v=pavgb(v,b);*/ \
   136   "pavgb %%mm2,%%mm4\n\t" \
   137   "pavgb %%mm1,%%mm5\n\t" \
   138   /*u^=ff; \
   139     a^=ff;*/ \
   140   "pxor %%mm7,%%mm4\n\t" \
   141   "pxor %%mm7,%%mm0\n\t" \
   142   /*m=u-v;*/ \
   143   "psubb %%mm5,%%mm4\n\t" \
   144   /*freed u, allocated m: (a, b, c, d, m, v, ll, ff)*/ \
   145   /*a^=m; \
   146     d^=m;*/ \
   147   "pxor %%mm4,%%mm0\n\t" \
   148   "pxor %%mm4,%%mm3\n\t" \
   149   /*t=pavgb(a,d);*/ \
   150   "pavgb %%mm3,%%mm0\n\t" \
   151   "psllw $7,%%mm7\n\t" \
   152   /*freed a, d, ff, allocated t, of: (t, b, c, ., m, v, ll, of)*/ \
   153   /*t^=m; \
   154     u=m+v;*/ \
   155   "pxor %%mm4,%%mm0\n\t" \
   156   "paddb %%mm5,%%mm4\n\t" \
   157   /*freed t, m, allocated f, u: (f, b, c, ., u, v, ll, of)*/ \
   158   /*f=pavgb(f,u); \
   159     of=128;*/ \
   160   "pavgb %%mm4,%%mm0\n\t" \
   161   "packsswb %%mm7,%%mm7\n\t" \
   162   /*freed u, ff, allocated ll: (f, b, c, ., ll, v, ll, of)*/ \
   163   /*f=pavgb(f,v);*/ \
   164   "pavgb %%mm5,%%mm0\n\t" \
   165   "movq %%mm7,%%mm3\n\t" \
   166   "movq %%mm6,%%mm4\n\t" \
   167   /*freed v, allocated of: (f, b, c, of, ll, ., ll, of)*/ \
   168   /*Now compute lflim of R_i=-(128+mm0) cf. Section 7.10 of the sepc.*/ \
   169   /*There's no unsigned byte+signed byte with unsigned saturation op code, so \
   170      we have to split things by sign (the other option is to work in 16 bits, \
   171      but staying in 8 bits gives much better parallelism).*/ \
   172   /*Instead of adding the offset of 128 in mm3, we use it to split mm0. \
   173     This is the same number of instructions as computing a mask and splitting \
   174      after the lflim computation, but has shorter dependency chains.*/ \
   175   /*mm0=R_i<0?-R_i:0 (denoted abs(R_i<0))\
   176     mm3=R_i>0?R_i:0* (denoted abs(R_i>0))*/ \
   177   "psubusb %%mm0,%%mm3\n\t" \
   178   "psubusb %%mm7,%%mm0\n\t" \
   179   /*mm6=255-max(2*L-abs(R_i<0),0) \
   180     mm4=255-max(2*L-abs(R_i>0),0)*/ \
   181   "paddusb %%mm3,%%mm4\n\t" \
   182   "paddusb %%mm0,%%mm6\n\t" \
   183   /*mm0=min(abs(R_i<0),max(2*L-abs(R_i<0),0)) \
   184     mm3=min(abs(R_i>0),max(2*L-abs(R_i>0),0))*/ \
   185   "paddusb %%mm4,%%mm3\n\t" \
   186   "paddusb %%mm6,%%mm0\n\t" \
   187   "psubusb %%mm4,%%mm3\n\t" \
   188   "psubusb %%mm6,%%mm0\n\t" \
   189   /*mm1={b0+lflim(R_0,L),...,b7+lflim(R_7,L)}*/ \
   190   /*mm2={c0-lflim(R_0,L),...,c7-lflim(R_7,L)}*/ \
   191   "paddusb %%mm3,%%mm1\n\t" \
   192   "psubusb %%mm3,%%mm2\n\t" \
   193   "psubusb %%mm0,%%mm1\n\t" \
   194   "paddusb %%mm0,%%mm2\n\t" \
   196 #define OC_LOOP_FILTER_V(_filter,_pix,_ystride,_ll) \
   197   do{ \
   198     ptrdiff_t ystride3__; \
   199     __asm__ __volatile__( \
   200       /*mm0={a0,...,a7}*/ \
   201       "movq (%[pix]),%%mm0\n\t" \
   202       /*ystride3=_ystride*3*/ \
   203       "lea (%[ystride],%[ystride],2),%[ystride3]\n\t" \
   204       /*mm3={d0,...,d7}*/ \
   205       "movq (%[pix],%[ystride3]),%%mm3\n\t" \
   206       /*mm1={b0,...,b7}*/ \
   207       "movq (%[pix],%[ystride]),%%mm1\n\t" \
   208       /*mm2={c0,...,c7}*/ \
   209       "movq (%[pix],%[ystride],2),%%mm2\n\t" \
   210       _filter \
   211       /*Write it back out.*/ \
   212       "movq %%mm1,(%[pix],%[ystride])\n\t" \
   213       "movq %%mm2,(%[pix],%[ystride],2)\n\t" \
   214       :[ystride3]"=&r"(ystride3__) \
   215       :[pix]"r"(_pix-_ystride*2),[ystride]"r"((ptrdiff_t)(_ystride)), \
   216        [ll]"r"(_ll) \
   217       :"memory" \
   218     ); \
   219   } \
   220   while(0)
   222 #define OC_LOOP_FILTER_H(_filter,_pix,_ystride,_ll) \
   223   do{ \
   224     unsigned char *pix__; \
   225     ptrdiff_t      ystride3__; \
   226     ptrdiff_t      d__; \
   227     pix__=(_pix)-2; \
   228     __asm__ __volatile__( \
   229       /*x x x x d0 c0 b0 a0*/ \
   230       "movd (%[pix]),%%mm0\n\t" \
   231       /*x x x x d1 c1 b1 a1*/ \
   232       "movd (%[pix],%[ystride]),%%mm1\n\t" \
   233       /*ystride3=_ystride*3*/ \
   234       "lea (%[ystride],%[ystride],2),%[ystride3]\n\t" \
   235       /*x x x x d2 c2 b2 a2*/ \
   236       "movd (%[pix],%[ystride],2),%%mm2\n\t" \
   237       /*x x x x d3 c3 b3 a3*/ \
   238       "lea (%[pix],%[ystride],4),%[d]\n\t" \
   239       "movd (%[pix],%[ystride3]),%%mm3\n\t" \
   240       /*x x x x d4 c4 b4 a4*/ \
   241       "movd (%[d]),%%mm4\n\t" \
   242       /*x x x x d5 c5 b5 a5*/ \
   243       "movd (%[d],%[ystride]),%%mm5\n\t" \
   244       /*x x x x d6 c6 b6 a6*/ \
   245       "movd (%[d],%[ystride],2),%%mm6\n\t" \
   246       /*x x x x d7 c7 b7 a7*/ \
   247       "movd (%[d],%[ystride3]),%%mm7\n\t" \
   248       /*mm0=d1 d0 c1 c0 b1 b0 a1 a0*/ \
   249       "punpcklbw %%mm1,%%mm0\n\t" \
   250       /*mm2=d3 d2 c3 c2 b3 b2 a3 a2*/ \
   251       "punpcklbw %%mm3,%%mm2\n\t" \
   252       /*mm3=d1 d0 c1 c0 b1 b0 a1 a0*/ \
   253       "movq %%mm0,%%mm3\n\t" \
   254       /*mm0=b3 b2 b1 b0 a3 a2 a1 a0*/ \
   255       "punpcklwd %%mm2,%%mm0\n\t" \
   256       /*mm3=d3 d2 d1 d0 c3 c2 c1 c0*/ \
   257       "punpckhwd %%mm2,%%mm3\n\t" \
   258       /*mm1=b3 b2 b1 b0 a3 a2 a1 a0*/ \
   259       "movq %%mm0,%%mm1\n\t" \
   260       /*mm4=d5 d4 c5 c4 b5 b4 a5 a4*/ \
   261       "punpcklbw %%mm5,%%mm4\n\t" \
   262       /*mm6=d7 d6 c7 c6 b7 b6 a7 a6*/ \
   263       "punpcklbw %%mm7,%%mm6\n\t" \
   264       /*mm5=d5 d4 c5 c4 b5 b4 a5 a4*/ \
   265       "movq %%mm4,%%mm5\n\t" \
   266       /*mm4=b7 b6 b5 b4 a7 a6 a5 a4*/ \
   267       "punpcklwd %%mm6,%%mm4\n\t" \
   268       /*mm5=d7 d6 d5 d4 c7 c6 c5 c4*/ \
   269       "punpckhwd %%mm6,%%mm5\n\t" \
   270       /*mm2=d3 d2 d1 d0 c3 c2 c1 c0*/ \
   271       "movq %%mm3,%%mm2\n\t" \
   272       /*mm0=a7 a6 a5 a4 a3 a2 a1 a0*/ \
   273       "punpckldq %%mm4,%%mm0\n\t" \
   274       /*mm1=b7 b6 b5 b4 b3 b2 b1 b0*/ \
   275       "punpckhdq %%mm4,%%mm1\n\t" \
   276       /*mm2=c7 c6 c5 c4 c3 c2 c1 c0*/ \
   277       "punpckldq %%mm5,%%mm2\n\t" \
   278       /*mm3=d7 d6 d5 d4 d3 d2 d1 d0*/ \
   279       "punpckhdq %%mm5,%%mm3\n\t" \
   280       _filter \
   281       /*mm2={b0+R_0'',...,b7+R_7''}*/ \
   282       "movq %%mm1,%%mm0\n\t" \
   283       /*mm1={b0+R_0'',c0-R_0'',...,b3+R_3'',c3-R_3''}*/ \
   284       "punpcklbw %%mm2,%%mm1\n\t" \
   285       /*mm2={b4+R_4'',c4-R_4'',...,b7+R_7'',c7-R_7''}*/ \
   286       "punpckhbw %%mm2,%%mm0\n\t" \
   287       /*[d]=c1 b1 c0 b0*/ \
   288       "movd %%mm1,%[d]\n\t" \
   289       "movw %w[d],1(%[pix])\n\t" \
   290       "psrlq $32,%%mm1\n\t" \
   291       "shr $16,%[d]\n\t" \
   292       "movw %w[d],1(%[pix],%[ystride])\n\t" \
   293       /*[d]=c3 b3 c2 b2*/ \
   294       "movd %%mm1,%[d]\n\t" \
   295       "movw %w[d],1(%[pix],%[ystride],2)\n\t" \
   296       "shr $16,%[d]\n\t" \
   297       "movw %w[d],1(%[pix],%[ystride3])\n\t" \
   298       "lea (%[pix],%[ystride],4),%[pix]\n\t" \
   299       /*[d]=c5 b5 c4 b4*/ \
   300       "movd %%mm0,%[d]\n\t" \
   301       "movw %w[d],1(%[pix])\n\t" \
   302       "psrlq $32,%%mm0\n\t" \
   303       "shr $16,%[d]\n\t" \
   304       "movw %w[d],1(%[pix],%[ystride])\n\t" \
   305       /*[d]=c7 b7 c6 b6*/ \
   306       "movd %%mm0,%[d]\n\t" \
   307       "movw %w[d],1(%[pix],%[ystride],2)\n\t" \
   308       "shr $16,%[d]\n\t" \
   309       "movw %w[d],1(%[pix],%[ystride3])\n\t" \
   310       :[pix]"+r"(pix__),[ystride3]"=&r"(ystride3__),[d]"=&r"(d__) \
   311       :[ystride]"r"((ptrdiff_t)(_ystride)),[ll]"r"(_ll) \
   312       :"memory" \
   313     ); \
   314   } \
   315   while(0)
   317 # endif
   318 #endif

mercurial