|
1 /* |
|
2 * Copyright (c) 2010 The WebM project authors. All Rights Reserved. |
|
3 * |
|
4 * Use of this source code is governed by a BSD-style license |
|
5 * that can be found in the LICENSE file in the root of the source |
|
6 * tree. An additional intellectual property rights grant can be found |
|
7 * in the file PATENTS. All contributing project authors may |
|
8 * be found in the AUTHORS file in the root of the source tree. |
|
9 */ |
|
10 |
|
11 |
|
12 #include "onyx_int.h" |
|
13 #include "mcomp.h" |
|
14 #include "vpx_mem/vpx_mem.h" |
|
15 #include "vpx_config.h" |
|
16 #include <stdio.h> |
|
17 #include <limits.h> |
|
18 #include <math.h> |
|
19 #include "vp8/common/findnearmv.h" |
|
20 |
|
21 #ifdef VP8_ENTROPY_STATS |
|
22 static int mv_ref_ct [31] [4] [2]; |
|
23 static int mv_mode_cts [4] [2]; |
|
24 #endif |
|
25 |
|
26 int vp8_mv_bit_cost(int_mv *mv, int_mv *ref, int *mvcost[2], int Weight) |
|
27 { |
|
28 /* MV costing is based on the distribution of vectors in the previous |
|
29 * frame and as such will tend to over state the cost of vectors. In |
|
30 * addition coding a new vector can have a knock on effect on the cost |
|
31 * of subsequent vectors and the quality of prediction from NEAR and |
|
32 * NEAREST for subsequent blocks. The "Weight" parameter allows, to a |
|
33 * limited extent, for some account to be taken of these factors. |
|
34 */ |
|
35 return ((mvcost[0][(mv->as_mv.row - ref->as_mv.row) >> 1] + mvcost[1][(mv->as_mv.col - ref->as_mv.col) >> 1]) * Weight) >> 7; |
|
36 } |
|
37 |
|
38 static int mv_err_cost(int_mv *mv, int_mv *ref, int *mvcost[2], int error_per_bit) |
|
39 { |
|
40 /* Ignore mv costing if mvcost is NULL */ |
|
41 if (mvcost) |
|
42 return ((mvcost[0][(mv->as_mv.row - ref->as_mv.row) >> 1] + |
|
43 mvcost[1][(mv->as_mv.col - ref->as_mv.col) >> 1]) |
|
44 * error_per_bit + 128) >> 8; |
|
45 return 0; |
|
46 } |
|
47 |
|
48 static int mvsad_err_cost(int_mv *mv, int_mv *ref, int *mvsadcost[2], int error_per_bit) |
|
49 { |
|
50 /* Calculate sad error cost on full pixel basis. */ |
|
51 /* Ignore mv costing if mvsadcost is NULL */ |
|
52 if (mvsadcost) |
|
53 return ((mvsadcost[0][(mv->as_mv.row - ref->as_mv.row)] + |
|
54 mvsadcost[1][(mv->as_mv.col - ref->as_mv.col)]) |
|
55 * error_per_bit + 128) >> 8; |
|
56 return 0; |
|
57 } |
|
58 |
|
59 void vp8_init_dsmotion_compensation(MACROBLOCK *x, int stride) |
|
60 { |
|
61 int Len; |
|
62 int search_site_count = 0; |
|
63 |
|
64 |
|
65 /* Generate offsets for 4 search sites per step. */ |
|
66 Len = MAX_FIRST_STEP; |
|
67 x->ss[search_site_count].mv.col = 0; |
|
68 x->ss[search_site_count].mv.row = 0; |
|
69 x->ss[search_site_count].offset = 0; |
|
70 search_site_count++; |
|
71 |
|
72 while (Len > 0) |
|
73 { |
|
74 |
|
75 /* Compute offsets for search sites. */ |
|
76 x->ss[search_site_count].mv.col = 0; |
|
77 x->ss[search_site_count].mv.row = -Len; |
|
78 x->ss[search_site_count].offset = -Len * stride; |
|
79 search_site_count++; |
|
80 |
|
81 /* Compute offsets for search sites. */ |
|
82 x->ss[search_site_count].mv.col = 0; |
|
83 x->ss[search_site_count].mv.row = Len; |
|
84 x->ss[search_site_count].offset = Len * stride; |
|
85 search_site_count++; |
|
86 |
|
87 /* Compute offsets for search sites. */ |
|
88 x->ss[search_site_count].mv.col = -Len; |
|
89 x->ss[search_site_count].mv.row = 0; |
|
90 x->ss[search_site_count].offset = -Len; |
|
91 search_site_count++; |
|
92 |
|
93 /* Compute offsets for search sites. */ |
|
94 x->ss[search_site_count].mv.col = Len; |
|
95 x->ss[search_site_count].mv.row = 0; |
|
96 x->ss[search_site_count].offset = Len; |
|
97 search_site_count++; |
|
98 |
|
99 /* Contract. */ |
|
100 Len /= 2; |
|
101 } |
|
102 |
|
103 x->ss_count = search_site_count; |
|
104 x->searches_per_step = 4; |
|
105 } |
|
106 |
|
107 void vp8_init3smotion_compensation(MACROBLOCK *x, int stride) |
|
108 { |
|
109 int Len; |
|
110 int search_site_count = 0; |
|
111 |
|
112 /* Generate offsets for 8 search sites per step. */ |
|
113 Len = MAX_FIRST_STEP; |
|
114 x->ss[search_site_count].mv.col = 0; |
|
115 x->ss[search_site_count].mv.row = 0; |
|
116 x->ss[search_site_count].offset = 0; |
|
117 search_site_count++; |
|
118 |
|
119 while (Len > 0) |
|
120 { |
|
121 |
|
122 /* Compute offsets for search sites. */ |
|
123 x->ss[search_site_count].mv.col = 0; |
|
124 x->ss[search_site_count].mv.row = -Len; |
|
125 x->ss[search_site_count].offset = -Len * stride; |
|
126 search_site_count++; |
|
127 |
|
128 /* Compute offsets for search sites. */ |
|
129 x->ss[search_site_count].mv.col = 0; |
|
130 x->ss[search_site_count].mv.row = Len; |
|
131 x->ss[search_site_count].offset = Len * stride; |
|
132 search_site_count++; |
|
133 |
|
134 /* Compute offsets for search sites. */ |
|
135 x->ss[search_site_count].mv.col = -Len; |
|
136 x->ss[search_site_count].mv.row = 0; |
|
137 x->ss[search_site_count].offset = -Len; |
|
138 search_site_count++; |
|
139 |
|
140 /* Compute offsets for search sites. */ |
|
141 x->ss[search_site_count].mv.col = Len; |
|
142 x->ss[search_site_count].mv.row = 0; |
|
143 x->ss[search_site_count].offset = Len; |
|
144 search_site_count++; |
|
145 |
|
146 /* Compute offsets for search sites. */ |
|
147 x->ss[search_site_count].mv.col = -Len; |
|
148 x->ss[search_site_count].mv.row = -Len; |
|
149 x->ss[search_site_count].offset = -Len * stride - Len; |
|
150 search_site_count++; |
|
151 |
|
152 /* Compute offsets for search sites. */ |
|
153 x->ss[search_site_count].mv.col = Len; |
|
154 x->ss[search_site_count].mv.row = -Len; |
|
155 x->ss[search_site_count].offset = -Len * stride + Len; |
|
156 search_site_count++; |
|
157 |
|
158 /* Compute offsets for search sites. */ |
|
159 x->ss[search_site_count].mv.col = -Len; |
|
160 x->ss[search_site_count].mv.row = Len; |
|
161 x->ss[search_site_count].offset = Len * stride - Len; |
|
162 search_site_count++; |
|
163 |
|
164 /* Compute offsets for search sites. */ |
|
165 x->ss[search_site_count].mv.col = Len; |
|
166 x->ss[search_site_count].mv.row = Len; |
|
167 x->ss[search_site_count].offset = Len * stride + Len; |
|
168 search_site_count++; |
|
169 |
|
170 |
|
171 /* Contract. */ |
|
172 Len /= 2; |
|
173 } |
|
174 |
|
175 x->ss_count = search_site_count; |
|
176 x->searches_per_step = 8; |
|
177 } |
|
178 |
|
179 /* |
|
180 * To avoid the penalty for crossing cache-line read, preload the reference |
|
181 * area in a small buffer, which is aligned to make sure there won't be crossing |
|
182 * cache-line read while reading from this buffer. This reduced the cpu |
|
183 * cycles spent on reading ref data in sub-pixel filter functions. |
|
184 * TODO: Currently, since sub-pixel search range here is -3 ~ 3, copy 22 rows x |
|
185 * 32 cols area that is enough for 16x16 macroblock. Later, for SPLITMV, we |
|
186 * could reduce the area. |
|
187 */ |
|
188 |
|
189 /* estimated cost of a motion vector (r,c) */ |
|
190 #define MVC(r,c) (mvcost ? ((mvcost[0][(r)-rr] + mvcost[1][(c) - rc]) * error_per_bit + 128 )>>8 : 0) |
|
191 /* pointer to predictor base of a motionvector */ |
|
192 #define PRE(r,c) (y + (((r)>>2) * y_stride + ((c)>>2) -(offset))) |
|
193 /* convert motion vector component to offset for svf calc */ |
|
194 #define SP(x) (((x)&3)<<1) |
|
195 /* returns subpixel variance error function. */ |
|
196 #define DIST(r,c) vfp->svf( PRE(r,c), y_stride, SP(c),SP(r), z,b->src_stride,&sse) |
|
197 #define IFMVCV(r,c,s,e) if ( c >= minc && c <= maxc && r >= minr && r <= maxr) s else e; |
|
198 /* returns distortion + motion vector cost */ |
|
199 #define ERR(r,c) (MVC(r,c)+DIST(r,c)) |
|
200 /* checks if (r,c) has better score than previous best */ |
|
201 #define CHECK_BETTER(v,r,c) IFMVCV(r,c,{thismse = DIST(r,c); if((v = (MVC(r,c)+thismse)) < besterr) { besterr = v; br=r; bc=c; *distortion = thismse; *sse1 = sse; }}, v=UINT_MAX;) |
|
202 |
|
203 int vp8_find_best_sub_pixel_step_iteratively(MACROBLOCK *x, BLOCK *b, BLOCKD *d, |
|
204 int_mv *bestmv, int_mv *ref_mv, |
|
205 int error_per_bit, |
|
206 const vp8_variance_fn_ptr_t *vfp, |
|
207 int *mvcost[2], int *distortion, |
|
208 unsigned int *sse1) |
|
209 { |
|
210 unsigned char *z = (*(b->base_src) + b->src); |
|
211 |
|
212 int rr = ref_mv->as_mv.row >> 1, rc = ref_mv->as_mv.col >> 1; |
|
213 int br = bestmv->as_mv.row * 4, bc = bestmv->as_mv.col * 4; |
|
214 int tr = br, tc = bc; |
|
215 unsigned int besterr; |
|
216 unsigned int left, right, up, down, diag; |
|
217 unsigned int sse; |
|
218 unsigned int whichdir; |
|
219 unsigned int halfiters = 4; |
|
220 unsigned int quarteriters = 4; |
|
221 int thismse; |
|
222 |
|
223 int minc = MAX(x->mv_col_min * 4, |
|
224 (ref_mv->as_mv.col >> 1) - ((1 << mvlong_width) - 1)); |
|
225 int maxc = MIN(x->mv_col_max * 4, |
|
226 (ref_mv->as_mv.col >> 1) + ((1 << mvlong_width) - 1)); |
|
227 int minr = MAX(x->mv_row_min * 4, |
|
228 (ref_mv->as_mv.row >> 1) - ((1 << mvlong_width) - 1)); |
|
229 int maxr = MIN(x->mv_row_max * 4, |
|
230 (ref_mv->as_mv.row >> 1) + ((1 << mvlong_width) - 1)); |
|
231 |
|
232 int y_stride; |
|
233 int offset; |
|
234 int pre_stride = x->e_mbd.pre.y_stride; |
|
235 unsigned char *base_pre = x->e_mbd.pre.y_buffer; |
|
236 |
|
237 |
|
238 #if ARCH_X86 || ARCH_X86_64 |
|
239 MACROBLOCKD *xd = &x->e_mbd; |
|
240 unsigned char *y_0 = base_pre + d->offset + (bestmv->as_mv.row) * pre_stride + bestmv->as_mv.col; |
|
241 unsigned char *y; |
|
242 int buf_r1, buf_r2, buf_c1; |
|
243 |
|
244 /* Clamping to avoid out-of-range data access */ |
|
245 buf_r1 = ((bestmv->as_mv.row - 3) < x->mv_row_min)?(bestmv->as_mv.row - x->mv_row_min):3; |
|
246 buf_r2 = ((bestmv->as_mv.row + 3) > x->mv_row_max)?(x->mv_row_max - bestmv->as_mv.row):3; |
|
247 buf_c1 = ((bestmv->as_mv.col - 3) < x->mv_col_min)?(bestmv->as_mv.col - x->mv_col_min):3; |
|
248 y_stride = 32; |
|
249 |
|
250 /* Copy to intermediate buffer before searching. */ |
|
251 vfp->copymem(y_0 - buf_c1 - pre_stride*buf_r1, pre_stride, xd->y_buf, y_stride, 16+buf_r1+buf_r2); |
|
252 y = xd->y_buf + y_stride*buf_r1 +buf_c1; |
|
253 #else |
|
254 unsigned char *y = base_pre + d->offset + (bestmv->as_mv.row) * pre_stride + bestmv->as_mv.col; |
|
255 y_stride = pre_stride; |
|
256 #endif |
|
257 |
|
258 offset = (bestmv->as_mv.row) * y_stride + bestmv->as_mv.col; |
|
259 |
|
260 /* central mv */ |
|
261 bestmv->as_mv.row *= 8; |
|
262 bestmv->as_mv.col *= 8; |
|
263 |
|
264 /* calculate central point error */ |
|
265 besterr = vfp->vf(y, y_stride, z, b->src_stride, sse1); |
|
266 *distortion = besterr; |
|
267 besterr += mv_err_cost(bestmv, ref_mv, mvcost, error_per_bit); |
|
268 |
|
269 /* TODO: Each subsequent iteration checks at least one point in common |
|
270 * with the last iteration could be 2 ( if diag selected) |
|
271 */ |
|
272 while (--halfiters) |
|
273 { |
|
274 /* 1/2 pel */ |
|
275 CHECK_BETTER(left, tr, tc - 2); |
|
276 CHECK_BETTER(right, tr, tc + 2); |
|
277 CHECK_BETTER(up, tr - 2, tc); |
|
278 CHECK_BETTER(down, tr + 2, tc); |
|
279 |
|
280 whichdir = (left < right ? 0 : 1) + (up < down ? 0 : 2); |
|
281 |
|
282 switch (whichdir) |
|
283 { |
|
284 case 0: |
|
285 CHECK_BETTER(diag, tr - 2, tc - 2); |
|
286 break; |
|
287 case 1: |
|
288 CHECK_BETTER(diag, tr - 2, tc + 2); |
|
289 break; |
|
290 case 2: |
|
291 CHECK_BETTER(diag, tr + 2, tc - 2); |
|
292 break; |
|
293 case 3: |
|
294 CHECK_BETTER(diag, tr + 2, tc + 2); |
|
295 break; |
|
296 } |
|
297 |
|
298 /* no reason to check the same one again. */ |
|
299 if (tr == br && tc == bc) |
|
300 break; |
|
301 |
|
302 tr = br; |
|
303 tc = bc; |
|
304 } |
|
305 |
|
306 /* TODO: Each subsequent iteration checks at least one point in common |
|
307 * with the last iteration could be 2 ( if diag selected) |
|
308 */ |
|
309 |
|
310 /* 1/4 pel */ |
|
311 while (--quarteriters) |
|
312 { |
|
313 CHECK_BETTER(left, tr, tc - 1); |
|
314 CHECK_BETTER(right, tr, tc + 1); |
|
315 CHECK_BETTER(up, tr - 1, tc); |
|
316 CHECK_BETTER(down, tr + 1, tc); |
|
317 |
|
318 whichdir = (left < right ? 0 : 1) + (up < down ? 0 : 2); |
|
319 |
|
320 switch (whichdir) |
|
321 { |
|
322 case 0: |
|
323 CHECK_BETTER(diag, tr - 1, tc - 1); |
|
324 break; |
|
325 case 1: |
|
326 CHECK_BETTER(diag, tr - 1, tc + 1); |
|
327 break; |
|
328 case 2: |
|
329 CHECK_BETTER(diag, tr + 1, tc - 1); |
|
330 break; |
|
331 case 3: |
|
332 CHECK_BETTER(diag, tr + 1, tc + 1); |
|
333 break; |
|
334 } |
|
335 |
|
336 /* no reason to check the same one again. */ |
|
337 if (tr == br && tc == bc) |
|
338 break; |
|
339 |
|
340 tr = br; |
|
341 tc = bc; |
|
342 } |
|
343 |
|
344 bestmv->as_mv.row = br * 2; |
|
345 bestmv->as_mv.col = bc * 2; |
|
346 |
|
347 if ((abs(bestmv->as_mv.col - ref_mv->as_mv.col) > (MAX_FULL_PEL_VAL<<3)) || |
|
348 (abs(bestmv->as_mv.row - ref_mv->as_mv.row) > (MAX_FULL_PEL_VAL<<3))) |
|
349 return INT_MAX; |
|
350 |
|
351 return besterr; |
|
352 } |
|
353 #undef MVC |
|
354 #undef PRE |
|
355 #undef SP |
|
356 #undef DIST |
|
357 #undef IFMVCV |
|
358 #undef ERR |
|
359 #undef CHECK_BETTER |
|
360 |
|
361 int vp8_find_best_sub_pixel_step(MACROBLOCK *x, BLOCK *b, BLOCKD *d, |
|
362 int_mv *bestmv, int_mv *ref_mv, |
|
363 int error_per_bit, |
|
364 const vp8_variance_fn_ptr_t *vfp, |
|
365 int *mvcost[2], int *distortion, |
|
366 unsigned int *sse1) |
|
367 { |
|
368 int bestmse = INT_MAX; |
|
369 int_mv startmv; |
|
370 int_mv this_mv; |
|
371 unsigned char *z = (*(b->base_src) + b->src); |
|
372 int left, right, up, down, diag; |
|
373 unsigned int sse; |
|
374 int whichdir ; |
|
375 int thismse; |
|
376 int y_stride; |
|
377 int pre_stride = x->e_mbd.pre.y_stride; |
|
378 unsigned char *base_pre = x->e_mbd.pre.y_buffer; |
|
379 |
|
380 #if ARCH_X86 || ARCH_X86_64 |
|
381 MACROBLOCKD *xd = &x->e_mbd; |
|
382 unsigned char *y_0 = base_pre + d->offset + (bestmv->as_mv.row) * pre_stride + bestmv->as_mv.col; |
|
383 unsigned char *y; |
|
384 |
|
385 y_stride = 32; |
|
386 /* Copy 18 rows x 32 cols area to intermediate buffer before searching. */ |
|
387 vfp->copymem(y_0 - 1 - pre_stride, pre_stride, xd->y_buf, y_stride, 18); |
|
388 y = xd->y_buf + y_stride + 1; |
|
389 #else |
|
390 unsigned char *y = base_pre + d->offset + (bestmv->as_mv.row) * pre_stride + bestmv->as_mv.col; |
|
391 y_stride = pre_stride; |
|
392 #endif |
|
393 |
|
394 /* central mv */ |
|
395 bestmv->as_mv.row <<= 3; |
|
396 bestmv->as_mv.col <<= 3; |
|
397 startmv = *bestmv; |
|
398 |
|
399 /* calculate central point error */ |
|
400 bestmse = vfp->vf(y, y_stride, z, b->src_stride, sse1); |
|
401 *distortion = bestmse; |
|
402 bestmse += mv_err_cost(bestmv, ref_mv, mvcost, error_per_bit); |
|
403 |
|
404 /* go left then right and check error */ |
|
405 this_mv.as_mv.row = startmv.as_mv.row; |
|
406 this_mv.as_mv.col = ((startmv.as_mv.col - 8) | 4); |
|
407 thismse = vfp->svf_halfpix_h(y - 1, y_stride, z, b->src_stride, &sse); |
|
408 left = thismse + mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); |
|
409 |
|
410 if (left < bestmse) |
|
411 { |
|
412 *bestmv = this_mv; |
|
413 bestmse = left; |
|
414 *distortion = thismse; |
|
415 *sse1 = sse; |
|
416 } |
|
417 |
|
418 this_mv.as_mv.col += 8; |
|
419 thismse = vfp->svf_halfpix_h(y, y_stride, z, b->src_stride, &sse); |
|
420 right = thismse + mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); |
|
421 |
|
422 if (right < bestmse) |
|
423 { |
|
424 *bestmv = this_mv; |
|
425 bestmse = right; |
|
426 *distortion = thismse; |
|
427 *sse1 = sse; |
|
428 } |
|
429 |
|
430 /* go up then down and check error */ |
|
431 this_mv.as_mv.col = startmv.as_mv.col; |
|
432 this_mv.as_mv.row = ((startmv.as_mv.row - 8) | 4); |
|
433 thismse = vfp->svf_halfpix_v(y - y_stride, y_stride, z, b->src_stride, &sse); |
|
434 up = thismse + mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); |
|
435 |
|
436 if (up < bestmse) |
|
437 { |
|
438 *bestmv = this_mv; |
|
439 bestmse = up; |
|
440 *distortion = thismse; |
|
441 *sse1 = sse; |
|
442 } |
|
443 |
|
444 this_mv.as_mv.row += 8; |
|
445 thismse = vfp->svf_halfpix_v(y, y_stride, z, b->src_stride, &sse); |
|
446 down = thismse + mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); |
|
447 |
|
448 if (down < bestmse) |
|
449 { |
|
450 *bestmv = this_mv; |
|
451 bestmse = down; |
|
452 *distortion = thismse; |
|
453 *sse1 = sse; |
|
454 } |
|
455 |
|
456 |
|
457 /* now check 1 more diagonal */ |
|
458 whichdir = (left < right ? 0 : 1) + (up < down ? 0 : 2); |
|
459 this_mv = startmv; |
|
460 |
|
461 switch (whichdir) |
|
462 { |
|
463 case 0: |
|
464 this_mv.as_mv.col = (this_mv.as_mv.col - 8) | 4; |
|
465 this_mv.as_mv.row = (this_mv.as_mv.row - 8) | 4; |
|
466 thismse = vfp->svf_halfpix_hv(y - 1 - y_stride, y_stride, z, b->src_stride, &sse); |
|
467 break; |
|
468 case 1: |
|
469 this_mv.as_mv.col += 4; |
|
470 this_mv.as_mv.row = (this_mv.as_mv.row - 8) | 4; |
|
471 thismse = vfp->svf_halfpix_hv(y - y_stride, y_stride, z, b->src_stride, &sse); |
|
472 break; |
|
473 case 2: |
|
474 this_mv.as_mv.col = (this_mv.as_mv.col - 8) | 4; |
|
475 this_mv.as_mv.row += 4; |
|
476 thismse = vfp->svf_halfpix_hv(y - 1, y_stride, z, b->src_stride, &sse); |
|
477 break; |
|
478 case 3: |
|
479 default: |
|
480 this_mv.as_mv.col += 4; |
|
481 this_mv.as_mv.row += 4; |
|
482 thismse = vfp->svf_halfpix_hv(y, y_stride, z, b->src_stride, &sse); |
|
483 break; |
|
484 } |
|
485 |
|
486 diag = thismse + mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); |
|
487 |
|
488 if (diag < bestmse) |
|
489 { |
|
490 *bestmv = this_mv; |
|
491 bestmse = diag; |
|
492 *distortion = thismse; |
|
493 *sse1 = sse; |
|
494 } |
|
495 |
|
496 |
|
497 /* time to check quarter pels. */ |
|
498 if (bestmv->as_mv.row < startmv.as_mv.row) |
|
499 y -= y_stride; |
|
500 |
|
501 if (bestmv->as_mv.col < startmv.as_mv.col) |
|
502 y--; |
|
503 |
|
504 startmv = *bestmv; |
|
505 |
|
506 |
|
507 |
|
508 /* go left then right and check error */ |
|
509 this_mv.as_mv.row = startmv.as_mv.row; |
|
510 |
|
511 if (startmv.as_mv.col & 7) |
|
512 { |
|
513 this_mv.as_mv.col = startmv.as_mv.col - 2; |
|
514 thismse = vfp->svf(y, y_stride, this_mv.as_mv.col & 7, this_mv.as_mv.row & 7, z, b->src_stride, &sse); |
|
515 } |
|
516 else |
|
517 { |
|
518 this_mv.as_mv.col = (startmv.as_mv.col - 8) | 6; |
|
519 thismse = vfp->svf(y - 1, y_stride, 6, this_mv.as_mv.row & 7, z, b->src_stride, &sse); |
|
520 } |
|
521 |
|
522 left = thismse + mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); |
|
523 |
|
524 if (left < bestmse) |
|
525 { |
|
526 *bestmv = this_mv; |
|
527 bestmse = left; |
|
528 *distortion = thismse; |
|
529 *sse1 = sse; |
|
530 } |
|
531 |
|
532 this_mv.as_mv.col += 4; |
|
533 thismse = vfp->svf(y, y_stride, this_mv.as_mv.col & 7, this_mv.as_mv.row & 7, z, b->src_stride, &sse); |
|
534 right = thismse + mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); |
|
535 |
|
536 if (right < bestmse) |
|
537 { |
|
538 *bestmv = this_mv; |
|
539 bestmse = right; |
|
540 *distortion = thismse; |
|
541 *sse1 = sse; |
|
542 } |
|
543 |
|
544 /* go up then down and check error */ |
|
545 this_mv.as_mv.col = startmv.as_mv.col; |
|
546 |
|
547 if (startmv.as_mv.row & 7) |
|
548 { |
|
549 this_mv.as_mv.row = startmv.as_mv.row - 2; |
|
550 thismse = vfp->svf(y, y_stride, this_mv.as_mv.col & 7, this_mv.as_mv.row & 7, z, b->src_stride, &sse); |
|
551 } |
|
552 else |
|
553 { |
|
554 this_mv.as_mv.row = (startmv.as_mv.row - 8) | 6; |
|
555 thismse = vfp->svf(y - y_stride, y_stride, this_mv.as_mv.col & 7, 6, z, b->src_stride, &sse); |
|
556 } |
|
557 |
|
558 up = thismse + mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); |
|
559 |
|
560 if (up < bestmse) |
|
561 { |
|
562 *bestmv = this_mv; |
|
563 bestmse = up; |
|
564 *distortion = thismse; |
|
565 *sse1 = sse; |
|
566 } |
|
567 |
|
568 this_mv.as_mv.row += 4; |
|
569 thismse = vfp->svf(y, y_stride, this_mv.as_mv.col & 7, this_mv.as_mv.row & 7, z, b->src_stride, &sse); |
|
570 down = thismse + mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); |
|
571 |
|
572 if (down < bestmse) |
|
573 { |
|
574 *bestmv = this_mv; |
|
575 bestmse = down; |
|
576 *distortion = thismse; |
|
577 *sse1 = sse; |
|
578 } |
|
579 |
|
580 |
|
581 /* now check 1 more diagonal */ |
|
582 whichdir = (left < right ? 0 : 1) + (up < down ? 0 : 2); |
|
583 |
|
584 this_mv = startmv; |
|
585 |
|
586 switch (whichdir) |
|
587 { |
|
588 case 0: |
|
589 |
|
590 if (startmv.as_mv.row & 7) |
|
591 { |
|
592 this_mv.as_mv.row -= 2; |
|
593 |
|
594 if (startmv.as_mv.col & 7) |
|
595 { |
|
596 this_mv.as_mv.col -= 2; |
|
597 thismse = vfp->svf(y, y_stride, this_mv.as_mv.col & 7, this_mv.as_mv.row & 7, z, b->src_stride, &sse); |
|
598 } |
|
599 else |
|
600 { |
|
601 this_mv.as_mv.col = (startmv.as_mv.col - 8) | 6; |
|
602 thismse = vfp->svf(y - 1, y_stride, 6, this_mv.as_mv.row & 7, z, b->src_stride, &sse);; |
|
603 } |
|
604 } |
|
605 else |
|
606 { |
|
607 this_mv.as_mv.row = (startmv.as_mv.row - 8) | 6; |
|
608 |
|
609 if (startmv.as_mv.col & 7) |
|
610 { |
|
611 this_mv.as_mv.col -= 2; |
|
612 thismse = vfp->svf(y - y_stride, y_stride, this_mv.as_mv.col & 7, 6, z, b->src_stride, &sse); |
|
613 } |
|
614 else |
|
615 { |
|
616 this_mv.as_mv.col = (startmv.as_mv.col - 8) | 6; |
|
617 thismse = vfp->svf(y - y_stride - 1, y_stride, 6, 6, z, b->src_stride, &sse); |
|
618 } |
|
619 } |
|
620 |
|
621 break; |
|
622 case 1: |
|
623 this_mv.as_mv.col += 2; |
|
624 |
|
625 if (startmv.as_mv.row & 7) |
|
626 { |
|
627 this_mv.as_mv.row -= 2; |
|
628 thismse = vfp->svf(y, y_stride, this_mv.as_mv.col & 7, this_mv.as_mv.row & 7, z, b->src_stride, &sse); |
|
629 } |
|
630 else |
|
631 { |
|
632 this_mv.as_mv.row = (startmv.as_mv.row - 8) | 6; |
|
633 thismse = vfp->svf(y - y_stride, y_stride, this_mv.as_mv.col & 7, 6, z, b->src_stride, &sse); |
|
634 } |
|
635 |
|
636 break; |
|
637 case 2: |
|
638 this_mv.as_mv.row += 2; |
|
639 |
|
640 if (startmv.as_mv.col & 7) |
|
641 { |
|
642 this_mv.as_mv.col -= 2; |
|
643 thismse = vfp->svf(y, y_stride, this_mv.as_mv.col & 7, this_mv.as_mv.row & 7, z, b->src_stride, &sse); |
|
644 } |
|
645 else |
|
646 { |
|
647 this_mv.as_mv.col = (startmv.as_mv.col - 8) | 6; |
|
648 thismse = vfp->svf(y - 1, y_stride, 6, this_mv.as_mv.row & 7, z, b->src_stride, &sse); |
|
649 } |
|
650 |
|
651 break; |
|
652 case 3: |
|
653 this_mv.as_mv.col += 2; |
|
654 this_mv.as_mv.row += 2; |
|
655 thismse = vfp->svf(y, y_stride, this_mv.as_mv.col & 7, this_mv.as_mv.row & 7, z, b->src_stride, &sse); |
|
656 break; |
|
657 } |
|
658 |
|
659 diag = thismse + mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); |
|
660 |
|
661 if (diag < bestmse) |
|
662 { |
|
663 *bestmv = this_mv; |
|
664 bestmse = diag; |
|
665 *distortion = thismse; |
|
666 *sse1 = sse; |
|
667 } |
|
668 |
|
669 return bestmse; |
|
670 } |
|
671 |
|
672 int vp8_find_best_half_pixel_step(MACROBLOCK *x, BLOCK *b, BLOCKD *d, |
|
673 int_mv *bestmv, int_mv *ref_mv, |
|
674 int error_per_bit, |
|
675 const vp8_variance_fn_ptr_t *vfp, |
|
676 int *mvcost[2], int *distortion, |
|
677 unsigned int *sse1) |
|
678 { |
|
679 int bestmse = INT_MAX; |
|
680 int_mv startmv; |
|
681 int_mv this_mv; |
|
682 unsigned char *z = (*(b->base_src) + b->src); |
|
683 int left, right, up, down, diag; |
|
684 unsigned int sse; |
|
685 int whichdir ; |
|
686 int thismse; |
|
687 int y_stride; |
|
688 int pre_stride = x->e_mbd.pre.y_stride; |
|
689 unsigned char *base_pre = x->e_mbd.pre.y_buffer; |
|
690 |
|
691 #if ARCH_X86 || ARCH_X86_64 |
|
692 MACROBLOCKD *xd = &x->e_mbd; |
|
693 unsigned char *y_0 = base_pre + d->offset + (bestmv->as_mv.row) * pre_stride + bestmv->as_mv.col; |
|
694 unsigned char *y; |
|
695 |
|
696 y_stride = 32; |
|
697 /* Copy 18 rows x 32 cols area to intermediate buffer before searching. */ |
|
698 vfp->copymem(y_0 - 1 - pre_stride, pre_stride, xd->y_buf, y_stride, 18); |
|
699 y = xd->y_buf + y_stride + 1; |
|
700 #else |
|
701 unsigned char *y = base_pre + d->offset + (bestmv->as_mv.row) * pre_stride + bestmv->as_mv.col; |
|
702 y_stride = pre_stride; |
|
703 #endif |
|
704 |
|
705 /* central mv */ |
|
706 bestmv->as_mv.row *= 8; |
|
707 bestmv->as_mv.col *= 8; |
|
708 startmv = *bestmv; |
|
709 |
|
710 /* calculate central point error */ |
|
711 bestmse = vfp->vf(y, y_stride, z, b->src_stride, sse1); |
|
712 *distortion = bestmse; |
|
713 bestmse += mv_err_cost(bestmv, ref_mv, mvcost, error_per_bit); |
|
714 |
|
715 /* go left then right and check error */ |
|
716 this_mv.as_mv.row = startmv.as_mv.row; |
|
717 this_mv.as_mv.col = ((startmv.as_mv.col - 8) | 4); |
|
718 thismse = vfp->svf_halfpix_h(y - 1, y_stride, z, b->src_stride, &sse); |
|
719 left = thismse + mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); |
|
720 |
|
721 if (left < bestmse) |
|
722 { |
|
723 *bestmv = this_mv; |
|
724 bestmse = left; |
|
725 *distortion = thismse; |
|
726 *sse1 = sse; |
|
727 } |
|
728 |
|
729 this_mv.as_mv.col += 8; |
|
730 thismse = vfp->svf_halfpix_h(y, y_stride, z, b->src_stride, &sse); |
|
731 right = thismse + mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); |
|
732 |
|
733 if (right < bestmse) |
|
734 { |
|
735 *bestmv = this_mv; |
|
736 bestmse = right; |
|
737 *distortion = thismse; |
|
738 *sse1 = sse; |
|
739 } |
|
740 |
|
741 /* go up then down and check error */ |
|
742 this_mv.as_mv.col = startmv.as_mv.col; |
|
743 this_mv.as_mv.row = ((startmv.as_mv.row - 8) | 4); |
|
744 thismse = vfp->svf_halfpix_v(y - y_stride, y_stride, z, b->src_stride, &sse); |
|
745 up = thismse + mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); |
|
746 |
|
747 if (up < bestmse) |
|
748 { |
|
749 *bestmv = this_mv; |
|
750 bestmse = up; |
|
751 *distortion = thismse; |
|
752 *sse1 = sse; |
|
753 } |
|
754 |
|
755 this_mv.as_mv.row += 8; |
|
756 thismse = vfp->svf_halfpix_v(y, y_stride, z, b->src_stride, &sse); |
|
757 down = thismse + mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); |
|
758 |
|
759 if (down < bestmse) |
|
760 { |
|
761 *bestmv = this_mv; |
|
762 bestmse = down; |
|
763 *distortion = thismse; |
|
764 *sse1 = sse; |
|
765 } |
|
766 |
|
767 /* now check 1 more diagonal - */ |
|
768 whichdir = (left < right ? 0 : 1) + (up < down ? 0 : 2); |
|
769 this_mv = startmv; |
|
770 |
|
771 switch (whichdir) |
|
772 { |
|
773 case 0: |
|
774 this_mv.as_mv.col = (this_mv.as_mv.col - 8) | 4; |
|
775 this_mv.as_mv.row = (this_mv.as_mv.row - 8) | 4; |
|
776 thismse = vfp->svf_halfpix_hv(y - 1 - y_stride, y_stride, z, b->src_stride, &sse); |
|
777 break; |
|
778 case 1: |
|
779 this_mv.as_mv.col += 4; |
|
780 this_mv.as_mv.row = (this_mv.as_mv.row - 8) | 4; |
|
781 thismse = vfp->svf_halfpix_hv(y - y_stride, y_stride, z, b->src_stride, &sse); |
|
782 break; |
|
783 case 2: |
|
784 this_mv.as_mv.col = (this_mv.as_mv.col - 8) | 4; |
|
785 this_mv.as_mv.row += 4; |
|
786 thismse = vfp->svf_halfpix_hv(y - 1, y_stride, z, b->src_stride, &sse); |
|
787 break; |
|
788 case 3: |
|
789 default: |
|
790 this_mv.as_mv.col += 4; |
|
791 this_mv.as_mv.row += 4; |
|
792 thismse = vfp->svf_halfpix_hv(y, y_stride, z, b->src_stride, &sse); |
|
793 break; |
|
794 } |
|
795 |
|
796 diag = thismse + mv_err_cost(&this_mv, ref_mv, mvcost, error_per_bit); |
|
797 |
|
798 if (diag < bestmse) |
|
799 { |
|
800 *bestmv = this_mv; |
|
801 bestmse = diag; |
|
802 *distortion = thismse; |
|
803 *sse1 = sse; |
|
804 } |
|
805 |
|
806 return bestmse; |
|
807 } |
|
808 |
|
809 #define CHECK_BOUNDS(range) \ |
|
810 {\ |
|
811 all_in = 1;\ |
|
812 all_in &= ((br-range) >= x->mv_row_min);\ |
|
813 all_in &= ((br+range) <= x->mv_row_max);\ |
|
814 all_in &= ((bc-range) >= x->mv_col_min);\ |
|
815 all_in &= ((bc+range) <= x->mv_col_max);\ |
|
816 } |
|
817 |
|
818 #define CHECK_POINT \ |
|
819 {\ |
|
820 if (this_mv.as_mv.col < x->mv_col_min) continue;\ |
|
821 if (this_mv.as_mv.col > x->mv_col_max) continue;\ |
|
822 if (this_mv.as_mv.row < x->mv_row_min) continue;\ |
|
823 if (this_mv.as_mv.row > x->mv_row_max) continue;\ |
|
824 } |
|
825 |
|
826 #define CHECK_BETTER \ |
|
827 {\ |
|
828 if (thissad < bestsad)\ |
|
829 {\ |
|
830 thissad += mvsad_err_cost(&this_mv, &fcenter_mv, mvsadcost, sad_per_bit);\ |
|
831 if (thissad < bestsad)\ |
|
832 {\ |
|
833 bestsad = thissad;\ |
|
834 best_site = i;\ |
|
835 }\ |
|
836 }\ |
|
837 } |
|
838 |
|
839 static const MV next_chkpts[6][3] = |
|
840 { |
|
841 {{ -2, 0}, { -1, -2}, {1, -2}}, |
|
842 {{ -1, -2}, {1, -2}, {2, 0}}, |
|
843 {{1, -2}, {2, 0}, {1, 2}}, |
|
844 {{2, 0}, {1, 2}, { -1, 2}}, |
|
845 {{1, 2}, { -1, 2}, { -2, 0}}, |
|
846 {{ -1, 2}, { -2, 0}, { -1, -2}} |
|
847 }; |
|
848 |
|
849 int vp8_hex_search |
|
850 ( |
|
851 MACROBLOCK *x, |
|
852 BLOCK *b, |
|
853 BLOCKD *d, |
|
854 int_mv *ref_mv, |
|
855 int_mv *best_mv, |
|
856 int search_param, |
|
857 int sad_per_bit, |
|
858 const vp8_variance_fn_ptr_t *vfp, |
|
859 int *mvsadcost[2], |
|
860 int *mvcost[2], |
|
861 int_mv *center_mv |
|
862 ) |
|
863 { |
|
864 MV hex[6] = { { -1, -2}, {1, -2}, {2, 0}, {1, 2}, { -1, 2}, { -2, 0} } ; |
|
865 MV neighbors[4] = {{0, -1}, { -1, 0}, {1, 0}, {0, 1}} ; |
|
866 int i, j; |
|
867 |
|
868 unsigned char *what = (*(b->base_src) + b->src); |
|
869 int what_stride = b->src_stride; |
|
870 int pre_stride = x->e_mbd.pre.y_stride; |
|
871 unsigned char *base_pre = x->e_mbd.pre.y_buffer; |
|
872 |
|
873 int in_what_stride = pre_stride; |
|
874 int br, bc; |
|
875 int_mv this_mv; |
|
876 unsigned int bestsad; |
|
877 unsigned int thissad; |
|
878 unsigned char *base_offset; |
|
879 unsigned char *this_offset; |
|
880 int k = -1; |
|
881 int all_in; |
|
882 int best_site = -1; |
|
883 int hex_range = 127; |
|
884 int dia_range = 8; |
|
885 |
|
886 int_mv fcenter_mv; |
|
887 fcenter_mv.as_mv.row = center_mv->as_mv.row >> 3; |
|
888 fcenter_mv.as_mv.col = center_mv->as_mv.col >> 3; |
|
889 |
|
890 /* adjust ref_mv to make sure it is within MV range */ |
|
891 vp8_clamp_mv(ref_mv, x->mv_col_min, x->mv_col_max, x->mv_row_min, x->mv_row_max); |
|
892 br = ref_mv->as_mv.row; |
|
893 bc = ref_mv->as_mv.col; |
|
894 |
|
895 /* Work out the start point for the search */ |
|
896 base_offset = (unsigned char *)(base_pre + d->offset); |
|
897 this_offset = base_offset + (br * (pre_stride)) + bc; |
|
898 this_mv.as_mv.row = br; |
|
899 this_mv.as_mv.col = bc; |
|
900 bestsad = vfp->sdf(what, what_stride, this_offset, in_what_stride, UINT_MAX) |
|
901 + mvsad_err_cost(&this_mv, &fcenter_mv, mvsadcost, sad_per_bit); |
|
902 |
|
903 #if CONFIG_MULTI_RES_ENCODING |
|
904 /* Lower search range based on prediction info */ |
|
905 if (search_param >= 6) goto cal_neighbors; |
|
906 else if (search_param >= 5) hex_range = 4; |
|
907 else if (search_param >= 4) hex_range = 6; |
|
908 else if (search_param >= 3) hex_range = 15; |
|
909 else if (search_param >= 2) hex_range = 31; |
|
910 else if (search_param >= 1) hex_range = 63; |
|
911 |
|
912 dia_range = 8; |
|
913 #endif |
|
914 |
|
915 /* hex search */ |
|
916 CHECK_BOUNDS(2) |
|
917 |
|
918 if(all_in) |
|
919 { |
|
920 for (i = 0; i < 6; i++) |
|
921 { |
|
922 this_mv.as_mv.row = br + hex[i].row; |
|
923 this_mv.as_mv.col = bc + hex[i].col; |
|
924 this_offset = base_offset + (this_mv.as_mv.row * in_what_stride) + this_mv.as_mv.col; |
|
925 thissad = vfp->sdf(what, what_stride, this_offset, in_what_stride, bestsad); |
|
926 CHECK_BETTER |
|
927 } |
|
928 }else |
|
929 { |
|
930 for (i = 0; i < 6; i++) |
|
931 { |
|
932 this_mv.as_mv.row = br + hex[i].row; |
|
933 this_mv.as_mv.col = bc + hex[i].col; |
|
934 CHECK_POINT |
|
935 this_offset = base_offset + (this_mv.as_mv.row * in_what_stride) + this_mv.as_mv.col; |
|
936 thissad = vfp->sdf(what, what_stride, this_offset, in_what_stride, bestsad); |
|
937 CHECK_BETTER |
|
938 } |
|
939 } |
|
940 |
|
941 if (best_site == -1) |
|
942 goto cal_neighbors; |
|
943 else |
|
944 { |
|
945 br += hex[best_site].row; |
|
946 bc += hex[best_site].col; |
|
947 k = best_site; |
|
948 } |
|
949 |
|
950 for (j = 1; j < hex_range; j++) |
|
951 { |
|
952 best_site = -1; |
|
953 CHECK_BOUNDS(2) |
|
954 |
|
955 if(all_in) |
|
956 { |
|
957 for (i = 0; i < 3; i++) |
|
958 { |
|
959 this_mv.as_mv.row = br + next_chkpts[k][i].row; |
|
960 this_mv.as_mv.col = bc + next_chkpts[k][i].col; |
|
961 this_offset = base_offset + (this_mv.as_mv.row * (in_what_stride)) + this_mv.as_mv.col; |
|
962 thissad = vfp->sdf(what, what_stride, this_offset, in_what_stride, bestsad); |
|
963 CHECK_BETTER |
|
964 } |
|
965 }else |
|
966 { |
|
967 for (i = 0; i < 3; i++) |
|
968 { |
|
969 this_mv.as_mv.row = br + next_chkpts[k][i].row; |
|
970 this_mv.as_mv.col = bc + next_chkpts[k][i].col; |
|
971 CHECK_POINT |
|
972 this_offset = base_offset + (this_mv.as_mv.row * (in_what_stride)) + this_mv.as_mv.col; |
|
973 thissad = vfp->sdf(what, what_stride, this_offset, in_what_stride, bestsad); |
|
974 CHECK_BETTER |
|
975 } |
|
976 } |
|
977 |
|
978 if (best_site == -1) |
|
979 break; |
|
980 else |
|
981 { |
|
982 br += next_chkpts[k][best_site].row; |
|
983 bc += next_chkpts[k][best_site].col; |
|
984 k += 5 + best_site; |
|
985 if (k >= 12) k -= 12; |
|
986 else if (k >= 6) k -= 6; |
|
987 } |
|
988 } |
|
989 |
|
990 /* check 4 1-away neighbors */ |
|
991 cal_neighbors: |
|
992 for (j = 0; j < dia_range; j++) |
|
993 { |
|
994 best_site = -1; |
|
995 CHECK_BOUNDS(1) |
|
996 |
|
997 if(all_in) |
|
998 { |
|
999 for (i = 0; i < 4; i++) |
|
1000 { |
|
1001 this_mv.as_mv.row = br + neighbors[i].row; |
|
1002 this_mv.as_mv.col = bc + neighbors[i].col; |
|
1003 this_offset = base_offset + (this_mv.as_mv.row * (in_what_stride)) + this_mv.as_mv.col; |
|
1004 thissad = vfp->sdf(what, what_stride, this_offset, in_what_stride, bestsad); |
|
1005 CHECK_BETTER |
|
1006 } |
|
1007 }else |
|
1008 { |
|
1009 for (i = 0; i < 4; i++) |
|
1010 { |
|
1011 this_mv.as_mv.row = br + neighbors[i].row; |
|
1012 this_mv.as_mv.col = bc + neighbors[i].col; |
|
1013 CHECK_POINT |
|
1014 this_offset = base_offset + (this_mv.as_mv.row * (in_what_stride)) + this_mv.as_mv.col; |
|
1015 thissad = vfp->sdf(what, what_stride, this_offset, in_what_stride, bestsad); |
|
1016 CHECK_BETTER |
|
1017 } |
|
1018 } |
|
1019 |
|
1020 if (best_site == -1) |
|
1021 break; |
|
1022 else |
|
1023 { |
|
1024 br += neighbors[best_site].row; |
|
1025 bc += neighbors[best_site].col; |
|
1026 } |
|
1027 } |
|
1028 |
|
1029 best_mv->as_mv.row = br; |
|
1030 best_mv->as_mv.col = bc; |
|
1031 |
|
1032 return bestsad; |
|
1033 } |
|
1034 #undef CHECK_BOUNDS |
|
1035 #undef CHECK_POINT |
|
1036 #undef CHECK_BETTER |
|
1037 |
|
1038 int vp8_diamond_search_sad_c |
|
1039 ( |
|
1040 MACROBLOCK *x, |
|
1041 BLOCK *b, |
|
1042 BLOCKD *d, |
|
1043 int_mv *ref_mv, |
|
1044 int_mv *best_mv, |
|
1045 int search_param, |
|
1046 int sad_per_bit, |
|
1047 int *num00, |
|
1048 vp8_variance_fn_ptr_t *fn_ptr, |
|
1049 int *mvcost[2], |
|
1050 int_mv *center_mv |
|
1051 ) |
|
1052 { |
|
1053 int i, j, step; |
|
1054 |
|
1055 unsigned char *what = (*(b->base_src) + b->src); |
|
1056 int what_stride = b->src_stride; |
|
1057 unsigned char *in_what; |
|
1058 int pre_stride = x->e_mbd.pre.y_stride; |
|
1059 unsigned char *base_pre = x->e_mbd.pre.y_buffer; |
|
1060 int in_what_stride = pre_stride; |
|
1061 unsigned char *best_address; |
|
1062 |
|
1063 int tot_steps; |
|
1064 int_mv this_mv; |
|
1065 |
|
1066 unsigned int bestsad; |
|
1067 unsigned int thissad; |
|
1068 int best_site = 0; |
|
1069 int last_site = 0; |
|
1070 |
|
1071 int ref_row; |
|
1072 int ref_col; |
|
1073 int this_row_offset; |
|
1074 int this_col_offset; |
|
1075 search_site *ss; |
|
1076 |
|
1077 unsigned char *check_here; |
|
1078 |
|
1079 int *mvsadcost[2]; |
|
1080 int_mv fcenter_mv; |
|
1081 |
|
1082 mvsadcost[0] = x->mvsadcost[0]; |
|
1083 mvsadcost[1] = x->mvsadcost[1]; |
|
1084 fcenter_mv.as_mv.row = center_mv->as_mv.row >> 3; |
|
1085 fcenter_mv.as_mv.col = center_mv->as_mv.col >> 3; |
|
1086 |
|
1087 vp8_clamp_mv(ref_mv, x->mv_col_min, x->mv_col_max, x->mv_row_min, x->mv_row_max); |
|
1088 ref_row = ref_mv->as_mv.row; |
|
1089 ref_col = ref_mv->as_mv.col; |
|
1090 *num00 = 0; |
|
1091 best_mv->as_mv.row = ref_row; |
|
1092 best_mv->as_mv.col = ref_col; |
|
1093 |
|
1094 /* Work out the start point for the search */ |
|
1095 in_what = (unsigned char *)(base_pre + d->offset + (ref_row * pre_stride) + ref_col); |
|
1096 best_address = in_what; |
|
1097 |
|
1098 /* Check the starting position */ |
|
1099 bestsad = fn_ptr->sdf(what, what_stride, in_what, in_what_stride, UINT_MAX) |
|
1100 + mvsad_err_cost(best_mv, &fcenter_mv, mvsadcost, sad_per_bit); |
|
1101 |
|
1102 /* search_param determines the length of the initial step and hence |
|
1103 * the number of iterations 0 = initial step (MAX_FIRST_STEP) pel : |
|
1104 * 1 = (MAX_FIRST_STEP/2) pel, 2 = (MAX_FIRST_STEP/4) pel... etc. |
|
1105 */ |
|
1106 ss = &x->ss[search_param * x->searches_per_step]; |
|
1107 tot_steps = (x->ss_count / x->searches_per_step) - search_param; |
|
1108 |
|
1109 i = 1; |
|
1110 |
|
1111 for (step = 0; step < tot_steps ; step++) |
|
1112 { |
|
1113 for (j = 0 ; j < x->searches_per_step ; j++) |
|
1114 { |
|
1115 /* Trap illegal vectors */ |
|
1116 this_row_offset = best_mv->as_mv.row + ss[i].mv.row; |
|
1117 this_col_offset = best_mv->as_mv.col + ss[i].mv.col; |
|
1118 |
|
1119 if ((this_col_offset > x->mv_col_min) && (this_col_offset < x->mv_col_max) && |
|
1120 (this_row_offset > x->mv_row_min) && (this_row_offset < x->mv_row_max)) |
|
1121 |
|
1122 { |
|
1123 check_here = ss[i].offset + best_address; |
|
1124 thissad = fn_ptr->sdf(what, what_stride, check_here, in_what_stride, bestsad); |
|
1125 |
|
1126 if (thissad < bestsad) |
|
1127 { |
|
1128 this_mv.as_mv.row = this_row_offset; |
|
1129 this_mv.as_mv.col = this_col_offset; |
|
1130 thissad += mvsad_err_cost(&this_mv, &fcenter_mv, |
|
1131 mvsadcost, sad_per_bit); |
|
1132 |
|
1133 if (thissad < bestsad) |
|
1134 { |
|
1135 bestsad = thissad; |
|
1136 best_site = i; |
|
1137 } |
|
1138 } |
|
1139 } |
|
1140 |
|
1141 i++; |
|
1142 } |
|
1143 |
|
1144 if (best_site != last_site) |
|
1145 { |
|
1146 best_mv->as_mv.row += ss[best_site].mv.row; |
|
1147 best_mv->as_mv.col += ss[best_site].mv.col; |
|
1148 best_address += ss[best_site].offset; |
|
1149 last_site = best_site; |
|
1150 } |
|
1151 else if (best_address == in_what) |
|
1152 (*num00)++; |
|
1153 } |
|
1154 |
|
1155 this_mv.as_mv.row = best_mv->as_mv.row << 3; |
|
1156 this_mv.as_mv.col = best_mv->as_mv.col << 3; |
|
1157 |
|
1158 return fn_ptr->vf(what, what_stride, best_address, in_what_stride, &thissad) |
|
1159 + mv_err_cost(&this_mv, center_mv, mvcost, x->errorperbit); |
|
1160 } |
|
1161 |
|
1162 int vp8_diamond_search_sadx4 |
|
1163 ( |
|
1164 MACROBLOCK *x, |
|
1165 BLOCK *b, |
|
1166 BLOCKD *d, |
|
1167 int_mv *ref_mv, |
|
1168 int_mv *best_mv, |
|
1169 int search_param, |
|
1170 int sad_per_bit, |
|
1171 int *num00, |
|
1172 vp8_variance_fn_ptr_t *fn_ptr, |
|
1173 int *mvcost[2], |
|
1174 int_mv *center_mv |
|
1175 ) |
|
1176 { |
|
1177 int i, j, step; |
|
1178 |
|
1179 unsigned char *what = (*(b->base_src) + b->src); |
|
1180 int what_stride = b->src_stride; |
|
1181 unsigned char *in_what; |
|
1182 int pre_stride = x->e_mbd.pre.y_stride; |
|
1183 unsigned char *base_pre = x->e_mbd.pre.y_buffer; |
|
1184 int in_what_stride = pre_stride; |
|
1185 unsigned char *best_address; |
|
1186 |
|
1187 int tot_steps; |
|
1188 int_mv this_mv; |
|
1189 |
|
1190 unsigned int bestsad; |
|
1191 unsigned int thissad; |
|
1192 int best_site = 0; |
|
1193 int last_site = 0; |
|
1194 |
|
1195 int ref_row; |
|
1196 int ref_col; |
|
1197 int this_row_offset; |
|
1198 int this_col_offset; |
|
1199 search_site *ss; |
|
1200 |
|
1201 unsigned char *check_here; |
|
1202 |
|
1203 int *mvsadcost[2]; |
|
1204 int_mv fcenter_mv; |
|
1205 |
|
1206 mvsadcost[0] = x->mvsadcost[0]; |
|
1207 mvsadcost[1] = x->mvsadcost[1]; |
|
1208 fcenter_mv.as_mv.row = center_mv->as_mv.row >> 3; |
|
1209 fcenter_mv.as_mv.col = center_mv->as_mv.col >> 3; |
|
1210 |
|
1211 vp8_clamp_mv(ref_mv, x->mv_col_min, x->mv_col_max, x->mv_row_min, x->mv_row_max); |
|
1212 ref_row = ref_mv->as_mv.row; |
|
1213 ref_col = ref_mv->as_mv.col; |
|
1214 *num00 = 0; |
|
1215 best_mv->as_mv.row = ref_row; |
|
1216 best_mv->as_mv.col = ref_col; |
|
1217 |
|
1218 /* Work out the start point for the search */ |
|
1219 in_what = (unsigned char *)(base_pre + d->offset + (ref_row * pre_stride) + ref_col); |
|
1220 best_address = in_what; |
|
1221 |
|
1222 /* Check the starting position */ |
|
1223 bestsad = fn_ptr->sdf(what, what_stride, in_what, in_what_stride, UINT_MAX) |
|
1224 + mvsad_err_cost(best_mv, &fcenter_mv, mvsadcost, sad_per_bit); |
|
1225 |
|
1226 /* search_param determines the length of the initial step and hence the |
|
1227 * number of iterations 0 = initial step (MAX_FIRST_STEP) pel : 1 = |
|
1228 * (MAX_FIRST_STEP/2) pel, 2 = (MAX_FIRST_STEP/4) pel... etc. |
|
1229 */ |
|
1230 ss = &x->ss[search_param * x->searches_per_step]; |
|
1231 tot_steps = (x->ss_count / x->searches_per_step) - search_param; |
|
1232 |
|
1233 i = 1; |
|
1234 |
|
1235 for (step = 0; step < tot_steps ; step++) |
|
1236 { |
|
1237 int all_in = 1, t; |
|
1238 |
|
1239 /* To know if all neighbor points are within the bounds, 4 bounds |
|
1240 * checking are enough instead of checking 4 bounds for each |
|
1241 * points. |
|
1242 */ |
|
1243 all_in &= ((best_mv->as_mv.row + ss[i].mv.row)> x->mv_row_min); |
|
1244 all_in &= ((best_mv->as_mv.row + ss[i+1].mv.row) < x->mv_row_max); |
|
1245 all_in &= ((best_mv->as_mv.col + ss[i+2].mv.col) > x->mv_col_min); |
|
1246 all_in &= ((best_mv->as_mv.col + ss[i+3].mv.col) < x->mv_col_max); |
|
1247 |
|
1248 if (all_in) |
|
1249 { |
|
1250 unsigned int sad_array[4]; |
|
1251 |
|
1252 for (j = 0 ; j < x->searches_per_step ; j += 4) |
|
1253 { |
|
1254 const unsigned char *block_offset[4]; |
|
1255 |
|
1256 for (t = 0; t < 4; t++) |
|
1257 block_offset[t] = ss[i+t].offset + best_address; |
|
1258 |
|
1259 fn_ptr->sdx4df(what, what_stride, block_offset, in_what_stride, sad_array); |
|
1260 |
|
1261 for (t = 0; t < 4; t++, i++) |
|
1262 { |
|
1263 if (sad_array[t] < bestsad) |
|
1264 { |
|
1265 this_mv.as_mv.row = best_mv->as_mv.row + ss[i].mv.row; |
|
1266 this_mv.as_mv.col = best_mv->as_mv.col + ss[i].mv.col; |
|
1267 sad_array[t] += mvsad_err_cost(&this_mv, &fcenter_mv, |
|
1268 mvsadcost, sad_per_bit); |
|
1269 |
|
1270 if (sad_array[t] < bestsad) |
|
1271 { |
|
1272 bestsad = sad_array[t]; |
|
1273 best_site = i; |
|
1274 } |
|
1275 } |
|
1276 } |
|
1277 } |
|
1278 } |
|
1279 else |
|
1280 { |
|
1281 for (j = 0 ; j < x->searches_per_step ; j++) |
|
1282 { |
|
1283 /* Trap illegal vectors */ |
|
1284 this_row_offset = best_mv->as_mv.row + ss[i].mv.row; |
|
1285 this_col_offset = best_mv->as_mv.col + ss[i].mv.col; |
|
1286 |
|
1287 if ((this_col_offset > x->mv_col_min) && (this_col_offset < x->mv_col_max) && |
|
1288 (this_row_offset > x->mv_row_min) && (this_row_offset < x->mv_row_max)) |
|
1289 { |
|
1290 check_here = ss[i].offset + best_address; |
|
1291 thissad = fn_ptr->sdf(what, what_stride, check_here, in_what_stride, bestsad); |
|
1292 |
|
1293 if (thissad < bestsad) |
|
1294 { |
|
1295 this_mv.as_mv.row = this_row_offset; |
|
1296 this_mv.as_mv.col = this_col_offset; |
|
1297 thissad += mvsad_err_cost(&this_mv, &fcenter_mv, |
|
1298 mvsadcost, sad_per_bit); |
|
1299 |
|
1300 if (thissad < bestsad) |
|
1301 { |
|
1302 bestsad = thissad; |
|
1303 best_site = i; |
|
1304 } |
|
1305 } |
|
1306 } |
|
1307 i++; |
|
1308 } |
|
1309 } |
|
1310 |
|
1311 if (best_site != last_site) |
|
1312 { |
|
1313 best_mv->as_mv.row += ss[best_site].mv.row; |
|
1314 best_mv->as_mv.col += ss[best_site].mv.col; |
|
1315 best_address += ss[best_site].offset; |
|
1316 last_site = best_site; |
|
1317 } |
|
1318 else if (best_address == in_what) |
|
1319 (*num00)++; |
|
1320 } |
|
1321 |
|
1322 this_mv.as_mv.row = best_mv->as_mv.row * 8; |
|
1323 this_mv.as_mv.col = best_mv->as_mv.col * 8; |
|
1324 |
|
1325 return fn_ptr->vf(what, what_stride, best_address, in_what_stride, &thissad) |
|
1326 + mv_err_cost(&this_mv, center_mv, mvcost, x->errorperbit); |
|
1327 } |
|
1328 |
|
1329 int vp8_full_search_sad_c(MACROBLOCK *x, BLOCK *b, BLOCKD *d, int_mv *ref_mv, |
|
1330 int sad_per_bit, int distance, |
|
1331 vp8_variance_fn_ptr_t *fn_ptr, int *mvcost[2], |
|
1332 int_mv *center_mv) |
|
1333 { |
|
1334 unsigned char *what = (*(b->base_src) + b->src); |
|
1335 int what_stride = b->src_stride; |
|
1336 unsigned char *in_what; |
|
1337 int pre_stride = x->e_mbd.pre.y_stride; |
|
1338 unsigned char *base_pre = x->e_mbd.pre.y_buffer; |
|
1339 int in_what_stride = pre_stride; |
|
1340 int mv_stride = pre_stride; |
|
1341 unsigned char *bestaddress; |
|
1342 int_mv *best_mv = &d->bmi.mv; |
|
1343 int_mv this_mv; |
|
1344 unsigned int bestsad; |
|
1345 unsigned int thissad; |
|
1346 int r, c; |
|
1347 |
|
1348 unsigned char *check_here; |
|
1349 |
|
1350 int ref_row = ref_mv->as_mv.row; |
|
1351 int ref_col = ref_mv->as_mv.col; |
|
1352 |
|
1353 int row_min = ref_row - distance; |
|
1354 int row_max = ref_row + distance; |
|
1355 int col_min = ref_col - distance; |
|
1356 int col_max = ref_col + distance; |
|
1357 |
|
1358 int *mvsadcost[2]; |
|
1359 int_mv fcenter_mv; |
|
1360 |
|
1361 mvsadcost[0] = x->mvsadcost[0]; |
|
1362 mvsadcost[1] = x->mvsadcost[1]; |
|
1363 fcenter_mv.as_mv.row = center_mv->as_mv.row >> 3; |
|
1364 fcenter_mv.as_mv.col = center_mv->as_mv.col >> 3; |
|
1365 |
|
1366 /* Work out the mid point for the search */ |
|
1367 in_what = base_pre + d->offset; |
|
1368 bestaddress = in_what + (ref_row * pre_stride) + ref_col; |
|
1369 |
|
1370 best_mv->as_mv.row = ref_row; |
|
1371 best_mv->as_mv.col = ref_col; |
|
1372 |
|
1373 /* Baseline value at the centre */ |
|
1374 bestsad = fn_ptr->sdf(what, what_stride, bestaddress, |
|
1375 in_what_stride, UINT_MAX) |
|
1376 + mvsad_err_cost(best_mv, &fcenter_mv, mvsadcost, sad_per_bit); |
|
1377 |
|
1378 /* Apply further limits to prevent us looking using vectors that |
|
1379 * stretch beyiond the UMV border |
|
1380 */ |
|
1381 if (col_min < x->mv_col_min) |
|
1382 col_min = x->mv_col_min; |
|
1383 |
|
1384 if (col_max > x->mv_col_max) |
|
1385 col_max = x->mv_col_max; |
|
1386 |
|
1387 if (row_min < x->mv_row_min) |
|
1388 row_min = x->mv_row_min; |
|
1389 |
|
1390 if (row_max > x->mv_row_max) |
|
1391 row_max = x->mv_row_max; |
|
1392 |
|
1393 for (r = row_min; r < row_max ; r++) |
|
1394 { |
|
1395 this_mv.as_mv.row = r; |
|
1396 check_here = r * mv_stride + in_what + col_min; |
|
1397 |
|
1398 for (c = col_min; c < col_max; c++) |
|
1399 { |
|
1400 thissad = fn_ptr->sdf(what, what_stride, check_here, in_what_stride, bestsad); |
|
1401 |
|
1402 this_mv.as_mv.col = c; |
|
1403 thissad += mvsad_err_cost(&this_mv, &fcenter_mv, |
|
1404 mvsadcost, sad_per_bit); |
|
1405 |
|
1406 if (thissad < bestsad) |
|
1407 { |
|
1408 bestsad = thissad; |
|
1409 best_mv->as_mv.row = r; |
|
1410 best_mv->as_mv.col = c; |
|
1411 bestaddress = check_here; |
|
1412 } |
|
1413 |
|
1414 check_here++; |
|
1415 } |
|
1416 } |
|
1417 |
|
1418 this_mv.as_mv.row = best_mv->as_mv.row << 3; |
|
1419 this_mv.as_mv.col = best_mv->as_mv.col << 3; |
|
1420 |
|
1421 return fn_ptr->vf(what, what_stride, bestaddress, in_what_stride, &thissad) |
|
1422 + mv_err_cost(&this_mv, center_mv, mvcost, x->errorperbit); |
|
1423 } |
|
1424 |
|
1425 int vp8_full_search_sadx3(MACROBLOCK *x, BLOCK *b, BLOCKD *d, int_mv *ref_mv, |
|
1426 int sad_per_bit, int distance, |
|
1427 vp8_variance_fn_ptr_t *fn_ptr, int *mvcost[2], |
|
1428 int_mv *center_mv) |
|
1429 { |
|
1430 unsigned char *what = (*(b->base_src) + b->src); |
|
1431 int what_stride = b->src_stride; |
|
1432 unsigned char *in_what; |
|
1433 int pre_stride = x->e_mbd.pre.y_stride; |
|
1434 unsigned char *base_pre = x->e_mbd.pre.y_buffer; |
|
1435 int in_what_stride = pre_stride; |
|
1436 int mv_stride = pre_stride; |
|
1437 unsigned char *bestaddress; |
|
1438 int_mv *best_mv = &d->bmi.mv; |
|
1439 int_mv this_mv; |
|
1440 unsigned int bestsad; |
|
1441 unsigned int thissad; |
|
1442 int r, c; |
|
1443 |
|
1444 unsigned char *check_here; |
|
1445 |
|
1446 int ref_row = ref_mv->as_mv.row; |
|
1447 int ref_col = ref_mv->as_mv.col; |
|
1448 |
|
1449 int row_min = ref_row - distance; |
|
1450 int row_max = ref_row + distance; |
|
1451 int col_min = ref_col - distance; |
|
1452 int col_max = ref_col + distance; |
|
1453 |
|
1454 unsigned int sad_array[3]; |
|
1455 |
|
1456 int *mvsadcost[2]; |
|
1457 int_mv fcenter_mv; |
|
1458 |
|
1459 mvsadcost[0] = x->mvsadcost[0]; |
|
1460 mvsadcost[1] = x->mvsadcost[1]; |
|
1461 fcenter_mv.as_mv.row = center_mv->as_mv.row >> 3; |
|
1462 fcenter_mv.as_mv.col = center_mv->as_mv.col >> 3; |
|
1463 |
|
1464 /* Work out the mid point for the search */ |
|
1465 in_what = base_pre + d->offset; |
|
1466 bestaddress = in_what + (ref_row * pre_stride) + ref_col; |
|
1467 |
|
1468 best_mv->as_mv.row = ref_row; |
|
1469 best_mv->as_mv.col = ref_col; |
|
1470 |
|
1471 /* Baseline value at the centre */ |
|
1472 bestsad = fn_ptr->sdf(what, what_stride, bestaddress, |
|
1473 in_what_stride, UINT_MAX) |
|
1474 + mvsad_err_cost(best_mv, &fcenter_mv, mvsadcost, sad_per_bit); |
|
1475 |
|
1476 /* Apply further limits to prevent us looking using vectors that stretch |
|
1477 * beyond the UMV border |
|
1478 */ |
|
1479 if (col_min < x->mv_col_min) |
|
1480 col_min = x->mv_col_min; |
|
1481 |
|
1482 if (col_max > x->mv_col_max) |
|
1483 col_max = x->mv_col_max; |
|
1484 |
|
1485 if (row_min < x->mv_row_min) |
|
1486 row_min = x->mv_row_min; |
|
1487 |
|
1488 if (row_max > x->mv_row_max) |
|
1489 row_max = x->mv_row_max; |
|
1490 |
|
1491 for (r = row_min; r < row_max ; r++) |
|
1492 { |
|
1493 this_mv.as_mv.row = r; |
|
1494 check_here = r * mv_stride + in_what + col_min; |
|
1495 c = col_min; |
|
1496 |
|
1497 while ((c + 2) < col_max) |
|
1498 { |
|
1499 int i; |
|
1500 |
|
1501 fn_ptr->sdx3f(what, what_stride, check_here, in_what_stride, sad_array); |
|
1502 |
|
1503 for (i = 0; i < 3; i++) |
|
1504 { |
|
1505 thissad = sad_array[i]; |
|
1506 |
|
1507 if (thissad < bestsad) |
|
1508 { |
|
1509 this_mv.as_mv.col = c; |
|
1510 thissad += mvsad_err_cost(&this_mv, &fcenter_mv, |
|
1511 mvsadcost, sad_per_bit); |
|
1512 |
|
1513 if (thissad < bestsad) |
|
1514 { |
|
1515 bestsad = thissad; |
|
1516 best_mv->as_mv.row = r; |
|
1517 best_mv->as_mv.col = c; |
|
1518 bestaddress = check_here; |
|
1519 } |
|
1520 } |
|
1521 |
|
1522 check_here++; |
|
1523 c++; |
|
1524 } |
|
1525 } |
|
1526 |
|
1527 while (c < col_max) |
|
1528 { |
|
1529 thissad = fn_ptr->sdf(what, what_stride, check_here, in_what_stride, bestsad); |
|
1530 |
|
1531 if (thissad < bestsad) |
|
1532 { |
|
1533 this_mv.as_mv.col = c; |
|
1534 thissad += mvsad_err_cost(&this_mv, &fcenter_mv, |
|
1535 mvsadcost, sad_per_bit); |
|
1536 |
|
1537 if (thissad < bestsad) |
|
1538 { |
|
1539 bestsad = thissad; |
|
1540 best_mv->as_mv.row = r; |
|
1541 best_mv->as_mv.col = c; |
|
1542 bestaddress = check_here; |
|
1543 } |
|
1544 } |
|
1545 |
|
1546 check_here ++; |
|
1547 c ++; |
|
1548 } |
|
1549 |
|
1550 } |
|
1551 |
|
1552 this_mv.as_mv.row = best_mv->as_mv.row << 3; |
|
1553 this_mv.as_mv.col = best_mv->as_mv.col << 3; |
|
1554 |
|
1555 return fn_ptr->vf(what, what_stride, bestaddress, in_what_stride, &thissad) |
|
1556 + mv_err_cost(&this_mv, center_mv, mvcost, x->errorperbit); |
|
1557 } |
|
1558 |
|
1559 int vp8_full_search_sadx8(MACROBLOCK *x, BLOCK *b, BLOCKD *d, int_mv *ref_mv, |
|
1560 int sad_per_bit, int distance, |
|
1561 vp8_variance_fn_ptr_t *fn_ptr, int *mvcost[2], |
|
1562 int_mv *center_mv) |
|
1563 { |
|
1564 unsigned char *what = (*(b->base_src) + b->src); |
|
1565 int what_stride = b->src_stride; |
|
1566 int pre_stride = x->e_mbd.pre.y_stride; |
|
1567 unsigned char *base_pre = x->e_mbd.pre.y_buffer; |
|
1568 unsigned char *in_what; |
|
1569 int in_what_stride = pre_stride; |
|
1570 int mv_stride = pre_stride; |
|
1571 unsigned char *bestaddress; |
|
1572 int_mv *best_mv = &d->bmi.mv; |
|
1573 int_mv this_mv; |
|
1574 unsigned int bestsad; |
|
1575 unsigned int thissad; |
|
1576 int r, c; |
|
1577 |
|
1578 unsigned char *check_here; |
|
1579 |
|
1580 int ref_row = ref_mv->as_mv.row; |
|
1581 int ref_col = ref_mv->as_mv.col; |
|
1582 |
|
1583 int row_min = ref_row - distance; |
|
1584 int row_max = ref_row + distance; |
|
1585 int col_min = ref_col - distance; |
|
1586 int col_max = ref_col + distance; |
|
1587 |
|
1588 DECLARE_ALIGNED_ARRAY(16, unsigned short, sad_array8, 8); |
|
1589 unsigned int sad_array[3]; |
|
1590 |
|
1591 int *mvsadcost[2]; |
|
1592 int_mv fcenter_mv; |
|
1593 |
|
1594 mvsadcost[0] = x->mvsadcost[0]; |
|
1595 mvsadcost[1] = x->mvsadcost[1]; |
|
1596 fcenter_mv.as_mv.row = center_mv->as_mv.row >> 3; |
|
1597 fcenter_mv.as_mv.col = center_mv->as_mv.col >> 3; |
|
1598 |
|
1599 /* Work out the mid point for the search */ |
|
1600 in_what = base_pre + d->offset; |
|
1601 bestaddress = in_what + (ref_row * pre_stride) + ref_col; |
|
1602 |
|
1603 best_mv->as_mv.row = ref_row; |
|
1604 best_mv->as_mv.col = ref_col; |
|
1605 |
|
1606 /* Baseline value at the centre */ |
|
1607 bestsad = fn_ptr->sdf(what, what_stride, |
|
1608 bestaddress, in_what_stride, UINT_MAX) |
|
1609 + mvsad_err_cost(best_mv, &fcenter_mv, mvsadcost, sad_per_bit); |
|
1610 |
|
1611 /* Apply further limits to prevent us looking using vectors that stretch |
|
1612 * beyond the UMV border |
|
1613 */ |
|
1614 if (col_min < x->mv_col_min) |
|
1615 col_min = x->mv_col_min; |
|
1616 |
|
1617 if (col_max > x->mv_col_max) |
|
1618 col_max = x->mv_col_max; |
|
1619 |
|
1620 if (row_min < x->mv_row_min) |
|
1621 row_min = x->mv_row_min; |
|
1622 |
|
1623 if (row_max > x->mv_row_max) |
|
1624 row_max = x->mv_row_max; |
|
1625 |
|
1626 for (r = row_min; r < row_max ; r++) |
|
1627 { |
|
1628 this_mv.as_mv.row = r; |
|
1629 check_here = r * mv_stride + in_what + col_min; |
|
1630 c = col_min; |
|
1631 |
|
1632 while ((c + 7) < col_max) |
|
1633 { |
|
1634 int i; |
|
1635 |
|
1636 fn_ptr->sdx8f(what, what_stride, check_here, in_what_stride, sad_array8); |
|
1637 |
|
1638 for (i = 0; i < 8; i++) |
|
1639 { |
|
1640 thissad = sad_array8[i]; |
|
1641 |
|
1642 if (thissad < bestsad) |
|
1643 { |
|
1644 this_mv.as_mv.col = c; |
|
1645 thissad += mvsad_err_cost(&this_mv, &fcenter_mv, |
|
1646 mvsadcost, sad_per_bit); |
|
1647 |
|
1648 if (thissad < bestsad) |
|
1649 { |
|
1650 bestsad = thissad; |
|
1651 best_mv->as_mv.row = r; |
|
1652 best_mv->as_mv.col = c; |
|
1653 bestaddress = check_here; |
|
1654 } |
|
1655 } |
|
1656 |
|
1657 check_here++; |
|
1658 c++; |
|
1659 } |
|
1660 } |
|
1661 |
|
1662 while ((c + 2) < col_max) |
|
1663 { |
|
1664 int i; |
|
1665 |
|
1666 fn_ptr->sdx3f(what, what_stride, check_here , in_what_stride, sad_array); |
|
1667 |
|
1668 for (i = 0; i < 3; i++) |
|
1669 { |
|
1670 thissad = sad_array[i]; |
|
1671 |
|
1672 if (thissad < bestsad) |
|
1673 { |
|
1674 this_mv.as_mv.col = c; |
|
1675 thissad += mvsad_err_cost(&this_mv, &fcenter_mv, |
|
1676 mvsadcost, sad_per_bit); |
|
1677 |
|
1678 if (thissad < bestsad) |
|
1679 { |
|
1680 bestsad = thissad; |
|
1681 best_mv->as_mv.row = r; |
|
1682 best_mv->as_mv.col = c; |
|
1683 bestaddress = check_here; |
|
1684 } |
|
1685 } |
|
1686 |
|
1687 check_here++; |
|
1688 c++; |
|
1689 } |
|
1690 } |
|
1691 |
|
1692 while (c < col_max) |
|
1693 { |
|
1694 thissad = fn_ptr->sdf(what, what_stride, check_here , in_what_stride, bestsad); |
|
1695 |
|
1696 if (thissad < bestsad) |
|
1697 { |
|
1698 this_mv.as_mv.col = c; |
|
1699 thissad += mvsad_err_cost(&this_mv, &fcenter_mv, |
|
1700 mvsadcost, sad_per_bit); |
|
1701 |
|
1702 if (thissad < bestsad) |
|
1703 { |
|
1704 bestsad = thissad; |
|
1705 best_mv->as_mv.row = r; |
|
1706 best_mv->as_mv.col = c; |
|
1707 bestaddress = check_here; |
|
1708 } |
|
1709 } |
|
1710 |
|
1711 check_here ++; |
|
1712 c ++; |
|
1713 } |
|
1714 } |
|
1715 |
|
1716 this_mv.as_mv.row = best_mv->as_mv.row * 8; |
|
1717 this_mv.as_mv.col = best_mv->as_mv.col * 8; |
|
1718 |
|
1719 return fn_ptr->vf(what, what_stride, bestaddress, in_what_stride, &thissad) |
|
1720 + mv_err_cost(&this_mv, center_mv, mvcost, x->errorperbit); |
|
1721 } |
|
1722 |
|
1723 int vp8_refining_search_sad_c(MACROBLOCK *x, BLOCK *b, BLOCKD *d, int_mv *ref_mv, |
|
1724 int error_per_bit, int search_range, |
|
1725 vp8_variance_fn_ptr_t *fn_ptr, int *mvcost[2], |
|
1726 int_mv *center_mv) |
|
1727 { |
|
1728 MV neighbors[4] = {{-1, 0}, {0, -1}, {0, 1}, {1, 0}}; |
|
1729 int i, j; |
|
1730 short this_row_offset, this_col_offset; |
|
1731 |
|
1732 int what_stride = b->src_stride; |
|
1733 int pre_stride = x->e_mbd.pre.y_stride; |
|
1734 unsigned char *base_pre = x->e_mbd.pre.y_buffer; |
|
1735 int in_what_stride = pre_stride; |
|
1736 unsigned char *what = (*(b->base_src) + b->src); |
|
1737 unsigned char *best_address = (unsigned char *)(base_pre + d->offset + |
|
1738 (ref_mv->as_mv.row * pre_stride) + ref_mv->as_mv.col); |
|
1739 unsigned char *check_here; |
|
1740 int_mv this_mv; |
|
1741 unsigned int bestsad; |
|
1742 unsigned int thissad; |
|
1743 |
|
1744 int *mvsadcost[2]; |
|
1745 int_mv fcenter_mv; |
|
1746 |
|
1747 mvsadcost[0] = x->mvsadcost[0]; |
|
1748 mvsadcost[1] = x->mvsadcost[1]; |
|
1749 fcenter_mv.as_mv.row = center_mv->as_mv.row >> 3; |
|
1750 fcenter_mv.as_mv.col = center_mv->as_mv.col >> 3; |
|
1751 |
|
1752 bestsad = fn_ptr->sdf(what, what_stride, best_address, |
|
1753 in_what_stride, UINT_MAX) |
|
1754 + mvsad_err_cost(ref_mv, &fcenter_mv, mvsadcost, error_per_bit); |
|
1755 |
|
1756 for (i=0; i<search_range; i++) |
|
1757 { |
|
1758 int best_site = -1; |
|
1759 |
|
1760 for (j = 0 ; j < 4 ; j++) |
|
1761 { |
|
1762 this_row_offset = ref_mv->as_mv.row + neighbors[j].row; |
|
1763 this_col_offset = ref_mv->as_mv.col + neighbors[j].col; |
|
1764 |
|
1765 if ((this_col_offset > x->mv_col_min) && (this_col_offset < x->mv_col_max) && |
|
1766 (this_row_offset > x->mv_row_min) && (this_row_offset < x->mv_row_max)) |
|
1767 { |
|
1768 check_here = (neighbors[j].row)*in_what_stride + neighbors[j].col + best_address; |
|
1769 thissad = fn_ptr->sdf(what, what_stride, check_here , in_what_stride, bestsad); |
|
1770 |
|
1771 if (thissad < bestsad) |
|
1772 { |
|
1773 this_mv.as_mv.row = this_row_offset; |
|
1774 this_mv.as_mv.col = this_col_offset; |
|
1775 thissad += mvsad_err_cost(&this_mv, &fcenter_mv, mvsadcost, error_per_bit); |
|
1776 |
|
1777 if (thissad < bestsad) |
|
1778 { |
|
1779 bestsad = thissad; |
|
1780 best_site = j; |
|
1781 } |
|
1782 } |
|
1783 } |
|
1784 } |
|
1785 |
|
1786 if (best_site == -1) |
|
1787 break; |
|
1788 else |
|
1789 { |
|
1790 ref_mv->as_mv.row += neighbors[best_site].row; |
|
1791 ref_mv->as_mv.col += neighbors[best_site].col; |
|
1792 best_address += (neighbors[best_site].row)*in_what_stride + neighbors[best_site].col; |
|
1793 } |
|
1794 } |
|
1795 |
|
1796 this_mv.as_mv.row = ref_mv->as_mv.row << 3; |
|
1797 this_mv.as_mv.col = ref_mv->as_mv.col << 3; |
|
1798 |
|
1799 return fn_ptr->vf(what, what_stride, best_address, in_what_stride, &thissad) |
|
1800 + mv_err_cost(&this_mv, center_mv, mvcost, x->errorperbit); |
|
1801 } |
|
1802 |
|
1803 int vp8_refining_search_sadx4(MACROBLOCK *x, BLOCK *b, BLOCKD *d, |
|
1804 int_mv *ref_mv, int error_per_bit, |
|
1805 int search_range, vp8_variance_fn_ptr_t *fn_ptr, |
|
1806 int *mvcost[2], int_mv *center_mv) |
|
1807 { |
|
1808 MV neighbors[4] = {{-1, 0}, {0, -1}, {0, 1}, {1, 0}}; |
|
1809 int i, j; |
|
1810 short this_row_offset, this_col_offset; |
|
1811 |
|
1812 int what_stride = b->src_stride; |
|
1813 int pre_stride = x->e_mbd.pre.y_stride; |
|
1814 unsigned char *base_pre = x->e_mbd.pre.y_buffer; |
|
1815 int in_what_stride = pre_stride; |
|
1816 unsigned char *what = (*(b->base_src) + b->src); |
|
1817 unsigned char *best_address = (unsigned char *)(base_pre + d->offset + |
|
1818 (ref_mv->as_mv.row * pre_stride) + ref_mv->as_mv.col); |
|
1819 unsigned char *check_here; |
|
1820 int_mv this_mv; |
|
1821 unsigned int bestsad; |
|
1822 unsigned int thissad; |
|
1823 |
|
1824 int *mvsadcost[2]; |
|
1825 int_mv fcenter_mv; |
|
1826 |
|
1827 mvsadcost[0] = x->mvsadcost[0]; |
|
1828 mvsadcost[1] = x->mvsadcost[1]; |
|
1829 fcenter_mv.as_mv.row = center_mv->as_mv.row >> 3; |
|
1830 fcenter_mv.as_mv.col = center_mv->as_mv.col >> 3; |
|
1831 |
|
1832 bestsad = fn_ptr->sdf(what, what_stride, best_address, |
|
1833 in_what_stride, UINT_MAX) |
|
1834 + mvsad_err_cost(ref_mv, &fcenter_mv, mvsadcost, error_per_bit); |
|
1835 |
|
1836 for (i=0; i<search_range; i++) |
|
1837 { |
|
1838 int best_site = -1; |
|
1839 int all_in = 1; |
|
1840 |
|
1841 all_in &= ((ref_mv->as_mv.row - 1) > x->mv_row_min); |
|
1842 all_in &= ((ref_mv->as_mv.row + 1) < x->mv_row_max); |
|
1843 all_in &= ((ref_mv->as_mv.col - 1) > x->mv_col_min); |
|
1844 all_in &= ((ref_mv->as_mv.col + 1) < x->mv_col_max); |
|
1845 |
|
1846 if(all_in) |
|
1847 { |
|
1848 unsigned int sad_array[4]; |
|
1849 const unsigned char *block_offset[4]; |
|
1850 block_offset[0] = best_address - in_what_stride; |
|
1851 block_offset[1] = best_address - 1; |
|
1852 block_offset[2] = best_address + 1; |
|
1853 block_offset[3] = best_address + in_what_stride; |
|
1854 |
|
1855 fn_ptr->sdx4df(what, what_stride, block_offset, in_what_stride, sad_array); |
|
1856 |
|
1857 for (j = 0; j < 4; j++) |
|
1858 { |
|
1859 if (sad_array[j] < bestsad) |
|
1860 { |
|
1861 this_mv.as_mv.row = ref_mv->as_mv.row + neighbors[j].row; |
|
1862 this_mv.as_mv.col = ref_mv->as_mv.col + neighbors[j].col; |
|
1863 sad_array[j] += mvsad_err_cost(&this_mv, &fcenter_mv, mvsadcost, error_per_bit); |
|
1864 |
|
1865 if (sad_array[j] < bestsad) |
|
1866 { |
|
1867 bestsad = sad_array[j]; |
|
1868 best_site = j; |
|
1869 } |
|
1870 } |
|
1871 } |
|
1872 } |
|
1873 else |
|
1874 { |
|
1875 for (j = 0 ; j < 4 ; j++) |
|
1876 { |
|
1877 this_row_offset = ref_mv->as_mv.row + neighbors[j].row; |
|
1878 this_col_offset = ref_mv->as_mv.col + neighbors[j].col; |
|
1879 |
|
1880 if ((this_col_offset > x->mv_col_min) && (this_col_offset < x->mv_col_max) && |
|
1881 (this_row_offset > x->mv_row_min) && (this_row_offset < x->mv_row_max)) |
|
1882 { |
|
1883 check_here = (neighbors[j].row)*in_what_stride + neighbors[j].col + best_address; |
|
1884 thissad = fn_ptr->sdf(what, what_stride, check_here , in_what_stride, bestsad); |
|
1885 |
|
1886 if (thissad < bestsad) |
|
1887 { |
|
1888 this_mv.as_mv.row = this_row_offset; |
|
1889 this_mv.as_mv.col = this_col_offset; |
|
1890 thissad += mvsad_err_cost(&this_mv, &fcenter_mv, mvsadcost, error_per_bit); |
|
1891 |
|
1892 if (thissad < bestsad) |
|
1893 { |
|
1894 bestsad = thissad; |
|
1895 best_site = j; |
|
1896 } |
|
1897 } |
|
1898 } |
|
1899 } |
|
1900 } |
|
1901 |
|
1902 if (best_site == -1) |
|
1903 break; |
|
1904 else |
|
1905 { |
|
1906 ref_mv->as_mv.row += neighbors[best_site].row; |
|
1907 ref_mv->as_mv.col += neighbors[best_site].col; |
|
1908 best_address += (neighbors[best_site].row)*in_what_stride + neighbors[best_site].col; |
|
1909 } |
|
1910 } |
|
1911 |
|
1912 this_mv.as_mv.row = ref_mv->as_mv.row * 8; |
|
1913 this_mv.as_mv.col = ref_mv->as_mv.col * 8; |
|
1914 |
|
1915 return fn_ptr->vf(what, what_stride, best_address, in_what_stride, &thissad) |
|
1916 + mv_err_cost(&this_mv, center_mv, mvcost, x->errorperbit); |
|
1917 } |
|
1918 |
|
1919 #ifdef VP8_ENTROPY_STATS |
|
1920 void print_mode_context(void) |
|
1921 { |
|
1922 FILE *f = fopen("modecont.c", "w"); |
|
1923 int i, j; |
|
1924 |
|
1925 fprintf(f, "#include \"entropy.h\"\n"); |
|
1926 fprintf(f, "const int vp8_mode_contexts[6][4] =\n"); |
|
1927 fprintf(f, "{\n"); |
|
1928 |
|
1929 for (j = 0; j < 6; j++) |
|
1930 { |
|
1931 fprintf(f, " { /* %d */\n", j); |
|
1932 fprintf(f, " "); |
|
1933 |
|
1934 for (i = 0; i < 4; i++) |
|
1935 { |
|
1936 int overal_prob; |
|
1937 int this_prob; |
|
1938 int count; |
|
1939 |
|
1940 /* Overall probs */ |
|
1941 count = mv_mode_cts[i][0] + mv_mode_cts[i][1]; |
|
1942 |
|
1943 if (count) |
|
1944 overal_prob = 256 * mv_mode_cts[i][0] / count; |
|
1945 else |
|
1946 overal_prob = 128; |
|
1947 |
|
1948 if (overal_prob == 0) |
|
1949 overal_prob = 1; |
|
1950 |
|
1951 /* context probs */ |
|
1952 count = mv_ref_ct[j][i][0] + mv_ref_ct[j][i][1]; |
|
1953 |
|
1954 if (count) |
|
1955 this_prob = 256 * mv_ref_ct[j][i][0] / count; |
|
1956 else |
|
1957 this_prob = 128; |
|
1958 |
|
1959 if (this_prob == 0) |
|
1960 this_prob = 1; |
|
1961 |
|
1962 fprintf(f, "%5d, ", this_prob); |
|
1963 } |
|
1964 |
|
1965 fprintf(f, " },\n"); |
|
1966 } |
|
1967 |
|
1968 fprintf(f, "};\n"); |
|
1969 fclose(f); |
|
1970 } |
|
1971 |
|
1972 /* MV ref count VP8_ENTROPY_STATS stats code */ |
|
1973 #ifdef VP8_ENTROPY_STATS |
|
1974 void init_mv_ref_counts() |
|
1975 { |
|
1976 vpx_memset(mv_ref_ct, 0, sizeof(mv_ref_ct)); |
|
1977 vpx_memset(mv_mode_cts, 0, sizeof(mv_mode_cts)); |
|
1978 } |
|
1979 |
|
1980 void accum_mv_refs(MB_PREDICTION_MODE m, const int ct[4]) |
|
1981 { |
|
1982 if (m == ZEROMV) |
|
1983 { |
|
1984 ++mv_ref_ct [ct[0]] [0] [0]; |
|
1985 ++mv_mode_cts[0][0]; |
|
1986 } |
|
1987 else |
|
1988 { |
|
1989 ++mv_ref_ct [ct[0]] [0] [1]; |
|
1990 ++mv_mode_cts[0][1]; |
|
1991 |
|
1992 if (m == NEARESTMV) |
|
1993 { |
|
1994 ++mv_ref_ct [ct[1]] [1] [0]; |
|
1995 ++mv_mode_cts[1][0]; |
|
1996 } |
|
1997 else |
|
1998 { |
|
1999 ++mv_ref_ct [ct[1]] [1] [1]; |
|
2000 ++mv_mode_cts[1][1]; |
|
2001 |
|
2002 if (m == NEARMV) |
|
2003 { |
|
2004 ++mv_ref_ct [ct[2]] [2] [0]; |
|
2005 ++mv_mode_cts[2][0]; |
|
2006 } |
|
2007 else |
|
2008 { |
|
2009 ++mv_ref_ct [ct[2]] [2] [1]; |
|
2010 ++mv_mode_cts[2][1]; |
|
2011 |
|
2012 if (m == NEWMV) |
|
2013 { |
|
2014 ++mv_ref_ct [ct[3]] [3] [0]; |
|
2015 ++mv_mode_cts[3][0]; |
|
2016 } |
|
2017 else |
|
2018 { |
|
2019 ++mv_ref_ct [ct[3]] [3] [1]; |
|
2020 ++mv_mode_cts[3][1]; |
|
2021 } |
|
2022 } |
|
2023 } |
|
2024 } |
|
2025 } |
|
2026 |
|
2027 #endif/* END MV ref count VP8_ENTROPY_STATS stats code */ |
|
2028 |
|
2029 #endif |