ffmpeg / libavcodec / ivi_dsp.c @ 05aec7bb
History  View  Annotate  Download (15 KB)
1 
/*


2 
* DSP functions for Indeo Video Interactive codecs (Indeo4 and Indeo5)

3 
*

4 
* Copyright (c) 2009 Maxim Poliakovski

5 
*

6 
* This file is part of FFmpeg.

7 
*

8 
* FFmpeg is free software; you can redistribute it and/or

9 
* modify it under the terms of the GNU Lesser General Public

10 
* License as published by the Free Software Foundation; either

11 
* version 2.1 of the License, or (at your option) any later version.

12 
*

13 
* FFmpeg is distributed in the hope that it will be useful,

14 
* but WITHOUT ANY WARRANTY; without even the implied warranty of

15 
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU

16 
* Lesser General Public License for more details.

17 
*

18 
* You should have received a copy of the GNU Lesser General Public

19 
* License along with FFmpeg; if not, write to the Free Software

20 
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 021101301 USA

21 
*/

22  
23 
/**

24 
* @file libavcodec/ivi_dsp.c

25 
* DSP functions (inverse transforms, motion compensation, wavelet recompostions)

26 
* for Indeo Video Interactive codecs.

27 
*/

28  
29 
#include "avcodec.h" 
30 
#include "dsputil.h" 
31 
#include "dwt.h" 
32 
#include "ivi_common.h" 
33 
#include "ivi_dsp.h" 
34  
35 
void ff_ivi_recompose53(const IVIPlaneDesc *plane, uint8_t *dst, 
36 
const int dst_pitch, const int num_bands) 
37 
{ 
38 
int x, y, indx;

39 
int32_t p0, p1, p2, p3, tmp0, tmp1, tmp2; 
40 
int32_t b0_1, b0_2, b1_1, b1_2, b1_3, b2_1, b2_2, b2_3, b2_4, b2_5, b2_6; 
41 
int32_t b3_1, b3_2, b3_3, b3_4, b3_5, b3_6, b3_7, b3_8, b3_9; 
42 
int32_t pitch, back_pitch; 
43 
const IDWTELEM *b0_ptr, *b1_ptr, *b2_ptr, *b3_ptr;

44  
45 
/* all bands should have the same pitch */

46 
pitch = plane>bands[0].pitch;

47  
48 
/* pixels at the position "y1" will be set to pixels at the "y" for the 1st iteration */

49 
back_pitch = 0;

50  
51 
/* get pointers to the wavelet bands */

52 
b0_ptr = plane>bands[0].buf;

53 
b1_ptr = plane>bands[1].buf;

54 
b2_ptr = plane>bands[2].buf;

55 
b3_ptr = plane>bands[3].buf;

56  
57 
for (y = 0; y < plane>height; y += 2) { 
58 
/* load storage variables with values */

59 
if (num_bands > 0) { 
60 
b0_1 = b0_ptr[0];

61 
b0_2 = b0_ptr[pitch]; 
62 
} 
63  
64 
if (num_bands > 1) { 
65 
b1_1 = b1_ptr[back_pitch]; 
66 
b1_2 = b1_ptr[0];

67 
b1_3 = b1_1  b1_2*6 + b1_ptr[pitch];

68 
} 
69  
70 
if (num_bands > 2) { 
71 
b2_2 = b2_ptr[0]; // b2[x, y ] 
72 
b2_3 = b2_2; // b2[x+1,y ] = b2[x,y]

73 
b2_5 = b2_ptr[pitch]; // b2[x ,y+1]

74 
b2_6 = b2_5; // b2[x+1,y+1] = b2[x,y+1]

75 
} 
76  
77 
if (num_bands > 3) { 
78 
b3_2 = b3_ptr[back_pitch]; // b3[x ,y1]

79 
b3_3 = b3_2; // b3[x+1,y1] = b3[x ,y1]

80 
b3_5 = b3_ptr[0]; // b3[x ,y ] 
81 
b3_6 = b3_5; // b3[x+1,y ] = b3[x ,y ]

82 
b3_8 = b3_2  b3_5*6 + b3_ptr[pitch];

83 
b3_9 = b3_8; 
84 
} 
85  
86 
for (x = 0, indx = 0; x < plane>width; x+=2, indx++) { 
87 
/* some values calculated in the previous iterations can */

88 
/* be reused in the next ones, so do appropriate copying */

89 
b2_1 = b2_2; // b2[x1,y ] = b2[x, y ]

90 
b2_2 = b2_3; // b2[x ,y ] = b2[x+1,y ]

91 
b2_4 = b2_5; // b2[x1,y+1] = b2[x ,y+1]

92 
b2_5 = b2_6; // b2[x ,y+1] = b2[x+1,y+1]

93 
b3_1 = b3_2; // b3[x1,y1] = b3[x ,y1]

94 
b3_2 = b3_3; // b3[x ,y1] = b3[x+1,y1]

95 
b3_4 = b3_5; // b3[x1,y ] = b3[x ,y ]

96 
b3_5 = b3_6; // b3[x ,y ] = b3[x+1,y ]

97 
b3_7 = b3_8; // vert_HPF(x1)

98 
b3_8 = b3_9; // vert_HPF(x )

99  
100 
p0 = p1 = p2 = p3 = 0;

101  
102 
/* process the LLband by applying LPF both vertically and horizontally */

103 
if (num_bands > 0) { 
104 
tmp0 = b0_1; 
105 
tmp2 = b0_2; 
106 
b0_1 = b0_ptr[indx+1];

107 
b0_2 = b0_ptr[pitch+indx+1];

108 
tmp1 = tmp0 + b0_1; 
109  
110 
p0 = tmp0 << 4;

111 
p1 = tmp1 << 3;

112 
p2 = (tmp0 + tmp2) << 3;

113 
p3 = (tmp1 + tmp2 + b0_2) << 2;

114 
} 
115  
116 
/* process the HLband by applying HPF vertically and LPF horizontally */

117 
if (num_bands > 1) { 
118 
tmp0 = b1_2; 
119 
tmp1 = b1_1; 
120 
b1_2 = b1_ptr[indx+1];

121 
b1_1 = b1_ptr[back_pitch+indx+1];

122  
123 
tmp2 = tmp1  tmp0*6 + b1_3;

124 
b1_3 = b1_1  b1_2*6 + b1_ptr[pitch+indx+1]; 
125  
126 
p0 += (tmp0 + tmp1) << 3;

127 
p1 += (tmp0 + tmp1 + b1_1 + b1_2) << 2;

128 
p2 += tmp2 << 2;

129 
p3 += (tmp2 + b1_3) << 1;

130 
} 
131  
132 
/* process the LHband by applying LPF vertically and HPF horizontally */

133 
if (num_bands > 2) { 
134 
b2_3 = b2_ptr[indx+1];

135 
b2_6 = b2_ptr[pitch+indx+1];

136  
137 
tmp0 = b2_1 + b2_2; 
138 
tmp1 = b2_1  b2_2*6 + b2_3;

139  
140 
p0 += tmp0 << 3;

141 
p1 += tmp1 << 2;

142 
p2 += (tmp0 + b2_4 + b2_5) << 2;

143 
p3 += (tmp1 + b2_4  b2_5*6 + b2_6) << 1; 
144 
} 
145  
146 
/* process the HHband by applying HPF both vertically and horizontally */

147 
if (num_bands > 3) { 
148 
b3_6 = b3_ptr[indx+1]; // b3[x+1,y ] 
149 
b3_3 = b3_ptr[back_pitch+indx+1]; // b3[x+1,y1] 
150  
151 
tmp0 = b3_1 + b3_4; 
152 
tmp1 = b3_2 + b3_5; 
153 
tmp2 = b3_3 + b3_6; 
154  
155 
b3_9 = b3_3  b3_6*6 + b3_ptr[pitch+indx+1]; 
156  
157 
p0 += (tmp0 + tmp1) << 2;

158 
p1 += (tmp0  tmp1*6 + tmp2) << 1; 
159 
p2 += (b3_7 + b3_8) << 1;

160 
p3 += b3_7  b3_8*6 + b3_9;

161 
} 
162  
163 
/* output four pixels */

164 
dst[x] = av_clip_uint8((p0 >> 6) + 128); 
165 
dst[x+1] = av_clip_uint8((p1 >> 6) + 128); 
166 
dst[dst_pitch+x] = av_clip_uint8((p2 >> 6) + 128); 
167 
dst[dst_pitch+x+1] = av_clip_uint8((p3 >> 6) + 128); 
168 
}// for x

169  
170 
dst += dst_pitch << 1;

171  
172 
back_pitch = pitch; 
173  
174 
b0_ptr += pitch; 
175 
b1_ptr += pitch; 
176 
b2_ptr += pitch; 
177 
b3_ptr += pitch; 
178 
} 
179 
} 
180  
181 
/** butterfly operation for the inverse slant transform */

182 
#define IVI_SLANT_BFLY(s1, s2, o1, o2, t) \

183 
t = s1  s2;\ 
184 
o1 = s1 + s2;\ 
185 
o2 = t;\ 
186  
187 
/** This is a reflection a,b = 1/2, 5/4 for the inverse slant transform */

188 
#define IVI_IREFLECT(s1, s2, o1, o2, t) \

189 
t = ((s1 + s2*2 + 2) >> 2) + s1;\ 
190 
o2 = ((s1*2  s2 + 2) >> 2)  s2;\ 
191 
o1 = t;\ 
192  
193 
/** This is a reflection a,b = 1/2, 7/8 for the inverse slant transform */

194 
#define IVI_SLANT_PART4(s1, s2, o1, o2, t) \

195 
t = s2 + ((s1*4  s2 + 4) >> 3);\ 
196 
o2 = s1 + ((s1  s2*4 + 4) >> 3);\ 
197 
o1 = t;\ 
198  
199 
/** inverse slant8 transform */

200 
#define IVI_INV_SLANT8(s1, s4, s8, s5, s2, s6, s3, s7,\

201 
d1, d2, d3, d4, d5, d6, d7, d8,\ 
202 
t0, t1, t2, t3, t4, t5, t6, t7, t8) {\ 
203 
IVI_SLANT_PART4(s4, s5, t4, t5, t0);\ 
204 
\ 
205 
IVI_SLANT_BFLY(s1, t5, t1, t5, t0); IVI_SLANT_BFLY(s2, s6, t2, t6, t0);\ 
206 
IVI_SLANT_BFLY(s7, s3, t7, t3, t0); IVI_SLANT_BFLY(t4, s8, t4, t8, t0);\ 
207 
\ 
208 
IVI_SLANT_BFLY(t1, t2, t1, t2, t0); IVI_IREFLECT (t4, t3, t4, t3, t0);\ 
209 
IVI_SLANT_BFLY(t5, t6, t5, t6, t0); IVI_IREFLECT (t8, t7, t8, t7, t0);\ 
210 
IVI_SLANT_BFLY(t1, t4, t1, t4, t0); IVI_SLANT_BFLY(t2, t3, t2, t3, t0);\ 
211 
IVI_SLANT_BFLY(t5, t8, t5, t8, t0); IVI_SLANT_BFLY(t6, t7, t6, t7, t0);\ 
212 
d1 = COMPENSATE(t1);\ 
213 
d2 = COMPENSATE(t2);\ 
214 
d3 = COMPENSATE(t3);\ 
215 
d4 = COMPENSATE(t4);\ 
216 
d5 = COMPENSATE(t5);\ 
217 
d6 = COMPENSATE(t6);\ 
218 
d7 = COMPENSATE(t7);\ 
219 
d8 = COMPENSATE(t8);} 
220  
221 
/** inverse slant4 transform */

222 
#define IVI_INV_SLANT4(s1, s4, s2, s3, d1, d2, d3, d4, t0, t1, t2, t3, t4) {\

223 
IVI_SLANT_BFLY(s1, s2, t1, t2, t0); IVI_IREFLECT (s4, s3, t4, t3, t0);\ 
224 
\ 
225 
IVI_SLANT_BFLY(t1, t4, t1, t4, t0); IVI_SLANT_BFLY(t2, t3, t2, t3, t0);\ 
226 
d1 = COMPENSATE(t1);\ 
227 
d2 = COMPENSATE(t2);\ 
228 
d3 = COMPENSATE(t3);\ 
229 
d4 = COMPENSATE(t4);} 
230  
231 
void ff_ivi_inverse_slant_8x8(const int32_t *in, int16_t *out, uint32_t pitch, const uint8_t *flags) 
232 
{ 
233 
int i;

234 
const int32_t *src;

235 
int32_t *dst; 
236 
int tmp[64]; 
237 
int t0, t1, t2, t3, t4, t5, t6, t7, t8;

238  
239 
#define COMPENSATE(x) (x)

240 
src = in; 
241 
dst = tmp; 
242 
for (i = 0; i < 8; i++) { 
243 
if (flags[i]) {

244 
IVI_INV_SLANT8(src[0], src[8], src[16], src[24], src[32], src[40], src[48], src[56], 
245 
dst[0], dst[8], dst[16], dst[24], dst[32], dst[40], dst[48], dst[56], 
246 
t0, t1, t2, t3, t4, t5, t6, t7, t8); 
247 
} else

248 
dst[0] = dst[8] = dst[16] = dst[24] = dst[32] = dst[40] = dst[48] = dst[56] = 0; 
249  
250 
src++; 
251 
dst++; 
252 
} 
253 
#undef COMPENSATE

254  
255 
#define COMPENSATE(x) ((x + 1)>>1) 
256 
src = tmp; 
257 
for (i = 0; i < 8; i++) { 
258 
if (!src[0] && !src[1] && !src[2] && !src[3] && !src[4] && !src[5] && !src[6] && !src[7]) { 
259 
memset(out, 0, 8*sizeof(out[0])); 
260 
} else {

261 
IVI_INV_SLANT8(src[0], src[1], src[2], src[3], src[4], src[5], src[6], src[7], 
262 
out[0], out[1], out[2], out[3], out[4], out[5], out[6], out[7], 
263 
t0, t1, t2, t3, t4, t5, t6, t7, t8); 
264 
} 
265 
src += 8;

266 
out += pitch; 
267 
} 
268 
#undef COMPENSATE

269 
} 
270  
271 
void ff_ivi_inverse_slant_4x4(const int32_t *in, int16_t *out, uint32_t pitch, const uint8_t *flags) 
272 
{ 
273 
int i;

274 
const int32_t *src;

275 
int32_t *dst; 
276 
int tmp[16]; 
277 
int t0, t1, t2, t3, t4;

278  
279 
#define COMPENSATE(x) (x)

280 
src = in; 
281 
dst = tmp; 
282 
for (i = 0; i < 4; i++) { 
283 
if (flags[i]) {

284 
IVI_INV_SLANT4(src[0], src[4], src[8], src[12], 
285 
dst[0], dst[4], dst[8], dst[12], 
286 
t0, t1, t2, t3, t4); 
287 
} else

288 
dst[0] = dst[4] = dst[8] = dst[12] = 0; 
289  
290 
src++; 
291 
dst++; 
292 
} 
293 
#undef COMPENSATE

294  
295 
#define COMPENSATE(x) ((x + 1)>>1) 
296 
src = tmp; 
297 
for (i = 0; i < 4; i++) { 
298 
if (!src[0] && !src[1] && !src[2] && !src[3]) { 
299 
out[0] = out[1] = out[2] = out[3] = 0; 
300 
} else {

301 
IVI_INV_SLANT4(src[0], src[1], src[2], src[3], 
302 
out[0], out[1], out[2], out[3], 
303 
t0, t1, t2, t3, t4); 
304 
} 
305 
src += 4;

306 
out += pitch; 
307 
} 
308 
#undef COMPENSATE

309 
} 
310  
311 
void ff_ivi_dc_slant_2d(const int32_t *in, int16_t *out, uint32_t pitch, int blk_size) 
312 
{ 
313 
int x, y;

314 
int16_t dc_coeff; 
315  
316 
dc_coeff = (*in + 1) >> 1; 
317  
318 
for (y = 0; y < blk_size; out += pitch, y++) { 
319 
for (x = 0; x < blk_size; x++) 
320 
out[x] = dc_coeff; 
321 
} 
322 
} 
323  
324 
void ff_ivi_row_slant8(const int32_t *in, int16_t *out, uint32_t pitch, const uint8_t *flags) 
325 
{ 
326 
int i;

327 
int t0, t1, t2, t3, t4, t5, t6, t7, t8;

328  
329 
#define COMPENSATE(x) ((x + 1)>>1) 
330 
for (i = 0; i < 8; i++) { 
331 
if (!in[0] && !in[1] && !in[2] && !in[3] && !in[4] && !in[5] && !in[6] && !in[7]) { 
332 
memset(out, 0, 8*sizeof(out[0])); 
333 
} else {

334 
IVI_INV_SLANT8( in[0], in[1], in[2], in[3], in[4], in[5], in[6], in[7], 
335 
out[0], out[1], out[2], out[3], out[4], out[5], out[6], out[7], 
336 
t0, t1, t2, t3, t4, t5, t6, t7, t8); 
337 
} 
338 
in += 8;

339 
out += pitch; 
340 
} 
341 
#undef COMPENSATE

342 
} 
343  
344 
void ff_ivi_dc_row_slant(const int32_t *in, int16_t *out, uint32_t pitch, int blk_size) 
345 
{ 
346 
int x, y;

347 
int16_t dc_coeff; 
348  
349 
dc_coeff = (*in + 1) >> 1; 
350  
351 
for (x = 0; x < blk_size; x++) 
352 
out[x] = dc_coeff; 
353  
354 
out += pitch; 
355  
356 
for (y = 1; y < blk_size; out += pitch, y++) { 
357 
for (x = 0; x < blk_size; x++) 
358 
out[x] = 0;

359 
} 
360 
} 
361  
362 
void ff_ivi_col_slant8(const int32_t *in, int16_t *out, uint32_t pitch, const uint8_t *flags) 
363 
{ 
364 
int i, row2, row4, row8;

365 
int t0, t1, t2, t3, t4, t5, t6, t7, t8;

366  
367 
row2 = pitch << 1;

368 
row4 = pitch << 2;

369 
row8 = pitch << 3;

370  
371 
#define COMPENSATE(x) ((x + 1)>>1) 
372 
for (i = 0; i < 8; i++) { 
373 
if (flags[i]) {

374 
IVI_INV_SLANT8(in[0], in[8], in[16], in[24], in[32], in[40], in[48], in[56], 
375 
out[0], out[pitch], out[row2], out[row2 + pitch], out[row4],

376 
out[row4 + pitch], out[row4 + row2], out[row8  pitch], 
377 
t0, t1, t2, t3, t4, t5, t6, t7, t8); 
378 
} else {

379 
out[0] = out[pitch] = out[row2] = out[row2 + pitch] = out[row4] =

380 
out[row4 + pitch] = out[row4 + row2] = out[row8  pitch] = 0;

381 
} 
382  
383 
in++; 
384 
out++; 
385 
} 
386 
#undef COMPENSATE

387 
} 
388  
389 
void ff_ivi_dc_col_slant(const int32_t *in, int16_t *out, uint32_t pitch, int blk_size) 
390 
{ 
391 
int x, y;

392 
int16_t dc_coeff; 
393  
394 
dc_coeff = (*in + 1) >> 1; 
395  
396 
for (y = 0; y < blk_size; out += pitch, y++) { 
397 
out[0] = dc_coeff;

398 
for (x = 1; x < blk_size; x++) 
399 
out[x] = 0;

400 
} 
401 
} 
402  
403 
void ff_ivi_put_pixels_8x8(const int32_t *in, int16_t *out, uint32_t pitch, 
404 
const uint8_t *flags)

405 
{ 
406 
int x, y;

407  
408 
for (y = 0; y < 8; out += pitch, in += 8, y++) 
409 
for (x = 0; x < 8; x++) 
410 
out[x] = in[x]; 
411 
} 
412  
413 
void ff_ivi_put_dc_pixel_8x8(const int32_t *in, int16_t *out, uint32_t pitch, 
414 
int blk_size)

415 
{ 
416 
int y;

417  
418 
out[0] = in[0]; 
419 
memset(out + 1, 0, 7*sizeof(out[0])); 
420 
out += pitch; 
421  
422 
for (y = 1; y < 8; out += pitch, y++) 
423 
memset(out, 0, 8*sizeof(out[0])); 
424 
} 
425  
426 
#define IVI_MC_TEMPLATE(size, suffix, OP) \

427 
void ff_ivi_mc_ ## size ##x## size ## suffix (int16_t *buf, const int16_t *ref_buf, \ 
428 
uint32_t pitch, int mc_type) \

429 
{ \ 
430 
int i, j; \

431 
const int16_t *wptr; \

432 
\ 
433 
switch (mc_type) { \

434 
case 0: /* fullpel (no interpolation) */ \ 
435 
for (i = 0; i < size; i++, buf += pitch, ref_buf += pitch) { \ 
436 
for (j = 0; j < size; j++) {\ 
437 
OP(buf[j], ref_buf[j]); \ 
438 
} \ 
439 
} \ 
440 
break; \

441 
case 1: /* horizontal halfpel interpolation */ \ 
442 
for (i = 0; i < size; i++, buf += pitch, ref_buf += pitch) \ 
443 
for (j = 0; j < size; j++) \ 
444 
OP(buf[j], (ref_buf[j] + ref_buf[j+1]) >> 1); \ 
445 
break; \

446 
case 2: /* vertical halfpel interpolation */ \ 
447 
wptr = ref_buf + pitch; \ 
448 
for (i = 0; i < size; i++, buf += pitch, wptr += pitch, ref_buf += pitch) \ 
449 
for (j = 0; j < size; j++) \ 
450 
OP(buf[j], (ref_buf[j] + wptr[j]) >> 1); \

451 
break; \

452 
case 3: /* vertical and horizontal halfpel interpolation */ \ 
453 
wptr = ref_buf + pitch; \ 
454 
for (i = 0; i < size; i++, buf += pitch, wptr += pitch, ref_buf += pitch) \ 
455 
for (j = 0; j < size; j++) \ 
456 
OP(buf[j], (ref_buf[j] + ref_buf[j+1] + wptr[j] + wptr[j+1]) >> 2); \ 
457 
break; \

458 
} \ 
459 
} \ 
460  
461 
#define OP_PUT(a, b) (a) = (b)

462 
#define OP_ADD(a, b) (a) += (b)

463  
464 
IVI_MC_TEMPLATE(8, _no_delta, OP_PUT);

465 
IVI_MC_TEMPLATE(8, _delta, OP_ADD);

466 
IVI_MC_TEMPLATE(4, _no_delta, OP_PUT);

467 
IVI_MC_TEMPLATE(4, _delta, OP_ADD);
