VirtualBox

source: vbox/trunk/src/libs/ffmpeg-20060710/libavcodec/sh4/dsputil_align.c@ 10184

最後變更 在這個檔案從10184是 5776,由 vboxsync 提交於 17 年 前

ffmpeg: exported to OSE

檔案大小: 14.2 KB
 
1/*
2 * aligned/packed access motion
3 *
4 * Copyright (c) 2001-2003 BERO <[email protected]>
5 *
6 * This library is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU Lesser General Public
8 * License as published by the Free Software Foundation; either
9 * version 2 of the License, or (at your option) any later version.
10 *
11 * This library is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 * Lesser General Public License for more details.
15 *
16 * You should have received a copy of the GNU Lesser General Public
17 * License along with this library; if not, write to the Free Software
18 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
19 */
20
21
22#include "../avcodec.h"
23#include "../dsputil.h"
24
25
26#define LP(p) *(uint32_t*)(p)
27
28
29#define UNPACK(ph,pl,tt0,tt1) do { \
30 uint32_t t0,t1; t0=tt0;t1=tt1; \
31 ph = ( (t0 & ~BYTE_VEC32(0x03))>>2) + ( (t1 & ~BYTE_VEC32(0x03))>>2); \
32 pl = (t0 & BYTE_VEC32(0x03)) + (t1 & BYTE_VEC32(0x03)); } while(0)
33
34#define rnd_PACK(ph,pl,nph,npl) ph + nph + (((pl + npl + BYTE_VEC32(0x02))>>2) & BYTE_VEC32(0x03))
35#define no_rnd_PACK(ph,pl,nph,npl) ph + nph + (((pl + npl + BYTE_VEC32(0x01))>>2) & BYTE_VEC32(0x03))
36
37/* little endian */
38#define MERGE1(a,b,ofs) (ofs==0)?a:( ((a)>>(8*ofs))|((b)<<(32-8*ofs)) )
39#define MERGE2(a,b,ofs) (ofs==3)?b:( ((a)>>(8*(ofs+1)))|((b)<<(32-8*(ofs+1))) )
40/* big
41#define MERGE1(a,b,ofs) (ofs==0)?a:( ((a)<<(8*ofs))|((b)>>(32-8*ofs)) )
42#define MERGE2(a,b,ofs) (ofs==3)?b:( ((a)<<(8+8*ofs))|((b)>>(32-8-8*ofs)) )
43*/
44
45
46#define put(d,s) d = s
47#define avg(d,s) d = rnd_avg32(s,d)
48
49#define OP_C4(ofs) \
50 ref-=ofs; \
51 do { \
52 OP(LP(dest),MERGE1(LP(ref),LP(ref+4),ofs)); \
53 ref+=stride; \
54 dest+=stride; \
55 } while(--height)
56
57#define OP_C40() \
58 do { \
59 OP(LP(dest),LP(ref)); \
60 ref+=stride; \
61 dest+=stride; \
62 } while(--height)
63
64
65#define OP put
66
67static void put_pixels4_c(uint8_t *dest,const uint8_t *ref, const int stride,int height)
68{
69 switch((int)ref&3){
70 case 0: OP_C40(); return;
71 case 1: OP_C4(1); return;
72 case 2: OP_C4(2); return;
73 case 3: OP_C4(3); return;
74 }
75}
76
77#undef OP
78#define OP avg
79
80static void avg_pixels4_c(uint8_t *dest,const uint8_t *ref, const int stride,int height)
81{
82 switch((int)ref&3){
83 case 0: OP_C40(); return;
84 case 1: OP_C4(1); return;
85 case 2: OP_C4(2); return;
86 case 3: OP_C4(3); return;
87 }
88}
89
90#undef OP
91
92#define OP_C(ofs,sz,avg2) \
93{ \
94 ref-=ofs; \
95 do { \
96 uint32_t t0,t1; \
97 t0 = LP(ref+0); \
98 t1 = LP(ref+4); \
99 OP(LP(dest+0), MERGE1(t0,t1,ofs)); \
100 t0 = LP(ref+8); \
101 OP(LP(dest+4), MERGE1(t1,t0,ofs)); \
102if (sz==16) { \
103 t1 = LP(ref+12); \
104 OP(LP(dest+8), MERGE1(t0,t1,ofs)); \
105 t0 = LP(ref+16); \
106 OP(LP(dest+12), MERGE1(t1,t0,ofs)); \
107} \
108 ref+=stride; \
109 dest+= stride; \
110 } while(--height); \
111}
112
113/* aligned */
114#define OP_C0(sz,avg2) \
115{ \
116 do { \
117 OP(LP(dest+0), LP(ref+0)); \
118 OP(LP(dest+4), LP(ref+4)); \
119if (sz==16) { \
120 OP(LP(dest+8), LP(ref+8)); \
121 OP(LP(dest+12), LP(ref+12)); \
122} \
123 ref+=stride; \
124 dest+= stride; \
125 } while(--height); \
126}
127
128#define OP_X(ofs,sz,avg2) \
129{ \
130 ref-=ofs; \
131 do { \
132 uint32_t t0,t1; \
133 t0 = LP(ref+0); \
134 t1 = LP(ref+4); \
135 OP(LP(dest+0), avg2(MERGE1(t0,t1,ofs),MERGE2(t0,t1,ofs))); \
136 t0 = LP(ref+8); \
137 OP(LP(dest+4), avg2(MERGE1(t1,t0,ofs),MERGE2(t1,t0,ofs))); \
138if (sz==16) { \
139 t1 = LP(ref+12); \
140 OP(LP(dest+8), avg2(MERGE1(t0,t1,ofs),MERGE2(t0,t1,ofs))); \
141 t0 = LP(ref+16); \
142 OP(LP(dest+12), avg2(MERGE1(t1,t0,ofs),MERGE2(t1,t0,ofs))); \
143} \
144 ref+=stride; \
145 dest+= stride; \
146 } while(--height); \
147}
148
149/* aligned */
150#define OP_Y0(sz,avg2) \
151{ \
152 uint32_t t0,t1,t2,t3,t; \
153\
154 t0 = LP(ref+0); \
155 t1 = LP(ref+4); \
156if (sz==16) { \
157 t2 = LP(ref+8); \
158 t3 = LP(ref+12); \
159} \
160 do { \
161 ref += stride; \
162\
163 t = LP(ref+0); \
164 OP(LP(dest+0), avg2(t0,t)); t0 = t; \
165 t = LP(ref+4); \
166 OP(LP(dest+4), avg2(t1,t)); t1 = t; \
167if (sz==16) { \
168 t = LP(ref+8); \
169 OP(LP(dest+8), avg2(t2,t)); t2 = t; \
170 t = LP(ref+12); \
171 OP(LP(dest+12), avg2(t3,t)); t3 = t; \
172} \
173 dest+= stride; \
174 } while(--height); \
175}
176
177#define OP_Y(ofs,sz,avg2) \
178{ \
179 uint32_t t0,t1,t2,t3,t,w0,w1; \
180\
181 ref-=ofs; \
182 w0 = LP(ref+0); \
183 w1 = LP(ref+4); \
184 t0 = MERGE1(w0,w1,ofs); \
185 w0 = LP(ref+8); \
186 t1 = MERGE1(w1,w0,ofs); \
187if (sz==16) { \
188 w1 = LP(ref+12); \
189 t2 = MERGE1(w0,w1,ofs); \
190 w0 = LP(ref+16); \
191 t3 = MERGE1(w1,w0,ofs); \
192} \
193 do { \
194 ref += stride; \
195\
196 w0 = LP(ref+0); \
197 w1 = LP(ref+4); \
198 t = MERGE1(w0,w1,ofs); \
199 OP(LP(dest+0), avg2(t0,t)); t0 = t; \
200 w0 = LP(ref+8); \
201 t = MERGE1(w1,w0,ofs); \
202 OP(LP(dest+4), avg2(t1,t)); t1 = t; \
203if (sz==16) { \
204 w1 = LP(ref+12); \
205 t = MERGE1(w0,w1,ofs); \
206 OP(LP(dest+8), avg2(t2,t)); t2 = t; \
207 w0 = LP(ref+16); \
208 t = MERGE1(w1,w0,ofs); \
209 OP(LP(dest+12), avg2(t3,t)); t3 = t; \
210} \
211 dest+=stride; \
212 } while(--height); \
213}
214
215#define OP_X0(sz,avg2) OP_X(0,sz,avg2)
216#define OP_XY0(sz,PACK) OP_XY(0,sz,PACK)
217#define OP_XY(ofs,sz,PACK) \
218{ \
219 uint32_t t2,t3,w0,w1; \
220 uint32_t a0,a1,a2,a3,a4,a5,a6,a7; \
221\
222 ref -= ofs; \
223 w0 = LP(ref+0); \
224 w1 = LP(ref+4); \
225 UNPACK(a0,a1,MERGE1(w0,w1,ofs),MERGE2(w0,w1,ofs)); \
226 w0 = LP(ref+8); \
227 UNPACK(a2,a3,MERGE1(w1,w0,ofs),MERGE2(w1,w0,ofs)); \
228if (sz==16) { \
229 w1 = LP(ref+12); \
230 UNPACK(a4,a5,MERGE1(w0,w1,ofs),MERGE2(w0,w1,ofs)); \
231 w0 = LP(ref+16); \
232 UNPACK(a6,a7,MERGE1(w1,w0,ofs),MERGE2(w1,w0,ofs)); \
233} \
234 do { \
235 ref+=stride; \
236 w0 = LP(ref+0); \
237 w1 = LP(ref+4); \
238 UNPACK(t2,t3,MERGE1(w0,w1,ofs),MERGE2(w0,w1,ofs)); \
239 OP(LP(dest+0),PACK(a0,a1,t2,t3)); \
240 a0 = t2; a1 = t3; \
241 w0 = LP(ref+8); \
242 UNPACK(t2,t3,MERGE1(w1,w0,ofs),MERGE2(w1,w0,ofs)); \
243 OP(LP(dest+4),PACK(a2,a3,t2,t3)); \
244 a2 = t2; a3 = t3; \
245if (sz==16) { \
246 w1 = LP(ref+12); \
247 UNPACK(t2,t3,MERGE1(w0,w1,ofs),MERGE2(w0,w1,ofs)); \
248 OP(LP(dest+8),PACK(a4,a5,t2,t3)); \
249 a4 = t2; a5 = t3; \
250 w0 = LP(ref+16); \
251 UNPACK(t2,t3,MERGE1(w1,w0,ofs),MERGE2(w1,w0,ofs)); \
252 OP(LP(dest+12),PACK(a6,a7,t2,t3)); \
253 a6 = t2; a7 = t3; \
254} \
255 dest+=stride; \
256 } while(--height); \
257}
258
259#define DEFFUNC(op,rnd,xy,sz,OP_N,avgfunc) \
260static void op##_##rnd##_pixels##sz##_##xy (uint8_t * dest, const uint8_t * ref, \
261 const int stride, int height) \
262{ \
263 switch((int)ref&3) { \
264 case 0:OP_N##0(sz,rnd##_##avgfunc); return; \
265 case 1:OP_N(1,sz,rnd##_##avgfunc); return; \
266 case 2:OP_N(2,sz,rnd##_##avgfunc); return; \
267 case 3:OP_N(3,sz,rnd##_##avgfunc); return; \
268 } \
269}
270
271#define OP put
272
273DEFFUNC(put, rnd,o,8,OP_C,avg2)
274DEFFUNC(put, rnd,x,8,OP_X,avg2)
275DEFFUNC(put,no_rnd,x,8,OP_X,avg2)
276DEFFUNC(put, rnd,y,8,OP_Y,avg2)
277DEFFUNC(put,no_rnd,y,8,OP_Y,avg2)
278DEFFUNC(put, rnd,xy,8,OP_XY,PACK)
279DEFFUNC(put,no_rnd,xy,8,OP_XY,PACK)
280DEFFUNC(put, rnd,o,16,OP_C,avg2)
281DEFFUNC(put, rnd,x,16,OP_X,avg2)
282DEFFUNC(put,no_rnd,x,16,OP_X,avg2)
283DEFFUNC(put, rnd,y,16,OP_Y,avg2)
284DEFFUNC(put,no_rnd,y,16,OP_Y,avg2)
285DEFFUNC(put, rnd,xy,16,OP_XY,PACK)
286DEFFUNC(put,no_rnd,xy,16,OP_XY,PACK)
287
288#undef OP
289#define OP avg
290
291DEFFUNC(avg, rnd,o,8,OP_C,avg2)
292DEFFUNC(avg, rnd,x,8,OP_X,avg2)
293DEFFUNC(avg,no_rnd,x,8,OP_X,avg2)
294DEFFUNC(avg, rnd,y,8,OP_Y,avg2)
295DEFFUNC(avg,no_rnd,y,8,OP_Y,avg2)
296DEFFUNC(avg, rnd,xy,8,OP_XY,PACK)
297DEFFUNC(avg,no_rnd,xy,8,OP_XY,PACK)
298DEFFUNC(avg, rnd,o,16,OP_C,avg2)
299DEFFUNC(avg, rnd,x,16,OP_X,avg2)
300DEFFUNC(avg,no_rnd,x,16,OP_X,avg2)
301DEFFUNC(avg, rnd,y,16,OP_Y,avg2)
302DEFFUNC(avg,no_rnd,y,16,OP_Y,avg2)
303DEFFUNC(avg, rnd,xy,16,OP_XY,PACK)
304DEFFUNC(avg,no_rnd,xy,16,OP_XY,PACK)
305
306#undef OP
307
308#define put_no_rnd_pixels8_o put_rnd_pixels8_o
309#define put_no_rnd_pixels16_o put_rnd_pixels16_o
310#define avg_no_rnd_pixels8_o avg_rnd_pixels8_o
311#define avg_no_rnd_pixels16_o avg_rnd_pixels16_o
312
313#define put_pixels8_c put_rnd_pixels8_o
314#define put_pixels16_c put_rnd_pixels16_o
315#define avg_pixels8_c avg_rnd_pixels8_o
316#define avg_pixels16_c avg_rnd_pixels16_o
317#define put_no_rnd_pixels8_c put_rnd_pixels8_o
318#define put_no_rnd_pixels16_c put_rnd_pixels16_o
319#define avg_no_rnd_pixels8_c avg_rnd_pixels8_o
320#define avg_no_rnd_pixels16_c avg_rnd_pixels16_o
321
322#define QPEL
323
324#ifdef QPEL
325
326#include "qpel.c"
327
328#endif
329
330void dsputil_init_align(DSPContext* c, AVCodecContext *avctx)
331{
332 c->put_pixels_tab[0][0] = put_rnd_pixels16_o;
333 c->put_pixels_tab[0][1] = put_rnd_pixels16_x;
334 c->put_pixels_tab[0][2] = put_rnd_pixels16_y;
335 c->put_pixels_tab[0][3] = put_rnd_pixels16_xy;
336 c->put_pixels_tab[1][0] = put_rnd_pixels8_o;
337 c->put_pixels_tab[1][1] = put_rnd_pixels8_x;
338 c->put_pixels_tab[1][2] = put_rnd_pixels8_y;
339 c->put_pixels_tab[1][3] = put_rnd_pixels8_xy;
340
341 c->put_no_rnd_pixels_tab[0][0] = put_no_rnd_pixels16_o;
342 c->put_no_rnd_pixels_tab[0][1] = put_no_rnd_pixels16_x;
343 c->put_no_rnd_pixels_tab[0][2] = put_no_rnd_pixels16_y;
344 c->put_no_rnd_pixels_tab[0][3] = put_no_rnd_pixels16_xy;
345 c->put_no_rnd_pixels_tab[1][0] = put_no_rnd_pixels8_o;
346 c->put_no_rnd_pixels_tab[1][1] = put_no_rnd_pixels8_x;
347 c->put_no_rnd_pixels_tab[1][2] = put_no_rnd_pixels8_y;
348 c->put_no_rnd_pixels_tab[1][3] = put_no_rnd_pixels8_xy;
349
350 c->avg_pixels_tab[0][0] = avg_rnd_pixels16_o;
351 c->avg_pixels_tab[0][1] = avg_rnd_pixels16_x;
352 c->avg_pixels_tab[0][2] = avg_rnd_pixels16_y;
353 c->avg_pixels_tab[0][3] = avg_rnd_pixels16_xy;
354 c->avg_pixels_tab[1][0] = avg_rnd_pixels8_o;
355 c->avg_pixels_tab[1][1] = avg_rnd_pixels8_x;
356 c->avg_pixels_tab[1][2] = avg_rnd_pixels8_y;
357 c->avg_pixels_tab[1][3] = avg_rnd_pixels8_xy;
358
359 c->avg_no_rnd_pixels_tab[0][0] = avg_no_rnd_pixels16_o;
360 c->avg_no_rnd_pixels_tab[0][1] = avg_no_rnd_pixels16_x;
361 c->avg_no_rnd_pixels_tab[0][2] = avg_no_rnd_pixels16_y;
362 c->avg_no_rnd_pixels_tab[0][3] = avg_no_rnd_pixels16_xy;
363 c->avg_no_rnd_pixels_tab[1][0] = avg_no_rnd_pixels8_o;
364 c->avg_no_rnd_pixels_tab[1][1] = avg_no_rnd_pixels8_x;
365 c->avg_no_rnd_pixels_tab[1][2] = avg_no_rnd_pixels8_y;
366 c->avg_no_rnd_pixels_tab[1][3] = avg_no_rnd_pixels8_xy;
367
368#ifdef QPEL
369
370#define dspfunc(PFX, IDX, NUM) \
371 c->PFX ## _pixels_tab[IDX][ 0] = PFX ## NUM ## _mc00_c; \
372 c->PFX ## _pixels_tab[IDX][ 1] = PFX ## NUM ## _mc10_c; \
373 c->PFX ## _pixels_tab[IDX][ 2] = PFX ## NUM ## _mc20_c; \
374 c->PFX ## _pixels_tab[IDX][ 3] = PFX ## NUM ## _mc30_c; \
375 c->PFX ## _pixels_tab[IDX][ 4] = PFX ## NUM ## _mc01_c; \
376 c->PFX ## _pixels_tab[IDX][ 5] = PFX ## NUM ## _mc11_c; \
377 c->PFX ## _pixels_tab[IDX][ 6] = PFX ## NUM ## _mc21_c; \
378 c->PFX ## _pixels_tab[IDX][ 7] = PFX ## NUM ## _mc31_c; \
379 c->PFX ## _pixels_tab[IDX][ 8] = PFX ## NUM ## _mc02_c; \
380 c->PFX ## _pixels_tab[IDX][ 9] = PFX ## NUM ## _mc12_c; \
381 c->PFX ## _pixels_tab[IDX][10] = PFX ## NUM ## _mc22_c; \
382 c->PFX ## _pixels_tab[IDX][11] = PFX ## NUM ## _mc32_c; \
383 c->PFX ## _pixels_tab[IDX][12] = PFX ## NUM ## _mc03_c; \
384 c->PFX ## _pixels_tab[IDX][13] = PFX ## NUM ## _mc13_c; \
385 c->PFX ## _pixels_tab[IDX][14] = PFX ## NUM ## _mc23_c; \
386 c->PFX ## _pixels_tab[IDX][15] = PFX ## NUM ## _mc33_c
387
388 dspfunc(put_qpel, 0, 16);
389 dspfunc(put_no_rnd_qpel, 0, 16);
390
391 dspfunc(avg_qpel, 0, 16);
392 /* dspfunc(avg_no_rnd_qpel, 0, 16); */
393
394 dspfunc(put_qpel, 1, 8);
395 dspfunc(put_no_rnd_qpel, 1, 8);
396
397 dspfunc(avg_qpel, 1, 8);
398 /* dspfunc(avg_no_rnd_qpel, 1, 8); */
399
400 dspfunc(put_h264_qpel, 0, 16);
401 dspfunc(put_h264_qpel, 1, 8);
402 dspfunc(put_h264_qpel, 2, 4);
403 dspfunc(avg_h264_qpel, 0, 16);
404 dspfunc(avg_h264_qpel, 1, 8);
405 dspfunc(avg_h264_qpel, 2, 4);
406
407#undef dspfunc
408 c->put_h264_chroma_pixels_tab[0]= put_h264_chroma_mc8_c;
409 c->put_h264_chroma_pixels_tab[1]= put_h264_chroma_mc4_c;
410 c->put_h264_chroma_pixels_tab[2]= put_h264_chroma_mc2_c;
411 c->avg_h264_chroma_pixels_tab[0]= avg_h264_chroma_mc8_c;
412 c->avg_h264_chroma_pixels_tab[1]= avg_h264_chroma_mc4_c;
413 c->avg_h264_chroma_pixels_tab[2]= avg_h264_chroma_mc2_c;
414
415 c->put_mspel_pixels_tab[0]= put_mspel8_mc00_c;
416 c->put_mspel_pixels_tab[1]= put_mspel8_mc10_c;
417 c->put_mspel_pixels_tab[2]= put_mspel8_mc20_c;
418 c->put_mspel_pixels_tab[3]= put_mspel8_mc30_c;
419 c->put_mspel_pixels_tab[4]= put_mspel8_mc02_c;
420 c->put_mspel_pixels_tab[5]= put_mspel8_mc12_c;
421 c->put_mspel_pixels_tab[6]= put_mspel8_mc22_c;
422 c->put_mspel_pixels_tab[7]= put_mspel8_mc32_c;
423
424 c->gmc1 = gmc1_c;
425 c->gmc = gmc_c;
426
427#endif
428}
注意: 瀏覽 TracBrowser 來幫助您使用儲存庫瀏覽器

© 2024 Oracle Support Privacy / Do Not Sell My Info Terms of Use Trademark Policy Automated Access Etiquette