Revision be449fca libavcodec/i386/vc1dsp_mmx.c

View differences:

libavcodec/i386/vc1dsp_mmx.c
74 74
                                       const uint8_t *src, x86_reg stride,
75 75
                                       int rnd, int64_t shift)
76 76
{
77
    asm volatile(
77
    __asm__ volatile(
78 78
        "mov       $3, %%"REG_c"           \n\t"
79 79
        LOAD_ROUNDER_MMX("%5")
80 80
        "movq      "MANGLE(ff_pw_9)", %%mm6 \n\t"
......
114 114

  
115 115
    src -= 1;
116 116
    rnd -= (-1+9+9-1)*1024; /* Add -1024 bias */
117
    asm volatile(
117
    __asm__ volatile(
118 118
        LOAD_ROUNDER_MMX("%4")
119 119
        "movq      "MANGLE(ff_pw_128)", %%mm6\n\t"
120 120
        "movq      "MANGLE(ff_pw_9)", %%mm5 \n\t"
......
155 155
                               x86_reg stride, int rnd, x86_reg offset)
156 156
{
157 157
    rnd = 8-rnd;
158
    asm volatile(
158
    __asm__ volatile(
159 159
        "mov       $8, %%"REG_c"           \n\t"
160 160
        LOAD_ROUNDER_MMX("%5")
161 161
        "movq      "MANGLE(ff_pw_9)", %%mm6\n\t"
......
264 264
{                                                                       \
265 265
    int h = 8;                                                          \
266 266
    src -= src_stride;                                                  \
267
    asm volatile(                                                       \
267
    __asm__ volatile(                                                       \
268 268
        LOAD_ROUNDER_MMX("%5")                                          \
269 269
        "movq      "MANGLE(ff_pw_53)", %%mm5\n\t"                       \
270 270
        "movq      "MANGLE(ff_pw_18)", %%mm6\n\t"                       \
......
320 320
    int h = 8;                                                          \
321 321
    src -= 1;                                                           \
322 322
    rnd -= (-4+58+13-3)*256; /* Add -256 bias */                        \
323
    asm volatile(                                                       \
323
    __asm__ volatile(                                                       \
324 324
        LOAD_ROUNDER_MMX("%4")                                          \
325 325
        "movq      "MANGLE(ff_pw_18)", %%mm6   \n\t"                    \
326 326
        "movq      "MANGLE(ff_pw_53)", %%mm5   \n\t"                    \
......
358 358
    int h = 8;                                                          \
359 359
    src -= offset;                                                      \
360 360
    rnd = 32-rnd;                                                       \
361
    asm volatile (                                                      \
361
    __asm__ volatile (                                                      \
362 362
        LOAD_ROUNDER_MMX("%6")                                          \
363 363
        "movq      "MANGLE(ff_pw_53)", %%mm5       \n\t"                \
364 364
        "movq      "MANGLE(ff_pw_18)", %%mm6       \n\t"                \
......
412 412
    static const vc1_mspel_mc_filter_8bits vc1_put_shift_8bits[] =
413 413
         { NULL, vc1_put_shift1_mmx, vc1_put_shift2_mmx, vc1_put_shift3_mmx };
414 414

  
415
    asm volatile(
415
    __asm__ volatile(
416 416
        "pxor %%mm0, %%mm0         \n\t"
417 417
        ::: "memory"
418 418
    );

Also available in: Unified diff