diff --git a/libavcodec/adpcm.c b/libavcodec/adpcm.c index c88db5ca15..1281acc33a 100644 --- a/libavcodec/adpcm.c +++ b/libavcodec/adpcm.c @@ -824,13 +824,13 @@ static int adpcm_decode_frame(AVCodecContext *avctx, coeff2r = ea_adpcm_table[(*src & 0x0F) + 4]; src++; - shift_left = (*src >> 4 ) + 8; - shift_right = (*src & 0x0F) + 8; + shift_left = 20 - (*src >> 4); + shift_right = 20 - (*src & 0x0F); src++; for (count2 = 0; count2 < 28; count2++) { - next_left_sample = (int32_t)((*src & 0xF0) << 24) >> shift_left; - next_right_sample = (int32_t)((*src & 0x0F) << 28) >> shift_right; + next_left_sample = sign_extend(*src >> 4, 4) << shift_left; + next_right_sample = sign_extend(*src, 4) << shift_right; src++; next_left_sample = (next_left_sample + @@ -861,13 +861,13 @@ static int adpcm_decode_frame(AVCodecContext *avctx, for(channel = 0; channel < avctx->channels; channel++) { for (i=0; i<2; i++) coeff[channel][i] = ea_adpcm_table[(*src >> 4) + 4*i]; - shift[channel] = (*src & 0x0F) + 8; + shift[channel] = 20 - (*src & 0x0F); src++; } for (count1 = 0; count1 < nb_samples / 2; count1++) { for(i = 4; i >= 0; i-=4) { /* Pairwise samples LL RR (st) or LL LL (mono) */ for(channel = 0; channel < avctx->channels; channel++) { - int32_t sample = (int32_t)(((*(src+channel) >> i) & 0x0F) << 0x1C) >> shift[channel]; + int32_t sample = sign_extend(src[channel] >> i, 4) << shift[channel]; sample = (sample + c->status[channel].sample1 * coeff[channel][0] + c->status[channel].sample2 * coeff[channel][1] + 0x80) >> 8; @@ -932,14 +932,14 @@ static int adpcm_decode_frame(AVCodecContext *avctx, } else { coeff1 = ea_adpcm_table[ *srcC>>4 ]; coeff2 = ea_adpcm_table[(*srcC>>4) + 4]; - shift = (*srcC++ & 0x0F) + 8; + shift = 20 - (*srcC++ & 0x0F); if (srcC > src_end - 14) break; for (count2=0; count2<28; count2++) { if (count2 & 1) - next_sample = (int32_t)((*srcC++ & 0x0F) << 28) >> shift; + next_sample = sign_extend(*srcC++, 4) << shift; else - next_sample = (int32_t)((*srcC & 0xF0) << 24) >> shift; + next_sample = sign_extend(*srcC >> 4, 4) << shift; next_sample += (current_sample * coeff1) + (previous_sample * coeff2); @@ -976,7 +976,7 @@ static int adpcm_decode_frame(AVCodecContext *avctx, for (n=0; n<4; n++, s+=32*avctx->channels) { for (i=0; i<2; i++) coeff[i][n] = ea_adpcm_table[(src[0]&0x0F)+4*i]; - shift[n] = (src[2]&0x0F) + 8; + shift[n] = 20 - (src[2] & 0x0F); for (s2=s, i=0; i<2; i++, src+=2, s2+=avctx->channels) s2[0] = (src[0]&0xF0) + (src[1]<<8); } @@ -985,7 +985,7 @@ static int adpcm_decode_frame(AVCodecContext *avctx, s = &samples[m*avctx->channels + channel]; for (n=0; n<4; n++, src++, s+=32*avctx->channels) { for (s2=s, i=0; i<8; i+=4, s2+=avctx->channels) { - int level = (int32_t)((*src & (0xF0>>i)) << (24+i)) >> shift[n]; + int level = sign_extend(*src >> (4 - i), 4) << shift[n]; int pred = s2[-1*avctx->channels] * coeff[0][n] + s2[-2*avctx->channels] * coeff[1][n]; s2[0] = av_clip_int16((level + pred + 0x80) >> 8); @@ -1149,18 +1149,18 @@ static int adpcm_decode_frame(AVCodecContext *avctx, /* Read in every sample for this channel. */ for (i = 0; i < nb_samples / 14; i++) { int index = (*src >> 4) & 7; - unsigned int exp = 28 - (*src++ & 15); + unsigned int exp = *src++ & 15; int factor1 = table[ch][index * 2]; int factor2 = table[ch][index * 2 + 1]; /* Decode 14 samples. */ for (n = 0; n < 14; n++) { int32_t sampledat; - if(n&1) sampledat= *src++ <<28; - else sampledat= (*src&0xF0)<<24; + if(n&1) sampledat = sign_extend(*src++, 4); + else sampledat = sign_extend(*src >> 4, 4); sampledat = ((prev[ch][0]*factor1 - + prev[ch][1]*factor2) >> 11) + (sampledat>>exp); + + prev[ch][1]*factor2) >> 11) + (sampledat << exp); *samples = av_clip_int16(sampledat); prev[ch][1] = prev[ch][0]; prev[ch][0] = *samples++; diff --git a/libavcodec/mpeg12.c b/libavcodec/mpeg12.c index 08d46d4ef9..3bc857dc5f 100644 --- a/libavcodec/mpeg12.c +++ b/libavcodec/mpeg12.c @@ -735,9 +735,8 @@ static void exchange_uv(MpegEncContext *s) #define MT_16X8 2 #define MT_DMV 3 -static int mpeg_decode_mb(Mpeg1Context *s1, DCTELEM block[12][64]) +static int mpeg_decode_mb(MpegEncContext *s, DCTELEM block[12][64]) { - MpegEncContext *s = &s1->mpeg_enc_ctx; int i, j, k, cbp, val, mb_type, motion_type; const int mb_block_count = 4 + (1 << s->chroma_format); @@ -1658,10 +1657,9 @@ static int mpeg_field_start(MpegEncContext *s, const uint8_t *buf, int buf_size) * @return DECODE_SLICE_ERROR if the slice is damaged
* DECODE_SLICE_OK if this slice is ok
*/ -static int mpeg_decode_slice(Mpeg1Context *s1, int mb_y, +static int mpeg_decode_slice(MpegEncContext *s, int mb_y, const uint8_t **buf, int buf_size) { - MpegEncContext *s = &s1->mpeg_enc_ctx; AVCodecContext *avctx = s->avctx; const int lowres = s->avctx->lowres; const int field_pic = s->picture_structure != PICT_FRAME; @@ -1750,7 +1748,7 @@ static int mpeg_decode_slice(Mpeg1Context *s1, int mb_y, if (CONFIG_MPEG_XVMC_DECODER && s->avctx->xvmc_acceleration > 1) ff_xvmc_init_block(s); // set s->block - if (mpeg_decode_mb(s1, s->block) < 0) + if (mpeg_decode_mb(s, s->block) < 0) return -1; if (s->current_picture.f.motion_val[0] && !s->encoding) { // note motion_val is normally NULL unless we want to extract the MVs @@ -1893,7 +1891,7 @@ static int slice_decode_thread(AVCodecContext *c, void *arg) uint32_t start_code; int ret; - ret = mpeg_decode_slice((Mpeg1Context*)s, mb_y, &buf, s->gb.buffer_end - buf); + ret = mpeg_decode_slice(s, mb_y, &buf, s->gb.buffer_end - buf); emms_c(); //av_log(c, AV_LOG_DEBUG, "ret:%d resync:%d/%d mb:%d/%d ts:%d/%d ec:%d\n", //ret, s->resync_mb_x, s->resync_mb_y, s->mb_x, s->mb_y, s->start_mb_y, s->end_mb_y, s->error_count); @@ -2308,7 +2306,7 @@ static int decode_chunks(AVCodecContext *avctx, } if (CONFIG_VDPAU && uses_vdpau(avctx)) - ff_vdpau_mpeg_picture_complete(s, buf, buf_size, s->slice_count); + ff_vdpau_mpeg_picture_complete(s2, buf, buf_size, s->slice_count); if (slice_end(avctx, picture)) { @@ -2500,7 +2498,7 @@ static int decode_chunks(AVCodecContext *avctx, } buf_ptr += 2; // FIXME add minimum number of bytes per slice } else { - ret = mpeg_decode_slice(s, mb_y, &buf_ptr, input_size); + ret = mpeg_decode_slice(s2, mb_y, &buf_ptr, input_size); emms_c(); if (ret < 0) { diff --git a/libavcodec/vdpau.c b/libavcodec/vdpau.c index 260d05d906..ce5103a6af 100644 --- a/libavcodec/vdpau.c +++ b/libavcodec/vdpau.c @@ -190,10 +190,9 @@ void ff_vdpau_h264_picture_complete(MpegEncContext *s) render->bitstream_buffers_used = 0; } -void ff_vdpau_mpeg_picture_complete(Mpeg1Context *s1, const uint8_t *buf, +void ff_vdpau_mpeg_picture_complete(MpegEncContext *s, const uint8_t *buf, int buf_size, int slice_count) { - MpegEncContext *s = &s1->mpeg_enc_ctx; struct vdpau_render_state *render, *last, *next; int i; diff --git a/libavcodec/vdpau_internal.h b/libavcodec/vdpau_internal.h index 2d8a5d8e6f..0a8d0b6b55 100644 --- a/libavcodec/vdpau_internal.h +++ b/libavcodec/vdpau_internal.h @@ -26,12 +26,11 @@ #include #include "mpegvideo.h" -#include "mpeg12.h" void ff_vdpau_add_data_chunk(MpegEncContext *s, const uint8_t *buf, int buf_size); -void ff_vdpau_mpeg_picture_complete(Mpeg1Context *s1, const uint8_t *buf, +void ff_vdpau_mpeg_picture_complete(MpegEncContext *s, const uint8_t *buf, int buf_size, int slice_count); void ff_vdpau_h264_picture_start(MpegEncContext *s);