diff options
author | Michael Niedermayer <michaelni@gmx.at> | 2006-01-27 13:19:13 +0000 |
---|---|---|
committer | Michael Niedermayer <michaelni@gmx.at> | 2006-01-27 13:19:13 +0000 |
commit | 07fc2b8278b81635e9e4de1dc18ce646116d8fe5 (patch) | |
tree | 87eb43a96ef0fcf9f8451eeb567ede8c63ccdbcb /libavcodec | |
parent | e435887cf6addd8ade091e0bfb5341567e4ddd94 (diff) | |
download | ffmpeg-07fc2b8278b81635e9e4de1dc18ce646116d8fe5.tar.gz |
fixing bframe strategy 2
bits vs. bytes factor of 8 error
16 byte offset error
some other minor things
Originally committed as revision 4902 to svn://svn.ffmpeg.org/ffmpeg/trunk
Diffstat (limited to 'libavcodec')
-rw-r--r-- | libavcodec/h263.c | 4 | ||||
-rw-r--r-- | libavcodec/mpegvideo.c | 42 | ||||
-rw-r--r-- | libavcodec/mpegvideo.h | 2 |
3 files changed, 33 insertions, 15 deletions
diff --git a/libavcodec/h263.c b/libavcodec/h263.c index 2e41c65e0f..f7369c18d5 100644 --- a/libavcodec/h263.c +++ b/libavcodec/h263.c @@ -997,7 +997,9 @@ void mpeg4_encode_mb(MpegEncContext * s, if(pic==NULL || pic->pict_type!=B_TYPE) break; - b_pic= pic->data[0] + offset + 16; //FIXME +16 + b_pic= pic->data[0] + offset; + if(pic->type != FF_BUFFER_TYPE_SHARED) + b_pic+= INPLACE_OFFSET; diff= s->dsp.sad[0](NULL, p_pic, b_pic, s->linesize, 16); if(diff>s->qscale*70){ //FIXME check that 70 is optimal s->mb_skipped=0; diff --git a/libavcodec/mpegvideo.c b/libavcodec/mpegvideo.c index 6ae89185ad..b0d9c27253 100644 --- a/libavcodec/mpegvideo.c +++ b/libavcodec/mpegvideo.c @@ -2058,7 +2058,6 @@ static int load_input_picture(MpegEncContext *s, AVFrame *pic_arg){ } alloc_picture(s, (Picture*)pic, 1); }else{ - int offset= 16; i= ff_find_unused_picture(s, 0); pic= (AVFrame*)&s->picture[i]; @@ -2066,9 +2065,9 @@ static int load_input_picture(MpegEncContext *s, AVFrame *pic_arg){ alloc_picture(s, (Picture*)pic, 0); - if( pic->data[0] + offset == pic_arg->data[0] - && pic->data[1] + offset == pic_arg->data[1] - && pic->data[2] + offset == pic_arg->data[2]){ + if( pic->data[0] + INPLACE_OFFSET == pic_arg->data[0] + && pic->data[1] + INPLACE_OFFSET == pic_arg->data[1] + && pic->data[2] + INPLACE_OFFSET == pic_arg->data[2]){ // empty }else{ int h_chroma_shift, v_chroma_shift; @@ -2082,7 +2081,7 @@ static int load_input_picture(MpegEncContext *s, AVFrame *pic_arg){ int w= s->width >>h_shift; int h= s->height>>v_shift; uint8_t *src= pic_arg->data[i]; - uint8_t *dst= pic->data[i] + offset; + uint8_t *dst= pic->data[i] + INPLACE_OFFSET; if(src_stride==dst_stride) memcpy(dst, src, src_stride*h); @@ -2147,13 +2146,18 @@ static int estimate_best_b_count(MpegEncContext *s){ AVCodecContext *c= avcodec_alloc_context(); AVFrame input[FF_MAX_B_FRAMES+2]; const int scale= s->avctx->brd_scale; - int i, j, out_size; + int i, j, out_size, p_lambda, b_lambda, lambda2; int outbuf_size= s->width * s->height; //FIXME uint8_t *outbuf= av_malloc(outbuf_size); ImgReSampleContext *resample; int64_t best_rd= INT64_MAX; int best_b_count= -1; - const int lambda2= s->lambda2; + +// emms_c(); + p_lambda= s->last_lambda_for[P_TYPE]; //s->next_picture_ptr->quality; + b_lambda= s->last_lambda_for[B_TYPE]; //p_lambda *ABS(s->avctx->b_quant_factor) + s->avctx->b_quant_offset; + if(!b_lambda) b_lambda= p_lambda; //FIXME we should do this somewhere else + lambda2= (b_lambda*b_lambda + (1<<FF_LAMBDA_SHIFT)/2 ) >> FF_LAMBDA_SHIFT; c->width = s->width >> scale; c->height= s->height>> scale; @@ -2175,6 +2179,16 @@ static int estimate_best_b_count(MpegEncContext *s){ for(i=0; i<s->max_b_frames+2; i++){ int ysize= c->width*c->height; int csize= (c->width/2)*(c->height/2); + Picture pre_input, *pre_input_ptr= i ? s->input_picture[i-1] : s->next_picture_ptr; + + if(pre_input_ptr) + pre_input= *pre_input_ptr; + + if(pre_input.type != FF_BUFFER_TYPE_SHARED && i){ + pre_input.data[0]+=INPLACE_OFFSET; + pre_input.data[1]+=INPLACE_OFFSET; + pre_input.data[2]+=INPLACE_OFFSET; + } avcodec_get_frame_defaults(&input[i]); input[i].data[0]= av_malloc(ysize + 2*csize); @@ -2185,7 +2199,7 @@ static int estimate_best_b_count(MpegEncContext *s){ input[i].linesize[2]= c->width/2; if(!i || s->input_picture[i-1]) - img_resample(resample, &input[i], i ? s->input_picture[i-1] : s->next_picture_ptr); + img_resample(resample, &input[i], &pre_input); } for(j=0; j<s->max_b_frames+1; j++){ @@ -2197,23 +2211,23 @@ static int estimate_best_b_count(MpegEncContext *s){ c->error[0]= c->error[1]= c->error[2]= 0; input[0].pict_type= I_TYPE; - input[0].quality= 2 * FF_QP2LAMBDA; + input[0].quality= 1 * FF_QP2LAMBDA; out_size = avcodec_encode_video(c, outbuf, outbuf_size, &input[0]); - rd += (out_size * lambda2) >> FF_LAMBDA_SHIFT; +// rd += (out_size * lambda2) >> FF_LAMBDA_SHIFT; for(i=0; i<s->max_b_frames+1; i++){ int is_p= i % (j+1) == j || i==s->max_b_frames; input[i+1].pict_type= is_p ? P_TYPE : B_TYPE; - input[i+1].quality= s->last_lambda_for[input[i+1].pict_type]; + input[i+1].quality= is_p ? p_lambda : b_lambda; out_size = avcodec_encode_video(c, outbuf, outbuf_size, &input[i+1]); - rd += (out_size * lambda2) >> FF_LAMBDA_SHIFT; + rd += (out_size * lambda2) >> (FF_LAMBDA_SHIFT - 3); } /* get the delayed frames */ while(out_size){ out_size = avcodec_encode_video(c, outbuf, outbuf_size, NULL); - rd += (out_size * lambda2) >> FF_LAMBDA_SHIFT; + rd += (out_size * lambda2) >> (FF_LAMBDA_SHIFT - 3); } rd += c->error[0] + c->error[1] + c->error[2]; @@ -2391,7 +2405,7 @@ no_output_pic: s->current_picture_ptr= s->reordered_input_picture[0]; for(i=0; i<4; i++){ - s->new_picture.data[i]+=16; + s->new_picture.data[i]+= INPLACE_OFFSET; } } copy_picture(&s->current_picture, s->current_picture_ptr); diff --git a/libavcodec/mpegvideo.h b/libavcodec/mpegvideo.h index c1b7024662..59746bea86 100644 --- a/libavcodec/mpegvideo.h +++ b/libavcodec/mpegvideo.h @@ -70,6 +70,8 @@ enum OutputFormat { #define MAX_MB_BYTES (30*16*16*3/8 + 120) +#define INPLACE_OFFSET 16 + typedef struct Predictor{ double coeff; double count; |