int frame_size_out_encode = out_fmt_ctx->streams[out_stream->index]->codec->frame_size; if(!frame_size_out_encode) frame_size_out_encode = put_into_fifo->nb_samples;
确定采样长度,先取得编码器frame_size,如果该值是0,比如pcm,则用输入Frame的nb_samples 作为采样长度。
AVCodecContext *acout = out_fmt_ctx->streams[0]->codec; AVFrame *frame_fifo; frame_fifo = av_frame_alloc(); frame_fifo->nb_samples = frame_size_out_encode; frame_fifo->channel_layout = acout->channel_layout; frame_fifo->channels = av_get_channel_layout_nb_channels(frame_fifo->channel_layout); frame_fifo->format = acout->sample_fmt; frame_fifo->sample_rate = acout->sample_rate; av_frame_get_buffer(frame_fifo, 0); av_samples_set_silence(frame_fifo->data, 0, frame_fifo->nb_samples, frame_fifo->channels, (AVSampleFormat)frame_fifo->format);
按channels,format等参数申请缓存
av_init_packet(&pkt_out); int ret = avcodec_encode_audio2(out_fmt_ctx->streams[0]->codec, &pkt_out, frame_fifo, &got_picture); if (got_picture ) { pkt_out.stream_index = out_stream->index; pkt_out.dts = av_rescale_q_rnd(pkt_out.dts, out_fmt_ctx->streams[out_stream->index]->codec->time_base, out_fmt_ctx->streams[out_stream->index]->time_base, (AVRounding)(AV_ROUND_NEAR_INF|AV_ROUND_PASS_MINMAX)); pkt_out.pts = av_rescale_q_rnd(pkt_out.pts, out_fmt_ctx->streams[out_stream->index]->codec->time_base, out_fmt_ctx->streams[out_stream->index]->time_base, (AVRounding)(AV_ROUND_NEAR_INF|AV_ROUND_PASS_MINMAX)); pkt_out.duration = av_rescale_q(pkt_out.duration, out_fmt_ctx->streams[out_stream->index]->codec->time_base, out_fmt_ctx->streams[out_stream->index]->time_base); av_log(NULL, AV_LOG_DEBUG, "Muxing frame "); /* mux encoded frame */ ret = av_interleaved_write_frame(out_fmt_ctx, &pkt_out); av_free_packet(&pkt_out); if (ret < 0) { printf("write a null frame failed! "); break; } printf("success write a null frame:index %d ", frame_index++); duration_sum += frame_size_out_encode; if(duration_sum * av_q2d(out_fmt_ctx->streams[out_stream->index]->codec->time_base) >= g_offset) break; }
编码,写入AVPacket,duration_sum记录总时长的信息,g_offset是静音时常