zoukankan      html  css  js  c++  java
  • windows ffmpeg 推送摄像头数据到rtmp服务

    文本主要讲述windows系统下如何利用ffmpeg获取摄像机流并推送到rtmp服务,命令的用法前文

    中有讲到过,这次是通过代码来实现。实现该项功能的基本流程如下:

                     图1 ffmpeg推流流程图

    较前面的文章的流程图而言,本流程图显的复杂些,稍微解释下:

    ffmpeg 打开摄像头跟打开普通的视频流方法一致,只是输入url是摄像头的名称。真正打开

    摄像头操作由dshow来完成,ffmpeg只是调用dshow相应的接口获取返回值;读取packet

    的API 依然是av_read_frame,返回的packet并不是编码后的视频包而是原始的图片帧,图片

    帧的格式是YUV422(我测试的机器是这样的)。拿到图片帧后并不能直接编码,需要做格式转换

    因为libx264不支持YUV422,所以有了上图中的格式转换这一环节。转换图片帧格式后将视频数据

    丢给编码器就行编码,编码的packet才可以写到输出流中。

    下面给出代码:

      1.初始化Ffmpeg

    void Init()
    {
         av_register_all();
         avfilter_register_all();
    avformat_network_init(); avdevice_register_all();
    av_log_set_level(AV_LOG_ERROR); }

     2.创建输入上下文,打开摄像头

    int OpenInput(char *fileName)
    {
    	 context = avformat_alloc_context();
    	 context->interrupt_callback.callback = interrupt_cb;
    	 AVInputFormat *ifmt = av_find_input_format("dshow");
    	 AVDictionary *format_opts =  nullptr;
    	 av_dict_set_int(&format_opts, "rtbufsize", 18432000  , 0);
    	int ret = avformat_open_input(&context, fileName, ifmt, &format_opts);
    	if(ret < 0)
    	{
    		return  ret;
    	}
    	ret = avformat_find_stream_info(context,nullptr);
    	av_dump_format(context, 0, fileName, 0);
    	if(ret >= 0) 
    	{
    		cout <<"open input stream successfully" << endl;
    	}
    	return ret;
    }
    

     3. 初始化编码器

    int InitOutputCodec(AVCodecContext** pOutPutEncContext, int iWidth, int iHeight)
    {
    	AVCodec *  pH264Codec = avcodec_find_encoder(AV_CODEC_ID_H264);
    	if(NULL == pH264Codec)
    	{
    		printf("%s", "avcodec_find_encoder failed");
    		return  0;
    	}
    	*pOutPutEncContext = avcodec_alloc_context3(pH264Codec);
    	(*pOutPutEncContext)->codec_id = pH264Codec->id;
    	(*pOutPutEncContext)->time_base.num =0;
    	(*pOutPutEncContext)->time_base.den = 1;
    	(*pOutPutEncContext)->pix_fmt            = AV_PIX_FMT_YUV420P;
    	(*pOutPutEncContext)->width              =  iWidth;
    	(*pOutPutEncContext)->height             = iHeight;
    	(*pOutPutEncContext)->has_b_frames = 0;
    	(*pOutPutEncContext)->max_b_frames = 0;
    
    	AVDictionary *options = nullptr;
    	(*pOutPutEncContext)->flags |= AV_CODEC_FLAG_GLOBAL_HEADER;
    	int ret = avcodec_open2(*pOutPutEncContext, pH264Codec, &options);
    	if (ret < 0)
    	{
    		printf("%s", "open codec failed");
    		return  ret;
    	}
    	return 1;
    }
    

     4.创建输出上下文以及输出流

    int OpenOutput(char *fileName)
    {
    	int ret = 0;
    	 ret  = avformat_alloc_output_context2(&outputContext, nullptr, "flv", fileName);
    	if(ret < 0)
    	{
    		goto Error;
    	}
    	ret = avio_open2(&outputContext->pb, fileName, AVIO_FLAG_READ_WRITE,nullptr, nullptr);	
    	if(ret < 0)
    	{
    		goto Error;
    	}
    
    	for(int i = 0; i < context->nb_streams; i++)
    	{
    		AVStream * stream = avformat_new_stream(outputContext, pOutPutEncContext->codec);
    		stream->codec = pOutPutEncContext;
    		if(ret < 0)
    		{
    			goto Error;
    		}
    	}
    	 av_dump_format(outputContext, 0, fileName, 1);
    	 ret = avformat_write_header(outputContext, nullptr);
    	if(ret < 0)
    	{
    		goto Error;
    	}
    	if(ret >= 0)
    	cout <<"open output stream successfully" << endl;
    	return ret ;
    Error:
    	if(outputContext)
    	{
    		for(int i = 0; i < outputContext->nb_streams; i++)
    		{
    			avcodec_close(outputContext->streams[i]->codec);
    		}
    		avformat_close_input(&outputContext);
    	}
    	return ret ;
    }
    

     4.从输入流读取视频包

    std::shared_ptr<AVPacket> ReadPacketFromSource()
    {
    	std::shared_ptr<AVPacket> packet(static_cast<AVPacket*>(av_malloc(sizeof(AVPacket))), [&](AVPacket *p) { av_packet_free(&p); av_freep(&p); });
    	av_init_packet(packet.get());
    	lastFrameRealtime = av_gettime();
    	int ret = av_read_frame(context, packet.get());
    	if(ret >= 0)
    	{
    		return packet;
    	}
    	else
    	{
    		return nullptr;
    	}
    }
    

      5.格式转换

    int YUV422To420(uint8_t *yuv422, uint8_t *yuv420, int width, int height)
    {
    	int s = width * height;
    	int i,j,k = 0;
    	for(i = 0; i < s;i++)
    	{
    		yuv420[i] = yuv422[i * 2];
    	}
    
    	for(i = 0; i < height; i++)
    	{
    		if(i%2 != 0) continue;
    		for(j = 0; j <(width /2); j++)
    		{
    			if(4 * j + 1 > 2 * width) break;
    			yuv420[s + k * 2 * width / 4 +j] = yuv422[i * 2 * width + 4 *j + 1];
    		}
    		k++;
    	}
    
    	k = 0;
    
    	for(i = 0; i < height; i++)
    	{
    		if(i % 2 == 0) continue;
    		for(j = 0; j < width / 2; j++)
    		{
    			if(4 * j + 3 > 2 * width) break;
    			yuv420[s + s / 4 + k * 2 *width / 4 + j] = yuv422[i *2 * width + 4 * j + 3];
    		}
    		k++;
    	}
    	return 1;
    };
    

      6. 简单实例

    	string fileInput="video=Integrated Webcam";
    	string fileOutput="rtmp://127.0.0.1/live/mystream";
    
    	Init();
    	if(OpenInput((char *)fileInput.c_str()) < 0)
    	{
    		cout << "Open file Input failed!" << endl;
    		this_thread::sleep_for(chrono::seconds(10));
    		return 0;
    	}
    	InitOutputCodec(&pOutPutEncContext,DstWidth,DstHeight);
    	if(OpenOutput((char *)fileOutput.c_str()) < 0)
    	{
    		cout << "Open file Output failed!" << endl;
    		this_thread::sleep_for(chrono::seconds(10));
    		return 0;
    	}
    	auto timebase = av_q2d(context->streams[0]->time_base);
    	int count = 0;
    	auto in_stream = context->streams[0];
    	auto out_stream = outputContext->streams[0];
    	int iGetPic = 0;
    	uint8_t *yuv420Buffer = (uint8_t *)malloc(DstWidth * DstHeight * 3 / 2);
    			yuv420Buffer[DstWidth * DstHeight * 3 / 2 - 1] = 0;
    	while(true)
    	{
    		auto packet = ReadPacketFromSource();
    		if(packet)
    		{
    			auto pSwsFrame = av_frame_alloc();
    			int numBytes=av_image_get_buffer_size(AV_PIX_FMT_YUYV422, DstWidth, DstHeight, 1);
    			YUV422To420(packet->data, yuv420Buffer, DstWidth, DstHeight);
    			av_image_fill_arrays((pSwsFrame)->data, (pSwsFrame)->linesize, yuv420Buffer, AV_PIX_FMT_YUV420P, DstWidth, DstHeight, 1);
    			AVPacket *pTmpPkt = (AVPacket *)av_malloc(sizeof(AVPacket));
    			av_init_packet(pTmpPkt);
    			pTmpPkt->data = NULL;
    			pTmpPkt->size = 0;
    
    			int iRet = avcodec_encode_video2(pOutPutEncContext, pTmpPkt, pSwsFrame, &iGetPic);
    			if(iRet >= 0 && iGetPic)
    			{
    				int ret = av_write_frame(outputContext, pTmpPkt);
    				cout << "ret:"<< ret <<endl;
    				//this_thread::sleep_for(std::chrono::milliseconds(40));
    			}
    			av_frame_free(&pSwsFrame);
    			av_packet_free(&pTmpPkt);
    		}
    	}
    	CloseInput();
    	CloseOutput();
    	cout <<"Transcode file end!" << endl;
    	this_thread::sleep_for(chrono::hours(10));
    	return 0;
    

      

    如需交流,可以加QQ群1038388075,766718184,或者QQ:350197870

    视频教程 播放地址: http://www.iqiyi.com/u/1426749687

     视频下载地址:http://www.chungen90.com/?news_3/

     Demo下载地址: http://www.chungen90.com/?news_2

      

     

     

  • 相关阅读:
    BSS Audio® Introduces Full-Bandwidth Acoustic Echo Cancellation Algorithm for Soundweb London Conferencing Processors
    转:虚拟运营商颠覆八大行业 170号码将成主流?
    移动通信调制技术的进展 转
    转:瑞利信道,莱斯信道和高斯信道模型
    转:Android开发之旅:环境搭建及HelloWorld
    web端视频直播网站的弊端和优势
    频域分辨率与DFT,DCT,MDCT理解
    转:超声波支付
    谈音频算法技术研发团队建立
    转:HTML5标准与性能之四:asm.js
  • 原文地址:https://www.cnblogs.com/wanggang123/p/6133212.html
Copyright © 2011-2022 走看看