2,543
社区成员
发帖
与我相关
我的任务
分享
#include "Cffmpeg.h"
#include <stdio.h>
#include <windows.h>
#include <QDateTime>
#include <QDebug>
#define __STDC_CONSTANT_MACROS
#pragma comment(lib, "avutil.lib")
#pragma comment(lib, "avcodec.lib")
#pragma comment(lib, "avformat.lib")
#pragma comment(lib, "swscale.lib")
//内存对齐可能对CPU效率有影响; 具体根据CPU的存取粒度;
int Cffmpeg::interrupt_cb(void *mpeg)
{
Cffmpeg *p_mpeg = (Cffmpeg *)mpeg;
if ((av_gettime()/1000 - p_mpeg->dwLastFrameRealtime) > 60*1000)
{
qDebug() << QStringLiteral("断开连接!");
return 1;
}
return 0;
}
Cffmpeg::Cffmpeg()
{
ffmpegInit();
QDateTime dateTime = QDateTime::currentDateTime();
dwLastFrameRealtime = dateTime.toMSecsSinceEpoch();
}
Cffmpeg::~Cffmpeg()
{
if (m_pDstFrame)
av_frame_free(&m_pDstFrame);
if (m_pImg_convert_ctx)
sws_freeContext(m_pImg_convert_ctx);
if (m_pCodecCtx)
avcodec_close(m_pCodecCtx);
if (m_pFormatCtx)
avformat_close_input(&m_pFormatCtx);
if (m_pCodec)
{
delete m_pCodec;
}
}
int Cffmpeg::ffmpegInit()
{
m_pFormatCtx = NULL;
m_pCodecCtx = NULL;
m_pCodec = NULL;
m_pDstFrame = NULL;
m_pImg_convert_ctx = NULL;
m_iVideoindex = -1;
isFormatCtxNull = 0;
m_QSfilepath.clear();
memset(m_errBuf, 0, sizeof(m_errBuf));
av_register_all();
avcodec_register_all();
avformat_network_init();
m_pFormatCtx = avformat_alloc_context();
m_pDstFrame = av_frame_alloc(); //申请一个AVFrame 并做一些初始化工作
return DJJ_SUCCESS;
}
int Cffmpeg::OpenVideoFile(QString filepath, int *linksocket)
{
int ret, i;
QByteArray qbFilepath = filepath.toLatin1();
char *szFilepath = qbFilepath.data();
m_pFormatCtx->interrupt_callback.callback = interrupt_cb;
m_pFormatCtx->interrupt_callback.opaque = this;
AVDictionary* options = NULL;
av_dict_set(&options, "rtsp_transport", "tcp", 0);
av_dict_set(&options, "bufsize", "655360", 0);
*linksocket = avformat_open_input(&m_pFormatCtx, szFilepath, NULL, &options);
if (0 != *linksocket)
{
av_strerror(*linksocket, m_errBuf, sizeof(m_errBuf));
printf("Could not open input stream, filename: %s. Error:%d(%s)\n", szFilepath, *linksocket, m_errBuf);
avformat_close_input(&m_pFormatCtx);
return DJJ_FAILURE;
}
m_pFormatCtx->probesize = 32;
ret = avformat_find_stream_info(m_pFormatCtx, NULL);
if (ret < 0)
{
av_strerror(ret, m_errBuf, sizeof(m_errBuf));
printf("Could not find stream information Error: %d(%s)\n", ret, m_errBuf);
return DJJ_FAILURE;
}
printf("------------------------File Information-------------------------\n");
av_dump_format(m_pFormatCtx, 0, szFilepath, 0); //获取文件的信息到 AVFormatContext
printf("------------------------------END--------------------------------\n");
return DJJ_SUCCESS;
}
int Cffmpeg::SwsVideo(int *w,int *h,int dstWidth/* =0 */, int dstHeight/* =0 */, AVPixelFormat pix_fmt /* = AV_PIX_FMT_YUV420P */)
{
int i, videoindex, ret;
uint8_t *out_buffer;
m_iVideoindex = -1;
if (!m_pFormatCtx)
{
return DJJ_FAILURE;
}
for (i = 0; i < m_pFormatCtx->nb_streams; i++)
{
if (m_pFormatCtx->streams[i]->codec->codec_type == AVMEDIA_TYPE_VIDEO)
{
m_iVideoindex = i;
break;
}
}
if (-1 == m_iVideoindex)
{
printf("Can not find a video stream.\n");
return DJJ_FAILURE;
}
m_pCodecCtx = m_pFormatCtx->streams[m_iVideoindex]->codec;
m_pCodec = avcodec_find_decoder(m_pCodecCtx->codec_id); //通过code ID查找一个已经注册的音视频编码器;
if (m_pCodec == NULL)
{
printf("Codec not found!\n");
return DJJ_FAILURE;
}
ret = avcodec_open2(m_pCodecCtx, m_pCodec, NULL); //使用给定的AvCodec初始化AVCodecContext
if (ret < 0)
{
av_strerror(ret, m_errBuf, sizeof(m_errBuf));
printf("Could not open codec. Error:%d(%s)\n", ret, m_errBuf);
avcodec_close(m_pCodecCtx);
m_pCodecCtx = NULL;
return DJJ_FAILURE;
}
out_buffer = (uint8_t *)av_malloc(avpicture_get_size(pix_fmt, m_pCodecCtx->width, m_pCodecCtx->height));//avpicture_get_size 解码之后一帧图像的大小;//av_malloc封装了malloc并做了一些安全性操作;
avpicture_fill((AVPicture *)m_pDstFrame, out_buffer, pix_fmt, m_pCodecCtx->width, m_pCodecCtx->height);//avpicture_fill 将pFrameYUV 按照 AV_PIX_FMT_YUV420P的格式“关联”到 out_buffer, 数据转换完之后 也自动保存到了out_buffer;
m_pImg_convert_ctx = sws_getContext(m_pCodecCtx->width,
m_pCodecCtx->height,
m_pCodecCtx->pix_fmt,
(dstWidth == 0 ? m_pCodecCtx->width:dstWidth),
(dstHeight == 0? m_pCodecCtx->height:dstHeight),
pix_fmt,
SWS_BICUBIC,
NULL,
NULL,
NULL);//获得并初始化一个SwsContext结构体,这是一个缩放及格式转换的函数 这里的缩放比例为1:1
*w = m_pCodecCtx->width;
*h = m_pCodecCtx->height;
return DJJ_SUCCESS;
}
int Cffmpeg::ReadPkt(AVPacket** pOutPkt,int count)
{
int ret;
if (!m_pFormatCtx)
{
return DJJ_FAILURE;
}
ret = av_read_frame(m_pFormatCtx, *pOutPkt);//从源文件容器中读取一个AVPacket数据包; (音频可能包含多帧),这个时候为AVpacket的数据分配内存
if (ret >= 0)
{
if ((*pOutPkt)->stream_index == m_iVideoindex) //如果是视频帧的话,进行解码
{
QDateTime dateTime = QDateTime::currentDateTime();
dwLastFrameRealtime = dateTime.toMSecsSinceEpoch();
return DJJ_SUCCESS;
}
}
else
{
if (count < 20)//读取数据出错计数器
{
av_strerror(ret, m_errBuf, sizeof(m_errBuf));
printf("Read Frame Error:%d(%s)\n", ret, m_errBuf);
return DJJ_FAILURE;
}
else
{
isFormatCtxNull = 1;
Sleep(3000);
emit breakLine();
return DJJ_FAILURE;
}
}
return DJJ_FAILURE;
}
int Cffmpeg::ProcessPkt(int *count, AVPacket* pPkt, uint8_t** out, int* linesize)
{
int ret, got_picture = 0;
AVFrame* testFrame;
testFrame = av_frame_alloc();
// printf("AVPacket->pts:%d\n", pPkt->pts);
ret = avcodec_decode_video2(m_pCodecCtx, testFrame, &got_picture, pPkt);
if (ret < 0)
{
(*count)++;//解码出错计数器(日志no frame问题,不用理会)
av_strerror(ret, m_errBuf, sizeof(m_errBuf));
printf("Decode Error:%d(%s)\n", ret, m_errBuf);
return DJJ_FAILURE;
}
(*count) = 0;//重置解码出错计数器
if (got_picture)
{
sws_scale(m_pImg_convert_ctx,
(const uint8_t* const*)testFrame->data,
testFrame->linesize,
0,
m_pCodecCtx->height,
m_pDstFrame->data,
m_pDstFrame->linesize);//图片像素转换、拉伸函数;
(*out) = (uint8_t*)malloc(avpicture_get_size(AV_PIX_FMT_YUV420P, m_pCodecCtx->width, m_pCodecCtx->height));
memcpy((*out), m_pDstFrame->data[0], avpicture_get_size(AV_PIX_FMT_YUV420P, m_pCodecCtx->width, m_pCodecCtx->height));
*linesize = m_pDstFrame->linesize[0];
}
av_frame_unref(testFrame);
av_frame_free(&testFrame);
if (got_picture)
{
return DJJ_SUCCESS;
}
else
{
return DJJ_FAILURE;
}
}
dwLastFrameRealtime = dateTime.toMSecsSinceEpoch()