/////////////////////////////////////////////////////////////////////////////////// // Copyright (C) 2018 F4HKW // // for F4EXB / SDRAngel // // // // This program is free software; you can redistribute it and/or modify // // it under the terms of the GNU General Public License as published by // // the Free Software Foundation as version 3 of the License, or // // // // This program is distributed in the hope that it will be useful, // // but WITHOUT ANY WARRANTY; without even the implied warranty of // // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // // GNU General Public License V3 for more details. // // // // You should have received a copy of the GNU General Public License // // along with this program. If not, see . // /////////////////////////////////////////////////////////////////////////////////// #include #include #include extern "C" { #include } #include "audio/audiofifo.h" #include "datvideorender.h" DATVideoRender::DATVideoRender(QWidget *parent) : TVScreen(true, parent), m_parentWidget(parent) { installEventFilter(this); m_isFullScreen = false; m_running = false; m_isFFMPEGInitialized = false; m_isOpen = false; m_formatCtx = nullptr; m_videoDecoderCtx = nullptr; m_audioDecoderCtx = nullptr; m_swsCtx = nullptr; m_audioFifo = nullptr; m_audioSWR = nullptr; m_audioSampleRate = 48000; m_audioFifoBufferIndex = 0; m_videoStreamIndex = -1; m_audioStreamIndex = -1; m_audioMute = false; m_videoMute = false; m_audioVolume = 0; m_updateAudioResampler = false; m_currentRenderWidth = -1; m_currentRenderHeight = -1; m_frame = nullptr; m_frameCount = -1; m_audioDecodeOK = false; m_videoDecodeOK = false; // for (int i = 0; i < m_audioFifoBufferSize; i++) // { // m_audioFifoBuffer[2*i] = 8192.0f * sin((M_PI * i)/(m_audioFifoBufferSize/1000.0f)); // m_audioFifoBuffer[2*i+1] = m_audioFifoBuffer[2*i]; // } } DATVideoRender::~DATVideoRender() { if (m_audioSWR) { swr_free(&m_audioSWR); } } bool DATVideoRender::eventFilter(QObject *obj, QEvent *event) { if (event->type() == QEvent::MouseButtonRelease) { SetFullScreen(false); return true; } else { // standard event processing return QObject::eventFilter(obj, event); } } void DATVideoRender::SetFullScreen(bool fullScreen) { if (m_isFullScreen == fullScreen) { return; } if (fullScreen == true) { qDebug("DATVideoRender::SetFullScreen: go to fullscreen"); // m_originalWindowFlags = this->windowFlags(); // m_originalSize = this->size(); // m_parentWidget->layout()->removeWidget(this); // //this->setParent(0); // this->setWindowFlags( Qt::FramelessWindowHint|Qt::WindowStaysOnTopHint); // m_parentWidget->show(); setWindowFlags(Qt::Window); setWindowState(Qt::WindowFullScreen); show(); m_isFullScreen = true; } else { qDebug("DATVideoRender::SetFullScreen: come back from fullscreen"); // //this->setParent(m_parentWidget); // this->resize(m_originalSize); // this->overrideWindowFlags(m_originalWindowFlags); // setWindowState(Qt::WindowNoState); // m_parentWidget->layout()->addWidget(this); // m_parentWidget->show(); setWindowFlags(Qt::Widget); setWindowState(Qt::WindowNoState); show(); m_isFullScreen = false; } } static int ReadFunction(void *opaque, uint8_t *buf, int buf_size) { QIODevice *stream = reinterpret_cast(opaque); int nbBytes = stream->read((char *)buf, buf_size); return nbBytes; } static int64_t SeekFunction(void *opaque, int64_t offset, int whence) { QIODevice *stream = reinterpret_cast(opaque); if (whence == AVSEEK_SIZE) { return -1; } if (stream->isSequential()) { return -1; } if (stream->seek(offset) == false) { return -1; } return stream->pos(); } void DATVideoRender::ResetMetaData() { MetaData.CodecID = -1; MetaData.PID = -1; MetaData.Program = ""; MetaData.Stream = ""; MetaData.Width = -1; MetaData.Height = -1; MetaData.BitRate = -1; MetaData.Channels = -1; MetaData.CodecDescription = ""; MetaData.OK_Decoding = false; MetaData.OK_TransportStream = false; MetaData.OK_VideoStream = false; emit onMetaDataChanged(&MetaData); } bool DATVideoRender::InitializeFFMPEG() { ResetMetaData(); if (m_isFFMPEGInitialized) { return false; } avcodec_register_all(); av_register_all(); av_log_set_level(AV_LOG_FATAL); //av_log_set_level(AV_LOG_ERROR); m_isFFMPEGInitialized = true; return true; } bool DATVideoRender::PreprocessStream() { AVDictionary *opts = nullptr; AVCodec *videoCodec = nullptr; AVCodec *audioCodec = nullptr; int intRet = -1; char *buffer = nullptr; //Identify stream if (avformat_find_stream_info(m_formatCtx, nullptr) < 0) { avformat_close_input(&m_formatCtx); m_formatCtx = nullptr; qDebug() << "DATVideoProcess::PreprocessStream cannot find stream info"; return false; } //Find video stream intRet = av_find_best_stream(m_formatCtx, AVMEDIA_TYPE_VIDEO, -1, -1, nullptr, 0); if (intRet < 0) { avformat_close_input(&m_formatCtx); qDebug() << "DATVideoProcess::PreprocessStream cannot find video stream"; return false; } m_videoStreamIndex = intRet; //Find audio stream intRet = av_find_best_stream(m_formatCtx, AVMEDIA_TYPE_AUDIO, -1, -1, nullptr, 0); if (intRet < 0) { qDebug() << "DATVideoProcess::PreprocessStream cannot find audio stream"; } m_audioStreamIndex = intRet; // Prepare Video Codec and extract meta data AVCodecParameters *parms = m_formatCtx->streams[m_videoStreamIndex]->codecpar; if (m_videoDecoderCtx) { avcodec_free_context(&m_videoDecoderCtx); } m_videoDecoderCtx = avcodec_alloc_context3(NULL); avcodec_parameters_to_context(m_videoDecoderCtx, parms); // m_videoDecoderCtx = m_formatCtx->streams[m_videoStreamIndex]->codec; // old style //Meta Data MetaData.PID = m_formatCtx->streams[m_videoStreamIndex]->id; MetaData.CodecID = m_videoDecoderCtx->codec_id; MetaData.OK_TransportStream = true; MetaData.Program = ""; MetaData.Stream = ""; if (m_formatCtx->programs) { buffer = nullptr; av_dict_get_string(m_formatCtx->programs[m_videoStreamIndex]->metadata, &buffer, ':', '\n'); if (buffer != nullptr) { MetaData.Program = QString("%1").arg(buffer); } } buffer = nullptr; av_dict_get_string(m_formatCtx->streams[m_videoStreamIndex]->metadata, &buffer, ':', '\n'); if (buffer != nullptr) { MetaData.Stream = QString("%1").arg(buffer); } emit onMetaDataChanged(&MetaData); //Decoder videoCodec = avcodec_find_decoder(m_videoDecoderCtx->codec_id); if (videoCodec == nullptr) { avformat_close_input(&m_formatCtx); m_formatCtx = nullptr; qDebug() << "DATVideoProcess::PreprocessStream cannot find associated video CODEC"; return false; } else { qDebug() << "DATVideoProcess::PreprocessStream: video CODEC found: " << videoCodec->name; } av_dict_set(&opts, "refcounted_frames", "1", 0); if (avcodec_open2(m_videoDecoderCtx, videoCodec, &opts) < 0) { avformat_close_input(&m_formatCtx); m_formatCtx = nullptr; qDebug() << "DATVideoProcess::PreprocessStream cannot open associated video CODEC"; return false; } //Allocate Frame m_frame = av_frame_alloc(); if (!m_frame) { avformat_close_input(&m_formatCtx); m_formatCtx = nullptr; qDebug() << "DATVideoProcess::PreprocessStream cannot allocate frame"; return false; } m_frameCount = 0; MetaData.Width = m_videoDecoderCtx->width; MetaData.Height = m_videoDecoderCtx->height; MetaData.BitRate = m_videoDecoderCtx->bit_rate; MetaData.Channels = m_videoDecoderCtx->channels; MetaData.CodecDescription = QString("%1").arg(videoCodec->long_name); MetaData.OK_VideoStream = true; emit onMetaDataChanged(&MetaData); // Prepare Audio Codec if (m_audioStreamIndex >= 0) { AVCodecParameters *parms = m_formatCtx->streams[m_audioStreamIndex]->codecpar; if (m_audioDecoderCtx) { avcodec_free_context(&m_audioDecoderCtx); } m_audioDecoderCtx = avcodec_alloc_context3(NULL); avcodec_parameters_to_context(m_audioDecoderCtx, parms); //m_audioDecoderCtx = m_formatCtx->streams[m_audioStreamIndex]->codec; // old style qDebug() << "DATVideoProcess::PreprocessStream: audio: " << " channels: " << m_audioDecoderCtx->channels << " channel_layout: " << m_audioDecoderCtx->channel_layout << " sample_rate: " << m_audioDecoderCtx->sample_rate << " sample_fmt: " << m_audioDecoderCtx->sample_fmt << " codec_id: "<< m_audioDecoderCtx->codec_id; audioCodec = avcodec_find_decoder(m_audioDecoderCtx->codec_id); if (audioCodec == nullptr) { qDebug() << "DATVideoProcess::PreprocessStream cannot find associated audio CODEC"; m_audioStreamIndex = -1; // invalidate audio } else { qDebug() << "DATVideoProcess::PreprocessStream: audio CODEC found: " << audioCodec->name; if (avcodec_open2(m_audioDecoderCtx, audioCodec, nullptr) < 0) { qDebug() << "DATVideoProcess::PreprocessStream cannot open associated audio CODEC"; m_audioStreamIndex = -1; // invalidate audio } else { setResampler(); } } } return true; } bool DATVideoRender::OpenStream(DATVideostream *device) { int ioBufferSize = 32768; unsigned char *ptrIOBuffer = nullptr; AVIOContext *ioCtx = nullptr; if (m_running) { return false; } if (device == nullptr) { qDebug() << "DATVideoProcess::OpenStream QIODevice is nullptr"; return false; } if (m_isOpen) { qDebug() << "DATVideoProcess::OpenStream already open"; return false; } if (device->bytesAvailable() <= 0) { qDebug() << "DATVideoProcess::OpenStream no data available"; MetaData.OK_Data = false; emit onMetaDataChanged(&MetaData); return false; } //Only once execution m_running = true; MetaData.OK_Data = true; emit onMetaDataChanged(&MetaData); InitializeFFMPEG(); if (!m_isFFMPEGInitialized) { qDebug() << "DATVideoProcess::OpenStream FFMPEG not initialized"; m_running = false; return false; } if (!device->open(QIODevice::ReadOnly)) { qDebug() << "DATVideoProcess::OpenStream cannot open QIODevice"; m_running = false; return false; } //Connect QIODevice to FFMPEG Reader m_formatCtx = avformat_alloc_context(); if (m_formatCtx == nullptr) { qDebug() << "DATVideoProcess::OpenStream cannot alloc format FFMPEG context"; m_running = false; return false; } ptrIOBuffer = (unsigned char *)av_malloc(ioBufferSize + AV_INPUT_BUFFER_PADDING_SIZE); ioCtx = avio_alloc_context(ptrIOBuffer, ioBufferSize, 0, reinterpret_cast(device), &ReadFunction, nullptr, &SeekFunction); m_formatCtx->pb = ioCtx; m_formatCtx->flags |= AVFMT_FLAG_CUSTOM_IO; if (avformat_open_input(&m_formatCtx, nullptr, nullptr, nullptr) < 0) { qDebug() << "DATVideoProcess::OpenStream cannot open stream"; m_running = false; return false; } if (!PreprocessStream()) { m_running = false; return false; } m_isOpen = true; m_running = false; return true; } bool DATVideoRender::RenderStream() { AVPacket packet; int gotFrame; bool needRenderingSetup; if (!m_isOpen) { qDebug() << "DATVideoProcess::RenderStream Stream not open"; return false; } if (m_running) { return false; } //Only once execution m_running = true; //********** Rendering ********** if (av_read_frame(m_formatCtx, &packet) < 0) { qDebug() << "DATVideoProcess::RenderStream reading packet error"; m_running = false; return false; } //Video channel if ((packet.stream_index == m_videoStreamIndex) && (!m_videoMute)) { memset(m_frame, 0, sizeof(AVFrame)); av_frame_unref(m_frame); gotFrame = 0; if (new_decode(m_videoDecoderCtx, m_frame, &gotFrame, &packet) >= 0) { m_videoDecodeOK = true; if (gotFrame) { //Rendering and RGB Converter setup needRenderingSetup = (m_frameCount == 0); needRenderingSetup |= (m_swsCtx == nullptr); if ((m_currentRenderWidth != m_frame->width) || (m_currentRenderHeight != m_frame->height)) { needRenderingSetup = true; } if (needRenderingSetup) { if (m_swsCtx != nullptr) { sws_freeContext(m_swsCtx); m_swsCtx = nullptr; } //Convertisseur YUV -> RGB m_swsCtx = sws_alloc_context(); av_opt_set_int(m_swsCtx, "srcw", m_frame->width, 0); av_opt_set_int(m_swsCtx, "srch", m_frame->height, 0); av_opt_set_int(m_swsCtx, "src_format", m_frame->format, 0); av_opt_set_int(m_swsCtx, "dstw", m_frame->width, 0); av_opt_set_int(m_swsCtx, "dsth", m_frame->height, 0); av_opt_set_int(m_swsCtx, "dst_format", AV_PIX_FMT_RGB24, 0); av_opt_set_int(m_swsCtx, "sws_flag", SWS_FAST_BILINEAR /* SWS_BICUBIC*/, 0); if (sws_init_context(m_swsCtx, nullptr, nullptr) < 0) { qDebug() << "DATVideoProcess::RenderStream cannont init video data converter"; m_swsCtx = nullptr; m_running = false; return false; } if ((m_currentRenderHeight > 0) && (m_currentRenderWidth > 0)) { //av_freep(&m_pbytDecodedData[0]); //av_freep(&m_pintDecodedLineSize[0]); } if (av_image_alloc(m_pbytDecodedData, m_pintDecodedLineSize, m_frame->width, m_frame->height, AV_PIX_FMT_RGB24, 1) < 0) { qDebug() << "DATVideoProcess::RenderStream cannont init video image buffer"; sws_freeContext(m_swsCtx); m_swsCtx = nullptr; m_running = false; return false; } //Rendering device setup resizeTVScreen(m_frame->width, m_frame->height); update(); resetImage(); m_currentRenderWidth = m_frame->width; m_currentRenderHeight = m_frame->height; MetaData.Width = m_frame->width; MetaData.Height = m_frame->height; MetaData.OK_Decoding = true; emit onMetaDataChanged(&MetaData); } //Frame rendering if (sws_scale(m_swsCtx, m_frame->data, m_frame->linesize, 0, m_frame->height, m_pbytDecodedData, m_pintDecodedLineSize) < 0) { qDebug() << "DATVideoProcess::RenderStream error converting video frame to RGB"; m_running = false; return false; } renderImage(m_pbytDecodedData[0]); av_frame_unref(m_frame); m_frameCount++; } } else { m_videoDecodeOK = false; // qDebug() << "DATVideoProcess::RenderStream video decode error"; } } // Audio channel else if ((packet.stream_index == m_audioStreamIndex) && (m_audioFifo) && (swr_is_initialized(m_audioSWR)) && (!m_audioMute)) { if (m_updateAudioResampler) { setResampler(); m_updateAudioResampler = false; } memset(m_frame, 0, sizeof(AVFrame)); av_frame_unref(m_frame); gotFrame = 0; if (new_decode(m_audioDecoderCtx, m_frame, &gotFrame, &packet) >= 0) { m_audioDecodeOK = true; if (gotFrame) { int16_t *audioBuffer; av_samples_alloc((uint8_t**) &audioBuffer, nullptr, 2, m_frame->nb_samples, AV_SAMPLE_FMT_S16, 0); int frame_count = swr_convert(m_audioSWR, (uint8_t**) &audioBuffer, m_frame->nb_samples, (const uint8_t**) m_frame->data, m_frame->nb_samples); // direct writing: // int ret = m_audioFifo->write((const quint8*) &audioBuffer[0], frame_count); // if (ret < frame_count) { // qDebug("DATVideoRender::RenderStream: audio frames missing %d vs %d", ret, frame_count); // } // buffered writing: // if (m_audioFifoBufferIndex + frame_count < m_audioFifoBufferSize) // { // std::copy(&audioBuffer[0], &audioBuffer[2*frame_count], &m_audioFifoBuffer[2*m_audioFifoBufferIndex]); // m_audioFifoBufferIndex += frame_count; // } // else // { // int remainder = m_audioFifoBufferSize - m_audioFifoBufferIndex; // std::copy(&audioBuffer[0], &audioBuffer[2*remainder], &m_audioFifoBuffer[2*m_audioFifoBufferIndex]); // m_audioFifo->write((const quint8*) &m_audioFifoBuffer[0], m_audioFifoBufferSize); // std::copy(&audioBuffer[2*remainder], &audioBuffer[2*frame_count], &m_audioFifoBuffer[0]); // m_audioFifoBufferIndex = frame_count - remainder; // } // Apply volume: for (int i = 0; i < frame_count; i++) { m_audioFifoBuffer[2*m_audioFifoBufferIndex] = m_audioVolume * audioBuffer[2*i]; m_audioFifoBuffer[2*m_audioFifoBufferIndex+1] = m_audioVolume * audioBuffer[2*i+1]; m_audioFifoBufferIndex++; if (m_audioFifoBufferIndex >= m_audioFifoBufferSize) { m_audioFifo->write((const quint8*) &m_audioFifoBuffer[0], m_audioFifoBufferSize); m_audioFifoBufferIndex = 0; } } } } else { m_audioDecodeOK = false; // qDebug("DATVideoRender::RenderStream: audio decode error"); } } av_packet_unref(&packet); //********** Rendering ********** m_running = false; return true; } void DATVideoRender::setAudioVolume(int audioVolume) { int audioVolumeConstrained = audioVolume < 0 ? 0 : audioVolume > 100 ? 100 : audioVolume; m_audioVolume = pow(10.0, audioVolumeConstrained*0.02 - 2.0); // .01 -> 1 log } void DATVideoRender::setResampler() { if (m_audioSWR) { swr_free(&m_audioSWR); } m_audioSWR = swr_alloc(); av_opt_set_int(m_audioSWR, "in_channel_count", m_audioDecoderCtx->channels, 0); av_opt_set_int(m_audioSWR, "out_channel_count", 2, 0); av_opt_set_int(m_audioSWR, "in_channel_layout", m_audioDecoderCtx->channel_layout, 0); av_opt_set_int(m_audioSWR, "out_channel_layout", AV_CH_LAYOUT_STEREO, 0); av_opt_set_int(m_audioSWR, "in_sample_rate", m_audioDecoderCtx->sample_rate, 0); av_opt_set_int(m_audioSWR, "out_sample_rate", m_audioSampleRate, 0); av_opt_set_sample_fmt(m_audioSWR, "in_sample_fmt", m_audioDecoderCtx->sample_fmt, 0); av_opt_set_sample_fmt(m_audioSWR, "out_sample_fmt", AV_SAMPLE_FMT_S16, 0); swr_init(m_audioSWR); qDebug() << "DATVideoRender::setResampler: " << " in_channel_count: " << m_audioDecoderCtx->channels << " out_channel_count: " << 2 << " in_channel_layout: " << m_audioDecoderCtx->channel_layout << " out_channel_layout: " << AV_CH_LAYOUT_STEREO << " in_sample_rate: " << m_audioDecoderCtx->sample_rate << " out_sample_rate: " << m_audioSampleRate << " in_sample_fmt: " << m_audioDecoderCtx->sample_fmt << " out_sample_fmt: " << AV_SAMPLE_FMT_S16; } bool DATVideoRender::CloseStream(QIODevice *device) { if (m_running) { return false; } if (!device) { qDebug() << "DATVideoProcess::CloseStream QIODevice is nullptr"; return false; } if (!m_isOpen) { qDebug() << "DATVideoProcess::CloseStream Stream not open"; return false; } if (!m_formatCtx) { qDebug() << "DATVideoProcess::CloseStream FFMEG Context is not initialized"; return false; } //Only once execution m_running = true; // maybe done in the avcodec_close // avformat_close_input(&m_formatCtx); // m_formatCtx=nullptr; if (m_videoDecoderCtx) { avcodec_close(m_videoDecoderCtx); m_videoDecoderCtx = nullptr; } if (m_frame) { av_frame_unref(m_frame); av_frame_free(&m_frame); } if (m_swsCtx != nullptr) { sws_freeContext(m_swsCtx); m_swsCtx = nullptr; } device->close(); m_isOpen = false; m_running = false; m_currentRenderWidth = -1; m_currentRenderHeight = -1; ResetMetaData(); emit onMetaDataChanged(&MetaData); return true; } /** * Replacement of deprecated avcodec_decode_video2 with the same signature * https://blogs.gentoo.org/lu_zero/2016/03/29/new-avcodec-api/ */ int DATVideoRender::new_decode(AVCodecContext *avctx, AVFrame *frame, int *got_frame, AVPacket *pkt) { int ret; *got_frame = 0; if (pkt) { ret = avcodec_send_packet(avctx, pkt); // In particular, we don't expect AVERROR(EAGAIN), because we read all // decoded frames with avcodec_receive_frame() until done. if (ret < 0) { return ret == AVERROR_EOF ? 0 : ret; } } ret = avcodec_receive_frame(avctx, frame); if (ret < 0 && ret != AVERROR(EAGAIN) && ret != AVERROR_EOF) { return ret; } if (ret >= 0) { *got_frame = 1; } return 0; }