diff options
author | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-10 18:07:22 +0000 |
---|---|---|
committer | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-10 18:07:22 +0000 |
commit | c04dcc2e7d834218ef2d4194331e383402495ae1 (patch) | |
tree | 7333e38d10d75386e60f336b80c2443c1166031d /xbmc/guilib/FFmpegImage.cpp | |
parent | Initial commit. (diff) | |
download | kodi-c04dcc2e7d834218ef2d4194331e383402495ae1.tar.xz kodi-c04dcc2e7d834218ef2d4194331e383402495ae1.zip |
Adding upstream version 2:20.4+dfsg.upstream/2%20.4+dfsg
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'xbmc/guilib/FFmpegImage.cpp')
-rw-r--r-- | xbmc/guilib/FFmpegImage.cpp | 754 |
1 files changed, 754 insertions, 0 deletions
diff --git a/xbmc/guilib/FFmpegImage.cpp b/xbmc/guilib/FFmpegImage.cpp new file mode 100644 index 0000000..fa25527 --- /dev/null +++ b/xbmc/guilib/FFmpegImage.cpp @@ -0,0 +1,754 @@ +/* + * Copyright (C) 2012-2018 Team Kodi + * This file is part of Kodi - https://kodi.tv + * + * SPDX-License-Identifier: GPL-2.0-or-later + * See LICENSES/README.md for more information. + */ +#include "FFmpegImage.h" + +#include "cores/FFmpeg.h" +#include "guilib/Texture.h" +#include "settings/AdvancedSettings.h" +#include "settings/Settings.h" +#include "settings/SettingsComponent.h" +#include "utils/log.h" + +#include <algorithm> + +extern "C" +{ +#include <libavformat/avformat.h> +#include <libavutil/imgutils.h> +#include <libavcodec/avcodec.h> +#include <libavutil/avutil.h> +#include <libswscale/swscale.h> +#include <libavutil/pixdesc.h> +} + +Frame::Frame(const Frame& src) : + m_delay(src.m_delay), + m_imageSize(src.m_imageSize), + m_height(src.m_height), + m_width(src.m_width) +{ + if (src.m_pImage) + { + m_pImage = (unsigned char*) av_malloc(m_imageSize); + memcpy(m_pImage, src.m_pImage, m_imageSize); + } +} + +Frame::~Frame() +{ + av_free(m_pImage); + m_pImage = nullptr; +} + +struct ThumbDataManagement +{ + uint8_t* intermediateBuffer = nullptr; // gets av_alloced + AVFrame* frame_input = nullptr; + AVFrame* frame_temporary = nullptr; + SwsContext* sws = nullptr; + AVCodecContext* avOutctx = nullptr; + AVCodec* codec = nullptr; + ~ThumbDataManagement() + { + av_free(intermediateBuffer); + intermediateBuffer = nullptr; + av_frame_free(&frame_input); + frame_input = nullptr; + av_frame_free(&frame_temporary); + frame_temporary = nullptr; + avcodec_free_context(&avOutctx); + avOutctx = nullptr; + sws_freeContext(sws); + sws = nullptr; + } +}; + +// valid positions are including 0 (start of buffer) +// and bufferSize -1 last data point +static inline size_t Clamp(int64_t newPosition, size_t bufferSize) +{ + return std::min(std::max((int64_t) 0, newPosition), (int64_t) (bufferSize -1)); +} + +static int mem_file_read(void *h, uint8_t* buf, int size) +{ + if (size < 0) + return -1; + + MemBuffer* mbuf = static_cast<MemBuffer*>(h); + int64_t unread = mbuf->size - mbuf->pos; + if (unread <= 0) + return AVERROR_EOF; + + size_t tocopy = std::min((size_t)size, (size_t)unread); + memcpy(buf, mbuf->data + mbuf->pos, tocopy); + mbuf->pos += tocopy; + return static_cast<int>(tocopy); +} + +static int64_t mem_file_seek(void *h, int64_t pos, int whence) +{ + MemBuffer* mbuf = static_cast<MemBuffer*>(h); + if (whence == AVSEEK_SIZE) + return mbuf->size; + + // we want to ignore the AVSEEK_FORCE flag and therefore mask it away + whence &= ~AVSEEK_FORCE; + + if (whence == SEEK_SET) + { + mbuf->pos = Clamp(pos, mbuf->size); + } + else if (whence == SEEK_CUR) + { + mbuf->pos = Clamp(((int64_t)mbuf->pos) + pos, mbuf->size); + } + else + CLog::LogF(LOGERROR, "Unknown seek mode: {}", whence); + + return mbuf->pos; +} + +CFFmpegImage::CFFmpegImage(const std::string& strMimeType) : m_strMimeType(strMimeType) +{ + m_hasAlpha = false; + m_pFrame = nullptr; + m_outputBuffer = nullptr; +} + +CFFmpegImage::~CFFmpegImage() +{ + av_frame_free(&m_pFrame); + // someone could have forgotten to call us + CleanupLocalOutputBuffer(); + if (m_fctx) + { + avcodec_free_context(&m_codec_ctx); + avformat_close_input(&m_fctx); + } + if (m_ioctx) + FreeIOCtx(&m_ioctx); + + m_buf.data = nullptr; + m_buf.pos = 0; + m_buf.size = 0; +} + +bool CFFmpegImage::LoadImageFromMemory(unsigned char* buffer, unsigned int bufSize, + unsigned int width, unsigned int height) +{ + + if (!Initialize(buffer, bufSize)) + { + //log + return false; + } + + av_frame_free(&m_pFrame); + m_pFrame = ExtractFrame(); + + return !(m_pFrame == nullptr); +} + +bool CFFmpegImage::Initialize(unsigned char* buffer, size_t bufSize) +{ + int bufferSize = 4096; + uint8_t* fbuffer = (uint8_t*)av_malloc(bufferSize + AV_INPUT_BUFFER_PADDING_SIZE); + if (!fbuffer) + { + CLog::LogF(LOGERROR, "Could not allocate buffer"); + return false; + } + m_buf.data = buffer; + m_buf.size = bufSize; + m_buf.pos = 0; + + m_ioctx = avio_alloc_context(fbuffer, bufferSize, 0, &m_buf, + mem_file_read, NULL, mem_file_seek); + + if (!m_ioctx) + { + av_free(fbuffer); + CLog::LogF(LOGERROR, "Could not allocate AVIOContext"); + return false; + } + + // signal to ffmepg this is not streaming protocol + m_ioctx->max_packet_size = bufferSize; + + m_fctx = avformat_alloc_context(); + if (!m_fctx) + { + FreeIOCtx(&m_ioctx); + CLog::LogF(LOGERROR, "Could not allocate AVFormatContext"); + return false; + } + + m_fctx->pb = m_ioctx; + + // Some clients have pngs saved as jpeg or ask us for png but are jpeg + // mythv throws all mimetypes away and asks us with application/octet-stream + // this is poor man's fallback to at least identify png / jpeg + bool is_jpeg = (bufSize > 2 && buffer[0] == 0xFF && buffer[1] == 0xD8 && buffer[2] == 0xFF); + bool is_png = (bufSize > 3 && buffer[1] == 'P' && buffer[2] == 'N' && buffer[3] == 'G'); + bool is_tiff = (bufSize > 2 && buffer[0] == 'I' && buffer[1] == 'I' && buffer[2] == '*'); + + AVInputFormat* inp = nullptr; + if (is_jpeg) + inp = av_find_input_format("image2"); + else if (m_strMimeType == "image/apng") + inp = av_find_input_format("apng"); + else if (is_png) + inp = av_find_input_format("png_pipe"); + else if (is_tiff) + inp = av_find_input_format("tiff_pipe"); + else if (m_strMimeType == "image/jp2") + inp = av_find_input_format("j2k_pipe"); + else if (m_strMimeType == "image/webp") + inp = av_find_input_format("webp_pipe"); + // brute force parse if above check already failed + else if (m_strMimeType == "image/jpeg" || m_strMimeType == "image/jpg") + inp = av_find_input_format("image2"); + else if (m_strMimeType == "image/png") + inp = av_find_input_format("png_pipe"); + else if (m_strMimeType == "image/tiff") + inp = av_find_input_format("tiff_pipe"); + else if (m_strMimeType == "image/gif") + inp = av_find_input_format("gif"); + + if (avformat_open_input(&m_fctx, NULL, inp, NULL) < 0) + { + CLog::Log(LOGERROR, "Could not find suitable input format: {}", m_strMimeType); + avformat_close_input(&m_fctx); + FreeIOCtx(&m_ioctx); + return false; + } + + if (m_fctx->nb_streams <= 0) + { + avformat_close_input(&m_fctx); + FreeIOCtx(&m_ioctx); + return false; + } + AVCodecParameters* codec_params = m_fctx->streams[0]->codecpar; + AVCodec* codec = avcodec_find_decoder(codec_params->codec_id); + m_codec_ctx = avcodec_alloc_context3(codec); + if (!m_codec_ctx) + { + avformat_close_input(&m_fctx); + FreeIOCtx(&m_ioctx); + return false; + } + + if (avcodec_parameters_to_context(m_codec_ctx, codec_params) < 0) + { + avformat_close_input(&m_fctx); + avcodec_free_context(&m_codec_ctx); + FreeIOCtx(&m_ioctx); + return false; + } + + if (avcodec_open2(m_codec_ctx, codec, NULL) < 0) + { + avformat_close_input(&m_fctx); + avcodec_free_context(&m_codec_ctx); + FreeIOCtx(&m_ioctx); + return false; + } + + return true; +} + +AVFrame* CFFmpegImage::ExtractFrame() +{ + if (!m_fctx || !m_fctx->streams[0]) + { + CLog::LogF(LOGERROR, "No valid format context or stream"); + return nullptr; + } + + AVPacket pkt; + AVFrame* frame = av_frame_alloc(); + int frame_decoded = 0; + int ret = 0; + ret = av_read_frame(m_fctx, &pkt); + if (ret < 0) + { + CLog::Log(LOGDEBUG, "Error [{}] while reading frame: {}", ret, strerror(AVERROR(ret))); + av_frame_free(&frame); + av_packet_unref(&pkt); + return nullptr; + } + + ret = DecodeFFmpegFrame(m_codec_ctx, frame, &frame_decoded, &pkt); + if (ret < 0 || frame_decoded == 0 || !frame) + { + CLog::Log(LOGDEBUG, "Error [{}] while decoding frame: {}", ret, strerror(AVERROR(ret))); + av_frame_free(&frame); + av_packet_unref(&pkt); + return nullptr; + } + //we need milliseconds + frame->pkt_duration = av_rescale_q(frame->pkt_duration, m_fctx->streams[0]->time_base, AVRational{ 1, 1000 }); + m_height = frame->height; + m_width = frame->width; + m_originalWidth = m_width; + m_originalHeight = m_height; + + const AVPixFmtDescriptor* pixDescriptor = av_pix_fmt_desc_get(static_cast<AVPixelFormat>(frame->format)); + if (pixDescriptor && ((pixDescriptor->flags & (AV_PIX_FMT_FLAG_ALPHA | AV_PIX_FMT_FLAG_PAL)) != 0)) + m_hasAlpha = true; + + AVDictionary* dic = frame->metadata; + AVDictionaryEntry* entry = NULL; + if (dic) + { + entry = av_dict_get(dic, "Orientation", NULL, AV_DICT_MATCH_CASE); + if (entry && entry->value) + { + int orientation = atoi(entry->value); + // only values between including 0 and including 8 + // http://sylvana.net/jpegcrop/exif_orientation.html + if (orientation >= 0 && orientation <= 8) + m_orientation = (unsigned int)orientation; + } + } + av_packet_unref(&pkt); + + return frame; +} + +AVPixelFormat CFFmpegImage::ConvertFormats(AVFrame* frame) +{ + switch (frame->format) { + case AV_PIX_FMT_YUVJ420P: + return AV_PIX_FMT_YUV420P; + break; + case AV_PIX_FMT_YUVJ422P: + return AV_PIX_FMT_YUV422P; + break; + case AV_PIX_FMT_YUVJ444P: + return AV_PIX_FMT_YUV444P; + break; + case AV_PIX_FMT_YUVJ440P: + return AV_PIX_FMT_YUV440P; + default: + return static_cast<AVPixelFormat>(frame->format); + break; + } +} + +void CFFmpegImage::FreeIOCtx(AVIOContext** ioctx) +{ + av_freep(&((*ioctx)->buffer)); + av_freep(ioctx); +} + +bool CFFmpegImage::Decode(unsigned char * const pixels, unsigned int width, unsigned int height, + unsigned int pitch, unsigned int format) +{ + if (m_width == 0 || m_height == 0 || format != XB_FMT_A8R8G8B8) + return false; + + if (pixels == nullptr) + { + CLog::Log(LOGERROR, "{} - No valid buffer pointer (nullptr) passed", __FUNCTION__); + return false; + } + + if (!m_pFrame || !m_pFrame->data[0]) + { + CLog::LogF(LOGERROR, "AVFrame member not allocated"); + return false; + } + + return DecodeFrame(m_pFrame, width, height, pitch, pixels); +} + +int CFFmpegImage::EncodeFFmpegFrame(AVCodecContext *avctx, AVPacket *pkt, int *got_packet, AVFrame *frame) +{ + int ret; + + *got_packet = 0; + + ret = avcodec_send_frame(avctx, frame); + if (ret < 0) + return ret; + + ret = avcodec_receive_packet(avctx, pkt); + if (!ret) + *got_packet = 1; + + if (ret == AVERROR(EAGAIN)) + return 0; + + return ret; +} + +int CFFmpegImage::DecodeFFmpegFrame(AVCodecContext *avctx, AVFrame *frame, int *got_frame, AVPacket *pkt) +{ + int ret; + + *got_frame = 0; + + if (pkt) + { + ret = avcodec_send_packet(avctx, pkt); + // In particular, we don't expect AVERROR(EAGAIN), because we read all + // decoded frames with avcodec_receive_frame() until done. + if (ret < 0) + return ret; + } + + ret = avcodec_receive_frame(avctx, frame); + if (ret < 0 && ret != AVERROR(EAGAIN) && ret != AVERROR_EOF) + return ret; + if (ret >= 0) // static code analysers would complain + *got_frame = 1; + + return 0; +} + +bool CFFmpegImage::DecodeFrame(AVFrame* frame, unsigned int width, unsigned int height, unsigned int pitch, unsigned char * const pixels) +{ + if (pixels == nullptr) + { + CLog::Log(LOGERROR, "{} - No valid buffer pointer (nullptr) passed", __FUNCTION__); + return false; + } + + AVFrame* pictureRGB = av_frame_alloc(); + if (!pictureRGB) + { + CLog::LogF(LOGERROR, "AVFrame could not be allocated"); + return false; + } + + // we align on 16 as the input provided by the Texture also aligns the buffer size to 16 + int size = av_image_fill_arrays(pictureRGB->data, pictureRGB->linesize, NULL, AV_PIX_FMT_RGB32, width, height, 16); + if (size < 0) + { + CLog::LogF(LOGERROR, "Could not allocate AVFrame member with {} x {} pixels", width, height); + av_frame_free(&pictureRGB); + return false; + } + + bool needsCopy = false; + int pixelsSize = pitch * height; + bool aligned = (((uintptr_t)(const void *)(pixels)) % (32) == 0); + if (!aligned) + CLog::Log(LOGDEBUG, "Alignment of external buffer is not suitable for ffmpeg intrinsics - please fix your malloc"); + + if (aligned && size == pixelsSize && (int)pitch == pictureRGB->linesize[0]) + { + // We can use the pixels buffer directly + pictureRGB->data[0] = pixels; + } + else + { + // We need an extra buffer and copy it manually afterwards + pictureRGB->format = AV_PIX_FMT_RGB32; + pictureRGB->width = width; + pictureRGB->height = height; + // we copy the data manually later so give a chance to intrinsics (e.g. mmx, neon) + if (av_frame_get_buffer(pictureRGB, 32) < 0) + { + CLog::LogF(LOGERROR, "Could not allocate temp buffer of size {} bytes", size); + av_frame_free(&pictureRGB); + return false; + } + needsCopy = true; + } + + // Especially jpeg formats are full range this we need to take care here + // Input Formats like RGBA are handled correctly automatically + AVColorRange range = frame->color_range; + AVPixelFormat pixFormat = ConvertFormats(frame); + + // assumption quadratic maximums e.g. 2048x2048 + float ratio = m_width / (float)m_height; + unsigned int nHeight = m_originalHeight; + unsigned int nWidth = m_originalWidth; + if (nHeight > height) + { + nHeight = height; + nWidth = (unsigned int)(nHeight * ratio + 0.5f); + } + if (nWidth > width) + { + nWidth = width; + nHeight = (unsigned int)(nWidth / ratio + 0.5f); + } + + struct SwsContext* context = sws_getContext(m_originalWidth, m_originalHeight, pixFormat, + nWidth, nHeight, AV_PIX_FMT_RGB32, SWS_BICUBIC, NULL, NULL, NULL); + + if (range == AVCOL_RANGE_JPEG) + { + int* inv_table = nullptr; + int* table = nullptr; + int srcRange, dstRange, brightness, contrast, saturation; + sws_getColorspaceDetails(context, &inv_table, &srcRange, &table, &dstRange, &brightness, &contrast, &saturation); + srcRange = 1; + sws_setColorspaceDetails(context, inv_table, srcRange, table, dstRange, brightness, contrast, saturation); + } + + sws_scale(context, frame->data, frame->linesize, 0, m_originalHeight, + pictureRGB->data, pictureRGB->linesize); + sws_freeContext(context); + + if (needsCopy) + { + int minPitch = std::min((int)pitch, pictureRGB->linesize[0]); + if (minPitch < 0) + { + CLog::LogF(LOGERROR, "negative pitch or height"); + av_frame_free(&pictureRGB); + return false; + } + const unsigned char *src = pictureRGB->data[0]; + unsigned char* dst = pixels; + + for (unsigned int y = 0; y < nHeight; y++) + { + memcpy(dst, src, minPitch); + src += pictureRGB->linesize[0]; + dst += pitch; + } + av_frame_free(&pictureRGB); + } + else + { + // we only lended the data so don't get it deleted + pictureRGB->data[0] = nullptr; + av_frame_free(&pictureRGB); + } + + // update width and height original dimensions are kept + m_height = nHeight; + m_width = nWidth; + + return true; +} + +bool CFFmpegImage::CreateThumbnailFromSurface(unsigned char* bufferin, unsigned int width, + unsigned int height, unsigned int format, + unsigned int pitch, + const std::string& destFile, + unsigned char* &bufferout, + unsigned int &bufferoutSize) +{ + // It seems XB_FMT_A8R8G8B8 mean RGBA and not ARGB + if (format != XB_FMT_A8R8G8B8) + { + CLog::Log(LOGERROR, "Supplied format: {} is not supported.", format); + return false; + } + + bool jpg_output = false; + if (m_strMimeType == "image/jpeg" || m_strMimeType == "image/jpg") + jpg_output = true; + else if (m_strMimeType == "image/png") + jpg_output = false; + else + { + CLog::Log(LOGERROR, "Output Format is not supported: {} is not supported.", destFile); + return false; + } + + ThumbDataManagement tdm; + + tdm.codec = avcodec_find_encoder(jpg_output ? AV_CODEC_ID_MJPEG : AV_CODEC_ID_PNG); + if (!tdm.codec) + { + CLog::Log(LOGERROR, "You are missing a working encoder for format: {}", + jpg_output ? "JPEG" : "PNG"); + return false; + } + + tdm.avOutctx = avcodec_alloc_context3(tdm.codec); + if (!tdm.avOutctx) + { + CLog::Log(LOGERROR, "Could not allocate context for thumbnail: {}", destFile); + return false; + } + + tdm.avOutctx->height = height; + tdm.avOutctx->width = width; + tdm.avOutctx->time_base.num = 1; + tdm.avOutctx->time_base.den = 1; + tdm.avOutctx->pix_fmt = jpg_output ? AV_PIX_FMT_YUVJ420P : AV_PIX_FMT_RGBA; + tdm.avOutctx->flags = AV_CODEC_FLAG_QSCALE; + tdm.avOutctx->mb_lmin = tdm.avOutctx->qmin * FF_QP2LAMBDA; + tdm.avOutctx->mb_lmax = tdm.avOutctx->qmax * FF_QP2LAMBDA; + unsigned int quality = + CServiceBroker::GetSettingsComponent()->GetAdvancedSettings()->m_imageQualityJpeg; + tdm.avOutctx->global_quality = + jpg_output ? quality * FF_QP2LAMBDA : tdm.avOutctx->qmin * FF_QP2LAMBDA; + + unsigned int internalBufOutSize = 0; + + int size = av_image_get_buffer_size(tdm.avOutctx->pix_fmt, tdm.avOutctx->width, tdm.avOutctx->height, 16); + if (size < 0) + { + CLog::Log(LOGERROR, "Could not compute picture size for thumbnail: {}", destFile); + CleanupLocalOutputBuffer(); + return false; + } + internalBufOutSize = (unsigned int) size; + + tdm.intermediateBuffer = (uint8_t*) av_malloc(internalBufOutSize); + if (!tdm.intermediateBuffer) + { + CLog::Log(LOGERROR, "Could not allocate memory for thumbnail: {}", destFile); + CleanupLocalOutputBuffer(); + return false; + } + + if (avcodec_open2(tdm.avOutctx, tdm.codec, NULL) < 0) + { + CLog::Log(LOGERROR, "Could not open avcodec context thumbnail: {}", destFile); + CleanupLocalOutputBuffer(); + return false; + } + + tdm.frame_input = av_frame_alloc(); + if (!tdm.frame_input) + { + CLog::Log(LOGERROR, "Could not allocate frame for thumbnail: {}", destFile); + CleanupLocalOutputBuffer(); + return false; + } + + // convert the RGB32 frame to AV_PIX_FMT_YUV420P - we use this later on as AV_PIX_FMT_YUVJ420P + tdm.frame_temporary = av_frame_alloc(); + if (!tdm.frame_temporary) + { + CLog::Log(LOGERROR, "Could not allocate frame for thumbnail: {}", destFile); + CleanupLocalOutputBuffer(); + return false; + } + + if (av_image_fill_arrays(tdm.frame_temporary->data, tdm.frame_temporary->linesize, tdm.intermediateBuffer, jpg_output ? AV_PIX_FMT_YUV420P : AV_PIX_FMT_RGBA, width, height, 16) < 0) + { + CLog::Log(LOGERROR, "Could not fill picture for thumbnail: {}", destFile); + CleanupLocalOutputBuffer(); + return false; + } + + uint8_t* src[] = { bufferin, NULL, NULL, NULL }; + int srcStride[] = { (int) pitch, 0, 0, 0}; + + //input size == output size which means only pix_fmt conversion + tdm.sws = sws_getContext(width, height, AV_PIX_FMT_RGB32, width, height, jpg_output ? AV_PIX_FMT_YUV420P : AV_PIX_FMT_RGBA, 0, 0, 0, 0); + if (!tdm.sws) + { + CLog::Log(LOGERROR, "Could not setup scaling context for thumbnail: {}", destFile); + CleanupLocalOutputBuffer(); + return false; + } + + // Setup jpeg range for sws + if (jpg_output) + { + int* inv_table = nullptr; + int* table = nullptr; + int srcRange, dstRange, brightness, contrast, saturation; + + if (sws_getColorspaceDetails(tdm.sws, &inv_table, &srcRange, &table, &dstRange, &brightness, &contrast, &saturation) < 0) + { + CLog::Log(LOGERROR, "SWS_SCALE failed to get ColorSpaceDetails for thumbnail: {}", destFile); + CleanupLocalOutputBuffer(); + return false; + } + dstRange = 1; // jpeg full range yuv420p output + srcRange = 0; // full range RGB32 input + if (sws_setColorspaceDetails(tdm.sws, inv_table, srcRange, table, dstRange, brightness, contrast, saturation) < 0) + { + CLog::Log(LOGERROR, "SWS_SCALE failed to set ColorSpace Details for thumbnail: {}", destFile); + CleanupLocalOutputBuffer(); + return false; + } + } + + if (sws_scale(tdm.sws, src, srcStride, 0, height, tdm.frame_temporary->data, tdm.frame_temporary->linesize) < 0) + { + CLog::Log(LOGERROR, "SWS_SCALE failed for thumbnail: {}", destFile); + CleanupLocalOutputBuffer(); + return false; + } + tdm.frame_input->pts = 1; + tdm.frame_input->quality = tdm.avOutctx->global_quality; + tdm.frame_input->data[0] = tdm.frame_temporary->data[0]; + tdm.frame_input->data[1] = tdm.frame_temporary->data[1]; + tdm.frame_input->data[2] = tdm.frame_temporary->data[2]; + tdm.frame_input->height = height; + tdm.frame_input->width = width; + tdm.frame_input->linesize[0] = tdm.frame_temporary->linesize[0]; + tdm.frame_input->linesize[1] = tdm.frame_temporary->linesize[1]; + tdm.frame_input->linesize[2] = tdm.frame_temporary->linesize[2]; + // this is deprecated but mjpeg is not yet transitioned + tdm.frame_input->format = jpg_output ? AV_PIX_FMT_YUVJ420P : AV_PIX_FMT_RGBA; + + int got_package = 0; + AVPacket* avpkt; + avpkt = av_packet_alloc(); + + int ret = EncodeFFmpegFrame(tdm.avOutctx, avpkt, &got_package, tdm.frame_input); + + if ((ret < 0) || (got_package == 0)) + { + CLog::Log(LOGERROR, "Could not encode thumbnail: {}", destFile); + CleanupLocalOutputBuffer(); + av_packet_free(&avpkt); + return false; + } + + bufferoutSize = avpkt->size; + m_outputBuffer = (uint8_t*) av_malloc(bufferoutSize); + if (!m_outputBuffer) + { + CLog::Log(LOGERROR, "Could not generate allocate memory for thumbnail: {}", destFile); + CleanupLocalOutputBuffer(); + av_packet_free(&avpkt); + return false; + } + // update buffer ptr for caller + bufferout = m_outputBuffer; + + // copy avpkt data into outputbuffer + memcpy(m_outputBuffer, avpkt->data, avpkt->size); + av_packet_free(&avpkt); + + return true; +} + +void CFFmpegImage::ReleaseThumbnailBuffer() +{ + CleanupLocalOutputBuffer(); +} + +void CFFmpegImage::CleanupLocalOutputBuffer() +{ + av_free(m_outputBuffer); + m_outputBuffer = nullptr; +} + +std::shared_ptr<Frame> CFFmpegImage::ReadFrame() +{ + AVFrame* avframe = ExtractFrame(); + if (avframe == nullptr) + return nullptr; + std::shared_ptr<Frame> frame(new Frame()); + frame->m_delay = (unsigned int)avframe->pkt_duration; + frame->m_pitch = avframe->width * 4; + frame->m_pImage = (unsigned char*) av_malloc(avframe->height * frame->m_pitch); + DecodeFrame(avframe, avframe->width, avframe->height, frame->m_pitch, frame->m_pImage); + av_frame_free(&avframe); + return frame; +} |