summaryrefslogtreecommitdiffstats
path: root/xbmc/guilib/FFmpegImage.cpp
diff options
context:
space:
mode:
authorDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-10 18:07:22 +0000
committerDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-10 18:07:22 +0000
commitc04dcc2e7d834218ef2d4194331e383402495ae1 (patch)
tree7333e38d10d75386e60f336b80c2443c1166031d /xbmc/guilib/FFmpegImage.cpp
parentInitial commit. (diff)
downloadkodi-c04dcc2e7d834218ef2d4194331e383402495ae1.tar.xz
kodi-c04dcc2e7d834218ef2d4194331e383402495ae1.zip
Adding upstream version 2:20.4+dfsg.upstream/2%20.4+dfsg
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'xbmc/guilib/FFmpegImage.cpp')
-rw-r--r--xbmc/guilib/FFmpegImage.cpp754
1 files changed, 754 insertions, 0 deletions
diff --git a/xbmc/guilib/FFmpegImage.cpp b/xbmc/guilib/FFmpegImage.cpp
new file mode 100644
index 0000000..fa25527
--- /dev/null
+++ b/xbmc/guilib/FFmpegImage.cpp
@@ -0,0 +1,754 @@
+/*
+ * Copyright (C) 2012-2018 Team Kodi
+ * This file is part of Kodi - https://kodi.tv
+ *
+ * SPDX-License-Identifier: GPL-2.0-or-later
+ * See LICENSES/README.md for more information.
+ */
+#include "FFmpegImage.h"
+
+#include "cores/FFmpeg.h"
+#include "guilib/Texture.h"
+#include "settings/AdvancedSettings.h"
+#include "settings/Settings.h"
+#include "settings/SettingsComponent.h"
+#include "utils/log.h"
+
+#include <algorithm>
+
+extern "C"
+{
+#include <libavformat/avformat.h>
+#include <libavutil/imgutils.h>
+#include <libavcodec/avcodec.h>
+#include <libavutil/avutil.h>
+#include <libswscale/swscale.h>
+#include <libavutil/pixdesc.h>
+}
+
+Frame::Frame(const Frame& src) :
+ m_delay(src.m_delay),
+ m_imageSize(src.m_imageSize),
+ m_height(src.m_height),
+ m_width(src.m_width)
+{
+ if (src.m_pImage)
+ {
+ m_pImage = (unsigned char*) av_malloc(m_imageSize);
+ memcpy(m_pImage, src.m_pImage, m_imageSize);
+ }
+}
+
+Frame::~Frame()
+{
+ av_free(m_pImage);
+ m_pImage = nullptr;
+}
+
+struct ThumbDataManagement
+{
+ uint8_t* intermediateBuffer = nullptr; // gets av_alloced
+ AVFrame* frame_input = nullptr;
+ AVFrame* frame_temporary = nullptr;
+ SwsContext* sws = nullptr;
+ AVCodecContext* avOutctx = nullptr;
+ AVCodec* codec = nullptr;
+ ~ThumbDataManagement()
+ {
+ av_free(intermediateBuffer);
+ intermediateBuffer = nullptr;
+ av_frame_free(&frame_input);
+ frame_input = nullptr;
+ av_frame_free(&frame_temporary);
+ frame_temporary = nullptr;
+ avcodec_free_context(&avOutctx);
+ avOutctx = nullptr;
+ sws_freeContext(sws);
+ sws = nullptr;
+ }
+};
+
+// valid positions are including 0 (start of buffer)
+// and bufferSize -1 last data point
+static inline size_t Clamp(int64_t newPosition, size_t bufferSize)
+{
+ return std::min(std::max((int64_t) 0, newPosition), (int64_t) (bufferSize -1));
+}
+
+static int mem_file_read(void *h, uint8_t* buf, int size)
+{
+ if (size < 0)
+ return -1;
+
+ MemBuffer* mbuf = static_cast<MemBuffer*>(h);
+ int64_t unread = mbuf->size - mbuf->pos;
+ if (unread <= 0)
+ return AVERROR_EOF;
+
+ size_t tocopy = std::min((size_t)size, (size_t)unread);
+ memcpy(buf, mbuf->data + mbuf->pos, tocopy);
+ mbuf->pos += tocopy;
+ return static_cast<int>(tocopy);
+}
+
+static int64_t mem_file_seek(void *h, int64_t pos, int whence)
+{
+ MemBuffer* mbuf = static_cast<MemBuffer*>(h);
+ if (whence == AVSEEK_SIZE)
+ return mbuf->size;
+
+ // we want to ignore the AVSEEK_FORCE flag and therefore mask it away
+ whence &= ~AVSEEK_FORCE;
+
+ if (whence == SEEK_SET)
+ {
+ mbuf->pos = Clamp(pos, mbuf->size);
+ }
+ else if (whence == SEEK_CUR)
+ {
+ mbuf->pos = Clamp(((int64_t)mbuf->pos) + pos, mbuf->size);
+ }
+ else
+ CLog::LogF(LOGERROR, "Unknown seek mode: {}", whence);
+
+ return mbuf->pos;
+}
+
+CFFmpegImage::CFFmpegImage(const std::string& strMimeType) : m_strMimeType(strMimeType)
+{
+ m_hasAlpha = false;
+ m_pFrame = nullptr;
+ m_outputBuffer = nullptr;
+}
+
+CFFmpegImage::~CFFmpegImage()
+{
+ av_frame_free(&m_pFrame);
+ // someone could have forgotten to call us
+ CleanupLocalOutputBuffer();
+ if (m_fctx)
+ {
+ avcodec_free_context(&m_codec_ctx);
+ avformat_close_input(&m_fctx);
+ }
+ if (m_ioctx)
+ FreeIOCtx(&m_ioctx);
+
+ m_buf.data = nullptr;
+ m_buf.pos = 0;
+ m_buf.size = 0;
+}
+
+bool CFFmpegImage::LoadImageFromMemory(unsigned char* buffer, unsigned int bufSize,
+ unsigned int width, unsigned int height)
+{
+
+ if (!Initialize(buffer, bufSize))
+ {
+ //log
+ return false;
+ }
+
+ av_frame_free(&m_pFrame);
+ m_pFrame = ExtractFrame();
+
+ return !(m_pFrame == nullptr);
+}
+
+bool CFFmpegImage::Initialize(unsigned char* buffer, size_t bufSize)
+{
+ int bufferSize = 4096;
+ uint8_t* fbuffer = (uint8_t*)av_malloc(bufferSize + AV_INPUT_BUFFER_PADDING_SIZE);
+ if (!fbuffer)
+ {
+ CLog::LogF(LOGERROR, "Could not allocate buffer");
+ return false;
+ }
+ m_buf.data = buffer;
+ m_buf.size = bufSize;
+ m_buf.pos = 0;
+
+ m_ioctx = avio_alloc_context(fbuffer, bufferSize, 0, &m_buf,
+ mem_file_read, NULL, mem_file_seek);
+
+ if (!m_ioctx)
+ {
+ av_free(fbuffer);
+ CLog::LogF(LOGERROR, "Could not allocate AVIOContext");
+ return false;
+ }
+
+ // signal to ffmepg this is not streaming protocol
+ m_ioctx->max_packet_size = bufferSize;
+
+ m_fctx = avformat_alloc_context();
+ if (!m_fctx)
+ {
+ FreeIOCtx(&m_ioctx);
+ CLog::LogF(LOGERROR, "Could not allocate AVFormatContext");
+ return false;
+ }
+
+ m_fctx->pb = m_ioctx;
+
+ // Some clients have pngs saved as jpeg or ask us for png but are jpeg
+ // mythv throws all mimetypes away and asks us with application/octet-stream
+ // this is poor man's fallback to at least identify png / jpeg
+ bool is_jpeg = (bufSize > 2 && buffer[0] == 0xFF && buffer[1] == 0xD8 && buffer[2] == 0xFF);
+ bool is_png = (bufSize > 3 && buffer[1] == 'P' && buffer[2] == 'N' && buffer[3] == 'G');
+ bool is_tiff = (bufSize > 2 && buffer[0] == 'I' && buffer[1] == 'I' && buffer[2] == '*');
+
+ AVInputFormat* inp = nullptr;
+ if (is_jpeg)
+ inp = av_find_input_format("image2");
+ else if (m_strMimeType == "image/apng")
+ inp = av_find_input_format("apng");
+ else if (is_png)
+ inp = av_find_input_format("png_pipe");
+ else if (is_tiff)
+ inp = av_find_input_format("tiff_pipe");
+ else if (m_strMimeType == "image/jp2")
+ inp = av_find_input_format("j2k_pipe");
+ else if (m_strMimeType == "image/webp")
+ inp = av_find_input_format("webp_pipe");
+ // brute force parse if above check already failed
+ else if (m_strMimeType == "image/jpeg" || m_strMimeType == "image/jpg")
+ inp = av_find_input_format("image2");
+ else if (m_strMimeType == "image/png")
+ inp = av_find_input_format("png_pipe");
+ else if (m_strMimeType == "image/tiff")
+ inp = av_find_input_format("tiff_pipe");
+ else if (m_strMimeType == "image/gif")
+ inp = av_find_input_format("gif");
+
+ if (avformat_open_input(&m_fctx, NULL, inp, NULL) < 0)
+ {
+ CLog::Log(LOGERROR, "Could not find suitable input format: {}", m_strMimeType);
+ avformat_close_input(&m_fctx);
+ FreeIOCtx(&m_ioctx);
+ return false;
+ }
+
+ if (m_fctx->nb_streams <= 0)
+ {
+ avformat_close_input(&m_fctx);
+ FreeIOCtx(&m_ioctx);
+ return false;
+ }
+ AVCodecParameters* codec_params = m_fctx->streams[0]->codecpar;
+ AVCodec* codec = avcodec_find_decoder(codec_params->codec_id);
+ m_codec_ctx = avcodec_alloc_context3(codec);
+ if (!m_codec_ctx)
+ {
+ avformat_close_input(&m_fctx);
+ FreeIOCtx(&m_ioctx);
+ return false;
+ }
+
+ if (avcodec_parameters_to_context(m_codec_ctx, codec_params) < 0)
+ {
+ avformat_close_input(&m_fctx);
+ avcodec_free_context(&m_codec_ctx);
+ FreeIOCtx(&m_ioctx);
+ return false;
+ }
+
+ if (avcodec_open2(m_codec_ctx, codec, NULL) < 0)
+ {
+ avformat_close_input(&m_fctx);
+ avcodec_free_context(&m_codec_ctx);
+ FreeIOCtx(&m_ioctx);
+ return false;
+ }
+
+ return true;
+}
+
+AVFrame* CFFmpegImage::ExtractFrame()
+{
+ if (!m_fctx || !m_fctx->streams[0])
+ {
+ CLog::LogF(LOGERROR, "No valid format context or stream");
+ return nullptr;
+ }
+
+ AVPacket pkt;
+ AVFrame* frame = av_frame_alloc();
+ int frame_decoded = 0;
+ int ret = 0;
+ ret = av_read_frame(m_fctx, &pkt);
+ if (ret < 0)
+ {
+ CLog::Log(LOGDEBUG, "Error [{}] while reading frame: {}", ret, strerror(AVERROR(ret)));
+ av_frame_free(&frame);
+ av_packet_unref(&pkt);
+ return nullptr;
+ }
+
+ ret = DecodeFFmpegFrame(m_codec_ctx, frame, &frame_decoded, &pkt);
+ if (ret < 0 || frame_decoded == 0 || !frame)
+ {
+ CLog::Log(LOGDEBUG, "Error [{}] while decoding frame: {}", ret, strerror(AVERROR(ret)));
+ av_frame_free(&frame);
+ av_packet_unref(&pkt);
+ return nullptr;
+ }
+ //we need milliseconds
+ frame->pkt_duration = av_rescale_q(frame->pkt_duration, m_fctx->streams[0]->time_base, AVRational{ 1, 1000 });
+ m_height = frame->height;
+ m_width = frame->width;
+ m_originalWidth = m_width;
+ m_originalHeight = m_height;
+
+ const AVPixFmtDescriptor* pixDescriptor = av_pix_fmt_desc_get(static_cast<AVPixelFormat>(frame->format));
+ if (pixDescriptor && ((pixDescriptor->flags & (AV_PIX_FMT_FLAG_ALPHA | AV_PIX_FMT_FLAG_PAL)) != 0))
+ m_hasAlpha = true;
+
+ AVDictionary* dic = frame->metadata;
+ AVDictionaryEntry* entry = NULL;
+ if (dic)
+ {
+ entry = av_dict_get(dic, "Orientation", NULL, AV_DICT_MATCH_CASE);
+ if (entry && entry->value)
+ {
+ int orientation = atoi(entry->value);
+ // only values between including 0 and including 8
+ // http://sylvana.net/jpegcrop/exif_orientation.html
+ if (orientation >= 0 && orientation <= 8)
+ m_orientation = (unsigned int)orientation;
+ }
+ }
+ av_packet_unref(&pkt);
+
+ return frame;
+}
+
+AVPixelFormat CFFmpegImage::ConvertFormats(AVFrame* frame)
+{
+ switch (frame->format) {
+ case AV_PIX_FMT_YUVJ420P:
+ return AV_PIX_FMT_YUV420P;
+ break;
+ case AV_PIX_FMT_YUVJ422P:
+ return AV_PIX_FMT_YUV422P;
+ break;
+ case AV_PIX_FMT_YUVJ444P:
+ return AV_PIX_FMT_YUV444P;
+ break;
+ case AV_PIX_FMT_YUVJ440P:
+ return AV_PIX_FMT_YUV440P;
+ default:
+ return static_cast<AVPixelFormat>(frame->format);
+ break;
+ }
+}
+
+void CFFmpegImage::FreeIOCtx(AVIOContext** ioctx)
+{
+ av_freep(&((*ioctx)->buffer));
+ av_freep(ioctx);
+}
+
+bool CFFmpegImage::Decode(unsigned char * const pixels, unsigned int width, unsigned int height,
+ unsigned int pitch, unsigned int format)
+{
+ if (m_width == 0 || m_height == 0 || format != XB_FMT_A8R8G8B8)
+ return false;
+
+ if (pixels == nullptr)
+ {
+ CLog::Log(LOGERROR, "{} - No valid buffer pointer (nullptr) passed", __FUNCTION__);
+ return false;
+ }
+
+ if (!m_pFrame || !m_pFrame->data[0])
+ {
+ CLog::LogF(LOGERROR, "AVFrame member not allocated");
+ return false;
+ }
+
+ return DecodeFrame(m_pFrame, width, height, pitch, pixels);
+}
+
+int CFFmpegImage::EncodeFFmpegFrame(AVCodecContext *avctx, AVPacket *pkt, int *got_packet, AVFrame *frame)
+{
+ int ret;
+
+ *got_packet = 0;
+
+ ret = avcodec_send_frame(avctx, frame);
+ if (ret < 0)
+ return ret;
+
+ ret = avcodec_receive_packet(avctx, pkt);
+ if (!ret)
+ *got_packet = 1;
+
+ if (ret == AVERROR(EAGAIN))
+ return 0;
+
+ return ret;
+}
+
+int CFFmpegImage::DecodeFFmpegFrame(AVCodecContext *avctx, AVFrame *frame, int *got_frame, AVPacket *pkt)
+{
+ int ret;
+
+ *got_frame = 0;
+
+ if (pkt)
+ {
+ ret = avcodec_send_packet(avctx, pkt);
+ // In particular, we don't expect AVERROR(EAGAIN), because we read all
+ // decoded frames with avcodec_receive_frame() until done.
+ if (ret < 0)
+ return ret;
+ }
+
+ ret = avcodec_receive_frame(avctx, frame);
+ if (ret < 0 && ret != AVERROR(EAGAIN) && ret != AVERROR_EOF)
+ return ret;
+ if (ret >= 0) // static code analysers would complain
+ *got_frame = 1;
+
+ return 0;
+}
+
+bool CFFmpegImage::DecodeFrame(AVFrame* frame, unsigned int width, unsigned int height, unsigned int pitch, unsigned char * const pixels)
+{
+ if (pixels == nullptr)
+ {
+ CLog::Log(LOGERROR, "{} - No valid buffer pointer (nullptr) passed", __FUNCTION__);
+ return false;
+ }
+
+ AVFrame* pictureRGB = av_frame_alloc();
+ if (!pictureRGB)
+ {
+ CLog::LogF(LOGERROR, "AVFrame could not be allocated");
+ return false;
+ }
+
+ // we align on 16 as the input provided by the Texture also aligns the buffer size to 16
+ int size = av_image_fill_arrays(pictureRGB->data, pictureRGB->linesize, NULL, AV_PIX_FMT_RGB32, width, height, 16);
+ if (size < 0)
+ {
+ CLog::LogF(LOGERROR, "Could not allocate AVFrame member with {} x {} pixels", width, height);
+ av_frame_free(&pictureRGB);
+ return false;
+ }
+
+ bool needsCopy = false;
+ int pixelsSize = pitch * height;
+ bool aligned = (((uintptr_t)(const void *)(pixels)) % (32) == 0);
+ if (!aligned)
+ CLog::Log(LOGDEBUG, "Alignment of external buffer is not suitable for ffmpeg intrinsics - please fix your malloc");
+
+ if (aligned && size == pixelsSize && (int)pitch == pictureRGB->linesize[0])
+ {
+ // We can use the pixels buffer directly
+ pictureRGB->data[0] = pixels;
+ }
+ else
+ {
+ // We need an extra buffer and copy it manually afterwards
+ pictureRGB->format = AV_PIX_FMT_RGB32;
+ pictureRGB->width = width;
+ pictureRGB->height = height;
+ // we copy the data manually later so give a chance to intrinsics (e.g. mmx, neon)
+ if (av_frame_get_buffer(pictureRGB, 32) < 0)
+ {
+ CLog::LogF(LOGERROR, "Could not allocate temp buffer of size {} bytes", size);
+ av_frame_free(&pictureRGB);
+ return false;
+ }
+ needsCopy = true;
+ }
+
+ // Especially jpeg formats are full range this we need to take care here
+ // Input Formats like RGBA are handled correctly automatically
+ AVColorRange range = frame->color_range;
+ AVPixelFormat pixFormat = ConvertFormats(frame);
+
+ // assumption quadratic maximums e.g. 2048x2048
+ float ratio = m_width / (float)m_height;
+ unsigned int nHeight = m_originalHeight;
+ unsigned int nWidth = m_originalWidth;
+ if (nHeight > height)
+ {
+ nHeight = height;
+ nWidth = (unsigned int)(nHeight * ratio + 0.5f);
+ }
+ if (nWidth > width)
+ {
+ nWidth = width;
+ nHeight = (unsigned int)(nWidth / ratio + 0.5f);
+ }
+
+ struct SwsContext* context = sws_getContext(m_originalWidth, m_originalHeight, pixFormat,
+ nWidth, nHeight, AV_PIX_FMT_RGB32, SWS_BICUBIC, NULL, NULL, NULL);
+
+ if (range == AVCOL_RANGE_JPEG)
+ {
+ int* inv_table = nullptr;
+ int* table = nullptr;
+ int srcRange, dstRange, brightness, contrast, saturation;
+ sws_getColorspaceDetails(context, &inv_table, &srcRange, &table, &dstRange, &brightness, &contrast, &saturation);
+ srcRange = 1;
+ sws_setColorspaceDetails(context, inv_table, srcRange, table, dstRange, brightness, contrast, saturation);
+ }
+
+ sws_scale(context, frame->data, frame->linesize, 0, m_originalHeight,
+ pictureRGB->data, pictureRGB->linesize);
+ sws_freeContext(context);
+
+ if (needsCopy)
+ {
+ int minPitch = std::min((int)pitch, pictureRGB->linesize[0]);
+ if (minPitch < 0)
+ {
+ CLog::LogF(LOGERROR, "negative pitch or height");
+ av_frame_free(&pictureRGB);
+ return false;
+ }
+ const unsigned char *src = pictureRGB->data[0];
+ unsigned char* dst = pixels;
+
+ for (unsigned int y = 0; y < nHeight; y++)
+ {
+ memcpy(dst, src, minPitch);
+ src += pictureRGB->linesize[0];
+ dst += pitch;
+ }
+ av_frame_free(&pictureRGB);
+ }
+ else
+ {
+ // we only lended the data so don't get it deleted
+ pictureRGB->data[0] = nullptr;
+ av_frame_free(&pictureRGB);
+ }
+
+ // update width and height original dimensions are kept
+ m_height = nHeight;
+ m_width = nWidth;
+
+ return true;
+}
+
+bool CFFmpegImage::CreateThumbnailFromSurface(unsigned char* bufferin, unsigned int width,
+ unsigned int height, unsigned int format,
+ unsigned int pitch,
+ const std::string& destFile,
+ unsigned char* &bufferout,
+ unsigned int &bufferoutSize)
+{
+ // It seems XB_FMT_A8R8G8B8 mean RGBA and not ARGB
+ if (format != XB_FMT_A8R8G8B8)
+ {
+ CLog::Log(LOGERROR, "Supplied format: {} is not supported.", format);
+ return false;
+ }
+
+ bool jpg_output = false;
+ if (m_strMimeType == "image/jpeg" || m_strMimeType == "image/jpg")
+ jpg_output = true;
+ else if (m_strMimeType == "image/png")
+ jpg_output = false;
+ else
+ {
+ CLog::Log(LOGERROR, "Output Format is not supported: {} is not supported.", destFile);
+ return false;
+ }
+
+ ThumbDataManagement tdm;
+
+ tdm.codec = avcodec_find_encoder(jpg_output ? AV_CODEC_ID_MJPEG : AV_CODEC_ID_PNG);
+ if (!tdm.codec)
+ {
+ CLog::Log(LOGERROR, "You are missing a working encoder for format: {}",
+ jpg_output ? "JPEG" : "PNG");
+ return false;
+ }
+
+ tdm.avOutctx = avcodec_alloc_context3(tdm.codec);
+ if (!tdm.avOutctx)
+ {
+ CLog::Log(LOGERROR, "Could not allocate context for thumbnail: {}", destFile);
+ return false;
+ }
+
+ tdm.avOutctx->height = height;
+ tdm.avOutctx->width = width;
+ tdm.avOutctx->time_base.num = 1;
+ tdm.avOutctx->time_base.den = 1;
+ tdm.avOutctx->pix_fmt = jpg_output ? AV_PIX_FMT_YUVJ420P : AV_PIX_FMT_RGBA;
+ tdm.avOutctx->flags = AV_CODEC_FLAG_QSCALE;
+ tdm.avOutctx->mb_lmin = tdm.avOutctx->qmin * FF_QP2LAMBDA;
+ tdm.avOutctx->mb_lmax = tdm.avOutctx->qmax * FF_QP2LAMBDA;
+ unsigned int quality =
+ CServiceBroker::GetSettingsComponent()->GetAdvancedSettings()->m_imageQualityJpeg;
+ tdm.avOutctx->global_quality =
+ jpg_output ? quality * FF_QP2LAMBDA : tdm.avOutctx->qmin * FF_QP2LAMBDA;
+
+ unsigned int internalBufOutSize = 0;
+
+ int size = av_image_get_buffer_size(tdm.avOutctx->pix_fmt, tdm.avOutctx->width, tdm.avOutctx->height, 16);
+ if (size < 0)
+ {
+ CLog::Log(LOGERROR, "Could not compute picture size for thumbnail: {}", destFile);
+ CleanupLocalOutputBuffer();
+ return false;
+ }
+ internalBufOutSize = (unsigned int) size;
+
+ tdm.intermediateBuffer = (uint8_t*) av_malloc(internalBufOutSize);
+ if (!tdm.intermediateBuffer)
+ {
+ CLog::Log(LOGERROR, "Could not allocate memory for thumbnail: {}", destFile);
+ CleanupLocalOutputBuffer();
+ return false;
+ }
+
+ if (avcodec_open2(tdm.avOutctx, tdm.codec, NULL) < 0)
+ {
+ CLog::Log(LOGERROR, "Could not open avcodec context thumbnail: {}", destFile);
+ CleanupLocalOutputBuffer();
+ return false;
+ }
+
+ tdm.frame_input = av_frame_alloc();
+ if (!tdm.frame_input)
+ {
+ CLog::Log(LOGERROR, "Could not allocate frame for thumbnail: {}", destFile);
+ CleanupLocalOutputBuffer();
+ return false;
+ }
+
+ // convert the RGB32 frame to AV_PIX_FMT_YUV420P - we use this later on as AV_PIX_FMT_YUVJ420P
+ tdm.frame_temporary = av_frame_alloc();
+ if (!tdm.frame_temporary)
+ {
+ CLog::Log(LOGERROR, "Could not allocate frame for thumbnail: {}", destFile);
+ CleanupLocalOutputBuffer();
+ return false;
+ }
+
+ if (av_image_fill_arrays(tdm.frame_temporary->data, tdm.frame_temporary->linesize, tdm.intermediateBuffer, jpg_output ? AV_PIX_FMT_YUV420P : AV_PIX_FMT_RGBA, width, height, 16) < 0)
+ {
+ CLog::Log(LOGERROR, "Could not fill picture for thumbnail: {}", destFile);
+ CleanupLocalOutputBuffer();
+ return false;
+ }
+
+ uint8_t* src[] = { bufferin, NULL, NULL, NULL };
+ int srcStride[] = { (int) pitch, 0, 0, 0};
+
+ //input size == output size which means only pix_fmt conversion
+ tdm.sws = sws_getContext(width, height, AV_PIX_FMT_RGB32, width, height, jpg_output ? AV_PIX_FMT_YUV420P : AV_PIX_FMT_RGBA, 0, 0, 0, 0);
+ if (!tdm.sws)
+ {
+ CLog::Log(LOGERROR, "Could not setup scaling context for thumbnail: {}", destFile);
+ CleanupLocalOutputBuffer();
+ return false;
+ }
+
+ // Setup jpeg range for sws
+ if (jpg_output)
+ {
+ int* inv_table = nullptr;
+ int* table = nullptr;
+ int srcRange, dstRange, brightness, contrast, saturation;
+
+ if (sws_getColorspaceDetails(tdm.sws, &inv_table, &srcRange, &table, &dstRange, &brightness, &contrast, &saturation) < 0)
+ {
+ CLog::Log(LOGERROR, "SWS_SCALE failed to get ColorSpaceDetails for thumbnail: {}", destFile);
+ CleanupLocalOutputBuffer();
+ return false;
+ }
+ dstRange = 1; // jpeg full range yuv420p output
+ srcRange = 0; // full range RGB32 input
+ if (sws_setColorspaceDetails(tdm.sws, inv_table, srcRange, table, dstRange, brightness, contrast, saturation) < 0)
+ {
+ CLog::Log(LOGERROR, "SWS_SCALE failed to set ColorSpace Details for thumbnail: {}", destFile);
+ CleanupLocalOutputBuffer();
+ return false;
+ }
+ }
+
+ if (sws_scale(tdm.sws, src, srcStride, 0, height, tdm.frame_temporary->data, tdm.frame_temporary->linesize) < 0)
+ {
+ CLog::Log(LOGERROR, "SWS_SCALE failed for thumbnail: {}", destFile);
+ CleanupLocalOutputBuffer();
+ return false;
+ }
+ tdm.frame_input->pts = 1;
+ tdm.frame_input->quality = tdm.avOutctx->global_quality;
+ tdm.frame_input->data[0] = tdm.frame_temporary->data[0];
+ tdm.frame_input->data[1] = tdm.frame_temporary->data[1];
+ tdm.frame_input->data[2] = tdm.frame_temporary->data[2];
+ tdm.frame_input->height = height;
+ tdm.frame_input->width = width;
+ tdm.frame_input->linesize[0] = tdm.frame_temporary->linesize[0];
+ tdm.frame_input->linesize[1] = tdm.frame_temporary->linesize[1];
+ tdm.frame_input->linesize[2] = tdm.frame_temporary->linesize[2];
+ // this is deprecated but mjpeg is not yet transitioned
+ tdm.frame_input->format = jpg_output ? AV_PIX_FMT_YUVJ420P : AV_PIX_FMT_RGBA;
+
+ int got_package = 0;
+ AVPacket* avpkt;
+ avpkt = av_packet_alloc();
+
+ int ret = EncodeFFmpegFrame(tdm.avOutctx, avpkt, &got_package, tdm.frame_input);
+
+ if ((ret < 0) || (got_package == 0))
+ {
+ CLog::Log(LOGERROR, "Could not encode thumbnail: {}", destFile);
+ CleanupLocalOutputBuffer();
+ av_packet_free(&avpkt);
+ return false;
+ }
+
+ bufferoutSize = avpkt->size;
+ m_outputBuffer = (uint8_t*) av_malloc(bufferoutSize);
+ if (!m_outputBuffer)
+ {
+ CLog::Log(LOGERROR, "Could not generate allocate memory for thumbnail: {}", destFile);
+ CleanupLocalOutputBuffer();
+ av_packet_free(&avpkt);
+ return false;
+ }
+ // update buffer ptr for caller
+ bufferout = m_outputBuffer;
+
+ // copy avpkt data into outputbuffer
+ memcpy(m_outputBuffer, avpkt->data, avpkt->size);
+ av_packet_free(&avpkt);
+
+ return true;
+}
+
+void CFFmpegImage::ReleaseThumbnailBuffer()
+{
+ CleanupLocalOutputBuffer();
+}
+
+void CFFmpegImage::CleanupLocalOutputBuffer()
+{
+ av_free(m_outputBuffer);
+ m_outputBuffer = nullptr;
+}
+
+std::shared_ptr<Frame> CFFmpegImage::ReadFrame()
+{
+ AVFrame* avframe = ExtractFrame();
+ if (avframe == nullptr)
+ return nullptr;
+ std::shared_ptr<Frame> frame(new Frame());
+ frame->m_delay = (unsigned int)avframe->pkt_duration;
+ frame->m_pitch = avframe->width * 4;
+ frame->m_pImage = (unsigned char*) av_malloc(avframe->height * frame->m_pitch);
+ DecodeFrame(avframe, avframe->width, avframe->height, frame->m_pitch, frame->m_pImage);
+ av_frame_free(&avframe);
+ return frame;
+}