summaryrefslogtreecommitdiffstats
path: root/xbmc/addons/Scraper.cpp
diff options
context:
space:
mode:
authorDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-10 18:07:22 +0000
committerDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-10 18:07:22 +0000
commitc04dcc2e7d834218ef2d4194331e383402495ae1 (patch)
tree7333e38d10d75386e60f336b80c2443c1166031d /xbmc/addons/Scraper.cpp
parentInitial commit. (diff)
downloadkodi-c04dcc2e7d834218ef2d4194331e383402495ae1.tar.xz
kodi-c04dcc2e7d834218ef2d4194331e383402495ae1.zip
Adding upstream version 2:20.4+dfsg.upstream/2%20.4+dfsg
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to '')
-rw-r--r--xbmc/addons/Scraper.cpp1470
1 files changed, 1470 insertions, 0 deletions
diff --git a/xbmc/addons/Scraper.cpp b/xbmc/addons/Scraper.cpp
new file mode 100644
index 0000000..14c555f
--- /dev/null
+++ b/xbmc/addons/Scraper.cpp
@@ -0,0 +1,1470 @@
+/*
+ * Copyright (C) 2005-2018 Team Kodi
+ * This file is part of Kodi - https://kodi.tv
+ *
+ * SPDX-License-Identifier: GPL-2.0-or-later
+ * See LICENSES/README.md for more information.
+ */
+
+#include "Scraper.h"
+
+#include "FileItem.h"
+#include "ServiceBroker.h"
+#include "URL.h"
+#include "Util.h"
+#include "addons/AddonManager.h"
+#include "addons/addoninfo/AddonInfo.h"
+#include "addons/addoninfo/AddonType.h"
+#include "addons/settings/AddonSettings.h"
+#include "filesystem/CurlFile.h"
+#include "filesystem/Directory.h"
+#include "filesystem/File.h"
+#include "filesystem/PluginDirectory.h"
+#include "guilib/LocalizeStrings.h"
+#include "music/Album.h"
+#include "music/Artist.h"
+#include "music/MusicDatabase.h"
+#include "music/infoscanner/MusicAlbumInfo.h"
+#include "music/infoscanner/MusicArtistInfo.h"
+#include "settings/AdvancedSettings.h"
+#include "settings/SettingsComponent.h"
+#include "settings/SettingsValueFlatJsonSerializer.h"
+#include "utils/CharsetConverter.h"
+#include "utils/ScraperParser.h"
+#include "utils/ScraperUrl.h"
+#include "utils/StringUtils.h"
+#include "utils/URIUtils.h"
+#include "utils/XMLUtils.h"
+#include "utils/log.h"
+#include "video/VideoDatabase.h"
+
+#include <algorithm>
+#include <sstream>
+
+#include <fstrcmp.h>
+
+using namespace XFILE;
+using namespace MUSIC_GRABBER;
+using namespace VIDEO;
+
+namespace ADDON
+{
+
+typedef struct
+{
+ const char *name;
+ CONTENT_TYPE type;
+ int pretty;
+} ContentMapping;
+
+static const ContentMapping content[] = {{"unknown", CONTENT_NONE, 231},
+ {"albums", CONTENT_ALBUMS, 132},
+ {"music", CONTENT_ALBUMS, 132},
+ {"artists", CONTENT_ARTISTS, 133},
+ {"movies", CONTENT_MOVIES, 20342},
+ {"tvshows", CONTENT_TVSHOWS, 20343},
+ {"musicvideos", CONTENT_MUSICVIDEOS, 20389}};
+
+std::string TranslateContent(const CONTENT_TYPE &type, bool pretty /*=false*/)
+{
+ for (const ContentMapping& map : content)
+ {
+ if (type == map.type)
+ {
+ if (pretty && map.pretty)
+ return g_localizeStrings.Get(map.pretty);
+ else
+ return map.name;
+ }
+ }
+ return "";
+}
+
+CONTENT_TYPE TranslateContent(const std::string &string)
+{
+ for (const ContentMapping& map : content)
+ {
+ if (string == map.name)
+ return map.type;
+ }
+ return CONTENT_NONE;
+}
+
+AddonType ScraperTypeFromContent(const CONTENT_TYPE& content)
+{
+ switch (content)
+ {
+ case CONTENT_ALBUMS:
+ return AddonType::SCRAPER_ALBUMS;
+ case CONTENT_ARTISTS:
+ return AddonType::SCRAPER_ARTISTS;
+ case CONTENT_MOVIES:
+ return AddonType::SCRAPER_MOVIES;
+ case CONTENT_MUSICVIDEOS:
+ return AddonType::SCRAPER_MUSICVIDEOS;
+ case CONTENT_TVSHOWS:
+ return AddonType::SCRAPER_TVSHOWS;
+ default:
+ return AddonType::UNKNOWN;
+ }
+}
+
+// if the XML root is <error>, throw CScraperError with enclosed <title>/<message> values
+static void CheckScraperError(const TiXmlElement *pxeRoot)
+{
+ if (!pxeRoot || StringUtils::CompareNoCase(pxeRoot->Value(), "error"))
+ return;
+ std::string sTitle;
+ std::string sMessage;
+ XMLUtils::GetString(pxeRoot, "title", sTitle);
+ XMLUtils::GetString(pxeRoot, "message", sMessage);
+ throw CScraperError(sTitle, sMessage);
+}
+
+CScraper::CScraper(const AddonInfoPtr& addonInfo, AddonType addonType)
+ : CAddon(addonInfo, addonType),
+ m_fLoaded(false),
+ m_requiressettings(false),
+ m_pathContent(CONTENT_NONE)
+{
+ m_requiressettings = addonInfo->Type(addonType)->GetValue("@requiressettings").asBoolean();
+
+ CDateTimeSpan persistence;
+ std::string tmp = addonInfo->Type(addonType)->GetValue("@cachepersistence").asString();
+ if (!tmp.empty())
+ m_persistence.SetFromTimeString(tmp);
+
+ switch (addonType)
+ {
+ case AddonType::SCRAPER_ALBUMS:
+ m_pathContent = CONTENT_ALBUMS;
+ break;
+ case AddonType::SCRAPER_ARTISTS:
+ m_pathContent = CONTENT_ARTISTS;
+ break;
+ case AddonType::SCRAPER_MOVIES:
+ m_pathContent = CONTENT_MOVIES;
+ break;
+ case AddonType::SCRAPER_MUSICVIDEOS:
+ m_pathContent = CONTENT_MUSICVIDEOS;
+ break;
+ case AddonType::SCRAPER_TVSHOWS:
+ m_pathContent = CONTENT_TVSHOWS;
+ break;
+ default:
+ break;
+ }
+
+ m_isPython = URIUtils::GetExtension(addonInfo->Type(addonType)->LibPath()) == ".py";
+}
+
+bool CScraper::Supports(const CONTENT_TYPE &content) const
+{
+ return Type() == ScraperTypeFromContent(content);
+}
+
+bool CScraper::SetPathSettings(CONTENT_TYPE content, const std::string &xml)
+{
+ m_pathContent = content;
+ if (!LoadSettings(false, false))
+ return false;
+
+ if (xml.empty())
+ return true;
+
+ CXBMCTinyXML doc;
+ doc.Parse(xml);
+ return SettingsFromXML(doc, false);
+}
+
+std::string CScraper::GetPathSettings()
+{
+ if (!LoadSettings(false, true))
+ return "";
+
+ std::stringstream stream;
+ CXBMCTinyXML doc;
+ SettingsToXML(doc);
+ if (doc.RootElement())
+ stream << *doc.RootElement();
+
+ return stream.str();
+}
+
+void CScraper::ClearCache()
+{
+ std::string strCachePath = URIUtils::AddFileToFolder(CServiceBroker::GetSettingsComponent()->GetAdvancedSettings()->m_cachePath, "scrapers");
+
+ // create scraper cache dir if needed
+ if (!CDirectory::Exists(strCachePath))
+ CDirectory::Create(strCachePath);
+
+ strCachePath = URIUtils::AddFileToFolder(strCachePath, ID());
+ URIUtils::AddSlashAtEnd(strCachePath);
+
+ if (CDirectory::Exists(strCachePath))
+ {
+ CFileItemList items;
+ CDirectory::GetDirectory(strCachePath, items, "", DIR_FLAG_DEFAULTS);
+ for (int i = 0; i < items.Size(); ++i)
+ {
+ // wipe cache
+ if (items[i]->m_dateTime + m_persistence <= CDateTime::GetCurrentDateTime())
+ CFile::Delete(items[i]->GetDynPath());
+ }
+ }
+ else
+ CDirectory::Create(strCachePath);
+}
+
+// returns a vector of strings: the first is the XML output by the function; the rest
+// is XML output by chained functions, possibly recursively
+// the CCurlFile object is passed in so that URL fetches can be canceled from other threads
+// throws CScraperError abort on internal failures (e.g., parse errors)
+std::vector<std::string> CScraper::Run(const std::string &function,
+ const CScraperUrl &scrURL,
+ CCurlFile &http,
+ const std::vector<std::string> *extras)
+{
+ if (!Load())
+ throw CScraperError();
+
+ std::string strXML = InternalRun(function, scrURL, http, extras);
+ if (strXML.empty())
+ {
+ if (function != "NfoUrl" && function != "ResolveIDToUrl")
+ CLog::Log(LOGERROR, "{}: Unable to parse web site", __FUNCTION__);
+ throw CScraperError();
+ }
+
+ CLog::Log(LOGDEBUG, "scraper: {} returned {}", function, strXML);
+
+ CXBMCTinyXML doc;
+ /* all data was converted to UTF-8 before being processed by scraper */
+ doc.Parse(strXML, TIXML_ENCODING_UTF8);
+ if (!doc.RootElement())
+ {
+ CLog::Log(LOGERROR, "{}: Unable to parse XML", __FUNCTION__);
+ throw CScraperError();
+ }
+
+ std::vector<std::string> result;
+ result.push_back(strXML);
+ TiXmlElement *xchain = doc.RootElement()->FirstChildElement();
+ // skip children of the root element until <url> or <chain>
+ while (xchain && strcmp(xchain->Value(), "url") && strcmp(xchain->Value(), "chain"))
+ xchain = xchain->NextSiblingElement();
+ while (xchain)
+ {
+ // <chain|url function="...">param</>
+ const char *szFunction = xchain->Attribute("function");
+ if (szFunction)
+ {
+ CScraperUrl scrURL2;
+ std::vector<std::string> extras;
+ // for <chain>, pass the contained text as a parameter; for <url>, as URL content
+ if (strcmp(xchain->Value(), "chain") == 0)
+ {
+ if (xchain->FirstChild())
+ extras.emplace_back(xchain->FirstChild()->Value());
+ }
+ else
+ scrURL2.ParseAndAppendUrl(xchain);
+ // Fix for empty chains. $$1 would still contain the
+ // previous value as there is no child of the xml node.
+ // since $$1 will always either contain the data from an
+ // url or the parameters to a chain, we can safely clear it here
+ // to fix this issue
+ m_parser.m_param[0].clear();
+ std::vector<std::string> result2 = RunNoThrow(szFunction, scrURL2, http, &extras);
+ result.insert(result.end(), result2.begin(), result2.end());
+ }
+ xchain = xchain->NextSiblingElement();
+ // continue to skip past non-<url> or <chain> elements
+ while (xchain && strcmp(xchain->Value(), "url") && strcmp(xchain->Value(), "chain"))
+ xchain = xchain->NextSiblingElement();
+ }
+
+ return result;
+}
+
+// just like Run, but returns an empty list instead of throwing in case of error
+// don't use in new code; errors should be handled appropriately
+std::vector<std::string> CScraper::RunNoThrow(const std::string &function,
+ const CScraperUrl &url,
+ XFILE::CCurlFile &http,
+ const std::vector<std::string> *extras)
+{
+ std::vector<std::string> vcs;
+ try
+ {
+ vcs = Run(function, url, http, extras);
+ }
+ catch (const CScraperError &sce)
+ {
+ assert(sce.FAborted()); // the only kind we should get
+ }
+ return vcs;
+}
+
+std::string CScraper::InternalRun(const std::string &function,
+ const CScraperUrl &scrURL,
+ CCurlFile &http,
+ const std::vector<std::string> *extras)
+{
+ // walk the list of input URLs and fetch each into parser parameters
+ const auto& urls = scrURL.GetUrls();
+ size_t i;
+ for (i = 0; i < urls.size(); ++i)
+ {
+ if (!CScraperUrl::Get(urls[i], m_parser.m_param[i], http, ID()) ||
+ m_parser.m_param[i].empty())
+ return "";
+ }
+ // put the 'extra' parameters into the parser parameter list too
+ if (extras)
+ {
+ for (size_t j = 0; j < extras->size(); ++j)
+ m_parser.m_param[j + i] = (*extras)[j];
+ }
+
+ return m_parser.Parse(function, this);
+}
+
+std::string CScraper::GetPathSettingsAsJSON()
+{
+ static const std::string EmptyPathSettings = "{}";
+
+ if (!LoadSettings(false, true))
+ return EmptyPathSettings;
+
+ CSettingsValueFlatJsonSerializer jsonSerializer;
+ auto json = jsonSerializer.SerializeValues(GetSettings()->GetSettingsManager());
+ if (json.empty())
+ return EmptyPathSettings;
+
+ return json;
+}
+
+bool CScraper::Load()
+{
+ if (m_fLoaded || m_isPython)
+ return true;
+
+ bool result = m_parser.Load(LibPath());
+ if (result)
+ {
+ //! @todo this routine assumes that deps are a single level, and assumes the dep is installed.
+ //! 1. Does it make sense to have recursive dependencies?
+ //! 2. Should we be checking the dep versions or do we assume it is ok?
+ auto deps = GetDependencies();
+ auto itr = deps.begin();
+ while (itr != deps.end())
+ {
+ if (itr->id == "xbmc.metadata")
+ {
+ ++itr;
+ continue;
+ }
+ AddonPtr dep;
+
+ bool bOptional = itr->optional;
+
+ if (CServiceBroker::GetAddonMgr().GetAddon((*itr).id, dep, ADDON::OnlyEnabled::CHOICE_YES))
+ {
+ CXBMCTinyXML doc;
+ if (dep->Type() == AddonType::SCRAPER_LIBRARY && doc.LoadFile(dep->LibPath()))
+ m_parser.AddDocument(&doc);
+ }
+ else
+ {
+ if (!bOptional)
+ {
+ result = false;
+ break;
+ }
+ }
+ ++itr;
+ }
+ }
+
+ if (!result)
+ CLog::Log(LOGWARNING, "failed to load scraper XML from {}", LibPath());
+ return m_fLoaded = result;
+}
+
+bool CScraper::IsInUse() const
+{
+ if (Supports(CONTENT_ALBUMS) || Supports(CONTENT_ARTISTS))
+ { // music scraper
+ CMusicDatabase db;
+ if (db.Open() && db.ScraperInUse(ID()))
+ return true;
+ }
+ else
+ { // video scraper
+ CVideoDatabase db;
+ if (db.Open() && db.ScraperInUse(ID()))
+ return true;
+ }
+ return false;
+}
+
+bool CScraper::IsNoop()
+{
+ if (!Load())
+ throw CScraperError();
+
+ return !m_isPython && m_parser.IsNoop();
+}
+
+// pass in contents of .nfo file; returns URL (possibly empty if none found)
+// and may populate strId, or throws CScraperError on error
+CScraperUrl CScraper::NfoUrl(const std::string &sNfoContent)
+{
+ CScraperUrl scurlRet;
+
+ if (IsNoop())
+ return scurlRet;
+
+ if (m_isPython)
+ {
+ std::stringstream str;
+ str << "plugin://" << ID() << "?action=NfoUrl&nfo=" << CURL::Encode(sNfoContent)
+ << "&pathSettings=" << CURL::Encode(GetPathSettingsAsJSON());
+
+ CFileItemList items;
+ if (!XFILE::CDirectory::GetDirectory(str.str(), items, "", DIR_FLAG_DEFAULTS))
+ return scurlRet;
+
+ if (items.Size() == 0)
+ return scurlRet;
+ if (items.Size() > 1)
+ CLog::Log(LOGWARNING, "{}: scraper returned multiple results; using first", __FUNCTION__);
+
+ CScraperUrl::SUrlEntry surl;
+ surl.m_type = CScraperUrl::UrlType::General;
+ surl.m_url = items[0]->GetDynPath();
+ scurlRet.AppendUrl(surl);
+ return scurlRet;
+ }
+
+ // scraper function takes contents of .nfo file, returns XML (see below)
+ std::vector<std::string> vcsIn;
+ vcsIn.push_back(sNfoContent);
+ CScraperUrl scurl;
+ CCurlFile fcurl;
+ std::vector<std::string> vcsOut = Run("NfoUrl", scurl, fcurl, &vcsIn);
+ if (vcsOut.empty() || vcsOut[0].empty())
+ return scurlRet;
+ if (vcsOut.size() > 1)
+ CLog::Log(LOGWARNING, "{}: scraper returned multiple results; using first", __FUNCTION__);
+
+ // parse returned XML: either <error> element on error, blank on failure,
+ // or <url>...</url> or <url>...</url><id>...</id> on success
+ for (size_t i = 0; i < vcsOut.size(); ++i)
+ {
+ CXBMCTinyXML doc;
+ doc.Parse(vcsOut[i], TIXML_ENCODING_UTF8);
+ CheckScraperError(doc.RootElement());
+
+ if (doc.RootElement())
+ {
+ /*
+ NOTE: Scrapers might return invalid xml with some loose
+ elements (eg. '<url>http://some.url</url><id>123</id>').
+ Since XMLUtils::GetString() is assuming well formed xml
+ with start and end-tags we're not able to use it.
+ Check for the desired Elements instead.
+ */
+ TiXmlElement* pxeUrl = nullptr;
+ TiXmlElement* pId = nullptr;
+ if (!strcmp(doc.RootElement()->Value(), "details"))
+ {
+ pxeUrl = doc.RootElement()->FirstChildElement("url");
+ pId = doc.RootElement()->FirstChildElement("id");
+ }
+ else
+ {
+ pId = doc.FirstChildElement("id");
+ pxeUrl = doc.FirstChildElement("url");
+ }
+ if (pId && pId->FirstChild())
+ scurlRet.SetId(pId->FirstChild()->ValueStr());
+
+ if (pxeUrl && pxeUrl->Attribute("function"))
+ continue;
+
+ if (pxeUrl)
+ scurlRet.ParseAndAppendUrl(pxeUrl);
+ else if (!strcmp(doc.RootElement()->Value(), "url"))
+ scurlRet.ParseAndAppendUrl(doc.RootElement());
+ else
+ continue;
+ break;
+ }
+ }
+ return scurlRet;
+}
+
+CScraperUrl CScraper::ResolveIDToUrl(const std::string &externalID)
+{
+ CScraperUrl scurlRet;
+
+ if (m_isPython)
+ {
+ std::stringstream str;
+ str << "plugin://" << ID() << "?action=resolveid&key=" << CURL::Encode(externalID)
+ << "&pathSettings=" << CURL::Encode(GetPathSettingsAsJSON());
+
+ CFileItem item("resolve me", false);
+
+ if (XFILE::CPluginDirectory::GetPluginResult(str.str(), item, false))
+ scurlRet.ParseFromData(item.GetDynPath());
+
+ return scurlRet;
+ }
+
+ // scraper function takes an external ID, returns XML (see below)
+ std::vector<std::string> vcsIn;
+ vcsIn.push_back(externalID);
+ CScraperUrl scurl;
+ CCurlFile fcurl;
+ std::vector<std::string> vcsOut = Run("ResolveIDToUrl", scurl, fcurl, &vcsIn);
+ if (vcsOut.empty() || vcsOut[0].empty())
+ return scurlRet;
+ if (vcsOut.size() > 1)
+ CLog::Log(LOGWARNING, "{}: scraper returned multiple results; using first", __FUNCTION__);
+
+ // parse returned XML: either <error> element on error, blank on failure,
+ // or <url>...</url> or <url>...</url><id>...</id> on success
+ for (size_t i = 0; i < vcsOut.size(); ++i)
+ {
+ CXBMCTinyXML doc;
+ doc.Parse(vcsOut[i], TIXML_ENCODING_UTF8);
+ CheckScraperError(doc.RootElement());
+
+ if (doc.RootElement())
+ {
+ /*
+ NOTE: Scrapers might return invalid xml with some loose
+ elements (eg. '<url>http://some.url</url><id>123</id>').
+ Since XMLUtils::GetString() is assuming well formed xml
+ with start and end-tags we're not able to use it.
+ Check for the desired Elements instead.
+ */
+ TiXmlElement* pxeUrl = nullptr;
+ TiXmlElement* pId = nullptr;
+ if (!strcmp(doc.RootElement()->Value(), "details"))
+ {
+ pxeUrl = doc.RootElement()->FirstChildElement("url");
+ pId = doc.RootElement()->FirstChildElement("id");
+ }
+ else
+ {
+ pId = doc.FirstChildElement("id");
+ pxeUrl = doc.FirstChildElement("url");
+ }
+ if (pId && pId->FirstChild())
+ scurlRet.SetId(pId->FirstChild()->ValueStr());
+
+ if (pxeUrl && pxeUrl->Attribute("function"))
+ continue;
+
+ if (pxeUrl)
+ scurlRet.ParseAndAppendUrl(pxeUrl);
+ else if (!strcmp(doc.RootElement()->Value(), "url"))
+ scurlRet.ParseAndAppendUrl(doc.RootElement());
+ else
+ continue;
+ break;
+ }
+ }
+ return scurlRet;
+}
+
+static bool RelevanceSortFunction(const CScraperUrl &left, const CScraperUrl &right)
+{
+ return left.GetRelevance() > right.GetRelevance();
+}
+
+template<class T>
+static T FromFileItem(const CFileItem &item);
+
+template<>
+CScraperUrl FromFileItem<CScraperUrl>(const CFileItem &item)
+{
+ CScraperUrl url;
+
+ url.SetTitle(item.GetLabel());
+ if (item.HasProperty("relevance"))
+ url.SetRelevance(item.GetProperty("relevance").asDouble());
+ CScraperUrl::SUrlEntry surl;
+ surl.m_type = CScraperUrl::UrlType::General;
+ surl.m_url = item.GetDynPath();
+ url.AppendUrl(surl);
+
+ return url;
+}
+
+template<>
+CMusicAlbumInfo FromFileItem<CMusicAlbumInfo>(const CFileItem &item)
+{
+ CMusicAlbumInfo info;
+ const std::string& sTitle = item.GetLabel();
+ std::string sArtist = item.GetProperty("album.artist").asString();
+ std::string sAlbumName;
+ if (!sArtist.empty())
+ sAlbumName = StringUtils::Format("{} - {}", sArtist, sTitle);
+ else
+ sAlbumName = sTitle;
+
+ CScraperUrl url;
+ url.AppendUrl(CScraperUrl::SUrlEntry(item.GetDynPath()));
+
+ info = CMusicAlbumInfo(sTitle, sArtist, sAlbumName, url);
+ if (item.HasProperty("relevance"))
+ info.SetRelevance(item.GetProperty("relevance").asFloat());
+
+ if (item.HasProperty("album.releasestatus"))
+ info.GetAlbum().strReleaseStatus = item.GetProperty("album.releasestatus").asString();
+ if (item.HasProperty("album.type"))
+ info.GetAlbum().strType = item.GetProperty("album.type").asString();
+ if (item.HasProperty("album.year"))
+ info.GetAlbum().strReleaseDate = item.GetProperty("album.year").asString();
+ if (item.HasProperty("album.label"))
+ info.GetAlbum().strLabel = item.GetProperty("album.label").asString();
+ info.GetAlbum().art = item.GetArt();
+
+ return info;
+}
+
+template<>
+CMusicArtistInfo FromFileItem<CMusicArtistInfo>(const CFileItem &item)
+{
+ CMusicArtistInfo info;
+ const std::string& sTitle = item.GetLabel();
+
+ CScraperUrl url;
+ url.AppendUrl(CScraperUrl::SUrlEntry(item.GetDynPath()));
+
+ info = CMusicArtistInfo(sTitle, url);
+ if (item.HasProperty("artist.genre"))
+ info.GetArtist().genre = StringUtils::Split(item.GetProperty("artist.genre").asString(),
+ CServiceBroker::GetSettingsComponent()->GetAdvancedSettings()->m_musicItemSeparator);
+ if (item.HasProperty("artist.disambiguation"))
+ info.GetArtist().strDisambiguation = item.GetProperty("artist.disambiguation").asString();
+ if (item.HasProperty("artist.type"))
+ info.GetArtist().strType = item.GetProperty("artist.type").asString();
+ if (item.HasProperty("artist.gender"))
+ info.GetArtist().strGender = item.GetProperty("artist.gender").asString();
+ if (item.HasProperty("artist.born"))
+ info.GetArtist().strBorn = item.GetProperty("artist.born").asString();
+
+ return info;
+}
+
+template<class T>
+static std::vector<T> PythonFind(const std::string &ID,
+ const std::map<std::string, std::string> &additionals)
+{
+ std::vector<T> result;
+ CFileItemList items;
+ std::stringstream str;
+ str << "plugin://" << ID << "?action=find";
+ for (const auto &it : additionals)
+ str << "&" << it.first << "=" << CURL::Encode(it.second);
+
+ if (XFILE::CDirectory::GetDirectory(str.str(), items, "", DIR_FLAG_DEFAULTS))
+ {
+ for (const auto& it : items)
+ result.emplace_back(std::move(FromFileItem<T>(*it)));
+ }
+
+ return result;
+}
+
+static std::string FromString(const CFileItem &item, const std::string &key)
+{
+ return item.GetProperty(key).asString();
+}
+
+static std::vector<std::string> FromArray(const CFileItem &item, const std::string &key, int sep)
+{
+ return StringUtils::Split(item.GetProperty(key).asString(),
+ sep ? CServiceBroker::GetSettingsComponent()->GetAdvancedSettings()->m_videoItemSeparator
+ : CServiceBroker::GetSettingsComponent()->GetAdvancedSettings()->m_musicItemSeparator);
+}
+
+static void ParseThumbs(CScraperUrl &scurl,
+ const CFileItem &item,
+ int nThumbs,
+ const std::string &tag)
+{
+ for (int i = 0; i < nThumbs; ++i)
+ {
+ std::stringstream prefix;
+ prefix << tag << i + 1;
+ std::string url = FromString(item, prefix.str() + ".url");
+ std::string aspect = FromString(item, prefix.str() + ".aspect");
+ std::string preview = FromString(item, prefix.str() + ".preview");
+ scurl.AddParsedUrl(url, aspect, preview);
+ }
+}
+
+static std::string ParseFanart(const CFileItem &item, int nFanart, const std::string &tag)
+{
+ std::string result;
+ TiXmlElement fanart("fanart");
+ for (int i = 0; i < nFanart; ++i)
+ {
+ std::stringstream prefix;
+ prefix << tag << i + 1;
+ std::string url = FromString(item, prefix.str() + ".url");
+ std::string preview = FromString(item, prefix.str() + ".preview");
+ TiXmlElement thumb("thumb");
+ thumb.SetAttribute("preview", preview);
+ TiXmlText text(url);
+ thumb.InsertEndChild(text);
+ fanart.InsertEndChild(thumb);
+ }
+ result << fanart;
+
+ return result;
+}
+
+template<class T>
+static void DetailsFromFileItem(const CFileItem &, T &);
+
+template<>
+void DetailsFromFileItem<CAlbum>(const CFileItem &item, CAlbum &album)
+{
+ album.strAlbum = item.GetLabel();
+ album.strMusicBrainzAlbumID = FromString(item, "album.musicbrainzid");
+ album.strReleaseGroupMBID = FromString(item, "album.releasegroupid");
+
+ int nArtists = item.GetProperty("album.artists").asInteger32();
+ album.artistCredits.reserve(nArtists);
+ for (int i = 0; i < nArtists; ++i)
+ {
+ std::stringstream prefix;
+ prefix << "album.artist" << i + 1;
+ CArtistCredit artistCredit;
+ artistCredit.SetArtist(FromString(item, prefix.str() + ".name"));
+ artistCredit.SetMusicBrainzArtistID(FromString(item, prefix.str() + ".musicbrainzid"));
+ album.artistCredits.push_back(artistCredit);
+ }
+
+ album.strArtistDesc = FromString(item, "album.artist_description");
+ album.genre = FromArray(item, "album.genre", 0);
+ album.styles = FromArray(item, "album.styles", 0);
+ album.moods = FromArray(item, "album.moods", 0);
+ album.themes = FromArray(item, "album.themes", 0);
+ album.bCompilation = item.GetProperty("album.compilation").asBoolean();
+ album.strReview = FromString(item, "album.review");
+ album.strReleaseDate = FromString(item, "album.releasedate");
+ if (album.strReleaseDate.empty())
+ album.strReleaseDate = FromString(item, "album.year");
+ album.strOrigReleaseDate = FromString(item, "album.originaldate");
+ album.strLabel = FromString(item, "album.label");
+ album.strType = FromString(item, "album.type");
+ album.strReleaseStatus = FromString(item, "album.releasestatus");
+ album.fRating = item.GetProperty("album.rating").asFloat();
+ album.iUserrating = item.GetProperty("album.user_rating").asInteger32();
+ album.iVotes = item.GetProperty("album.votes").asInteger32();
+
+ /* Scrapers fetch a list of possible art but do not set the current images used because art
+ selection depends on other preferences so is handled by CMusicInfoScanner
+ album.art = item.GetArt();
+ */
+
+ int nThumbs = item.GetProperty("album.thumbs").asInteger32();
+ ParseThumbs(album.thumbURL, item, nThumbs, "album.thumb");
+}
+
+template<>
+void DetailsFromFileItem<CArtist>(const CFileItem &item, CArtist &artist)
+{
+ artist.strArtist = item.GetLabel();
+ artist.strMusicBrainzArtistID = FromString(item, "artist.musicbrainzid");
+ artist.strDisambiguation = FromString(item, "artist.disambiguation");
+ artist.strType = FromString(item, "artist.type");
+ artist.strGender = FromString(item, "artist.gender");
+ artist.genre = FromArray(item, "artist.genre", 0);
+ artist.styles = FromArray(item, "artist.styles", 0);
+ artist.moods = FromArray(item, "artist.moods", 0);
+ artist.yearsActive = FromArray(item, "artist.years_active", 0);
+ artist.instruments = FromArray(item, "artist.instruments", 0);
+ artist.strBorn = FromString(item, "artist.born");
+ artist.strFormed = FromString(item, "artist.formed");
+ artist.strBiography = FromString(item, "artist.biography");
+ artist.strDied = FromString(item, "artist.died");
+ artist.strDisbanded = FromString(item, "artist.disbanded");
+
+ /* Scrapers fetch a list of possible art but do not set the current images used because art
+ selection depends on other preferences so is handled by CMusicInfoScanner
+ artist.art = item.GetArt();
+ */
+
+ int nAlbums = item.GetProperty("artist.albums").asInteger32();
+ artist.discography.reserve(nAlbums);
+ for (int i = 0; i < nAlbums; ++i)
+ {
+ std::stringstream prefix;
+ prefix << "artist.album" << i + 1;
+ CDiscoAlbum discoAlbum;
+ discoAlbum.strAlbum = FromString(item, prefix.str() + ".title");
+ discoAlbum.strYear = FromString(item, prefix.str() + ".year");
+ discoAlbum.strReleaseGroupMBID = FromString(item, prefix.str() + ".musicbrainzreleasegroupid");
+ artist.discography.emplace_back(discoAlbum);
+ }
+
+ int nThumbs = item.GetProperty("artist.thumbs").asInteger32();
+ ParseThumbs(artist.thumbURL, item, nThumbs, "artist.thumb");
+
+ // Support deprecated fanarts property, add to artist.thumbURL
+ int nFanart = item.GetProperty("artist.fanarts").asInteger32();
+ if (nFanart > 0)
+ {
+ CFanart fanart;
+ fanart.m_xml = ParseFanart(item, nFanart, "artist.fanart");
+ fanart.Unpack();
+ for (unsigned int i = 0; i < fanart.GetNumFanarts(); i++)
+ artist.thumbURL.AddParsedUrl(fanart.GetImageURL(i), "fanart", fanart.GetPreviewURL(i));
+ }
+}
+
+template<>
+void DetailsFromFileItem<CVideoInfoTag>(const CFileItem &item, CVideoInfoTag &tag)
+{
+ if (item.HasVideoInfoTag())
+ tag = *item.GetVideoInfoTag();
+}
+
+template<class T>
+static bool PythonDetails(const std::string &ID,
+ const std::string &key,
+ const std::string &url,
+ const std::string &action,
+ const std::string &pathSettings,
+ T &result)
+{
+ std::stringstream str;
+ str << "plugin://" << ID << "?action=" << action << "&" << key << "=" << CURL::Encode(url);
+ str << "&pathSettings=" << CURL::Encode(pathSettings);
+
+ CFileItem item(url, false);
+
+ if (!XFILE::CPluginDirectory::GetPluginResult(str.str(), item, false))
+ return false;
+
+ DetailsFromFileItem(item, result);
+ return true;
+}
+
+// fetch list of matching movies sorted by relevance (may be empty);
+// throws CScraperError on error; first called with fFirst set, then unset if first try fails
+std::vector<CScraperUrl> CScraper::FindMovie(XFILE::CCurlFile &fcurl,
+ const std::string &movieTitle, int movieYear,
+ bool fFirst)
+{
+ // prepare parameters for URL creation
+ std::string sTitle, sYear;
+ if (movieYear < 0)
+ {
+ std::string sTitleYear;
+ CUtil::CleanString(movieTitle, sTitle, sTitleYear, sYear, true /*fRemoveExt*/, fFirst);
+ }
+ else
+ {
+ sTitle = movieTitle;
+ sYear = std::to_string( movieYear );
+ }
+
+ CLog::Log(LOGDEBUG,
+ "{}: Searching for '{}' using {} scraper "
+ "(path: '{}', content: '{}', version: '{}')",
+ __FUNCTION__, sTitle, Name(), Path(), ADDON::TranslateContent(Content()),
+ Version().asString());
+
+ std::vector<CScraperUrl> vcscurl;
+ if (IsNoop())
+ return vcscurl;
+
+ if (!fFirst)
+ StringUtils::Replace(sTitle, '-', ' ');
+
+ if (m_isPython)
+ {
+ std::map<std::string, std::string> additionals{{"title", sTitle}};
+ if (!sYear.empty())
+ additionals.insert({"year", sYear});
+ additionals.emplace("pathSettings", GetPathSettingsAsJSON());
+ return PythonFind<CScraperUrl>(ID(), additionals);
+ }
+
+ std::vector<std::string> vcsIn(1);
+ g_charsetConverter.utf8To(SearchStringEncoding(), sTitle, vcsIn[0]);
+ vcsIn[0] = CURL::Encode(vcsIn[0]);
+ if (fFirst && !sYear.empty())
+ vcsIn.push_back(sYear);
+
+ // request a search URL from the title/filename/etc.
+ CScraperUrl scurl;
+ std::vector<std::string> vcsOut = Run("CreateSearchUrl", scurl, fcurl, &vcsIn);
+ if (vcsOut.empty())
+ {
+ CLog::Log(LOGDEBUG, "{}: CreateSearchUrl failed", __FUNCTION__);
+ throw CScraperError();
+ }
+ scurl.ParseFromData(vcsOut[0]);
+
+ // do the search, and parse the result into a list
+ vcsIn.clear();
+ vcsIn.push_back(scurl.GetFirstThumbUrl());
+ vcsOut = Run("GetSearchResults", scurl, fcurl, &vcsIn);
+
+ bool fSort(true);
+ std::set<std::string> stsDupeCheck;
+ bool fResults(false);
+ for (std::vector<std::string>::const_iterator i = vcsOut.begin(); i != vcsOut.end(); ++i)
+ {
+ CXBMCTinyXML doc;
+ doc.Parse(*i, TIXML_ENCODING_UTF8);
+ if (!doc.RootElement())
+ {
+ CLog::Log(LOGERROR, "{}: Unable to parse XML", __FUNCTION__);
+ continue; // might have more valid results later
+ }
+
+ CheckScraperError(doc.RootElement());
+
+ TiXmlHandle xhDoc(&doc);
+ TiXmlHandle xhResults = xhDoc.FirstChild("results");
+ if (!xhResults.Element())
+ continue;
+ fResults = true; // even if empty
+
+ // we need to sort if returned results don't specify 'sorted="yes"'
+ if (fSort)
+ {
+ const char *sorted = xhResults.Element()->Attribute("sorted");
+ if (sorted != nullptr)
+ fSort = !StringUtils::EqualsNoCase(sorted, "yes");
+ }
+
+ for (TiXmlElement *pxeMovie = xhResults.FirstChild("entity").Element(); pxeMovie;
+ pxeMovie = pxeMovie->NextSiblingElement())
+ {
+ TiXmlNode *pxnTitle = pxeMovie->FirstChild("title");
+ TiXmlElement *pxeLink = pxeMovie->FirstChildElement("url");
+ if (pxnTitle && pxnTitle->FirstChild() && pxeLink && pxeLink->FirstChild())
+ {
+ CScraperUrl scurlMovie;
+ auto title = pxnTitle->FirstChild()->ValueStr();
+ std::string id;
+ if (XMLUtils::GetString(pxeMovie, "id", id))
+ scurlMovie.SetId(id);
+
+ for (; pxeLink && pxeLink->FirstChild(); pxeLink = pxeLink->NextSiblingElement("url"))
+ scurlMovie.ParseAndAppendUrl(pxeLink);
+
+ // calculate the relevance of this hit
+ std::string sCompareTitle = scurlMovie.GetTitle();
+ StringUtils::ToLower(sCompareTitle);
+ std::string sMatchTitle = sTitle;
+ StringUtils::ToLower(sMatchTitle);
+
+ /*
+ * Identify the best match by performing a fuzzy string compare on the search term and
+ * the result. Additionally, use the year (if available) to further refine the best match.
+ * An exact match scores 1, a match off by a year scores 0.5 (release dates can vary between
+ * countries), otherwise it scores 0.
+ */
+ std::string sCompareYear;
+ XMLUtils::GetString(pxeMovie, "year", sCompareYear);
+
+ double yearScore = 0;
+ if (!sYear.empty() && !sCompareYear.empty())
+ yearScore =
+ std::max(0.0, 1 - 0.5 * abs(atoi(sYear.c_str()) - atoi(sCompareYear.c_str())));
+
+ scurlMovie.SetRelevance(fstrcmp(sMatchTitle.c_str(), sCompareTitle.c_str()) + yearScore);
+
+ // reconstruct a title for the user
+ if (!sCompareYear.empty())
+ title += StringUtils::Format(" ({})", sCompareYear);
+
+ std::string sLanguage;
+ if (XMLUtils::GetString(pxeMovie, "language", sLanguage) && !sLanguage.empty())
+ title += StringUtils::Format(" ({})", sLanguage);
+
+ // filter for dupes from naughty scrapers
+ if (stsDupeCheck.insert(scurlMovie.GetFirstThumbUrl() + " " + title).second)
+ {
+ scurlMovie.SetTitle(title);
+ vcscurl.push_back(scurlMovie);
+ }
+ }
+ }
+ }
+
+ if (!fResults)
+ throw CScraperError(); // scraper aborted
+
+ if (fSort)
+ std::stable_sort(vcscurl.begin(), vcscurl.end(), RelevanceSortFunction);
+
+ return vcscurl;
+}
+
+// find album by artist, using fcurl for web fetches
+// returns a list of albums (empty if no match or failure)
+std::vector<CMusicAlbumInfo> CScraper::FindAlbum(CCurlFile &fcurl,
+ const std::string &sAlbum,
+ const std::string &sArtist)
+{
+ CLog::Log(LOGDEBUG,
+ "{}: Searching for '{} - {}' using {} scraper "
+ "(path: '{}', content: '{}', version: '{}')",
+ __FUNCTION__, sArtist, sAlbum, Name(), Path(), ADDON::TranslateContent(Content()),
+ Version().asString());
+
+ std::vector<CMusicAlbumInfo> vcali;
+ if (IsNoop())
+ return vcali;
+
+ if (m_isPython)
+ return PythonFind<CMusicAlbumInfo>(ID(),
+ {{"title", sAlbum}, {"artist", sArtist}, {"pathSettings", GetPathSettingsAsJSON()}});
+
+ // scraper function is given the album and artist as parameters and
+ // returns an XML <url> element parseable by CScraperUrl
+ std::vector<std::string> extras(2);
+ g_charsetConverter.utf8To(SearchStringEncoding(), sAlbum, extras[0]);
+ g_charsetConverter.utf8To(SearchStringEncoding(), sArtist, extras[1]);
+ extras[0] = CURL::Encode(extras[0]);
+ extras[1] = CURL::Encode(extras[1]);
+ CScraperUrl scurl;
+ std::vector<std::string> vcsOut = RunNoThrow("CreateAlbumSearchUrl", scurl, fcurl, &extras);
+ if (vcsOut.size() > 1)
+ CLog::Log(LOGWARNING, "{}: scraper returned multiple results; using first", __FUNCTION__);
+
+ if (vcsOut.empty() || vcsOut[0].empty())
+ return vcali;
+ scurl.ParseFromData(vcsOut[0]);
+
+ // the next function is passed the contents of the returned URL, and returns
+ // an empty string on failure; on success, returns XML matches in the form:
+ // <results>
+ // <entity>
+ // <title>...</title>
+ // <url>...</url> (with the usual CScraperUrl decorations like post or spoof)
+ // <artist>...</artist>
+ // <year>...</year>
+ // <relevance [scale="..."]>...</relevance> (scale defaults to 1; score is divided by it)
+ // </entity>
+ // ...
+ // </results>
+ vcsOut = RunNoThrow("GetAlbumSearchResults", scurl, fcurl);
+
+ // parse the returned XML into a vector of album objects
+ for (std::vector<std::string>::const_iterator i = vcsOut.begin(); i != vcsOut.end(); ++i)
+ {
+ CXBMCTinyXML doc;
+ doc.Parse(*i, TIXML_ENCODING_UTF8);
+ TiXmlHandle xhDoc(&doc);
+
+ for (TiXmlElement *pxeAlbum = xhDoc.FirstChild("results").FirstChild("entity").Element();
+ pxeAlbum; pxeAlbum = pxeAlbum->NextSiblingElement())
+ {
+ std::string sTitle;
+ if (XMLUtils::GetString(pxeAlbum, "title", sTitle) && !sTitle.empty())
+ {
+ std::string sArtist;
+ std::string sAlbumName;
+ if (XMLUtils::GetString(pxeAlbum, "artist", sArtist) && !sArtist.empty())
+ sAlbumName = StringUtils::Format("{} - {}", sArtist, sTitle);
+ else
+ sAlbumName = sTitle;
+
+ std::string sYear;
+ if (XMLUtils::GetString(pxeAlbum, "year", sYear) && !sYear.empty())
+ sAlbumName = StringUtils::Format("{} ({})", sAlbumName, sYear);
+
+ // if no URL is provided, use the URL we got back from CreateAlbumSearchUrl
+ // (e.g., in case we only got one result back and were sent to the detail page)
+ TiXmlElement *pxeLink = pxeAlbum->FirstChildElement("url");
+ CScraperUrl scurlAlbum;
+ if (!pxeLink)
+ scurlAlbum.ParseFromData(scurl.GetData());
+ for (; pxeLink && pxeLink->FirstChild(); pxeLink = pxeLink->NextSiblingElement("url"))
+ scurlAlbum.ParseAndAppendUrl(pxeLink);
+
+ if (!scurlAlbum.HasUrls())
+ continue;
+
+ CMusicAlbumInfo ali(sTitle, sArtist, sAlbumName, scurlAlbum);
+
+ TiXmlElement *pxeRel = pxeAlbum->FirstChildElement("relevance");
+ if (pxeRel && pxeRel->FirstChild())
+ {
+ const char *szScale = pxeRel->Attribute("scale");
+ float flScale = szScale ? float(atof(szScale)) : 1;
+ ali.SetRelevance(float(atof(pxeRel->FirstChild()->Value())) / flScale);
+ }
+
+ vcali.push_back(ali);
+ }
+ }
+ }
+ return vcali;
+}
+
+// find artist, using fcurl for web fetches
+// returns a list of artists (empty if no match or failure)
+std::vector<CMusicArtistInfo> CScraper::FindArtist(CCurlFile &fcurl, const std::string &sArtist)
+{
+ CLog::Log(LOGDEBUG,
+ "{}: Searching for '{}' using {} scraper "
+ "(file: '{}', content: '{}', version: '{}')",
+ __FUNCTION__, sArtist, Name(), Path(), ADDON::TranslateContent(Content()),
+ Version().asString());
+
+ std::vector<CMusicArtistInfo> vcari;
+ if (IsNoop())
+ return vcari;
+
+ if (m_isPython)
+ return PythonFind<CMusicArtistInfo>(ID(),
+ {{"artist", sArtist}, {"pathSettings", GetPathSettingsAsJSON()}});
+
+ // scraper function is given the artist as parameter and
+ // returns an XML <url> element parseable by CScraperUrl
+ std::vector<std::string> extras(1);
+ g_charsetConverter.utf8To(SearchStringEncoding(), sArtist, extras[0]);
+ extras[0] = CURL::Encode(extras[0]);
+ CScraperUrl scurl;
+ std::vector<std::string> vcsOut = RunNoThrow("CreateArtistSearchUrl", scurl, fcurl, &extras);
+
+ if (vcsOut.empty() || vcsOut[0].empty())
+ return vcari;
+ scurl.ParseFromData(vcsOut[0]);
+
+ // the next function is passed the contents of the returned URL, and returns
+ // an empty string on failure; on success, returns XML matches in the form:
+ // <results>
+ // <entity>
+ // <title>...</title>
+ // <year>...</year>
+ // <genre>...</genre>
+ // <disambiguation>...</disambiguation>
+ // <url>...</url> (with the usual CScraperUrl decorations like post or spoof)
+ // </entity>
+ // ...
+ // </results>
+ vcsOut = RunNoThrow("GetArtistSearchResults", scurl, fcurl);
+
+ // parse the returned XML into a vector of artist objects
+ for (std::vector<std::string>::const_iterator i = vcsOut.begin(); i != vcsOut.end(); ++i)
+ {
+ CXBMCTinyXML doc;
+ doc.Parse(*i, TIXML_ENCODING_UTF8);
+ if (!doc.RootElement())
+ {
+ CLog::Log(LOGERROR, "{}: Unable to parse XML", __FUNCTION__);
+ return vcari;
+ }
+ TiXmlHandle xhDoc(&doc);
+ for (TiXmlElement *pxeArtist = xhDoc.FirstChild("results").FirstChild("entity").Element();
+ pxeArtist; pxeArtist = pxeArtist->NextSiblingElement())
+ {
+ TiXmlNode *pxnTitle = pxeArtist->FirstChild("title");
+ if (pxnTitle && pxnTitle->FirstChild())
+ {
+ CScraperUrl scurlArtist;
+
+ TiXmlElement *pxeLink = pxeArtist->FirstChildElement("url");
+ if (!pxeLink)
+ scurlArtist.ParseFromData(scurl.GetData());
+ for (; pxeLink && pxeLink->FirstChild(); pxeLink = pxeLink->NextSiblingElement("url"))
+ scurlArtist.ParseAndAppendUrl(pxeLink);
+
+ if (!scurlArtist.HasUrls())
+ continue;
+
+ CMusicArtistInfo ari(pxnTitle->FirstChild()->Value(), scurlArtist);
+ std::string genre;
+ XMLUtils::GetString(pxeArtist, "genre", genre);
+ if (!genre.empty())
+ ari.GetArtist().genre =
+ StringUtils::Split(genre, CServiceBroker::GetSettingsComponent()->GetAdvancedSettings()->m_musicItemSeparator);
+ XMLUtils::GetString(pxeArtist, "disambiguation", ari.GetArtist().strDisambiguation);
+ XMLUtils::GetString(pxeArtist, "year", ari.GetArtist().strBorn);
+
+ vcari.push_back(ari);
+ }
+ }
+ }
+ return vcari;
+}
+
+// fetch list of episodes from URL (from video database)
+EPISODELIST CScraper::GetEpisodeList(XFILE::CCurlFile &fcurl, const CScraperUrl &scurl)
+{
+ EPISODELIST vcep;
+ if (!scurl.HasUrls())
+ return vcep;
+
+ CLog::Log(LOGDEBUG,
+ "{}: Searching '{}' using {} scraper "
+ "(file: '{}', content: '{}', version: '{}')",
+ __FUNCTION__, scurl.GetFirstThumbUrl(), Name(), Path(),
+ ADDON::TranslateContent(Content()), Version().asString());
+
+ if (m_isPython)
+ {
+ std::stringstream str;
+ str << "plugin://" << ID()
+ << "?action=getepisodelist&url=" << CURL::Encode(scurl.GetFirstThumbUrl())
+ << "&pathSettings=" << CURL::Encode(GetPathSettingsAsJSON());
+
+ CFileItemList items;
+ if (!XFILE::CDirectory::GetDirectory(str.str(), items, "", DIR_FLAG_DEFAULTS))
+ return vcep;
+
+ for (int i = 0; i < items.Size(); ++i)
+ {
+ EPISODE ep;
+ const auto& tag = *items[i]->GetVideoInfoTag();
+ ep.strTitle = tag.m_strTitle;
+ ep.iSeason = tag.m_iSeason;
+ ep.iEpisode = tag.m_iEpisode;
+ ep.cDate = tag.m_firstAired;
+ ep.iSubepisode = items[i]->GetProperty("video.sub_episode").asInteger();
+ CScraperUrl::SUrlEntry surl;
+ surl.m_type = CScraperUrl::UrlType::General;
+ surl.m_url = items[i]->GetURL().Get();
+ ep.cScraperUrl.AppendUrl(surl);
+ vcep.push_back(ep);
+ }
+
+ return vcep;
+ }
+
+ std::vector<std::string> vcsIn;
+ vcsIn.push_back(scurl.GetFirstThumbUrl());
+ std::vector<std::string> vcsOut = RunNoThrow("GetEpisodeList", scurl, fcurl, &vcsIn);
+
+ // parse the XML response
+ for (std::vector<std::string>::const_iterator i = vcsOut.begin(); i != vcsOut.end(); ++i)
+ {
+ CXBMCTinyXML doc;
+ doc.Parse(*i);
+ if (!doc.RootElement())
+ {
+ CLog::Log(LOGERROR, "{}: Unable to parse XML", __FUNCTION__);
+ continue;
+ }
+
+ TiXmlHandle xhDoc(&doc);
+ for (TiXmlElement *pxeMovie = xhDoc.FirstChild("episodeguide").FirstChild("episode").Element();
+ pxeMovie; pxeMovie = pxeMovie->NextSiblingElement())
+ {
+ EPISODE ep;
+ TiXmlElement *pxeLink = pxeMovie->FirstChildElement("url");
+ std::string strEpNum;
+ if (pxeLink && XMLUtils::GetInt(pxeMovie, "season", ep.iSeason) &&
+ XMLUtils::GetString(pxeMovie, "epnum", strEpNum) && !strEpNum.empty())
+ {
+ CScraperUrl &scurlEp(ep.cScraperUrl);
+ size_t dot = strEpNum.find('.');
+ ep.iEpisode = atoi(strEpNum.c_str());
+ ep.iSubepisode = (dot != std::string::npos) ? atoi(strEpNum.substr(dot + 1).c_str()) : 0;
+ std::string title;
+ if (!XMLUtils::GetString(pxeMovie, "title", title) || title.empty())
+ title = g_localizeStrings.Get(10005); // Not available
+ scurlEp.SetTitle(title);
+ std::string id;
+ if (XMLUtils::GetString(pxeMovie, "id", id))
+ scurlEp.SetId(id);
+
+ for (; pxeLink && pxeLink->FirstChild(); pxeLink = pxeLink->NextSiblingElement("url"))
+ scurlEp.ParseAndAppendUrl(pxeLink);
+
+ // date must be the format of yyyy-mm-dd
+ ep.cDate.SetValid(false);
+ std::string sDate;
+ if (XMLUtils::GetString(pxeMovie, "aired", sDate) && sDate.length() == 10)
+ {
+ tm tm;
+ if (strptime(sDate.c_str(), "%Y-%m-%d", &tm))
+ ep.cDate.SetDate(1900 + tm.tm_year, tm.tm_mon + 1, tm.tm_mday);
+ }
+ vcep.push_back(ep);
+ }
+ }
+ }
+
+ return vcep;
+}
+
+// takes URL; returns true and populates video details on success, false otherwise
+bool CScraper::GetVideoDetails(XFILE::CCurlFile &fcurl,
+ const CScraperUrl &scurl,
+ bool fMovie /*else episode*/,
+ CVideoInfoTag &video)
+{
+ CLog::Log(LOGDEBUG,
+ "{}: Reading {} '{}' using {} scraper "
+ "(file: '{}', content: '{}', version: '{}')",
+ __FUNCTION__, fMovie ? MediaTypeMovie : MediaTypeEpisode, scurl.GetFirstThumbUrl(),
+ Name(), Path(), ADDON::TranslateContent(Content()), Version().asString());
+
+ video.Reset();
+
+ if (m_isPython)
+ return PythonDetails(ID(), "url", scurl.GetFirstThumbUrl(),
+ fMovie ? "getdetails" : "getepisodedetails", GetPathSettingsAsJSON(), video);
+
+ std::string sFunc = fMovie ? "GetDetails" : "GetEpisodeDetails";
+ std::vector<std::string> vcsIn;
+ vcsIn.push_back(scurl.GetId());
+ vcsIn.push_back(scurl.GetFirstThumbUrl());
+ std::vector<std::string> vcsOut = RunNoThrow(sFunc, scurl, fcurl, &vcsIn);
+
+ // parse XML output
+ bool fRet(false);
+ for (std::vector<std::string>::const_iterator i = vcsOut.begin(); i != vcsOut.end(); ++i)
+ {
+ CXBMCTinyXML doc;
+ doc.Parse(*i, TIXML_ENCODING_UTF8);
+ if (!doc.RootElement())
+ {
+ CLog::Log(LOGERROR, "{}: Unable to parse XML", __FUNCTION__);
+ continue;
+ }
+
+ TiXmlHandle xhDoc(&doc);
+ TiXmlElement *pxeDetails = xhDoc.FirstChild("details").Element();
+ if (!pxeDetails)
+ {
+ CLog::Log(LOGERROR, "{}: Invalid XML file (want <details>)", __FUNCTION__);
+ continue;
+ }
+ video.Load(pxeDetails, true /*fChain*/);
+ fRet = true; // but don't exit in case of chaining
+ }
+ return fRet;
+}
+
+// takes a URL; returns true and populates album on success, false otherwise
+bool CScraper::GetAlbumDetails(CCurlFile &fcurl, const CScraperUrl &scurl, CAlbum &album)
+{
+ CLog::Log(LOGDEBUG,
+ "{}: Reading '{}' using {} scraper "
+ "(file: '{}', content: '{}', version: '{}')",
+ __FUNCTION__, scurl.GetFirstThumbUrl(), Name(), Path(),
+ ADDON::TranslateContent(Content()), Version().asString());
+
+ if (m_isPython)
+ return PythonDetails(ID(), "url", scurl.GetFirstThumbUrl(),
+ "getdetails", GetPathSettingsAsJSON(), album);
+
+ std::vector<std::string> vcsOut = RunNoThrow("GetAlbumDetails", scurl, fcurl);
+
+ // parse the returned XML into an album object (see CAlbum::Load for details)
+ bool fRet(false);
+ for (std::vector<std::string>::const_iterator i = vcsOut.begin(); i != vcsOut.end(); ++i)
+ {
+ CXBMCTinyXML doc;
+ doc.Parse(*i, TIXML_ENCODING_UTF8);
+ if (!doc.RootElement())
+ {
+ CLog::Log(LOGERROR, "{}: Unable to parse XML", __FUNCTION__);
+ return false;
+ }
+ fRet = album.Load(doc.RootElement(), i != vcsOut.begin());
+ }
+ return fRet;
+}
+
+// takes a URL (one returned from FindArtist), the original search string, and
+// returns true and populates artist on success, false on failure
+bool CScraper::GetArtistDetails(CCurlFile &fcurl,
+ const CScraperUrl &scurl,
+ const std::string &sSearch,
+ CArtist &artist)
+{
+ if (!scurl.HasUrls())
+ return false;
+
+ CLog::Log(LOGDEBUG,
+ "{}: Reading '{}' ('{}') using {} scraper "
+ "(file: '{}', content: '{}', version: '{}')",
+ __FUNCTION__, scurl.GetFirstThumbUrl(), sSearch, Name(), Path(),
+ ADDON::TranslateContent(Content()), Version().asString());
+
+ if (m_isPython)
+ return PythonDetails(ID(), "url", scurl.GetFirstThumbUrl(),
+ "getdetails", GetPathSettingsAsJSON(), artist);
+
+ // pass in the original search string for chaining to search other sites
+ std::vector<std::string> vcIn;
+ vcIn.push_back(sSearch);
+ vcIn[0] = CURL::Encode(vcIn[0]);
+
+ std::vector<std::string> vcsOut = RunNoThrow("GetArtistDetails", scurl, fcurl, &vcIn);
+
+ // ok, now parse the xml file
+ bool fRet(false);
+ for (std::vector<std::string>::const_iterator i = vcsOut.begin(); i != vcsOut.end(); ++i)
+ {
+ CXBMCTinyXML doc;
+ doc.Parse(*i, TIXML_ENCODING_UTF8);
+ if (!doc.RootElement())
+ {
+ CLog::Log(LOGERROR, "{}: Unable to parse XML", __FUNCTION__);
+ return false;
+ }
+
+ fRet = artist.Load(doc.RootElement(), i != vcsOut.begin());
+ }
+ return fRet;
+}
+
+bool CScraper::GetArtwork(XFILE::CCurlFile &fcurl, CVideoInfoTag &details)
+{
+ if (!details.HasUniqueID())
+ return false;
+
+ CLog::Log(LOGDEBUG,
+ "{}: Reading artwork for '{}' using {} scraper "
+ "(file: '{}', content: '{}', version: '{}')",
+ __FUNCTION__, details.GetUniqueID(), Name(), Path(), ADDON::TranslateContent(Content()),
+ Version().asString());
+
+ if (m_isPython)
+ return PythonDetails(ID(), "id", details.GetUniqueID(),
+ "getartwork", GetPathSettingsAsJSON(), details);
+
+ std::vector<std::string> vcsIn;
+ CScraperUrl scurl;
+ vcsIn.push_back(details.GetUniqueID());
+ std::vector<std::string> vcsOut = RunNoThrow("GetArt", scurl, fcurl, &vcsIn);
+
+ bool fRet(false);
+ for (std::vector<std::string>::const_iterator it = vcsOut.begin(); it != vcsOut.end(); ++it)
+ {
+ CXBMCTinyXML doc;
+ doc.Parse(*it, TIXML_ENCODING_UTF8);
+ if (!doc.RootElement())
+ {
+ CLog::Log(LOGERROR, "{}: Unable to parse XML", __FUNCTION__);
+ return false;
+ }
+ fRet = details.Load(doc.RootElement(), it != vcsOut.begin());
+ }
+ return fRet;
+}
+}