984 lines
29 KiB
C++
984 lines
29 KiB
C++
// -*- mode: cpp; mode: fold -*-
|
|
// Description /*{{{*/
|
|
/* ######################################################################
|
|
|
|
HTTP and HTTPS share a lot of common code and these classes are
|
|
exactly the dumping ground for this common code
|
|
|
|
##################################################################### */
|
|
/*}}}*/
|
|
// Include Files /*{{{*/
|
|
#include <config.h>
|
|
|
|
#include <apt-pkg/configuration.h>
|
|
#include <apt-pkg/debversion.h>
|
|
#include <apt-pkg/error.h>
|
|
#include <apt-pkg/fileutl.h>
|
|
#include <apt-pkg/strutl.h>
|
|
|
|
#include <cctype>
|
|
#include <csignal>
|
|
#include <cstdio>
|
|
#include <cstdlib>
|
|
#include <ctime>
|
|
#include <iostream>
|
|
#include <limits>
|
|
#include <map>
|
|
#include <string>
|
|
#include <string_view>
|
|
#include <vector>
|
|
#include <sys/stat.h>
|
|
#include <sys/time.h>
|
|
#include <unistd.h>
|
|
|
|
#include "basehttp.h"
|
|
|
|
#include <apti18n.h>
|
|
/*}}}*/
|
|
using namespace std;
|
|
|
|
string BaseHttpMethod::FailFile;
|
|
int BaseHttpMethod::FailFd = -1;
|
|
time_t BaseHttpMethod::FailTime = 0;
|
|
|
|
// Number of successful requests in a pipeline needed to continue
|
|
// pipelining after a connection reset.
|
|
constexpr int PIPELINE_MIN_SUCCESSFUL_ANSWERS_TO_CONTINUE = 3;
|
|
|
|
// ServerState::RunHeaders - Get the headers before the data /*{{{*/
|
|
// ---------------------------------------------------------------------
|
|
/* Returns 0 if things are OK, 1 if an IO error occurred and 2 if a header
|
|
parse error occurred */
|
|
ServerState::RunHeadersResult ServerState::RunHeaders(RequestState &Req,
|
|
const std::string &Uri)
|
|
{
|
|
Owner->Status(_("Waiting for headers"));
|
|
do
|
|
{
|
|
string Data;
|
|
if (ReadHeaderLines(Data) == false)
|
|
continue;
|
|
|
|
if (Owner->Debug == true)
|
|
clog << "Answer for: " << Uri << endl << Data;
|
|
|
|
for (string::const_iterator I = Data.begin(); I < Data.end(); ++I)
|
|
{
|
|
string::const_iterator J = I;
|
|
for (; J != Data.end() && *J != '\n' && *J != '\r'; ++J);
|
|
if (Req.HeaderLine(string(I,J)) == false)
|
|
return RUN_HEADERS_PARSE_ERROR;
|
|
I = J;
|
|
}
|
|
|
|
// 100 Continue is a Nop...
|
|
if (Req.Result == 100)
|
|
continue;
|
|
|
|
// Tidy up the connection persistence state.
|
|
if (Req.Encoding == RequestState::Closes && Req.haveContent == HaveContent::TRI_TRUE)
|
|
Persistent = false;
|
|
|
|
return RUN_HEADERS_OK;
|
|
} while (LoadNextResponse(false, Req) == ResultState::SUCCESSFUL);
|
|
|
|
return RUN_HEADERS_IO_ERROR;
|
|
}
|
|
/*}}}*/
|
|
bool RequestState::HeaderLine(string const &Line) /*{{{*/
|
|
{
|
|
if (Line.empty() == true)
|
|
return true;
|
|
|
|
if (Result == 0 && Line.size() > 4 && stringcasecmp(Line.data(), Line.data() + 4, "HTTP") == 0)
|
|
{
|
|
// Evil servers return no version
|
|
if (Line[4] == '/')
|
|
{
|
|
int const elements = sscanf(Line.c_str(),"HTTP/%3u.%3u %3u%359[^\n]",&Major,&Minor,&Result,Code);
|
|
if (elements == 3)
|
|
{
|
|
Code[0] = '\0';
|
|
if (Owner != NULL && Owner->Debug == true)
|
|
clog << "HTTP server doesn't give Reason-Phrase for " << std::to_string(Result) << std::endl;
|
|
}
|
|
else if (elements != 4)
|
|
return _error->Error(_("The HTTP server sent an invalid reply header"));
|
|
}
|
|
else
|
|
{
|
|
Major = 0;
|
|
Minor = 9;
|
|
if (sscanf(Line.c_str(),"HTTP %3u%359[^\n]",&Result,Code) != 2)
|
|
return _error->Error(_("The HTTP server sent an invalid reply header"));
|
|
}
|
|
auto const CodeLen = strlen(Code);
|
|
auto const CodeEnd = std::remove_if(Code, Code + CodeLen, [](char c) { return isprint(c) == 0; });
|
|
*CodeEnd = '\0';
|
|
|
|
/* Check the HTTP response header to get the default persistence
|
|
state. */
|
|
if (Major < 1)
|
|
Server->Persistent = false;
|
|
else
|
|
{
|
|
if (Major == 1 && Minor == 0)
|
|
{
|
|
Server->Persistent = false;
|
|
}
|
|
else
|
|
{
|
|
Server->Persistent = true;
|
|
if (Server->PipelineAllowed)
|
|
Server->Pipeline = true;
|
|
}
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
// Blah, some servers use "connection:closes", evil.
|
|
// and some even send empty header fields…
|
|
string::size_type Pos = Line.find(':');
|
|
if (Pos == string::npos)
|
|
return _error->Error(_("Bad header line"));
|
|
++Pos;
|
|
|
|
// Parse off any trailing spaces between the : and the next word.
|
|
string::size_type Pos2 = Pos;
|
|
while (Pos2 < Line.length() && isspace_ascii(Line[Pos2]) != 0)
|
|
Pos2++;
|
|
|
|
string const Tag(Line,0,Pos);
|
|
string const Val(Line,Pos2);
|
|
|
|
if (stringcasecmp(Tag,"Content-Length:") == 0)
|
|
{
|
|
auto ContentLength = strtoull(Val.c_str(), NULL, 10);
|
|
if (ContentLength == 0)
|
|
{
|
|
haveContent = HaveContent::TRI_FALSE;
|
|
return true;
|
|
}
|
|
if (Encoding == Closes)
|
|
Encoding = Stream;
|
|
haveContent = HaveContent::TRI_TRUE;
|
|
|
|
unsigned long long * DownloadSizePtr = &DownloadSize;
|
|
if (Result == 416 || (Result >= 300 && Result < 400))
|
|
DownloadSizePtr = &JunkSize;
|
|
|
|
*DownloadSizePtr = ContentLength;
|
|
if (*DownloadSizePtr >= std::numeric_limits<unsigned long long>::max())
|
|
return _error->Errno("HeaderLine", _("The HTTP server sent an invalid Content-Length header"));
|
|
else if (*DownloadSizePtr == 0)
|
|
haveContent = HaveContent::TRI_FALSE;
|
|
|
|
// On partial content (206) the Content-Length less than the real
|
|
// size, so do not set it here but leave that to the Content-Range
|
|
// header instead
|
|
if(Result != 206 && TotalFileSize == 0)
|
|
TotalFileSize = DownloadSize;
|
|
|
|
return true;
|
|
}
|
|
|
|
if (stringcasecmp(Tag,"Content-Type:") == 0)
|
|
{
|
|
if (haveContent == HaveContent::TRI_UNKNOWN)
|
|
haveContent = HaveContent::TRI_TRUE;
|
|
return true;
|
|
}
|
|
|
|
// The Content-Range field only has a meaning in HTTP/1.1 for the
|
|
// 206 (Partial Content) and 416 (Range Not Satisfiable) responses
|
|
// according to RFC7233 "Range Requests", §4.2, so only consider it
|
|
// for such responses.
|
|
if ((Result == 416 || Result == 206) && stringcasecmp(Tag,"Content-Range:") == 0)
|
|
{
|
|
if (haveContent == HaveContent::TRI_UNKNOWN)
|
|
haveContent = HaveContent::TRI_TRUE;
|
|
|
|
// §14.16 says 'byte-range-resp-spec' should be a '*' in case of 416
|
|
if (Result == 416 && sscanf(Val.c_str(), "bytes */%llu",&TotalFileSize) == 1)
|
|
; // we got the expected filesize which is all we wanted
|
|
else if (sscanf(Val.c_str(),"bytes %llu-%*u/%llu",&StartPos,&TotalFileSize) != 2)
|
|
return _error->Error(_("The HTTP server sent an invalid Content-Range header"));
|
|
if (StartPos > TotalFileSize)
|
|
return _error->Error(_("This HTTP server has broken range support"));
|
|
|
|
// figure out what we will download
|
|
DownloadSize = TotalFileSize - StartPos;
|
|
return true;
|
|
}
|
|
|
|
if (stringcasecmp(Tag,"Transfer-Encoding:") == 0)
|
|
{
|
|
if (haveContent == HaveContent::TRI_UNKNOWN)
|
|
haveContent = HaveContent::TRI_TRUE;
|
|
if (stringcasecmp(Val,"chunked") == 0)
|
|
Encoding = Chunked;
|
|
return true;
|
|
}
|
|
|
|
if (stringcasecmp(Tag,"Connection:") == 0)
|
|
{
|
|
if (stringcasecmp(Val,"close") == 0)
|
|
{
|
|
Server->Persistent = false;
|
|
Server->Pipeline = false;
|
|
/* Some servers send error pages (as they are dynamically generated)
|
|
for simplicity via a connection close instead of e.g. chunked,
|
|
so assuming an always closing server only if we get a file + close */
|
|
if (Result >= 200 && Result < 300 && Server->PipelineAnswersReceived < PIPELINE_MIN_SUCCESSFUL_ANSWERS_TO_CONTINUE)
|
|
{
|
|
Server->PipelineAllowed = false;
|
|
Server->PipelineAnswersReceived = 0;
|
|
}
|
|
}
|
|
else if (stringcasecmp(Val,"keep-alive") == 0)
|
|
Server->Persistent = true;
|
|
return true;
|
|
}
|
|
|
|
if (stringcasecmp(Tag,"Last-Modified:") == 0)
|
|
{
|
|
if (RFC1123StrToTime(Val, Date) == false)
|
|
return _error->Error(_("Unknown date format"));
|
|
return true;
|
|
}
|
|
|
|
if (stringcasecmp(Tag,"Location:") == 0)
|
|
{
|
|
Location = Val;
|
|
return true;
|
|
}
|
|
|
|
if (Server->RangesAllowed && stringcasecmp(Tag, "Accept-Ranges:") == 0)
|
|
{
|
|
std::string ranges = ',' + Val + ',';
|
|
ranges.erase(std::remove(ranges.begin(), ranges.end(), ' '), ranges.end());
|
|
if (ranges.find(",bytes,") == std::string::npos)
|
|
Server->RangesAllowed = false;
|
|
return true;
|
|
}
|
|
|
|
if (stringcasecmp(Tag, "Retry-After:") == 0)
|
|
{
|
|
unsigned long _retry_after_s;
|
|
if (RFC1123StrToTime(Val, RetryAfter))
|
|
{
|
|
return true;
|
|
}
|
|
if (StrToNum(Val.c_str(), _retry_after_s, 4, 10) == 1)
|
|
{
|
|
RetryAfter = time(nullptr) + _retry_after_s;
|
|
return true;
|
|
}
|
|
return _error->Error(_("Unknown date format"));
|
|
}
|
|
|
|
if (Server->RangesAllowed && stringcasecmp(Tag, "Via:") == 0)
|
|
{
|
|
auto const parts = VectorizeString(Val, ' ');
|
|
std::string_view const varnish{"(Varnish/"};
|
|
if (parts.size() != 3 || parts[1] != "varnish" || parts[2].empty() ||
|
|
not APT::String::Startswith(parts[2], std::string{varnish}) ||
|
|
parts[2].back() != ')')
|
|
return true;
|
|
auto const version = parts[2].substr(varnish.length(), parts[2].length() - (varnish.length() + 1));
|
|
if (version.empty())
|
|
return true;
|
|
std::string_view const varnishsupport{"6.4~"};
|
|
if (debVersioningSystem::CmpFragment(version.data(), version.data() + version.length(),
|
|
varnishsupport.begin(), varnishsupport.end()) < 0)
|
|
Server->RangesAllowed = false;
|
|
return true;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
/*}}}*/
|
|
// ServerState::ServerState - Constructor /*{{{*/
|
|
ServerState::ServerState(URI Srv, BaseHttpMethod *Owner) :
|
|
ServerName(Srv), TimeOut(30), Owner(Owner)
|
|
{
|
|
Reset();
|
|
}
|
|
/*}}}*/
|
|
bool RequestState::AddPartialFileToHashes(FileFd &File) /*{{{*/
|
|
{
|
|
File.Truncate(StartPos);
|
|
return Server->GetHashes()->AddFD(File, StartPos);
|
|
}
|
|
/*}}}*/
|
|
void ServerState::Reset() /*{{{*/
|
|
{
|
|
Persistent = false;
|
|
Pipeline = false;
|
|
PipelineAllowed = true;
|
|
PipelineAnswersReceived = 0;
|
|
}
|
|
/*}}}*/
|
|
|
|
// BaseHttpMethod::DealWithHeaders - Handle the retrieved header data /*{{{*/
|
|
// ---------------------------------------------------------------------
|
|
/* We look at the header data we got back from the server and decide what
|
|
to do. Returns DealWithHeadersResult (see http.h for details).
|
|
*/
|
|
static std::string fixURIEncoding(std::string const &part)
|
|
{
|
|
// if the server sends a space this is not an encoded URI
|
|
// so other clients seem to encode it and we do it as well
|
|
if (part.find_first_of(" ") != std::string::npos)
|
|
return aptMethod::URIEncode(part);
|
|
return part;
|
|
}
|
|
BaseHttpMethod::DealWithHeadersResult
|
|
BaseHttpMethod::DealWithHeaders(FetchResult &Res, RequestState &Req)
|
|
{
|
|
// Not Modified
|
|
if (Req.Result == 304)
|
|
{
|
|
RemoveFile("server", Queue->DestFile);
|
|
Res.IMSHit = true;
|
|
Res.LastModified = Queue->LastModified;
|
|
Res.Size = 0;
|
|
return IMS_HIT;
|
|
}
|
|
|
|
/* Note that it is only OK for us to treat all redirection the same
|
|
because we *always* use GET, not other HTTP methods.
|
|
Codes not mentioned are handled as errors later as required by the
|
|
HTTP spec to handle unknown codes the same as the x00 code. */
|
|
constexpr unsigned int RedirectCodes[] = {
|
|
301, // Moved Permanently
|
|
302, // Found
|
|
303, // See Other
|
|
307, // Temporary Redirect
|
|
308, // Permanent Redirect
|
|
};
|
|
if (AllowRedirect && std::find(std::begin(RedirectCodes), std::end(RedirectCodes), Req.Result) != std::end(RedirectCodes))
|
|
{
|
|
if (Req.Location.empty() == true)
|
|
;
|
|
else if (Req.Location[0] == '/' && Queue->Uri.empty() == false)
|
|
{
|
|
URI Uri(Queue->Uri);
|
|
if (Uri.Host.empty() == false)
|
|
NextURI = URI::SiteOnly(Uri);
|
|
else
|
|
NextURI.clear();
|
|
if (_config->FindB("Acquire::Send-URI-Encoded", false))
|
|
NextURI.append(fixURIEncoding(Req.Location));
|
|
else
|
|
NextURI.append(DeQuoteString(Req.Location));
|
|
if (Queue->Uri == NextURI)
|
|
{
|
|
SetFailReason("RedirectionLoop");
|
|
_error->Error("Redirection loop encountered");
|
|
if (Req.haveContent == HaveContent::TRI_TRUE)
|
|
return ERROR_WITH_CONTENT_PAGE;
|
|
return ERROR_UNRECOVERABLE;
|
|
}
|
|
return TRY_AGAIN_OR_REDIRECT;
|
|
}
|
|
else
|
|
{
|
|
bool const SendURIEncoded = _config->FindB("Acquire::Send-URI-Encoded", false);
|
|
if (not SendURIEncoded)
|
|
Req.Location = DeQuoteString(Req.Location);
|
|
URI tmpURI(Req.Location);
|
|
if (SendURIEncoded)
|
|
tmpURI.Path = fixURIEncoding(tmpURI.Path);
|
|
if (tmpURI.Access.find('+') != std::string::npos)
|
|
{
|
|
_error->Error("Server tried to trick us into using a specific implementation: %s", tmpURI.Access.c_str());
|
|
if (Req.haveContent == HaveContent::TRI_TRUE)
|
|
return ERROR_WITH_CONTENT_PAGE;
|
|
return ERROR_UNRECOVERABLE;
|
|
}
|
|
NextURI = tmpURI;
|
|
URI Uri(Queue->Uri);
|
|
if (Binary.find('+') != std::string::npos)
|
|
{
|
|
auto base = Binary.substr(0, Binary.find('+'));
|
|
if (base != tmpURI.Access)
|
|
{
|
|
tmpURI.Access = base + '+' + tmpURI.Access;
|
|
if (tmpURI.Access == Binary)
|
|
{
|
|
std::swap(tmpURI.Access, Uri.Access);
|
|
NextURI = tmpURI;
|
|
std::swap(tmpURI.Access, Uri.Access);
|
|
}
|
|
else
|
|
NextURI = tmpURI;
|
|
}
|
|
}
|
|
if (Queue->Uri == NextURI)
|
|
{
|
|
SetFailReason("RedirectionLoop");
|
|
_error->Error("Redirection loop encountered");
|
|
if (Req.haveContent == HaveContent::TRI_TRUE)
|
|
return ERROR_WITH_CONTENT_PAGE;
|
|
return ERROR_UNRECOVERABLE;
|
|
}
|
|
Uri.Access = Binary;
|
|
// same protocol redirects are okay
|
|
if (tmpURI.Access == Uri.Access)
|
|
return TRY_AGAIN_OR_REDIRECT;
|
|
// as well as http to https
|
|
else if ((Uri.Access == "http" || Uri.Access == "https+http") && tmpURI.Access == "https")
|
|
return TRY_AGAIN_OR_REDIRECT;
|
|
else
|
|
{
|
|
auto const tmpplus = tmpURI.Access.find('+');
|
|
if (tmpplus != std::string::npos && tmpURI.Access.substr(tmpplus + 1) == "https")
|
|
{
|
|
auto const uriplus = Uri.Access.find('+');
|
|
if (uriplus == std::string::npos)
|
|
{
|
|
if (Uri.Access == tmpURI.Access.substr(0, tmpplus)) // foo -> foo+https
|
|
return TRY_AGAIN_OR_REDIRECT;
|
|
}
|
|
else if (Uri.Access.substr(uriplus + 1) == "http" &&
|
|
Uri.Access.substr(0, uriplus) == tmpURI.Access.substr(0, tmpplus)) // foo+http -> foo+https
|
|
return TRY_AGAIN_OR_REDIRECT;
|
|
}
|
|
}
|
|
_error->Error("Redirection from %s to '%s' is forbidden", Uri.Access.c_str(), NextURI.c_str());
|
|
}
|
|
/* else pass through for error message */
|
|
}
|
|
// the server is not supporting ranges as much as we would like. Retry without ranges
|
|
else if (not Server->RangesAllowed && (Req.Result == 416 || Req.Result == 206))
|
|
{
|
|
RemoveFile("server", Queue->DestFile);
|
|
NextURI = Queue->Uri;
|
|
return TRY_AGAIN_OR_REDIRECT;
|
|
}
|
|
// retry after an invalid range response without partial data
|
|
else if (Req.Result == 416)
|
|
{
|
|
struct stat SBuf;
|
|
if (stat(Queue->DestFile.c_str(),&SBuf) >= 0 && SBuf.st_size > 0)
|
|
{
|
|
bool partialHit = false;
|
|
if (Queue->ExpectedHashes.usable() == true)
|
|
{
|
|
Hashes resultHashes(Queue->ExpectedHashes);
|
|
FileFd file(Queue->DestFile, FileFd::ReadOnly);
|
|
Req.TotalFileSize = file.FileSize();
|
|
Req.Date = file.ModificationTime();
|
|
resultHashes.AddFD(file);
|
|
HashStringList const hashList = resultHashes.GetHashStringList();
|
|
partialHit = (Queue->ExpectedHashes == hashList);
|
|
}
|
|
else if ((unsigned long long)SBuf.st_size == Req.TotalFileSize)
|
|
partialHit = true;
|
|
if (partialHit == true)
|
|
{
|
|
// the file is completely downloaded, but was not moved
|
|
if (Req.haveContent == HaveContent::TRI_TRUE)
|
|
{
|
|
// nuke the sent error page
|
|
Server->RunDataToDevNull(Req);
|
|
Req.haveContent = HaveContent::TRI_FALSE;
|
|
}
|
|
Req.StartPos = Req.TotalFileSize;
|
|
Req.Result = 200;
|
|
}
|
|
else if (RemoveFile("server", Queue->DestFile))
|
|
{
|
|
NextURI = Queue->Uri;
|
|
return TRY_AGAIN_OR_REDIRECT;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* We have a reply we don't handle. This should indicate a perm server
|
|
failure */
|
|
if (Req.Result < 200 || Req.Result >= 300)
|
|
{
|
|
if (_error->PendingError() == false)
|
|
{
|
|
std::string err;
|
|
strprintf(err, "HttpError%u", Req.Result);
|
|
SetFailReason(err);
|
|
_error->Error("%u %s", Req.Result, Req.Code);
|
|
}
|
|
if (Req.haveContent == HaveContent::TRI_TRUE)
|
|
return ERROR_WITH_CONTENT_PAGE;
|
|
return ERROR_UNRECOVERABLE;
|
|
}
|
|
|
|
// This is some sort of 2xx 'data follows' reply
|
|
Res.LastModified = Req.Date;
|
|
Res.Size = Req.TotalFileSize;
|
|
return FILE_IS_OPEN;
|
|
}
|
|
/*}}}*/
|
|
// BaseHttpMethod::SigTerm - Handle a fatal signal /*{{{*/
|
|
// ---------------------------------------------------------------------
|
|
/* This closes and timestamps the open file. This is necessary to get
|
|
resume behavior on user abort */
|
|
void BaseHttpMethod::SigTerm(int)
|
|
{
|
|
if (FailFd == -1)
|
|
_exit(100);
|
|
|
|
struct timeval times[2];
|
|
times[0].tv_sec = FailTime;
|
|
times[1].tv_sec = FailTime;
|
|
times[0].tv_usec = times[1].tv_usec = 0;
|
|
utimes(FailFile.c_str(), times);
|
|
close(FailFd);
|
|
|
|
_exit(100);
|
|
}
|
|
/*}}}*/
|
|
// BaseHttpMethod::Fetch - Fetch an item /*{{{*/
|
|
// ---------------------------------------------------------------------
|
|
/* This adds an item to the pipeline. We keep the pipeline at a fixed
|
|
depth. */
|
|
bool BaseHttpMethod::Fetch(FetchItem *)
|
|
{
|
|
if (Server == nullptr || QueueBack == nullptr)
|
|
return true;
|
|
|
|
// If pipelining is disabled, we only queue 1 request
|
|
auto const AllowedDepth = Server->Pipeline ? PipelineDepth : 0;
|
|
// how deep is our pipeline currently?
|
|
decltype(PipelineDepth) CurrentDepth = 0;
|
|
for (FetchItem const *I = Queue; I != QueueBack; I = I->Next)
|
|
++CurrentDepth;
|
|
if (CurrentDepth > AllowedDepth)
|
|
return true;
|
|
|
|
do {
|
|
// Make sure we stick with the same server
|
|
if (Server->Comp(URI(QueueBack->Uri)) == false)
|
|
break;
|
|
|
|
bool const UsableHashes = QueueBack->ExpectedHashes.usable();
|
|
// if we have no hashes, do at most one such request
|
|
// as we can't fixup pipeling misbehaviors otherwise
|
|
if (CurrentDepth != 0 && UsableHashes == false)
|
|
break;
|
|
|
|
if (UsableHashes && FileExists(QueueBack->DestFile))
|
|
{
|
|
FileFd partial(QueueBack->DestFile, FileFd::ReadOnly);
|
|
Hashes wehave(QueueBack->ExpectedHashes);
|
|
if (QueueBack->ExpectedHashes.FileSize() == partial.FileSize())
|
|
{
|
|
if (wehave.AddFD(partial) &&
|
|
wehave.GetHashStringList() == QueueBack->ExpectedHashes)
|
|
{
|
|
FetchResult Res;
|
|
Res.Filename = QueueBack->DestFile;
|
|
Res.ResumePoint = QueueBack->ExpectedHashes.FileSize();
|
|
URIStart(Res);
|
|
// move item to the start of the queue as URIDone will
|
|
// always dequeued the first item in the queue
|
|
if (Queue != QueueBack)
|
|
{
|
|
FetchItem *Prev = Queue;
|
|
for (; Prev->Next != QueueBack; Prev = Prev->Next)
|
|
/* look for the previous queue item */;
|
|
Prev->Next = QueueBack->Next;
|
|
QueueBack->Next = Queue;
|
|
Queue = QueueBack;
|
|
QueueBack = Prev->Next;
|
|
}
|
|
Res.TakeHashes(wehave);
|
|
URIDone(Res);
|
|
continue;
|
|
}
|
|
else
|
|
RemoveFile("Fetch-Partial", QueueBack->DestFile);
|
|
}
|
|
}
|
|
auto const Tmp = QueueBack;
|
|
QueueBack = QueueBack->Next;
|
|
SendReq(Tmp);
|
|
++CurrentDepth;
|
|
} while (CurrentDepth <= AllowedDepth && QueueBack != nullptr);
|
|
|
|
return true;
|
|
}
|
|
/*}}}*/
|
|
// BaseHttpMethod::Loop - Main loop /*{{{*/
|
|
int BaseHttpMethod::Loop()
|
|
{
|
|
signal(SIGTERM,SigTerm);
|
|
signal(SIGINT,SigTerm);
|
|
|
|
Server = 0;
|
|
|
|
int FailCounter = 0;
|
|
while (1)
|
|
{
|
|
// We have no commands, wait for some to arrive
|
|
if (Queue == 0)
|
|
{
|
|
if (WaitFd(STDIN_FILENO) == false)
|
|
return 0;
|
|
}
|
|
|
|
/* Run messages, we can accept 0 (no message) if we didn't
|
|
do a WaitFd above.. Otherwise the FD is closed. */
|
|
int Result = Run(true);
|
|
if (Result != -1 && (Result != 0 || Queue == 0))
|
|
{
|
|
if(FailReason.empty() == false ||
|
|
ConfigFindB("DependOnSTDIN", true) == true)
|
|
return 100;
|
|
else
|
|
return 0;
|
|
}
|
|
|
|
if (Queue == 0)
|
|
continue;
|
|
|
|
// Connect to the server
|
|
if (Server == 0 || Server->Comp(URI(Queue->Uri)) == false)
|
|
{
|
|
if (!Queue->Proxy().empty())
|
|
{
|
|
URI uri(Queue->Uri);
|
|
_config->Set("Acquire::" + uri.Access + "::proxy::" + uri.Host, Queue->Proxy());
|
|
}
|
|
Server = CreateServerState(URI(Queue->Uri));
|
|
setPostfixForMethodNames(::URI(Queue->Uri).Host.c_str());
|
|
AllowRedirect = ConfigFindB("AllowRedirect", true);
|
|
PipelineDepth = ConfigFindI("Pipeline-Depth", 10);
|
|
Server->RangesAllowed = ConfigFindB("AllowRanges", true);
|
|
Debug = DebugEnabled();
|
|
}
|
|
|
|
/* If the server has explicitly said this is the last connection
|
|
then we pre-emptively shut down the pipeline and tear down
|
|
the connection. This will speed up HTTP/1.0 servers a tad
|
|
since we don't have to wait for the close sequence to
|
|
complete */
|
|
if (Server->Persistent == false)
|
|
Server->Close();
|
|
|
|
// Reset the pipeline
|
|
if (Server->IsOpen() == false) {
|
|
QueueBack = Queue;
|
|
Server->PipelineAnswersReceived = 0;
|
|
}
|
|
|
|
// Connect to the host
|
|
switch (Server->Open())
|
|
{
|
|
case ResultState::FATAL_ERROR:
|
|
Fail(false);
|
|
Server = nullptr;
|
|
continue;
|
|
case ResultState::TRANSIENT_ERROR:
|
|
Fail(true);
|
|
Server = nullptr;
|
|
continue;
|
|
case ResultState::SUCCESSFUL:
|
|
break;
|
|
}
|
|
|
|
// Fill the pipeline.
|
|
Fetch(0);
|
|
|
|
RequestState Req(this, Server.get());
|
|
// Fetch the next URL header data from the server.
|
|
switch (Server->RunHeaders(Req, Queue->Uri))
|
|
{
|
|
case ServerState::RUN_HEADERS_OK:
|
|
break;
|
|
|
|
// The header data is bad
|
|
case ServerState::RUN_HEADERS_PARSE_ERROR:
|
|
{
|
|
_error->Error(_("Bad header data"));
|
|
Fail(true);
|
|
Server->Close();
|
|
RotateDNS();
|
|
continue;
|
|
}
|
|
|
|
// The server closed a connection during the header get..
|
|
default:
|
|
case ServerState::RUN_HEADERS_IO_ERROR:
|
|
{
|
|
FailCounter++;
|
|
_error->Discard();
|
|
Server->Close();
|
|
Server->Pipeline = false;
|
|
Server->PipelineAllowed = false;
|
|
|
|
if (FailCounter >= 2)
|
|
{
|
|
Fail(_("Connection failed"),true);
|
|
FailCounter = 0;
|
|
}
|
|
|
|
RotateDNS();
|
|
continue;
|
|
}
|
|
};
|
|
|
|
// Decide what to do.
|
|
FetchResult Res;
|
|
Res.Filename = Queue->DestFile;
|
|
switch (DealWithHeaders(Res, Req))
|
|
{
|
|
// Ok, the file is Open
|
|
case FILE_IS_OPEN:
|
|
{
|
|
URIStart(Res);
|
|
|
|
// Run the data
|
|
ResultState Result = ResultState::SUCCESSFUL;
|
|
|
|
// ensure we don't fetch too much
|
|
// we could do "Server->MaximumSize = Queue->MaximumSize" here
|
|
// but that would break the clever pipeline messup detection
|
|
// so instead we use the size of the biggest item in the queue
|
|
Req.MaximumSize = FindMaximumObjectSizeInQueue();
|
|
|
|
if (Req.haveContent == HaveContent::TRI_TRUE)
|
|
{
|
|
/* If the server provides Content-Length we can figure out with it if
|
|
this satisfies any request we have made so far (in the pipeline).
|
|
If not we can kill the connection as whatever file the server is trying
|
|
to send to us would be rejected with a hashsum mismatch later or triggers
|
|
a maximum size error. We don't run the data to /dev/null as this can be MBs
|
|
of junk data we would waste bandwidth on and instead just close the connection
|
|
to reopen a fresh one which should be more cost/time efficient */
|
|
if (Req.DownloadSize > 0)
|
|
{
|
|
decltype(Queue->ExpectedHashes.FileSize()) const filesize = Req.StartPos + Req.DownloadSize;
|
|
bool found = false;
|
|
for (FetchItem const *I = Queue; I != 0 && I != QueueBack; I = I->Next)
|
|
{
|
|
auto const fs = I->ExpectedHashes.FileSize();
|
|
if (fs == 0 || fs == filesize)
|
|
{
|
|
found = true;
|
|
break;
|
|
}
|
|
}
|
|
if (found == false)
|
|
{
|
|
SetFailReason("MaximumSizeExceeded");
|
|
_error->Error(_("File has unexpected size (%llu != %llu). Mirror sync in progress?"),
|
|
filesize, Queue->ExpectedHashes.FileSize());
|
|
Result = ResultState::FATAL_ERROR;
|
|
}
|
|
}
|
|
if (Result == ResultState::SUCCESSFUL)
|
|
Result = Server->RunData(Req);
|
|
}
|
|
|
|
/* If the server is sending back sizeless responses then fill in
|
|
the size now */
|
|
if (Res.Size == 0)
|
|
Res.Size = Req.File.Size();
|
|
|
|
// Close the file, destroy the FD object and timestamp it
|
|
FailFd = -1;
|
|
Req.File.Close();
|
|
|
|
// Timestamp
|
|
struct timeval times[2];
|
|
times[0].tv_sec = times[1].tv_sec = Req.Date;
|
|
times[0].tv_usec = times[1].tv_usec = 0;
|
|
utimes(Queue->DestFile.c_str(), times);
|
|
|
|
// Send status to APT
|
|
if (Result == ResultState::SUCCESSFUL)
|
|
{
|
|
Hashes * const resultHashes = Server->GetHashes();
|
|
HashStringList const hashList = resultHashes->GetHashStringList();
|
|
if (PipelineDepth != 0 && Queue->ExpectedHashes.usable() == true && Queue->ExpectedHashes != hashList)
|
|
{
|
|
// we did not get the expected hash… mhhh:
|
|
// could it be that server/proxy messed up pipelining?
|
|
FetchItem * BeforeI = Queue;
|
|
for (FetchItem *I = Queue->Next; I != 0 && I != QueueBack; I = I->Next)
|
|
{
|
|
if (I->ExpectedHashes.usable() == true && I->ExpectedHashes == hashList)
|
|
{
|
|
// yes, he did! Disable pipelining and rewrite queue
|
|
if (Server->Pipeline == true)
|
|
{
|
|
std::string msg;
|
|
strprintf(msg, _("Automatically disabled %s due to incorrect response from server/proxy. (man 5 apt.conf)"), "Acquire::http::Pipeline-Depth");
|
|
Warning(std::move(msg));
|
|
Server->Pipeline = false;
|
|
Server->PipelineAllowed = false;
|
|
// we keep the PipelineDepth value so that the rest of the queue can be fixed up as well
|
|
}
|
|
Rename(Res.Filename, I->DestFile);
|
|
Res.Filename = I->DestFile;
|
|
BeforeI->Next = I->Next;
|
|
I->Next = Queue;
|
|
Queue = I;
|
|
break;
|
|
}
|
|
BeforeI = I;
|
|
}
|
|
}
|
|
if (Server->Pipeline == true)
|
|
{
|
|
Server->PipelineAnswersReceived++;
|
|
}
|
|
Res.TakeHashes(*resultHashes);
|
|
URIDone(Res);
|
|
}
|
|
else
|
|
{
|
|
if (not Server->IsOpen())
|
|
{
|
|
// Reset the pipeline
|
|
QueueBack = Queue;
|
|
Server->PipelineAnswersReceived = 0;
|
|
}
|
|
|
|
Server->Close();
|
|
FailCounter = 0;
|
|
switch (Result)
|
|
{
|
|
case ResultState::TRANSIENT_ERROR:
|
|
Fail(true);
|
|
break;
|
|
case ResultState::FATAL_ERROR:
|
|
case ResultState::SUCCESSFUL:
|
|
Fail(false);
|
|
break;
|
|
}
|
|
}
|
|
break;
|
|
}
|
|
|
|
// IMS hit
|
|
case IMS_HIT:
|
|
{
|
|
URIDone(Res);
|
|
break;
|
|
}
|
|
|
|
// Hard server error, not found or something
|
|
case ERROR_UNRECOVERABLE:
|
|
{
|
|
Fail();
|
|
break;
|
|
}
|
|
|
|
// Hard internal error, kill the connection and fail
|
|
case ERROR_NOT_FROM_SERVER:
|
|
{
|
|
Fail();
|
|
RotateDNS();
|
|
Server->Close();
|
|
break;
|
|
}
|
|
|
|
// We need to flush the data, the header is like a 404 w/ error text
|
|
case ERROR_WITH_CONTENT_PAGE:
|
|
{
|
|
Server->RunDataToDevNull(Req);
|
|
constexpr unsigned int TransientCodes[] = {
|
|
408, // Request Timeout
|
|
429, // Too Many Requests
|
|
500, // Internal Server Error
|
|
502, // Bad Gateway
|
|
503, // Service Unavailable
|
|
504, // Gateway Timeout
|
|
599, // Network Connect Timeout Error
|
|
};
|
|
if (std::find(std::begin(TransientCodes), std::end(TransientCodes), Req.Result) != std::end(TransientCodes))
|
|
{
|
|
if (Req.RetryAfter)
|
|
{
|
|
std::unordered_map<std::string, std::string> fields;
|
|
fields["Retry-After"] = std::to_string(Req.RetryAfter);
|
|
SetFailReason("TooManyRequests");
|
|
FailWithContext(std::string("TooManyRequests"), true, fields);
|
|
}
|
|
else
|
|
{
|
|
Fail(true);
|
|
}
|
|
}
|
|
else
|
|
Fail();
|
|
break;
|
|
}
|
|
|
|
// Try again with a new URL
|
|
case TRY_AGAIN_OR_REDIRECT:
|
|
{
|
|
// Clear rest of response if there is content
|
|
if (Req.haveContent == HaveContent::TRI_TRUE)
|
|
Server->RunDataToDevNull(Req);
|
|
Redirect(NextURI);
|
|
break;
|
|
}
|
|
|
|
default:
|
|
Fail(_("Internal error"));
|
|
break;
|
|
}
|
|
|
|
FailCounter = 0;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
/*}}}*/
|
|
unsigned long long BaseHttpMethod::FindMaximumObjectSizeInQueue() const /*{{{*/
|
|
{
|
|
unsigned long long MaxSizeInQueue = 0;
|
|
for (FetchItem *I = Queue; I != 0 && I != QueueBack; I = I->Next)
|
|
{
|
|
if (I->MaximumSize == 0)
|
|
return 0;
|
|
MaxSizeInQueue = std::max(MaxSizeInQueue, I->MaximumSize);
|
|
}
|
|
return MaxSizeInQueue;
|
|
}
|
|
/*}}}*/
|
|
BaseHttpMethod::BaseHttpMethod(std::string &&Binary, char const *const Ver, unsigned long const Flags) /*{{{*/
|
|
: aptAuthConfMethod(std::move(Binary), Ver, Flags), Server(nullptr),
|
|
AllowRedirect(false), Debug(false), PipelineDepth(10)
|
|
{
|
|
}
|
|
/*}}}*/
|
|
bool BaseHttpMethod::Configuration(std::string Message) /*{{{*/
|
|
{
|
|
if (aptAuthConfMethod::Configuration(Message) == false)
|
|
return false;
|
|
|
|
_config->CndSet("Acquire::tor::Proxy",
|
|
"socks5h://apt-transport-tor@127.0.0.1:9050");
|
|
return true;
|
|
}
|
|
/*}}}*/
|
|
bool BaseHttpMethod::AddProxyAuth(URI &Proxy, URI const &Server) /*{{{*/
|
|
{
|
|
MaybeAddAuthTo(Proxy);
|
|
if (std::find(methodNames.begin(), methodNames.end(), "tor") != methodNames.end() &&
|
|
Proxy.User == "apt-transport-tor" && Proxy.Password.empty())
|
|
{
|
|
std::string pass = Server.Host;
|
|
pass.erase(std::remove_if(pass.begin(), pass.end(), [](char const c) { return std::isalnum(c) == 0; }), pass.end());
|
|
if (pass.length() > 255)
|
|
Proxy.Password = pass.substr(0, 255);
|
|
else
|
|
Proxy.Password = std::move(pass);
|
|
}
|
|
return true;
|
|
}
|
|
/*}}}*/
|