mirror of
https://github.com/NixOS/nix.git
synced 2025-11-09 03:56:01 +01:00
Merge pull request #13881 from xokdvium/pass-url-verbatim
lib{store,fetchers}: Pass URLs specified directly verbatim to FileTra…
This commit is contained in:
commit
3e0fb3f8d2
8 changed files with 84 additions and 14 deletions
|
|
@ -43,7 +43,7 @@ DownloadFileResult downloadFile(
|
||||||
if (cached && !cached->expired)
|
if (cached && !cached->expired)
|
||||||
return useCached();
|
return useCached();
|
||||||
|
|
||||||
FileTransferRequest request(parseURL(url));
|
FileTransferRequest request(ValidURL{url});
|
||||||
request.headers = headers;
|
request.headers = headers;
|
||||||
if (cached)
|
if (cached)
|
||||||
request.expectedETag = getStrAttr(cached->value, "etag");
|
request.expectedETag = getStrAttr(cached->value, "etag");
|
||||||
|
|
@ -109,13 +109,13 @@ DownloadFileResult downloadFile(
|
||||||
static DownloadTarballResult downloadTarball_(
|
static DownloadTarballResult downloadTarball_(
|
||||||
const Settings & settings, const std::string & urlS, const Headers & headers, const std::string & displayPrefix)
|
const Settings & settings, const std::string & urlS, const Headers & headers, const std::string & displayPrefix)
|
||||||
{
|
{
|
||||||
auto url = parseURL(urlS);
|
ValidURL url = urlS;
|
||||||
|
|
||||||
// Some friendly error messages for common mistakes.
|
// Some friendly error messages for common mistakes.
|
||||||
// Namely lets catch when the url is a local file path, but
|
// Namely lets catch when the url is a local file path, but
|
||||||
// it is not in fact a tarball.
|
// it is not in fact a tarball.
|
||||||
if (url.scheme == "file") {
|
if (url.scheme() == "file") {
|
||||||
std::filesystem::path localPath = renderUrlPathEnsureLegal(url.path);
|
std::filesystem::path localPath = renderUrlPathEnsureLegal(url.path());
|
||||||
if (!exists(localPath)) {
|
if (!exists(localPath)) {
|
||||||
throw Error("tarball '%s' does not exist.", localPath);
|
throw Error("tarball '%s' does not exist.", localPath);
|
||||||
}
|
}
|
||||||
|
|
@ -166,7 +166,7 @@ static DownloadTarballResult downloadTarball_(
|
||||||
|
|
||||||
/* Note: if the download is cached, `importTarball()` will receive
|
/* Note: if the download is cached, `importTarball()` will receive
|
||||||
no data, which causes it to import an empty tarball. */
|
no data, which causes it to import an empty tarball. */
|
||||||
auto archive = !url.path.empty() && hasSuffix(toLower(url.path.back()), ".zip") ? ({
|
auto archive = !url.path().empty() && hasSuffix(toLower(url.path().back()), ".zip") ? ({
|
||||||
/* In streaming mode, libarchive doesn't handle
|
/* In streaming mode, libarchive doesn't handle
|
||||||
symlinks in zip files correctly (#10649). So write
|
symlinks in zip files correctly (#10649). So write
|
||||||
the entire file to disk so libarchive can access it
|
the entire file to disk so libarchive can access it
|
||||||
|
|
|
||||||
|
|
@ -37,7 +37,7 @@ static void builtinFetchurl(const BuiltinBuilderContext & ctx)
|
||||||
|
|
||||||
auto fetch = [&](const std::string & url) {
|
auto fetch = [&](const std::string & url) {
|
||||||
auto source = sinkToSource([&](Sink & sink) {
|
auto source = sinkToSource([&](Sink & sink) {
|
||||||
FileTransferRequest request(parseURL(url));
|
FileTransferRequest request(ValidURL{url});
|
||||||
request.decompress = false;
|
request.decompress = false;
|
||||||
|
|
||||||
auto decompressor = makeDecompressionSink(unpack && hasSuffix(mainUrl, ".xz") ? "xz" : "none", sink);
|
auto decompressor = makeDecompressionSink(unpack && hasSuffix(mainUrl, ".xz") ? "xz" : "none", sink);
|
||||||
|
|
|
||||||
|
|
@ -784,7 +784,7 @@ struct curlFileTransfer : public FileTransfer
|
||||||
|
|
||||||
void enqueueItem(std::shared_ptr<TransferItem> item)
|
void enqueueItem(std::shared_ptr<TransferItem> item)
|
||||||
{
|
{
|
||||||
if (item->request.data && item->request.uri.scheme != "http" && item->request.uri.scheme != "https")
|
if (item->request.data && item->request.uri.scheme() != "http" && item->request.uri.scheme() != "https")
|
||||||
throw nix::Error("uploading to '%s' is not supported", item->request.uri.to_string());
|
throw nix::Error("uploading to '%s' is not supported", item->request.uri.to_string());
|
||||||
|
|
||||||
{
|
{
|
||||||
|
|
@ -801,11 +801,11 @@ struct curlFileTransfer : public FileTransfer
|
||||||
void enqueueFileTransfer(const FileTransferRequest & request, Callback<FileTransferResult> callback) override
|
void enqueueFileTransfer(const FileTransferRequest & request, Callback<FileTransferResult> callback) override
|
||||||
{
|
{
|
||||||
/* Ugly hack to support s3:// URIs. */
|
/* Ugly hack to support s3:// URIs. */
|
||||||
if (request.uri.scheme == "s3") {
|
if (request.uri.scheme() == "s3") {
|
||||||
// FIXME: do this on a worker thread
|
// FIXME: do this on a worker thread
|
||||||
try {
|
try {
|
||||||
#if NIX_WITH_S3_SUPPORT
|
#if NIX_WITH_S3_SUPPORT
|
||||||
auto parsed = ParsedS3URL::parse(request.uri);
|
auto parsed = ParsedS3URL::parse(request.uri.parsed());
|
||||||
|
|
||||||
std::string profile = parsed.profile.value_or("");
|
std::string profile = parsed.profile.value_or("");
|
||||||
std::string region = parsed.region.value_or(Aws::Region::US_EAST_1);
|
std::string region = parsed.region.value_or(Aws::Region::US_EAST_1);
|
||||||
|
|
|
||||||
|
|
@ -71,7 +71,7 @@ extern const unsigned int RETRY_TIME_MS_DEFAULT;
|
||||||
|
|
||||||
struct FileTransferRequest
|
struct FileTransferRequest
|
||||||
{
|
{
|
||||||
ParsedURL uri;
|
ValidURL uri;
|
||||||
Headers headers;
|
Headers headers;
|
||||||
std::string expectedETag;
|
std::string expectedETag;
|
||||||
bool verifyTLS = true;
|
bool verifyTLS = true;
|
||||||
|
|
@ -85,8 +85,8 @@ struct FileTransferRequest
|
||||||
std::string mimeType;
|
std::string mimeType;
|
||||||
std::function<void(std::string_view data)> dataCallback;
|
std::function<void(std::string_view data)> dataCallback;
|
||||||
|
|
||||||
FileTransferRequest(ParsedURL uri)
|
FileTransferRequest(ValidURL uri)
|
||||||
: uri(uri)
|
: uri(std::move(uri))
|
||||||
, parentAct(getCurActivity())
|
, parentAct(getCurActivity())
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -341,4 +341,63 @@ ParsedURL fixGitURL(const std::string & url);
|
||||||
*/
|
*/
|
||||||
bool isValidSchemeName(std::string_view scheme);
|
bool isValidSchemeName(std::string_view scheme);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Either a ParsedURL or a verbatim string, but the string must be a valid
|
||||||
|
* ParsedURL. This is necessary because in certain cases URI must be passed
|
||||||
|
* verbatim (e.g. in builtin fetchers), since those are specified by the user.
|
||||||
|
* In those cases normalizations performed by the ParsedURL might be surprising
|
||||||
|
* and undesirable, since Nix must be a universal client that has to work with
|
||||||
|
* various broken services that might interpret URLs in quirky and non-standard ways.
|
||||||
|
*
|
||||||
|
* One of those examples is space-as-plus encoding that is very widespread, but it's
|
||||||
|
* not strictly RFC3986 compliant. We must preserve that information verbatim.
|
||||||
|
*
|
||||||
|
* Though we perform parsing and validation for internal needs.
|
||||||
|
*/
|
||||||
|
struct ValidURL : private ParsedURL
|
||||||
|
{
|
||||||
|
std::optional<std::string> encoded;
|
||||||
|
|
||||||
|
ValidURL(std::string str)
|
||||||
|
: ParsedURL(parseURL(str, /*lenient=*/false))
|
||||||
|
, encoded(std::move(str))
|
||||||
|
{
|
||||||
|
}
|
||||||
|
|
||||||
|
ValidURL(std::string_view str)
|
||||||
|
: ValidURL(std::string{str})
|
||||||
|
{
|
||||||
|
}
|
||||||
|
|
||||||
|
ValidURL(ParsedURL parsed)
|
||||||
|
: ParsedURL{std::move(parsed)}
|
||||||
|
{
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Get the encoded URL (if specified) verbatim or encode the parsed URL.
|
||||||
|
*/
|
||||||
|
std::string to_string() const
|
||||||
|
{
|
||||||
|
return encoded.or_else([&]() -> std::optional<std::string> { return ParsedURL::to_string(); }).value();
|
||||||
|
}
|
||||||
|
|
||||||
|
const ParsedURL & parsed() const &
|
||||||
|
{
|
||||||
|
return *this;
|
||||||
|
}
|
||||||
|
|
||||||
|
std::string_view scheme() const &
|
||||||
|
{
|
||||||
|
return ParsedURL::scheme;
|
||||||
|
}
|
||||||
|
|
||||||
|
const auto & path() const &
|
||||||
|
{
|
||||||
|
return ParsedURL::path;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
std::ostream & operator<<(std::ostream & os, const ValidURL & url);
|
||||||
|
|
||||||
} // namespace nix
|
} // namespace nix
|
||||||
|
|
|
||||||
|
|
@ -434,4 +434,10 @@ bool isValidSchemeName(std::string_view s)
|
||||||
return std::regex_match(s.begin(), s.end(), regex, std::regex_constants::match_default);
|
return std::regex_match(s.begin(), s.end(), regex, std::regex_constants::match_default);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
std::ostream & operator<<(std::ostream & os, const ValidURL & url)
|
||||||
|
{
|
||||||
|
os << url.to_string();
|
||||||
|
return os;
|
||||||
|
}
|
||||||
|
|
||||||
} // namespace nix
|
} // namespace nix
|
||||||
|
|
|
||||||
|
|
@ -105,7 +105,7 @@ std::tuple<StorePath, Hash> prefetchFile(
|
||||||
|
|
||||||
FdSink sink(fd.get());
|
FdSink sink(fd.get());
|
||||||
|
|
||||||
FileTransferRequest req(parseURL(url));
|
FileTransferRequest req(ValidURL{url});
|
||||||
req.decompress = false;
|
req.decompress = false;
|
||||||
getFileTransfer()->download(std::move(req), sink);
|
getFileTransfer()->download(std::move(req), sink);
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -88,3 +88,8 @@ requireDaemonNewerThan "2.20"
|
||||||
expected=100
|
expected=100
|
||||||
if [[ -v NIX_DAEMON_PACKAGE ]]; then expected=1; fi # work around the daemon not returning a 100 status correctly
|
if [[ -v NIX_DAEMON_PACKAGE ]]; then expected=1; fi # work around the daemon not returning a 100 status correctly
|
||||||
expectStderr $expected nix-build --expr '{ url }: builtins.derivation { name = "nix-cache-info"; system = "x86_64-linux"; builder = "builtin:fetchurl"; inherit url; outputHashMode = "flat"; }' --argstr url "file://$narxz" 2>&1 | grep 'must be a fixed-output or impure derivation'
|
expectStderr $expected nix-build --expr '{ url }: builtins.derivation { name = "nix-cache-info"; system = "x86_64-linux"; builder = "builtin:fetchurl"; inherit url; outputHashMode = "flat"; }' --argstr url "file://$narxz" 2>&1 | grep 'must be a fixed-output or impure derivation'
|
||||||
|
|
||||||
|
requireDaemonNewerThan "2.32.0pre20250831"
|
||||||
|
|
||||||
|
expect 1 nix-build --expr 'import <nix/fetchurl.nix>' --argstr name 'name' --argstr url "file://authority.not.allowed/fetchurl.sh?a=1&a=2" --no-out-link |&
|
||||||
|
grepQuiet "error: file:// URL 'file://authority.not.allowed/fetchurl.sh?a=1&a=2' has unexpected authority 'authority.not.allowed'"
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue