mirror of
https://github.com/astral-sh/uv.git
synced 2025-07-07 05:15:00 +00:00
Cache downloaded wheel when range requests aren't supported (#5089)
## Summary When range requests aren't supported, we fall back to streaming the wheel, stopping as soon as we hit a `METADATA` file. This is a small optimization, but the downside is that we don't get to cache the resulting wheel... We don't know whether `METADATA` will be at the beginning or end of the wheel, but it _seems_ like a better tradeoff to download and cache the entire wheel? Closes: https://github.com/astral-sh/uv/issues/5088. Sort of a revert of: https://github.com/astral-sh/uv/pull/1792.
This commit is contained in:
parent
0e7df9623d
commit
2ff3b380b1
3 changed files with 63 additions and 114 deletions
|
@ -4,18 +4,17 @@ use std::path::{Path, PathBuf};
|
||||||
use std::str::FromStr;
|
use std::str::FromStr;
|
||||||
|
|
||||||
use async_http_range_reader::AsyncHttpRangeReader;
|
use async_http_range_reader::AsyncHttpRangeReader;
|
||||||
use futures::{FutureExt, TryStreamExt};
|
use futures::FutureExt;
|
||||||
use http::HeaderMap;
|
use http::HeaderMap;
|
||||||
use reqwest::{Client, Response, StatusCode};
|
use reqwest::{Client, Response, StatusCode};
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
use tokio::io::AsyncReadExt;
|
use tokio_util::compat::TokioAsyncReadCompatExt;
|
||||||
use tokio_util::compat::{FuturesAsyncReadCompatExt, TokioAsyncReadCompatExt};
|
|
||||||
use tracing::{info_span, instrument, trace, warn, Instrument};
|
use tracing::{info_span, instrument, trace, warn, Instrument};
|
||||||
use url::Url;
|
use url::Url;
|
||||||
|
|
||||||
use distribution_filename::{DistFilename, SourceDistFilename, WheelFilename};
|
use distribution_filename::{DistFilename, SourceDistFilename, WheelFilename};
|
||||||
use distribution_types::{BuiltDist, File, FileLocation, IndexUrl, IndexUrls, Name};
|
use distribution_types::{BuiltDist, File, FileLocation, IndexUrl, IndexUrls, Name};
|
||||||
use install_wheel_rs::metadata::{find_archive_dist_info, is_metadata_entry};
|
use install_wheel_rs::metadata::find_archive_dist_info;
|
||||||
use pep440_rs::Version;
|
use pep440_rs::Version;
|
||||||
use pep508_rs::MarkerEnvironment;
|
use pep508_rs::MarkerEnvironment;
|
||||||
use platform_tags::Platform;
|
use platform_tags::Platform;
|
||||||
|
@ -598,8 +597,7 @@ impl RegistryClient {
|
||||||
.instrument(info_span!("read_metadata_range_request", wheel = %filename))
|
.instrument(info_span!("read_metadata_range_request", wheel = %filename))
|
||||||
};
|
};
|
||||||
|
|
||||||
let result = self
|
self.cached_client()
|
||||||
.cached_client()
|
|
||||||
.get_serde(
|
.get_serde(
|
||||||
req,
|
req,
|
||||||
&cache_entry,
|
&cache_entry,
|
||||||
|
@ -607,67 +605,8 @@ impl RegistryClient {
|
||||||
read_metadata_range_request,
|
read_metadata_range_request,
|
||||||
)
|
)
|
||||||
.await
|
.await
|
||||||
.map_err(crate::Error::from);
|
|
||||||
|
|
||||||
match result {
|
|
||||||
Ok(metadata) => return Ok(metadata),
|
|
||||||
Err(err) => {
|
|
||||||
if err.is_http_range_requests_unsupported() {
|
|
||||||
// The range request version failed. Fall back to streaming the file to search
|
|
||||||
// for the METADATA file.
|
|
||||||
warn!("Range requests not supported for {filename}; streaming wheel");
|
|
||||||
} else {
|
|
||||||
return Err(err);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
// Create a request to stream the file.
|
|
||||||
let req = self
|
|
||||||
.uncached_client()
|
|
||||||
.get(url.clone())
|
|
||||||
.header(
|
|
||||||
// `reqwest` defaults to accepting compressed responses.
|
|
||||||
// Specify identity encoding to get consistent .whl downloading
|
|
||||||
// behavior from servers. ref: https://github.com/pypa/pip/pull/1688
|
|
||||||
"accept-encoding",
|
|
||||||
reqwest::header::HeaderValue::from_static("identity"),
|
|
||||||
)
|
|
||||||
.build()
|
|
||||||
.map_err(ErrorKind::from)?;
|
|
||||||
|
|
||||||
// Stream the file, searching for the METADATA.
|
|
||||||
let read_metadata_stream = |response: Response| {
|
|
||||||
async {
|
|
||||||
let reader = response
|
|
||||||
.bytes_stream()
|
|
||||||
.map_err(|err| self.handle_response_errors(err))
|
|
||||||
.into_async_read();
|
|
||||||
|
|
||||||
read_metadata_async_stream(filename, url.to_string(), reader).await
|
|
||||||
}
|
|
||||||
.instrument(info_span!("read_metadata_stream", wheel = %filename))
|
|
||||||
};
|
|
||||||
|
|
||||||
self.cached_client()
|
|
||||||
.get_serde(req, &cache_entry, cache_control, read_metadata_stream)
|
|
||||||
.await
|
|
||||||
.map_err(crate::Error::from)
|
.map_err(crate::Error::from)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Handle a specific `reqwest` error, and convert it to [`io::Error`].
|
|
||||||
fn handle_response_errors(&self, err: reqwest::Error) -> std::io::Error {
|
|
||||||
if err.is_timeout() {
|
|
||||||
std::io::Error::new(
|
|
||||||
std::io::ErrorKind::TimedOut,
|
|
||||||
format!(
|
|
||||||
"Failed to download distribution due to network timeout. Try increasing UV_HTTP_TIMEOUT (current value: {}s).", self.timeout()
|
|
||||||
),
|
|
||||||
)
|
|
||||||
} else {
|
|
||||||
std::io::Error::new(std::io::ErrorKind::Other, err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Read a wheel's `METADATA` file from a zip file.
|
/// Read a wheel's `METADATA` file from a zip file.
|
||||||
|
@ -708,50 +647,6 @@ async fn read_metadata_async_seek(
|
||||||
Ok(metadata)
|
Ok(metadata)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Like [`read_metadata_async_seek`], but doesn't use seek.
|
|
||||||
async fn read_metadata_async_stream<R: futures::AsyncRead + Unpin>(
|
|
||||||
filename: &WheelFilename,
|
|
||||||
debug_source: String,
|
|
||||||
reader: R,
|
|
||||||
) -> Result<Metadata23, Error> {
|
|
||||||
let reader = futures::io::BufReader::with_capacity(128 * 1024, reader);
|
|
||||||
let mut zip = async_zip::base::read::stream::ZipFileReader::new(reader);
|
|
||||||
|
|
||||||
while let Some(mut entry) = zip
|
|
||||||
.next_with_entry()
|
|
||||||
.await
|
|
||||||
.map_err(|err| ErrorKind::Zip(filename.clone(), err))?
|
|
||||||
{
|
|
||||||
// Find the `METADATA` entry.
|
|
||||||
let path = entry
|
|
||||||
.reader()
|
|
||||||
.entry()
|
|
||||||
.filename()
|
|
||||||
.as_str()
|
|
||||||
.map_err(|err| ErrorKind::Zip(filename.clone(), err))?;
|
|
||||||
|
|
||||||
if is_metadata_entry(path, filename) {
|
|
||||||
let mut reader = entry.reader_mut().compat();
|
|
||||||
let mut contents = Vec::new();
|
|
||||||
reader.read_to_end(&mut contents).await.unwrap();
|
|
||||||
|
|
||||||
let metadata = Metadata23::parse_metadata(&contents).map_err(|err| {
|
|
||||||
ErrorKind::MetadataParseError(filename.clone(), debug_source, Box::new(err))
|
|
||||||
})?;
|
|
||||||
return Ok(metadata);
|
|
||||||
}
|
|
||||||
|
|
||||||
// Close current file to get access to the next one. See docs:
|
|
||||||
// https://docs.rs/async_zip/0.0.16/async_zip/base/read/stream/
|
|
||||||
zip = entry
|
|
||||||
.skip()
|
|
||||||
.await
|
|
||||||
.map_err(|err| ErrorKind::Zip(filename.clone(), err))?;
|
|
||||||
}
|
|
||||||
|
|
||||||
Err(ErrorKind::MetadataNotFound(filename.clone(), debug_source).into())
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(
|
#[derive(
|
||||||
Default, Debug, Serialize, Deserialize, rkyv::Archive, rkyv::Deserialize, rkyv::Serialize,
|
Default, Debug, Serialize, Deserialize, rkyv::Archive, rkyv::Deserialize, rkyv::Serialize,
|
||||||
)]
|
)]
|
||||||
|
|
|
@ -208,6 +208,31 @@ impl<'a, Context: BuildContext> DistributionDatabase<'a, Context> {
|
||||||
hashes: archive.hashes,
|
hashes: archive.hashes,
|
||||||
filename: wheel.filename.clone(),
|
filename: wheel.filename.clone(),
|
||||||
}),
|
}),
|
||||||
|
Err(Error::Client(err)) if err.is_http_streaming_unsupported() => {
|
||||||
|
warn!(
|
||||||
|
"Streaming unsupported for {dist}; downloading wheel to disk ({err})"
|
||||||
|
);
|
||||||
|
|
||||||
|
// If the request failed because streaming is unsupported, download the
|
||||||
|
// wheel directly.
|
||||||
|
let archive = self
|
||||||
|
.download_wheel(
|
||||||
|
url,
|
||||||
|
&wheel.filename,
|
||||||
|
wheel.file.size,
|
||||||
|
&wheel_entry,
|
||||||
|
dist,
|
||||||
|
hashes,
|
||||||
|
)
|
||||||
|
.await?;
|
||||||
|
|
||||||
|
Ok(LocalWheel {
|
||||||
|
dist: Dist::Built(dist.clone()),
|
||||||
|
archive: self.build_context.cache().archive(&archive.id),
|
||||||
|
hashes: archive.hashes,
|
||||||
|
filename: wheel.filename.clone(),
|
||||||
|
})
|
||||||
|
}
|
||||||
Err(Error::Extract(err)) => {
|
Err(Error::Extract(err)) => {
|
||||||
if err.is_http_streaming_unsupported() {
|
if err.is_http_streaming_unsupported() {
|
||||||
warn!(
|
warn!(
|
||||||
|
@ -293,6 +318,36 @@ impl<'a, Context: BuildContext> DistributionDatabase<'a, Context> {
|
||||||
filename: wheel.filename.clone(),
|
filename: wheel.filename.clone(),
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
Err(Error::Extract(err)) => {
|
||||||
|
if err.is_http_streaming_unsupported() {
|
||||||
|
warn!(
|
||||||
|
"Streaming unsupported for {dist}; downloading wheel to disk ({err})"
|
||||||
|
);
|
||||||
|
} else if err.is_http_streaming_failed() {
|
||||||
|
warn!("Streaming failed for {dist}; downloading wheel to disk ({err})");
|
||||||
|
} else {
|
||||||
|
return Err(Error::Extract(err));
|
||||||
|
}
|
||||||
|
|
||||||
|
// If the request failed because streaming is unsupported, download the
|
||||||
|
// wheel directly.
|
||||||
|
let archive = self
|
||||||
|
.download_wheel(
|
||||||
|
wheel.url.raw().clone(),
|
||||||
|
&wheel.filename,
|
||||||
|
None,
|
||||||
|
&wheel_entry,
|
||||||
|
dist,
|
||||||
|
hashes,
|
||||||
|
)
|
||||||
|
.await?;
|
||||||
|
Ok(LocalWheel {
|
||||||
|
dist: Dist::Built(dist.clone()),
|
||||||
|
archive: self.build_context.cache().archive(&archive.id),
|
||||||
|
hashes: archive.hashes,
|
||||||
|
filename: wheel.filename.clone(),
|
||||||
|
})
|
||||||
|
}
|
||||||
Err(err) => Err(err),
|
Err(err) => Err(err),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -390,11 +445,11 @@ impl<'a, Context: BuildContext> DistributionDatabase<'a, Context> {
|
||||||
|
|
||||||
match result {
|
match result {
|
||||||
Ok(metadata) => Ok(ArchiveMetadata::from_metadata23(metadata)),
|
Ok(metadata) => Ok(ArchiveMetadata::from_metadata23(metadata)),
|
||||||
Err(err) if err.is_http_streaming_unsupported() => {
|
Err(err) if err.is_http_range_requests_unsupported() => {
|
||||||
warn!("Streaming unsupported when fetching metadata for {dist}; downloading wheel directly ({err})");
|
warn!("Range requests unsupported when fetching metadata for {dist}; downloading wheel directly ({err})");
|
||||||
|
|
||||||
// If the request failed due to an error that could be resolved by
|
// If the request failed due to an error that could be resolved by downloading the
|
||||||
// downloading the wheel directly, try that.
|
// wheel directly, try that.
|
||||||
let wheel = self.get_wheel(dist, hashes).await?;
|
let wheel = self.get_wheel(dist, hashes).await?;
|
||||||
let metadata = wheel.metadata()?;
|
let metadata = wheel.metadata()?;
|
||||||
let hashes = wheel.hashes;
|
let hashes = wheel.hashes;
|
||||||
|
|
|
@ -3227,7 +3227,6 @@ fn no_stream() -> Result<()> {
|
||||||
|
|
||||||
----- stderr -----
|
----- stderr -----
|
||||||
Resolved 1 package in [TIME]
|
Resolved 1 package in [TIME]
|
||||||
Prepared 1 package in [TIME]
|
|
||||||
Installed 1 package in [TIME]
|
Installed 1 package in [TIME]
|
||||||
+ hashb-foxglove-protocolbuffers-python==25.3.0.1.20240226043130+465630478360
|
+ hashb-foxglove-protocolbuffers-python==25.3.0.1.20240226043130+465630478360
|
||||||
"###
|
"###
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue