Add progress bar for uv cache clean (#8857)

## Summary

Closes #8786
This commit is contained in:
Jo 2024-11-07 00:43:43 +08:00 committed by GitHub
parent 987d778631
commit dcd24b7c42
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
9 changed files with 171 additions and 127 deletions

View file

@ -320,8 +320,8 @@ impl Cache {
}
/// Clear the cache, removing all entries.
pub fn clear(&self) -> Result<Removal, io::Error> {
rm_rf(&self.root)
pub fn clear(&self, reporter: Option<&dyn CleanReporter>) -> Result<Removal, io::Error> {
rm_rf(&self.root, reporter)
}
/// Remove a package from the cache.
@ -379,7 +379,7 @@ impl Cache {
let path = fs_err::canonicalize(entry.path())?;
if !after.contains(&path) && before.contains(&path) {
debug!("Removing dangling cache entry: {}", path.display());
summary += rm_rf(path)?;
summary += rm_rf(path, None)?;
}
}
}
@ -409,13 +409,13 @@ impl Cache {
if CacheBucket::iter().all(|bucket| entry.file_name() != bucket.to_str()) {
let path = entry.path();
debug!("Removing dangling cache bucket: {}", path.display());
summary += rm_rf(path)?;
summary += rm_rf(path, None)?;
}
} else {
// If the file is not a marker file, remove it.
let path = entry.path();
debug!("Removing dangling cache bucket: {}", path.display());
summary += rm_rf(path)?;
summary += rm_rf(path, None)?;
}
}
@ -427,7 +427,7 @@ impl Cache {
let entry = entry?;
let path = fs_err::canonicalize(entry.path())?;
debug!("Removing dangling cache environment: {}", path.display());
summary += rm_rf(path)?;
summary += rm_rf(path, None)?;
}
}
Err(err) if err.kind() == io::ErrorKind::NotFound => (),
@ -444,7 +444,7 @@ impl Cache {
let path = fs_err::canonicalize(entry.path())?;
if path.is_dir() {
debug!("Removing unzipped wheel entry: {}", path.display());
summary += rm_rf(path)?;
summary += rm_rf(path, None)?;
}
}
}
@ -472,10 +472,10 @@ impl Cache {
if path.is_dir() {
debug!("Removing unzipped built wheel entry: {}", path.display());
summary += rm_rf(path)?;
summary += rm_rf(path, None)?;
} else if path.is_symlink() {
debug!("Removing unzipped built wheel entry: {}", path.display());
summary += rm_rf(path)?;
summary += rm_rf(path, None)?;
}
}
}
@ -505,7 +505,7 @@ impl Cache {
let path = fs_err::canonicalize(entry.path())?;
if !references.contains(&path) {
debug!("Removing dangling cache archive: {}", path.display());
summary += rm_rf(path)?;
summary += rm_rf(path, None)?;
}
}
}
@ -517,6 +517,15 @@ impl Cache {
}
}
pub trait CleanReporter: Send + Sync {
/// Called after one file or directory is removed.
fn on_clean(&self);
/// Called after a package is cleaned.
fn on_clean_package(&self, _package: &str, _removal: &Removal) {}
/// Called after all files and directories are removed.
fn on_complete(&self);
}
/// The different kinds of data in the cache are stored in different bucket, which in our case
/// are subdirectories of the cache root.
#[derive(Debug, Clone, Copy, Eq, PartialEq, Hash)]
@ -800,32 +809,32 @@ impl CacheBucket {
Self::Wheels => {
// For `pypi` wheels, we expect a directory per package (indexed by name).
let root = cache.bucket(self).join(WheelCacheKind::Pypi);
summary += rm_rf(root.join(name.to_string()))?;
summary += rm_rf(root.join(name.to_string()), None)?;
// For alternate indices, we expect a directory for every index (under an `index`
// subdirectory), followed by a directory per package (indexed by name).
let root = cache.bucket(self).join(WheelCacheKind::Index);
for directory in directories(root) {
summary += rm_rf(directory.join(name.to_string()))?;
summary += rm_rf(directory.join(name.to_string()), None)?;
}
// For direct URLs, we expect a directory for every URL, followed by a
// directory per package (indexed by name).
let root = cache.bucket(self).join(WheelCacheKind::Url);
for directory in directories(root) {
summary += rm_rf(directory.join(name.to_string()))?;
summary += rm_rf(directory.join(name.to_string()), None)?;
}
}
Self::SourceDistributions => {
// For `pypi` wheels, we expect a directory per package (indexed by name).
let root = cache.bucket(self).join(WheelCacheKind::Pypi);
summary += rm_rf(root.join(name.to_string()))?;
summary += rm_rf(root.join(name.to_string()), None)?;
// For alternate indices, we expect a directory for every index (under an `index`
// subdirectory), followed by a directory per package (indexed by name).
let root = cache.bucket(self).join(WheelCacheKind::Index);
for directory in directories(root) {
summary += rm_rf(directory.join(name.to_string()))?;
summary += rm_rf(directory.join(name.to_string()), None)?;
}
// For direct URLs, we expect a directory for every URL, followed by a
@ -834,7 +843,7 @@ impl CacheBucket {
let root = cache.bucket(self).join(WheelCacheKind::Url);
for url in directories(root) {
if directories(&url).any(|version| is_match(&version, name)) {
summary += rm_rf(url)?;
summary += rm_rf(url, None)?;
}
}
@ -844,7 +853,7 @@ impl CacheBucket {
let root = cache.bucket(self).join(WheelCacheKind::Path);
for path in directories(root) {
if directories(&path).any(|version| is_match(&version, name)) {
summary += rm_rf(path)?;
summary += rm_rf(path, None)?;
}
}
@ -855,7 +864,7 @@ impl CacheBucket {
for repository in directories(root) {
for sha in directories(repository) {
if is_match(&sha, name) {
summary += rm_rf(sha)?;
summary += rm_rf(sha, None)?;
}
}
}
@ -863,20 +872,20 @@ impl CacheBucket {
Self::Simple => {
// For `pypi` wheels, we expect a rkyv file per package, indexed by name.
let root = cache.bucket(self).join(WheelCacheKind::Pypi);
summary += rm_rf(root.join(format!("{name}.rkyv")))?;
summary += rm_rf(root.join(format!("{name}.rkyv")), None)?;
// For alternate indices, we expect a directory for every index (under an `index`
// subdirectory), followed by a directory per package (indexed by name).
let root = cache.bucket(self).join(WheelCacheKind::Index);
for directory in directories(root) {
summary += rm_rf(directory.join(format!("{name}.rkyv")))?;
summary += rm_rf(directory.join(format!("{name}.rkyv")), None)?;
}
}
Self::FlatIndex => {
// We can't know if the flat index includes a package, so we just remove the entire
// cache entry.
let root = cache.bucket(self);
summary += rm_rf(root)?;
summary += rm_rf(root, None)?;
}
Self::Git => {
// Nothing to do.

View file

@ -5,11 +5,13 @@
use std::io;
use std::path::Path;
use crate::CleanReporter;
/// Remove a file or directory and all its contents, returning a [`Removal`] with
/// the number of files and directories removed, along with a total byte count.
pub fn rm_rf(path: impl AsRef<Path>) -> io::Result<Removal> {
pub fn rm_rf(path: impl AsRef<Path>, reporter: Option<&dyn CleanReporter>) -> io::Result<Removal> {
let mut removal = Removal::default();
removal.rm_rf(path.as_ref())?;
removal.rm_rf(path.as_ref(), reporter)?;
Ok(removal)
}
@ -28,7 +30,7 @@ pub struct Removal {
impl Removal {
/// Recursively remove a file or directory and all its contents.
fn rm_rf(&mut self, path: &Path) -> io::Result<()> {
fn rm_rf(&mut self, path: &Path, reporter: Option<&dyn CleanReporter>) -> io::Result<()> {
let metadata = match fs_err::symlink_metadata(path) {
Ok(metadata) => metadata,
Err(err) if err.kind() == io::ErrorKind::NotFound => return Ok(()),
@ -47,6 +49,8 @@ impl Removal {
remove_file(path)?;
}
reporter.map(CleanReporter::on_clean);
return Ok(());
}
@ -61,7 +65,7 @@ impl Removal {
if set_readable(dir).unwrap_or(false) {
// Retry the operation; if we _just_ `self.rm_rf(dir)` and continue,
// `walkdir` may give us duplicate entries for the directory.
return self.rm_rf(path);
return self.rm_rf(path, reporter);
}
}
}
@ -88,8 +92,12 @@ impl Removal {
}
remove_file(entry.path())?;
}
reporter.map(CleanReporter::on_clean);
}
reporter.map(CleanReporter::on_complete);
Ok(())
}
}