mirror of
https://github.com/astral-sh/uv.git
synced 2025-07-07 13:25:00 +00:00
![]() ## Summary This seems to be one of the most consistent benchmark cases we have in terms of standard deviation: ``` ➜ hyperfine "target/profiling/main pip compile scripts/requirements/airflow.in" --runs 200 Benchmark 1: target/profiling/main pip compile scripts/requirements/airflow.in Time (mean ± σ): 292.6 ms ± 6.6 ms [User: 414.1 ms, System: 194.2 ms] Range (min … max): 282.7 ms … 320.1 ms 200 runs ``` For smaller benchmarks, scispacy and dtlssocket seem to be a bit more consistent than our current jupyter benchmark, but it hasn't given us any problems so I'll leave it for now. |
||
---|---|---|
.. | ||
bench | ||
benchmarks | ||
links | ||
packages | ||
popular_packages | ||
requirements | ||
scenarios | ||
workspaces | ||
check_cache_compat.py | ||
check_embedded_python.py | ||
check_system_python.py | ||
release.sh | ||
sync_scenarios.sh | ||
transform_readme.py | ||
update_schemastore.py |