
Hello there! First real docs PR for uv. 1. I expect this will be rewritten a gazillion times to have a consistent tone with the rest of the docs, despite me trying to stick to it as best as I could. Feel free to edit! 2. I went super on the verbose mode, while also providing a callout with a TLDR on top. Scrap anything you feel it's redundant! 3. I placed the guide under `integrations` since Charlie added the FastAPI integration there. ## Summary <!-- What's the purpose of the change? What does it do, and why? --> Addresses #5945 ## Test Plan <!-- How was it tested? --> I just looked at the docs on the dev server of mkdocs if it looked nice. **I could not test the commands that I wrote work** outside of macOS. If someone among contributors has a Windows/Linux laptop, it should be enough, even for the GPU-supported versions: I expect the installation will just break once torch checks for CUDA (perhaps even at runtime). --------- Co-authored-by: Santiago Castro <bryant1410@gmail.com> Co-authored-by: Charlie Marsh <charlie.r.marsh@gmail.com>
6.2 KiB
Installing PyTorch with uv
PyTorch is a popular open-source deep learning framework that has
first-class support for acceleration via GPUs. Installation, however, can be complex, as you won't
find all the wheels for PyTorch on PyPI and you have to manage this through external indexes. This
guide aims to help you set up a pyproject.toml
file using uv
indexes features.
!!! info "Available PyTorch indexes"
By default, from PyPI you will download:
* CPU-only wheels on Windows and macOS.
* CUDA (i.e. GPU) on Linux. At the time of writing, for PyTorch stable version (2.5.0) this defaults to CUDA 12.4. On older versions, this might be different.
If you want to install CPU-only PyTorch wheels on Linux, or a PyTorch version that supports a different CUDA version, you need to resort to external indexes:
* https://download.pytorch.org/whl/cu118
* https://download.pytorch.org/whl/cu121
* https://download.pytorch.org/whl/cu124
* https://download.pytorch.org/whl/cpu
* https://download.pytorch.org/whl/rocm6.2 (AMD GPUs, only available on Linux)
The PyTorch website offers a simple interface to
determine what pip
command you should run to install PyTorch. This guide should help you do the
same with uv
. If the following instructions fail, you might want to check the link for any
difference, and open an issue to report the discrepancy. In this case, or if you run into any other
issue, please refer to Getting Help.
Initialise a project
Create a new project with uv init
:
uv init
!!! tip "Supported Python versions"
Make sure to use a Python version that is supported by PyTorch. You can find the compatibility matrix [here](https://github.com/pytorch/pytorch/blob/main/RELEASE.md#release-compatibility-matrix).
Add PyTorch indexes
Open the pyproject.toml
file, and create a custom index matching the channel you want to use (CPU,
NVIDIA or AMD), to instruct uv
where to find the PyTorch wheels.
=== "CPU-only"
```toml
[[tool.uv.index]]
name = "pytorch-cpu"
url = "https://download.pytorch.org/whl/cpu"
explicit = true
```
=== "CUDA 11.8"
```toml
[[tool.uv.index]]
name = "pytorch-cu118"
url = "https://download.pytorch.org/whl/cu118"
explicit = true
```
=== "CUDA 12.1"
```toml
[[tool.uv.index]]
name = "pytorch-cu121"
url = "https://download.pytorch.org/whl/cu121"
explicit = true
```
=== "CUDA 12.4"
```toml
[[tool.uv.index]]
name = "pytorch-cu124"
url = "https://download.pytorch.org/whl/cu124"
explicit = true
```
=== "ROCm6"
```toml
[[tool.uv.index]]
name = "pytorch-rocm"
url = "https://download.pytorch.org/whl/rocm6.2"
explicit = true
```
Note the explicit
option: this prevents packages from being installed from that index unless
explicitly pinned to it (see the step below). This means that only PyTorch will be installed from
this index, while all other packages will be looked up on PyPI (the default index).
Pin PyTorch to the custom index
Now you need to pin specific PyTorch versions to the appropriate indexes. The tool.uv.sources
table in the pyproject.toml
is a mapping that matches a package to a list of indexes inside of
which to look. Note that you need to create a new entry for every library you want to install. In
other words, if your project depends on both PyTorch and torchvision, you need to do as follows:
=== "CPU-only"
Note that we use the `platform_system` marker to instruct `uv` to look into this index on Linux and Windows.
```toml
[tool.uv.sources]
torch = [
{ index = "pytorch-cpu", marker = "platform_system != 'Darwin'"},
]
torchvision = [
{ index = "pytorch-cpu", marker = "platform_system != 'Darwin'"},
]
```
=== "CUDA 11.8"
Note that we use the `platform_system` marker to instruct `uv` to look into this index on Linux and Windows.
```toml
[tool.uv.sources]
torch = [
{ index = "pytorch-cu118", marker = "platform_system != 'Darwin'"},
]
torchvision = [
{ index = "pytorch-cu118", marker = "platform_system != 'Darwin'"},
]
```
=== "CUDA 12.1"
Note that we use the `platform_system` marker to instruct `uv` to look into this index on Linux and Windows.
```toml
[tool.uv.sources]
torch = [
{ index = "pytorch-cu121", marker = "platform_system != 'Darwin'"},
]
torchvision = [
{ index = "pytorch-cu121", marker = "platform_system != 'Darwin'"},
]
```
=== "CUDA 12.4"
Note that we use the `platform_system` marker to instruct `uv` to look into this index on Linux and Windows.
```toml
[tool.uv.sources]
torch = [
{ index = "pytorch-cu124", marker = "platform_system != 'Darwin'"},
]
torchvision = [
{ index = "pytorch-cu124", marker = "platform_system != 'Darwin'"},
]
```
=== "ROCm6"
Note that we use the `platform_system` marker to instruct `uv` to look into this index when on Linux only.
```toml
[tool.uv.sources]
torch = [
{ index = "pytorch-rocm", marker = "platform_system == 'Linux'"},
]
torchvision = [
{ index = "pytorch-rocm", marker = "platform_system == 'Linux'"},
]
```
Add PyTorch to your dependencies
Finally, we can add PyTorch to the project.dependencies
section of the pyproject.toml
. To
install PyTorch and torchvision, run:
uv add torch torchvision
However, if you want to be more explicit, you could also:
[project]
dependencies = [
"torch==2.5.0 ; platform_system == 'Darwin'",
"torch==2.5.0+cu124 ; platform_system != 'Darwin'",
"torchvision==0.20.0 ; platform_system == 'Darwin'",
"torchvision==0.20.0+cu124 ; platform_system != 'Darwin'",
]
This will install PyTorch 2.5.0 and torchvision 0.19 on macOS, and PyTorch 2.5.0+cu124 with torchvision 0.20.0+cu124 on Linux and Windows.
!!! warning "PyTorch on Intel Macs"
Note that the last minor version to support Intel Macs is PyTorch 2.2. In other words, if you try to install PyTorch 2.3.0 or more on an Intel Mac, you will get an error. Alternatively, you can install PyTorch with conda