Skip to main content

Tensors and Dynamic neural networks in Python with strong GPU acceleration

Project description

PyTorch Logo


PyTorch is a Python package that provides two high-level features:

  • Tensor computation (like NumPy) with strong GPU acceleration
  • Deep neural networks built on a tape-based autograd system

You can reuse your favorite Python packages such as NumPy, SciPy, and Cython to extend PyTorch when needed.

Our trunk health (Continuous Integration signals) can be found at hud.pytorch.org.

More About PyTorch

Learn the basics of PyTorch

At a granular level, PyTorch is a library that consists of the following components:

Component Description
torch A Tensor library like NumPy, with strong GPU support
torch.autograd A tape-based automatic differentiation library that supports all differentiable Tensor operations in torch
torch.jit A compilation stack (TorchScript) to create serializable and optimizable models from PyTorch code
torch.nn A neural networks library deeply integrated with autograd designed for maximum flexibility
torch.multiprocessing Python multiprocessing, but with magical memory sharing of torch Tensors across processes. Useful for data loading and Hogwild training
torch.utils DataLoader and other utility functions for convenience

Usually, PyTorch is used either as:

  • A replacement for NumPy to use the power of GPUs.
  • A deep learning research platform that provides maximum flexibility and speed.

Elaborating Further:

A GPU-Ready Tensor Library

If you use NumPy, then you have used Tensors (a.k.a. ndarray).

Tensor illustration

PyTorch provides Tensors that can live either on the CPU or the GPU and accelerates the computation by a huge amount.

We provide a wide variety of tensor routines to accelerate and fit your scientific computation needs such as slicing, indexing, mathematical operations, linear algebra, reductions. And they are fast!

Dynamic Neural Networks: Tape-Based Autograd

PyTorch has a unique way of building neural networks: using and replaying a tape recorder.

Most frameworks such as TensorFlow, Theano, Caffe, and CNTK have a static view of the world. One has to build a neural network and reuse the same structure again and again. Changing the way the network behaves means that one has to start from scratch.

With PyTorch, we use a technique called reverse-mode auto-differentiation, which allows you to change the way your network behaves arbitrarily with zero lag or overhead. Our inspiration comes from several research papers on this topic, as well as current and past work such as torch-autograd, autograd, Chainer, etc.

While this technique is not unique to PyTorch, it's one of the fastest implementations of it to date. You get the best of speed and flexibility for your crazy research.

Dynamic graph

Python First

PyTorch is not a Python binding into a monolithic C++ framework. It is built to be deeply integrated into Python. You can use it naturally like you would use NumPy / SciPy / scikit-learn etc. You can write your new neural network layers in Python itself, using your favorite libraries and use packages such as Cython and Numba. Our goal is to not reinvent the wheel where appropriate.

Imperative Experiences

PyTorch is designed to be intuitive, linear in thought, and easy to use. When you execute a line of code, it gets executed. There isn't an asynchronous view of the world. When you drop into a debugger or receive error messages and stack traces, understanding them is straightforward. The stack trace points to exactly where your code was defined. We hope you never spend hours debugging your code because of bad stack traces or asynchronous and opaque execution engines.

Fast and Lean

PyTorch has minimal framework overhead. We integrate acceleration libraries such as Intel MKL and NVIDIA (cuDNN, NCCL) to maximize speed. At the core, its CPU and GPU Tensor and neural network backends are mature and have been tested for years.

Hence, PyTorch is quite fast — whether you run small or large neural networks.

The memory usage in PyTorch is extremely efficient compared to Torch or some of the alternatives. We've written custom memory allocators for the GPU to make sure that your deep learning models are maximally memory efficient. This enables you to train bigger deep learning models than before.

Extensions Without Pain

Writing new neural network modules, or interfacing with PyTorch's Tensor API was designed to be straightforward and with minimal abstractions.

You can write new neural network layers in Python using the torch API or your favorite NumPy-based libraries such as SciPy.

If you want to write your layers in C/C++, we provide a convenient extension API that is efficient and with minimal boilerplate. No wrapper code needs to be written. You can see a tutorial here and an example here.

Installation

Binaries

Commands to install binaries via Conda or pip wheels are on our website: https://pytorch.org/get-started/locally/

NVIDIA Jetson Platforms

Python wheels for NVIDIA's Jetson Nano, Jetson TX1/TX2, Jetson Xavier NX/AGX, and Jetson AGX Orin are provided here and the L4T container is published here

They require JetPack 4.2 and above, and @dusty-nv and @ptrblck are maintaining them.

From Source

Prerequisites

If you are installing from source, you will need:

  • Python 3.9 or later
  • A compiler that fully supports C++17, such as clang or gcc (gcc 9.4.0 or newer is required, on Linux)
  • Visual Studio or Visual Studio Build Tool (Windows only)

* PyTorch CI uses Visual C++ BuildTools, which come with Visual Studio Enterprise, Professional, or Community Editions. You can also install the build tools from https://visualstudio.microsoft.com/visual-cpp-build-tools/. The build tools do not come with Visual Studio Code by default.

An example of environment setup is shown below:

  • Linux:
$ source <CONDA_INSTALL_DIR>/bin/activate
$ conda create -y -n <CONDA_NAME>
$ conda activate <CONDA_NAME>
  • Windows:
$ source <CONDA_INSTALL_DIR>\Scripts\activate.bat
$ conda create -y -n <CONDA_NAME>
$ conda activate <CONDA_NAME>
$ call "C:\Program Files\Microsoft Visual Studio\<VERSION>\Community\VC\Auxiliary\Build\vcvarsall.bat" x64

A conda environment is not required. You can also do a PyTorch build in a standard virtual environment, e.g., created with tools like uv, provided your system has installed all the necessary dependencies unavailable as pip packages (e.g., CUDA, MKL.)

NVIDIA CUDA Support

If you want to compile with CUDA support, select a supported version of CUDA from our support matrix, then install the following:

Note: You could refer to the cuDNN Support Matrix for cuDNN versions with the various supported CUDA, CUDA driver and NVIDIA hardware

If you want to disable CUDA support, export the environment variable USE_CUDA=0. Other potentially useful environment variables may be found in setup.py. If CUDA is installed in a non-standard location, set PATH so that the nvcc you want to use can be found (e.g., export PATH=/usr/local/cuda-12.8/bin:$PATH).

If you are building for NVIDIA's Jetson platforms (Jetson Nano, TX1, TX2, AGX Xavier), Instructions to install PyTorch for Jetson Nano are available here

AMD ROCm Support

If you want to compile with ROCm support, install

  • AMD ROCm 4.0 and above installation
  • ROCm is currently supported only for Linux systems.

By default the build system expects ROCm to be installed in /opt/rocm. If ROCm is installed in a different directory, the ROCM_PATH environment variable must be set to the ROCm installation directory. The build system automatically detects the AMD GPU architecture. Optionally, the AMD GPU architecture can be explicitly set with the PYTORCH_ROCM_ARCH environment variable AMD GPU architecture

If you want to disable ROCm support, export the environment variable USE_ROCM=0. Other potentially useful environment variables may be found in setup.py.

Intel GPU Support

If you want to compile with Intel GPU support, follow these

If you want to disable Intel GPU support, export the environment variable USE_XPU=0. Other potentially useful environment variables may be found in setup.py.

Get the PyTorch Source

git clone https://github.com/pytorch/pytorch
cd pytorch
# if you are updating an existing checkout
git submodule sync
git submodule update --init --recursive

Install Dependencies

Common

conda install cmake ninja
# Run this command from the PyTorch directory after cloning the source code using the “Get the PyTorch Source“ section below
pip install -r requirements.txt

On Linux

pip install mkl-static mkl-include
# CUDA only: Add LAPACK support for the GPU if needed
# magma installation: run with active conda environment. specify CUDA version to install
.ci/docker/common/install_magma_conda.sh 12.4

# (optional) If using torch.compile with inductor/triton, install the matching version of triton
# Run from the pytorch directory after cloning
# For Intel GPU support, please explicitly `export USE_XPU=1` before running command.
make triton

On MacOS

# Add this package on intel x86 processor machines only
pip install mkl-static mkl-include
# Add these packages if torch.distributed is needed
conda install pkg-config libuv

On Windows

pip install mkl-static mkl-include
# Add these packages if torch.distributed is needed.
# Distributed package support on Windows is a prototype feature and is subject to changes.
conda install -c conda-forge libuv=1.39

Install PyTorch

On Linux

If you're compiling for AMD ROCm then first run this command:

# Only run this if you're compiling for ROCm
python tools/amd_build/build_amd.py

Install PyTorch

export CMAKE_PREFIX_PATH="${CONDA_PREFIX:-'$(dirname $(which conda))/../'}:${CMAKE_PREFIX_PATH}"
python setup.py develop

On macOS

python3 setup.py develop

On Windows

If you want to build legacy python code, please refer to Building on legacy code and CUDA

CPU-only builds

In this mode PyTorch computations will run on your CPU, not your GPU.

python setup.py develop

Note on OpenMP: The desired OpenMP implementation is Intel OpenMP (iomp). In order to link against iomp, you'll need to manually download the library and set up the building environment by tweaking CMAKE_INCLUDE_PATH and LIB. The instruction here is an example for setting up both MKL and Intel OpenMP. Without these configurations for CMake, Microsoft Visual C OpenMP runtime (vcomp) will be used.

CUDA based build

In this mode PyTorch computations will leverage your GPU via CUDA for faster number crunching

NVTX is needed to build Pytorch with CUDA. NVTX is a part of CUDA distributive, where it is called "Nsight Compute". To install it onto an already installed CUDA run CUDA installation once again and check the corresponding checkbox. Make sure that CUDA with Nsight Compute is installed after Visual Studio.

Currently, VS 2017 / 2019, and Ninja are supported as the generator of CMake. If ninja.exe is detected in PATH, then Ninja will be used as the default generator, otherwise, it will use VS 2017 / 2019.
If Ninja is selected as the generator, the latest MSVC will get selected as the underlying toolchain.

Additional libraries such as Magma, oneDNN, a.k.a. MKLDNN or DNNL, and Sccache are often needed. Please refer to the installation-helper to install them.

You can refer to the build_pytorch.bat script for some other environment variables configurations

cmd

:: Set the environment variables after you have downloaded and unzipped the mkl package,
:: else CMake would throw an error as `Could NOT find OpenMP`.
set CMAKE_INCLUDE_PATH={Your directory}\mkl\include
set LIB={Your directory}\mkl\lib;%LIB%

:: Read the content in the previous section carefully before you proceed.
:: [Optional] If you want to override the underlying toolset used by Ninja and Visual Studio with CUDA, please run the following script block.
:: "Visual Studio 2019 Developer Command Prompt" will be run automatically.
:: Make sure you have CMake >= 3.12 before you do this when you use the Visual Studio generator.
set CMAKE_GENERATOR_TOOLSET_VERSION=14.27
set DISTUTILS_USE_SDK=1
for /f "usebackq tokens=*" %i in (`"%ProgramFiles(x86)%\Microsoft Visual Studio\Installer\vswhere.exe" -version [15^,17^) -products * -latest -property installationPath`) do call "%i\VC\Auxiliary\Build\vcvarsall.bat" x64 -vcvars_ver=%CMAKE_GENERATOR_TOOLSET_VERSION%

:: [Optional] If you want to override the CUDA host compiler
set CUDAHOSTCXX=C:\Program Files (x86)\Microsoft Visual Studio\2019\Community\VC\Tools\MSVC\14.27.29110\bin\HostX64\x64\cl.exe

python setup.py develop

Intel GPU builds

In this mode PyTorch with Intel GPU support will be built.

Please make sure the common prerequisites as well as the prerequisites for Intel GPU are properly installed and the environment variables are configured prior to starting the build. For build tool support, Visual Studio 2022 is required.

Then PyTorch can be built with the command:

:: CMD Commands:
:: Set the CMAKE_PREFIX_PATH to help find corresponding packages
:: %CONDA_PREFIX% only works after `conda activate custom_env`

if defined CMAKE_PREFIX_PATH (
    set "CMAKE_PREFIX_PATH=%CONDA_PREFIX%\Library;%CMAKE_PREFIX_PATH%"
) else (
    set "CMAKE_PREFIX_PATH=%CONDA_PREFIX%\Library"
)

python setup.py develop
Adjust Build Options (Optional)

You can adjust the configuration of cmake variables optionally (without building first), by doing the following. For example, adjusting the pre-detected directories for CuDNN or BLAS can be done with such a step.

On Linux

export CMAKE_PREFIX_PATH="${CONDA_PREFIX:-'$(dirname $(which conda))/../'}:${CMAKE_PREFIX_PATH}"
CMAKE_ONLY=1 python setup.py build
ccmake build  # or cmake-gui build

On macOS

export CMAKE_PREFIX_PATH="${CONDA_PREFIX:-'$(dirname $(which conda))/../'}:${CMAKE_PREFIX_PATH}"
MACOSX_DEPLOYMENT_TARGET=10.9 CC=clang CXX=clang++ CMAKE_ONLY=1 python setup.py build
ccmake build  # or cmake-gui build

Docker Image

Using pre-built images

You can also pull a pre-built docker image from Docker Hub and run with docker v19.03+

docker run --gpus all --rm -ti --ipc=host pytorch/pytorch:latest

Please note that PyTorch uses shared memory to share data between processes, so if torch multiprocessing is used (e.g. for multithreaded data loaders) the default shared memory segment size that container runs with is not enough, and you should increase shared memory size either with --ipc=host or --shm-size command line options to nvidia-docker run.

Building the image yourself

NOTE: Must be built with a docker version > 18.06

The Dockerfile is supplied to build images with CUDA 11.1 support and cuDNN v8. You can pass PYTHON_VERSION=x.y make variable to specify which Python version is to be used by Miniconda, or leave it unset to use the default.

make -f docker.Makefile
# images are tagged as docker.io/${your_docker_username}/pytorch

You can also pass the CMAKE_VARS="..." environment variable to specify additional CMake variables to be passed to CMake during the build. See setup.py for the list of available variables.

make -f docker.Makefile

Building the Documentation

To build documentation in various formats, you will need Sphinx and the pytorch_sphinx_theme2.

Before you build the documentation locally, ensure torch is installed in your environment. For small fixes, you can install the nightly version as described in Getting Started.

For more complex fixes, such as adding a new module and docstrings for the new module, you might need to install torch from source. See Docstring Guidelines for docstring conventions.

cd docs/
pip install -r requirements.txt
make html
make serve

Run make to get a list of all available output formats.

If you get a katex error run npm install katex. If it persists, try npm install -g katex

[!NOTE] If you installed nodejs with a different package manager (e.g., conda) then npm will probably install a version of katex that is not compatible with your version of nodejs and doc builds will fail. A combination of versions that is known to work is node@6.13.1 and katex@0.13.18. To install the latter with npm you can run npm install -g katex@0.13.18

[!NOTE] If you see a numpy incompatibility error, run:

pip install 'numpy<2'

When you make changes to the dependencies run by CI, edit the .ci/docker/requirements-docs.txt file.

Building a PDF

To compile a PDF of all PyTorch documentation, ensure you have texlive and LaTeX installed. On macOS, you can install them using:

brew install --cask mactex

To create the PDF:

  1. Run:

    make latexpdf
    

    This will generate the necessary files in the build/latex directory.

  2. Navigate to this directory and execute:

    make LATEXOPTS="-interaction=nonstopmode"
    

    This will produce a pytorch.pdf with the desired content. Run this command one more time so that it generates the correct table of contents and index.

[!NOTE] To view the Table of Contents, switch to the Table of Contents view in your PDF viewer.

Previous Versions

Installation instructions and binaries for previous PyTorch versions may be found on our website.

Getting Started

Three-pointers to get you started:

Resources

Communication

Releases and Contributing

Typically, PyTorch has three minor releases a year. Please let us know if you encounter a bug by filing an issue.

We appreciate all contributions. If you are planning to contribute back bug-fixes, please do so without any further discussion.

If you plan to contribute new features, utility functions, or extensions to the core, please first open an issue and discuss the feature with us. Sending a PR without discussion might end up resulting in a rejected PR because we might be taking the core in a different direction than you might be aware of.

To learn more about making a contribution to Pytorch, please see our Contribution page. For more information about PyTorch releases, see Release page.

The Team

PyTorch is a community-driven project with several skillful engineers and researchers contributing to it.

PyTorch is currently maintained by Soumith Chintala, Gregory Chanan, Dmytro Dzhulgakov, Edward Yang, and Nikita Shulga with major contributions coming from hundreds of talented individuals in various forms and means. A non-exhaustive but growing list needs to mention: Trevor Killeen, Sasank Chilamkurthy, Sergey Zagoruyko, Adam Lerer, Francisco Massa, Alykhan Tejani, Luca Antiga, Alban Desmaison, Andreas Koepf, James Bradbury, Zeming Lin, Yuandong Tian, Guillaume Lample, Marat Dukhan, Natalia Gimelshein, Christian Sarofeen, Martin Raison, Edward Yang, Zachary Devito.

Note: This project is unrelated to hughperkins/pytorch with the same name. Hugh is a valuable contributor to the Torch community and has helped with many things Torch and PyTorch.

License

PyTorch has a BSD-style license, as found in the LICENSE file.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distributions

torch-2.8.0-cp313-none-macosx_11_0_arm64.whl (73.6 MB view details)

Uploaded CPython 3.13macOS 11.0+ ARM64

torch-2.8.0-cp313-cp313t-win_amd64.whl (241.6 MB view details)

Uploaded CPython 3.13tWindows x86-64

torch-2.8.0-cp313-cp313t-manylinux_2_28_x86_64.whl (888.0 MB view details)

Uploaded CPython 3.13tmanylinux: glibc 2.28+ x86-64

torch-2.8.0-cp313-cp313t-manylinux_2_28_aarch64.whl (102.0 MB view details)

Uploaded CPython 3.13tmanylinux: glibc 2.28+ ARM64

torch-2.8.0-cp313-cp313t-macosx_14_0_arm64.whl (73.9 MB view details)

Uploaded CPython 3.13tmacOS 14.0+ ARM64

torch-2.8.0-cp313-cp313-win_amd64.whl (241.3 MB view details)

Uploaded CPython 3.13Windows x86-64

torch-2.8.0-cp313-cp313-manylinux_2_28_x86_64.whl (887.9 MB view details)

Uploaded CPython 3.13manylinux: glibc 2.28+ x86-64

torch-2.8.0-cp313-cp313-manylinux_2_28_aarch64.whl (101.9 MB view details)

Uploaded CPython 3.13manylinux: glibc 2.28+ ARM64

torch-2.8.0-cp312-none-macosx_11_0_arm64.whl (73.6 MB view details)

Uploaded CPython 3.12macOS 11.0+ ARM64

torch-2.8.0-cp312-cp312-win_amd64.whl (241.3 MB view details)

Uploaded CPython 3.12Windows x86-64

torch-2.8.0-cp312-cp312-manylinux_2_28_x86_64.whl (887.9 MB view details)

Uploaded CPython 3.12manylinux: glibc 2.28+ x86-64

torch-2.8.0-cp312-cp312-manylinux_2_28_aarch64.whl (101.9 MB view details)

Uploaded CPython 3.12manylinux: glibc 2.28+ ARM64

torch-2.8.0-cp311-none-macosx_11_0_arm64.whl (73.6 MB view details)

Uploaded CPython 3.11macOS 11.0+ ARM64

torch-2.8.0-cp311-cp311-win_amd64.whl (241.4 MB view details)

Uploaded CPython 3.11Windows x86-64

torch-2.8.0-cp311-cp311-manylinux_2_28_x86_64.whl (888.1 MB view details)

Uploaded CPython 3.11manylinux: glibc 2.28+ x86-64

torch-2.8.0-cp311-cp311-manylinux_2_28_aarch64.whl (102.1 MB view details)

Uploaded CPython 3.11manylinux: glibc 2.28+ ARM64

torch-2.8.0-cp310-none-macosx_11_0_arm64.whl (73.6 MB view details)

Uploaded CPython 3.10macOS 11.0+ ARM64

torch-2.8.0-cp310-cp310-win_amd64.whl (241.4 MB view details)

Uploaded CPython 3.10Windows x86-64

torch-2.8.0-cp310-cp310-manylinux_2_28_x86_64.whl (888.0 MB view details)

Uploaded CPython 3.10manylinux: glibc 2.28+ x86-64

torch-2.8.0-cp310-cp310-manylinux_2_28_aarch64.whl (102.1 MB view details)

Uploaded CPython 3.10manylinux: glibc 2.28+ ARM64

torch-2.8.0-cp39-none-macosx_11_0_arm64.whl (73.6 MB view details)

Uploaded CPython 3.9macOS 11.0+ ARM64

torch-2.8.0-cp39-cp39-win_amd64.whl (241.2 MB view details)

Uploaded CPython 3.9Windows x86-64

torch-2.8.0-cp39-cp39-manylinux_2_28_x86_64.whl (888.0 MB view details)

Uploaded CPython 3.9manylinux: glibc 2.28+ x86-64

torch-2.8.0-cp39-cp39-manylinux_2_28_aarch64.whl (102.1 MB view details)

Uploaded CPython 3.9manylinux: glibc 2.28+ ARM64

File details

Details for the file torch-2.8.0-cp313-none-macosx_11_0_arm64.whl.

File metadata

File hashes

Hashes for torch-2.8.0-cp313-none-macosx_11_0_arm64.whl
Algorithm Hash digest
SHA256 7b677e17f5a3e69fdef7eb3b9da72622f8d322692930297e4ccb52fefc6c8211
MD5 9c71ec4081d4ef6c20f975c06bd1ee3f
BLAKE2b-256 046e650bb7f28f771af0cb791b02348db8b7f5f64f40f6829ee82aa6ce99aabe

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp313-cp313t-win_amd64.whl.

File metadata

  • Download URL: torch-2.8.0-cp313-cp313t-win_amd64.whl
  • Upload date:
  • Size: 241.6 MB
  • Tags: CPython 3.13t, Windows x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.5

File hashes

Hashes for torch-2.8.0-cp313-cp313t-win_amd64.whl
Algorithm Hash digest
SHA256 a7242b86f42be98ac674b88a4988643b9bc6145437ec8f048fea23f72feb5eca
MD5 9aaccc06d4dfcd1c3b405ff6f1742a83
BLAKE2b-256 b4a86a04e4b54472fc5dba7ca2341ab219e529f3c07b6941059fbf18dccac31f

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp313-cp313t-manylinux_2_28_x86_64.whl.

File metadata

File hashes

Hashes for torch-2.8.0-cp313-cp313t-manylinux_2_28_x86_64.whl
Algorithm Hash digest
SHA256 8f0a9d617a66509ded240add3754e462430a6c1fc5589f86c17b433dd808f97a
MD5 9cbc5504ddf6da07c515a8b5451438c7
BLAKE2b-256 4a155e488ca0bc6162c86a33b58642bc577c84ded17c7b72d97e49b5833e2d73

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp313-cp313t-manylinux_2_28_aarch64.whl.

File metadata

File hashes

Hashes for torch-2.8.0-cp313-cp313t-manylinux_2_28_aarch64.whl
Algorithm Hash digest
SHA256 83c13411a26fac3d101fe8035a6b0476ae606deb8688e904e796a3534c197def
MD5 067aa02737160e7785cd73ac6437a918
BLAKE2b-256 150e8a800e093b7f7430dbaefa80075aee9158ec22e4c4fc3c1a66e4fb96cb4f

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp313-cp313t-macosx_14_0_arm64.whl.

File metadata

File hashes

Hashes for torch-2.8.0-cp313-cp313t-macosx_14_0_arm64.whl
Algorithm Hash digest
SHA256 1a62a1ec4b0498930e2543535cf70b1bef8c777713de7ceb84cd79115f553767
MD5 a340318c092f9984273eea2030c2229a
BLAKE2b-256 de698b7b13bba430f5e21d77708b616f767683629fc4f8037564a177d20f90ed

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp313-cp313-win_amd64.whl.

File metadata

  • Download URL: torch-2.8.0-cp313-cp313-win_amd64.whl
  • Upload date:
  • Size: 241.3 MB
  • Tags: CPython 3.13, Windows x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.5

File hashes

Hashes for torch-2.8.0-cp313-cp313-win_amd64.whl
Algorithm Hash digest
SHA256 659df54119ae03e83a800addc125856effda88b016dfc54d9f65215c3975be16
MD5 47949b18f79c7991fb3db360fb7a3904
BLAKE2b-256 e354941ea0a860f2717d86a811adf0c2cd01b3983bdd460d0803053c4e0b8649

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp313-cp313-manylinux_2_28_x86_64.whl.

File metadata

File hashes

Hashes for torch-2.8.0-cp313-cp313-manylinux_2_28_x86_64.whl
Algorithm Hash digest
SHA256 65616ca8ec6f43245e1f5f296603e33923f4c30f93d65e103d9e50c25b35150b
MD5 fc0c298db2eb18b8ee1414b0020d330c
BLAKE2b-256 16823948e54c01b2109238357c6f86242e6ecbf0c63a1af46906772902f82057

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp313-cp313-manylinux_2_28_aarch64.whl.

File metadata

File hashes

Hashes for torch-2.8.0-cp313-cp313-manylinux_2_28_aarch64.whl
Algorithm Hash digest
SHA256 2b2f96814e0345f5a5aed9bf9734efa913678ed19caf6dc2cddb7930672d6128
MD5 ca142b486d7a20cd5387e3e3593b6a9b
BLAKE2b-256 104e469ced5a0603245d6a19a556e9053300033f9c5baccf43a3d25ba73e189e

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp312-none-macosx_11_0_arm64.whl.

File metadata

File hashes

Hashes for torch-2.8.0-cp312-none-macosx_11_0_arm64.whl
Algorithm Hash digest
SHA256 619c2869db3ada2c0105487ba21b5008defcc472d23f8b80ed91ac4a380283b0
MD5 30025ea379e5dc9d3ba3f60ad071778c
BLAKE2b-256 be665c9a321b325aaecb92d4d1855421e3a055abd77903b7dab6575ca07796db

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp312-cp312-win_amd64.whl.

File metadata

  • Download URL: torch-2.8.0-cp312-cp312-win_amd64.whl
  • Upload date:
  • Size: 241.3 MB
  • Tags: CPython 3.12, Windows x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.5

File hashes

Hashes for torch-2.8.0-cp312-cp312-win_amd64.whl
Algorithm Hash digest
SHA256 2f4ac52f0130275d7517b03a33d2493bab3693c83dcfadf4f81688ea82147d2e
MD5 5a68c736d16cfe4db451d0e26b078318
BLAKE2b-256 af8a5c87f08e3abd825c7dfecef5a0f1d9aa5df5dd0e3fd1fa2f490a8e512402

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp312-cp312-manylinux_2_28_x86_64.whl.

File metadata

File hashes

Hashes for torch-2.8.0-cp312-cp312-manylinux_2_28_x86_64.whl
Algorithm Hash digest
SHA256 b2aca0939fb7e4d842561febbd4ffda67a8e958ff725c1c27e244e85e982173c
MD5 38231c9d184973328f7d07cdf9c4fb36
BLAKE2b-256 99a86acf48d48838fb8fe480597d98a0668c2beb02ee4755cc136de92a0a956f

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp312-cp312-manylinux_2_28_aarch64.whl.

File metadata

File hashes

Hashes for torch-2.8.0-cp312-cp312-manylinux_2_28_aarch64.whl
Algorithm Hash digest
SHA256 e2fab4153768d433f8ed9279c8133a114a034a61e77a3a104dcdf54388838705
MD5 d6e749d0ef43782f3a2d91dc1545385a
BLAKE2b-256 490c2fd4df0d83a495bb5e54dca4474c4ec5f9c62db185421563deeb5dabf609

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp311-none-macosx_11_0_arm64.whl.

File metadata

File hashes

Hashes for torch-2.8.0-cp311-none-macosx_11_0_arm64.whl
Algorithm Hash digest
SHA256 5ae0524688fb6707c57a530c2325e13bb0090b745ba7b4a2cd6a3ce262572916
MD5 5c436f0842a3cfae1d91b58c557026a5
BLAKE2b-256 a45e05a5c46085d9b97e928f3f037081d3d2b87fb4b4195030fc099aaec5effc

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp311-cp311-win_amd64.whl.

File metadata

  • Download URL: torch-2.8.0-cp311-cp311-win_amd64.whl
  • Upload date:
  • Size: 241.4 MB
  • Tags: CPython 3.11, Windows x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.5

File hashes

Hashes for torch-2.8.0-cp311-cp311-win_amd64.whl
Algorithm Hash digest
SHA256 8c7ef765e27551b2fbfc0f41bcf270e1292d9bf79f8e0724848b1682be6e80aa
MD5 a5bdec4099e17fe258c653a22453ff53
BLAKE2b-256 84572f64161769610cf6b1c5ed782bd8a780e18a3c9d48931319f2887fa9d0b1

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp311-cp311-manylinux_2_28_x86_64.whl.

File metadata

File hashes

Hashes for torch-2.8.0-cp311-cp311-manylinux_2_28_x86_64.whl
Algorithm Hash digest
SHA256 c12fa219f51a933d5f80eeb3a7a5d0cbe9168c0a14bbb4055f1979431660879b
MD5 30808715527e3643ace4d0825e0666d8
BLAKE2b-256 5a634fdc45a0304536e75a5e1b1bbfb1b56dd0e2743c48ee83ca729f7ce44162

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp311-cp311-manylinux_2_28_aarch64.whl.

File metadata

File hashes

Hashes for torch-2.8.0-cp311-cp311-manylinux_2_28_aarch64.whl
Algorithm Hash digest
SHA256 220a06fd7af8b653c35d359dfe1aaf32f65aa85befa342629f716acb134b9710
MD5 a5be40b5bb2744db97191e760cdc6e6a
BLAKE2b-256 8fc43e7a3887eba14e815e614db70b3b529112d1513d9dae6f4d43e373360b7f

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp310-none-macosx_11_0_arm64.whl.

File metadata

File hashes

Hashes for torch-2.8.0-cp310-none-macosx_11_0_arm64.whl
Algorithm Hash digest
SHA256 a3f16a58a9a800f589b26d47ee15aca3acf065546137fc2af039876135f4c760
MD5 82f055ce589517dd0f7430b7f053fb58
BLAKE2b-256 efd6e6d4c57e61c2b2175d3aafbfb779926a2cfd7c32eeda7c543925dceec923

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp310-cp310-win_amd64.whl.

File metadata

  • Download URL: torch-2.8.0-cp310-cp310-win_amd64.whl
  • Upload date:
  • Size: 241.4 MB
  • Tags: CPython 3.10, Windows x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.5

File hashes

Hashes for torch-2.8.0-cp310-cp310-win_amd64.whl
Algorithm Hash digest
SHA256 e8e5bf982e87e2b59d932769938b698858c64cc53753894be25629bdf5cf2f46
MD5 9b6fdc39decde1787055b0d683da0069
BLAKE2b-256 7087f69752d0dd4ba8218c390f0438130c166fa264a33b7025adb5014b92192c

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp310-cp310-manylinux_2_28_x86_64.whl.

File metadata

File hashes

Hashes for torch-2.8.0-cp310-cp310-manylinux_2_28_x86_64.whl
Algorithm Hash digest
SHA256 89aa9ee820bb39d4d72b794345cccef106b574508dd17dbec457949678c76011
MD5 cb4550b1130b8447c32197392c9115ad
BLAKE2b-256 701c58da560016f81c339ae14ab16c98153d51c941544ae568da3cb5b1ceb572

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp310-cp310-manylinux_2_28_aarch64.whl.

File metadata

File hashes

Hashes for torch-2.8.0-cp310-cp310-manylinux_2_28_aarch64.whl
Algorithm Hash digest
SHA256 0be92c08b44009d4131d1ff7a8060d10bafdb7ddcb7359ef8d8c5169007ea905
MD5 b1288dac1bddd9e4797a31a592e3ae07
BLAKE2b-256 6328110f7274254f1b8476c561dada127173f994afa2b1ffc044efb773c15650

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp39-none-macosx_11_0_arm64.whl.

File metadata

File hashes

Hashes for torch-2.8.0-cp39-none-macosx_11_0_arm64.whl
Algorithm Hash digest
SHA256 e9f071f5b52a9f6970dc8a919694b27a91ae9dc08898b2b988abbef5eddfd1ae
MD5 8db58a34cd26c7ad7e93a4565260f88f
BLAKE2b-256 ae95ae26263aceb3d57b821179f827d0e321373ed49423e603dd5906ab14a730

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp39-cp39-win_amd64.whl.

File metadata

  • Download URL: torch-2.8.0-cp39-cp39-win_amd64.whl
  • Upload date:
  • Size: 241.2 MB
  • Tags: CPython 3.9, Windows x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.5

File hashes

Hashes for torch-2.8.0-cp39-cp39-win_amd64.whl
Algorithm Hash digest
SHA256 5128fe752a355d9308e56af1ad28b15266fe2da5948660fad44de9e3a9e36e8c
MD5 0aeb01c7fe2839508dfea2bfe67352ae
BLAKE2b-256 b9dc1f1f621afe15e3c496e1e8f94f8903f75f87e7d642d5a985e92210cc208d

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp39-cp39-manylinux_2_28_x86_64.whl.

File metadata

File hashes

Hashes for torch-2.8.0-cp39-cp39-manylinux_2_28_x86_64.whl
Algorithm Hash digest
SHA256 06fcee8000e5c62a9f3e52a688b9c5abb7c6228d0e56e3452983416025c41381
MD5 195762e01550324c5212acabf3878032
BLAKE2b-256 fddd1630cb51b10d3d2e97db95e5a84c32def81fc26b005bce6fc880b0e6db81

See more details on using hashes here.

File details

Details for the file torch-2.8.0-cp39-cp39-manylinux_2_28_aarch64.whl.

File metadata

File hashes

Hashes for torch-2.8.0-cp39-cp39-manylinux_2_28_aarch64.whl
Algorithm Hash digest
SHA256 da6afa31c13b669d4ba49d8a2169f0db2c3ec6bec4af898aa714f401d4c38904
MD5 9e324187fcbda8fa9cb990c667dad641
BLAKE2b-256 5bb0a321f27270049baa12f5c3fb0d6ceea005634787e3af9a8d75dce8306b0a

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page
pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy