Skip to content

Commit 3a4cb85

Browse files
authored
Merge branch 'site' into fix-previous-versions
2 parents 15da966 + 56a6530 commit 3a4cb85

11 files changed

+992
-290
lines changed

_events/pt-26-live-q-a.md

+20
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,20 @@
1+
---
2+
category: event
3+
title: "PyTorch 2.6 Live Q&A"
4+
date: February 7, 2025
5+
poster: assets/images/ai-programming.png
6+
---
7+
8+
**Date**: February 7, 10 am PST
9+
10+
<a href="/pt-26-live-q-a">
11+
<img style="width:100%" src="/assets/images/pt-26-live-q-a.png" alt="PyTorch 2.6 Live Q&A">
12+
</a>
13+
14+
Wondering what's new in the recent PyTorch 2.6 release? Do you have questions? Join us for a live Q&A on PyTorch 2.6 with PyTorch Core Maintainer, Nikita Shulga (Meta).
15+
16+
Nikita is a Software Engineer at Meta where he is, among other things, responsible for PyTorch releases and continuous integration. Nikita is committed to uplifting the developer community and continuously improving PyTorch. He earned his Master’s degree in Applied Mathematics from the Moscow Institute of Physics and Technology (MIPT).
17+
18+
Bring your PyTorch 2.6 questions for Nikita during this live Q&A session.
19+
20+
[Register now to join the event](/pt-26-live-q-a)

_get_started/previous-versions.md

+49
Original file line numberDiff line numberDiff line change
@@ -17,6 +17,55 @@ your convenience.
1717

1818
## Commands for Versions >= 1.0.0
1919

20+
### v2.5.1
21+
22+
#### Conda
23+
24+
##### OSX
25+
26+
```
27+
# conda
28+
conda install pytorch==2.5.1 torchvision==0.20.0 torchaudio==2.5.0 -c pytorch
29+
```
30+
31+
##### Linux and Windows
32+
33+
```
34+
# CUDA 11.8
35+
conda install pytorch==2.5.1 torchvision==0.20.0 torchaudio==2.5.0 pytorch-cuda=11.8 -c pytorch -c nvidia
36+
# CUDA 12.1
37+
conda install pytorch==2.5.1 torchvision==0.20.0 torchaudio==2.5.0 pytorch-cuda=12.1 -c pytorch -c nvidia
38+
# CUDA 12.4
39+
conda install pytorch==2.5.1 torchvision==0.20.0 torchaudio==2.5.0 pytorch-cuda=12.4 -c pytorch -c nvidia
40+
# CPU Only
41+
conda install pytorch==2.5.1 torchvision==0.20.0 torchaudio==2.5.0 cpuonly -c pytorch
42+
```
43+
44+
#### Wheel
45+
46+
##### OSX
47+
48+
```
49+
pip install torch==2.5.1 torchvision==0.20.0 torchaudio==2.5.0
50+
```
51+
52+
##### Linux and Windows
53+
54+
```
55+
# ROCM 6.1 (Linux only)
56+
pip install torch==2.5.1 torchvision==0.20.0 torchaudio==2.5.0 --index-url https://download.pytorch.org/whl/rocm6.1
57+
# ROCM 6.2 (Linux only)
58+
pip install torch==2.5.1 torchvision==0.20.0 torchaudio==2.5.0 --index-url https://download.pytorch.org/whl/rocm6.2
59+
# CUDA 11.8
60+
pip install torch==2.5.1 torchvision==0.20.0 torchaudio==2.5.0 --index-url https://download.pytorch.org/whl/cu118
61+
# CUDA 12.1
62+
pip install torch==2.5.1 torchvision==0.20.0 torchaudio==2.5.0 --index-url https://download.pytorch.org/whl/cu121
63+
# CUDA 12.4
64+
pip install torch==2.5.1 torchvision==0.20.0 torchaudio==2.5.0 --index-url https://download.pytorch.org/whl/cu124
65+
# CPU only
66+
pip install torch==2.5.1 torchvision==0.20.0 torchaudio==2.5.0 --index-url https://download.pytorch.org/whl/cpu
67+
```
68+
2069
### v2.5.0
2170

2271
#### Conda
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,25 @@
1+
---
2+
layout: blog_detail
3+
title: "2025 Priorities for the PyTorch Technical Advisory Council (TAC)"
4+
author: "Luca Antiga, PyTorch TAC Chair"
5+
---
6+
7+
![social share](/assets/images/1738166706211.jpg){:style="max-width:600px; width:100%; display: block; margin-left: auto; margin-right: auto"}
8+
9+
10+
[2024 has been a year of incredible growth for PyTorch](https://pytorch.org/blog/2024-year-in-review/). As that continues in 2025, the PyTorch Foundation has made important steps towards evolving the governance of the project under the Linux Foundation’s vendor-neutral umbrella.
11+
12+
An important piece of governance for PyTorch is represented by the Technical Advisory Council (TAC). The TAC acts as a bridge between the industry, including but not limited to the PyTorch Foundation members, the community, and the PyTorch core development team.
13+
14+
Operating with transparency and inclusivity, the TAC gathers input, facilitates collaboration, and drives initiatives that enhance the experience for everyone who relies on PyTorch.
15+
16+
In 2025, the TAC will focus on four key areas:
17+
18+
1. **Build Open, Multi-Cloud Continuous Integration (CI):** Building on the groundwork from 2024, the TAC will oversee the transition to an open, community-driven CI infrastructure. In addition to ensuring the extremely high bar for correctness that PyTorch has, PyTorch’s CI is complex with a high-quality bar including many automated functional and performance daily test runs. In 2025, PyTorch’s CI infrastructure will be fully open sourced and extended to support multiple compute providers, enabling broader contribution and participation to the effort from organizations benefitting from PyTorch.
19+
2. **Support more Accelerators:** The TAC is committed to creating a level playing field for the growing landscape of AI accelerators. By gathering industry players and PyTorch developers, the TAC will facilitate efforts towards third-party device support and provide levels of integration of external CI systems with the main PyTorch CI. This will make it easier for emerging hardware to gain adoption within the PyTorch ecosystem, and for users to experiment with diverse compute options for training and inference.
20+
3. **Create a High-Quality, User-Centric Ecosystem:** A big focus for the TAC in early 2025 is on improving the experience and discoverability of the PyTorch ecosystem. With many projects growing organically, users often face challenges navigating projects of different scope and quality within the rapidly changing AI landscape. To solve this, a newly curated ecosystem landscape tool will be launched soon on the PyTorch website. We will also introduce lightweight, open processes to improve projects and ensure users a predictable, high-quality experience. In many ways, the experience with PyTorch is as good as its ecosystem.
21+
4. **Gather Feedback from Industry and the Community:** PyTorch has widespread adoption across research labs, startups, and enterprises. Striking the right balance between expressiveness and performance across the board is a very challenging task, so the TAC set out to be one of the several ways the Core development team receives signals. During our monthly TAC meetings, we provide the opportunity to PyTorch Foundation members from industry and academia, as well as non-member organizations to present their use case, their challenges and discuss them directly with appropriate members of the Core team. This feedback loop helps prioritize improvements, ensuring the framework stays relevant in a fast-evolving AI landscape.
22+
23+
By focusing on these priorities, the TAC aims to maintain PyTorch’s position as the leading deep learning framework, while ensuring it remains open, accessible, and responsive to the needs of its diverse community.
24+
25+
As members of the TAC, we’re extremely excited to contribute to the success of PyTorch and to the impact it’s having in the real world. If you are a PyTorch user or developer, consider [participating in our monthly calls](https://zoom-lfx.platform.linuxfoundation.org/meetings/pytorch?__hstc=132719121.a26416c161ac91bef494ffc19f91a62e.1723036593114.1738082449904.1738088158683.375&__hssc=132719121.1.1738088158683&__hsfp=810579359) (they are open to everyone, and the recordings are available [here](https://lists.pytorch.org/g/tac)). Also, if you develop or maintain a project based on PyTorch, consider contributing it to the new PyTorch ecosystem ([instructions](https://pytorch.org/ecosystem/join)).

_posts/2025-01-29-pytorch2-6.md

+146
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,146 @@
1+
---
2+
layout: blog_detail
3+
title: "PyTorch 2.6 Release Blog"
4+
---
5+
6+
We are excited to announce the release of PyTorch® 2.6 ([release notes](https://github.com/pytorch/pytorch/releases/tag/v2.6.0))! This release features multiple improvements for PT2: `torch.compile` can now be used with Python 3.13; new performance-related knob `torch.compiler.set_stance`; several AOTInductor enhancements. Besides the PT2 improvements, another highlight is FP16 support on X86 CPUs.
7+
8+
NOTE: Starting with this release we are not going to publish on Conda, please see [[Announcement] Deprecating PyTorch’s official Anaconda channel](https://github.com/pytorch/pytorch/issues/138506) for the details.
9+
10+
For this release the experimental Linux binaries shipped with CUDA 12.6.3 (as well as Linux Aarch64, Linux ROCm 6.2.4, and Linux XPU binaries) are built with CXX11_ABI=1 and are [using the Manylinux 2.28 build platform](https://dev-discuss.pytorch.org/t/pytorch-linux-wheels-switching-to-new-wheel-build-platform-manylinux-2-28-on-november-12-2024/2581). If you build PyTorch extensions with custom C++ or CUDA extensions, please update these builds to use CXX_ABI=1 as well and report any issues you are seeing. For the next PyTorch 2.7 release we plan to switch all Linux builds to Manylinux 2.28 and CXX11_ABI=1, please see [[RFC] PyTorch next wheel build platform: manylinux-2.28](https://github.com/pytorch/pytorch/issues/123649) for the details and discussion.
11+
12+
Also in this release as an important security improvement measure we have changed the default value for `weights_only` parameter of `torch.load`. This is a backward compatibility-breaking change, please see [this forum post](https://dev-discuss.pytorch.org/t/bc-breaking-change-torch-load-is-being-flipped-to-use-weights-only-true-by-default-in-the-nightlies-after-137602/2573) for more details.
13+
14+
This release is composed of 3892 commits from 520 contributors since PyTorch 2.5. We want to sincerely thank our dedicated community for your contributions. As always, we encourage you to try these out and report any issues as we improve PyTorch. More information about how to get started with the PyTorch 2-series can be found at our [Getting Started](https://pytorch.org/get-started/pytorch-2.0/) page.
15+
16+
17+
<table class="table table-bordered">
18+
<tr>
19+
<td>Beta
20+
</td>
21+
<td>Prototype
22+
</td>
23+
</tr>
24+
<tr>
25+
<td>torch.compiler.set_stance
26+
</td>
27+
<td>Improved PyTorch user experience on Intel GPUs
28+
</td>
29+
</tr>
30+
<tr>
31+
<td>torch.library.triton_op
32+
</td>
33+
<td>FlexAttention support on X86 CPU for LLMs
34+
</td>
35+
</tr>
36+
<tr>
37+
<td>torch.compile support for Python 3.13
38+
</td>
39+
<td>Dim.AUTO
40+
</td>
41+
</tr>
42+
<tr>
43+
<td>New packaging APIs for AOTInductor
44+
</td>
45+
<td>CUTLASS and CK GEMM/CONV Backends for AOTInductor
46+
</td>
47+
</tr>
48+
<tr>
49+
<td>AOTInductor: minifier
50+
</td>
51+
<td>
52+
</td>
53+
</tr>
54+
<tr>
55+
<td>AOTInductor: ABI-compatible mode code generation
56+
</td>
57+
<td>
58+
</td>
59+
</tr>
60+
<tr>
61+
<td>FP16 support for X86 CPUs
62+
</td>
63+
<td>
64+
</td>
65+
</tr>
66+
</table>
67+
68+
69+
*To see a full list of public feature submissions click [here](https://docs.google.com/spreadsheets/d/1TzGkWuUMF1yTe88adz1dt2mzbIsZLd3PBasy588VWgk/edit?usp=sharing).
70+
71+
72+
## BETA FEATURES
73+
74+
75+
### [Beta] torch.compiler.set_stance
76+
77+
This feature enables the user to specify different behaviors (“stances”) that `torch.compile` can take between different invocations of compiled functions. One of the stances, for example, is
78+
79+
“eager_on_recompile”, that instructs PyTorch to code eagerly when a recompile is necessary, reusing cached compiled code when possible.
80+
81+
For more information please refer to the [set_stance documentation](https://pytorch.org/docs/2.6/generated/torch.compiler.set_stance.html#torch.compiler.set_stance) and the [Dynamic Compilation Control with torch.compiler.set_stance](https://pytorch.org/tutorials/recipes/torch_compiler_set_stance_tutorial.html) tutorial.
82+
83+
### [Beta] torch.library.triton_op
84+
85+
`torch.library.triton_op` offers a standard way of creating custom operators that are backed by user-defined triton kernels.
86+
87+
When users turn user-defined triton kernels into custom operators, `torch.library.triton_op` allows `torch.compile` to peek into the implementation, enabling `torch.compile` to optimize the triton kernel inside it.
88+
89+
For more information please refer to the [triton_op documentation](https://pytorch.org/docs/2.6/library.html#torch.library.triton_op) and the[ Using User-Defined Triton Kernels with torch.compile](https://pytorch.org/tutorials/recipes/torch_compile_user_defined_triton_kernel_tutorial.html) tutorial.
90+
91+
### [Beta] torch.compile support for Python 3.13
92+
93+
`torch.compile` previously only supported Python up to version 3.12. Users can now optimize models with `torch.compile` in Python 3.13.
94+
95+
### [Beta] New packaging APIs for AOTInductor
96+
97+
A new package format, “[PT2 archive](https://docs.google.com/document/d/1RQ4cmywilnFUT1VE-4oTGxwXdc8vowCSZsrRgo3wFA8/edit?usp=sharing)”, has been introduced. This essentially contains a zipfile of all the files that need to be used by AOTInductor, and allows users to send everything needed to other environments. There is also functionality to package multiple models into one artifact, and to store additional metadata inside of the package.
98+
99+
For more details please see the updated [torch.export AOTInductor Tutorial for Python runtime](https://pytorch.org/tutorials/recipes/torch_export_aoti_python.html).
100+
101+
### [Beta] AOTInductor: minifier
102+
103+
If a user encounters an error while using AOTInductor APIs, AOTInductor Minifier allows creation of a minimal nn.Module that reproduces the error.
104+
105+
For more information please see the [AOTInductor Minifier documentation](https://pytorch.org/docs/2.6/torch.compiler_aot_inductor_minifier.html).
106+
107+
### [Beta] AOTInductor: ABI-compatible mode code generation
108+
109+
AOTInductor-generated model code has dependency on Pytorch cpp libraries. As Pytorch evolves quickly, it’s important to make sure previously AOTInductor compiled models can continue to run on newer Pytorch versions, i.e. AOTInductor is backward compatible.
110+
111+
In order to guarantee application binary interface (ABI) backward compatibility, we have carefully defined a set of stable C interfaces in libtorch and make sure AOTInductor generates code that only refers to the specific set of APIs and nothing else in libtorch. We will keep the set of C APIs stable across Pytorch versions and thus provide backward compatibility guarantees for AOTInductor-compiled models.
112+
113+
### [Beta] FP16 support for X86 CPUs (both eager and Inductor modes)
114+
115+
Float16 datatype is commonly used for reduced memory usage and faster computation in AI inference and training. CPUs like the recently launched [Intel® Xeon® 6 with P-Cores](https://www.intel.com/content/www/us/en/products/details/processors/xeon/xeon6-p-cores.html) support Float16 datatype with native accelerator [AMX](https://www.intel.com/content/www/us/en/products/docs/accelerator-engines/advanced-matrix-extensions/overview.html). Float16 support on X86 CPUs was introduced in PyTorch 2.5 as a prototype feature, and now it has been further improved for both eager mode and Torch.compile + Inductor mode, making it Beta level feature with both functionality and performance verified with a broad scope of workloads.
116+
117+
118+
## PROTOTYPE FEATURES
119+
120+
### [Prototype] Improved PyTorch user experience on Intel GPUs
121+
122+
PyTorch user experience on Intel GPUs is further improved with simplified installation steps, Windows release binary distribution and expanded coverage of supported GPU models including the latest Intel® Arc™ B-Series discrete graphics. Application developers and researchers seeking to fine-tune, inference and develop with PyTorch models on [Intel® Core™ Ultra AI PCs ](https://www.intel.com/content/www/us/en/products/docs/processors/core-ultra/ai-pc.html)and [Intel® Arc™ discrete graphics](https://www.intel.com/content/www/us/en/products/details/discrete-gpus/arc.html) will now be able to directly install PyTorch with binary releases for Windows, Linux and Windows Subsystem for Linux 2.
123+
124+
125+
126+
* Simplified Intel GPU software stack setup to enable one-click installation of the torch-xpu PIP wheels to run deep learning workloads in an out of the box fashion, eliminating the complexity of installing and activating Intel GPU development software bundles.
127+
* Windows binary releases for torch core, torchvision and torchaudio have been made available for Intel GPUs, and the supported GPU models have been expanded from Intel® Core™ Ultra Processors with Intel® Arc™ Graphics, [Intel® Core™ Ultra Series 2 with Intel® Arc™ Graphics](https://www.intel.com/content/www/us/en/products/details/processors/core-ultra.html) and [Intel® Arc™ A-Series Graphics](https://www.intel.com/content/www/us/en/products/docs/discrete-gpus/arc/desktop/a-series/overview.html) to the latest GPU hardware [Intel® Arc™ B-Series graphics](https://www.intel.com/content/www/us/en/products/docs/discrete-gpus/arc/desktop/b-series/overview.html).
128+
* Further enhanced coverage of Aten operators on Intel GPUs with SYCL* kernels for smooth eager mode execution, as well as bug fixes and performance optimizations for torch.compile on Intel GPUs.
129+
130+
For more information regarding Intel GPU support, please refer to [Getting Started Guide](https://pytorch.org/docs/main/notes/get_start_xpu.html).
131+
132+
### [Prototype] FlexAttention support on X86 CPU for LLMs
133+
134+
FlexAttention was initially introduced in PyTorch 2.5 to provide optimized implementations for Attention variants with a flexible API. In PyTorch 2.6, X86 CPU support for FlexAttention was added through TorchInductor CPP backend. This new feature leverages and extends current CPP template abilities to support broad attention variants (e.x.: PageAttention, which is critical for LLMs inference) based on the existing FlexAttention API, and brings optimized performance on x86 CPUs. With this feature, it’s easy to use FlexAttention API to compose Attention solutions on CPU platforms and achieve good performance.
135+
136+
### [Prototype] Dim.AUTO
137+
138+
`Dim.AUTO` allows usage of automatic dynamic shapes with `torch.export`. Users can export with `Dim.AUTO `and “discover” the dynamic behavior of their models, with min/max ranges, relations between dimensions, and static/dynamic behavior being automatically inferred.
139+
140+
This is a more user-friendly experience compared to the existing named-Dims approach for specifying dynamic shapes, which requires the user to fully understand the dynamic behavior of their models at export time. `Dim.AUTO` allows users to write generic code that isn’t model-dependent, increasing ease-of-use for exporting with dynamic shapes.
141+
142+
Please see [torch.export tutorial](https://pytorch.org/tutorials/intermediate/torch_export_tutorial.html#constraints-dynamic-shapes) for more information.
143+
144+
### [Prototype] CUTLASS and CK GEMM/CONV Backends for AOTInductor
145+
146+
The CUTLASS and CK backend adds kernel choices for GEMM autotuning in Inductor. This is now also available in AOTInductor which can run in C++ runtime environments. A major improvement to the two backends is improved compile-time speed by eliminating redundant kernel binary compilations and dynamic shapes support.

assets/images/1738166706211.jpg

71.8 KB
Loading

assets/images/pt-26-live-q-a.png

320 KB
Loading

assets/quick-start-module.js

+3-3
Large diffs are not rendered by default.

index.html

+3-3
Original file line numberDiff line numberDiff line change
@@ -31,11 +31,11 @@ <h3>Blog</h3>
3131
</p>
3232
</div>
3333
<div class="col-md">
34-
<h3>PyTorch 2.5</h3>
34+
<h3>PyTorch 2.6</h3>
3535
<p>
36-
Featuring a new CuDNN backend for SDPA, improvements to TorchDynamo, regional compilation of torch.compile, and more.
36+
Featuring Python 3.13 support for torch.compile, several AOTInductor enhancements, FP16 support on X86 CPUs, and more.
3737
<p>
38-
<a href="/blog/pytorch2-5/">Learn more</a>
38+
<a href="/blog/pytorch2-6/">Learn more</a>
3939
</p>
4040
</div>
4141
<div class="col-md">

0 commit comments

Comments
 (0)