We’re excited to announce the 2026 PyTorch Docathon May 5-19! Refine technical docs, test tutorials in CI, and accelerate the transition from research to production. Open to all skill levels with support on Discord.
RSVP now: https://bit.ly/4sTVLYb
#PyTorch #OpenSource #AI
Posts by PyTorch
CFP is OPEN for PyTorch Conference 2026 in San Jose!
Share advancements in Core PyTorch, vLLM, DeepSpeed, and Ray. Blog: https://bit.ly/4c1ShM1
Deadlines:
Sessions: June 7
Posters: July 26
Save with Super Early Bird rates through April 10.
#PyTorchCon
The PyTorch Ecosystem Working Group welcomes PhysicsNeMo, Unsloth, ONNX, and KTransformers to the Landscape.
This map highlights innovative projects that extend, integrate with, or build upon PyTorch.
Read more: pytorch.org/blog/pytorch-ecosystem-l...
#PyTorch #OpenSource #AI
"PyTorch is probably the most important piece of open source software most enterprise technology leaders have never had a governance conversation about."
Mark Collier at KubeCon on why neutral governance is AI's path to market. Full diginomica.com article: https://bit.ly/4tpIUNa
PyTorch 2.11 Release Live Q&A w/ Andrey Talman & Nikita Shulga on Tuesday, March 31, 10 AM PT.
-Differentiable Collectives
-FlexAttention: FlashAttention-4 on Hopper/Blackwell
-MPS Operator expansion
-RNN/LSTM GPU Export
-XPU Graph
Register: https://pytorch.org/event/pytorch-2-11-release-live-qa/
#NCCL watchdog timeouts are often misunderstood. Meta’s analysis shows >60% are caused by CPU-side stuckness or divergence, not the network. This guide explains using #FlightRecorder to trace collective states and fix hangs
Read: https://bit.ly/4bCqItC #OpenSourceAI #PyTorch
Paris ML Systems Hackathon on April 9
Join #PyTorch Foundation and GPU MODE for a day-long build:
- Distributed training and inference tracks
- B300 and H200 access
- Prizes: GB300 NVL72 rack access
- Talks: PyTorch (Helion), vLLM, Prime Intellect
Register: https://bit.ly/4bSdKqE
PyTorch and Nebius collaborated to speed up DeepSeek-V3 pre-training (16B & 671B) on 256 NVIDIA B200 GPUs. Combining MXFP8 via TorchAO and DeepEP yielded +41% throughput vs BF16.
Full blog:
https://bit.ly/4uN3yIJ
PyTorch 2.11 features improvements for distributed training and hardware operator support. Join Andrey Talman and Nikita Shulga on Tuesday, March 31st at 10 am for a live update and Q&A.
Register: pytorch.org/event/pytorc...
#PyTorch #OpenSource #AI
PyTorch 2.11 is now available, featuring 2,723 commits from 432 contributors. Highlights: FlashAttention-4 for Blackwell/Hopper, Differentiable Collectives, XPU Graph for Intel GPUs, and expanded MPS support.
Release notes: pytorch.org/blog/pytorch...
PyTorch 2.10 is now optimized for Intel Core Ultra Series 3 processors to bring high-performance AI to the PC and edge.
Read our latest blog from the Intel PyTorch and Client AI SW teams for the full technical deep dive and benchmarks:
https://pytorch.org/blog/pytorch-2-10torchao/
TorchSpec and Mooncake teams introduce TorchSpec: a torch-native framework for speculative decoding training. By streaming hidden states via Mooncake, it enables disaggregated pipelines where inference and training scale independently.
https://bit.ly/47eBfIR
Build differentiable computational physics with NVIDIA Warp. It bridges CUDA and Python for high-performance GPU kernels with native auto-diff. Interoperable with PyTorch, JAX, and NumPy.
https://bit.ly/4uG78UQ
GDPA introduces an attention kernel for RecSys, replacing softmax with flexible activations. Deployed in Meta’s GEM model, it achieves 1,145 BF16 TFLOPs (97% utilization) on NVIDIA B200, outperforming FA4 by 3.5× in short K/V settings.
https://bit.ly/418LQl8
#ExecuTorch addresses fragmented native deployment for #AI agents as a #PyTorch native platform. It enables voice models across CPU, GPU, and NPU on Android, iOS, Linux, macOS & Windows
🔗 pytorch.org/blog/building-voice-agen...
Before we head to Paris for PyTorch Conference EU 2026, we’re looking back on 2025 keynotes from visionary AI leaders.
Starting with Eli Uriegas (@_seemethere) from Meta: 11k commits and 794M minutes of CI/CD compute.
Watch: https://youtu.be/xWjXsP1E5mQ?si=JRIVHQ06s3IvYPDq
#PyTorch #OpenSourceAI
MXFP8 training for MoEs on GB200s enables a 1.3x speedup with equivalent convergence versus BF16:
🔗 pytorch.org/blog/mxfp8-training-for-...
PyTorch Foundation is attending Optimized AI Conference in Atlanta, April 14-16. Join 100+ experts to discuss #LLM operations, #RAG, and #InferenceOptimization.
Get 20% off with code: OAIC-20.
Details: oaiconference.com.
#PyTorch #AIInfrastructure #OpenSourceAI
DeepNVMe just got faster and more flexible:
✅ Gen5 NVMe support
✅ 20X faster model checkpointing
✅ Cost-efficient SGLang inference via ZeRO-Inference
✅ CPU-only pinned memory support
📘 pytorch.org/blog/deepnvm...
#PyTorch #DeepSpeed #AIInfrastructure
The #PyTorchFoundation newsletter is your go-to source for the latest updates, events, and community insights to build and innovate with #PyTorch—all in support of accelerating #OpenSourceAI.
📬 Subscribe: pytorch.org/newsletter/
📖 June: pytorch.org/newsletter/j...
Update from the PyTorch ecosystem: The latest NVIDIA
DALI release adds DALI Proxy—making it easier to accelerate parts of your PyTorch DataLoader pipeline without a full refactor.
Learn more
🔗 developer.nvidia.com/blog/unlock-...
#PyTorch #OpenSourceAI #DataPipelines #DeepLearning
🧠 Responsible AI is a design decision—and a strategic edge.
This new guide shows how to build a Yellow Teaming assistant using PyTorch and AWS Graviton4 to surface risks early and build more accountable systems.
🔗 pytorch.org/blog/build-r...
#ResponsibleAI #LLM #PyTorch #builtonArm
⏳ Just a few days left to apply for the PyTorch Ambassador Program.
If you're making an impact with PyTorch through research, code, education, or community work, now’s your chance to join a global network of ML leaders.
📅 Deadline: June 7
🔗 pytorch.org/programs/amb...
#PyTorch #AICommunity
Join us at #GTC25Paris25 for the session “10x Your GPU Power with #Python: Python for Programming the GPU”
Learn how Python now matches the performance and control of C++ #CUDA.
Explore #PyTorch, CuPy, RAPIDS, cuda.parallel, numba.cuda, cuTile, etc.
🔗 www.nvidia.com/en-eu/gtc/se...
Mixture-of-Experts (MoE) is a popular #LLM architecture that reduces computation by activating fewer parameters per token. But it brings memory, communication, & control challenges.
💡We introduce MetaShuffling, enabling efficient Llama 4 model inference in production. 🔗 pytorch.org/blog/metashu...
The PyTorch Foundation is a Gold Sponsor of #MLSys2025 this week in Santa Clara.
Visit the booth and explore talks from Soumith Chintala, Ion Stoica, and Exec Dir Matt White on open source AI and scalable ML systems.
🔗 pytorch.org/blog/pytorch...
#PyTorch #OpenSourceAI #AIInfrastructure
🖇️: pytorch.org/event/toward...
PyTorch Foundation has expanded into an umbrella foundation.
vLLM and DeepSpeed have been accepted as hosted projects, advancing community-driven AI across the full lifecycle.
Quotes from AMD, AWS, Arm, Huawei, HuggingFace, IBM, Intel, LightningAI, Meta.
Read more: pytorch.org/blog/press-r...
Can language model systems autonomously complete entire tasks end-to-end?
In our next Expert Exchange webinar, Ofir Press explores autonomous LM systems for software engineering, featuring SWE-bench & SWE-agent—used by OpenAI, Meta, & more.
🔗 pytorch.org/autonomous-l...
#PyTorch #AI #OpenSource
TODAY: Join PyTorch Core Maintainers Piotr Bialecki (NVIDIA) and Nikita Shulga (Meta) for a live Q&A session on the #PyTorch 2.7 release at 12 PM PST.
Have questions? Drop them below, & we'll share them during the webinar.
🔗 More info: pytorch.org/pt-27-releas...
#MachineLearning #OpenSourceAI