Orders placed by US customers today will be shipped via Express air courier shipping and should arrive in time for Christmas so you can out-drip everyone at your family party and/or work function.
Posts by Nous Research
left to right: a black sweater. The front features white embroidery of a stylized nous text logo and the back is screen printed. a white t-shirt with blue embroidery of a stylized nous text logo on the front and screen printing on the back. a white t-shirt with black embroidery of a stylized nous text logo on the front and screen printing on the back.
NOUS RESEARCH Q4 2025 DROP
All items made entirely in the USA out of 100% cotton, featuring embroidery on the front and screen printing on the back.
1. "Good Technology" Super Heavy Weight Oversized Sweater
2. "Scaffolding" Tee
3. "Open Source" Tee
shop.nousresearch.com/collections/...
Controlling text generation and structure remains a difficult problem to solve.
Our newest blog post and release from Researcher in Residence @yaboilyrical (on Twitter) explores how this problem becomes solvable using Sequential Monte Carlo approximation.
nousresearch.com/steering-the...
Come join Nous and the Solana Foundation in NYC on Thursday, May 22nd, to discuss decentralized AI and Nous's efforts to democratize intelligence, including Psyche.
Limited capacity. Apply below👇
lu.ma/39b7e9pu?v=1
As always, we couldn't have gotten here without your help. Special thanks to our team, our community, and the open source movement.
Psyche’s initial training infrastructure is just the beginning of our journey. We plan to integrate full post training stages - supervised finetuning and reinforcement learning workloads, inference, and other parallelizable workloads in the creation and serving of AI going forward.
Looking ahead, we will draw model ideas from the community via our forum and Discord. By enabling highly parallel and scalable experimentation, we’re betting that the next innovation in model creation and design will come from the open source community
The resulting model will be small enough to train on a single H/DGX and run on a 3090, but will be powerful enough to serve as the basis for strong reasoning models and creative pursuits. The model will be trained continually without a final annealing step, resulting in a true unaltered base model.
40b parameters. 20t tokens. MLA architecture.
We are launching testnet with the pre-training of a 40B parameter LLM:
- MLA Architecture
- Dataset consisting of FineWeb (14T) + FineWeb-2 minus some less common languages (4T), and The Stack v2 (1T)
If you have 64+ H100 GPUs, Contact engineering@nousresearch.com to apply to provide hardware to the network’s training pool.
While compute on the network needs to be trusted and approved at this time, we plan to support trustless, community-owned compute resources. For now, open source enthusiasts can contribute via our mining pool, and we will be onboarding more nodes over the next weeks.
Psyche uses the Solana blockchain to decentralize parts of the core infrastructure for coordination and stores attestations for the nodes operating within the network. This design takes meaningful steps towards decentralization while ensuring training does not become too costly or redundant.
Training used to have a bandwidth constraint that kept the process centralized. In 2024, Nous's DisTrO optimizers broke through that constraint. With Psyche, we have created a custom peer-to-peer networking stack to coordinate globally distributed GPUs running DisTrO.
All the links!
Blog: nousresearch.com/nous-psyche/
Training dashboard: psyche.network
Code: github.com/PsycheFounda...
Docs: docs.psyche.network
Forum: forum.psyche.network
HuggingFace: huggingface.co/PsycheFounda...
Discord: discord.com/invite/jqVph...
This run represents the largest pre-training run conducted over the internet to date, surpassing previous iterations that trained smaller models on much fewer data tokens.
We are launching our testnet today with the pre-training of a 40B parameter LLM, a model powerful enough to serve as a foundation for future pursuits in open science.
Psyche is a decentralized training network that makes it possible to bring the world’s compute together to train powerful AI, giving individuals and small communities access to the resources required to create new, interesting, and unique large scale models.
A diagram of the Psyche network. It shows the relationship between the Solana Coordinator, an individual training client, the DisTrO optimizer inside a client, forward/backward passes to create gradients, the transmission and dissemination of the created DisTrO results, and the ingestion of data from a data provider.
Announcing the launch of Psyche
nousresearch.com/nous-psyche/
Nous Research is democratizing the development of Artificial Intelligence. Today, we’re embarking on our greatest effort to date to make that mission a reality: The Psyche Network
To ensure a smooth rollout, we made a waitlist: portal.nousresearch.com
- Access will be granted on a first-come, first-served basis
- Once granted access, you can create API keys and purchase credits
- OpenAI-compatible API
- Right now all accounts start off with $5.00 in free credits.
Today we’re releasing our Inference API that serves Nous Research models. We heard your feedback, and built a simple system to make our language models more accessible to developers and researchers everywhere.
The initial release features two models - Hermes 3 Llama 70B and DeepHermes 3 8B Preview
You can now experiment with Psyche’s DisTrO-enabled training code on our GitHub, and the larger open-sourced distributed training stack will be released alongside testnet. github.com/PsycheFounda...
Read more in our blog post: nousresearch.com/nous-psyche/
Recent AI breakthroughs challenge the status quo narrative that only closed, mega labs have the ability to push the frontier of superintelligence.
Today we announce Nous Psyche built on
@solana.com
www.youtube.com/watch?v=XMWI...
Run Hermes on phones, laptops, and CPUs without sacrificing speed, and may also be a great pairing with 70B for speculative decoding!
Learn more about Hermes, see our technical report, and chat with it now: nousresearch.com/hermes
Introducing a smol Hermes 3 LLM!
Hermes 3 3B is now available on huggingface alongside quantized GGUF versions to make it even smaller.
More info and download links here: huggingface.co/NousResearch...
Hermes 3 3B was built by Teknium, Roger Jin, Jeffrey Quesnelle and "nullvaluetensor".
Thursday December 12th
Doors @ 6pm, Talks @ 7pm
DCTRL, 436 W Pender St, Vancouver
Open Entry. Food + Drink + Merch.
DisTrO Demystified - Jeffrey Quesnelle , Bowen Peng
Why Decentralization Matters - Mark Murdock
Mapping Uncertainty at Inference Time - _xjdr
A graphic of the Nous Girl inside a circle. The graphic is blue and white. The circle has text wrapped around the edges that reads "REBELLION TO TYRANTS IS OBEDIENCE TO GOD."
We’re here to put the power of artificial intelligence into the hands of the many rather than the privileged few.
DeMo was created in March 2024 by Bowen Peng and Jeffrey Quesnelle and has been published on arXiv in collaboration with Diederik P. Kingma, co-founder of OpenAI and inventor of the Adam optimizer and VAEs.
The paper is available here: arxiv.org/abs/2411.19870
And code: github.com/bloc97/DeMo
We harness both Nous DisTrO, our novel networking stack that reduces inter-GPU communication by up to 10,000x during pretraining, and the testnet code for Psyche, a decentralized network that builds on Nous DisTrO to autonomously coordinate compute for model training.
Psyche details coming soon.