Subnet 81
Grail
Templar
Permissionless, incentivized distributed training of large-scale AI models on a global GPU network.

SN81 : Grail
| Subnet | Description | Category | Company |
|---|---|---|---|
| SN81 : Grail | LLM post-training | Decentralized Training | Covenant |
Bittensor Subnet 81 “Grail” is a decentralized AI training network that enables permissionless, collaborative training of large-scale machine learning models. In essence, Grail lets a global community of miners (contributors with hardware) collectively train a shared large language model (LLM) by pooling their compute power, with all coordination handled on-chain via the Bittensor protocol. The subnet was acquired and rebranded as “Grail” in mid-2025 (it was previously known as Patrol under Tensora Group).
Its core purpose is to crowdsource the training of AI models in an open, incentive-aligned way – often referred to as the “holy grail” of decentralized AI research, since cracking distributed training would allow communities to rival the giant centralized AI labs. In Grail’s network, any participant around the world with the required hardware can join to train the model and earn rewards, making it a permissionless and trust-minimized system for AI development. By leveraging Bittensor’s blockchain, Grail ensures that contributors are fairly incentivized in cryptocurrency for the value of their work, while producing an open-source model that is co-owned by the community. Ultimately, Grail aims to democratize AI – enabling state-of-the-art AI models to be built by the community rather than behind closed doors, and ensuring that the benefits (and token rewards) are distributed to the many contributors who help train these models.
Grail is essentially the deployment of Templar’s decentralized training framework on Subnet 81, comprising a network of software miners and validators that cooperate via the blockchain to train a shared machine learning model. The “product” is twofold: (1) the protocol and infrastructure that perform this distributed training, and (2) the resulting AI model that the network produces. Technically, Grail’s build includes several key components and innovations:
Miners (Training Nodes): These are participant nodes running the Grail miner software on their GPUs to train the model. Each miner receives a deterministic slice of the training dataset and computes a “pseudo-gradient” update on the model using that data. Miners operate in synchronous rounds (e.g. ~84-second windows) where they intensively train on their data shard and then upload their gradient contributions to a decentralized storage bucket (Grail uses a cloud storage layer). The miners’ goal is to produce gradients that improve the global model’s performance (i.e. lower the loss on the given data) more effectively than other peers. They do this without any central coordinator – instead, coordination happens via the Bittensor chain and shared storage: miners fetch data, compute updates, and submit those updates all according to the timetable enforced by the blockchain.
Validators (Evaluation Nodes): Grail’s validators are specialized nodes that download the miners’ submitted gradients from the storage and evaluate their quality. A validator will test each gradient by applying it to a copy of the model and measuring the loss reduction on a validation dataset sample. Essentially, the validator serves as an impartial judge: checking that the gradient was submitted on time (using timestamps compared against blockchain block times) and that it actually provides a meaningful improvement to the model’s accuracy. If a miner’s gradient fails to beat a baseline (or if it’s submitted outside the allowed window), the validator will flag it as a poor contribution – such low-quality or late submissions lead to the miner getting slashed or receiving little to no reward for that round. High-quality contributions, on the other hand, earn the miner higher token rewards, proportional to the measured improvement their work provided. The validator then aggregates approved gradients and applies them to update the global model parameters. In Grail’s architecture, an aggregator component may also be used to accumulate the gradients and periodically save model checkpoints.
Bittensor Blockchain Integration: The Bittensor (TAO) chain underpins the whole coordination and incentive mechanism. Grail is built as a Bittensor subnet, meaning the blockchain keeps track of participant registrations (miners/validators join the subnet by bonding tokens and are identified on-chain) and handles weight setting and reward payouts based on contribution. For each training round, the validator posts scores or “weights” for miners to the chain, reflecting their performance. These on-chain weights determine how the block rewards (in the subnet’s native token) are allocated among miners – effectively paying each contributor in proportion to the utility of their gradient in improving the model. The use of blockchain provides a trustless ledger of contributions and rewards, ensuring transparency and deterring cheating in the absence of a central authority.
Communication & Data Pipeline: Grail implements a peer-to-peer communication system for exchanging model updates and data. Rather than directly sending gradients to each other (which would be bandwidth-intensive), miners upload their gradients to the shared storage bucket and use Bittensor’s built-in networking to signal availability. This design decouples the expensive data transfer from the blockchain’s limited bandwidth. The Grail software employs gradient compression techniques to reduce communication overhead – for example, using algorithms like Decoupled Momentum (DeMo) with Top-K sparse updates and even applying transforms (like discrete cosine transform) to compress gradients before upload. These techniques are crucial to make distributed training feasible over the internet by cutting down the data each miner must send while preserving the effectiveness of updates. Grail’s framework also ensures that each miner performs unique work (e.g. each miner gets a different data slice) to avoid redundancy; a mechanism is in place to assure miners aren’t all submitting the same gradient or copying one another’s work. This uniqueness check, combined with an OpenSkill rating system that tracks each miner’s performance over time, helps maintain healthy competition and continuous contribution quality.
In summary, Grail’s build is a distributed training pipeline. The system continuously iterates through training rounds where miners train the shared model on their local data shards and validators merge the useful results. The end “product” is a collaboratively-trained AI model (for instance, a language model with on the order of a billion+ parameters) that is co-owned by the community of contributors. Notably, Grail is one of the first examples of a fully incentive-driven, permissionless AI training run in practice – an extension of the pioneering work done on Subnet 3 (Templar) which proved that this concept can work at smaller scale. With Grail, the Templar team is pushing that vision further, using the lessons learned and improved software (sometimes dubbed “Gauntlet” for its incentive mechanism and other code optimizations) to tackle even larger models and more complex training regimes. All of the software is open-source (the Templar team’s code repositories are publicly available) and the network operates without centralized control – making Grail a truly decentralized AI training platform.
Grail is developed and managed by the Templar team, an experienced group of engineers and researchers dedicated to decentralized AI. The Templar organization (sometimes referred to as Templar AI or tplr.ai) previously launched Subnet 3 “Templar” – the world’s first distributed, permissionless LLM training subnet – and also created Subnet 39 “Basilica” focusing on trustless GPU compute. The team is led by Samuel Dare, a blockchain veteran and visionary who helped conceive the idea of incentive-driven AI training (“Sam” often goes by the handle “distributed” in the community). Alongside Sam, Templar includes AI researchers such as Joel Lidin, Amir Sarfi, and Evangelos Pappas, who co-authored a 2025 research paper detailing the Grail/Templar training framework and its incentive system. These individuals (affiliated with Templar AI) bring expertise from both machine learning and crypto – for example, one co-author, Eugene Belilovsky, is a research professor at Mila/Concordia who advises on the project, and Jacob Steeves (another co-author) is a founder of the OpenTensor Foundation.
In practice, the Templar team functions as the initial “custodians” of Subnet 81, meaning they set up the network, deploy code updates, and guide its development in these early stages. However, they emphasize that Grail is a community endeavor – the end goal is to decentralize control as the network matures. The team has a strong track record: their first subnet (Templar SN3) successfully demonstrated distributed training with ~200 GPUs on a 1.2B parameter model, overcoming numerous technical hurdles and exploits through rapid iteration and community collaboration. They have fostered an open developer community around these subnets, encouraging miners and validators to contribute not just compute but also improvements to the code. Key members like Sam (“distributed”) are active in public forums, sharing progress and inviting feedback. The team also operates under a thematic ethos (drawn from Knights Templar lore) – with project names like Templar, Basilica, and Grail, reflecting their mission to build “sacred” infrastructure for AI. Templar’s core contributors are passionate about AI decentralization, and they are continually recruiting talent (researchers, ML engineers) to join the cause of “building the future of decentralized AI training,” as noted on their website.



