Subnet 64

Chutes

Rayon Labs

The Serverless AI Compute subnet enables quick deployment, running, and scaling of AI models via API.

SN64 : Chutes

SubnetDescriptionCategoryCompany
SN64 : ChutesGeneral purpose computeCompute
Generative Al
Rayon Labs

The Serverless AI Compute subnet allows users to deploy, run, and scale any AI model in seconds. Whether directly through the platform or via a simple API, this subnet offers a seamless, fast, and efficient solution for AI model management.

Images

Images are Docker images on which all applications (called “chutes”) run within the platform. These images must meet specific requirements, including having a CUDA installation (preferably version 12.2-12.6) and Python 3.10+ installed.

Chutes

A “chute” is an application running on top of an image. Each chute is essentially a FastAPI application, designed to perform specific tasks.

Cords

A “cord” is a single function within the chute, analogous to a route and method in FastAPI. Each cord performs a specific action or process within the chute.

Graval

Graval is a middleware library that ensures the GPUs being used are authentic and meet certain performance standards. It runs within the chute to validate hardware claims and perform additional checks, such as VRAM capacity and device information verification.

Registration

To use the platform, users must register with a Bittensor wallet and hotkey. Registration allows users to create API keys for secure access to the platform and services.

Creating API Keys

After registration, users can create API keys with different levels of access. These keys allow control over specific aspects of the platform, such as image management or chute deployment.

Validators and Subnet Owners

Validators and subnet owners on Bittensor can link their keys to a chute account, granting them free access and developer roles. This process ensures that validators and subnet owners have special privileges within the platform.

Developer Role and Deposit

To create and deploy chutes, users must first deposit a refundable amount of Tao, which serves as a developer deposit. This helps prevent abuse on the platform. After seven days, users can request their deposit back.

Building an Image

To deploy an application, users must first build an image. The platform provides a base image to start with, including Python 3.12.7 and necessary CUDA packages. Users can customize the image to include their desired dependencies and applications.

Deploying a Chute

Once an image is built and ready, users can deploy it as a chute. Deployment allows applications to run in a scalable, distributed environment, with various configurations to ensure optimal performance (e.g., GPU count, VRAM requirements).

Custom Chutes

Users can build custom chutes tailored to their needs. Chutes are flexible and can support various functions, including complex processes and arbitrary web servers. Functions within the chutes are decorated with special annotations (cords) to define their behavior.

Local Testing

Before deploying an image or chute, users can test them locally. This process allows users to verify functionality and debug issues in a controlled environment before deployment.

This platform provides users with the tools to build, deploy, and manage AI applications in a decentralized and scalable manner, allowing for flexibility and efficiency.

Namoray – Chief Incentive Mechanism Officer