Transparent, Scalable Pricing for your AI Workloads
From development to deployment, our pricing adapts to your needs

AI Workspace
Tailored for your AI Workloads, With Flexible Instances.
AI Lab
Designed for your educational needs, purchase as per daily lab requirements.
Bare Metal
Raw power at a predictable cost—optimized for your high-performance AI workloads.
Shakti Clusters
Scale seamlessly with transparent pricing for powerful, AI-optimized clusters.
AI Endpoints
Deploy AI models effortlessly with cost-effective, per-usage pricing.
Serverless
Maximize efficiency with no upfront costs—only pay for what your AI workloads consume.
Microsoft Azure AI Services
Experience Azure AI on Yotta Shakti Cloud with high-performance AI solutions.
Yotta Sarvam AI Services
Yotta Sarvam AI delivers scalable models, seamless deployment, and flexible pricing.
Add-On Services
Explore our add-on services to enhance performance.
AI Workspace Pricing
Shakti Cloud AI Workspace delivers a seamless experience with flexible configurations, offering Virtual Machine to meet diverse AI workload demands.
Medha VM
-
GPU : 1 x Nvidia H100 (80GB)
-
vCPU : 24
-
RAM : 240 GB
-
Block Storage : 2 TB
-
Unlimited Data Ingress & Egress
Nakshatra VM
-
GPU : 1 x Nvidia L40S (48GB)
-
vCPU : 24
-
RAM : 240 GB
-
Block Storage : 2 TB
-
Unlimited Data Ingress & Egress
Application Gateway
-
Distributes incoming traffic across targets like web traffic (HTTP/HTTPS)- Layer 7
Per GB
Cloud App Service
-
Platform for deploying and managing web apps with infrastructure abstraction.
Per Unit
Cloud DDOS Protection
-
ALWAYS on DDOS Protection against DDoSAttacks - Upto Clean Traffic per Mbps
Per Unit
Cloud DNS
-
DNS Service (per 1 million DNS Queries)
Per Mbps
Cloud Database for MySQL Community
-
Managed MySQL DB with Community subscription. Built on open-source binaries. Supports monitoring, patching, and backup. VM Infra to be subscribed seperately.
Per Unit
Cloud Database for MySQL Enterprise
-
Managed MySQL DB with enterprise subscription(Enterpise Edition 2 core license, minimum subscription is 4 core). Includes enterprise binaries, updates, backup, and monitoring. VM Infra to be subscribed seperately
Per Unit
Cloud Database for PostgreSQL Community
-
Managed PostgreSQL DB with open-source community edition license. Includes binaries, support for backup, patching, and configuration. VM Infra to be subscribed seperately
Per Unit
Cloud Database for PostgreSQL Enterprise
-
Fully managed PostgreSQL DB with enterprise subscription (min 4 cores for 3 years or 8 cores for 1 year). Includes binaries, licensing, maintenance, patching, and backup. VM Infra to be subscribed seperately
Per Unit
Redis Enterprise
-
Fully managed, enterprise-grade Redis service offering in-memory performance with high availability, clustering, and persistence. Ideal for caching, real-time analytics, and low-latency workloads. VM Infra to be subscribed seperately.
Per Unit
Redis Community
-
Managed Redis instance (in-memory key-value store) with community support. Suitable for caching, session state, and analytics. VM Infra to be subscribed seperately.
Per Unit
SQL Database Standard
-
Managed MS SQL Enterprise edition DB (licensed in 2-core packs, min 4 cores). Includes license and updates. VM Infra to be subscribed seperately
Per Unit
SQL Database Enterprise
-
Managed MS SQL Enterprise edition DB (licensed in 2-core packs, min 4 cores). Includes license and updates. VM Infra to be subscribed seperately
Per Unit
Cloud Kubernetes Service
-
Management service for orchestrating, scaling, and securing Kubernetes clusters. (Container infra to be subscribed seperately)
Per Unit
Cloud Monitoring & Notifications
-
Comprehensive monitoring platform for resources, infrastructure and workloads etc.
Data Ingested Per GB, Per API Calls
Backup, Recovery and Archival Service
-
Safe and Secure Backup Recovery of Virtual Machine
-
Disk Backup with Unlimited Data transfer Per Instance
-
If multiple VM backup required then need to subscribe in multiple units or instances
-
Default Backup Policy implemented DIWF, 14 Days Retention Period,But customer can customise as per his organisational policy.
-
Restoration on the same server is included, If other than same server then the resource and efforts are charged separately.
-
Includes Snapshot, File System, and Database backups in VMs
-
Disk Space for backup needs to be subscribed separately as per consumption requirement
Per Instance
Internet Bandwidth
-
Internet Bandwidth
Per Mbps
Content Delivery Network
-
Globally distributed network for fast, secure content delivery, reducing latency and load times for websites, apps, and media.
Per GB
Container Registry
-
Managed Registry offering centralized storage and distribution of container images used by native resources.
Supports versioning, tagging, and seamless integration with Kubernetes environments. Provides global accessibility for containerized applications.
Per GB
Insight and Analytics
-
Platform providing analytics and insights into log and telemetry data collected from resources
Per GB
Key Vault
-
Securely manage up to 10 encryption keys with our Key Vault Service.
-
KMS protects your data, controls access, and simplifies compliance from a centralized interface.
Per Unit
Network Load Balancer – Layer 4
-
Distribute traffic efficiently across your backend resources with our high-performance Layer 4 Load Balancer. Ensure low latency, high availability, and seamless scalability for your applications.
Per Mbps
AI Lab Pricing
Select the AI Lab that fits your budget and training needs — opt for unlimited hours or flexible time slots
Instant AI Lab, zero hassle. Just subscribe to get GPU access and the AI Lab platform—with unlimited user onboarding. Bring your own storage or add it separately. Comes with pre-configured ML/DL containers, so your students can start learning from day one.
Workstation Name |
Description |
GPU Memory |
Monthly price |
---|---|---|---|
AI-Lab Workstation |
|
8 GB L40S |
₹ 30,000 |
AI-Lab Workstation |
|
16 GB L40S |
₹ 45,000 |
AI-Lab Workstation |
|
40 GB H100 |
₹ 90,000 |
AI-Lab Workstation |
|
80 GB H100 |
₹ 1,40,000 |
AI-Lab Workstation |
|
160 GB H100 |
₹ 2,80,000 |
AI-Lab Workstation |
|
320 GB H100 |
₹ 5,60,000 |
Total control, ultimate flexibility. Choose your GPUs, scale users as needed and run the AI Lab platform your way. Create multiple GPU profiles—let students use GPUs on weekdays and researchers take over on weekends. Includes storage options and pre-configured ML/DL containers for a seamless setup.
Workstation Name |
Description |
|
Monthly Price Per GPU |
---|---|---|---|
GPU |
Select the GPU capacity that fits your AI Lab needs |
Click Here |
|
AI Lab Platform Access |
|
₹ 70,000 |
Bare Metal Pricing
Our budget-friendly plans are designed to provide the best value, tailored to meet your training needs
Plan Description |
Monthly Contract |
6 Months Contract |
12 Months Contract |
24 Months Contract |
36 Months Contract |
48 Months Contract |
---|---|---|---|---|---|---|
Bare Metal 8 x HGX H100
|
₹ 325 |
₹ 313 |
₹ 243 |
₹ 234 |
₹ 226 |
₹ 217 |
Plan Description |
Monthly Contract |
6 Months Contract |
12 Months Contract |
24 Months Contract |
36 Months Contract |
48 Months Contract |
---|---|---|---|---|---|---|
Bare Metal 4 x L40S
|
₹ 182 |
₹ 161 |
₹ 147 |
₹ 145 |
₹ 143 |
₹ 141 |
Shakti GPU Cluster Pricing
Our Budget-friendly plans are designed to provide the best value, tailored to meet your training needs.
Plan Description |
Monthly Contract |
6 Months Contract |
12 Months Contract |
24 Months Contract |
36 Months Contract |
48 Months Contract |
---|---|---|---|---|---|---|
Shakti Cloud – HGX H100 Cluster
|
₹ 357 |
₹ 345 |
₹ 267 |
₹ 257 |
₹ 249 |
₹ 239 |
Plan Description |
Monthly Contract |
6 Months Contract |
12 Months Contract |
24 Months Contract |
36 Months Contract |
48 Months Contract |
---|---|---|---|---|---|---|
Shakti Cloud – L40S Cluster
|
₹ 200 |
₹ 177 |
₹ 162 |
₹ 160 |
₹ 158 |
₹ 156 |
AI Endpoint Pricing
Select the right endpoint for your project and enjoy transparent, token-based pricing designed to meet your unique AI needs.
Meta/Llama3-8b-instruct
-
NVIDIA NIM for GPU accelerated Llama 3 8B inference through OpenAI compatible APIs
Llama-3.1-8b-instruct
-
NVIDIA NIM for GPU accelerated Llama 3.1 8B inference through OpenAI compatible APIs
Meta/Llama3-70b-instruct
-
NVIDIA NIM for GPU accelerated Llama 3 70B inference through OpenAI compatible APIs
Llama-3.1-405b-instruct
-
NVIDIA NIM for GPU accelerated Llama 3.1 405B inference through OpenAI compatible APIs
Llama-3.1-70b-instruct
-
NVIDIA NIM for GPU accelerated Llama 3.1 70B inference through OpenAI compatible APIs
Mixtral-8x7B-Instruct-v0.1
-
NVIDIA NIM for GPU accelerated Mixtral-8x7B-Instruct-v0.1 inference through OpenAI compatible APIs
Llama-3.1-8b-base
-
NVIDIA NIM for GPU accelerated Llama 3.1 8B inference through OpenAI compatible APIs
Mixtral-8x22B-Instruct-v0.1
-
NVIDIA NIM for GPU accelerated Mixtral-8x22B-Instruct-v0.1 inference through OpenAI compatible APIs
meta-llama-2-13b-chat
-
NVIDIA NIM for GPU accelerated Llama 2 13B inference through OpenAI compatible APIs
meta-llama-2-70b-chat
-
NVIDIA NIM for GPU accelerated Llama 2 70B inference through OpenAI compatible APIs
Llama-3-Taiwan-70B-Instruct
-
NVIDIA NIM for GPU accelerated Llama-3-Taiwan-70B-Instruct inference through OpenAI compatible APIs
nemotron-4-340b-instruct
-
NVIDIA NIM for GPU accelerated Nemotron-4-340B-Instruct inference through OpenAI compatible APIs
meta-llama-2-7b-chat
-
NVIDIA NIM for GPU accelerated Llama 2 7B inference through OpenAI compatible APIs
Mistral-7B-Instruct-v0.3
-
NVIDIA NIM for GPU accelerated Mistral-7B-Instruct-v0.3 inference through OpenAI compatible APIs
Nemotron-4-340B-Reward
-
NVIDIA NIM for GPU accelerated Nemotron-4-340B-Reward inference through OpenAI compatible APIs
Llama-3-Swallow-70B-Instruct-v0.1
-
NVIDIA NIM for GPU accelerated Llama-3-Swalow-70B-Instruct-v0.1 inference through OpenAI compatible APIs
Meta/Llama3-8b-instruct
-
NVIDIA NIM for GPU accelerated Llama 3 8B inference through OpenAI compatible APIs
Llama-3.1-8b-instruct
-
NVIDIA NIM for GPU accelerated Llama 3.1 8B inference through OpenAI compatible APIs
Meta/Llama3-70b-instruct
-
NVIDIA NIM for GPU accelerated Llama 3 70B inference through OpenAI compatible APIs
Llama-3.1-405b-instruct
-
NVIDIA NIM for GPU accelerated Llama 3.1 405B inference through OpenAI compatible APIs
Llama-3.1-70b-instruct
-
NVIDIA NIM for GPU accelerated Llama 3.1 70B inference through OpenAI compatible APIs
Mixtral-8x7B-Instruct-v0.1
-
NVIDIA NIM for GPU accelerated Mixtral-8x7B-Instruct-v0.1 inference through OpenAI compatible APIs
Mixtral-8x22B-Instruct-v0.1
-
NVIDIA NIM for GPU accelerated Mixtral-8x22B-Instruct-v0.1 inference through OpenAI compatible APIs
meta-llama-2-13b-chat
-
NVIDIA NIM for GPU accelerated Llama 2 13B inference through OpenAI compatible APIs
meta-llama-2-70b-chat
-
NVIDIA NIM for GPU accelerated Llama 2 70B inference through OpenAI compatible APIs
Llama-3-Taiwan-70B-Instruct
-
NVIDIA NIM for GPU accelerated Llama-3-Taiwan-70B-Instruct inference through OpenAI compatible APIs
nemotron-4-340b-instruct
-
NVIDIA NIM for GPU accelerated Nemotron-4-340B-Instruct inference through OpenAI compatible APIs
meta-llama-2-7b-chat
-
NVIDIA NIM for GPU accelerated Llama 2 7B inference through OpenAI compatible APIs
Mistral-7B-Instruct-v0.3
-
NVIDIA NIM for GPU accelerated Mistral-7B-Instruct-v0.3 inference through OpenAI compatible APIs
Nemotron-4-340B-Reward
-
NVIDIA NIM for GPU accelerated Nemotron-4-340B-Reward inference through OpenAI compatible APIs
Llama-3-Swallow-70B-Instruct-v0.1
-
NVIDIA NIM for GPU accelerated Llama-3-Swalow-70B-Instruct-v0.1 inference through OpenAI compatible APIs
Meta/Llama3-8b-instruct
-
NVIDIA NIM for GPU accelerated Llama 3 8B inference through OpenAI compatible APIs
Llama-3.1-8b-instruct
-
NVIDIA NIM for GPU accelerated Llama 3.1 8B inference through OpenAI compatible APIs
Meta/Llama3-70b-instruct
-
NVIDIA NIM for GPU accelerated Llama 3 70B inference through OpenAI compatible APIs
Llama-3.1-405b-instruct
-
NVIDIA NIM for GPU accelerated Llama 3.1 405B inference through OpenAI compatible APIs
Llama-3.1-70b-instruct
-
NVIDIA NIM for GPU accelerated Llama 3.1 70B inference through OpenAI compatible APIs
Mixtral-8x7B-Instruct-v0.1
-
NVIDIA NIM for GPU accelerated Mixtral-8x7B-Instruct-v0.1 inference through OpenAI compatible APIs
Mixtral-8x22B-Instruct-v0.1
-
NVIDIA NIM for GPU accelerated Mixtral-8x22B-Instruct-v0.1 inference through OpenAI compatible APIs
meta-llama-2-13b-chat
-
NVIDIA NIM for GPU accelerated Llama 2 13B inference through OpenAI compatible APIs
meta-llama-2-70b-chat
-
NVIDIA NIM for GPU accelerated Llama 2 70B inference through OpenAI compatible APIs
Llama-3-Taiwan-70B-Instruct
-
NVIDIA NIM for GPU accelerated Llama-3-Taiwan-70B-Instruct inference through OpenAI compatible APIs
nemotron-4-340b-instruct
-
NVIDIA NIM for GPU accelerated Nemotron-4-340B-Instruct inference through OpenAI compatible APIs
meta-llama-2-7b-chat
-
NVIDIA NIM for GPU accelerated Llama 2 7B inference through OpenAI compatible APIs
Mistral-7B-Instruct-v0.3
-
NVIDIA NIM for GPU accelerated Mistral-7B-Instruct-v0.3 inference through OpenAI compatible APIs
Nemotron-4-340B-Reward
-
NVIDIA NIM for GPU accelerated Nemotron-4-340B-Reward inference through OpenAI compatible APIs
NVIDIA Retrieval QA E5 Embedding v5
-
NVIDIA NIM for GPU accelerated NVIDIA Retrieval QA E5 Embedding v5 inference
NVIDIA Retrieval QA Mistral 4B Reranking v3
-
NVIDIA NIM for GPU accelerated NVIDIA Retrieval QA Mistral 4B Reranking v3 inference
NVIDIA Retrieval QA Mistral 7B Embedding v2
-
NVIDIA NIM for GPU accelerated NVIDIA Retrieval QA Mistral 7B Embedding v2 inference
Llama-3-Swallow-70B-Instruct-v0.1
-
NVIDIA NIM for GPU accelerated Llama-3-Swalow-70B-Instruct-v0.1 inference through OpenAI compatible APIs
snowflake Arctic Embed Large Embedding
-
NVIDIA NIM for GPU accelerated Snowflake Arctic Embed Large Embedding inference
MolMIM
-
MolMIM is a transformer-based model developed by NVIDIA for controlled small molecule generation.
DiffDock
-
Diffdock predicts the 3D structure of the interaction between a molecule and a protein.
ProteinMPNN
-
Predicts amino acid sequences from 3D structure of proteins.
AlphaFold2
-
A widely used model for predicting the 3D structures of proteins from their amino acid sequences.
ASR Parakeet CTC Riva 1.1b
-
RIVA ASR NIM delivers accurate English speech-to-text transcription and enables easy-to-use optimized ASR inference for large scale deployments.
TTS FastPitch HifiGAN Riva
-
RIVA TTS NIM provide easy access to state-of-the-art text to speech models, capable of synthesizing English speech from text
NMT Megatron Riva 1b
-
Riva NMT NIM provide easy access to state-of-the-art neural machine translation (NMT) models, capable of translating text from one language to another with exceptional accuracy.
Serverless Pricing
Leverage on-demand compute power with pay-as-you-go pricing, perfect for scaling AI applications seamlessly.
Shakti Cloud Serverless 1 x NVIDIA H100 80GB Per Sec
-
Shakti Cloud Serverless Runtime scaling Instance with
-
GPU: 80 GB H100
-
CPU: 14 Cores
-
RAM: 256GB
-
Storage: 500GB
-
Unlimited ingress and egress.
Shakti Cloud Serverless with 1 x 40 GB H100 Per Sec
-
Shakti Cloud Serverless Runtime scaling Instance with
-
GPU: 40 GB H100
-
CPU: 7 Cores
-
RAM: 128GB RAM
-
Storage: 250GB
-
Unlimited ingress and egress.
Shakti Cloud Serverless with 1 x 48 GB L40S Per Sec
-
Shakti Cloud Serverless Runtime scaling Instance with
-
GPU: 48 GB L40S
-
CPU: 16 Cores
-
RAM: 256GB
-
Storage: 500GB
-
Unlimited ingress and egress
Shakti Cloud Serverless with - 1 x 16 GB L40S Per Sec
-
Shakti Cloud Serverless Runtime scaling Instance with
-
GPU: 16 GB L40S
-
CPU: 4 Cores
-
RAM: 64GB
-
Storage: 128GB
-
Unlimited ingress and egress
Shakti Cloud Serverless with 1 x 24 GB L40S Per Sec
-
Shakti Cloud Serverless Runtime scaling Instance with
-
GPU: 24 GB L40S
-
CPU: 8 Cores
-
RAM: 128GB
-
Storage: 258GB
-
Unlimited ingress and egress
Shakti Cloud Serverless 1 x NVIDIA H100 80GB Per Sec
-
Shakti Cloud Serverless Runtime scaling Instance with
-
GPU: 80 GB H100
-
CPU: 14 Cores
-
RAM: 256GB
-
Storage: 500GB
-
Unlimited ingress and egress.
Shakti Cloud Serverless with 1 x 40 GB H100 Per Sec
-
Shakti Cloud Serverless Runtime scaling Instance with
-
GPU: 40 GB H100
-
CPU: 7 Cores
-
RAM: 128GB RAM
-
Storage: 250GB
-
Unlimited ingress and egress.
Shakti Cloud Serverless with 1 x 48 GB L40S Per Sec
-
Shakti Cloud Serverless Runtime scaling Instance with
-
GPU: 48 GB L40S
-
CPU: 16 Cores
-
RAM: 256GB
-
Storage: 500GB
-
Unlimited ingress and egress
Shakti Cloud Serverless with - 1 x 16 GB L40S Per Sec
-
Shakti Cloud Serverless Runtime scaling Instance with
-
GPU: 16 GB L40S
-
CPU: 4 Cores
-
RAM: 64GB
-
Storage: 128GB
-
Unlimited ingress and egress
Shakti Cloud Serverless with 1 x 24 GB L40S Per Sec
-
Shakti Cloud Serverless Runtime scaling Instance with
-
GPU: 24 GB L40S
-
CPU: 8 Cores
-
RAM: 128GB
-
Storage: 258GB
-
Unlimited ingress and egress
Microsoft Azure AI Services
Experience the power of Azure AI services on Yotta Shakti Cloud with cost-effective pricing designed for optimized performance for your AI workloads.
Product Name |
Service |
Price (Starting from) |
Explore Solutions |
---|---|---|---|
Shakti Cloud Azure ML studio |
Azure ML Studio running on Shakti Cloud with H100 GPUs. |
₹ 264 / GPU / Hr |
Know More |
Shakti Cloud Azure Database Services |
SQL Managed Instance I General Purpose (PaaS) |
₹ 23.54 / Core / Hr |
Know More |
SQL Managed Instance I Business Critical (PaaS) |
₹ 62.48 / Core / Hr |
Know More | |
PostgreSQL (preview) (General Purpose) |
₹ 11.08 / Core / Hr |
Know More | |
PostgreSQL (preview) (Business Critical) |
₹ 12.6 / Core / Hr |
Know More | |
Azure Arc-enabled SQL Server (Standard Edition) |
₹ 8.8 / Core / Hr *(VMs + Licensing cost will be additional) |
Know More | |
Azure Arc-enabled SQL Server (Enterprise Edition) |
₹ 33 / Core / Hr *(VMs + Licensing cost will be additional) |
Know More |
Services |
|
Price (Starting from) |
Explore Solutions |
---|---|---|---|
Azure Open AI |
₹ 13.2 / million inputs token |
Know More | |
Relational Database |
₹ 11 / Core / Hour |
Know More | |
Azure Open Datasets |
₹9.68 / GB |
Know More | |
Internet of Things |
₹ 1232 / Hub Unit (400,000 messages /day of 4 KB size) |
Know More | |
Open-Source Relational Database |
₹ 11 / Core / Hr |
Know More | |
Azure Analytics |
₹ 12.68 / vCore /Hr |
Know More | |
Integration |
₹ 52.8 / Million Operations |
Know More | |
Azure Web Services |
₹ 792 / app /month |
Know More |
Yotta Sarvam AI Services
Harness the capabilities of Yotta Sarvam AI Services with transparent pricing crafted for advanced AI model development, seemless deployment, and scalable performance tailored to your business need.
Name |
|
Price / Per Min |
|
---|---|---|---|
Sarvam Agents |
₹ 7.00 |
Name |
|
Price |
|
---|---|---|---|
Speech- to-text : Saarika |
₹ 30.00 per hour of audio |
||
Text-to-speech: Bulbul |
₹ 15.00 per 10,000 character |
||
ASR-Translate: Saaras |
₹ 30.00 per hour of audio |
||
Mayura |
₹ 20.00 per 10,000 characters |
Name |
|
Price |
|
---|---|---|---|
Parsing |
₹ 10.00 Per Page |
||
Call Analytics |
₹ 2.50 Per Minute |
Name |
|
Price |
|
---|---|---|---|
Legal Research |
₹ 100.00 Per Query |
||
Doc Ops |
₹ 3.00 Per Page |
||
Document Chat |
₹ 10.00 Per Query |
||
Data Analyst |
₹ 100.00 Per Query |
Add On Services
High Speed Storage
-
High Speed Storage for faster read and write data from local and remote
Object Storage
-
Object Storage Buckets to read and transmit data from local and remote
Public IP
-
Public IP for connnectivity
VPN
-
Virtual Private Network for Security
Accelerate AI with Shakti Cloud
