Transparent, Scalable Pricing for your AI Workloads
From development to deployment, our pricing adapts to your needs

-
AI Workspace
Tailored for your AI projects, with flexible Instance.
-
AI Lab
Designed for your educational needs, purchase as per daily lab requirements.
-
Bare Metal
Raw power at a predictable cost—optimized for your high-performance AI workloads.
-
Shakti Clusters
Scale seamlessly with transparent pricing for powerful, AI-optimized clusters.
-
AI Endpoints
Deploy AI models effortlessly with cost-effective, per-usage pricing.
-
Serverless
Maximize efficiency with no upfront costs—only pay for what your AI workloads consume.
-
Add-On Services
AI Workspace Pricing
Shakti Cloud AI Workspace delivers a seamless experience with flexible configurations, offering Virtual Machine to meet diverse AI workload demands.
-
GPU : 1 x Nvidia H100 (80GB)
-
vCPU : 24
-
RAM : 240 GB
-
Block Storage : 2 TB
-
Unlimited Data Ingress & Egress
-
GPU : 1 x Nvidia L40S (48GB)
-
vCPU : 24
-
RAM : 240 GB
-
Block Storage : 2 TB
-
Unlimited Data Ingress & Egress
-
Distributes incoming traffic across targets like web traffic (HTTP/HTTPS) - Layer 7
(Per Gateway per GB)
-
Platform for deploying and managing web apps with infrastructure abstraction.
(Per Instance)
Apps
-
Platform to run containers without managing infrastructure.
(Per vCPU per Memory)
Protection
-
ALWAYS on DDOS Protection against DDoSAttacks -
Upto Clean Traffic per Mbps
(Per Mbps)
-
DNS Service (per 1 million DNS Queries)
(Per Unit)
MySQL Community
-
Open-source relational MySQL database management with continuous
updates and patches with community support to handle workloads
efficiently. (Includes license plus support cost only)
(Per Instance)
MySQL Enterprise
-
Open-source relational MySQL database management service with
continuous updates and patches with enterprise grade support to handle
mission-critical workloads efficiently.
(1 License upto 16vcpu,Includes license plus support cost only)
(Per license)
for PostgreSQL
Community
-
Open-source object-relational database management service with
continuous updates and patches with community support, which is Ideal for
small and medium-scale transactional and analytical workloads,
ensuring fast data processing with minimal overheads. (Includes license plus support cost only)
(Per Instance)
for PostgreSQL
Enterprise
-
Open-source object-relational database management service with
continuous updates and patches with enterprise-grade support, which is
Ideal for large and medium-scale transactional and analytical
workloads, ensuring fast data processing with minimal overheads.
(Per core license cost, Includes license plus support cost only)
(Per license)
Service
-
Managed Kubernetes Service (Container infra needs to be subscribed additional)
(Per Unit)
-
Comprehensive monitoring platform for resources, infrastructure, etc.
(Data Ingested Per GB,
Per API Calls)
and Recovery
-
Back-up as a service
(Per GB)
Bandwidth
-
Internet Bandwidth
(Per Mbps)
Delivery Network
-
Globally distributed network for fast, secure content delivery,
reducing latency and load times for websites, apps, and media.
(Data Transfer per GB,
Requests)
-
Managed registry service for storing container
images used by resources.
(Per Storage GB )
Analytics
-
Platform providing analytics and insights into log and telemetry data
collected from resources
(Data Ingested Per GB)
-
Key Management System as a Service (KMS) for 10 Keys
(Per Unit)
-
Network Load Balancer – Layer 4
(Per Mbps)
-
Collects, analyzes, and visualizes logs and telemetry information.
(Per GB)
-
NAT Gateway
(Per Unit)
-
Enterprise-grade Redis service which comes with enterprise support,
making it perfect for mission-critical distributed data systems.
(Per Licence Unit)
Standard
-
This service offers Microsoft SQL Standard Edition, along with support.
Best for standard SQL workloads with moderate complexity and data
requirements. (Per License, per 2 core,Includes license plus support
(cost only)
(Per Licence Unit)
Enterprise
-
Enterprise-grade Microsoft SQL with Enterprise support and advanced
features for analytics, business intelligence, and large-scale
transactional systems. (Per License, per 2 core, Includes license plus
support cost only)
(Per Licence Unit)
-
Low-cost object storage - 100GB
(Per Unit)
- Windows
-
2 vCPU, 4GB RAM, 100GB Storage,vNIC, Up to 10G Network Performance
(Per Instance)
Linux (ubuntu)
-
2 vCPU, 4GB RAM, 100GB Storage,vNIC, Up to 10G Network Performance
(Per Instance)
-
Virtual Network (Vnet / VPC)
(Per Unit)
-
Cloud-based solution for automating workflows, infrastructure
management, and repetitive tasks across environments.
(Per Job Execution)
-
Managed service that enables secure RDP/SSH access to Azure VMs via a
web browser, eliminating public IP exposure.
(Per Instance)
Factory
-
Fully managed ETL (Extract, Transform, Load) service for integrating
data from multiple sources into cloud storage or databases.
(Pipeline Activity Run,
Data Movement Per GB)
-
Scalable data engineering and analytics service built for real-time
processing and collaboration.
(Per DBU -
Databricks Unit)
Service
-
Global application delivery network providing traffic acceleration,
security, and load balancing for web applications.
(Outbound Data
Transfer Per GB)
-
A set of pre-built AI models that enable applications to perform
intelligent tasks without deep AI expertise.
(Per Transactions,
API Calls)
-
Serverless container service that allows businesses to deploy
applications quickly without provisioning or maintaining VMs.
(Per vCPU Per
GB Memory)
-
Scalable email delivery service for transactional and marketing emails
with built-in security and compliance features.
(Per Mail)
-
Managed event ingestion service designed to handle 1 millions of events
per second for real-time data streaming.
(Per Unit)
-
Cloud-based workflow automation service that connects multiple
applications and services for streamlined business processes.
(Per Task)
for Cloud
-
Security posture management and threat protection service that helps
detect vulnerabilities and attacks in cloud environments.
(Per Endpoints)
-
Reliable Kafka cloud messaging service for decoupling applications and
enabling asynchronous communication.
(Per Instance)
AI Lab Pricing
Select the AI Lab that fits your budget and training needs — opt for unlimited hours or flexible time slots
Instant AI Lab, zero hassle. Just subscribe to get GPU access and the AI Lab platform—with unlimited user onboarding. Bring your own storage or add it separately. Comes with pre-configured ML/DL containers, so your students can start learning from day one.
Price per workstation |
Description |
GPU Memory |
Monthly price |
---|---|---|---|
AI Lab Workstation - |
|
8GB L40s |
₹ 30,000 |
AI Lab Workstation - |
|
16 GB L40s |
₹ 45,000 |
AI Lab Workstation - |
|
40 GB H100 |
₹ 90,000 |
AI Lab Workstation - |
|
80 GB H100 |
₹ 1,40,000 |
AI Lab Workstation - |
|
160 GB H100 |
₹ 2,80,000 |
AI Lab Workstation - |
|
320 GB H100 |
₹ 5,60,000 |
Total control, ultimate flexibility. Choose your GPUs, scale users as needed and run the AI Lab platform your way. Create multiple GPU profiles—let students use GPUs on weekdays and researchers take over on weekends. Includes storage options and pre-configured ML/DL containers for a seamless setup.
Workstation |
Description |
Price per Month |
---|---|---|
GPU |
Select the GPU capacity that fits your AI Lab needs |
|
AI Lab |
|
₹ 70,000 |
Bare Metal Pricing
Our budget-friendly plans are designed to provide the best value, tailored to meet your training needs
Plan |
monthly contract |
6 months contract |
12 months contract |
24 months contract |
36 months contract |
48 months contract |
---|---|---|---|---|---|---|
Baremetal 8 x H100 HGX
|
325 |
313 |
243 |
234 |
226 |
217 |
Plan |
monthly contract |
6 months contract |
12 months contract |
24 months contract |
36 months contract |
48 months contract |
---|---|---|---|---|---|---|
Bare Metal 4 x L40S
|
182 |
161 |
147 |
145 |
143 |
141 |
Shakti GPU Clusters Pricing
Our budget-friendly plans are designed to provide the best value, tailored to meet your training needs
Plan |
monthly contract |
6 months contract |
12 months contract |
24 months contract |
36 months contract |
48 months contract |
---|---|---|---|---|---|---|
SHAKTI CLOUD - H100 HGX Cluster
|
357 |
345 |
267 |
257 |
249 |
239 |
Plan |
monthly contract |
6 months contract |
12 months contract |
24 months contract |
36 months contract |
48 months contract |
---|---|---|---|---|---|---|
SHAKTI CLOUD - L40S Cluster
|
200 |
177 |
162 |
160 |
158 |
156 |
AI Endpoints Pricing
Select the right endpoint for your project and enjoy transparent, token-based pricing designed to meet your unique AI needs.
Model Name |
Description |
Use cases |
Pricing |
---|---|---|---|
Meta/Llama3-8b-instruct |
NVIDIA NIM for GPU accelerated Llama 3 8B inference through OpenAI compatible APIs |
|
20 / 1M Tokens |
Llama-3.1-8b-instruct |
NVIDIA NIM for GPU accelerated Llama 3.1 8B inference through OpenAI compatible APIs |
|
20 / 1M Tokens |
Meta/Llama3-70b-instruct |
NVIDIA NIM for GPU accelerated Llama 3 70B inference through OpenAI compatible APIs |
|
140 / 1M Tokens |
Llama-3.1-405b-instruct |
NVIDIA NIM for GPU accelerated Llama 3.1 405B inference through OpenAI compatible APIs |
|
705 / 1M Tokens |
Llama-3.1-70b-instruct |
NVIDIA NIM for GPU accelerated Llama 3.1 70B inference through OpenAI compatible APIs |
|
140 / 1M Tokens |
Mixtral-8x7B-Instruct-v0.1 |
NVIDIA NIM for GPU accelerated Mixtral-8x7B-Instruct-v0.1 inference through OpenAI compatible APIs |
|
90 / 1M Tokens |
Llama-3.1-8b-base |
NVIDIA NIM for GPU accelerated Llama 3.1 8B inference through OpenAI compatible APIs |
|
20 / 1M Tokens |
Mixtral-8x22B-Instruct-v0.1 |
NVIDIA NIM for GPU accelerated Mixtral-8x22B-Instruct-v0.1 inference through OpenAI compatible APIs |
|
140 / 1M Tokens |
MolMIM |
MolMIM is a transformer-based model developed by NVIDIA for controlled small molecule generation. |
|
350 / Per Request |
meta-llama-2-13b-chat |
NVIDIA NIM for GPU accelerated Llama 2 13B inference through OpenAI compatible APIs |
|
65 / 1M Tokens |
meta-llama-2-70b-chat |
NVIDIA NIM for GPU accelerated Llama 2 70B inference through OpenAI compatible APIs |
|
140 / 1M Tokens |
DiffDock |
Diffdock predicts the 3D structure of the interaction between a molecule and a protein. |
|
440 / Per Request |
Llama-3-Taiwan-70B-Instruct |
NVIDIA NIM for GPU accelerated Llama-3-Taiwan-70B-Instruct inference through OpenAI compatible APIs |
|
140 / 1M Tokens |
ASR Parakeet CTC Riva 1.1b |
RIVA ASR NIM delivers accurate English speech-to-text transcription and enables easy-to-use optimized ASR inference for large scale deployments. |
|
175 / Per Request |
ProteinMPNN |
Predicts amino acid sequences from 3D structure of proteins. |
|
350 / Per Request |
nemotron-4-340b-instruct |
NVIDIA NIM for GPU accelerated Nemotron-4-340B-Instruct inference through OpenAI compatible APIs |
|
705 / 1M Tokens |
AlphaFold2 |
A widely used model for predicting the 3D structures of proteins from their amino acid sequences. |
|
175 / Per Request |
TTS FastPitch HifiGAN Riva |
RIVA TTS NIM provide easy access to state-of-the-art text to speech models, capable of synthesizing English speech from text |
|
90 / Per Request |
NMT Megatron Riva 1b |
Riva NMT NIM provide easy access to state-of-the-art neural machine translation (NMT) models, capable of translating text from one language to another with exceptional accuracy. |
|
45 / 1M Tokens |
meta-llama-2-7b-chat |
NVIDIA NIM for GPU accelerated Llama 2 7B inference through OpenAI compatible APIs |
|
20 / 1M Tokens |
Mistral-7B-Instruct-v0.3 |
NVIDIA NIM for GPU accelerated Mistral-7B-Instruct-v0.3 inference through OpenAI compatible APIs |
|
20 / 1M Tokens |
Nemotron-4-340B-Reward |
NVIDIA NIM for GPU accelerated Nemotron-4-340B-Reward inference through OpenAI compatible APIs |
|
705 / 1M Tokens |
NVIDIA Retrieval QA E5 Embedding v5 |
NVIDIA NIM for GPU accelerated NVIDIA Retrieval QA E5 Embedding v5 inference |
|
175 / 1M Tokens |
NVIDIA Retrieval QA Mistral 4B Reranking v3 |
NVIDIA NIM for GPU accelerated NVIDIA Retrieval QA Mistral 4B Reranking v3 inference |
|
45 / 1M Tokens |
NVIDIA Retrieval QA Mistral 7B Embedding v2 |
NVIDIA NIM for GPU accelerated NVIDIA Retrieval QA Mistral 7B Embedding v2 inference |
|
65 / 1M Tokens |
Llama-3-Swallow-70B-Instruct-v0.1 |
NVIDIA NIM for GPU accelerated Llama-3-Swalow-70B-Instruct-v0.1 inference through OpenAI compatible APIs |
|
140 / 1M Tokens |
snowflake Arctic Embed Large Embedding |
NVIDIA NIM for GPU accelerated Snowflake Arctic Embed Large Embedding inference |
|
45 / 1M Tokens |
Serverless Pricing
Leverage on-demand compute power with pay-as-you-go pricing, perfect for scaling AI applications seamlessly.
-
GPU: 80 GB H100
-
CPU: 14 Cores
-
RAM: 256GB
-
Storage: 500GB
-
Unlimited ingress and egress
-
GPU: 40 GB H100
-
CPU: 7 Cores
-
RAM: 128GB RAM
-
Storage: 250GB
-
Unlimited ingress and egress
-
GPU: 48 GB L40S
-
CPU: 16 Cores
-
RAM: 256GB
-
Storage: 500GB
-
Unlimited ingress and egress
-
GPU: 16 GB L40S
-
CPU: 4 Cores
-
RAM: 64GB
-
Storage: 128GB
-
Unlimited ingress and egress
-
GPU: 24 GB L40S
-
CPU: 8 Cores
-
RAM: 128GB
-
Storage: 258GB
-
Unlimited ingress and egress
-
GPU: 80 GB H100
-
CPU: 14 Cores
-
RAM: 256GB
-
Storage: 500GB
-
Unlimited ingress and egress
-
GPU: 40 GB H100
-
CPU: 7 Cores
-
RAM: 128GB
-
Storage: 250GB
-
Unlimited ingress and egress
-
GPU: 48 GB L40S
-
CPU: 16 Cores
-
RAM: 256GB
-
Storage: 500GB
-
Unlimited ingress and egress
-
GPU: 16 GB L40S
-
CPU: 4 Cores
-
RAM: 64GB
-
Storage: 128GB
-
Unlimited ingress and egress
-
GPU: 24 GB L40S
-
CPU: 8 Cores
-
RAM: 128GB
-
Storage: 258GB
-
Unlimited ingress and egress
Add-On Services
Storage High Speed
Storage for faster read and write data from local and remote ₹ 5.80 Per GB
Storage Object Storage Buckets to read and transmit data from local and remote ₹ 2.50 Per GB
IP Public IP for connectivity ₹ 528 Per IP
MRC ₹ 660
Accelerate AI with
Shakti Cloud
Get Started
