Skip to main content
Screenshot2026 01 23at3 36 57PM

Infrastructure Requirements

Service CategoryService TypeDescriptionQuantity
ComputeKubernetes Service - System Node Pool (Masters)4 vCPUs, 16 GB RAM, Linux3
ComputeKubernetes Service User Node Pool (workers)16 vCPUs, 128 GB RAM, Linux4 - 6
StoragePersistent StoreProvisioned SSD (Premium), 3,500 IOPS, 150 MiB/sec throughput1 TB
GPU (Hosted LLM)H100 / H200 family160 – 320 GB vRAM as per the concurrency requirements. 256GB RAM. 8 Core. 1 TB Storage3 - 4
External LLMsAPI / URLNo need of hardwareMin - 2
*** Note: **Above infra is for ~500 concurrent user with external LLM via API (e.g. ChatGPT/Claude).
For Hosted GPU concurrency is ~50-60
*** Note: **Additional components (e.g., firewall, load balancer) required per infrastructure policies.* Note: Capacity is subject to increase based on the usage load

Responsibility Matrix

#TaskTurintonClient / Partner
1Platform deploymentResponsible
2Platform Training SupportResponsible
3Platform UpdatesResponsible
4Building POC Use CaseResponsibleSupport
5Ontology / Context Graph developmentResponsibleResponsible
6Logging & Observability SetupResponsibleResponsible
7New LLM model deployment (using vLLM)ResponsibleResponsible
8Building & Distributing New Use Case (After POC)SupportResponsible
9Infrastructure ProvisioningSupportResponsible
10Hardware RequirementsResponsible
11Data AccessResponsible
12Monitor InfrastructureResponsible
13Security & Compliance ManagementResponsible
14Disaster Recovery / Backup StrategyResponsible