One bucket where every object is searchable by meaning, transformable by any AI pipeline, and analyzable across millions of siblings.
The gap isn’t model capability. It’s the data plumbing underneath every agent.
McKinsey puts AI’s annual productivity opportunity at $4.4T. AI infra spend is the fastest-growing segment in cloud — and the wedge for an Augmented S3 that does RAG, transform, and analyze natively is exactly where K3 lives.
Gartner forecasts total worldwide AI spending at $2.5T in 2026, with $1.37T concentrated in AI infrastructure — compute, storage, networking, and platform services purpose-built for AI workloads.
K3's wedge sits at the intersection of vector databases ($3.2B in 2026) and RAG infrastructure ($3.3B in 2026) — the two segments K3 collapses into one S3-compatible bucket. Combined growth rate: 25%+ CAGR through 2030.
Bottom-up: 5 paying customers + 5 named enterprise POCs (Crédit Agricole, PepsiCo, NBE, Rubix) + 40+ early-access pipeline, against a $13.5B MEA sovereign cloud market and €100B European sovereign cloud opportunity by 2031.
Storage, compute, orchestration, and retrieval — designed to compose into one bucket. Built on the same control plane, the same auth, and the same multi-tenant isolation.
An S3-compatible bucket that searches, transforms, and analyzes every object — RAG, AI pipelines on every upload, and cross-bucket analysis, all from one bucket.
A FaaS runtime designed so AI agents can author, test, and call their own tools. Python and Rust, MCP-enabled, ~200ms warm cold starts, massively parallel by default.
A purpose-built DSL for LLM-driven workflows — 15 composable primitives, full type checking, resumable threads, first-class human-in-the-loop. The language K3 transforms run on.
Drop-in pymilvus compatibility, instant serverless pools, automated backups, multi-tenant isolation. The vector index every K3 retrieval call hits.
K3 is Augmented S3 — vector, transform, analyze, and an embedded warehouse pipelines offload into (HTAP soon). Open it in a spreadsheet or ask the built-in support agent.
Hybrid semantic search across your bucket — dense vectors, BM25, multimodal queries (text, image, audio, video), and reranking. The full RAG stack, just sitting there waiting for a query.
Trigger a Scriptum pipeline on every object that matches your rules — summarize, redact, caption, transcribe, classify. Outputs land back in the same bucket under a derived prefix like `summaries/`.
Run Scriptum pipelines across many objects at once — outputs offload into K3's embedded warehouse (HTAP soon), queryable in SQL or the Excel interface alongside the bucket itself.
Six industries. Same bucket. Same fifteen pipelines, all shipping day one.
Hybrid search catches semantic clause variants keyword review misses entirely.
Every answer carries a citation chain: policy → regulation → guidance.
Auto-PII redaction and fraud signal scoring run inline at ingest, not weeks later.
Defect patterns auto-cluster across years of FMEA reports — root causes that no single engineer would have spotted.
Multi-tenant + sovereign by default — your enterprise customers get data residency for free.
Foreign-language source material auto-translated and indexed alongside English; multi-classification handling out of one bucket.
The same control plane runs three more products that compose with it — compute for the agents that build on it, the language they orchestrate in, and the vector index every retrieval call hits.
A FaaS runtime designed so AI agents can author, test, and call their own tools — Python and Rust, ~200ms warm cold starts, MCP-enabled, massively parallel by default.
A purpose-built DSL with 15 composable primitives, full compile-time type checking, resumable threads, and first-class human-in-the-loop. The language K3 uses for every Transform and Analyze pipeline.
Drop-in pymilvus compatibility, instant serverless pools, automated backups, multi-tenant isolation. Powered by Milvus 2.6 and exposed inside every K3 bucket as the underlying vector index.
The capabilities investors care about, scored across the patterns customers actually compare us to.
| Capability | Dodil K3 + Ignite + Scriptum + VBase | AWS S3 + Pinecone + Lambda Hyperscaler patchwork | MongoDB Atlas Vector Managed DB vendor | Vercel + LangChain + Supabase AI app stack | Self-built (S3 + Milvus + glue) DIY |
|---|---|---|---|---|---|
S3-compatible storage Drop-in for any existing S3 SDK | |||||
Hybrid semantic search Dense + BM25 + reranking, in the bucket | |||||
AI transform on every upload Summarize, redact, caption, classify, extract | |||||
Cross-bucket analyze pipelines Anomaly detection, batch scoring, datasets | |||||
Workflow language with type checking Compile-time safety for LLM workflows | |||||
Multi-tenant by design Org isolation across storage, search, processing, credentials | |||||
Sovereign / on-prem deploy Run the full stack in your own datacenter |
Our Cloud
Your data plane runs in your datacenter. Our control plane is in London — expanding to the Middle East and Asia.
K3 is the same Augmented S3 either way. Drop the data plane into your own datacenter, keep every object in-country, and we manage the upgrades, observability, and pipelines. The control plane sits in London today; the data plane sits wherever your compliance officer wants it.
Traction
Five paying customers. Forty-plus on the early-access list. Five enterprise POCs across four countries — all on the sovereign K3 deployment path.
Compliance research + auditor-grade citation chains across policies, regulations, and guidance — running on K3 in their own jurisdiction.
Defect intelligence and supply-chain knowledge retrieval across decades of QA reports — local data plane, no upstream egress.
AI-powered customer service and fraud signal scoring — sovereign deployment for the National Bank of Egypt.
Multi-tenant K3 deployments across Rubix's regional client base — one platform, many sovereign data planes.
Investment Opportunity
We have more demand than we can fulfill. 5 Enterprise POCs, 40+ early access signups, and a growing waitlist. This round is about scaling infrastructure to capture the opportunity — not finding product-market fit.
Building Augmented S3 for the next decade of AI.