Three-founder team's common thread: managed Postgres-as-a-service across Heroku, Citus, Azure, and Crunchy Bridge. Ubicloud is Daniel Farina's 4th managed-cloud control plane.
Where they came from
Founder
Role
Prior
Umur Cubukcu
Co-founder, Co-CEO
Citus Data co-founder/CEO (YC S11), 4y Azure Postgres lead, YC Visiting Partner 2023
Ozgun Erdogan
Co-founder, Co-CEO / CTO
Citus Data co-founder/CTO, Amazon distributed systems, 4y Azure engineering lead
Ubicloud as the 4th iteration of a 13-year Postgres-as-a-service arc. Ruby chosen for infra orchestration because REPL + mature libraries = productivity advantage for a small team.
Representative prices (Germany region, 2026) — Ubicloud vs hyperscaler on-demand:
Service
Ubicloud
Hyperscaler
Savings
VM: 2 vCPU / 8 GB
~$26 / mo
AWS ~$69, Azure ~$65, GCP ~$62
~60–65%
VM: 32 vCPU / 128 GB
linear scaling
AWS ~$1,104 / mo
~60–65%
Burstable 1 vCPU
$6.65 / mo
—
—
Managed Postgres Hobby
$12.41 / mo
—
—
Managed Postgres Standard (2 vCPU)
$49 / mo
AWS RDS ~$200 / mo
~67%
Managed Kubernetes (dev)
$46 / mo
EKS control + EC2 variable
~73%
GitHub Actions 2 vCPU Linux
$0.0008 / min
GitHub $0.0080 / min
10x (90%)
AI inference (Qwen2.5-VL-72B)
$0.80 / M tokens (in+out)
—
—
AI inference (Qwen3-Embedding-8B)
$0.05 / M input tokens
—
—
Public IPv4: $3/mo. Egress: free up to ~0.625 TB per 2 vCPUs, then $3/TB (≈30x cheaper than hyperscaler egress). Free tier on inference: 500k tokens/month. Per-token pricing for most chat models (Llama 3.3, Mistral Small 3, DeepSeek V3/R1) is dashboard-only.
Performance Claims (Postgres)
Self-published TPC-C benchmarks — Ubicloud vs AWS (no independent third-party verification found):
TPC-C (transactional): 1.4x more TPS than Aurora at 5.8x lower cost; 4.6x more TPS than RDS at 2.8x lower cost
Latency: 1.91x lower than Aurora, 7.65x lower than RDS
TPC-H (analytical): 2.42x faster than Aurora; 2.96x faster than RDS
Headline:"9x price/performance" vs RDS/Aurora
Driver: SPDK + local NVMe + Cloud Hypervisor = less I/O overhead per dollar
Caveat: all numbers sourced from ubicloud.com — no external benchmark surfaced
Competitive Positioning
Vs hyperscalers (AWS / GCP / Azure)
3x–10x cheaper, open source, portable
Opinionated and narrow — targets the 10% of services that drive 80% of spend; explicitly no Lambda/DynamoDB/SageMaker equivalents
Vs open-source cloud (OpenStack etc.)
Offers a first-party managed service
Opinionated stack vs pluggable-everything
Modern components (Cloud Hypervisor, SPDK) post-dating OpenStack's design era
Cubukcu: "OpenStack takes an army of people"
Vs bare-metal VPS (Hetzner, DO, Linode, Vultr, Scaleway, OVH)
Adds managed PaaS layer (Postgres, K8s, runners, inference) they lack
CoreWeave, Lambda — Ubicloud exited this race (GPU rental deprecated Dec 2025); pivot to inference-as-PaaS
Key Customers & Partnerships
Anchor partnerClickHouse native managed Postgres, powered by Ubicloud · Jan 22, 2026 · private preview
ClickHouse (Jan 22, 2026) — strategic wedge. ClickHouse launched its own native managed Postgres service in private preview, powered entirely by Ubicloud. Coincided with ClickHouse's $400M Series D (Dragoneer-led). ClickHouse engineers now contribute upstream. Shifts Ubicloud toward B2B2B infrastructure play.
Direct customers with public stories: Felt, Hatchet (formal case studies); Resmo, Windmill, PeerDB (homepage logos)
AudienceKey — cited by third-party research as achieving 50% DB cost reduction post-migration (not independently verified on Ubicloud's site)
Claimed scale: ~400 paying customers per a Reddit-sourced figure — unverified
No public Turkish enterprise, government, or bank customers announced
Office & Data Center Footprint
Offices
Office
Address
San Francisco (HQ)
450 Townsend St., SF, CA 94107
Amsterdam / Amstelveen
Turfschip 267, 1186XK, Amstelveen NL
Istanbul
Esentepe Mah. Talatpaşa Cad. No:5/1, Levent
Production data center regions
🇩🇪
Falkenstein, DE
eu-central-h1 · Hetzner
🇫🇮
Helsinki, FI
eu-north-h1 · Hetzner
🇺🇸
Manassas, VA
us-east-a2 · Leaseweb
🇹🇷
Istanbul, TR
Private · B200 · on request
Marketing materials reference future regions (Frankfurt, Oregon, Singapore, São Paulo) and additional bare-metal partners (OVHcloud, Latitude.sh, AWS Bare Metal). No broader MENA or APAC presence. Ubicloud owns no physical hardware.
Recent Developments (2025)
ARM64 VMs and ARM GitHub Actions runners GA; "100x price/performance" on certain ARM CI workloads
Managed Kubernetes moved to public beta (Germany + Virginia); UbiCSI local-NVMe PV driver in preview
Postgres dashboard overhaul (June 2025)
AI Inference Endpoints — OpenAI-compatible API on vLLM with open-weight models, managed multi-GPU
SOC 2 Type II certified (Feb 2025 changelog)
Deprecated raw GPU VM runners (effective Dec 31, 2025) — strategic exit from CapEx-heavy GPU race
B200 HGX GPU launched in Türkiye (Istanbul) Private Location (Oct 2025); 4- and 8-GPU partitions added Nov 2025
B200 HGX GPU virtualization (Dec 15, 2025) — deep technical post on QEMU 10.1+, VFIO-PCI, NVIDIA Fabric Manager, Shared NVSwitch Multitenancy; HN front page
Recent Developments (2026)
ClickHouse partnership (Jan 22, 2026) — ClickHouse native Postgres powered by Ubicloud; private preview; engineering cross-contributions; tied to ClickHouse's $400M Series D
Blog output — LLM coding practices, VLM-based OCR, documentation automation, CPU-performance myths ("Does MHz still matter?"), AI Coding sober review
EuroGPT Enterprise continuing to scale (launched Nov 2024) — privacy-first ChatGPT Enterprise alternative, €19/user/mo, Llama 3.1 405B hosted in Germany
No new funding round publicly disclosed — most recent remains the Mar 2024 seed
Execution Cadence · 2023 → 2026
Company, fundraising, product, and partnership milestones on one axis.
Cadence signals
Monthly public changelog with named feature-level shipments
~11 model additions to inference endpoints across 2025
3 generations of GPU hardware productized (A100 → H100 → HGX B200) in under 12 months
Strategic inflection points
EuroGPT (Nov 2024) — moved up-stack from IaaS to SaaS
B200 in Türkiye (Oct 2025) — first non-EU/US region, founder-network origination
ClickHouse (Jan 2026) — B2B2B pivot validated
EU/EMEA Regulatory Posture — the credible parts
Dual-entity controller structure: Ubicloud B.V. (NL) and Ubicloud Inc. (US) — Schrems-II-aware
EEA-only storage of Customer Account Data (personal data of customers themselves)
Transfer basis: Article 45(1) adequacy + Article 46(2)(c) Standard Contractual Clauses
SOC 2 Type II confirmed (Feb 2025 changelog; dedicated /docs/security/soc2 URL currently 404s)
Matomo for analytics (not Google Analytics) — GDPR-friendlier choice
Penetration test referenced, available on request
Proactive engagement on EU Data Act — Nov 2023 blog post welcoming cloud-switching/portability provisions is their most substantive regulatory communication
EuroGPT residency guarantee: all GPU processing stays in Germany; no customer data used for training
EU/EMEA Regulatory Posture — the gaps
Silent or not-yet-claimed despite their EU sovereignty pitch:
No ISO 27001 / 27017 / 27018
No C5 (German BSI — often required for Bundesverwaltung procurement, conspicuous given the German region)
No SecNumCloud (France / ANSSI)
No ENS (Spain)
No EUCS claim, no Gaia-X participation
No public DORA posture — notable given ClickHouse partnership targets financial services; DORA in force since Jan 17, 2025
No public NIS2 posture — Ubicloud's IaaS would normally be in scope
No public EU AI Act role classification — despite operating EuroGPT and inference APIs
No published BAA process for HIPAA — ToS prohibits PHI absent separate written agreement
No public SLA posted
Short version: GDPR/SOC 2 baseline is credible; certification stack is light relative to the "sovereign, open, portable" pitch.
Contract Gotchas (Terms of Service)
Governing law: California
Data residency not contractually guaranteed by default — ToS permits Ubicloud to move Services Content between regions at its sole discretion absent a written addendum (EuroGPT is a named exception)
No SLA in the ToS — no uptime commitment, no service-credit regime
Backups are the customer's responsibility — "Ubicloud does not promise to retain any preservations or backups"
Termination at sole discretion, with or without notice; may result in immediate data destruction
PHI and GDPR Article 9 special-category data prohibited without separate written agreement
Inference endpoints, model catalog, vLLM internals, EuroGPT, B200
Briefing · April 2026
Ubicloud AI
Open-source inference endpoints, EuroGPT Enterprise, and B200 virtualization
TL;DR — AI Strategy
Pivoted from raw GPU rentals to managed inference PaaS — GPU GitHub Actions runners deprecated Dec 31, 2025; GPU VMs repositioned as private/enterprise-only
Open-weight only — no Claude/GPT/Gemini re-hosting; every model on the platform is open-weight
Production runtime: vLLM V1 with FlashAttention-3, FlashInfer, speculative decoding, prefix caching
Signature technical work: open-source virtualization of NVIDIA HGX B200 using QEMU 10.1+ + Fabric Manager Shared NVSwitch Multitenancy
AI footprint: Germany (Falkenstein, Helsinki, EuroGPT processing) + Türkiye Istanbul Private Location for B200
Product Surface
Two API surfaces:
Surface
Base URL
Purpose
Auth
Management
https://api.ubicloud.com
Manage API keys, endpoints, projects
Bearer JWT
Inference data plane
https://{model}.ai.ubicloud.com/v1
OpenAI-compatible inference
Bearer API key
Per-model subdomain pattern — each model gets its own hostname (e.g. llama-3-3-70b-turbo.ai.ubicloud.com/v1). There is no unified inference host.
SDK support: any OpenAI-compatible SDK (Python openai, JS); first-party Ruby SDK + ubi CLI (beta).
Free tier: 500,000 tokens / month.
OpenAI Compatibility
Documented and working against the per-model base URL:
POST /v1/chat/completions — non-streaming
POST /v1/chat/completions with stream=True — SSE streaming
POST /v1/chat/completions with response_format={"type":"json_object"} — JSON mode
POST /v1/chat/completions with tools=[...], tool_choice="auto" — function/tool calling
/v1/embeddings — implied by Qwen3-Embedding-8B launch (endpoint path not explicitly documented)
Not documented or not offered:/v1/completions (legacy), /v1/models on data plane, audio, image, batch API, fine-tuning API.
Model Catalog (Confirmed Public)
Model ID
Family
Role
First seen
llama-3-3-70b-turbo
Llama 3.3 70B
Chat
Feb 2025
mistral-small-3
Mistral Small 3 (24B)
Chat
Feb 2025
ds-r1-qwen-32b
DeepSeek-R1-Distill-Qwen-32B
Reasoning
Feb–Mar 2025
DeepSeek V3
DeepSeek V3
Chat
Jun 2025
DeepSeek R1
DeepSeek R1
Reasoning
Jun 2025
Qwen2.5-VL-72B
Qwen 2.5 VL
Vision-language
Jul 2025
Qwen3 VL
Qwen 3 VL
Vision-language
Oct 2025
Qwen3-Embedding-8B
Qwen 3 Embedding
Text embeddings
Mar 2026
Llama Guard 3
Meta
Moderation (EuroGPT)
Nov 2024
Llama 3.1 405B
Meta
Chat (EuroGPT)
Nov 2024
Open-weight only. No Llama 4 in public materials. Context windows and quantization not published per-model.
Public Pricing
Per-token pricing is dashboard-only for most chat models. Only two models are publicly priced on web:
Model
Price
Notes
Qwen2.5-VL-72B
$0.80 / M tokens (input + output)
Jul 2025
Qwen3-Embedding-8B
$0.05 / M input tokens
Mar 2026
Free tier
500k tokens / month
Feb 2025
March 2026 addition: new GET /project/{id}/inference-endpoint API returns full price table programmatically with separate per_million_prompt_tokens and per_million_completion_tokens.
Positioning claims (Ubicloud-authored): "3–10x lower than comparable offerings" for cloud overall; "3x lower than US alternatives" for EuroGPT. "10x cheaper than OpenAI" is NOT a Ubicloud claim — that phrasing came from third-party research.
Hardware Stack
GPU siliconA100 · H100 · HGX B200 · RTX PRO 6000
GPU
Status
First public mention
NVIDIA A100
Preview (Germany)
May 2025
NVIDIA H100
Production (prior GPU VMs)
—
NVIDIA HGX B200
Production (Türkiye Istanbul, on request)
Oct 2025
NVIDIA RTX PRO 6000
On request
Dec 2025
Not offered in public materials: H200, L40S, MI300X.
B200 partitioning via Shared NVSwitch Multitenancy
Partition size
When added
1-GPU, 2-GPU
Oct 2025 launch
4-GPU, 8-GPU
Nov 2025
Inside a partition: full NVLink/NVSwitch bandwidth. Across partitions: isolated. Fabric Manager enforces routing.
B200 Virtualization — Signature Tech Work
Ubicloud wrote the "missing manual" on open-source virtualization of NVIDIA HGX B200. Single 8-GPU node partitioned via Shared NVSwitch Multitenancy:
QEMU 10.1+ (not Cloud Hypervisor) — B200 needs multi-level PCIe topology that Cloud Hypervisor's flat topology can't produce; 10.1 added BAR-mapping optimizations critical for B200's 256 GB Region 2 BAR per GPU
Embeddings: Mistral E5 7B for RAG with private knowledge base
Web search: DuckDuckGo (privacy-preserving)
Data residency:"Data remains in Germany, including all GPU processing"
Training:"No customer data or metadata used for training purposes"
Security: encryption in transit + envelope encryption at rest, key rotation, file upload
SSO: OIDC at platform level (Jul 2025); EuroGPT-specific SSO not explicitly documented
Not disclosed: quantization of the 405B deployment. Not offered: private API for EuroGPT — raw API consumers use Inference Endpoints directly.
Strategic Pivot: GPU Rentals → Inference PaaS
Before (2024)
Offered raw GPU rentals (RTX 4000 Ada / H100) as GitHub Actions runners and GPU VMs.
Inflection (2025)
Recognized the CapEx-heavy raw-GPU race against CoreWeave, Lambda, AWS P5, Azure NDv5 as structurally unviable for a seed-stage company. Moved up-stack to managed inference PaaS + dedicated enterprise GPU (private locations).
After (Dec 31, 2025)
GPU GitHub Actions runners deprecated
GPU VMs repositioned as private/enterprise deployments (B200, RTX PRO 6000 on request)
Open-weight inference endpoints become the primary AI front door
EuroGPT Enterprise becomes the productized SaaS face
Implication: Ubicloud is no longer competing on GPU-hours — they're competing on tokens and on the quality of the managed inference stack.
Positioning
OpenAI
Anthropic
Gemini
Meta Llama
CoreWeave / Lambda
Together / Fireworks
Competitor class
Examples
Ubicloud's angle
Closed-model LLM vendors
OpenAI, Anthropic
Open-weight only; lower price; EU residency; no training use
Fast-inference specialists
Groq, Together, Fireworks, DeepInfra
Same model class; adds full IaaS underneath + EuroGPT SaaS on top
GPU clouds
CoreWeave, Lambda, AWS P5
Open-source B200 virtualization; control plane on GitHub; BYOC option
GPU-on-demand
RunPod, Vast.ai
Managed-first; GDPR-native; EuroGPT SaaS
European sovereign AI
Mistral-La Plateforme, Aleph Alpha
Broader IaaS (compute + K8s + Postgres) beyond just models