Nebius Group (NASDAQ: NBIS) is the fastest-growing AI cloud, having transformed from the remnants of Yandex N.V. into a $20B+ market cap AI infrastructure company in under 18 months.[5] The company posted $529.8M in FY2025 revenue[1], signed $20.4B in hyperscaler contracts (Microsoft $17.4B[2] + Meta $3B[3]), and guides $3.0–$3.4B revenue for 2026 with ~40% adjusted EBITDA margins.[6]
| Company | Focus | Status |
|---|---|---|
| Nebius AI Cloud | Full-stack AI infrastructure (core business) | Core Revenue Driver |
| Avride | Autonomous vehicles & delivery robots | R&D Stage |
| TripleTen | EdTech for tech career training | Revenue Generating |
| Toloka | AI data labeling & model development | Partner Ecosystem |
| ClickHouse | Open-source analytics database | Minority Stake |
| Metric | FY2024 | Q1 2025 | Q2 2025 | Q3 2025 | Q4 2025 | FY2025 |
|---|---|---|---|---|---|---|
| Group Revenue | $91.5M[1] | ~$50M | ~$106M | $146M[17] | $228M[1] | $529.8M[1] |
| YoY Growth | — | — | — | 355%[17] | 547%[1] | 478%[1] |
| QoQ Growth | — | — | — | 39%[17] | 56%[1] | — |
| Metric | Target | Source |
|---|---|---|
| Revenue | $3.0–$3.4B | [6] |
| ARR (End of 2026) | $7–$9B | [6] |
| Adj. EBITDA Margin | ~40% | [6] |
| CapEx | $16–$20B | [6] |
| Contracted Power | >3 GW | [6] |
| Connected Power | 800 MW–1 GW | [6] |
| Customer | Value | Duration | Scope |
|---|---|---|---|
| Microsoft | $17.4B[2] | 5 years | Dedicated GPU capacity from Vineland, NJ data center |
| Meta | $3.0B[3] | 5 years | AI infrastructure, deployed within 3 months |
| Israel Innovation Authority | $140M[15] | Multi-year | National AI supercomputer (4,000 B200 GPUs, 16,000 PFLOPS[18]) |
| Date | Type | Amount | Key Investors |
|---|---|---|---|
| Dec 2024 | Private Placement | $700M[12] | Accel, NVIDIA, Orbis Investments |
| Sep 2025 | Public Offering | $1.0B[13] | Public markets ($92.50/share) |
| Sep 2025 | Convertible Notes | Additional | Concurrent with equity offering[13] |
Arkady Volozh's quarterly shareholder letters reveal a consistent strategic narrative and escalating ambition:
| Quarter | Key CEO Quote / Insight | Strategic Signal |
|---|---|---|
| Q1 2025 | "On track to achieve $750M–$1B ARR by year-end." Revenue: $55.3M (+385% YoY). Expected EBITDA positive in H2 2025.[51] | Early validation; conservative guidance setting |
| Q2 2025 | Revenue grew 625% YoY and 106% QoQ. Achieved positive adj. EBITDA "sooner than expected." Raised ARR guidance to $900M–$1.1B.[52] | Demand exceeding plans; securing >1 GW power |
| Q3 2025 | "Every time we bring capacity online, we sell it." Running completely sold out, forcing rejection of enterprise customers. Meta $3B deal announced. Gross margins: 71%.[53] | Demand-constrained, not supply-limited; hyperscaler validation |
| Q4 2025 | "Everything we build, we sell. We are in the very early days of one of the biggest industrial and technological revolutions." Sold out Q3, Q4, and Q1 2026. Core cloud EBITDA margin: 24%.[54] | Operating leverage emerging; 9 new DCs announced; Rubin commitment |
| Metric | Q3 2025 | Q4 2025 | Notes |
|---|---|---|---|
| Cost of Revenue | — | 31% of revenue[56] | Implies ~69% gross margin |
| R&D Expense | — | 33% of revenue ($177.3M FY)[56] | Heavy investment in platform |
| SG&A Expense | — | 71% of revenue ($378.5M FY)[56] | Includes heavy GTM buildout |
| Operating Cash Flow | — | $834M (Q4)[55] | $401.9M FY from continuing ops |
| Cash Position | $4.9B[53] | $3.7B[55] | Spending down on CapEx |
| Convertible Notes | — | $4.16B issued[56] | Significant debt instrument |
| Shares Outstanding | — | 253M total[56] | 219.5M Class A + 33.6M Class B |
| Location | Type | Capacity | GPU Hardware | Status |
|---|---|---|---|---|
| Finland (Mantsala) | Owned | 25 MW active, expanding to 75 MW[27] | H100/H200 supercomputer clusters | Operational |
| New Jersey (Vineland) | Partnership (DataOne) | 300 MW[28] | Dedicated Microsoft capacity | Launching 2025 |
| Missouri (Kansas City) | Colocation (Patmos) | 5 MW initial, 40 MW expandable[24] | Blackwell & Hopper H200s | Operational |
| UK (Longcross, Surrey) | Colocation (Ark Data) | Liquid-ready, high-density | 4,000 Blackwell Ultra GPUs[29] | Q4 2025 |
| Iceland (Keflavik) | Colocation (Verne) | 10 MW[24] | 100% renewable energy | Operational |
| France (Paris) | Colocation (Equinix) | Urban campus | Early H200 adopters[24] | Operational |
| GPU Platform | Availability | Key Spec |
|---|---|---|
| NVIDIA HGX H100 | Available now | Foundation compute, cost-effective[22] |
| NVIDIA HGX H200 | Available now | Extended GPU memory[22] |
| NVIDIA HGX B200 | Available now | Blackwell architecture for LLMs[22] |
| NVIDIA GB200 NVL72 | Available now | Rack-scale liquid-cooled[22] |
| NVIDIA HGX B300 | Q4 2025[14] | AI reasoning architecture |
| NVIDIA GB300 NVL72 | Q4 2025[14] | 800 Gbps InfiniBand (first globally) |
| NVIDIA Vera Rubin NVL72 | H2 2026[16] | Next-gen; among first cloud providers |
Token Factory is Nebius's flagship managed inference product, launched November 2025.[7] It directly competes with The platform's Inference-as-a-Service strategy.
| Metric | Specification |
|---|---|
| Uptime SLA | 99.9% on dedicated endpoints[32] |
| Latency Target | Sub-second[32] |
| Throughput Capacity | 100M+ tokens per minute[32] |
| API | OpenAI-compatible[32] |
| Data Residency | EU (Finland, France) or US[32] |
| Zero-Retention Mode | No data stored or used for training[32] |
DeepSeek R1/V3, Llama 3.1/3.3, Qwen 2.5/3, GPT-OSS 120B/20B, Mistral, NVIDIA Nemotron, Gemma, GLM-4.5, Kimi-K2, Hermes, and more. Image generation via Flux and SDXL.
| Model | Tier | Input | Output |
|---|---|---|---|
| DeepSeek-R1-0528 | fast | $2.00 | $6.00 |
| DeepSeek-R1-0528 | base | $0.80 | $2.40 |
| DeepSeek-V3-0324 | base | $0.50 | $1.50 |
| Llama 3.1 405B | base | $1.00 | $3.00 |
| Llama 3.3 70B | base | $0.13 | $0.40 |
| Qwen3-235B-Instruct | base | $0.20 | $0.60 |
| Qwen3-30B-Instruct | base | $0.10 | $0.30 |
| GPT-OSS 120B | base | $0.15 | $0.60 |
| GPT-OSS 20B | base | $0.05 | $0.20 |
| Kimi-K2-Instruct | base | $0.50 | $2.40 |
| GLM-4.5 | base | $0.60 | $2.20 |
| Gemma-2-9b-it | base | $0.03 | $0.09 |
| Batch inference | 50% of real-time price | ||
| GPU | On-Demand ($/hr) | Reserved ($/hr) | Discount |
|---|---|---|---|
| NVIDIA HGX H100 | $2.95[35] | $2.00[35] | ~32% |
| NVIDIA HGX H200 | $3.50[35] | $2.30[35] | ~34% |
| NVIDIA HGX B200 | $5.50[35] | Contact | — |
| NVIDIA GB200 NVL72 | Contact[35] | Contact | — |
| NVIDIA L40S (AMD) | From $1.82[35] | — | — |
| NVIDIA L40S (Intel) | From $1.55[35] | — | — |
| Explorer Tier (H100) | $1.50 (first 1K hrs/mo)[36] | — | ~49% off on-demand |
Reserved clusters: up to 35% discount on on-demand rates for multi-month commitments.[35]
| Period | Headcount | Note |
|---|---|---|
| Dec 2023 (pre-restructuring) | 26,361[44] | Included all Yandex Russia employees |
| Dec 2024 (post-restructuring) | ~1,371[44] | Non-Russian operations only |
| Feb 2026 (current) | ~2,000+ (estimated) | 265 open roles suggest rapid growth[4] |
| Department | Est. Open Roles | Key Signal |
|---|---|---|
| Sales & Business Development | ~40+ | Global GTM (APJ, DACH, France, Middle East, US)[4] |
| Customer Experience / Solutions Architecture | ~30+ | Enterprise land-and-expand, HPC specialists[4] |
| Hardware Infrastructure & Data Centers | ~35+ | Massive physical buildout across 6+ sites[4] |
| Engineering (Backend, SRE, ML, HPC) | ~50+ | Platform maturity, C++, storage virtualization, YDB[4] |
| Product Management | ~8–10 | K8s, Security, Marketplace, API, Cluster Experience[4] |
| Security & Compliance | ~10+ | SOC, incident response, privacy, AI governance[4] |
| Recruiting & HR | ~15+ | Scaling the scaling team (exec recruiter, ML recruiter)[4] |
| Marketing & Developer Relations | ~20+ | Brand, community, 3 DevRel hires[4] |
| Finance & Legal | ~15+ | Public company infrastructure (audit, payroll, treasury)[4] |
| Operations & Strategy | ~15+ | GTM ops, pipeline management, strategic programs[4] |
| Role | Focus Area | Location | Key Requirements |
|---|---|---|---|
| PM, Kubernetes[41] | Managed K8s service, IaaS stack integration | Amsterdam, Remote EU | K8s expertise as user and DevOps engineer; IaaS knowledge (cloud networking, virtualization) |
| PM, Security[42] | Cloud security features, compliance | Amsterdam, Berlin, Finland, London, Remote EU | Security product experience, enterprise compliance |
| PM, Public API & Tools[4] | Developer experience, CLI/API | Amsterdam, Belgrade, Berlin, Prague, Remote EU | API design, developer tooling |
| PM, Marketplace[4] | ISV/partner ecosystem | Multiple EU locations | Platform marketplace, partner ecosystem |
| TPM, Cluster Experience[43] | GPU cluster UX for training/inference | Netherlands, Remote | HPC/GPU cluster management experience |
| Role | 25th Percentile | Median | 75th Percentile | 90th Percentile |
|---|---|---|---|---|
| Senior Product Manager[45] | $181K | $222K | $277K | $336K |
| Certification | Status | Scope |
|---|---|---|
| SOC 2 Type II | Achieved[46] | Full platform |
| HIPAA | Achieved[46] | Object storage audit logs, data-plane[14] |
| ISO 27001 | Achieved[46] | Information security management |
| ISO 27799 | Achieved[32] | Healthcare data protection |
| GDPR | Compliant[46] | EU data centers, zero-retention |
| Microsoft Entra ID | Integrated[14] | Enterprise SSO |
| FedRAMP | Not Announced | Likely roadmap item given Israel/US gov deals |
Security features include VPC Security Groups[14], per-object access controls, granular service roles, and FOCUS-compliant billing exports for FinOps.[14]
| Metric | Nebius | CoreWeave | Crusoe | Lambda |
|---|---|---|---|---|
| FY2025 Revenue | $530M[1] | ~$4.2B[47] | ~$600M* | ~$500M[47] |
| Connected Power | 220 MW[25] | 590 MW[47] | 370 MW | ~200 MW |
| Deployed GPUs | ~60,000[25] | ~600,000[47] | ~45,000 | ~40,000 |
| Backlog | $20.4B[2][3] | $55B+[48] | N/A | N/A |
| ClusterMAX Rating | Gold[26] | Platinum[49] | Gold | Silver |
| Managed Inference | Token Factory[7] | Inference Workloads | MemoryAlloy | Lambda Inference |
| Software Stack | Full (ex-Yandex)[30] | Growing | Growing | Basic |
| Key Advantage | Price + full stack | Scale + NVIDIA priority[47] | Energy + inference tech | Simplicity |
| H100 On-Demand | $2.95/hr[35] | $2.62/hr | $2.49/hr | $2.49/hr |
| Dimension | Nebius | Crusoe |
|---|---|---|
| Origin | Yandex cloud spinoff (decades of R&D)[10] | Clean energy Bitcoin miner pivot |
| Software Maturity | Hyperscaler-grade (K8s, storage, networking, DB)[30] | Building from scratch |
| Inference Product | Token Factory (disaggregated, 60+ models)[7] | MemoryAlloy (KV-cache fabric, Rust/CUDA) |
| Enterprise Contracts | Microsoft $17.4B + Meta $3B[2][3] | Growing enterprise base |
| Data Centers | 6+ locations (EU + US)[24] | 4+ US locations |
| Sovereign AI | Israel supercomputer, EU residency[15] | US-focused |
| GPU Provisioning | 45–60 sec container pull, <2 day cluster[26] | Fast provisioning (ClusterMAX Gold) |
| NVIDIA Relationship | Cloud Partner, early Rubin access[16] | Cloud Partner |
| CapEx (2026) | $16–$20B[6] | Not disclosed (smaller scale) |
| # | Action | Priority | Rationale |
|---|---|---|---|
| 1 | Benchmark against Token Factory pricing | Critical | Ensure The platform's inference costs compete with $0.13–$3.00/M token range[33] |
| 2 | Accelerate managed K8s and storage | Critical | Table-stakes for enterprise AI cloud buyers; Nebius has years of head start[23] |
| 3 | Pursue US government/defense AI deals | High | Follow Nebius's sovereign playbook; The platform has US energy + infrastructure advantage |
| 4 | Publish independent inference benchmarks | High | Nebius uses Artificial Analysis verification[9]; The platform needs similar third-party validation |
| 5 | Build developer ecosystem | Medium | Nebius has 3 DevRel hires + Head of Developer Community[4]; The platform needs equivalent investment |
| 6 | Target Nebius's customer concentration gap | Medium | Mid-market enterprises need alternatives to hyperscaler-dependent AI clouds |