Skip to content
FarmGPU Blog
OCP 9 min read

Scaling AI Clusters at Neoclouds: How FarmGPU Is Helping Build the Open Playbook

Scaling AI Clusters at Neoclouds: How FarmGPU Is Helping Build the Open Playbook

Notes from the new OCP FTI workstream, the first 19 community survey responses, a deep dive on hardware management, and what FarmGPU is contributing in 2026.

The Open Compute Project (OCP) was founded on a simple idea: when hyperscalers share their designs, the rest of the industry stops re-inventing the same rack, server, and switch over and over. That worked. The problem in 2026 is that "the rest of the industry" now includes a new tier of operators — the neoclouds — who are deploying AI clusters that look hyperscale on the inside but live on neocloud-sized teams, balance sheets, and timelines on the outside.

That's the gap the new OCP Future Technologies Initiative (FTI) workstream, Scaling AI Clusters at Neoclouds, exists to close. I'm chairing the workstream together with co-leads Yann-Guirec Manac'h (Scaleway) and Kyle McCrindle (Denvr Dataworks), and FarmGPU is one of the operators contributing real designs, real survey data, and real production experience back into the open community.

Scaling AI Clusters at Neoclouds — OCP Community
The official OCP Future Technologies Initiative workstream page. Charter, meeting cadence, member list, and how to join.

Part 1 — What the workgroup is trying to do

The workstream was officially launched at OCP Global Summit 2025 and is part of the FTI portfolio (alongside Open Systems for AI and Open Cluster Designs for AI). It's an open forum — neoclouds, OEMs, networking and power vendors, colos, and end users — focused on the unique challenges of scaling xPU IaaS/PaaS infrastructure from a single pod to multi-site, hyperscaler-adjacent capacity.

The four goals

  1. Foster knowledge sharing and best practices. A requested forum for operators to discuss OCP adoption, contribution patterns, and cross-project coordination — both online and at Summits.
  2. Promote interoperability across scales. Drive integration between small edge AI deployments and large centralized clouds so hybrid neocloud models actually work for training and inference workloads.
  3. Enhance energy efficiency and share sourcing strategies. Optimize power, cooling, and resource utilization, and share what's working globally for sourcing energy and facilities — first-party and colo.
  4. Develop open standards for AI infrastructure modularity. Identify the gaps where neocloud-specific patterns aren't yet captured by Open Systems for AI or Open Cluster Designs for AI, and feed those gaps back into the right OCP projects.
The four "getting" problems. What we hear from operators every week boils down to:
Getting facilities — most traditional DCs and colos weren't built for 30–100kW racks, liquid cooling distribution, or AI-grade power topologies.
Getting electrons — grid availability, co-generation, medium-voltage AC vs. high-voltage DC distribution; the data center boundary is moving upstream.
Getting hardware (fast enough) — accelerator allocations, generational upgrade timing, and the management/automation surface area required to actually operate it.
Getting customers — visibility, trust signals, and channels (OCP Marketplace, Experience Centers, Summit panels) that help operators be discovered.

The workstream doesn't traditionally produce specifications itself — that's what OCP Projects do — but it's where neocloud-specific requirements get organized and routed into the right project, particularly Open Systems for AI and Open Cluster Designs for AI.


Part 2 — What the community is telling us

The 2026 Focus Survey

We ran a community survey to set 2026 priorities and got 19 responses from a broadly representative cross-section: nine neocloud operators (Crusoe, Lambda, Scaleway, Denvr Dataworks, LightningAI/Voltage Park, FarmGPU, Edinburgh IDF, and others), three networking vendors, plus power and facilities vendors (Vertiv, Hitachi Energy, AirTrunk), OEMs, OOB management vendors, and consultants.

Twelve of nineteen are already using OCP hardware or designs in some form ("Yes" or "Some"), and eight of nineteen describe themselves as actively scaling across multiple sites or in rapid expansion.

First round of survey results

What 2026 focus areas matter most

Focus area Importance (avg) OCP leverage (avg)
Power sourcing + distribution architectures 4.21 3.68
Benchmarking + performance assurance 4.05 3.79
Interoperability (BMC/BIOS/firmware/telemetry) 3.95 3.84
Sustainability + efficiency practices 3.95 3.79
Networking patterns + tuning guidance 3.89 3.89
Facilities + cooling readiness 3.63 3.47
Open rack/pod reference patterns (30–100kW) 3.58 4.11
Storage patterns for AI workloads 3.47 3.42

Read those two columns together: power and benchmarking are the most important problems, but the area where the community thinks OCP can help most is open rack and pod reference patterns for 30–100kW racks — exactly the modular building-block work the Open Cluster Designs project is producing.

The most pressing operator challenges

  1. Networking architecture (RoCE/IB, topology, congestion control, NIC tuning) — 9 responses
  2. Management & automation (provisioning, firmware, orchestration, Day-2 ops) — 8
  3. Interoperability & vendor lock-in (BMC/BIOS/firmware consistency, telemetry) — 8
  4. Power availability / grid constraints / energy sourcing strategy — 7
  5. High-power rack & pod patterns (30–100kW racks; modular pods below 1MW) — 6

The signal is consistent: at neocloud scale, the operational surface area — networking tuning, multi-vendor firmware/BMC/telemetry, day-2 provisioning — eats more time than any single piece of hardware.

OCP FTI Survey — Scaling AI Clusters at Neoclouds (2026 Focus): Full Results
Official, anonymized results from the 2026 community survey: importance scores, top challenges, deliverable rankings, subgroup interest, and free-text responses.

That data is now driving how the workstream organizes its 2026 calendar — including the deep dives.


Hardware management deep dive — April 13, 2026

The interoperability and management theme above isn't theoretical, so we ran a focused OCP Hardware Management deep dive with the community. To prepare, FarmGPU built a microsite that indexes every OCP hardware-management work area (Manageability Profiles, Redfish, OpenBMC, SPDM, Liquid Cooling Management, etc.) into a single navigable map of specs, work groups, and pain points so neocloud engineers don't have to dig through eight separate project pages to figure out where to plug in.

OCP Hardware Management — Microsite (built by FarmGPU)
One-page navigable index of OCP hardware-management work: Manageability Profiles, Redfish, OpenBMC, SPDM, Liquid Cooling Management, and the gap map for neoclouds.

OCP hardware management experts — including Jeff Hilland (OCP HW Mgmt project chair), Jeff Autor (HPE / Redfish Forum), and Mike Rainier — joined to answer questions live.

Recording — OCP Hardware Management deep dive (Apr 13, 2026)
Full session video. Walkthrough of the OCP HW Management project, Redfish profiles, multi-vendor pain points, and the case for a neocloud-specific Redfish profile.

What came out of the discussion:


Part 3 — How FarmGPU is using OCP today

We're not just chairing the workstream — FarmGPU is putting OCP to work in production, and contributing the patterns back. Three areas worth calling out.

1) The first neocloud running an OCP backend fabric

FarmGPU is the first neocloud to deploy on the OCP Open Cluster Designs–aligned AI Training Fabric Reference Architecture, which we co-developed with Hedgehog and Celestica as a contribution to OCP. The RA is a network-centric, prescriptive complement to the OCP Open Cluster Designs OPG-M (Open Pod Group of M xPUs) and XOC-N (xPU Open Cluster of N xPUs) system architecture papers.

Schematic showing two OPG-128 pods plus a faded third composing into an XOC cluster under a shared spine, with XOC-256/512/1024 size labels.OPG-M building blocks compose into XOC-N clusters under a shared spine. The Training Fabric RA delivers 36 ready-to-deploy composition variants spanning OPG-64/128/256/512 and XOC-256/512/1024 in air- and liquid-cooled densities.

Key elements of the RA:

For the inference side, the companion OCP AI Inference Fabric Reference Architecture (also Hedgehog + Celestica) targets multi-tenant inference at the 32–256 xPU scale with two profiles — Minimal Inference (no back-end RDMA fabric) and Distributed Inference (adds a dedicated RDMA back-end with one-toggle RoCE enablement) — plus a distributed open-source Hedgehog Gateway on x86 for NAT/PAT and ingress/egress.

The point of contributing both RAs is not "buy this exact stack." It's giving every neocloud — and every OCP-aligned colo and OEM partner — a vendor-neutral, declarative, reproducible starting point for AI fabric deployment that doesn't lock anyone into a single vendor's reference network design.

2) Adopting the OCP NVMe SSD specification — and building a predictive failure engine on top of it

FarmGPU standardizes on the OCP Datacenter NVMe SSD Specification for our storage layer. The OCP NVMe SSD spec is one of the cleanest examples of how operator-driven OCP work compounds: it started as Meta + Microsoft requirements, was adopted by Dell, HP, and the broader OEM ecosystem, and is now a de-facto datacenter SSD baseline — better health telemetry, sanitization behavior, latency outliers, power management, and security than vanilla NVMe. We use it as the spec we hold our SSD vendors to.

The richer telemetry surface is exactly what makes proactive drive replacement possible at neocloud scale. We've built an internal FarmGPU Predictive Failure Engine that consumes OCP-spec health/SMART logs across the fleet, scores drives for NAND degradation, ECC escalation, and recoverable/uncoverable error trends, and surfaces actionable recommendations (replace, watch, monitor) per drive — across all hosts, GPU servers and storage nodes alike.

FarmGPU Fleet Predictive screen showing drive risk scores, NAND degradation, ECC escalation, and recommendations across multiple hosts

🌱
The code behind the engine isn't open source yet — but FarmGPU expects to open-source the predictive failure engine so other neoclouds can pick it up directly and contribute back. This is the "neocloud profile" model in miniature: write down what we need, get the OEMs to ship it through OCP, and then open-source the operator tooling that turns the spec into operational value.

3) Helping organize the neocloud playbook in 2026

Internally, FarmGPU is committing to:


Get involved

If you build, operate, sell into, or buy from neoclouds, the workstream is open. We meet monthly (with deep-dive sessions every two weeks for active topics), and the next focus areas — driven directly by the survey results above — are power architectures, facilities readiness, networking, and management/automation.

Join the OCP Scaling AI Clusters at Neoclouds workstream
Charter, meeting cadence, and how to participate. Open to neoclouds, OEMs, networking and power vendors, colos, and end users.

Questions or want to help? Email me at [email protected].

OCP's slogan is community-driven hyperscale innovation for all. In the AI era, "all" has to include the neoclouds — and the operators who are pickier about real, measurable performance than anyone else in the industry. That's the bar we're aiming at.