AI Infrastructure Management from Day 0 to Day N.
UEC & UAlink Ready. Built for GPUaaS and NeoCloud environments.
Unlock the full potential of your AI infrastructure with dynamic deployment of composed platforms.
Cruz Open Compute Orchestrator (COCO) by Dorado revolutionizes AI infrastructure management, setting a new standard for efficiency, scalability, and ROI. Designed to enable organizations to deploy and operate GPU-as-a-Service (GPUaaS) platforms and NeoCloud environments, COCO leverages open compute standards and UEC/UAlink readiness to deliver a unified control plane for the entire AI infrastructure lifecycle—from initial deployment (Day 0) to ongoing operations (Day N). By transforming infrastructure management into a streamlined, single-pane-of-glass experience, COCO breaks down operational silos between NetOps, CloudOps, and AIOps, simplifies the deployment of networks, compute clusters, and storage, and integrates AI Tenancy aligned with business needs. With real-time visibility into job status and resource utilization, seamless integration with leading orchestration tools like Slurm and Kubernetes, and AI-powered recommendations for optimizing performance, COCO maximizes GPU utilization rates, empowering organizations to unlock the full potential of their AI investments.
AI Infrastructure Orchestration
Comprehensive platform for managing AI infrastructure from bare metal to orchestrating AI Compute Clusters, Tenancy and Interconnect — initial deployment (Day 0) to ongoing operations (Day N).
Built for GPUaaS and NeoCloud environments.
Day 0 Deployment
Rapid infrastructure provisioning and configuration. Deploy complete AI infrastructure in hours
Day 1-N Operations
Continuous monitoring, scaling, and optimization. Automated management throughout infrastructure lifetime
AI Optimized
Purpose-buit for GPU-accelerated workloads with optimized networking and compute designs
Unify compute, networking, and storage management under one control plane.
GPUaaS
Transform GPU infrastructure into a shared service platform with multi-tenant support. Enable multiple teams and tenants to efficiently use GPU resources while monitoring isolution, security, and performance.
Maximize ROI.
Resource Isolation
Secure tenant separation with namespace isolation, network policies, and quota management
Usage Metering
Detailed tracking of GPU hours, memory usage, and compute cycles per tenant
Fair Scheduling
Intelligent workload distribution to maximize utilization and minimize idol time
At the heart of COCO is the Cruz AI Fabric Controller—a unified management layer that orchestrates all aspects of AI infrastructure.

Scale Up and Scale Out Fabrics
Built on open standards with support for next-generation AI networking technologies.
AI-Optimized Performance
Purpose-built for GPU workloads with optimized networking, storage, and compute configurations for training and inference.
Maximize ROI
GPUaaS multi-tenacy and intelligent scheduling achieves improved utlization, maximixing return on GPU infrastructure investment.
Reduce OPEX
Proactive monitoring and intelligent resource allocation dramatically reduces operational costs.
Rapid Deployment
Day O to production in hours with automated provisioning, configuration, and valiation of compute infrastructure.
