Eliminate wasteful compute by automatically suspending and resuming workloads based on activity. This reduces resource fragmentation and the need to reserve capacity based on static demand estimates.
CPU/GPU Workload utilization. Use cases includes serverless infrastructure, dev tools, generative AI, scientific computing and hyperscalers.
From webservers to 30GB+ LLMs models, accelerate boots by 3x or more. Eliminate the boot process by resuming from a fully initialized state.
Save, migrate and resume (SMR) transparently across CPU (Intel, ARM) and GPU (nVidia) workloads. One API seamlessly provides SMR capabilities across different hardware. No code-modifications required.
We’ve deployed a test cluster for you to play with where you can interact and experiment with the system.
Learn more about how Cedana is transforming compute orchestration and how we can help your organization.
From deploying on your cluster, to market, to GPU Checkpointing, learn our system and get started quickly.