Why Cedana?
Reduce compute costs by 20%-80%
Eliminate idle compute. Automatically suspends and resumes your workloads based on activity. Automatically bin-packs containers across instances, freeing up resources at fine-grain resolution.
Never lose work — even if hardware fails
Upon hardware or OOM failure, automatically resume workload on a new instance without losing work.
3x your performance
Accelerate cold start and time to first token by resuming your CPU/GPU workload from it's previous state. Eliminate boot time, initialization and other steps.
Use Cases
01
Reduce Cloud Cost by 20%-80%
Reduce Cloud Cost by 20%-80%
Cedana automatically suspends and resumes your workloads based on idle or active status. Automatically bin-packs containers across instances, freeing up resources at very fine-grain resolution.
02
Improve Start Times 2x
Improve Start Times 2x
Cedana automatically suspends and resumes your workloads based on idle or active status. Automatically bin-packs containers across instances, freeing up resources at very fine-grain resolution.
Improve Performance and SLA Automatically
Save, Migrate, and Resume stateful GPU workloads. Live-migrate workloads dynamically based on user load and resource availability to improve performance and utilization. Policy-based automation enables workload-level SLAs to be enforced.
Speed Time to First Token 2x
Accelerate your inferencing time-to-first-token by eliminating the cold boot process including library and model initialization and optimization. Leverage your existing model optimizations. Significantly reduce network bottlenecks.
01
Reduce Compute Cost
Reduce Compute Cost by 20%-80%
Automatically suspends and resumes your workloads based on idle or active status. Automatically bin-packs containers across instances, freeing up resourcesat very fine-grain resolution.
02
Increase Throughput 2x
Increase Throughput 2x
Automation increases throughput and capacity. When a GPU fails, workloads are automatically resumed on a new instance without losing work. This reduces the significant manual overhead of restarting jobs on new instances. GPU utilization is automatically increased by treating the entire pool of available GPUs as a single, logical shared cluster, avoiding resource fragmentation or static reservation of capacity. These capabilities automatically increase throughput and capacity.
01
Minimize Downtime
Minimize Downtime by 90%
Databases, analytics, webservers and other stateful workloads continue without losing work even through node (instance) failures. Avoid costly over-provisioning to meet SLA requirements.
02
Reduce Costs
Reduce Cloud Cost by 5x
Long-running jobs can be run on spot instances.
When a spot instance is revoked, your job is
automatically continued on a new instance
without any manual intervention.
How it Works?
Save a process or container using our API. Saves the complete state of the workload including process and filesystem state, open network connections, in-memory (RAM and VRAM), data, namespaces and everything in between
Migrate the workload onto another instance.
Resume workloads as new process/container on another instance. Realtime performance with not service disruption.
Use Save, Migrate, Resume (SMR) to implement policy-based automation. Cedana automatically suspends and resumes workloads based on activity, enabling fine-grained bin packing of containers. This saves up to 80% of compute costs.