Stay up to date with the latest changes in our product
Production-ready core engine with improved isolation and portability
2025/12/31
up_limit >= 100 (passthrough mode).Partitioned scheduling, non-locking device extension, and hard isolation milestone
2025/11/30
integral_decay_factor for PID controller to forget old clamp data.Computing isolation modes (soft/hard/shared), elastic rate limiter, and VRAM hard-isolation
2025/10/31
nvidia.com/gpu limits.ld.so.preload is not an ELF file bug by changing conf path.ld.so.conf.d for dynamic libs rather than LD_LIBRARY_PATH.so.1 fallback for apps detecting libcuda.so.1.Autoscaling, Karpenter node expansion, and GPU worker preemption
2025/09/30
up_limit >= 100 for passthrough scenarios.Large-scale benchmarking, RDMA support, and hypervisor probes
2025/08/31
Karpenter integration, progressive migration, and hypervisor/TUI enhancements
2025/07/31
dlsym hooking and NVML hook for device limiting (instead of env var).Scheduler framework refactor, alerting integration, and NVIDIA remoting milestone
2025/06/30
Multi-GPU requests, GPU model filtering, and per-GPU limiting via UUIDs
2025/05/31
TensorFusionWorkload lifecycle signaling (status/conditions).Canary rollout support and deeper limiter foundations (memory hooks, runtime env utilities)
2025/04/30
TFLOPs-based limiting, workload lifecycle controls, and richer GPU device metrics
2025/03/31
Cluster reconciliation hardening and control-plane stability improvements
2025/02/28
Metrics foundations and early scheduling/observability building blocks
2025/01/31