An end-to-end intelligent serverless computing platform integrating an ML-based request scheduler, custom gateway, load balancer, and autoscaler. Achieves 1.84× cost reduction and 97.2% SLA compliance benchmarked against AWS Power Tuning and standard Kubernetes HPA.
1.84×
Cost reduction vs. baseline
97.2%
SLA compliance rate
~0.1 ms
Scheduler inference latency
3
Industry baselines beaten
Benchmarked against AWS Power Tuning, standard HPA, and baseline OpenFaaS.
Advanced ML models forecast workload patterns and resource requirements ahead of arrival, enabling proactive warm-up and eliminating cold starts under predictable traffic.
Integer Linear Programming (ILP) solver computes mathematically optimal resource allocation, maximising utility subject to SLA and capacity constraints.
Custom OpenFaaS gateway with idle-first pod selection, real-time request routing, and Kubernetes Custom Resource Definitions (CRDs) for function lifecycle management.
Prometheus-backed multi-dimensional telemetry with cost analytics, per-function SLA tracking, and real-time rebalancing triggers.
Saarthi: An End-to-End Intelligent Platform for Optimising Distributed Serverless Workloads
Agarwal, S. et al. — arXiv:2511.06599
View on arXiv