AI that learns the cheapest way to charge your fleet
Instead of following fixed rules or relying on price forecasts, our AI learns the best charging schedule from your real-world data — and keeps improving over time.
Based on peer-reviewed reinforcement learning research from TU Delft and ETH Zurich
Three RL approaches
Each algorithm tackles the charging problem differently — from exploration-driven learning to constraint-aware optimization.
CPO — Constrained Policy Optimization
Key strengths
- Built-in safety guarantees during training
- Proven mathematical convergence properties
- Directly handles constraints without penalty tuning
- Monotonic policy improvement at each step
Best for
When you have unlimited training data and time — but impractical for real-world deployment.
DDPG / SAC — with manual penalty tuning
Key strengths
- Sample-efficient — learns from historical data
- Precise, continuous power control per socket
- DDPG: deterministic, reproducible schedules
- SAC: robust to uncertainty via entropy regularization
Best for
When you have time to manually tune penalty coefficients — but risky without expert oversight.
AL-SAC — Augmented Lagrangian SAC
Key strengths
- Zero constraint violations — guaranteed safety
- No manual penalty tuning — fully automated
- Data-efficient — reuses past experience for fast learning
- Lowest cost of all RL methods in comparative testing
Best for
The best choice for real-world deployment — safe, efficient, and fully automated.
How the AI learns your optimal schedule
The system interacts with a simulated version of your charging plaza, learning from every decision it makes.
Heuristic vs MPC vs RL
How do reinforcement learning approaches compare to the rule-based and optimization methods already in the simulator?
| Dimension | Heuristic | MPC | RL NEW |
|---|---|---|---|
| Approach | Hand-crafted rules | Mathematical optimization | Learned from experience |
| Needs forecasts? | No | Yes — accuracy critical | No — learns implicitly |
| Adaptability | None — static rules | Re-solve per horizon | Continuous adaptation |
| Setup effort | Low | Medium — model required | High — training required |
| Constraint handling | Manual checks | Built into solver | Learned (AL-SAC) |
| Computation | Instant | Seconds per horizon | Hours to train, ms to infer |
| Optimality | Feasible, not optimal | Locally optimal | Near-optimal with enough training |
| Uncertainty handling | Poor | Stochastic MPC possible | Natural — trained on noise |
Let AI optimize your charging costs
We're building AI-powered charging optimization for real-world deployments. Get in touch to discuss a pilot with your data.