Hi! PARIS Summer School 2025Speaker Insight – Aymeric Dieuleveut, École polytechnique As machine learning systems become embedded in critical decisions, from finance to infrastructure, the need for trustworthy, interpretable predictions has never been greater. Aymeric Dieuleveut, Professor of Statistics and Machine Learning at École polytechnique and scientific co-director of the Hi! PARIS Center, believes the key lies not in the models themselves, but in how we communicate their uncertainty. At this year’s Hi! PARIS Summer School, Dieuleveut introduced participants to conformal prediction a statistical framework designed to make machine learning outputs more transparent, reliable, and ready for real-world deployment. Key Takeaways Conformal prediction provides a flexible way to quantify uncertainty around machine learning predictions, offering guarantees that are easy to interpret. Rather than replacing existing models, conformal methods build on top of any trained predictor, including black-box models, probabilistic forecasts, or quantile regressors. Several trade-offs structure the design of conformal methods, especially between computational efficiency and statistical robustness. This approach has already been deployed in real-world applications, such as energy price forecasting at EDF. Conformal prediction is part of a broader ecosystem of methods, alongside privacy, decentralization, and robustness, that aim to build public trust in AI systems. Aymeric Dieuleveut at the Hi! PARIS Summer School 2025 Moving beyond the single prediction At its core, conformal prediction challenges a basic assumption in machine learning: that a model should produce a single best guess. Instead, it offers prediction sets, ranges or intervals, with statistical guarantees that the true value lies within. For Dieuleveut, this marks a shift not only in method, but in mindset. “When we make predictions with black-box models, we often don’t know how reliable the outputs are,” he explained. “Conformal prediction helps us go beyond that, to actually measure the uncertainty in a principled way.” Exploring the trade-offs During his tutorial, Dieuleveut walked participants through the two key trade-offs involved in designing conformal prediction methods. The first is between computational cost and statistical efficiency. Some variants, such as split conformal prediction, are simple and fast. Others offer stronger guarantees but require more intensive computation. The second trade-off concerns the strength of the guarantee. Most conformal methods ensure what’s known as marginal validity, meaning the coverage guarantee holds on average. But newer methods are moving toward conditional validity, where the coverage depends on specific conditions or inputs. “This is a subtle but important evolution,” Dieuleveut noted. “It brings us closer to more personalized, context-aware uncertainty estimates.” Aymeric Dieuleveut at the Hi! PARIS Summer School 2025 From energy markets to model deployment Conformal prediction isn’t just a theoretical construct, it’s already in use. One example Dieuleveut highlighted comes from the PhD work of Margaux Zaffran, conducted with EDF. By applying conformal methods to electricity price forecasts, her work helped quantify uncertainty in a domain where stakes are high and volatility is common. As Dieuleveut emphasized, this is one of the most compelling strengths of conformal prediction: it’s model-agnostic and ready to plug into existing systems. “People don’t want to retrain their entire model pipeline just to estimate uncertainty,” he said. “Conformal prediction allows them to add this layer on top.” Part of a broader trust ecosystem In a broader sense, conformal prediction is one piece of a larger puzzle. Alongside techniques focused on privacy, robustness, and decentralization, it contributes to building trust in AI systems. Each of these methods tackles a different dimension, privacy protects data, robustness handles adversaries, decentralization enables learning across networks, but all share a common goal: making machine learning models more reliable and aligned with real-world constraints. Dieuleveut also noted that, methodologically, these areas are deeply connected. Many draw from shared optimization principles and can be applied using overlapping toolkits. Compatible, not competitive One misconception Dieuleveut addressed during his session is the idea that conformal prediction is at odds with Bayesian or probabilistic approaches. In fact, the opposite is true. Conformal methods are often complementary, enhancing existing models rather than replacing them. “You can apply conformal prediction to virtually any trained model,” he explained. “That’s why it’s so powerful, it doesn’t throw away years of progress in other domains. It builds on them.” In a landscape where model reuse is critical and deployment pipelines are complex, that kind of adaptability isn’t just convenient, it’s essential.
Hi!ckathon 2022 challenge, opening on March 4.
The Career Fair event organized by Hi! PARIS Center offers our students the opportunity to identify potential future internships and job opportunities in AI and Data Science, receive career advice, and engage in discussions with the participating companies and startups.
Optimal Transport for Machine Learning is in the spotlight of the Hi! PARIS Reading groups in October-December 2025, a scientific networking action gathering affiliates and corporate donors around important topics of the moment!
The Hi! PARIS Meet Up! event organized by Hi! PARIS is about the “AI Act”. Save-the date: March 25, 2025 (6:00-8:30 pm), hosted by our corporate donor VINCI.
Transfer Learning is in the spotlight of the Hi! PARIS Reading groups in February-April 2025, a scientific networking action gathering affiliates and corporate donors around important topics of the moment!
In echo of UNESCO’s International Day of Women and Girls in Science, the Hi! PARIS Center is organizing a round table on February 14, 2025. The focus is “Breaking biases early: Empowering girls to envision themselves in STEM”