A different problem at a different scale. The 2019–2021 platform created the infrastructure substrate. By 2021, the question shifted: how do you govern an AI capability that has become shared infrastructure for 10+ data science teams across an industrial group — spanning energy retail, grid operations, commercial analytics, and customer experience — while continuing to push the frontier toward Generative AI and autonomous agent systems?
This is not a continuation of the previous role. It is a different problem: platform governance at enterprise scale, technology horizon management, and the translation of fast-moving AI research into production-viable architecture — while keeping 100+ daily jobs running at 99.9% uptime.
Platform governance and engineering leadership. Defined and maintained the GenAI/ML platform roadmap for EDF Group. Led the Architecture Review Board: produced reusable reference blueprints — RAG pipeline patterns, agentic workflow contracts, model serving standards, feature store schemas — adopted across business units without mandating them. Quarterly technical strategy presented to C-level, translated into investment priorities and team roadmaps.
The hardest governance problem was not technical standardization. It was managing the tension between platform stability (10+ teams on shared infrastructure) and technology velocity (LLM tooling evolving faster than any previous stack). Decision: stable infrastructure contracts at the platform layer, swappable tooling at the application layer. Teams adopted LangChain or LlamaIndex without the platform changing. The serving infrastructure, observability contracts, and deployment pipelines stayed consistent across all of them.
GenAI and Agentic AI in production. Moved LLM integration beyond proof-of-concept into operational systems: RAG pipelines with evaluated chunking strategies and retrieval quality metrics; pgvector for semantic search at document corpus scale; LangGraph state machines for multi-step agentic workflows with human-in-the-loop checkpoints.
Flagship delivery: AutoCons-Radar — an LLM-powered eligibility engine processing French public procurement data (BOAMP, TED/JOUE, DECP) to identify photovoltaic self-consumption projects 12–24 months before Enedis registration, enabling EDF commercial teams to anticipate market opportunities ahead of competitors.
Led the Gemini Code Assist rollout across engineering teams: adoption strategy, guardrail configuration, cost optimization, and the organizational change management that determined whether the tool was actually used or became shelfware.
LLMOps as a first-class discipline. Established LLMOps practices for production LLM systems: prompt versioning, evaluation frameworks (RAGAS + domain-specific evals), per-team cost monitoring, hallucination guardrails, trace logging for audit and debugging. The principle from 2019 carried forward: if you can’t observe it, you can’t operate it. Applied to LLM outputs, evaluation became an operational concern — not a pre-deployment checkpoint.
MLOps at sustained scale.
| Metric | Result |
|---|---|
| Platform uptime SLA | 99.9% across 100+ daily training jobs |
| Infrastructure cost | −30% via FinOps on multi-M€ AWS bill |
| Incident resolution | <90 min median (from 4h at platform launch) |
| Teams served | 10+ data science & product teams |
Stack — Python · LangChain · LangGraph · LlamaIndex · pgvector · Mistral · Gemini · Ollama · MLflow · Evidently · FastAPI · Kafka · Airflow · Kubernetes · OpenShift · AWS · Terraform · Prometheus · Grafana · ELK