DXG is Wolters Kluwer's central product technology organization.
We are a team of more than 7,000 professionals united under one mission: To drive faster, customer-focused innovation through next-generation technology and lean product development.
Digital eXperience Group
The Wolters Kluwer Digital eXperience Group (DXG), formerly Global Platform Organization, co-creates state-of-the-art digital solutions with our businesses around the world. The DXG mandate is to grow revenue in the company’s digital products through innovation in, and adoption of, advanced technologies and tools to meet and further anticipate customer needs. The group drives innovation in Wolters Kluwer through its user experience center of excellence, focused on customer-centric product development, and its artificial intelligence center of excellence, applying cutting-edge technologies for the next generation of expert solutions.
Senior LLM Engineer – Production AI Systems
Role Overview
We are hiring a Senior LLM Engineer to design, build, and operate LLM-powered systems in production.
This role focuses on system-level LLM engineering, including architecture, evaluation, latency and cost control, and operational reliability.
Responsibilities
Design and implement LLM-based systems for production use cases.
Participate in architectural decisions around model selection, prompt designs, data flows, and system constraints.
Optimize LLM usage for latency, cost, and reliability, using techniques such as prompting strategies, caching, and system-level optimizations.
Establish and own the evaluation framework for AI systems, from defining metrics to implementing automated evaluation pipelines.
Architect, build, and rigorously evaluate RAG pipelines, with a strong focus on retrieval and generation metrics.
Implement guardrails and constraints to improve the reliability and safety of model behavior, and to mitigate known failure modes.
Required Skills & Experience
Strong Python skills, with production-grade engineering standards.
Proven experience building and operating LLM systems in production, including performance and latency trade-offs.
Deep understanding of LLM failure modes (hallucinations, drift, prompt sensitivity) coupled with a rigorous, metric-driven approach to evaluating and mitigating them.
Ability to work autonomously and make sound engineering trade-offs under real-world constraints.
Professional-level English communication skills.
Nice to Have
Hands-on experience deploying, serving, and optimizing open-source large language models (e.g., Llama 3, Mistral, Mixtral) in production.
Familiarity with LLM application frameworks and orchestration layers (e.g., LangChain, LlamaIndex, or custom implementations).
Strong DevOps / MLOps background.
Location: Bois-Colombes (France) or Bari (Italy) - hybrid
Our Interview Practices
To maintain a fair and genuine hiring process, we kindly ask that all candidates participate in interviews without the assistance of AI tools or external prompts. Our interview process is designed to assess your individual skills, experiences, and communication style. We value authenticity and want to ensure we’re getting to know you—not a digital assistant. To help maintain this integrity, we ask to remove virtual backgrounds and include in-person interviews in our hiring process. Please note that use of AI-generated responses or third-party support during interviews will be grounds for disqualification from the recruitment process.
Applicants may be required to appear onsite at a Wolters Kluwer office as part of the recruitment process.