New York

October 15–17, 2025

Berlin

November 3–4, 2025

London

June 16–17, 2025

Navigating LLM Deployment: Tips, tricks, and techniques

Unlock best practices for deploying self-hosted LLMs—optimize performance, ensure reliability, and tackle real-world challenges in critical industries

Register or log in to access this video

Create an account to access our free engineering leadership content, free online events and to receive our weekly email newsletter. We will also keep you up to date with LeadDev events.

Register with google

We have linked your account and just need a few more details to complete your registration:

Terms and conditions

 

 

Enter your email address to reset your password.

 

A link has been emailed to you - check your inbox.



Don't have an account? Click here to register
June 25, 2025

Self-hosted Language Models are going to power the next generation of applications in critical industries like financial services, healthcare, and defense. Self-hosting LLMs, as opposed to using API-based models, comes with its own host of challenges – as well as needing to solve business problems, engineers need to wrestle with the intricacies of model inference, deployment and infrastructure. In this talk we are going to discuss the best practices in model optimisation, serving and monitoring – with practical tips and real case-studies.