- This event has passed.
State of the art – Practical insights for LLM fine-tuning and evaluation
[The workshop will be dispense in english ]
Machine Learning Level :
**Good knowledge of Machine Learning
*Basic skills in Python
If vendors almost announce to sell AGI-as-as-service through API to enterprise client, the reality after trying to use proprietary LLM as a service on a specific use case is often different. Non relevant generation, out-of-context answer, misunderstanding of the user queries and more broadly lacking subject matter expertise starts to erode the users and shareholders confidence of the potential transformative power of deploying LLM-enhanced business workflow across your organisation. You’re not alone in this journey.
In this talk, we’ll explore the landscape of fine-tuning solutions for open-source LLM, weighing their pros and cons. We’ll delve into the data required and how to design a robust evaluation framework to systematically assess your in-house model’s performance.
We’ll deep dive on the subtle differences between the Parameter Efficient Finetuning Methods PEFT), the reinforcement learning approaches, what to keep in mind when considering which one to use.
This talk is a synthesis of deploying LLM capabilities at various organisations, from startup to corporate environments. It’s a blend of insights from research papers and pragmatic experiences. We won’t go onto the details of the mathematical operations under the hood for each fine-tuning approach, instead our goal is to share the intuition of those concepts, equipping you to design an effective roadmap for fine-tuning an LLM for your specific business use case.
Slides in pdf will be made available for free on the speaker twitter at the end of the talk @fpaupier.
François Paupier, machine learning engineer, fpaupier engineering services