Best Practices For Fine Tuning Mistral

  Рет қаралды 315

Hamel Husain

Hamel Husain

Күн бұрын

Sophia Yang discusses best practices for fine-tuning Mistral models. We will cover topics like: (1) The permissive Mistral ToS and how it's perfect for fine tuning smaller models from bigger ones (2) How should people collect data (3) Domain specific evals (4) Use cases & examples (5) Common mistakes
This is a talk from Mastering LLMs: A survey course on applied topics for Large Language Models.
For more info and resources related to this talk, see:: parlance-labs.com/talks/fine_...
My personal site: hamel.dev/
My twitter: x.com/HamelHusain
Parlance Labs: parlance-labs.com/
00:00 Introduction
Sophia Yang introduces herself and provides an overview of the talk, which will cover Mistral models, their fine-tuning API, and demos.
0:35 Mistral's History and Model Offerings
Sophia discusses Mistral's history, from their founding to the release of various models, including open-source and enterprise-grade models, as well as specialized models like CodeStraw.
02:52 Customization and Fine-Tuning
Mistral recently released a fine-tuning codebase and API, allowing users to customize their models using LoRa fine-tuning. Sophia compares the performance of LoRa fine-tuning to full fine-tuning.
04:22 Prompting vs. Fine-Tuning
Sophia discusses the advantages and use cases for prompting and fine-tuning, emphasizing the importance of considering prompting before fine-tuning for specific tasks.
05:35 Fine-Tuning Demos
Sophia demonstrates how to use fine-tuned models shared by colleagues, as well as models fine-tuned on specific datasets like research paper abstracts and medical chatbots.
10:57 Developer Examples and Real-World Use Cases
Sophia showcases real-world examples of startups and developers using Mistral's fine-tuning API for various applications, such as information retrieval, medical domain, and legal co-pilots.
12:09 Using Mistral's Fine-Tuning API
Sophia walks through an end-to-end example of using Mistral's Fine-Tuning API on a custom dataset, including data preparation, uploading, creating fine-tuning jobs, and using the fine-tuned model.
19:10 Open-Source Fine-Tuning with Mistral
Sophia demonstrates how to fine-tune Mistral models using their open-source codebase, including installing dependencies, preparing data, and running the training process locally.

Пікірлер
Language models on the command-line w/ Simon Willison
1:07:05
Hamel Husain
Рет қаралды 1,6 М.
🌊Насколько Глубокий Океан ? #shorts
00:42
ОДИН ДЕНЬ ИЗ ДЕТСТВА❤️ #shorts
00:59
BATEK_OFFICIAL
Рет қаралды 7 МЛН
버블티로 체감되는 요즘 물가
00:16
진영민yeongmin
Рет қаралды 76 МЛН
A Kubernetes Primer For Data Scientists, Part 1
32:58
Hamel Husain
Рет қаралды 1,5 М.
QLoRA-How to Fine-tune an LLM on a Single GPU (w/ Python Code)
36:58
Build an SQL Agent with Llama 3 | Langchain | Ollama
20:28
TheAILearner
Рет қаралды 2,3 М.
How To Evaluate MLOps Tools
27:18
Hamel Husain
Рет қаралды 8 М.
What Makes Large Language Models Expensive?
19:20
IBM Technology
Рет қаралды 63 М.
nbdev live coding with Hamel Husain
1:08:00
Hamel Husain
Рет қаралды 3 М.
LangChain Explained in 13 Minutes | QuickStart Tutorial for Beginners
12:44
Generative AI in a Nutshell - how to survive and thrive in the age of AI
17:57
Prompt Engineering, RAG, and Fine-tuning: Benefits and When to Use
15:21
Intro to Amazon EMR - Big Data Tutorial using Spark
22:02
jayzern
Рет қаралды 18 М.
WHO DO I LOVE MOST?
0:20
dednahype
Рет қаралды 2 МЛН
Let me show you (P1)
0:21
Discovery Boy
Рет қаралды 2,7 МЛН
это самое вкусное блюдо
0:12
Katya Klon
Рет қаралды 2,1 МЛН
How to get convenience store snack for free
1:00
Mykoreandic
Рет қаралды 43 МЛН