No video

Chunking Best Practices for RAG Applications

  Рет қаралды 9,753

KX

KX

8 ай бұрын

Join our livestream chat on Chunking Best Practices for Retrieval Augmented Generation. In this session, Data Scientist Ryan Siegler will overview how to choose the best chunk size and different chunking methods, including:
- Naive Chunking
- Structural Chunkers
- Summarization
- Extraction
- and Multi-Modal Chunking
As we observe increasing context windows for LLMs, surfacing the most relevant information to the LLM will still yield you the best results. Come learn how to optimize your RAG pipelines with these chunking tips.

Пікірлер: 13
@mauriciolopes8502
@mauriciolopes8502 6 ай бұрын
Thank you, Ryan! Awesome lecture.
@deepaksingh9318
@deepaksingh9318 5 ай бұрын
Thanks .. It was A very good content and full of Details..
@reiniervaneijk
@reiniervaneijk 4 ай бұрын
Good job guys, valuable talk thnx
@tizulis2
@tizulis2 5 ай бұрын
excellent presentation!
@user-cf7hs6by9w
@user-cf7hs6by9w 8 ай бұрын
Keep up the good work!
@tonylv6119
@tonylv6119 5 ай бұрын
Sometimes, document has some images and figures inside, i think that's hard part to deal with that for RAG.😊
@Jonathan-rm6kt
@Jonathan-rm6kt 7 ай бұрын
Hi, thanks for the video, really covered a lot of relevant questions for me. Open question to the community- I have been struggling with the retrieval relevance for relatively small chunks using ada-002 (OpenAI embedding). For example, I search do a similarity Search on a key word ("sea slug") I know only appears a few times, and the top k result doesn't even include either parts words. It appears in the text as "sea-slug", but this feels extremely brittle and like something the embeddings should capture. Is this somewhat expected? Hence the need for more complicated retrieval?
@RyanSieglerAI
@RyanSieglerAI 6 ай бұрын
Since the embeddings capture the context of a chunk, it isn't focused on specific words (this is where hybrid search can come into play). My thought is the embedding model doesn't know much context around a word like "sea-slug" so potentially finetuning the embedding model with some examples using that phrase, or using a hybrid search method would help.
@maryamashraf6370
@maryamashraf6370 6 ай бұрын
Great video, learnt allot! Had a question. What should be the chunking approach for a RAG application scraping the Internet for context? Since the documents would be web pages I get that you'd start off with the html splitter, but what approach should you use to try to get as much relevant context as possible while limiting the number of pages you embed? Especially considering that embeddings will be made in real time, trying to make the process as fast as possible etc. Would the approach be very different from using an offline document corpus?
@vijaybrock
@vijaybrock 3 ай бұрын
Hi Sir, Can you suggest me the best chunking strategy for 10K reports (pdfs) to chat with?
@soren81
@soren81 6 ай бұрын
Great video! I have a question about chunk decoupling. Shouldn't the vector storage embedding do pretty much the same abstraktion with the large text, as the summary does? I mean, wouldn't the summary and the original end up i the same place in the vector space, rendering the summary more or less pointless?
@RyanSieglerAI
@RyanSieglerAI 6 ай бұрын
Thanks for the question! In this context, the summary should highlight the key points and concepts in the original document, which should make retrieval more accurate especially in cases where there are documents covering similar/adjacent concepts. This is because in a full document there could be unnecessary information that could throw off vector search. The quality of the summary needs to be high for this to work. If the quality of the summary is not good and does not present the key points of the original document then yes it would be better to just embed original document as a whole.
Semantic Chunking for RAG
1:03:16
AI Makerspace
Рет қаралды 5 М.
Lehanga 🤣 #comedy #funny
00:31
Micky Makeover
Рет қаралды 27 МЛН
لااا! هذه البرتقالة مزعجة جدًا #قصير
00:15
One More Arabic
Рет қаралды 51 МЛН
Why Is He Unhappy…?
00:26
Alan Chikin Chow
Рет қаралды 101 МЛН
managed to catch #tiktok
00:16
Анастасия Тарасова
Рет қаралды 16 МЛН
The 5 Levels Of Text Splitting For Retrieval
1:09:00
Greg Kamradt (Data Indy)
Рет қаралды 63 М.
How to Build a Real-Time Multimodal RAG Application in Minutes
1:00:40
Merantix AI Campus
Рет қаралды 2,3 М.
Semantic Chunking for RAG
29:56
James Briggs
Рет қаралды 21 М.
How to set up RAG - Retrieval Augmented Generation (demo)
19:52
Don Woodlock
Рет қаралды 24 М.
Advanced RAG Techniques with @LlamaIndex
48:35
Timescale
Рет қаралды 3 М.
Why Agent Frameworks Will Fail (and what to use instead)
19:21
Dave Ebbelaar
Рет қаралды 43 М.
Lessons Learned on LLM RAG Solutions
34:31
Prolego
Рет қаралды 23 М.
Mastering Retrieval for LLMs - BM25, Fine-tuned Embeddings, and Re-Rankers
1:50:09
LlamaIndex Sessions: 12 RAG Pain Points and Solutions
37:57
LlamaIndex
Рет қаралды 13 М.
Lehanga 🤣 #comedy #funny
00:31
Micky Makeover
Рет қаралды 27 МЛН