Completely LOCAL and OFFLINE RAG (Retrieval Augmented Generation) Using LM Studio and LangChain!

  Рет қаралды 6,676

Sacred Scriptorium

Sacred Scriptorium

4 ай бұрын

This is my implementation of Local RAG for local document querying. Here is the script! github.com/ruddythor/localrag

Пікірлер: 21
@aufkeinsten7883
@aufkeinsten7883 3 ай бұрын
Hey man, appreciate you sharing this. Quick question; If I enter the prompts not in the command line but in LMStudio, the RAG-Pipeline will not be used, correct? Is there a way to write the prompts in LMStudio using this RAG-Pipeline or would I need to build my own UI (like a Web-UI or a small custom application) if I wanted to write prompts in a bit more user-friendly way?
@scriptoriumscribe
@scriptoriumscribe 3 ай бұрын
I don’t think there’s a way to use rag with on studio yet unfortunately. But you can join their discord and request it as a feature! I think there is demand for that. Gpt4all has a built in rag solution but it is only okay
@aufkeinsten7883
@aufkeinsten7883 3 ай бұрын
@@scriptoriumscribe Thanks for the quick reply! Will see whether a self-built RAG-pipeline can be integrated with ollama and open-webUI, will definitely check out gpt4all, appreciate the hint :) Do you know if the gpt4all-rag will be outperformed by something like shown in the video? Do you have an idea what this performance would depend on?
@petec737
@petec737 4 ай бұрын
Looks great, let us know when you can post the links, would love to give it a test. Out of curiosity, have you tested how much text you can give it as input?
@scriptoriumscribe
@scriptoriumscribe 4 ай бұрын
done! github.com/ruddythor/localrag
@scriptoriumscribe
@scriptoriumscribe 4 ай бұрын
also, i haven't tested how much i can give it as input, but that is next on my list to try when i get back around to rag!
@sharduljadhav2269
@sharduljadhav2269 2 ай бұрын
Great Job man! This cleared alot of my doubts regarding this topic. how can we use this in case of ollama? I want my document to be excel , so i used pandas for it, but it eventually gave me alot of errors. can you take input as an excel file ?
@scriptoriumscribe
@scriptoriumscribe 2 ай бұрын
I’m on holiday in Asia now and will try to do another video on this when I get back. But yes you should be able to use an excel file, though you would likely change the data loader. Thanks for the idea on the next video!
@mple2836
@mple2836 4 ай бұрын
how do we make it work ? I have lm studio, trying to do something similar to use my obsidan notes as reference for the llm.
@scriptoriumscribe
@scriptoriumscribe 4 ай бұрын
Hmm. Maybe I’ll put some better readme stuff in the repo. Let me do that tonight. Thanks for the comment!
@mple2836
@mple2836 4 ай бұрын
@@scriptoriumscribe thks man
@LIGTH-BIT
@LIGTH-BIT Ай бұрын
Is it necessary to use the call you use?
@Alpha1200
@Alpha1200 2 ай бұрын
Genuine question because I have absolutely no idea what I'm doing: I've downloaded LM Studio. I downloaded LLama2 13B Chat. I have a very large document (300.000 words) that I would like to use this RAG for. What do I do now, exactly? Like you linked the github, but what do I do with it? Do I just copy-paste the script to somewhere? To where?
@scriptoriumscribe
@scriptoriumscribe Ай бұрын
oh no, i've just now seen this! let me see if i can make some kind of blog post. but there's a few things you'd have to do to get something similar working. i'll see if i can make either a video or document explaining how to do this better. thanks for the comment! tldr, i think you'd have to download/clone/copy the script, then get your python environment set up for your computer, then install the python dependencies (pip install), then update the directory to point toward the correct directory. also worth noting you'd have to download LM studio and a couple ai models too. does this help some?
@lovelynxo54
@lovelynxo54 13 күн бұрын
@@scriptoriumscribe did u ever make a post for this? I'm a beginner and lost on how to set everything up as well? I appreciate it.
@portalpacific4500
@portalpacific4500 4 ай бұрын
What is the max word count of context you can add?
@scriptoriumscribe
@scriptoriumscribe 3 ай бұрын
I actually don’t know to be honest. Sorry.
@AIgen-ct3hz
@AIgen-ct3hz 2 ай бұрын
why the model is showing as gpt-3.5 turbo in the logs?
@scriptoriumscribe
@scriptoriumscribe 2 ай бұрын
Hm. Do you know the timestamp? Might have been from a previous run where I ran it with an open ai key. It can run both ways.
@emmanuelkolawole6720
@emmanuelkolawole6720 4 ай бұрын
Hello beyond the binary, please can you create another one just like this for CSV files?
@cristianaguilar4253
@cristianaguilar4253 22 күн бұрын
video no sound its maybe afraid microphone
RAG + Langchain Python Project: Easy AI/Chat For Your Docs
16:42
What is RAG? (Retrieval Augmented Generation)
11:37
Don Woodlock
Рет қаралды 86 М.
UFC 302 : Махачев VS Порье
02:54
Setanta Sports UFC
Рет қаралды 1,4 МЛН
Заметили?
00:11
Double Bubble
Рет қаралды 3,5 МЛН
Ну Лилит))) прода в онк: завидные котики
00:51
The Worlds Most Powerfull Batteries !
00:48
Woody & Kleiny
Рет қаралды 27 МЛН
Python's 5 Worst Features
19:44
Indently
Рет қаралды 84 М.
Reliable, fully local RAG agents with LLaMA3
21:19
LangChain
Рет қаралды 88 М.
Local RAG using Ollama and Anything LLM
15:07
GovBotics
Рет қаралды 10 М.
What is Retrieval-Augmented Generation (RAG)?
6:36
IBM Technology
Рет қаралды 533 М.
Run your own AI (but private)
22:13
NetworkChuck
Рет қаралды 1,1 МЛН
I Analyzed My Finance With Local LLMs
17:51
Thu Vu data analytics
Рет қаралды 408 М.
"Mastering LM-Studio: Unleashing LLMs locally | OffGrid-AI"
23:58
UFC 302 : Махачев VS Порье
02:54
Setanta Sports UFC
Рет қаралды 1,4 МЛН