The most comprehensive and clear explanation of RAG that I have ever seen. Thank you for your work. I sincerely wish you success and prosperity!
@genai-level-up24 күн бұрын
This breakdown of vector stores and retrievers clicked for me at 00:00:27 . Explaining embeddings as coordinates in multi-dimensional space is such a brilliant analogy! It finally makes sense how similar chunks of information can be grouped together. Thanks for making this complex topic so digestible!
@bhaveshdavda653526 күн бұрын
This is how I love learning when I'm short on time! Just the right depth and breath to get a better grasp of a topic I've read numerous articles about but never understood at the level this video made me understand. Thank you!
@majjikishore868927 күн бұрын
Thanks for this amazing gold mine content, keep on making videos sir
@acastanob5 ай бұрын
Came here out of curiosity, and ended up watching the full video. Thanks for taking the time to explain the very basics. Learned a lot!
@felipeclavijo17365 ай бұрын
YOU ARE JUST INCREDIBLE !!!!!!!!!! keep them coming. you are pretty much my main teacher.
@koushik7604Ай бұрын
Omg! This is a real gem. Thank you so much.
@RajaRahamathullah045 ай бұрын
This was amazing! Thank you very much for all the hard work - you’re incredible! Keep up the great content. My humble request to continue with UI interface for this RAG application.
@nicoloferrari59093 ай бұрын
That's incredible how good you're at explaining this though argument! Thanks a lot for your work. Really appreciate it
@ic4roswings5 ай бұрын
You are the man, I have watched god knows how many videos about rag and i finally get it, Thank you very much
@gisdevelopr3 ай бұрын
Very beautiful explain each step and make it so simple to understand, thanks for providing this video.
@nmstoker5 ай бұрын
Great video and nice that it's possible to run entirely locally, all with open source 🎉
@iiboo15 ай бұрын
Fantastic! You've really made my day by explaining it so clearly. Thank you!
@NumericLee5 ай бұрын
outstanding. for next video, I would love to see how LLMs are applied to mine unstructured data
@davidtindell9505 ай бұрын
Thank You. I have done many different RAG apps. A future vid suggestion: comparing results from completely local RAG to using remote embeddings, say OpenAI. I am finding that the remote embeddings are consistently better quality and the Q&A responses are better. To me, it is pointless to do a RAG app if the embeddings are poor and the answers are mediocre !
@R-f3u6z4 ай бұрын
Thanks for the fantastic explanation!
@pedrolima-lr3lu2 ай бұрын
Best explanation ! Thank You
@マーベリックテック5 ай бұрын
Absolutely loved it, Thank you for your efforts 🙂
@iamliam12415 ай бұрын
Very clear and useful. Thank you!
@Ali-km8kn3 ай бұрын
Great video and explanation! Thank you. I've a question. Will the context variable be inputted to the model through the prompt as embeddings of each page of the four pages or it will be converted back to string? Thank you in advanced.
@uwegenosdude5 ай бұрын
Thanks a lot for your very interesting video. It's great that all the time your code is working out of the box. Only langchain-ollama was missing in the requirements.txt. And unfortunately faiss-gpu is not supported on Windows 11 (AFAIK?). Great stuff you are offering to all of us all the time. Your explanations are always so good to understand. Amazing !! Please keep going !
@bald_ai_dev5 ай бұрын
is it better to use specialized embedding models like nomic-embed-text or llama3.1 itself as an embedding model? also can you please do a tutorial on some of the major rag ideas like building a self correcting rag (CRAG) and the compare the results with naive rag using an evaluating framework like ragas, giskard etc?
@HassanAllaham5 ай бұрын
This is a very good comment and a very useful request.. I would like the model to respond not just with the answer but also with the source of that answer (file name + page number = make sure the model is not drunk one 🧐). I believe this would be great if we add a re-ranker model 💯
@rayofvictory5 ай бұрын
@@HassanAllaham that will be great! @underfitted can you please chime in?
@kamalkhanal4095 ай бұрын
@@HassanAllaham For that, you need to keep the retriever output in a variable or a list while executing.
@thedevmachineАй бұрын
Santiago, I loved the video. Very clear explanation. I have a question. As you know, there is a limit to passing a prompt. For example, if I want a summary of a whole document, in theory, I have to pass the whole document to the LLM so it can create a summary of it. But this won’t fit in the context window. Chatgpt has I think 128K limit on the api but OLLAMA does not have this I think. Also I have no idea if 128K is enough for any LLM's. If I already stored a large document on my vector database how could I pass the whole document to LLM to summarize it? I cant just add whole document in the prompt. Thanks
@sirishkumar-m5z5 ай бұрын
RAG is a powerful tool for working with open-source models. It's a good idea to explore alternative tools as well, ensuring you choose the best fit for your specific needs.
@kamalkhanal4095 ай бұрын
Love the video! Could you please create a video showing how to export a Jupyter notebook into a proper project structure and deploy it on the cloud?
@mehershahzad-n5s4 ай бұрын
You well explained RAG
@sam-uw3gf5 ай бұрын
Great video
@andrewsperspective5 ай бұрын
This was perfect thank you
@fushumang17165 ай бұрын
next video is how to query if documents have images. Can LLMs describe or get context from images
@Lucky-op7qz5 ай бұрын
superrr amazinggggg,explaination
@AkshayRanchod-p4e4 ай бұрын
Apparently import langchain_ollama does not exist. I keep getting this error when trying to run the model
@Nabad0775 ай бұрын
You are really really good
@underfitted5 ай бұрын
🙏🏻
@treakfreak34663 ай бұрын
Hey there thanks , your videos are really helpful. I am student creating project around rag I want a video how can I make interface oriented or easy full stack rag bot without the large GPU
@papalevies5 ай бұрын
faiss-gpu only supports up to python 3.10, is there an alternative?
@JohnSanabria4 ай бұрын
It is possible to run the jupyter notebook on Google Colab? How it could be?
@bharathjpv93343 ай бұрын
Upload it from local
@mdsoykot29325 ай бұрын
I'm facing an issue trying to install faiss-gpu on a Mac with an M3 Pro chip. Is anyone else having this problem?
@sharangkulkarni17595 ай бұрын
good video
@bharasiva965 ай бұрын
Gravenberch was my motm
@ChronicleContent5 ай бұрын
*"An unnecessarily complicated introduction to RAG that only works locally.". There I fixed it for you.
@rayofvictory5 ай бұрын
May I know what is unnecessarily complicated? He is taking the time to go step by step for users to scale this solution for our use cases.
@tecnopadre5 ай бұрын
People always have to criticize, whatever it is.@@rayofvictory
@o_glethorpe5 ай бұрын
If his explanation is too complex for you maybe this subject is not for you.
@ChronicleContent5 ай бұрын
@@o_glethorpe I am not talking about me. This is supposed to be introduction. You don't need any of these to create a RAG.
@ChronicleContent5 ай бұрын
@@rayofvictory you don't need langchain or any other of these tools to create a RAG. Also he mentions "a user employee asks" but all this is local so that's not true.