RAG vs. Fine Tuning

  Рет қаралды 59,441

IBM Technology

IBM Technology

Күн бұрын

Пікірлер: 70
@Kk-ed1gr
@Kk-ed1gr 2 ай бұрын
Thank you for the clarification, I had this question in mind last week, and I am glad that you have provided the answers I need.
@IBMTechnology
@IBMTechnology 2 ай бұрын
Glad it was helpful!
@yusufersayyem7242
@yusufersayyem7242 2 ай бұрын
35 minutes after downloading the clip, I received a notification, perhaps due to the weak internet in my country.... Finally, I would like to thank you sir for this wonderful explanation
@IBMTechnology
@IBMTechnology 2 ай бұрын
You're welcome!
@FauziFayyad
@FauziFayyad 2 ай бұрын
I have just watched the 1 years ago, then it updated today. Amazingg 🎉
@educationrepublic9273
@educationrepublic9273 2 ай бұрын
Love IBM's short and sharp explainers! Thank you for an excellent video once again :)
@florentromanet5439
@florentromanet5439 2 ай бұрын
I wanted to scream "WHY NOT BOTH⁉️) until 7:35 😂
@sabeensadaf4926
@sabeensadaf4926 14 күн бұрын
Thanks a lot, that was a good one to understand both RAG and fine tuning.
@RaushanKumar-qb3de
@RaushanKumar-qb3de 17 күн бұрын
Wow... combination is great. Thanks dear! For information
@EmaPython
@EmaPython 8 күн бұрын
Great video, thanks. It was useful for me
@emont
@emont Ай бұрын
your video didn't included EDA, LLM answers based on pre loaded info, a future evolution is LLM answering based on real time information
@MukeshKala
@MukeshKala 2 ай бұрын
Great explanation ❤
@CalvHobbes
@CalvHobbes Ай бұрын
Thank you, this is very useful. I'm curious about how the volume of data might affect the choice of FT vs RAG. If we tune the model again with the new data, would it become much larger over time? On the other hand, if we use RAG, would the restrictions on context length hold us back (i.e. if we don't want a very expensive model)?
@joeyjiron06
@joeyjiron06 Ай бұрын
Love the video! I’m building an app that empowers users to generate landing pages with a prompt using AI. I’m planning on building many custom components/sections that I want the model to use when generating a response. I want the model to choose the right sections to use and fill out the copy in the components to fit the users prompt. What would be the best way to handle this in the model? Use rag, fine tuning, both, neither, something else?
@sandeeppatil5925
@sandeeppatil5925 Ай бұрын
wonderful explanation ... will love to also know the choice of use from TCO or cost POV
@shrutisingh9801
@shrutisingh9801 Ай бұрын
can u make a video about reinforcement learning and performance evaliation of llm models?
@johannvgeorge8393
@johannvgeorge8393 Ай бұрын
Thank you for this helpful video🙂. Could you please explain the implementation of how we can update the RAG system with the latest information?
@kalcavaleiro6993
@kalcavaleiro6993 Ай бұрын
updating the database? cz RAG using such as vector database that updated regularly using similarity from the prompt to the database content to retrieve and then augmented the prompt to the LLM
@mark-lq4rk
@mark-lq4rk 2 ай бұрын
Thank you for the fascinating presentation. Assume certain conditions are similar, how would the cost of rag and fine-tuning differ?
@IBMTechnology
@IBMTechnology 2 ай бұрын
RAG is generally more cost-efficient than fine-tuning because it limits resource costs by leveraging existing data and eliminating the need for extensive training stages.
@scycer
@scycer Күн бұрын
True, but over scale I would assume the excessive context needed to be provided at runtime via RAG to answer questions may outweigh the initial cost in fine tuning no? Like it seems fine tuning is an investment choice in training costs while RAG is an ongoing cost of additional tokens. Obviously in relation to context that can be train ahead of time.
@bharathYerukola-gt7vt
@bharathYerukola-gt7vt 2 ай бұрын
Make a vedio on termonolgioes are often used on ai like benchmark and art of the state and etcc ❤
@steveyy3567
@steveyy3567 Ай бұрын
clear clarificaion, great job!
@davidrivera2946
@davidrivera2946 Ай бұрын
Like Developer i was working and create system with RAG patern and its fine, but have problems something with specefic documents, i mean when you play with tons of documents the RAG system get more complex, and you depend in strength way of prompts, im not yet play with fine-tunning but is something near to do for experiment, nice video thanks
@shahraanhussain7465
@shahraanhussain7465 15 күн бұрын
How would i get to know which model is using RAG in it or Not?
@bharathYerukola-gt7vt
@bharathYerukola-gt7vt 2 ай бұрын
Nice vedio and also make a vedio on neural networks in deep like how neiral network is interlinked with deep learning and machine learning and what is actaully neuarl network and architecuture and why architectute is inporatnt fir neural networks and what is neural network actalkuy like a technique or mathematical expression or anything else so make a vedio on all these
@giovannispillo5176
@giovannispillo5176 Ай бұрын
Fantastic Technology for value Great Lesson
@IBMTechnology
@IBMTechnology Ай бұрын
Glad it was helpful!
@user-rw6iw8jg2t
@user-rw6iw8jg2t Ай бұрын
Yes that's true LLM' are generalistic by default.
@ggggdyeye
@ggggdyeye 2 ай бұрын
sir can you tell me how to make the vectorstore and store it in a specific file to use it every time.
@youssefsayed4378
@youssefsayed4378 28 күн бұрын
Use Case: if i have a huge online library of books and i need to use llm to answer questions based on these books and research papers i guess we will use RAG but the point is can use it with a really HUGE amount of data (books and PDFs) and what if there multiple answers for the same question but from different resources and each resource has its own opinion which could be in a different direction than the other resource. what will happened?
@nickbobrowski
@nickbobrowski 26 күн бұрын
Great use case, Youssef! When you use RAG, it provides the model with multiple snippets of documents from your database. It's important to adjust the chunk size and the number of snippets injected into the context along with the user prompt. Typically, what I do with my clients is start with creating a set of evaluations for the system. These look like example prompts and example outputs. Any change I make to the system - I always run evals to see if the performance improves or gets worse. Once we have evals that measure how close the actual outputs are to the target outputs, we can work backwards and optimize the chunk size and number of snippets provided to the LLM. This way, it will get a balanced selection of relevant documents from your database. In some cases, it requires careful engineering to write proper search queries. Finally, the way the model writes the final response based on the retrieved information can be steered by instructions and fine-tuning. If you're interested in AI Development, feel free to contact me!
@BibekMishra84
@BibekMishra84 10 күн бұрын
I have a question. Is the LLM retrained on the new information during fine-tuning ?
@jignareshamwala3401
@jignareshamwala3401 3 күн бұрын
Yes, LLM model is retrained for fine-tuning. For efficient fine-tuning check out Parameter-efficient fine-tuning (PEFT). In PEFT a small set of parameters are trained while preserving most of the large pre trained model’s structure, PEFT saves time and computational resources.
@infotainmentunlimitedbyrohit
@infotainmentunlimitedbyrohit 2 ай бұрын
Thank you 🙏💛
@andiglazkov4915
@andiglazkov4915 2 ай бұрын
Thanks ☺️
@IBMTechnology
@IBMTechnology 2 ай бұрын
You're welcome!
@rfflduck
@rfflduck 2 ай бұрын
Great video!
@IBMTechnology
@IBMTechnology 2 ай бұрын
Thanks for the visit
@Siapanpeteellis
@Siapanpeteellis 2 ай бұрын
What happens to a model when it is fine-tuned? do you use a database for RAG?
@cloudnativecedric
@cloudnativecedric 2 ай бұрын
Good question! So with fine-tuning, using an approach like PEFT (Parameter-Efficient Fine-Tuning) which only updates a subset of the full model's parameters, we have new model weights and biases, which could then shared, deployed on a server, etc. for model inferencing with AI-enabled applications. For RAG, yes indeed, the most common method is with a vector database and turning your data into embeddings to search for similarity when using the LLM. But, there's other ways of setting up RAG pipelines too :)
@jasonrhtx
@jasonrhtx 2 ай бұрын
@@cloudnativecedricWhen would it make sense to first use PEFT, then apply RAG? Do both PEFT and RAG assign/label semantic relationships to the texts of user-added corpora and store these in a graph database?
@salehmir9205
@salehmir9205 Ай бұрын
this is gold
@cho7official55
@cho7official55 2 ай бұрын
I thought the retriever was on the far right, and llm in the middle of both, was I wrong, partially, is that schematic representation doesn't fathom all of the architecture, I'd like to go deeper on that matter.
@cloudnativecedric
@cloudnativecedric 2 ай бұрын
There are a lot of variances with the RAG approach that can lead to different architectures, but there's a full video on the IBM Technology channel that dives into RAG as well!
@ElaraArale
@ElaraArale 2 ай бұрын
Thank you~!
@IBMTechnology
@IBMTechnology 2 ай бұрын
You're welcome!
@hi5wifi-s567
@hi5wifi-s567 2 ай бұрын
Using “Fine Tuning” , then machine ( accounting software) can be a bookkeeper to prepare financial records for …?
@cloudnativecedric
@cloudnativecedric 2 ай бұрын
Just some ideas from the top of my head for fine-tuning with financial records are preparing financial statements, tax preparation (fine tuning on region-specific tax rules and historical data), expense tracking & categorization, etc.
@memehub2002
@memehub2002 2 ай бұрын
cool
@ridwanajibari4443
@ridwanajibari4443 Ай бұрын
so the concept of rag is like you attach file in gpt and asked question based on the attached file. isn’t it?
@Criszusep
@Criszusep 2 ай бұрын
Euro 2024 World Championship. Nice... of course the LLM could't give a response 😂
@umakrishnamarineni3520
@umakrishnamarineni3520 2 ай бұрын
The RAG isn't updated with new tournament 😂😅
@Datasciencewithsheddy
@Datasciencewithsheddy 8 күн бұрын
You’ll always have to use a combination of both RAG and FT.
@scycer
@scycer Күн бұрын
Kinda, we will always have to choose the ideal model for the use case (off the shelf or finer tuned) and what context is provided to the model (rag and other data). Really, it's all about context, whether it's engrained in the model or added as part of the prompt.
@GG-uz8us
@GG-uz8us Ай бұрын
I would like to see a real app that is in production with RAG and fine-tuning.
@atanasmatev9600
@atanasmatev9600 2 ай бұрын
Large Language model is "LMM"?
@cloudnativecedric
@cloudnativecedric 2 ай бұрын
Whoops! Good catch, sometimes I mess up when speaking and writing at the same time, it should be “LLM”.
@harryli7557
@harryli7557 Ай бұрын
Large Manguage Model! 2:08
@RajeshKumar-sz6ef
@RajeshKumar-sz6ef 29 күн бұрын
You did not tell about cost difference :)
@Robert-zc8hr
@Robert-zc8hr 28 күн бұрын
Obviously you need both, duh? No, seriously, they are not mutually exclusive. Fine tuning is learning, RAG is gathering requirements for a specific project. An expert needs to do both, he needs to learn in order to specialize, and he needs to be able to gather information for the specific task at hand.
@anshumanpatel9212
@anshumanpatel9212 23 сағат бұрын
Did he just write “LMM”, instead of “LMM”?
@SandraGarcia-t1k
@SandraGarcia-t1k Ай бұрын
Garcia Michelle Thompson Scott Martinez Ronald
@byankko
@byankko 2 күн бұрын
wtf
@SandraGarcia-t1k
@SandraGarcia-t1k Ай бұрын
Garcia Kimberly Lopez Karen Hall Mark
@einjim
@einjim 2 ай бұрын
So, You are all told to wear your watch on your right hand right?!
@SandraGarcia-t1k
@SandraGarcia-t1k Ай бұрын
White Deborah Wilson Susan Garcia Cynthia
@JamilaJibril-e8h
@JamilaJibril-e8h 2 ай бұрын
Uhhh okay i see you .....😂😂😂
@Zaid-st6wn
@Zaid-st6wn Ай бұрын
LMM lol
What are AI Agents?
12:29
IBM Technology
Рет қаралды 670 М.
How to Improve LLMs with RAG (Overview + Python Code)
21:41
Shaw Talebi
Рет қаралды 76 М.
HELP!!!
00:46
Natan por Aí
Рет қаралды 75 МЛН
Disrespect or Respect 💔❤️
00:27
Thiago Productions
Рет қаралды 41 МЛН
ТЫ В ДЕТСТВЕ КОГДА ВЫПАЛ ЗУБ😂#shorts
00:59
BATEK_OFFICIAL
Рет қаралды 4,2 МЛН
What are Word Embeddings?
8:38
IBM Technology
Рет қаралды 17 М.
Transformers (how LLMs work) explained visually | DL5
27:14
3Blue1Brown
Рет қаралды 3,7 МЛН
Fine-Tuning Large Language Models (LLMs)
1:16:12
Oren Sultan
Рет қаралды 2,3 М.
Prompt Engineering, RAG, and Fine-tuning: Benefits and When to Use
15:21
Llama: The Open-Source AI Model that's Changing How We Think About AI
8:46
What is LangChain?
8:08
IBM Technology
Рет қаралды 250 М.
RAG Explained
8:03
IBM Technology
Рет қаралды 114 М.
What is Prompt Tuning?
8:33
IBM Technology
Рет қаралды 219 М.
What is Agentic RAG?
5:42
IBM Technology
Рет қаралды 28 М.
EASIEST Way to Fine-Tune a LLM and Use It With Ollama
5:18
warpdotdev
Рет қаралды 131 М.