Does alpha have any role during the fine-tuning process; or is it only used during merging to scale the lora-weights?
@ArjoRoy-pe6tf24 күн бұрын
This is the first, and probably the last video I would ever need to watch, to get the intuition behind LoRA, thanks to the inventor. 🫡
@MoonlitRitual26 күн бұрын
Thank you sir for creating the best thing since slice bread
@jmr646828 күн бұрын
You brought so much value by publishing this video, thanks !
@RalphDratmanАй бұрын
This is tremendously helpful.
@harriehausenman8623Ай бұрын
Anyone else thinking of Futurama's 'Brain Slugs' when hearing LoRa, or is it just me 😄
@harriehausenman8623Ай бұрын
Thanks for doing this video! An interesting 'check-point' in history, so to speak 🤗
@LBSbransАй бұрын
Pls share the new video with details!❤
@ernestofootfighterАй бұрын
Great teaser. I would love if you delved deeper into the theoretical side.
@dusky41512 ай бұрын
Im going to use a comparison but tell me if ive got the general idea right: Lets pretend im playing an MMORPG and I choose "Archer" as my class. A LoRA is like fine tuning the armor of my archer, or maybe his bow, so that he is more specialized for a particular battle. A full checkpoint fine-tuning however is like changing class from "Archer" to "Warrior"
@schurrle272 ай бұрын
Amazing explanation! Though expected when coming from the founder of course
@HadbbdbdDhhdbd2 ай бұрын
Really helpful and brief explanation. ty.
@AaronGoldman2 ай бұрын
Do we need the base model? Would it make sense to use a panel of experts? The final model just being the sum of many LoRA and leave them as decomposed matrices for cheaper matrix multiplies?
@saharshayegan2 ай бұрын
This was very helpful! Thank you, Edward!
@KiWelten2 ай бұрын
Thank you for all your work!
@umeranwaar2 ай бұрын
I am literally blown away by the quality of your explanation! I am a AI researcher myself, so I can really appreciate the beauty of explaining the technical concepts in "simple" language while not making it "simpler". 🙂
@AmanBansil3 ай бұрын
It’s not often that I find the inventor of a technique explaining the technique. This is incredibly helpful. Thank you
@nowcastthedieАй бұрын
It's even rarer than someone can actually do both effectively, very talented guy & great explanation.
@MichaelBrody-o8g3 ай бұрын
Katarina Bridge
@DanielMoore-s3j3 ай бұрын
Timmy Square
@DavidGreen-h2i4 ай бұрын
Larson Parkway
@phdperson4 ай бұрын
This is amazing and very valuable. Thank you!!!
@Jhonnyzilla4 ай бұрын
That is such a good explanation, thanks!
@EleaseNiebergall-e7g4 ай бұрын
Jeanne Station
@shklbor4 ай бұрын
Awesome explanation and kudos for a great contribution to DL, please make a followup video on QLoRA
@tuhinmailme5 ай бұрын
These things existed for a lot time in vision research. Like only finetuning classifiers of large models on new tasks
@jimshtepa54235 ай бұрын
thank you for a great presentation. I am new to llm and would like to try to run the code on github. is my local machine (macbook m1) can handle it? or is it something for large enterprises with massive compute inventory?
@sorooshsohangir5 ай бұрын
Great Job!!!
@EsZfW5f6 ай бұрын
Thanks!
@Krishna1729-z8v6 ай бұрын
I have worked on Markov chain Monte Carlo algorithm, it took me 1 hour to map the posterior distribution and that’s not even close….looking forward to use this Gflownets
@shibohao89307 ай бұрын
Great video! Looking forward to your video explaining the relation between GFN and Max-Entropy RL
@tonywang79337 ай бұрын
3:26 That is the best explanation!!
@tectract7 ай бұрын
Very cool. I know some of these words.
@redthunder61837 ай бұрын
Thank you so much for explaining this clearly, everything I watch on KZbin is made by ppl who have no idea how the tech works, or don’t even know how to code outside of copy/paste/change inputs, but pretend like they do. Furthermore, there’s just so many useless libraries around LLMs that ppl claim are the next big thing, but in reality, they create code bloat, introduce more unknowns, make the code harder to work with since u now gotta learn the library, and don’t work as well as if u just wrote everything urself.
@ph10m8 ай бұрын
This was a great intuitive explanation of it. I wish more people took the adaptability of lora seriously, though: everyone (and their dog) upload full models after doing small fine-tunes *with* lora, instead of just the adapters. Not only would it help experimentation, but time too, as we have to download unnecessary base models over and over...
@jett_royce8 ай бұрын
LoRA is such an unlock for resource-constrained creators looking to leverage models for specific domains. Thank you for this amazing work!
@houbenbub8 ай бұрын
Awesome video, thanks for making it :)
@BruceChar0078 ай бұрын
能不能继续在微调后的LoRA模型上面微调,效果怎么样
@lophyre13809 ай бұрын
Very informative video, but please get a better mic
@user-wp8yx9 ай бұрын
Trying to teach a mistral7b model sanskrit. It already has Sanskrit characters as tokens and is the best performing 7b llama based model I can find. You seem like a knowledgable person in this area. Do you have any advice for lora? Rank, alpha? How about targeting of q,k,v? Other strategies? I have about 3gb of datasets that range from translations, corpus, to data tables. I wonder if I should use different strategies for different data types?
@bobbyparikh56909 ай бұрын
Fantastic video Edward! In case someone wants a quick refresher on low-rank decomposition of matrices, here's a great video: kzbin.info/www/bejne/aKDKlaqmfalmjJo&ab_channel=ritvikmath
@arnoldpalmer-fv7pf9 ай бұрын
So much groundbreaking research broken down into an easy to follow 7 minute video, I love it 🙏
@nathangonzales-hess65699 ай бұрын
That was great. Thanks for sharing. I really appreciate the simple style, no distracting animated plots or fancy editing. Look forward to more!
@justinpresent9 ай бұрын
thanks edward for the gentle intro!
@ellielikesmath10 ай бұрын
i was trying to come up with something like this, in that i wanted to train a generator which would be the inverse of a classifier, and the classifier gave a score to how good a solution was drawn from some range. this looks miles and miles more sophisticated than what i was doing with tf and pytorch, but i definitely understand, at least on that level of abstraction, why such a development is necessary. i look forward to trying this, cheers.
@Bbb7865110 ай бұрын
Thank you so much for the video Edward. Its an inspo seeing you make videos and take off. Im currently a masters of science student in data science and Im always excited about NN architectures and new ML algos. Whenever easy, could you please share 1-2 tips for writing good research papers in ML? I recently started in a lab that does neuroscience-ML, and rlly want to make an impact there
@faizanjaved144310 ай бұрын
Hey there! Can we talk about Q*, the AGI developed by Sam Altman? I'm excited to discuss this with you, it's one of the most interesting topics for me after Sora.
@candrewlee1410 ай бұрын
This was fantastic! Thank you, it’s great to hear from a real expert in this AI mega-hype cycle.
@DB-Barrelmaker10 ай бұрын
The audio is terrible, it's noticeable when you're dealing with a complex subject containing alot of niche phrases
@DigitalAlligator10 ай бұрын
Shit, you invented LoRA 😮? How you come up with that idea works so good?!