NEW: LoRA Models override Pre-trained Knowledge (MIT)

  Рет қаралды 2,605

Discover AI

Discover AI

Күн бұрын

Пікірлер: 14
@SashaBaych
@SashaBaych 13 күн бұрын
My favorite data science youtuber these days! Thank you So many channels now are about pure hype of delivering AI news with no substance... But you are an inspiration. Damn I want to read a paper a day at least now!
@code4AI
@code4AI 9 күн бұрын
Do it! Smile ....
@deter3
@deter3 11 күн бұрын
how do you measure model's generalization capability , this is a really fuzzy and vogue concepts and we keep using it while do not have clear measurement .
@monologtr_
@monologtr_ 13 күн бұрын
hows fine tuning vision low mıdel with ocr, vqa custom datasets
@rikhendrix261
@rikhendrix261 13 күн бұрын
What determines if the task is the same? Is it the instruction prompt? And what defines the size of a task which is correct for LoRA?
@novantha1
@novantha1 12 күн бұрын
Your intuition, basically. It’s tricky because some tasks will be in distribution, even when dealing with unique data, while some tasks will explicitly not be in distribution. Here’s a couple of things to consider: For simple math, let’s say addition, subtraction, multiplication and division, do you think that a new equation outside of the example equations is in-distribution or out of distribution? For logical reasoning problems, do you think that a problem with a similar structure to a problem in the training set is in distribution or out of distribution? For creative writing, do you think that a model being asked to write stories in the same genres as the training examples is in distribution or out of distribution? It gets really nuanced, and I think the only way to really understand this is to approach them on a model-by-model and dataset-by-dataset basis.
@vladimirnadvornik8254
@vladimirnadvornik8254 13 күн бұрын
If I understand it correctly, then doing full fine tuning and running SVD on the difference between the finetuned and the original model would create a LoRA that does not suffer from this problem. Is it correct?
@EvanGoodwin-bl7zq
@EvanGoodwin-bl7zq 12 күн бұрын
Could you train LORA's at different ranks, scaling up and measuring performance? Then when you reach an acceptable level of performance you determine - or improvement falls below a certain level - you stop the process. It might involve some upfront costs, but I assume you would save on inference down the line because the 'acceptable' LORA would be computational more efficient than the full trained model. It would depend on the use case. If you are doing lots of inference, it would definitely payoff down the line. It would be interesting to see the costs of training multiple LORA's in this way vs full training.
@vladimirnadvornik8254
@vladimirnadvornik8254 11 күн бұрын
LORA is not more efficient for inference. Either you can merge the LoRA into the model, then it is exactly the same or you can compute the LoRA separately and then it is less efficient.
@EvanGoodwin-bl7zq
@EvanGoodwin-bl7zq 8 күн бұрын
@@vladimirnadvornik8254 Ok, then perhaps a better approach would be to train a LORA on different model sizes - 1B, 3B, 8B (which are computationally more efficient - and stop when acceptable accuracy is reached or improvement falls below a certain level.
@jonmichaelgalindo
@jonmichaelgalindo 12 күн бұрын
What about undertraining loras on each block and merging as you go? You update all the parameters, and no single lora "overpowers" the original data vectors.
@code4AI
@code4AI 9 күн бұрын
??? If you "undertrain" a fine-tuning mechanism, then you have a broken fine-tuned weight tensor structure. Why merge something that is not working into the pre-trained model?
@NLPprompter
@NLPprompter 13 күн бұрын
I'm guessing Lamini AI company doing something like this to achieve what they said better than RAG...
NEW 3.5 SONNET V2 Has a LOGIC BUG: Reasoning ERROR
12:48
Discover AI
Рет қаралды 2 М.
Decoding on Graphs: Empower LLMs with KGs (MIT)
29:11
Discover AI
Рет қаралды 5 М.
Человек паук уже не тот
00:32
Miracle
Рет қаралды 4 МЛН
Amazing remote control#devil  #lilith #funny #shorts
00:30
Devil Lilith
Рет қаралды 15 МЛН
Каха и лужа  #непосредственнокаха
00:15
Wait for it 😂
00:19
ILYA BORZOV
Рет қаралды 11 МЛН
Inside the VLM: NEW "Task Vectors" emerge (UC Berkeley)
22:37
Discover AI
Рет қаралды 2,4 М.
LoRA explained (and a bit about precision and quantization)
17:07
We Fell For The Oldest Lie On The Internet
13:08
Kurzgesagt – In a Nutshell
Рет қаралды 4,3 МЛН
NEW AI Models Solve Hamiltonian Mechanics
21:52
Discover AI
Рет қаралды 3,7 М.
How Physicists Broke the Solar Efficiency Record
20:47
Dr Ben Miles
Рет қаралды 690 М.
New VLM Post-Training Layer Scaling: No more Forgetting #ai
20:54
Discover AI
Рет қаралды 1,8 М.
NEW Knowledge-Graph Adaptive Reasoning: Plan-on-Graph LLM
23:33
Discover AI
Рет қаралды 4,9 М.
AI Prompt Programming: Beyond DSPy & TextGrad: PDL
32:09
Discover AI
Рет қаралды 2,6 М.
Человек паук уже не тот
00:32
Miracle
Рет қаралды 4 МЛН