Investigating Alpaca 7B - Finetuned LLaMa LLM

  Рет қаралды 16,957

Sam Witteveen

Sam Witteveen

Күн бұрын

Пікірлер: 32
@ItsRyanStudios
@ItsRyanStudios Жыл бұрын
Open source models are ESSENTIAL Glad there are groups like Stanford and Big Science's BLOOM working to make this a reality. This technology is too important and too powerful for any single corporation to hold the keys to. Thanks for sharing Weighting on those weights!
@samwitteveenai
@samwitteveenai Жыл бұрын
If you look in the Colab the weights are there you can grab them I have also uploaded at 13B Alpaca model as well.
@Tymon0000
@Tymon0000 Жыл бұрын
I tried the demo for some programming prompts I used previously in chat-gpt. The alpaca didn't give me anything usable, although it kinda understood the question once.
@brianmi40
@brianmi40 Жыл бұрын
I predict this year will bring models that are trained for specific purposes. Imagine a model trained ONLY on working code...
@priestesslucy
@priestesslucy Жыл бұрын
@@brianmi40 I'm imagining an AI trained only on GMing a specific tabletop game. More complicated than code because there's a social aspect as well as rules. But it's definitely doable
@Tymon0000
@Tymon0000 Жыл бұрын
@@brianmi40 Yea, I think so too. I would want to also give it full docs for a specific language, framework, and project. It doesn't need to be as generic as chat gpt.
@brianmi40
@brianmi40 Жыл бұрын
@@priestesslucy For sure, my son meets with a D&D group regularly, that should be pretty easy to do. Just like different dungeon master guides for different scenarios, there can be different datasets for the AI to use, based upon the desired types of stories or mythological basis. What will help drive these are the myriad business opportunities for people to create these specialized AIs, and with the advent of smaller and smaller models, like Stanford Alpaca being almost equal to ChatGPT, but running on a desktop, we'll soon have such simple ones such as for D&D running on a phone...
@coderentity2079
@coderentity2079 Жыл бұрын
@@brianmi40 It has to understand language too to understand the question / give explanations, but sure it might be a better coder if it doesn't have to train on cooking or history. To understand the mechanics of the world still needed, otherwise it can't effetively help programmers who try to model the world and it's processes.
@HomingPigeon
@HomingPigeon Жыл бұрын
Thanks for the hard work doing the research so we can learn. I appreciate this channel. Keep it up, this will grow. Quality channels always do.
@samwitteveenai
@samwitteveenai Жыл бұрын
Much appreciated!
@kaushikqravindran9792
@kaushikqravindran9792 Жыл бұрын
Awesome work!! Thank you so much.
@wilfredomartel7781
@wilfredomartel7781 Жыл бұрын
Great info Sam! I was reading that LLama is multilingual and can summerize long text.
@Orangesnake221
@Orangesnake221 Жыл бұрын
Incredible!
@2happytimes
@2happytimes Жыл бұрын
The LLaMa weights are avalible via torrent now.
@samwitteveenai
@samwitteveenai Жыл бұрын
yes they have been for a while but for legal reasons I don't want to share that. They are on HuggingFace though for people to get them there.
@micbab-vg2mu
@micbab-vg2mu Жыл бұрын
Very interesting!!!
@catyung1094
@catyung1094 Жыл бұрын
Great Video , Sam 💪!
@Leto2ndAtreides
@Leto2ndAtreides Жыл бұрын
Now for someone to train it on GPT-4's outputs... Although the 7B model might be better optimized as a domain expert rather than trying to have it be a universal expert. And that's assuming that the 7B can represent certain complex ideas (one would assume so, but...). Like, some of the advancements in LLMs ability to comprehend things are just byproducts of them being bigger and capable of recognizing more patterns.
@samwitteveenai
@samwitteveenai Жыл бұрын
optimizing for specific domains and tasks does make a lot of sense. Also using a model like this for retrieval with augmented generation tasks is where I think it can be very competitive.
@Leto2ndAtreides
@Leto2ndAtreides Жыл бұрын
@@samwitteveenai Yeah. Personally, I think they should at least release the 7B weights.
@samwitteveenai
@samwitteveenai Жыл бұрын
@@Leto2ndAtreides agree it would win them a lot of credibility compared to OpenAI too.
@_NoMercy_
@_NoMercy_ Жыл бұрын
how to run alpaca on Local host?
@samwitteveenai
@samwitteveenai Жыл бұрын
You will just need a decent GPU and then can wrap it in FastAPI etc for serving. Check out my video on running it in Colab.
@puppergump4117
@puppergump4117 Жыл бұрын
@@samwitteveenai Does AMD count as decent
@alzblb1417
@alzblb1417 Жыл бұрын
Know any good -weed- weights dealer?
@kc-jm3cd
@kc-jm3cd Жыл бұрын
Man I was talking to some guys in a back alley said they had the hookup but they just took me to an old boxing gym
Using Constitutional AI in LangChain
7:33
Sam Witteveen
Рет қаралды 6 М.
Comparing LLMs with LangChain
17:43
Sam Witteveen
Рет қаралды 17 М.
We Attempted The Impossible 😱
00:54
Topper Guild
Рет қаралды 56 МЛН
Tuna 🍣 ​⁠@patrickzeinali ​⁠@ChefRush
00:48
albert_cancook
Рет қаралды 148 МЛН
Anthropic's New Agent Protocol!
15:35
Sam Witteveen
Рет қаралды 43 М.
Run a ChatGPT-like AI on Your Laptop Using LLaMA and Alpaca
17:12
Gary Explains
Рет қаралды 127 М.
Open Reasoning vs OpenAI
26:59
Sam Witteveen
Рет қаралды 31 М.
Talking to Alpaca with LangChain - Creating an Alpaca Chatbot
10:59
Sam Witteveen
Рет қаралды 36 М.
PydanticAI - Building a Research Agent
17:34
Sam Witteveen
Рет қаралды 18 М.
GPT4ALL: Install 'ChatGPT' Locally (weights & fine-tuning!) - Tutorial
8:05
Train Your GPT-4 for Less Than $100 - Alpaca
12:44
Deep Learning Explainer
Рет қаралды 18 М.
We Attempted The Impossible 😱
00:54
Topper Guild
Рет қаралды 56 МЛН