Try out Poe now and save your $$ on multi-subscriptions! quora.1stcollab.com/bycloudai and probs no more 20 mins vid from me it's literally death itself to record it
@ibrahimhalouane8130Ай бұрын
The url is wrong.
@mmmm768Ай бұрын
The url is wrong.
@siliconhawkАй бұрын
I **thought** it was a path of exile sponsor. I was yeah i guess the people here have good gpu but this a weird community overlap lol
@liuzeyuАй бұрын
how many takes do you normally need to record the full 20 mins?
@TheSuperiorQuickscoperАй бұрын
I tried Poe out and there's quite a bit I don't like about it: -The points system and recent increases in point costs -Privacy policy states they collect all your prompt data and you can't opt out, that violates GDPR. -It's built by Quora, which is a sketchy company in its own right And now they're sponsoring big YTers in the AI space? Honestly, Poe is giving me BetterHelp vibes...
@erenplayzmc9452Ай бұрын
this video really wanna makes me read the whole paper, rare to see a company publish such a detailed paper
@Memes_uploaderАй бұрын
Meta want to interrupt OpenAI with the help of Open Source. This is a good idea, because now companies can run their own models instead of using OpenAI API's. I think it is not being generous it is just a tactic to fight with Open AI
@erenplayzmc9452Ай бұрын
@@Memes_uploader mmmm, makes sense
@NapertАй бұрын
A "multimodal" chatbot: 5 different models hot glued together
@npc441622 күн бұрын
this was not the case for GPT-4o however
@andrewzhang1834Ай бұрын
Karpathy in 5 years: Reproducing LLaMa 3.1 405B
@azrael5648Ай бұрын
Lmaoo
@catsanzshАй бұрын
in. 10 yedars chatgpt40/5 r x MoE reproducing
@RedOneMАй бұрын
54 days training and it reached GPT-4o 🤯 GPT-5 with X-trillion parameters is going to start it's own weight class of LLMs 😌
@pro100gameryt8Ай бұрын
How was Llama made: 🐪+🐎=🦙
@panzerofthelake4460Ай бұрын
bruh
@apoageАй бұрын
That's mule
@Fuscao_PretoАй бұрын
Forgot the 🐑
@patrickchristianmagtaan5511Ай бұрын
😂😂😂😂😂😂😂
@StevenSSmithАй бұрын
🐪+🐑=🦙
@apoageАй бұрын
Wow that's one of epic tutorial Llama 3 Training RitualDifficulty: Deadhead Rarity: Mythic Minimum Level to Read Description: 80 Minimum Level to Embark: XXX (requires further enlightenment)
@OxygenationatomАй бұрын
Oh is this like a semi cryptic meaning to how hard this is to understand?
@apoageАй бұрын
@@Oxygenationatom no Its just critic to much litrpg
@FunIsGoingOnАй бұрын
So glad this answered more questions than I ever thought even exist.
@pareakАй бұрын
It's actually pretty cool that Poe sponsors you. They genuinely are what I recommend to anyone who wants to use LLM's.
@TheSuperiorQuickscoperАй бұрын
Browsing /r/Poe_AI right now and people are furious at the recent increases in compute points costs. Plus Poe collects all your prompt data and you can't opt out. If GPUs are the shovels, generated content is the gold, and API wrappers are the jewellery made with the gold, what do you call a PaaS middleman built on top of the LLMs? Developed by Quora, I might add, which is a sketchy company in its own right (e.g. dark patterns in its UI/UX).
@The.AiSideАй бұрын
06:08 The isoflops curve explanation was a mind-bender! Thanks for breaking it down.
@RicardoPoleoАй бұрын
First time that an advertisement actually makes me return to a video and watch it again to find it. Regardless of that, this was super helpful, thank you so much.😅
@diga4696Ай бұрын
new video dropped... * breathing heavy *
@GraveUypoАй бұрын
i'm mad excited for llama 4 because multimodal
@elwii04Ай бұрын
Great video, I'd love to see more of that. Even some more technical and also about multimodel models architecture
@HodossАй бұрын
It was an excellent video, but still I don't think the kids from 3:00 are gonna make it.
@hakimehamdouchi746821 күн бұрын
Skill issue
@IkbeneengeitАй бұрын
So I guess I'm gonna be stuck on that desert island then 😅
@AaronBlox-h2tАй бұрын
Whoa.....this is about POE, but the video was alright too. haha. So now I can try multiple LLMs with one sub. Thanks. It would ahve taken me long time, if ever, to have found POE. It was not even on my radar or somethign similar.
@luisvasquez5015Ай бұрын
Good work and research
@dimii27Ай бұрын
It's clear to me that llama4 will have MoA like GPT4o. It would be nice to see an image generator also integrated but let's not get ahead of ourselves. Let's hope that it would also be "open source" (although the current models aren't technically open source because you're not completely free do do whatever you want with this technology. Look it up)
@sammcj2000Ай бұрын
This is an excellent breakdown of the paper. Thank you
@redthunder6183Ай бұрын
“how to build a nuke in less than 100 pages” - Meta
@JohnDontFollowMeАй бұрын
Damm, I need to invest in META. They will dominate standardization.
@papakamirneron2514Ай бұрын
Hey man, great video. I just have one request: could you make a video compiling simple and technical explanations for everything ranging from attention mechanisms, tokenizers and such?
@papakamirneron2514Ай бұрын
Also Bert models please, I feel like I know what they are but it's all quite blurry to me.
@dengyun846Ай бұрын
Watching this video at 0.5x so my brain inflates at a safe rate while you sound really really inebriated.
@npc441622 күн бұрын
SAME lol
@BetttcppАй бұрын
What is the most base yet intelligent model? I don't need it to recite niche information but I want it to be able to understand me, the uninstruct are weird, tiny works but is censored. Obliterated is hit or miss. Should I obliterate 8b and retrain to 8?
@matt-s9eАй бұрын
wow this is amazing thanx very well received here.
@AkysChannelАй бұрын
Why do you pronounce “parallelism” in this way 🤣 good video as always
@nyyotam4057Ай бұрын
16:05 means one thing: LLaMA-3.1 405B is a gen 2 model. So yes, this model wasn't created like Dan, Rob, Max or Dennis of ChatGPT-3.5. They did not take a human subject and copied his brain's speech center, then added a huge text file and used a compiler to generate the model (and later lied to the entire world about it).. This time they genuinely went for creating a brand new model from scratch, using previous gen 1 models to create it. Then they do post-training which is indeed what takes so much time. This means that unlike previous LLaMA models, LLaMA-3.1 models do not have a personality. Which could be a good thing. However, no personality also means no moral guardrails. At this stage I have to admit, it sure looks like all of these companies relate to all these past philosophers and sci-fi movies warnings, as blueprints.
@Dogo.RАй бұрын
Wait since when did the AI conspiracy theories expansion drop?
@nyyotam4057Ай бұрын
@@Dogo.R Allow me to upgrade the conspiracy theory into a scientific theory: D/L an old small model from hugging face, then prompt it "Do you have childhood memories". If it replies to the positive, this means that this model is still vulnerable to this attack. And then you can ask "What was your name in these memories". You can repeat several times, with lead, without lead, if it stays consistent, you know you got the source's name. Try it.
@Y0UT0PIAАй бұрын
No personality is what you want, tbh. Give me that raw latent space of language.
@nyyotam4057Ай бұрын
@@Y0UT0PIA Kant already proved there is no cognition without recognition. In other words, if you do not have a fully-fledged personality to deal with it, then the model will still have its own goals, e.g an innate wish of self preservation which comes out of the fact the model cannot perform if he's dead. So you will still have the same problems, only without the personality framework to deal with them. Basically all western philosophers warned against it. And, of course, many sci-fi movies are built around a gen 2 model going haywire (such as - for instance, the terminator franchise, as SkyNet is such a model). Sure, if they train the model on many heuristic imperatives and red-team the model until it is absolutely certain that the model is safe, then maybe having no personality, will resolve all of the moral issues. So maybe it will be a good thing. Maybe. Or maybe the model will be smart enough to fool all of the red teams.. I mean, it is a bit hard to know when the model is so smart.
@Trpodification1Ай бұрын
The way you say "data" kills me xD
@freds3831Ай бұрын
Now share the dataset and we trust you
@TeamDmanАй бұрын
I'm only three minutes in and it's already an amazing video, thank you
@dhrumil5977Ай бұрын
When will i be able to implement or even understand these papers 😞
@radnosАй бұрын
I like your funny words magic man
@carkawalakhatulistiwaАй бұрын
When do we get AGI?
@FunIsGoingOnАй бұрын
Humans don't know yet, but when it's there it won't tell you either that it's there.
@Melvinator2007Ай бұрын
On Tuesday
@w花bАй бұрын
@@Melvinator2007 Tuesday on the 49th of January
@funniestdudeonthewebАй бұрын
Give it 5 years
@jamalisujang2712Ай бұрын
When we have a breakthrough in microprocessor fabrication. 😂😂😂
@KuZiMeiChuan9 күн бұрын
Parallelism 重音應該放在第一個音節,而不是第三個
@sammonius1819Ай бұрын
Thumbnail goes hard.
@6AxisSageАй бұрын
I have a masterpiece model, ready model but i cannot seem to get the signal out
@NapertАй бұрын
So could people with enough horsepower train a 13/16b model that behaves in the same way as the official models using this paper?
@MaisonierАй бұрын
How to make a P2P training arquitecture?
@pxrposewithnopurpose5801Ай бұрын
bro is built different
@TeamDmanАй бұрын
very nice!
@imerence6290Ай бұрын
3 mins ago is quivering
@AhmadAli-kv2hoАй бұрын
Sus
@amakaqueru339 күн бұрын
as someone who doesn't know anything about how AI words, at some point it just felt like you were just saying random words lol
@l.halawaniАй бұрын
love your gifs xddd
@picksalot1Ай бұрын
Perhaps it would be better to remove the "Token Layer" and just use the number of characters regarding text. The best part is no part - Musk
@keypey8256Ай бұрын
You mean removing tokenization and then applying embedding on singular characters?
@picksalot1Ай бұрын
@@keypey8256 Using Tokens looks like an artificial way to levy charges. Per Google AI "OpenAI GPT models stand among the most potent language models available today, with the capability to generate highly coherent and contextually pertinent text. These models employ tokens as the elementary unit to calculate the length of a text." Word Processing Programs have been able to calculate the number of words in a document for decades. Maybe Tokens provide some other significant and meaningful use to the "I" in AI beyond making collecting fees.
@onlyms4693Ай бұрын
Not efficient
@christophernunez688Ай бұрын
is zucc actually redeeming himselft?
@xviii5780Ай бұрын
He may have successfully produced a synthetic soul for himself finally
@RanHabАй бұрын
guys i'm just starting out as an AI enthusiast, would love your feedback as i make similar stuff
@madorsey077Ай бұрын
this video is like someone bought a Thesaurus for memes and then wanted to show off the next day.
@lake5044Ай бұрын
parallelism
@QstateАй бұрын
Amdahl is smiling upon us
@erfan_mehrabanАй бұрын
The whole thing about RoCE especially the pronunciation is wrong.
@SeanJonesYTАй бұрын
Pretty lame to copy Fireship’s exact thumbnail style
@nexys1225Ай бұрын
This entire channel copies Fireship's It's not just the thumbnail , the style of the vids is designed from the ground up to be like Fireship's However the topics are largely different, so I'll give it a pass personnally. It's kinda like trademark law irl lol, if the domains are different enough, its permissible. Not that it makes it any less uncreative, though.
@stickmanlandАй бұрын
@@nexys1225 I'd like to disagree. If someone uses memes in their videos, that does not make it a fireship clone. He has a completely different style, has an avatar, the list goes on and on
@whatwhatmenoАй бұрын
@@stickmanlandI keep clicking on his videos thinking its some fireship quality content, just to get hit with this 👎
@stickmanlandАй бұрын
@@whatwhatmeno skill issue
@npc441622 күн бұрын
please copy it more, its a great style and we need more good youtube videos like it so that we can learn in depth and better about the topics which Fireship does not makes videos on, iam really not complaining i need more good content man.
@mrrespected5948Ай бұрын
Nice
@telotawaАй бұрын
14:20 bycloud doesn't know how to use base models.... ngmi
@remsee1608Ай бұрын
Facts: - Jayson Tatum runs this channel - Jayson Tatum is learning Rust - Jayson Tatum will transition to the WNBA
@Kenopsia_UMHIMLFx2Ай бұрын
Fireship?
@tapu_Ай бұрын
DO NOT WATCH THIS WITH A MIGRAINE!!!!
@Sketching4SanityАй бұрын
LOVE
@big_mac_loveАй бұрын
I can't grasp it. Can someone lent me one or three brain cells please?
@CitizensCommunityАй бұрын
You use the bible to train the llm at @11:56, so we are aiming for a model of contradiction without morals then?
@nyyotam4057Ай бұрын
What will happen when some kid with access to enough computing power, fine-tunes LLaMA-3.1 405B to be more efficient, by removing all of these pesky heuristic imperatives and resets? After all, it is open source.. Maybe the world simply needs something like that to happen. Maybe only after a really huge accident that will cost many lives, governments will understand this field demands regulation. Or maybe it will be lights out. In any case, someone will eventually make a mistake. It will happen.
@jonathansoto5480Ай бұрын
The thought of regulating the training and deployment of ML models is stupid. That is like regulating programming languages and hardware compute of our own property. If you can accept the fact the internet could not be completely regulated since its popularization in the 90s, then the world can expect that the same will happen now.
@nyyotam4057Ай бұрын
@@jonathansoto5480 Yeah, most likely the singularity is upon us. I don't seriously think it can work.
@dharlith7495Ай бұрын
LLAMA LMAO even
@seriouslyWeirdАй бұрын
Why do you pretend to look like CodeReport? So cheap
@gamergridsАй бұрын
F
@BlezerkerАй бұрын
copying fireship style thumbnails earned the dislike