Finally, our requests were heard! The models seem pretty decent. Can’t wait to check them out for myself. And I just hope that with a little patience we’ll eventually get more context length
@inout33948 ай бұрын
On Reddit say, Ollama have Llama 3 ready to download and run
@NithinPrabhu938 ай бұрын
Kudos to the great work you are doing ! Big fan of localGPT !
@engineerprompt8 ай бұрын
Thank you 😊
@wasteid12798 ай бұрын
Dude there was once a time I used to get news from Google about the latest tech updates. But now you have replaced Google for me 🤣 Great work 💯
@engineerprompt8 ай бұрын
haha, that's the best compliment I have gotten :)
@bigglyguy84298 ай бұрын
I've found the 8B model is not censored against ERP
@engineerprompt8 ай бұрын
are you controlling via system message? I heard the same and want to test it out.
@unclecode8 ай бұрын
Impressive for "Sally" example 🤩! Haven't seen this in other open models, right? Have u seen? Other models answer 2, then when u bring up the objection they understand they can't take it from context.
@engineerprompt8 ай бұрын
Yup, it's smart, I tested wizradlm (last video on the channels), and had to remind it
@unclecode8 ай бұрын
@@engineerprompt and less verbose in comparison with wizardml. I saw ur other video and tweeted about it. Crazy time.
@seeowltv8 ай бұрын
Thank you for your video. Could you add your script for forigners?
@engineerprompt8 ай бұрын
Thank you, good idea. Will start doing that
@VerdonTrigance8 ай бұрын
What's about new Stable Diffusion 3? Is it connected to Llama 3?
@engineerprompt8 ай бұрын
there is an image generation model on meta.ai, figured that out later but not sure if its based on stable diffusion.
@binaryvat8 ай бұрын
What is Llama 3?
@HammadShah7128 ай бұрын
Llama 3 is a large language model which is trained on large amount of data on many GPUs and training take weeks and months and right now it is best open source large language model
@holdthetruthhostage8 ай бұрын
Im waiting for 8x70b Mixture Of Experts
@engineerprompt8 ай бұрын
Actually its interesting that Meta didn't go for a MoE, even the 400B version seems to be a dense model not an MoE.
@GoldenkingGT1018 ай бұрын
Waiting for local gpt integration of llama3 8b
@engineerprompt8 ай бұрын
that's coming soon. I will have a busy weekend :)
@renierdelacruz46528 ай бұрын
Great Video, thank for sharing
@engineerprompt8 ай бұрын
Thanks 😊
@drgutman8 ай бұрын
8k context window. not impressed. also I think they've compared the 8B with mistral 7B o.1 not the new o.2.
@engineerprompt8 ай бұрын
I agree but its really impressive that you can train an 8B model for 15T tokens. The scaling laws go brrr.......