Wanna build your own AI Startup? Go here: www.skool.com/new-society
@startingoverpodcastАй бұрын
Why aren't you using Msty?
@aaaaaaaaoooooooАй бұрын
Wait, my data is not private with o1? I didn't know that. Where can I check this? Where is this notified to the user, or did they bury it in small text?
@indiemusicvideoblogАй бұрын
Great! Now build a local agent with lama that can control your computer like Antropic
@orthodox_gentlemanАй бұрын
Very doable with Open-Interpreter which is open source and free
@BllakezАй бұрын
@@orthodox_gentleman How much should I pay someone to setup for me?
@alexrayoalvАй бұрын
I literally did this 6 months ago.
@anubisaiАй бұрын
You build it.😂
@marilynlucas5128Ай бұрын
Skyvern!
@samimejri8079Ай бұрын
I just used Llama 3.2 locally and asked about starting a 3d printing business as a 3D beginner. It gave a similar output of what you spent a good time building in this video... Maybe do it the next time, show a before and after response from an LLM.
@bossgd100Ай бұрын
😂
@DCinziАй бұрын
There is a model called Llama3.3B-Overthinker. I think it would fit the task quite nicely.
@JackGamerEuphoriaDevАй бұрын
Is there available in Ollama or hugging face? If you don't mind the question. Thanks by the way for giving directions..
@chrystofferaugusto1194Ай бұрын
Btw, the concept you reached in this video of undetermined number of agents is far superior than it was from a video from 5 days ago. Really awesome 👏🏻
@godned74Ай бұрын
You could try "When providing responses, use concise and primary representations. However, include additional details only when needed to ensure clarity and completeness of the task" and you should get short response's with out compromising the chain of thought.
@eviv8010Ай бұрын
nice clickbait
@kylev.8248Ай бұрын
It’s not clickbait tho
@bruce_x_offiАй бұрын
@@kylev.8248 You must be King of fools
@mihaitanitaАй бұрын
So, you've used Claude 3.5 (2024 october update) within Cursor AI Editor to develop a (simple) python script that run some agenting on a 70b model on ollama? Where's the o1 in here?
@DancoliioАй бұрын
o1 is a reasoning model which kept their reasoning 'recipe' private. This is his take (which resonates with the average user of locally owned open source models) to kind of hack the way the 70b model works and simulate reasoning to enhance the final output> a simple method which actually does provide better replies.
@BikramAdhikari89Ай бұрын
He is not sharing his research paper published in arxiv my man.
@MrMoonsilverАй бұрын
Cool new format with the presentation man
@bsiix1576Ай бұрын
Maybe I missed it, but what hardware is needed for that nemotron - it is 43GB? Doesn't that mean you need at least that much VRAM? And here I thought I was a baller with my 16GB vram...
@mariomanca7546Ай бұрын
If you instruct the agent to use the fewest possible lines, it's likely to eliminate comments, which is suboptimal but expected.
@foxusmusicus2929Ай бұрын
Great video. Which hardware specs do you have? :-)
@AK-ox3mvАй бұрын
How much you'r local O1 results has more accuracy in comparison to original nemotron 70b and llama 3 3b without uaing chain of thought? Was there any improvement in bechmarks like Humaneval and MMLU?
@Plife-507Ай бұрын
I want to build an agent swarm to do coin margined futures btc trading. With each agent handing a serpearte part, ta, market sentiment, execution, risk tolerance, is there a way to keep each model small and only train it to focus on its task?
@FrankDecker-n9eАй бұрын
@DavidOndrej, what is your Mac specs? I have a Macbook Pro M3 Max 48 GB..
@KiranMohan-dpthinkrАй бұрын
Hey David, how can we reassure clients that their data is secure and won't be shared with the LLM provider for internal training purposes? What steps can we take to ensure their data privacy and address any concerns they might have?
@cdunne1620Ай бұрын
You d to ask that in David’s classroom at skoool
@KiranMohan-dpthinkrАй бұрын
@@cdunne1620 Sure
@haljohnson6947Ай бұрын
He mentions that in the video like four times
@KiranMohan-dpthinkrАй бұрын
@@haljohnson6947 can you mention the specific timeline where he described about it.
@KiranMohan-dpthinkrАй бұрын
@@haljohnson6947 pls mention the timeline where he mentioned it.
@michaeltse321Ай бұрын
You downloade nemotron and not the 70b version which is why you had the error
@qkb3128Ай бұрын
Would have loved to check this out yet I don’t have that kinda money to spend to see the code. Good luck to ya .
@costatattooz840Ай бұрын
locally what hardware do you need to run this at minimum? i have a 64gb ram + 3060 12gb
@ticketforlife2103Ай бұрын
Watch the video
@H3XM0SАй бұрын
You'll need over 40gb vram so like 2 x rtx 4090 might be a good option. No idea what hardware is being used in the video. Anyone saying 'watch the video' should provide a timestamp.
@bollvigblackАй бұрын
this guys is rich. not even joking so
@chrystofferaugusto1194Ай бұрын
He is on a MacBook Pro bro…
@skeyenettАй бұрын
64GB RAM + 4070 Ti Super (16 VRAM) = Run Nemotron-70b-instruct-q2_K
@LuxciumАй бұрын
😂 I love the way you have called out your mistake 4:00 it was just so delightful to see you handle it like a boss that I have had to replay it more than 3 times to enjoy the moment... You are definitely a smart man!!! I am eager to see the evolution over time!!! 😅
@dark_cobaltАй бұрын
Already have it lol. Running it on my RX 7900XTX with q4m, but i think ill buy myself 1-2 Radeon W7900 Pro to gain a lot more performance. Alsp you don't need Ollama for it, because it's available in LM Studio and it's downloading from Huggingface. Btw what PC hardware specs do you have?
@rhadiemАй бұрын
He's clearly using a 128gb Macbook Pro which can use the memory as vram. He's running un-quantized. How much vram do you have on your gaming gpu? Nobody asked about your hardware bro.
@dark_cobaltАй бұрын
@@rhadiem Every PC can use the RAM as VRAM. It's how computers work. It's called virtual memory. If the VRAM fills up, the computer uses the RAM as backup memory, to stay stable and not crash. But the RAM is waaaaaaay slower than the VRAM, that's why I am asking him what specs he has. My GPU has 24GB of VRAM and even with the Quant 4M (around 32GB) model of Nemotron 70B my VRAM gets filled up completely and my RAM also to 50GB, which slows down the model to such an amount, that it's painfully slow. He is using a way bigger model, without any issues. If he has a GPU with this huge amount of VRAM, this would be totally understandable, but with the RAM? I don't understand why lol. 😄
@TheDarkLordAngelАй бұрын
That mark on your nose-it’s almost like a signature, something that’s so naturally you.🖖👍
@orthodox_gentlemanАй бұрын
Dude, there are very few people that can run nemotron locally….
@hrarungАй бұрын
awesome video David! How to train this model based on my dataset? and How to give it a nice UI?
@aaaaaaaaoooooooАй бұрын
Are my prompts on o1-preview used to train the AI even if I opt out? Where do I find this information?
@TheAsianDude9999Ай бұрын
What vscode extension are you using for your ai?
@borick2024Ай бұрын
Have you had a chance to compare your results against GPT4o?
@MrAndrew535Ай бұрын
I want to preserve a million-word dialogue between myself and my ChatGPT on multiple threads while upgrading to your recommendations. How do I achieve that?
@szebikeАй бұрын
Nice, your contribution to the open source community is awesome!
@ysh7713Ай бұрын
opensource?
@szebikeАй бұрын
@@ysh7713 Well kind of ~ better than giving all you data to a faceless big company who wills steal your data 100%.
@hotlineoperatorАй бұрын
I have test o1 - and it is not so smart. People still need to quide its selections. Big problem with models is censorship, someone else have select what you can do and not to do with these tools.
@VisualifeАй бұрын
You should use Anything LLM and docker / Open WebUI
@rafaelortega1376Ай бұрын
No repo to share the code?
@EtH-xf6brАй бұрын
What a beast Macbook you need to have to get such a fast response. I have 7800x3D and 4080 rtx and its waaay slower.
@danieleduardo9800Ай бұрын
How’d you get composer in the sidebar?
@Bakobiibizo26 күн бұрын
A terminal?! I'm freaking out man
@gauravrewaliya3269Ай бұрын
How to make local ai with backpropogation feature ( if got wrong stuff, CEO instruct what's wrong and it improve sub local agent by time )
@devbites77Ай бұрын
Inspiring stuff. Cheers!
@MiNiD33Ай бұрын
"Comments are apologies in code." - Robert C Martin. Cursor is helping you. Also for the price of the spec of this machine, you can buy an insane number if tokens from anthropic or openai. It might be worth getting people started using a hosted service.
@jefferystartm9442Ай бұрын
Brooooo , there are tools you are behind on . Agent s and Claude computer use?? E2B has an open source version tooo 😊 stay blessed Ondrej
@VinceOmondiАй бұрын
Good stuff, Ondrej!
@SjarMenaceАй бұрын
why do you have that thing on your nose?
@babyjvadakkan5300Ай бұрын
For correcting the nasal path/nose bridge (or something like that
@INeedMemeАй бұрын
More oxygen bro
@cdunne1620Ай бұрын
Soccer players used to wear them years ago for example Robbie Fowler for Liverpool
@AGINews-TogethWithAIАй бұрын
exactly what I needed thank you so much David🎉
@FuZZbaLLbeeАй бұрын
You can also use the ollama streaming output to generate text. This way you know what’s the generator is doing. Also I think that GPT o1 does more then split up a task and let agents fix the individual tasks. But nevertheless, a nice tutorial on making agents.
@zechariahprince5671Ай бұрын
We have had AGI for over a year.
@11metatron11Ай бұрын
Not a chance with my elderly MacBook Pro. Looks like I need some new gear…
@jayhu6075Ай бұрын
What a great explanation. Thnx
@eado9440Ай бұрын
🎉 you actually made it. Thanks
@gaelfalezАй бұрын
Missing the comparison between result using multiple agents and result using just 1.... Disappointing. We Don t even know if it is worth the work....
@skulltrickАй бұрын
Very inspiring! Thanks
@aatheraj1667Ай бұрын
Yet, we don't one that could trade Nasdaq futures.
@olivert.7177Ай бұрын
There is also an nemotron-mini model which is only 4b.
@samuelgarcia1802Ай бұрын
How good it is? In hugging face I saw nematron was in a bad place
@orthodox_gentlemanАй бұрын
Really??? Omg that is great
@aljosja3353Ай бұрын
Which computer u can use for local llm
@Gamatoto2038Ай бұрын
strong pc
@MrMoonsilverАй бұрын
Also, I hope the bruise on your nose heals soon. Been a long time now.
@TetardoАй бұрын
I think it’s a medical device that helps him breathe
@avi7278Ай бұрын
Oh yeah im sure openai is quaking in their boots, bro.
@chrystofferaugusto1194Ай бұрын
You should have a discord community to people share projects and business
@chrystofferaugusto1194Ай бұрын
Never mind, now I got the business model on skool. Nice call, thinking about joining it
@SCHaworthАй бұрын
No. Not quite. You have to split the turns.
@claxvii177th6Ай бұрын
1 token per second is too slow for any pratical use...
@sltАй бұрын
Dadusak!
@supermandemАй бұрын
Bro llama is nowhere near o1 wtf
@sushilsharma1621Ай бұрын
clickbait or misleading title
@blasterzmАй бұрын
Lol, that's not how O1 works. You can't tell it in the system prompt
@themax2goАй бұрын
modern day sham(mer) 👍
@dorukkurtogluАй бұрын
27:36 LOL🤪
@immortalityIMTАй бұрын
Cool!
@Álvaro-o5eАй бұрын
99% of free stuff sucks. One of them is this video. 20 minutes to answer "why is the sky blue?"
@overunityinventorАй бұрын
free stuff has a learning curve, it's not everyone's cup of tea
@tomwawer5714Ай бұрын
99% of paid software sucks and it hurts your wallet
@adithyansreeni7491Ай бұрын
i fkin slep bro
@gustavramedies2901Ай бұрын
David i would like to create sales agents,lead generators,receptionist,appointment setters and I want to sell them.Can you help 😢
@ShishuSudАй бұрын
😇
@EduardoAlarconGalloАй бұрын
Title is misleading. You are using Llama which is a LLM but not a Reasoner model
@surendarreddys7298Ай бұрын
1st one to comment 😄
@HimaLoubiАй бұрын
😂 you need a graphic card with a price of a Tesla car to run that module locally ; btw you talk like 10.000word/min , 😅
@TheBhushanJPawarАй бұрын
I am getting following error: bhushan@Bhushans-MacBook-Pro ~ % ollama run nemotron Error: llama runner process has terminated: signal: killed
@TheBhushanJPawarАй бұрын
After clearing some memory now it's started working...