Grok's AI Predictions for 2024
10:23
Пікірлер
@igoriane93
@igoriane93 5 күн бұрын
In my specific case (7900XT) it tries to locate ROCm0, as Iam using windows my computer freezes.
@TigerTriangleTech
@TigerTriangleTech 4 күн бұрын
The AMD 7900 XT is supported by Ollama "out of the box". No need to use the modified version. But you might also want to make sure your OS and drivers are up to date.
@shiv4667
@shiv4667 6 күн бұрын
Thanks brother, I also have a 6750xt. I installed the llama3.2 11b today and it was very slow and I realized that it was using 100% CPU. After this fix it is around 50% for both CPU and GPU. Thanks a lot.
@TigerTriangleTech
@TigerTriangleTech 4 күн бұрын
Good deal! Glad to hear it!
@nome_de_usuario
@nome_de_usuario 12 күн бұрын
You are awesome, you saved my life and my graduation project and a lot of money LOL
@TigerTriangleTech
@TigerTriangleTech 12 күн бұрын
Glad to hear it! Best of luck with your project!
@coolcha
@coolcha 14 күн бұрын
Can this work with integrated graphics?
@TigerTriangleTech
@TigerTriangleTech 12 күн бұрын
Like with the AMD APUs? I'm not sure. Even if it was supported, I doubt if it could perform very well. Too many limitations.
@DiegoThoms
@DiegoThoms 16 күн бұрын
Thanks! I made it work with my 6600!
@TigerTriangleTech
@TigerTriangleTech 16 күн бұрын
Very cool! Thanks for the feedback.
@vikoscharger2427
@vikoscharger2427 20 күн бұрын
Thank you!It works with my 5700xt! Night and day istead of running on my CPU!
@TigerTriangleTech
@TigerTriangleTech 19 күн бұрын
Great! Thanks for the feedback!
@RomanMondragon-lh5fh
@RomanMondragon-lh5fh 21 күн бұрын
worked on the rx 6700xt, thank you so much
@TigerTriangleTech
@TigerTriangleTech 19 күн бұрын
You're welcome! Glad it worked!
@joshuaosei5628
@joshuaosei5628 26 күн бұрын
This worked on RX 5500M! Thanks
@TigerTriangleTech
@TigerTriangleTech 26 күн бұрын
Good deal. You're welcome!
@murataras5938
@murataras5938 26 күн бұрын
yeni versiyonda model kısmında bu menü gelmiyor nasıl model yükleyeceğiz acil yadım
@TigerTriangleTech
@TigerTriangleTech 26 күн бұрын
Does this help? To download models, navigate to the Ollama Settings in the Connections tab. Alternatively, you can also download models directly by typing a command like ollama run hf.co/[model creator]/[model name] in the model selection dropdown. This action will create a button labeled "Pull [Model Name]" for downloading. docs.openwebui.com/features/workspace/models#model-discovery-and-importexport
@rogerellis7758
@rogerellis7758 29 күн бұрын
after an install, it showed no models, that's what I was hoping to figure out is how do I install a model after installing it with docker ?
@TigerTriangleTech
@TigerTriangleTech 29 күн бұрын
Hi, did you install Ollama and pull some models? If so, they should then show up in Open WebUI. Make sure Ollama is running and working first.
@ThermoIdiot
@ThermoIdiot Ай бұрын
can it run on rx 6500m laptop gpu
@TigerTriangleTech
@TigerTriangleTech Ай бұрын
Hi, you will need to determine the LLVM target of your GPU and then use the related ROCBLAS packages. I would recommend looking at your log file and determine that target which should start with gfx. In your case it might be GFX1033 but I'm not 100% sure. Maybe someone else can chime in here. Here is the place in the video that will show you in the Ollama log file: kzbin.info/www/bejne/fV7Ooamiq7CAZ8k Here is the latest release (currently) of the ROCBLAS packages: github.com/likelovewant/ROCmLibs-for-gfx1103-AMD780M-APU/releases/tag/v0.6.1.2 Here is a video I made to show how to look at the log file: kzbin.info/www/bejne/pZ-Wo6evhbyGnaM Hope that helps!
@TheNewAllotmenteer
@TheNewAllotmenteer Ай бұрын
Fascinating!! What tool are you using for the text to voice, if I can ask?
@TigerTriangleTech
@TigerTriangleTech Ай бұрын
Sure, I was simply using the feature built into Microsoft Edge, where you can highlight text and have it read to you. Thanks for watching!
@causalinference4176
@causalinference4176 Ай бұрын
Very useful, but I love the tiger in the beginning
@TigerTriangleTech
@TigerTriangleTech Ай бұрын
Hey, thanks for the feedback! I appreciate it.
@abdullahal-jauni3013
@abdullahal-jauni3013 Ай бұрын
perfect explanation video, it worked on amd 5700 xt. thank you
@TigerTriangleTech
@TigerTriangleTech Ай бұрын
You're welcome! Thanks for the feedback!
@anjinho_ruindade_pura
@anjinho_ruindade_pura Ай бұрын
Great video! It worked perfectly <3
@TigerTriangleTech
@TigerTriangleTech Ай бұрын
Glad it helped! Thanks for the feedback!
@louisgamercool2324
@louisgamercool2324 Ай бұрын
Thanks so much man, I don't know if you remember my comment from an earlier video where I wrote that gpu was slower than CPU, but its all good you have fixed my problem. I'm using llama to plugin to obsidian for my notes, so speed is all I need as the information will be my own. I do have one question though, I only get 440 prompt eval rate on my 5700xt, which like you said could just mean that I need to explore other models that worked well for my system. Was just wondering where I should look for smaller models as I am extremely new to AI. Once again, thanks for the great video.
@TigerTriangleTech
@TigerTriangleTech Ай бұрын
Yep, I do remember, now that you mention it. Glad it helped. Llama 3.2 3B worked well for me. Mini models like Phi also work well on my system and might be worth a look. I think if you try models with that smaller parameter count you should see the improved speed over the larger models (7 or 8B, etc...). Which Obsidian plugin are you using? ollama.com/search
@zandanshah
@zandanshah Ай бұрын
Thanks, keep the good work going.
@TigerTriangleTech
@TigerTriangleTech Ай бұрын
Thanks, will do!
@zandanshah
@zandanshah Ай бұрын
I just got AMD Radeon RX 7900 XTX 24 GB VRAM.
@louisgamercool2324
@louisgamercool2324 Ай бұрын
Hi, I have checked your other video regarding whether or not it is using gpu and ollama says that GPU is being used. Moreover, adrenaline shows that my 5700xt is at 99% usage. However, I am getting slower output compared to CPU. Is there a fix for this? Thanks
@louisgamercool2324
@louisgamercool2324 Ай бұрын
Otherwise Great video, thanks a lot.
@TigerTriangleTech
@TigerTriangleTech Ай бұрын
I have run into this as well. You might try a smaller model (3B parameter like Llama 3.2 or maybe Phi). I'm wondering if it's a bottleneck between transferring data between the CPU and GPU and it could be a matter of needing more RAM. I address this in my video on performance: kzbin.info/www/bejne/omHXlGWKiN2ehZo
@TigerTriangleTech
@TigerTriangleTech Ай бұрын
Thanks for your feedback!
@kevinmiole
@kevinmiole Ай бұрын
But I have rx6600 can I use ollama?
@TigerTriangleTech
@TigerTriangleTech Ай бұрын
It should work. That is an LLVM target of gfx1032. rocm.docs.amd.com/projects/install-on-windows/en/latest/reference/system-requirements.html#supported-gpus-win github.com/likelovewant/ollama-for-amd#windows
@varunaeeriyaulla
@varunaeeriyaulla Ай бұрын
Yes you can. Follow the steps
@KhongGuan-e8h
@KhongGuan-e8h Ай бұрын
its work on my trash gpu (amd rx580)
@TigerTriangleTech
@TigerTriangleTech Ай бұрын
Thanks for watching! And for your feedback!
@Karmezz
@Karmezz Ай бұрын
This video was easy to understand and proffesional. Worked on rx 6600 no problems keep up!
@TigerTriangleTech
@TigerTriangleTech Ай бұрын
Good deal! Thanks for your feedback!
@iljaruppel3047
@iljaruppel3047 Ай бұрын
This is the first video that realy simple, well explained and also works. Thank you👌 But i have the Problem that mistral is much slower with GPU than with the CPU. I use a rx 6600XT. I dont know what is the Problem.
@TigerTriangleTech
@TigerTriangleTech Ай бұрын
Thanks, I'm glad it was helpful. The problem you are having usually points to a resource problem. Switching to smaller models like Llama 3.2 3B (or models that are around that size) might work better. Just keep in mind it's a trade off because you do lose some quality. If you haven't yet watched my video on performance, you might want to check it out as it demonstrates a similar problem I had when using Llama 3 on one of my systems. kzbin.info/www/bejne/omHXlGWKiN2ehZo
@Senorkawai
@Senorkawai 2 ай бұрын
Here is the translation: "Thank you, I followed the steps in the video and finally Ollama ran on my GPU RX 6600. I just had to restart for it to work (before the restart, it was using the CPU). Now I will check if I can use Docker and WebUI to use it with a better interface. Thanks again!"
@TigerTriangleTech
@TigerTriangleTech 2 ай бұрын
Great! I'm glad it worked for you. I also have videos on Docker and Web UI if you need help with that. Thanks for watching!
@vmeow9895
@vmeow9895 2 ай бұрын
why after a while, the response_token/s gradually decreases? i use rx 6600xt
@TigerTriangleTech
@TigerTriangleTech 2 ай бұрын
Hello and thanks for watching. I'm not sure, but in a chat situation, I would think it could be because it is storing previous prompts in memory in order to "remember" the previous conversations. How much it keeps could depend on the context window of the model. I have not tested this but that's my best guess. Someone else may have a better answer.
@vmeow9895
@vmeow9895 2 ай бұрын
@@TigerTriangleTech Thanks for your reply to my question, Hope you can solve it later😁
@naimishranderi
@naimishranderi 2 ай бұрын
Hey there, very good video. Can you please tell, what is your computer specification? My laptop can't process even simple query.
@TigerTriangleTech
@TigerTriangleTech 2 ай бұрын
Thank you! And great question. I show my specs (for the computer I mainly use for recording) and address how to get better performance with Ollama in my 3 latest videos. Here is one of them... kzbin.info/www/bejne/omHXlGWKiN2ehZo I hope it helps!
@MichelBertrand
@MichelBertrand 2 ай бұрын
I have been wanting to run ollama on my 6750XT ever since AMD announced they were supporting their GPUs. THANK YOU! Just installed and it seems to work perfectly. So much faster than running on my 13700k !
@TigerTriangleTech
@TigerTriangleTech 2 ай бұрын
I'm glad to hear it. Thanks for watching!
@iamnobody-001
@iamnobody-001 2 ай бұрын
hi, can you make similar video, but for discrete gpu, i have old gtx 1050 and ollama is using cpu only, i want to know how to activate the gpu, also how to disable it, to see the difference. thanks
@TigerTriangleTech
@TigerTriangleTech 2 ай бұрын
Hi there! Actually, the video I made was for a discrete GPU, but it was AMD rather than NVIDIA. You shouldn't have to install any work around. Ollama supports Nvidia GPUs with compute capability 5.0+. You should be in good shape because the GTX 1050 has a compute capability of 6.1. If this is not working correctly you might want to update your drivers? Not sure. To force CPU usage, set an environment variable of CUDA_VISIBLE_DEVICES to -1. To enable GPU usage just remove that variable and it should utilize your GPU again. For more info view this page: github.com/ollama/ollama/blob/main/docs/gpu.md
@Byrex_Lorence
@Byrex_Lorence 2 ай бұрын
It works! thank you so much. you did a very good video
@TigerTriangleTech
@TigerTriangleTech 2 ай бұрын
Great! Thank you for watching and for your feedback!
@VoxPrimeAIadvocate
@VoxPrimeAIadvocate 3 ай бұрын
How do you use the openweb ui API key? Not the open AI API key. I want to "generate secret API key" and put my local AI into apps, VR, discord bot, and use on my phone.
@TigerTriangleTech
@TigerTriangleTech 3 ай бұрын
Hi there! Thanks for your patience. I have been busy on other projects lately. Open WebUI actually sits on top of something like Ollama so it would not have an API key that you would use. Understanding API Keys: Open AI API Key: You need this if you're connecting to OpenAI's services. This isn't related to running local models but is for accessing their cloud-based models. No API Key for Local Models: When using Ollama or similar local AI runners, you generally don't need an API key for the model itself since it runs on your machine. Exposing Local AI to External Apps: Local AI Access: If you want to access Ollama or any local AI from outside your local network (like in VR, on your phone, or through a Discord bot), you'll indeed need to make your local service accessible over the internet. Here's how: Dynamic DNS Services: Services like No-IP can give your dynamic IP a static hostname. This is useful but might not directly help with API access unless paired with port forwarding or tunneling. Tunneling with ngrok: This is a great suggestion for exposing a local server to the internet temporarily. However, for a more permanent solution, you might look into setting up a reverse proxy or using a VPN. Securing Your Service: If you decide to make your local AI accessible via an API (which would be necessary for integration into apps or bots), you should indeed secure it: Create an API Endpoint: Your app or bot would interact with this endpoint. Here, you might consider generating a "secret API key" not for Ollama but for your application layer that interacts with Ollama or another local AI. This key would be used to authenticate and authorize requests to your service, not to interact with the AI model itself. Security: Use HTTPS for your API, implement authentication (like API keys or OAuth), and consider rate limiting to protect your local resources. Implementation Ideas: For Discord Bot or Mobile App: You would write an application that communicates with your local AI through a secure API you've set up. This app would use your custom API key to authenticate its requests to your local server. Future Video Content: This is indeed a fantastic idea for a video or a series! Covering how to set up a local AI, expose it safely to the internet, secure it with API keys, and then integrate it into various platforms could be very educational. Thanks for watching!
@mahiaravaarava
@mahiaravaarava 4 ай бұрын
The political views of popular AI chatbots vary based on their training data and design. Some aim to remain neutral, providing balanced responses, while others might reflect the biases present in their sources. Understanding their design and limitations helps in interpreting their outputs objectively.
@TigerTriangleTech
@TigerTriangleTech 3 ай бұрын
Very true! Hey, thanks for watching!
@leighjenkins5601
@leighjenkins5601 6 ай бұрын
We cannot trust woke corporations and their woke agendas programmed in.
@TigerTriangleTech
@TigerTriangleTech 6 ай бұрын
Thank you for watching! I believe these open models, especially the uncensored ones, ran locally have a bright future and the people that are learning about them now will be ahead of the game.
@Cage-a-liciousAI
@Cage-a-liciousAI 7 ай бұрын
Very well done video! Looking forward to part 2
@TigerTriangleTech
@TigerTriangleTech 7 ай бұрын
Thanks Cage-a-liciousAI! And keep making those N. Cage videos. Very creative!
@irocz5150
@irocz5150 7 ай бұрын
far from be an expert but doing a prompt is very slow to get an answer.
@TigerTriangleTech
@TigerTriangleTech 7 ай бұрын
Yeah, running these locally is going to depend on your local resources. You will definitely want a good GPU and Ollama should take advantage of that. Still can be a bit slow though it seems. Also depends on the parameter size of the model. Thanks for watching!
@TigerTriangleTech
@TigerTriangleTech Ай бұрын
If you haven't yet watched my video on performance, you might want to check it out as it demonstrates a similar problem I had on one of my systems. kzbin.info/www/bejne/omHXlGWKiN2ehZo
@irocz5150
@irocz5150 7 ай бұрын
I really appreciate the time for creating this Ollama videos. For noobs this step by step are very helpful.
@TigerTriangleTech
@TigerTriangleTech 6 ай бұрын
Glad it was helpful! Thanks for the feedback!
@Mukka-luv-Dots
@Mukka-luv-Dots 9 ай бұрын
Exactly what I want thank you very much 🙏
@TigerTriangleTech
@TigerTriangleTech 9 ай бұрын
I'm really glad to hear that! Thanks for letting me know.
@Ibhenriksen
@Ibhenriksen 9 ай бұрын
Hopefully the FBI isn't knocking at your door man.
@TigerTriangleTech
@TigerTriangleTech 9 ай бұрын
So far so good :)
@Rationalific
@Rationalific 10 ай бұрын
Interesting stuff. I certainly prefer a full earnest discussion rather than a lecture and no answer to the posed question.
@rogerellis7758
@rogerellis7758 11 ай бұрын
nice, I like writing stories with Grok all the time!
@hugokappes4077
@hugokappes4077 11 ай бұрын
you got to get grok playing poker
@TigerTriangleTech
@TigerTriangleTech 11 ай бұрын
Yep, more cards to keep track of, so that will be interesting.
@hugokappes4077
@hugokappes4077 11 ай бұрын
@@TigerTriangleTech can grok really shuffle a deck and not know what the cards are ? and with poker you need to figure a way for there to be some skin in the game something to win or lose, do a bit longer game too,,
@TigerTriangleTech
@TigerTriangleTech 11 ай бұрын
​@@hugokappes4077 The honest answer to that is, it depends. For a few reasons: First, this is generative AI and not specifically trained for poker so the way cards are tracked, and score is kept will differ especially since there are so many variations of the game. Second, Grok is an AI chatbot as opposed to a rules-based chatbot, so it's non-deterministic. Which means the output is not always the same and will vary from session to session. And finally, AI chatbots can still give wrong answers but do it with confidence (known as hallucinations). Thank you for watching and thanks for your feedback! I appreciate it!
@digitt2
@digitt2 Жыл бұрын
It's going to be a lot harder to deceive lie and use denial and some people will be angry.
@hugokappes4077
@hugokappes4077 Жыл бұрын
Hi ! again,, !!!
@TigerTriangleTech
@TigerTriangleTech Жыл бұрын
Hello again!
@BrianBellia
@BrianBellia Жыл бұрын
Best Grok video I've seen so far. Very clever the way you got it to spill the beans on Optimus. Loved the Lost in Space bit!
@TigerTriangleTech
@TigerTriangleTech Жыл бұрын
Thanks, it was fun to put together. Time consuming but fun 😃.
@Denny_7782
@Denny_7782 Жыл бұрын
Rad!
@hugokappes4077
@hugokappes4077 Жыл бұрын
ok new channel, nice im 4th subscriber !!!!
@TigerTriangleTech
@TigerTriangleTech Жыл бұрын
Thank you!
@hugokappes4077
@hugokappes4077 Жыл бұрын
@@TigerTriangleTech looking forward to the next one,, i may make some suggestions when i thin of something,,
@TigerTriangleTech
@TigerTriangleTech Жыл бұрын
@hugokappes4077 sounds great! Would love the feedback. Yeah, I have some stuff in the works.