Deploy ANY Open-Source LLM with Ollama on an AWS EC2 + GPU in 10 Min (Llama-3.1, Gemma-2 etc.)

  Рет қаралды 5,235

Developers Digest

Developers Digest

Күн бұрын

Пікірлер: 21
@DevelopersDigest
@DevelopersDigest 3 ай бұрын
The best way to support this channel? Comment, like, and subscribe!
@hpongpong
@hpongpong 3 ай бұрын
Great concise presentation. Thank you so much!
@DevelopersDigest
@DevelopersDigest 3 ай бұрын
Thank you! 🙏
@ryanroman6589
@ryanroman6589 3 ай бұрын
this is super valuable. awesome vid!
@DevelopersDigest
@DevelopersDigest 3 ай бұрын
Thank you! 🙏
@rembautimes8808
@rembautimes8808 3 ай бұрын
Thanks very nice tutorial
@DevelopersDigest
@DevelopersDigest 3 ай бұрын
Thank you
@brunozwietisch
@brunozwietisch 25 күн бұрын
I’m looking to learn how to use Llama. Do you have a minimum configuration for the 8B version to run? Because here in Brazil, the dollar-to-real exchange rate is 6 to 1, and by the end of the month, the budget gets tight for those who want to learn.
@DevelopersDigest
@DevelopersDigest 24 күн бұрын
Groq has a free tier for llama as well as cloudflare!
@alejandrogallardo1414
@alejandrogallardo1414 3 ай бұрын
for models at ~70b, i am getting timeout issues using vanilla ollama. It works with the first pull/run, but times out when i need to reload model. Do you have any recommendations for persistently keeping the same model running?
@DevelopersDigest
@DevelopersDigest 3 ай бұрын
github.com/ollama/ollama/pull/2146
@rehanshaikh2708
@rehanshaikh2708 Ай бұрын
how can i use this endpoint in langchain chatollama?
@nexuslux
@nexuslux 3 ай бұрын
Can you use open web ui?
@danielgannage8109
@danielgannage8109 3 ай бұрын
This is very informative! Thanks :) Curious why you used a g4dn.xlarge GPU ($300/month) instead of a t3.medium CPU ($30/month)? I assumed the 8 Billion parameter model was out of reach with regular hardware. What max model size works with the g4dn.xlarge GPU? To put into perspective, I have a $4K macbook (16gb ram) that can really only run the large (150 million) or medium (100 million parameter) sized model, which i think the t3.medium CPU on AWS can only run the 50 million param (small model).
@dylanv3044
@dylanv3044 3 ай бұрын
maybe a dumb question. how do you turn the stream data you received into readable sentences
@DevelopersDigest
@DevelopersDigest 3 ай бұрын
You could accumulate tokens and split by the end of sentences . ! ? Etc and then send resp after grouping function like that
@ConAim
@ConAim 2 ай бұрын
Stay away from AWS, it will cost you arms and legs in a long run..
@DevelopersDigest
@DevelopersDigest 2 ай бұрын
Which vendors do you prefer? 🙂
@BeCodeless-dot-net
@BeCodeless-dot-net 3 ай бұрын
nice explaination
@DevelopersDigest
@DevelopersDigest 3 ай бұрын
Thank you!
Qwen Just Casually Started the Local AI Revolution
16:05
Cole Medin
Рет қаралды 87 М.
host ALL your AI locally
24:20
NetworkChuck
Рет қаралды 1,3 МЛН
How To Choose Mac N Cheese Date Night.. 🧀
00:58
Jojo Sim
Рет қаралды 93 МЛН
Thank you Santa
00:13
Nadir Show
Рет қаралды 28 МЛН
AI's 7 Levels Exposed: Are You Ready?
31:28
The Professor
Рет қаралды 16
Deploy Ollama and OpenWebUI on Amazon EC2 GPU Instances
45:18
StratusGrid
Рет қаралды 1,1 М.
Ollama on Kubernetes: ChatGPT for free!
18:29
Mathis Van Eetvelde
Рет қаралды 6 М.
All You Need To Know About Running LLMs Locally
10:30
bycloud
Рет қаралды 175 М.
Expert Guide: Installing Ollama LLM with GPU on AWS in Just 10 Mins
10:14
Fast and Simple Development
Рет қаралды 9 М.
Llama 3.1 is ACTUALLY really good! (and open source)
7:05
ForrestKnight
Рет қаралды 48 М.
Do NOT Learn Kubernetes Without Knowing These Concepts...
13:01
Travis Media
Рет қаралды 324 М.
Llama 3.2 Vision + Ollama: Chat with Images LOCALLY
10:30
Leon van Zyl
Рет қаралды 12 М.
How To Choose Mac N Cheese Date Night.. 🧀
00:58
Jojo Sim
Рет қаралды 93 МЛН