Ollama on Kubernetes: ChatGPT for free!

  Рет қаралды 12,740

Mathis Van Eetvelde

Mathis Van Eetvelde

Күн бұрын

Пікірлер: 31
@oadm_beats
@oadm_beats 2 ай бұрын
Excelente video! Me encantó cómo explicaste el proceso de implementar Ollama en Kubernetes. Gracias por compartir tu conocimiento!
@SteveHarris-mi9ou
@SteveHarris-mi9ou 2 ай бұрын
It's a cool setup to use to run a RAG setup locally as well - nice going.
@zulhilmizainudin
@zulhilmizainudin 2 ай бұрын
Looking forward for the next video!
@mathisve
@mathisve 2 ай бұрын
You can find the video here: kzbin.info/www/bejne/j6DQoGV6o7FshKM
@zulhilmizainudin
@zulhilmizainudin 2 ай бұрын
@@mathisve thanks!
@JoeIrizarry88
@JoeIrizarry88 Ай бұрын
Great Vid. In the future I recommend adding a license to your GitHub repo such as MIT. 😉 👍🏻
@Techonsapevole
@Techonsapevole 2 ай бұрын
I use docker compose but i was curious about k8s
@beastmastern159
@beastmastern159 Ай бұрын
nice video, nvidia container toolkit only suported in lunux distribution so i installed a ubuntu booted whit my windows to make this work whit my gpu into kubernetes xdd but was good think i love this.
@Fayaz-Rehman
@Fayaz-Rehman Ай бұрын
Thank you very much for sharing. Can we save the model llamaL3.5 file in a persistent volume?
@Sentientforce
@Sentientforce 2 ай бұрын
Can you please advise how to run ollama in k3d cluster in wsl2- windows 11 and docker desktop environment. The issue I’m not able to solve is making gpu visible in a node.
@HosseinOjvar
@HosseinOjvar 2 ай бұрын
Helpful tutorial thank you
@MuhammadRehanAbbasi-j5w
@MuhammadRehanAbbasi-j5w 2 ай бұрын
Would really like the video on how to add a GPU to this, both locally and on the cloud.
@mathisve
@mathisve 2 ай бұрын
Stay tuned for that video! I'm working on it as we speak, should be out later this week!
@mathisve
@mathisve 2 ай бұрын
I have uploaded a new video on how to run Ollama on a GPU in the cloud: kzbin.info/www/bejne/j6DQoGV6o7FshKM
@samson-olusegun
@samson-olusegun 2 ай бұрын
Would using a k8s job to make the pull API call suffice?
@mathisve
@mathisve 2 ай бұрын
Yes and no! On paper, if you only had one pod this could work. But the API call needs to be made every time a new Ollama pod is scheduled (unless you're using a PVC mounted to the pod to store the model). As far as I'm aware it's not possible to start a Kubernetes job at the creation of a new pod without using an operator.
@ganeshnaiknavare2815
@ganeshnaiknavare2815 2 күн бұрын
I want to deploy the ollama model free can you please suggest me platform
@Tudorabil
@Tudorabil 2 ай бұрын
Would a postStart command `ollama pull llama3.2` work for the ollama pod, so you don't have to create the curl pod?
@mathisve
@mathisve 2 ай бұрын
I haven't tried it out but I think it might. Not sure how I didn't think of this
@domagalsky
@domagalsky 2 ай бұрын
Why don't you just prebuild docker image with embeded llama model in it? It saves spin up time of the image, reduces complexity and amount of api calls to get the model per pod deploy.
@mathisve
@mathisve 2 ай бұрын
That's a possibility for sure. One downside of this is that depending on the size of the LLM, the images can get very big. In my opinion it's much more pragmatic to mount a PVC to the pods so they only have to download it once.
@domagalsky
@domagalsky 2 ай бұрын
@@mathisve having shared volume ReadOnlyMany sounds the best option!
@enzanto
@enzanto 2 ай бұрын
Is there a reason why you do not use volume mounts? no need to DL the model each time
@mathisve
@mathisve 2 ай бұрын
That would indeed be the ideal scenario. Attach a rwx pvc to all ollama pods, that way they only have to pull the model once. I didn't include this in the video for simplicity sake
@enzanto
@enzanto 2 ай бұрын
@mathisve good to know, I "feared" that it was some other reason I could not think of. Going to set it up in kubernetes my self... A day I get the tile for it
@MahatmaLevolence
@MahatmaLevolence Ай бұрын
I'd love something like this it'd be very helpful for me but being a cybertard i wouldn't even know where to start. Have fun techies!
@unclesam007
@unclesam007 2 ай бұрын
here i cant deploy a simple laravel app on k8s🤒
@mathisve
@mathisve 2 ай бұрын
Do you need help with deploying Laravel on Kubernetes?
Ollama with GPU on Kubernetes: 70 Tokens/sec !
20:19
Mathis Van Eetvelde
Рет қаралды 2,2 М.
Wazuh! Powerful, Open Source Endpoint Security Monitoring!
50:18
Awesome Open Source
Рет қаралды 12 М.
“Don’t stop the chances.”
00:44
ISSEI / いっせい
Рет қаралды 62 МЛН
人是不能做到吗?#火影忍者 #家人  #佐助
00:20
火影忍者一家
Рет қаралды 20 МЛН
Transformers (how LLMs work) explained visually | DL5
27:14
3Blue1Brown
Рет қаралды 4,6 МЛН
Anthropic MCP with Ollama, No Claude? Watch This!
29:55
Chris Hay
Рет қаралды 19 М.
Stop Using Docker. Use Open Source Instead
12:40
DevOps Toolbox
Рет қаралды 292 М.
host ALL your AI locally
24:20
NetworkChuck
Рет қаралды 1,6 МЛН
NVIDIA CEO Jensen Huang's Vision for the Future
1:03:03
Cleo Abram
Рет қаралды 595 М.
Turn ANY Website into LLM Knowledge in SECONDS
18:44
Cole Medin
Рет қаралды 166 М.
docker stack is my new favorite way to deploy to a VPS
27:47
Dreams of Code
Рет қаралды 159 М.