How to Use Open Source LLMs in AutoGen Powered by vLLM

  Рет қаралды 5,337

Yeyu Lab

Yeyu Lab

Күн бұрын

Пікірлер: 19
@joeclacher445
@joeclacher445 Жыл бұрын
Unreal video! Looking forward to testing various models instead of GPT!
@yeyulab
@yeyulab Жыл бұрын
Thanks, would like to see the performances as well.
@JDWilsonJr
@JDWilsonJr Жыл бұрын
This is spot on. Thank you for making the video and explaining so well.
@yeyulab
@yeyulab Жыл бұрын
Thanks
@Beenee_AI
@Beenee_AI Жыл бұрын
Great ! You are really good at what you do!
@yeyulab
@yeyulab Жыл бұрын
Thank you!
@jaoltr
@jaoltr Жыл бұрын
Terrific video! Thank you for sharing your knowledge.
@yeyulab
@yeyulab 11 ай бұрын
Glad it was helpful!
@KodandocomFaria
@KodandocomFaria Жыл бұрын
Is it possible to use powerinfer instead of vllm? If possible which one would be faster ? Perhaps a good video to make by comparing those two inference tools
@yeyulab
@yeyulab 11 ай бұрын
Looks Powerinfer is a pretty new inference tool. It cannot be supported in Autogen directly right now but If you can run Uvicorn to serve its inference, there maybe a chance. Thanks for the recommendation.
@truliapro7112
@truliapro7112 Жыл бұрын
How to use autogen with aws bedrock models ?
@current.undone
@current.undone Жыл бұрын
thanks for sharing. Can vLLM be installed on Mac? Please help if it can as Max Studio has all the musles needed to do the heavylifting 🙂
@yeyulab
@yeyulab Жыл бұрын
vLLM does not support MAC backend at the moment and I guess the reason is that they want to maximize the throughput of generation by V100/H100 GPUs.
@shubhamgarg5007
@shubhamgarg5007 Жыл бұрын
Hey, thanks for the in depth explanation. While its great that we can use Autogen along with open source models using vLLM, is there any chance we could use Gemini API along with autogen?
@yeyulab
@yeyulab 11 ай бұрын
There is an on-going branch of AutoGen working on Gemini integration. Soon you can use it I think. github.com/microsoft/autogen/tree/gemini
@shubhamgarg5007
@shubhamgarg5007 11 ай бұрын
@@yeyulab Yeah, I checked it but it has no commits since the last 2 weeks and I doubt its one of their top priorities as of right now. I couldn' find any online resources to use Gemini's free api with autogen either.
@yeyulab
@yeyulab 11 ай бұрын
Free Gemini API is really useful I agree. Let me check with their team.
@Nick_With_A_Stick
@Nick_With_A_Stick 11 ай бұрын
I don’t believe you have it in your youtube tag’s but you should fill our your youtube tags with things like “ vLLM tutorial” as k looked for one and came up very very short, and this would’ve been much mote useful. Thanks for the vid!
@yeyulab
@yeyulab 11 ай бұрын
Good suggestion, thanks!
vLLM on Kubernetes in Production
27:31
Kubesimplify
Рет қаралды 4 М.
Quando eu quero Sushi (sem desperdiçar) 🍣
00:26
Los Wagners
Рет қаралды 15 МЛН
BAYGUYSTAN | 1 СЕРИЯ | bayGUYS
36:55
bayGUYS
Рет қаралды 1,9 МЛН
It works #beatbox #tiktok
00:34
BeatboxJCOP
Рет қаралды 41 МЛН
vLLM - Turbo Charge your LLM Inference
8:55
Sam Witteveen
Рет қаралды 17 М.
How to Ask Questions to LLMs
56:35
Phi-AI
Рет қаралды 68
How to Create APIs for AutoGen
23:37
Yeyu Lab
Рет қаралды 1 М.
Talk to Your Documents, Powered by Llama-Index
17:32
Prompt Engineering
Рет қаралды 88 М.
Ollama and Python for Local AI LLM Systems (Ollama, Llama2, Python)
30:10
Eli the Computer Guy
Рет қаралды 9 М.
Quando eu quero Sushi (sem desperdiçar) 🍣
00:26
Los Wagners
Рет қаралды 15 МЛН