No video

Getting Started with Groq API | Making Near Real Time Chatting with LLMs Possible

  Рет қаралды 31,846

Prompt Engineering

Prompt Engineering

Күн бұрын

Let's have a look at the Groq API that offers access to their Language Processing Units (LPUs) for free (for a limited time). The API can provide close to 500 tokens per second with Mixtral-8x7B. Enjoy :)
🦾 Discord: / discord
☕ Buy me a Coffee: ko-fi.com/prom...
|🔴 Patreon: / promptengineering
💼Consulting: calendly.com/e...
📧 Business Contact: engineerprompt@gmail.com
Become Member: tinyurl.com/y5h...
💻 Pre-configured localGPT VM: bit.ly/localGPT (use Code: PromptEngineering for 50% off).
LINKS:
Signup for API Access: groq.com/
Colab Notebook: tinyurl.com/2nx...
Streamlit Chatbot: tinyurl.com/3f4...
TIMESTAMPS:
[00:00] Introduction
[00:34] How to Access the Groq API?
[01:06] API Playground
[03:15] Getting Started with Groq API
[05:33] Near real-time speed
[07:24] Advanced API Features: Streaming and Stop Sequences
[11:02] Building a Chatbot with Streamlit and Grok API
All Interesting Videos:
Everything LangChain: • LangChain
Everything LLM: • Large Language Models
Everything Midjourney: • MidJourney Tutorials
AI Image Generation: • AI Image Generation Tu...

Пікірлер: 45
@martg0
@martg0 5 ай бұрын
Thanks for the video! I will start testing this API with a POC I am working now to learn.
@thierry-le-frippon
@thierry-le-frippon 5 ай бұрын
They should sell their LPUs instead and compete with Nvidia. They would surely get lots of backup and investments. They will probably be copied instead othetwise and fade away quickly.
@jonoburcham4059
@jonoburcham4059 5 ай бұрын
Great video! Can ou make a voice chatbot using groq in one of your next videos please? I would also love to see if you do this on streamlit or if it's too slow and you use something else. Thanks so much for your videos
@engineerprompt
@engineerprompt 5 ай бұрын
Planning on making that. For voice chatbot, might just do cli though
@KOTAGIRISIVAKUMAR
@KOTAGIRISIVAKUMAR 5 ай бұрын
why you cant use the conversational retrieval chain instead of the conversation chain Because it can handle the memory by default no need maintain externally? @prompt Engineering
@user-yz8tf3uv4y
@user-yz8tf3uv4y 5 ай бұрын
Please Create a step-by-step video guide on using the Groq API with Streamlit.
@engineerprompt
@engineerprompt 5 ай бұрын
That's coming soon
@sausage4mash
@sausage4mash 5 ай бұрын
did someone say free
@engineerprompt
@engineerprompt 5 ай бұрын
For the time being :)
@dhruvpatel2554
@dhruvpatel2554 5 ай бұрын
Awesome stuff !!!!
@osamaa.h.altameemi5592
@osamaa.h.altameemi5592 5 ай бұрын
this is next level. OpenAI got some serious competition.
@DestanBegu
@DestanBegu 5 ай бұрын
Thanks for your content! I´m using Streamlit as well and want to give Content as the System role. For Example "answer me in short sentences in italian" so it will do this for each prompt i do. Where can i do this in the code? I used the Streamlit Chatbot Repo. Thanks in advance
@hmsfaceface8925
@hmsfaceface8925 5 ай бұрын
how can the groq fpga use mixtral 8x7b with just 250gigs of vram?
@coyoteq
@coyoteq 5 ай бұрын
Bcoz of groq tpu...
@vishnuprabhaviswanathan546
@vishnuprabhaviswanathan546 5 ай бұрын
How to control the output of LLM for a single input?
@ConnectorIQ
@ConnectorIQ 16 күн бұрын
almost a baby version of a quantum computer if you can actually perfect a model based on speed of responses to your questions and using the groq gpu...
@shaheerabdullah6738
@shaheerabdullah6738 3 ай бұрын
Very Helpful.
@ramimithalouni6592
@ramimithalouni6592 5 ай бұрын
what is the time to receive the first chunk in streaming?
@easy-dashboard
@easy-dashboard 5 ай бұрын
Depends on the amount of input tokens. With a one line instructions it's below 1 second. If you include context of a RAG-System it will go up to 3 seconds to start the first token (30k tokens of context)
@jmay3230
@jmay3230 5 ай бұрын
If temp can adjust to minus what is impact on generation ( consider it as hypothetical if case don't exist )
@engineerprompt
@engineerprompt 5 ай бұрын
It will be same as setting it zero :) basically if you set it zero, it will pick the next most probable token. If you set a higher value, it can to sample among the most probable tokens
@benben2846
@benben2846 5 ай бұрын
tu est fort man ^^👍
@jesusleguizamon6566
@jesusleguizamon6566 4 ай бұрын
Hi, this api have function calling? regards
@mickelodiansurname9578
@mickelodiansurname9578 5 ай бұрын
Heres the question, can Groq cards also work on inference for art and audio and voice models? or is it just LLM inference specific? It is like, well superfast... the only worry is literally the latency from you to the endpoint... so if its say, a streaming interruptible feed you are giving the model then the use cases for TTS and Speech applications just went through the damn roof!
@engineerprompt
@engineerprompt 5 ай бұрын
I am not sure but I was listening to Chamath (who is an investor in Groq) and he was talking about the initial use cases of the hardware. Seems like they were focused on vision so it might have the ability
@engineerprompt
@engineerprompt 5 ай бұрын
I am trying to put together an example for end to end speech conversation, let's see how that goes
@bobsmithy3103
@bobsmithy3103 5 ай бұрын
Can it run other models?
@Francotujk
@Francotujk 4 ай бұрын
What are the rate limits of the free api? Is it necessary to provide credit card?
@engineerprompt
@engineerprompt 4 ай бұрын
It's free at the moment and there is a rate limit as well. Seems to keep changing. Last time I checked, it was around 20 messages per minute
@ranaayushmansingh2368
@ranaayushmansingh2368 2 ай бұрын
can we fine tune this and use it?
@engineerprompt
@engineerprompt 2 ай бұрын
You can't fine-tune via their api yet.
@prestonmccauley43
@prestonmccauley43 5 ай бұрын
I tried a few things with this and it is incredibly fast.
@engineerprompt
@engineerprompt 5 ай бұрын
I agree!
@siriyakcr
@siriyakcr 5 ай бұрын
Wow
@CharlesDonboscoA
@CharlesDonboscoA 5 ай бұрын
Hi whether it's free or paid ?
@engineerprompt
@engineerprompt 5 ай бұрын
Free at the moment
@ZombieJig
@ZombieJig 5 ай бұрын
Fuck all these cloud only AI services, release the cards!
@thierry-le-frippon
@thierry-le-frippon 5 ай бұрын
Yes, otherwise they will fade away quickly. Their window of opportunity is small. Money is looking at eating in the nvidia cake now not tomorrow.
@conciousaizielia
@conciousaizielia 5 ай бұрын
Grok is not a llm it can run a llm
@TheJscriptor09
@TheJscriptor09 5 ай бұрын
YALLM ... it is almost becoming daily news ... Yet Another LLM.
@savire.ergheiz
@savire.ergheiz 5 ай бұрын
Fast but useless. These oss models still way far behind cgpt4.
@manishadeshmukh2519
@manishadeshmukh2519 4 ай бұрын
Bro groq outsmarts GPT-4 in 70B model
@manishadeshmukh2519
@manishadeshmukh2519 4 ай бұрын
It is super faster than gpt 4
@geo4design
@geo4design 5 ай бұрын
This is an AD
Chat with Documents is Now Crazy Fast thanks to Groq API and Streamlit
12:18
World’s Fastest Talking AI: Deepgram + Groq
11:45
Greg Kamradt (Data Indy)
Рет қаралды 43 М.
ISSEI & yellow girl 💛
00:33
ISSEI / いっせい
Рет қаралды 18 МЛН
Harley Quinn's desire to win!!!#Harley Quinn #joker
00:24
Harley Quinn with the Joker
Рет қаралды 9 МЛН
Survive 100 Days In Nuclear Bunker, Win $500,000
32:21
MrBeast
Рет қаралды 156 МЛН
I Analyzed My Finance With Local LLMs
17:51
Thu Vu data analytics
Рет қаралды 464 М.
Groq's AI Chip Breaks Speed Records
7:58
Groq
Рет қаралды 152 М.
CrewAI + Groq Tutorial: Crash Course for Beginners
32:10
codewithbrandon
Рет қаралды 22 М.
INSANELY FAST Talking AI: Powered by Groq & Deepgram
12:11
Prompt Engineering
Рет қаралды 10 М.
this is the fastest AI chip in the world: Groq explained
6:30
morethisdayinai
Рет қаралды 10 М.
How to build a real-time AI assistant (with voice and vision)
20:43
Conversation with Groq CEO Jonathan Ross
34:57
Social Capital
Рет қаралды 31 М.
Build Anything with Llama 3 Agents, Here’s How
12:23
David Ondrej
Рет қаралды 135 М.