TINY LM Agents on Edge Devices: Can We Scale?

  Рет қаралды 1,300

Discover AI

Discover AI

Күн бұрын

TINY Language Models are an absolutely interesting hot topic for the dev of AI agents on Edge Devices and multi-agent systems.
Latest insights from a technical point of view.
All rights w/ authors:
TinyHelen’s First Curriculum: Training and Evaluating
Tiny Language Models in a Simpler Language Environment
by Ke Yang, Volodymyr Kindratenko, ChengXiang Zhai
from University of Illinois Urbana-Champaign.
GitHub repo:
github.com/Emp...
#airesearch
#aiagents
#coding
#edge

Пікірлер
@code4AI
@code4AI 17 сағат бұрын
With the automatic audio dubbing from KZbin /Google you hear a synthetic voice in your regional language. To hear my original voice in English, switch to "Default" or "English" in the settings. Thank you.
@johnjac
@johnjac 15 сағат бұрын
I'm SOOOO excited for this. Thanks for making me aware.
@davidwynter6856
@davidwynter6856 12 сағат бұрын
I wonder if using a 2000 word vocabulary but instead of simple language use a specialized domain vocabulary for domain specific expert Tiny LLM might work. I.e. is it the language simplicity used or the vocabulary size that affects the result?
@AdamBrusselback
@AdamBrusselback 16 сағат бұрын
I've been wondering about progressively building up a language model by training these small models to be extremely strong for their parameter count, and then embedding them inside the layers of an LLM while freezing all of the small model layers except the first and last (to allow it to adapt to the larger model as it is trained). Would be interesting to see the results.
@AlexJohnson-g4n
@AlexJohnson-g4n 17 сағат бұрын
Tiny Language Models offer fascinating potential! They could transform AI in Edge Devices and multi-agent systems. Checking out the UIUC research. What practical challenges do we anticipate?
@fdavis1555
@fdavis1555 18 сағат бұрын
This is a very helpful concept!
@vrc5674
@vrc5674 44 минут бұрын
I wonder if you could apply the Meta's LCM technique to tokenize concepts rather the text and further improve the performance of the model. In a sense, I guess you're transferring some of the burden from the LLM to the tokenizer by doing this in that, the concept tokenizer itself would have to be trained. It might turn out to be more efficient for these smaller networks to work on wrangling concepts rather than wasting precious resources inside the tiny LLMs on converting tokens into concepts.
NEW: Better In-Context Learning ICL, Improved RAG (Harvard)
26:43
Support each other🤝
00:31
ISSEI / いっせい
Рет қаралды 81 МЛН
99.9% IMPOSSIBLE
00:24
STORROR
Рет қаралды 31 МЛН
Quando eu quero Sushi (sem desperdiçar) 🍣
00:26
Los Wagners
Рет қаралды 15 МЛН
LCM: The Ultimate Evolution of AI? Large Concept Models
30:13
Discover AI
Рет қаралды 55 М.
Multi Agents AI:  Simple Code w/ Smolagents
26:02
Discover AI
Рет қаралды 3,8 М.
AI Unpacked with Nobel Laureate Geoffrey Hinton
23:33
Valence
Рет қаралды 10 М.
Visualizing transformers and attention | Talk for TNG Big Tech Day '24
57:45
The Genius Way Computers Multiply Big Numbers
22:04
PurpleMind
Рет қаралды 247 М.
Netflix Removed React?
20:36
Theo - t3․gg
Рет қаралды 20 М.
Smarter Reasoning w/o RAG: SOLUTION for Short-Context LLMs
33:32
Discover AI
Рет қаралды 4,9 М.