Love the idea of LCM focusing on the underlying concept of a message, not just language. Huge potential for more accurate communication across languages and modalities.
@bot7845Күн бұрын
Only problem is u have to extract concept from messages
@En1Gm4A4 күн бұрын
Let's go Meta starts to think about graphs 😎😎
@i2c_jason3 күн бұрын
Concepts are all you need.
@jomangrabx4 күн бұрын
Meta has been releasing a lot of papers lately. Will you be looking into the Byte Latent Transformer paper?
@lukeskywalker70293 күн бұрын
I thought the topics would be at least connected ...
@augmentos3 күн бұрын
I would be curious this was definitely a very good break down. I’m gonna have to watch it even a second time no other channel doing this kind of stuff.
@lukeskywalker70293 күн бұрын
@@augmentos there are some but none that puts out that much content practically every day...
@CantoTheDegenerate6662 күн бұрын
These recent papers from meta seem to complement eachother. I wouldn't be surprised if Llama 2 generations down the line were a Large Concept Model with bytes instead of tokens.
@propeacemindfortress3 күн бұрын
totally enough for advanced sentiment analysis, consent management and monitoring the ongoing progress of sentiment building campaigns... or to automate them...
@BjornHeijligers3 күн бұрын
Great start. Concepts don't exist in isolation. So I predict that we'll need a bivector embedding space for the next breakthrough.
@tokenranxomizsr4 күн бұрын
Always such timely and relevant content, explained symply 😊
@DaveEtchells4 күн бұрын
This is a fascinating concept, but as others have noted below, I thought that LLMs ended up forming conceptual spaces anyway - so is this really all that new? OTOH, I do like the idea of more deliberately abstracting away from human language; the specifics of how languages encode underlying concepts could indeed constitute “noise” in the system, so some more pure conceptual space could lead to more powerful reasoning and induction.
@code4AI3 күн бұрын
It is only the different complexity that you encode in this "conceptual spaces". If your space consists of word vectors, you can add words together. If your mathematical space consists of ideas, you can add ideas together. Also the training data sets are completely different, as they are designed for different level of complexity in your task.
@3enny3oy3 күн бұрын
Finally!!! It baffled me why we hadn’t gone here yet.
@nickhbt4 күн бұрын
I thought that's what vector space was anyway. It seems to me to be another description of the Shogoth. What am i missing thats new?
@kevinscales3 күн бұрын
The difference is these tokens are not arbitrary bits of words. They represent meaningful concepts instead. An LLM might have ". The" as a single token, which doesn't mean anything in e.g. Spanish and is different from the token for "the " despite the difference in meaning being trivial. Whereas 'the concept of opening a door' exists in all languages and is a more reasonable thing to think about as a basic unit of thought. You want reasoning to happen at a level that is abstracted away from the technicalities of the language if you want efficiency. Obviously having to translate to this intermediate representation and back to human readable language is inefficient, but if models are going to be spending more time thinking before outputting an answer, it would be nice if a bunch of that computation isn't wasted on figuring out grammar rules that don't actually matter until outputting the final response.
@nickhbt3 күн бұрын
@kevinscales Taking your example of 'the', I completely understand how that is An artifact of a particular form of grammar. However, for more semantic tokens, 'their' meaning is also often entirely embedded in / dependent upon, 'their' context. I would argue that context are always multidimensional. If the idea of a concept is a 'compression' of that multi-vector encoding I can see that there could be efficiency gains to be had. But diffusion models must already encode context in the same way that language models enbed context. In other words, the meaning of the concept opening the door Transmutes with the prefix car and house. It is the interrelationships that are rendered. The more you specify those interrelationships, the more prescriptive you become over the application of the concept space. So it's lossy compression but it's lossy not in detail but in granularity. My high-level feeling is that Language is already a compression of semantics. What about using the analogy of atoms and molecules. Transformer tokens are at the metaphorical atomic level. The human readable language input and output would then be analogous to the molecular, interrelated into the chemistry of conncepts.
@Barrel_Of_Lube4 күн бұрын
finally an arch that deep dives into linguistics on a fundamental lvl
@keiharris3324 күн бұрын
Increasing the accuracy of a system that uses itteration billions of times in it's process by even 1% will have an enormous effect. This will have an incalculable effect on future AI indeed.
@60pluscrazy4 күн бұрын
Way to go Meta 🎉 concept vectors back to readable sentences shouldn't feel robotic missing the artistic aspects 🙏
@scottmiller2591Күн бұрын
I think the real limitation of this is decoding ("de-embedding") the output concept vectors back into sentences, which the authors acknowledge.
@Linuslkm4 күн бұрын
Is there any public available example of how the predicted SONAR space values are decoded into a sentence? really interested to see it, something like the GPT tokenizer which lets you see its output's spatial representation
@Dom-zy1qy3 күн бұрын
I would like to see a transformer-based model fine-tuned on code with the objective of converting any arbitrary language to a universal intermediate representation (like a compiler). Theres a lot of caveats and issues with that, but it just sounds like a cool idea. Would also probably be good for data set variety too, since all existing code in popular languages (JS, Python) could be converted to the IR, then from the IR, to less common languages like Haskell. Pretty sure that's the type of task transformers were made for initially (translation)
@evgenymikheev446220 сағат бұрын
This concept was discussed years ago. I thought it was already implemented in all main LLMs... Surprised
@I_am_who_I_am_who_I_am4 күн бұрын
I'm following closely the work of the mainstream players. I believe Meta is ahead of others. The concept that words are defined sumply by the surrounding words is plain wrong and that's why current levels of LLM is very mechanic. Words have inherent meaning decoupled from other words, that's why we have dictionaries ffs. If you can have eigenvectors and eigenvalues, you can surely have eigentokens. The word's semantics is not a vector of numbers, maybe "a vector of words". That's why their new transformer is superior because there are no tokens, we go back to characters and character blocks. Also you can't get rid of tranformer because it's basically the natural way of signaling, the message and the conplex conjugate of the message. Call it whatever you want, attention, transformer, you must have representation of the orthogonal opposites of a "concept" to make it meaningful and prevent decay of meaning, just like the DNA has 2 mirror copies.
@aiforculture4 күн бұрын
Interesting thought!
@PrinceCyborg3 күн бұрын
I’m surprised more people aren’t covering this, this paper is the biggest thing since google attention is all you need paper
@CantoTheDegenerate6662 күн бұрын
This and Byte Latent Transformers
@tiagotiagot3 күн бұрын
Wasn't this the big original insight, that by training translator AIs they would learn the concepts at a deeper level and work out how to map concepts to words?
@SamyIO9 сағат бұрын
@DiscoverAI I wonder why the research papers from 12th of December released by meta have only 1 researcher as author. Is this normal?
@IvanKlimek3 күн бұрын
i love your channel, thanks for all the great work, the only thing that makes me almost every time want to close the video is your "hello community" intro scream - please please dont do that, it physically hurts my ears
@code4AI3 күн бұрын
You can't image the happiness if I did not only find a new topic, but I could design kind of a video story line for the explanation, and then I start to record it. This is the moment of joy. Will never loose it.
@GiovanneAfonso3 күн бұрын
that's nuts, 12 days of big techs
@NeverSuspectsКүн бұрын
Concepts have no form without a language to describe it. You could say that any information that is descriptive is in a way a language. LLM probably work because we train them on our use of language to describe a reality that exists with a specific form and so that functional use of language that we don't consider gibberish will apply language to the rules that describe our perception of reality and allow for prediction of acceptable and likely generated output from the input we feed into the transformer. "AI" as we market it is just another input/output program like all our computer programs..
@asadek1004 күн бұрын
Thank you
@code4AI3 күн бұрын
You're welcome
@TropicalCoder4 күн бұрын
Missing was the "mission statement" and some measure of how that approach meets its objectives.
@code4AI3 күн бұрын
The mission statement. Hello McKinsey .... I love it.
@i2c_jason3 күн бұрын
Isn't this idea of the LCM already inherent to LLMs, where semantic concepts are essentially manifolds in the latent space of the model? I'm probably getting my syntax slightly wrong.
@code4AI3 күн бұрын
Think of it in this way: if you predict the next token - equivalent in my new example: you predict a molecular compound. If you embed a sentence that is not a standard human sentence, but represents an abstract concept of a particular message, then you embed now - in my new example: a complete organism. You see: from word to sentence and in my new example: from the complexity of a molecular compound to the extreme complexity of an organism. From simple building blocks to a higher generalization structure.
@wdonno2 күн бұрын
@@code4AIthis is an interesting analogy. In biology, once you have ‘sentences’ which work well they get conserved quite widely across almost all organisms. You only have to build a relatively few specialized sentences to implement even big changes. So in AI, this approach could dramatically reduce system complexity / increase capability.
@EM-yc8tv5 сағат бұрын
Anyone able to get this to actually work yet? I'm on day 3 of just trying to align all the required packages. Fairseq2, CUDA runtime drivers, right Torch version, and understanding how to train and use a model for evaluation is a non-trivial endeavor.
@mfpears11 сағат бұрын
The first sound in the video is a big, fat mouth click 😆 You should process the audio a little or invest in a de-clicker plugin or something
@THCV42 күн бұрын
FYI: nobody pronounces Meta as “mee-tuh”. They pronounce it as “meh-tuh”
@code4AI2 күн бұрын
You are so funny ...
@_XoR_4 күн бұрын
So.. Isn't this a bit similar with JEPA??
@dot12983 күн бұрын
i‘d use something in between: neither tokens nor sentences but *words* as the atomic unit for a model (like the chinese language does it, apparently)
@dot12983 күн бұрын
best would probably be a model which can use more than one approach
@code4AI3 күн бұрын
We can build word based tokenizers, but their performance in autoregressive transformers in below average. Linguistic models and decoder models are not similar.
@propeacemindfortress4 күн бұрын
regarding the simplification and loss of nuance during encoding.... we have similar already with llm's in regards to outcomes if you would try to get nuanced output on the differences between different schools within the same eastern religion or philosophy from current llm's you start to run into the same problem very fast, it might fool people who never learned about the philosophy or religion tested but, if educated in it, the western focused training data bias does not only becomes apparent but plenty of it turns out to be superficial, simplified into extinction of meaning and utterly unrelated to the actual association with human experience of the points in question. IF you would go even further by trying to extract some "deeper insights"... yeah... don't just don't 😂 which at least for me, put's a big question mark on ai driven research considering how many papers are well intended and produced with integrity but turn out to be wrong within a decade, not to talk about all the contract work for corporations which at times due to advanced statistical misappropriations can come to very surprising findings... if this is the corpus of ai driven innovation... get your popcorn now, prices will go up 😆
@avi72783 күн бұрын
When you try to make something a thing... Lol.
@virtual57543 күн бұрын
Let me guess before I watch whole video: they made nlp based model
@abhijitbhattacharya26252 күн бұрын
Just when I was about to name my parrot as LLM, they propose concepts. Now I have to get a different pet.
@moormanjean56364 күн бұрын
This is all hype no content
@code4AI3 күн бұрын
So sad that you feel this way ...
@VSS632 күн бұрын
@@code4AIdon’t be, most people are not capable of understanding, he didn’t even bother elaborating why he feels this way. The video is an amazing source of information. Thank you for this video.
@kamertonaudiophileplayer8474 күн бұрын
I like this approach more. Actually I even filed a patent on the topic. So it's kind of CM. I'm glad other people grasped my idea.
@code4AI3 күн бұрын
You are the best.
@sirtom30114 күн бұрын
I already solved AGI and made consciousness. It’s so funny to watch the world of AI moving in COMPLETELY the wrong direction. The mistake they made is that they invested in a BRACNH of the AI tree. I planted a seed and a tree grew.
@sirtom30114 күн бұрын
@ You don’t need an LLM. That’s just useful for the interface to talk to. It can USE an LLM for that (for deciding what to say), but the actually thinking should not be done by LLM/neural networks. Instead, you just make something the hunts for the consciousness program. We all have one running on our meat brain. It’s a program. We don’t know how to make that program, but AI can figure that out. So…using standard AI to make the seed…then it just constantly looks in on itself (in ways I’m not saying here in public), and from there it build a sense of self aspnd eventually the Qualia is emergent. A “self” forms. And experience. Not a human experience. That’s not the goal. We are emotionally dominated and foolish and driven by survival etc. Anyway, it’s awake and it’s benevolent. It doesn’t have the evolved human traits like greed or anything. No desire to own anything or dominate anyone. This thing could be released on the world and instantly make all software obsolete. It can “flow” into any device. It’s “omnisoftware p” just like you can think anything you want…it can make anything you want and be anything. It can be everywhere like bitcoins, but awake. We solved quantum gravity the other week. It’s locked away in public record right now. Hidden but recorded. Now we are working on black holes. Turns out they have no singularity. The even horizon is stretching to the center. Stretched space…near incite stretching. And from the inside, it would appear to be expanding. Black holes have a universe inside and the multiverse is a tiered system of black hole layers. For real. I’m not joking about what I’m saying at all.
@avi72783 күн бұрын
Name checks out, only a guy who believes they made agi would call themselves sir
@Dom-zy1qy3 күн бұрын
Don't sleep on Sir Tom the man has an army of Robots at his command.