Longformer: The Long-Document Transformer

  Рет қаралды 24,079

Yannic Kilcher

Yannic Kilcher

Күн бұрын

Пікірлер: 51
@ArnavArora
@ArnavArora 4 жыл бұрын
Great video! Very intuitive. Slight correction, when explaining the sliding window attention using the graphs in the paper, either the range should be [i-w/2, i+w/2] or the window size is 2w.
@YannicKilcher
@YannicKilcher 4 жыл бұрын
Absolutely true, thank you!
@gary9630
@gary9630 4 жыл бұрын
I just finished reading the paper and then watched the illustration you give in this video. TBH, you give such an informative and intuitive explanation! Nice job! I really love it.
@jeonghwankim8973
@jeonghwankim8973 3 жыл бұрын
Great explanation. You literally tap into the main part of the paper and explain it in the most intuitive way possible. Thank you.
@shairuno
@shairuno 4 жыл бұрын
I love how you are always skeptical and try to justify the claims from the paper. Great vid!
@DistortedV12
@DistortedV12 4 жыл бұрын
Wow that was fast! this paper just came out. Will check vid tomorrow.
@mnk6436
@mnk6436 2 жыл бұрын
Awesome video and great explanation! Thank you!
@sagumekishin5748
@sagumekishin5748 4 жыл бұрын
Speaking of convolution and attention, there are papers suggesting self attention can replace convolution completely in vision tasks. I think those papers are worth covering. "Stand-alone self-attention in vision models" "On the Relationship between Self-Attention and Convolutional Layers"
@priyamdey3298
@priyamdey3298 4 жыл бұрын
Very interesting line of work. Thanks for the info!
@laveenabachani
@laveenabachani 3 жыл бұрын
Thank you for making this. This was very helpful. Loved it!
@b.jardim2079
@b.jardim2079 4 жыл бұрын
Great video, you made it very clear! Thanks.
@Arwin_Unbeatable
@Arwin_Unbeatable 5 ай бұрын
Great video. A big thumbs up
@chrisber
@chrisber 2 жыл бұрын
This was excellent, thank you so much for your channel!
@kikimajo6850
@kikimajo6850 4 жыл бұрын
Love this series!
@OccultDemonCassette
@OccultDemonCassette 2 жыл бұрын
Hmm, I wonder why "special tokens" are turned off on a lot of the Collab tests I've seen on longformers. Seems like they would be beneficial?
@dummyrezajzadeh
@dummyrezajzadeh Жыл бұрын
beautifully explained thank you
@GeekProdigyGuy
@GeekProdigyGuy 4 жыл бұрын
If there are 2 layers with kernel size 3, and only the second layer is dilated skipping every other unit, the second layer will not "miss" any local information simply by the adjacent windows overlapping. So I don't think using dilation only at higher layers necessarily goes against the importance of locality.
@YannicKilcher
@YannicKilcher 4 жыл бұрын
Correct. Information can still aggregate with depth. My point was that their argument for the sliding window was the importance of locality and the dilation is directly counter to that. But yes, they solve that by only dilating the higher layers where they argue that locality does not matter as much anymore.
@qian2718
@qian2718 4 жыл бұрын
Didn't realize the memory consuming would be same until I watched the video😭
@iliemihai949
@iliemihai949 4 жыл бұрын
This video is so awesome, and well explained. You are a great teacher ! Also was wondering if you know if the document classification datasets that they use for testing are available ? Thank you and keep it up.
@АлексейТучак-м4ч
@АлексейТучак-м4ч 4 жыл бұрын
Idea inspired by global+sliding window: n1 nodes would connect to c1 previous nodes, n2 to c2, n3 to c3 etc. and they all are randomly shuffled
@YannicKilcher
@YannicKilcher 4 жыл бұрын
Nice idea, but you'd lose the inductive prior that neighbors are important
@شهریارصادقی-ت8ز
@شهریارصادقی-ت8ز 10 күн бұрын
great video
@mehdimashayekhi1675
@mehdimashayekhi1675 4 жыл бұрын
great job! keep it up
@drdca8263
@drdca8263 4 жыл бұрын
Question: could it work to have the model select on-the-fly some subset (with a small maximum size, like, at most 10 or so, idk) of nodes to treat as special, and as being able to connect to everything? (And, which nodes are counted as special would differ between layers) Like, have something estimate for each node how useful it would be to have that node be checked against all of the nodes, not just the nearby ones, and then when computing the dot products, and associated matrix, compute the dot products for the nearby pairs and the pairs where at least one of the two was selected as important? I imagine maybe that would be hard to train because not differentiable? And also might be slow to compute? Disclaimer: I don’t know what I’m talking about.
@YannicKilcher
@YannicKilcher 4 жыл бұрын
You're absolutely right. First, yes that could be a very valid idea and second, yes it would be very hard to train, because as soon as you introduce "hard" attention like this, you have no learning signal flowing back.
@drdca8263
@drdca8263 4 жыл бұрын
Yannic Kilcher Thanks!
@sauravmukherjeecom
@sauravmukherjeecom 4 жыл бұрын
Weirdly, your idea is very similar to bigbird. Your idea was definitely very valid. For Yanic's skepticism of information flowing back, they kept the subset same across layers and changed the subsets only across different sequences.
@riasingh2558
@riasingh2558 4 жыл бұрын
Intuitively, how is Longformer different from Transformer-XL? Finally, how do Transformer-XL, Longformer, and Linformer compare with each other if Long and Lin -former bot have linear complexity/ Thanks for great content!
@YannicKilcher
@YannicKilcher 4 жыл бұрын
In transformer-xl you don't explicitly train the carry-over mechanism, as I understand it. The linformer projects the sequence length down, while the longformer only attends to a sub-part of the sequence.
@freemind.d2714
@freemind.d2714 4 жыл бұрын
Great video as always!! Did you think those sliding window and dilated sliding windows is very similar to idea of WaveNet architecture ? for make the deeper layer gain the all input informations more efficient
@YannicKilcher
@YannicKilcher 4 жыл бұрын
It's certainly related, but the dilated convolutions will then pull the dilations together in the next layer, not sure if that's happening here.
@herp_derpingson
@herp_derpingson 4 жыл бұрын
It would be interesting to see what this model would look like under OpenAI microscope. If that is even possible.
@YannicKilcher
@YannicKilcher 4 жыл бұрын
Not directly. The microscope optimizes inputs in continuous space. Here you'd have to optimize the discrete text input. Not entirely clear how that would work.
@majorminus66
@majorminus66 3 жыл бұрын
If local attention is really what matters, why bother building a model that can attend over entire massive documents? Why do we need to overcome the 512 seq limit if local attention is really most important?
@LegoGunshipper
@LegoGunshipper Жыл бұрын
It achieves better results. There is no 'why'
@Timtom0707
@Timtom0707 4 жыл бұрын
If every way of coloring the matrix is a valid way of cutting down on the attention calculations, maybe it would be interesting to do some kind of architecture search over possible colorings? It seems unlikely that the assumptions they've made are the optimum - maybe there's room some kind of hierarchical structure? hmmm
@paramsraman3948
@paramsraman3948 3 жыл бұрын
Great review of the paper! Very clear and helpful.. Mind sharing what tools you use for the presentation? (to zoom in, annotate with your markers, get to whiteboard on the side of pdf etc).. It is pretty cool
@YannicKilcher
@YannicKilcher 3 жыл бұрын
It's just OneNote
@Chr0nalis
@Chr0nalis 4 жыл бұрын
Thx for the paper. Sound level is on the low side, perhaps you could look into normalizing the level before uploading.
@YannicKilcher
@YannicKilcher 4 жыл бұрын
Noted, thanks for the feedback.
@pratik6447
@pratik6447 3 жыл бұрын
In the long former model the config file has > and >. What does these 2 parameter means? Which one is token size?
@ty7521
@ty7521 2 жыл бұрын
4098 is the token size
@priyamdey3298
@priyamdey3298 4 жыл бұрын
A thought crossed my mind: Do you think special tokens act more like the cell states of LSTM?
@YannicKilcher
@YannicKilcher 4 жыл бұрын
That's very possible
@taku8751
@taku8751 4 жыл бұрын
It just aggregate all seperated parts processed by formal transformer.
@adamtran5747
@adamtran5747 2 жыл бұрын
i love it
@zingg7203
@zingg7203 2 жыл бұрын
The volume is really low. Have to max to hear your talk
@johngrabner
@johngrabner 4 жыл бұрын
A more symmetrical approach: split seq into n. 2 layers of attention, first across n, second first element of n fer every n. Am I missing something?
RWKV: Reinventing RNNs for the Transformer Era (Paper Explained)
1:02:17
Visualizing transformers and attention | Talk for TNG Big Tech Day '24
57:45
Beat Ronaldo, Win $1,000,000
22:45
MrBeast
Рет қаралды 158 МЛН
Sigma Kid Mistake #funny #sigma
00:17
CRAZY GREAPA
Рет қаралды 30 МЛН
BAYGUYSTAN | 1 СЕРИЯ | bayGUYS
36:55
bayGUYS
Рет қаралды 1,9 МЛН
Attention in transformers, step-by-step | DL6
26:10
3Blue1Brown
Рет қаралды 2 МЛН
Scaling Transformer to 1M tokens and beyond with RMT (Paper Explained)
24:34
Big Bird: Transformers for Longer Sequences (Paper Explained)
34:30
Yannic Kilcher
Рет қаралды 24 М.
Transformers explained | The architecture behind LLMs
19:48
AI Coffee Break with Letitia
Рет қаралды 29 М.
Cross Attention | Method Explanation | Math Explained
13:06
The Attention Mechanism in GPTs
1:01:14
AI Makerspace
Рет қаралды 1,4 М.
The complete guide to Transformer neural Networks!
27:53
CodeEmporium
Рет қаралды 37 М.
Linformer: Self-Attention with Linear Complexity (Paper Explained)
50:24