Sentence Transformers - EXPLAINED!

  Рет қаралды 26,253

CodeEmporium

CodeEmporium

2 жыл бұрын

Let's talk about Recurrent Networks, Transformer Neural Networks, BERT Networks and Sentence Transformers all in one video!
Follow me on M E D I U M: towardsdatascience.com/likeli...
Joins us on D I S C O R D: / discord
Please like and S U B S C R I B E: / codeemporium
INVESTING
[1] Webull (You can get 3 free stocks setting up a webull account today): a.webull.com/8XVa1znjYxio6ESdff
REFERENCES
[1] The Main paper that introduced Sentence Transformers: arxiv.org/pdf/1908.10084.pdf
[2] The paper that introduced BERT: arxiv.org/pdf/1810.04805.pdf
[3] How BERT is different from transformers: ai.stackexchange.com/question...
[4] The triplet loss and architecture described in page 5: openreview.net/pdf?id=hga6dk7...
[5] AWS K-NN elastic Search: docs.aws.amazon.com/opensearc...
[6] Approximate Nearest Neighbors algorithm used by spotify (ANNOY): github.com/spotify/annoy
[7] Colah’s blog that very well breaks down LSTM networks: colah.github.io/posts/2015-08...
[8] My video on LSTMs, explaining what you need to know: • LSTM Networks - EXPLAI...
[9] My original video on Transformer Neural Networks: • Transformer Neural Net...

Пікірлер: 39
@CodeEmporium
@CodeEmporium 2 жыл бұрын
Hey Everyone! Hope you're all doing super well. This video will give you everything you need to know about Transformer Neural Networks, BERT Networks and Sentence Transformers - or at least all that we can cover in 17 minutes. Hoping we all understand why these Architectures were developed the way they were, painting the picture as a fluid story. I'm trying another teaching style here. If you like this kind of video, please do let me know in the comments. Put a lot of effort into this, so I hope you think this is good! Enjoy! And Cheers!
@LiaAnggraini1
@LiaAnggraini1 2 жыл бұрын
Thank you! This is what I need for my thesis
@manash.b4892
@manash.b4892 9 ай бұрын
Wow. Thanks a lot for all these videos. I am self-studying beginner and your videos have been a boon. Keep up the good work, man!
@prasannabiswas2727
@prasannabiswas2727 2 жыл бұрын
Really the best info out. Thank you.
@HazemAzim
@HazemAzim 8 ай бұрын
really neat . Thank you , I was looking for nice stuff on SBERT with decent depth
@Daniel-gy1rc
@Daniel-gy1rc 2 жыл бұрын
dude you are amazing. Hope you keep this work up! Explaining complex things in an easy-to-follow and examplified way is a great skill!
@CodeEmporium
@CodeEmporium 2 жыл бұрын
Thanks a ton Daniel! Much appreciated complements :)
@simoneparvizi775
@simoneparvizi775 2 жыл бұрын
Hey man huge fan! Would you do a video about the "vanishing gradient problem"? Tbh I've been looking for a good video on it, but they're just not on point as you are....I'd really like your explanation on such argument! Keep up with the great work
@kevon217
@kevon217 Жыл бұрын
Excellent overview!
@TheHamoodz
@TheHamoodz Жыл бұрын
This channel has orders of magnitude more views than it deserves
@RaghavendraK458
@RaghavendraK458 2 жыл бұрын
Great video. Thanks
@WhatsAI
@WhatsAI 2 жыл бұрын
Amazing overview !
@user-vu4or4ih8p
@user-vu4or4ih8p 4 ай бұрын
Great stuff. Thanks
@JJ-dz2ne
@JJ-dz2ne 9 ай бұрын
Very informative, thank you!
@CodeEmporium
@CodeEmporium 9 ай бұрын
You are very welcome! Thanks for watching and commenting
@masteronepiece6559
@masteronepiece6559 2 жыл бұрын
Nice overview
@TheShadyStudios
@TheShadyStudios 2 жыл бұрын
Great choice!
@freedmoresidume
@freedmoresidume 2 жыл бұрын
Great video, thanks a lot
@CodeEmporium
@CodeEmporium 2 жыл бұрын
Welcome:)
@moslehmahamud9574
@moslehmahamud9574 2 жыл бұрын
this is good!
@clairewang8370
@clairewang8370 Жыл бұрын
This is 🔥!!!😍😍😍😍😍
@PritishMishra
@PritishMishra 2 жыл бұрын
Great video!!! Can we get some project videos on Transformer? As you showed in this video about the text-similarity with BERT so do you have any plan to create a video to do this with python?
@thekarthikbharadwaj
@thekarthikbharadwaj 2 жыл бұрын
Yes, really needed. Internet is lacking with an exact project developed using Transformers with proper backend information
@miriamramstudio3982
@miriamramstudio3982 Жыл бұрын
Great video. One part I didn't completely understood is the NLI part. Do you mean that after that NLI step, the mean pooling sentence vector of the newly trained BERT not be "poor" anymore? Thanks.
@NicholasRenotte
@NicholasRenotte 2 жыл бұрын
Oooooooh, this is so freaking cool!! When are we teaming up to build something?!
@CodeEmporium
@CodeEmporium 2 жыл бұрын
Dude. I will reach out ma guy (sorry i didn't before) :)
@NicholasRenotte
@NicholasRenotte 2 жыл бұрын
@@CodeEmporium ayyy no problemo man!
@keerthana2354
@keerthana2354 2 жыл бұрын
Can we use this for comparing two web articles?
@kestonsmith1354
@kestonsmith1354 2 жыл бұрын
My favourite model to train is T5. So much better . I don't like encoder models . I rather use a model that uses both encoder and decoder rather than either/or.
@shoukatali5671
@shoukatali5671 2 жыл бұрын
Приятно
@Han-ve8uh
@Han-ve8uh 2 жыл бұрын
Could you explain these 2 points in more detail? 3:21 transformers weren't designed to be language models + 16:35 transformers not complex enough to train a language model 1. What are language models supposed to do that transformers can't? My interpretation is that transformers do seq-seq tasks like translation, and translation needs a language model, so transformers are language models. Anything wrong with this thinking? 2. Can I say transformers are only invented to parallelize RNN family of models with attention? Any other obvious general or task specific benefits of transformers?
@mizoru_
@mizoru_ 2 жыл бұрын
I guess he means that they get better through improved pretraining (thus understand language better) From Papers with code: "BERT improves upon standard Transformers by removing the unidirectionality constraint by using a masked language model (MLM) pre-training objective. The masked language model randomly masks some of the tokens from the input, and the objective is to predict the original vocabulary id of the masked word based only on its context."
@Slayer-dan
@Slayer-dan Жыл бұрын
Thank you sir.
@norlesh
@norlesh 9 ай бұрын
5:11 Bidirectional Encoder Representation FROM Transformer (not of Transformers)
@InquilineKea
@InquilineKea Жыл бұрын
QUORAAAAA
@CodeEmporium
@CodeEmporium Жыл бұрын
AAAHHH
@roccococolombo2044
@roccococolombo2044 2 жыл бұрын
It is spelled chien nor chein
BERT Transformers for Language - EXPLAINED!
35:28
CodeEmporium
Рет қаралды 17 М.
Transformer Neural Networks - EXPLAINED! (Attention is all you need)
13:05
Smart Sigma Kid #funny #sigma #comedy
00:19
CRAZY GREAPA
Рет қаралды 11 МЛН
СНЕЖКИ ЛЕТОМ?? #shorts
00:30
Паша Осадчий
Рет қаралды 7 МЛН
The spelled-out intro to neural networks and backpropagation: building micrograd
2:25:52
CS480/680 Lecture 19: Attention and Transformer Networks
1:22:38
Pascal Poupart
Рет қаралды 339 М.
MIT 6.S191: Recurrent Neural Networks, Transformers, and Attention
1:01:31
Let's build GPT: from scratch, in code, spelled out.
1:56:20
Andrej Karpathy
Рет қаралды 4,4 МЛН
Transformer Neural Networks, ChatGPT's foundation, Clearly Explained!!!
36:15
StatQuest with Josh Starmer
Рет қаралды 593 М.
L6 Diffusion Models (SP24)
2:22:24
Pieter Abbeel
Рет қаралды 10 М.
Intro to Sentence Embeddings with Transformers
31:06
James Briggs
Рет қаралды 24 М.
LSTM Networks - EXPLAINED!
16:12
CodeEmporium
Рет қаралды 299 М.
Stanford CS25: V2 I Introduction to Transformers w/ Andrej Karpathy
1:11:41