Let's Recreate Google Translate! | Model Training

  Рет қаралды 5,366

Edan Meyer

Edan Meyer

Күн бұрын

Пікірлер: 25
@shenchiang3837
@shenchiang3837 3 жыл бұрын
Thanks for making this fantastic video, would love to learn more about that model.generate function!
@aswin0825
@aswin0825 2 жыл бұрын
I was wondering if you could help us with mc4 dataset in huggingface. Thanks for your effort! We really appreciate your efforts!
@neoncommander
@neoncommander 3 жыл бұрын
Great videos!! Greetings from Spain, you have earned a new subscriber.
@eah2886
@eah2886 2 жыл бұрын
it is possible to make this model on vs code or any other IDE by tweaking the code a little bit?
@梁大可-l5h
@梁大可-l5h 3 ай бұрын
Thank you!
@jayrcastro10
@jayrcastro10 3 жыл бұрын
How do the model knows that it will convert english to japanese by just adding it as a token?
@EdanMeyer
@EdanMeyer 3 жыл бұрын
That is because we format the training data that way. We can choose to do it in any other way we like. For example, we could start every sentence off with "Translate this from English to Japanese: ...", or we could use a completely random string of numbers to represent a translation pair. The important thing is that we are consistent throughout all the examples in our training and testing. So long as it is consistent, the model should learn the scheme we have come up with. The tags I used here could easily be changed and were somewhat arbitrary.
@jayrcastro10
@jayrcastro10 3 жыл бұрын
@@EdanMeyer Understood! thank you so much for explaining it!
@keburius2631
@keburius2631 Жыл бұрын
hello, thanks for this tutorials, very helpful. I have 1 question. if i set model max_length to 256 or 512 before training. will the result model have opportunity to return translated text with this lengths? if not how can i get this result? thank you in advance
@moseshu6917
@moseshu6917 2 жыл бұрын
Thanks for making this fantastic video, Compared with facebook fairseq framework, which model is best?
@titaniumiwnl
@titaniumiwnl Жыл бұрын
How do you deal with unkown tokens? Or tokens which are out if vocabulary for the model
@johndonne2170
@johndonne2170 2 жыл бұрын
Hi, Edan and sorry for my English! Please suggest whether it is possible to create a model for a language for which there are few parallel translations (about 10,000 lines)? How effective will the model trained on such small data be?
@dharaanv785
@dharaanv785 2 жыл бұрын
Hey Eden!, we really appreciate ur efforts and it was a really useful but i'm just curious to try it with mc4 Dataset in huggingface can u help us with the same!
@seyedmortezakamali2597
@seyedmortezakamali2597 2 жыл бұрын
Hi thanks for your content how can I make a convertor for programming languages for example python to c++
@vanduytran5490
@vanduytran5490 8 ай бұрын
thanks you sooo much
@umarlawan5092
@umarlawan5092 2 жыл бұрын
nice project, i realy like it
@charleenlozi4775
@charleenlozi4775 Жыл бұрын
thanks so much for this
@dislike__button
@dislike__button 2 жыл бұрын
Awesome content! Will it work if you train the model on a dataset with languages different from the ones the model was pretrained on?
@person-no2zs
@person-no2zs 2 жыл бұрын
My question too. Have you figured it out yet?
@EdanMeyer
@EdanMeyer 2 жыл бұрын
The bad news is that unfortunately it will almost certainly not work. The good news is that almost all the big language models have public checkpoints pretrained on multilingual data, so you can use one of those.
@dislike__button
@dislike__button 2 жыл бұрын
@@EdanMeyer thanks for the answer. The problem is I want to train it on a low resource language that is not used in any ml models.
@cupnoodles5956
@cupnoodles5956 Жыл бұрын
@@dislike__button I'm a year late but have you ever come up to a solution on this? If you don't mind I really wanted to learn and work on low resource language as well
@dislike__button
@dislike__button Жыл бұрын
@@cupnoodles5956 I ended up fine tuning Meta's m2m100 model, replacing one of the low resource languages it was trained on (Zulu) with my own. It worked pretty well.
Let's Recreate Google Translate! | Multilingual Data
28:53
Edan Meyer
Рет қаралды 8 М.
7 PyTorch Tips You Should Know
17:12
Edan Meyer
Рет қаралды 21 М.
iPhone or Chocolate??
00:16
Hungry FAM
Рет қаралды 42 МЛН
Spongebob ate Michael Jackson 😱 #meme #spongebob #gmod
00:14
Mr. LoLo
Рет қаралды 10 МЛН
Let's Recreate Google Translate! | Neural Machine Translation
15:34
NLP Project in 20 Minutes Using BERT
20:13
Edan Meyer
Рет қаралды 9 М.
LSTM is dead. Long Live Transformers!
28:48
Seattle Applied Deep Learning
Рет қаралды 528 М.
How Google Translate Works - The Machine Learning Algorithm Explained!
15:03
How to win a argument
9:28
ajaxkmr
Рет қаралды 576 М.
Inverse Reinforcement Learning Explained
19:59
Edan Meyer
Рет қаралды 10 М.
The spelled-out intro to language modeling: building makemore
1:57:45
Andrej Karpathy
Рет қаралды 713 М.
How to Make a Language Translator - Intro to Deep Learning #11
8:48
Building a Translator with Transformers
17:50
CodeEmporium
Рет қаралды 24 М.
Fine-tuning Large Language Models (LLMs) | w/ Example Code
28:18
Shaw Talebi
Рет қаралды 324 М.