I wonder . You are beautiful and intelligent speak english well Impressed by you
@sadface74572 жыл бұрын
🥳 thank you w&b for bring us this content
@jiashuhan90112 жыл бұрын
8:19 It's correct. Support from china! Love your videos ❤
@DerPylz2 жыл бұрын
My message to Google: "Dataset or it didn't happen!"
@sadface74572 жыл бұрын
does matter once the dataset is in petabyte range?
@DerPylz2 жыл бұрын
@@sadface7457 of course it does. If the dataset is open source, people can use the parts that are relevant for them. No need to waste the energy and work hours needed to generate it, again. Publishing a "scientific" paper and not providing the data is so weird to me. Especially when we're living in the age of the reproducibility crisis. Maybe I'm an idealist, but I find science is not science if it's not for the progress of society. It's just internal R&D of a company.
@dinhanhx2 жыл бұрын
@@DerPylz it's called opendata, not opensource.
@DerPylz2 жыл бұрын
@@dinhanhx of course, thanks for pointing out my typo.
@niofer72472 жыл бұрын
Great summary. Love your videos, always great explanations!!!
@fahimfaisal19562 жыл бұрын
thanks Letitia for the nice explanation......keep going strong :) :)
@marcocipriano59222 жыл бұрын
I love when she says "yes, you guess correctly!" Me still sleepy in bed: ...did I?
@Aldraz2 жыл бұрын
Would be great if Google improved the already implemented languages. Translating from Japanese to English is still a huge pain and a complete nonsense most of times.
@nourhesham8942 жыл бұрын
Love your videos! Wondering if you can explain the SegFormer paper in easy terms!
@AICoffeeBreak2 жыл бұрын
Writing it onto THE list. Did not hear of the paper before, will have a look.
@harumambaru2 жыл бұрын
8:19 - I hope this is correct. Comment below if not
@AICoffeeBreak2 жыл бұрын
Let's check how well Google Translate works for Chinese. 😅
@harumambaru2 жыл бұрын
@@AICoffeeBreak Lets wait for a viewer with at least 2 languages, English and Chinese to find out the truth
@AICoffeeBreak2 жыл бұрын
⌚⏳
@ANTIMONcom2 жыл бұрын
Perhaps now google translate will be a bit less shit for all translations that are not to or from English 😑
@Skinishh2 жыл бұрын
Great video once again! 👏 Question: How does the model know the target language to translate to? Is there a token in the input that states the task at hand? or maybe the first token in the decoder
@AICoffeeBreak2 жыл бұрын
Hi, thanks for the question. I would have replied earlier, but I was in vacation while reading the comment and couldn't check the paper. Now I did check it. Yes, your intuition is right: they append a language token. Papers usually do not agree on whether to add (1) the source language token in the encoder and the target language to the decoder, or to (2) append a target language to the source sequence (encoder). On page 13, it says: "Different from Siddhant et al. (2022), in addition to the token that was prepended to the source sequence to signify the target language for both translation and MASS tasks, we add a token ( for the translation task, and for the MASS task) that specifies the task to be performed by the model. We find this to be critical for zero-resource performance, especially when model sizes are scaled up. In the absence of this task token, our models learnt to ‘infer’ the task from the source languages instead of relying on the token, resulting in copying sentences when provided zero-resource language sentences with the token."
@Skinishh2 жыл бұрын
@@AICoffeeBreak thanks! 🙏
@VV-fl8fi2 жыл бұрын
As a Chinese, the sentence appeared at 8:20 is correct hahaha
@AICoffeeBreak2 жыл бұрын
Phew, good to know that Google Translate worked in this case. 😅