ELMo | Lecture 57 (Part 1) | Applied Deep Learning

  Рет қаралды 942

Maziar Raissi

Maziar Raissi

Күн бұрын

Пікірлер: 1
@AITechTalks
@AITechTalks Жыл бұрын
Hello Dr. Raissi, Great videos by the way! You said that to get the Nearest Neighbors of the "play" vector in the biLM model we would have to push it through the LSTMs and produce the hidden state activations h_k,j for each layer of the LSTM. Then we get the h_k,j for that particular word play in the sentence. But, wouldn't it still need to be trained on a specific task to get the ELMo_k word representations for it and hence the nearest neighbors for the other sentences in the SemCor dataset they used in the paper? Since the ELMo_k^task has to have the trained gamma^task and s^task parameters I would think you would need to fine tune this to the specific task to get these nearest neighbor sentences with the "play" word vector.
GPT-1 | Lecture 57 (Part 2) | Applied Deep Learning
9:20
Maziar Raissi
Рет қаралды 2,1 М.
The LeetCode Fallacy
6:08
NeetCode
Рет қаралды 594 М.
Vector databases are so hot right now. WTF are they?
3:22
Fireship
Рет қаралды 1 МЛН
When Computers Write Proofs, What's the Point of Mathematicians?
6:34
Quanta Magazine
Рет қаралды 416 М.
Doc2Vec | Lecture 49 (Part 2) | Applied Deep Learning
14:08
Maziar Raissi
Рет қаралды 7 М.