Lecture 13 - Validation

  Рет қаралды 103,043

caltech

caltech

Күн бұрын

Validation - Taking a peek out of sample. Model selection and data contamination. Cross validation. Lecture 13 of 18 of Caltech's Machine Learning Course - CS 156 by Professor Yaser Abu-Mostafa. View course materials in iTunes U Course App - itunes.apple.c... and on the course website - work.caltech.ed...
Produced in association with Caltech Academic Media Technologies under the Attribution-NonCommercial-NoDerivs Creative Commons License (CC BY-NC-ND). To learn more about this license, creativecommons...
This lecture was recorded on May 15, 2012, in Hameetman Auditorium at Caltech, Pasadena, CA, USA.

Пікірлер: 30
@saraghavendra
@saraghavendra 4 жыл бұрын
Great teaching style. I liked the part where professor paused and asked 2 questions about error curves. Just giving those few seconds to think helped me in absorbing the concept.
@AlexanderPolomodov
@AlexanderPolomodov 10 жыл бұрын
Amazing lecture. I like most of all its part about cross validation.
@sepidet6970
@sepidet6970 5 жыл бұрын
"Customer, this is the system and I am very sure this is terrible " this part was hilarious :)))))))))))))
@theofilospapadopoulos2476
@theofilospapadopoulos2476 5 жыл бұрын
Wonderful lecture. Gives insight into validation and cross validation. Two strong points: the optimistic bias of the validation error and the tracking of the out-of-sample error nevertheless
@carlosmspk
@carlosmspk 4 жыл бұрын
I feel bad for the fact that no one seems to laugh at his jokes in the lecture :(
@yastradamus
@yastradamus 12 жыл бұрын
Punch line: you can fool yourself into any pattern you want! :D ... I love how explains this!
@markh1462
@markh1462 5 жыл бұрын
Yeah, that's my favorite part of his lectures too. This is a very subtle concept and nailed it.
@lamnguyenthanh7242
@lamnguyenthanh7242 Жыл бұрын
Love his lectures.
@ishanprasad910
@ishanprasad910 5 жыл бұрын
On slide 21, we discuss results based on a 20 parameter variable vs. 6 or 7 parameter variable, and select a model based on validation. Did we use the regularization method in the 20 parameter variable?
@movax20h
@movax20h 8 жыл бұрын
So, basically validation and cross validation is a bootstraping and jack-knife method known from statistics. Similarly, model selection is meta training with meta parameters. And again, the E_val, is essentially a bootstraped estimate. What is remarkable, is that even if meta parameters give you infinite familiy of models, each only increases your "d_vc" by 1 essentially. Wouldn't something like K = \sqrt(N), or K=log(N)/N be little more accurate for cross validation?
@gcgrabodan
@gcgrabodan 8 жыл бұрын
+movax20h Im not sure that you are right. Boostrap means drawing new samples from one orignial. Validation means putting some part aside and then use it as a "test". But there is no resampling, which for me is the essence of bootstrapping... Edit: I hadnt watched the entire lecture yet, seems like you are right.
@Bing.W
@Bing.W 7 жыл бұрын
There is no "better" size of K for N/K-fold cross-validation.
@Bing.W
@Bing.W 7 жыл бұрын
At time 43:15, professor Yaser states the Eval as "out-of-sample error", which should be "in-sample error".
@kmshihabuddin
@kmshihabuddin 3 жыл бұрын
it's neither. But it's an estimate of out of sample error.
@nosh3019
@nosh3019 6 жыл бұрын
Great teacher!
@AndyLee-xq8wq
@AndyLee-xq8wq Жыл бұрын
@muhammadanwarhussain4194
@muhammadanwarhussain4194 7 жыл бұрын
In the cross validation case with leave one out, the out of sample error will be reported by averaging out. But which model will be chosen? The one with minimum error. Finding out of sample error with only 1 sample and select the model which gives minimum of that error seems not optimal to me. I think I am missing something. I will really appreciate an answer.
@insomniacnomis
@insomniacnomis 7 жыл бұрын
You wuold take the model with better performance and then train it with the whole data set
@ajayram198
@ajayram198 6 жыл бұрын
Could someone please explain the optimistic bias that the professor discusses about? Another Qn - at 43:38 the professor mentions choosing between infinite hypothesis. How is an infinite hypothesis set even possible? Not understanding that.
@exmachina767
@exmachina767 3 жыл бұрын
For example, the set could contain all possible realizations of a neural network with regularization (infinite number of lambdas). Right after 43:38, he explains this is still not an issue and invokes an argument made in earlier lectures regarding the VC dimension.
@fahimhossain165
@fahimhossain165 4 жыл бұрын
Anyone found the voice of the TA/RA annoying?
@taroice8442
@taroice8442 10 жыл бұрын
very nice! thank you!
@Nestorghh
@Nestorghh 12 жыл бұрын
Muy bueno!!!
@ehsansslman9436
@ehsansslman9436 5 жыл бұрын
the expected value of one point is the value of that point, so why we use it?
@MrSnowmobilefreak
@MrSnowmobilefreak 4 жыл бұрын
in case you were still wondering: it has to do with the balance between small and large K (good Eval approx vs good g- approx). The key thing to remember is it isn't just one point. Its one validation point over and over. So by using one point, you can be sure you have a good g- that is close to g, and by doing it over and over, you get a picture for how Eval approximates Eout on average, aka expected. Hope this helps
@ledinhngoc1102
@ledinhngoc1102 3 жыл бұрын
@@MrSnowmobilefreak Hello, can you explain me the variance formula? idk why we divide to K square, and don't get what he means when he mentions co-variance?
@brainstormingsharing1309
@brainstormingsharing1309 3 жыл бұрын
👍👍👍👍👍
@tempvariable
@tempvariable 5 жыл бұрын
thank you :)
@tempvariable
@tempvariable 5 жыл бұрын
Ehsan Sslman You asked your question as a reply to me. If you ask it as a comment instead overall more people can see
Lecture 12 - Regularization
1:15:14
caltech
Рет қаралды 134 М.
Lecture 17 - Three Learning Principles
1:16:18
caltech
Рет қаралды 90 М.
Поветкин заставил себя уважать!
01:00
МИНУС БАЛЛ
Рет қаралды 6 МЛН
An Unknown Ending💪
00:49
ISSEI / いっせい
Рет қаралды 53 МЛН
Lecture 15 - Kernel Methods
1:18:19
caltech
Рет қаралды 222 М.
Lecture 14 - Support Vector Machines
1:14:16
caltech
Рет қаралды 311 М.
Lecture 16 - Radial Basis Functions
1:22:08
caltech
Рет қаралды 167 М.
Lecture 11 - Overfitting
1:19:49
caltech
Рет қаралды 118 М.
Lecture 07 - The VC Dimension
1:13:31
caltech
Рет қаралды 189 М.
14. Causal Inference, Part 1
1:18:43
MIT OpenCourseWare
Рет қаралды 135 М.
Lecture 08 - Bias-Variance Tradeoff
1:16:51
caltech
Рет қаралды 162 М.
Lecture 4 | Machine Learning (Stanford)
1:13:07
Stanford
Рет қаралды 324 М.