IR20.8 Learning to rank with an SVM
2:32
IR20.9 Learning to rank: features
1:11
IR20.5 SVM explained visually
10:01
9 жыл бұрын
IR20.2 Large margin classification
6:01
IR20.1 Centroid classifier
6:16
9 жыл бұрын
LM.9 Jelinek-Mercer smoothing
1:07
9 жыл бұрын
LM.7 Good-Turing estimate
11:03
9 жыл бұрын
LM.4 The unigram model (urn model)
1:45
LM.14 Issues to consider
2:39
9 жыл бұрын
LM.2 What is a language model?
2:07
9 жыл бұрын
LM.10 Dirichlet smoothing
2:21
9 жыл бұрын
LM.13 Language model ranking formula
0:54
LM.11 Leave-one-out smoothing
2:30
9 жыл бұрын
LM.5 Zero-frequency problem
2:03
9 жыл бұрын
LM.3 Query likelihood ranking
5:03
9 жыл бұрын
LM.1 Overview
2:17
9 жыл бұрын
BIR.17 Modelling term frequency
3:12
BIR.16 Linked dependence assumption
14:00
BIR.12 Example
3:52
9 жыл бұрын
BIR.3 Probability of relevance
2:14
9 жыл бұрын
Пікірлер
@thetrainoflife8327
@thetrainoflife8327 7 күн бұрын
Thank you Professor. It was super helpful !
@shubhayandas5127
@shubhayandas5127 8 күн бұрын
Score of document 3 should be 13 instead of 11.
@yourkidelias
@yourkidelias 19 күн бұрын
Great vid, thank you very much!
@thecheekychinaman6713
@thecheekychinaman6713 Ай бұрын
Excellent, clear introduction, thank you!
@Faisal-jo5vk
@Faisal-jo5vk Ай бұрын
how do i write this in python??
@Faisal-jo5vk
@Faisal-jo5vk Ай бұрын
the formula
@DavidRodriguez-j9b
@DavidRodriguez-j9b 2 ай бұрын
Clara Parkways
@TommyMichaelia-b3u
@TommyMichaelia-b3u 2 ай бұрын
Rozella Ways
@hayakanaan246
@hayakanaan246 2 ай бұрын
Amazing!
@sanketmole4314
@sanketmole4314 2 ай бұрын
Thank you sir 🙏
@ahmadjaradat3011
@ahmadjaradat3011 2 ай бұрын
Such a nice explanation!!!
@srinivastadinada4846
@srinivastadinada4846 3 ай бұрын
THANK YOUUUUU
@mkaberli614
@mkaberli614 3 ай бұрын
I wish those who make these videos would pay attention to the sound level. I can't complain too much because, other than the time invested, I didn't pay anything to watch it
@leeris19
@leeris19 3 ай бұрын
Now this is how you teach. Teaching should be an adventure filled with teeny tiny realizations to produce an outcome.
@Rrccccooo
@Rrccccooo 3 ай бұрын
i am algortithm
@erictao8396
@erictao8396 3 ай бұрын
Great explanation, thanks!
@RyeCA
@RyeCA 3 ай бұрын
excellent, thank you
@nikilragav
@nikilragav 4 ай бұрын
2:32 but that point (7,4) is actually closest to (5,4) so this k-dimensional tree to find nearest neighbor actually gives you the wrong answer?
@ishitaraj7723
@ishitaraj7723 4 ай бұрын
Very smooth explanation. Loved it!
@none4517
@none4517 4 ай бұрын
Good video. You should have used the arrow / laser tool. Which points are you talking about ?
@albertofabbri1543
@albertofabbri1543 4 ай бұрын
Very nice explanation! Thank you very much!
@Eonz
@Eonz 4 ай бұрын
I love data.
@Eonz
@Eonz 4 ай бұрын
K means this video.
@InvictusForever
@InvictusForever 4 ай бұрын
So helpful. Really lucky to have found this goldmine!!
@ApexPredator283
@ApexPredator283 4 ай бұрын
What happens if one of the centeroids does not get any datapoint assigned to it in the first clusterin round?
@sabbirneplumpstein334
@sabbirneplumpstein334 4 ай бұрын
You´re amazing
@kagan770
@kagan770 5 ай бұрын
# SUMMARY A discussion on web search algorithms, focusing on the impact of data quantity and link analysis techniques like PageRank. # IDEAS: - Web search engines handle staggering amounts of information, making architecture maintenance a significant challenge. - Google’s architecture processed 20 petabytes of data per day five years ago. - Large data volumes make computational tasks harder but simplify algorithmic processes. - A random subset of web pages is used to build search engine indexes. - Precision at rank 10 measures the accuracy of the top 10 search results. - Competitors with larger data sets can achieve higher precision in search results. - Distribution of scores for relevant and non-relevant documents remains unchanged with more data. - Precision at a fixed rank improves with increased data volume. - Search engines can improve rankings by increasing the amount of crawled data. - Larger data sets can outperform better algorithms if the latter have less data. - The density of relevant documents at the top of rankings affects precision improvements. - Historical example: Quill had an index size four times larger than Google’s. - Larger indexes lead to better search results if algorithms are comparable. - Precision as a function of rank generally decreases, with more relevant documents at the top. - More data in the index leads to better performance for free. - Link analysis techniques like PageRank are crucial for ranking web pages. - PageRank evaluates the importance of web pages based on link structure. - HITS algorithm identifies hubs and authorities in web content. - Combining large data sets with effective link analysis improves search engine performance. - Search engines must balance computational challenges with algorithmic efficiency. # INSIGHTS: - Large data volumes simplify algorithmic processes despite increasing computational challenges. - Precision at a fixed rank improves significantly with increased data volume. - Larger data sets can outperform better algorithms with less data. - The density of relevant documents at the top of rankings is crucial for precision improvements. - Combining large data sets with effective link analysis enhances search engine performance. # QUOTES: - "Google's architecture was churning through about 20 petabytes of data per day." - "Having that much data actually makes some things a lot easier." - "You can never get the entire web; nobody has the entire web." - "Precision at rank 10 would be 40%." - "The overall distribution of scores shouldn't change because you're just getting four times the data." - "Precision at a fixed rank will actually go up." - "The accuracy of the top page of your results depends on how much data you've crawled." - "Quill's index size was four times as big as Google's." - "If you have the same algorithms but four times as much data, you'll do better." - "Precision as a function of rank generally decreases." # HABITS: - Regularly update and maintain large-scale data architectures to handle vast information volumes. - Continuously gather and analyze large random samples of web pages for indexing. - Focus on improving both algorithmic processes and data collection efforts. # FACTS: - Google processed 20 petabytes of data daily five years ago. - No search engine has access to the entire web. - Larger data sets lead to higher precision in search results. - Quill had an index size four times larger than Google’s. # REFERENCES: - PageRank - HITS algorithm - Quill search engine # ONE-SENTENCE TAKEAWAY Increasing the amount of crawled data significantly improves search engine precision and performance. # RECOMMENDATIONS: - Regularly update and maintain large-scale data architectures for handling vast information volumes. - Continuously gather and analyze large random samples of web pages for indexing. - Focus on improving both algorithmic processes and data collection efforts. - Invest in gathering more data to enhance search engine precision and performance. - Combine large data sets with effective link analysis techniques like PageRank.%
@paedrufernando2351
@paedrufernando2351 5 ай бұрын
u speak like Jordan Belfort...lol
@DataWiseDiscoveries
@DataWiseDiscoveries 5 ай бұрын
Great collection of videos, Thoroughly loved it..
@archismanghosh7283
@archismanghosh7283 5 ай бұрын
You just cleared every doubts on this topic, it's 10 days before my exam watching your video and getting everything cleared
@glitchAI
@glitchAI 6 ай бұрын
why does the covariance matrix rotates the vectors towards the greatest variance?
@AI-HOMELAB
@AI-HOMELAB 6 ай бұрын
great explanation, simple and visualized. Thanks! =)
@amalalmuarik5160
@amalalmuarik5160 6 ай бұрын
THANKS, you've answered a lot of questions in my mind with your amazing explanation!!!!
@ebesko24
@ebesko24 6 ай бұрын
you sound like Gale Boetticher from breaking bad
@NickLilovich
@NickLilovich 7 ай бұрын
This video has (by far) the highest knowledge/time of any other video on this topic on KZbin. Clear explanation of the math and the iterative method, along with analogy to the simpler algorithm (k-means). Thanks Victor!
@ankitkusumakar7237
@ankitkusumakar7237 7 ай бұрын
Content is good, but please amplify audio.
@bunny_4_4_
@bunny_4_4_ 7 ай бұрын
When andrew tate explaining Math
@raihanpahlevi6870
@raihanpahlevi6870 7 ай бұрын
sir we cant see your cursor omg
@raihanpahlevi6870
@raihanpahlevi6870 7 ай бұрын
how to know the value of P(b) and P(a)
@wajahatmehdi
@wajahatmehdi 7 ай бұрын
Excellent explanation
@DereC519
@DereC519 8 ай бұрын
ty
@tazanteflight8670
@tazanteflight8670 8 ай бұрын
Its amazing this works at all, because the first step is to take a 2d image that makes sense, into a 1d image that has lost ALL spatial information. A 1d stream of pixels is not an image.
@deepakjoshi7730
@deepakjoshi7730 8 ай бұрын
Splendid. Example very well portrays the algorithm stepwise!
@nkapila6
@nkapila6 8 ай бұрын
Thanks for this. Your video helped bring clarity to the problem statement.
@saunakroychowdhury5990
@saunakroychowdhury5990 8 ай бұрын
but is not projection (y .e)e where y = x - mew
@raoufkeskes7965
@raoufkeskes7965 9 ай бұрын
at 3:08 the variance estimator shoud be divided by (nb-1) as corrected estimation and not nb .. that's what we call Bessel's correction
@yeah6732
@yeah6732 9 ай бұрын
Great tutorial! But why the slop of two eigenvectors are expected to be the same?!
@DrKnowsMore
@DrKnowsMore 10 ай бұрын
Outstanding!
@johanesalberto6136
@johanesalberto6136 11 ай бұрын
thanks brother
@azuriste8856
@azuriste8856 11 ай бұрын
Great Explanation Sir. I don't know why it motivated me to appreciate and comment on the video.
@samarthpardhi7307
@samarthpardhi7307 Жыл бұрын
Andrew Tate of machine learning