Victor Lavrenko
Victor Lavrenko
  • 421
  • 9 949 592

Відео

IR20.8 Learning to rank with an SVM
Переглядів 10 тис.8 років тому
IR20.8 Learning to rank with an SVM
IR20.10 Learning to rank with click data
Переглядів 6 тис.8 років тому
IR20.10 Learning to rank with click data
IR20.9 Learning to rank: features
Переглядів 3,5 тис.8 років тому
IR20.9 Learning to rank: features
IR20.7 Learning to rank for Information Retrieval
Переглядів 7 тис.8 років тому
IR20.7 Learning to rank for Information Retrieval
IR20.3 Passive-aggressive algorithm (PA)
Переглядів 10 тис.8 років тому
IR20.3 Passive-aggressive algorithm (PA)
IR20.5 SVM explained visually
Переглядів 11 тис.8 років тому
IR20.5 SVM explained visually
IR20.2 Large margin classification
Переглядів 6 тис.8 років тому
IR20.2 Large margin classification
IR20.1 Centroid classifier
Переглядів 15 тис.8 років тому
IR20.1 Centroid classifier
IR20.4 Convergence of the PA algorithm
Переглядів 3,2 тис.8 років тому
IR20.4 Convergence of the PA algorithm
IR20.6 Sequential minimal optimization (SMO)
Переглядів 16 тис.8 років тому
IR20.6 Sequential minimal optimization (SMO)
LM.9 Jelinek-Mercer smoothing
Переглядів 6 тис.8 років тому
LM.9 Jelinek-Mercer smoothing
LM.7 Good-Turing estimate
Переглядів 6 тис.8 років тому
LM.7 Good-Turing estimate
LM.4 The unigram model (urn model)
Переглядів 8 тис.8 років тому
LM.4 The unigram model (urn model)
LM.14 Issues to consider
Переглядів 1,2 тис.8 років тому
LM.14 Issues to consider
LM.8 Interpolation with background model
Переглядів 2,4 тис.8 років тому
LM.8 Interpolation with background model
LM.2 What is a language model?
Переглядів 21 тис.8 років тому
LM.2 What is a language model?
LM.10 Dirichlet smoothing
Переглядів 5 тис.8 років тому
LM.10 Dirichlet smoothing
LM.13 Language model ranking formula
Переглядів 2,2 тис.8 років тому
LM.13 Language model ranking formula
LM.11 Leave-one-out smoothing
Переглядів 1,4 тис.8 років тому
LM.11 Leave-one-out smoothing
LM.5 Zero-frequency problem
Переглядів 6 тис.8 років тому
LM.5 Zero-frequency problem
LM.3 Query likelihood ranking
Переглядів 8 тис.8 років тому
LM.3 Query likelihood ranking
LM.1 Overview
Переглядів 9 тис.8 років тому
LM.1 Overview
LM.6 Laplace correction and absolute discounting
Переглядів 6 тис.8 років тому
LM.6 Laplace correction and absolute discounting
LM.12 Smoothing and inverse document frequency
Переглядів 2,6 тис.8 років тому
LM.12 Smoothing and inverse document frequency
BIR.10 Estimation with relevant examples
Переглядів 2,4 тис.8 років тому
BIR.10 Estimation with relevant examples
BIR.17 Modelling term frequency
Переглядів 9288 років тому
BIR.17 Modelling term frequency
BIR.16 Linked dependence assumption
Переглядів 7358 років тому
BIR.16 Linked dependence assumption
BIR.12 Example
Переглядів 2,4 тис.8 років тому
BIR.12 Example
BIR.3 Probability of relevance
Переглядів 4,7 тис.8 років тому
BIR.3 Probability of relevance

КОМЕНТАРІ

  • @kagan770
    @kagan770 12 днів тому

    # SUMMARY A discussion on web search algorithms, focusing on the impact of data quantity and link analysis techniques like PageRank. # IDEAS: - Web search engines handle staggering amounts of information, making architecture maintenance a significant challenge. - Google’s architecture processed 20 petabytes of data per day five years ago. - Large data volumes make computational tasks harder but simplify algorithmic processes. - A random subset of web pages is used to build search engine indexes. - Precision at rank 10 measures the accuracy of the top 10 search results. - Competitors with larger data sets can achieve higher precision in search results. - Distribution of scores for relevant and non-relevant documents remains unchanged with more data. - Precision at a fixed rank improves with increased data volume. - Search engines can improve rankings by increasing the amount of crawled data. - Larger data sets can outperform better algorithms if the latter have less data. - The density of relevant documents at the top of rankings affects precision improvements. - Historical example: Quill had an index size four times larger than Google’s. - Larger indexes lead to better search results if algorithms are comparable. - Precision as a function of rank generally decreases, with more relevant documents at the top. - More data in the index leads to better performance for free. - Link analysis techniques like PageRank are crucial for ranking web pages. - PageRank evaluates the importance of web pages based on link structure. - HITS algorithm identifies hubs and authorities in web content. - Combining large data sets with effective link analysis improves search engine performance. - Search engines must balance computational challenges with algorithmic efficiency. # INSIGHTS: - Large data volumes simplify algorithmic processes despite increasing computational challenges. - Precision at a fixed rank improves significantly with increased data volume. - Larger data sets can outperform better algorithms with less data. - The density of relevant documents at the top of rankings is crucial for precision improvements. - Combining large data sets with effective link analysis enhances search engine performance. # QUOTES: - "Google's architecture was churning through about 20 petabytes of data per day." - "Having that much data actually makes some things a lot easier." - "You can never get the entire web; nobody has the entire web." - "Precision at rank 10 would be 40%." - "The overall distribution of scores shouldn't change because you're just getting four times the data." - "Precision at a fixed rank will actually go up." - "The accuracy of the top page of your results depends on how much data you've crawled." - "Quill's index size was four times as big as Google's." - "If you have the same algorithms but four times as much data, you'll do better." - "Precision as a function of rank generally decreases." # HABITS: - Regularly update and maintain large-scale data architectures to handle vast information volumes. - Continuously gather and analyze large random samples of web pages for indexing. - Focus on improving both algorithmic processes and data collection efforts. # FACTS: - Google processed 20 petabytes of data daily five years ago. - No search engine has access to the entire web. - Larger data sets lead to higher precision in search results. - Quill had an index size four times larger than Google’s. # REFERENCES: - PageRank - HITS algorithm - Quill search engine # ONE-SENTENCE TAKEAWAY Increasing the amount of crawled data significantly improves search engine precision and performance. # RECOMMENDATIONS: - Regularly update and maintain large-scale data architectures for handling vast information volumes. - Continuously gather and analyze large random samples of web pages for indexing. - Focus on improving both algorithmic processes and data collection efforts. - Invest in gathering more data to enhance search engine precision and performance. - Combine large data sets with effective link analysis techniques like PageRank.%

  • @paedrufernando2351
    @paedrufernando2351 14 днів тому

    u speak like Jordan Belfort...lol

  • @DataWiseDiscoveries
    @DataWiseDiscoveries 19 днів тому

    Great collection of videos, Thoroughly loved it..

  • @archismanghosh7283
    @archismanghosh7283 21 день тому

    You just cleared every doubts on this topic, it's 10 days before my exam watching your video and getting everything cleared

  • @glitchAI
    @glitchAI 24 дні тому

    why does the covariance matrix rotates the vectors towards the greatest variance?

  • @TheTechPhilosopherTTPVLOGS
    @TheTechPhilosopherTTPVLOGS Місяць тому

    great explanation, simple and visualized. Thanks! =)

  • @amalalmuarik5160
    @amalalmuarik5160 Місяць тому

    THANKS, you've answered a lot of questions in my mind with your amazing explanation!!!!

  • @ebesko24
    @ebesko24 Місяць тому

    you sound like Gale Boetticher from breaking bad

  • @NickLilovich
    @NickLilovich Місяць тому

    This video has (by far) the highest knowledge/time of any other video on this topic on UA-cam. Clear explanation of the math and the iterative method, along with analogy to the simpler algorithm (k-means). Thanks Victor!

  • @ankitkusumakar7237
    @ankitkusumakar7237 2 місяці тому

    Content is good, but please amplify audio.

  • @bunny_4_4_
    @bunny_4_4_ 2 місяці тому

    When andrew tate explaining Math

  • @raihanpahlevi6870
    @raihanpahlevi6870 2 місяці тому

    sir we cant see your cursor omg

  • @raihanpahlevi6870
    @raihanpahlevi6870 2 місяці тому

    how to know the value of P(b) and P(a)

  • @wajahatmehdi
    @wajahatmehdi 2 місяці тому

    Excellent explanation

  • @DereC519
    @DereC519 2 місяці тому

    ty

  • @tazanteflight8670
    @tazanteflight8670 2 місяці тому

    Its amazing this works at all, because the first step is to take a 2d image that makes sense, into a 1d image that has lost ALL spatial information. A 1d stream of pixels is not an image.

  • @deepakjoshi7730
    @deepakjoshi7730 3 місяці тому

    Splendid. Example very well portrays the algorithm stepwise!

  • @theclockmaster
    @theclockmaster 3 місяці тому

    Thanks for this. Your video helped bring clarity to the problem statement.

  • @saunakroychowdhury5990
    @saunakroychowdhury5990 3 місяці тому

    but is not projection (y .e)e where y = x - mew

  • @raoufkeskes7965
    @raoufkeskes7965 4 місяці тому

    at 3:08 the variance estimator shoud be divided by (nb-1) as corrected estimation and not nb .. that's what we call Bessel's correction

  • @yeah6732
    @yeah6732 4 місяці тому

    Great tutorial! But why the slop of two eigenvectors are expected to be the same?!

  • @DrKnowsMore
    @DrKnowsMore 5 місяців тому

    Outstanding!

  • @johanesalberto6136
    @johanesalberto6136 6 місяців тому

    thanks brother

  • @azuriste8856
    @azuriste8856 6 місяців тому

    Great Explanation Sir. I don't know why it motivated me to appreciate and comment on the video.

  • @samarthpardhi7307
    @samarthpardhi7307 6 місяців тому

    Andrew Tate of machine learning

  • @guyteigh3375
    @guyteigh3375 6 місяців тому

    This course / lecture series has been staggeringly useful, thank you. It was also explained in a way that I could understand easily - and I have been struggling with the eplanations from others. You simplified things superbly. I did get lost when we started talking about mathmatical functions et al, but the information I needed was more to do with concepts and ideas - so i could safely let the maths part slip by, though noting different efficiencies of course. Thank you. Sincerely appreciate you sharing your work.

  • @virgenalosveinte5915
    @virgenalosveinte5915 6 місяців тому

    Amazing, thank you!

  • @mtushar
    @mtushar 7 місяців тому

    Enlightening, thank you!

  • @martijnhuijnen1
    @martijnhuijnen1 7 місяців тому

    Thanks so much! I will refer my students to your webpage!

  • @m07hcn62
    @m07hcn62 7 місяців тому

    This is awesome explanation. Thanks !

  • @Bulgogi_Haxen
    @Bulgogi_Haxen 7 місяців тому

    Studying at TUM. I admire german students who are following the lecture contents from the uni. Taking ML course atm, but here, the lecture is just like dumping only the whole concepts, regardless of whether students can understand them or not... So nice explanations in every video in ML related playlist.. I fcking regret that I did not choose the UK to study my master's.

  • @JD-rx8vq
    @JD-rx8vq 8 місяців тому

    Wow, you explain very well, thank you! I was having a hard time understanding my professor's explanation in our class.

  • @jonathanfiscus3221
    @jonathanfiscus3221 8 місяців тому

    Thanks for posting these Victor. I'm working on understanding the prior bias of precision and this helped. I hope things are going well!

  • @ArjunSK
    @ArjunSK 9 місяців тому

    Great tutorial!

  • @dinar.mingaliev
    @dinar.mingaliev 9 місяців тому

    Incredibly insightful! Your teaching style, peppered with examples, made complex topics like inverted index data structure and MapReduce algorithms easy to grasp. The way you broke down the compression techniques was particularly eye-opening, and I gained a newfound appreciation for the mechanics behind large-scale search engines and big data management. Before watching your lectures, I was quite overwhelmed by these concepts. However, your clear and structured approach has removed that uncertainty and replaced it with genuine interest and understanding. Thank you for your dedication to spreading knowledge. Your work has had a significant impact on my learning journey, and I am truly grateful for that. Please continue to share your wisdom; you are making a real difference in the lives of your viewers!

  • @josmbolomnyoi2498
    @josmbolomnyoi2498 9 місяців тому

    how do we block the japanese hack

  • @razakpapi
    @razakpapi 9 місяців тому

    andrew tate?

  • @sachinshettyvs2848
    @sachinshettyvs2848 9 місяців тому

    Thankyou😃

  • @mightyduckyo
    @mightyduckyo 10 місяців тому

    Step 1 is center, should we also scale so variance = 1

  • @user-ns8rn8fu3z
    @user-ns8rn8fu3z 10 місяців тому

    Hi sir is k means and kneighborhood algorithms are same ?

  • @Isomorphist
    @Isomorphist 10 місяців тому

    Great playlist.

  • @michael_bryant
    @michael_bryant 11 місяців тому

    This is the first time that PCA has actually made sense mathematically. Great video

  • @conradsnowman
    @conradsnowman 11 місяців тому

    I cant help but notice the middle dotted line looks like a logistic regression curve. I should know this.. But is there any relation?

  • @sedgeleyp
    @sedgeleyp 11 місяців тому

    Sounds just like Andrew Tate

    • @gulsumyldrm4039
      @gulsumyldrm4039 2 місяці тому

      Thanks ı dont want to listen more ….

  • @anubratadas1
    @anubratadas1 11 місяців тому

    As mentioned by @omidmo7554, I was exactly in a similar situation. you have explained it so lucidly. Thank you so much Victor!

  • @jospijkers1003
    @jospijkers1003 11 місяців тому

    SVD 3:08

  • @vinaykumardaivajna5260
    @vinaykumardaivajna5260 Рік тому

    Great Explination

  • @rodrigoma7422
    @rodrigoma7422 Рік тому

    O chatgpt me mandou aqui 😳

  • @pereeia9048
    @pereeia9048 Рік тому

    Amazing video, perfectly explained the concepts without getting bogged down in the math/technical details.

  • @jacobrafati4200
    @jacobrafati4200 Рік тому

    Andrew Tate?