PageRank algorithm: how it works

Поділитися
Вставка
  • Опубліковано 18 лис 2024

КОМЕНТАРІ • 27

  • @oliver5356
    @oliver5356 6 років тому +2

    We add random hops so as to avoid spider traps and dead ends. 0.84 is the typical (empirical) value used for choosing to pick a neighbor or randomly hop.
    Good clear explanation of the algorithm but not too much as to the reasons behind its formulation.

  • @xisreal
    @xisreal 7 років тому +6

    Thanks for the great video with nice and clear explanations and visuals. A couple of questions came up.
    1. I believe, (1-lambda)/N must represent the probability of starting at page x, not hoping to it from another node.Hopping from another node is captured by the second summand of PR(x). Besides, if hoping from another node (with N-1 nodes around) we would have N-1 in denominator.
    2. lambda can be defined more clearly (and mathematically). You bring a good example of lambda as a parameter of a Bernoulli distribution, but what random variable follows Bern(lambda) in this model? Which brings to the next comment :)
    3. From what I understood, it's the lambda that is initialized to 1/N, not the PR(x). Then PR(x) is initialized with lambda, where PR(x) changes with each iteration, but lambda remains fixed at initial value. :)
    4. It makes sense to introduce subscripted PR since the start.
    5. Finally, from basic probability, coin flipping is associated with 1/2 parameter in students' mind. Since lambda is vastly different from this, another example of Bernoulli random variable may make more sense. For example, it could come from card playing or die throwing :)
    Thanks for this and other videos!

    • @Elrossss
      @Elrossss 7 років тому +1

      Oleg, if I'm not wrong I believe the (1-lambda)/N represents the surfer randomly jumping to another url (including itself) rather than following any particular link. This is stated in the wikipedia page for the pagerank algorithm.

  • @yuanzhibao7078
    @yuanzhibao7078 8 років тому +1

    Wonderful video!!!!! Help me solve a lot misunderstandings about pagerank. Thank you so much.

  • @Amine-gp8im
    @Amine-gp8im 6 років тому +1

    Hello,
    is the sparse list representation the same as adjacency list ?
    Thank you

  • @zz-sm5ei
    @zz-sm5ei 4 роки тому

    How get the value of the 0.18 and 0.82 ?

  • @Amine-gp8im
    @Amine-gp8im 6 років тому

    I hope someone can answer my question:
    in the video he said that if we are implementing the algorithm in a single core we have to pull the pagerank and don't push it to the neighbors.
    Why did he only mention single core implementation doesn't this work for parallel implementation?

  • @officialkanyau
    @officialkanyau 10 років тому +1

    Amazing series, thanks so much Victor.
    I have a question though. I was wondering whether the pagerank algorithm you gave at 1:25 is complete. Please correct me if I'm wrong. In the example you gave beginning at 7:37, _PR(B) = 0.18 * 9.1+..._, my understanding is the 0.18 is the probability that someone will get to page B through randomly hoping to the page rather than by clicking a link that brings them to page B. i.e. 0.18 = 1 - λ
    You have multiplied this 0.18 with 9.1, which is the probability that one would get to page B through hoping if there were no links between the pages. That makes sense, but in the original pagerank equation, _PR(B) = (1 - λ) * (1 / N )..._, it doesn't seem to me like the 0.18 is included. I would have expected the original equation to start this was; _PR(B) = (1 - λ)* (1 - λ)* (1 / N)_ If I'm correct, is there a reason you didn't include the 0.18 (1 - λ) in the original equation? Thanks

    • @vlavrenko
      @vlavrenko  10 років тому +3

      0.18 is the same as (1-λ), there's no need to include it twice.
      Think of it this way: how can I arrive at node B?
      Either I take a random hop (with probability (1-λ) = 0.18), or I follow some link (with probability λ = 0.82).
      If it was a hop, I randomly pick one of the N nodes in my graph, and the chance I pick B is 1/N = 1/11 = 0.091
      So the joint probability of taking a random hop AND landing on B is (1-λ) / N = 0.18*0.091
      Regards.

    • @officialkanyau
      @officialkanyau 10 років тому

      Victor Lavrenko Thanks so much Victor. You're awesome.

    • @gonzalesjoseph1720
      @gonzalesjoseph1720 8 років тому

      +fckingkim reverseosmosissystemguides.com/category/water-filter/

  • @Ryuzakiv91
    @Ryuzakiv91 10 років тому

    Shouldn't the sum PR(A) + PR(B) + ... = 1 or 100 in every step? How can we verify if the algorith is correct?

    • @xisreal
      @xisreal 7 років тому

      Good point, this may need further clarification. I think he meant that at every iteration the probabilities are rescaled to add up to 1. I suppose, that's why he insists on using two arrays, not 1 - so that we rescale at the completion of each iteration and not interim.

  • @DenysonData
    @DenysonData 9 років тому

    Funny accent, flawless explanations!! Thanks!! :)

  • @seattle-bfg
    @seattle-bfg 5 років тому

    The fact that you keep a static image up that does not actually have a node X makes.. no sense.

  • @jydk37
    @jydk37 7 років тому

    Is lambda a Google trade secret? Without understanding how lambda is generated, I fail to see any value in this presentation.

    • @serhan6157
      @serhan6157 6 років тому

      Not exactly. Google left this formula ages ago. Well, not left but improved lets say, hence they have more sophisticated algorithm.
      Anyway, if we get back to your question, you can set lambda as 0.8-0.9. Practically this is the logical value because it makes 5 jumps on average with this value

  • @Faisal-jo5vk
    @Faisal-jo5vk Місяць тому

    how do i write this in python??

  • @rahulthankachan
    @rahulthankachan 9 років тому +1

    how did you calculate lamda? Thank you

    • @yuanzhibao7078
      @yuanzhibao7078 8 років тому

      +Rahul Thankachan You dont need to calculate lamda. It's from the truth. LOL

  • @suhasnayak4704
    @suhasnayak4704 6 років тому

    It would be more clear if you can explain things with some animations or if you can record the videos while you explain the things on white board.

    • @Ayoub-adventures
      @Ayoub-adventures 2 роки тому

      For me only audio was cristal clear. He explains clearly

  • @mh8316
    @mh8316 9 років тому +9

    You need to slow down and explain everything in slower steps.

  • @tiennguyenhuu3072
    @tiennguyenhuu3072 7 років тому

    please translate to Vietnamese thank you