Це відео не доступне.
Перепрошуємо.

How to Download Wikipedia

Поділитися
Вставка
  • Опубліковано 19 кві 2021
  • Did you know you can collect data from Wikipedia?
    Wikipedia has very comprehensive information on nearly every topic. It looks like unstructured free text but most of the information on wikipedia is linked to each other and it is structured on a platform called Wikidata. By using a query language, you can collect data on any domain you might want.
    In this video, we look into how to collect data on directors who were nominated or awarded an Academy Award for best director.
    👇Learn how to complete your first real-world data science project
    Hands-on Data Science: Complete Your First Portfolio Project
    www.misraturp....
    🐼 All you need to know about Pandas in one place! Download my Pandas Cheat Sheet (free) - misraturp.gumr...

КОМЕНТАРІ • 28

  • @misraturp
    @misraturp  3 роки тому +1

    👇Learn how to complete your first real-world data science project
    Hands-on Data Science: Complete Your First Portfolio Project
    www.misraturp.com/hods

  • @hagalhagal9989
    @hagalhagal9989 Рік тому +1

    Best source of information on how to download information from Wikipedia!

  • @tasosvemmos2420
    @tasosvemmos2420 Місяць тому

    Nice video. Thank you!

  • @anirbanbhowmik1416
    @anirbanbhowmik1416 2 роки тому +1

    Thank you very much for this video!! Maybe I won't use it right at this moment, but I can very well imagine the value of this knowledge. Often, people don't understand that they really need to keep in mind that most of their viewers would likely be 'dummies', just like me! And, this video is really perfect! ❤️

    • @misraturp
      @misraturp  2 роки тому

      That's great to hear Anirban! You are very welcome. :)

    • @user-di3kf3nf7w
      @user-di3kf3nf7w 3 місяці тому

      B​@@misraturp

  • @ColinTorrible
    @ColinTorrible Рік тому

    You rock! Thank you so much for thia video. This is exactly what I was looking to learn and you nailed it.

  • @Avra_b
    @Avra_b 3 роки тому +1

    Ah this is pretty cool , thanks . Your videos on streamlit was useful as well :)

    • @misraturp
      @misraturp  3 роки тому

      Glad you like them. Thank you!

  • @josevanderlleegs299
    @josevanderlleegs299 3 роки тому +1

    its nice...its pritty cool and helpful...great video!!

  • @yoyovatsa2179
    @yoyovatsa2179 3 роки тому +3

    It went from 0 to 100 real quick when you pasted that query. One doubt though, how do we get to know what columns are there, for some other data we might need, this seems like a very good alternative to scraping and then spending endless time cleaning the data. Also the link in the description isn't working. Nice video overall, thanks

    • @misraturp
      @misraturp  3 роки тому +3

      Hey, thanks for the heads-up. I fixed the link. The main way to figure out what features or columns you can get is to explore the topic on wikidata.org. I go clicking around and seeing what attributes each page has and use what I found in my queries.

    • @yoyovatsa2179
      @yoyovatsa2179 3 роки тому +2

      @@misraturp ok thank you I will try it out later.

  • @latafatali7335
    @latafatali7335 3 роки тому +1

    I need to scrap Edit history of a wikipedia page. Help !!

    • @misraturp
      @misraturp  3 роки тому

      Hey Latafat, can't say I've done that before. Have you found a way?

  • @tuxhome3686
    @tuxhome3686 Рік тому

    great video! very informative.

  • @sushrutdhiman1776
    @sushrutdhiman1776 3 роки тому

    Hi! I'm facing a problem, rows with empty columns are getting dropped.
    Please tell me if there's way to avoid that

    • @misraturp
      @misraturp  3 роки тому

      Hey Sushrut,
      You can wrap the part that might have some empty values in an "Optional" wrapper. That way all the results should be returned. It looks like this:
      OPTIONAL{
      // whatever you'd like to query
      }

    • @sushrutdhiman1776
      @sushrutdhiman1776 3 роки тому

      @@misraturp Hi Thank you! That did sort it out thanks but then I got a query timeout error which is even worse. Is there something we can do about it?

    • @misraturp
      @misraturp  3 роки тому

      I get that quite often too. Unfortunately, there is a limit to how long a query can run on the query interface of Wikidata. You can try to make your query be more efficient or divide it into multiple queries if possible to make it run faster.
      There are some tips and tricks on how to make queries more efficient here: www.wikidata.org/wiki/Wikidata:SPARQL_query_service/query_optimization

  • @hikaroto2791
    @hikaroto2791 2 роки тому +1

    ill try on biological related data

    • @misraturp
      @misraturp  2 роки тому

      That's cool! There is a lot of detail in biological domain on wikidata.

  • @alizhadigerov9599
    @alizhadigerov9599 2 роки тому

    Hi! How to scrap exact text from Wikipedia?

    • @misraturp
      @misraturp  2 роки тому

      For that you'd need to create a scraper using Python. It is not possible to do it through wikidata and SparQL.

  • @amandas3014
    @amandas3014 8 місяців тому

    You are very beautiful