XGBoost in Python from Start to Finish

Поділитися
Вставка
  • Опубліковано 14 січ 2025

КОМЕНТАРІ • 728

  • @statquest
    @statquest  4 роки тому +50

    NOTE: You can support StatQuest by purchasing the Jupyter Notebook and Python code seen in this video here: statquest.gumroad.com/l/uroxo
    Support StatQuest by buying my books The StatQuest Illustrated Guide to Machine Learning, The StatQuest Illustrated Guide to Neural Networks and AI, or a Study Guide or Merch!!! statquest.org/statquest-store/

    • @PetStuBa
      @PetStuBa 4 роки тому +2

      Dear Josh ... I have a request for new videos or livechats ... could you explain us these tests maybe ? ... Tukey, Bonferroni and Scheffé , it's hard for me to understand , you explain everything so well ... could be very helpful for a lot of people out there ... have a nice day , greetings from Europe

  • @znull3356
    @znull3356 3 роки тому +108

    Please keep doing these long-form Python tutorials on the various ideas we've covered in earlier 'Quests. They're great for those of us working in Python, and they give me another way to support the channel. It has been a more-than-pleasant surprise that as I've grown from learning the basics of stats to machine learning and eventually deep learning, StatQuest has grown along with me into those very same fields.
    Thanks Josh.

  • @kelvinhsueh5434
    @kelvinhsueh5434 3 роки тому +27

    You are amazing. Can't imagine how much work you put into those step-by-step tutorials.
    Just bought the Jupyter Notebook code and it's beyond worth it! Thank you :)

    • @statquest
      @statquest  3 роки тому +3

      Thank you very much for your support! :)

  • @zheyizhao4865
    @zheyizhao4865 4 роки тому +15

    Hey Josh, I just purchased all of your 3 Jupyter Notebook! I transferred from Econ major to Data Science, it was a nightmare before I find your channel. Your channel shed the light upon my academic career! Look forward to more of the 'Python from Start to Finish' series, and I will definitely support it!

  • @cszthomas
    @cszthomas 2 роки тому +1

    Thank you for the great work!

    • @statquest
      @statquest  2 роки тому +1

      Wow! Thank you so much for supporting StatQuest!!! BAM! :)

  • @tantalumCRAFT
    @tantalumCRAFT 2 роки тому +4

    This is hands down the best Python tutorial on UA-cam.. not just for XGBoost, but overall Python logic and syntax. Nice work, subscribed!!

  • @abhinaym5923
    @abhinaym5923 4 роки тому +5

    I am purchasing the Jupiter notebook to contribute to your work! Thanks a lot for this video! You are awesome! Will be very very happy to have more ML tutorials and thank you Josh!

    • @statquest
      @statquest  4 роки тому

      Thank you very much! :)

  • @samxu5320
    @samxu5320 3 роки тому +2

    Your pronunciation is the most authentic and clearest that I have ever heard

  • @sudheerrao07
    @sudheerrao07 4 роки тому +7

    Wow. Finally I see a face for the name. Your previous videos have had immensely helpul. I assumed you are a very senior person. I am not measuring your age. I mean, your way of explaining seemed like a professor with half a century of experience. But in reality, you are quite young. Thank you for all your simple-yet-detailed videos. No words to quantify how much I appreciate them. 🙏

  • @felixzhao3435
    @felixzhao3435 2 роки тому +1

    Thanks!

    • @statquest
      @statquest  2 роки тому

      WOW! Thank you so much for supporting StatQuest!!! BAM! :)

  • @OlgaW_Lavender
    @OlgaW_Lavender 4 місяці тому +1

    Outstanding! Complete with the thinking path - how to analyze variables in a logical way - and with common errors. Just purchased the Notebook. Thank you for all of your work on this channel.

    • @statquest
      @statquest  4 місяці тому

      Triple bam!!! Thank you very much for supporting StatQuest! :)

  • @RahulEdvin
    @RahulEdvin 4 роки тому +10

    Josh, you’re well and truly phenomenal ! Love from Madras !

    • @prashanthb6521
      @prashanthb6521 4 роки тому +1

      Chennai

    • @statquest
      @statquest  4 роки тому +2

      BAM! Thank you very much!!!

    • @starmerf
      @starmerf 4 роки тому +6

      Hi Rahul I taught atIIT-madras 19192-1993 lived on campus across from post office josh visited us there

    • @RahulEdvin
      @RahulEdvin 4 роки тому +1

      Frank Starmer Hello Frank, wow! That’s great to know ! :) I’m sure you must have had a good time here. Cheers :)

    • @prashanthb6521
      @prashanthb6521 4 роки тому +1

      @@starmerf wow the world is a small place ☺

  • @markrauschkolb5370
    @markrauschkolb5370 3 роки тому +10

    Extremeley helpful - would love to see more from the "start to finish" series

  • @AdamsJamsYouTube
    @AdamsJamsYouTube 2 роки тому +5

    Josh, this video is epic and really helped me understand the actual process of tuning hyperparameters, something that had been a bit of a black box until I saw this video. Your channel is awesome too - great jingles as well :D

  • @fernandes1431
    @fernandes1431 2 роки тому +3

    Can't thank you enough for the clearest and best explanation on UA-cam

  • @socksdealer
    @socksdealer Місяць тому +1

    I understand you better than if it would be explained in my native language) Thank you for your work!

  • @navyasreepinjala1582
    @navyasreepinjala1582 2 роки тому +2

    I love your teaching style. Extremely helpful for a beginner like me. Really helped me a lot in my exams. No words. You are the best!!!!

  • @jinwooseong2862
    @jinwooseong2862 3 роки тому +1

    I watched your all video for XGBoost. It helps me a lot. very appreciated!

  • @julieirwin3288
    @julieirwin3288 4 роки тому +4

    What did we do to deserve a great guy like Josh ? Thank you Josh!

  • @maurosobreira8695
    @maurosobreira8695 3 роки тому +1

    A true, real Master Class - You got my support!

  • @Krath1988
    @Krath1988 4 роки тому +112

    Liked, favorited, recommended, shared, and sacrificed my first-born to this video.

  • @romanroman5226
    @romanroman5226 3 роки тому +1

    Awesome video! The cleanest xgboost explanation a have ever seen.

  • @sane7263
    @sane7263 Рік тому +3

    That's the Best video I've ever seen. Period.
    TRIPLE BAM! :)

  • @RahulVarshney_
    @RahulVarshney_ 4 роки тому +3

    "25:36" that's what i was waiting for from the beginning...Truly amazing.. You are providing precious information..CHEERS

    • @statquest
      @statquest  4 роки тому +1

      Glad it was helpful!

    • @RahulVarshney_
      @RahulVarshney_ 4 роки тому

      @@statquest one small request..can you provide some valuable information through a video like which model to chose for different datasets..how do we decide what model we should chose...thanks in advance

    • @statquest
      @statquest  4 роки тому

      @@RahulVarshney_ I'll keep that in mind. In the mean time, check out: scikit-learn.org/stable/tutorial/machine_learning_map/index.html

    • @RahulVarshney_
      @RahulVarshney_ 4 роки тому +1

      @@statquest that is amazing ...i will complete it today itself thanks again for your prompt reply
      Can i get your email

  • @danielmagical6298
    @danielmagical6298 4 роки тому +2

    Hi Josh, great job really helpful material as I'm discovering XGBoost just now.
    Thank you and keep you great work!

    • @statquest
      @statquest  4 роки тому

      Thank you very much! :)

  • @minseong4644
    @minseong4644 3 роки тому +1

    Such an amazing job Josh.. Couldn't find any better explanation than this!
    Mesmerizing!

  • @josephhayes9152
    @josephhayes9152 3 роки тому +1

    Thanks for the great tutorial! You covered a lot of details (mostly data cleaning) that are often overlooked or skipped as 'trivial' steps.

    • @statquest
      @statquest  3 роки тому +1

      Thank you! Yes, "data cleaning" is 95% of the job.

  • @francovega7089
    @francovega7089 2 роки тому +1

    I really appreciate your content Josh. Thanks for your time

  • @thomsondcruz
    @thomsondcruz Рік тому

    Absolutely loved this video Josh. It breaks down everything into understandable chunks. Thank you and God bless. BAM! The only thing I missed (and its very minor) was taking in a new data row and making an actual prediction by using the model.

    • @statquest
      @statquest  Рік тому +1

      Thanks! For new data, you just call clf_xgb.predict() with the row of new data.

  • @VarunKumar-pz5si
    @VarunKumar-pz5si 3 роки тому +1

    I'm very grateful to have you as my teacher.

  • @keyurshah8451
    @keyurshah8451 2 роки тому +1

    Hey Mate, amazing tutorial. Very complex problem explained in really simple and effective way. I am using XGBOOST for one of the classification model and after watching your video it made me realise I can further improve my model. So thank you again and keep making those videos. Kudos to you and long live data science 🙏🙏

  • @SergioPolimante
    @SergioPolimante 3 роки тому +1

    This kind of content is SUPER HARD to produce. I really understand and appreciate your effort here. Thanks and congratulations.

  • @muskanroxx22
    @muskanroxx22 3 роки тому +1

    You're a very kind human being Josh!! Thank you so much for making these videos. Your content is gold!!! I am new to data science and this is exactly what I needed!! :)
    Much love from India!

    • @statquest
      @statquest  3 роки тому +1

      Glad you like my videos!! BAM! :)

    • @muskanroxx22
      @muskanroxx22 3 роки тому

      @@statquest Hey Josh! I am learning about Bayesian Optimizer and I don't seem to get it even after watching tons of tutorials, can you suggest where I should learn it from please? I couldn't find a video on your channel on this.

    • @statquest
      @statquest  3 роки тому

      @@muskanroxx22 Unfortunately I don't know of a good source for that.

  • @KukaKaz
    @KukaKaz 4 роки тому +5

    Yes pls more videos with python❤thank u for the webinar

  • @parismollo7016
    @parismollo7016 4 роки тому +3

    I haven't watched it yet but I know this will be great!!!!!!!! Thank you Josh.

  • @janskovajsa237
    @janskovajsa237 5 місяців тому +1

    41:21 In my version of xgboost were parameters early_stopping_rounds=10 and eval_metric='aucpr' moved to XGBClassifier, so if it is not working I suggest trying this.
    Although I really appreciate value of StatQuest videos, I have to admit I really hate all that singing and Bams. Makes me feel like attending school for slower children

    • @statquest
      @statquest  5 місяців тому

      Sorry to hear that.

    • @ovrava
      @ovrava Місяць тому

      @@statquest i like it. its cool.

  • @chiragpalan9780
    @chiragpalan9780 4 роки тому +5

    This guy is amazing.
    DOUBLE BAM 💥 💥

  • @PradeepMahato007
    @PradeepMahato007 4 роки тому +2

    BAMMMMM !!!
    This is awesome 👍 Josh !! Thank you for your contribution, really helpful for new learners.😊😊😊

  • @godoren
    @godoren 4 роки тому +2

    Thank you for your job, the explanation of the topic is very clear and transparent.

    • @statquest
      @statquest  4 роки тому

      Thank you very much! :)

  • @darksoul1381
    @darksoul1381 4 роки тому +1

    I was wondering how to find stuff regarding dealing with actual churn data and sampling issues. The tutorial addressed a lot of them. Thanks!

  • @codinghighlightswithsadra7343
    @codinghighlightswithsadra7343 Рік тому +2

    Thank you so much for the work that you used in step by step tutorial. it was amazing.

  • @mdaroza
    @mdaroza 3 роки тому +2

    Amazingly organized and well explained!

  • @marceloherdy2379
    @marceloherdy2379 4 роки тому +1

    Man, this video is awesome! Congratulations!

  • @trendytrenessh462
    @trendytrenessh462 3 роки тому +2

    It is really lovely to be able to put a face to the "Hooray!", "BAM !!!" and "Note:"s 😄❤

  • @haskycrawford
    @haskycrawford 4 роки тому +1

    I love the channel! Eu aprendo + aqui do que a Graduação! You great josh!

  • @sreejaysreedharan4085
    @sreejaysreedharan4085 4 роки тому +1

    Lovely and priceless video Josh...BAM BAM BAM as usual !! :) God bless. .

    • @statquest
      @statquest  4 роки тому

      Thank you very much! :)

  • @daniloyukihara2143
    @daniloyukihara2143 3 роки тому +1

    hurray, i picture you totally different!
    Thanks a lot for all the videos!

  • @aksharkottuvada
    @aksharkottuvada 2 роки тому +1

    Thank you Josh. Needed this tutorial to better solve a ML Problem as part of my internship :)

  • @dillonmears6696
    @dillonmears6696 2 роки тому +1

    Great video! You did a wonderful job of explaining the process. Thanks!

  • @marekslazak1003
    @marekslazak1003 2 роки тому +4

    Jesus, i just learned more over 10 minutes of this than i did throughtout an entire semester of a similar subject on CS. ++ tutorial

  • @andrewxie9896
    @andrewxie9896 3 роки тому +1

    you are simply an amazing human being, also the notebooks are great! :D

  • @dagma3437
    @dagma3437 4 роки тому

    I'm so glad you are a bad-ass stats guru and a teacher waaaaaaaaay before a singer and a guitarist ...Thank you! ;)

    • @statquest
      @statquest  4 роки тому +1

      joshuastarmer.bandcamp.com/

    • @dagma3437
      @dagma3437 4 роки тому

      StatQuest with Josh Starmer ...not bad. A poor man’s Jack Johnson 🤔

    • @dagma3437
      @dagma3437 4 роки тому +1

      Just pulling your leg. Thanks for all the content on stats

  • @kd6600xt
    @kd6600xt 2 роки тому +5

    Quick note: At 43:30 instead of using the plot_confusion_matrix() which is now depreciated, you need to use ConfusionMatrixDisplay.from_estimator(). This can be done as follows:
    Include: from sklearn.metrics import ConfusionMatrixDisplay at the start with the other imports. Then when printing the confusion matrix you need to use the line: ConfusionMatrixDisplay.from_estimator(clf_xgb, X_test, y_test, display_labels=["Did not leave", "Left"])

    • @statquest
      @statquest  2 роки тому +3

      Thank you very much! I really appreciate it.

    • @kd6600xt
      @kd6600xt 2 роки тому +1

      @@statquest no problem!

    • @emilioluissaenzguillen5719
      @emilioluissaenzguillen5719 2 роки тому

      I am getting this error following what you say:
      XGBoostError: [17:07:52] c:\buildkite-agent\builds\buildkite-windows-cpu-autoscaling-group-i-08de971ced8a8cdc6-1\xgboost\xgboost-ci-windows\src\c_api\c_api_utils.h:167: Invalid missing value: null
      Do you know why it might be? Thanks.

    • @emilioluissaenzguillen5719
      @emilioluissaenzguillen5719 2 роки тому +3

      I solved the above error by setting missing=0 on the above code as follows:
      clf_xgb = xgb.XGBClassifier(objective='binary:logistic',
      missing=0,
      seed=42,
      ## the next three arguments set up early stopping.
      eval_metric='aucpr',
      early_stopping_rounds=10)

    • @benchen9910
      @benchen9910 Рік тому

      @@emilioluissaenzguillen5719 thanks it works

  • @miguelbarajas9892
    @miguelbarajas9892 2 роки тому +1

    Freaking amazing! You explain everything so well. Thank you!

  • @coolmusic4meyee
    @coolmusic4meyee 11 місяців тому +1

    Great explanation and walk-through, big thanks!

    • @statquest
      @statquest  11 місяців тому

      Glad you enjoyed it!

  • @felixwhise4165
    @felixwhise4165 4 роки тому +1

    just here to say thank you! will come back in a month when I have time to watch it. :)

  • @williamTjS
    @williamTjS 2 роки тому +1

    Amazing! Thanks so much for the detailed video

  • @jessehe9286
    @jessehe9286 4 роки тому +6

    Great video! Love it!
    request that you do a comparison of XGBoost, CatBoost, and LightGBM, and a quest on ensemble learning.

    • @statquest
      @statquest  4 роки тому +5

      I'll keep those topics in mind.

  • @Fressia94
    @Fressia94 4 роки тому +1

    many thanks to your great and so understandable video. It literaly helps me a lot in Python and XGBoost package

  • @danielpinzon9284
    @danielpinzon9284 4 роки тому +1

    Love u Josh.... you are a TRIPLE BAM!!! Greetings from Bogotá, Colombia.

  • @viniantunes5944
    @viniantunes5944 4 роки тому +4

    Josh, you're the didactic in person form.
    Thanks!

  • @nehabalani7290
    @nehabalani7290 4 роки тому +2

    Good to also see you sing rather than just hear :).. i had to comment this even before starting the training

  • @marcelocoip7275
    @marcelocoip7275 2 роки тому +1

    Hard work here, I'ts funny how the responsabile scientist and the funny guy coexist, very useful lesson, thanks!

  • @jimmyrico5364
    @jimmyrico5364 4 роки тому +1

    This is a great piece of work, thanks for sharing it!
    Maybe the only additional piece I'd add which I've found useful on the documentation of XGBoost is that one can take advantage of parallel computing (more cores or using a graphic card your machine or you could have on the cloud) by simply passing the parameter (n_jobs = -1) while doing both, the RandomizedSearchCV stage and the setting the XGB regressor type (XGBRegressor for example).

  • @saeedesmailii
    @saeedesmailii 4 роки тому +2

    It was extremely helpful. Please continue making these videos. I suggest making a video to explain the clustering with unlabeled data, and predicting the future trend in time-series data.

    • @statquest
      @statquest  4 роки тому +1

      I'll keep that in mind. :)

  • @aaltinozz
    @aaltinozz 4 роки тому +1

    all week searched for this thank u very much

  • @gisleberge4363
    @gisleberge4363 2 роки тому +1

    Appreciate the Python related videos...helps to manoeuvre the code when I try to replicate the method later on...easy to follow the whole thing, also for beginners... 🙂

    • @statquest
      @statquest  2 роки тому +1

      Thanks! There will be a lot more python stuff soon.

  • @ketanshetye5029
    @ketanshetye5029 4 роки тому +1

    could not help u with money right now , but i watched all the adds in video , hope that helps u financially . love u videos . keep up!!

  • @harshbordekar8564
    @harshbordekar8564 2 роки тому +1

    Thank you for the awesomeness!!

  • @CharlotteWilson-j9d
    @CharlotteWilson-j9d Рік тому

    First you’ve saved me this is super clear! I love all your videos so much 😊
    I do have two questions…
    1. How would you handle a classification problem with time series data?
    2. Is there any other evaluation test you should or could do to evaluate the effectiveness of your model?

    • @statquest
      @statquest  Рік тому

      1. I've never used XGBoost with time series (or done much of any time series stuff before), so I can't answer this question.
      2. There are lots of ways to evaluate a model. I only present a few, but there are many more, and they really depend on what you want your model to do. Just google it.

  • @henkhbit5748
    @henkhbit5748 4 роки тому +1

    Greatly appreciated this videoLike you said, telcos should gives more effort to tie the current customers. In real practice you want to know what the probability is that a current customer will no longer renew the subscription. You should then try to bind the customer with a high risk with incentives.

  • @miloszpabis
    @miloszpabis 9 місяців тому +1

    Love these Python tutorials after watching theory videos:D

    • @statquest
      @statquest  9 місяців тому

      Glad you like them!

  • @shazm4020
    @shazm4020 3 роки тому +1

    Thank you so much Josh Starmer! BAM!

  • @jwxdxd
    @jwxdxd 3 роки тому +1

    You are amazing! Thank you so much !!

  • @abdulkayumshaikh5411
    @abdulkayumshaikh5411 3 роки тому +1

    Hello josh, you are doing amazing work keep doing

  • @azingo2313
    @azingo2313 Рік тому +1

    This man deserves Nobel Prize for peace of mind ❤❤

  • @pacificbloom1
    @pacificbloom1 3 роки тому +1

    Wonderful video josh.....pleasee pleasee pleasee make more videos on start to finish on python for different models.....i havr actually submitted my assignments using your techniques and got better results than what i have learned in my class
    Waiting for more to come especially on python :)

    • @statquest
      @statquest  3 роки тому

      Thanks! There should be more python coming out soon.

  • @jameswilliamson1726
    @jameswilliamson1726 Рік тому +1

    Another great tutorial. Thx

  • @FlexCrush1981
    @FlexCrush1981 4 роки тому +3

    Very enjoyable webinar Josh. Thanks for posting. I'm not 100% sure how to interpret the leaves. The largest leaf value is 0.188 where Dependents_No

    • @statquest
      @statquest  4 роки тому +2

      The leaves are how much to increase or decrease the log(odds) for one category. For more details, see: ua-cam.com/video/8b1JEDvenQU/v-deo.html

  • @arijitdas4504
    @arijitdas4504 4 роки тому +9

    If "Stay Cool" had a face, it'd be you :)

  • @k44zackie
    @k44zackie 3 роки тому +1

    Thank you very much for nice video! Very helpful for me.

    • @statquest
      @statquest  3 роки тому +1

      Glad it was helpful!

  • @NLarsen1989
    @NLarsen1989 3 роки тому +1

    Yikes, if I ever understand something enough to explain it as succinctly as you do then I'd be very happy. I've been smashing through a lot of your videos the last few days after spending countless months on python, sklearn and all the usual plug and play solutions and it's not been until I've started watching these that I've started to feel things click into place

    • @statquest
      @statquest  3 роки тому

      Awesome! I'm glad my videos are helpful! :)

  • @yurimartins1499
    @yurimartins1499 4 роки тому +1

    Thank you Josh!!
    As a suggestion, you could do a StatQuest explaining the measures in market basket analysis?

    • @statquest
      @statquest  4 роки тому +1

      I'll keep that in mind.

  • @kandiahchandrakumaran8521
    @kandiahchandrakumaran8521 Рік тому +1

    Amazing. Wonderful videos. I started only 3 months ago and with your videos I am very confident to do nalysis with Python. Manny thanks.
    Is it possible to create a video for Nomogram for competic#ing risks for Time-Event (survival analysis) based on CPH outputs?

    • @statquest
      @statquest  Рік тому +1

      I'll keep that topic in mind.

  • @azecem6187
    @azecem6187 4 роки тому +1

    Thanks a lot Josh!

    • @statquest
      @statquest  4 роки тому +1

      Any time! And thanks for your support! :)

  • @HardikShah17
    @HardikShah17 2 роки тому +1

    Excellent Video @StatQuest ! Can we please have more Start to Finish python videos? Like Lightgbm maybe?

    • @statquest
      @statquest  2 роки тому

      I'll keep that in mind! :)

  • @hollyching
    @hollyching 4 роки тому

    Thanks Josh for another GREAT video! Just some sharing and minor questions.
    1. try pandas_profiling when doing EDA. I personally love it. :)
    2. some features are highly correlated (eg: city name and zip code). Do we need to handle that before running XGB?
    3. Why choose 10 for early_stopping_rounds
    4. What’s the difference between
    - df.loc[df['Total_Charges']==' ']
    - df[df['Total_Charges']==' ']
    5. What’s the difference between
    - y=df['Churn_Value'].copy
    - y=df['Churn_Value']
    Many thanks in advance!
    H

    • @statquest
      @statquest  4 роки тому +1

      1) Thanks for the tip on pandas_profiling.
      2) No.
      3) It's a commonly used number
      4) I don't know.
      5) I believe the former is copy by value and the latter is copy by reference.

  • @erichganz4605
    @erichganz4605 2 роки тому +1

    This guy is just amazing

  • @CDALearningHub
    @CDALearningHub 4 роки тому +1

    Thanks for sharing! Informative.

  • @BiffBifford
    @BiffBifford 4 роки тому +12

    I am not a math geek. I am here strictly for the intro song!

    • @magtazeum4071
      @magtazeum4071 4 роки тому +2

      same here

    • @statquest
      @statquest  4 роки тому +2

      DOUBLE BAM! :)

    • @rishavpaudel7591
      @rishavpaudel7591 4 роки тому +2

      @@statquest Sir your double and triple bam has really taught lots of things for me to be honest. Me as a student doing Post-Graduate in AI,
      lots of love form Nepal.

  • @eeera-op8vw
    @eeera-op8vw 3 місяці тому +1

    awesome!! I hope you can do a video about XGBoost with regression too.

    • @statquest
      @statquest  3 місяці тому

      I'll keep that in mind.

  • @Toyotaman
    @Toyotaman 2 роки тому

    38:05 stratify=y is not for yes is for dependent variable y. if you have a different variable, you gotta pass your response variable's name to stratify

    • @statquest
      @statquest  Рік тому

      Oops! thanks for catching that.

  • @nepalm222
    @nepalm222 3 роки тому +1

    Great Content, subscribed

    • @nepalm222
      @nepalm222 3 роки тому +1

      Also, single best python package run through Ive seen.

    • @statquest
      @statquest  3 роки тому

      Thank you very much! :)

  • @nikhilshaganti5585
    @nikhilshaganti5585 2 роки тому

    Thank you for this great tutorial Josh! Your videos have immensely helped in understanding some of the complex topics. One thing I noticed while watching this tutorial is the handling of categorical features. I think the explanation you gave for "Why not to use LabelEncoding?" is applicable for models like Linear Regression, SVM, NN but not for Trees because they only focus on the order of the feature values. For example, in a set of [1,2,3,4], threshold < 1.5 would be equivalent to threshold == 1. Please let me know if my thought process in wrong.

    • @statquest
      @statquest  2 роки тому

      To be honest, I'm not really sure I understand your question or your example. If we have a categorical feature with 4 colors, red, blue, green, and black, and we give them numbers, like 1, 2, 3 and 4, then a threshold < 2.5, would not make much sense and, based on how trees are implemented, there would be no options for threshold == 2 or threshold == 3. So we wouldn't be able to separate colors very well.

    • @nikhilshaganti5585
      @nikhilshaganti5585 2 роки тому

      right, in your example, the threshold of threshold

    • @statquest
      @statquest  2 роки тому

      @@nikhilshaganti5585 Sure, you can continue to separate things in later branches - but the greedy nature of the algorithm doesn't ensure that you'll get to those later branches. So you start by making a guess that it makes sense to group red and blue together.

  • @brandonterrell9680
    @brandonterrell9680 Рік тому +1

    # very helpful and informative, thank you!

  • @lahirulowe4752
    @lahirulowe4752 2 роки тому +1

    it would be a great idea to have an entire series of "start to finish"

  • @thiagotanure2212
    @thiagotanure2212 3 роки тому +3

    amazing tutorial Josh! Shared with my friends =D
    Could you do one of these about pygam? It would be amazing :)

    • @statquest
      @statquest  3 роки тому +1

      I'll keep that in mind.

  • @TheSoonAnn
    @TheSoonAnn Рік тому +1

    very good explaination

  • @yoniziv
    @yoniziv 4 роки тому +1

    Triple Bam! thanks for your great tutorial

  • @andrewnguyen5881
    @andrewnguyen5881 4 роки тому

    Again another quality video, I was following along with your every word, which did bring up come questions:
    1. When XGBoost deals with missing data, does it ever consider splitting the missing data in half?
    --Using your example, would it ever do 1 blue and 1 green? What would happen if XGBoost encountered a data set with alot of missing values?
    2. When you ran your Cross-Validation, was there a reason you only used 3 values for each hyperparameter? Could you have done more if you wanted to?
    3. When I ran my Cross-Validation, my scale_pos_weight didn't change even though I used the same parameters you did. What do you think the problem could be?

    • @statquest
      @statquest  4 роки тому

      1. Not that I know of. If there was a lot of missing values, it would still proceed just as described.
      2. I wanted the cross validation to run in a short period of time, so I picked 3 values for each hyperparameter. If I had more time on my hand, or a cluster of computers, I might have considered trying more.
      3. I'm not sure.

    • @andrewnguyen5881
      @andrewnguyen5881 4 роки тому

      @@statquest Gotcha! Thank you for actually answering my questions haha I really appreciate the help as someone getting into more Machine Learning. Not sure what your upcoming videos will be but I think some great videos would be:
      -How you did the Cross Validation for the hyper parameters in this video? ( I have watched your Cross-Validation video, but I want to learn more about actually doing it) [I also still can get my first run to output the same values you had lol]
      -XGBoost for Regression in Python
      -How to select the best algorithm based on the scenario for Regression or Classification
      Just some thoughts :)

    • @statquest
      @statquest  4 роки тому

      @@andrewnguyen5881 I'll keep those topics in mind.