- 48
- 244 003
Machine Learning Mastery
India
Приєднався 11 чер 2020
Implementing Bayesian Optimization - Step by Step Coding - Part 2
This is continuation of the Part1 video and I take you through all the building blocks for implementing of Bayesian Optimization step by step including key coding blocks.
Переглядів: 273
Відео
Implementing Bayesian Optimization - Step by Step Coding - Part 1
Переглядів 1832 місяці тому
In this video, I take you through all the building blocks for implementing of Bayesian Optimization step by step including key coding blocks.
Kalman Filter Simplified - Algorithm explained with Examples
Переглядів 1282 місяці тому
Kalman Filter Simplified - Algorithm explained with Examples
How to Manage Train vs Test Divergences
Переглядів 292 місяці тому
In practical applications, training vs testing set divergences are inevitable. In this video, I explain how to manage such divergences. You can score them using methods that I explained in my other videos.
Fixing Model Probability - Why this matters? How to do it?
Переглядів 1162 місяці тому
Many models including Support Vectors, Decision Trees, Random Forests, XGBoost, CatBoost, LightGBM provide you "Predicted Probability" which are usually INCORRECT. There is a statistical reason why they are incorrect and I explain a sound approach to correct them.
Nested Cross Validation - Algorithm Explained
Переглядів 1422 місяці тому
Learn when to use Nested Cross Validation and How it adds value to understand the variance bias play.
What is KFold Cross Validation? When NOT to use it? How to use it with modifications for your data
Переглядів 45010 місяців тому
KFold cross validation plays a very important role in understanding the variance play in your model. Most people take it for granted and don't use its full potential. I explain how to use it right, how to read its variance play and also highlight when NOT to use vanilla KFold ? But, rather use its extensions as implemented in SKLearn. My AI and Generative AI Courses are details here: ai.generat...
How to really find if my Test Data is diverging from my Training dataset? This WORKS!
Переглядів 53910 місяців тому
Adversarial Validation is a practical method used for finding if testset (seen in production) has started to diverge from training set. We detail the scoring function and how you can implement this. Very effective for mixed tabular data usecases. My AI and Generative AI Courses are details here: ai.generativeminds.co To get a FREE invite to our classes, fill below link: invite.generativeminds.co
Use CentralLimit Theorem to turn any distribution to Normal ? Really?
Переглядів 22511 місяців тому
Central Limit Theorem defines the law of large numbers. We list exactly what the law defines and how empirically non-gaussian distributions can be handled using this theorem for our applications. My AI and Generative AI Courses are details here: ai.generativeminds.co To get a FREE invite to our classes, fill below link: invite.generativeminds.co
How Bootstrapping helps with scoring your Train Test Divergences?
Переглядів 20611 місяців тому
How do you score Train Test Divergences? Bootstrapping is one simple approach to hep you get a grip on this topic. Relying on random sampling methods, its statistically valid and practically a good reference point to be used along side Adversarial scoring techniques. My AI and Generative AI Courses are details here: ai.generativeminds.co To get a FREE invite to our classes, fill below link: inv...
How I built Generative AI for Retail in 60 Days
Переглядів 528Рік тому
Below is the link to an FREE interactive video where I explain the step by step path to building your own Generative AI for your businesses within 60 Days. Just follow the steps and you will get RESULTS !! WATCH it FREE here : how-to-llm.generativeminds.co/ My AI and Generative AI Courses are details here: ai.generativeminds.co To get a FREE invite to our classes, fill below link: invite.genera...
Bayesian Optimization - Math and Algorithm Explained
Переглядів 54 тис.3 роки тому
Learn the algorithmic behind Bayesian optimization, Surrogate Function calculations and Acquisition Function (Upper Confidence Bound). Visualize a scratch implementation on how the approximation works iteratively. Finally, understand how to use scikit-optimize package todo hyperparameter tuning using bayesian optimization. My AI and Generative AI Courses are details here: ai.generativeminds.co ...
Decision Tree Hyperparam Tuning
Переглядів 3,8 тис.3 роки тому
Learn how to use Training and Validation dataset to find the optimum values for your hyperparameters of your decision Tree. Demonstrated for - Max Tree Depth and Min Sample Leaves hyper parameters. My AI and Generative AI Courses are details here: ai.generativeminds.co To get a FREE invite to our classes, fill below link: invite.generativeminds.co
Decision Tree Cost Pruning - Hands On
Переглядів 2,4 тис.3 роки тому
In this handson video you will Learn how to find the right Cost Pruning Alpha parameter for your decision tree. My AI and Generative AI Courses are details here: ai.generativeminds.co To get a FREE invite to our classes, fill below link: invite.generativeminds.co
Gradient Boosting Hands-On Step by Step from Scratch
Переглядів 2,8 тис.3 роки тому
Learn how to write gradient boosting tree algorithm from scratch. Learn how the Loss function is derived and applied into python code as part of your boosting iteration. Learn a trick to present your charts as interpretable categorical values rather than encoded numerical values. (This is done a lot in practice) My AI and Generative AI Courses are details here: ai.generativeminds.co To get a FR...
Hyperparameters - Introduction & Search
Переглядів 5 тис.3 роки тому
Hyperparameters - Introduction & Search
Feature Importance Formulation of Decision Trees
Переглядів 6 тис.3 роки тому
Feature Importance Formulation of Decision Trees
How to Regularize with Dropouts | Deep Learning Hands On
Переглядів 6393 роки тому
How to Regularize with Dropouts | Deep Learning Hands On
How to Regularizing with Weight & Activation Regularizations | Deep Learning
Переглядів 5983 роки тому
How to Regularizing with Weight & Activation Regularizations | Deep Learning
How to Fix Vanishing & Exploding Gradient Problems | Deep Learning
Переглядів 3 тис.3 роки тому
How to Fix Vanishing & Exploding Gradient Problems | Deep Learning
How to Accelerate training with Batch Normalization? | Deep Learning
Переглядів 7984 роки тому
How to Accelerate training with Batch Normalization? | Deep Learning
What is a Perceptron Learning Algorithm - Step By Step Clearly Explained using Python
Переглядів 22 тис.4 роки тому
What is a Perceptron Learning Algorithm - Step By Step Clearly Explained using Python
How to Tune Learning Rate for your Architecture? | Deep Learning
Переглядів 1,6 тис.4 роки тому
How to Tune Learning Rate for your Architecture? | Deep Learning
How to Find the Right number of Layers/Neurons for your Neural Network?
Переглядів 13 тис.4 роки тому
How to Find the Right number of Layers/Neurons for your Neural Network?
How to Configure and Tune Batch Size for your Neural Network?
Переглядів 2,8 тис.4 роки тому
How to Configure and Tune Batch Size for your Neural Network?
Back Propagation Math Step By Step Detailed with an Example | Deep Learning
Переглядів 2,7 тис.4 роки тому
Back Propagation Math Step By Step Detailed with an Example | Deep Learning
Back Propagation Concept Math Step By Step for a Two Layer Feed Forward Network
Переглядів 4674 роки тому
Back Propagation Concept Math Step By Step for a Two Layer Feed Forward Network
How Gradient Descent finds the weights? Gradient Descent Math Step By Step with Example | Neural Net
Переглядів 12 тис.4 роки тому
How Gradient Descent finds the weights? Gradient Descent Math Step By Step with Example | Neural Net
How to use Gaussian Mixture Models, EM algorithm for Clustering? | Machine Learning Step By Step
Переглядів 19 тис.4 роки тому
How to use Gaussian Mixture Models, EM algorithm for Clustering? | Machine Learning Step By Step
Principal Component Analysis (PCA) Maths Explained with Implementation from Scratch
Переглядів 6744 роки тому
Principal Component Analysis (PCA) Maths Explained with Implementation from Scratch
image for undersamping and over sampling makes no sense. Can you explain what is the orange part and what is the blue part???so confusing...
Where or how do you get the initial 50 data points?
For those who want to understand more about how we get Lagrangian and Dual Formulation - Appendix C called SVM Dual Problem from "Hands On Machine Learning" book might be helpful. Explains Lagrangian and some maths
So much value. Thank you
"Would you be able to provide the references for this material, please? Thank you very much!"
Tutorial en castellano de optimizacion bayesiana, por si a alguien le interesa: ua-cam.com/video/nNRGOfneMdA/v-deo.html
Excellent way to teach❤
Thank you! 😃
man, your videos are amazing. i literally just remembered how great this video is but couldn't find your channel for almost an hour. glad i did. honestly one of my favourites about KFold
Glad you like them :)
Best explanation on GMM algorithm, thanks !
You are welcome!
Awesome and well explained
Thank you !!
Best explained... millions of thanks
Very Glad you found it helpful
can i get the code ?
Very well explained. I got the complete intuition of GMMs in a go.
Nice ! Have you thought how the equations change when its no more normal distribution but is t-dist ? (T-dist is practically more robust)
😀
Very well simplified explanation. Thank you
Glad it was helpful!
best ever explanation for GMM. Thanks a lot for such a helpful video.
Most welcome!
Hello, I’ve been watching your videos. Thank you! They are great. Is there a way to contact you directly?
How can I help you?
Which reference book do you use?
Hello there, Can you give the material notes
Nice sir
well done in <15 min!!! highly recommended
Glad it was helpful.
is the end of the video missing?
Content is not missing. Its a bit short , That's all.
This seems to be correct if and only if you assume that your three models are independant. This is fine, but I think this does not say much in practical cases, as it is very unlikely that you will have 3 base learners that are not correlated. In general, it seems pretty complicated to come up with a "comprehensive" formula that takes into account the respective covariances of these three models with each others and expresses the probabilistic advantage ensembling has.
The formulation has been the premise of why variance reduces theoretically when ensembling is in place compared to independent models. From practical standpoint, it works well which is why random forest is such a star with so many hyperparams to ensure you get different trees as much possible across 100s of features faced in real applications.
More on Statistics.
very clear explanation for decision tree features importance, thanks
You are welcome
You should mention the normalized method. I kill myself to find out how to normalize those numbers
Which normalization would you like to see? The wgt computation in each iteration is normalized. Could you clarify.
can you provide these slides
@machinelearningmaster Great video, thanks! Could you please share the dataset used so that we can replicate this?
Thanks a lot for the video
You are welcome
by far the clearest explanation of bayesian optimization, great work, thanks man!
Glad it was helpful!
Wonderful explaination! Thanks professor.
You are welcome!
is there a mistake in 9:10 ? there is 1 f(x) too much i think. Has to be N(f(x_1), ... (x_n) l o, C*)) / N(f(x_1), ... (x_n) l o, C)). Can anyone confirm this? ty
what a video!!! simple and straight forward
I am glad it was helpful.
Very intuitive explanation!! Thank you so much! I found this gem of a channel today!
I am glad. It was helpful.
thanks but there is a typo in y=-1 wtx+b<= -1 not wtx+b<= 1
This entire series is one of the most comprehensive explanations I have found for SVMs. Extremely grateful for it
I am glad. It was helpful.
Wonderful video, many thanks. Perhaps it would be nice if you made the code available for us learners to play with. Thanks.
Very clear and informative. Thanks!
Glad you found it helpful.
is there any repo containing the codes for the entire playlist
Why don't we inherit from the keras.model class? Thanks alot for your tutorials
Yes, you can do that and make it easier to use in multiple places.
extremely not good explanations
Thanks
Best explanation find the most, thank-you
Glad it was helpful!
You taught better than my instructor 🙂
Thanks so much, the best explanation of xgBoost I´ve seen so far, most people doesnt matter about the math intuition!
Good stuff but consider sharing the Notebooks in the future
Can you please tell me if I have 10 classes of training data, what number of epochs should I use?
In general, I recommend that we set epochs to very large value say 50,000. Then in your code you setup early exit logic as part of training. This will work best for most cases since the training fit will automatically exit when convergence has happened. Hope that helps.
I am a silent internet user, what I usually do when I like a content is just by pressing the like button and save the link on the txt file, but this time is an exception, your content is very simple and completely what I'm looking for. I write you a comment, as the first commentator in this video 😀 You have a new subscriber now. Keep up the good work
Thanks for sharing! I am glad the videos are helpful :)
Wow!!! Excellent lecture!!
Glad you liked it!
Why do you add the mean of the predicted points back to the predicted points?
Lets see if can correlate it with a hypotheses that humans would do to learn. Lets say we are in a Forest & searching for trails of human foot marks to get out of it. Every time we find a footprint, we valid & learn about surroundings, vegetation, terrain,etc. Over a period of time we learn ehat leads to exit And what doen't. That precisely the idea here. Hope that helps.
@@machinelearningmastery I'm sorry but I still don't get it. You can explain it with more math. What I don't get is after predicting a miu, why do we need to add omega? Like what does omega do where?