Thank you for such a valuable explanation. The practical example revealed the potential of these methodologies and your charisma made the video easy to follow. Cheers!
Unsupervised learning is not limited to deep learning. The classic ML method k-means clustering is already able to discover the similar patterns given the samples. I would say that the bright side of deep learning is on the feature extraction. In the old days, we do a lot of work to discover useful features: feature engineering. With deep learning, now we only need to supply the most basic features to the model, pixels for images, raw waveform or spectrogram for speech. This saves my days.
@@estring123 labeled data will still be valuable for some tasks, especially for fine-tuning models, validating performance, and solving new and specific problems. On top of that, having labeled data is critical for certain applications where high accuracy and interpretability are required for example medical diagnosis or safety-critical systems. Depending on the specific machine learning task and the type of data available, the balance between labeled and unlabeled data will vary.
@@estring123 Yes, you will still need labbeled data, the example given in the video is very bad and very wrong, deep learning models are a form of Supervised learning because like in the Video you might ask what in an image of a Pizza makes the algorithm know its a Pizza? The label Pizza is an arbitrary name given by people to it, you need the label to train the network. Back propgation isnt just going backwards like the video suggest, its the algorithm that actually make this Neural networks feasable from a computational possible other with it would be too slow. So why can this deep networks can "learn". The root of it is convolutional Neural networks, the convolutional layer take sections of image and isolants features, where previously the feature selection was crucial for success. Knowing the correct set of convolutional Layers on the other hand is not easy, so it was the combination with Genetic Optimization algorithms that have made them effective. However the output layer will still need labels, unsupervised learning is only useful to find useful features. But a classification problem needs labels this should be obvious, otherwise you cant classify.
I'm pretty sure he just wrote on a tablet of some kind, recorded the screen he was writing on, keyed out the background in a video editor and overlaid and flipped during editing.
Lightboard is a glass chalkboard pumped full of light. It's for recording video lecture topics. You face toward your viewers, and your writing glows in front of you.
Actually, the example is IMHO not well-suited for explaining ML and/pr DL as the aspect of "learning" (which is actually an optimization) is not really addressed by it. So it remains unclear a) what learning actually IS in terms of the example, and b) how the decision making can benefit from the learning aspect of the model.
I hate the idea of weighting variables because if you change them you change the answer. Which to me suggests there is no right or wrong answer - but if you get it right for your business or problem it says to me figuring out how to weight the variables is actually where the true problem and data is.
Introduces bias, which, depending on the scope, would include not just personal bias, but company bias, industry bias, and political bias. Weights and models have this issue.
I love these videos. I just had a tech exec at a Fortune 200 company ask me for any podcasts that could help him stay abreast of current/emerging technology. I didn't have a great answer for him, but I did mention this series. He was looking for more audio-centric content though. Food for thought, @IBM Technology!
We're glad you like the videos! As for a podcast, it's definitely something we're interested in, make sure you're subscribed, we'll be sure to announce it here, if and when it happens.
I don't completely agree on deep learning explanation, because for weight training, labelling is required. Yes pattern/feature extraction can be debated, but labelled data is required
Lol I know this guy from his beer brewing channel. I had to double check if it's actually him. So here I am, learning both how to brew beer and both Deep Learning. Crazy coincidence haha
Thanks for breaking it down in a way that's easy to understand! Your explanation was engaging, and informative, and showed the power of AI and ML. You made a complex topic feel accessible.
It was a great explanation for ML and DL. That Neural Network was a key detail for understanding The difference between ML and DL and their Fundamentals.
To determine the ideal threshold, you need to consider the problem you're trying to solve, the available data, and the desired behavior of the model. In this case, the person wants to make a decision about ordering pizza based on three factors: saving time, losing weight, and saving money. Here's a step-by-step process to help determine an appropriate threshold: Understand the problem: The main goal is to make a reasonable decision about ordering pizza based on the given factors. Analyze the factors and weights: Saving time (X1) has the highest weight (5), so it's the most important factor. Losing weight (X2) has a weight of 3, so it's moderately important. Saving money (X3) has the lowest weight (2), so it's the least important factor. Consider the possible scenarios: Best case: X1 = 1, X2 = 1, X3 = 1 Worst case: X1 = 0, X2 = 0, X3 = 0 Other scenarios: Various combinations of X1, X2, and X3 values Evaluate the weighted sums for each scenario: Best case: (1 * 5) + (1 * 3) + (1 * 2) = 10 Worst case: (0 * 5) + (0 * 3) + (0 * 2) = 0 Other scenarios: Weighted sums will range between 0 and 10 Choose a threshold that aligns with the desired behavior: If the person wants to order pizza only when all factors are favorable (best case), they could set the threshold to 10. If the person wants to order pizza more easily, even if some factors are not favorable, they could set the threshold to a lower value, like 5 or even 0. If the person wants to be more selective and only order pizza when the most important factors are favorable, they could set the threshold to a value between 5 and 10, like 7 or 8. Test and refine the threshold: After setting the initial threshold, the person should test the model with different input combinations and see if the decisions align with their expectations. If the model is making decisions that don't seem reasonable, they can adjust the threshold accordingly. In the given example, a threshold of 5 seems to be a good balance. It allows the model to decide to order pizza when the most important factor (saving time) is favorable, even if the other factors are not. However, it still requires a minimum level of overall favorability to make the decision.
Pizza, Burgers, Tacos. A distinguishing factor between them is their breading. So a NN can learn from human supervision that a pizza image is labeled pizza because of its' carbohydrate type by visually being taught that this breading is pizza dough. Then you can train the NN to identify the ingredients of the dough by training the NN on features of labeled dough (dough such as white flour or almond flour) like their color and texture or unsupervisingly training the NN to distinguish each food's breading by their color or other external features like texture looks (golden bubbles or white flour on the breading) Then the NN can identify a pizza's carbohydrate type from being fed images of pizza dough that's labeled pizza dough manually In the example of pizza, burgers, and tacos, if a NN is fed a labeled data set of different yellows and gold, and outputs the correct answer to the human input, then the NN's training on colors can be used to identify the dough type, My reflections
TL;DR: If an NN has more than 3 layers, it's considered a DNN. DL finds patterns on its own without human supervision, and learns from them. It's a more specific type of ML.
I've always thought that supervised learning is classification, and unsupervised is clustering. Thus DL is always a supervised learning because it still needs a labeled learning set. The differentiation between NN and DL is only in the feature extraction part, NN and "classic" ML require expert knowledge to shape input features, which are computed from the raw data and often normalized. In other words, DL doesn't require labeled features but still needs labeled data to learn from. Also, ML is not only NN but also rule induction algorithms (decision trees, Bayesian rules).
📝 Summary of Key Points: 📌 Deep learning is a subset of machine learning, with neural networks forming the backbone of deep learning algorithms. 🧐 Machine learning uses structured labeled data to make predictions, while deep learning can handle unstructured data without the need for human intervention in labeling. 🚀 Deep neural networks consist of more than three layers, including input and output layers, and can automatically determine distinguishing features in data without human supervision. 💡 Additional Insights and Observations: 💬 Quotable Moments: "Neural networks are the foundation of both machine learning and deep learning, considered subfields of AI." 📊 Data and Statistics: The threshold for decision-making in the example model was set at 5, with weighted inputs influencing the output. 🌐 References and Sources: The video emphasizes the role of neural networks in both machine learning and deep learning, highlighting their importance in AI research. 📣 Concluding Remarks: The video effectively explains the relationship between machine learning and deep learning, showcasing how neural networks play a crucial role in both fields. Understanding the distinctions in layer depth and human intervention provides valuable insights into the evolving landscape of AI technologies. Made with Talkbud
So I do like this series, but this confused me because he switched from one output - "should I buy pizza" - to another output - "is this a pizza or a taco". Is this a fundamental difference in what DL vs ML is able to do? Or that the first output doesn't require as many layers to become a neural network so therefore would always sit at a DL level? Sorry, I think I need to do more study and come back to this video
Dear Martin Keen, I really liked your video and find it extremely useful. However, I wanted also discuss about activation function so the formula you used is - (x1*w1)+(x2*w2)+(x3w3)-threshold. As I understood the threshold is a biggest number used, so that's why you took number 5? Also our w2 is equals to 0, so if the w2 even would be 999999999 (like for us it is super important to be fit) the answer for whole equation would still be positive. So this is my concern about formula if w2 id more prevalent than other options, why in any possible situation we are only capable to have the answer YES ORDER PIZZA. and even x1 and x2 would be 0, but x3=1, with w1 and w2 equaled to 899796 or any other big number we will still get positive outcome. This really baffled me, so I would happy to read your response!
Hello, I think this can be solved by replacing the number '0' with a '-1'. By doing so I guess it will be a more fair output based on our preferences. Good question btw 👍
4:30 but if your interest in staying lean is 10000, the equation still comes to the same conclusion. shouldnt X2 therefore be a choice between -1 and +1?
IBM teachers make a real effort to making learning fun, relatable and digestible.
Facts.
Thank you for such a valuable explanation. The practical example revealed the potential of these methodologies and your charisma made the video easy to follow. Cheers!
Unsupervised learning is not limited to deep learning. The classic ML method k-means clustering is already able to discover the similar patterns given the samples.
I would say that the bright side of deep learning is on the feature extraction. In the old days, we do a lot of work to discover useful features: feature engineering. With deep learning, now we only need to supply the most basic features to the model, pixels for images, raw waveform or spectrogram for speech. This saves my days.
so do you think the need for labelled data will decrease or increase?
@@estring123 labeled data will still be valuable for some tasks, especially for fine-tuning models, validating performance, and solving new and specific problems. On top of that, having labeled data is critical for certain applications where high accuracy and interpretability are required for example medical diagnosis or safety-critical systems. Depending on the specific machine learning task and the type of data available, the balance between labeled and unlabeled data will vary.
@@estring123 Yes, you will still need labbeled data, the example given in the video is very bad and very wrong, deep learning models are a form of Supervised learning because like in the Video you might ask what in an image of a Pizza makes the algorithm know its a Pizza? The label Pizza is an arbitrary name given by people to it, you need the label to train the network.
Back propgation isnt just going backwards like the video suggest, its the algorithm that actually make this Neural networks feasable from a computational possible other with it would be too slow.
So why can this deep networks can "learn". The root of it is convolutional Neural networks, the convolutional layer take sections of image and isolants features, where previously the feature selection was crucial for success. Knowing the correct set of convolutional Layers on the other hand is not easy, so it was the combination with Genetic Optimization algorithms that have made them effective. However the output layer will still need labels, unsupervised learning is only useful to find useful features. But a classification problem needs labels this should be obvious, otherwise you cant classify.
@@pedrorequio5515 With all these explanation and terms did you also understand this as a beginner back then heh
I'm still wondering how he wrote all of those from the opposite projection from us.
He wrote in his natural wiriting direction and the video was flipped left to right during video production before uploading.
I had the similar query. It's amazing btw.
I checked just to see if it was the first comment...
I'm pretty sure he just wrote on a tablet of some kind, recorded the screen he was writing on, keyed out the background in a video editor and overlaid and flipped during editing.
Lightboard is a glass chalkboard pumped full of light. It's for recording video lecture topics. You face toward your viewers, and your writing glows in front of you.
So next time I can't figure out what to have for dinner I just need to build a neural network?
Wow! I am impressed how good you are at explanation such things. I was struggling with it. Thank you.
Actually, the example is IMHO not well-suited for explaining ML and/pr DL as the aspect of "learning" (which is actually an optimization) is not really addressed by it. So it remains unclear a) what learning actually IS in terms of the example, and b) how the decision making can benefit from the learning aspect of the model.
I hate the idea of weighting variables because if you change them you change the answer. Which to me suggests there is no right or wrong answer - but if you get it right for your business or problem it says to me figuring out how to weight the variables is actually where the true problem and data is.
Introduces bias, which, depending on the scope, would include not just personal bias, but company bias, industry bias, and political bias. Weights and models have this issue.
I love these videos. I just had a tech exec at a Fortune 200 company ask me for any podcasts that could help him stay abreast of current/emerging technology. I didn't have a great answer for him, but I did mention this series. He was looking for more audio-centric content though. Food for thought, @IBM Technology!
We're glad you like the videos! As for a podcast, it's definitely something we're interested in, make sure you're subscribed, we'll be sure to announce it here, if and when it happens.
I don't completely agree on deep learning explanation, because for weight training, labelling is required. Yes pattern/feature extraction can be debated, but labelled data is required
\|
Lol I know this guy from his beer brewing channel. I had to double check if it's actually him. So here I am, learning both how to brew beer and both Deep Learning. Crazy coincidence haha
Thanks for breaking it down in a way that's easy to understand! Your explanation was engaging, and informative, and showed the power of AI and ML. You made a complex topic feel accessible.
It was a great explanation for ML and DL. That Neural Network was a key detail for understanding The difference between ML and DL and their Fundamentals.
the way he explained ! Boommed my mind
you are the master of simplicity .thank you so much
I like your style... you IBM people are smart....
Superficial explanation.
From 1-10 this is 20!! Thanks!
Such an awesome explanation, I finally understand the differences between all these different technologies. Thank you so much!
Glad it was helpful!
the way Aliagents structures their AI agents is groundbreaking, can’t wait to see what’s next
Awesome videos! Love your teaching method!
I programmed a 8080 to Jump Non Zero at times. Full Machine code to make side street and main street traffic lights. Worked first time with no bugs.
How did u determine the threshold? How did u come up with -5? Please explain this concept. Thanx!
I also want to know that
did you find out
To determine the ideal threshold, you need to consider the problem you're trying to solve, the available data, and the desired behavior of the model. In this case, the person wants to make a decision about ordering pizza based on three factors: saving time, losing weight, and saving money.
Here's a step-by-step process to help determine an appropriate threshold:
Understand the problem: The main goal is to make a reasonable decision about ordering pizza based on the given factors.
Analyze the factors and weights:
Saving time (X1) has the highest weight (5), so it's the most important factor.
Losing weight (X2) has a weight of 3, so it's moderately important.
Saving money (X3) has the lowest weight (2), so it's the least important factor.
Consider the possible scenarios:
Best case: X1 = 1, X2 = 1, X3 = 1
Worst case: X1 = 0, X2 = 0, X3 = 0
Other scenarios: Various combinations of X1, X2, and X3 values
Evaluate the weighted sums for each scenario:
Best case: (1 * 5) + (1 * 3) + (1 * 2) = 10
Worst case: (0 * 5) + (0 * 3) + (0 * 2) = 0
Other scenarios: Weighted sums will range between 0 and 10
Choose a threshold that aligns with the desired behavior:
If the person wants to order pizza only when all factors are favorable (best case), they could set the threshold to 10.
If the person wants to order pizza more easily, even if some factors are not favorable, they could set the threshold to a lower value, like 5 or even 0.
If the person wants to be more selective and only order pizza when the most important factors are favorable, they could set the threshold to a value between 5 and 10, like 7 or 8.
Test and refine the threshold:
After setting the initial threshold, the person should test the model with different input combinations and see if the decisions align with their expectations.
If the model is making decisions that don't seem reasonable, they can adjust the threshold accordingly.
In the given example, a threshold of 5 seems to be a good balance. It allows the model to decide to order pizza when the most important factor (saving time) is favorable, even if the other factors are not. However, it still requires a minimum level of overall favorability to make the decision.
Your smile made me really enjoy the whole video! Thank you for the wonderful video : )
Thank you! It was a great explanation!
00:20 No Sir - won't do that. Can't learn while digesting pizza.
I appreciate you for broadening my horizons on the subject.
Pizza, Burgers, Tacos. A distinguishing factor between them is their breading. So a NN can learn from human supervision that a pizza image is labeled pizza because of its' carbohydrate type by visually being taught that this breading is pizza dough. Then you can train the NN to identify the ingredients of the dough by training the NN on features of labeled dough (dough such as white flour or almond flour) like their color and texture or unsupervisingly training the NN to distinguish each food's breading by their color or other external features like texture looks (golden bubbles or white flour on the breading) Then the NN can identify a pizza's carbohydrate type from being fed images of pizza dough that's labeled pizza dough manually
In the example of pizza, burgers, and tacos, if a NN is fed a labeled data set of different yellows and gold, and outputs the correct answer to the human input, then the NN's training on colors can be used to identify the dough type,
My reflections
Great. I was always thinking NN and DL are two words for the same thing.
Charismatic presentation...
TL;DR:
If an NN has more than 3 layers, it's considered a DNN.
DL finds patterns on its own without human supervision, and learns from them. It's a more specific type of ML.
Thank you so much. You have explained it brilliantly ❤
very easy well explained thanks!
Bless UA-cams play speed feature.
I've always thought that supervised learning is classification, and unsupervised is clustering. Thus DL is always a supervised learning because it still needs a labeled learning set. The differentiation between NN and DL is only in the feature extraction part, NN and "classic" ML require expert knowledge to shape input features, which are computed from the raw data and often normalized. In other words, DL doesn't require labeled features but still needs labeled data to learn from. Also, ML is not only NN but also rule induction algorithms (decision trees, Bayesian rules).
I was hungry ignorant before this video. Now I learned about deep learning but it made me deeply hungry.
Hi there. Very helpful. Thank you.
Content is great. Audio is too low on these videos.
The presentation was amazing!
Excellent video. Thank you
Phenomenal easy explanation ❤
You're making education engaging and accessible for everyone. #NurserytoVarsity
Thank you! Summary: deep learning is not so deep after all!
I think there is a confusion between feature extraction and unsupervised learning. Hope that you can revise it
Fantastic distillation of the concepts.
Are the presenters mirror images to make their writing appear the way it does or is it another tech trick?
That was very interesting and a great explanation of machine and deep learning.
Thank you for valuable information 🙏🙏
Great explanation
TY
Thank you
📝 Summary of Key Points:
📌 Deep learning is a subset of machine learning, with neural networks forming the backbone of deep learning algorithms.
🧐 Machine learning uses structured labeled data to make predictions, while deep learning can handle unstructured data without the need for human intervention in labeling.
🚀 Deep neural networks consist of more than three layers, including input and output layers, and can automatically determine distinguishing features in data without human supervision.
💡 Additional Insights and Observations:
💬 Quotable Moments: "Neural networks are the foundation of both machine learning and deep learning, considered subfields of AI."
📊 Data and Statistics: The threshold for decision-making in the example model was set at 5, with weighted inputs influencing the output.
🌐 References and Sources: The video emphasizes the role of neural networks in both machine learning and deep learning, highlighting their importance in AI research.
📣 Concluding Remarks:
The video effectively explains the relationship between machine learning and deep learning, showcasing how neural networks play a crucial role in both fields. Understanding the distinctions in layer depth and human intervention provides valuable insights into the evolving landscape of AI technologies.
Made with Talkbud
Maravilhoso! Amei o vídeo, nota 1000000...
humanities are fundamental in this proccesses! now the funny example is pizza, what about human rights? who's feeding the bias to the algorithms???
This is awesome, thank you ♥
Thanks for the information given to me.
Thank you for your useful video🙏🏻
You did well explaining mate, no idea what they’re talking about
I'm impressed how he can write backwards so good haha
See ibm.biz/write-backwards
It's a recruiting criteria for IBM
Nice, loved it.
Steve Brunton style is becoming a genre...
Great job 👏
I didn't know that Gordon Ramsay gives lessons about Machine learning and deep learning. for real tho the video was amazing and very helpful
Good one
Thanks!
Thank you for your amazing and funny explanation
Great Video. I would like to understand more details about the layers. What are layers from a logical and technical prospective?
They are computing processes, I think.
It's so interesting.
Hmm so how did the subtraction of 5 threshold play an impact? What does that even mean at all? Simple and Great explanantion!!!!
Honestly, you had me at pizza
where are all the neurons, weights and biases stored? in ram, in a database? what datastructure is used?
"It's time for lunch!" lol. I love this video. Thanks so much!
Please provide, is multi layer neural network a deep learning model ? If not, please provide me an example of deep learning model.
How it's possible,so risk trade but you win so amazing. I from bangladesh.i want do it same to you
Thanks a lot
Thank you :)
Homebrew Challange guy!
bro managed to make an example about pizza... and i was eating it while watching this video 💀
5:08 classical ML human intervention
😂😂😂 hunger
Very nice bhai 👌🏻
greate!
UPSC aspirant from India 😊
Academically speaking, should AI not be a subset of DL? I think you’ve done a commercial magic trick here.
No.
I loved it❤️
Adam Gilchrist in the house
good one!
Beautifulllllllll ❤️❤️❤️😊
In back propagation. The error is synonymous for weight?
the moment he said pizza, i just pause and ordered one and resume when i got pizza.
I don't understand how you suddenly use 1(yes) and 0(no) as numbers to calculate with?
Yes, any weight attributed to x2 will result in 0. Can someone please explain this?
Plot twist, most people were eating while watching this video.
So I do like this series, but this confused me because he switched from one output - "should I buy pizza" - to another output - "is this a pizza or a taco". Is this a fundamental difference in what DL vs ML is able to do? Or that the first output doesn't require as many layers to become a neural network so therefore would always sit at a DL level? Sorry, I think I need to do more study and come back to this video
Free pizza? There's nothing to calculate there. That pizza is mine!
Great video. Thank you
You used threshold as 5 what actually threshold means according to your example of pizza ?
Dear Martin Keen, I really liked your video and find it extremely useful. However, I wanted also discuss about activation function so the formula you used is - (x1*w1)+(x2*w2)+(x3w3)-threshold. As I understood the threshold is a biggest number used, so that's why you took number 5? Also our w2 is equals to 0, so if the w2 even would be 999999999 (like for us it is super important to be fit) the answer for whole equation would still be positive. So this is my concern about formula if w2 id more prevalent than other options, why in any possible situation we are only capable to have the answer YES ORDER PIZZA. and even x1 and x2 would be 0, but x3=1, with w1 and w2 equaled to 899796 or any other big number we will still get positive outcome. This really baffled me, so I would happy to read your response!
Hello, I think this can be solved by replacing the number '0' with a '-1'.
By doing so I guess it will be a more fair output based on our preferences.
Good question btw 👍
My question as well.
respect for writing backwards so the camera sees normal
This video was...pizzawesome! 🍕
nice
4:30 but if your interest in staying lean is 10000, the equation still comes to the same conclusion. shouldnt X2 therefore be a choice between -1 and +1?
Top tip: the worst thing you can do when you’re learning is eat food beforehand :)