Data Scaling in Neural Network | Feature Scaling in ANN | End to End Deep Learning Course

Поділитися
Вставка
  • Опубліковано 1 лип 2024
  • Data scaling is a recommended pre-processing step when working with deep learning neural networks. Data scaling can be achieved by normalizing or standardizing real-valued input and output variables.
    Code - colab.research.google.com/dri...
    Video on Standardization: • Feature Scaling - Stan...
    Video on Normalization: • Feature Scaling - Norm...
    ============================
    Do you want to learn from me?
    Check my affordable mentorship program at : learnwith.campusx.in
    ============================
    📱 Grow with us:
    CampusX' LinkedIn: / campusx-official
    CampusX on Instagram for daily tips: / campusx.official
    My LinkedIn: / nitish-singh-03412789
    Discord: / discord
    👍If you find this video helpful, consider giving it a thumbs up and subscribing for more educational videos on data science!
    💭Share your thoughts, experiences, or questions in the comments below. I love hearing from you!
    ⌚Time Stamps⌚
    00:00 - Intro
    03:15 - Code Demo/Feature Scaling
    06:00 - Intuition

КОМЕНТАРІ • 41

  • @maanuiitd
    @maanuiitd 2 місяці тому +1

    Thank you so much for this exemplary video. Just wanted to add:
    1. Normalization: It can be used when data does contain outliers. Scikit-Learn provides a transformer called MinMaxScaler for this. It has a feature_range hyperparameter that lets you change the range if, for some reason, you don’t want 0-1 (e.g., neural networks work best with zero-mean inputs, so a range of -1 to 1 is preferable).
    2. Standardization: Unlike minmax scaling, standardization does not restrict values to a specific range. However, standardization is much less affected by outliers.
    3. Data with a heavy tail: When a feature’s distribution has a heavy tail (i.e., when values far from the mean are not exponentially rare), both min-max scaling and standardization will squash most values into a small range. Models generally don’t like this at all. So before you scale the feature, you should first transform it to shrink the heavy tail, and if possible to make the distribution roughly symmetrical. For example, a common way to do this for positive features with a heavy tail to the right is to replace the feature with its square root (or raise the feature to a power between 0 and 1). If the feature has a really long and heavy tail, such as a power law distribution, then replacing the feature with its logarithm may help. Another approach to handle heavy-tailed features consists in bucketizing the feature. This means chopping its distribution into roughly equal-sized buckets, and replacing each feature value with the index of the bucket it belongs to.
    4. Data with multiple peaks: When a feature has a multimodal distribution (i.e., with two or more clear peaks, called modes), it can also be helpful to bucketize it, but this time treating the bucket IDs as categories, rather than as numerical values. This means that the bucket indices must be encoded, for example using a OneHotEncoder (so you usually don’t want to use too many buckets). This approach will allow the model to more easily learn different rules for different ranges of this feature value. Another approach to transforming multimodal distributions is to add a feature for each of the modes (at least the main ones). The similarity measure is typically computed using a radial basis function (RBF)-any function that depends only on the distance between the input value and a fixed point. The most commonly used RBF is the Gaussian RBF, whose output value decays exponentially as the input value moves away from the fixed point.
    Reference: 3rd Edition Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow written by Aurelien Geron

  • @CODEToGetHer-rq2nf
    @CODEToGetHer-rq2nf 7 місяців тому +7

    God level teacher ❤️🤌🏻

  • @VarunSharma-ym2ns
    @VarunSharma-ym2ns 2 роки тому +5

    Nicely explained in hindi to understand......Helpful in interviews

  • @nayeem634
    @nayeem634 2 роки тому +13

    Sir, kindly add KNN and SVM algorithms to your 100 days of ML playlist thus it becomes more impressive.
    And a request to continue the deep learning regularly.

    • @sidindian1982
      @sidindian1982 Рік тому +1

      yess KNN is pending,,. , but SVM you can fetch old vedios on concepts & codes - guess its 2 yrs old one... just recheck ..

  • @pravinshende.DataScientist
    @pravinshende.DataScientist 2 роки тому +4

    hi sir before wathcing your vedio I am very happy because i know after watching i will get some new insight so i am thankful before watching this vedio bc you have created such a moment for us!!F

  • @SurajitDas-gk1uv
    @SurajitDas-gk1uv Рік тому +3

    Very good tutorial. Thank you very much for such a good tutorial.

  • @narendraparmar1631
    @narendraparmar1631 4 місяці тому +1

    Thanks for this easy explanation

  • @bhupendersharma0428
    @bhupendersharma0428 Рік тому +3

    Amazing Explanation SIr

  • @sujithsaikalakonda4863
    @sujithsaikalakonda4863 8 місяців тому +2

    very well explained sir.

  • @nationfirst-worldaffairs6410
    @nationfirst-worldaffairs6410 2 роки тому +4

    Well done teaching sir .....🤗🤗

  • @nishitaverma8805
    @nishitaverma8805 2 роки тому +2

    Waiting for the next video in the playlist

  • @ParthivShah
    @ParthivShah 2 місяці тому +1

    Thank You Sir.

  • @zkhan2023
    @zkhan2023 2 роки тому +2

    Guys sub like kero take sir bhi motivated rahain

  • @roshankumargupta46
    @roshankumargupta46 2 роки тому +4

    Your videos and content quality is way too good instead of interview oriented. Thanks.
    Any planning for projects on deep learning?

  • @zkhan2023
    @zkhan2023 2 роки тому +2

    good content sir jee

  • @barunkaushik7015
    @barunkaushik7015 Рік тому +1

    Superb

  • @shanu9494
    @shanu9494 Рік тому +2

    Sir, please continue Deep Learning and NLP playlists

  • @rb4754
    @rb4754 25 днів тому

    Well explained... Amazing!!!!!!!!!

  • @narendravarma4363
    @narendravarma4363 3 місяці тому

    I am doing pcos ultrasonic image classification whether infected or not infected.I collected the dataset from kaggle.During the training,I trained my model by rescaling the image.Do I need to rescale the test data also.

  • @MohiuddinShojib
    @MohiuddinShojib Рік тому +2

    If I have date type data in my features , do I need to also scaling ?

  • @akshaykanavaje474
    @akshaykanavaje474 2 роки тому +1

    thank u sir

  • @sandipansarkar9211
    @sandipansarkar9211 Рік тому +2

    finished watching

  • @bhushanpatel3902
    @bhushanpatel3902 Рік тому +1

    tooo good.....

  • @yashjain6372
    @yashjain6372 Рік тому +2

    best

  • @user-wp7zk9ee5z
    @user-wp7zk9ee5z 6 місяців тому

    Great video sir..!
    One correction is that after applying StandardScaler the values doesnt range between -1 to 1. Even in your code after applying scaling technique there are values more than 1 and -1.

  • @AmanSahu-vj9uu
    @AmanSahu-vj9uu Рік тому +1

    I have a question sir.
    Should we scale the dummy variables that are obtained when encoding categorical variables in the dataset, for Artificial Neural Networks models specifically?
    Some say that dummy variables should not be scaled as their values (0 & 1) already lie between -3 and 3 and by scaling them, the variables actually lose their interpretation.
    But some say that in ANN specifically it is ABSOLUTELY necessary to scale all the features including the encoded categorical variables.
    I'm confused as to what is right and why.
    I'm hoping to know your remarks on this.

    • @maanuiitd
      @maanuiitd 2 місяці тому

      IMO, you can scale other variables to match the scale of the dummy variables.

  • @jabed.akhtar
    @jabed.akhtar Рік тому +1

    Sir ji, kya aap please ye pura OneNote share karsake hai?

  • @surajprusty6904
    @surajprusty6904 3 місяці тому

    Here we can see the values are actually going beyond -2 to +2 instead of -1 to +1. Why is that?

  • @DarkShadow00972
    @DarkShadow00972 5 місяців тому

    Hi sir how can I get ur one notes DL.

  • @znyd.
    @znyd. Місяць тому

    💛

  • @kindaeasy9797
    @kindaeasy9797 2 дні тому

    Niceee

  • @bollarapuphanindra1264
    @bollarapuphanindra1264 2 роки тому +1

    sir pls provide link to the neural network performance word file that you created thank you

    • @campusx-official
      @campusx-official  2 роки тому +1

      docs.google.com/document/d/1YjuG6xj10ALmltbEqmxu1wOBWMIV5BU2DmVP05xlZ3c/edit?usp=drivesdk

  • @ajaykulkarni576
    @ajaykulkarni576 2 роки тому +1

    what digital pen are you using?

  • @mr.deep.
    @mr.deep. 2 роки тому +3

    Sir activation function pe video kab aaye ga

  • @ShubhamVerma-wf3vc
    @ShubhamVerma-wf3vc 2 роки тому +1

    Bhaiya interview preparation pai video kab aayega?