Active Learning in ML

Поділитися
Вставка
  • Опубліковано 18 гру 2024

КОМЕНТАРІ • 29

  • @lakshay510
    @lakshay510 3 роки тому +2

    Just did this yesterday in my organisation. Kudos to you guys for posting videos like this.

  • @saivarma7495
    @saivarma7495 3 роки тому +2

    For smart picking: We can apply any of the clustering algorithms on Du. Then pick small number of points from each cluster and provide them to experts for labelling

    • @AppliedAICourse
      @AppliedAICourse  3 роки тому +4

      But, picking points randomly from each cluster in Du may not always guarantee improvement with M1 as most of the points so picked could be very similar to points in DL.

  • @chathere2359
    @chathere2359 3 роки тому

    I really trying hard to arrange money for your course.
    Ofcourse u r greatest teacher sir, reason is there are many institutions does charge lakhs 😅

  • @abhinayabhi91
    @abhinayabhi91 2 роки тому

    I pick this video smartly so that I learn manually and written in exams easily
    Active learning in my view 😊

  • @shyamshankarmathurjayashan4910
    @shyamshankarmathurjayashan4910 3 роки тому +1

    For multiclass classification(say we have k classes), can we use a softmax logic and choose those points which give out probabilities around 0.5 for atleast k/2 classes (since it is confused for half of the classes now, such points could be a good pick to our subset)?

  • @RajeshSharma-bd5zo
    @RajeshSharma-bd5zo 3 роки тому +1

    For multi-class classification, we can go ahead with human labelling of class for which model is not behaving up to the mark or in other words making more % of errors in some classes. By introducing human labelled data for such classes we can bring some confidence in our new model M1.

    • @AppliedAICourse
      @AppliedAICourse  3 роки тому

      But, we don’t have class labels for points in Du. So, how can we determine if a point is erroneously classified or not.

  • @GPravinshankar
    @GPravinshankar 3 роки тому +1

    For multiclass classification( for example: 3 classes), we have to choose data points for which the model is giving roughly equal probability(0.33) for all the 3 classes.

    • @AppliedAICourse
      @AppliedAICourse  3 роки тому +1

      This thinking is in the right direction. Would you not pick those points where the class probabilities in a three class classification are .4, .5, .1? In this case, the model is confused between two of the three classes.

    • @GPravinshankar
      @GPravinshankar 3 роки тому

      @@AppliedAICourse yes, we should consider this case as well.

  • @priyushkumar9689
    @priyushkumar9689 3 роки тому

    for multiclass classification , we will set a threshold for probability that if the max probability among all the classes is less then certain threshold value (< 0.7) we will choose that data point for human labelling.

  • @shreymishra646
    @shreymishra646 3 роки тому +1

    Hi please correct me if I am wrong , for multiclass setting we can calculate entropy , since the entropy works on a similar principle where the spread of points is almost equal then the entropy should be high as compared to high seperation , so we can filter all the low entropy points

    • @AppliedAICourse
      @AppliedAICourse  3 роки тому

      Perfect, Entropy is a very popular metric that can be used to numerically quantify the uncertainty in class labels. That's why we use cross-entropy as the loss function in multi-class classification.

    • @devanshverma5395
      @devanshverma5395 3 роки тому

      Excellent approach

  • @anujsali3171
    @anujsali3171 3 роки тому

    A doubt: Why cant we use Unsupervised Techniques like Clustering to label the large Unlabelled data? Just curious🤔

    • @AppliedAICourse
      @AppliedAICourse  3 роки тому

      But, picking points randomly from each cluster in Du may not always guarantee improvement with M1 as most of the points so picked could be very similar to points in DL.

  • @dwarikaprasadteli1030
    @dwarikaprasadteli1030 3 роки тому

    Could you please make a explanation for "Extreme multi label classification" problem, Which is also covered in course (StackOverflow Tag prediction problem) but in that case we limited ourselves to use few labels. what kind of solution we can apply for these type of problem..

  • @CRTagadiya
    @CRTagadiya 3 роки тому

    the method you mentioned might not work if we don't have probabilistic model ( the model which only gives class not the probability). what should we do now?

    • @AppliedAICourse
      @AppliedAICourse  3 роки тому

      Most machine learning and deep learning models can be slightly modified to obtain class probabilities. Hence, this is not a major issue in the real world.

  • @anujsali3171
    @anujsali3171 3 роки тому

    My opinion to how to extend it to multi-class setting: In ML, if we're given with a k-class classification problem, we can solve it using k binary classifiers. So I think One Vs Rest we can employ. Just my thinking.

    • @AppliedAICourse
      @AppliedAICourse  3 роки тому

      While this is a possible solution, can you think of alternative and simpler methods where you don’t have to build k binary-classifiers.

  • @ehshankhan7003
    @ehshankhan7003 3 роки тому

    Suppose we have three class classifications, our threshold could be 0.33. But if we get a probability like 0.1, 0.4 and 0.5. We can choose both the 0.4 and 0.5 data as a sample for labelling.

    • @AppliedAICourse
      @AppliedAICourse  3 роки тому

      Note that there is only one xi that has these three probabilities for the three classes. It’s just a single point, not two or three. But, you are true that we will pick can xi’s and here the model is less certain based on the fact that no single class has a high probability like 0.9 or so.

  • @aakashverma1622
    @aakashverma1622 3 роки тому

    What i think is ....we can do 1 vs all.... let's say we have 10 classes....and we can get probability whether a point belong to that class or not...if the probability of a point belonging to a class is very high...we will leave that point.....and we will do same for all classes......in the end we will be left with points that has no high probability of belonging to any classes....we will choose those points

  • @True_Feelingsss...
    @True_Feelingsss... 3 роки тому

    Go with One vs Rest approach.

    • @AppliedAICourse
      @AppliedAICourse  3 роки тому

      While this is a possible solution, can you think of alternative and simpler methods where you don’t have to build k binary-classifiers.

    • @True_Feelingsss...
      @True_Feelingsss... 3 роки тому

      @@AppliedAICourse Method 1: Sample points from Du such that wherever model will fail to achieve 0.9 probability for particular class
      Method 2: For given point, Pick two highest class probabilities if there difference is less than some alpha e.g (0.3) then that particular point is not sure point.

  • @shyamshankarmathurjayashan4910
    @shyamshankarmathurjayashan4910 3 роки тому +3

    For multiclass classification(say we have k classes), can we use a softmax logic and choose those points which give out probabilities around 0.5 for atleast k/2 classes (since it is confused for half of the classes now, such points could be a good pick to our subset)?