13.3.1 L1-regularized Logistic Regression as Embedded Feature Selection (L13: Feature Selection)

Поділитися
Вставка
  • Опубліковано 7 лис 2024

КОМЕНТАРІ • 5

  • @amrelsayeh4446
    @amrelsayeh4446 4 місяці тому

    @sebastian At 13:20, why is the solution between the global minimum and the penalty minimum lie somewhere where one of the weights is zero. In other words, why it should lie at the corner of the penalty function not just at the line. between the global minimum and the penalty minimum.

  • @AyushSharma-jm6ki
    @AyushSharma-jm6ki Рік тому

    @sebastian amazing video. Thanks for sharing. I am getting deeper understanding of these topics with your videos.

  • @arunthiru6729
    @arunthiru6729 2 роки тому +3

    @sebastian I think using Logistic Regression directly for feature selection based on respective weights/coefficients means we are assuming all dimensions/features are independent. I understand this is not the correct way to do this. Pls advise.

    • @SebastianRaschka
      @SebastianRaschka  2 роки тому +1

      Yes, this assumption is correct. ML is full of trade-offs 😅. If you cannot make this assumption, I recommend the sequential feature selection approach

    • @wayne7936
      @wayne7936 6 місяців тому

      Thanks for pointing this out!