Text Vectorization NLP | Vectorization using Python | Bag Of Words | Machine Learning

Поділитися
Вставка
  • Опубліковано 28 гру 2024

КОМЕНТАРІ • 14

  • @abdullahkabir6129
    @abdullahkabir6129 7 місяців тому

    Cool explanation

  • @kirandeepmarala5541
    @kirandeepmarala5541 4 роки тому +3

    Nice Ranjan...Good explanation..You have good Teaching Skills.. Your channel will have nice growth in Future

    • @RanjanSharma
      @RanjanSharma  4 роки тому

      Thank you so much 🙂 Glad you liked it.

    • @CaughtBehindOriginal
      @CaughtBehindOriginal Рік тому

      @@RanjanSharma Why you stopped making videos? Just before the boom 😶

  • @jadhavpruthviraj5788
    @jadhavpruthviraj5788 2 роки тому

    Excellent Explations

  • @crazypahadan1
    @crazypahadan1 4 роки тому

    Congratulations 1000 subcribers good going 👌 keep it up

    • @RanjanSharma
      @RanjanSharma  4 роки тому

      Thank you so much 😀 Many thanks.

  • @smcgpra3874
    @smcgpra3874 Рік тому

    I was checking infused on steps
    Rwa to token text cleaning then vector and then apply ml models or algo
    Thank you for clearing

  • @rontackaberry8202
    @rontackaberry8202 Рік тому

    Great explanation but I'm confused why it only tokenized 11 words instead of the actual 12 words. It seems to be missing the token for the word "I".

  • @souvikganguly3236
    @souvikganguly3236 Рік тому

    In your theoritical explanation "I" has considered as unique one but whenever it comes to practical implementation then "I" has not considered as unique one. Can you please explain why does it occur?

  • @smudgepost
    @smudgepost Рік тому

    A very good explanation thank you. I'd like to see you demonstrate a matrix where lemmatization is used so I can visualize how the text pre-processing isn't just reduced to a numerical value but also see how sentiment is assessed, for cases where it is needed.

  • @teacode2280
    @teacode2280 3 роки тому +2

    i got error in line no. 42 in your code
    sentenses = tokenize.sent_tokenize(corpus)
    NameError: name 'corpus' is not defined

    • @kirusihaansathiyapalan1939
      @kirusihaansathiyapalan1939 2 роки тому +1

      the variable should be there: paragraph
      sentenses = tokenize.sent_tokenize(paragraph)