Hello GPT-4o-mini & Mistral NeMo!!!

Поділитися
Вставка
  • Опубліковано 13 жов 2024

КОМЕНТАРІ • 43

  • @dubesor
    @dubesor 2 місяці тому +1

    I tested both and in my findings, gpt-4o mini is about DeepSeek-V2 Chat level (which is half the price still).
    Nemo -12b is about Claude-3-Sonnet level (for 2-3% the price).
    More options are good, and small efficient models are welcomed, but I really wanna see a few more large flagships.

  • @thenoblerot
    @thenoblerot 2 місяці тому +5

    Native audio/video *input AND output* is huge. Looking forward to that!
    I'm gonna try 4o-mini in my robot project. Currently using Claude Haiku, but as much as I love Claude, Haiku doesn't do well with function calling.

  • @marcfruchtman9473
    @marcfruchtman9473 2 місяці тому +1

    Interesting. It looks like it is much lower cost, and still maintains 2nd place in MMLU, GPQA, DROP, MGSM, MATH, HumanEval, and MMU, with 3rd place in MathVista... overall, it is very impressive for the cost. (Will definitely try it out).
    Thanks for the video.

  • @Kalaanoo
    @Kalaanoo 2 місяці тому +1

    Thanks for the Video bro. One Question, am I correct that Gemini Flash still has that free-price tier, right? Since my workload is high and I was considering if makes sense empty out the free tier from Flash for easier high volume jobs?
    Am I missing something from here? Thanks!!

    • @1littlecoder
      @1littlecoder  2 місяці тому

      That's absolutely true. Gemini Flash has got a free Tier (with rate limits) but very good to prototype.

  • @dievas_
    @dievas_ 2 місяці тому +2

    Great news video, my dude, thank you!

  • @henkhbit5748
    @henkhbit5748 2 місяці тому

    Thanks for the update. How many languages does it support? For example does it support turkish, dutch, polish, thai and ...? Good for us, prices will drop to other competitors also like google and anthropic. Hopefully Mistral or Meta will release an open source competitive multimodal llm.

  • @unclecode
    @unclecode 2 місяці тому +1

    Hahaha agreed, Sam Altman likely did this intentionally! This is indeed cost effective! The funny thing about Gemeni flash context caching is that the content must consist of at least 32,769 tokens. This means that in many real-world cases, like in RAG where the context is fetched from a larger datastore, and it's smaller than 32,769 tokens, you won't experience the caching benefits. This seems more like a sales/marketing constraint rather than a technical one, reminiscent of those pesky terms and conditions with small font sizes that you overlook when purchasing the product.

    • @1littlecoder
      @1littlecoder  2 місяці тому +1

      That is a very interesting detail that I might have overlooked as well. Thanks for sharing

  • @AleksaMilic-d2e
    @AleksaMilic-d2e 2 місяці тому +1

    Can chatgpt 4o mini browser the internet?? Thanks for the answer

  • @98Shorty_Story
    @98Shorty_Story 2 місяці тому

    Should I think of the gpt4 mini as a model in between this, better than the gpt4 but less performing than the gpt4o?

  • @noorahmadharal
    @noorahmadharal 2 місяці тому +5

    Amazing news 🎉

  • @carlkim2577
    @carlkim2577 2 місяці тому

    Could you do a comparison bus between this and Gemini flash? Especially for vision. Take on image recognition and data extraction.

  • @100_IQ_EQ
    @100_IQ_EQ 2 місяці тому

    Could you make a video on what kind of apps one can build on top of these and make money? Also, how can one build apps on these? Also, how one can build skills to deliver requirements for genAI and other LLM based jobs on freelance websites like upwork?
    I have played tekken. Characters like Haihachi, Yoshimitsu. It also has a movie on primevideo.

  • @saksham3
    @saksham3 2 місяці тому +2

    Using the gpt-4o-mini API, when I ask, "Which model are you?" I get the reply, "I am based on OpenAI's GPT-3 model, which is designed to..." 🤔
    Even in the playground, the same thing happens. However, when I select gpt-4o-mini in the Android app and ask the same question, it says it is based on the gpt-4 architecture.
    This is very strange! 😱

    • @MrKrzysiek9991
      @MrKrzysiek9991 2 місяці тому +3

      gpt-4o-mini is very bad in my tests it was performing worst than LLama 3 8B :D

    • @RyluRocky
      @RyluRocky 2 місяці тому

      This isn’t a fair test, OpenAI (for whatever reason) never programs the model to know its own name, which is simply a knowledge issue and has little to do with intelligence.

  • @Sri_Harsha_Electronics_Guthik
    @Sri_Harsha_Electronics_Guthik 2 місяці тому +4

    Tekken 7 now i think. used to pay 2 rs per coin for gon!

    • @1littlecoder
      @1littlecoder  2 місяці тому +2

      Exactly that arcade Machine. I don't remember how much I used to put in.

  • @ps3301
    @ps3301 2 місяці тому +1

    I tried to switch to mini version. It wouldn't let me!!

  • @ojikutu
    @ojikutu 2 місяці тому +7

    RIP togetherAI, DeepInfra, OctoAI and all the inferencing as a service providers.

    • @1littlecoder
      @1littlecoder  2 місяці тому +2

      Wiped!

    • @blisphul8084
      @blisphul8084 2 місяці тому +1

      4o mini is still way slower than models running in Groq. I'll stick to Gemma 2 9b on Groq for now, since it's 500 t/s and works well enough.

    • @SloanMosley
      @SloanMosley 2 місяці тому +1

      They still have unmoderated and fine-tuned models. They do still have cheaper models as well, but not as smart at that price.

    • @ravishmahajan9314
      @ravishmahajan9314 2 місяці тому

      Why?

    • @MrKrzysiek9991
      @MrKrzysiek9991 2 місяці тому +1

      Better you test it first ...

  • @MichealScott24
    @MichealScott24 2 місяці тому +3

    wohuu lets go it would be fun now :) to utilise the API like for poor broke people like me or student currently and who have limited budget!❤

    • @1littlecoder
      @1littlecoder  2 місяці тому +2

      @@MichealScott24 absolutely

    • @MichealScott24
      @MichealScott24 2 місяці тому +2

      - gpt 3.5 retired which means for free users they get data-analysis - vision model/rag & web search which would be cool

    • @randomdreamer1619
      @randomdreamer1619 2 місяці тому

      Yeah and you better use the free api for now, gpt4o mini doesnt let me to put pdf on it in web or app version.

    • @randomdreamer1619
      @randomdreamer1619 2 місяці тому

      Despite it says its able to process pdf..... and now u can stop using copilot that love to dodge ur question.

  • @serkhetreo2489
    @serkhetreo2489 2 місяці тому +1

    Is it going to be on the site

    • @1littlecoder
      @1littlecoder  2 місяці тому +1

      @@serkhetreo2489 that's what they mentioned. But it's not yet

    • @serkhetreo2489
      @serkhetreo2489 2 місяці тому

      @@1littlecoder okay 👍

  • @AngeloXification
    @AngeloXification 2 місяці тому +1

    Your turn Anthropic

  • @FushigiMigi
    @FushigiMigi 2 місяці тому

    They need a better and more free fine tuning ability (more user friendly) imo.

  • @ravishmahajan9314
    @ravishmahajan9314 2 місяці тому

    Why people are talking about Microsoft Phi model. It's jusr 2 B model.
    Are there any use cases thay can be done via such small models?
    One use case i found was sentiment analysis & ticket classification.
    But are there any other use cases?
    Also can we use 2B modle using my normal laptop with 8GB ram?

    • @ickorling7328
      @ickorling7328 2 місяці тому

      Yes. It will run slow, but yes. An iGPU on windows allocates RAM as vRAM dynamically.

  • @riser9644
    @riser9644 2 місяці тому +1

    azure openai is so slow