КОМЕНТАРІ •

  • @big_zzzzz
    @big_zzzzz 4 місяці тому +1

    Priceless info!

  • @KarenWeissKarwei
    @KarenWeissKarwei 4 місяці тому +1

    Great video, informative and understandable. Thank you!

  • @abdshomad
    @abdshomad 4 місяці тому +3

    As always, the content is well delivered. Thank you for always share the knowledge 👍

    • @SkalskiP
      @SkalskiP 4 місяці тому

      my pleasure!

  • @sumukharaghavanm6466
    @sumukharaghavanm6466 4 місяці тому +1

    Great solution for students
    Thanks a lot!!!!

  • @Codewello
    @Codewello 4 місяці тому +1

    Awesome as always! I have learned a lot from you, especially about supervision Also, I love the thumbnail.
    You look like you're saying 'come at me, bro 😁😁

    • @Roboflow
      @Roboflow 4 місяці тому

      Glad to hear it!

  • @uttamdwivedi7709
    @uttamdwivedi7709 4 місяці тому +7

    Great work !!! Could you provide a tutorial on how to train (finetune) this YOLO-World model on specific type of data?

    • @Roboflow
      @Roboflow 4 місяці тому +8

      I'll think about it. If enough people are interested we could at least write a blog.

    • @elhadjmeguellati3031
      @elhadjmeguellati3031 4 місяці тому

      interested and thanks for the very usefull content
      @@Roboflow

    • @scottsharp142
      @scottsharp142 4 місяці тому

      Yes, would love too see this as well. Thanks for great content.

    • @smartfusion8799
      @smartfusion8799 3 місяці тому

      Yes please, is it possible to run a fine tuned /light version on a edge device?

  • @LukasSmith827
    @LukasSmith827 4 місяці тому +2

    the best to ever do it

    • @SkalskiP
      @SkalskiP 4 місяці тому

      haha you are too nice! But thanks!

  • @elvenkim
    @elvenkim 4 місяці тому +1

    Hi Pieter! Great delivery, love the final video on YOLO + SAM. May I check with you on how do we extract the coordinate of the bounding box?

    • @Roboflow
      @Roboflow 4 місяці тому +1

      In my code just access detections.xyxy :)

    • @elvenkim
      @elvenkim 4 місяці тому

      @@Roboflow many thanks Pieter!

  • @jimshtepa5423
    @jimshtepa5423 Місяць тому

    have you done any video on training a model for custom dataset?

  • @nidalidais9999
    @nidalidais9999 Місяць тому

    hi man , good work , what the difference between YOLO-World and T-REX model , and how to compare between models usually

  • @wolpumba4099
    @wolpumba4099 Місяць тому +1

    *YOLO-World Explained: A Bullet List Summary with Timestamps*
    *What is YOLO-World? (**0:00**)*
    * A cutting-edge, zero-shot object detection model that's 20x faster than predecessors. (0:24)
    * Uses a "prompt-then-detect" paradigm to achieve speed, encoding prompts offline and reusing embeddings. (2:26)
    * Leverages a faster CNN backbone and streamlined architecture for increased efficiency. (2:57)
    * Outperforms previous zero-shot detectors (like GroundingDINO) in terms of speed while maintaining accuracy. (2:12)
    *Advantages of YOLO-World:*
    * No need for custom dataset training for object detection. (0:42)
    * Real-time video processing capabilities (up to 50 FPS on powerful GPUs). (9:22)
    * Can incorporate color and position references in prompts for refined detection. (10:16)
    *Limitations of YOLO-World (**13:16**):*
    * Still slower than traditional real-time object detectors. (13:34)
    * May be less accurate than models trained on custom datasets, especially in uncontrolled environments. (13:51)
    * Can misclassify objects, particularly with low-resolution images or videos. (14:19)
    *Using YOLO-World Effectively (**5:33**):*
    * Experiment with different confidence thresholds for optimal results. (7:14)
    * Utilize non-max suppression (NMS) to eliminate duplicate detections. (8:07)
    * Filter detections based on relative area to remove unwanted large bounding boxes. (11:04)
    * Combine with FastSAM or EfficientSAM for zero-shot segmentation tasks. (15:21)
    *Beyond the Basics (**15:08**):*
    * YOLO-World opens possibilities for open-vocabulary video processing and edge deployment. (15:10)
    * Potential for advanced use cases like background removal, replacement, and object manipulation in video. (15:43)
    i used gemini 1.5 pro to summarize the transcript.

    • @Roboflow
      @Roboflow Місяць тому

      Curious how did you do it

    • @wolpumba4099
      @wolpumba4099 Місяць тому

      @@Roboflow I used the prompt "create bullet list summary: ". Then another prompt "add starting (not stopping) timestamps".

  • @g.s.3389
    @g.s.3389 4 місяці тому +1

    wow

  • @TUSHARGOPALKA-nj7jx
    @TUSHARGOPALKA-nj7jx Місяць тому

    Do we have a yolo-v8 model trained on the ade20k dataset? If not, how would one do it?

  • @froukehermens2176
    @froukehermens2176 4 місяці тому +2

    Can you use YOLO-world + SAM to annotate images for training a (faster) object detector? (or image segmentation - maybe even pose estimation?).

    • @Roboflow
      @Roboflow 4 місяці тому +1

      Yes you can! Some time ago we showed how to do it with Grounding DINO + SAM combo: ua-cam.com/video/oEQYStnF2l8/v-deo.htmlsi=JzsB_leYOXbGtiGL

    • @Amir-vn2wx
      @Amir-vn2wx 4 місяці тому +1

      @@Roboflow This is awesome!

  • @99develop80
    @99develop80 3 місяці тому +1

    Thank you for the video! I have a question. What do you call the technology that uses YOLO-world + Efficient SAM in the back of the video to switch from detection to segmentation along the baseline? Or is there a way to implement it?

    • @Roboflow
      @Roboflow 3 місяці тому

      I use Gradio library to build those interactive demos.

  • @nazaruddinnurcharis598
    @nazaruddinnurcharis598 2 місяці тому

    Good information, whether this Yolo can be used to detect objects in realtime using a camera?, because I am in a project to develop Yolo for use in realtime cameras that I plan to use on my farm to detect predators.

  • @misaeldavidlinareswarthon190
    @misaeldavidlinareswarthon190 4 місяці тому +1

    Impressive !!!! ... I have a quiestion
    So for maximun speed I still have to use Yolov8 or yolo-world have less latency with coustom dataset

    • @Roboflow
      @Roboflow 4 місяці тому

      If you need a model that runs in real-time or faster you still need to train object detector on custom datasets. It does not need to be YOLOv8.

  • @rajeshktym
    @rajeshktym 4 місяці тому +1

    Hi, is it a good suggestion to use YOLO-World for apple grade detection? A global shutter 2MP camera will capture 5 apples in the same position in a single frame (apple cup conveyor with trigger). We need to find bounding box of each apple and the classification result like grade A or grade B. What may be the maximum time required to obtain grade and boundary box information for each apple using jetson Nano.

    • @Roboflow
      @Roboflow 4 місяці тому

      I think you can always spend few minutes to try. Like I said in the video: don’t be afraid to experiment, but be prepared that in your use case you might still need to train model on custom dataset.
      During my tests conveyor object detection usually worked really well. At least if objects do not occlude each other. That’s why I feel quite confident that detection part will work. I’m worried about classification.

  • @potobill
    @potobill Місяць тому

    is there a C++ version? Is the C++ version faster or the same speed?

  • @avamaeva7999
    @avamaeva7999 4 місяці тому +1

    This is a game changer, but it needs to work on mobile to be of real use in my setting? Two questions please:
    1 - Can quantizisation be used on this model to make it much quicker, perhaps to a level where it will work in real time (at least 10fps) on state of the art phones (eg iPhone 15)?
    2 - Can the model be run through the TFLite Converter? If not, any ideas whether that facility might be introduced?
    Many Thanks

    • @Roboflow
      @Roboflow 4 місяці тому +1

      Good questions. As far as I know no quantized version was yet released. I’ll try to reach out to authors and ask.

  • @nourabdou4118
    @nourabdou4118 4 місяці тому +1

    Thank you, very informative. I've a question regarding the prompts, Does it support and understands things like "Red Zones" or "Grey Areas" ?
    I've tried to use it on maps and I was trying to identify grey areas or red areas but it doesn't work. Is there any workaround? thank you again!

    • @Roboflow
      @Roboflow 4 місяці тому +1

      hard to say without looking at the exact image. zone or area sounds very general :/ Is there any chance you could look for a gray rectangle or circle? I'm thinking of something more precise. And I assume you need a very low confidence threshold to do it anyway.

    • @nourabdou4118
      @nourabdou4118 4 місяці тому

      @@Roboflow It works and obviously it's not correct 100% but It works which's good, thank you so much

  • @vipulpardeshi2868
    @vipulpardeshi2868 4 місяці тому +2

    Hey , I just want to know , Is there any method to use Roboflow models on Offline Projects . Because by using API inferencing is very slow and I want fast detections.Is there any way to save the model .pt file and use it later without alsways importing Roboflow workspace. Thanks❤

    • @Roboflow
      @Roboflow 4 місяці тому +1

      Absolutely! You can use inference pip package to run any model from Roboflow on your local machine. You only need internet during the first run to download it. Then it is cached locally and you can run it offline.

    • @vipulpardeshi2868
      @vipulpardeshi2868 4 місяці тому

      Ok thanks for the reply , you guys are the best

  • @TUSHARGOPALKA-nj7jx
    @TUSHARGOPALKA-nj7jx Місяць тому

    Would Yolo-world-m or s version run in ms on a CPU?

  • @alaaalmazroey3226
    @alaaalmazroey3226 4 місяці тому +1

    Hi, Does yolo-world + SAM work well to segment all the cars and trucks perfectly in the video scenes when there is a very crowded in the road? If not what do you suggest? Thsnks

    • @Roboflow
      @Roboflow 4 місяці тому

      If you plan to detect cars, just use any of models pre trained on COCO. You do not need zero shot detection to find cars :)

    • @DDDprinting
      @DDDprinting Місяць тому

      ​@@RoboflowDo you have a recommendation for a camera for this kind of work?

  • @alaaalmazroey3226
    @alaaalmazroey3226 4 місяці тому +1

    Can YOLO-world detect the road area from dash camera accurately? As i need to detected for autonomous vehicle

    • @Roboflow
      @Roboflow 4 місяці тому

      I recommend you try with your own images here: huggingface.co/spaces/stevengrove/YOLO-World

  • @richarddjarbeng7093
    @richarddjarbeng7093 3 місяці тому +1

    Cool tutorial. I have 2 questions.
    1. Is there a list of classes that the model can detect? For instance if I want to detect 'yellow tricycles' but I am not sure if the model knows tricycles where can I check this.
    2. How do you use this for semantic segmentation? You showed this briefly for the suitcases and croissants but you didn't go into the details.

    • @Roboflow
      @Roboflow 3 місяці тому

      There is no list… You need to experiment. But that’s easy. All you need to do is use HF space: huggingface.co/spaces/stevengrove/YOLO-World
      You need to use boxes coming from YOLO-World to prompt SAM. Take a look at the code here. Few months ago we showed how to combo GroundongDINO + SAM combo: ua-cam.com/video/oEQYStnF2l8/v-deo.html

    • @richarddjarbeng7093
      @richarddjarbeng7093 3 місяці тому

      @@Roboflow Will check it out. Thanks for the quick response

  • @baseerfarooqui5897
    @baseerfarooqui5897 3 місяці тому +1

    hi very informatic video i am getting this error while running code "AttributeError: type object 'Detections' has no attribute 'from_inference. i am using on my local system

    • @Roboflow
      @Roboflow 3 місяці тому

      What version of supervision you have installed?

  • @sreekanthreddy6979
    @sreekanthreddy6979 3 місяці тому +1

    how to do this with web camera ?

  • @alaaalmazroey3226
    @alaaalmazroey3226 4 місяці тому +1

    Hi, does YOLO-world can detect object (e.g. houses) perfectly from geospatial images?

    • @Roboflow
      @Roboflow 4 місяці тому

      I tested. I’m afraid not ;/

  • @user-vv8my4lj9i
    @user-vv8my4lj9i 4 місяці тому +1

    Hi, is there a way to count the time of objects in zone

    • @Roboflow
      @Roboflow 4 місяці тому

      Yup. It is on out list of videos that are coming really soon!

  • @iconolk7338
    @iconolk7338 5 днів тому +1

    I want to use this project. It works on the hugging face, but strangely it doesn't fit my environment, it doesn't work on my PC.
    I want to "clone" that on the hugging face, is there a way?

    • @Roboflow
      @Roboflow 5 днів тому

      Yes. HF Spaces work like git. You can clone entire project to your local.

  • @isaac10231
    @isaac10231 4 місяці тому +1

    Can this be run locally on an rtx card? Or at least, how do we run this locally,?

    • @Roboflow
      @Roboflow 4 місяці тому

      Absolutely! I think you can easily run it on RTX.

  • @user-bh3rc8rq3x
    @user-bh3rc8rq3x 4 місяці тому +1

    Is this helpful in detecting the damaged object in real time??

    • @Roboflow
      @Roboflow 4 місяці тому

      Probably depends on type of object and type of damage, but I think yes.

    • @user-bh3rc8rq3x
      @user-bh3rc8rq3x 4 місяці тому +1

      ​@@Roboflow Thank you. Let's consider the example of suitcases and backpacks shown in the video. Can this technology be useful for detecting damage in them?

    • @Roboflow
      @Roboflow 4 місяці тому

      @@user-bh3rc8rq3x I'll try to answare this question during community session

  • @khalidalsinan3768
    @khalidalsinan3768 3 місяці тому +1

    in the huggingface website, when i upload a video, it outputs a video of 2 seconds only. Anyone knows how to fix this?

    • @Roboflow
      @Roboflow 3 місяці тому

      We need to prevent long video processing , because it makes other users wait longer.

    • @Roboflow
      @Roboflow 3 місяці тому

      You would need to clone the space and make it process longer files.

    • @KhalidAlsinan
      @KhalidAlsinan 3 місяці тому

      @@Roboflowhow do I “clone” it?

  • @rafaelsetyan1755
    @rafaelsetyan1755 4 місяці тому +1

    Has anybody tried this model in UAV/Drone data, is it accurate? It might be possible to export onnx and to do inference in C++, isn't it?

    • @Roboflow
      @Roboflow 4 місяці тому +1

      The only test I made on drone footage was "lake detection". But that was a large object; you are probably considering detecting smaller objects.
      As for ONNX export, yes, export is possible, but (as far as I know) once you export your text prompt is frozen.

  • @abdshomad
    @abdshomad 4 місяці тому

    Yesterday I tried to detect red, yellow, green traffic light. It still did not recognize the color. Any specific guide on how to identify color?

    • @atharvpatawar8346
      @atharvpatawar8346 4 місяці тому +1

      If it’s able to detect the individual traffic lights, get the bounding boxes and use clustering to find the majority colour within that box

    • @abdshomad
      @abdshomad 4 місяці тому

      @@atharvpatawar8346 currently it can't. It will detect the whole lights. Even I tried to change the prompt to : circle, box, bulb, still not possible. Maybe have to apply 2nd classifier?

    • @SkalskiP
      @SkalskiP 4 місяці тому +1

      @@abdshomad I'd say iy you need to use YOLO-World and second level classifier it is probably not wort this.

    • @SkalskiP
      @SkalskiP 4 місяці тому +1

      @@abdshomad which version of model did you used?

  • @paulpolizzi3421
    @paulpolizzi3421 4 місяці тому +1

    can this work on my kids soccer videos?

    • @Roboflow
      @Roboflow 4 місяці тому

      It probably can. But soccer is pretty standard use-case. YOLOv8 or other typical detector is probably a much better choice for you.

  • @novandaardhi7867
    @novandaardhi7867 4 місяці тому +1

    can this integrated with ros2 using Nvidia Jetson Nano?

    • @Roboflow
      @Roboflow 4 місяці тому

      We are going to test Jetson deployments internally soon, but I can already tell you that it will be pretty hard to run it on the Nano board. Xavier / Orin sounds a lot more realistic.

    • @novandaardhi7867
      @novandaardhi7867 4 місяці тому

      thanks, maybe I can consider using Orin to run it, I'll wait for you to do a test on Jetson

  • @zdong2483
    @zdong2483 3 місяці тому

    report issue when running note book on Mar 23, 2023, have to use !pip install -q ultralytics==8.1.30, otherwise fail.

    • @Roboflow
      @Roboflow 3 місяці тому

      I’m not sure what you mean, but I just tested the code and everything works.

  • @polnapanda4934
    @polnapanda4934 4 місяці тому

    After couple of hours working on google colab It cuts almost all performance, deletes data and says that i can buy gpu power

    • @Roboflow
      @Roboflow 4 місяці тому

      Sorry to hear that. Google Colab is free, but only up to a certain point :/

    • @polnapanda4934
      @polnapanda4934 4 місяці тому +1

      @@RoboflowYep :c i was training my model and it deleted all progress after 4 hours of training

  • @chandanchakma2875
    @chandanchakma2875 13 днів тому

    i want to learn AI .please make a playlist ..

  • @user-ur3ml4cp2l
    @user-ur3ml4cp2l 4 місяці тому +2

    It is not working well when object size is less, GROUDING DINO Working well than Yolo-World.

    • @Roboflow
      @Roboflow 4 місяці тому +1

      I think it all depends on specific cases. What do you meant by “object size is less”?

    • @user-ur3ml4cp2l
      @user-ur3ml4cp2l 4 місяці тому +1

      @@Roboflow I mean when object is far away in image. Yolo world could not detect as many objects as GROUNDING DINO Could in such situation.

    • @Roboflow
      @Roboflow 4 місяці тому +1

      @@user-ur3ml4cp2l have you tried lower confidence threshold?

    • @user-ur3ml4cp2l
      @user-ur3ml4cp2l 4 місяці тому

      @Roboflow yes tried that too, but still, the performance of GROUNDING DINO was superior. It could detect objects on more images than Yolo-world.

    • @science.20246
      @science.20246 4 місяці тому +1

      groundino is more accurate

  • @netq254
    @netq254 3 місяці тому

    "Cheap Nvidia T4" £1000 is not cheap bro

    • @Roboflow
      @Roboflow 3 місяці тому +2

      Compared to A100 or H100 it is ;) but what I meant is just using T4 on AWS.

    • @netq254
      @netq254 3 місяці тому

      @@Roboflow Holy hell you're right! I didn't realise how expensive these cards are!