- 234
- 461 289
Anton Maltsev
Russia
Приєднався 6 вер 2013
consulting - cvml.rembrain.ai/
email - anton@rembrain.ai
email - anton@rembrain.ai
Is the Radxa Rock5C Lite (RK3582) one of the best boards for hobby Computer Vision Right now?
My LinkedIn - www.linkedin.com/in/maltsevanton/
My Telegram channel - t.me/CVML_team
e-mail: anton@rembrain.ai
Twitter - Serious_WK
Video about:
Hailo on different boards - ua-cam.com/video/cUXSZcNgSz8/v-deo.html
RK3576 - ua-cam.com/video/RXDwdOtfog4/v-deo.html + medium.com/@zlodeibaal/how-to-run-hailo-on-arm-boards-d2ad599311fa
RK3588 max speed - ua-cam.com/video/mDRfXNuIMBE/v-deo.html
RK3566 - ua-cam.com/video/e_BtAdS5M70/v-deo.html
RK3568 - ua-cam.com/video/NHVPxPlY2lI/v-deo.html
My Telegram channel - t.me/CVML_team
e-mail: anton@rembrain.ai
Twitter - Serious_WK
Video about:
Hailo on different boards - ua-cam.com/video/cUXSZcNgSz8/v-deo.html
RK3576 - ua-cam.com/video/RXDwdOtfog4/v-deo.html + medium.com/@zlodeibaal/how-to-run-hailo-on-arm-boards-d2ad599311fa
RK3588 max speed - ua-cam.com/video/mDRfXNuIMBE/v-deo.html
RK3566 - ua-cam.com/video/e_BtAdS5M70/v-deo.html
RK3568 - ua-cam.com/video/NHVPxPlY2lI/v-deo.html
Переглядів: 310
Відео
Choosing a 2d camera for Computer Vision product: The Сomprehensive Guide
Переглядів 452День тому
Article - medium.com/@zlodeibaal/how-to-choose-a-camera-for-a-computer-vision-project-1da6490bb3dd Pavel - www.linkedin.com/in/pavelryabenko/ Yuriy - www.linkedin.com/in/yuriy-bukhtiyarov-74790b20/ Andrey - www.linkedin.com/in/andreyragimov/ 00:00:00 - Intro. 00:02:18 - General approach. Physics, technical questions 00:05:45 - A first table for comparison 00:06:14 - Camera connection type: USB,...
Using big pre-trained models for prototyping
Переглядів 994Місяць тому
00:00:00 - Intro. 00:01:07 - Big VLM: ChatGPT, Qwen2 VL, LLAMA/LLAVA, NVLM, Molmo, etc. 00:03:38 - Grounding Dino, Grounding SAM, Florence 2, etc. 00:06:25 - SAM, SAM2 00:07:51 - CLIP, DINOv2, etc. 00:10:38 - From prototype to production. My LinkedIn - www.linkedin.com/in/maltsevanton/ My Telegram channel - t.me/CVML_team e-mail: anton@rembrain.ai Twitter - Serious_WK
A short sample: Classic stereo Depth vs. Neural stereo Depth vs. Monocular depth.
Переглядів 505Місяць тому
00:00:00 - Intro. 00:01:52 - The first example 00:02:36 - The second example Stereo on the Edge - ua-cam.com/video/mEs_8vxxSqI/v-deo.html Stereo vs mono - ua-cam.com/video/UQBIKsUQZOc/v-deo.html My LinkedIn - www.linkedin.com/in/maltsevanton/ My Telegram channel - t.me/CVML_team e-mail: anton@rembrain.ai Twitter - Serious_WK
How fast is Hailo-8L with boards other than RPi5?
Переглядів 997Місяць тому
00:00:00 - Intro 00:00:48 - How easy is it to run Hailo on a random board? 00:01:07 - How is the speed related to PCLe lines and protocol? 00:01:25 - Is the inference speed related to the CPU? 00:01:48 - Our players: RPi5, NanoPC-t6 (RK3588), RockPi 3A, Khadas VIM3, OrangePi Ai Pro 00:03:40 - A few assumptions of the test 00:05:33 - Speed comparison: batch size 1 00:08:47 - Speed comparison: ba...
Depth estimation. From the theory to the Edge.
Переглядів 642Місяць тому
My big video about the comparison of different neural networks - ua-cam.com/video/JmZdSGtJHNw/v-deo.html My big article about different approaches of stereo reconstruction for 3D cameras - medium.com/@zlodeibaal/3d-cameras-in-2022-choosing-a-camera-for-cv-project-6eb6fcc67948 00:00:00 - Intro 00:02:34 - How stereo cameras work (principle, main formula, etc.) 00:07:44 - Disparity estimation of l...
Orbbec Gemini 335L. Let's check how it's working.
Переглядів 402Місяць тому
00:00:00 - Intro 00:01:52 - What are the direct competitors of 335L? 00:11:02 - Installing on Linux 00:12:15 - Complex objects overview (dark, curved, shiny, complex light, windows) 00:14:40 - What about reconstruction with neural networks? If you are interested in this camera, you can check in here - store.orbbec.com/collections/gemini-330-stereo-depth-cameras?Anton (Yes, the link is with stat...
Depth Pro - monocular network from Apple. But can it do this?!
Переглядів 1,5 тис.2 місяці тому
00:00:00 - Depth Pro: Sharp Monocular Metric Depth in Less Than a Second 00:02:44 - How stable is it? 00:06:00 - Comparison with OAK-D. 00:07:51 - Summary. My LinkedIn - www.linkedin.com/in/maltsevanton/ My Telegram channel - t.me/CVML_team e-mail: anton@rembrain.ai Twitter - Serious_WK
Choosing AI Edge board in 2024 / 2025
Переглядів 1,2 тис.2 місяці тому
Main article - medium.com/@zlodeibaal/cookbook-for-edge-ai-boards-2024-2025-b9d7dcad73d6 00:00:00 - Intro. 00:02:25 - Main criteria for comparison: product, engineering, and scientific 00:14:33 - Nvidia Jetsons: Orin, Xavier, NX, AGX, Nano, etc. 00:19:03 - x86 CPUs for AI: Intel, AMD. OpenVino, etc. 00:23:43 - AI for CPUs from ARM and RISK-V 00:25:14 - RockChips NPUs: RK3588, RK3576, RK3399, RK...
Ultralytics Yolo (Yolov11). Do you need it?
Переглядів 2,8 тис.2 місяці тому
00:00:00 - Intro. 00:00:50 - The only comparison we have 00:01:53 - License. AGPL-3. 00:02:39 - OpenVino, TensorRT. But what about Qualcomm, RockChip, etc? 00:03:44 - Nvidia T4. The sad standard for comparison 00:05:25 - Code Base 00:06:50 - Augmentation, Loss function, New model. Is it worth it? 00:07:20 - Augmentation - a tricky improvement 00:10:15 - Loss function? 00:11:45 - Do I need to us...
Имеет ли смысл идти в Computer Vision и ML в 2024?
Переглядів 7 тис.3 місяці тому
Прошлое видео на тему - ua-cam.com/video/A3HdPnw8wN4/v-deo.html My LinkedIn - www.linkedin.com/in/maltsevanton/ My Telegram channel - t.me/CVML_team e-mail: anton@rembrain.ai Twitter - Serious_WK
Does this even work?! - Radxa NIO 12L (MediaTek Genio 1200)
Переглядів 8943 місяці тому
00:00:00 - Intro. 00:01:00 - The main problem with Radxa NIO 12L (Only TFlite) 00:03:56 - Documentation from Radxa and from MediaTek 00:04:40 - TFlite - CPU delegate, GPU delegate, ArmNN delegate. NPU?... 00:05:00 - Running NPU inference 00:06:31 - Yolov8, Yolov5, YoloX, etc. 00:07:28 - Two available commands: ncc-tflite, neuronrt (export and inference) 00:10:21 - What is "should work, but not ...
Computer Vision and AI for NXP (Debix model A)
Переглядів 4243 місяці тому
Computer Vision and AI for NXP (Debix model A)
LLMs for RockChip. Guide for RKLLM. RK3588 vs RK3576 comparision
Переглядів 2,4 тис.4 місяці тому
LLMs for RockChip. Guide for RKLLM. RK3588 vs RK3576 comparision
Data or Models? A lot of AI researchers struggle with this!
Переглядів 4864 місяці тому
Data or Models? A lot of AI researchers struggle with this!
Everything about OpenCV, OpenCV.ai in 2024
Переглядів 1,6 тис.4 місяці тому
Everything about OpenCV, OpenCV.ai in 2024
Segment Anything 2 (SAM 2): how to start + limitations
Переглядів 5 тис.4 місяці тому
Segment Anything 2 (SAM 2): how to start limitations
Albumentations with Vladimir: everything about augmentations
Переглядів 8155 місяців тому
Albumentations with Vladimir: everything about augmentations
The five main reasons your Computer Vision system will not work
Переглядів 1,2 тис.5 місяців тому
The five main reasons your Computer Vision system will not work
OrangePi AiPro. A comprehensive guide: setting up, model export, and overview (Huawei Ascend 310 B4)
Переглядів 2 тис.5 місяців тому
OrangePi AiPro. A comprehensive guide: setting up, model export, and overview (Huawei Ascend 310 B4)
Do you need a new ML model in production?
Переглядів 2685 місяців тому
Do you need a new ML model in production?
Data Science. Чем отличается новичок от эксперта.
Переглядів 3,8 тис.5 місяців тому
Data Science. Чем отличается новичок от эксперта.
MASA tracker. Is it hype or a pretty nice tracker?
Переглядів 1 тис.6 місяців тому
MASA tracker. Is it hype or a pretty nice tracker?
Unlocking Hailo: Expert Insights and Practical Tricks (works on Raspberry Pi!)
Переглядів 3,5 тис.7 місяців тому
Unlocking Hailo: Expert Insights and Practical Tricks (works on Raspberry Pi!)
Computer Vision on NPU - all you need to know
Переглядів 2,9 тис.7 місяців тому
Computer Vision on NPU - all you need to know
Milk-V DUO. Is it good for computer vision? How to start neural network on it?
Переглядів 2,3 тис.7 місяців тому
Milk-V DUO. Is it good for computer vision? How to start neural network on it?
Computer Vision Inference on Embedded World 2024
Переглядів 1 тис.8 місяців тому
Computer Vision Inference on Embedded World 2024
STM32MP2 - A short interview about technical details on Embedded World.
Переглядів 5098 місяців тому
STM32MP2 - A short interview about technical details on Embedded World.
how can rk3582(4 tops) is faster than rk3588(6 tops)?
awesome)
КРАСАВА
Great video! At 2:57 what model were you running and how were you running it? Was it the "flusk api" you mention later? Why do you think 76 was faster?
We can use Polarize filters and IR Projects with the camera to avoid reflections and glares. Please give your insights.
Like Zed X Cameras with Polarize Lens
Hi, Sami! Thanks for the feedback! 1) Yes, if you are creating your own camera, it's super important to choose: a) Optical schema polarization, filters, lens coating, etc.) b) Light schema With both of them, you can improve quality a lot. And there are a lot of tricks here. 2) In general, polarized lenses may decrease image quality. For example, if you need to look at a monitor, sky, water, etc., 90% of the time, they increase quality and 10% decrease it.
Great Comparison
my clients would like to run visual recognization on this chip for 1080p images, expecting 2 seconds a picture, would this be realistic? I tried yolov5m on my raspberry pi 5 which is quite a bit faster than the rk35xx and it took it average 12 seconds to do that..
I think so because RK3588 NPU is much faster than RPi. But of course, it requires a little bit of testing and looking deeper into models/inference pipelines, etc.
Nice overview about the things which to take into account to choose a right camera. And there were many details I didnt know before. Thanks!
try Ultra version, and use USB camera, builtin one is garbage has miss-soldred MPIS port, found that the hard way, That said its a €50 ai enabled SBC, with PoE capablities, its not bad for the price range, but would go with the RK3999 Chipset if I got budget
Nicely done and gives good overview of camera selection issue. In my experience, camera driver compatibility for RockChip SBCs was the most painful.
Hi can you help to use yolo11 on rp5 by with the HAILO M.2 ML kit for a webcam and ip-cam? Any codes available that you know? Thx!
Hi Anton! Thank you for this video. You said that in the end of the day it doesn't matter witch stereocam is used. Could this conclusion be also applied to cameras with structured light? Stereo vision used in realsense d or orbecc gemini series uses active light to project IR pattern. However there is no prior knowledge in the pattern, it just helps to detect feature points. Could you please review or provide your opinion about orbbec astra cameras? I also checked your old video about 3d reconstruction and there realsense d and orbec astra were presented just as cameras using the active light. Maybe there is no significant difference between binocular stereo vision and structured light when speaking about mentioned devices? Thank you for sharing your expertise, much appreciated!
Привет, Айдар! With modern neural networks, this structured light is not something critical. It can increase cuality in some specific conditions (very big uniform planes, low light conditions, etc.). This camera has structured light. But for 90% conditions, I don't think it's critical.
U kinda sound like viktor from arcane 😁😁
Ничего непонятно, но очень интересно. Загорелся думал, что можно будет вставить в Nvme М2 SSD И на винде запускать stable diffusion. А тут опять командная строка наверное... И картинки будет рисовать в командной строке... Когда уже доработают эти технологии. Вроде бы востребованы искусственный интеллект и где он? Начинаешь что-то искать кругом только для игр. Видеокарты. А вот такие вот интегрированные модули для нейросетей днём с огнём не сыщешь.
I would appriciate an interface outside of the demo without that watermark and bigger resolution.
You can use code from here - github.com/facebookresearch/sam2 Here is my COLAB demo, but I am not sure it still will run - colab.research.google.com/drive/1Un09HITLLM-ljkG1Ehn9cJjdwk8FVI_1
Годно
Thank you.
CRE stereo could be used for metric distance estimation or what could best distance measurement system for self driving car is streo is accurate enough does it provide distance range
Another interesting board worth mentioning is the AMD Zynq 7000 series, which works with the open source Pynq project, which comes with Jupyter Notebook and tons of examples. Though it is not an NPU per se, it does accelerate a lot of ML utility functions and could be a great pre-processor in a low power ML pipeline.
From my experience, sometimes could be very useful to build prototype in unity and prepare some virtual test environment before building robot
Yolo is legendary. Just hope this can run fast on CPU
Наивысшую услугу родине окажет только ЛИБЕРАЛ. А человечесво спасут только специалисты по машинному зрению.
файнали, ин шорт термс, коллект йоур овн датасетс, трейн зе сота моделс бай юселф, энд хев а фан, экчуалли. Мени сенкс.
What is the model of the top section video?
It's original output from OrbBec scanner. I expect it should be a little bit modified StereoBM algorithm
@AntonMaltsev did u used monocular webcam or stereo one? I found it interested to try 😄
Please, could you add the links to the models? or papers?
For UniMatch I used this version - github.com/fateshelled/unimatch_onnx (original one is here - github.com/autonomousvision/unimatch ) For Depth Anything I used this one - huggingface.co/spaces/depth-anything/Depth-Anything-V2
I have experienced a lot of stereo, ToF Cameras and also with Deep Stereo algorithms, I would say that, None of these two can surpass accuracy of Neural Stereo Depth, RAFT Stereo, High Frequency Stereo Matching Network, Stereoformer etc are the whole on another level when comparing with Classical and Mono.
Yes, totally agree. I thinkl that for some super specific cases, the plain math can be a little bit more accurate. But for 99% of real tasks, it doesn't matter.
ua-cam.com/video/24iwwksDfE4/v-deo.html - Second sample ua-cam.com/video/YDzV87feewA/v-deo.html - First sample
Why do people always say their names at the beginning of the video, who cares about the name, what does it matter, what value does it add to the video?
any video updates regarding of this?
Thank you so much for testing our camera and making this awesome video! Your review means a lot to us and really helps others see what the product can do in real-life scenarios. We truly appreciate your support! !!
"Thank you again for the video! I just wanted to mention that the product name in the video title might have a small typo - our product name is actually 'Orbbec Gemini 335L.' Just a friendly note in case you'd like to update it!".
Thank you!
Где тогда получить этот опыт начальный для вката в машинной обучение?
Кто-то говорит что ии всех скоро, ии же очень выгоден для бизнеса но вакансий в сфере машинного обучения очень мало. Почему так? Должен быть скачок
Fantastic info! Will probably look towards another solution for edge inference now
Here is the link to the article on how to run everything - medium.com/@zlodeibaal/how-to-run-hailo-on-arm-boards-d2ad599311fa
What would you use for the desk scene for pick and place?
Nice question! It depends on a lot of questions: What is the gripper? If it's some vacuum suction cup, all cameras are OK for objects with plain surfaces. If it's a human-like arm, I may train my network on raw stereo data. For example we did it like this here - ua-cam.com/video/Js6_eA6c1Nw/v-deo.html What is my budget? With an unlimited budget - I would prefer Mech Mind or some Zivid cameras. It's much easier for a complex objects - ua-cam.com/video/kv0U6F7kEhg/v-deo.html Can you install a 3D camera on your hand? D405 can be nice at a close distance. Also, I would think if I have enough computational power to run stereo reconstruction myself - ua-cam.com/video/mEs_8vxxSqI/v-deo.html Sometimes it can give the best quality
I'm a researcher from Belgium at university. Your channel is a true gold mine. Keep up the brilliant work.
Thanks for the great experiment, I was just planning on buying the new Pi-AI-Hat but I'll have to change my mind.
It will probably boost your speed a bit, but not for all models and probably not as much as you expect. I saw some reviews that usual boost is 30-40%, and specifically some slow models it could be almost x2.
Thanks for the nice video. Is it possible to connect two cameras to the RK3588 and have depth estimation? I mean, does it require image synchronization?
Hi! 1) It is better to have a stereo camera with onboard synchronization. Without it, you can't work with moving objects. If they are static, then it should be OK. 2) It's hard to run some workable network on RK3588. It was pretty big part of this video:)
Nice video, as always! I would be happy to see also something about mono cameras here, which would be used in local robot setups, for example with robot arms using ROS or used in other areas like small rooms. There eare so much camaras ou there that it's hard to find good ones.
Thank you! It's a nice idea, but it's pretty hard to do. There are a lot of them, and it's super hard to do a good comparison. I have a plan, but definitely not in the next few months.
друг, почитай про пленоптические камеры и вообще про камеры светового потока И анубисом тебя заклинаю, поправь произношение
Норм, только надо шрифт покрупнее чтобы был, и большая часть по времени таких видео должна быть занята схемами, примерами, это всё-таки для технарей. Норм ❤
Отличный ролик и канал, спасибо, Антон!
Спасибо, очень интересно! Подписываюсь :) И поддерживаю запрос на видео по компетенциям.
Занимаюсь задачей синтеза речи, кое- каким ресерчем и агентами. Но, разумеется, не в РФ..
У меня валяется одна подобная камера OAK-D. На мой взгляд это скорее игрушка. Максимум для пылесоса годится. Расстояние настолько незначительное, что в ней нет никакого смысла.
would love to see tests of computer vision speed between the two chips!
Сними обзор на русском так как к железке есть интерес но сложно понимать что то на английском
Thank you for review of this product. Do you know the result of CPU score of this board like UNIXBENCH? I'm interested in CPU performance of this board.
Here is the table with a few results for the stress-ng test that I ran - docs.google.com/spreadsheets/d/1BMj8WImysOSuiT-6O3g15gqHnYF-pUGUhi8VmhhAat4/edit?gid=0#gid=0
hmm how well does it work with sequences?
There is just a single frame input.
Anton, Very informative video! what kind of board do you suggest I am looking for a security camera based detection system in realtime .It has to be battery powered to detect friend or foe .I am looking at 1000-1500 quantity for production also ! Thank you
Hi! Depending on what kind of neural networks you have, how much battery life you need, etc.
@@AntonMaltsev We want to use YOLO v8 or above and battery life of approx 12 hours I am looking at
@@PavanSBV-y3u I recommend you consult an AI researcher who can choose the algorithm based on your dataset and algorithm. There is almost no difference between different YOLO networks, and YOLOv8 is not the best network from an export point. Also, there are many questions about inference speed, latency, etc. But if your input is correct, RockChip (3566,3568), some NXP boards, MediaTek, Sophon and Qualcomm are okay. But next, you need to go through my guide and consider the board based on your location, industry, type of development, etc.