Interactive Eye Projection with Webcam or Kinect - TouchDesigner Tutorial 009
Вставка
- Опубліковано 12 тра 2024
- In this TouchDesigner tutorial, you learn how to create an interactive projection, where an eye follows people passing by. I use a webcam and the bodytrackCHOP to track persons, but you can also use a Kinect or other data like mouse cursor position, noise or any other object.
With free AI image generation tools like Bing Image Creator, you can create your own texture, I used PromeAI: www.promeai.pro/?vsource=i_8f...
On my Patreon you find animated textures and the project files of all my tutorials: / reflekkt
Instagram / reflekkt_
0:00 Intro & project overview
1:36 Setup bullet solver environment & actor
2:36 Basic render setup
3:00 BulletsolverCHOP & feedback data
3:45 Setup tracking data
4:46 Look at tracked object
5:17 Calculate velocity
6:30 Add random rotation & movements
7:41 Integrate Webcam / Kinect for tracking
8:58 Automatic switching when persons enter installation
9:50 Outro
please upload tutorial again 😊 thanks
yo it looks so cool
Hello! I'm on a mac, and BodyTrack COMP is unfortunately not compatible with a MAC! What would you recommend for subbing in the bodytracking aspect with a webcam on a MAC, so I can get my program to work like you have it here? Thank for the wonderful tutorial!
I second this
Except I’m not on a Mac
But the “body track” still doesn’t work
How can I get it to work with the Kinect without that chop?
@@TheClumsyAcrobat let me know if you find a solution pls!
Hello! How could i track two or more person? And how to add new sphere/eye for new person in the scene?
it is amazing 👏👏👏. i would like to know how to use it in fullscreen
You can use a windowCOMP and drag and drop the last null in it, then open as performance window 👍
Hello, it's excellent, but I have the jump when the person leaves the frame and returns, the eye makes a sudden movement, is there a way to soften the switch? thanks
Put the lag chop
There’s a problem that the eyes would run out of the screen. It just flow out. It seems like its not fully following my neck. Btw Thanks for your tutorial!
You need to limit the position data to the boundaries of your screen. Find out the boundaries by using the mouseinCHOP, move your mouse all the way to the left, note the x value, then all the way to the right, note the x value, all the way down, note the y value, and all the way up, note the y value. Now do the same with your camera or Kinect input, noting the position of your neck for example at the boundaries of the screen. With a mathCHOP just before the "positionNULL", use the Range function to convert your data.
Lets say, the neck data all the way on the left is 0.2 and all the way to the right 1, and the screen position (which you checked with mouseinCHOP) is left 0 and right 1.5, then change the range from 0.2 and 1 to 0 and 1.5. That way, the position gets mapped properly on the output. If you move your camera or use a different output screen, you have to readjust these values. So for using it in an installation, make sure the camera won't move after setting it up correctly.
@@reflekkt_netI tried this and so far have not been successful at keeping the eye on screen
How do I know what values to input for the range?
It's a good and important question. Check out my reply under ztime8038's comment, that should get you in the right direction.
I think I'll also create a little video on that topic soon.
ouuuu...this is so hot
Can I use laptop web cam?
Yes you can 👍
For some reason my eye doesnt move. It follows my mouse but it doesn't move from its spot.
Did you initialize and start the simulation? To do that, go to the bsolverCOMP that you created in the beginning (navigate one level up by pressing U or scroll up with mousewheel) and click on Initialize Sim and Collision Shapes, and then Start Sim.
@@reflekkt_net Thankl you so much, lol silly me