why you are so underrated man? I found your tutorial to be one of the best in fact the best one for me over here. Please keep up the good work. I was soo lost in the project Idea that I had in mind and following your this project, I have learned a lot and have been working on my project idea after. THANK YOU
The world need more people like you. Love your practical approach to problems. Only suggestion is a external microphone. The voice will be great will a little more richness.
Dear Mr. Taha, I thank you from my heart to create such a great content. Your open source well-documented code and detailed walkthrough video of the implementation made me learn a lot of practical stuffs. I am working on a project that uses ConvLSTM to detect emotion from frontal face video and I found your video + code extremely helpful for my project, as I can carry forward the things I have learned from your video. The world needs more people like you! Please keep on making these wonderful videos and open source code. You don't know how much you are helping people in the field of Deep learning and Computer Vision! ❤
Dear Efran, I came across your comment while reviewing the discussion thread. I am currently working on a similar project, specifically developing a Video Facial Emotion Recognition model using ConvLSTM. I was wondering if you could share your findings with me. Thank you! Best regards, Sebastian
@@BleedAIAcademy I didn't get email for code after clicking on given link and varifying the email-id as mentioned in description. Can you help me please.
That is just the video display error caused because the video size is bigger than the output limit set in the colab notebook, you do not have to solve it to run the application, you can just save/download the output video in your system and play it in any video player you like.
Excellent explanation of human activity recognition using deep learning, one of the best videos I've seen on the subject! You had shown pose estimation in the lead up to ConvLSTM and LRCM but did not use it. in your two examples Can the frame sequence of pose vectors (comprised of skeletal joint locations) be used with an LSTM to improve the accuracy even further. Although, there are some recent papers on using Graph Convolutional Networks for human activity recognition. How does the accuracy of GCN-based human activity recognition compare with LRCN?
Thank you Vijay for the complements, For the first part of your question, yes, you can try sending the skeletal points to further improvement, although I"m not sure if the increase in computation will be worth the accuracy gains, some people opt for optical flow and then feed the motion vectors to the model to capture motion information. For the second part, I've not done such a comparison so can't say about that.
Very well explained video, I have searched this topic a lot, you really did it. Thank you very much. I have a question, Can we apply this to Real-Time Scenarios such as detecting and classifying plant diseases?
Having unexpected trouble from ValueError: Failed to convert a NumPy array to a Tensor (Unsupported object type generator) from my video dataset, do you have any suffestion?
What kind of adjustement can be made to optimize the following to having multiple individuals in a scene and each getting detected and classified by the action they are doing.
A basic technique is to run person detection on the video, and then inside each person crop do action recognition. Optimize this by doing action recognition only after set intervals, since no one will be changing their actions every next frame, also track people instead of detecting on each frame. There are dedicated models for this specific problem in paperswithcode, this problem Is called: Spatio Temporal Detection, here's a nice dataset for it: arxiv.org/abs/1705.08421
@@BleedAIAcademy Thank you for the reference I will make sure to look it up. Will you be covering something similar anytime soon? For example action recognition in Sports.
@@alimchogle3876 apparently there are still ufc50 datasets from kaggle u can find. and I believe u can use them. I downloaded just to see how the structure is, so that I can properly organize my data directory
super explanation. sir want to detect the play going on clips such passing the ball and running with ball on the field with no. of players are playing can i use this approach.
hello sir thanks for your time to explore the action recognition in videos,but i have one query if any action not existing in dataset but am testing that activity its not give proper resut how to solve this problem sir
HI , Dear Taha ,is there a possibility to apply concept of single action detection and generalize it to detect more than one action ? Example video contains ( swing , running , sliding , jumping ) can we apply the same concept to your code and detect these movements in same video ?
Yeah, so there are dedicated approaches for that, but a hack to repurpose this approach for what you're trying to achieve would be something like this: Use people detection to get bounding boxes of all the people, then adjust the boxes for input size you need for the model, pass in the all the boxes to the model to get predictions. Now obviously this would be slow but to make it a bit fast, use an object tracker to detect activity every nth second or frame.
I am trying to use this code on a smaller dataset but it doesn’t provide high accuracy at the end and there is overfitting. What shall I do to prevent this?
hi i have tried downloading the video using youtube-dl but it shows an error in the uploader id as youtube has removed youtube-dl is there any alternate way to download the youtube video and saving it in mp4 format??? anyone?
Hi, can you please explain the benefit of using a deep learning model and training it to detect the set of actions, when you could have easily used heuristics to classify them?
Really well explained video, I have search alot for this topic you really nailed it. Thank you so much. I have a question , Can we implement this on a Real-Time Scenario like detecting a specific or multiple actions that we wanted to detect?
Thank you for your comments and Yes, Please do read the associated blogpost and I've written another one where I do this for a video too. But if you're looking to monitor CCTV footage and actions in real time of multiple people then this can quickly get complex, because when you're adding a temporal domain to the model the latency will take a hit. So do check pros and cons of different approaches out there to figure what's best for your particular use case.
@@bytebardicbro forget about the function that download video from UA-cam. Just chose the video you want to test, download it manually from UA-cam or somewhere else and put it in the correct path (i think was folder Test Video). Then you just run the next session of the code without running the downloading part. The program will take our video and create another one with the results
@@Elti1998 can you help me bro? i just saw your solution i upladed video in folder Test Video, but the error says 'there is not defined video title how did you fix it?
You'll need to setup such a system, I have a tutorial where I show how to read a live IP camera stream in OpenCV but yes it's possible to connect the two systems but some optimizations are needed.
Thank you very much for your efforts, it is really a great work, but I have a question, how can I get the code? I waited a lot on the email, I did not receive any response
We used pafy library to download the videos from youtube, you can download those clips from youtube. Just search for the target keyword, you can further use the filter tools to fine-tune search results.
Hi sir appreciate if you could share your insights on my project. I am having 4 scenarios/action for recognition, and would like to develop a video analytics system (real-time) to ensure all 4 actions are executed with correct order. Will this CNN+LSTM method be suitable for this task? Thanks in advanced.
This dataset given is not running, I tried extracting it and also tried entering the folder path but it is not running. Can anyone help to run this dataset?
hi. very well explained. i wanted to know if i can use this approach for a sign language project..? i have the video dataset of 50 classes.. i want to use mediapipe holistic so can u pls tell if i can use this approach alongwith it?
Excellent explanation I appreciate Your Efforts and Time. I have question if i ran the model Local on my device Can I use WebCam to the Test Model as when i try i found error called Invalid Input dimension as Expected Only three Dimensions not 4 where i am passing The Sequence Number So any help
@@A_Y_Elarabawy Hi, you'll need to do segmentation prediction, for e.g predict the activity after fixed time intervals over N frames. If the action recognition approach was really fast then I would suggest using a window approach where you're predicting every N frames and the buffer keeps filling up with new frames.
@@BleedAIAcademy thank you very much. And one more thing, can i use your code for my project? If that's okay with you, I don't want to get any copyright infringements.
You can decrease img size, reduce step frames from 20 to a lower number, get rid of few layers or reduce filters.. but all these things will also reduce performance, two strategies for making this work for real time scenarios is 1: Use better CNN algorithm designs, e.g. mobilenetv2 variants etc or use a complete different architecture that doesn't use LSTM. and 2) design a smart pipeline that reduces unnecessary prediction calls during production. For an in-depth talk about this, you can call me here: pensight.com/x/tahaatbleedai You can also check this blogpost that I wrote, this approach works in real-time, but it's as efficient as above: learnopencv.com/introduction-to-video-classification-and-human-activity-recognition/
@@sohaibshaikh5534 Just noticed that link was wrong, so you can download from here for now: bleedai.com/human-activity-recognition-using-tensorflow-cnn-lstm/ Click the download code button on top.
it took me 4 hours to understand taking video frames and storing it in list .. after that i tried by myself. it worked.. so just try to get concept thereotically first and then try to do it by urself.. just dont look code to copy. take pen and paper and write step by step what u need to do.. for eg. if u want to extract frames then write like this.. : to get into main dir i need to set path variable. then if i have 10 class then i need to loop through it. to get into first class . take its index. then list all dir..then select video , then extract.. like this .. to understand it properly .. just take one class and only 1 video first and outline what to do.btw. thanks sir for the turorial
Dear Taha, thank you for giving such a great explanation. However, i have a question. is it possible for me to use the combined algorithm to predict (lets say) a growth of a tree by feeding it images from one week to another
@@BleedAIAcademy i’m trying to train on kth dataset however the keenel is crashing while model evaluation and i reduced my data set after it model evaluation was working fine but on the video classification part the kernel crashed again? I’m running this on local machine jupyter RAM:- 12 GB ,Nvidia 1650 4 GB , AMD RYZEN 5 4600H Any possible solution????
Hi Thank you for the amazing content. I wanted to know how the second model you described, the LRCN approach, compares to state of the art slow fast models. Are they even comparable? Is it better to use the LRCN instead of slowfast for small dataset for recognizing a single action?
Complete and patient training. 👍 Thank you very much for this very useful content.😊 I just verified the email, but I could not receive the codes. Can you please guide me?🙏
hello,the Download Code is Real-Time 3D Hands Landmarks Detection & Hands Classification,not Human Activity Recognition using CNN + LSTM maybe something went wrong
@@BleedAIAcademy the link you have provided to download the code is wrong it is for Real time 3D Hands landmarks detection & classification not for the HAR
I found error at Construct the required LRCN model here: LRCN_model = create_LRCN_model() it shows NotImplementedError:Cannot convert a symbolic Tensor(lstm_5/strided_slice:0) to a numpy array.This error may indicate that you're trying to pass a Tensor to a Numpy call, which is not supported.
Can I use this code for two category classification for forged and pristine on a custom dataset ?Here the data is having the forgery by removing an object usually a person classify as forged and where the person is not removed labeled as pristine. Will it work for this project by creating a bounding around the and comparing the frames with its previous one?
Hi Sir, OSError: MoviePy error: the file test_videos/Comparison of Four Styles of Tai Chi-Output-SeqLen20.mp4 could not be found! how to solve this error!!
Hi, the link to the paper for using ConvLSTM layer is in the notebook, please download the code or go over the blogpost to get the exact link. Also the paper explains the layer, the architecture of the model itself in the tutorial is custom.
You are a great teacher. I selected this video very accidentally. Everything was clearly explained. Thank you very much.
why you are so underrated man? I found your tutorial to be one of the best in fact the best one for me over here. Please keep up the good work. I was soo lost in the project Idea that I had in mind and following your this project, I have learned a lot and have been working on my project idea after. THANK YOU
The world need more people like you. Love your practical approach to problems. Only suggestion is a external microphone. The voice will be great will a little more richness.
Thank you brother, this is the most comprehensive video on this subject out here!
Thank you
Dear Mr. Taha, I thank you from my heart to create such a great content. Your open source well-documented code and detailed walkthrough video of the implementation made me learn a lot of practical stuffs. I am working on a project that uses ConvLSTM to detect emotion from frontal face video and I found your video + code extremely helpful for my project, as I can carry forward the things I have learned from your video. The world needs more people like you! Please keep on making these wonderful videos and open source code. You don't know how much you are helping people in the field of Deep learning and Computer Vision! ❤
But I didn't get email for code after clicking on given link and varifying the email-id as mentioned in description. Can you help me pls
Thank you Erfan, I'm humbled by your praise 💖
Dear Efran,
I came across your comment while reviewing the discussion thread. I am currently working on a similar project, specifically developing a Video Facial Emotion Recognition model using ConvLSTM. I was wondering if you could share your findings with me.
Thank you!
Best regards, Sebastian
@@sebastianjitaru7163 Same here as well using ConvLSTM for an ASL sign language recognition system
@@BleedAIAcademy I didn't get email for code after clicking on given link and varifying the email-id as mentioned in description. Can you help me please.
The link site does not send the source code
do u still not got the link?
@@tanvish4293 I also did not get the code , please can you help me
Your introduction is so good and informative already
Very well illustrated. Thanks for putting this together.
Thank you
Hello the blogpost link is not working properly. Can you please check the link? It says the page not found
Thanks for pointing this out. Please check the link now.
@@BleedAIAcademy Thanks a lot
Can this model be applied to the ucf-crime datset
Really helped sir. There are very few videos on video processing
This video should be famous!
Thank you Kunal :)
@@BleedAIAcademy Hello sir! Do you have any video in which I can use AGG16 with LSTM? For video classification?
@@kunaltulsidasani6766 Nope
@@kunaltulsidasani6766 Not exactly
@@BleedAIAcademy what all architecture can I use which will take 4 inputs as shown here
Taha u have done an amazing job..
Iam not able to download the test video from the code u provided guved me regexnit found error
Thanks a lot Mr. Taha for this wonderful tuto and explaination
very well-explained video. Your code is so easy to understand. It's really helpful for my assignment
This explained video is very well, but I can’t understand how to solve the ValueError : max_duration at the end of the video ?
That is just the video display error caused because the video size is bigger than the output limit set in the colab notebook, you do not have to solve it to run the application, you can just save/download the output video in your system and play it in any video player you like.
Great work man! keep sharing with us
Thank you Bukhari Jee
amazing video , very detailed and organized .Can you use the same model to detect real -time using camera instead of uploading videos from youtube ?
Very well explained. More power to you.
Thank you
u r a gem. Loved the content.
how can i get the source code? I am not able to receive the email, after i put the email address. plz help
Excellent explanation of human activity recognition using deep learning, one of the best videos I've seen on the subject! You had shown pose estimation in the lead up to ConvLSTM and LRCM but did not use it. in your two examples Can the frame sequence of pose vectors (comprised of skeletal joint locations) be used with an LSTM to improve the accuracy even further. Although, there are some recent papers on using Graph Convolutional Networks for human activity recognition. How does the accuracy of GCN-based human activity recognition compare with LRCN?
Thank you Vijay for the complements, For the first part of your question, yes, you can try sending the skeletal points to further improvement, although I"m not sure if the increase in computation will be worth the accuracy gains, some people opt for optical flow and then feed the motion vectors to the model to capture motion information. For the second part, I've not done such a comparison so can't say about that.
Very well explained video, I have searched this topic a lot, you really did it. Thank you very much. I have a question, Can we apply this to Real-Time Scenarios such as detecting and classifying plant diseases?
Thank you, and You don't need a video classification solution for this sort of task.
Having unexpected trouble from ValueError: Failed to convert a NumPy array to a Tensor (Unsupported object type generator) from my video dataset, do you have any suffestion?
Hi james, as mentioned in the error message, the data type you are using is not supported.
This video is really cool!, I could understand ConvLSTM2D very clearly, I subscribed your channel :)
What kind of adjustement can be made to optimize the following to having multiple individuals in a scene and each getting detected and classified by the action they are doing.
A basic technique is to run person detection on the video, and then inside each person crop do action recognition. Optimize this by doing action recognition only after set intervals, since no one will be changing their actions every next frame, also track people instead of detecting on each frame.
There are dedicated models for this specific problem in paperswithcode, this problem Is called: Spatio Temporal Detection, here's a nice dataset for it: arxiv.org/abs/1705.08421
@@BleedAIAcademy Thank you for the reference I will make sure to look it up. Will you be covering something similar anytime soon? For example action recognition in Sports.
@@rolandsalloum8797 No plans at the moment.
Sadly the dataset is no longer available. Hope you can check this again, because I think this video may be of help still. (I will be watching this!)
Yes the dataset is no longer available and I am facing problem for that
@@alimchogle3876 apparently there are still ufc50 datasets from kaggle u can find. and I believe u can use them. I downloaded just to see how the structure is, so that I can properly organize my data directory
This video helps me a lot! Thank you so so much....
hello can we use same model using WEB CAMERA to identify actions ? Thanks
Thank you so much! This will most certainly help my case study.
You're welcome
Thank you
You're Welcome :)
super explanation. sir want to detect the play going on clips such passing the ball and running with ball on the field with no. of players are playing can i use this approach.
is it possible to do in live footage say for example live football match
Thanks man really helped for my work
when is the site going to start working again
Thank you sir for the beautiful explanation. Sir, can you help me in choosing the right approach for anomaly detection. Kindly make a video on it too
Thanks for sharing such a quality content. ANd one one thing I want to ask is "how to convert these saved model to onnx or tensorrt".
Hi, I would recommend that you go to the onnx GitHub repo they have many tutorials on conversion from one package format to another
Really really excellent video! Helped me in my final year project to understand how exactly to train and setup model parameters 💯👍🏻
hey bro what were the datasets that you have used?
Can this be used in Lane departure warning system?
Can you help with, human action prediction problems. The model will predict human action before it happens.
thank you very much sir, can we use this approach for fall detection ? please help me
very good job 👍
Thanks.
Can i do this with realtime video
hello sir thanks for your time to explore the action recognition in videos,but i have one query if any action not existing in dataset but am testing that activity its not give proper resut how to solve this problem sir
I cant download the you tube vedios from the website
HI , Dear Taha ,is there a possibility to apply concept of single action detection and generalize it to detect more than one action ?
Example video contains ( swing , running , sliding , jumping ) can we apply the same concept to your code and detect these movements in same video ?
Yeah, so there are dedicated approaches for that, but a hack to repurpose this approach for what you're trying to achieve would be something like this:
Use people detection to get bounding boxes of all the people, then adjust the boxes for input size you need for the model, pass in the all the boxes to the model to get predictions.
Now obviously this would be slow but to make it a bit fast, use an object tracker to detect activity every nth second or frame.
Sir how to take input from a live camera or a video we provide?
Excellent video, thank you!
This is really helpful, but how can i use a dataset from my device not from any online source?
Thank you.
Can you draw the confusion matrix for the 4 classifications. Can anyone help with the code
if i want to do it with cars and not people, what would you recomend guys, behavior of cars while driving
very helpful,Thank you so much!!!
Thanks a lot for your video and great explanation in such an easy and understandable way
it helped me a lot. appreciating your efforts,
keep this up
Thank you
I am trying to use this code on a smaller dataset but it doesn’t provide high accuracy at the end and there is overfitting. What shall I do to prevent this?
I didn't get the email for the code after clicking on the given link and varifying the email-id as mentioned in the description.
hi i have tried downloading the video using youtube-dl but it shows an error in the uploader id as youtube has removed youtube-dl is there any alternate way to download the youtube video and saving it in mp4 format??? anyone?
Did you find solution for it
Please help with this similar problem. Suggest any other library
hello can i use this on real time cctv video? for detecting Recognition of ppls on real time?
Hi, can you please explain the benefit of using a deep learning model and training it to detect the set of actions, when you could have easily used heuristics to classify them?
Heuristics for this kind of problem is not a good solution, it will not generalize to all variations of the action.
How we can evaluate the model on the videos downloaded from youtype ?
Really well explained video, I have search alot for this topic you really nailed it. Thank you so much. I have a question , Can we implement this on a Real-Time Scenario like detecting a specific or multiple actions that we wanted to detect?
Thank you for your comments and Yes, Please do read the associated blogpost and I've written another one where I do this for a video too. But if you're looking to monitor CCTV footage and actions in real time of multiple people then this can quickly get complex, because when you're adding a temporal domain to the model the latency will take a hit. So do check pros and cons of different approaches out there to figure what's best for your particular use case.
Dear sir can we use the data as violence detection key Framing
the link is not send source code, can you give us the notebook? thank you sir
Hi, thank you for this amazing video!!
Is it possible to train this model with other datasets?
did you get the correct output?
Actually, I'm getting error in downloading the video with link
@@bytebardicsame for me. Just upload a random video in the correct path and you are done!
@@Elti1998 bro i am getting error like " Unable to extract uploader id;"
could u help to resolve this error?
@@bytebardicbro forget about the function that download video from UA-cam. Just chose the video you want to test, download it manually from UA-cam or somewhere else and put it in the correct path (i think was folder Test Video). Then you just run the next session of the code without running the downloading part. The program will take our video and create another one with the results
@@Elti1998 can you help me bro? i just saw your solution i upladed video in folder Test Video, but the error says 'there is not defined video title how did you fix it?
Hi, by mentioning email address in the given Link doesnot send link of source code. Please send working URL to access code. Thanks
how do i connect the database please help?
While prediction can we pass the link for cctv camera and do live activity recognition
You'll need to setup such a system, I have a tutorial where I show how to read a live IP camera stream in OpenCV but yes it's possible to connect the two systems but some optimizations are needed.
hello, i am doing it in a jupyter note book in that i am getting video_path error can you explain y it is giving this error?
Thank you very much for your efforts, it is really a great work, but I have a question, how can I get the code? I waited a lot on the email, I did not receive any response
where can i get more testing videos like the one used in this code
We used pafy library to download the videos from youtube, you can download those clips from youtube. Just search for the target keyword, you can further use the filter tools to fine-tune search results.
Hi sir appreciate if you could share your insights on my project.
I am having 4 scenarios/action for recognition, and would like to develop a video analytics system (real-time) to ensure all 4 actions are executed with correct order. Will this CNN+LSTM method be suitable for this task? Thanks in advanced.
This dataset given is not running, I tried extracting it and also tried entering the folder path but it is not running. Can anyone help to run this dataset?
hi. very well explained. i wanted to know if i can use this approach for a sign language project..? i have the video dataset of 50 classes.. i want to use mediapipe holistic so can u pls tell if i can use this approach alongwith it?
I too need to know if i can implement it for hand gestures
Excellent explanation I appreciate Your Efforts and Time. I have question if i ran the model Local on my device Can I use WebCam to the Test Model as when i try i found error called Invalid Input dimension as Expected Only three Dimensions not 4 where i am passing The Sequence Number So any help
I mean Can I apply this Hybrid Model In real time to predict the Action while Running not just Saving It in another video after termination
@@A_Y_Elarabawy Hi, you'll need to do segmentation prediction, for e.g predict the activity after fixed time intervals over N frames. If the action recognition approach was really fast then I would suggest using a window approach where you're predicting every N frames and the buffer keeps filling up with new frames.
I'm not able to get the code? I tried three or four times.
Code Link: drive.google.com/drive/folders/1zt3Kt_KogaewsPBIbUPONQCScQWyBVKF?usp=sharing
@@BleedAIAcademy thank you very much. And one more thing, can i use your code for my project? If that's okay with you, I don't want to get any copyright infringements.
Hello, if i want to use this model in Real-time what should i do ?
You can decrease img size, reduce step frames from 20 to a lower number, get rid of few layers or reduce filters.. but all these things will also reduce performance, two strategies for making this work for real time scenarios is 1: Use better CNN algorithm designs, e.g. mobilenetv2 variants etc or use a complete different architecture that doesn't use LSTM. and 2) design a smart pipeline that reduces unnecessary prediction calls during production. For an in-depth talk about this, you can call me here: pensight.com/x/tahaatbleedai
You can also check this blogpost that I wrote, this approach works in real-time, but it's as efficient as above: learnopencv.com/introduction-to-video-classification-and-human-activity-recognition/
@@BleedAIAcademy Can i have the code or link of the collab or Github, I've registered but still in a waitinglist '
@@sohaibshaikh5534 If you're talking about the new upcoming course, then it's not launched yet, at the moment you can only join the waiting list.
@@BleedAIAcademy How can i have the code of the above HAR project you have developed?
@@sohaibshaikh5534 Just noticed that link was wrong, so you can download from here for now: bleedai.com/human-activity-recognition-using-tensorflow-cnn-lstm/ Click the download code button on top.
this will work for general video classification right ? like classifier such as "sports", "news" ,"nature" or not.
Yes it will
@@BleedAIAcademy how can i get my own dataset output video , instead of youtube videos....please help me
it took me 4 hours to understand taking video frames and storing it in list .. after that i tried by myself. it worked.. so just try to get concept thereotically first and then try to do it by urself.. just dont look code to copy. take pen and paper and write step by step what u need to do.. for eg. if u want to extract frames then write like this.. : to get into main dir i need to set path variable. then if i have 10 class then i need to loop through it. to get into first class . take its index. then list all dir..then select video , then extract.. like this .. to understand it properly .. just take one class and only 1 video first and outline what to do.btw. thanks sir for the turorial
Can anyone download the code?
drive.google.com/drive/folders/1zt3Kt_KogaewsPBIbUPONQCScQWyBVKF?usp=drive_link
@@BleedAIAcademy 👌
how to get code?
it is not getting downloaded
Dear Taha, thank you for giving such a great explanation. However, i have a question. is it possible for me to use the combined algorithm to predict (lets say) a growth of a tree by feeding it images from one week to another
dear i did not get code into my mail
Code Drive Link: drive.google.com/drive/folders/1zt3Kt_KogaewsPBIbUPONQCScQWyBVKF?usp=drive_link
@@BleedAIAcademy Thank you sir
Which algorithm is used here for CNN and LSTM
Hi you do not continue the channel?
Hey, you talked about the reasearch paper in your video can you please give me the link of that research paper
Hi, that paper is linked in the blogpost and the jupyter notebook for this tutorial.
@@BleedAIAcademy i’m trying to train on kth dataset however the keenel is crashing while model evaluation and i reduced my data set after it model evaluation was working fine but on the video classification part the kernel crashed again?
I’m running this on local machine jupyter RAM:- 12 GB ,Nvidia 1650 4 GB , AMD RYZEN 5 4600H
Any possible solution????
permision to learn sir. how i can downlaod your dataset?
thanks before
After saving model the download method and predict method are not compiling and running. Plz help me do this really want to complete this
Hi, what error are you facing, have you tried running the colab notebook without any modifications?
Hi Thank you for the amazing content. I wanted to know how the second model you described, the LRCN approach, compares to state of the art slow fast models. Are they even comparable? Is it better to use the LRCN instead of slowfast for small dataset for recognizing a single action?
Even I have the same doubt
Complete and patient training. 👍
Thank you very much for this very useful content.😊
I just verified the email, but I could not receive the codes. Can you please guide me?🙏
Please share your email address
Thanks
Cannot download the code.
After filling the email, The link is broken
Eran
hello sir, how should i get the code ??
hello,the Download Code is Real-Time 3D Hands Landmarks Detection & Hands Classification,not Human Activity Recognition using CNN + LSTM
maybe something went wrong
It's working fine. What problem you're facing?
@@BleedAIAcademy the file I download from the above URL is your other project, not suitable for what this video explains
@@BleedAIAcademy the file I download from the above URL is "Real Time 3D Hands Landmarks Detection & Classification"
@@BleedAIAcademy the link you have provided to download the code is wrong it is for Real time 3D Hands landmarks detection & classification not for the HAR
@@sohaibshaikh5534 Oh, just checked by clicking the link, you're right.
I found error at Construct the required LRCN model here: LRCN_model = create_LRCN_model() it shows NotImplementedError:Cannot convert a symbolic Tensor(lstm_5/strided_slice:0) to a numpy array.This error may indicate that you're trying to pass a Tensor to a Numpy call, which is not supported.
Can I use this code for two category classification for forged and pristine on a custom dataset ?Here the data is having the forgery by removing an object usually a person classify as forged and where the person is not removed labeled as pristine. Will it work for this project by creating a bounding around the and comparing the frames with its previous one?
Hi there is no connection here, both are completely separate problems.
Hi Sir,
OSError: MoviePy error: the file test_videos/Comparison of Four Styles of Tai Chi-Output-SeqLen20.mp4 could not be found!
how to solve this error!!
Hi, make sure this exact file is present in the directory specified, try renaming the file and double check the path.
Can u tell me which paper did u follow to implement ConvLSTM model?
Hi, the link to the paper for using ConvLSTM layer is in the notebook, please download the code or go over the blogpost to get the exact link. Also the paper explains the layer, the architecture of the model itself in the tutorial is custom.
@@BleedAIAcademy can u tell me what learning rate are you considering?
Why after applying first convolution , out put is 64 ? It should be 62*62*16 right
Hello, nice work! Should the number of frames in each video be the same or we can work on videos with different sequence lengths ?
Yes, you can do that, the script will automatically extract a predefined number of frames from the whole video.
@@BleedAIAcademy hi can u please help me in accessing the dataset
hi can u please help me ,please
I am unable to download the code by link (not receiving the email). Please help me.
Hi, how can we implement these models for real-time detection, thanks