Hamidreza Kasaei
Hamidreza Kasaei
  • 31
  • 4 196
VITAL: Visual Teleoperation to Enhance Robot Learning through Human-in-the-Loop Corrections
Imitation Learning (IL) has emerged as a powerful approach in robotics, allowing robots to acquire new skills by mimicking human actions. Despite its potential, the data collection process for IL remains a significant challenge due to the logistical difficulties and high costs associated with obtaining high-quality demonstrations. To address these issues, we propose a low-cost visual teleoperation system for bimanual manipulation tasks, called VITAL. Our approach leverages affordable hardware and visual processing techniques to collect demonstrations, which are then augmented to create extensive training datasets for imitation learning. By utilizing both real and simulated environments, along with human-in-the-loop corrections, we enhance the generalizability and robustness of the learned policies. We evaluated our method through several rounds of experiments in both simulated and real-robot settings, focusing on tasks of varying complexity, including bottle collecting, stacking objects, and hammering. Our experimental results validate the effectiveness of our approach in learning robust robot policies from simulated data, significantly improved by human-in-the-loop corrections and real-world data integration. Additionally, we demonstrate the framework's capability to generalize to new tasks, such as setting a drink tray, showcasing its adaptability and potential for handling a wide range of real-world bimanual manipulation tasks.
Переглядів: 123

Відео

Simulated Teacher - progress visualization
Переглядів 182 місяці тому
The main idea is to emulate the interactions of a robot with the surrounding environment over a long period in a single context scenario (office, kitchen, etc.), where a robot would be expected to learn continually and adaptively. In such a human-in-the-loop open-ended learning scenario, the interaction between the human user and the robot unfolds through a sequence of actions: teach, ask, and ...
Ensemble of Deep Features for Continual Few-shot Object Recognition
Переглядів 164 місяці тому
Ensemble of Deep Features for Continual Few-shot Object Recognition
Harnessing the Synergy between Pushing, Grasping, and Throwing for Object Manipulation in Cluttered
Переглядів 1696 місяців тому
Harnessing the Synergy between Pushing, Grasping, and Throwing to Enhance Object Manipulation in Cluttered Scenarios. (ICRA 2024) In this work, we delve into the intricate synergy among non-prehensile actions like pushing, and prehensile actions such as grasping and throwing, within the domain of robotic manipulation. We introduce an innovative approach to learning these synergies by leveraging...
GraspCaps: A Capsule Network Approach for Familiar 6DoF Object Grasping
Переглядів 659 місяців тому
As robots become more widely available outside industrial settings, the need for reliable object grasping and manipulation is increasing. In such environments, robots must be able to grasp and manipulate novel objects in various situations. This paper presents GraspCaps, a novel architecture based on Capsule Networks for generating per-point 6D grasp configurations for familiar objects. GraspCa...
Active perception for manipulating a toy car from side to side
Переглядів 6110 місяців тому
Active perception for manipulating a toy car from side to side
Early or Late Fusion Matters: Efficient RGBD Fusion in Vision Transformers for 3D Object Recognition
Переглядів 106Рік тому
Early or Late Fusion Matters: Efficient RGB-D Fusion in Vision Transformers for 3D Object Recognition, by Georgios Tziafas, Hamidreza Kasaei read more about this work: arxiv.org/pdf/2210.00843.pdf
Enhancing Fine-Grained 3D Object Recognition using Hybrid Multi-Modal Vision Transformer-CNN Models
Переглядів 29Рік тому
find more about this work: Paper: arxiv.org/pdf/2210.04613.pdf Authors: Songsong Xiong, Georgios Tziafas, Hamidreza Kasaei
Simultaneous Multi-View Object Recognition and Grasping in Open-Ended Domains
Переглядів 196Рік тому
Paper: arxiv.org/abs/2106.01866 IRL-Lab: www.ai.rug.nl/irl-lab/ To aid humans in everyday tasks, robots need to know which objects exist in the scene, where they are, and how to grasp and manipulate them in different situations. Therefore, object recognition and grasping are two key functionalities for autonomous robots. Most state-of-the-art approaches treat object recognition and grasping as ...
MVGrasp: Real-Time Multi-View 3D Object Grasping in HighlyCluttered Environments
Переглядів 352Рік тому
In this paper, we propose a multi-view deep learning approach to handle robust object grasping in human-centric domains. In particular, our approach takes a point cloud of an arbitrary object as an input, and then, generates orthographic views of the given object. The obtained views are finally used to estimate pixel-wise grasp synthesis for each object. We train the model end-to-end using a sy...
Throwing Objects into A Moving Basket While Avoiding Obstacles
Переглядів 541Рік тому
The capabilities of a robot will be increased significantly by exploiting throwing behavior. In particular, throwing will enable robots to rapidly place the object into the target basket, located outside its feasible kinematic space, without traveling to the desired location. In previous approaches, the robot often learned a parameterized throwing kernel through analytical approaches, imitation...
MVGrasp: Real-Time Multi-View 3D Object Grasping in Highly Cluttered Environments
Переглядів 2722 роки тому
Nowadays robots play an increasingly important role in our daily life. In human-centered environments, robots often encounter piles of objects, packed items, or isolated objects. Therefore, a robot must be able to grasp and manipulate different objects in various situations to help humans with daily tasks. In this paper, we propose a multi-view deep learning approach to handle robust object gra...
Lifelong Ensemble Learning based on Multiple Representations for Few-Shot Object Recognition
Переглядів 832 роки тому
Hamidreza Kasaei*, Songsong Xiong* (*equal contributions) Department of Artificial Intelligence, University of Groningen, The Netherlands
Real-Time Multi-View 3D Object Grasping in Highly Cluttered Environments
Переглядів 632 роки тому
Real-Time Multi-View 3D Object Grasping in Highly Cluttered Environments
Dual arm manipulation
Переглядів 3712 роки тому
Most of the service robots are painstakingly coded and trained in advance to perform tasks correctly. The knowledge of such robots is fixed after the training phase, and any changes in the environment require complicated re-programming by expert users. Such limitations prevent the successful deployment of service robots in human-centered environments. Lifelong learning and continuous improvemen...
Simultaneous Multi-View Object Recognition and Grasping in Open-Ended Domains
Переглядів 892 роки тому
Simultaneous Multi-View Object Recognition and Grasping in Open-Ended Domains
Generating bounding box, local reference frame, and three orthographic views for two sample objects
Переглядів 342 роки тому
Generating bounding box, local reference frame, and three orthographic views for two sample objects
Simultaneous Multi-View Object Recognition and Grasping in Open-Ended Domains
Переглядів 1093 роки тому
Simultaneous Multi-View Object Recognition and Grasping in Open-Ended Domains
Fine-Grained 3D Object Categorization by Combining Shape & Texture Features in Multiple Colorspaces
Переглядів 463 роки тому
Fine-Grained 3D Object Categorization by Combining Shape & Texture Features in Multiple Colorspaces
Importance of Shape Features, Color Constancy, Similarity Measures in Open-Ended Object Recognition
Переглядів 223 роки тому
Importance of Shape Features, Color Constancy, Similarity Measures in Open-Ended Object Recognition
Investigating the Importance of Color and Similarity Measures in Open-Ended 3D Object Recognition
Переглядів 1224 роки тому
Investigating the Importance of Color and Similarity Measures in Open-Ended 3D Object Recognition
Clear table scenario using UR5e robot
Переглядів 1944 роки тому
Clear table scenario using UR5e robot
system demonstration : serve a drink scenario
Переглядів 905 років тому
system demonstration : serve a drink scenario
OrthographicNet -- system demonstration using Imperial College Domestic Environment dataset
Переглядів 455 років тому
OrthographicNet system demonstration using Imperial College Domestic Environment dataset
OrthographicNet - a real-time system demonstration (serve_a_drink scenario)
Переглядів 745 років тому
OrthographicNet - a real-time system demonstration (serve_a_drink scenario)
A short summary of my experiences in robotics
Переглядів 1656 років тому
A short summary of my experiences in robotics
Kinesthetic Teaching
Переглядів 1586 років тому
Kinesthetic Teaching
Real Robot Demonstration : Clear Table Task
Переглядів 1096 років тому
Real Robot Demonstration : Clear Table Task
System Demonstration using Washington RGB-D Scene Dataset
Переглядів 2176 років тому
System Demonstration using Washington RGB-D Scene Dataset
Evaluation of Object Affordance Detection
Переглядів 1736 років тому
Evaluation of Object Affordance Detection