Ongoing Thesis

Masterarbeiten

Deep learning based collaborative task prediction and execution for the mobile Yumi robot

Stichworte:
Human-robot collaboration, human-object interaction, Deep Learning, assistive robot

Beschreibung

The participation of assistive robots in current industrial, as well as daily life scenarios, is increasing significantly. To better enable robots to perform the tasks that people are performing, or intend to perform, we want to investigate a deep learning-based vision solution. 

Voraussetzungen

- Deep Learning 

- Python3, Pytorch

- Computer Vision

- C++ and ROS

 

Kontakt

- Yuankai Wu

- Arne-christoph Hildebrandt (ABB(Asea Brown Boveri))

Betreuer:

Yuankai Wu - Arne-christoph Hildebrandt (ABB(Asea Brown Boveri))

Real-time multiple human tracking with camera-radar fusion for safe autonomous transportation

Stichworte:
Deep Learning, data fusion, multi-human tracking

Beschreibung

This work aims to develop low-cost real-time 3D multiple object tracking frameworks based on camera-radar fusion for multiple people tracking. The purpose is to ensure the safety of AGV navigations in industrial autonomous transportation scenarios, most importantly, to avoid collisions with people.

Voraussetzungen

- Basic knowledge of computer vision

- Basic knowledge of deep learning-based tracker model

- Python 3. Pytorch

- C++

Kontakt

Yuankai Wu (yuankai.wu@tum.de)

Sophie Roscher (sophie.roscher@sick.de)

Betreuer:

Yuankai Wu - Sophie Roscher (Sick Ag (Sick))

Semi-automatic Objects Annotation for Human Acitivity Datasets

Stichworte:
Object annotation, semi-automatic, object detection, object tracking, optical flow

Beschreibung

Most of the current human activity datasets focus on end-to-end models, i.e., they directly use RGB images as input and ignore the semantic information of human-object interaction. On the one hand, this model is easier to build, but on the other hand, it is also because of the huge amount of work and expense required to annotate the objects. This work focuses on annotating objects in the human behavior dataset to the extent that it allows our object detection models and human activity prediction models to provide valid semantic information.

Voraussetzungen

  • Very high motivation
  • Computer Vision Fundamentals
  • Programming in python
  • Basic understanding of deep learning

Kontakt

yuankai.wu@tum.de

Betreuer:

Yuankai Wu

Forschungspraxis (Research Internships)

Human-robot interaction using vision-based human-object interaction prediction

Kurzbeschreibung:
Human-object interaction, human-robot interaction

Beschreibung

We use vision solution to locate the target object for the robot and send the desired object back to the operator to complete the whole process of human-robot interaction.

Voraussetzungen

- Panda arm

- Computer vision

- Human-object interaction prediction

-Grasping

Betreuer:

Yuankai Wu

A 3D human-object interaction dataset for assistive robot

Stichworte:
Human-robot interaction, human-object predition, 3D features

Beschreibung

Collect a dataset available for assistive robots in a 3D environment. Meet the visual perception system under human-object interaction detection to serve the robot.

Voraussetzungen

- Basics of Realsense 

- Basics of ROS

- Basics of computer vision

Betreuer:

Yuankai Wu