[go: up one dir, main page]

Skip to content

Fast object detection learning with CNN and kernel based architecture

Notifications You must be signed in to change notification settings

robotology/online-detection-demo

Repository files navigation

On-line Detection Application

In this repository, we collect the source code of the On-line Detection Application, a pipeline for efficiently training an object detection system on a humanoid robot. This allows to iteratively adapt an object detection model to novel scenarios, by exploiting: (i) a teacher-learner pipeline, (ii) weakly supervised learning techniques to reduce the human labeling effort and (iii) an on-line learning approach for fast model re-training.

The diagram of the final application is as follows:

image

The proposed pipeline allows therefore to train a detection model in few seconds, within two different modalities:

  • Teacher-learner modality: the user shows the object of interest to the robot under different viewposes, handling it in hand (Fig.2 a).
  • Exploration modality: the robot autonomously explores the sorrounding scenario (e.g. a table-top) acquiring images with self supervision or by asking questions to the human in case of doubts (Fig.2 b).

While these modalities can be freely interchanged, a possible use could be to initially train the detection model with the Teacher-learner modality. The goal of this initial interaction is that of “bootstrapping” the system, with an initial object detection model. After that, the robot relies on this initial knowledge to adapt to new settings, by actively exploring the environment and asking for limited human intervention, with the Exploration modality. During this phase, it iteratively builds new training sets by using both (i) high confidence predictions of the current models with a self-supervision strategy and (ii) asking to a human expert to refine/correct low confidence predicitions. In this latter case, the user rovides refined annotations using a graphical interface on a tablet.

The application is composed by several modules, each one accounting for the different components of the pipeline. We tested it on the R1 humanoid robot.

Slide1

Usage

In order to use this application, you first need to clone it in your system and execute the installation instructions and the system setup actions that you can find at this link.

After setting up your system, you can choose among several versions of this application, according to your needs. Here, we will cover the complete version, comprising all types of interaction with the robot, but others are available as xml applications in the folder app/scripts. To execute the complete version you should:

  • Run all the modules of the Weakly Supervised Online Detection application, represented by the app/scripts/WS_onlineDetection.xml file, except the detection_demo.lua module.
  • Run the matlab detection module (i.e., you should run the maltab script Detection_main.m , from the folder modules/modules_matlab)
  • Connect alla available ports from the Weakly Supervised Online Detection application
  • Run the detection_demo.lua and connect all again

Once everything is running, you can interact with the application by one of the following options:

  • Giving commands using the terminal (you can find the list of possible commands at this link)
  • Giving commands using the speech interface. The list of possible commands is reported in the following table.
Command Action
"Have a look at this object_name" The robot replies "Let me have a look at the object_name" and the learning phase begins. There are a few seconds of images acquisition, where the user needs to show and hold the object in his/her hand and show it to the robot. At the end of the acquisition (the actual number of frames can be set and modified by the user) a new detection model is trained
"Forget the object_name" it makes the model delete the classifier and the rls of that particular object.
"Forget all objects" it deletes all the models.
"Explore the table" this command makes the robot start the refinement of the detection model with the exploration of the environment. It also starts the interactive phase, where the robot asks the annotation of doubtful images.
"Stop refinement" this command makes the robot start the refinement of the detection model with the exploration of the environment. It also starts the interactive phase, where the robot asks the annotation of doubtful images.
"Look at the object_name" to accomplish this command, the robot looks at the desired object if it is present in the list of the detected objects.
"Look around" to accomplish this command the robot changes periodically the fixation point, moving the gaze, alternating randomly between the different detected objects in the scene.
"Where is the object_name"? to accomplish this command the robot replies with the list of objects that are close to the mentioned one (within a specified radius).
"What is close to the object_name"? to accomplish this command the robot replies with the name of the closest object to the one requested.

Related publications

This application is the result of the implementation of three works, rispectively reported in the following publications:

About

Fast object detection learning with CNN and kernel based architecture

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published