Skip to content

Latest commit

 

History

History
117 lines (72 loc) · 6.08 KB

README.md

File metadata and controls

117 lines (72 loc) · 6.08 KB

Sensorimotor Cross-Behavior Knowledge Transfer for Grounded Category Recognition

Abstract:

Humans use exploratory behaviors coupled with multi-modal perception to learn about the objects around them. Research in robotics has shown that robots too can use such behaviors (e.g., grasping, pushing, shaking) to infer object properties that cannot always be detected using visual input alone. However, such learned representations are specific to each individual robot and cannot be directly transferred to another robot with different actions, sensors, and morphology. To address this challenge, we propose a framework for knowledge transfer across different behaviors and modalities that enables a source robot to transfer knowledge about objects to a target robot that has never interacted with them. The intuition behind our approach is that if two robots interact with a shared set of objects, the produced sensory data can be used to learn a mapping between the two robots' feature spaces. We evaluate the framework on a category recognition task using a dataset containing 9 robot behaviors performed multiple times on a set of 100 objects. The results show that the proposed framework can enable a target robot to perform category recognition on a set of novel objects and categories without the need to physically interact with the objects to learn the categorization model.

Gyan Tatiya, Ramtin Hosseini, Michael C. Hughes and Jivko Sinapov, "Sensorimotor Cross-Behavior Knowledge Transfer for Grounded Category Recognition", IEEE 9th International Conference on Development and Learning and Epigenetic Robotics (ICDL-EpiRob), Oslo, Norway, Aug. 19-22, 2019.

Poster:

drawing

BibTeX:

@inproceedings{tatiya2019sensorimotor,
  title={Sensorimotor Cross-Behavior Knowledge Transfer for Grounded Category Recognition},
  author={Tatiya, Gyan and Hosseini, Ramtin and Hughes, Michael C and Sinapov, Jivko},
  booktitle={2019 Joint IEEE 9th International Conference on Development and Learning and Epigenetic Robotics (ICDL-EpiRob)},
  pages={1--6},
  year={2019},
  organization={IEEE}
}

Development Environment

For our research, we used 64-bit Ubuntu 16.04 based computer with 16 GB RAM, Intel Core i7-7700 CPU (3.20 GHz x 8 cores) and NVIDIA GeForce GTX 1060 (3GB RAM, 1280 CUDA Cores). The neural networks were implemented in widely used deep learning framework TensorFlow 1.12 with GPU support (cuDNN 7, CUDA 9).

Dependencies

Python 3.5.6 is used for development and following packages are required to run the code:

pip install tensorflow-gpu==1.12.0
pip install sklearn==0.20.0
pip install matplotlib==3.0.0
pip install numpy==1.15.3

How to run the code?

Run: python main.py [mapping] [classifier]

mapping: A2A, A2H, H2A, H2H
classifier: KNN, SVM-RBF

Example: python main.py H2H SVM-RBF

Experiment Pipeline

  • The source robot interacts with all the 20 categories (highlighted in solid red line), but the target robot interacts with only 15 categories (highlighted in solid blue line).

drawing

  • The objects of 15 categories shared by both the robots are used to train the encoder-decoder network that learns to project the sensory signal of the source robot to the target robot.

drawing

  • Subsequently, the trained encoder-decoder network is used to generate “reconstructed” sensory signals for the other 5 object categories (highlighted in dashed blue line) that the target robot did not interact with by projecting the sensory signal of the source robot.

drawing

  • Once the features are projected, an category recognition classifier is trained using the projected data from the source context (i.e., how well it would do if it transfered knowledge from the source robot).

drawing

  • Additional category recognition classifier is trained using the ground truth data produced by the target robot (i.e., the best the target robot could do if it had explored all the objects) for comparison.

drawing

Results

Support Vector Machine (SVM-RBF)

Accuracy

drawing

drawing

RMSE loss vs Accuracy

drawing

Accuracy Delta

drawing

drawing

drawing

drawing

RMSE loss vs Accuracy Delta

drawing

k-Nearest Neighbors (3-NN)

Accuracy

drawing

drawing

RMSE loss vs Accuracy

drawing

Accuracy Delta

drawing

drawing

drawing

drawing

RMSE loss vs Accuracy Delta

drawing