Boosting Self-localization with Graph Convolutional Neural Networks

Takeda Koji, Tanaka Kanji

Abstract

Scene graph representation has recently merited attention for being flexible and descriptive where visual robot self-localization is concerned. In a typical self-localization application, the objects, object features and object relationships of the environment map are projected as nodes, node features and edges, respectively, on to the scene graph and subsequently mapped to a query scene graph using a graph matching engine. However, the computational, storage, and communication overhead costs of such a system are directly proportional to the number of feature dimensionalities of the graph nodes, often significant in large-scale applications. In this study, we demonstrate the feasibility of a graph convolutional neural network (GCN) to train and predict alongside a graph matching engine. However, visual features do not often translate well into graph features in modern graph convolution models, thereby affecting their performance. Therefore, we developed a novel knowledge transfer framework that introduces an arbitrary self-localization model as the teacher to train the GCN-based self-localization system i.e., the student. The framework, additionally, facilitated lightweight storage and communication by formulating the compact output signals from the teacher model as training data. Results on the Oxford RobotCar datasets reveal that the proposed method outperforms existing comparative methods and teacher self-localization systems.

Download


Paper Citation


in Harvard Style

Koji T. and Kanji T. (2021). Boosting Self-localization with Graph Convolutional Neural Networks.In Proceedings of the 16th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications - Volume 4: VISAPP, ISBN 978-989-758-488-6, pages 861-868. DOI: 10.5220/0010212908610868


in Bibtex Style

@conference{visapp21,
author={Takeda Koji and Tanaka Kanji},
title={Boosting Self-localization with Graph Convolutional Neural Networks},
booktitle={Proceedings of the 16th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications - Volume 4: VISAPP,},
year={2021},
pages={861-868},
publisher={SciTePress},
organization={INSTICC},
doi={10.5220/0010212908610868},
isbn={978-989-758-488-6},
}


in EndNote Style

TY - CONF

JO - Proceedings of the 16th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications - Volume 4: VISAPP,
TI - Boosting Self-localization with Graph Convolutional Neural Networks
SN - 978-989-758-488-6
AU - Koji T.
AU - Kanji T.
PY - 2021
SP - 861
EP - 868
DO - 10.5220/0010212908610868