Abstract

This work proposes a fast deployment pipeline for visually-servoed robots which does not assume anything about either the robot - e.g. sizes, colour or the presence of markers - or the deployment environment. Specifically, we apply a learning based approach to reliably estimate the pose of a robot in the image frame of a 2D camera upon which a visual servoing control system can be deployed. To alleviate the time-consuming process of labelling image data, we propose a weakly supervised pipeline that can produce a vast amount of data in a small amount of time. We evaluate our approach on a dataset of remote camera images captured in various indoor environments demonstrating high tracking performances when integrated into a fully-autonomous pipeline with a simple controller. With this, we then analyse the data requirement of our approach, showing how it is possible to deploy a new robot in a new environment in fewer than 30.00 min.

Paper: https://ieeexplore.ieee.org/abstract/document/10341260

Bibtex

@INPROCEEDINGS{robinson2023,
  author={Robinson, Luke and De Martini, Daniele and Gadd, Matthew and Newman, Paul},
  booktitle={2023 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)}, 
  title={Visual Servoing on Wheels: Robust Robot Orientation Estimation in Remote Viewpoint Control}, 
  year={2023},
  volume={},
  number={},
  pages={6364-6370},
  keywords={Training;Visualization;Pipelines;Wheels;Cameras;Visual servoing;Mobile robots;Cloud Robotics;Visual Servoing;Deep Learning},
  doi={10.1109/IROS55552.2023.10341260}}