Abstract
Nowadays, the utilization of multiple sensors on every robotic platform is a reality due to their low cost, small size and light weight. Multi Sensor Fusion (MSF) algorithms are required to take advance of all the given measurements in a robust and complete manner. These high demanding developed algorithms need to be tested under real and challenging situations and environments. To the knowledge of the authors, none of the available datasets fulfills are fully suitable to be used for MSF applied to Robotics, due to the lack of multiple sensor measurements provided by light weight, small size and low cost sensors; due to their restricted motions (typically planar movements); or to their limited environmental conditions.
The contributions of the paper are twofold. First, a low-cost portable and versatile testbed has been developed for MSF research with various types of sensors. Second, a group of datasets for MSF research for Robotics have been made public as a common framework for algorithm testing after a comparison with the existing databases in the state of the art, highlighting the differences and advantages of the one presented in this paper, that are: low-cost sensors for general use on Robotics, fully 3 dimensional movements (six degrees of freedom), as well as challenging indoor and outdoor small and large environments.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Abeywardena, D., Dissanayake, G.: Tightly-coupled model aided visual-inertial fusion for quadrotor micro air vehicles. In: Field and Service Robotics, pp. 153–166. Springer (2015)
Abeywardena, D., Wang, Z., Kodagoda, S., Dissanayake, G.: Visual-inertial fusion for quadrotor micro air vehicles with improved scale observability. In: 2013 IEEE International Conference on Robotics and Automation, pp. 3148–3153. IEEE (2013)
Bartling, S., Friesike, S., Bartling, S., Friesike, S.: Opening Science. Springer (2014)
Blanco, J.-L., Moreno, F.-A., Gonzalez, J.: A collection of outdoor robotic datasets with centimeter-accuracy ground truth. Autonomous Robots 27(4), 327–351 (2009)
Blanco, J.-L., Moreno, F.-A., González-Jiménez, J.: The málaga urban dataset: High-rate stereo and lidars in a realistic urban scenario. International Journal of Robotics Research 33(2), 207–214 (2014)
Ciftcioglu, O., Sariyildiz, I.S.: Data sensor fusion for autonomous robotics (2012)
Fang, Z., Yang, S., Jain, S., Dubey, G., Maeta, S., Roth, S., Scherer, S., Zhang, Y., Nuske, S.: Robust autonomous flight in constrained and visually degraded environments
Garrido-Jurado, S., Muñoz-Salinas, R., Madrid-Cuevas, F.J., MarÃn-Jiménez, M.J.: Automatic generation and detection of highly reliable fiducial markers under occlusion. Pattern Recognition 47(6), 2280–2292 (2014)
Geiger, A., Lenz, P., Stiller, C., Urtasun, R.: Vision meets robotics: The kitti dataset. International Journal of Robotics Research (2013)
Guerrier, S.: Improving accuracy with multiple sensors: study of redundant MEMS-IMU/GPS configurations. In: Proceedings of the 22nd International Technical Meeting of The Satellite Division of the Institute of Navigation, pp. 3114–3121 (2009)
Guivant, J., Nebot, E.: Simultaneous localization and map building: test case for outdoor applications. In: IEEE Int. Conference on Robotics and Automation (2002)
Huber, D., Badino, H., Kanade, T.: The CMU visual localization data set
Klein, G.S.W., Drummond, T.W.: Tightly integrated sensor fusion for robust visual tracking. Image and Vision Computing 22(10), 769–776 (2004)
Kohli, P., Silberman, N., Hoiem, D., Fergus, R.: Indoor segmentation and support inference from RGBD images. In: ECCV (2012)
Silberman, N., Fergus, R.: Indoor scene segmentation using a structured light sensor. In: Proceedings of the International Conference on Computer Vision - Workshop on 3D Representation and Recognition (2011)
Smith, M., Baldwin, I., Churchill, W., Paul, R., Newman, P.: The new college vision and laser data set. The International Journal of Robotics Research 28(5), 595–599 (2009)
Sturm, J., Engelhard, N., Endres, F., Burgard, W., Cremers, D.: A benchmark for the evaluation of RGB-D SLAM systems. In: Proc. of the International Conference on Intelligent Robot Systems, October 2012
Waegli, A., Guerrier, S., Skaloud, J.: Redundant MEMS-IMU integrated with GPS for performance assessment in sports. In: 2008 IEEE/ION Position, Location and Navigation Symposium, pp. 1260–1268. IEEE (2008)
Weiss, S., Achtelik, M.W., Chli, M., Siegwart, R.: Versatile distributed pose estimation and sensor self-calibration for an autonomous MAV. In: 2012 IEEE International Conference on Robotics and Automation, pp. 31–38. IEEE (2012)
Weiss, S., Siegwart, R.: Real-time metric state estimation for modular vision-inertial systems. In: 2011 IEEE International Conference on Robotics and Automation, pp. 4531–4537. IEEE (2011)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing Switzerland
About this paper
Cite this paper
Sanchez-Lopez, J.L., Fu, C., Campoy, P. (2016). FuSeOn: A Low-Cost Portable Multi Sensor Fusion Research Testbed for Robotics. In: Reis, L., Moreira, A., Lima, P., Montano, L., Muñoz-Martinez, V. (eds) Robot 2015: Second Iberian Robotics Conference. Advances in Intelligent Systems and Computing, vol 417. Springer, Cham. https://doi.org/10.1007/978-3-319-27146-0_5
Download citation
DOI: https://doi.org/10.1007/978-3-319-27146-0_5
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-27145-3
Online ISBN: 978-3-319-27146-0
eBook Packages: Computer ScienceComputer Science (R0)