Abstract
Sensors have been developed and applied in a wide range of fields such as robotics and autonomous vehicle navigation (AVN). Due to the inability of a single sensor to fully sense its surroundings, multiple sensors based on individual specialties are commonly used in order to complement the shortcomings and enrich perception. However, it is challenging to integrate the heterogeneous types of sensory information and produce useful results. This research aims to achieve a high degree of accuracy with a minimum false-positive and false-negative rate for the sake of reliability and safety. This paper introduces a likelihood-based data fusion model, which integrates information from various sensors, maps it into the integrated data space and generates the solution considering all the information from the sensors. Two distinct sensors: an optical camera and a LIght Detection And Range (Lidar) sensor were used for the experiment. The experimental results showed the usefulness of the proposed model in comparison with single sensor outcomes.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Khaleghi, B., Khamis, A., Karray, F.O., Razavi, S.N.: Multisensor data fusion: a review of the state-of-the-art. Inf. Fusion (2011)
How, J.P., Bethke, B., Frank, A., Dale, D.,Vian, J.: Real-time indoor autonomous vehicle test environment. IEEE Control Syst. Mag. 51–64 (2008)
Subramaniana, V., Burksa, T.F., Arroyob, A.A.: Development of machine vision and laser radar based autonomous vehicle guidance systems for citrus grove navigation. Comput. Electron. Agric. 53(2), 130–143 (2006)
Sun, J., Wu, Z., Pan, G.: Context-aware smart vehicle: from model to prototype. J. Zhejiang Univ. Sci. A 10(7), 1049–1059 (2009)
Hillel, A.B., Lerner, R., Levi, D., Raz, G.: Recent progress in road and lane detection: a survey. Mach. Vis. Appl. 1–19 (2012)
Huang, A.S., Moore, D., Antone, M., Olson, E., Teller, S.: Finding multiple lanes in urban road networks with vision and lidar. Auton. Robots 26(2–3), 103–122 (2009)
Cheng, Y.: Mean shift, mode seeking, and clustering. IEEE Trans. Pattern Anal. Mach. Intell. 17(8), 790–799 (1995)
Kodagoda, K., Ge, S.S., Wijesoma, W.S., Balasuriya, A.P.: IMMPDAF approach for road-boundary tracking. IEEE Trans. Veh. Technol. 56(2), 478–486 (2007)
Garcia, F., Musleh, B., de la Escalera, A., Armingol, J.M.: Fusion procedure for pedestrian detection based on laser scanner and computer vision. In: 2011 14th International IEEE Conference on Intelligent Transportation Systems (ITSC), pp. 1325–1330 (2011)
Morrison, R.: Fiducial marker detection and pose estimation from lidar range data. Technical Report. DTIC Document (2010)
Adhikari, S.P., Kim, H.: Dynamic programming and curve fitting based road boundary detection. In: 9th WSEAS International Conference on Computational Intelligence, pp. 236–240 (2010)
Jo, J.H., Tsunoda, Y., Sullivan, T., Lennon, M., Jo, T., Chun, Y.: BINS: blackboard-based intelligent navigation system for multiple sensory data integration. In: The 17th International Conference on Image Processing, Computer Vision, and Pattern Recognition, Las Vegas, USA (2013)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer International Publishing Switzerland
About this paper
Cite this paper
Jo, J., Tsunoda, Y., Stantic, B., Liew, A.WC. (2017). A Likelihood-Based Data Fusion Model for the Integration of Multiple Sensor Data: A Case Study with Vision and Lidar Sensors. In: Kim, JH., Karray, F., Jo, J., Sincak, P., Myung, H. (eds) Robot Intelligence Technology and Applications 4. Advances in Intelligent Systems and Computing, vol 447. Springer, Cham. https://doi.org/10.1007/978-3-319-31293-4_39
Download citation
DOI: https://doi.org/10.1007/978-3-319-31293-4_39
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-31291-0
Online ISBN: 978-3-319-31293-4
eBook Packages: EngineeringEngineering (R0)