|Title||Multi-sensor mapping for low contrast, quasi-dynamic, large objects|
|Publication Type||Journal Article|
|Year of Publication||2020|
|Authors||Shah V., Schild K., Lindeman M., Duncan D., Sutherland D., Cenedese C., Straneo F, Singh H.|
|Type of Article||Article|
|Keywords||localization; mapping; Marine robotics; melt; Robotics; visual navigation|
This paper proposes a systems level solution for addressing the problem of mapping large moving targets with slow but complicated dynamics. Our approach utilizes the complementary nature of multiple sensing modalities. While this work is applicable to other domains we focus our efforts on mapping rotating and translating icebergs. Our solution involves a rigidly coupled combination of a line scan sensor - a subsurface multibeam sonar, with an area scan sensor - an optical camera. This allows the system to exploit the optical camera information to perform iceberg relative navigation which can directly be used by the multibeam sonar to map the iceberg underwater. This paper details the algorithm required to compute the scale of the navigation solution and corrections to find iceberg centric navigation and thus an accurate iceberg reconstruction. This approach is successfully demonstrated on real world iceberg data collected during the 2018 Sermilik campaign in Eastern Greenland. Due to the availability of iceberg mounted GPS observations during this research expedition we could also groundtruth our navigation and thus our systems level mapping efforts.