PhD Pilot Blog

How a Multi-Sensor, Multi-Robot System Works for Environmental Monitoring

DIWA PhD Pilot blog Haizhou Zhang

Haizhou Zhang, University of Turku. hazhan@utu.fi

Have you ever wondered how robots can help protect our environment?

Imagine drones flying above rivers and autonomous boats sailing smoothly on the water, all equipped with cutting-edge sensors working together. This isn’t science fiction—it’s today’s technology revolutionizing environmental monitoring. My research at the University of Turku is all about making this futuristic vision a reality by combining multiple robots equipped with diverse sensors to better understand and protect our natural surroundings.

Photo: Haizhou Zhang.

Why Multi-Sensor Integration Matters

Traditional environmental monitoring methods often rely on single-sensor systems, such as cameras or simple water sampling devices. While these methods have served well, they have clear limitations. For example, cameras alone might struggle to detect underwater features, and water sensors cannot monitor large areas quickly.

This is where multi-sensor systems excel. By integrating various sensors like LiDAR (laser scanning), high-resolution cameras, and inertial measurement units (IMUs), we achieve a far more detailed and comprehensive view of the environment. For instance, LiDAR can map terrain accurately, while cameras provide visual detail and IMUs ensure precise navigation even in challenging conditions.

How Robots Work Together

In the sketch shown above, we envisioned a collaborative robot system: Unmanned Aerial Vehicles (UAVs)—essentially drones—and Unmanned Surface Vehicles (USVs)—autonomous boats—cooperate. UAVs provide aerial coverage and can rapidly survey large areas, identifying critical points of interest from above. Meanwhile, USVs operate closely on water, carrying out detailed inspections and deploying underwater sensors.

Through collaborative frameworks, these robots share information instantly. For example, drones identify suitable areas for boats to investigate further, directing their attention effectively. This teamwork significantly enhances situational awareness, ensuring accurate data collection and quick responses to environmental changes.

Real-Life Applications and Benefits

Consider river monitoring: Traditionally, surveying a riverbed or tracking fish populations requires significant resources. With our robotic systems, UAVs quickly identify regions of interest, while USVs equipped with LiDAR and underwater cameras investigate detailed conditions beneath the surface, such as riverbed erosion or fish habitats.

Additionally, deploying these multi-robot systems in remote areas or during environmental emergencies, like oil spills or natural disasters, can provide rapid and comprehensive data, helping authorities respond faster and better informed.

Key Technical Components

1. multi-modal Sensor Fusion

Integrating LiDAR point clouds with camera images and IMU data builds a unified 3D model of the environment. Accurate calibration and time synchronization are critical—misaligned data can lead to false positives in change detection or object identification.

2. Deep Learning for perception

We leverage convolutional neural networks for detection. For example, on Unmanned Aerial Vehicles (UAVs), segmentation highlights shoreline boundaries; on USVs, it identifies underwater conditions, like vegetation and debris. These techniques convert raw sensor feeds into actionable insights for environment assessment.

Looking Ahead

By combining multi-sensor systems on multiple robots, environmental monitoring can become more adaptive, scalable, and secure. Future work will extend this approach to quadruped robots for shoreline inspections and explore satellite data integration for wider-area context. The potential impact spans freshwater conservation, marine ecology, and disaster response.


References

Neuvition. (21.4.2023). Enhancing Environmental Management with LiDAR System Technology. https://www.neuvition.com/media/enhancing-environmental-management-with-lidar-system-technology.html

Zhang, H., Yu, X., Ha, S., & Westerlund, T. (2023). LiDAR-Generated Images Derived Keypoints Assisted Point Cloud Registration Scheme in Odometry Estimation. Remote Sensing15(20), 5074. https://doi.org/10.3390/rs15205074

16.5.2025.

Share the Post:

Related Posts

Digitaaliset vedetpodcast

Uusi “Digitaaliset vedet” podcast sukeltaa syvälle Digitaaliset vedet -lippulaivan tutkimukseen

Tutkimusyhteisö saa uuden äänen, kun kuudesta jaksosta koostuva Digitaaliset vedet –podcast julkaisee ensimmäisen jaksonsa ensi viikolla. Podcast pureutuu DIWA-lippulaivan tutkimukseen, tarjoten kuulijoille puhetta digitaalisista vesistä, vesitutkimuksesta ja DIWA-lippulaivasta.   Ensimmäinen jakso julkaistaan torstaina 19.2., ja se on kuunneltavissa Spotifyssa sekä katsottavissa YouTubessa. Podcastin tavoitteena on tehdä DIWA-lippulaivan tutkimus saavutettavaksi kaikille ja herättää keskustelua vesitutkimuksen ympärillä.  Seuraa podcast-sivuamme ja

NHC2026

Nordic Hydrological Conference on August 18th to 20th, 2026, in Åland Islands, Finland! 

We are proud to announce the upcoming Nordic Hydrological Conference (NHC2026), which brings together experts, researchers, and practitioners to explore new directions in “Digital solutions in hydrology and sustainable water management”.   The conference is organized in cooperation with the Nordic Association for Hydrology (NHF) and Digital Waters Flagship. Conference