CMS: Enabling Real-world Cooperative Multi-modal Sensing
Bo Wu Jerry Li Ruoshen Mo Justin Yue Hang Qiu
University of California, Riverside
Paper | Code | Demo | Bibtex
Abstract
CMS, is a open-source, cooperative multi-modal sensing platform. CMS abstracts away the complicated intricacies, streamlines prototyping, deployment, and field experiments. Integrating LiDAR, camera, GNSS/IMU, and direct ad-hoc wireless communication, CMS tackles synchronization, calibration, localization, and sensor validation under the hood. This video demonstrates the capabilities of CMS. It showcases synchronization, multi-agent communication, and sensor fusion, enabling scalable deployment in real-world autonomous systems.
![]() |
CMS Overview
CMS integrates LiDAR, Camera, GNSS with a power-over-ethernet (PoE) switch, which forwards the data to a central ROS node (running on a laptop). The laptop and sensors are synchronized with GNSS time, and all intrinsic and extrinsic parameters are calibrated for all sensors. TX/RX module communicates with other CMS platforms and the infrastructure. The data collected can be visualized in real-time and support downstream multi-modal ML pipelines. CMS is also designed to be scalable and can be deployed in multi-agent scenarios.

Evaluations
We evaluate CMS across areas of synchronization, calibration, communication, and localization as they are all necessary for a successful multi agent system. Below are some evaluations while the rest can be found in our paper.



Citation
@inbook{10.1145/3715014.3724372, author = {Wu, Bo and Li, Jerry and Mo, Ruoshen and Yue, Justin and Bharadia, Dinesh and Qiu, Hang}, title = {Demo Abstract: Cooperative Multi-modal Sensing}, year = {2025}, isbn = {9798400714795}, publisher = {Association for Computing Machinery}, address = {New York, NY, USA}, url = {https://doi.org/10.1145/3715014.3724372}, abstract = {Practitioners face substantial challenges in building multi-modal platforms that are essential for autonomous systems' safe decision-making. Those complications, including synchronization, calibration, and tedious sensor validation, hinder user adoption for real-world applications. We present CMS, a Cooperative Multi-modal Sensing Platform. CMS provides one consistent interface, integrating LiDAR, camera, RaDAR, and GNSS/IMU, streamlines these processes and makes the intricacies transparent to users and applications. Our demonstration shows that CMS can obtain high-quality multi-modal sensor data, paving the way toward real-world prototypes of cooperative autonomous systems.}, booktitle = {Proceedings of the 23rd ACM Conference on Embedded Networked Sensor Systems}, pages = {712–713}, numpages = {2} }