<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "http://jats.nlm.nih.gov/publishing/1.1/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xml:lang="en" article-type="review-article" dtd-version="1.1">
<front>
<journal-meta>
<journal-id journal-id-type="pmc">CMC</journal-id>
<journal-id journal-id-type="nlm-ta">CMC</journal-id>
<journal-id journal-id-type="publisher-id">CMC</journal-id>
<journal-title-group>
<journal-title>Computers, Materials &#x0026; Continua</journal-title>
</journal-title-group>
<issn pub-type="epub">1546-2226</issn>
<issn pub-type="ppub">1546-2218</issn>
<publisher>
<publisher-name>Tech Science Press</publisher-name>
<publisher-loc>USA</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">71599</article-id>
<article-id pub-id-type="doi">10.32604/cmc.2025.071599</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Review</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Sensor Fusion Models in Autonomous Systems: A Review</article-title>
<alt-title alt-title-type="left-running-head">Sensor Fusion Models in Autonomous Systems: A Review</alt-title>
<alt-title alt-title-type="right-running-head">Sensor Fusion Models in Autonomous Systems: A Review</alt-title>
</title-group>
<contrib-group>
<contrib id="author-1" contrib-type="author">
<name name-style="western"><surname>Mittal</surname><given-names>Sangeeta</given-names></name><xref ref-type="aff" rid="aff-1">1</xref></contrib>
<contrib id="author-2" contrib-type="author">
<name name-style="western"><surname>Gupta</surname><given-names>Chetna</given-names></name><xref ref-type="aff" rid="aff-1">1</xref></contrib>
<contrib id="author-3" contrib-type="author" corresp="yes">
<name name-style="western"><surname>Gupta</surname><given-names>Varun</given-names></name><xref ref-type="aff" rid="aff-2">2</xref><xref ref-type="aff" rid="aff-3">3</xref><email>varun.iit13@gmail.com</email></contrib>
<aff id="aff-1"><label>1</label><institution>School of Computer Science Engineering and Technology, Bennett University</institution>, <addr-line>Greater Noida, 201310</addr-line>, <country>India</country></aff>
<aff id="aff-2"><label>2</label><institution>Multidisciplinary Research Centre for Innovations in SMEs (MrciS), Gisma University of Applied Sciences</institution>, <addr-line>Potsdam, 14469</addr-line>, <country>Germany</country></aff>
<aff id="aff-3"><label>3</label><institution>Department of Economics and Business Administration, Universidad de Alcal&#x00E1;</institution>, <addr-line>Madrid, 28801</addr-line>, <country>Spain</country></aff>
</contrib-group>
<author-notes>
<corresp id="cor1"><label>&#x002A;</label>Corresponding Author: Varun Gupta. Email: <email>varun.iit13@gmail.com</email></corresp>
</author-notes>
<pub-date date-type="collection" publication-format="electronic">
<year>2026</year>
</pub-date>
<pub-date date-type="pub" publication-format="electronic">
<day>10</day><month>2</month><year>2026</year>
</pub-date>
<volume>87</volume>
<issue>1</issue>
<elocation-id>6</elocation-id>
<history>
<date date-type="received">
<day>08</day>
<month>08</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>23</day>
<month>10</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2026 The Authors.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Published by Tech Science Press.</copyright-holder>
<license xlink:href="https://creativecommons.org/licenses/by/4.0/">
<license-p>This work is licensed under a <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution 4.0 International License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<self-uri content-type="pdf" xlink:href="TSP_CMC_71599.pdf"></self-uri>
<abstract>
<p>This survey presents a comprehensive examination of sensor fusion research spanning four decades, tracing the methodological evolution, application domains, and alignment with classical hierarchical models. Building on this long-term trajectory, the foundational approaches such as probabilistic inference, early neural networks, rule-based methods, and feature-level fusion established the principles of uncertainty handling and multi-sensor integration in the 1990s. The fusion methods of 2000s marked the consolidation of these ideas through advanced Kalman and particle filtering, Bayesian&#x2013;Dempster&#x2013;Shafer hybrids, distributed consensus algorithms, and machine learning ensembles for more robust and domain-specific implementations. From 2011 to 2020, the widespread adoption of deep learning transformed the field driving some major breakthroughs in the autonomous vehicles domain. A key contribution of this work is the assessment of contemporary methods against the JDL model, revealing gaps at higher levels- especially in situation and impact assessment. Contemporary methods offer only limited implementation of higher-level fusion. The survey also reviews the benchmark multi-sensor datasets, noting their role in advancing the field while identifying major shortcomings like the lack of domain diversity and hierarchical coverage. By synthesizing developments across decades and paradigms, this survey provides both a historical narrative and a forward-looking perspective. It highlights unresolved challenges in transparency, scalability, robustness, and trustworthiness, while identifying emerging paradigms such as neuromorphic fusion and explainable AI as promising directions. This paves the way forward for advancing sensor fusion towards transparent and adaptive next-generation autonomous systems.</p>
</abstract>
<kwd-group kwd-group-type="author">
<kwd>Sensor fusion</kwd>
<kwd>autonomous systems</kwd>
<kwd>artificial intelligence</kwd>
<kwd>machine learning</kwd>
<kwd>sensor data integration</kwd>
<kwd>intelligent systems</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="s1">
<label>1</label>
<title>Introduction</title>
<p>Sensor fusion represents one of the most transformative technologies in modern autonomous systems, enabling machines to perceive and interpret their environment with unprecedented accuracy and reliability. The fundamental principle underlying sensor fusion is the combination of data from multiple sensors to create a more comprehensive understanding than would be possible using individual sensors alone [<xref ref-type="bibr" rid="ref-1">1</xref>]. This technological paradigm has evolved from simple data combination techniques to applications of classical mathematical models to sophisticated Artificial Intelligence-driven approaches capable of real-time decision-making in complex, dynamic environments. The historical development of sensor fusion can be traced back to the 1950s, when military applications first demonstrated the potential of combining multiple radar systems for enhanced target detection [<xref ref-type="bibr" rid="ref-2">2</xref>]. The concept gained significant momentum in the 1960s when mathematicians developed algorithmic frameworks for multi-sensor data integration, laying the groundwork for modern fusion architectures. The establishment of the Joint Directors of Laboratories (JDL) Data Fusion Subpanel in 1986 marked a pivotal moment in the field, introducing standardized models and terminology that continue to influence contemporary research [<xref ref-type="bibr" rid="ref-3">3</xref>].</p>
<p>Modern autonomous systems in diverse domains such as autonomous vehicles and unmanned aerial systems, healthcare monitoring, and defense applications rely heavily on sensor fusion to achieve reliable operation in real-world environments [<xref ref-type="bibr" rid="ref-4">4</xref>,<xref ref-type="bibr" rid="ref-5">5</xref>]. These systems rely on a multitude of sensors to perceive their environment and make informed decisions. The integration of sensors such as LiDAR, cameras, radar, Inertial Measurement Units (IMUs), and GPS enables these systems to overcome the inherent limitations of individual sensors while capitalizing on their complementary strengths [<xref ref-type="bibr" rid="ref-6">6</xref>,<xref ref-type="bibr" rid="ref-7">7</xref>] Sensor data fusion is essential for these systems to integrate heterogeneous, high-volume, real-time data and derive a coherent understanding of surroundings [<xref ref-type="bibr" rid="ref-8">8</xref>,<xref ref-type="bibr" rid="ref-9">9</xref>]. However, an interesting finding is that the rise of powerful AI methods has overshadowed the original spirit of hierarchical fusion, with little substantive advancement occurring at the higher fusion levels.</p>
<p>The emergence of Explainable AI (XAI) to address the black-box nature of deep learning systems by providing interpretable insights into fusion decisions has benefitted sensor fusion for reliable decision-making. Visual explanations are being developed for autonomous systems [<xref ref-type="bibr" rid="ref-10">10</xref>,<xref ref-type="bibr" rid="ref-11">11</xref>]. This development is particularly significant for autonomous vehicles and medical applications, where understanding the reasoning behind system decisions is essential. However, the complexity of explainability methods needs to be reduced for producing explanations in real-time [<xref ref-type="bibr" rid="ref-10">10</xref>,<xref ref-type="bibr" rid="ref-12">12</xref>]. Contemporary research trends also emphasize edge AI deployment and neuromorphic computing as promising directions to achieve ultra-efficient sensor fusion with minimal latency. These approaches enable real-time processing directly on the sensor nodes, reducing communication overhead and improving the responsiveness of the system while maintaining low power consumption [<xref ref-type="bibr" rid="ref-13">13</xref>,<xref ref-type="bibr" rid="ref-14">14</xref>].</p>
<p>This review paper has been written with the objective of putting into context the evolution of sensor fusion methods. It seeks to trace how foundational models developed at a time when computational and sensing resources were limited have set the stage for contemporary approaches. The paper highlights the evolution in the design philosophies, techniques and application domains of sensor fusion. The study revealed that the probabilistic and rule-based models are largely being replaced by machine learning approaches. Due to this, sensor fusion research has now moved from concept-driven formulations to data-driven, adaptive, and context-aware systems.</p>
<p>Unlike prior reviews, this work offers a multi-era synthesis that traces the evolution of sensor fusion from early probabilistic and rule-based frameworks to modern AI/ML-based architectures. Explicitly, the fusion layers of traditional models are applied to contemporary AI/ML pipelines, exposing gaps at higher fusion levels where implementation remains limited. Furthermore, this survey provides broad coverage across domains that include transportation, healthcare, defense, agriculture, industry, and smart cities, far beyond the narrower focus of earlier surveys. Finally, the emerging paradigms such as neuromorphic computing, edge AI, and explainable AI are also positioned as promising directions for next-generation sensor fusion systems. These contributions distinguish our review from existing literature and are useful both for researchers and practitioners. This survey spans four decades of sensor-fusion research&#x2014;from probabilistic and rule-based methods in the 1980s&#x2013;1990s, through Bayesian and filtering approaches of the 2000s, to deep-learning and transformer approaches of the 2010s&#x2013;post-2021, and neuromorphic paradigms in recent times. The classical data-fusion frameworks have been integrated to adopt a unified reference hierarchical sensor-fusion framework with following levels: Level 0&#x2014;signal preprocessing; Level 1&#x2014;object refinement, Level 2&#x2014;situation assessment, Level 3&#x2014;impact/threat assessment, and Level 4&#x2014;process refinement. A key contribution of this survey is to situate fusion methods within their historical foundations, highlighting existing challenges and future opportunities.</p>
<p>This discussion has been organized in the remainder of the paper as follows. The research method adopted for the review is presented in <xref ref-type="sec" rid="s2">Section 2</xref>. It also outlines the limitations of existing reviews in this area. In <xref ref-type="sec" rid="s3">Section 3</xref>, the types of sensors used in various application areas of autonomous systems has been explained, and the characteristics of the sensor data has been described, leading to challenges in sensor fusion. The four-decade evolution of sensor fusion from the early 1980s to the current day has been described in <xref ref-type="sec" rid="s4">Section 4</xref>. This section reviews the evolution of sensor fusion models, analyzes their effectiveness, and charts the way forward. <xref ref-type="sec" rid="s5">Section 5</xref> describes the development of layered fusion models and details the methods mapped to each level. <xref ref-type="sec" rid="s6">Section 6</xref> discusses the future research directions within this field. The paper is concluded in <xref ref-type="sec" rid="s7">Section 7</xref>.</p>
</sec>
<sec id="s2">
<label>2</label>
<title>Research Method</title>
<p>In line with systematic review practices [<xref ref-type="bibr" rid="ref-15">15</xref>], the search and selection process followed a structured PRISMA-style workflow. Publications spanning from the early 1980s through May 2025 were retrieved from major scientific databases, including IEEE Xplore, Scopus, Web of Science, ScienceDirect, SpringerLink, and arXiv. Seminal surveys and foundational works on classical sensor fusion were used as anchors to expand the search, ensuring both breadth and depth of coverage.</p>
<p><bold>Identification.</bold> A total of 135 records were initially identified through database searches. These encompassed peer-reviewed journals, highly cited conference proceedings, authoritative book chapters, and selected arXiv preprints. Benchmark surveys and seminal contributions were also incorporated to establish the initial reference base.</p>
<p><bold>Screening.</bold> Following the removal of 15 duplicates, 120 records were screened at the title and abstract level. At this stage, priority was given to studies addressing sensor fusion in autonomous system domains such as transportation, healthcare, defense, robotics, agriculture, and smart cities. Thirty-six records were excluded as irrelevant, leaving 84 studies for full-text assessment.</p>
<p><bold>Eligibility.</bold> Full-text evaluation was then performed on these 84 studies to ensure that each (a) proposed, applied, or critically reviewed sensor fusion models or techniques; (b) documented applications in autonomous domains; and (c) addressed either classical (model-driven) or contemporary (AI/ML-based) approaches. Studies lacking methodological rigor, technical clarity, or empirical results were excluded. This led to the removal of 3 articles that failed to meet eligibility criteria.</p>
<p><bold>Inclusion.</bold> A final set of 81 studies was included in the review corpus. These works collectively support a multi-era synthesis mapping classical models to modern AI/ML pipelines and incorporating emerging paradigms such as neuromorphic and quantum-inspired fusion. Extracted content was consolidated into thematic tables covering chronological and technological milestones, application-specific deployments and challenges, and comparative insights across classical and modern approaches.</p>
<p>The overall workflow is summarized in <xref ref-type="table" rid="table-1">Table 1</xref>.</p>
<table-wrap id="table-1">
<label>Table 1</label>
<caption>
<title>Summary of the search and selection process (PRISMA-style)</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col align="center"/> </colgroup>
<thead>
<tr>
<th>Stage</th>
<th>Number of records</th>
</tr>
</thead>
<tbody>
<tr>
<td>Identified through databases</td>
<td>135</td>
</tr>
<tr>
<td>Duplicates removed</td>
<td>15</td>
</tr>
<tr>
<td>Records screened (titles/abstracts)</td>
<td>120</td>
</tr>
<tr>
<td>Records excluded (domains with lesser scope of multisensor fusion)</td>
<td>32</td>
</tr>
<tr>
<td>Full-text articles studied</td>
<td>88</td>
</tr>
<tr>
<td>Full-text articles excluded (lack of rigor/duplication)</td>
<td>3</td>
</tr>
<tr>
<td><bold>Total Studies finally included in review</bold></td>
<td><bold>85</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>While every effort was made to ensure comprehensiveness, limitations remain. The review is constrained by the availability of published results only, without inclusion of unpublished industrial reports, internal datasets, or simulations. Given the vastness of the field, some subdomains may not have been fully represented. Despite these constraints, the structured and transparent approach adopted here ensures both analytical rigor and reproducibility, offering a panoramic yet critical view of the sensor fusion landscape. This foundation informs researchers, practitioners, and policymakers by situating contemporary developments within their historical and methodological contexts.</p>
<p><xref ref-type="table" rid="table-2">Table 2</xref> summarizes salient features of recent state-of-the-art surveys and demonstrates how the present review advances the literature through a unique multi-era synthesis, explicit mapping from fusion layers to AI/ML pipelines, and broad cross-domain coverage.</p>
<table-wrap id="table-2">
<label>Table 2</label>
<caption>
<title>Comparison of recent review papers on sensor fusion in autonomous systems and the present study</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col align="center"/>
<col align="center"/> </colgroup>
<thead>
<tr>
<th>Paper and year</th>
<th>Main focus</th>
<th>Contribution highlights</th>
</tr>
</thead>
<tbody>
<tr>
<td>Zhao et al., 2024 [<xref ref-type="bibr" rid="ref-16">16</xref>]</td>
<td>Review of 3D object detection methods for multi-sensor fusion with emphasis on LiDAR&#x2013;camera integration</td>
<td>Detailed taxonomy of fusion strategies (early, feature, and late fusion); analysis of datasets and metrics; focused on 3D perception in autonomous vehicles</td>
</tr>
<tr>
<td>Wang et al., 2024 [<xref ref-type="bibr" rid="ref-17">17</xref>]</td>
<td>Survey of sensor fusion and localization methods, spanning Kalman filters, particle filters, and ML-based approaches</td>
<td>Proposes hybrid ML&#x2013;classical frameworks; addresses challenges in GPS-denied and noisy environments; emphasizes scalability for real-world deployment</td>
</tr>
<tr>
<td>Mehta et al., 2025 [<xref ref-type="bibr" rid="ref-18">18</xref>]</td>
<td>Comprehensive overview of sensor fusion in autonomous vehicles, UAVs, and robotics across fusion levels</td>
<td>Provides systematic classification across data-, feature-, and decision-level fusion; highlights safety, reliability, and integration of diverse sensing modalities</td>
</tr>
<tr>
<td>Yeong et al., 2025 [<xref ref-type="bibr" rid="ref-10">10</xref>]</td>
<td>Structured review of multi-sensor fusion with focus on deep learning methods in autonomous driving</td>
<td>Presents formal mathematical formulations; incorporates vision&#x2013;language models and large language models into fusion; discusses emerging AI-driven trends</td>
</tr>
<tr>
<td><bold>Present review</bold></td>
<td>Evolution of sensor fusion in autonomous systems, bridging classical models and AI/ML-driven architectures across domains</td>
<td>Advances beyond prior surveys by integrating neuromorphic computing, edge AI, and explainable AI; offers multi-era synthesis, explicit layer-to-pipeline mapping, and broad cross-domain coverage (transportation, healthcare, defense, agriculture, industry, and smart cities)</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s3">
<label>3</label>
<title>Characteristics of Sensor Data</title>
<p>Sensor data in autonomous systems tends to be multi-modal (originating from different sensor types), high-dimensional, voluminous, and generated in real time. These data often contain noise and uncertainties specific to each sensor. Moreover, the data rates can be extremely high, reaching gigabytes per second. Thus, efficient pre-processing (filtering, calibration, compression) is needed prior to fusion. Data may also be heterogeneous in format and scale, requiring transformation into common representations or extraction of intermediate features.</p>
<p>Another important aspect is the context-dependence and non-stationarity of sensor data. Sensors operate under varying conditions (day/night, clear/rainy, highway/city), which directly affect data quality. Fusion systems must be robust to such variations, for example by dynamically weighting sensor contributions (e.g., relying more on radar in heavy rain). Synchronization among sensors is equally critical, as misaligned timestamps can propagate into significant fusion errors.</p>
<p>To illustrate the diversity of sensor data used in autonomous systems, <xref ref-type="table" rid="table-3">Table 3</xref> provides examples of benchmark datasets across domains such as autonomous driving, wearable computing, and remote sensing. These datasets exemplify the modalities and scenarios available to researchers for developing and evaluating fusion methods.</p>
<table-wrap id="table-3">
<label>Table 3</label>
<caption>
<title>Some publicly available multi-modal sensor datasets and application domains</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/> </colgroup>
<thead>
<tr>
<th>Dataset name</th>
<th>Key characteristics</th>
<th>Sensor modalities</th>
<th>Application domain</th>
</tr>
</thead>
<tbody>
<tr>
<td>WISDM (2023) [<xref ref-type="bibr" rid="ref-19">19</xref>]</td>
<td>Motion sensor data from smartphones for activity classification</td>
<td>Smartphone accelerometer and gyroscope</td>
<td>Human activity recognition</td>
</tr>
<tr>
<td>Argoverse 2 (2021) [<xref ref-type="bibr" rid="ref-20">20</xref>]</td>
<td>Large-scale multi-sensor driving dataset with 3D tracking annotations</td>
<td>LiDAR, multiple cameras, GPS</td>
<td>Autonomous driving</td>
</tr>
<tr>
<td>RarePlanes (2021) [<xref ref-type="bibr" rid="ref-21">21</xref>]</td>
<td>Aerial imagery combining real and synthetic data for object detection</td>
<td>Satellite RGB imagery (real &#x002B; synthetic)</td>
<td>Remote sensing and surveillance</td>
</tr>
<tr>
<td>UrbanLoco (2020) [<xref ref-type="bibr" rid="ref-22">22</xref>]</td>
<td>Urban localization dataset in challenging environments</td>
<td>LiDAR, IMU, GPS, Cameras</td>
<td>Autonomous navigation</td>
</tr>
<tr>
<td>nuScenes (2019) [<xref ref-type="bibr" rid="ref-23">23</xref>]</td>
<td><inline-formula id="ieqn-1"><mml:math id="mml-ieqn-1"><mml:msup><mml:mi>360</mml:mi><mml:mrow><mml:mo>&#x2218;</mml:mo></mml:mrow></mml:msup></mml:math></inline-formula> sensor coverage with multimodal annotations for detection and tracking</td>
<td>LiDAR, cameras, radar, GPS, IMU</td>
<td>Autonomous driving</td>
</tr>
<tr>
<td>A9 Highway Dataset (2018) [<xref ref-type="bibr" rid="ref-24">24</xref>]</td>
<td>Multi-modal dataset of highway driving scenarios</td>
<td>Cameras, LiDAR, radar, GPS</td>
<td>Autonomous driving</td>
</tr>
<tr>
<td>ExtraSensory (2016) [<xref ref-type="bibr" rid="ref-25">25</xref>]</td>
<td>Crowd-sourced multi-modal recordings for activity recognition</td>
<td>Smartphone and smartwatch sensors (accelerometer, gyroscope, audio, GPS)</td>
<td>Wearable IoT and computing</td>
</tr>
<tr>
<td>OPPORTUNITY (2011) [<xref ref-type="bibr" rid="ref-26">26</xref>]</td>
<td>Sensor recordings in a home environment for activity recognition</td>
<td>Body-worn inertial, object, and ambient environmental sensors</td>
<td>Human activity recognition</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The datasets included in <xref ref-type="table" rid="table-3">Table 3</xref> were chosen because they represent widely recognized benchmarks within their respective domains and are frequently cited in state-of-the-art sensor fusion studies. Their inclusion ensures that the survey reflects the most commonly used testbeds against which fusion methods are evaluated, while also highlighting their limitations in representativeness and domain coverage.</p>

<p>While these benchmark datasets provide valuable testbeds, their utility varies significantly depending on the target application. For instance, large-scale autonomous driving datasets such as Argoverse 2 [<xref ref-type="bibr" rid="ref-20">20</xref>], nuScenes [<xref ref-type="bibr" rid="ref-23">23</xref>], and the A9 Highway dataset [<xref ref-type="bibr" rid="ref-24">24</xref>] are rich in multimodal coverage and support complex perception tasks, but they are often biased toward urban traffic conditions in developed regions. This limits their generalizability to rural or less-structured environments. Similarly, UrbanLoco [<xref ref-type="bibr" rid="ref-22">22</xref>] offers challenging urban localization scenarios but is geographically constrained and may not fully capture cross-regional variations such as GPS multipath in dense high-rise cities. In contrast, human activity and mobile health benchmarks such as WISDM [<xref ref-type="bibr" rid="ref-19">19</xref>], OPPORTUNITY [<xref ref-type="bibr" rid="ref-26">26</xref>], and ExtraSensory [<xref ref-type="bibr" rid="ref-25">25</xref>] demonstrate strong utility for wearable and IoT-driven fusion research. However, many of these datasets are collected in controlled or semi-structured environments, which may not reflect the noise and variability encountered in real-world deployments. They also tend to have limited subject diversity, raising questions about demographic generalizability in healthcare applications. Remote sensing benchmarks such as RarePlanes [<xref ref-type="bibr" rid="ref-21">21</xref>] highlight another dimension: the fusion of synthetic and real data for training. While this enables large-scale dataset generation, it also introduces a domain gap between simulated and operational settings, complicating transferability of models trained exclusively on such data. Overall, <xref ref-type="table" rid="table-3">Table 3</xref> underscores both the breadth of sensor modalities represented and the uneven distribution of benchmarks across domains. Autonomous driving enjoys abundant and well-annotated datasets, while healthcare, smart cities, and industrial domains remain comparatively underrepresented. This imbalance constrains cross-domain fusion research and highlights a critical need for more diverse, standardized, and globally representative datasets. Without such resources, fusion models risk overfitting to narrow operational conditions and may fail when transferred to new environments. Overall, <xref ref-type="table" rid="table-3">Table 3</xref> underscores both the breadth of sensor modalities represented and the uneven distribution of benchmarks across domains. Autonomous driving enjoys abundant and well-annotated datasets, while healthcare, smart cities, and industrial domains remain comparatively underrepresented. This imbalance constrains cross-domain fusion research and highlights a critical need for more diverse, standardized, and globally representative datasets. Without such resources, fusion models risk overfitting to narrow operational conditions and may fail when transferred to new environments. A further consideration is the inherent trade-offs among these datasets. Large-scale benchmarks such as nuScenes and Argoverse 2 provide extensive multimodal coverage but sacrifice diversity across geographic and environmental contexts. Conversely, smaller datasets like OPPORTUNITY and ExtraSensory capture rich multimodal signals in daily-life settings but lack the scale needed for training data-intensive models. Synthetic-enhanced datasets such as RarePlanes expand coverage at low cost yet introduce a domain gap that complicates real-world transferability. These trade-offs between scale, diversity, realism, and generalizability must therefore be carefully weighed when selecting benchmarks for evaluating sensor fusion methods.</p>

</sec>
<sec id="s4">
<label>4</label>
<title>Evolution of Sensor Fusion: A Four-Decade Perspective</title>
<p>The evolution of sensor fusion techniques over the past four decades represents a remarkable journey from basic mathematical algorithms to sophisticated AI-driven systems. Recent proliferation of complex sensor arrays and the need for real-time and adaptive fusion have driven the adoption of deep learning, transformer architectures, and energy-efficient neuromorphic computing, enabling autonomous systems to achieve new levels of perception and autonomy. Over the years, numerous sensor fusion models have been proposed. In this section, a decade-wise review of the evolution of sensor fusion is given. A hybrid perspective integrating traditional fusion architectures to a unified hierarchical model is also proposed.</p>
<sec id="s4_1">
<label>4.1</label>
<title>Some Popular Early Fusion Models and Frameworks</title>
<p>The concept of multisensor data fusion dates back to the 1970s in the context of robotics and defense. Early work focused on establishing theoretical foundations and lexicons for combining data from multiple sources within permissible time frames. A widely cited definition by Hall and McMullen described data fusion as a &#x2019;hierarchical transformation of observed data from multiple sources into a form that enables decision making&#x2019; [<xref ref-type="bibr" rid="ref-27">27</xref>]. In practice, many initial fusion systems were deterministic or rule-based aimed at achieving specific fusion goals under hardware constraints of the time.</p>
<p>One pioneering framework was the Joint Directors of Laboratories (JDL) Data Fusion Model [<xref ref-type="bibr" rid="ref-3">3</xref>]. Developed in the military community in the 1980s, the JDL model defined a taxonomy of fusion across levels 0 to 4: from raw data alignment, to object refinement (state estimation), situation assessment, impact assessment, and process refinement. It emphasized combining sensor observations to estimate object identity and position, originally for surveillance/tracking applications. Steinberg et al. later revised the model in 1999 to refine these levels and generalize it to broader situations [<xref ref-type="bibr" rid="ref-28">28</xref>]. The layered approach of the JDL framework influenced many subsequent system designs, ensuring that each level of fusion produces outputs at increasing levels of abstraction. Although conceived for military sensing, the concepts of the JDL model are applicable to any multisensor system.</p>
<p>Around the same time, reference [<xref ref-type="bibr" rid="ref-29">29</xref>] proposed a simpler three-level architecture for sensor fusion. The lowest level dealt with raw signal fusion (often requiring training to learn correlations between sensors); the intermediate &#x201C;evidence&#x201D; level fused features or evidence using statistical methods (with spatial/temporal alignment as a preprocessing step); the highest &#x201C;dynamics&#x201D; level fused information in the context of system dynamics or models. Reference [<xref ref-type="bibr" rid="ref-29">29</xref>] introduced performance indicators such as the quality of fused information and robustness to uncertainties, which foreshadowed later work on fusion confidence and uncertainty estimation.</p>
<p>Another influential early framework was by Luo and Kay [<xref ref-type="bibr" rid="ref-30">30</xref>], who distinguished between multi-sensor integration (using multiple sensors to reach one decision) and multi-sensor fusion. They proposed hierarchical structure with distributed fusion centers, highlighting that fusion could occur at different hierarchy levels of a system. The data collected at the sensor level is integrated at the fusion centers, where the actual fusion is done. After processing all sensors, domain-specific high-level information of interest is obtained. The fusion process is supported by relevant databases and libraries. In the process of fusion, raw signals from individual sensors are abstracted to symbolic information. This idea of performing some fusion locally (sensor node level) and some globally (central level) is reflected in today&#x2019;s edge vs. cloud fusion split in IoT systems.</p>
<p>Harris et al described another example of hierarchical fusion called the waterfall model [<xref ref-type="bibr" rid="ref-31">31</xref>]. The hierarchical levels are similar in essence as the earlier models with an emphasis on the processing functions of the lower levels. Sensors pre-processing is done at at level 1 while feature extraction and pattern processing in level 2. It is followed by situation assessment and decision making being done at level 3. Conceptually, the processed signal from level 1 are converted to fetaures in level 2 that leads to state description and querying in Level 3 of the model.</p>
<p>Bedworth and Brien [<xref ref-type="bibr" rid="ref-32">32</xref>] described a hybrid framework called the Omnibus model. This process model was inspired by conceptual OODA (Observe, Orient, Decide and Act) cycle called Boyd loop and the waterfall model. Various tasks in data fusion and its functional objectives are realized in different modules. Three levels of data fusion, that is, data, feature and decision level have been defined. Separate modules implement various level-wise tasks and meet their functional objectives.</p>
<p>Several foundational frameworks emerged during the 1970s&#x2013;2000s that established the theoretical and architectural basis for multisensor data fusion. <xref ref-type="table" rid="table-4">Table 4</xref> summarizes these early models, highlighting their central ideas and lasting contributions. The JDL model remains one of the most influential, while subsequent frameworks such as the three-level architecture, waterfall model, and Omnibus model introduced alternative perspectives emphasizing hierarchy, distributed processing, and hybrid design. Collectively, these models shaped the evolution of modern sensor fusion approaches.</p>
<table-wrap id="table-4">
<label>Table 4</label>
<caption>
<title>Early multisensor data fusion frameworks</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col align="center"/>
<col align="center"/> </colgroup>
<thead>
<tr>
<th>Framework/Model</th>
<th>Key idea</th>
<th>Notable features/Contributions</th>
</tr>
</thead>
<tbody>
<tr>
<td>Hall &#x0026; McMullen (1970s&#x2013;1990s) [<xref ref-type="bibr" rid="ref-27">27</xref>]</td>
<td>Formal definition of data fusion as hierarchical transformation of multi-source observations.</td>
<td>Established theoretical foundations and lexicon; early implementations were rule-based or deterministic under hardware constraints.</td>
</tr>
<tr>
<td>JDL Model (1980s) [<xref ref-type="bibr" rid="ref-3">3</xref>,<xref ref-type="bibr" rid="ref-28">28</xref>]</td>
<td>Taxonomy of fusion across Levels 0&#x2013;4.</td>
<td>Levels: raw data alignment, object refinement, situation assessment, impact assessment, process refinement. Revised in 1999 for broader generality; widely influential in military and civilian systems.</td>
</tr>
<tr>
<td>Thomopoulos three-level architecture (1990) [<xref ref-type="bibr" rid="ref-29">29</xref>]</td>
<td>Three levels: signal, evidence, dynamics.</td>
<td>Introduced preprocessing for spatial/temporal alignment, robustness and quality measures for fused information; foreshadowed later work on uncertainty estimation.</td>
</tr>
<tr>
<td>Luo &#x0026; Kay hierarchical model (2002) [<xref ref-type="bibr" rid="ref-30">30</xref>]</td>
<td>Distinguished between integration and fusion; proposed distributed hierarchical fusion centers.</td>
<td>Fusion occurs at both local (sensor) and global (central) levels; supported by databases/libraries. Anticipates modern edge vs. cloud fusion in IoT systems.</td>
</tr>
<tr>
<td>Harris waterfall model (1998) [<xref ref-type="bibr" rid="ref-31">31</xref>]</td>
<td>Hierarchical fusion in sequential stages.</td>
<td>Level 1: sensor preprocessing; Level 2: feature extraction/pattern processing; Level 3: situation assessment and decision making.</td>
</tr>
<tr>
<td>Bedworth &#x0026; Brien omnibus model (2000) [<xref ref-type="bibr" rid="ref-32">32</xref>]</td>
<td>Hybrid model inspired by OODA loop and waterfall model.</td>
<td>Defined data, feature, and decision-level fusion; separate modules implement level-wise tasks to meet functional objectives.</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>A common feature of all models discussed till now is the hierarchical transformation of data. This higher-level integration of locally processed sensor data at intermediate or final nodes can be suitably applied to modern autonomous systems also. Interestingly, although these models were proposed more than a decade apart, they embody the same fundamental principle of <italic>hierarchical fusion</italic>, where both the level of cognizance about the system and the granularity of information progressively increase across successive layers. To systematically map the actual fusion models deployed in autonomous systems, we first developed a unified architecture that aligns the levels proposed by different frameworks and examined the extent of consensus among them. The rationale for this integration lies in the observation that, despite differences in terminology and chronology, all major fusion frameworks embody a common principle of hierarchical refinement: data is progressively transformed from raw sensor measurements into higher-level situational understanding and decision support. By aligning these frameworks, we expose the underlying consensus of established fusion frameworks. The unified model draws upon the Joint Directors of Laboratories (JDL) data fusion model, which has been briefly introduced earlier, it is elaborated here to provide the rationale for harmonizing different models. The JDL levels can be summarized as follows:
<list list-type="simple">
<list-item><label>1.</label><p><bold>Level 0: Sub-Object Data Assessment (Source Preprocessing)</bold>&#x2014;Deals with raw sensor data (signals, features, pixels, etc.), encompassing tasks such as noise filtering, feature extraction, registration, and alignment. <italic>Example: Cleaning raw radar or camera data and synchronizing sensing rates before applying detection algorithms</italic>.</p></list-item>
<list-item><label>2.</label><p><bold>Level 1: Object Refinement</bold>&#x2014;Integrates features to form objects/entities and estimate their states, involving tasks such as detection, tracking, identification, and classification. <italic>Example: Detecting and tracking a vehicle using fused radar and camera data</italic>.</p></list-item>
<list-item><label>3.</label><p><bold>Level 2: Situation Assessment</bold>&#x2014;Develops an understanding of relationships among objects and the environment, covering tasks such as scene analysis, intent recognition, and context modeling. <italic>Example: Recognizing that multiple vehicles are forming a traffic jam</italic>.</p></list-item>
<list-item><label>4.</label><p><bold>Level 3: Impact (Threat) Assessment</bold>&#x2014;Focuses on predicting the future state and potential consequences of the situation, including threat assessment, risk prediction, and decision support. <italic>Example: Predicting that a speeding car may cause a collision</italic>.</p></list-item>
<list-item><label>5.</label><p><bold>Level 4: Process Refinement (Resource Management)</bold>&#x2014;Controls and improves the fusion process itself, with tasks such as sensor management, adaptive fusion strategies, and feedback optimization. <italic>Example: Directing a drone to collect additional data in regions of high uncertainty</italic>.</p></list-item>
<list-item><label>6.</label><p><bold>Level 5: User/Cognitive Refinement (extension)</bold>&#x2014;Accounts for human&#x2013;machine interaction, including visualization, operator decision support, and incorporating human feedback.</p></list-item>
</list></p>
<p>By aligning the JDL model with other popular frameworks, an integrated multisensor fusion model has been shown in <xref ref-type="fig" rid="fig-1">Fig. 1</xref>.</p>
<fig id="fig-1">
<label>Figure 1</label>
<caption>
<title>A hierarchical unified multisensor fusion model</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_71599-fig-1.tif"/>
</fig>
<p>Traditional models established the architectural blueprints and terminology for sensor fusion. These can also be related to modern autonomous systems, as raw data from multiple sensors of autonomous systems must undergo several processing stages to become actionable information. In the next sections, we will discuss the historical trajectory of sensor fusion over the past four decades, examining how each decade introduced new sensors, fusion strategies, and application domains that collectively shaped the current state-of-the-art.</p>
</sec>
<sec id="s4_2">
<label>4.2</label>
<title>Foundational Fusion Approaches of the 1990s</title>
<p>The 1990s marked a decisive transition in sensor fusion research, moving from conceptual discussions to increasingly concrete algorithmic implementations. Researchers explored a broad range of strategies to address uncertainty, adaptability, and system integration, with varying degrees of success. Key categories included probabilistic inference methods, neural network&#x2013;based approaches, rule-based and evidence-theoretic reasoning, feature-level integration, and modular hybrid architectures. Collectively, these approaches laid important foundations, though they were frequently constrained by computational power and often tailored to narrow application contexts.</p>
<p><italic>Probabilistic Fusion Methods:</italic> Probabilistic inference emerged as a rigorous framework for handling uncertainty. Cox et al. (1992) applied Bayesian inference to stereo vision, showing that probabilistic depth estimation could outperform deterministic triangulation by integrating evidence from stereo pairs [<xref ref-type="bibr" rid="ref-33">33</xref>]. Larkin (1998) used Bayesian networks to classify acoustic signals, explicitly capturing dependencies among features [<xref ref-type="bibr" rid="ref-34">34</xref>]. Shahbazian et al. (1993) introduced multi-hypothesis tracking, allowing radar and sonar to jointly maintain competing target hypotheses [<xref ref-type="bibr" rid="ref-35">35</xref>]. While these approaches formalized uncertainty propagation, their computational cost scaled poorly with the number of sensors, preventing real-time deployment in dynamic environments.</p>
<p><italic>Neural Network-Based Methods:</italic> The growing availability of computational resources encouraged early use of neural networks for adaptive fusion. Perlovsky and McManus (1991) presented a maximum-likelihood neural network that adaptively classified sensor inputs, blending statistical estimation with learning [<xref ref-type="bibr" rid="ref-36">36</xref>]. Davis and Stentz (1995) demonstrated neural networks for autonomous outdoor navigation, where fused vision and range inputs were mapped to obstacle recognition in unstructured terrains [<xref ref-type="bibr" rid="ref-37">37</xref>]. Similarly, Cao and Hall (1998) applied neural networks to autonomous guided vehicles (AGVs) for vision-based navigation [<xref ref-type="bibr" rid="ref-38">38</xref>]. These methods demonstrated adaptability and the ability to capture nonlinear inter-sensor dependencies, but were shallow by modern standards, trained on limited data, and lacked interpretability, restricting their robustness in diverse conditions.</p>
<p><italic>Rule-Based and Evidence-Theoretic Approaches:</italic> Rule-driven frameworks also played an important role. McKee (1993) proposed a taxonomy of &#x201C;what can be fused&#x201D; for vision systems, providing systematic guidance for constructing integration pipelines [<xref ref-type="bibr" rid="ref-40">40</xref>]. Blasch and Hong (1999) implemented Dempster&#x2013;Shafer evidence theory in a &#x201C;belief filtering&#x201D; mechanism for target tracking [<xref ref-type="bibr" rid="ref-43">43</xref>]. This enabled reasoning under partial or conflicting evidence without requiring strict prior probabilities. Rule-based systems were transparent and interpretable, but generalization was limited, and belief combination rules were difficult to tune when ambiguity was high.</p>
<p><italic>Feature-Level Fusion:</italic> Several works moved beyond raw data integration to focus on fusing intermediate representations. Tang and Lee (1992) proposed geometric feature relation graphs to preserve spatial consistency across multi-camera vision sensors [<xref ref-type="bibr" rid="ref-39">39</xref>]. Peli et al. (1999) introduced unified feature-level fusion before classification, improving recognition accuracy in multisensor vision systems [<xref ref-type="bibr" rid="ref-44">44</xref>]. These approaches demonstrated the utility of fusing more compact representations rather than raw data, reducing computational demands. However, they relied on precise calibration and were sensitive to occlusion, noise, and sensor failures.</p>
<p><italic>Application-Specific Architectures:</italic> The decade also produced domain-tailored modular architectures. Kam, Zhu, and Kalata (1997) developed one of the earliest multi-sensor frameworks for mobile robots, integrating sonar, vision, and dead-reckoning [<xref ref-type="bibr" rid="ref-42">42</xref>]. Alag, Goebel, and Agogino (1995) proposed a supervisory fusion framework for Intelligent Vehicle Highway Systems (IVHS), emphasizing fault detection, redundancy, and supervisory control [<xref ref-type="bibr" rid="ref-41">41</xref>]. Beyond robotics, Mandenius et al. (1997) applied fusion in industrial bioprocessing, combining chemical and process sensors for real-time monitoring [<xref ref-type="bibr" rid="ref-45">45</xref>]. These architectures highlighted the feasibility of embedding fusion into control pipelines, but remained tightly coupled to specific sensor suites, limiting scalability and cross-domain applicability.</p>
<p>Overall, the 1990s advanced sensor fusion by formalizing uncertainty modeling, exploring adaptive neural approaches, and embedding fusion into practical systems. As summarized in <xref ref-type="table" rid="table-5">Table 5</xref>, most methods remained constrained by high computational demands, narrow scope, and lack of real-time generalizability. Yet, they established enduring design principles&#x2014;probabilistic rigor, adaptive learning, interpretable reasoning, and modular integration&#x2014;that continue to influence sensor fusion research today.</p>
<table-wrap id="table-5">
<label>Table 5</label>
<caption>
<title>Representative sensor fusion works (1991&#x2013;2000)</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/> </colgroup>
<thead>
<tr>
<th>Reference (Year)</th>
<th>Fusion strategy</th>
<th>Application domain</th>
<th>Representative contribution</th>
<th>Limitations</th>
</tr>
</thead>
<tbody>
<tr>
<td>Perlovsky &#x0026; McManus (1991) [<xref ref-type="bibr" rid="ref-36">36</xref>]</td>
<td>Neural network (maximum likelihood)</td>
<td>Classification (general)</td>
<td>Hybrid neural-probabilistic model enabling adaptive classification of sensor inputs</td>
<td>Computationally expensive; shallow network with limited scalability</td>
</tr>
<tr>
<td>Cox et al. (1992) [<xref ref-type="bibr" rid="ref-33">33</xref>]</td>
<td>Bayesian inference</td>
<td>Computer vision (stereo)</td>
<td>Bayesian fusion improved stereo depth estimation by integrating multiple cues</td>
<td>Assumed static scenes; not scalable to large sensor sets</td>
</tr>
<tr>
<td>Tang &#x0026; Lee (1992) [<xref ref-type="bibr" rid="ref-39">39</xref>]</td>
<td>Geometric feature relation graph</td>
<td>Computer vision (multi-camera)</td>
<td>Ensured spatial consistency in multi-camera fusion through feature graphs</td>
<td>Relied on structured environments; sensitive to occlusion and dynamics</td>
</tr>
<tr>
<td>Shahbazian et al. (1993) [<xref ref-type="bibr" rid="ref-35">35</xref>]</td>
<td>Multi-hypothesis tracking (Bayesian)</td>
<td>Radar/sonar tracking</td>
<td>Maintained competing hypotheses across heterogeneous sensors for robust tracking</td>
<td>Very high computation; lacked real-time demonstration</td>
</tr>
<tr>
<td>McKee (1993) [<xref ref-type="bibr" rid="ref-40">40</xref>]</td>
<td>Conceptual taxonomy</td>
<td>Computer vision (general)</td>
<td>Systematic taxonomy outlining &#x201C;what can be fused&#x201D; in vision pipelines</td>
<td>Conceptual; lacked experimental implementation</td>
</tr>
<tr>
<td>Alag, Goebel &#x0026; Agogino (1995) [<xref ref-type="bibr" rid="ref-41">41</xref>]</td>
<td>Rule-based &#x002B; supervisory control</td>
<td>Intelligent vehicles (IVHS)</td>
<td>Framework for sensor validation and supervisory fusion enhancing safety and redundancy</td>
<td>Knowledge engineering complexity; limited handling of drift and long-term variability</td>
</tr>
<tr>
<td>Davis &#x0026; Stentz (1995) [<xref ref-type="bibr" rid="ref-37">37</xref>]</td>
<td>Neural network fusion</td>
<td>Autonomous navigation</td>
<td>Neural networks fused sensor streams for obstacle recognition in unstructured terrain</td>
<td>Limited training data; interpretability issues</td>
</tr>
<tr>
<td>Kam, Zhu &#x0026; Kalata (1997) [<xref ref-type="bibr" rid="ref-42">42</xref>]</td>
<td>Modular fusion architecture</td>
<td>Mobile robotics</td>
<td>Integrated sonar, vision, and dead-reckoning in a cohesive multi-sensor framework</td>
<td>Lacked principled uncertainty modeling; performance tuned to specific sensors</td>
</tr>
<tr>
<td>Cao &#x0026; Hall (1998) [<xref ref-type="bibr" rid="ref-38">38</xref>]</td>
<td>Neural networks</td>
<td>Autonomous guided vehicles</td>
<td>Neural fusion applied to visual navigation in AGVs; early application to vehicles</td>
<td>Black-box model; limited generalization beyond controlled tests</td>
</tr>
<tr>
<td>Larkin (1998) [<xref ref-type="bibr" rid="ref-34">34</xref>]</td>
<td>Bayesian networks</td>
<td>Acoustic classification</td>
<td>Bayesian network captured dependencies in acoustic feature classification</td>
<td>Computationally heavy; not demonstrated in real-time systems</td>
</tr>
<tr>
<td>Blasch &#x0026; Hong (1999) [<xref ref-type="bibr" rid="ref-43">43</xref>]</td>
<td>Dempster&#x2013;Shafer (belief filtering)</td>
<td>Target tracking</td>
<td>Applied evidence theory for uncertainty reasoning in multi-sensor target tracking</td>
<td>Performance degraded with ambiguous/conflicting evidence</td>
</tr>
<tr>
<td>Peli et al. (1999) [<xref ref-type="bibr" rid="ref-44">44</xref>]</td>
<td>Feature-level fusion</td>
<td>Multisensor vision</td>
<td>Unified features across sensors before classification, improving recognition performance</td>
<td>Dependent on calibration; fragile under sensor faults</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>By the end of the decade, sensor fusion had matured from theoretical constructs to operational prototypes in vision, robotics, defense, and industrial applications. However, most systems were specialized, computationally demanding, and lacked generalizable frameworks. The subsequent decade (2000&#x2013;2010) witnessed increasing convergence and the emergence of machine learning as a unifying tool for sensor fusion across domains.</p>
</sec>
<sec id="s4_3">
<label>4.3</label>
<title>Contemporary Models: 2001&#x2013;2010&#x2014;Early Machine Learning Era</title>
<p>Building upon the foundational probabilistic, neural, and rule-based strategies of the 1990s, sensor fusion research in the 2000s advanced toward greater methodological rigor and broader applicability. Several developments characterized this decade: the extension of probabilistic filters for nonlinear and non-Gaussian systems, the integration of Bayesian and evidence-theoretic reasoning, the rise of distributed consensus schemes for multi-agent settings, and the early adoption of machine learning techniques to learn fusion mappings from data rather than relying solely on handcrafted rules. Collectively, these advances established many of the algorithmic templates that would later be scaled and generalized in the deep learning era.</p>
<p><italic>Probabilistic Filtering and Bayesian Extensions:</italic> Kalman filter variants dominated this period, particularly in navigation and tracking tasks. The Extended Kalman Filter (EKF) and Unscented Kalman Filter (UKF) were widely adopted for fusing inertial measurement units (IMUs) with GPS data, enabling more reliable state estimation under nonlinear dynamics [<xref ref-type="bibr" rid="ref-46">46</xref>]. Cou&#x00E9; et al. (2002) applied Bayesian programming to automotive state estimation, demonstrating how Bayes filters could flexibly combine sonar, odometry, and other modalities under uncertainty [<xref ref-type="bibr" rid="ref-47">47</xref>]. Particle filters emerged as an important alternative, addressing limitations of Kalman-based methods by accommodating non-Gaussian noise and multimodal posterior distributions. These probabilistic methods significantly improved robustness in early autonomous robots and vehicles, although they remained computationally demanding for high-dimensional state spaces.</p>
<p><italic>Rule-Based Evolution and Evidence-Theoretic Integration:</italic> Rule-based fusion was refined into more mathematically grounded frameworks. Koks and Challa (2003) proposed combining Bayesian methods with Dempster&#x2013;Shafer (D&#x2013;S) evidence theory, providing a hybrid reasoning scheme capable of fusing probabilistic estimates with uncertain or incomplete evidence [<xref ref-type="bibr" rid="ref-48">48</xref>]. This integration allowed richer representations of belief states but incurred high computational cost as sensor sets scaled. Distributed consensus algorithms also gained prominence in this era. Xiao et al. (2005) introduced a consensus-based scheme that enabled sensor networks or multi-robot systems to achieve agreement on global state estimates despite each node holding only partial information [<xref ref-type="bibr" rid="ref-49">49</xref>]. These approaches were particularly important for wireless sensor networks (WSNs), where centralized fusion was often infeasible.</p>
<p><italic>Machine Learning for Fusion Mappings:</italic> A key innovation was the move toward learning fusion rules directly from data. Faceli et al. (2004) proposed a hybrid intelligent framework combining neural networks, fuzzy inference, and decision trees, allowing the system to adaptively determine fusion weights and mappings [<xref ref-type="bibr" rid="ref-50">50</xref>]. While deep learning was not yet viable, simpler models such as multilayer perceptrons and fuzzy systems demonstrated the feasibility of training adaptive fusion models. These early machine learning&#x2013;driven systems reduced reliance on handcrafted rules, though their learning capacity was limited by data availability and computational constraints.</p>
<p><italic>Decision-Level and Classifier Fusion:</italic> Decision-level fusion became increasingly popular for classification tasks. Instead of integrating raw signals, systems combined outputs of independent classifiers trained on individual sensor modalities. For instance, in wearable human activity recognition (HAR), classifiers based on accelerometers and gyroscopes could be fused via voting or weighted averaging to yield more robust predictions. This ensemble approach improved resilience to sensor failures and noise. While Chavez-Garcia and Aycard (2015) [<xref ref-type="bibr" rid="ref-51">51</xref>] formally studied multisensor decision fusion slightly after 2010, their work synthesized principles already established in the late 2000s, particularly in intelligent vehicle perception.</p>
<p><italic>Application-Specific Advances:</italic> The decade also saw growing application diversity. Choi et al. (2011) applied hierarchical fusion of RFID and odometry for indoor robot localization, building on techniques developed in the late 2000s [<xref ref-type="bibr" rid="ref-52">52</xref>]. Lu and Michaels (2009) fused ultrasonic sensor data for structural health monitoring under varying conditions, addressing robustness challenges in safety-critical applications [<xref ref-type="bibr" rid="ref-53">53</xref>]. In agriculture, Huang et al. (2007) integrated multiple sensing modalities for precision farming, reflecting the growing role of sensor fusion in environmental and industrial domains [<xref ref-type="bibr" rid="ref-54">54</xref>]. Each domain imposed distinct requirements&#x2014;low power consumption for wearable devices, high accuracy for aircraft navigation, or resilience to noise and environmental variation for outdoor robotics&#x2014;driving tailored fusion solutions.</p>
<p>The representative works summarized in <xref ref-type="table" rid="table-6">Table 6</xref> illustrate how this decade broadened the methodological toolkit. Unlike the 1990s, where most systems were rigidly rule-based or narrowly probabilistic, the 2000s emphasized flexibility through probabilistic generalizations, distributed consensus, and adaptive machine learning. While these advances greatly expanded the scope of sensor fusion, limitations remained, particularly in computational scalability, dependence on expert tuning, and restricted ability to automatically learn complex feature hierarchies. These constraints would soon motivate the adoption of deep learning approaches in the following decade.</p>
<table-wrap id="table-6">
<label>Table 6</label>
<caption>
<title>Representative sensor fusion studies (2001&#x2013;2010)</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/> </colgroup>
<thead>
<tr>
<th>Reference (Year)</th>
<th>Fusion strategy</th>
<th>Sensor modalities</th>
<th>Application domain</th>
<th>Representative contribution</th>
<th>Limitations</th>
</tr>
</thead>
<tbody>
<tr>
<td>Cou&#x00E9; et al. (2002) [<xref ref-type="bibr" rid="ref-47">47</xref>]</td>
<td>Bayesian programming</td>
<td>Sonar, odometry, automotive sensors</td>
<td>Automotive state estimation</td>
<td>Demonstrated Bayesian filters for robust vehicle state estimation under uncertainty</td>
<td>Computationally intensive; limited real-time deployment</td>
</tr>
<tr>
<td>Koks &#x0026; Challa (2003) [<xref ref-type="bibr" rid="ref-48">48</xref>]</td>
<td>Bayesian &#x002B; Dempster&#x2013;Shafer integration</td>
<td>Heterogeneous multi-sensor inputs</td>
<td>General multi-sensor fusion</td>
<td>Hybridized probabilistic and evidence-based reasoning for uncertainty management</td>
<td>High computational complexity; scalability issues</td>
</tr>
<tr>
<td>Faceli et al. (2004) [<xref ref-type="bibr" rid="ref-50">50</xref>]</td>
<td>Hybrid AI (NNs &#x002B; fuzzy logic &#x002B; decision trees)</td>
<td>Simulated sensors</td>
<td>General sensor fusion</td>
<td>Introduced adaptive ensemble-based fusion learned from data</td>
<td>Limited by shallow models and lack of real-world datasets</td>
</tr>
<tr>
<td>Tan (2004) [<xref ref-type="bibr" rid="ref-55">55</xref>]</td>
<td>Cognitive architecture (FALCON)</td>
<td>Simulated neural agents</td>
<td>Cognitive /adaptive systems</td>
<td>Demonstrated adaptive learning in fusion within cognitive architectures</td>
<td>Proof-of-concept; evaluated only on simple tasks</td>
</tr>
<tr>
<td>Xiao et al. (2005) [<xref ref-type="bibr" rid="ref-49">49</xref>]</td>
<td>Distributed consensus</td>
<td>Networked sensors (WSN)</td>
<td>Wireless sensor networks/multi-robot systems</td>
<td>Consensus algorithm enabling distributed nodes to agree on state estimates</td>
<td>Dependent on reliable communication links; latency under large networks</td>
</tr>
<tr>
<td>Upcroft et al. (2005) [<xref ref-type="bibr" rid="ref-56">56</xref>]</td>
<td>Decentralized probabilistic fusion</td>
<td>Acoustic, radar (UAV sensors)</td>
<td>UAV perception</td>
<td>Applied probabilistic fusion for UAV navigation with heterogeneous sensors</td>
<td>Assumed reliable inter-agent data exchange; computationally heavy</td>
</tr>
<tr>
<td>Huang et al. (2007) [<xref ref-type="bibr" rid="ref-54">54</xref>]</td>
<td>Feature/data-level fusion</td>
<td>Soil, crop, agricultural sensors</td>
<td>Precision farming</td>
<td>Improved agricultural decision-making via multi-sensor integration</td>
<td>Sensitive to calibration; limited to structured farming conditions</td>
</tr>
<tr>
<td>Klausner et al. (2007) [<xref ref-type="bibr" rid="ref-57">57</xref>]</td>
<td>Feature &#x002B; decision fusion</td>
<td>Smart cameras</td>
<td>Intelligent traffic systems</td>
<td>Demonstrated distributed vehicle classification with fused smart-camera inputs</td>
<td>Required reliable inter-camera synchronization; narrow domain</td>
</tr>
<tr>
<td>Seraji &#x0026; Serrano (2009) [<xref ref-type="bibr" rid="ref-58">58</xref>]</td>
<td>Rule-based decision fusion</td>
<td>Terrain safety sensors</td>
<td>Planetary rover navigation</td>
<td>Combined multiple terrain-safety detectors for robust rover navigation</td>
<td>Rule-based; lacked adaptability to unforeseen conditions</td>
</tr>
<tr>
<td>Lu &#x0026; Michaels (2009) [<xref ref-type="bibr" rid="ref-53">53</xref>]</td>
<td>Feature-level fusion under environment variability</td>
<td>Ultrasonic sensors</td>
<td>Structural health monitoring</td>
<td>Developed fusion for damage detection robust to environmental changes</td>
<td>Limited sensor types; application-specific framework</td>
</tr>
<tr>
<td>Gross et al. (2010) [<xref ref-type="bibr" rid="ref-59">59</xref>]</td>
<td>Comparative filtering (EKF, UKF, PF)</td>
<td>GPS, INS</td>
<td>Navigation systems</td>
<td>Provided comparative evaluation of probabilistic filters under navigation uncertainty</td>
<td>Trade-offs: particle filters accurate but computationally costly; EKF efficient but less robust</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>In summary, the early 2000s represented an era of methodological consolidation and gradual transition from handcrafted fusion rules toward data-driven adaptability. Probabilistic frameworks were extended to handle nonlinearities and non-Gaussian noise, distributed consensus schemes emerged for networked systems, and hybrid AI methods showcased the potential of learned fusion. While the computational and data limitations of the period constrained progress, this decade equipped the field with versatile building blocks&#x2014;Kalman filter variants, Bayesian/evidence hybrids, consensus protocols, and ensemble learning&#x2014;that directly informed the deep learning&#x2013;driven breakthroughs of the 2010s. These advances thus represent the logical evolution of the 1990s prototypes into more flexible, scalable, and domain-diverse fusion frameworks.</p>
</sec>
<sec id="s4_4">
<label>4.4</label>
<title>Contemporary Models: 2011&#x2013;2020&#x2014;Transformative Fusion Works</title>
<p>Building upon the probabilistic, rule-based, and early machine learning approaches of the 2000s, the period from 2011 to 2020 marked a decisive transformation in sensor fusion. This shift was driven by three converging factors: the availability of large-scale multimodal datasets, rapid advances in deep learning, and growing deployment of autonomous systems in safety-critical contexts. Fusion models moved from handcrafted pipelines and shallow learners toward end-to-end trainable architectures capable of learning cross-modal representations directly from data. Research during this decade spanned autonomous vehicles, UAVs, precision agriculture, infrastructure monitoring, and wearable human activity recognition, demonstrating both methodological diversity and domain-specific innovation.</p>
<p><italic>Deep Learning&#x2013;Based Fusion Architectures:</italic> One of the most transformative advances was the adoption of deep neural networks for multi-sensor fusion. In autonomous driving, vision and LiDAR fusion evolved from late fusion of independent detections to early and mid-level feature fusion within deep networks. Approaches such as PointNet&#x002B;&#x002B; and multimodal convolutional fusion architectures enabled learned feature representations across modalities, significantly improving detection and localization accuracy [<xref ref-type="bibr" rid="ref-60">60</xref>,<xref ref-type="bibr" rid="ref-61">61</xref>]. Unlike handcrafted pipelines, these models could discover optimal cross-modal mappings, albeit at the cost of requiring large annotated datasets and high computational resources.</p>
<p><italic>Distributed and Cooperative Fusion:</italic> Another major development was the emergence of cooperative and distributed fusion frameworks, especially for connected autonomous vehicles and IoT-driven systems. Cooperative perception (V2X) allowed vehicles to exchange sensor data, extending situational awareness beyond line-of-sight occlusions. Liu et al. (2023) [<xref ref-type="bibr" rid="ref-62">62</xref>] reviewed this paradigm, which was conceptually established in the late 2010s through simulation-based studies. These works emphasized the need for synchronization, low-latency communication, and consensus protocols, anticipating real-world multi-agent fusion systems.</p>
<p><italic>Domain Diversification and Application-Specific Fusion:</italic> Fusion research extended into healthcare, agriculture, and infrastructure. In wearable HAR, Banos et al. (2012) combined accelerometer, gyroscope, and contextual sensors to mitigate noise sensitivity and improve recognition reliability [<xref ref-type="bibr" rid="ref-63">63</xref>]. In precision agriculture, Maimaitijiang et al. (2020) fused UAV imagery, satellite data, and ground-based sensors using machine learning for crop monitoring, enabling multiscale environmental insights [<xref ref-type="bibr" rid="ref-64">64</xref>]. Infrastructure monitoring adopted sensor fusion of accelerometers, strain gauges, and vibration sensors to detect anomalies in bridges and civil structures. These application-specific systems demonstrated that the core fusion principles&#x2014;robust uncertainty handling, redundancy, and adaptive learning&#x2014;could generalize across domains.</p>
<p><italic>Reliability, Explainability, and Adversarial Concerns:</italic> By the late 2010s, researchers recognized that fusion systems in safety-critical domains required not only empirical accuracy but also transparency and robustness. Explainable AI (XAI) techniques were explored to interpret multimodal fusion decisions, particularly in healthcare and autonomous driving. Simultaneously, adversarial studies revealed vulnerabilities, such as perturbations or physical artifacts that could mislead fused perception systems. This highlighted the need for redundancy-driven architectures, formal verification of fusion pipelines, and design of fail-operational strategies for safety-critical deployments.</p>
<p>Representative studies from this decade are summarized in <xref ref-type="table" rid="table-7">Table 7</xref>, which highlights the methodologies, fusion granularity, application domains, and technical contributions.</p>
<table-wrap id="table-7">
<label>Table 7</label>
<caption>
<title>Representative sensor fusion studies (2011&#x2013;2020)</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/> </colgroup>
<thead>
<tr>
<th>Reference (Year)</th>
<th>Fusion methodology</th>
<th>Fusion granularity</th>
<th>Application domain</th>
<th>Representative contribution</th>
</tr>
</thead>
<tbody>
<tr>
<td>Choi et al. (2011) [<xref ref-type="bibr" rid="ref-52">52</xref>]</td>
<td>Hierarchical fusion (RFID &#x002B; odometry)</td>
<td>Feature-level</td>
<td>Indoor mobile robots</td>
<td>Improved indoor localization by combining absolute and relative sensors</td>
</tr>
<tr>
<td>Banos et al. (2012) [<xref ref-type="bibr" rid="ref-63">63</xref>]</td>
<td>Multisensor fusion for noise reduction</td>
<td>Feature-level</td>
<td>Human Activity Recognition</td>
<td>Enhanced HAR robustness by integrating inertial and contextual signals</td>
</tr>
<tr>
<td>Fagiano et al. (2013) [<xref ref-type="bibr" rid="ref-65">65</xref>]</td>
<td>Kalman filter extensions</td>
<td>Data-level</td>
<td>Airborne wind energy systems</td>
<td>Real-time estimation of wind states using fused sensor streams</td>
</tr>
<tr>
<td>Chavez-Garcia &#x0026; Aycard (2015) [<xref ref-type="bibr" rid="ref-51">51</xref>]</td>
<td>Ensemble of classifiers</td>
<td>Decision/feature-level</td>
<td>Autonomous driving</td>
<td>Combined multiple classifiers for robust object detection and tracking</td>
</tr>
<tr>
<td>Chen et al. (2016) [<xref ref-type="bibr" rid="ref-66">66</xref>]</td>
<td>Multimodal sensor network with custom fusion hardware</td>
<td>Feature-level</td>
<td>Road surface monitoring</td>
<td>Low-cost, multimodal fusion for pothole detection</td>
</tr>
<tr>
<td>Guo et al. (2017) [<xref ref-type="bibr" rid="ref-67">67</xref>]</td>
<td>Fault-tolerant fusion scheme</td>
<td>Data-level</td>
<td>UAV navigation</td>
<td>Airspeed sensor fault detection through redundancy and fusion</td>
</tr>
<tr>
<td>Tsinganos &#x0026; Skodras (2018) [<xref ref-type="bibr" rid="ref-68">68</xref>]</td>
<td>Classifier fusion (comparative study)</td>
<td>Decision-level</td>
<td>Wearable fall detection</td>
<td>Empirical comparison of sensor-specific vs. fused classifiers</td>
</tr>
<tr>
<td>Barbier et al. (2019) [<xref ref-type="bibr" rid="ref-69">69</xref>]</td>
<td>Statistical model checking</td>
<td>Decision-level</td>
<td>Autonomous driving</td>
<td>Validation of fused decision outputs against formal safety criteria</td>
</tr>
<tr>
<td>Maimaitijiang et al. (2020) [<xref ref-type="bibr" rid="ref-64">64</xref>]</td>
<td>Multiscale data fusion with ML</td>
<td>Feature-level</td>
<td>Precision agriculture</td>
<td>Crop monitoring by integrating UAV, satellite, and ground sensors</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>To complement these representative studies, <xref ref-type="table" rid="table-8">Table 8</xref> presents a focused technical comparison of widely used classical approaches&#x2014;Extended Kalman Filters (EKF), Particle Filters (PF), Dempster&#x2013;Shafer (D&#x2013;S) theory, and Bayesian inference&#x2014;under challenging conditions such as non-Gaussian noise and conflicting sensor evidence. This highlights the continued relevance of classical filters alongside modern learning-based approaches.</p>
<table-wrap id="table-8">
<label>Table 8</label>
<caption>
<title>Comparative analysis of classical fusion approaches under challenging conditions</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col align="center"/>
<col align="center"/> </colgroup>
<thead>
<tr>
<th>Aspect</th>
<th>EKF vs. Particle Filter (PF) in non-Gaussian noise</th>
<th>D&#x2013;S Theory vs. Bayesian Inference under sensor conflict</th>
</tr>
</thead>
<tbody>
<tr>
<td><bold>Model assumptions</bold></td>
<td>EKF assumes Gaussian noise and local linearization; diverges in multimodal distributions. PF models arbitrary distributions through sampling.</td>
<td>Bayesian inference requires priors; D&#x2013;S assigns belief without precise priors.</td>
</tr>
<tr>
<td><bold>Noise handling</bold></td>
<td>PF accommodates heavy-tailed and multimodal noise; EKF fragile under outliers.</td>
<td>Bayesian inference may overweight conflicting likelihoods; D&#x2013;S can represent ignorance explicitly.</td>
</tr>
<tr>
<td><bold>Robustness</bold></td>
<td>PF robust under strong nonlinearities if sufficient particles are used; EKF brittle under nonlinearity.</td>
<td>D&#x2013;S maintains robustness under conflicting evidence; Bayesian updates may yield counterintuitive posteriors.</td>
</tr>
<tr>
<td><bold>Computation</bold></td>
<td>EKF computationally efficient; PF scales poorly with high-dimensional states.</td>
<td>Bayesian inference efficient in structured models; D&#x2013;S expensive in large frames of discernment.</td>
</tr>
<tr>
<td><bold>Uncertainty representation</bold></td>
<td>EKF outputs covariance estimates; PF yields posterior distributions (richer).</td>
<td>D&#x2013;S distinguishes belief, plausibility, and ignorance; Bayesian yields single posterior.</td>
</tr>
<tr>
<td><bold>Best-Suited applications</bold></td>
<td>EKF: real-time navigation (SLAM, GPS&#x2013;INS) under Gaussian noise. PF: UAV/UGV localization in cluttered, uncertain environments.</td>
<td>D&#x2013;S: heterogeneous, fault-tolerant fusion (e.g., radar&#x2013;camera&#x2013;LiDAR). Bayesian: structured domains with reliable priors (e.g., GNSS&#x2013;IMU integration).</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>In summary, the 2011&#x2013;2020 decade marked the transition from handcrafted, model-driven fusion toward data-driven and learned fusion paradigms. Deep learning architectures enabled joint feature representations across heterogeneous modalities, cooperative fusion expanded to multi-agent systems, and application domains diversified beyond vehicles and robots to healthcare, agriculture, and infrastructure. Despite these advances, classical methods such as Kalman filtering and Bayesian inference remained essential, particularly in constrained environments or where formal guarantees were required. The coexistence of classical and AI-driven approaches underscores the versatility of sensor fusion, while ongoing challenges in scalability, robustness, and verifiability continue to motivate research in the current decade.</p>
</sec>
<sec id="s4_5">
<label>4.5</label>
<title>Recent Advances: 2021&#x2013;2025&#x2014;Toward Robust and Scalable Fusion</title>
<p>Extending the deep learning&#x2013;driven breakthroughs of the 2010s, sensor fusion research from 2021 onward has accelerated toward tackling real-world deployment challenges. Models are no longer expected to perform well only in controlled datasets but must generalize across environments, sensor suites, and tasks with minimal reconfiguration. This decade has also been marked by the emergence of transformer-based architectures, context-aware dynamic fusion, and practical demonstrations of cooperative perception in multi-agent systems. Fusion has become increasingly pervasive, appearing in domains as varied as autonomous firefighting robots, UAV-based wildlife monitoring, intelligent transportation infrastructure, and healthcare wearables.</p>
<p><italic>Generalizability and Cross-Domain Adaptation:</italic> A central focus of this era is improving the robustness and scalability of fusion models. Systems trained on one platform (e.g., a specific vehicle type or city) are increasingly adapted to new conditions with minimal retraining, using transfer learning, domain adaptation, and synthetic-to-real approaches. High-fidelity simulators are leveraged to generate rare or safety-critical scenarios, with adaptation methods ensuring real-world applicability. Physics-informed neural networks emerged as a hybrid approach, embedding sensor physics into learning pipelines to reduce data requirements and enforce physical consistency.</p>
<p><italic>Transformer-Based and Attention Mechanisms:</italic> Transformers and attention-based architectures became central to fusion pipelines. Chitta et al. (2022) proposed TransFuser, a transformer-based model that jointly encodes LiDAR and camera streams for autonomous driving [<xref ref-type="bibr" rid="ref-70">70</xref>]. These architectures enable multi-task and cross-modal learning, allowing a single network to perform detection, segmentation, and tracking simultaneously. However, they remain computationally heavy and require large training datasets. HydraFusion (Malawade et al., 2022) extended this by incorporating attention-driven context selection, dynamically weighting sensors depending on environmental conditions [<xref ref-type="bibr" rid="ref-71">71</xref>]. Such adaptive mechanisms improve resilience but increase training complexity.</p>
<p><italic>Edge&#x2013;Cloud Hybrid Fusion Architectures:</italic> The push toward deployment in connected and resource-constrained environments led to hybrid strategies. Edge devices handle low-latency, safety-critical decisions (e.g., obstacle avoidance), while cloud or roadside servers manage computationally intensive tasks such as global route planning or cooperative perception. This split addresses both responsiveness and scalability, though it introduces latency-management and bandwidth-allocation challenges.</p>
<p><italic>Self-Calibration and Fault Tolerance:</italic> Autonomous systems now integrate self-diagnostic routines to detect and respond to sensor degradation (e.g., blocked LiDARs, degraded cameras). Multi-sensor redundancy allows systems to isolate and exclude faulty sensors or re-calibrate them dynamically. Tommingas et al. (2025) demonstrated fusion of UWB and GNSS with ML-based uncertainty modeling, highlighting the need for adaptable frameworks capable of self-healing in diverse environments [<xref ref-type="bibr" rid="ref-72">72</xref>].</p>
<p>In the domain of robust navigation, a GNSS/IMU/VO fusion framework with multipath inflation factor has been proposed to explicitly mitigate the challenges of urban multipath interference. By leveraging real-time IMU and VO inputs, the system dynamically adjusts GNSS weighting and adaptively updates VO velocity variance within a robust extended Kalman filter. Field tests in dense urban areas demonstrated 63.4% and 56.1% improvements in horizontal and 3D positioning accuracy, respectively, over conventional fusion schemes [<xref ref-type="bibr" rid="ref-73">73</xref>]. This work highlights the importance of incorporating environment-aware weighting models for next-generation positioning, navigation, and timing (PNT) systems. Beyond terrestrial navigation, recent work has demonstrated the value of multi-sensor association for high-precision space target localization. By fusing visible light and infrared detection with laser ranging under a Gaussian mixture TPHD framework, this approach achieves great accuracy, outperforming binary star angular-only methods [<xref ref-type="bibr" rid="ref-74">74</xref>]. This highlights how sensor fusion enables unprecedented precision in space situational awareness and orbital tracking.</p>
<p><italic>Diversified Applications:</italic> Healthcare, smart cities, and environmental monitoring benefited significantly. Rashid et al. (2023) developed SELF-CARE, a wearable fusion framework for stress detection, combining multiple biosignals with context identification [<xref ref-type="bibr" rid="ref-75">75</xref>]. Hasanujjaman et al. (2023) fused autonomous vehicle and CCTV camera data for smart traffic management [<xref ref-type="bibr" rid="ref-76">76</xref>]. In addition, advances in embedded ultra-precision sensing have expanded the scope of sensor fusion to metrology and industrial domains. A recent study introduced a fiber microprobe interference-based displacement measurement system capable of measuring ranges up to 700 mm with subnanometer accuracy. Unlike conventional interferometers, this approach enables compact, embedded measurements in confined spaces, supporting applications in high-end equipment manufacturing and biomedical robotics [<xref ref-type="bibr" rid="ref-77">77</xref>]. Aguilar-Lazcano et al. (2023) surveyed sensor fusion in wildlife monitoring, highlighting challenges of sparse data and field deployment [<xref ref-type="bibr" rid="ref-78">78</xref>]. These illustrate how the principles of redundancy, adaptability, and interpretability are increasingly tailored to domain-specific constraints. In intelligent transportation and scene understanding, multi-modal remote perception learning frameworks have been introduced to integrate object detection with contextual scene semantics. For example, a Deep Fused Network (DFN) combines multi-object detection and semantic analysis, yielding improvement on SUN-RGB-D and on NYU-Dv2 compared to existing approaches [<xref ref-type="bibr" rid="ref-79">79</xref>]. These results underline the growing role of context-aware multimodal fusion for complex environments in autonomous driving and robotics. Industrial monitoring and predictive maintenance are also benefiting from self-supervised representation learning. A recently proposed multihead attention self-supervised (MAS) model learns robust features from multidimensional industrial sensor data using contrastive augmentation strategies. Applied to a real-world water circulation system, MAS improved anomaly detection performance without reliance on large labeled datasets [<xref ref-type="bibr" rid="ref-80">80</xref>]. Such approaches demonstrate the promise of representation learning in industrial sensor fusion for fault detection and equipment health monitoring.</p>
<p>Representative works from this period are summarized in <xref ref-type="table" rid="table-9">Table 9</xref>, capturing the methodologies, AI/ML integration, fusion granularity, application domains, and limitations.</p>
<table-wrap id="table-9">
<label>Table 9</label>
<caption>
<title>Representative sensor fusion studies (2021&#x2013;2025)</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/> </colgroup>
<thead>
<tr>
<th>Reference (Year)</th>
<th>Fusion strategy</th>
<th>AI/ML integration</th>
<th>Fusion granularity</th>
<th>Application domain</th>
<th>Limitations</th>
</tr>
</thead>
<tbody>
<tr>
<td>Malawade et al. (2022) [<xref ref-type="bibr" rid="ref-71">71</xref>]</td>
<td>Context-aware selective fusion</td>
<td>Attention-based deep learning</td>
<td>Feature /Decision</td>
<td>Autonomous driving</td>
<td>Training complexity; context detector tuning</td>
</tr>
<tr>
<td>Zhang et al. (2022) [<xref ref-type="bibr" rid="ref-81">81</xref>]</td>
<td>Hybrid rule-based &#x002B; ML</td>
<td>Rules trigger ML control</td>
<td>Feature-level</td>
<td>Robotics (firefighting)</td>
<td>Limited adaptability; false alarm sensitivity</td>
</tr>
<tr>
<td>Chitta et al. (2022) [<xref ref-type="bibr" rid="ref-70">70</xref>]</td>
<td>Transformer-based fusion</td>
<td>Transformer networks</td>
<td>Deep feature fusion</td>
<td>Autonomous driving</td>
<td>Extremely data- and compute-intensive</td>
</tr>
<tr>
<td>Xiang et al. (2023) [<xref ref-type="bibr" rid="ref-82">82</xref>]</td>
<td>Cooperative multi-agent fusion</td>
<td>N/A (survey)</td>
<td>Multi-agent</td>
<td>Connected vehicles</td>
<td>Lack of datasets; absence of unified benchmarks</td>
</tr>
<tr>
<td>Ignatious et al. (2023) [<xref ref-type="bibr" rid="ref-83">83</xref>]</td>
<td>Multi-level fusion pipeline</td>
<td>CNN-based detection &#x002B; rule logic</td>
<td>Sensor /Decision</td>
<td>Autonomous driving</td>
<td>Static fusion strategy; limited adaptability</td>
</tr>
<tr>
<td>Rashid et al. (2023) [<xref ref-type="bibr" rid="ref-75">75</xref>]</td>
<td>Context-aware wearable fusion (SELF-CARE)</td>
<td>Ensemble models &#x002B; context ID</td>
<td>Feature-level</td>
<td>Healthcare (stress detection)</td>
<td>Requires personalized calibration; context-labeling overhead</td>
</tr>
<tr>
<td>Hasanujjaman et al. (2023) [<xref ref-type="bibr" rid="ref-76">76</xref>]</td>
<td>AV&#x2013;CCTV multi-source fusion</td>
<td>Deep learning</td>
<td>Feature-level</td>
<td>Smart city traffic</td>
<td>Bandwidth overhead; privacy risks with video data</td>
</tr>
<tr>
<td>Aguilar-Lazcano et al. (2023) [<xref ref-type="bibr" rid="ref-78">78</xref>]</td>
<td>ML-based survey of sensor fusion</td>
<td>N/A (review)</td>
<td>Feature /Decision (review)</td>
<td>Wildlife monitoring</td>
<td>Sparse datasets; limited annotated field data</td>
</tr>
<tr>
<td>Liu et al. (2024) [<xref ref-type="bibr" rid="ref-84">84</xref>]</td>
<td>Bird&#x2019;s-eye-view (BEV) multi-task fusion</td>
<td>Transformer-based ML</td>
<td>Multi-task/multi-level</td>
<td>Autonomous driving</td>
<td>BEV transformation errors at range; flat-terrain assumption</td>
</tr>
<tr>
<td>Tommingas et al. (2025) [<xref ref-type="bibr" rid="ref-72">72</xref>]</td>
<td>UWB &#x002B; GNSS with ML-based uncertainty</td>
<td>Probabilistic ML</td>
<td>Sensor-level</td>
<td>High-precision positioning</td>
<td>Retraining required for new sensors/environments</td>
</tr>
</tbody>
</table>
</table-wrap>
<p><italic>Key Trends and Challenges:</italic> The works in <xref ref-type="table" rid="table-9">Table 9</xref> reflect several defining directions. Transformer-based models and attention mechanisms (e.g., TransFuser, HydraFusion) dominate high-performance fusion pipelines but remain resource-intensive. Context-aware frameworks demonstrate adaptability but raise challenges in calibration and scalability. Application diversification is notable&#x2014;ranging from autonomous driving to stress detection and wildlife monitoring&#x2014;yet many domains suffer from data scarcity and lack of standardized benchmarks. Cooperative perception moved from conceptual discussions to initial real-world demonstrations, though interoperability and evaluation metrics remain unresolved.</p>

<p>Another important theme is hybridization: combining learning-based adaptability with model-driven rigor. Physics-informed neural networks, domain adaptation, and simulation-based training address limitations of purely data-driven methods. Similarly, hybrid edge&#x2013;cloud fusion architectures balance real-time responsiveness with global situational analysis, though at the cost of latency management and secure communication. Finally, fault tolerance and self-calibration have become indispensable, marking a shift toward self-healing, resilient fusion pipelines capable of long-term deployment.</p>
<p>In summary, the 2021&#x2013;2025 period marks the consolidation of deep learning and transformer-based architectures, the practical emergence of cooperative fusion, and the diversification of sensor fusion into new domains. The emphasis has shifted from achieving accuracy in benchmark datasets to ensuring robustness, scalability, and adaptability in highly dynamic real-world conditions. These trends set the stage for future research on verifiable, resource-efficient, and generalizable sensor fusion frameworks.</p>
</sec>
</sec>
<sec id="s5">
<label>5</label>
<title>Mapping the Hierarchical Integrated Model with Contemporary Fusion Methods</title>
<p>Classical sensor fusion frameworks were remarkably forward-looking, often articulating layered capabilities that exceeded the computational and sensing resources available at their time of conception. These models established a conceptual hierarchy&#x2014;signal acquisition, feature extraction, state estimation, decision-making, and refinement&#x2014;that continues to underpin modern multi-sensor fusion architectures. To assess how contemporary systems align with these expectations, we map representative works in autonomous navigation onto a level-wise framework, spanning three decades of research.</p>
<p>The mapping process involved systematic extraction of the operational pipeline from each selected study. For each work, the sensor inputs, the fusion operations, and the resulting outputs were identified and aligned with a hierarchical integrated model (see <xref ref-type="fig" rid="fig-1">Fig. 1</xref>). In this model, Level 0 corresponds to preprocessing and signal conditioning (e.g., filtering, synchronization, calibration); Level 1 captures per-sensor or object-level estimation; Level 2 concerns scene-level integration (data association, global context, or unified representations); Level 3 involves decision-making and control outputs; and Level 4 corresponds to refinement and adaptivity.</p>

<p>The mapping in <xref ref-type="table" rid="table-10">Table 10</xref> consolidates representative studies across three decades to show how fusion practices in the autonomous navigation domain have progressively aligned with the layered structure of the JDL framework. At Level 0, early works employed handcrafted preprocessing pipelines, while recent methods rely on modality-specific neural encoders for denoising and synchronization. At Level 1, probabilistic inference and classical classifiers gave way to deep architectures such as BEVFusion, which directly learn object-level representations from multimodal inputs. Level 2 has similarly evolved from rule-based association and evidential reasoning toward unified spatial embeddings, such as occupancy grids and bird&#x2019;s-eye view projections, that support multi-task perception.</p>
<table-wrap id="table-10">
<label>Table 10</label>
<caption>
<title>Mapping of JDL fusion levels to AI, ML, and DL-based approaches</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/> </colgroup>
<thead>
<tr>
<th>JDL level</th>
<th>AI-based methods</th>
<th>ML-based methods</th>
<th>Deep learning methods</th>
</tr>
</thead>
<tbody>
<tr>
<td>Level 0&#x2014;Sensors/Preprocessing</td>
<td>Rule-based filtering, calibration [<xref ref-type="bibr" rid="ref-41">41</xref>]</td>
<td>Feature selection, statistical weighting [<xref ref-type="bibr" rid="ref-57">57</xref>]</td>
<td>CNN/autoencoder preprocessing, modality encoders [<xref ref-type="bibr" rid="ref-71">71</xref>]</td>
</tr>
<tr>
<td>Level 1&#x2014;Object/Feature refinement</td>
<td>Bayesian inference, D&#x2013;S reasoning [<xref ref-type="bibr" rid="ref-33">33</xref>,<xref ref-type="bibr" rid="ref-35">35</xref>]</td>
<td>SVMs, decision trees, fuzzy logic, RFID&#x002B;ultrasonic fusion [<xref ref-type="bibr" rid="ref-52">52</xref>]</td>
<td>Camera&#x2013;LiDAR joint detection, Transfuser, BEVFusion [<xref ref-type="bibr" rid="ref-70">70</xref>,<xref ref-type="bibr" rid="ref-84">84</xref>]</td>
</tr>
<tr>
<td>Level 2&#x2014;Situation assessment</td>
<td>Rule-based scene reasoning, evidential fusion [<xref ref-type="bibr" rid="ref-43">43</xref>]</td>
<td>Consensus algorithms in WSN, ensemble tracking [<xref ref-type="bibr" rid="ref-49">49</xref>]</td>
<td>Learned scene embeddings, occupancy grids, BEV(Bird&#x2019;s Eye View) maps [<xref ref-type="bibr" rid="ref-69">69</xref>,<xref ref-type="bibr" rid="ref-84">84</xref>]</td>
</tr>
<tr>
<td>Level 3&#x2014;Threat/Decision Assessment</td>
<td>Supervisory control (IVHS) [<xref ref-type="bibr" rid="ref-41">41</xref>]</td>
<td>Ensemble decision fusion (HAR, fall detection) [<xref ref-type="bibr" rid="ref-68">68</xref>]</td>
<td>End-to-end decision pipelines only (HydraFusion, driving intent), not implemented through deep learning [<xref ref-type="bibr" rid="ref-71">71</xref>]</td>
</tr>
<tr>
<td>Level 4&#x2014;Process refinement and control</td>
<td>Adaptive weighting, supervisory recalibration [<xref ref-type="bibr" rid="ref-47">47</xref>]</td>
<td>Online learning [<xref ref-type="bibr" rid="ref-51">51</xref>]</td>
<td>Rely on Statistical methods for Uncertainty prediction, self-calibration [<xref ref-type="bibr" rid="ref-62">62</xref>,<xref ref-type="bibr" rid="ref-72">72</xref>]</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>In contrast, Levels 3 and 4 remain largely underdeveloped in deep learning pipelines. Whereas classical and machine learning approaches introduced decision-level fusion, supervisory control, and adaptive reliability discounting, contemporary deep networks typically collapse decision-making and process refinement into end-to-end models without explicit reasoning layers. As a result, deep learning systems are effective at perception but do not yet provide interpretable situation assessment or proactive impact evaluation. This diagnostic gap highlights a structural divergence: while empirical accuracy has improved dramatically, the modularity and transparency of classical models have been lost.</p>
<p>Three principal inferences follow from this mapping. First, there is a clear methodological progression: handcrafted features and Bayesian estimators in the 1990s and 2000s gave way to evidential and hierarchical reasoning in the 2010s, and most recently to representation-centric deep fusion pipelines such as HydraFusion and BEVFusion. Second, representational practice has shifted from object-centric and feature-centric fusion toward spatially unified forms that directly support downstream tasks such as detection, segmentation, and planning. Third, uncertainty modeling and validation have re-emerged as central concerns, either through explicit probabilistic frameworks or through hybrid ML&#x2013;classical pipelines where learned uncertainty predictors feed adaptive filters.</p>
<p>A key implication is that while deep learning has advanced perception-oriented levels of fusion, it has not extended the hierarchy upward into situation assessment or impact evaluation. This finding directly motivates the discussion in the following section to reconcile the performance of modern end-to-end fusion with the interpretability and rigor of classical frameworks. For better understanding, a corpus was drawn from autonomous navigation, a domain where multisensor fusion has been both intensively researched and practically deployed. Early works focused on indoor mobile robots and Automated Guided Vehicles, where modular sensor suites and structured environments enabled interpretable designs [<xref ref-type="bibr" rid="ref-38">38</xref>,<xref ref-type="bibr" rid="ref-52">52</xref>]. Over time, emphasis shifted toward high-speed, safety-critical vehicular contexts requiring robustness to adverse weather, dynamic traffic, and uncertain environments. Representative works include [<xref ref-type="bibr" rid="ref-47">47</xref>,<xref ref-type="bibr" rid="ref-51">51</xref>,<xref ref-type="bibr" rid="ref-69">69</xref>,<xref ref-type="bibr" rid="ref-71">71</xref>,<xref ref-type="bibr" rid="ref-84">84</xref>]. These works exemplify the progressive alignment of practical implementations with the layered classical models. The level-wise mapping is summarized in <xref ref-type="table" rid="table-11">Table 11</xref>.</p>
<table-wrap id="table-11">
<label>Table 11</label>
<caption>
<title>Evolution of layer-wise fusion in the autonomous navigation domain</title>
</caption>
<table>
<colgroup>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/> </colgroup>
<tbody>
<tr>
<td><bold>Paper (Year)</bold></td>
<td><bold>Level 0&#x2014;Sensors /Preprocessing</bold></td>
<td><bold>Level 1&#x2014;Object/Feature refinement</bold></td>
<td><bold>Level 2&#x2014;Situation assessment</bold></td>
<td><bold>Level 3&#x2014;Threat/Decision assessment</bold></td>
<td><bold>Level 4&#x2014;Process refinement and control</bold></td>
</tr>
<tr>
<td>Cao &#x0026; Hall (1998) [<xref ref-type="bibr" rid="ref-38">38</xref>]</td>
<td>Sonar <inline-formula id="ieqn-2"><mml:math id="mml-ieqn-2"><mml:mo stretchy="false">&#x2192;</mml:mo></mml:math></inline-formula> distances; camera <inline-formula id="ieqn-3"><mml:math id="mml-ieqn-3"><mml:mo stretchy="false">&#x2192;</mml:mo></mml:math></inline-formula> centroids; encoder <inline-formula id="ieqn-4"><mml:math id="mml-ieqn-4"><mml:mo stretchy="false">&#x2192;</mml:mo><mml:mi mathvariant="normal">&#x0394;</mml:mi><mml:mi>S</mml:mi><mml:mo>,</mml:mo><mml:mi mathvariant="normal">&#x0394;</mml:mi><mml:mi>&#x03B8;</mml:mi></mml:math></inline-formula> (with noise model).</td>
<td>Concatenated state vector from multiple sensors</td>
<td>&#x2014;</td>
<td>Neural network outputs steering and wheel velocities</td>
<td>No explicit uncertainty modeling or adaptive refinement.</td>
</tr>
<tr>
<td>Cou&#x00E9; et al. (2002) [<xref ref-type="bibr" rid="ref-47">47</xref>]</td>
<td>Odometry <inline-formula id="ieqn-5"><mml:math id="mml-ieqn-5"><mml:mo stretchy="false">&#x2192;</mml:mo></mml:math></inline-formula> displacement; gyroscope <inline-formula id="ieqn-6"><mml:math id="mml-ieqn-6"><mml:mo stretchy="false">&#x2192;</mml:mo></mml:math></inline-formula> angular rate; vision <inline-formula id="ieqn-7"><mml:math id="mml-ieqn-7"><mml:mo stretchy="false">&#x2192;</mml:mo></mml:math></inline-formula> landmarks.</td>
<td>Bayesian pose estimation fusing GPS, odometry, and gyroscope</td>
<td>Digital map priors fused with vision-derived landmarks</td>
<td>&#x2014;</td>
<td>Adaptive weighting for unreliable sensors.</td>
</tr>
<tr>
<td>Klausner et al. (2007) [<xref ref-type="bibr" rid="ref-57">57</xref>]</td>
<td>Audio <inline-formula id="ieqn-8"><mml:math id="mml-ieqn-8"><mml:mo stretchy="false">&#x2192;</mml:mo></mml:math></inline-formula> spectral features; images <inline-formula id="ieqn-9"><mml:math id="mml-ieqn-9"><mml:mo stretchy="false">&#x2192;</mml:mo></mml:math></inline-formula> Haar-like gradients.</td>
<td>Per-sensor classifiers using LS-SVM</td>
<td>Feature fusion via Genetic Algorithms; cross-modal selection</td>
<td>Decision fusion via Dempster&#x2013;Shafer theory</td>
<td>Adaptive switching between feature- and decision-level fusion.</td>
</tr>
<tr>
<td>Choi et al. (2011) [<xref ref-type="bibr" rid="ref-52">52</xref>]</td>
<td>RFID decoding <inline-formula id="ieqn-10"><mml:math id="mml-ieqn-10"><mml:mo stretchy="false">&#x2192;</mml:mo></mml:math></inline-formula> coordinates; ultrasonics <inline-formula id="ieqn-11"><mml:math id="mml-ieqn-11"><mml:mo stretchy="false">&#x2192;</mml:mo></mml:math></inline-formula> time-of-flight ranges.</td>
<td>Global Pose Estimation (RFID) &#x002B; Local Environment Cognition (ultrasonics)</td>
<td>Hierarchical matching aligns local ultrasonic maps to RFID-based global pose</td>
<td>Refined pose used for robot navigation</td>
<td>Iterative feedback between global and local estimators.</td>
</tr>
<tr>
<td>Chavez-Garcia &#x0026; Aycard (2016) [<xref ref-type="bibr" rid="ref-51">51</xref>]</td>
<td>LiDAR <inline-formula id="ieqn-12"><mml:math id="mml-ieqn-12"><mml:mo stretchy="false">&#x2192;</mml:mo></mml:math></inline-formula> point clouds; radar <inline-formula id="ieqn-13"><mml:math id="mml-ieqn-13"><mml:mo stretchy="false">&#x2192;</mml:mo></mml:math></inline-formula> velocity targets; camera <inline-formula id="ieqn-14"><mml:math id="mml-ieqn-14"><mml:mo stretchy="false">&#x2192;</mml:mo></mml:math></inline-formula> HOG features.</td>
<td>Decision fusion using Mahalanobis distance</td>
<td>Multi-object tracking via MCMC association</td>
<td>&#x2014;</td>
<td>Online belief updates and reliability discounting.</td>
</tr>
<tr>
<td>Barbier et al. (2019) [<xref ref-type="bibr" rid="ref-69">69</xref>]</td>
<td>Sensor data <inline-formula id="ieqn-15"><mml:math id="mml-ieqn-15"><mml:mo stretchy="false">&#x2192;</mml:mo></mml:math></inline-formula> per-cell occupancy probabilities.</td>
<td>Monte Carlo (particle) tracking of occupancy over time</td>
<td>Bayesian per-cell fusion for future occupancy maps</td>
<td>Risk thresholds for collision avoidance decisions</td>
<td>KPI checks with statistical model validation.</td>
</tr>
<tr>
<td>HydraFusion (2022) [<xref ref-type="bibr" rid="ref-71">71</xref>]</td>
<td>CNN encoders for each modality</td>
<td>Intermediate feature fusion at decision layers</td>
<td>Driving context inferred from multimodal representations</td>
<td>Planning and tracking tasks conditioned on context</td>
<td>&#x2014;</td>
</tr>
<tr>
<td><bold>Paper (Year)</bold></td>
<td><bold>Level 0 &#x2014; Preprocessing / Signal Conditioning</bold></td>
<td><bold>Level 1 &#x2014; Object-Level Estimation</bold></td>
<td><bold>Level 2 &#x2014; Scene-Level Integration</bold></td>
<td><bold>Level 3 &#x2014; Decision and Control</bold></td>
<td><bold>Level 4 &#x2014; Refinement and Adaptivity</bold></td>
</tr>
<tr>
<td>BEVFusion (2022) [<xref ref-type="bibr" rid="ref-84">84</xref>]</td>
<td>Camera and LiDAR encoders produce feature maps</td>
<td>Features projected into Bird&#x2019;s-Eye View (BEV) grids</td>
<td>Concatenated BEV representation for unified scene understanding</td>
<td>Multi-task heads for detection and segmentation</td>
<td>&#x2014;</td>
</tr>
<tr>
<td>Xiang et al. (2022) [<xref ref-type="bibr" rid="ref-82">82</xref>]</td>
<td>YOLOv5 generates semantic regions</td>
<td>Semantic point cloud from fused labels</td>
<td>DBSCAN clustering for 3D bounding boxes and classes</td>
<td>Broadcast blind-spot warnings to other vehicles</td>
<td>&#x2014;</td>
</tr>
<tr>
<td>Tommingas et al. (2025) [<xref ref-type="bibr" rid="ref-72">72</xref>]</td>
<td>UWB multilateration; GNSS <inline-formula id="ieqn-16"><mml:math id="mml-ieqn-16"><mml:mo stretchy="false">&#x2192;</mml:mo></mml:math></inline-formula> ENU coordinates</td>
<td>Per-sensor position estimates with diagnostics</td>
<td>Extreme Gradient Boosting regression for integrated position</td>
<td>Adaptive KF produces final fused state</td>
<td>&#x2014;</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Three principal inferences emerge from this mapping. First, a methodological progression is evident: early systems emphasized handcrafted features and direct neural control [<xref ref-type="bibr" rid="ref-33">33</xref>] or structured Bayesian inference [<xref ref-type="bibr" rid="ref-47">47</xref>] while mid-era work incorporated evidential and hierarchical reasoning [<xref ref-type="bibr" rid="ref-57">57</xref>,<xref ref-type="bibr" rid="ref-52">52</xref>] and recent contributions prioritize deep, representation-centric pipelines such as BEVFusion [<xref ref-type="bibr" rid="ref-84">84</xref>]and HydraFusion [<xref ref-type="bibr" rid="ref-71">71</xref>] or hybrid ML&#x2013;classical models like [<xref ref-type="bibr" rid="ref-72">72</xref>]. Second, representational practice has shifted away from object and feature-centric fusion and towards spatially unified forms like dynamic occupancy grids and Bird&#x2019;s-Eye Views that support simultaneous detection, segmentation, and planning. Third, uncertainty modeling and validation have re-emerged as central concerns: either through explicit probabilistic and evidential frameworks [<xref ref-type="bibr" rid="ref-51">51</xref>,<xref ref-type="bibr" rid="ref-69">69</xref>] or through learned uncertainty predictors feeding classical estimators [<xref ref-type="bibr" rid="ref-72">72</xref>].</p>
<p>Another diagnostic gap exposed by this mapping is that many deep fusion architectures collapse classical Level 0&#x2013;Level 4 distinctions into monolithic networks. These systems interleave preprocessing, per-sensor encoding, scene integration, and decision heads, making it difficult to isolate errors or provide component-level guarantees. While empirically effective, this consolidation impairs explainability and makes fault localization harder. Moreover, comparability and certifiability also become limited as safety-critical validation requires modular evidence. To reconcile the empirical power of modern end-to-end fusion with the interpretability and rigor of classical frameworks, we propose adopting <italic>level-aware practices</italic>: (1) Publish per-level diagnostics and artifacts alongside end-to-end metrics. For instance, Level 0 signal quality measures, Level 1 covariances, Level 2 association maps, Level 3 decision triggers (2) Design explicit interfaces within learned pipelines like exposing calibrated per-sensor estimates and uncertainty tensors (3) Develop benchmark suites stressing level-specific degradations of sensor noise, occlusion and association ambiguity (4) Pursue hybrid architectures where learned models provide uncertainty estimates or feature embeddings to principled filters and planners, as demonstrated in recent UWB&#x2013;GNSS fusion with ML-informed adaptive Kalman filtering.</p>
<p>These practices offer a pathway to preserve the accuracy and adaptability of modern learning-based fusion while restoring the modular transparency, comparability, and verifiability envisioned in the original hierarchical models. This synthesis illustrates that the conceptual clarity of classical architectures remains essential, even as fusion methods evolve into highly integrated deep networks.</p>
</sec>
<sec id="s6">
<label>6</label>
<title>Future Research Directions: The Way Forward for Sensor Fusion</title>
<p>The preceding mapping highlights a persistent gap in sensor fusion research: while Levels 0&#x2013;2 of the JDL framework involving signal conditioning, object estimation, and scene-level integration are well represented in modern methods, higher-level reasoning of Levels 3 and 4, remains underdeveloped. Current deep learning pipelines excel at perception but provide limited support for situation assessment like inter-object relationships, intent prediction. and impact/threat assessment. The lack of this involves risk analysis and proactive decision-making. This limitation is exacerbated by the scarcity of hierarchical datasets encompassing all JDL levels, preventing systematic training and benchmarking of higher-level inference. Consequently, although deep fusion models achieve high empirical accuracy, their opacity and lack of causal reasoning hinder deployment in safety-critical contexts.</p>
<sec id="s6_1">
<label>6.1</label>
<title>Explainability and Trustworthiness</title>
<p>To address this limitation, explainable AI (XAI) has become central to sensor fusion research. By exposing how models approximate higher-level reasoning, XAI can bridge the trust gap between opaque neural fusion and stakeholder accountability. In autonomous driving, trustworthy deployment hinges on transparent fusion pipelines with interpretable decision-making at multiple abstraction levels [<xref ref-type="bibr" rid="ref-11">11</xref>]. Similarly, in medical contexts, opaque multi-modal fusion undermines clinical reliability; interpretable frameworks are increasingly recognized as prerequisites for adoption [<xref ref-type="bibr" rid="ref-12">12</xref>]. The absence of hierarchical, explainable fusion is therefore both a technical and socio-ethical barrier. Recent surveys [<xref ref-type="bibr" rid="ref-10">10</xref>] underscore that progress remains incremental, and much work is needed before interpretable and certifiable fusion frameworks can be reliably deployed in safety-critical environments.</p>
</sec>
<sec id="s6_2">
<label>6.2</label>
<title>Future Research Priorities</title>
<p>Several research priorities emerge for bridging this gap:
<list list-type="bullet">
<list-item>
<p><bold>Unified evaluation frameworks</bold> and context-aware benchmarks are needed to standardize interpretability metrics in autonomous domains.</p></list-item>
<list-item>
<p><bold>Computationally efficient real-time XAI</bold> methods must be developed to ensure safety-critical explainability without introducing decision delays.</p></list-item>
<list-item>
<p><bold>Causal reasoning integration</bold> should illuminate cause&#x2013;effect relations in multimodal fusion, improving transparency and prediction of rare events.</p></list-item>
<list-item>
<p><bold>Scalable fusion algorithms</bold> are required to process heterogeneous, high-volume sensor data streams while maintaining robustness and interpretability.</p></list-item>
<list-item>
<p><bold>Ethical and regulatory compliance</bold> must be embedded into design aligning with global frameworks.</p></list-item>
<list-item>
<p><bold>Large Language Models (LLMs)</bold> may serve as adaptive explanation translators to generate stakeholder-specific justifications of fusion outputs.</p></list-item>
<list-item>
<p><bold>Adversarial robustness and security</bold> must be prioritized to guard against spoofing, sensor jamming, and multimodal adversarial attacks.</p></list-item>
<list-item>
<p><bold>Human&#x2013;AI collaboration and training</bold> will be critical to build trust, requiring education of engineers, regulators, and end-users in interpreting sensor fusion pipelines. Collectively, these directions define a roadmap toward transparent, resilient, and standardized sensor fusion for autonomous systems.</p></list-item>
</list></p>
</sec>
<sec id="s6_3">
<label>6.3</label>
<title>Neuromorphic Fusion as a Future Paradigm</title>
<p>Beyond deep learning, neuromorphic sensor fusion offers a promising path toward energy-efficient and inherently interpretable models. Ceolini et al. (2020) introduced one of the first multimodal neuromorphic benchmarks, integrating event-based vision (DVS) with electromyography (EMG) signals [<xref ref-type="bibr" rid="ref-85">85</xref>]. Using delta modulation, continuous EMG signals were converted into spike trains compatible with spiking neural networks (SNNs), while DVS provided native event-driven input. Fusion was achieved via late concatenation in the penultimate layer, followed by retraining of the output classifier across neuromorphic hardware platforms (Intel Loihi, ODIN&#x002B;MorphIC). The released dataset comprised 15,750 samples from 21 subjects performing five static hand gestures, making it a pioneering benchmark for multimodal neuromorphic fusion. Results showed accuracy comparable to GPU baselines, while achieving energy-delay product (EDP) improvements of up to <inline-formula id="ieqn-17"><mml:math id="mml-ieqn-17"><mml:mn>600</mml:mn><mml:mo>&#x00D7;</mml:mo></mml:math></inline-formula> on MorphIC and <inline-formula id="ieqn-18"><mml:math id="mml-ieqn-18"><mml:mn>30</mml:mn></mml:math></inline-formula>&#x2013;<inline-formula id="ieqn-19"><mml:math id="mml-ieqn-19"><mml:mn>50</mml:mn><mml:mo>&#x00D7;</mml:mo></mml:math></inline-formula> on Loihi. Although inference latency increased modestly, the efficiency gains highlight the feasibility of neuromorphic fusion for embedded, low-power autonomous platforms.</p>
<p>This study opens several technical directions for neuromorphic fusion. First, encoding fidelity remains an open challenge: spike conversion from continuous signals risks discarding fine-grained information, motivating adaptive or learned encoding schemes co-designed with SNN architectures. Second, hardware&#x2013;algorithm co-design is critical: current neuromorphic platforms face constraints such as limited neuron counts, fixed precision, and inefficient crossbar operations. Progress will require sparsity-aware SNN topologies and novel hardware primitives capable of handling dense multimodal streams. Third, standardized benchmarks are urgently needed. While Ceolini&#x2019;s dataset is valuable, broader benchmarks reflecting dynamic driving, UAV navigation, or healthcare monitoring are necessary for systematic evaluation across modalities and platforms.</p>
<p><italic>Explainability and Security in Neuromorphic Fusion:</italic> Neuromorphic systems also offer opportunities for explainability and robustness. The temporal and event-driven nature of SNNs makes causal reasoning more tractable, as spike timing and event sequences can be directly linked to decision outcomes. Developing XAI tailored for neuromorphic pipelines could deliver transparent reasoning for safety-critical systems such as AV perception or prosthetic control. Security is equally pressing: while neuromorphic fusion may resist conventional adversarial perturbations, it introduces new vulnerabilities such as spoofed event streams, requiring adversary-aware design and validation.</p>
<p><italic>Generalization Across Domains:</italic> A key long-term challenge is extending neuromorphic fusion beyond static benchmarks to dynamic, heterogeneous domains. Late-fusion architectures demonstrated for DVS&#x002B;EMG can be generalized to LiDAR, radar, inertial, and acoustic signals, supporting low-latency, always-on fusion in energy-constrained platforms. Hybrid pipelines&#x2014;where neuromorphic encoders perform low-level, energy-efficient fusion before passing to deep learning or symbolic reasoning modules&#x2014;could combine efficiency with semantic richness. Such hybridization points toward a future in which neuromorphic front-ends complement AI-driven back-ends, delivering scalable, interpretable, and trustworthy sensor fusion for autonomous systems.</p>
<p>In summary, future research must simultaneously advance the explainability of classical deep learning fusion systems and explore emerging paradigms such as neuromorphic computing. Together, these trajectories aim to reconcile the empirical success of modern AI with the interpretability, efficiency, and trustworthiness demanded by safety-critical autonomous deployments.</p>
</sec>
</sec>
<sec id="s7">
<label>7</label>
<title>Conclusions</title>
<p>This survey set out to provide a critical and structured examination of sensor fusion research spanning more than three decades, with the dual objectives of tracing the methodological evolution of fusion techniques and assessing their alignment with classical hierarchical models. These objectives have been met by systematically reviewing representative studies across different periods, analyzing their methodologies, applications, and limitations, and mapping them to the JDL framework.</p>
<p>The survey has documented how early work in the 1990s established the foundational principles of probabilistic inference, neural network&#x2013;based fusion, rule-based systems, and application-specific frameworks. These studies demonstrated the feasibility of multi-sensor integration under uncertainty, albeit within constrained computational and application settings. The review of the 2000s highlighted the maturation of probabilistic filters, the emergence of distributed consensus schemes, and the first uses of machine learning ensembles for fusion, marking a shift from theoretical constructs to robust, domain-specific implementations.</p>
<p>In analyzing the period from 2011 to 2020, the survey has shown how deep learning fundamentally transformed sensor fusion by enabling scalable, feature-level integration of high-dimensional multimodal data. Benchmarks such as nuScenes, Argoverse, and OPPORTUNITY were shown to play a pivotal role in standardizing evaluation and accelerating progress, particularly in autonomous driving and human activity recognition. The discussion also emphasized how decision-level ensembles, cooperative fusion concepts, and robustness studies broadened the applicability of fusion beyond narrowly engineered pipelines.</p>
<p>For the most recent period, from 2021 onward, the survey has demonstrated how research is moving toward real-world deployment and scalability. Contributions such as transformer-based fusion models, physics-informed learning, hybrid edge&#x2013;cloud architectures, and cooperative vehicle&#x2013;infrastructure systems reflect an emphasis on adaptability, fault tolerance, and generalization. By including representative studies across emerging application domains such as healthcare, smart cities, and environmental monitoring, the survey has highlighted the growing breadth of sensor fusion research.</p>
<p>The mapping exercise comparing classical hierarchical models to contemporary methods has achieved the objective of clarifying both continuity and divergence. It showed how classical pipelines, with explicit level-wise structure, anticipated many capabilities that are now realized in modern deep fusion systems, while also exposing critical gaps at higher JDL levels where reasoning, intent prediction, and impact assessment remain underdeveloped.</p>
<p>Through this systematic review, the survey has achieved its intended goals. It has established a coherent historical narrative, provided a comparative analysis of methods and benchmarks, and identified both strengths and limitations across decades of research. It has also articulated the open challenges of explainability, robustness, and trustworthiness, thereby framing the agenda for future research. In doing so, this work contributes not only a consolidation of prior knowledge but also a roadmap for advancing sensor fusion toward transparent, scalable, and safety-critical deployment in autonomous systems.</p>
</sec>
</body>
<back>
<ack>
<p>Not applicable.</p>
</ack>
<sec>
<title>Funding Statement</title>
<p>Not applicable.</p>
</sec>
<sec>
<title>Author Contributions</title>
<p>Conceptualization: Sangeeta Mittal, Chetna Gupta, Varun Gupta; Data curation: Sangeeta Mittal, Chetna Gupta; Formal Analysis: Sangeeta Mittal, Chetna Gupta; Investigation: Sangeeta Mittal, Chetna Gupta; Methodology: Sangeeta Mittal, Chetna Gupta, Varun Gupta; Project administration: Sangeeta Mittal, Chetna Gupta; Resources: Sangeeta Mittal, Chetna Gupta, Varun Gupta; Software: Sangeeta Mittal, Chetna Gupta, Varun Gupta; Supervision, Validation: Sangeeta Mittal, Chetna Gupta; Visualization: Sangeeta Mittal, Chetna Gupta, Varun Gupta; Writing&#x2014;original draft: Sangeeta Mittal, Chetna Gupta, Varun Gupta; Writing&#x2014;review &#x0026; editing: Sangeeta Mittal, Chetna Gupta, Varun Gupta. All authors reviewed the results and approved the final version of the manuscript.</p>
</sec>
<sec sec-type="data-availability">
<title>Availability of Data and Materials</title>
<p>Not applicable.</p>
</sec>
<sec>
<title>Ethics Approval</title>
<p>Not applicable.</p>
</sec>
<sec sec-type="COI-statement">
<title>Conflicts of Interest</title>
<p>The authors declare no conflicts of interest to report regarding the present study.</p>
</sec>
<ref-list content-type="authoryear">
<title>References</title>
<ref id="ref-1"><label>[1]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>Elmenreich</surname> <given-names>W</given-names></string-name></person-group>. <source>An introduction to sensor fusion</source>. Vol. <volume>502</volume>, No. 1&#x2013;28. <publisher-loc>Vienna, Austria</publisher-loc>: <publisher-name>Vienna University of Technology</publisher-name>; <year>2002</year>. <fpage>37</fpage> p.</mixed-citation></ref>
<ref id="ref-2"><label>[2]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Hughes</surname> <given-names>T</given-names></string-name></person-group>. <article-title>Sensor fusion in a military avionics environment</article-title>. <source>Measur Cont</source>. <year>1989</year>;<volume>22</volume>(<issue>7</issue>):<fpage>203</fpage>&#x2013;<lpage>5</lpage>. doi:<pub-id pub-id-type="doi">10.1177/002029408902200703</pub-id>.</mixed-citation></ref>
<ref id="ref-3"><label>[3]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>White</surname> <given-names>F</given-names></string-name>, <string-name><surname>Lexicon</surname> <given-names>FDF</given-names></string-name></person-group>. <article-title>Joint directors of laboratories-technical panel for C3I, data fusion sub-panel. San Diego, CA, USA: Naval Ocean Systems Center</article-title>. <year>1987</year> <comment>[cited 2025 Oct 20]</comment>. Available from: <ext-link ext-link-type="uri" xlink:href="https://apps.dtic.mil/sti/pdfs/ADA391672.pdf">https://apps.dtic.mil/sti/pdfs/ADA391672.pdf</ext-link>.</mixed-citation></ref>
<ref id="ref-4"><label>[4]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Dawar</surname> <given-names>N</given-names></string-name>, <string-name><surname>Kehtarnavaz</surname> <given-names>N</given-names></string-name></person-group>. <article-title>A convolutional neural network-based sensor fusion system for monitoring transition movements in healthcare applications</article-title>. In: <conf-name>2018 IEEE 14th International Conference on Control and Automation (ICCA)</conf-name>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2018</year>. p. <fpage>482</fpage>&#x2013;<lpage>5</lpage>.</mixed-citation></ref>
<ref id="ref-5"><label>[5]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Yeong</surname> <given-names>DJ</given-names></string-name>, <string-name><surname>Velasco-Hernandez</surname> <given-names>G</given-names></string-name>, <string-name><surname>Barry</surname> <given-names>J</given-names></string-name>, <string-name><surname>Walsh</surname> <given-names>J</given-names></string-name></person-group>. <article-title>Sensor and sensor fusion technology in autonomous vehicles: a review</article-title>. <source>Sensors</source>. <year>2021</year>;<volume>21</volume>(<issue>6</issue>):<fpage>2140</fpage>. doi:<pub-id pub-id-type="doi">10.3390/s21062140</pub-id>; <pub-id pub-id-type="pmid">33803889</pub-id></mixed-citation></ref>
<ref id="ref-6"><label>[6]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Cinar</surname> <given-names>E</given-names></string-name></person-group>. <article-title>A sensor fusion method using transfer learning models for equipment condition monitoring</article-title>. <source>Sensors</source>. <year>2022</year>;<volume>22</volume>(<issue>18</issue>):<fpage>6791</fpage>. doi:<pub-id pub-id-type="doi">10.3390/s22186791</pub-id>; <pub-id pub-id-type="pmid">36146139</pub-id></mixed-citation></ref>
<ref id="ref-7"><label>[7]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Issa</surname> <given-names>ME</given-names></string-name>, <string-name><surname>Helmi</surname> <given-names>AM</given-names></string-name>, <string-name><surname>Al-Qaness</surname> <given-names>MA</given-names></string-name>, <string-name><surname>Dahou</surname> <given-names>A</given-names></string-name>, <string-name><surname>Elaziz</surname> <given-names>MA</given-names></string-name>, <string-name><surname>Dama&#x0161;evi&#x010D;ius</surname> <given-names>R</given-names></string-name></person-group>. <article-title>Human activity recognition based on embedded sensor data fusion for the internet of healthcare things</article-title>. <source>Healthcare</source>. <year>2022 Jun</year>;<volume>10</volume>(<issue>6</issue>):<fpage>1084</fpage>. doi:<pub-id pub-id-type="doi">10.3390/healthcare10061084</pub-id>; <pub-id pub-id-type="pmid">35742136</pub-id></mixed-citation></ref>
<ref id="ref-8"><label>[8]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Abdelmoneem</surname> <given-names>RM</given-names></string-name>, <string-name><surname>Shaaban</surname> <given-names>E</given-names></string-name>, <string-name><surname>Benslimane</surname> <given-names>A</given-names></string-name></person-group>. <article-title>A survey on multi-sensor fusion techniques in IoT for healthcare</article-title>. In: <conf-name>2018 13th International Conference on Computer Engineering and Systems (ICCES)</conf-name>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2018</year>. p. <fpage>157</fpage>&#x2013;<lpage>62</lpage>.</mixed-citation></ref>
<ref id="ref-9"><label>[9]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Conway</surname> <given-names>M</given-names></string-name>, <string-name><surname>Reily</surname> <given-names>B</given-names></string-name>, <string-name><surname>Reardon</surname> <given-names>C</given-names></string-name></person-group>. <article-title>Learned sensor fusion for robust human activity recognition in challenging environments</article-title>. In: <conf-name>2024 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)</conf-name>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2024</year>. p. <fpage>11537</fpage>&#x2013;<lpage>43</lpage>.</mixed-citation></ref>
<ref id="ref-10"><label>[10]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Yeong</surname> <given-names>DJ</given-names></string-name>, <string-name><surname>Panduru</surname> <given-names>K</given-names></string-name>, <string-name><surname>Walsh</surname> <given-names>J</given-names></string-name></person-group>. <article-title>Exploring the unseen: a survey of multi-sensor fusion and the role of explainable AI (XAI) in autonomous vehicles</article-title>. <source>Sensors</source>. <year>2025</year>;<volume>25</volume>(<issue>3</issue>):<fpage>856</fpage>. doi:<pub-id pub-id-type="doi">10.3390/s25030856</pub-id>; <pub-id pub-id-type="pmid">39943496</pub-id></mixed-citation></ref>
<ref id="ref-11"><label>[11]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>De Jong Yeong</surname> <given-names>KP</given-names></string-name>, <string-name><surname>Walsh</surname> <given-names>J</given-names></string-name></person-group>. <chapter-title>Building trustworthy autonomous vehicles: the role of multi-sensor fusion and explainable AI (xAI) in on-road and off-road scenarios</chapter-title>. In: <source>Sensors and electronic instrumentation advances</source>. Vol. <volume>145</volume>. <publisher-loc>Oslo, Norway</publisher-loc>: <publisher-name>IFSA Publishing</publisher-name>; <year>2024</year>. doi:<pub-id pub-id-type="doi">10.3390/s25030856</pub-id>.</mixed-citation></ref>
<ref id="ref-12"><label>[12]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Yang</surname> <given-names>G</given-names></string-name>, <string-name><surname>Ye</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Xia</surname> <given-names>J</given-names></string-name></person-group>. <article-title>Unbox the black-box for the medical explainable AI via multi-modal and multi-centre data fusion: a mini-review, two showcases and beyond</article-title>. <source>Inf Fusion</source>. <year>2022</year>;<volume>77</volume>:<fpage>29</fpage>&#x2013;<lpage>52</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.inffus.2021.07.016</pub-id>; <pub-id pub-id-type="pmid">34980946</pub-id></mixed-citation></ref>
<ref id="ref-13"><label>[13]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Fu</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Tian</surname> <given-names>D</given-names></string-name>, <string-name><surname>Duan</surname> <given-names>X</given-names></string-name>, <string-name><surname>Zhou</surname> <given-names>J</given-names></string-name>, <string-name><surname>Lang</surname> <given-names>P</given-names></string-name>, <string-name><surname>Lin</surname> <given-names>C</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>A camera-radar fusion method based on edge computing</article-title>. In: <conf-name>2020 IEEE International Conference on Edge Computing (EDGE)</conf-name>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2020</year>. p. <fpage>9</fpage>&#x2013;<lpage>14</lpage>.</mixed-citation></ref>
<ref id="ref-14"><label>[14]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Davies</surname> <given-names>M</given-names></string-name>, <string-name><surname>Wild</surname> <given-names>A</given-names></string-name>, <string-name><surname>Orchard</surname> <given-names>G</given-names></string-name>, <string-name><surname>Sandamirskaya</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Guerra</surname> <given-names>GAF</given-names></string-name>, <string-name><surname>Joshi</surname> <given-names>P</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Advancing neuromorphic computing with Loihi: a survey of results and outlook</article-title>. <source>Proc IEEE</source>. <year>2021</year>;<volume>109</volume>(<issue>5</issue>):<fpage>911</fpage>&#x2013;<lpage>34</lpage>. doi:<pub-id pub-id-type="doi">10.1109/jproc.2021.3067593</pub-id>.</mixed-citation></ref>
<ref id="ref-15"><label>[15]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Kitchenham</surname> <given-names>BA</given-names></string-name>, <string-name><surname>Charters</surname> <given-names>S</given-names></string-name></person-group>. <article-title>Guidelines for performing systematic literature reviews in software engineering. EBSE Technical Report. Keele, UK: Keele University. Durham, UK: University of Durham</article-title>; <year>2007</year>.</mixed-citation></ref>
<ref id="ref-16"><label>[16]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Zhao</surname> <given-names>J</given-names></string-name>, <string-name><surname>Li</surname> <given-names>L</given-names></string-name>, <string-name><surname>Dai</surname> <given-names>J</given-names></string-name></person-group>. <article-title>A review of multi-sensor fusion 3D object detection for autonomous driving</article-title>. In: <conf-name>Eleventh International Symposium on Precision Mechanical Measurements</conf-name>. Vol. 13178. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2024</year>. p. <fpage>667</fpage>&#x2013;<lpage>85</lpage>.</mixed-citation></ref>
<ref id="ref-17"><label>[17]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Wang</surname> <given-names>S</given-names></string-name>, <string-name><surname>Ahmad</surname> <given-names>NS</given-names></string-name></person-group>. <article-title>A comprehensive review on sensor fusion techniques for localization of a dynamic target in GPS-denied environments</article-title>. <source>IEEE Access</source>. <year>2025</year>;<volume>13</volume>:<fpage>2252</fpage>&#x2013;<lpage>85</lpage>. doi:<pub-id pub-id-type="doi">10.1109/access.2024.3519874</pub-id>.</mixed-citation></ref>
<ref id="ref-18"><label>[18]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Mehta</surname> <given-names>M</given-names></string-name></person-group>. <article-title>Sensor fusion techniques in autonomous systems: a review of methods and applications</article-title>. <source>Int Res J Eng Technol (IRJET)</source>. <year>2025</year>;<volume>12</volume>(<issue>4</issue>):<fpage>1902</fpage>&#x2013;<lpage>8</lpage>.</mixed-citation></ref>
<ref id="ref-19"><label>[19]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Heydarian</surname> <given-names>M</given-names></string-name>, <string-name><surname>Doyle</surname> <given-names>TE</given-names></string-name></person-group>. <article-title>rwisdm: repaired wisdm, a public dataset for human activity recognition</article-title>. <comment>arXiv:2305.10222. 2023</comment>.</mixed-citation></ref>
<ref id="ref-20"><label>[20]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Wilson</surname> <given-names>B</given-names></string-name>, <string-name><surname>Qi</surname> <given-names>W</given-names></string-name>, <string-name><surname>Agarwal</surname> <given-names>T</given-names></string-name>, <string-name><surname>Lambert</surname> <given-names>J</given-names></string-name>, <string-name><surname>Singh</surname> <given-names>J</given-names></string-name>, <string-name><surname>Khandelwal</surname> <given-names>S</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Argoverse 2: next generation datasets for self-driving perception and forecasting</article-title>. <comment>arXiv:2301.00493. 2023</comment>.</mixed-citation></ref>
<ref id="ref-21"><label>[21]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Shermeyer</surname> <given-names>J</given-names></string-name>, <string-name><surname>Hossler</surname> <given-names>T</given-names></string-name>, <string-name><surname>Van Etten</surname> <given-names>A</given-names></string-name>, <string-name><surname>Hogan</surname> <given-names>D</given-names></string-name>, <string-name><surname>Lewis</surname> <given-names>R</given-names></string-name>, <string-name><surname>Kim</surname> <given-names>D</given-names></string-name></person-group>. <article-title>Rareplanes: synthetic data takes flight</article-title>. In: <conf-name>Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision</conf-name>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2021</year>. p. <fpage>207</fpage>&#x2013;<lpage>17</lpage>.</mixed-citation></ref>
<ref id="ref-22"><label>[22]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Wen</surname> <given-names>W</given-names></string-name>, <string-name><surname>Zhou</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>G</given-names></string-name>, <string-name><surname>Fahandezh-Saadi</surname> <given-names>S</given-names></string-name>, <string-name><surname>Bai</surname> <given-names>X</given-names></string-name>, <string-name><surname>Zhan</surname> <given-names>W</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>UrbanLoco: a full sensor suite dataset for mapping and localization in urban scenes</article-title>. In: <conf-name>2020 IEEE International Conference on Robotics and Automation (ICRA)</conf-name>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2020</year>. p. <fpage>2310</fpage>&#x2013;<lpage>6</lpage>.</mixed-citation></ref>
<ref id="ref-23"><label>[23]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Caesar</surname> <given-names>H</given-names></string-name>, <string-name><surname>Bankiti</surname> <given-names>V</given-names></string-name>, <string-name><surname>Lang</surname> <given-names>AH</given-names></string-name>, <string-name><surname>Vora</surname> <given-names>S</given-names></string-name>, <string-name><surname>Liong</surname> <given-names>VE</given-names></string-name>, <string-name><surname>Xu</surname> <given-names>Q</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Nuscenes: a multimodal dataset for autonomous driving</article-title>. In: <conf-name>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</conf-name>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2020</year>. p. <fpage>11621</fpage>&#x2013;<lpage>31</lpage>.</mixed-citation></ref>
<ref id="ref-24"><label>[24]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Cre&#x00DF;</surname> <given-names>C</given-names></string-name>, <string-name><surname>Zimmer</surname> <given-names>W</given-names></string-name>, <string-name><surname>Strand</surname> <given-names>L</given-names></string-name>, <string-name><surname>Fortkord</surname> <given-names>M</given-names></string-name>, <string-name><surname>Dai</surname> <given-names>S</given-names></string-name>, <string-name><surname>Lakshminarasimhan</surname> <given-names>V</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>A9-dataset: multi-sensor infrastructure-based dataset for mobility research</article-title>. In: <conf-name>2022 IEEE Intelligent Vehicles Symposium (IV)</conf-name>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2022</year>. p. <fpage>965</fpage>&#x2013;<lpage>70</lpage>.</mixed-citation></ref>
<ref id="ref-25"><label>[25]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Vaizman</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Ellis</surname> <given-names>K</given-names></string-name>, <string-name><surname>Lanckriet</surname> <given-names>G</given-names></string-name></person-group>. <article-title>Recognizing detailed human context in the wild from smartphones and smartwatches</article-title>. <source>IEEE Perv Comput</source>. <year>2017</year>;<volume>16</volume>(<issue>4</issue>):<fpage>62</fpage>&#x2013;<lpage>74</lpage>. doi:<pub-id pub-id-type="doi">10.1109/mprv.2017.3971131</pub-id>.</mixed-citation></ref>
<ref id="ref-26"><label>[26]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Sagha</surname> <given-names>H</given-names></string-name>, <string-name><surname>Digumarti</surname> <given-names>ST</given-names></string-name>, <string-name><surname>del R Mill&#x00E1;n</surname> <given-names>J</given-names></string-name>, <string-name><surname>Chavarriaga</surname> <given-names>R</given-names></string-name>, <string-name><surname>Calatroni</surname> <given-names>A</given-names></string-name>, <string-name><surname>Roggen</surname> <given-names>D</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Benchmarking classification techniques using the Opportunity human activity dataset</article-title>. In: <conf-name>2011 IEEE International Conference on Systems, Man, and Cybernetics</conf-name>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2011</year>. p. <fpage>36</fpage>&#x2013;<lpage>40</lpage>.</mixed-citation></ref>
<ref id="ref-27"><label>[27]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>Hall</surname> <given-names>DL</given-names></string-name>, <string-name><surname>McMullen</surname> <given-names>SAH</given-names></string-name></person-group>. <source>Mathematical techniques in multisensor data fusion</source>. <edition>2nd ed</edition>. <publisher-loc>Boston, MA, USA</publisher-loc>: <publisher-name>Artech House</publisher-name>; <year>2004</year>.</mixed-citation></ref>
<ref id="ref-28"><label>[28]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Steinberg</surname> <given-names>AN</given-names></string-name>, <string-name><surname>Bowman</surname> <given-names>CL</given-names></string-name></person-group>. <article-title>Rethinking the JDL data fusion levels</article-title>. <source>Nssdf Jhapl</source>. <year>2004</year>;<volume>38</volume>:<fpage>39</fpage>.</mixed-citation></ref>
<ref id="ref-29"><label>[29]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Thomopoulos</surname> <given-names>SC</given-names></string-name></person-group>. <article-title>Sensor integration and data fusion</article-title>. <source>J Robot Syst</source>. <year>1990</year>;<volume>7</volume>(<issue>3</issue>):<fpage>337</fpage>&#x2013;<lpage>72</lpage>.</mixed-citation></ref>
<ref id="ref-30"><label>[30]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Luo</surname> <given-names>RC</given-names></string-name>, <string-name><surname>Kay</surname> <given-names>MG</given-names></string-name></person-group>. <article-title>Multisensor integration and fusion in intelligent systems</article-title>. <source>IEEE Trans Syst Man Cybernet</source>. <year>2002</year>;<volume>19</volume>(<issue>5</issue>):<fpage>901</fpage>&#x2013;<lpage>31</lpage>. doi:<pub-id pub-id-type="doi">10.1109/21.44007</pub-id>.</mixed-citation></ref>
<ref id="ref-31"><label>[31]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Harris</surname> <given-names>C</given-names></string-name>, <string-name><surname>Bailey</surname> <given-names>A</given-names></string-name>, <string-name><surname>Dodd</surname> <given-names>T</given-names></string-name></person-group>. <article-title>Multi-sensor data fusion in defence and aerospace</article-title>. <source>Aeronaut J</source>. <year>1998</year>;<volume>102</volume>(<issue>1015</issue>):<fpage>229</fpage>&#x2013;<lpage>44</lpage>. doi:<pub-id pub-id-type="doi">10.1017/s0001924000065271</pub-id>.</mixed-citation></ref>
<ref id="ref-32"><label>[32]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Bedworth</surname> <given-names>M</given-names></string-name>, <string-name><surname>O&#x2019;Brien</surname> <given-names>J</given-names></string-name></person-group>. <article-title>The Omnibus model: a new model of data fusion?</article-title> <source>IEEE Aerosp Electr Syst Magaz</source>. <year>2000</year>;<volume>15</volume>(<issue>4</issue>):<fpage>30</fpage>&#x2013;<lpage>6</lpage>. doi:<pub-id pub-id-type="doi">10.1109/62.839632</pub-id>.</mixed-citation></ref>
<ref id="ref-33"><label>[33]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Cox</surname> <given-names>IJ</given-names></string-name>, <string-name><surname>Hingorani</surname> <given-names>S</given-names></string-name>, <string-name><surname>Maggs</surname> <given-names>BM</given-names></string-name>, <string-name><surname>Rao</surname> <given-names>SB</given-names></string-name></person-group>. <article-title>Stereo without disparity gradient smoothing: a Bayesian sensor fusion solution</article-title>. In: <conf-name>BMVC92: Proceedings of the British Machine Vision Conference</conf-name>. <publisher-loc>Cham, Switzerland</publisher-loc>: <publisher-name>Springer</publisher-name>; <year>1992</year>. p. <fpage>337</fpage>&#x2013;<lpage>46</lpage>.</mixed-citation></ref>
<ref id="ref-34"><label>[34]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Larkin</surname> <given-names>MJ</given-names></string-name></person-group>. <article-title>Sensor fusion and classification of acoustic signals using Bayesian networks</article-title>. In: <conf-name>Conference Record of Thirty-Second Asilomar Conference on Signals, Systems and Computers</conf-name>. Vol. 2. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>1998</year>. p. <fpage>1359</fpage>&#x2013;<lpage>62</lpage>. doi:<pub-id pub-id-type="doi">10.1109/acssc.1998.751395</pub-id>.</mixed-citation></ref>
<ref id="ref-35"><label>[35]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>Shahbazian</surname> <given-names>E</given-names></string-name>, <string-name><surname>Simard</surname> <given-names>MA</given-names></string-name>, <string-name><surname>Bourassa</surname> <given-names>S</given-names></string-name></person-group>. <chapter-title>Implementation strategies for the central-level multihypothesis tracking fusion with multiple dissimilar sensors</chapter-title>. In: <source>Signal processing, sensor fusion, and target recognition II</source>. Vol. <volume>1955</volume>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>1993</year>. p. <fpage>78</fpage>&#x2013;<lpage>89</lpage>. doi: <pub-id pub-id-type="doi">10.1117/12.155003</pub-id>.</mixed-citation></ref>
<ref id="ref-36"><label>[36]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Perlovsky</surname> <given-names>LI</given-names></string-name>, <string-name><surname>McManus</surname> <given-names>MM</given-names></string-name></person-group>. <article-title>Maximum likelihood neural networks for sensor fusion and adaptive classification</article-title>. <source>Neur Netw</source>. <year>1991</year>;<volume>4</volume>(<issue>1</issue>):<fpage>89</fpage>&#x2013;<lpage>102</lpage>. doi: <pub-id pub-id-type="doi">10.1016/0893-6080(91)90035-4</pub-id>.</mixed-citation></ref>
<ref id="ref-37"><label>[37]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Davis</surname> <given-names>IL</given-names></string-name>, <string-name><surname>Stentz</surname> <given-names>A</given-names></string-name></person-group>. <article-title>Sensor fusion for autonomous outdoor navigation using neural networks</article-title>. In: <conf-name>Proceedings of the 1995 IEEE/RSJ International Conference on Intelligent Robots and Systems. Human Robot Interaction and Cooperative Robots</conf-name>. Vol. 3. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>1995</year>. p. <fpage>338</fpage>&#x2013;<lpage>43</lpage>.</mixed-citation></ref>
<ref id="ref-38"><label>[38]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>Cao</surname> <given-names>J</given-names></string-name>, <string-name><surname>Hall</surname> <given-names>EL</given-names></string-name></person-group>. <chapter-title>Sensor fusion for the navigation of an autonomous guided vehicle using neural networks</chapter-title>. In: <source>Intelligent robots and computer vision XVII: algorithms, techniques, and active vision</source>. Vol. <volume>3522</volume>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>1998</year>. p. <fpage>286</fpage>&#x2013;<lpage>94</lpage>. doi: <pub-id pub-id-type="doi">10.1117/12.360289</pub-id>.</mixed-citation></ref>
<ref id="ref-39"><label>[39]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Tang</surname> <given-names>YC</given-names></string-name>, <string-name><surname>Lee</surname> <given-names>CG</given-names></string-name></person-group>. <article-title>A geometric feature relation graph formulation for consistent sensor fusion</article-title>. <source>IEEE Trans Syst, Man Cybernet</source>. <year>1992</year>;<volume>22</volume>(<issue>1</issue>):<fpage>115</fpage>&#x2013;<lpage>29</lpage>. doi: <pub-id pub-id-type="doi">10.1109/21.141316</pub-id>.</mixed-citation></ref>
<ref id="ref-40"><label>[40]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>McKee</surname> <given-names>GT</given-names></string-name></person-group>. <chapter-title>What can be fused?</chapter-title> In: <source>Multisensor fusion for computer vision</source>. <publisher-loc>Berlin/Heidelberg, Germany</publisher-loc>: <publisher-name>Springer</publisher-name>; <year>1993</year>. p. <fpage>71</fpage>&#x2013;<lpage>84</lpage>.</mixed-citation></ref>
<ref id="ref-41"><label>[41]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Alag</surname> <given-names>S</given-names></string-name>, <string-name><surname>Goebel</surname> <given-names>K</given-names></string-name>, <string-name><surname>Agogino</surname> <given-names>A</given-names></string-name></person-group>. <article-title>A framework for intelligent sensor validation, sensor fusion, and supervisory control of automated vehicles in IVHS</article-title>. In: <conf-name>Proceedings of the 1995 Annual Meeting of ITS America; 1995 Mar 19&#x2013;22</conf-name>; <publisher-loc>Los Angeles, CA, USA</publisher-loc>. p. <fpage>77</fpage>&#x2013;<lpage>87</lpage>.</mixed-citation></ref>
<ref id="ref-42"><label>[42]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Kam</surname> <given-names>M</given-names></string-name>, <string-name><surname>Zhu</surname> <given-names>X</given-names></string-name>, <string-name><surname>Kalata</surname> <given-names>P</given-names></string-name></person-group>. <article-title>Sensor fusion for mobile robot navigation</article-title>. <source>Proc IEEE</source>. <year>1997</year>;<volume>85</volume>(<issue>1</issue>):<fpage>108</fpage>&#x2013;<lpage>19</lpage>. doi: <pub-id pub-id-type="doi">10.1109/jproc.1997.554212</pub-id>.</mixed-citation></ref>
<ref id="ref-43"><label>[43]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>Blasch</surname> <given-names>EP</given-names></string-name>, <string-name><surname>Hong</surname> <given-names>L</given-names></string-name></person-group>. <chapter-title>Sensor fusion cognition using belief filtering for tracking and identification</chapter-title>. In: <source>Sensor fusion: architectures, algorithms, and applications III</source>. Vol. <volume>3719</volume>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>1999</year>. p. <fpage>250</fpage>&#x2013;<lpage>9</lpage>.</mixed-citation></ref>
<ref id="ref-44"><label>[44]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>Peli</surname> <given-names>T</given-names></string-name>, <string-name><surname>Young</surname> <given-names>M</given-names></string-name>, <string-name><surname>Knox</surname> <given-names>R</given-names></string-name>, <string-name><surname>Ellis</surname> <given-names>KK</given-names></string-name>, <string-name><surname>Bennett</surname> <given-names>F</given-names></string-name></person-group>. <chapter-title>Feature-level sensor fusion</chapter-title>. In: <source>Sensor fusion: architectures, algorithms, and applications III</source>. Vol. <volume>3719</volume>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>1999</year>. p. <fpage>332</fpage>&#x2013;<lpage>9</lpage>.</mixed-citation></ref>
<ref id="ref-45"><label>[45]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Mandenius</surname> <given-names>CF</given-names></string-name>, <string-name><surname>Ekl&#x00F6;v</surname> <given-names>T</given-names></string-name>, <string-name><surname>Lundstr&#x00F6;m</surname> <given-names>I</given-names></string-name></person-group>. <article-title>Sensor fusion with on-line gas emission multisensor arrays and standard process measuring devices in a baker&#x2019;s yeast manufacturing process</article-title>. <source>Biotechnol Bioeng</source>. <year>1997</year>;<volume>55</volume>(<issue>2</issue>):<fpage>427</fpage>&#x2013;<lpage>38</lpage>. doi: <pub-id pub-id-type="doi">10.1002/(sici)1097-0290(19970720)55:2&#x003C;427::aid-bit20&#x003E;3.3.co;2-j</pub-id>.</mixed-citation></ref>
<ref id="ref-46"><label>[46]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>St-Pierre</surname> <given-names>M</given-names></string-name>, <string-name><surname>Gingras</surname> <given-names>D</given-names></string-name></person-group>. <article-title>Comparison between the unscented kalman filter and the extended kalman filter for the position estimation module of an integrated navigation information system</article-title>. In: <conf-name>Proceedings of the IEEE Intelligent Vehicles Symposium, 2004</conf-name>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2004</year>. p. <fpage>831</fpage>&#x2013;<lpage>5</lpage>.</mixed-citation></ref>
<ref id="ref-47"><label>[47]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Cou&#x00E9;</surname> <given-names>C</given-names></string-name>, <string-name><surname>Fraichard</surname> <given-names>T</given-names></string-name>, <string-name><surname>Bessi&#x00E8;re</surname> <given-names>P</given-names></string-name>, <string-name><surname>Mazer</surname> <given-names>E</given-names></string-name></person-group>. <article-title>Multi-sensor data fusion using bayesian programming: an automotive application</article-title>. In: <conf-name>Proceedings of the 2002 IEEE Intelligent Vehicle Symposium</conf-name>. Vol. 2. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2002</year>. p. <fpage>442</fpage>&#x2013;<lpage>7</lpage>.</mixed-citation></ref>
<ref id="ref-48"><label>[48]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Koks</surname> <given-names>D</given-names></string-name>, <string-name><surname>Challa</surname> <given-names>S</given-names></string-name></person-group>. <article-title>An introduction to bayesian and dempster&#x2013;shafer data fusion. Edinburgh, Australia: DSTO Systems Sciences Laboratory. DSTO-TR-1436</article-title>. <year>2003</year> <comment>[cited 2025 Oct 15]</comment>. Available from: <ext-link ext-link-type="uri" xlink:href="https://www.dst.defence.gov.au/publication/introduction-bayesian-and-dempster-shafer-data-fusion">https://www.dst.defence.gov.au/publication/introduction-bayesian-and-dempster-shafer-data-fusion</ext-link>.</mixed-citation></ref>
<ref id="ref-49"><label>[49]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Xiao</surname> <given-names>L</given-names></string-name>, <string-name><surname>Boyd</surname> <given-names>S</given-names></string-name>, <string-name><surname>Lall</surname> <given-names>S</given-names></string-name></person-group>. <article-title>A scheme for robust distributed sensor fusion based on average consensus</article-title>. In: <conf-name>Proceedings of the Fourth International Symposium on Information Processing in Sensor Networks (IPSN 2005)</conf-name>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2005</year>. p. <fpage>63</fpage>&#x2013;<lpage>70</lpage>.</mixed-citation></ref>
<ref id="ref-50"><label>[50]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Faceli</surname> <given-names>K</given-names></string-name>, <string-name><surname>De Carvalho</surname> <given-names>ADSB</given-names></string-name>, <string-name><surname>Rezende</surname> <given-names>SO</given-names></string-name></person-group>. <article-title>Combining intelligent techniques for sensor fusion</article-title>. <source>Appl Intell</source>. <year>2004</year>;<volume>20</volume>(<issue>3</issue>):<fpage>199</fpage>&#x2013;<lpage>213</lpage>. doi:<pub-id pub-id-type="doi">10.1023/b:apin.0000021413.05467.20</pub-id>.</mixed-citation></ref>
<ref id="ref-51"><label>[51]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Chavez-Garcia</surname> <given-names>RO</given-names></string-name>, <string-name><surname>Aycard</surname> <given-names>O</given-names></string-name></person-group>. <article-title>Multiple sensor fusion and classification for moving object detection and tracking</article-title>. <source>IEEE Trans Intell Transport Syst</source>. <year>2015</year>;<volume>17</volume>(<issue>2</issue>):<fpage>525</fpage>&#x2013;<lpage>34</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tits.2015.2479925</pub-id>.</mixed-citation></ref>
<ref id="ref-52"><label>[52]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Choi</surname> <given-names>BS</given-names></string-name>, <string-name><surname>Lee</surname> <given-names>JW</given-names></string-name>, <string-name><surname>Lee</surname> <given-names>JJ</given-names></string-name>, <string-name><surname>Park</surname> <given-names>KT</given-names></string-name></person-group>. <article-title>A hierarchical algorithm for indoor mobile robot localization using RFID sensor fusion</article-title>. <source>IEEE Trans Indust Elect</source>. <year>2011</year>;<volume>58</volume>(<issue>6</issue>):<fpage>2226</fpage>&#x2013;<lpage>35</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tie.2011.2109330</pub-id>.</mixed-citation></ref>
<ref id="ref-53"><label>[53]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Lu</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Michaels</surname> <given-names>JE</given-names></string-name></person-group>. <article-title>Feature extraction and sensor fusion for ultrasonic structural health monitoring under changing environmental conditions</article-title>. <source>IEEE Sens J</source>. <year>2009</year>;<volume>9</volume>(<issue>11</issue>):<fpage>1462</fpage>&#x2013;<lpage>71</lpage>. doi:<pub-id pub-id-type="doi">10.1109/jsen.2009.2019339</pub-id>.</mixed-citation></ref>
<ref id="ref-54"><label>[54]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Huang</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Lan</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Hoffmann</surname> <given-names>WC</given-names></string-name>, <string-name><surname>Lacey</surname> <given-names>RE</given-names></string-name></person-group>. <article-title>Multisensor data fusion for high quality data analysis and processing in measurement and instrumentation</article-title>. <source>J Bionic Eng</source>. <year>2007</year>;<volume>4</volume>(<issue>1</issue>):<fpage>53</fpage>&#x2013;<lpage>62</lpage>. doi:<pub-id pub-id-type="doi">10.1016/s1672-6529(07)60013-4</pub-id>.</mixed-citation></ref>
<ref id="ref-55"><label>[55]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Tan</surname> <given-names>AH</given-names></string-name></person-group>. <article-title>FALCON: a fusion architecture for learning, cognition, and navigation</article-title>. In: <conf-name>Proceedings of the 2004 IEEE International Joint Conference on Neural Networks (IJCNN)</conf-name>. Vol. 4. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2004</year>. p. <fpage>3297</fpage>&#x2013;<lpage>302</lpage>. doi:<pub-id pub-id-type="doi">10.1109/ijcnn.2004.1381208</pub-id>.</mixed-citation></ref>
<ref id="ref-56"><label>[56]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Upcroft</surname> <given-names>B</given-names></string-name>, <string-name><surname>Ong</surname> <given-names>LL</given-names></string-name>, <string-name><surname>Kumar</surname> <given-names>S</given-names></string-name>, <string-name><surname>Ridley</surname> <given-names>M</given-names></string-name>, <string-name><surname>Bailey</surname> <given-names>T</given-names></string-name>, <string-name><surname>Sukkarieh</surname> <given-names>S</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Rich probabilistic representations for bearing-only decentralised data fusion</article-title>. In: <conf-name>Proceedings of the 2005 International Conference on Information Fusion</conf-name>. Vol. 2. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2005</year>. p. <fpage>8</fpage>&#x2013;<lpage>15</lpage>.</mixed-citation></ref>
<ref id="ref-57"><label>[57]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Klausner</surname> <given-names>A</given-names></string-name>, <string-name><surname>Tengg</surname> <given-names>A</given-names></string-name>, <string-name><surname>Rinner</surname> <given-names>B</given-names></string-name></person-group>. <article-title>Vehicle classification on multi-sensor smart cameras using feature- and decision-fusion</article-title>. In: <conf-name>2007 First ACM/IEEE International Conference on Distributed Smart Cameras</conf-name>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2007</year>. p. <fpage>67</fpage>&#x2013;<lpage>74</lpage>.</mixed-citation></ref>
<ref id="ref-58"><label>[58]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Seraji</surname> <given-names>H</given-names></string-name>, <string-name><surname>Serrano</surname> <given-names>N</given-names></string-name></person-group>. <article-title>A multisensor decision fusion system for terrain safety assessment</article-title>. <source>IEEE Trans Robot</source>. <year>2009</year>;<volume>25</volume>(<issue>1</issue>):<fpage>99</fpage>&#x2013;<lpage>108</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tro.2008.2006705</pub-id>.</mixed-citation></ref>
<ref id="ref-59"><label>[59]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Gross</surname> <given-names>J</given-names></string-name>, <string-name><surname>Gu</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Gururajan</surname> <given-names>S</given-names></string-name>, <string-name><surname>Seanor</surname> <given-names>B</given-names></string-name>, <string-name><surname>Napolitano</surname> <given-names>M</given-names></string-name></person-group>. <article-title>A comparison of extended kalman filter, sigma-point kalman filter, and particle filter in GPS/INS sensor fusion</article-title>. In: <conf-name>Proceedings of the AIAA Guidance, Navigation, and Control Conference; 2010 Aug 2&#x2013;5</conf-name>; <publisher-loc>Toronto, ON, Canada</publisher-loc>.</mixed-citation></ref>
<ref id="ref-60"><label>[60]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Fayyad</surname> <given-names>J</given-names></string-name>, <string-name><surname>Jaradat</surname> <given-names>MA</given-names></string-name>, <string-name><surname>Gruyer</surname> <given-names>D</given-names></string-name>, <string-name><surname>Najjaran</surname> <given-names>H</given-names></string-name></person-group>. <article-title>Deep learning sensor fusion for autonomous vehicle perception and localization: a review</article-title>. <source>Sensors</source>. <year>2020</year>;<volume>20</volume>(<issue>15</issue>):<fpage>4220</fpage>. doi:<pub-id pub-id-type="doi">10.3390/s20154220</pub-id>; <pub-id pub-id-type="pmid">32751275</pub-id></mixed-citation></ref>
<ref id="ref-61"><label>[61]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Velasco-Hernandez</surname> <given-names>G</given-names></string-name>, <string-name><surname>Barry</surname> <given-names>J</given-names></string-name>, <string-name><surname>Walsh</surname> <given-names>J</given-names></string-name></person-group>. <article-title>Autonomous driving architectures, perception and data fusion: a review</article-title>. In: <conf-name>2020 IEEE 16th International Conference on Intelligent Computer Communication and Processing (ICCP)</conf-name>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2020</year>. p. <fpage>315</fpage>&#x2013;<lpage>21</lpage>.</mixed-citation></ref>
<ref id="ref-62"><label>[62]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><surname>Liu</surname> <given-names>S</given-names></string-name>, <string-name><surname>Gao</surname> <given-names>C</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Peng</surname> <given-names>X</given-names></string-name>, <string-name><surname>Kong</surname> <given-names>X</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>K</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Towards vehicle-to-everything autonomous driving: a survey on collaborative perception</article-title>. <comment>arXiv:2308.16714. 2023</comment>.</mixed-citation></ref>
<ref id="ref-63"><label>[63]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Banos</surname> <given-names>O</given-names></string-name>, <string-name><surname>Damas</surname> <given-names>M</given-names></string-name>, <string-name><surname>Pomares</surname> <given-names>H</given-names></string-name>, <string-name><surname>Rojas</surname> <given-names>I</given-names></string-name></person-group>. <article-title>On the use of sensor fusion to reduce the impact of rotational and additive noise in human activity recognition</article-title>. <source>Sensors</source>. <year>2012</year>;<volume>12</volume>(<issue>6</issue>):<fpage>8039</fpage>&#x2013;<lpage>54</lpage>. doi:<pub-id pub-id-type="doi">10.3390/s120608039</pub-id>; <pub-id pub-id-type="pmid">22969386</pub-id></mixed-citation></ref>
<ref id="ref-64"><label>[64]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Maimaitijiang</surname> <given-names>M</given-names></string-name>, <string-name><surname>Sagan</surname> <given-names>V</given-names></string-name>, <string-name><surname>Sidike</surname> <given-names>P</given-names></string-name>, <string-name><surname>Daloye</surname> <given-names>AM</given-names></string-name>, <string-name><surname>Erkbol</surname> <given-names>H</given-names></string-name>, <string-name><surname>Fritschi</surname> <given-names>FB</given-names></string-name></person-group>. <article-title>Crop monitoring using satellite/UAV data fusion and machine learning</article-title>. <source>Remote Sens</source>. <year>2020</year>;<volume>12</volume>(<issue>9</issue>):<fpage>1357</fpage>. doi:<pub-id pub-id-type="doi">10.3390/rs12091357</pub-id>.</mixed-citation></ref>
<ref id="ref-65"><label>[65]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Fagiano</surname> <given-names>L</given-names></string-name>, <string-name><surname>Huynh</surname> <given-names>K</given-names></string-name>, <string-name><surname>Bamieh</surname> <given-names>B</given-names></string-name>, <string-name><surname>Khammash</surname> <given-names>M</given-names></string-name></person-group>. <article-title>On sensor fusion for airborne wind energy systems</article-title>. <source>IEEE Trans Cont Syst Technol</source>. <year>2013</year>;<volume>22</volume>(<issue>3</issue>):<fpage>930</fpage>&#x2013;<lpage>43</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tcst.2013.2269865</pub-id>.</mixed-citation></ref>
<ref id="ref-66"><label>[66]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Chen</surname> <given-names>YL</given-names></string-name>, <string-name><surname>Jahanshahi</surname> <given-names>MR</given-names></string-name>, <string-name><surname>Manjunatha</surname> <given-names>P</given-names></string-name>, <string-name><surname>Gan</surname> <given-names>W</given-names></string-name>, <string-name><surname>Abdelbarr</surname> <given-names>M</given-names></string-name>, <string-name><surname>Masri</surname> <given-names>SF</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Inexpensive multimodal sensor fusion system for autonomous data acquisition of road surface conditions</article-title>. <source>IEEE Sens J</source>. <year>2016</year>;<volume>16</volume>(<issue>21</issue>):<fpage>7731</fpage>&#x2013;<lpage>43</lpage>. doi:<pub-id pub-id-type="doi">10.1109/jsen.2016.2602871</pub-id>.</mixed-citation></ref>
<ref id="ref-67"><label>[67]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Guo</surname> <given-names>D</given-names></string-name>, <string-name><surname>Zhong</surname> <given-names>M</given-names></string-name>, <string-name><surname>Zhou</surname> <given-names>D</given-names></string-name></person-group>. <article-title>Multisensor data-fusion-based approach to airspeed measurement fault detection for unmanned aerial vehicles</article-title>. <source>IEEE Trans Instrument Measur</source>. <year>2017</year>;<volume>67</volume>(<issue>2</issue>):<fpage>317</fpage>&#x2013;<lpage>27</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tim.2017.2735663</pub-id>.</mixed-citation></ref>
<ref id="ref-68"><label>[68]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Tsinganos</surname> <given-names>P</given-names></string-name>, <string-name><surname>Skodras</surname> <given-names>A</given-names></string-name></person-group>. <article-title>On the comparison of wearable sensor data fusion to a single sensor machine learning technique in fall detection</article-title>. <source>Sensors</source>. <year>2018</year>;<volume>18</volume>(<issue>2</issue>):<fpage>592</fpage>. doi:<pub-id pub-id-type="doi">10.3390/s18020592</pub-id>; <pub-id pub-id-type="pmid">29443923</pub-id></mixed-citation></ref>
<ref id="ref-69"><label>[69]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Barbier</surname> <given-names>M</given-names></string-name>, <string-name><surname>Renzaglia</surname> <given-names>A</given-names></string-name>, <string-name><surname>Quilbeuf</surname> <given-names>J</given-names></string-name>, <string-name><surname>Rummelhard</surname> <given-names>L</given-names></string-name>, <string-name><surname>Paigwar</surname> <given-names>A</given-names></string-name>, <string-name><surname>Laugier</surname> <given-names>C</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Validation of perception and decision-making systems for autonomous driving via statistical model checking</article-title>. In: <conf-name>2019 IEEE Intelligent Vehicles Symposium (IV)</conf-name>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2019</year>. p. <fpage>252</fpage>&#x2013;<lpage>9</lpage>.</mixed-citation></ref>
<ref id="ref-70"><label>[70]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Chitta</surname> <given-names>K</given-names></string-name>, <string-name><surname>Prakash</surname> <given-names>A</given-names></string-name>, <string-name><surname>Geiger</surname> <given-names>A</given-names></string-name></person-group>. <article-title>Transfuser: imitation with transformer-based sensor fusion for autonomous driving</article-title>. In: <conf-name>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</conf-name>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2022</year>. p. <fpage>6929</fpage>&#x2013;<lpage>38</lpage>.</mixed-citation></ref>
<ref id="ref-71"><label>[71]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Malawade</surname> <given-names>AV</given-names></string-name>, <string-name><surname>Mortlock</surname> <given-names>T</given-names></string-name>, <string-name><surname>AlFaruque</surname> <given-names>MA</given-names></string-name></person-group>. <article-title>HydraFusion: context-aware selective sensor fusion for robust and efficient autonomous vehicle perception</article-title>. In: <conf-name>Proceedings of the 2022 ACM/IEEE 13th International Conference on Cyber-Physical Systems (ICCPS)</conf-name>. <publisher-loc>New York, NY, USA</publisher-loc>: <publisher-name>ACM</publisher-name>; <year>2022</year>. p. <fpage>68</fpage>&#x2013;<lpage>79</lpage>.</mixed-citation></ref>
<ref id="ref-72"><label>[72]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Tommingas</surname> <given-names>M</given-names></string-name>, <string-name><surname>Laadung</surname> <given-names>T</given-names></string-name>, <string-name><surname>Varbla</surname> <given-names>S</given-names></string-name>, <string-name><surname>M&#x00FC;&#x00FC;rsepp</surname> <given-names>I</given-names></string-name>, <string-name><surname>Alam</surname> <given-names>MM</given-names></string-name></person-group>. <article-title>UWB and GNSS sensor fusion using ml-based positioning uncertainty estimation</article-title>. <source>IEEE Open J Commun Soc</source>. <year>2025</year>;<volume>6</volume>:<fpage>2177</fpage>&#x2013;<lpage>89</lpage>. doi:<pub-id pub-id-type="doi">10.1109/ojcoms.2025.3571303</pub-id>.</mixed-citation></ref>
<ref id="ref-73"><label>[73]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Wang</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Sun</surname> <given-names>R</given-names></string-name>, <string-name><surname>Jiang</surname> <given-names>L</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>H</given-names></string-name>, <string-name><surname>Mao</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Ochieng</surname> <given-names>WY</given-names></string-name></person-group>. <article-title>Multipath inflation factor for robust GNSS/IMU/VO fusion based navigation in urban areas</article-title>. <source>IEEE Internet Things J</source>. <year>2025</year>;<volume>12</volume>(<issue>11</issue>):<fpage>16256</fpage>&#x2013;<lpage>65</lpage>. doi:<pub-id pub-id-type="doi">10.1109/jiot.2025.3535819</pub-id>.</mixed-citation></ref>
<ref id="ref-74"><label>[74]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Wang</surname> <given-names>L</given-names></string-name>, <string-name><surname>Fu</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Zhu</surname> <given-names>R</given-names></string-name>, <string-name><surname>Liu</surname> <given-names>N</given-names></string-name>, <string-name><surname>Shi</surname> <given-names>H</given-names></string-name>, <string-name><surname>Liu</surname> <given-names>Z</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Research on high precision localization of space target with multi-sensor association</article-title>. <source>Opt Lasers Eng</source>. <year>2025</year>;<volume>184</volume>:<fpage>108553</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.optlaseng.2024.108553</pub-id>.</mixed-citation></ref>
<ref id="ref-75"><label>[75]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Rashid</surname> <given-names>N</given-names></string-name>, <string-name><surname>Mortlock</surname> <given-names>T</given-names></string-name>, <string-name><surname>Al Faruque</surname> <given-names>MA</given-names></string-name></person-group>. <article-title>Stress detection using context-aware sensor fusion from wearable devices</article-title>. <source>IEEE Internet Things J</source>. <year>2023</year>;<volume>10</volume>(<issue>16</issue>):<fpage>14114</fpage>&#x2013;<lpage>27</lpage>. doi:<pub-id pub-id-type="doi">10.1109/jiot.2023.3265768</pub-id>.</mixed-citation></ref>
<ref id="ref-76"><label>[76]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Hasanujjaman</surname> <given-names>M</given-names></string-name>, <string-name><surname>Chowdhury</surname> <given-names>MZ</given-names></string-name>, <string-name><surname>Jang</surname> <given-names>YM</given-names></string-name></person-group>. <article-title>Sensor fusion in autonomous vehicle with traffic surveillance camera system: detection, localization, and AI networking</article-title>. <source>Sensors</source>. <year>2023</year>;<volume>23</volume>(<issue>6</issue>):<fpage>3335</fpage>. doi:<pub-id pub-id-type="doi">10.3390/s23063335</pub-id>; <pub-id pub-id-type="pmid">36992043</pub-id></mixed-citation></ref>
<ref id="ref-77"><label>[77]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Zhang</surname> <given-names>C</given-names></string-name>, <string-name><surname>Dong</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Hu</surname> <given-names>P</given-names></string-name>, <string-name><surname>Fu</surname> <given-names>H</given-names></string-name>, <string-name><surname>Yang</surname> <given-names>H</given-names></string-name>, <string-name><surname>Yang</surname> <given-names>R</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Large-range displacement measurement in narrow space scenarios: fiber microprobe sensor with subnanometer accuracy</article-title>. <source>Photon Res</source>. <year>2024</year>;<volume>12</volume>(<issue>9</issue>):<fpage>1877</fpage>&#x2013;<lpage>89</lpage>. doi:<pub-id pub-id-type="doi">10.1364/prj.525667</pub-id>.</mixed-citation></ref>
<ref id="ref-78"><label>[78]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Aguilar-Lazcano</surname> <given-names>CA</given-names></string-name>, <string-name><surname>Espinosa-Curiel</surname> <given-names>IE</given-names></string-name>, <string-name><surname>R&#x00ED;os-Mart&#x00ED;nez</surname> <given-names>JA</given-names></string-name>, <string-name><surname>Madera-Ram&#x00ED;rez</surname> <given-names>FA</given-names></string-name>, <string-name><surname>P&#x00E9;rez-Espinosa</surname> <given-names>H</given-names></string-name></person-group>. <article-title>Machine learning-based sensor data fusion for animal monitoring: a scoping review</article-title>. <source>Sensors</source>. <year>2023</year>;<volume>23</volume>(<issue>12</issue>):<fpage>5732</fpage>. doi:<pub-id pub-id-type="doi">10.3390/s23125732</pub-id>; <pub-id pub-id-type="pmid">37420896</pub-id></mixed-citation></ref>
<ref id="ref-79"><label>[79]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Almujally</surname> <given-names>NA</given-names></string-name>, <string-name><surname>Rafique</surname> <given-names>AA</given-names></string-name>, <string-name><surname>Al Mudawi</surname> <given-names>N</given-names></string-name>, <string-name><surname>Alazeb</surname> <given-names>A</given-names></string-name>, <string-name><surname>Alonazi</surname> <given-names>M</given-names></string-name>, <string-name><surname>Algarni</surname> <given-names>A</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Multi-modal remote perception learning for object sensory data</article-title>. <source>Front Neurorobot</source>. <year>2024</year>;<volume>18</volume>:<fpage>1427786</fpage>. doi:<pub-id pub-id-type="doi">10.3389/fnbot.2024.1427786</pub-id>; <pub-id pub-id-type="pmid">39377028</pub-id></mixed-citation></ref>
<ref id="ref-80"><label>[80]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Qiao</surname> <given-names>Y</given-names></string-name>, <string-name><surname>L&#x00FC;</surname> <given-names>J</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>T</given-names></string-name>, <string-name><surname>Liu</surname> <given-names>K</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>B</given-names></string-name>, <string-name><surname>Snoussi</surname> <given-names>H</given-names></string-name></person-group>. <article-title>A multihead attention self-supervised representation model for industrial sensors anomaly detection</article-title>. <source>IEEE Trans Indust Inform</source>. <year>2024</year>;<volume>20</volume>(<issue>2</issue>):<fpage>2190</fpage>&#x2013;<lpage>9</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tii.2023.3280337</pub-id>.</mixed-citation></ref>
<ref id="ref-81"><label>[81]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Zhang</surname> <given-names>S</given-names></string-name>, <string-name><surname>Yao</surname> <given-names>J</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>R</given-names></string-name>, <string-name><surname>Liu</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Ma</surname> <given-names>C</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>Y</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Design of intelligent fire-fighting robot based on multi-sensor fusion and experimental study on fire scene patrol</article-title>. <source>Robot Auton Syst</source>. <year>2022</year>;<volume>154</volume>:<fpage>104122</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.robot.2022.104122</pub-id>.</mixed-citation></ref>
<ref id="ref-82"><label>[82]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Xiang</surname> <given-names>C</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>L</given-names></string-name>, <string-name><surname>Xie</surname> <given-names>X</given-names></string-name>, <string-name><surname>Zhao</surname> <given-names>L</given-names></string-name>, <string-name><surname>Ke</surname> <given-names>X</given-names></string-name>, <string-name><surname>Niu</surname> <given-names>Z</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Multi-sensor fusion algorithm in cooperative vehicle-infrastructure system for blind spot warning</article-title>. <source>Int J Distrib Sens Netw</source>. <year>2022</year>;<volume>18</volume>(<issue>5</issue>):<fpage>15501329221100412</fpage>. doi:<pub-id pub-id-type="doi">10.1177/15501329221100412</pub-id>.</mixed-citation></ref>
<ref id="ref-83"><label>[83]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Ignatious</surname> <given-names>HA</given-names></string-name>, <string-name><surname>Khan</surname> <given-names>M</given-names></string-name></person-group>. <article-title>An overview of sensors in autonomous vehicles</article-title>. <source>Procedia Comput Sci</source>. <year>2022</year>;<volume>198</volume>:<fpage>736</fpage>&#x2013;<lpage>41</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.procs.2021.12.315</pub-id>.</mixed-citation></ref>
<ref id="ref-84"><label>[84]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Liu</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Tang</surname> <given-names>H</given-names></string-name>, <string-name><surname>Amini</surname> <given-names>A</given-names></string-name>, <string-name><surname>Yang</surname> <given-names>X</given-names></string-name>, <string-name><surname>Mao</surname> <given-names>H</given-names></string-name>, <string-name><surname>Rus</surname> <given-names>DL</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>BEVFusion: multi-task multi-sensor fusion with unified bird&#x2019;s-eye view representation</article-title>. In: <conf-name>2023 IEEE International Conference on Robotics and Automation (ICRA)</conf-name>. <publisher-loc>Piscataway, NJ, USA</publisher-loc>: <publisher-name>IEEE</publisher-name>; <year>2023</year>. p. <fpage>2774</fpage>&#x2013;<lpage>81</lpage>.</mixed-citation></ref>
<ref id="ref-85"><label>[85]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Ceolini</surname> <given-names>E</given-names></string-name>, <string-name><surname>Frenkel</surname> <given-names>C</given-names></string-name>, <string-name><surname>Shrestha</surname> <given-names>SB</given-names></string-name>, <string-name><surname>Taverni</surname> <given-names>G</given-names></string-name>, <string-name><surname>Khacef</surname> <given-names>L</given-names></string-name>, <string-name><surname>Payvand</surname> <given-names>M</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Hand-gesture recognition based on EMG and event-based camera sensor fusion: a benchmark in neuromorphic computing</article-title>. <source>Front Neurosci</source>. <year>2020</year>;<volume>14</volume>:<fpage>637</fpage>. doi:<pub-id pub-id-type="doi">10.3389/fnins.2020.00637</pub-id>; <pub-id pub-id-type="pmid">32903824</pub-id></mixed-citation></ref>
</ref-list>
</back></article>