[go: up one dir, main page]

Next Article in Journal
Comparison of Faster R-CNN, YOLO, and SSD for Third Molar Angle Detection in Dental Panoramic X-rays
Previous Article in Journal
YOLO-BFRV: An Efficient Model for Detecting Printed Circuit Board Defects
Previous Article in Special Issue
Anthropogenic Object Localization: Evaluation of Broad-Area High-Resolution Imagery Scans Using Deep Learning in Overhead Imagery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Regulating Modality Utilization within Multimodal Fusion Networks

1
Department of Electrical & Microelectronic Engineering, Rochester Institute of Technology, Rochester, NY 14623, USA
2
Department of Electrical & Computer Engineering and Department of Computer Science, The University of Texas at San Antonio, San Antonio, TX 78249, USA
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(18), 6054; https://doi.org/10.3390/s24186054
Submission received: 30 May 2024 / Revised: 30 August 2024 / Accepted: 17 September 2024 / Published: 19 September 2024
(This article belongs to the Special Issue Deep Learning Methods for Aerial Imagery)
Figure 1
<p>Computing modality utilization by randomly shuffling a modality <math display="inline"><semantics> <msub> <mi>M</mi> <mi>i</mi> </msub> </semantics></math> within the dataset to break the association between the input modality <math display="inline"><semantics> <msub> <mi>M</mi> <mi>i</mi> </msub> </semantics></math> and the output <span class="html-italic">Y</span>.</p> ">
Figure 2
<p>Modality utilization-based training targets the decision layers while using pre-trained feature extractors with frozen weights.</p> ">
Figure 3
<p>Visualization of multimodal fusion network’s gradient descent on the loss surface of the fusion network task. Optimizing <math display="inline"><semantics> <msub> <mi>L</mi> <mrow> <mi>m</mi> <mi>u</mi> </mrow> </msub> </semantics></math> from the very beginning can push the network in the local minima.</p> ">
Figure 4
<p>Clipped exponential function-based for loss factor warm-up for MU-based training.</p> ">
Figure 5
<p>NTIRE 2021 Multimodal Aerial View Object Classification Challenge Dataset [<a href="#B15-sensors-24-06054" class="html-bibr">15</a>].</p> ">
Figure 6
<p>NTIRE 2021 Multimodal Aerial View Object Classification Network with ResNet18 as the backbone.</p> ">
Figure 7
<p>Visualization of NTIRE21 dataset classification using Multimodal Aerial View Object Classification Network.</p> ">
Figure 8
<p>MCubeS Multimodal Material Segmentation Dataset [<a href="#B16-sensors-24-06054" class="html-bibr">16</a>].</p> ">
Figure 9
<p>MCubeS Multimodal Material Segmentation Network with UNet as the backbone.</p> ">
Figure 10
<p>Visualization of MCubeS dataset image segmentation using the Multimodal Material Segmentation Network.</p> ">
Figure 11
<p>Effects of different target utilization <math display="inline"><semantics> <mrow> <mi>M</mi> <msub> <mi>U</mi> <mrow> <mi>t</mi> <mi>a</mi> <mi>r</mi> <mi>g</mi> <mi>e</mi> <mi>t</mi> </mrow> </msub> </mrow> </semantics></math> on modality utilization and classification accuracy with modality utilization-based training method in the NTIRE21 dataset. Loss factor <math display="inline"><semantics> <mrow> <msub> <mi>λ</mi> <mi>L</mi> </msub> <mo>=</mo> <mn>100.0</mn> </mrow> </semantics></math> with SAR as the focus modality.</p> ">
Figure 12
<p>Effects of the loss factor <math display="inline"><semantics> <msub> <mi>λ</mi> <mi>L</mi> </msub> </semantics></math> on modality utilization and classification accuracy with modality utilization-based training method in NTIRE21 dataset. Target utilization <math display="inline"><semantics> <mrow> <mi>M</mi> <msub> <mi>U</mi> <mrow> <mi>t</mi> <mi>a</mi> <mi>r</mi> <mi>g</mi> <mi>e</mi> <mi>t</mi> </mrow> </msub> <mo>=</mo> <mn>50</mn> </mrow> </semantics></math>% with SAR as the focus modality.</p> ">
Figure 13
<p>Effects of the loss factor buildup rate <math display="inline"><semantics> <mi>β</mi> </semantics></math> on modality utilization and classification accuracy with modality utilization-based training method in the NTIRE21 dataset. Target utilization <math display="inline"><semantics> <mrow> <mi>M</mi> <msub> <mi>U</mi> <mrow> <mi>t</mi> <mi>a</mi> <mi>r</mi> <mi>g</mi> <mi>e</mi> <mi>t</mi> </mrow> </msub> <mo>=</mo> <mn>50</mn> </mrow> </semantics></math>%, Maximum Loss Factor <math display="inline"><semantics> <mrow> <msub> <mi>λ</mi> <mrow> <mi>L</mi> <mo>_</mo> <mi>m</mi> <mi>a</mi> <mi>x</mi> </mrow> </msub> <mo>=</mo> <mn>100</mn> </mrow> </semantics></math>, and buildup delay <math display="inline"><semantics> <mrow> <mi>δ</mi> <mo>=</mo> <mn>0</mn> </mrow> </semantics></math> with SAR as the focus modality.</p> ">
Figure 14
<p>Effects of Gaussian noise with mean = 0 and variance = <math display="inline"><semantics> <mrow> <mo>{</mo> <mn>0.06</mn> <mo>,</mo> <mn>0.09</mn> <mo>,</mo> <mn>0.12</mn> <mo>}</mo> </mrow> </semantics></math> in the EO modality, the SAR modality, and both modalities during inference on networks trained with different levels of SAR utilization.</p> ">
Review Reports Versions Notes

Abstract

:
Multimodal fusion networks play a pivotal role in leveraging diverse sources of information for enhanced machine learning applications in aerial imagery. However, current approaches often suffer from a bias towards certain modalities, diminishing the potential benefits of multimodal data. This paper addresses this issue by proposing a novel modality utilization-based training method for multimodal fusion networks. The method aims to guide the network’s utilization on its input modalities, ensuring a balanced integration of complementary information streams, effectively mitigating the overutilization of dominant modalities. The method is validated on multimodal aerial imagery classification and image segmentation tasks, effectively maintaining modality utilization within ± 10 % of the user-defined target utilization and demonstrating the versatility and efficacy of the proposed method across various applications. Furthermore, the study explores the robustness of the fusion networks against noise in input modalities, a crucial aspect in real-world scenarios. The method showcases better noise robustness by maintaining performance amidst environmental changes affecting different aerial imagery sensing modalities. The network trained with 75.0% EO utilization achieves significantly better accuracy (81.4%) in noisy conditions (noise variance = 0.12) compared to traditional training methods with 99.59% EO utilization (73.7%). Additionally, it maintains an average accuracy of 85.0% across different noise levels, outperforming the traditional method’s average accuracy of 81.9%. Overall, the proposed approach presents a significant step towards harnessing the full potential of multimodal data fusion in diverse machine learning applications such as robotics, healthcare, satellite imagery, and defense applications.

1. Introduction

Continued advancements in technology, data availability, and algorithmic innovation are set to propel the ongoing rise of machine learning. Utilizing statistics to identify and exploit patterns in data is the essence of machine learning. The amount of information in data has a huge impact on how well the machine learning algorithm learns these patterns. Data are also not limited to a single stream of information such as images, audio signals, or text. Multimodal data provide complementary information about the same phenomenon through multiple modalities or information streams. Information captured through multiple modalities can be beneficial in fields such as autonomous driving for image segmentation and object recognition [1], aerospace for activity recognition from aerial imagery [2], robotics for SLAM [3], human–robot collaboration [4], healthcare for diagnosing diseases from medical imagery [5], and defense applications [6]. Certain scenarios may favor one modality over others. For instance, in dark conditions, a near-infrared camera image is more useful for autonomous driving cars than an RGB camera image.
The information from different modalities needs to be merged or fused for a multimodal fusion network to utilize the information from all the different modalities efficiently. Applications within aerial imagery, such as video surveillance, meteorological analysis, vehicle navigation, land segmentation, and activity detection, heavily rely on a diverse array of data sources [7,8,9,10]. These sources encompass various modalities such as electro-optical imaging, synthetic aperture radar (SAR), hyperspectral imaging, and more, each offering unique perspectives and advantages depending on environmental conditions and observational requirements. The data fusion can take place at the data level, network level, and/or decision level [11]. This study focuses on network-level fusion for multimodal remote sensing data, allowing the network to make decisions using data from diverse modalities while learning feature embeddings independently for each modality. This enhances decision-making by exploiting complementary information effectively, resulting in improved accuracy and robustness.
Network-level fusion-based multimodal networks trained end-to-end may inadvertently prioritize one modality over others, resulting in the heavy utilization of a single modality, often neglecting the complementary information offered by others [12,13]. This limits the effectiveness of the multimodal system, especially in scenarios where all modalities contribute equally, or where certain modalities are crucial for accurate inference. During adverse weather conditions like heavy cloud cover, for instance, Electro-Optical imagery might be less effective, while synthetic aperture radar (SAR) could provide clearer insights due to its ability to penetrate clouds. Neglecting SAR data in such conditions could lead to incomplete or inaccurate assessments, highlighting the necessity of balanced modality utilization in aerial imagery applications to ensure robust and reliable outcomes. Furthermore, optimizing multimodal networks in aerial imagery requires the meticulous tuning of hyperparameters tailored to each modality. Variations in learning rates, regularization strengths, and network architectures for different modalities are essential for achieving optimal fusion outcomes [12]. The improper adjustment of hyperparameters for each modality can lead to suboptimal fusion, even causing multimodal networks to underperform compared to their unimodal counterparts.
These issues can be identified by observing the extent to which the network utilizes each of the input data modalities using a modality utilization metric [14]. Recognizing these issues and addressing them is crucial to leverage the multimodal system effectively. This work presents a modality utilization-based training method that can regulate a multimodal network’s utilization of its input data modalities. This helps alleviate the problem of the overutilization of a singular dominant modality. The first research question investigated in this study is, can we leverage the modality utilization metrics during training to regulate a network’s reliance on a dominant modality? The method is validated on a multimodal aerial imagery classification task [15] and a multimodal image segmentation task [16], showcasing its versatility in various multimodal applications. Furthermore, the study investigates the impact of modality utilization-based training on enhancing network robustness to noise, particularly relevant in aerial imagery where environmental factors like weather conditions and time of day can affect different sensing modalities differently. The second research question investigated in this study is, does the modality utilization-based training method improve the overall noise robustness of multimodal fusion networks? The key contributions of this research are as follows:
  • A modality utilization-based training method for multimodal fusion networks to regulate the network’s modality utilization;
  • Demonstrated that regulating modality utilization within a network improves overall noise robustness;
  • A heuristic approach for selecting target utilization-based on unimodal network performance.
The rest of the paper is organized as follows: Section 2 presents related work in the field of multimodal data fusion. Section 3 reviews the modality utilization metric, and presents the modality utilization-based training method. Section 4 lays down the details of the datasets and network architectures used to validate the presented approach. Section 5 presents the experimental results followed by a discussion in Section 6. Section 7 summarizes the findings and future work.

2. Related Work

Data fusion in multimodal systems can take place at three different levels: (i) early fusion or data-level fusion; (ii) intermediate fusion or network-level fusion; and (iii) late fusion or decision-level fusion [11]. Data-level fusion consists of combining information or features from different modalities at the raw input level to obtain a better representation of the data prior to the machine learning model [17,18]. Network-level fusion combines information from different modalities within the machine learning model via various mechanisms [19]. Concatenating high-level features or feature embeddings from different modalities within the model is one of the most common methods of network-level fusion. Decision-level fusion takes place at the decision level, where the decisions from multiple unimodal machine learning models are fused into a common decision [19,20]. This work focuses on network-level fusion because it can effectively exploit complementary information from different modalities, learning a joint representation of the multimodal data while being flexible enough to incorporate modality-specific features.
Object detection [21,22,23], image segmentation [24,25,26], and classification [27,28,29] are some of the common tasks in the field of aerial imagery. These tasks heavily rely on multimodal approaches, focusing on applications such as video surveillance, vehicle navigation, land segmentation, and activity detection. These applications use multiple heterogeneous image sources containing inter-modality and cross-modality information such as LiDAR, electro-optical imaging, synthetic aperture radar, hyperspectral imaging, and near-infrared imaging, which need to be combined to enhance the overall information about the phenomenon under observation [30]. Many advances have been made in the field of aerial imagery in recent years, promoted by various global contests such as the IEEE GRSS Data Fusion Contests [8], SpaceNet Challenges [9], and NTIRE challenges [10]. A multimodal knowledge distillation method was proposed by Z. Huang et al. [31] to develop a lightweight CNN model for arbitrary-oriented object detection. This contributes towards the deployment of lightweight models for remote sensing where computational resources are limited. Addressing the limited availability of paired multimodal data, S. Singh et al. [32] used a two-stage training approach for limited multimodal data fusion. Y. Xiang et al. [33] used edge-guided multimodal transformers to detect changes while monitoring land during natural disasters or cloud/fog occlusions based on heterogeneous satellite and aerial image modalities.
Network-level fusion in multimodal networks has access to more information than their unimodal counterparts. Multimodal fusion networks, trained end-to-end, tend to have an imbalance in the utilization of their input modalities due to their greedy nature [34,35]. This bias often results in the overutilization of a single modality, leading to the neglect of the valuable complementary information provided by other modalities [12,13]. Imbalance in multimodal systems limits their effectiveness, especially when all the modalities contribute equally or are crucial for accurate inference. Each modality branch in a multimodal network may require different hyperparameters (e.g., learning rates and regularization strengths) for optimal tuning. Failure to adjust these parameters can lead to suboptimal fusion, causing multimodal networks to underperform compared to their unimodal counterparts. Thus, addressing these challenges is crucial for realizing the full potential of network-level fusion in multimodal systems. M. Ghahremani and C. Wachinger [36] proposed multimodal batch normalization with regularization (RegBN) to tackle the bias and variance issues introduced by heterogeneous modalities. Similarly, I. Gat et al. [37] introduced a regularization term based on functional entropy to address this problem. H. Ma et al. [38] also approached the issue by introducing a regularization term to calibrate predictive confidence when one or more modalities are missing. Relying on multiple modalities also increases the computational complexity of the network. Y. Cao et al. [39] reduced the model complexity of using multiple modalities by learning the common information among the modalities via channel switching and spatial attention. Redundant modalities or modalities with low utility can also be removed based on the learning utility of each modality as demonstrated by Y. He et al. [40]. Another approach to tackle the hyperparameter mismatching across different modalities is to balance the learning of the various modality branches based on an adaptive tracking factor [41]. N. Wu et al. [12] used a conditional utilization rate for each modality based on unimodal performance to tackle the hyperparameter mismatch across different modalities; however, the conditional utilization rate cannot be computed during training time. Since the modality utilization metric can be effectively computed during training [14], the work presented in this paper leverages the modality utilization metric to balance the utilization of the input modalities.
Another rapidly growing approach in the field of multimodal data fusion is transformer architectures, which excel at capturing intricate relationships and long-range dependencies within data, making them suitable for multimodal applications [42]. These architectures are adept at learning the complex interactions among different modalities. Transformers can selectively focus on relevant parts of the input data from each modality by utilizing attention mechanisms [43], thus extracting the most informative features from every source [44]. This allows for more effective fusion of multimodal information, leading to enhanced performance in tasks such as image classification, segmentation, and sequence modeling across modalities. A multimodal fusion transformer (MFT) network featuring a multihead cross-patch attention mechanism was proposed by S. K. Roy et al. [45] for hyperspectral image-based land-cover classification augmented with data from other modalities such as LiDAR and synthetic aperture radar (SAR). Transformers have also been used with heterogeneous modalities such as audio–visual modalities for emotion recognition [46]. Using a CNN encoder and transformer decoder for feature extraction, L. Boussioux et al. [47] developed a tropical cyclone tracking and intensity estimation system using visual data from a reanalysis dataset and historical statistical data. Y. Luo et al. [48] used mixed-attention operations to utilize relatively dominant modality RGB-Thermal tracking in varying environmental conditions to achieve more robust tracking performance compared to single modality tracking systems. This shows that attention mechanisms can enable modality-specific processing, allowing the model to assign varying importance levels to different modalities based on task relevance.
Despite the advancements in multimodal transformer models, visual transformers suitable for aerial imagery require significantly more data compared to their CNN counterparts [49,50]. This poses a limitation on the applicability of transformers for smaller datasets. This study addresses this challenge by introducing a modality utilization-based training method for traditional multimodal neural network architectures. This method aims to regulate the utilization of various input modalities during training, effectively addressing issues such as modality utilization imbalance and hyperparameter mismatch between modalities. Moreover, the field of multimodal aerial imagery suffers from modality utilization imbalance [14,35,51], which is addressed in this work.

3. Method

The aim of the proposed method is to regulate a multimodal network’s utilization of its input modalities. This can be achieved by measuring the current utilization of each modality and minimizing the difference between the current and a target utilization for a modality.

3.1. Modality Utilization Metric

The modality utilization (MU) metric, proposed in [14], provides a method to quantify a network’s utilization of each modality. The MU metric was inspired by the permutation feature importance [52,53]. Breaking the association between an input modality and the network’s output, the modality utilization is calculated by observing the discrepancies in the multimodal fusion network output when compared to the original dataset.
Assume that there is a dataset D with M modalities and N samples, and a trained multimodal network F θ trained on this dataset. Figure 1 shows an example of the dataset and multimodal network with 4 modalities. A forward pass with dataset D on the network F θ will return the loss (L), which is the expected loss L of the network F θ :
L = E ( L { F θ , D } )
The modality utilization score for the i t h modality ( S c o r e i ) is computed by breaking the association between the input modality M i and the network output Y. This is performed by randomly shuffling the samples of the modality M i within the dataset D while keeping the samples of the remaining modalities ( M j where j i ) unchanged. A forward pass with the modified dataset D i on the network F θ will return a new loss ( L i ) of the network F θ :
L i = E ( L { F θ , D i } )
The modality utilization score for the i t h modality can then be calculated by observing the discrepancy between the loss L and the new loss L i :
S c o r e i = | L L i |
A larger MU score implies the network performance changed significantly due to changes in the modality M i , indicating that the network heavily relies on this modality. A smaller MU score implies that the network performance did not change significantly due to changes in the modality M i , suggesting that the network has low utilization of this modality. The MU score can be calculated for each modality and normalized to obtain the modality utilization metric for each modality. The modality utilization M U i ranges from 0.0 to 1.0, providing a percentage utilization of the network F θ on the i t h modality:
M U i = S c o r e i j = 1 M S c o r e j = | L L i | j = 1 M | L L j |
Algorithm 1 summarizes the modality utilization computation process. The MU metric computation is further explained in more detail with validation and ablation studies in [14]. This metric has also been extended to the reinforcement learning domain, providing valuable insights into the reinforcement learning agents’ action policies [54].
Algorithm 1: Modality Utilization
Initialize the multimodal fusion network F θ , learned model parameters θ , task dataset D ;
Compute the network loss L via forward pass with dataset D , Equation (1);
Sensors 24 06054 i001

3.2. Modality Utilization-Based Training

The modality utilization metric can be leveraged to regulate a multimodal fusion network’s utilization of certain modalities. A network is trained by minimizing a loss function that encapsulates how well the network is performing on a given task, such as classification, segmentation, object detection, and regression. This loss function can be augmented with a second term that minimizes the mean squared error of the current modality utilization ( M U c u r r ) of the focus modality and the set target modality utilization ( M U t a r g ). For a generic multimodal fusion network task, the loss function is as follows:
L T o t a l = L t a s k + λ L L m u
L T o t a l = L t a s k + λ L M S E ( M U c u r r , M U t a r g )
where L t a s k is the task loss, L m u is the modality utilization loss, and λ L is a scaling loss factor. Higher λ L emphasizes maintaining target modality utilization for a focus modality, while lower values prioritize solving the fusion network task. Extremely high λ L may overemphasize maintaining modality utilization at the expense of solving the fusion task, while λ L = 0 trains the multimodal fusion network conventionally, without modality utilization-based training.
The modality utilization-based training method targets the decision layers of the fusion network to regulate the utilization of the fusion network on its input modalities. Thus, pre-trained frozen weights from the unimodal models are used as feature extractors F E i ( M i ) (highlighted in red in Figure 2), and only the decision layers D L are trained (highlighted in green in Figure 2). This approach ensures that the quality of the feature embeddings from each modality is not affected by the modified loss function while allowing for a change in network utilization of its various input modalities.
Algorithm 2 summarizes the modality utilization-based training for the multimodal fusion network. The proposed method compares the current modality utilization of the focus modality with the target modality utilization and optimizes the network to maintain the target utilization while solving the fusion network task.
Algorithm 2: Modality utilization-based training
Initialize the multimodal fusion network F θ , pre-trained feature extractors ( F E i ), task dataset D , focus modality m f , loss factor λ L , and target modality utilization M U t a r g ;
Load the parameters for pre-trained feature extractors ( F E i ) for each modality;
Freeze the pre-trained feature extractors ( F E i );
Sensors 24 06054 i002

3.3. Loss Factor Warm-Up

The proposed training method leverages pre-trained feature extractors to independently extract feature embeddings from each modality. These embeddings are then concatenated and utilized to train the decision layers initialized randomly as illustrated in Figure 3. The initialization of the network plays a crucial role in its convergence towards the global minimum. Initially, both the task loss ( L t a s k ) and the modality utilization loss ( L m u ) are expected to be high. This early stage makes the training process susceptible to being pushed in a local minimum due to the high modality utilization loss ( L m u ) compared to the task-specific loss.
This issue can be resolved by adopting a slow start or a warm-up phase for the loss factor. Setting the loss factor to zero in the initial phase of training allows the network to focus solely on the task loss, gradually ramping up the loss factor after a certain number of training steps stabilizes the modality utilization-based training. A clipped exponential function can provide the desired behavior for the loss factor. The loss factor λ L ( i ) for the i t h training step is as follows:
λ L ( i ) = m a x ( 0 , λ L _ m a x ( 1 e β ( δ i ) )
where λ L _ m a x is the maximum value of the loss factor, β is the buildup rate of the loss factor from 0 to λ L _ m a x , and δ is the buildup delay when the loss factor starts increasing exponentially from zero. Figure 4 shows a visualization of the exponential-based loss factor function and hyperparameters. An understanding of the complexity of the task and the architecture of the fusion network can guide the selection of the hyperparameter values.
The maximum value of the loss factor λ L _ m a x dictates the amount of emphasis placed on maintaining the target modality utilization. Given that the computed M U c u r r and M U t a r g range between 0.0 and 1.0 , L m u (the mean squared error between M U c u r r and M U t a r g ) also ranges between 0.0 and 1.0 . Therefore, the maximum value of the loss factor λ L _ m a x must be selected such that the scaled modality utilization loss term λ L _ m a x × L m u is comparable to the range of the loss of the fusion network task L t a s k . A much larger value of λ L _ m a x can render L t a s k insignificant, causing the network to fail to converge properly. Conversely, a much smaller value of λ L _ m a x will make the scaled modality utilization loss term λ L _ m a x × L m u insignificant, essentially resulting in behavior similar to traditional machine learning training methods.
The buildup rate of the loss factor β dictates how quickly the value of λ L ( i ) climbs from 0 to λ L _ m a x . An aggressive value of β = 0.5 is generally a good starting point, as minimal differences in performance are observed with different values of β .
The buildup delay δ is the training step at which the loss factor starts increasing exponentially from 0 to λ L _ m a x . The ideal value of δ heavily depends on the complexity of the task and the architecture of the fusion network. The fusion network’s decision layers D L must be trained sufficiently to avoid falling into a local minimum before targeting a modality utilization for the focus modality.

3.4. Research Questions

The main research questions for this study are as follows:
  • RQ1: Can we leverage the modality utilization metrics during training to regulate a network’s reliance on a dominant modality?
  • RQ2: Does the modality utilization-based training method improve the overall noise robustness of multimodal fusion networks?
Two hypotheses are proposed to address these questions:
Hypothesis 1 (H1.)
The fusion network, trained with modality utilization-based methods, will effectively maintain the utilization of the focus modality within a margin of error of ± 10 % in relation to the target utilization.
Hypothesis 2 (H2.)
The noise robustness of a fusion network, trained using a modality utilization-based training method, will vary depending on the target utilization levels of input modalities when noise is introduced to the input modalities.

4. Experimental Design

4.1. Datasets and Network Architecture

The modality utilization-based training method for multimodal fusion networks has been validated on a classification task using the NTIRE21 dataset [15]. The presented method is not limited to aerial imagery classification problems and can be employed to different domains and machine learning problems. The versatility of the method is showcased on an image segmentation task using the MCubeS dataset [16].

4.1.1. Classification Task

The NTIRE21 dataset [15] presents a classification problem in the domain of aerial imagery. The dataset consists of aerial views of 10 classes of vehicles: sedan, SUV, pickup truck, van, box truck, motorcycle, flatbed truck, bus, pickup truck with trailer, and flatbed truck with trailer. The multimodal dataset features two modalities: (i) Electro-Optical imagery (EO), and (ii) synthetic aperture radar imagery (SAR). Electro-Optical imagery (EO) is a still image photographic sensing, where the incoming light is converted into electrical signals. Synthetic aperture radar (SAR) is an active imaging method, where the sensor uses microwave radar signals emitted towards Earth to capture surface properties. Figure 5 provides a preview of the NTIRE21 dataset.
The EO modality can capture more information during the day with clear skies; however, the data from the SAR modality may be more reliable during cloudy days or at nighttime. Information from different modalities may be more useful in certain scenarios, and the overutilization of a single dominant modality has the potential to drastically degrade the multimodal network’s performance with noisy data.
Unimodal networks for EO and SAR modalities were trained with ResNet18 [55] as the backbone of the network to obtain the pre-trained feature extractors for the multimodal fusion network. The multimodal fusion network (see Figure 6) consisted of two ResNet18-based pre-trained feature extractor branches for each modality, followed by a flattening layer, concatenation, and a fully connected layer with 512 neurons and a softmax activation function for classification. The network was trained over 250 epochs with an Adam optimizer and a learning rate of 0.001. A significant class imbalance exists in the NTIRE21 dataset, with the Sedan class consisting of 234,209 samples while the Bus class consisting of only 624 samples. To address this, only the first 624 samples from each class were used for this study. Specifically, 524 samples from each class were used for training, while 100 samples were used for testing. Models with the highest classification accuracy and the lowest modality utilization loss ( L m u ) are saved for further analysis. Additionally, early stopping is employed to prevent overfitting. Figure 7 provides an intuitive demonstration of the predicted classification on the NTIRE21 dataset using the Multimodal Aerial View Object Classification Network.

4.1.2. Image Segmentation Task

The MCubeS dataset [16] presents an image segmentation task in the domain of material segmentation in street scenes. The dataset consists of street scenes from a viewpoint on a road, pavement, or sidewalk containing 20 classes of materials: asphalt, concrete, metal, road marking, fabric, glass, plaster, plastic, rubber, sand, gravel, ceramic, cobblestone, brick, grass, wood, leaf, water, human body, and sky. The multimodal dataset features four modalities: (i) RGB camera (RGB), (ii) Angle of Polarization (AoLP), (iii) Degree of Polarization (DoLP), and (iv) Near Infrared (NIR). Figure 8 provides a preview of the MCubeS dataset. The RGB modality alone cannot capture the necessary information to identify different materials in an image. Different lighting conditions may make certain materials indistinguishable. Thus, other modalities aid in enhancing the overall performance of the multimodal image segmentation network.
Unimodal networks for the RGB, AoLP, DoLP, and NIR modalities were trained with U-Net [56] as the backbone of the network to obtain the pre-trained feature extractors for the multimodal fusion network. The multimodal fusion network (see Figure 9) consisted of four UNet-based pre-trained feature extractor branches for each modality, followed by a concatenation layer, two batch normalization and 2D convolution layers alternately, and a ReLU activation function for image segmentation. The two 2D convolution layers comprised a 3 × 3 × 300 and 1 × 1 × 20 filter size, respectively, with a stride of 1. Training lasted 500 epochs with SGD optimizer (lr: 0.05, momentum: 0.9) on a dataset split into 302 training, 96 validation, and 102 testing samples. Models with the highest mean Intersection over Union (mIoU) and the lowest modality utilization loss ( L m u ) are saved for further analysis. Additionally, early stopping is employed to prevent overfitting. Figure 10 provides an intuitive demonstration of the image segmentation on the MCubeS dataset using the Multimodal Material Segmentation Network.

5. Results

5.1. Ablation Studies

The presented modality utilization-based training method is first validated on the NTIRE21 dataset and MCubeS dataset without a loss factor warm-up. All experiments were conducted on Rochester Institute of Technology’s research computing server [57] equipped with an Intel Xeon Gold 6150 CPU running at 2.7 GHz, 32 GB of RAM and NVIDIA P4 and A100 GPU’s for NTIRE21 dataset and MCubeS dataset, respectively. The software environment included Red Hat Enterprise Linux 7 as the operating system, Python 3.8.11, and PyTorch 1.10.1 for the deep learning framework. When trained using the traditional method, the multimodal fusion network in NTIRE21 relies heavily on the EO modality, making it the dominant modality (see Table 1) [14]. The utilization of the fusion network in the MCubeS dataset is more evenly distributed when trained using the traditional training method, with RGB as the most utilized modality, while NIR is the least utilized modality. These modality utilization measures are used as the baseline for this study. The NTIRE21 network overemphasizes the EO modality while not utilizing the SAR modality at all. Such a utilization imbalance can be rectified using the presented modality utilization-based training method.
Given a target utilization M U t a r g and loss factor λ L , the utilization of a multimodal fusion network on the input focus modality m f can be manipulated using Algorithm 2. The classification network for NTIRE21 dataset is trained with SAR as the focus modality ( m f = S A R ), and loss factor λ L = 100 for different values of target utilization M U t a r g , shown in Figure 11. The network can be observed maintaining the SAR modality utilization M U S A R (green line) close to the target utilization M U t a r g (black dashed line) while maintaining a performance similar to the baseline methods. The network performance converges to the baseline performance quickly since pre-trained frozen feature extractors are used for this study. Table 2 summarizes the performance and the modality utilization of the network on EO and SAR modalities achieved for different M U t a r g . A drop in performance is observed when the network starts to rely heavily on the non-dominant modality. The performance is bounded by the unimodal SAR performance at M U t a r g = 100.0 % . A small difference can be noticed between target modality utilization SAR M U t a r g and M U S A R as the modality utilization-based training method minimizes the difference between M U t a r g and M U c u r r while also solving the fusion network task. A mean difference of 3.73% can be observed between SAR M U t a r g and M U S A R with a maximum of 7.10% and minimum of 0.59% difference.
Since the measure of MU is a percentage utilization with respect to all the input modalities, the utilization of the EO modality reduces with increasing utilization of the SAR modality. Similar network behavior can be observed in Table 3, where the experiment is repeated with EO as the focus modality ( m f = E O ). A mean difference of 3.76% can be observed between EO M U t a r g and M U E O with a maximum of 6.90% and minimum of 0.50% difference. The effects of information redundancy in the different modalities can also be observed in Table 2 and Table 3. As the utilization of the dominant modality decreases, the utilization of the non-dominant modality increases; however, the accuracy of the model stays unchanged for M U E O > 50 % . Redundant information allows for the reduction in the utilization of one modality up to the point where the unique information in that modality begins to be compromised.
The value of the loss factor λ L determines the extent to which the MU-based training method emphasizes solving the fusion network task and maintaining the current modality utilization M U c u r r close to the target utilization M U t a r g . The effects of the loss factor λ L on the modality utilization and network performance are demonstrated in Figure 12. The NTIRE21 fusion network was trained with m f = S A R , M U t a r g = 50.0 % , and λ L = 0.0 , 20.0 , 100.0 , 10,000.0 . The MU-based training method works like the traditional machine learning training method with λ L = 0 , eliminating the modality utilization loss term from Equation (5). The NTIRE21 fusion network thus behaves similarly to the baseline network with λ L = 0 . As the value of λ L is increased, the network maintains the M U c u r r closer to M U t a r g . However, a really high value of λ L can lead to catastrophic failure, as the network only focuses on maintaining modality utilization, completely ignoring the task-specific loss as seen in Figure 12 with an extreme value of λ L = 10,000.0 .
The modality utilization-based training method can be applied to a diverse range of multimodal machine learning applications. The method is also validated on the MCubeS image segmentation dataset. The multimodal network is trained for target utilization M U t a r g e t with a loss factor λ L = 100.0 , and RGB (the dominant modality) as the focus modality while observing the mean Intersection over Union (mIoU) and modality utilization of the four input modalities. The results in Table 4 show that the presented method is able to drive the utilization of the network close to the set target for the RGB while maintaining good performance. Since RGB is used as the focus modality, the utilization of the other three modalities is decided by the multimodal network. A mean difference of 4.61% can be observed between RGB M U t a r g and M U R G B with a maximum of 9.90% and minimum of 1.20% difference. Contrary to modality utilization results from the previous study [14] in Table 1, AoLP appears to be the non-dominant modality instead of the NIR modality with consistent low utilization. Across Table 2, Table 3 and Table 4, a mean difference of 4.03% can be observed between M U t a r g and M U c u r r with a maximum of 9.90% and minimum of 0.50% difference.
The results presented in this section were generated over a single fold without modulating λ L via loss factor warm-up. As the study is scaled to multifold validation, instability in the modality utilization-based training method can be observed. A five-fold validation for modality utilization-based training methods on the NTIRE21 dataset with m f = S A R , λ L = 100.0 , and M U t a r g = 0.0 reveals that the training method can become unstable with random utilization and data splits, highlighted in red in Table 5.
An untrained network is prone to getting pushed into a local minima by the modality utilization loss term in Equation (5). The loss factor warm-up presented in Section 3.3 becomes necessary to stabilize the modality utilization-based training method.

5.2. Validation Loss Factor Warm-Up

The network parameters with pre-trained feature extractors initially have a smaller task-specific loss, while they may have a larger modality utilization loss. A much higher modality utilization loss in the beginning can push the network away from the minima. The loss factor warm-up allows the network to improve the performance on the network task by keeping λ L = 0 prior to exponentially increasing the λ L value. The five-fold study on the NTIRE21 dataset shown in Table 5 is repeated with the loss factor warm-up to validate the loss factor warm start-up. Since the NTIRE21 dataset uses pre-trained frozen data, the network parameters can quickly converge to the global minimum. Thus, a buildup rate of β = 0.5 and a buildup delay of δ = 0.0 are used, achieving stable performance with consistent results, demonstrated in Table 6.
The buildup delay δ initially suppresses the modality utilization-based training, focusing on the fusion network. Since the feature extractors in the study are pre-trained, a value of δ = 0 is used. The buildup rate β dictates how quickly the loss factor λ L builds up from 0 to λ L _ m a x in Equation (7). A large value would aggressively drive λ L to the maximum value, making the network focus on maintaining the target utilization early. A smaller value would allow the network to focus on the fusion network task for longer; however, it focuses less on maintaining the target utilization. Figure 13 shows the effects of different buildup rates β on the NTIRE21 classification dataset. A slower building with β = 0.01 shows a weaker tendency to maintain the target utilization at SAR M U t a r g = 50 % .

5.3. Studying Noise Robustness Properties of the Modality Utilization-Based Training Method

The presence of noise in the data can negatively impact the performance of the multimodal network. Noise in one or more modalities may be introduced due to environmental changes or sensor failure. The overutilization of a singular dominant modality makes the multimodal network susceptible to noise in the dominant modality while underutilizing the non-dominant modality. In the case of the NTIRE21 dataset, the EO modality is the dominant modality; however, it is a passive sensing modality that cannot provide usable data during nighttime or on a cloudy day. A more balanced utilization of the input modalities can provide robustness against noise in a singular dominant modality. Modality utilization-based training provides a method to guide the utilization of the multimodal network on its input modalities, improving the network’s overall robustness against noise.
A noise ablation study was conducted with the NTIRE21 dataset by adding Gaussian noise with mean 0 and variance of { 0.06 , 0.09 , 0.12 } to the EO modality, the SAR modality, and both modalities during inference. The effects of noise were studied on networks trained with different levels of SAR utilization, i.e., 0.0%, 12.5%, 25.00%, 37.5%, 50.0%, 62.5%, 75.0%, 87.5%, and 100.0%. The results in Figure 14 show that the overall accuracy is the highest with Clean EO and Clean SAR modalities (indicated by blue) and the lowest with Noisy EO and Noisy SAR (indicated by red). This behavior is expected, as the addition of noise in the data degrades the performance of the network.
When noise is added to only the SAR modality (indicated by green), the performance of models trained with higher SAR utilization is worse than that of the models trained to have higher utilization with the EO modality. Similarly, when noise is added to only the EO modality (indicated by orange), the performance of models trained with higher SAR utilization is better than that of the models trained to have higher utilization with the EO modality. Figure 14 further reveals that network with M U E O = 75 % and M U S A R = 25 % performs almost the same as the Noisy-EO/Clean-SAR and Clean-EO/Noisy-SAR input modalities. The network exhibits better robustness against noise in either of the modalities compared to networks with other utilizations without significant degradation in performance, indicated by the clean EO and clean SAR performance. This property can be noticed across the different noise levels introduced to the data during inference time.
The fusion network trained with the traditional method self-optimizes to utilize 99.59% of the dominant EO modality and 0.40% of SAR non-dominant SAR modality as indicated in Table 1. Table 7 demonstrates that when noise is present in the dominant modality during inference time, a 75.0% to 87.5% utilization of the EO modality offers better robustness towards noisy dominant modality with minimal loss in accuracy. Compared to traditional training methods that achieve 99.59% EO utilization with 73.7% accuracy under heavy noise conditions (noise variance = 0.12), a network utilizing 75.0% EO performs significantly better with 81.4% accuracy. The network (EO M U t a r g = 75.0 % ) also achieves an average accuracy of 85.0% across various noise levels, exceeding the traditional method’s average accuracy of 81.9%.

5.4. Determining Target Modality Utilization

The target modality utilization M U t a r g for the focus modality m f is a critical hyperparameter, as it dictates how much the trained network will rely on the focus modality m f . Overutilization of the non-dominant modality or underutilization of the dominant modality can lead to network performance degradation. On the other hand, overutilization of the dominant modality or underutilization of the non-dominant modality can affect the robustness to noise in the dominant modality. Thus, choosing an appropriate M U t a r g is critical; however, it is a non-trivial problem to determine an appropriate target utilization M U t a r g value without observing the behavior of the network.
Problems such as mismatched hyperparameters across different multimodal branches or overutilization of a single dominant modality are not present in unimodal networks, which are trained on data from a single modality. However, their performance is limited to the information present in those modalities. Thus, unimodal network performance can provide heuristic insights into how well a modality represents information in the context of the fusion network task. Additionally, a multimodal network trained using traditional training methods can be used to determine the multimodal network’s utilization tendencies using the modality utilization metric as seen in Table 1.
Observing unimodal and multimodal network performance that is trained using traditional training methods provides insight into network utilization and information in each modality as viewed by the fusion network. This can guide the selection of the target modality utilization M U t a r g for the focus modality m f while balancing the utilization of the dominant and the non-dominant modalities. Furthermore, an array of noise robustness tests can reveal certain target utilizations, where the network may be more robust against noise present in one or multiple modalities as shown in Section 5.3. This is crucial for aerial imagery, where varying weather conditions can affect modality usability, especially in datasets lacking diverse weather conditions during training.

6. Discussion

The modality utilization-based training method aims to regulate the utilization of multimodal networks on their input modalities. Hypothesis H1 predicts that the fusion network, trained with modality utilization-based methods, will effectively maintain the utilization of the focus modality within a margin of error of ± 10 % in relation to the target utilization. In the results in Table 2, Table 3 and Table 4, a mean difference of 4.03% can be observed between M U t a r g and M U c u r r with a maximum of 9.90% and minimum of 0.50% difference, thus supporting hypothesis H1. The presented method successfully leverages the modality utilization metric to encourage the multimodal network to minimize the mean squared error between the current utilization and maintain a set target for a focus modality. The method alleviates the problem of the overutilization of a singular dominant modality by balancing the utilization among the different input modalities.
However, there needs to be an equilibrium between the network focusing on solving the fusion network task and maintaining the target modality utilization, dictated by a loss factor λ L . The network optimizes over two different goals, which can drive the network parameters in two different directions. The initial phase of the training process is particularly sensitive, and an instability is observed in the modality utilization-based training method as shown in the results in Table 5. A delayed implementation of the modality utilization-based training method through a warm startup stabilizes the training process, achieving consistent performance.
The aerial imagery classification task with the NTIRE21 dataset presents a case study where the dataset inherently causes a multimodal network to utilize only the dominant EO modality while ignoring the non-dominant SAR modality. The EO modality can provide more information during the daytime with clear sky. However, the SAR modality is more optimal for cloudy weather conditions since the active sensing method can capture information through the clouds. The absence of cloudy data samples leaves out crucial information from the network, driving the network to rely on a singular modality. Expert designers can employ the modality utilization-based training method to train the network correctly even with an imperfect dataset.
The method allows the multimodal network to be more robust towards the presence of noise, such as clouds and other weather conditions, where the EO modality becomes suboptimal. Hypothesis H2 predicts that the noise robustness of a fusion network, trained using a modality utilization-based training method, will vary depending on the target utilization levels of input modalities when noise is introduced to the input modalities. The results in Figure 14 demonstrate that the presented method can be used to make a network more robust to the presence of noise in one or more modalities by making it less reliant on a dominant modality. The network trained for lower SAR modality utilization performs better than that with higher SAR modality utilization when noise is present in the SAR modality. A similar trend is observed with the EO modality when noise is present in it. This is further validated by Table 7, where a decrease in the utilization of the dominant EO modality decreases the network accuracy when no noise is present. However, the network exhibits a greater degradation in accuracy with higher EO utilization when noise is present in the dominant EO modality. The average optimal EO utilization is revealed to be in the range of 75–87.5%, leveraging the information from the dominant modality while avoiding the overutilization of it. This is an improvement over the traditional training method, which self-optimizes its utilization to be 99.59% utilization of the dominant EO modality, shown in Table 1, making the network susceptible to noise. The array of tests reveals that the network’s performance is degraded by different levels based on the target utilization set for that network, thus supporting hypothesis H2. This indicates that there exists an EO-SAR utilization ratio where the network will be equally robust to the presence of noise in the EO or SAR modality, making it the ideal point to improve the overall network noise robustness.
Choosing the optimal target utilization for a modality is critical, as higher utilization of a non-dominant modality can result in poor performance, while overutilization of a singular dominant modality can undermine the information gain from multiple modalities. The unimodal performance of the network for each modality provides heuristic insight into the amount of relevant information in that modality. Furthermore, a multimodal network trained on all the modalities with traditional training methods can reveal how much the network tends to rely on one or the other modality using the modality utilization metric. The modality utilization metric, along with the unimodal performance, offers a holistic understanding of the network and guides the decision of choosing the appropriate target utilization. The aim while choosing the target utilization is to mitigate any overutilization of the network on a single modality, and the target utilization selection process must be motivated by this aim. The target utilization selection can also be made based on the noise robustness properties as demonstrated in Section 5.3; however, this requires conducting an array of tests whose design may be impacted by the number of modalities in the dataset.
The presented method was also validated on an image segmentation dataset in the domain of multimodal material segmentation. This showcases the versatility and applicability of the modality utilization-based training method to a diverse set of multimodal machine learning applications. Furthermore, this method is not limited to multimodal supervised learning applications and can be extended to the domain of reinforcement learning, offering promising avenues for future research and practical implementation in various domains. Overall, this study contributes significantly to advancing multimodal fusion networks, enhancing the utilization of diverse data modalities in machine learning applications.

7. Conclusions and Future Work

In conclusion, this work presented a modality utilization-based training method that can be employed to guide the utilization of a multimodal fusion network on its input modalities. The method leverages the modality utilization metric and introduces a modality utilization loss term to minimize the error between the current utilization of a focus modality and a set target. The method was validated on an aerial imagery classification dataset and an image segmentation dataset. The results showed that the presented method can successfully influence a multimodal network’s utilization of its input modalities, effectively maintains modality utilization within ± 10 % of the user-defined target utilization. Moreover, the network’s robustness against noise in the input modalities was studied, a prevalent challenge in practical scenarios. The method demonstrated higher resilience to input noise affecting different sensing modalities in the NTIRE21 dataset, further enhancing its practical utility. Specifically, networks trained with 75.0% EO utilization exhibited better accuracy (81.4%) under noisy conditions (noise variance = 0.12) compared to traditional methods utilizing 99.59% EO utilization (73.7%). Furthermore, the network maintained an average accuracy of 85.0% across varying noise levels, outperforming the traditional method’s average accuracy of 81.9%. Key contributions include the development of a modality utilization-based training framework, tailored to address utilization imbalances in multimodal fusion networks. The study also offers insights into enhancing network robustness against input noise, advancing the practical utility of multimodal systems.
Future work will validate the approach on networks trained end to end, without relying on pre-trained feature extractors. Additionally, conducting thorough validations on datasets with more than two modalities will enhance the understanding of the interactions between modality utilization and the information content in each modality. Exploring the method’s application with target utilization for multiple modalities and extending it to reinforcement learning are promising avenues for further research. Overall, the findings of this study represent a significant step towards realizing the full potential of multimodal data fusion in machine learning applications, offering promising avenues for future research and practical implementation in diverse domains.

Author Contributions

Conceptualization, S.S., E.S., P.P.M. and J.H.; methodology, S.S.; software, S.S.; validation, S.S.; formal analysis, S.S.; investigation, S.S.; resources, E.S., P.P.M. and J.H.; writing—original draft preparation, S.S.; writing—review and editing, E.S., P.P.M. and J.H.; visualization, S.S. and J.H.; supervision, E.S., P.P.M. and J.H.; project administration, J.H.; funding acquisition, E.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by an academic grant from the National Geospatial-Intelligence Agency (Grant No. HM04761912014, Project Title: Target Detection/Tracking and Activity Recognition from Multimodal Data.) Any opinions, findings and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of NGA, DoD, or the US government. Approved for public release, NGA-U-2024-01753. This material is based upon work supported in part by the Air Force Office of Scientific Research (AFOSR) under award FA9550-20-1-0039. This material is based upon work supported by the National Science Foundation under Award Numbers DGE-2125362 and 2332744. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were used in this study. NTIRE21 Dataset: Access can be requested from here: https://competitions.codalab.org/competitions/28095 (accessed on 19 May 2022); MCubeS Dataset: Can be accessed from here: https://github.com/kyotovision-public/multimodal-material-segmentation (accessed on 19 September 2022).

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
MUModality Utilization
EOElectro-Optical
SARSynthetic Aperture Radar
RGBRed, Green and Blue
AoLPAngle of Polarization
DoLPDegree of Polarization
NIRNear Infra-red
targTarget
currCurrent
FEFeature Extractor
m f Focus Modality
DLDecision Layer
MModality
LLoss
D Dataset
F θ Network parameters

References

  1. Xiao, Y.; Codevilla, F.; Gurram, A.; Urfalioglu, O.; López, A.M. Multimodal end-to-end autonomous driving. IEEE Trans. Intell. Transp. Syst. 2020, 23, 537–547. [Google Scholar] [CrossRef]
  2. Sharma, M.; Dhanaraj, M.; Karnam, S.; Chachlakis, D.G.; Ptucha, R.; Markopoulos, P.P.; Saber, E. YOLOrs: Object detection in multimodal remote sensing imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 14, 1497–1508. [Google Scholar] [CrossRef]
  3. Doherty, K.; Fourie, D.; Leonard, J. Multimodal semantic slam with probabilistic data association. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 2419–2425. [Google Scholar]
  4. Papanastasiou, S.; Kousi, N.; Karagiannis, P.; Gkournelos, C.; Papavasileiou, A.; Dimoulas, K.; Baris, K.; Koukas, S.; Michalos, G.; Makris, S. Towards seamless human robot collaboration: Integrating multimodal interaction. Int. J. Adv. Manuf. Technol. 2019, 105, 3881–3897. [Google Scholar] [CrossRef]
  5. Guo, Z.; Li, X.; Huang, H.; Guo, N.; Li, Q. Deep learning-based image segmentation on multimodal medical imaging. IEEE Trans. Radiat. Plasma Med. Sci. 2019, 3, 162–169. [Google Scholar] [CrossRef]
  6. He, M.; Hohil, M.; LaPeruta, T.; Nashed, K.; Lawrence, V.; Yao, Y.D. Performance evaluation of multimodal deep learning: Object identification using uav dataset. In Proceedings of the Artificial Intelligence and Machine Learning for Multi-Domain Operations Applications III, Online, 12–16 April 2021; SPIE: Bellingham, WA, USA, 2021; Volume 11746, pp. 602–608. [Google Scholar]
  7. Chen, W.; Li, X.; Wang, L. Multimodal Remote sensing science and technology. In Remote Sensing Intelligent Interpretation for Mine Geological Environment: From land Use and Land Cover Perspective; Springer: Berlin/Heidelberg, Germany, 2022; pp. 7–32. [Google Scholar]
  8. Liu, G.; Peng, B.; Liu, T.; Zhang, P.; Yuan, M.; Lu, C.; Cao, N.; Zhang, S.; Huang, S.; Wang, T.; et al. Large-Scale Fine-Grained Building Classification and Height Estimation for Semantic Urban Reconstruction: Outcome of the 2023 IEEE GRSS Data Fusion Contest. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 11194–11207. [Google Scholar] [CrossRef]
  9. Hänsch, R.; Arndt, J.; Lunga, D.; Pedelose, T.; Boedihardjo, A.; Pfefferkorn, J.; Petrie, D.; Bacastow, T.M. SpaceNet 8: Winning Approaches to Multi-Class Feature Segmentation from Satellite Imagery for Flood Disasters. In Proceedings of the IGARSS 2023–2023 IEEE International Geoscience and Remote Sensing Symposium, Pasadena, CA, USA, 16–21 July 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1241–1244. [Google Scholar]
  10. Low, S.; Nina, O.; Sappa, A.D.; Blasch, E.; Inkawhich, N. Multi-modal aerial view object classification challenge results-PBVS 2023. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 412–421. [Google Scholar]
  11. Meng, T.; Jing, X.; Yan, Z.; Pedrycz, W. A survey on machine learning for data fusion. Inf. Fusion 2020, 57, 115–129. [Google Scholar] [CrossRef]
  12. Wu, N.; Jastrzebski, S.; Cho, K.; Geras, K.J. Characterizing and overcoming the greedy nature of learning in multi-modal deep neural networks. In Proceedings of the International Conference on Machine Learning, Baltimore, MD, USA, 17–23 July 2022; PMLR: London, UK, 2022; pp. 24043–24055. [Google Scholar]
  13. Li, J.; Hong, D.; Gao, L.; Yao, J.; Zheng, K.; Zhang, B.; Chanussot, J. Deep learning in multimodal remote sensing data fusion: A comprehensive review. Int. J. Appl. Earth Obs. Geoinf. 2022, 112, 102926. [Google Scholar] [CrossRef]
  14. Singh, S.; Markopoulos, P.P.; Saber, E.; Lew, J.D.; Heard, J. Measuring Modality Utilization in Multi-Modal Neural Networks. In Proceedings of the 2023 IEEE Conference on Artificial Intelligence (CAI), Santa Clara, CA, USA, 5–6 June 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 11–14. [Google Scholar]
  15. Pérez-Pellitero, E.; Catley-Chandar, S.; Leonardis, A.; Timofte, R. NTIRE 2021 challenge on high dynamic range imaging: Dataset, methods and results. In Proceedings of the 2021 IEEE/CVF Conf. on Computer Vision and Pattern Recognition Workshops, CVPRW 2021, Online, 19–25 June 2021; IEEE Computer Society: Washington, DC, USA, 2021; pp. 691–700. [Google Scholar] [CrossRef]
  16. Liang, Y.; Wakaki, R.; Nobuhara, S.; Nishino, K. Multimodal Material Segmentation. In Proceedings of the IEEE/CVF Conf. on Comput. Vision and Patt. Recogn. (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 19800–19808. [Google Scholar]
  17. Chakraborty, J.; Stolinski, M. Signal-Level Fusion Approach for Embedded Ultrasonic Sensors in Damage Detection of Real RC Structures. Mathematics 2022, 10, 724. [Google Scholar] [CrossRef]
  18. Cai, H.; Qu, Z.; Li, Z.; Zhang, Y.; Hu, X.; Hu, B. Feature-level fusion approaches based on multimodal EEG data for depression recognition. Inf. Fusion 2020, 59, 127–138. [Google Scholar] [CrossRef]
  19. Boulahia, S.Y.; Amamra, A.; Madi, M.R.; Daikh, S. Early, intermediate and late fusion strategies for robust deep learning-based multimodal action recognition. Mach. Vis. Appl. 2021, 32, 121. [Google Scholar] [CrossRef]
  20. Su, Y.; Zhang, K.; Wang, J.; Madani, K. Environment sound classification using a two-stream CNN based on decision-level fusion. Sensors 2019, 19, 1733. [Google Scholar] [CrossRef] [PubMed]
  21. Ding, J.; Xue, N.; Xia, G.S.; Bai, X.; Yang, W.; Yang, M.Y.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; et al. Object detection in aerial images: A large-scale benchmark and challenges. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 7778–7796. [Google Scholar] [CrossRef] [PubMed]
  22. Yang, F.; Fan, H.; Chu, P.; Blasch, E.; Ling, H. Clustered object detection in aerial images. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8311–8320. [Google Scholar]
  23. Chen, C.; Zhong, J.; Tan, Y. Multiple-oriented and small object detection with convolutional neural networks for aerial image. Remote Sens. 2019, 11, 2176. [Google Scholar] [CrossRef]
  24. Hatamizadeh, A.; Sengupta, D.; Terzopoulos, D. End-to-end trainable deep active contour models for automated image segmentation: Delineating buildings in aerial imagery. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XII 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 730–746. [Google Scholar]
  25. Yue, K.; Yang, L.; Li, R.; Hu, W.; Zhang, F.; Li, W. TreeUNet: Adaptive tree convolutional neural networks for subdecimeter aerial image segmentation. ISPRS J. Photogramm. Remote Sens. 2019, 156, 1–13. [Google Scholar] [CrossRef]
  26. Guan, Z.; Miao, X.; Mu, Y.; Sun, Q.; Ye, Q.; Gao, D. Forest fire segmentation from aerial imagery data using an improved instance segmentation model. Remote Sens. 2022, 14, 3159. [Google Scholar] [CrossRef]
  27. Kyrkou, C.; Theocharides, T. Deep-Learning-Based Aerial Image Classification for Emergency Response Applications Using Unmanned Aerial Vehicles. In Proceedings of the CVPR Workshops, Long Beach, CA, USA, 16–20 June 2019; pp. 517–525. [Google Scholar]
  28. Zheng, X.; Yuan, Y.; Lu, X. A deep scene representation for aerial scene classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 4799–4809. [Google Scholar] [CrossRef]
  29. Cheng, G.; Xie, X.; Han, J.; Guo, L.; Xia, G.S. Remote sensing image scene classification meets deep learning: Challenges, methods, benchmarks, and opportunities. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 3735–3756. [Google Scholar] [CrossRef]
  30. Gómez-Chova, L.; Tuia, D.; Moser, G.; Camps-Valls, G. Multimodal classification of remote sensing images: A review and future directions. Proc. IEEE 2015, 103, 1560–1584. [Google Scholar] [CrossRef]
  31. Huang, Z.; Li, W.; Tao, R. Multimodal knowledge distillation for arbitrary-oriented object detection in aerial images. In Proceedings of the ICASSP 2023—2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; IEEE: Piscaway, NJ, USA, 2023; pp. 1–5. [Google Scholar]
  32. Singh, S.; Sharma, M.; Heard, J.; Lew, J.D.; Saber, E.; Markopoulos, P.P. Multimodal aerial view object classification with disjoint unimodal feature extraction and fully connected-layer fusion. In Proceedings of the Big Data V: Learning, Analytics, and Applications, Orlando, FL, USA, 1 April–5 May 2023; SPIE: Bellingham, WA, USA, 2023; Volume 12522, p. 1252206. [Google Scholar]
  33. Xiang, Y.; Tian, X.; Xu, Y.; Guan, X.; Chen, Z. EGMT-CD: Edge-Guided Multimodal Transformers Change Detection from Satellite and Aerial Images. Remote Sens. 2023, 16, 86. [Google Scholar] [CrossRef]
  34. Huang, Y.; Lin, J.; Zhou, C.; Yang, H.; Huang, L. Modality competition: What makes joint training of multi-modal network fail in deep learning? (provably). In Proceedings of the International Conference on Machine Learning, Baltimore, MD, USA, 17–23 July 2022; PMLR: London, UK, 2022; pp. 9226–9259. [Google Scholar]
  35. Hafner, S.; Ban, Y.; Nascetti, A. Investigating Imbalances Between SAR and Optical Utilization for Multi-Modal Urban Mapping. In Proceedings of the 2023 Joint Urban Remote Sensing Event (JURSE), Heraklion, Greece, 17–19 May 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–4. [Google Scholar]
  36. Ghahremani Boozandani, M.; Wachinger, C. RegBN: Batch Normalization of Multimodal Data with Regularization. In Proceedings of the Advances in Neural Information Processing Systems 36, New Orleans, LA, USA, 10–16 December 2023. [Google Scholar]
  37. Gat, I.; Schwartz, I.; Schwing, A.; Hazan, T. Removing bias in multi-modal classifiers: Regularization by maximizing functional entropies. In Proceedings of the Advances in Neural Information Processing Systems 33, Online, 6–12 December 2020; pp. 3197–3208. [Google Scholar]
  38. Ma, H.; Zhang, Q.; Zhang, C.; Wu, B.; Fu, H.; Zhou, J.T.; Hu, Q. Calibrating multimodal learning. In Proceedings of the International Conference on Machine Learning, Honolulu, HI, USA, 23–29 July 2023; PMLR: London, UK; pp. 23429–23450. [Google Scholar]
  39. Cao, Y.; Bin, J.; Hamari, J.; Blasch, E.; Liu, Z. Multimodal object detection by channel switching and spatial attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 403–411. [Google Scholar]
  40. He, Y.; Cheng, R.; Balasubramaniam, G.; Tsai, Y.H.H.; Zhao, H. Efficient Modality Selection in Multimodal Learning. J. Mach. Learn. Res. 2024, 25, 1–39. [Google Scholar]
  41. Sun, Y.; Mai, S.; Hu, H. Learning to balance the learning rates between various modalities via adaptive tracking factor. IEEE Signal Process. Lett. 2021, 28, 1650–1654. [Google Scholar] [CrossRef]
  42. Khan, S.; Naseer, M.; Hayat, M.; Zamir, S.W.; Khan, F.S.; Shah, M. Transformers in vision: A survey. ACM Comput. Surv. (CSUR) 2022, 54, 1–41. [Google Scholar] [CrossRef]
  43. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems 30, Long Beach, CA, USA, 4–7 December 2017. [Google Scholar]
  44. Xu, P.; Zhu, X.; Clifton, D.A. Multimodal learning with transformers: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 12113–12132. [Google Scholar] [CrossRef] [PubMed]
  45. Roy, S.K.; Deria, A.; Hong, D.; Rasti, B.; Plaza, A.; Chanussot, J. Multimodal fusion transformer for remote sensing image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5515620. [Google Scholar] [CrossRef]
  46. Huang, J.; Tao, J.; Liu, B.; Lian, Z.; Niu, M. Multimodal transformer fusion for continuous emotion recognition. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 3507–3511. [Google Scholar]
  47. Boussioux, L.; Zeng, C.; Guénais, T.; Bertsimas, D. Hurricane forecasting: A novel multimodal machine learning framework. Weather. Forecast. 2022, 37, 817–831. [Google Scholar] [CrossRef]
  48. Luo, Y.; Guo, X.; Dong, M.; Yu, J. Learning Modality Complementary Features with Mixed Attention Mechanism for RGB-T Tracking. Sensors 2023, 23, 6609. [Google Scholar] [CrossRef]
  49. Ivanov, A.; Dryden, N.; Ben-Nun, T.; Li, S.; Hoefler, T. Data movement is all you need: A case study on optimizing transformers. Proc. Mach. Learn. Syst. 2021, 3, 711–732. [Google Scholar]
  50. Wang, W.; Zhang, J.; Cao, Y.; Shen, Y.; Tao, D. Towards data-efficient detection transformers. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2022; pp. 88–105. [Google Scholar]
  51. Quan, Y.; Zhang, R.; Li, J.; Ji, S.; Guo, H.; Yu, A. Learning SAR-Optical Cross Modal Features for Land Cover Classification. Remote Sens. 2024, 16, 431. [Google Scholar] [CrossRef]
  52. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  53. Fisher, A.; Rudin, C.; Dominici, F. All models are wrong, but many are useful: Learning a variable’s importance by studying an entire class of prediction models simultaneously. J. Mach. Learn. Res. JMLR 2019, 20, 177. [Google Scholar]
  54. Singh, S.; Heard, J. Measuring State Utilization During Decision Making in Human-Robot Teams. In Proceedings of the Companion of the 2024 ACM/IEEE International Conference on Human-Robot Interaction, Boulder, CO, USA, 11–15 March 2024; pp. 985–989. [Google Scholar]
  55. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  56. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18. Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  57. Rochester Institute of Technology. Research Computing Services. 2023. Available online: https://www.rit.edu/researchcomputing/ (accessed on 19 May 2022). [CrossRef]
Figure 1. Computing modality utilization by randomly shuffling a modality M i within the dataset to break the association between the input modality M i and the output Y.
Figure 1. Computing modality utilization by randomly shuffling a modality M i within the dataset to break the association between the input modality M i and the output Y.
Sensors 24 06054 g001
Figure 2. Modality utilization-based training targets the decision layers while using pre-trained feature extractors with frozen weights.
Figure 2. Modality utilization-based training targets the decision layers while using pre-trained feature extractors with frozen weights.
Sensors 24 06054 g002
Figure 3. Visualization of multimodal fusion network’s gradient descent on the loss surface of the fusion network task. Optimizing L m u from the very beginning can push the network in the local minima.
Figure 3. Visualization of multimodal fusion network’s gradient descent on the loss surface of the fusion network task. Optimizing L m u from the very beginning can push the network in the local minima.
Sensors 24 06054 g003
Figure 4. Clipped exponential function-based for loss factor warm-up for MU-based training.
Figure 4. Clipped exponential function-based for loss factor warm-up for MU-based training.
Sensors 24 06054 g004
Figure 5. NTIRE 2021 Multimodal Aerial View Object Classification Challenge Dataset [15].
Figure 5. NTIRE 2021 Multimodal Aerial View Object Classification Challenge Dataset [15].
Sensors 24 06054 g005
Figure 6. NTIRE 2021 Multimodal Aerial View Object Classification Network with ResNet18 as the backbone.
Figure 6. NTIRE 2021 Multimodal Aerial View Object Classification Network with ResNet18 as the backbone.
Sensors 24 06054 g006
Figure 7. Visualization of NTIRE21 dataset classification using Multimodal Aerial View Object Classification Network.
Figure 7. Visualization of NTIRE21 dataset classification using Multimodal Aerial View Object Classification Network.
Sensors 24 06054 g007
Figure 8. MCubeS Multimodal Material Segmentation Dataset [16].
Figure 8. MCubeS Multimodal Material Segmentation Dataset [16].
Sensors 24 06054 g008
Figure 9. MCubeS Multimodal Material Segmentation Network with UNet as the backbone.
Figure 9. MCubeS Multimodal Material Segmentation Network with UNet as the backbone.
Sensors 24 06054 g009
Figure 10. Visualization of MCubeS dataset image segmentation using the Multimodal Material Segmentation Network.
Figure 10. Visualization of MCubeS dataset image segmentation using the Multimodal Material Segmentation Network.
Sensors 24 06054 g010
Figure 11. Effects of different target utilization M U t a r g e t on modality utilization and classification accuracy with modality utilization-based training method in the NTIRE21 dataset. Loss factor λ L = 100.0 with SAR as the focus modality.
Figure 11. Effects of different target utilization M U t a r g e t on modality utilization and classification accuracy with modality utilization-based training method in the NTIRE21 dataset. Loss factor λ L = 100.0 with SAR as the focus modality.
Sensors 24 06054 g011
Figure 12. Effects of the loss factor λ L on modality utilization and classification accuracy with modality utilization-based training method in NTIRE21 dataset. Target utilization M U t a r g e t = 50 % with SAR as the focus modality.
Figure 12. Effects of the loss factor λ L on modality utilization and classification accuracy with modality utilization-based training method in NTIRE21 dataset. Target utilization M U t a r g e t = 50 % with SAR as the focus modality.
Sensors 24 06054 g012
Figure 13. Effects of the loss factor buildup rate β on modality utilization and classification accuracy with modality utilization-based training method in the NTIRE21 dataset. Target utilization M U t a r g e t = 50 %, Maximum Loss Factor λ L _ m a x = 100 , and buildup delay δ = 0 with SAR as the focus modality.
Figure 13. Effects of the loss factor buildup rate β on modality utilization and classification accuracy with modality utilization-based training method in the NTIRE21 dataset. Target utilization M U t a r g e t = 50 %, Maximum Loss Factor λ L _ m a x = 100 , and buildup delay δ = 0 with SAR as the focus modality.
Sensors 24 06054 g013
Figure 14. Effects of Gaussian noise with mean = 0 and variance = { 0.06 , 0.09 , 0.12 } in the EO modality, the SAR modality, and both modalities during inference on networks trained with different levels of SAR utilization.
Figure 14. Effects of Gaussian noise with mean = 0 and variance = { 0.06 , 0.09 , 0.12 } in the EO modality, the SAR modality, and both modalities during inference on networks trained with different levels of SAR utilization.
Sensors 24 06054 g014
Table 1. Performance and modality utilization (MU) for the NTIRE21 and MCubeS datasets [14].
Table 1. Performance and modality utilization (MU) for the NTIRE21 and MCubeS datasets [14].
DatasetModalityPerformanceModality Utilization (MU) (%)
Accuracy (%)EOSAR
NTIRE21EO97.5100.0-
SAR84.9-100.0
EO-SAR97.899.590.40
mIoURBGAoLPDoLPNIR
MCubeSRGB0.318100.0---
AoLP0.266-100.0--
DoLP0.262--100.0-
NIR0.270---100.0
RGB-AoLP-DoLP-NIR0.37434.519.030.915.6
AoLP-DoLP-NIR0.351-67.321.011.7
Table 2. Multimodal fusion network trained on the NTIRE21 dataset for different target utilization M U t a r g e t with SAR as the focus modality. Highest value is represented by a bold values.
Table 2. Multimodal fusion network trained on the NTIRE21 dataset for different target utilization M U t a r g e t with SAR as the focus modality. Highest value is represented by a bold values.
SAR MU targ
(%)
Acc.
(%)
MU EO
(%)
MU SAR
(%)
0.097.199.40.6
12.597.692.87.2
25.097.782.117.9
37.597.763.536.5
50.097.247.053.0
62.596.831.768.3
75.095.319.780.3
87.592.213.186.9
100.084.44.895.1
Table 3. Multimodal fusion network trained on NTIRE21 dataset for different target utilization M U t a r g e t with EO as the focus modality. Highest value is represented by a bold values.
Table 3. Multimodal fusion network trained on NTIRE21 dataset for different target utilization M U t a r g e t with EO as the focus modality. Highest value is represented by a bold values.
EO MU targ
(%)
Acc.
(%)
MU EO
(%)
MU SAR
(%)
0.084.43.996.1
12.592.014.385.7
25.095.319.580.5
37.596.830.669.4
50.097.247.053.0
62.597.663.536.5
75.097.779.420.6
87.597.594.45.6
100.097.199.50.5
Table 4. Multimodal fusion network trained on MCubeS dataset for different target utilization M U t a r g e t with loss factor λ L = 100.0 and RGB as the focus modality. Highest value is represented by a bold values.
Table 4. Multimodal fusion network trained on MCubeS dataset for different target utilization M U t a r g e t with loss factor λ L = 100.0 and RGB as the focus modality. Highest value is represented by a bold values.
RGB MU targ (%)mIoU (%) MU RGB (%) MU AoLP (%) MU DoLP (%) MU NIR (%)
0.00.4001.24.254.440.2
12.50.4036.70.734.258.4
25.00.38818.610.87.363.3
37.50.39339.410.816.533.3
50.00.39754.915.90.129.1
62.50.39468.910.58.112.5
75.00.38784.95.40.59.2
87.50.40789.25.12.63.1
100.00.40396.71.00.91.4
Table 5. Five-fold validation for modality utilization-based training methods without loss factor warm-up on NTIRE21 dataset with m f = S A R , λ L = 100.0 , and M U t a r g = 0.0 . Instability can be observed in folds 2, 3, and 5. Catastrophic failures in training are represented by red values.
Table 5. Five-fold validation for modality utilization-based training methods without loss factor warm-up on NTIRE21 dataset with m f = S A R , λ L = 100.0 , and M U t a r g = 0.0 . Instability can be observed in folds 2, 3, and 5. Catastrophic failures in training are represented by red values.
SAR MU targ (%)FoldAcc. (%) MU EO (%) MU SAR (%)
0.0197.799.90.1
0.0244.557.842.2
0.0351.155.544.5
0.0498.499.90.1
0.0548.855.944.1
Table 6. Five-fold validation for modality utilization-based training methods stabilized with loss factor warm-up on NTIRE21 dataset with m f = S A R , λ L = 100.0 , and M U t a r g = 0.0 .
Table 6. Five-fold validation for modality utilization-based training methods stabilized with loss factor warm-up on NTIRE21 dataset with m f = S A R , λ L = 100.0 , and M U t a r g = 0.0 .
SAR MU targ (%)FoldAcc. (%) MU EO (%) MU SAR (%)
0.0197.799.40.6
0.0296.999.20.8
0.0397.699.30.7
0.0497.299.40.6
0.0597.099.60.4
Table 7. Effects of Gaussian noise with mean = 0 and variance = { 0.06 , 0.09 , 0.12 } in the EO modality (dominant modality) during inference on networks’ trained performance accuracy with different levels of EO utilization.
Table 7. Effects of Gaussian noise with mean = 0 and variance = { 0.06 , 0.09 , 0.12 } in the EO modality (dominant modality) during inference on networks’ trained performance accuracy with different levels of EO utilization.
EO MU targ No NoiseNoise
Var. = 0.06
Noise
Var. = 0.09
Noise
Var. = 0.12
Average
Acc.
100.095.184.474.473.781.9
87.594.686.874.980.384.2
75.092.885.780.181.485.0
62.591.883.780.282.184.5
50.089.880.882.182.383.8
37.589.183.878.980.983.2
25.091.283.979.581.183.9
12.592.286.581.281.785.4
0.090.385.983.383.085.6
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Singh, S.; Saber, E.; Markopoulos, P.P.; Heard, J. Regulating Modality Utilization within Multimodal Fusion Networks. Sensors 2024, 24, 6054. https://doi.org/10.3390/s24186054

AMA Style

Singh S, Saber E, Markopoulos PP, Heard J. Regulating Modality Utilization within Multimodal Fusion Networks. Sensors. 2024; 24(18):6054. https://doi.org/10.3390/s24186054

Chicago/Turabian Style

Singh, Saurav, Eli Saber, Panos P. Markopoulos, and Jamison Heard. 2024. "Regulating Modality Utilization within Multimodal Fusion Networks" Sensors 24, no. 18: 6054. https://doi.org/10.3390/s24186054

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop