Please use this identifier to cite or link to this item:
http://hdl.handle.net/10397/108800
| DC Field | Value | Language |
|---|---|---|
| dc.contributor | Department of Electrical and Electronic Engineering | - |
| dc.creator | Guo, Q | - |
| dc.creator | Liao, Y | - |
| dc.creator | Li, Z | - |
| dc.creator | Liang, S | - |
| dc.date.accessioned | 2024-08-27T04:40:40Z | - |
| dc.date.available | 2024-08-27T04:40:40Z | - |
| dc.identifier.uri | http://hdl.handle.net/10397/108800 | - |
| dc.language.iso | en | en_US |
| dc.publisher | MDPI AG | en_US |
| dc.rights | © 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). | en_US |
| dc.rights | The following publication Guo Q, Liao Y, Li Z, Liang S. Multi-Modal Representation via Contrastive Learning with Attention Bottleneck Fusion and Attentive Statistics Features. Entropy. 2023; 25(10):1421 is available at https://doi.org/10.3390/e25101421. | en_US |
| dc.subject | Attention bottleneck fusion | en_US |
| dc.subject | Attentive statistics features | en_US |
| dc.subject | Contrastive learning | en_US |
| dc.subject | Multimodal representation | en_US |
| dc.title | Multi-modal representation via contrastive learning with attention bottleneck fusion and attentive statistics features | en_US |
| dc.type | Journal/Magazine Article | en_US |
| dc.identifier.volume | 25 | - |
| dc.identifier.issue | 10 | - |
| dc.identifier.doi | 10.3390/e25101421 | - |
| dcterms.abstract | The integration of information from multiple modalities is a highly active area of research. Previous techniques have predominantly focused on fusing shallow features or high-level representations generated by deep unimodal networks, which only capture a subset of the hierarchical relationships across modalities. However, previous methods are often limited to exploiting the fine-grained statistical features inherent in multimodal data. This paper proposes an approach that densely integrates representations by computing image features’ means and standard deviations. The global statistics of features afford a holistic perspective, capturing the overarching distribution and trends inherent in the data, thereby facilitating enhanced comprehension and characterization of multimodal data. We also leverage a Transformer-based fusion encoder to effectively capture global variations in multimodal features. To further enhance the learning process, we incorporate a contrastive loss function that encourages the discovery of shared information across different modalities. To validate the effectiveness of our approach, we conduct experiments on three widely used multimodal sentiment analysis datasets. The results demonstrate the efficacy of our proposed method, achieving significant performance improvements compared to existing approaches. | - |
| dcterms.accessRights | open access | en_US |
| dcterms.bibliographicCitation | Entropy, Oct. 2023, v. 25, no. 10, 1421 | - |
| dcterms.isPartOf | Entropy | - |
| dcterms.issued | 2023-10 | - |
| dc.identifier.scopus | 2-s2.0-85175373728 | - |
| dc.identifier.eissn | 1099-4300 | - |
| dc.identifier.artn | 1421 | - |
| dc.description.validate | 202408 bcch | - |
| dc.description.oa | Version of Record | en_US |
| dc.identifier.FolderNumber | OA_Scopus/WOS | en_US |
| dc.description.fundingSource | Others | en_US |
| dc.description.fundingText | National Key Research and Development Program of China | en_US |
| dc.description.pubStatus | Published | en_US |
| dc.description.oaCategory | CC | en_US |
| Appears in Collections: | Journal/Magazine Article | |
Files in This Item:
| File | Description | Size | Format | |
|---|---|---|---|---|
| entropy-25-01421.pdf | 2.24 MB | Adobe PDF | View/Open |
Page views
67
Citations as of Nov 10, 2025
Downloads
22
Citations as of Nov 10, 2025
SCOPUSTM
Citations
6
Citations as of Dec 19, 2025
Google ScholarTM
Check
Altmetric
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.



