The IRMA Community
Newsletters
Research IRM
Click a keyword to search titles using our InfoSci-OnDemand powered search:
|
An Image-Text Matching Method for Multi-Modal Robots
|
Author(s): Ke Zheng (Hunan Biological and Electromechanical Polytechnic, China)and Zhou Li (Hunan Biological and Electromechanical Polytechnic, China)
Copyright: 2024
Volume: 36
Issue: 1
Pages: 21
Source title:
Journal of Organizational and End User Computing (JOEUC)
Editor(s)-in-Chief: Sangbing (Jason) Tsai (Wuyi University, China & International Engineering and Technology Institute (IETI), Hong Kong)and Wei Liu (Qingdao University, China)
DOI: 10.4018/JOEUC.334701
Purchase
|
Abstract
With the rapid development of artificial intelligence and deep learning, image-text matching has gradually become an important research topic in cross-modal fields. Achieving correct image-text matching requires a strong understanding of the correspondence between visual and textual information. In recent years, deep learning-based image-text matching methods have achieved significant success. However, image-text matching requires a deep understanding of intra-modal information and the exploration of fine-grained alignment between image regions and textual words. How to integrate these two aspects into a single model remains a challenge. Additionally, reducing the internal complexity of the model and effectively constructing and utilizing prior knowledge are also areas worth exploring, therefore addressing the issues of excessive computational complexity in existing fine-grained matching methods and the lack of multi-perspective matching.
Related Content
Ke Zheng, Zhou Li.
© 2024.
21 pages.
|
Weihui Han, Tianshuo Zhang, Jamal Khan, Lujian Wang, Chao Tu.
© 2024.
22 pages.
|
Chen Quan, Baoli Lu.
© 2024.
22 pages.
|
Peijin Li, Xinyi Peng, Chonghui Zhang, Tomas Baležentis.
© 2024.
25 pages.
|
Lei Zhao, Bowen Deng, Liang Wu, Chang Liu, Min Guo, Youjia Guo.
© 2024.
27 pages.
|
Xiaoye Ma, Yanyan Li, Muhammad Asif.
© 2024.
29 pages.
|
Hao Wu, Zhiyi Zhang, Zhilin Zhu.
© 2024.
12 pages.
|
|
|