In this special issue, we invite the original contributions from diverse research fields, developing new deep cross-media neural model for understanding and generating image descriptions, which aims to reduce the gap between image understanding and natural language descriptions.
Call For Paper
The topics of interest include, but are not limited to:
Attention guided UGID
Visual relationship in UGID
Compositional architectures for UGID
Multimodal learning for UGID
Describing novel objects in UGID
Natural language processing model
New datasets for UGID
Novel encoder-decoder based architecture
Deep cross-media neural model with applications of UGID, e.g., early childhood education, medical image analysis, assisted blinding and news automation, etc.