Liu Yongmei, Yang Fan, Yu Linsen. Automatic image annotation based on scene semantic trees[J]. Journal of Image and Graphics, 2013, 18(5): 529-536. DOI: 10.11834/jig.20130506.
获得了优于TM(translation model)、CMRM(cross media relevance model)、CRM(continous-space relevance model)、PLSA-GMM(概率潜在语义分析-高期混合模型)等模型的标注结果。
Abstract
Automatic image annotation(AIA) is an important and challenging job for image analysis and understanding such as Content-Based Image Retrieval(CBIR). In AIA
a model for annotation which represents the relationship between the semantic concept space and the visual feature space
is constructed by learning from annotated image datasets. The performance of the AIA is still poor because the relationship between the key words and visual features is too compli-cated due to the semantic gap. However
with constrains under the scenes
the correlation between them becomes simpler and clearer for better annotation results. In this paper
a method of automatic image annotation based on scene semantic tree is proposed. Image scenes are obtained through the annotated words using PLSA. Scene semantic trees are constructed for each image scene. Un-annotated images are classified into certain scenes with visual features and are annotated using the corresponding scene semantic tree. Using the visual features provided by Duygulu
the experiments get the more effective results on Corel5K database than TM