*Improving Natural Language Queries Search and Retrieval through Semantic Image Annotation Understanding

Document Type : Original Article

Authors

1 Faculty of Computer and Information Technology, Egyptian E-Learning University, Cairo, Egypt

2 Ain Shams University

3 Computer Engineering and Systems Department, Faculty of Engineering, Helwan University, Cairo, Egypt

4 Information Systems Department, Faculty of Computer and Information Sciences, Ain Shams University, Cairo, Egypt

Abstract

Retrieving images using detailed natural language queries remains a difficult challenge. Traditional annotation-based image retrieval systems using word matching techniques cannot efficiently support such query types. Significant improvements for this problem can be achieved with a semantic understanding for those query sentences and image annotations. This paper presents a two-stage semantic understanding approach for natural language query sentences. At the first stage, the Stanford parser and a designed rule-based relation extraction tool are used in triple extraction process to efficiently extract the objects attributes, instances and natural language annotations relationships involving these objects. The second stage integrates the extracted relations with external commonsense knowledge source in a mapping process to provide high-level semantic meanings to the extracted triples. Experiments are conducted for evaluating the benefit of the proposed semantic understanding against a testing set of natural language sentences from the Flickr8k dataset. The results show that the proposed approach succeeds to extract relational triples with average accuracy value of 97% for the different types of annotations relationships: attributes and instance relations, multiword dependence relations, and semantic relations.

Keywords