Neural Textual Features Composition for CBIR
Aboali, Mohamed A.; Elmaddah, Islam; Hossam El DIn Hassan Abdelmunim;
Abstract
Content Based Image Retrieval, CBIR, is a highly active leading research field with numerous applications that are currently expanding beyond traditional CBIR methodologies. This paper presents a novel CBIR methodology that addresses these growing demands. Query inputs of the proposed methodology are an image and a text. For instance, having an image, a user would like to obtain a similar one while also incorporating modifications described in text format that we refer to as a text-modifier. The proposed methodology uses a set of neural networks that operate in feature space and perform feature composition in a uniform-known domain which is the textual feature domain. In this methodology, ResNet is used to extract image features and LSTM to extract text features to form query inputs. The proposed methodology uses a set of three single-hidden-layer non-linear feedforward networks in a cascading structure labeled NetA, NetC, and NetB. NetA maps image features into corresponding textual features. NetC composes the textual features produced by NetA with text-modifier features to form target image textual features. Finally, NetB maps target textual features to target image features that are used to recall the target image from the image-base based on cosine similarity. The proposed architecture was tested using ResNet 18, 50 and 152 for extracting image features. The testing results are promising and can compete with the most recent approaches to our knowledge.
Other data
| Title | Neural Textual Features Composition for CBIR | Authors | Aboali, Mohamed A.; Elmaddah, Islam; Hossam El DIn Hassan Abdelmunim | Keywords | Computer vision;deep learning;feature extraction;image retrieval | Issue Date | 1-Jan-2023 | Journal | IEEE Access | ISSN | 2169-3536 | DOI | 10.1109/ACCESS.2023.3259737 | Scopus ID | 2-s2.0-85151502434 |
Recommend this item
Similar Items from Core Recommender Database
Items in Ain Shams Scholar are protected by copyright, with all rights reserved, unless otherwise indicated.