Search results
1 – 10 of over 6000Wei Lu, Heng Ding and Jiepu Jiang
The purpose of this paper is to utilize document expansion techniques for improving image representation and retrieval. This paper proposes a concise framework for tag-based image…
Abstract
Purpose
The purpose of this paper is to utilize document expansion techniques for improving image representation and retrieval. This paper proposes a concise framework for tag-based image retrieval (TBIR).
Design/methodology/approach
The proposed approach includes three core components: a strategy of selecting expansion (similar) images from the whole corpus (e.g. cluster-based or nearest neighbor-based); a technique for assessing image similarity, which is adopted for selecting expansion images (text, image, or mixed); and a model for matching the expanded image representation with the search query (merging or separate).
Findings
The results show that applying the proposed method yields significant improvements in effectiveness, and the method obtains better performance on the top of the rank and makes a great improvement on some topics with zero score in baseline. Moreover, nearest neighbor-based expansion strategy outperforms the cluster-based expansion strategy, and using image features for selecting expansion images is better than using text features in most cases, and the separate method for calculating the augmented probability P(q|RD) is able to erase the negative influences of error images in RD.
Research limitations/implications
Despite these methods only outperform on the top of the rank instead of the entire rank list, TBIR on mobile platforms still can benefit from this approach.
Originality/value
Unlike former studies addressing the sparsity, vocabulary mismatch, and tag relatedness in TBIR individually, the approach proposed by this paper addresses all these issues with a single document expansion framework. It is a comprehensive investigation of document expansion techniques in TBIR.
Details
Keywords
Aabid Hussain, Sumeer Gul, Tariq Ahmad Shah and Sheikh Shueb
The purpose of this study is to explore the retrieval effectiveness of three image search engines (ISE) – Google Images, Yahoo Image Search and Picsearch in terms of their image…
Abstract
Purpose
The purpose of this study is to explore the retrieval effectiveness of three image search engines (ISE) – Google Images, Yahoo Image Search and Picsearch in terms of their image retrieval capability. It is an effort to carry out a Cranfield experiment to know how efficient the commercial giants in the image search are and how efficient an image specific search engine is.
Design/methodology/approach
The keyword search feature of three ISEs – Google images, Yahoo Image Search and Picsearch – was exploited to make search with keyword captions of photos as query terms. Selected top ten images were used to act as a testbed for the study, as images were searched in accordance with features of the test bed. Features to be looked for included size (1200 × 800), format of images (JPEG/JPG) and the rank of the original image retrieved by ISEs under study. To gauge the overall retrieval effectiveness in terms of set standards, only first 50 result hits were checked. Retrieval efficiency of select ISEs were examined with respect to their precision and relative recall.
Findings
Yahoo Image Search outscores Google Images and Picsearch both in terms of precision and relative recall. Regarding other criteria – image size, image format and image rank in search results, Google Images is ahead of others.
Research limitations/implications
The study only takes into consideration basic image search feature, i.e. text-based search.
Practical implications
The study implies that image search engines should focus on relevant descriptions. The study evaluated text-based image retrieval facilities and thereby offers a choice to users to select best among the available ISEs for their use.
Originality/value
The study provides an insight into the effectiveness of the three ISEs. The study is one of the few studies to gauge retrieval effectiveness of ISEs. Study also produced key findings that are important for all ISE users and researchers and the Web image search industry. Findings of the study will also prove useful for search engine companies to improve their services.
Details
Keywords
Chao Lu, Chengzhi Zhang and Daqing He
In the era of social media, users all over the world annotate books with social tags to express their preferences and interests. The purpose of this paper is to explore different…
Abstract
Purpose
In the era of social media, users all over the world annotate books with social tags to express their preferences and interests. The purpose of this paper is to explore different tagging behaviours by analysing the book tags in different languages.
Design/methodology/approach
This investigation collected nearly 56,000 tags of 1,200 books from one Chinese and two English online bookmarking systems; it combined content analysis and machine-processing methods to evaluate the similarities and differences between different tagging systems from a cross-lingual perspective. Jaccard’s coefficient was adopted to evaluate the similarity level.
Findings
The results show that the similarity between mono-lingual tags of the same books is higher than that of cross-lingual tags in different systems and the similarity between tags of books written for specialties is higher than that of books written for the general public.
Research limitations/implications
Those who have more in common annotate books with more similar tags. The similarity between users in tagging systems determines the similarity of the tag sets.
Practical implications
The results and conclusion of this study will benefit users’ cross-lingual information retrieval and cross-lingual book recommendation for online bookmarking systems.
Originality/value
This study may be one of the first to compare cross-lingual tags. Its methodology can be applied to tag comparison between any two languages. The insights of this study will help develop cross-lingual tagging systems and improve information retrieval.
Details
Keywords
Ziming Zeng, Shouqiang Sun, Jingjing Sun, Jie Yin and Yueyan Shen
Dunhuang murals are rich in cultural and artistic value. The purpose of this paper is to construct a novel mobile visual search (MVS) framework for Dunhuang murals, enabling users…
Abstract
Purpose
Dunhuang murals are rich in cultural and artistic value. The purpose of this paper is to construct a novel mobile visual search (MVS) framework for Dunhuang murals, enabling users to efficiently search for similar, relevant and diversified images.
Design/methodology/approach
The convolutional neural network (CNN) model is fine-tuned in the data set of Dunhuang murals. Image features are extracted through the fine-tuned CNN model, and the similarities between different candidate images and the query image are calculated by the dot product. Then, the candidate images are sorted by similarity, and semantic labels are extracted from the most similar image. Ontology semantic distance (OSD) is proposed to match relevant images using semantic labels. Furthermore, the improved DivScore is introduced to diversify search results.
Findings
The results illustrate that the fine-tuned ResNet152 is the best choice to search for similar images at the visual feature level, and OSD is the effective method to search for the relevant images at the semantic level. After re-ranking based on DivScore, the diversification of search results is improved.
Originality/value
This study collects and builds the Dunhuang mural data set and proposes an effective MVS framework for Dunhuang murals to protect and inherit Dunhuang cultural heritage. Similar, relevant and diversified Dunhuang murals are searched to meet different demands.
Details
Keywords
As a relatively new computing paradigm, crowdsourcing has gained enormous attention in the recent decade. Its compliance with the Web 2.0 principles, also, puts forward…
Abstract
Purpose
As a relatively new computing paradigm, crowdsourcing has gained enormous attention in the recent decade. Its compliance with the Web 2.0 principles, also, puts forward unprecedented opportunities to empower the related services and mechanisms by leveraging humans’ intelligence and problem solving abilities. With respect to the pivotal role of search engines in the Web and information community, this paper aims to investigate the advantages and challenges of incorporating people – as intelligent agents – into search engines’ workflow.
Design/methodology/approach
To emphasize the role of the human in computational processes, some specific and related areas are studied. Then, through studying the current trends in the field of crowd-powered search engines and analyzing the actual needs and requirements, the perspectives and challenges are discussed.
Findings
As the research on this topic is still in its infancy, it is believed that this study can be considered as a roadmap for future works in the field. In this regard, current status and development trends are delineated through providing a general overview of the literature. Moreover, several recommendations for extending the applicability and efficiency of next generation of crowd-powered search engines are presented. In fact, becoming aware of different aspects and challenges of constructing search engines of this kind can shed light on the way of developing working systems with respect to essential considerations.
Originality/value
The present study was aimed to portrait the big picture of crowd-powered search engines and possible challenges and issues. As one of the early works that provided a comprehensive report on different aspects of the topic, it can be regarded as a reference point.
Details
Keywords
Image classification is becoming a supporting technology in several image-processing tasks. Due to rich semantic information contained in the images, it is very popular for an…
Abstract
Purpose
Image classification is becoming a supporting technology in several image-processing tasks. Due to rich semantic information contained in the images, it is very popular for an image to have several labels or tags. This paper aims to develop a novel multi-label classification approach with superior performance.
Design/methodology/approach
Many multi-label classification problems share two main characteristics: label correlations and label imbalance. However, most of current methods are devoted to either model label relationship or to only deal with unbalanced problem with traditional single-label methods. In this paper, multi-label classification problem is regarded as an unbalanced multi-task learning problem. Multi-task least-squares support vector machine (MTLS-SVM) is generalized for this problem, renamed as multi-label LS-SVM (ML2S-SVM).
Findings
Experimental results on the emotions, scene, yeast and bibtex data sets indicate that the ML2S-SVM is competitive with respect to the state-of-the-art methods in terms of Hamming loss and instance-based F1 score. The values of resulting parameters largely influence the performance of ML2S-SVM, so it is necessary for users to identify proper parameters in advance.
Originality/value
On the basis of MTLS-SVM, a novel multi-label classification approach, ML2S-SVM, is put forward. This method can overcome the unbalanced problem but also explicitly models arbitrary order correlations among labels by allowing multiple labels to share a subspace. In addition, the multi-label classification approach has a wider range of applications. That is to say, it is not limited to the field of image classification.
Details
Keywords
Yaolin Zhou, Zhaoyang Zhang, Xiaoyu Wang, Quanzheng Sheng and Rongying Zhao
The digitalization of archival management has rapidly developed with the maturation of digital technology. With data's exponential growth, archival resources have transitioned…
Abstract
Purpose
The digitalization of archival management has rapidly developed with the maturation of digital technology. With data's exponential growth, archival resources have transitioned from single modalities, such as text, images, audio and video, to integrated multimodal forms. This paper identifies key trends, gaps and areas of focus in the field. Furthermore, it proposes a theoretical organizational framework based on deep learning to address the challenges of managing archives in the era of big data.
Design/methodology/approach
Via a comprehensive systematic literature review, the authors investigate the field of multimodal archive resource organization and the application of deep learning techniques in archive organization. A systematic search and filtering process is conducted to identify relevant articles, which are then summarized, discussed and analyzed to provide a comprehensive understanding of existing literature.
Findings
The authors' findings reveal that most research on multimodal archive resources predominantly focuses on aspects related to storage, management and retrieval. Furthermore, the utilization of deep learning techniques in image archive retrieval is increasing, highlighting their potential for enhancing image archive organization practices; however, practical research and implementation remain scarce. The review also underscores gaps in the literature, emphasizing the need for more practical case studies and the application of theoretical concepts in real-world scenarios. In response to these insights, the authors' study proposes an innovative deep learning-based organizational framework. This proposed framework is designed to navigate the complexities inherent in managing multimodal archive resources, representing a significant stride toward more efficient and effective archival practices.
Originality/value
This study comprehensively reviews the existing literature on multimodal archive resources organization. Additionally, a theoretical organizational framework based on deep learning is proposed, offering a novel perspective and solution for further advancements in the field. These insights contribute theoretically and practically, providing valuable knowledge for researchers, practitioners and archivists involved in organizing multimodal archive resources.
Details
Keywords
Hanadi Buarki and Bashaer Alkhateeb
This paper aims to find out how people use hashtags as a medium of information retrieval and dissemination, and how they are used in social media tools, such as Instagram.
Abstract
Purpose
This paper aims to find out how people use hashtags as a medium of information retrieval and dissemination, and how they are used in social media tools, such as Instagram.
Design/methodology/approach
A quantitative question estimated the participants’ use of the hashtags during the workshop. Statistical data of the participants and their posts were collected from social network analysis tools. The posts that included the workshop’s designated hashtags were retrieved, recorded, coded and analysed to collect qualitative data.
Findings
In total, 74 (46 per cent) participants used the workshop’s hashtags to share posts, the retrieval of the hashtags declined by time and Google search engine retrieved the maximum results. It was found that a hashtag would be common when associated with descriptors, and that its use depends on its popularity, followers and its survival time. Finally, hashtags connect people, allow them to express their enthusiasm to reveal common interests and networks them through social media tools such as Instagram.
Research limitations/implications
The research limitations were in relation to the participants’ demographic information, the non-identification of their gender and hashtags being misspelt.
Practical implications
The research project summarises the experiences that social media has made connecting easier through the right use of hashtags by providing 24/7 free feedback, the possibility to exchange ideas and by their involvement in promoting and organising events. It also indicates interaction among people sharing the same interest by retrieving subject-based hashtags.
Originality/value
When retrieving information related to hashtags, it is recommended that multi-retrieval systems, social media tools and search engines should be consulted and not depend on a solo system or tool. Future research is recommended in search for a multi-retrieval social media and search engine tool that standardises the use of hashtags and will retrieve information from different platforms.
Details
Keywords
Shengli Deng, Anqi Zhao, Ruhua Huang and Haiping Zhao
This study aims to examine why users search for images, how users describe their image needs and what the images are used for by analysing questions obtained from two Chinese…
Abstract
Purpose
This study aims to examine why users search for images, how users describe their image needs and what the images are used for by analysing questions obtained from two Chinese social Q&A sites, Zhihu and Baidu Zhidao.
Design/methodology/approach
A total of 1,402 image questions were collected from Zhihu and Baidu Zhidao. Both quantitative analysis and qualitative content analysis were performed to identify user image needs and the potential differences on the two social Q&A sites.
Findings
Question-asker’s intention varies in different platforms. Zhihu users asked questions mainly aiming at a promotion of subsequent discussion, whereas users of Baidu Zhidao often did so to seek information. Syntactic attributes were not frequently used in both two sites. Zhihu users were more likely to express subjective evaluations on images (concept, emotion, theme and style) in their questions than users of Baidu Zhidao. In contrast, questions from Baidu Zhidao showed a tendency to more frequently include descriptive metadata (rights, format, size, quality and authenticity) and semantic attributes (generic activity, specific people, fashion and text) of the images than questions from Zhihu. Learning was an important use on social Q&A sites, especially on Baidu Zhidao. In addition, the images were primarily used to trigger emotion or served a persuasive purpose in Zhihu.
Practical implications
This study contributes to a better understanding of user image search behaviour, and the findings could be used to develop better image services on social Q&A sites. Meanwhile, the image attributes extracted from the questions are conducive to the improvement of image retrieval systems.
Originality/value
This study explored the features of image needs on social Q&A sites, especially considering image use specified in the question. The difference of image needs between two Chinese social Q&A sites (Zhihu and Baidu Zhidao) was identified.
Details
Keywords
This paper surveys theoretical and practical issues associated with a particular type of information retrieval problem, namely that where the information need is pictorial. The…
Abstract
This paper surveys theoretical and practical issues associated with a particular type of information retrieval problem, namely that where the information need is pictorial. The paper is contextualised by the notion of a visually stimulated society, in which the ease of record creation and transmission in the visual medium is contrasted with the difficulty of gaining effective subject access to the world's stores of such records. The technological developments which, in casting the visual image in electronic form, have contributed so significantly to its availability are reviewed briefly, as a prelude to the main thrust of the paper. Concentrating on still and moving pictorial forms of the visual image, the paper dwells on issues related to the subject indexing of pictorial material and discusses four models of pictorial information retrieval corresponding with permutations of the verbal and visual modes for the representation of picture content and of information need.