Rozprawy doktorskie na temat „IMAGE RETRIEVAL TECHNIQUES”
Utwórz poprawne odniesienie w stylach APA, MLA, Chicago, Harvard i wielu innych
Sprawdź 50 najlepszych rozpraw doktorskich naukowych na temat „IMAGE RETRIEVAL TECHNIQUES”.
Przycisk „Dodaj do bibliografii” jest dostępny obok każdej pracy w bibliografii. Użyj go – a my automatycznie utworzymy odniesienie bibliograficzne do wybranej pracy w stylu cytowania, którego potrzebujesz: APA, MLA, Harvard, Chicago, Vancouver itp.
Możesz również pobrać pełny tekst publikacji naukowej w formacie „.pdf” i przeczytać adnotację do pracy online, jeśli odpowiednie parametry są dostępne w metadanych.
Przeglądaj rozprawy doktorskie z różnych dziedzin i twórz odpowiednie bibliografie.
Shaffrey, Cian William. "Multiscale techniques for image segmentation, classification and retrieval". Thesis, University of Cambridge, 2003. https://www.repository.cam.ac.uk/handle/1810/272033.
Pełny tekst źródłaYang, Cheng 1974. "Image database retrieval with multiple-instance learning techniques". Thesis, Massachusetts Institute of Technology, 1998. http://hdl.handle.net/1721.1/50505.
Pełny tekst źródłaIncludes bibliographical references (p. 81-82).
In this thesis, we develop and test an approach to retrieving images from an image database based on content similarity. First, each picture is divided into many overlapping regions. For each region, the sub-picture is filtered and converted into a feature vector. In this way, each picture is represented by a number of different feature vectors. The user selects positive and negative image examples to train the system. During the training, a multiple-instance learning method known as the Diverse Density algorithm is employed to determine which feature vector in each image best represents the user's concept, and which dimensions of the feature vectors are important. The system tries to retrieve images with similar feature vectors from the remainder of the database. A variation of the weighted correlation statistic is used to determine image similarity. The approach is tested on a large database of natural scenes as well as single- and multiple-object images. Comparisons are made against a previous approach, and the effects of tuning various training parameters, as well as that of adjusting algorithmic details, are also studied.
by Cheng Yang.
S.M.
Carswell, James. "Using Raster Sketches for Digital Image Retrieval". Fogler Library, University of Maine, 2000. http://www.library.umaine.edu/theses/pdf/CarswellJD2000.pdf.
Pełny tekst źródłaZhang, Dengsheng 1963. "Image retrieval based on shape". Monash University, School of Computing and Information Technology, 2002. http://arrow.monash.edu.au/hdl/1959.1/8688.
Pełny tekst źródłaLim, Suryani. "Feature extraction, browsing and retrieval of images". Monash University, School of Computing and Information Technology, 2005. http://arrow.monash.edu.au/hdl/1959.1/9677.
Pełny tekst źródłaGoncalves, Pinheiro Antonio Manuel. "Shape approximation and retrieval using scale-space techniques". Thesis, University of Essex, 2001. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.391661.
Pełny tekst źródłaLi, Yuanxi. "Semantic image similarity based on deep knowledge for effective image retrieval". HKBU Institutional Repository, 2014. https://repository.hkbu.edu.hk/etd_oa/99.
Pełny tekst źródłaWong, Chun Fan. "Automatic semantic image annotation and retrieval". HKBU Institutional Repository, 2010. http://repository.hkbu.edu.hk/etd_ra/1188.
Pełny tekst źródłaLing, Haibin. "Techniques for image retrieval deformation insensitivity and automatic thumbnail cropping /". College Park, Md. : University of Maryland, 2006. http://hdl.handle.net/1903/3859.
Pełny tekst źródłaThesis research directed by: Computer Science. Title from t.p. of PDF. Includes bibliographical references. Published by UMI Dissertation Services, Ann Arbor, Mich. Also available in paper.
Liu, Danzhou. "EFFICIENT TECHNIQUES FOR RELEVANCE FEEDBACK PROCESSING IN CONTENT-BASED IMAGE RETRIEVAL". Doctoral diss., University of Central Florida, 2009. http://digital.library.ucf.edu/cdm/ref/collection/ETD/id/2991.
Pełny tekst źródłaPh.D.
School of Electrical Engineering and Computer Science
Engineering and Computer Science
Computer Science PhD
Yu, Ning. "Techniques for boosting the performance in content-based image retrieval systems". Doctoral diss., University of Central Florida, 2011. http://digital.library.ucf.edu/cdm/ref/collection/ETD/id/4726.
Pełny tekst źródłaID: 030646264; System requirements: World Wide Web browser and PDF reader.; Mode of access: World Wide Web.; Thesis (Ph.D.)--University of Central Florida, 2011.; Includes bibliographical references (p. 81-91).
Ph.D.
Doctorate
Computer Science
Engineering and Computer Science
Computer Science
Aboaisha, Hosain. "The optimisation of elementary and integrative content-based image retrieval techniques". Thesis, University of Huddersfield, 2015. http://eprints.hud.ac.uk/id/eprint/26164/.
Pełny tekst źródłaYoon, Janghyun. "A network-aware semantics-sensitive image retrieval system". Diss., Available online, Georgia Institute of Technology, 2004:, 2003. http://etd.gatech.edu/theses/available/etd-04082004-180459/unrestricted/yoon%5fjanghyun%5f200312%5fphd.pdf.
Pełny tekst źródłaVila, Duran Marius. "Information theory techniques for multimedia data classification and retrieval". Doctoral thesis, Universitat de Girona, 2015. http://hdl.handle.net/10803/302664.
Pełny tekst źródłaEns trobem a l’era de la informació on la majoria de les dades s’emmagatzemen en format digital. Per tant, la gestió de documents i vídeos digitals requereix el desenvolupament de tècniques eficients per a l’anàlisi automàtic. Entre elles, la captura de la similitud o dissimilitud entre diferents imatges de documents o fotogrames de vídeo és extremadament important. En aquesta tesi, analitzem, a diverses resolucions d’imatge, el comportament de tres famílies diferents de mesures basades en similitud d’imatges i aplicades a la classificació de factures. En aquests tres conjunt de mesures, el càlcul de la similitud entre dues imatges es basa, respectivament, en les diferències d’intensitat, en la informació mútua, i en la distància de compressió normalitzada. Degut a que els millors resultats s’obtenen amb les mesures basades en la informació mútua, es procedeix a investigar l’aplicació de tres generalitzacions de la informació mútua basades en Tsallis en diferents índexs entròpics. Aquestes tres generalitzacions es deriven respectivament de la distància de Kullback-Leibler, la diferència entre l’entropia i entropia condicional, i la divergència de Jensen-Shannon. En relació al processament de vídeo digital, proposem dos enfocaments diferents de teoria de la informació basats respectivament en la informació mútua de Tsallis i en la divergència de Jensen-Tsallis, per detectar els límits d’un pla cinematogràfic en una seqüència de vídeo i per seleccionar el fotograma clau més representatiu de cada pla. Finalment, l’entropia de Shannon s’ha utilitzat habitualment per quantificar la informativitat d’una imatge. El principal inconvenient d’aquesta mesura és que no té en compte la distribució espacial dels píxels. En aquesta tesi, s’analitzen quatre mesures de teoria de la informació que superen aquesta limitació. Tres d’elles (entropy rate, excess entropy i erasure entropy) consideren la imatge com un procés estocàstic estacionari, mentre que la quarta (partitional information) es basa en un canal d’informació entre les regions d’una imatge i els intervals de l’histograma
Tew, Kevin. "Skuery : manipulation of S-expressions using Xquery techniques /". Diss., CLICK HERE for online access, 2006. http://contentdm.lib.byu.edu/ETD/image/etd1677.pdf.
Pełny tekst źródłaVoulgaris, Georgios. "Techniques for content-based image characterization in wavelets domain". Thesis, University of South Wales, 2008. https://pure.southwales.ac.uk/en/studentthesis/techniques-for-contentbased-image-characterization-in-wavelets-domain(14c72275-a91e-4ba7-ada8-bdaee55de194).html.
Pełny tekst źródłaTeng, Shyh Wei 1973. "Image indexing and retrieval based on vector quantization". Monash University, Gippsland School of Computing and Information Technology, 2003. http://arrow.monash.edu.au/hdl/1959.1/5764.
Pełny tekst źródłaConser, Erik Timothy. "Improved Scoring Models for Semantic Image Retrieval Using Scene Graphs". PDXScholar, 2017. https://pdxscholar.library.pdx.edu/open_access_etds/3879.
Pełny tekst źródłaHuang, Ranxi. "Semi-automated techniques for the retrieval of dermatological condition in color skin images /". Online version of thesis, 2009. http://hdl.handle.net/1850/11355.
Pełny tekst źródłaFaichney, Jolon. "Content-Based Retrieval of Digital Video". Thesis, Griffith University, 2005. http://hdl.handle.net/10072/365697.
Pełny tekst źródłaThesis (PhD Doctorate)
Doctor of Philosophy (PhD)
School of Information Technology
Science, Environment, Engineering and Technology
Full Text
PIRAS, LUCA. "Interactive search techniques for content-based retrieval from archives of images". Doctoral thesis, Università degli Studi di Cagliari, 2011. http://hdl.handle.net/11584/266315.
Pełny tekst źródłaMoreux, Jean-Philippe, i Guillaume Chiron. "Image Retrieval in Digital Libraries: A Large Scale Multicollection Experimentation of Machine Learning techniques". Sächsische Landesbibliothek - Staats- und Universitätsbibliothek Dresden, 2017. https://slub.qucosa.de/id/qucosa%3A16444.
Pełny tekst źródłaSi historiquement, les bibliothèques numériques patrimoniales furent d’abord alimentées par des images, elles profitèrent rapidement de la technologie OCR pour indexer les collections imprimées afin d’améliorer périmètre et performance du service de recherche d’information offert aux utilisateurs. Mais l’accès aux ressources iconographiques n’a pas connu les mêmes progrès et ces dernières demeurent dans l’ombre : indexation manuelle lacunaire, hétérogène et non viable à grande échelle ; silos documentaires par genre iconographique ; recherche par le contenu (CBIR, content-based image retrieval) encore peu opérationnelle sur les collections patrimoniales. Aujourd’hui, il serait pourtant possible de mieux valoriser ces ressources, en particulier en exploitant les énormes volumes d’OCR produits durant les deux dernières décennies (tant comme descripteur textuel que pour l’identification automatique des illustrations imprimées). Et ainsi mettre en valeur ces gravures, dessins, photographies, cartes, etc. pour leur valeur propre mais aussi comme point d’entrée dans les collections, en favorisant découverte et rebond de document en document, de collection à collection. Cet article décrit une approche ETL (extract-transform-load) appliquée aux images d’une bibliothèque numérique à vocation encyclopédique : identifier et extraire l’iconographie partout où elle se trouve (dans les collections image mais aussi dans les imprimés : presse, revue, monographie) ; transformer, harmoniser et enrichir ses métadonnées descriptives grâce à des techniques d’apprentissage machine – machine learning – pour la classification et l’indexation automatiques ; charger ces données dans une application web dédiée à la recherche iconographique (ou dans d’autres services de la bibliothèque). Approche qualifiée de pragmatique à double titre, puisqu’il s’agit de valoriser des ressources numériques existantes et de mettre à profit des technologies (quasiment) mâtures.
Bosilj, Petra. "Image indexing and retrieval using component trees". Thesis, Lorient, 2016. http://www.theses.fr/2016LORIS396/document.
Pełny tekst źródłaThis thesis explores component trees, hierarchical structures from Mathematical Morphology, and their application to image retrieval and related tasks. The distinct component trees are analyzed and a novel classification into two superclasses is proposed, as well as a contribution to indexing and representation of the hierarchies using dendrograms. The first contribution to the field of image retrieval is in developing a novel feature detector, built upon the well-established MSER detection. The tree-based implementation of the MSER detector allows for changing the underlying tree in order to produce features of different stability properties. This resulted in the Tree of Shapes based Maximally Stable Region detector, leading to improvements over MSER in retrieval performance. Focusing on feature description, we extend the concept of 2D pattern spectra and adapt their global variant to more powerful, local schemes. Computed on the components of Min/Max-tree, they are histograms holding the information on distribution of image region attributes. The rotation and translation invariance is preserved from the global descriptor, while special attention is given to achieving scale invariance. We report comparable results to SIFT in image classification, as well as outperforming Morphology-based descriptors in satellite image retrieval, with a descriptor shorter than SIFT. Finally, a preprocessing or simplification technique for component trees is also presented, allowing the user to reevaluate the measures of region level of aggregation imposed on a component tree. The thesis is concluded by outlining the future perspectives based on the content of the thesis
Goodrum, Abby A. (Abby Ann). "Evaluation of Text-Based and Image-Based Representations for Moving Image Documents". Thesis, University of North Texas, 1997. https://digital.library.unt.edu/ark:/67531/metadc500441/.
Pełny tekst źródłaZlatoff, Nicolas. "Indexation d'images 2D : vers une reconnaissance d'objets multi-critèresContent-based image retrieval : On the way to object features". Lyon, INSA, 2006. http://theses.insa-lyon.fr/publication/2006ISAL0039/these.pdf.
Pełny tekst źródłaHuge volume of numeric images has recently led to strong needs for indexing and retrieval tools. Indexing an image consists in extracting a signature from it. Then, retrieving an image from an image database implies to compare several signatures together. We call content-based image retrieval systems those which build a signature from image low-level signal features such as color or texture. Such systems face a crucial limitation today. As a matter of fact, they allow to retrieve an image based on signal point of view, while users usually seek a more semantic-based search, related to what the image depicts (objects for instance). In this thesis, we have proposed an indexing system which may allow to bridge the gap between low-level features and semantic. First, the user has to formulate a kind of model (prototype) for the object sought. Then, while comparing this model which each image from the database, several features are considered, such as shape but also structural relationships between some regions of interest. The extraction of those regions remains an open and challenging problem. Segmentation approaches are often error-prone, because of artifacts from tight variations in illumination of the scene. That is why we do not describe an image with one unique segmentation, but rather with a hierarchy of segmentations. This represents the image at several levels of detail. It is build by iterative perceptual groupings on regions, considering both low-level and geometric features. When comparing a model with an image, we use one-to-one matching between model parts and regions from image, instead of considering the model in its whole. More precisely, comparison is based on shape similarity (through Angular Radial Transform and Curvature Scale Space) and on structural relationships among parts of object. All these features are then combined together, using Dempster-Shafer theory of belief, in order to derive one single similarity measure
O'Connor, Maureen J. Patillo Paul J. "Reengineering human performance and fatigue research through use of physiological monitoring devices, web-based and mobile device data collection methods, and integrated data storage techniques /". Monterey, Calif. : Springfield, Va. : Naval Postgraduate School ; Available from National Technical Information Service, 2003. http://library.nps.navy.mil/uhtbin/hyperion-image/03Dec%5FO'Connor.pdf.
Pełny tekst źródłaThesis advisor(s): Nita L. Miller, Thomas J. Housel. Includes bibliographical references (p. 115-117). Also available online.
Cáceres, Sheila Maricela Pinto. "Técnicas de visualização para sistemas de recuperação de imagens por conteúdo". [s.n.], 2010. http://repositorio.unicamp.br/jspui/handle/REPOSIP/275783.
Pełny tekst źródłaDissertação (mestrado) - Universidade Estadual de Campinas, Instituto de Computação
Made available in DSpace on 2018-08-17T03:09:27Z (GMT). No. of bitstreams: 1 Caceres_SheilaMaricelaPinto_M.pdf: 5019224 bytes, checksum: cf87ae15c741c8322d4d398124abee74 (MD5) Previous issue date: 2010
Resumo: Um sistema de Recuperação de Imagens por Conteúdo (CBIR) oferece mecanismos necessários para busca e recuperação de imagens baseando-se em propriedades visuais como cor, textura, forma, etc. Em um processo de busca de imagens, a apresentação de resultados é um componente essencial, na medida em que a obtenção desses resultados é o motivo da existência do sistema. Consequentemente, o uso de técnicas de visualização apropriadas pode determinar o sucesso ou o fracasso de um sistema CBIR. Técnicas de visualização são valiosas ferramentas na exploração de grandes quantidades de dados, como coleções de imagens. Contudo, técnicas para visualizar imagens retornadas por sistemas CBIR têm sido pobremente exploradas. Este trabalho apresenta um estudo comparativo e avaliação de várias técnicas de visualização para sistemas CBIR. Como resultado desse estudo, propõe-se um conjunto de técnicas originais que tentam suprir algumas das limitações identificadas em métodos da literatura. Dentre as características das técnicas propostas, destacam-se o enfoque baseado no centro e o uso de técnicas de agrupamento de dados para representar a similaridade intrínseca entre as imagens retornadas. Resultados experimentais mostram que os métodos propostos superam outras estratégias de visualização, considerando-se diversos critérios, como adequação para mostrar resultados em sistemas CBIR, quantidade de informação oferecida, satisfação de usuário, etc. As principais contribuições deste trabalho são: (i) estudo comparativo e análise de sete técnicas de visualização, quatro delas existentes na literatura e três técnicas novas propostas; (ii) avaliação de duas técnicas da literatura nunca antes avaliadas: anéis concêntricos e espiral; (iii) especificação e implementação de três novas técnicas de visualização baseadas em agrupamento; (iv) especificação e implementação de um framework para desenvolvimento de novas estruturas visuais para sistemas CBIR no qual foram implementadas as técnicas de visualização estudadas
Abstract: A Content-Based Image Retrieval (CBIR) system offers mechanisms needed to search and retrieve images based on visual properties such as color, texture, shape, etc. In an image search process, the presentation of results is an essential component as the retrieval of relevant images is the reason of the system existence. Consequently, the use of appropriate visualization techniques may determine the success of a CBIR system. Visualization techniques are valuable tools for the exploration of a great quantity of data, such as images collections. However, techniques for visualizing images in CBIR systems have been poorly explored. This work presents a comparative study of several visualization techniques for CBIR systems. As a result of this study, several original techniques were proposed trying to fulfill some of the absent characteristics in existing methods, such as the central-based focus and the use of clustering approaches to represent the intrinsic similarity between retrieved images. Experimental results show that the proposed methods overcome other visualization strategies by considering several criteria such as adaptation to show CBIR results, information load, user satisfaction, etc. The main contributions of this work are: (i) comparative study and analysis of seven visualization techniques, four of them from the literature and three new ones; (ii) validation of two techniques never evaluated before: concentric rings and spiral; (iii) specification and implementation of three new techniques of visualization based on clustering; (iv) specification and implementation of a framework for developing new visual structures for content-based image retrieval systems. The studied techniques were implemented by using this framework
Mestrado
Ciência da Computação
Mestre em Ciência da Computação
Le, Huu Ton. "Improving image representation using image saliency and information gain". Thesis, Poitiers, 2015. http://www.theses.fr/2015POIT2287/document.
Pełny tekst źródłaNowadays, along with the development of multimedia technology, content based image retrieval (CBIR) has become an interesting and active research topic with an increasing number of application domains: image indexing and retrieval, face recognition, event detection, hand writing scanning, objects detection and tracking, image classification, landmark detection... One of the most popular models in CBIR is Bag of Visual Words (BoVW) which is inspired by Bag of Words model from Information Retrieval field. In BoVW model, images are represented by histograms of visual words from a visual vocabulary. By comparing the images signatures, we can tell the difference between images. Image representation plays an important role in a CBIR system as it determines the precision of the retrieval results.In this thesis, image representation problem is addressed. Our first contribution is to propose a new framework for visual vocabulary construction using information gain (IG) values. The IG values are computed by a weighting scheme combined with a visual attention model. Secondly, we propose to use visual attention model to improve the performance of the proposed BoVW model. This contribution addresses the importance of saliency key-points in the images by a study on the saliency of local feature detectors. Inspired from the results from this study, we use saliency as a weighting or an additional histogram for image representation.The last contribution of this thesis to CBIR shows how our framework enhances the BoVP model. Finally, a query expansion technique is employed to increase the retrieval scores on both BoVW and BoVP models
Li, Honglin. "Hierarchical video semantic annotation the vision and techniques /". Connect to this title online, 2003. http://rave.ohiolink.edu/etdc/view?acc%5Fnum=osu1071863899.
Pełny tekst źródłaTitle from first page of PDF file. Document formatted into pages; contains xv, 146 p.; also includes graphics. Includes bibliographical references (p. 136-146).
Bahga, Arshdeep. "Technologies for context based video search". Thesis, Georgia Institute of Technology, 2010. http://hdl.handle.net/1853/33824.
Pełny tekst źródłaXiong, Li. "Resilient Reputation and Trust Management: Models and Techniques". Diss., Georgia Institute of Technology, 2005. http://hdl.handle.net/1853/7483.
Pełny tekst źródłaTao, Cui. "Schema Matching and Data Extraction over HTML Tables". Diss., CLICK HERE for online access, 2003. http://contentdm.lib.byu.edu/ETD/image/etd279.pdf.
Pełny tekst źródłaRasheed, Zeeshan. "Video categorization using semantics and semiotics". Doctoral diss., University of Central Florida, 2003. http://digital.library.ucf.edu/cdm/ref/collection/RTD/id/2888.
Pełny tekst źródłaThere is a great need to automatically segment, categorize, and annotate video data, and to develop efficient tools for browsing and searching. We believe that the categorization of videos can be achieved by exploring the concepts and meanings of the videos. This task requires bridging the gap between low-level content and high-level concepts (or semantics). Once a relationship is established between the low-level computable features of the video and its semantics, .the user would be able to navigate through videos through the use of concepts and ideas (for example, a user could extract only those scenes in an action film that actually contain fights) rat her than sequentially browsing the whole video. However, this relationship must follow the norms of human perception and abide by the rules that are most often followed by the creators (directors) of these videos. These rules are called film grammar in video production literature. Like any natural language, this grammar has several dialects, but it has been acknowledged to be universal. Therefore, the knowledge of film grammar can be exploited effectively for the understanding of films. To interpret an idea using the grammar, we need to first understand the symbols, as in natural languages, and second, understand the rules of combination of these symbols to represent concepts. In order to develop algorithms that exploit this film grammar, it is necessary to relate the symbols of the grammar to computable video features.
Ph.D.
Doctorate;
School of Electrical Engineering and Computer Science
Engineering and Computer Science
Electrical Engineering and Computer Science
120 p.
xix, 120 leaves, bound : ill., (some col.) ; 28 cm.
Awad, Dounia. "Vers un système perceptuel de reconnaissance d'objets". Thesis, La Rochelle, 2014. http://www.theses.fr/2014LAROS017/document.
Pełny tekst źródłaThe main objective of this thesis is to propose a pipeline for an object recognition algorithm, near to human perception, and at the same time, address the problems of Content Based image retrieval (CBIR) algorithm complexity : query run time and memory allocation. In this context, we propose a filter based on visual attention system to select salient points according to human interests from the interest points extracted by a traditionnal interest points detectors. The test of our approach, using Perreira Da Silva’s system as filter, on VOC 2005 databases, demonstrated that we can maintain approximately the same performance of a object recognition system by selecting only 40% of interest points (extracted by Harris-Laplace and Laplacian), while having an important gain in complexity (40% gain in query-run time and 60% in complexity). Furthermore, we address the problem of high dimensionality of descriptor in object recognition system. We proposed a new hybrid texture descriptor, representing the spatial frequency of some perceptual features extracted by a visual attention system. This descriptor has the advantage of being lower dimension vs. traditional descriptors. Evaluating our descriptor with an object recognition system (interest points detectors are Harris-Laplace & Laplacian) on VOC 2007 databases showed a slightly decrease in the performance (with 5% loss in Average Precision) compared to the original system, based on SIFT descriptor (with 50% complexity gain). In addition, we evaluated our descriptor using a visual attention system as interest point detector, on VOC 2005 databases. The experiment showed a slightly decrease in performance (with 3% loss in performance), meanwhile we reduced drastically the complexity of the system (with 50% gain in run-query time and 70% in complexity)
Vemulapalli, Smita. "Audio-video based handwritten mathematical content recognition". Diss., Georgia Institute of Technology, 2012. http://hdl.handle.net/1853/45958.
Pełny tekst źródłaAndrade, Felipe dos Santos Pinto de 1986. "Combinação de descritores locais e globais para recuperação de imagens e vídeos por conteúdo". [s.n.], 2012. http://repositorio.unicamp.br/jspui/handle/REPOSIP/275668.
Pełny tekst źródłaDissertação (mestrado) - Universidade Estadual de Campinas, Instituto de Computação
Made available in DSpace on 2018-08-22T03:54:24Z (GMT). No. of bitstreams: 1 Andrade_FelipedosSantosPintode_M.pdf: 3172421 bytes, checksum: cf06d3683b1828f22508de3f77ed1c00 (MD5) Previous issue date: 2012
Resumo: Recentemente, a fusão de descritores tem sido usada para melhorar o desempenho de sistemas de busca em tarefas de recuperação de imagens e vídeos. Descritores podem ser globais ou locais, dependendo de como analisam o conteúdo visual. A maioria dos trabalhos existentes tem se concentrado na fusão de um tipo de descritor. Este trabalho objetiva analisar o impacto da combinação de descritores locais e globais. Realiza-se um estudo comparativo de diferentes tipos de descritores e todas suas possíveis combinações. Além disso, investigam-se modelos para extração e a comparação das características globais e locais para recuperação de imagens e vídeos e estuda-se a utilização da técnica de programação genética para combinar esses descritores. Experimentos extensivos baseados em um projeto experimental rigoroso mostram que descritores locais e globais complementam-se quando combinados. Além disso, esta combinação produz resultados superiores aos observados para outras combinações e ao uso dos descritores individualmente
Abstract: Recently, fusion of descriptors has become a trend for improving the performance in image and video retrieval tasks. Descriptors can be global or local, depending on how they analyze visual content. Most of existing works have focused on the fusion of a single type of descriptor. Different from all of them, this work aims at analyzing the impact of combining global and local descriptors. Here, we perform a comparative study of different types of descriptors and all of their possible combinations. Furthermore, we investigate different models for extracting and comparing local and global features of images and videos, and evaluate the use of genetic programming as a suitable alternative for combining local and global descriptors. Extensive experiments following a rigorous experimental design show that global and local descriptors complement each other, such that, when combined, they outperform other combinations or single descriptors
Mestrado
Ciência da Computação
Mestre em Ciência da Computação
Gorisse, David. "Passage à l’échelle des méthodes de recherche sémantique dans les grandes bases d’images". Thesis, Cergy-Pontoise, 2010. http://www.theses.fr/2010CERG0519/document.
Pełny tekst źródłaIn this last decade, would the digital revolution and its ancillary consequence of a massive increases in digital picture quantities. The database size grow much faster than the processing capacity of computers. The current search engine which conceived for small data volumes do not any more allow to make searches in these new corpus with acceptable response times for users.In this thesis, we propose scalable content-based image retrieval engines.At first, we considered automatic search engines where images are indexed with global histograms. Secondly, we were interested in more sophisticated engines allowing to improve the search quality by working with bag of feature. In a last time, we proposed a strategy to reduce the complexity of interactive search engines. These engines allow to improve the results by using labels which the users supply to the system during the search sessions
Borba, Gustavo Benvenutti. "Automatic extraction of regions of interest from images based on visual attention models". Universidade Tecnológica Federal do Paraná, 2010. http://repositorio.utfpr.edu.br/jspui/handle/1/1295.
Pełny tekst źródłaEsta tese apresenta um método para a extração de regiões de interesse (ROIs) de imagens. No contexto deste trabalho, ROIs são definidas como os objetos semânticos que se destacam em uma imagem, podendo apresentar qualquer tamanho ou localização. O novo método baseia-se em modelos computacionais de atenção visual (VA), opera de forma completamente bottom-up, não supervisionada e não apresenta restrições com relação à categoria da imagem de entrada. Os elementos centrais da arquitetura são os modelos de VA propostos por Itti-Koch-Niebur e Stentiford. O modelo de Itti-Koch-Niebur considera as características de cor, intensidade e orientação da imagem e apresenta uma resposta na forma de coordenadas, correspondentes aos pontos de atenção (POAs) da imagem. O modelo Stentiford considera apenas as características de cor e apresenta a resposta na forma de áreas de atenção na imagem (AOAs). Na arquitetura proposta, a combinação de POAs e AOAs permite a obtenção dos contornos das ROIs. Duas implementações desta arquitetura, denominadas 'primeira versão' e 'versão melhorada' são apresentadas. A primeira versão utiliza principalmente operações tradicionais de morfologia matemática. Esta versão foi aplicada em dois sistemas de recuperação de imagens com base em regiões. No primeiro, as imagens são agrupadas de acordo com as ROIs, ao invés das características globais da imagem. O resultado são grupos de imagens mais significativos semanticamente, uma vez que o critério utilizado são os objetos da mesma categoria contidos nas imagens. No segundo sistema, á apresentada uma combinação da busca de imagens tradicional, baseada nas características globais da imagem, com a busca de imagens baseada em regiões. Ainda neste sistema, as buscas são especificadas através de mais de uma imagem exemplo. Na versão melhorada da arquitetura, os estágios principais são uma análise de coerência espacial entre as representações de ambos modelos de VA e uma representação multi-escala das AOAs. Se comparada à primeira versão, esta apresenta maior versatilidade, especialmente com relação aos tamanhos das ROIs presentes nas imagens. A versão melhorada foi avaliada diretamente, com uma ampla variedade de imagens diferentes bancos de imagens públicos, com padrões-ouro na forma de bounding boxes e de contornos reais dos objetos. As métricas utilizadas na avaliação foram presision, recall, F1 e area of overlap. Os resultados finais são excelentes, considerando-se a abordagem exclusivamente bottom-up e não-supervisionada do método.
This thesis presents a method for the extraction of regions of interest (ROIs) from images. By ROIs we mean the most prominent semantic objects in the images, of any size and located at any position in the image. The novel method is based on computational models of visual attention (VA), operates under a completely bottom-up and unsupervised way and does not present con-straints in the category of the input images. At the core of the architecture is de model VA proposed by Itti, Koch and Niebur and the one proposed by Stentiford. The first model takes into account color, intensity, and orientation features and provides coordinates corresponding to the points of attention (POAs) in the image. The second model considers color features and provides rough areas of attention (AOAs) in the image. In the proposed architecture, the POAs and AOAs are combined to establish the contours of the ROIs. Two implementations of this architecture are presented, namely 'first version' and 'improved version'. The first version mainly on traditional morphological operations and was applied in two novel region-based image retrieval systems. In the first one, images are clustered on the basis of the ROIs, instead of the global characteristics of the image. This provides a meaningful organization of the database images, since the output clusters tend to contain objects belonging to the same category. In the second system, we present a combination of the traditional global-based with region-based image retrieval under a multiple-example query scheme. In the improved version of the architecture, the main stages are a spatial coherence analysis between both VA models and a multiscale representation of the AOAs. Comparing to the first one, the improved version presents more versatility, mainly in terms of the size of the extracted ROIs. The improved version was directly evaluated for a wide variety of images from different publicly available databases, with ground truth in the form of bounding boxes and true object contours. The performance measures used were precision, recall, F1 and area overlap. Experimental results are of very high quality, particularly if one takes into account the bottom-up and unsupervised nature of the approach.
Gbehounou, Syntyche. "Indexation de bases d'images : évaluation de l'impact émotionnel". Thesis, Poitiers, 2014. http://www.theses.fr/2014POIT2295/document.
Pełny tekst źródłaThe goal of this work is to propose an efficient approach for emotional impact recognition based on CBIR techniques (descriptors, image representation). The main idea relies in classifying images according to their emotion which can be "Negative", "Neutral" or "Positive". Emotion is related to the image content and also to the personnal feelings. To achieve our goal we firstly need a correct assessed image database. Our first contribution is about this aspect. We proposed a set of 350 diversifed images rated by people around the world. Added to our choice to use CBIR methods, we studied the impact of visual saliency for the subjective evaluations and interest region segmentation for classification. The results are really interesting and prove that the CBIR methods are usefull for emotion recognition. The chosen desciptors are complementary and their performance are consistent on the database we have built and on IAPS, reference database for the analysis of the image emotional impact
Allani, Atig Olfa. "Une approche de recherche d'images basée sur la sémantique et les descripteurs visuels". Thesis, Paris 8, 2017. http://www.theses.fr/2017PA080032.
Pełny tekst źródłaImage retrieval is a very active search area. Several image retrieval approaches that allow mapping between low-level features and high-level semantics have been proposed. Among these, one can cite object recognition, ontologies, and relevance feedback. However, their main limitation concern their high dependence on reliable external resources and lack of capacity to combine semantic and visual information.This thesis proposes a system based on a pattern graph combining semantic and visual features, relevant visual feature selection for image retrieval and improvement of results visualization. The idea is (1) build a pattern graph composed of a modular ontology and a graph-based model, (2) to build visual feature collections to guide feature selection during online retrieval phase and (3) improve the retrieval results visualization with the integration of semantic relations.During the pattern graph building, ontology modules associated to each domain are automatically built using textual corpuses and external resources. The region's graphs summarize the visual information in a condensed form and classify it given its semantics. The pattern graph is obtained using modules composition. In visual features collections building, association rules are used to deduce the best practices on visual features use for image retrieval. Finally, results visualization uses the rich information on images to improve the results presentation.Our system has been tested on three image databases. The results show an improvement in the research process, a better adaptation of the visual features to the domains and a richer visualization of the results
Cantalloube, Faustine. "Détection et caractérisation d'exoplanètes dans des images à grand contraste par la résolution de problème inverse". Thesis, Université Grenoble Alpes (ComUE), 2016. http://www.theses.fr/2016GREAY017/document.
Pełny tekst źródłaDirect imaging of exoplanets provides valuable information about the light they emit, their interactions with their host star environment and their nature. In order to image such objects, advanced data processing tools adapted to the instrument are needed. In particular, the presence of quasi-static speckles in the images, due to optical aberrations distorting the light from the observed star, prevents planetary signals from being distinguished. In this thesis, I present two innovative image processing methods, both based on an inverse problem approach, enabling the disentanglement of the quasi-static speckles from the planetary signals. My work consisted of improving these two algorithms in order to be able to process on-sky images.The first one, called ANDROMEDA, is an algorithm dedicated to point source detection and characterization via a maximum likelihood approach. ANDROMEDA makes use of the temporal diversity provided by the image field rotation during the observation, to recognize the deterministic signature of a rotating companion over the stellar halo. From application of the original version on real data, I have proposed and qualified improvements in order to deal with the non-stable large scale structures due to the adaptative optics residuals and with the remaining level of correlated noise in the data. Once ANDROMEDA became operational on real data, I analyzed its performance and its sensitivity to the user-parameters proving the robustness of the algorithm. I also conducted a detailed comparison to the other algorithms widely used by the exoplanet imaging community today showing that ANDROMEDA is a competitive method with practical advantages. In particular, it is the only method that allows a fully unsupervised detection. By the numerous tests performed on different data set, ANDROMEDA proved its reliability and efficiency to extract companions in a rapid and systematic way (with only one user parameter to be tuned). From these applications, I identified several perspectives whose implementation could significantly improve the performance of the pipeline.The second algorithm, called MEDUSAE, consists in jointly estimating the aberrations (responsible for the speckle field) and the circumstellar objects by relying on a coronagraphic image formation model. MEDUSAE exploits the spectral diversity provided by multispectral data. In order to In order to refine the inversion strategy and probe the most critical parameters, I applied MEDUSAE on a simulated data set generated with the model used in the inversion. To investigate further the impact of the discrepancy between the image model used and the real images, I applied the method on realistic simulated images. At last, I applied MEDUSAE on real data and from the preliminary results obtained, I identified the important input required by the method and proposed leads that could be followed to make this algorithm operational to process on-sky data
Dang, Quoc Bao. "Information spotting in huge repositories of scanned document images". Thesis, La Rochelle, 2018. http://www.theses.fr/2018LAROS024/document.
Pełny tekst źródłaThis work aims at developing a generic framework which is able to produce camera-based applications of information spotting in huge repositories of heterogeneous content document images via local descriptors. The targeted systems may take as input a portion of an image acquired as a query and the system is capable of returning focused portion of database image that match the query best. We firstly propose a set of generic feature descriptors for camera-based document images retrieval and spotting systems. Our proposed descriptors comprise SRIF, PSRIF, DELTRIF and SSKSRIF that are built from spatial space information of nearest keypoints around a keypoints which are extracted from centroids of connected components. From these keypoints, the invariant geometrical features are considered to be taken into account for the descriptor. SRIF and PSRIF are computed from a local set of m nearest keypoints around a keypoint. While DELTRIF and SSKSRIF can fix the way to combine local shape description without using parameter via Delaunay triangulation formed from a set of keypoints extracted from a document image. Furthermore, we propose a framework to compute the descriptors based on spatial space of dedicated keypoints e.g SURF or SIFT or ORB so that they can deal with heterogeneous-content camera-based document image retrieval and spotting. In practice, a large-scale indexing system with an enormous of descriptors put the burdens for memory when they are stored. In addition, high dimension of descriptors can make the accuracy of indexing reduce. We propose three robust indexing frameworks that can be employed without storing local descriptors in the memory for saving memory and speeding up retrieval time by discarding distance validating. The randomized clustering tree indexing inherits kd-tree, kmean-tree and random forest from the way to select K dimensions randomly combined with the highest variance dimension from each node of the tree. We also proposed the weighted Euclidean distance between two data points that is computed and oriented the highest variance dimension. The secondly proposed hashing relies on an indexing system that employs one simple hash table for indexing and retrieving without storing database descriptors. Besides, we propose an extended hashing based method for indexing multi-kinds of features coming from multi-layer of the image. Along with proposed descriptors as well indexing frameworks, we proposed a simple robust way to compute shape orientation of MSER regions so that they can combine with dedicated descriptors (e.g SIFT, SURF, ORB and etc.) rotation invariantly. In the case that descriptors are able to capture neighborhood information around MSER regions, we propose a way to extend MSER regions by increasing the radius of each region. This strategy can be also applied for other detected regions in order to make descriptors be more distinctive. Moreover, we employed the extended hashing based method for indexing multi-kinds of features from multi-layer of images. This system are not only applied for uniform feature type but also multiple feature types from multi-layers separated. Finally, in order to assess the performances of our contributions, and based on the assessment that no public dataset exists for camera-based document image retrieval and spotting systems, we built a new dataset which has been made freely and publicly available for the scientific community. This dataset contains portions of document images acquired via a camera as a query. It is composed of three kinds of information: textual content, graphical content and heterogeneous content
Basunia, Mahmudunnabi. "A Recursive Phase Retrieval Technique Using Transport of Intensity: Reconstruction of Imaged Phase and 3D Surfaces". University of Dayton / OhioLINK, 2016. http://rave.ohiolink.edu/etdc/view?acc_num=dayton1481049563470488.
Pełny tekst źródłaFaessel, Nicolas. "Indexation et interrogation de pages web décomposées en blocs visuels". Thesis, Aix-Marseille 3, 2011. http://www.theses.fr/2011AIX30014/document.
Pełny tekst źródłaThis thesis is about indexing and querying Web pages. We propose a new model called BlockWeb, based on the decomposition of Web pages into a hierarchy of visual blocks. This model takes in account the visual importance of each block as well as the permeability of block's content to their neighbor blocks on the page. Splitting up a page into blocks has several advantages in terms of indexing and querying. It allows to query the system with a finer granularity than the whole page: the most similar blocks to the query can be returned instead of the whole page. A page is modeled as a directed acyclic graph, the IP graph, where each node is associated with a block and is labeled by the coefficient of importance of this block and each arc is labeled by the coefficient of permeability of the target node content to the source node content. In order to build this graph from the bloc tree representation of a page, we propose a new language : XIML (acronym for XML Indexing Management Language), a rule based language like XSLT. The model has been assessed on two distinct dataset: finding the best entry point in a dataset of electronic newspaper articles, and images indexing and querying in a dataset drawn from web pages of the ImagEval 2006 campaign. We present the results of these experiments
Tarafdar, Arundhati. "Wordspotting from multilingual and stylistic documents". Thesis, Tours, 2017. http://www.theses.fr/2017TOUR4022/document.
Pełny tekst źródłaWord spotting in graphical documents is a very challenging task. To address such scenarios this thesis deals with developing a word spotting system dedicated to geographical documents with Bangla and English (Roman) scripts. In the proposed system, at first, text-graphics layers are separated using filtering, clustering and self-reinforcement through classifier. Additionally, instead of using binary decision we have used probabilistic measurement to represent the text components. Subsequently, in the text layer, character segmentation approach is applied using water-reservoir based method to extract individual character from the document. Then recognition of these isolated characters is done using rotation invariant feature, coupled with SVM classifier. Well recognized characters are then grouped based on their sizes. Initial spotting is started to find a query word among those groups of characters. In case if the system could spot a word partially due to any noise, SIFT is applied to identify missing portion of that partial spotting. Experimental results on Roman and Bangla scripts document images show that the method is feasible to spot a location in text labeled graphical documents. Experiments are done on an annotated dataset which was developed for this work. We have made this annotated dataset available publicly for other researchers
SIDHU, ARUNIMA. "ANALYSIS OF IMAGE RETRIEVAL TECHNIQUES". Thesis, 2016. http://dspace.dtu.ac.in:8080/jspui/handle/repository/15273.
Pełny tekst źródłaLin, Hui-Chuan, i 林惠娟. "A Study of Information Hiding and Image Retrieval Techniques for compressed Images". Thesis, 2006. http://ndltd.ncl.edu.tw/handle/47958365589227005176.
Pełny tekst źródła國立臺中技術學院
資訊科技與應用研究所
95
This thesis focuses on information steganography, copyright protection and image retrieval for compressed images. There are three schemes proposed in this thesis. The first proposed scheme is tree growth based watermarking technique. To safeguard the image rightful ownerships, a representative logo or owner information could be hidden in the host image. The operation of this scheme is to divide the codewords into two groups by tree growing structure. The copyright information is embedded during the vector quantization compression. This scheme is simple and robust to protect the copyright efficiently. The second proposed scheme in the thesis is a steganography technique based on the palette method. The secret information is hidden in a cover image to ensure the transmission security. In this method, the palette colors are divided into two groups using the de-clustering scheme. The median edge detector (MED) predictor and flags are applied in information hiding. It not only increases the hidden capacity, when retrieving the message, but the original image is also recovered at the same time. This method can solve the problem that the palette image has distortion after data hiding. On the other hand, as a result of rapid Internet growth, the amount of multimedia circulation almost increases by geometric series acceleration. Many content based image retrieval (CBIR) technologies have been proposed in literatures. But few for palette-based image were proposed. However, the palette-based images have been widely used on Internet. In this thesis, a new image retrieval scheme based on palette images is proposed. The palette color (PC) is used as the first index directly and the edge-direction histogram (EDH) as the second index. The best advantage is to leave out large computation. Furthermore, this kind of compression format may save two-thirds of the image space in the database.
Huang, Shiuan, i 黃暄. "Multi-Query Image Retrieval using CNN and SIFT Techniques". Thesis, 2016. http://ndltd.ncl.edu.tw/handle/7e8v85.
Pełny tekst źródła國立交通大學
電子工程學系 電子研究所
104
Due to the rapid growth of image number, the content-based image retrieval for a large database becomes an essential tool in image processing. Although there are many published studies on this topic, it is still a challenge to do an advanced search, for example, retrieving a specific building using a view different from the photographing angles in the database. In addition, if the user can provide additional images as the second or the third queries, how do we combine the information provided by these multiple queries? Thus, we propose a multi-query fusion method to achieve a better accuracy. In this study, we test two different types of features designed for retrieval purpose. We adopt the Scale-Invariant Feature Transform (SIFT) feature as the low-level feature and the Convolutional Neural Network (CNN) feature as the high-level feature for retrieval. In using the SIFT features, the Bag-of-Word is implemented using the Term Frequency–Inverse Document Frequency (TF–IDF) retrieval algorithm. The AlexNet is adopted as our CNN model and it is modified to the Siamese-Triplet Network to match the image retrieval purpose. The Network weights are pre-trained by ImageNet and are fine-tuned using specific landmark datasets in retrieval. We use the CNN as the feature extractor instead of the image classifier. The loss function calculates the similarity between the query and the similar images or dissimilar images. Several levels of data fusion methods are proposed. The first one is combining the features of the 6th layer features and the 7th layer features derived from CNN. The second one is combine the information provided by the SIFT features and the CNN features. The third one is combining the information provided by multiple queries. When appropriate, we try both early fusion concept and the late fusion concept. Our best proposed method can exceed most of the state-of-the-art retrieval methods for a single query. The multi-query retrieval can further increase the retrieval accuracy.
Tsai, Tsung Ting, i 蔡宗廷. "Content-Based Image Retrieval based on Search Engine Techniques". Thesis, 2008. http://ndltd.ncl.edu.tw/handle/13568655121103117364.
Pełny tekst źródła國立暨南國際大學
資訊工程學系
96
Content-based image retrieval has been an important research topic for a long time. However, to reduce the search time in a large image database remains a challenge problem. The information retrieval (IR), which is the core of the text search engine techniques, has some well-known and efficient methods which can be applied to search information in a large database. Therefore, our solution simply extracts some "visual words" from images, these are analogies to the "words" in articles, and we can apply those methods in the IR domain directly. Our method can be divided into the following three parts: (1) Extract visual words. We recursively divide an image into four equal-sized blocks, and then two methods are proposed to extract visual words from these blocks. (2) Build index. Create index between visual words and images, and the associated TF-IDF weights to the database. The key method in this part is the inverted index, which can reduce the time and computing resources when we searching words using the index. (3) Search images. We search the similar images of the query image on the created index. Two methods, (a) Count Match (b) Vector Model, are proposed to estimate the similarities between query image and images in the database. We have evaluated the proposed methods on the image databases crawled from the auction webpages.
Liu, Yi-Min, i 柳依旻. "Adaptive Relevance Feedback Techniques for Content-Based Image Retrieval". Thesis, 2007. http://ndltd.ncl.edu.tw/handle/18126839365646713051.
Pełny tekst źródła國立暨南國際大學
資訊管理學系
95
Due to the popularity of Internet and the growing demand of image access, the volume of image databases is exploding. Hence, we need a more efficient and effective image searching technology. Relevance feedback (RF) is an interaction process between the user and the system such that the user’s information need is satisfied by the retrievals from the system. Traditional RF techniques use the same system parameter values for all types of query images. It is questionable that the best performance can be obtained through such setting. Hence, we propose self-adapting parameterization for the traditional relevance feedback approaches including the query vector modification (QVM) and the feature relevance estimation (FRE) methods using the particle swarm optimization. As such different system parameter values can be used to handle various types of queries, the retrieval system is thus more efficient and effective.