Dissertations / Theses on the topic 'Recherche automatique'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 dissertations / theses for your research on the topic 'Recherche automatique.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Pradilla, Magdalena. "Recherche de descripteurs en indexation automatique des documents /." Paris : Centre de recherche en informatique appliquée aux sciences sociales, 1987. http://catalogue.bnf.fr/ark:/12148/cb34973730h.
Full textMoreau, Fabienne. "Revisiter le couplage traitement automatique des langues et recherche d'information." Phd thesis, Université Rennes 1, 2006. http://tel.archives-ouvertes.fr/tel-00524514.
Full textRiadh, Ksantini. "Analyse multirésolution et recherche d'images." Sherbrooke : Université de Sherbrooke, 2003.
Find full textHaddoud, Mounia. "Indexation et extraction de termes-clés pour la recherche d'information." Rouen, 2016. http://www.theses.fr/2016ROUES010.
Full textIn this thesis, I focused on two text mining problems : automatic keyphrase extraction in text documents and texts classification. For the automatic keyphrase extraction problem, i define the document phrase maximality index (DPM-index), a new measure to discriminate overlapping keyphrase candidates in a text document. As an application i developed a supervised learning system which uses 18 statistical features, among them the DPMindex and 5 other new features. I experimentally compare my results to those of 21 keyphrase extraction methods on SemEval-2010/Task-5 scientific articles corpus. When all the systems extract 10 keyphrases per document, my method enhances by 13% the F-Score of the best system. In particular, the DPM-index feature increases the F-Score of my keyphrase extraction system by a rate of 9%. This makes the DPM-index contribution comparable to that of the well-known TFIDF measure on such a system. For the text classification problem, i propose 80 metrics never used for the term weighting problem and compare them to 16 functions of the literature. A large number of these metrics were initially proposed for other data mining problems : feature selection, classification rules and term collocations. While many previous works have shown the merits of using a particular metric, my experience suggests that the results obtained by such metrics can be highly dependent on the label distribution on the corpus and on the performance measures used (microaveraged or macroaveraged F1-Score). The solution I propose consists in combining the metrics in order to improve the classification. More precisely, i show that using a SVM classifier which combines the outputs of SVM classifiers that utilize different metrics performs well in all situations. The second main contribution is an extended term representation for the vector space model that improves significantly the prediction of the text classifier
Claveau, Vincent. "Acquisition automatique de lexiques sémantiques pour la recherche d'information." Phd thesis, Université Rennes 1, 2003. http://tel.archives-ouvertes.fr/tel-00524646.
Full textMorneau, Maxime. "Recherche d'information sémantique et extraction automatique d'ontologie du domaine." Thesis, Université Laval, 2006. http://www.theses.ulaval.ca/2006/23828/23828.pdf.
Full textIt can prove to be diffcult, even for a small size organization, to find information among hundreds, even thousands of electronic documents. Most often, the methods employed by search engines on the Internet are used by companies wanting to improve information retrieval on their intranet. These techniques rest on statistical methods and do not make it possible neither to evaluate the semantics contained in the user requests, nor in the documents. Certain methods were developed to extract this semantics and thus, to improve the answer given to requests. On the other hand, the majority of these techniques were conceived to be applied on the entire World Wide Web and not on a particular field of knowledge, like corporative data. It could be interesting to use domain specific ontologies in trying to link a specific query to related documents and thus, to be able to better answer these queries. This thesis presents our approach which proposes the use of the Text-To-Onto software to automatically create an ontology describing a particular field. Thereafter, this ontology is used by the Sesei software, which is a semantic filter for conventional search engines. This method makes it possible to improve the relevance of documents returned to the user.
Méger, Nicolas. "Recherche automatique des fenêtres temporelles optimales des motifs séquentiels." Lyon, INSA, 2004. http://theses.insa-lyon.fr/publication/2004ISAL0095/these.pdf.
Full textThis work addresses the problem of mining patterns under constraints in event sequences. Extracted patterns are episode rules. Our main contribution is an automatic search for optimal time window of each one of the episode rules. We propose to extract only rules having such an optimal time window. These rules are termed FLM-rules. We present an algorithm, WinMiner, that aims to extract FLM-rules, given a minimum support threshold, a minimum confidence threshold and a maximum gap constraint. Proofs of the correctness of this algorithm are supplied. We also propose a dedicated interest measure that aims to select FLM-rules such that their heads and bodies can be considered as dependant. Two applications are described. The first one is about mining medical datasets while the other one deals with seismic datasets
Méger, Nicolas Boulicaut Jean-François Rigotti Christophe. "Recherche automatique des fenêtres temporelles optimales des motifs séquentiels." Villeurbanne : Doc'INSA, 2005. http://docinsa.insa-lyon.fr/these/pont.php?id=meger.
Full textGrivolla, Jens. "Apprentissage et décision automatique en recherche documentaire : prédiction de difficulté de requêtes et sélection de modèle de recherche." Avignon, 2006. http://www.theses.fr/2006AVIG0142.
Full textThis thesis is centered around the subject of information retrieval, with a focus on those queries that are particularly difficult to handle for current retrieval systems. In the application and evaluation settings we were concerned with, a user expresses his information need as a natural language query. There are different approaches for treating those queries, but current systems typically use a single approach for all queries, without taking into account the specific properties of each query. However, it has been shown that the performance of one strategy relative to another can vary greatly depending on the query. We have approached this problem by proposing methods that will permit to automatically identify those queries that will pose particular difficulties to the retrieval system, in order to allow for a specific treatment. This research topic was very new and barely starting to be explored at the beginning of my work, but has received much attention these last years. We have developed a certain number of quality predictor functions that obtain results comparable to those published recently by other research teams. However, the ability of individual predictors to accurately classify queries by their level of difficulty remains rather limited. The major particularity and originality of our work lies in the combination of those different measures. Using methods of automatic classification with corpus-based training, we have been able to obtain quite reliable predictions, on the basis of measures that individually are far less discriminant. We have also adapted our approach to other application settings, with very encouraging results. We have thus developed a method for the selective application of query expansion techniques, as well as the selection of the most appropriate retrieval model for each query
Koutchoukali, Mohamed Mehdi. "Configuration automatique du moteur de recherche d'images via la sélection des paramètres de la recherche." Thèse, Université du Québec à Trois-Rivières, 2014. http://depot-e.uqtr.ca/7343/1/030673542.pdf.
Full textCandillier, Laurent Gilleron Rémi. "Apprentissage automatique de profils de lecteurs." [S.l.] : [s.n.], 2001. http://www.univ-lille1.fr/bustl-grisemine/pdf/memoires/A2001-6.pdf.
Full textLoupy, Claude de. "Evaluation de l'apport de connaissances linguistiques en desambigui͏̈sation sémantique et recherche documentaire." Avignon, 2000. http://www.theses.fr/2000AVIGA001.
Full textHajj, Hassan Ali. "Détection multidimensionnelle au test paramétrique avec recherche automatique des causes." Thesis, Grenoble, 2014. http://www.theses.fr/2014GRENM085/document.
Full textNowadays, control of manufacturing process is an essential task to ensure production of high quality. At the end of the semiconductor manufacturing process, an electric test, called Parametric Test (PT), is performed. The PT aims at detecting wafers whose electrical behavior is abnormal, based on a set of static electrical parameters measured on multiple sites of each wafer. The purpose of this thesis is to develop a dynamic detection system at PT level to detect abnormal wafers from a recent history of electrical measurements. For this, we develop a real time detection system based on an optimized learning technique, where training data and detection model are updated through a moving temporal window. The detection scheme is based on one class Support Vector Machines (1-SVM), a variant of the statistical learning algorithm SVM widely used for binary classification. 1-SVM was introduced in the context of one class classification problems for anomaly detection. In order to improve the predictive performance of the 1-SVM classification algorithm, two variable selection methods are developed. The first one is a filter method based on a calculated score with MADe filter, a robust approach for univariate outlier detection. The second one is of wrapper type that adapts the SVM Recursive Feature Elimination method (SVM-RFE) to the 1-SVM algorithm. For detected abnormal wafers, we propose a method to determine their multidimensional signatures to identify the electrical parameters responsible for the anomaly. Finally, we evaluate our proposed system on real datasets of STMicroelecronics and compare it to the detection system based on Hotelling's T2 test, one of the most known detection systems in the literature. The results show that our system yields very good performance and can provide an efficient way for real-time detection
Moreau, Fabienne Sébillot Pascale. "Revisiter le couplage traitement automatique des langues et recherche d'information." [S.l.] : [s.n.], 2006. ftp://ftp.irisa.fr/techreports/theses/2006/moreau.pdf.
Full textGurtner, Karine. "Extraction automatique de connaissances à partir de corpus de textes." Paris 7, 2000. http://www.theses.fr/2000PA077104.
Full textVinot, Romain. "Classification automatique de textes dans des catégories non thématiques." Phd thesis, Télécom ParisTech, 2004. http://pastel.archives-ouvertes.fr/pastel-00000812.
Full textBilhaut, Frédérik. "Analyse automatique de structures thématiques discursives - Application à la recherche d'information." Phd thesis, Université de Caen, 2006. http://tel.archives-ouvertes.fr/tel-00258766.
Full textEl, Jed Olfa. "WebSum : système de résumé automatique de réponses des moteurs de recherche." Toulouse 3, 2006. http://www.theses.fr/2006TOU30145.
Full textThis thesis lies within the general framework of the information retrieval and more precisely, within the framework of the web document classification and organization. Our objective is to develop a system of automatic summarizing of the search engine answers in the encyclopaedic style (WebSum). This type of summary aims at classifying the search engine answers according to the various topics or what we call in our work, facets of the user query. To carry out this objective, we propose : - A method of identification of the facets of a given query based on the generative lexicon; - An approach of classification of the search engine answers under this various facets; - And a method of evaluation of the relevance of the web pages
Bouzayani, Abdessalem. "Extension automatique de l'annotation d'images pour la recherche et la classification." Thesis, Université de Lorraine, 2018. http://www.theses.fr/2018LORR0045/document.
Full textThis thesis deals the problem of image annotation extension. Indeed, the fast growth of available visual contents has led a need for indexing and searching of multimedia information methods. Image annotation allows indexing and searching in a large collection of images in an easy and fast way. We wish, from partially manually annotated images databases, complete automatically the annotation of these sets, in order to make methods of research and / or classification of images more efficient. For automatic image annotation extension, we use probabilistic graphical models. The proposed model is based on a mixture of multinomial distributions and mixtures of Gaussian where we have combined visual and textual characteristics. To reduce the cost of manual annotation and improve the quality of the annotation obtained, we have incorporated user feedback into our model. User feedback was done using learning in learning, incremental learning and active learning. To reduce the semantic gap problem and to enrich the image annotation, we use a semantic hierarchy by modeling many semantic relationships between keywords. We present a semi-automatic method to build a semantic hierarchy from a set of keywords. After building the hierarchy, we integrate it into our image annotation model. The model obtained with this hierarchy is a mixture of Bernoulli distributions and Gaussian mixtures
Bueno, Steve. "L'activation automatique de la mémoire sémantique." Aix-Marseille 1, 2002. http://www.theses.fr/2002AIX10068.
Full textLambolez, Pierre-Yves. "Recherche d'informations pour la maintenance logicielle." Toulouse 3, 1994. http://www.theses.fr/1994TOU30201.
Full textAbdul, Rauf Sadaf. "Sélection de corpus en traduction automatique statistique." Phd thesis, Université du Maine, 2012. http://tel.archives-ouvertes.fr/tel-00732984.
Full textKermad, Chafik. "Segmentation d'image: recherche d'une mise en oeuvre automatique par coopération de méthodes." Phd thesis, Université Rennes 1, 1997. http://tel.archives-ouvertes.fr/tel-00008781.
Full textKERMAD, CHAFIK DJALAL. "Segmentation d'images : recherche d'une mise en uvre automatique par cooperation de methodes." Rennes 1, 1997. http://www.theses.fr/1997REN10109.
Full textLAMMALI, NASSIMA. "Recherche automatique du reseau hydrographique a l'aide du modele numerique de terrain." Toulouse 3, 1989. http://www.theses.fr/1989TOU30006.
Full textSoulé-Dupuy, Chantal. "Systèmes de recherche d'informations : le système videotex Infodiab : mécanismes d'indexation et d'interrogation." Toulouse 3, 1990. http://www.theses.fr/1990TOU30019.
Full textTawbe, Bilal. "Recherche d'images couleur à partir de leurs histogrammes." Sherbrooke : Université de Sherbrooke, 1999.
Find full textTirilly, Pierre. "Traitement automatique des langues pour l'indexation d'images." Phd thesis, Université Rennes 1, 2010. http://tel.archives-ouvertes.fr/tel-00516422.
Full textTollari, Sabrina. "Indexation et recherche d'images par fusion d'informations textuelles et visuelles." Toulon, 2006. http://www.theses.fr/2006TOUL0013.
Full textBecause of the exponential growing number of images, Image Retrieval Systems flmRS) must be scalable while reducing as much as possible the semantic gap. Usual web ImRS scan the associated text which only ensures scalability, on the other hand content based systems do not optimize semantic similarity. Robust ImRS may extract and merge textual arid visual informations from generally iniss-labeled and miss-segmented images. Therefore we propose two scalable approximation methods tested on COREL database (10K images, 150 words). We first train with miss-labeled images a simple bayesian image auto-annotation model using visual index based on Vector Approximation Files. This model is very fast and is better than priors (+29%). Second, we propose an adaptive reduction of the visual dimension for miss-labeled images. It relies on the hypothesis that each concept in a general image database is included in various visual contexts. We show theoritically under this weak assumption that we can approximate the Linear Discriminant Analysis (ALDA) and classification results using a simple ascendant hierarchical classification. Results demonstrate ALDA efficiency enhancing classification by 59% while selec¬ting only 10% of the visual dimensions. Approximation of Maximal Marginal Diversity (AMMD) is also proposed. This thesis demonstrates the interest of considering concept dependant visual features, and gives methods to extract them from real miss-labeled and miss-segmented large image databases, opening new perpectives for textuo-visual web analyses
Kherfi, Mohammed Lamine. "Rôle du contre-exemple dans le retour de pertinence en recherche d'images." Sherbrooke : Université de Sherbrooke, 2002.
Find full textFarooque, Mahfuza. "Techniques de déduction automatique vues comme recherche de preuve en calcul des séquents." Phd thesis, Ecole Polytechnique X, 2013. http://pastel.archives-ouvertes.fr/pastel-00961344.
Full textOuddan, Mohammed Amine. "Indexation et recherche des documents code source basées sur une caractérisation structuro-sémantique : application à la détection de plagiats." Université de Marne-la-Vallée, 2007. http://www.theses.fr/2007MARN0340.
Full textSource code characterization is a very complex task due the amount of similarity between computer science assignments. The various transformations that occur within a plagiarized code make the plagiarism detection more difficult. We propose a multilanguage source code retrieval system for plagiarism detection which is based on twolevel characterization approach. The first level reflects the syntactic feature of the code allowing a structural characterization of its content, and the second level relates to its functional feature allowing a semantic characterization. Our approach is based on the concept of Grammar with Actions which consists to assign significance to the parsing process in a context of characterization, and at the same time, allowing access to the structural and semantic content of the code using the grammar of its programming language. The aim idea is to translate the source code into a set of symbols sequences called characteristic sequences. In the first level of characterization we talk about structural sequences and in the second level we talk about genetic sequences. In order to quantify the similarity between characteristic sequences, we use sequence alignment techniques where the similarity rate is considered as an abstraction of the plagiarism rate between the characterized codes
Notin, Jean-Marc. "Recherche et construction de preuves en logique non-commutative." Nancy 1, 2004. http://www.theses.fr/2004NAN10183.
Full textPartially commutative logics allow to express properties mixing concurency and sequentiality. Thus, the logic NL extends linear logic with non-commutative connectives. The characteristic of NL comes from the interactions between commutative and non-commutative connectives. A first study led us to analyze these interactions within the framework of proof nets. Taking such interactions into account during top-down proof search (proof nets construction) requires the introduction of specific structures (labels, dependency sets). Thus, we propose several algorithms for building proof nets in the multiplicative fragment of NL (MNL). Another studied approach is bottom-up proof search, in particular within the framework of connection methods. By using labels associated with the subformulas, and constraints expressed on these labels, we propose a connection characterization for MNL. The associated connection method can be seen like a new algorithm for proof nets construction in MNL
Hallab, Mohamed. "Hypertextualisation automatique multilingue à partir des fréquences de N-grammes." Paris 8, 2001. http://www.theses.fr/2001PA082019.
Full textBougouin, Adrien. "Indexation automatique par termes-clés en domaines de spécialité." Nantes, 2015. https://archive.bu.univ-nantes.fr/pollux/show/show?id=50037b84-7248-4a67-a1c0-ecc3f737a44f.
Full textKeyphrases are words or multi-word expressions that represent the content of a document. Keyphrases give a synoptic view of a document and help to index it for information retrieval. This Ph. D thesis focuses on domain-specific automatic keyphrase annotation. Automatic keyphrase annotation is still a difficult task, and current systems do not achieve satisfactory results. Our work is divided in two steps. First, we propose a keyphrase candidate selection method that focuses on the categories of adjectives relevant within keyphrases and propose a method to rank them according to their importance within the document. This method, TopicRank, is a graph-based method that clusters keyphrase candidates into topics, ranks the topics and extracts one keyphrase per important topic. Our experiments show that TopicRank significantly outperforms other graph-basedmethods for automatic keyphrase annotation. Second, we focus on domain-specific documents and adapt our previous work. We study the best practice of manual keyphrase annotation by professional indexers andmimic it with a newmethod, TopicCoRank. TopicCoRank adds a new graph representing the specific domain to the topic graph of TopicRank. Leveraging this second graph, TopicCoRank possesses the rare ability to provide keyphrases that do not occur within documents. Applied on four corpora of four specific domains, TopicCoRank significantly outperforms TopicRank
Bigi, Brigitte. "Contribution à la modélisation du langage pour des applications de recherche documentaire et de traitement de la parole." Avignon, 2000. http://www.theses.fr/2000AVIG0125.
Full textSimon, François. "Evaluation de la performabilité des systèmes de production et des systèmes temps réel par réseaux de Petri stochastiques géneralisé." Mulhouse, 1996. http://www.theses.fr/1996MULH0456.
Full textLongo, Laurence. "Vers des moteurs de recherche "intelligents" : un outil de détection automatique de thèmes : méthode basée sur l'identification automatique des chaînes de référence." Phd thesis, Université de Strasbourg, 2013. http://tel.archives-ouvertes.fr/tel-00939243.
Full textVallès-Parlangeau, Nathalie. "Recherche d'événements articulatoires pertinents pour l'annotation automatique de la base de données multisensorielles ACCOR." Toulouse 3, 1998. http://www.theses.fr/1998TOU30295.
Full textSabri, Mohamed. "Filtrage et restauration en traitement des images numériques : recherche d'une mise en œuvre automatique." Rennes 1, 1991. http://www.theses.fr/1991REN10027.
Full textSakji, Saoussen. "Recherche d'information et indexation automatique des médicaments à l'aide de plusieurs terminologies de santé." Rouen, 2010. http://www.theses.fr/2010ROUES032.
Full textRoussey, Catherine. "Une Méthode d'indexation sémantique adaptée aux corpus multilingues." Lyon, INSA, 2001. http://theses.insa-lyon.fr/publication/2001ISAL0059/these.pdf.
Full textThis thesis deals with indexing problems of a multilingual corpus in an information retrieval system. Indexing procedure identifies the knowledge related to a text and represents it by keywords called descriptors. However, multilinguality increases the complexity of the indexing procedure because some "translation" is necessary to represent documents and queries in the same indexing space. To find a good translation of a term, the concept denoted by this term should be identified. Currently, pertinent information retrieval aims at representing the document semantics by concepts instead of terms. This thesis proposes a semantic indexing method for XML-encoded documents based on knowledge describing the document content, whatever the document language is. Our method indexes documents using two types of knowledge: - Domain knowledge, which composes a pivot language, used to represent documents and queries in the same indexing space not dependent on the languages. - Terminological knowledge, which is organized in several vocabularies (one per language). Terminological knowledge is related to domain knowledge and constitutes some presentation languages used to visualize the domain knowledge in several languages. In order to manipulate our indices, we define our own knowledge representation model entitled the semantic graphs. This model is an enrichment of the Sowa model of conceptual graphs by differentiating domain knowledge from terminological knowledge, but also by proposing a comparison function of graphs more adapted to the purpose information retrieval. Our indexing method can be used as well in a multilingual information retrieval system as in a knowledge hypertext. That's why we claim that our indexing method is generic. To validate our proposition, a prototype, called SyDoM, was developed, dedicated for the needs of virtual library
Roussey, Catherine Pinon Jean-Marie. "Une Méthode d'indexation sémantique adaptée aux corpus multilingues." Villeurbanne : Doc'INSA, 2005. http://docinsa.insa-lyon.fr/these/pont.php?id=roussey.
Full textOmhover, Jean-François. "Recherche d'images par similarité de contenus régionaux." Paris 6, 2004. http://www.theses.fr/2004PA066254.
Full textPessiot, Jean-François. "Apprentissage automatique pour l'extraction de caractéristiques : application au partitionnement de documents, au résumé automatique et au filtrage collaboratif." Paris 6, 2008. http://www.theses.fr/2008PA066218.
Full textJoly, Alexis. "Recherche par similarité statistique dans une grande base de signatures locales pour l'identification rapide d'extraits vidéo." La Rochelle, 2005. http://www.theses.fr/2005LAROS144.
Full textContent-based video indexing deals with techniques used to analyse and to exploit video databases without needs of any additional textual description. The work presented in this report is focused more precisely on content-based video copy detection, which is one of the emerging multimedia applications for which there is a need of a concerted effort from the database community and the computer vision community. To overcome the difficulties due to the use of very large databases, both in terms of robustness and speed, we propose a complete original and efficient strategy. The first part of this report presents the particular context of copy detection and the signatures used to describe the content of the videos. The originality of our method is that it is based both on local signatures and on a global similarity measure computed after the search in the signatures database. This similarity measure is not only a vote like other classical local approaches but it includes a registration step between candidate objects and objects retrieved by the search. The second part presents the main contribution of the thesis: A new indexing and retrieval technique belonging to the approximate similarity search techniques family. Recent works shows that trading quality for time can be widely profitable to speed-up descriptors similarity search. Whereas all other approximate techniques deal with K Nearest Neighbors search, the principle of our method is to extend the approximate paradigm to range queries. The main originality consists in determining relevant regions of the space according a theoritical model for the distortions undergone by the signatures. The method allows to determine the optimal region of the space with a high controlled probability to contain the good answer. This search paradigm is called statistical query. In practice, to simplify the access to signatures, the relevant regions are determined by using an Hilbert space filling curve and the space partition that induces. The experiments show that the technique is sublinear in database size with an assymptotically linear behavior (but only for huge databases) and that the quality performances are stable. Furthermore, they highlight that statistical queries provide a very high speed-up compared to classical exact range queries. The third part is focused on the global system assessment and the description of three applications. The experiments show that the simple theoretical distortion model is efficient enough to control the effective probability to retrieve a descriptor. They also point out that approximate similarity search is particularly profitable when using local signatures since the lost of some search results does not affect the global robustness of the detection. Furthermore, the detection results are almost invariant to strong database size growing (three orders of magnitude). The proposed approach was integrated in a difered real-time TV monitoring system which is able to control 40 000 hours of videos. The high quantity and variability of the results of this system open new data mining perspectives
Toma, Adrian. "Classes d'objets et recherche d'information." Paris 13, 2007. http://www.theses.fr/2007PA131008.
Full textThe aim of this thesis is the study of the improvement of the information retrieval process on the Internet, using a new linguistic theory of the semantico-syntactical classes. The main linguistic problems confronting the information retrieval are the polymorphism and the polysemy of natural languages. Same topic documents use different expressions, while there are several keyword variants for the same given topic. Due to search keys ambiguity, the intended sense of the request is not always reflected by the retrieved documents. The implementation in informatics of the linguistic concept of classes has to go through a first phase where a comparison is drawn between these notions in computer science and in linguistics. We also suggest the use of hierarchy of classes as well as of domains and of domains’hierarchies. The classes are structured in databases and the hierarchies are encoded in graphs. The importance of the document ranking draws our interest towards the use of the linguistic classes in automatic semantic content analysis in order to improve document’s estimated relevance. We show how automatic query reformulation and expansion based on semantico-syntactical classes improves the retrieval by focusing the process on the request’s inferred topic and by adding to the initial keywords other relevant concepts. The software we developed uses the hierarchy of the classes to detect the request’s topic; we then show how to use the dictionary based on linguistic classes to expand the query. We show examples on how cross-lingual query expansion benefits from the use of linguistic classes. Since the linguistic problems of the information retrieval are also rooted in the document indexing, we suggest and analyse a new indexing technique that has semantics at its core. The index entry is no longer the decontextualized and therefore meaningless word but a new concept of minimal indexing structure issued from the way predicates are described in the linguistic classes’theory
Lutfi, Rania. "Indexation intelligente et recherche par le contenu de l'audio." Nantes, 2003. http://www.theses.fr/2003NANT2028.
Full textKessler, Rémy. "Traitement automatique d'informations appliqué aux ressources humaines." Phd thesis, Université d'Avignon, 2009. http://tel.archives-ouvertes.fr/tel-00453642.
Full textDomingues, Catherine. "Etude d'outils informatiques et linguistiques pour l'aide à la recherche automatique d'information dans un corpus documentaire." Marne-la-Vallée, 2001. http://www.theses.fr/2001MARN0104.
Full text