Tesis sobre el tema "Traitement de langage naturel"
Crea una cita precisa en los estilos APA, MLA, Chicago, Harvard y otros
Consulte los 50 mejores tesis para su investigación sobre el tema "Traitement de langage naturel".
Junto a cada fuente en la lista de referencias hay un botón "Agregar a la bibliografía". Pulsa este botón, y generaremos automáticamente la referencia bibliográfica para la obra elegida en el estilo de cita que necesites: APA, MLA, Harvard, Vancouver, Chicago, etc.
También puede descargar el texto completo de la publicación académica en formato pdf y leer en línea su resumen siempre que esté disponible en los metadatos.
Explore tesis sobre una amplia variedad de disciplinas y organice su bibliografía correctamente.
Ameli, Samila. "Construction d'un langage de dictionnaire conceptuel en vue du traitement du langage naturel : application au langage médical". Compiègne, 1989. http://www.theses.fr/1989COMPD226.
Texto completoThis study deals with the realisation of a « new generation » information retrieval system, taking consideration of texts signification. This system compares texts (questions and documents) by their content. A knowledge base being indispensable for text “comprehension”, a dictionary of concepts has been designed in which are defined the concepts and their mutual relations thru a user friendly language called SUMIX. SUMIX enables us (1) to solve ambiguities due to polysemia by considering context dependencies, (2) to make use of property inheritance and so can largely help cogniticiens in the creation of the knowledge and inference base, (3) to define subject dependant relation between concepts which make possible metaknowledge handling. The dictionary of concepts is essentially used (1) to index concepts (and not characters string) which enables us to select a wide range of documents in the conceptual extraction phase, (2) to filter the previously selected documents by comparing the structure of each document with that of the query in the structural analysis phase
Gayral, Françoise. "Sémantique du langage naturel et profondeur variable : Une première approche". Paris 13, 1992. http://www.theses.fr/1992PA132004.
Texto completoBelabbas, Azeddine. "Satisfaction de contraintes et validation des grammaires du langage naturel". Paris 13, 1996. http://www.theses.fr/1996PA132044.
Texto completoNazarenko, Adeline. "Compréhension du langage naturel : le problème de la causalité". Paris 13, 1994. http://www.theses.fr/1994PA132007.
Texto completoCiortuz, Liviu-Virgil. "Programmation concurrente par contraintes et traitement du langage naturel : le système DF". Lille 1, 1996. http://www.theses.fr/1996LIL10145.
Texto completoNous avons implémenté un prototype du système df en oz, le langage concurrent multi-paradigme développé au dfki, en mettant en oeuvre une alternative typée et orientée-objet de son sous-système d'articles ouverts. Le système df est appliqué au traitement du langage naturel : analyse, génération et traduction automatique. Nous avons entrepris la conception d'un noyau hpsg pour le roumain, avec une implémentation concurrente. La définitude (définiteness), la topique (topic) et la modification (adjunction) du groupe nominal roumain sont analysées et la fonctionnalité des pronoms clitiques roumains sont expliquées à partir de la linéarisation (linearization) du groupe verbal transitif. Nous avons défini deux méta-schémas au dessus des schémas de la règle de dominance immédiate (id) dans la théorie de hpsg : le méta-schéma id 1, qui concerne les sujets multiples (par exemple les déterminants dans le groupe nominal roumain) et le méta-schéma id 2/6, pour la corrélation des composants localement ordonnés en dépendance non-bornée (comme les clitiques dans le groupe verbal transitif roumain). Les deux méta-schémas id font marcher la concurrence dans le cadre de la théorie des grammaires hpsg
Kupsc, Anna. "Une grammaire hpsg des clitiques polonais". Paris 7, 2000. http://www.theses.fr/2000PA070086.
Texto completoPopesco, Liana. "Analyse et génération de textes à partir d'un seul ensemble de connaissances pour chaque langue naturelle et de meta-règles de structuration". Paris 6, 1986. http://www.theses.fr/1986PA066138.
Texto completoMaire-Reppert, Daniele. "L'imparfait de l'indicatif en vue d'un traitement informatique du français". Paris 4, 1990. http://www.theses.fr/1990PA040039.
Texto completoMy approach to the French imperfect is in keeping with the methodology of expert systems. I have first identified and then given a topological representation of the values of the imperfect, i. E. : descriptive state, permanent state, new state, progressive process, habit, possibility, hypothetical, politeness and hypocoristic. We have then defined its constant in order to distinguish the imperfect from the other tenses. Finally, we have worked out a set of heuristic rules (about a hundred of production rules), the function of which is to associate a semantic value to a temporal morpheme according to the context. This contextual research has been led at the level of the text, of the sentence and the archetype of the verb. Our study of the imperfect has been completed by an example of insertion of the "heuristic rules" module in architecture of natural languages treatment and by short analysis of the contribution of such an approach in teaching foreign languages
Smart, John Ferguson. "L' analyse et la représentation de compte-rendus médicaux". Aix-Marseille 2, 1996. http://www.theses.fr/1996AIX22095.
Texto completoFouqueré, Christophe. "Systèmes d'analyse tolérante du langage naturel". Paris 13, 1988. http://www.theses.fr/1988PA132003.
Texto completoRAMMAL, MAHMOUD. "Une interface conceptuelle pour le traitement du langage naturel. Application au langage medical dans le systeme adm". Compiègne, 1993. http://www.theses.fr/1993COMP594S.
Texto completoLaskri, Mohamed Tayeb. "Approche de l'automatisation de thésaurus : étude de la sémantique adaptée du langage naturel". Aix-Marseille 2, 1987. http://www.theses.fr/1987AIX22076.
Texto completoFort, Karën. "Les ressources annotées, un enjeu pour l’analyse de contenu : vers une méthodologie de l’annotation manuelle de corpus". Paris 13, 2012. http://scbd-sto.univ-paris13.fr/intranet/edgalilee_th_2012_fort.pdf.
Texto completoManual corpus annotation has become a key issue for Natural Langage Processing (NLP), as manually annotated corpora are used both to create and to evaluate NLP tools. However, the process of manual annotation remains underdescribed and the tools used to support it are often misused. This situation prevents the campaign manager from evaluating and guarantying the quality of the annotation. We propose in this work a unified vision of manual corpus annotation for NLP. It results from our experience of annotation campaigns, either as a manager or as a participant, as well as from collaborations with other researchers. We first propose a global methodology for managing manual corpus annotation campaigns, that relies on two pillars: an organization for annotation campaigns that puts evaluation at the heart of the process and an innovative grid for the analysis of the complexity dimensions of an annotation campaign. A second part of our work concerns the tools of the campaign manager. We evaluated the precise influence of automatic pre-annotation on the quality and speed of the correction by humans, through a series of experiments on part-of-speech tagging for English. Furthermore, we propose practical solutions for the evaluation of manual annotations, that proche che vide the campaign manager with the means to select the most appropriate measures. Finally, we brought to light the processes and tools involved in an annotation campaign and we instantiated the methodology that we described
Alain, Pierre. "Contributions à l'évaluation des modèles de langage". Rennes 1, 2007. http://www.theses.fr/2007REN1S003.
Texto completoThis work deals with the evaluation of language models independently of any applicative task. A comparative study between several language models is generally related to the role that a model has into a complete system. Our objective consists in being independant of the applicative system, and thus to provide a true comparison of language models. Perplexity is a widely used criterion as to comparing language models without any task assumptions. However, the main drawback is that perplexity supposes probability distributions and hence cannot compare heterogeneous models. As an evaluation framework, we went back to the definition of the Shannon's game which is based on model prediction performance using rank based statistics. Our methodology is able to predict joint word sequences that are independent of the task or model assumptions. Experiments are carried out on French and English modeling with large vocabularies, and compare different kinds of language models
Sileo, Damien. "Représentations sémantiques et discursives pour la compréhension automatique du langage naturel". Thesis, Toulouse 3, 2019. http://www.theses.fr/2019TOU30201.
Texto completoComputational models for automatic text understanding have gained a lot of interest due to unusual performance gains over the last few years, some of them leading to super-human scores. This success reignited some grandeur claims about artificial intelligence, such as universal sentence representation. In this thesis, we question these claims through two complementary angles. Firstly, are neural networks and vector representations expressive enough to process text and perform a wide array of complex tasks? In this thesis, we will present currently used computational neural models and their training techniques. We propose a criterion for expressive compositions and show that a popular evaluation suite and sentence encoders (SentEval/InferSent) have an expressivity bottleneck; minor changes can yield new compositions that are expressive and insightful, but might not be sufficient, which may justify the paradigm shift towards newer Transformers-based models. Secondly, we will discuss the question of universality in sentence representation: what actually lies behind these universality claims? We delineate a few theories of meaning, and in a subsequent part of this thesis, we argue that semantics (unsituated, literal content) as opposed to pragmatics (meaning as use) is preponderant in the current training and evaluation data of natural language understanding models. To alleviate that problem, we show that discourse marker prediction (classification of hidden discourse markers between sentences) can be seen as a pragmatics-centered training signal for text understanding. We build a new discourse marker prediction dataset that yields significantly better results than previous work. In addition, we propose a new discourse-based evaluation suite that could incentivize researchers to take into account pragmatic considerations when evaluating text understanding models
Dégremont, Jean-François. "Ethnométhodologie et innovation technologique : le cas du traitement automatique des langues naturelles". Paris 7, 1989. http://www.theses.fr/1989PA070043.
Texto completoThe thesis begins with a short historical reminder of ethnomethodology, considered as a scientific field, since the whole beginners during the 30's until the 1967 explosion in US and Europe. The first part is an explication of the main concepts of ethnomethodology. They are developped from the pariseptist school theoretical point of view, which tries to associate the strongest refuse of inductions and the indifference principle, mainly when natural languages, considered as well as studies objects and communication tools, are used. The second part of the thesis is devoted to the concrete application of these theoretical concepts in the field of technological strategies which have been elaborated in France in the area of natural language processing. Three studies successively describe the ethnomethods and rational properties of practical activities which are used in an administrative team, the elaboration of a technology policy and indexical descriptions of the language industry field. The conclusion tries to show how the concepts and methods developped by ethnomethodology can increase, in this field, the efficacy of strategical analysis and the quality of research and development programs
PARK, SE YOUNG. "Un algorithme efficace pour l'analyse du langage naturel : application aux traitements des erreurs et aux grammaires discontinues". Paris 7, 1989. http://www.theses.fr/1989PA077214.
Texto completoMazahreh, Mazhar. "Recherche et analyse informatique des expressions du langage naturel correspondant à des questions sur les bases de données". Paris, EHESS, 1990. http://www.theses.fr/1990EHES0059.
Texto completoHarrathi, Farah. "Extraction de concepts et de relations entre concepts à partir des documents multilingues : approche statistique et ontologique". Lyon, INSA, 2009. http://theses.insa-lyon.fr/publication/2009ISAL0073/these.pdf.
Texto completoThe research work of this thesis is related to the problem of document search indexing and more specifically in that of the extraction of semantic descriptors for document indexing. Information Retrieval System (IRS) is a set of models and systems for selecting a set of documents satisfying user needs in terms of information expressed as a query. In IR, a query is composed mainly of two processes for representation and retrieval. The process of representation is called indexing, it allows to represent documents and query descriptors, or indexes. These descriptors reflect the contents of documents. The retrieval process consists on the comparison between documents representations and query representation. In the classical IRS, the descriptors used are words (simple or compound). These IRS consider the document as a set of words, often called a "bag of words". In these systems, the words are considered as graphs without semantics. The only information used for these words is their occurrence frequency in the documents. These systems do not take into account the semantic relationships between words. For example, it is impossible to find documents represented by a word synonymous with M1 word M2, where the request is represented by M2. Also, in a classic IRS document indexed by the term "bus" will never be found by a query indexed by the word "taxi", yet these are two words that deal with the same subject "means of transportation. " To address these limitations, several studies were interested taking into account of the semantic indexing terms. This type of indexing is called semantic or conceptual indexing. These works take into account the notion of concept in place of notion of word. In this work the terms denoting concepts are extracted from the document by using statistical techniques. These terms are then projected onto resource of semantics such as: ontology, thesaurus and so on to extract the concepts involved
Fourour, Nordine. "Identification et catégorisation automatique des entités nommées dans les textes français". Nantes, 2004. http://www.theses.fr/2004NANT2126.
Texto completoNamed Entity (NE) Recognition is a recurring problem in the different domain of Natural Language Processing. As a result of, a linguistic investigation allowing to set-up operational parameters defining the concept of named entity, a state of art of the domain, and a corpus investigation using referential and graphical criteria, we present Nemesis - a French named entity recognizer. This system analyzes the internal and external evidences by using grammar rules and trigger word lexicons, and includes a learning process. With these processes, Nemesis performance achieves about 90% of precision and 80% of recall. To increase the recall, we put forward optional modules (analysis of the wide context and utilization of the Web as a source of new contexts) and investigate in setting up a disambiguation and grammar rules inference module
Tartier, Annie. "Analyse automatique de l'évolution terminologique : variations et distances". Nantes, 2004. http://www.theses.fr/2004NANT2040.
Texto completoThe aim of this thesis is to work out automatic methods for uncovering any evolutionary phenomena within terms extracted from diachronic corpora of scientific or technical texts. The first research axis concerns the nature of changes. It is based on a terminological variation typology aiming to define a distance between two terminological forms. That distance allows us to easily put together the variants of a term and to define measures from sets of studied terms. The second axis concerns time structuration and proposes several diachronic examination modes in order to distinguish ephemeral changes from durable ones which could be the signs of an evolution. These ideas are implemented in a prototype which first proposes temporal profiles, then some information about stable, old or new terms, information given for exact forms or to the nearest variant
Balicco, Laurence. "Génération de repliques en français dans une interface homme-machine en langue naturelle". Grenoble 2, 1993. http://www.theses.fr/1993GRE21025.
Texto completoThis research takes place in the context of natural language generation. This field has benn neglected for a long time because it seemed a much easier phase that those of analysis. The thesis corresponds to a first work on generation in the criss team and places the problem of generation in the context of a manmachine dialogue in natural language. Some of its consequences are : generation from a logical content to be translated into natural language, this translation of the original content kept as close as possible,. . . After the study of the different works that have been done, we decided to create our own generation system, resusing when it is possible, the tools elaborated during the analyzing process. This generation process is based on a linguistic model, which uses syntactic and morphologic information and in which linguistic transformations called operations are defined (coodination, anaphorisation, thematisation,. . . ). These operations can be given by the dialogue or calulated during the generation process. The model allows the creation of several of the same utterance and therefore a best adaptation for different users. This thesis presents the studied works, essentially on the french and the english languages, the linguistic model developped, the computing model used, and a brief presentation of an european project which offers a possible application of ou
Alsandouk, Fatima. "Grammaire de scene : processus de comprehension de textes de description geometrique". Toulouse 2, 1990. http://www.theses.fr/1990TOU20058.
Texto completoBoulaknadel, Siham. "Traitement automatique des langues et recherche d'information en langue arabe dans un domaine de spécialité : apport des connaissanaces morphologiques et syntaxiques pour l'indexation". Nantes, 2008. http://www.theses.fr/2008NANT2052.
Texto completoInformation retrieval aims to provide to an user an easy access to information. To achieve this goal, an information retrieval system (IRS) must represent, store and organize information, then provide to the user the elements corresponding to the need for information expressed by his query. Most of information retrieval systems (IRS) use simple terms to index and retrieve documents. However, this representation is not precise enough to represent the contents of documents and queries, because of the ambiguity of terms isolated from their context. A solution to this problem is to use multi-word terms to replace simple term. This approach is based on the assumption that a multi-word term is less ambiguous than a simple term. Our thesis is part of the information retrieval in Arabic specific domain. The objective of our work was on the one hand, identifying a multi-word terms present in queries and documents. On the other hand, exploiting the richness of language by combining several linguistic knowledge belonging at the morphological and syntax level, and showing how the contribution of syntactic and morphological knowledge helps to improve access to information. Thus, we proposed a platform integrating various components in the public domain; it leads to show significant contribution of these components. In addition, we have defined linguistically a multi-word term in Arabic and we developed a system of identification of multi-word terms which is based on a mixed approach combining statistical model and linguistic data
Ouersighni, Riadh Hassoun Mohamed Dichy Joseph. "La conception et la réalisation d'un système d'analyse morpho-syntaxique robuste pour l'arabe utilisation pour la détection et le diagnostic des fautes d'accord /". Lyon : Université Lumière Lyon 2, 2002. http://demeter.univ-lyon2.fr:8080/sdx/theses/lyon2/2002/ouersighni_r.
Texto completoCaucheteux, Charlotte. "Language representations in deep learning algorithms and the brain". Electronic Thesis or Diss., université Paris-Saclay, 2023. http://www.theses.fr/2023UPASG031.
Texto completoRecent deep language models -- like GPT-3 and ChatGPT -- are capable to produce text that closely resembles that of humans. Such similarity raises questions about how the brain and deep models process language, the mechanisms they use, and the internal representations they construct. In this thesis, I compare the internal representations of the brain and deep language models, with the goal of identifying their similarities and differences. To this aim, I analyze functional resonance imaging (fMRI) and magnetoencephalography (MEG) recordings of participants listening to and reading sentences, and compare them to the activations of thousands of language algorithms corresponding to these same sentences.Our results first highlight high-level similarities between the internal representations of the brain and deep language models. We find that deep nets' activations significantly predict brain activity across subjects for different cohorts (>500 participants), recording modalities (MEG and fMRI), stimulus types (isolated words, sentences, and natural stories), stimulus modalities (auditory and visual presentation), languages (Dutch, English and French), and deep language models. This alignment is maximal in brain regions repeatedly associated with language, for the best-performing algorithms and for participants who best understand the stories. Critically, we evidence a similar processing hierarchy between the two systems. The first layers of the algorithms align with low-level processing regions in the brain, such as auditory areas and the temporal lobe, while the deep layers align with regions associated with higher-level processing, such fronto-parietal areas.We then show how such similarities can be leveraged to build better predictive models of brain activity and better decompose several linguistic processes in the brain, such as syntax and semantics. Finally, we explore the differences between deep language models and the brain's activations. We find that the brain predicts distant and hierarchical representations, unlike current language models that are mostly trained to make short-term and word-level predictions. Overall, modern algorithms are still far from processing language in the same way that humans do. However, the direct links between their inner workings and that of the brain provide an promising platform for better understanding both systems, and pave the way for building better algorithms inspired by the human brain
Krit, Hatem. "Locadelane : un langage objet d'aide à la compréhension automatique du discours exprimé en langage naturel et écri". Toulouse 3, 1990. http://www.theses.fr/1990TOU30008.
Texto completoDenand, Nicolas. "Traitement automatique de phrases locatives statiques du français". Aix-Marseille 2, 2004. http://www.theses.fr/2004AIX22035.
Texto completoTannier, Xavier. "Extraction et recherche d'information en langage naturel dans les documents semi-structurés". Phd thesis, Ecole Nationale Supérieure des Mines de Saint-Etienne, 2006. http://tel.archives-ouvertes.fr/tel-00121721.
Texto completo(écrits en XML en pratique) combine des aspects de la RI
traditionnelle et ceux de l'interrogation de bases de données. La
structure a une importance primordiale, mais le besoin d'information
reste vague. L'unité de recherche est variable (un paragraphe, une
figure, un article complet\dots). Par ailleurs, la flexibilité du
langage XML autorise des manipulations du contenu qui provoquent
parfois des ruptures arbitraires dans le flot naturel du texte.
Les problèmes posés par ces caractéristiques sont nombreux, que ce
soit au niveau du pré-traitement des documents ou de leur
interrogation. Face à ces problèmes, nous avons étudié les solutions
spécifiques que pouvait apporter le traitement automatique de la
langue (TAL). Nous avons ainsi proposé un cadre théorique et une
approche pratique pour permettre l'utilisation des techniques
d'analyse textuelle en faisant abstraction de la structure. Nous avons
également conçu une interface d'interrogation en langage naturel pour
la RI dans les documents XML, et proposé des méthodes tirant profit de
la structure pour améliorer la recherche des éléments pertinents.
Véronis, Jean. "Contribution à l'étude de l'erreur dans le dialogue homme-machine en langage naturel". Aix-Marseille 3, 1988. http://www.theses.fr/1988AIX30043.
Texto completoPerraud, Freddy. "Modélisation du langage naturel appliquée à la reconnaissance de l'écriture manuscrite en-ligne". Nantes, 2005. http://www.theses.fr/2005NANT2112.
Texto completoBelec, Yves. "Des règles expertes pour une méthode applicative d'analyse ou de génération du langage naturel". Toulouse 3, 1990. http://www.theses.fr/1990TOU30136.
Texto completoMela, Augusta. "Traitement automatique de la coordination par et". Paris 13, 1992. http://www.theses.fr/1992PA132040.
Texto completoArias, Aguilar José Anibal. "Méthodes spectrales pour le traitement automatique de documents audio". Toulouse 3, 2008. http://thesesups.ups-tlse.fr/436/.
Texto completoThe disfluencies are a frequently occurring phenomenon in any spontaneous speech production; it consists of the interruption of the normal flow of speech. They have given rise to numerous studies in Natural Language Processing. Indeed, their study and precise identification are essential, both from a theoretical and applicative perspective. However, most of the researches about the subject relate to everyday uses of language: "small talk" dialogs, requests for schedule, speeches, etc. But what about spontaneous speech production made in a restrained framework? To our knowledge, no study has ever been carried out in this context. However, we know that using a "language specialty" in the framework of a given task leads to specific behaviours. Our thesis work is devoted to the linguistic and computational study of disfluencies within such a framework. These dialogs concern air traffic control, which entails both pragmatic and linguistic constraints. We carry out an exhaustive study of disfluencies phenomena in this context. At first we conduct a subtle analysis of these phenomena. Then we model them to a level of abstraction, which allows us to obtain the patterns corresponding to the different configurations observed. Finally we propose a methodology for automatic processing. It consists of several algorithms to identify the different phenomena, even in the absence of explicit markers. It is integrated into a system of automatic processing of speech. Eventually, the methodology is validated on a corpus of 400 sentences
EL, HAROUCHY ZAHRA. "Dictionnaire et grammaire pour le traitement automatique des ambiguites morphologiques des mots simples en francais". Besançon, 1997. http://www.theses.fr/1997BESA1010.
Texto completoWhen carrying out the automatic analysis of a text, one of the first stages consists in determining the grammatical categories of the words. In order to do this, a dictionary has been designed which recognises the one or several grammatical categories of non-compound words from their endings. This dictionary, which we have called automatic dictionary, is a collection of general rules (which can consist of sub- rules). A general rule sets forth an ending. An operator (the one or several grammatical categories) is associated with each rule. For example, we have the following general rule: +words ending in 'able' are adjectives;. Examples of exceptions to (or sub-rules) of this general rule are nouns such as (+cartable ;), conjugated verbs like (+ accable ;), and morphological ambiguities such as + noun and conjugated verb (like +sable;, +table. . . ;), and ambiguities such as + adjectival nouns ;(like, for example, + comptable ;. . . ) consequently, this sort of dictionary gives prominence to those words posessing several grammatical categories. When the automatic dictionary detects a word posessing several categories, the grammar system is consulted,of which the role is to pick out the morphological ambiguities by studying the immediate context. The rules in the grammar system work like a group of possible combinations of elements capable of going after and-or before the ambiguous form ( for example, a rule states that an ambiguous form such as + pronoun or article ; preceded by + a cause de ; is, in fact, an article)
Culioli-Atwood, Marie-Hélène. "Operations referentielles. Analyse de la determination en francais en vue d'un traitement informatise". Paris 7, 1992. http://www.theses.fr/1992PA070014.
Texto completoThe purpose of the thesis is (1) to gather a maximun of systematic and detailed observations concerning the occurence of determiners in french ( in the pattern det. + n ); (2) to build a system of metalinguistic representation enabling the modelling of facts; (3) to build procedures of reasoning having in mind an algorithmic treatment whether in generation or in analysis. The work gives the conceptual basis for modelling both on a formal and a semantic level. The thesis is made up of three parts: analysis of the problems in relation to the paraphrastic manipulations; study of groups of nominalised predicates based on semantic classifications; study of determiners in prepositional phrases. This work of research builds the preliminary steps of any computerized treatment of determination as used in a french text
Oh, Hyun-Gum. "Représentation des valeurs sémantiques du passé composé français en vue d'un traitement informatique". Paris 4, 1991. http://www.theses.fr/1991PA040070.
Texto completoWe present a model solving the problem of tense and aspect in the French language, detail research of "passé compose". There are three parties in this thesis: first, generality and theoric concept second ; values of "passé compose" in French; third: strategie of contextual exploration. Its shows that a natural language processing the aim of which is to build semantic representations of tenses is possible thanks to only linguistic data without using any other knowledge of the world. This linguistic approach has been implemented by a generator of expert system called "snark"
Haddad, Afifa. "Traitement des nominalisations anaphoriques en indexation automatique". Lyon 2, 2001. http://theses.univ-lyon2.fr/documents/lyon2/2001/haddad_a.
Texto completoThis thesis proposes en indexation method for integral texts based on anaphoric noun phrases. The motivation is to take advantage from the wide context of an anaphora relation in order to build a rich descriptor ? and to get consequently a performant index. The main contribution here is the design of a complete method enabling the systematic reconstitution of all arguments of each anaphoric nominalization encountered in the text. A completely resolved noun phrase constitutes a rich descriptor that is then added to the index. The resolution a nominal anaphora makes use the results of other preliminarly activities. These consists in collecting the syntactic structures of the possible noun phrase corresponding to a nominalization and, identifying a set of the anaphoric noun phrase and the form of its precedent. The feasibility of the proposed has been demonstrated through an application to a real-life corpus
Haddad, Afifa Le Guern Michel. "Traitement des nominalisations anaphoriques en indexation automatique". [S.l.] : [s.n.], 2001. http://theses.univ-lyon2.fr/sdx/theses/lyon2/intranet/haddad_a.
Texto completoSilva-Julia, Rita da. "Un système hybride pour le traitement du langage naturel et pour la récupération de l'information". Toulouse 3, 1995. http://www.theses.fr/1995TOU30255.
Texto completoRajman, Martin. "Apports d'une approche a base de corpus aux techniques de traitement automatique du langage naturel". Paris, ENST, 1995. http://www.theses.fr/1995ENST0043.
Texto completoRajman, Martin. "Apports d'une approche à base de corpus aux techniques de traitement automatique du langage naturel /". Paris : École nationale supérieure des télécommunications, 1997. http://catalogue.bnf.fr/ark:/12148/cb36707722b.
Texto completoLevrat, Bernard. "Le problème du sens dans les sytèmes de traitement du langage naturel : Une approche alternative au travers de la paraphrase". Paris 13, 1993. http://www.theses.fr/1993PA132023.
Texto completoN'Guéma, Sylvain Abraham. "Intégration de paramètres formels d'intonation à l'analyse syntaxique automatique dans une perspective d'aide à la désambigui͏̈sation syntaxique". Avignon, 1998. http://www.theses.fr/1998AVIG0121.
Texto completoPoibeau, Thierry. "Extraction d'information à base de connaissances hybrides". Paris 13, 2002. http://www.theses.fr/2002PA132001.
Texto completoAmoia, Marilisa Gardent Claire Pinkal Manfred. "Reconnaissance d'implications textuelles à forte composante linguistique". S. l. : S. l. : Nancy 1 ; Universität des Saarlandes, 2008. http://www.scd.uhp-nancy.fr/docnum/SCD_T_2008_0099_AMOIA.pdf.
Texto completoKoroleva, Anna. "Assisted authoring for avoiding inadequate claims in scientific reporting". Thesis, université Paris-Saclay, 2020. http://www.theses.fr/2020UPASS021.
Texto completoIn this thesis, we report on our work on developing Natural Language Processing (NLP) algorithms to aid readers and authors of scientific (biomedical) articles in detecting spin (distorted presentation of research results). Our algorithm focuses on spin in abstracts of articles reporting Randomized Controlled Trials (RCTs). We studied the phenomenon of spin from the linguistic point of view to create a description of its textual features. We annotated a set of corpora for the key tasks of our spin detection pipeline: extraction of declared (primary) and reported outcomes, assessment of semantic similarity of pairs of trial outcomes, and extraction of relations between reported outcomes and their statistical significance levels. Besides, we anno-tated two smaller corpora for identification of statements of similarity of treatments and of within-group comparisons. We developed and tested a number of rule-based and machine learning algorithmsforthe key tasksof spindetection(outcome extraction,outcome similarity assessment, and outcome-significance relation extraction). The best performance was shown by a deep learning approach that consists in fine-tuning deep pre-trained domain-specific language representations(BioBERT and SciBERT models) for our downstream tasks. This approach was implemented in our spin detection prototype system, called De-Spin, released as open source code. Our prototype includes some other important algorithms, such as text structure analysis (identification of the abstract of an article, identification of sections within the abstract), detection of statements of similarity of treatments and of within-group comparisons, extraction of data from trial registries. Identification of abstract sections is performed with a deep learning approach using the fine-tuned BioBERT model, while other tasks are performed using a rule-based approach. Our prototype system includes a simple annotation and visualization interface
Nelakanti, Anil Kumar. "Modélisation du langage à l'aide de pénalités structurées". Phd thesis, Université Pierre et Marie Curie - Paris VI, 2014. http://tel.archives-ouvertes.fr/tel-01001634.
Texto completoOuldja, Hadj. "Réalisation d'une interface en langage naturel et son application à l'enseignement assisté par ordinateur". Paris 6, 1988. http://www.theses.fr/1988PA066456.
Texto completoFernández, Sabido Silvia Fidelina Berche Bertrand Torres Moreno Juan-Manuel. "Applications exploratoires des modèles de spins au traitement automatique de la langue". S. l. : S. n, 2009. http://www.scd.uhp-nancy.fr/docnum/SCD_T_2009_0055_FERNANDEZ-SABIDO.pdf.
Texto completo