Academic literature on the topic 'Traitement du Langage Naturel (NLP)'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Traitement du Langage Naturel (NLP).'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Journal articles on the topic "Traitement du Langage Naturel (NLP)"
de Bonis, M., JD Guelfi, and M. Somogyi. "Observations psychiatriques en langage naturel, statistique textuelle et classification des dépressions." Psychiatry and Psychobiology 5, no. 1 (1990): 1–11. http://dx.doi.org/10.1017/s0767399x0000331x.
Full textSabatier, Paul. "Interfaces en langage naturel : du traitement du non attendu à la composition de phrases assistée." Annales Des Télécommunications 44, no. 1-2 (January 1989): 77–84. http://dx.doi.org/10.1007/bf02999879.
Full textAbeillé, Anne. "Review of Miller & Torris (1990): Formalismes syntaxiques pour le traitement automatique du langage naturel." Lingvisticæ Investigationes. International Journal of Linguistics and Language Resources 15, no. 2 (January 1, 1991): 429–32. http://dx.doi.org/10.1075/li.15.2.13abe.
Full textFouqueré, Christophe, and Fabrice Issac. "Corpus issus du Web : constitution et analyse informationnelle." Revue québécoise de linguistique 32, no. 1 (February 20, 2006): 111–34. http://dx.doi.org/10.7202/012246ar.
Full textGrosjean, François, and Alain Matthey. "L'apport potentiel de l'intelligence artificielle et du traitement automatique du langage naturel à une nouvelle version d'Hector." Travaux neuchâtelois de linguistique, no. 18 (September 1, 1992): 51–66. http://dx.doi.org/10.26034/tranel.1992.2466.
Full textKempf, E., S. Priou, B. Dura, J. Calderaro, C. Brones, P. Wasjbürt, L. Bennani, and X. Tannier. "Structuration des critères histopronostiques tumoraux par traitement automatique du langage naturel - Une comparaison entre apprentissage machine et règles." Journal of Epidemiology and Population Health 72 (March 2024): 202301. http://dx.doi.org/10.1016/j.jeph.2024.202301.
Full textLangevin, Christian. "Les technologies de l’intelligence artificielle au service des médias et des éditeurs de contenus : Traitement du langage naturel (TAL)." I2D - Information, données & documents 1, no. 1 (July 11, 2022): 30–37. http://dx.doi.org/10.3917/i2d.221.0030.
Full textDündar, Oğuz İbrahim. "Utilisation Potentielle De Chatgpt Dans L'apprentissage Des Langues Etrangères : Exploration Des Possibilités Selon Les Niveaux Langagiers Du CECRL." Kahramanmaraş Sütçü İmam Üniversitesi Sosyal Bilimler Dergisi 21, no. 1 (April 30, 2024): 63–75. http://dx.doi.org/10.33437/ksusbd.1384040.
Full textZeghari, Z., N. Bennani Mechita, FZ Benfouila, T. Benamar, M. Merabet, M. Youbi, J. Belayachi, and R. Abouqal. "P35 - Quelles sont les observations des professionnels de santé marocains au sujet de la crise sanitaire du COVID-19 ? Traitement du langage naturel de la question ouverte d'une enquête à grande échelle." Journal of Epidemiology and Population Health 72 (May 2024): 202475. http://dx.doi.org/10.1016/j.jeph.2024.202475.
Full textDalbin, Sylvie. "Compréhension des langues et interaction . Sous la direction de Gérard Sabah, Hermès Science Publications : Lavoisier, 2006. – 400 p. – (Traité IC2, série Cognition et traitement de l’information). – ISBN 2-7462-1256-0 : 120 €. Sémantique et traitement automatique du langage naturel . Sous la direction de Patrice Enjalbert Paris : Hermès Science Publications : Lavoisier, 2005. – 410 p. – (Traité IC2, série Cognition et traitement de l’information). – ISBN 2-7462-1126-2 : 120 €." Documentaliste-Sciences de l'Information Vol. 44, no. 1 (February 28, 2007): II. http://dx.doi.org/10.3917/docsi.441.0096b.
Full textDissertations / Theses on the topic "Traitement du Langage Naturel (NLP)"
Moncla, Ludovic. "Automatic Reconstruction of Itineraries from Descriptive Texts." Thesis, Pau, 2015. http://www.theses.fr/2015PAUU3029/document.
Full textThis PhD thesis is part of the research project PERDIDO, which aims at extracting and retrieving displacements from textual documents. This work was conducted in collaboration with the LIUPPA laboratory of the university of Pau (France), the IAAA team of the university of Zaragoza (Spain) and the COGIT laboratory of IGN (France). The objective of this PhD is to propose a method for establishing a processing chain to support the geoparsing and geocoding of text documents describing events strongly linked with space. We propose an approach for the automatic geocoding of itineraries described in natural language. Our proposal is divided into two main tasks. The first task aims at identifying and extracting information describing the itinerary in texts such as spatial named entities and expressions of displacement or perception. The second task deal with the reconstruction of the itinerary. Our proposal combines local information extracted using natural language processing and physical features extracted from external geographical sources such as gazetteers or datasets providing digital elevation models. The geoparsing part is a Natural Language Processing approach which combines the use of part of speech and syntactico-semantic combined patterns (cascade of transducers) for the annotation of spatial named entities and expressions of displacement or perception. The main contribution in the first task of our approach is the toponym disambiguation which represents an important issue in Geographical Information Retrieval (GIR). We propose an unsupervised geocoding algorithm that takes profit of clustering techniques to provide a solution for disambiguating the toponyms found in gazetteers, and at the same time estimating the spatial footprint of those other fine-grain toponyms not found in gazetteers. We propose a generic graph-based model for the automatic reconstruction of itineraries from texts, where each vertex represents a location and each edge represents a path between locations. %, combining information extracted from texts and information extracted from geographical databases. Our model is original in that in addition to taking into account the classic elements (paths and waypoints), it allows to represent the other elements describing an itinerary, such as features seen or mentioned as landmarks. To build automatically this graph-based representation of the itinerary, our approach computes an informed spanning tree on a weighted graph. Each edge of the initial graph is weighted using a multi-criteria analysis approach combining qualitative and quantitative criteria. Criteria are based on information extracted from the text and information extracted from geographical sources. For instance, we compare information given in the text such as spatial relations describing orientation (e.g., going south) with the geographical coordinates of locations found in gazetteers. Finally, according to the definition of an itinerary and the information used in natural language to describe itineraries, we propose a markup langugage for encoding spatial and motion information based on the Text Encoding and Interchange guidelines (TEI) which defines a standard for the representation of texts in digital form. Additionally, the rationale of the proposed approach has been verified with a set of experiments on a corpus of multilingual hiking descriptions (French, Spanish and Italian)
Lauly, Stanislas. "Exploration des réseaux de neurones à base d'autoencodeur dans le cadre de la modélisation des données textuelles." Thèse, Université de Sherbrooke, 2016. http://hdl.handle.net/11143/9461.
Full textBourgeade, Tom. "Interprétabilité a priori et explicabilité a posteriori dans le traitement automatique des langues." Thesis, Toulouse 3, 2022. http://www.theses.fr/2022TOU30063.
Full textWith the advent of Transformer architectures in Natural Language Processing a few years ago, we have observed unprecedented progress in various text classification or generation tasks. However, the explosion in the number of parameters, and the complexity of these state-of-the-art blackbox models, is making ever more apparent the now urgent need for transparency in machine learning approaches. The ability to explain, interpret, and understand algorithmic decisions will become paramount as computer models start becoming more and more present in our everyday lives. Using eXplainable AI (XAI) methods, we can for example diagnose dataset biases, spurious correlations which can ultimately taint the training process of models, leading them to learn undesirable shortcuts, which could lead to unfair, incomprehensible, or even risky algorithmic decisions. These failure modes of AI, may ultimately erode the trust humans may have otherwise placed in beneficial applications. In this work, we more specifically explore two major aspects of XAI, in the context of Natural Language Processing tasks and models: in the first part, we approach the subject of intrinsic interpretability, which encompasses all methods which are inherently easy to produce explanations for. In particular, we focus on word embedding representations, which are an essential component of practically all NLP architectures, allowing these mathematical models to process human language in a more semantically-rich way. Unfortunately, many of the models which generate these representations, produce them in a way which is not interpretable by humans. To address this problem, we experiment with the construction and usage of Interpretable Word Embedding models, which attempt to correct this issue, by using constraints which enforce interpretability on these representations. We then make use of these, in a simple but effective novel setup, to attempt to detect lexical correlations, spurious or otherwise, in some popular NLP datasets. In the second part, we explore post-hoc explainability methods, which can target already trained models, and attempt to extract various forms of explanations of their decisions. These can range from diagnosing which parts of an input were the most relevant to a particular decision, to generating adversarial examples, which are carefully crafted to help reveal weaknesses in a model. We explore a novel type of approach, in parts allowed by the highly-performant but opaque recent Transformer architectures: instead of using a separate method to produce explanations of a model's decisions, we design and fine-tune an architecture which jointly learns to both perform its task, while also producing free-form Natural Language Explanations of its own outputs. We evaluate our approach on a large-scale dataset annotated with human explanations, and qualitatively judge some of our approach's machine-generated explanations
Michalon, Olivier. "Modèles statistiques pour la prédiction de cadres sémantiques." Thesis, Aix-Marseille, 2017. http://www.theses.fr/2017AIXM0221/document.
Full textIn natural language processing, each analysis step has improved the way in which language can be modeled by machines. Another step of analysis still poorly mastered resides in semantic parsing. This type of analysis can provide information which would allow for many advances, such as better human-machine interactions or more reliable translations. There exist several types of meaning representation structures, such as PropBank, AMR and FrameNet. FrameNet corresponds to the frame semantic framework whose theory has been described by Charles Fillmore (1971). In this theory, each prototypical situation and each different elements involved are represented in such a way that two similar situations are represented by the same object, called a semantic frame. The work that we will describe here follows the work already developed for machine prediction of frame semantic representations. We will present four prediction systems, and each one of them allowed to validate another hypothesis on the necessary properties for effective prediction. We will show that semantic parsing can also be improved by providing prediction models with refined information as input of the system, with firstly a syntactic analysis where deep links are made explicit and secondly vectorial representations of the vocabulary learned beforehand
Cousot, Kévin. "Inférences et explications dans les réseaux lexico-sémantiques." Thesis, Montpellier, 2019. http://www.theses.fr/2019MONTS108.
Full textThanks to the democratization of new communication technologies, there is a growing quantity of textual resources, making Automatic Natural Language Processing (NLP) a discipline of crucial importance both scientifically and industrially. Easily available, these data offer unprecedented opportunities and, from opinion analysis to information research and semantic text analysis, there are many applications.However, this textual data cannot be easily exploited in its raw state and, in order to carry out such tasks, it seems essential to have resources describing semantic knowledge, particularly in the form of lexico-semantic networks such as that of the JeuxDeMots project. However, the constitution and maintenance of such resources remain difficult operations, due to their large size but also because of problems of polysemy and semantic identification. Moreover, their use can be tricky because a significant part of the necessary information is not directly accessible in the resource but must be inferred from the data of the lexico-semantic network.Our work seeks to demonstrate that lexico-semantic networks are, by their connexionic nature, much more than a collection of raw facts and that more complex structures such as interpretation paths contain more information and allow multiple inference operations to be performed. In particular, we will show how to use a knowledge base to provide explanations to high-level facts. These explanations allow at least to validate and memorize new information.In doing so, we can assess the coverage and relevance of the database data and consolidate it. Similarly, the search for paths is useful for classification and disambiguation problems, as they are justifications for the calculated results.In the context of the recognition of named entities, they also make it possible to type entities and disambiguate them (is the occurrence of the term Paris a reference to the city, and which one, or to a starlet?) by highlighting the density of connections between ambiguous entities, their context and their possible type.Finally, we propose to turn the large size of the JeuxDeMots network to our advantage to enrich the database with new facts from a large number of comparable examples and by an abduction process on the types of semantic relationships that can connect two given terms. Each inference is accompanied by explanations that can be validated or invalidated, thus providing a learning process
Manishina, Elena. "Data-driven natural language generation using statistical machine translation and discriminative learning." Thesis, Avignon, 2016. http://www.theses.fr/2016AVIG0209/document.
Full textThe humanity has long been passionate about creating intellectual machines that can freely communicate with us in our language. Most modern systems communicating directly with the user share one common feature: they have a dialog system (DS) at their base. As of today almost all DS components embraced statistical methods and widely use them as their core models. Until recently Natural Language Generation (NLG) component of a dialog system used primarily hand-coded generation templates, which represented model phrases in a natural language mapped to a particular semantic content. Today data-driven models are making their way into the NLG domain. In this thesis, we follow along this new line of research and present several novel data-driven approaches to natural language generation. In our work we focus on two important aspects of NLG systems development: building an efficient generator and diversifying its output. Two key ideas that we defend here are the following: first, the task of NLG can be regarded as the translation between a natural language and a formal meaning representation, and therefore, can be performed using statistical machine translation techniques, and second, corpus extension and diversification which traditionally involved manual paraphrasing and rule crafting can be performed automatically using well-known and widely used synonym and paraphrase extraction methods. Concerning our first idea, we investigate the possibility of using NGRAM translation framework and explore the potential of discriminative learning, notably Conditional Random Fields (CRF) models, as applied to NLG; we build a generation pipeline which allows for inclusion and combination of different generation models (NGRAM and CRF) and which uses an efficient decoding framework (finite-state transducers' best path search). Regarding the second objective, namely corpus extension, we propose to enlarge the system's vocabulary and the set of available syntactic structures via integrating automatically obtained synonyms and paraphrases into the training corpus. To our knowledge, there have been no attempts to increase the size of the system vocabulary by incorporating synonyms. To date most studies on corpus extension focused on paraphrasing and resorted to crowd-sourcing in order to obtain paraphrases, which then required additional manual validation often performed by system developers. We prove that automatic corpus extension by means of paraphrase extraction and validation is just as effective as crowd-sourcing, being at the same time less costly in terms of development time and resources. During intermediate experiments our generation models showed a significantly better performance than the phrase-based baseline model and appeared to be more robust in handling unknown combinations of concepts than the current in-house rule-based generator. The final human evaluation confirmed that our data-driven NLG models is a viable alternative to rule-based generators
Annouz, Hamid. "Traitement morphologique des unités linguistiques du kabyle à l’aide de logiciel NooJ : Construction d’une base de données." Thesis, Paris, INALCO, 2019. http://www.theses.fr/2019INAL0022.
Full textThis work introduces the Kabyle language to the field of Natural Language Processing by giving it a database for the NooJ software that allows the automatic recognition of linguistic units in a written corpus.We have divided the work in four parts. The first part is the place to give a snapshot on the history of formal linguistics, to present the field of NLP and the NooJ software and the linguistic units that have been treated. The second part is devoted to the description of the process that has been followed for the treatment and the integration of Kabyle verbs in NooJ. We have built a dictionary that contains 4508 entries and 8762 derived components and some models of flexion for each type which have been linked with each entry. In the third part, we have explained the processing of nouns and other units. We have built, for the nouns, a dictionary (3508 entries, 501 derived components) that have been linked to the models of flexion and for the other units (870 entries including adverbs, prepositions, conjunctions, interrogatives, personal pronouns, etc.). The second and third part are completed by examples of applications on a text, this procedure has allowed us to show with various sort of annotations the ambiguities.Regarding the last part we have devoted it to ambiguities, after having identified a list of various types of amalgams, we have tried to show, with the help of some examples of syntactic grammars, some of the tools used by NooJ for disambiguation
Neme, Alexis. "An arabic language resource for computational morphology based on the semitic model." Thesis, Paris Est, 2020. http://www.theses.fr/2020PESC2013.
Full textWe developed an original approach to Arabic traditional morphology, involving new concepts in Semitic lexicology, morphology, and grammar for standard written Arabic. This new methodology for handling the rich and complex Semitic languages is based on good practices in Finite-State technologies (FSA/FST) by using Unitex, a lexicon-based corpus processing suite. For verbs (Neme, 2011), I proposed an inflectional taxonomy that increases the lexicon readability and makes it easier for Arabic speakers and linguists to encode, correct, and update it. Traditional grammar defines inflectional verbal classes by using verbal pattern-classes and root-classes. In our taxonomy, traditional pattern-classes are reused, and root-classes are redefined into a simpler system. The lexicon of verbs covered more than 99% of an evaluation corpus. For nouns and adjectives (Neme, 2013), we went one step further in the adaptation of traditional morphology. First, while this tradition is based on derivational rules, we found our description on inflectional ones. Next, we keep the concepts of root and pattern, which is the backbone of the traditional Semitic model. Still, our breakthrough lies in the reversal of the traditional root-and-pattern Semitic model into a pattern-and-root model, which keeps small and orderly the set of pattern classes and root sub-classes. I elaborated a taxonomy for broken plural containing 160 inflectional classes, which simplifies ten times the encoding of broken plural. Since then, I elaborated comprehensive resources for Arabic. These resources are described in Neme and Paumier (2019). To take into account all aspects of the rich morphology of Arabic, I have completed our taxonomy with suffixal inflexional classes for regular plurals, adverbs, and other parts of speech (POS) to cover all the lexicon. In all, I identified around 1000 Semitic and suffixal inflectional classes implemented with concatenative and non-concatenative FST devices.From scratch, I created 76000 fully vowelized lemmas, and each one is associated with an inflectional class. These lemmas are inflected by using these 1000 FSTs, producing a fully inflected lexicon with more than 6 million forms. I extended this fully inflected resource using agglutination grammars to identify words composed of up to 5 segments, agglutinated around a core inflected verb, noun, adjective, or particle. The agglutination grammars extend the recognition to more than 500 million valid delimited word forms, partially or fully vowelized. The flat file size of 6 million forms is 340 megabytes (UTF-16). It is compressed then into 11 Mbytes before loading to memory for fast retrieval. The generation, compression, and minimization of the full-form lexicon take less than one minute on a common Unix laptop. The lexical coverage rate is more than 99%. The tagger speed is 5000 words/second, and more than 200 000 words/s, if the resources are preloaded/resident in the RAM. The accuracy and speed of our tools result from our systematic linguistic approach and from our choice to embrace the best practices in mathematical and computational methods. The lookup procedure is fast because we use Minimal Acyclic Deterministic Finite Automaton (Revuz, 1992) to compress the full-form dictionary, and because it has only constant strings and no embedded rules. The breakthrough of our linguistic approach remains principally on the reversal of the traditional root-and-pattern Semitic model into a pattern-and-root model.Nonetheless, our computational approach is based on good practices in Finite-State technologies (FSA/FST) as all the full-forms were computed in advance for accurate identification and to get the best from the FSA compression for fast and efficient lookups
Mars, Mourad. "Analyse morphologique robuste de l'arabe et applications pédagogiques." Thesis, Grenoble, 2012. http://www.theses.fr/2012GRENL046.
Full textL'auteur n'a pas fourni de résumé en anglais
Zhou, Rongyan. "Exploration of opportunities and challenges brought by Industry 4.0 to the global supply chains and the macroeconomy by integrating Artificial Intelligence and more traditional methods." Electronic Thesis or Diss., université Paris-Saclay, 2021. http://www.theses.fr/2021UPAST037.
Full textIndustry 4.0 is a significant shift and a tremendous challenge for every industrial segment, especially for the manufacturing industry that gave birth to the new industrial revolution. The research first uses literature analysis to sort out the literature, and focuses on the use of “core literature extension method” to enumerate the development direction and application status of different fields, which devotes to showing a leading role for theory and practice of industry 4.0. The research then explores the main trend of multi-tier supply in Industry 4.0 by combining machine learning and traditional methods. Next, the research investigates the relationship of industry 4.0 investment and employment to look into the inter-regional dependence of industry 4.0 so as to present a reasonable clustering based on different criteria and make suggestions and analysis of the global supply chain for enterprises and organizations. Furthermore, our analysis system takes a glance at the macroeconomy. The combination of natural language processing in machine learning to classify research topics and traditional literature review to investigate the multi-tier supply chain significantly improves the study's objectivity and lays a solid foundation for further research. Using complex networks and econometrics to analyze the global supply chain and macroeconomic issues enriches the research methodology at the macro and policy level. This research provides analysis and references to researchers, decision-makers, and companies for their strategic decision-making
Books on the topic "Traitement du Langage Naturel (NLP)"
Philip, Miller, and Torris Thérèse, eds. Formalismes syntaxiques pour le traitement automatique du langage naturel. Paris: Hermes, 1990.
Find full textL' intelligence artificielle et le langage. 2nd ed. Paris: Hermes, 1990.
Find full textJacobs, Ceriel J. H., 1955-, ed. Parsing techniques: A practical guide. New York: Ellis Horwood, 1990.
Find full textPowers, David M. W. Machine learning of natural language. London: Springer-Verlag, 1989.
Find full textGérard, Ligozat, ed. Outils logiques pour le traitement du temps: De la linguistique à l'intelligence artificielle. Paris: Masson, 1989.
Find full textAsker, Zadeh Lotfi, and Kacprzyk Janusz, eds. Computing with words in information/intelligent systems. New York: Physica-Verlag, 1999.
Find full textGoshawke, Walter. Computer translation of natural language. Wilmslow, United Kingdom: Sigma Press, 1987.
Find full textInternational Conference on Applications of Natural Language to Information Systems (9th 2004 Salford, England). Natural language processing and information systems: 9th International Conference on Applications of Natural Language to Information Systems, NLDB 2004, Salford, UK, June 23-25, 2004 : proceedings. Berlin: Springer, 2004.
Find full textIntroduction to natural language processing. Reston: Reston Publishing Company, 1985.
Find full textHarris, Mary Dee. Introduction to natural language processing. Reston, Va: Reston Pub. Co., 1985.
Find full textBook chapters on the topic "Traitement du Langage Naturel (NLP)"
Gicquel, Quentin, Denys Proux, Pierre Marchal, Caroline Hagége, Yasmina Berrouane, Stéfan J. Darmoni, Suzanne Pereira, Frédérique Segond, and Marie-Héléne Metzger. "Évaluation d’un outil d’aide á l’anonymisation des documents médicaux basé sur le traitement automatique du langage naturel." In Informatique et Santé, 165–76. Paris: Springer Paris, 2011. http://dx.doi.org/10.1007/978-2-8178-0285-5_15.
Full textFLEURY SOARES, Gustavo, and Induraj PUDHUPATTU RAMAMURTHY. "Comparaison de modèles d’apprentissage automatique et d’apprentissage profond." In Optimisation et apprentissage, 153–71. ISTE Group, 2023. http://dx.doi.org/10.51926/iste.9071.ch6.
Full textConference papers on the topic "Traitement du Langage Naturel (NLP)"
ORLIANGES, Jean-Christophe, Younes El Moustakime, Aurelian Crunteanu STANESCU, Ricardo Carrizales Juarez, and Oihan Allegret. "Retour vers le perceptron - fabrication d’un neurone synthétique à base de composants électroniques analogiques simples." In Les journées de l'interdisciplinarité 2023. Limoges: Université de Limoges, 2024. http://dx.doi.org/10.25965/lji.761.
Full text