Letteratura scientifica selezionata sul tema "Large Language Models (LLMs)"
Cita una fonte nei formati APA, MLA, Chicago, Harvard e in molti altri stili
Consulta la lista di attuali articoli, libri, tesi, atti di convegni e altre fonti scientifiche attinenti al tema "Large Language Models (LLMs)".
Accanto a ogni fonte nell'elenco di riferimenti c'è un pulsante "Aggiungi alla bibliografia". Premilo e genereremo automaticamente la citazione bibliografica dell'opera scelta nello stile citazionale di cui hai bisogno: APA, MLA, Harvard, Chicago, Vancouver ecc.
Puoi anche scaricare il testo completo della pubblicazione scientifica nel formato .pdf e leggere online l'abstract (il sommario) dell'opera se è presente nei metadati.
Articoli di riviste sul tema "Large Language Models (LLMs)"
Zhang, Tianyi, Faisal Ladhak, Esin Durmus, Percy Liang, Kathleen McKeown e Tatsunori B. Hashimoto. "Benchmarking Large Language Models for News Summarization". Transactions of the Association for Computational Linguistics 12 (2024): 39–57. http://dx.doi.org/10.1162/tacl_a_00632.
Testo completoHamaniuk, Vita A. "The potential of Large Language Models in language education". Educational Dimension 5 (9 dicembre 2021): 208–10. http://dx.doi.org/10.31812/ed.650.
Testo completoYang, Jidong. "Large language models privacy and security". Applied and Computational Engineering 76, n. 1 (16 luglio 2024): 177–88. http://dx.doi.org/10.54254/2755-2721/76/20240584.
Testo completoHuang, Dawei, Chuan Yan, Qing Li e Xiaojiang Peng. "From Large Language Models to Large Multimodal Models: A Literature Review". Applied Sciences 14, n. 12 (11 giugno 2024): 5068. http://dx.doi.org/10.3390/app14125068.
Testo completoKumar, Deepak, Yousef Anees AbuHashem e Zakir Durumeric. "Watch Your Language: Investigating Content Moderation with Large Language Models". Proceedings of the International AAAI Conference on Web and Social Media 18 (28 maggio 2024): 865–78. http://dx.doi.org/10.1609/icwsm.v18i1.31358.
Testo completoPendyala, Vishnu S., e Christopher E. Hall. "Explaining Misinformation Detection Using Large Language Models". Electronics 13, n. 9 (26 aprile 2024): 1673. http://dx.doi.org/10.3390/electronics13091673.
Testo completoCheng, Jerome. "Applications of Large Language Models in Pathology". Bioengineering 11, n. 4 (31 marzo 2024): 342. http://dx.doi.org/10.3390/bioengineering11040342.
Testo completoChu, Zhibo, Zichong Wang e Wenbin Zhang. "Fairness in Large Language Models: A Taxonomic Survey". ACM SIGKDD Explorations Newsletter 26, n. 1 (24 luglio 2024): 34–48. http://dx.doi.org/10.1145/3682112.3682117.
Testo completoLin, Hsiao-Ying, e Jeffrey Voas. "Lower Energy Large Language Models (LLMs)". Computer 56, n. 10 (ottobre 2023): 14–16. http://dx.doi.org/10.1109/mc.2023.3278160.
Testo completoLong, Robert. "Introspective Capabilities in Large Language Models". Journal of Consciousness Studies 30, n. 9 (30 settembre 2023): 143–53. http://dx.doi.org/10.53765/20512201.30.9.143.
Testo completoTesi sul tema "Large Language Models (LLMs)"
Labeau, Matthieu. "Neural language models : Dealing with large vocabularies". Thesis, Université Paris-Saclay (ComUE), 2018. http://www.theses.fr/2018SACLS313/document.
Testo completoThis work investigates practical methods to ease training and improve performances of neural language models with large vocabularies. The main limitation of neural language models is their expensive computational cost: it depends on the size of the vocabulary, with which it grows linearly. Despite several training tricks, the most straightforward way to limit computation time is to limit the vocabulary size, which is not a satisfactory solution for numerous tasks. Most of the existing methods used to train large-vocabulary language models revolve around avoiding the computation of the partition function, ensuring that output scores are normalized into a probability distribution. Here, we focus on sampling-based approaches, including importance sampling and noise contrastive estimation. These methods allow an approximate computation of the partition function. After examining the mechanism of self-normalization in noise-contrastive estimation, we first propose to improve its efficiency with solutions that are adapted to the inner workings of the method and experimentally show that they considerably ease training. Our second contribution is to expand on a generalization of several sampling based objectives as Bregman divergences, in order to experiment with new objectives. We use Beta divergences to derive a set of objectives from which noise contrastive estimation is a particular case. Finally, we aim at improving performances on full vocabulary language models, by augmenting output words representation with subwords. We experiment on a Czech dataset and show that using character-based representations besides word embeddings for output representations gives better results. We also show that reducing the size of the output look-up table improves results even more
Zervakis, Georgios. "Enriching large language models with semantic lexicons and analogies". Electronic Thesis or Diss., Université de Lorraine, 2023. http://www.theses.fr/2023LORR0039.
Testo completoRecent advances in deep learning and neural networks have made it possible to address complex natural language processing tasks, which find application in a plethora of real-world problems ranging from smart assistants in mobile devices to the prediction of cancer. Nonetheless, modern systems based on these frameworks exhibit various limitations that may compromise their performance and trustworthiness, render them unfair towards minorities, or subject them to privacy leakage. It is our belief that integrating symbolic knowledge and reasoning into the deep learning framework is a necessary step towards addressing the aforementioned limitations. For example, lexical resources can enrich deep neural networks with semantic or syntactic knowledge, and logical rules can provide learning and reasoning mechanisms. Therefore, the scope of this thesis is to develop and evaluate ways of integrating different types of symbolic knowledge and reasoning into a widely used language model, Bidirectional Encoder Representations from Transformers (BERT). ln a first stage, we consider retrofitting, a simple and popular technique for refining distributional word embeddings based on relations coming from a semantic lexicon. Inspired by this technique, we present two methods for incorporating this knowledge into BERT contextualized embeddings. We evaluate these methods on three biomedical datasets for relation extraction and one movie review dataset for sentiment analysis, and show that they do not substantially impact the performance for these tasks. Furthermore, we conduct a qualitative analysis to provide further insights on this negative result. ln a second stage, we integrate analogical reasoning with BERT as a means to improve its performance on the target sense verification task, and make it more robust. To do so, we reformulate target sense verification as an analogy detection task. We present a hybrid model that combines BERT to encode the input data into quadruples and a convolutional neural classifier to decide whether they constitute valid analogies. We test our system on a benchmark dataset, and show that it can outperform existing approaches. Our empirical study shows the importance of the input encoding for BERT, and how this dependence gets alleviated by integrating the axiomatic properties of analogies during training, while preserving performance and improving robustness
Chadha, Vikrampal. "Simulation of large-scale system-level models". Thesis, This resource online, 1994. http://scholar.lib.vt.edu/theses/available/etd-12162009-020334/.
Testo completoHittner, Brian Edward. "Rendering large-scale terrain models and positioning objects in relation to 3D terrain". Thesis, Monterey, Calif. : Springfield, Va. : Naval Postgraduate School ; Available from National Technical Information Service, 2003. http://library.nps.navy.mil/uhtbin/hyperion-image/03Dec%5FHittner.pdf.
Testo completoThesis advisor(s): Don Brutzman, Curt Blais. Includes bibliographical references (p. 117-118). Also available online.
Kropff, Emilio. "Statistical and dynamical properties of large cortical network models: insights into semantic memory and language". Doctoral thesis, SISSA, 2007. http://hdl.handle.net/20.500.11767/4639.
Testo completoZhao, Ying, e ying zhao@rmit edu au. "Effective Authorship Attribution in Large Document Collections". RMIT University. Computer Science and Information Technology, 2008. http://adt.lib.rmit.edu.au/adt/public/adt-VIT20080730.162501.
Testo completoPan, Bi-Yu. "Hierarchical test generation for VHDL behavioral models". Thesis, This resource online, 1992. http://scholar.lib.vt.edu/theses/available/etd-09052009-040449/.
Testo completoWest, James F. "An examination of the application of design metrics to the development of testing strategies in large-scale SDL models". Virtual Press, 2000. http://liblink.bsu.edu/uhtbin/catkey/1191725.
Testo completoDepartment of Computer Science
Kapoor, Shekhar. "Process level test generation for VHDL behavioral models". Thesis, This resource online, 1994. http://scholar.lib.vt.edu/theses/available/etd-05022009-040753/.
Testo completoNarayanaswamy, Sathyanarayanan. "Development of VHDL behavioral models with back annotated timing". Thesis, This resource online, 1994. http://scholar.lib.vt.edu/theses/available/etd-06112009-063442/.
Testo completoLibri sul tema "Large Language Models (LLMs)"
Amaratunga, Thimira. Understanding Large Language Models. Berkeley, CA: Apress, 2023. http://dx.doi.org/10.1007/979-8-8688-0017-7.
Testo completoMartra, Pere. Large Language Models Projects. Berkeley, CA: Apress, 2024. http://dx.doi.org/10.1007/979-8-8688-0515-8.
Testo completoKucharavy, Andrei, Octave Plancherel, Valentin Mulder, Alain Mermoud e Vincent Lenders, a cura di. Large Language Models in Cybersecurity. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7.
Testo completoKamath, Uday, Kevin Keenan, Garrett Somers e Sarah Sorenson. Large Language Models: A Deep Dive. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-65647-7.
Testo completoGrigorov, Dilyan. Introduction to Python and Large Language Models. Berkeley, CA: Apress, 2024. http://dx.doi.org/10.1007/979-8-8688-0540-0.
Testo completoTörnberg, Petter. How to Use Large-Language Models for Text Analysis. 1 Oliver’s Yard, 55 City Road, London EC1Y 1SP United Kingdom: SAGE Publications Ltd, 2024. http://dx.doi.org/10.4135/9781529683707.
Testo completoKing, Susan. Waking the princess. Waterville, Me: Wheeler Pub., 2004.
Cerca il testo completoKing, Susan. La princesa dormida. Barcelona: Ediciones Urano, 2010.
Cerca il testo completoKing, Susan. Waking the princess. New York: New American Library, 2003.
Cerca il testo completoLewis, Carroll. Through the looking glass. San Diego, CA: ICON Classics, 2005.
Cerca il testo completoCapitoli di libri sul tema "Large Language Models (LLMs)"
Amaratunga, Thimira. "Popular LLMs". In Understanding Large Language Models, 119–30. Berkeley, CA: Apress, 2023. http://dx.doi.org/10.1007/979-8-8688-0017-7_5.
Testo completoRuiu, Dragos. "LLMs Red Teaming". In Large Language Models in Cybersecurity, 213–23. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7_24.
Testo completoAmaratunga, Thimira. "What Makes LLMs Large?" In Understanding Large Language Models, 81–117. Berkeley, CA: Apress, 2023. http://dx.doi.org/10.1007/979-8-8688-0017-7_4.
Testo completoMartra, Pere. "Vector Databases and LLMs". In Large Language Models Projects, 31–62. Berkeley, CA: Apress, 2024. http://dx.doi.org/10.1007/979-8-8688-0515-8_2.
Testo completoKucharavy, Andrei. "Overview of Existing LLM Families". In Large Language Models in Cybersecurity, 31–44. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7_3.
Testo completoWürsch, Maxime, Dimitri Percia David e Alain Mermoud. "Monitoring Emerging Trends in LLM Research". In Large Language Models in Cybersecurity, 153–61. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7_17.
Testo completoKamath, Uday, Kevin Keenan, Garrett Somers e Sarah Sorenson. "Multimodal LLMs". In Large Language Models: A Deep Dive, 375–421. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-65647-7_9.
Testo completoKucharavy, Andrei. "Fundamental Limitations of Generative LLMs". In Large Language Models in Cybersecurity, 55–64. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7_5.
Testo completoMeier, Raphael. "LLM-Aided Social Media Influence Operations". In Large Language Models in Cybersecurity, 105–12. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7_11.
Testo completoSchillaci, Zachary. "LLM Adoption Trends and Associated Risks". In Large Language Models in Cybersecurity, 121–28. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7_13.
Testo completoAtti di convegni sul tema "Large Language Models (LLMs)"
Coppolillo, Erica, Francesco Calimeri, Giuseppe Manco, Simona Perri e Francesco Ricca. "LLASP: Fine-tuning Large Language Models for Answer Set Programming". In 21st International Conference on Principles of Knowledge Representation and Reasoning {KR-2023}, 834–44. California: International Joint Conferences on Artificial Intelligence Organization, 2024. http://dx.doi.org/10.24963/kr.2024/78.
Testo completoZhang, Duzhen, Yahan Yu, Jiahua Dong, Chenxing Li, Dan Su, Chenhui Chu e Dong Yu. "MM-LLMs: Recent Advances in MultiModal Large Language Models". In Findings of the Association for Computational Linguistics ACL 2024, 12401–30. Stroudsburg, PA, USA: Association for Computational Linguistics, 2024. http://dx.doi.org/10.18653/v1/2024.findings-acl.738.
Testo completoZolnai-Lucas, Aaron, Jack Boylan, Chris Hokamp e Parsa Ghaffari. "STAGE: Simplified Text-Attributed Graph Embeddings using Pre-trained LLMs". In Proceedings of the 1st Workshop on Knowledge Graphs and Large Language Models (KaLLM 2024), 92–104. Stroudsburg, PA, USA: Association for Computational Linguistics, 2024. http://dx.doi.org/10.18653/v1/2024.kallm-1.10.
Testo completoBarek, MD Abdul, Md Mostafizur Rahman, Shapna Akter, A. B. M. Kamrul Islam Riad, Md Abdur Rahman, Hossain Shahriar, Akond Rahman e Fan Wu. "Mitigating Insecure Outputs in Large Language Models(LLMs): A Practical Educational Module". In 2024 IEEE 48th Annual Computers, Software, and Applications Conference (COMPSAC), 2424–29. IEEE, 2024. http://dx.doi.org/10.1109/compsac61105.2024.00389.
Testo completoKotkar, Aishwarya D., Radhakrushna S. Mahadik, Piyush G. More e Sandeep A. Thorat. "Comparative Analysis of Transformer-based Large Language Models (LLMs) for Text Summarization". In 2024 1st International Conference on Advanced Computing and Emerging Technologies (ACET), 1–7. IEEE, 2024. http://dx.doi.org/10.1109/acet61898.2024.10730348.
Testo completoGurgurov, Daniil, Mareike Hartmann e Simon Ostermann. "Adapting Multilingual LLMs to Low-Resource Languages with Knowledge Graphs via Adapters". In Proceedings of the 1st Workshop on Knowledge Graphs and Large Language Models (KaLLM 2024), 63–74. Stroudsburg, PA, USA: Association for Computational Linguistics, 2024. http://dx.doi.org/10.18653/v1/2024.kallm-1.7.
Testo completoWasi, Azmine Toushik. "HRGraph: Leveraging LLMs for HR Data Knowledge Graphs with Information Propagation-based Job Recommendation". In Proceedings of the 1st Workshop on Knowledge Graphs and Large Language Models (KaLLM 2024), 56–62. Stroudsburg, PA, USA: Association for Computational Linguistics, 2024. http://dx.doi.org/10.18653/v1/2024.kallm-1.6.
Testo completoCao, Ye. "Optimizing Style Recognition Algorithm for Digital Art Images Using Large Language Models (LLMs)". In 2024 5th International Conference on Electronics and Sustainable Communication Systems (ICESC), 1536–42. IEEE, 2024. http://dx.doi.org/10.1109/icesc60852.2024.10689850.
Testo completoAlipour, Hanieh, Nick Pendar e Kohinoor Roy. "ChatGPT Alternative Solutions: Large Language Models Survey". In 5th International Conference on Networks, Blockchain and Internet of Things. Academy & Industry Research Collaboration Center, 2024. http://dx.doi.org/10.5121/csit.2024.1405114.
Testo completoMa, Kevin, Daniele Grandi, Christopher McComb e Kosa Goucher-Lambert. "Conceptual Design Generation Using Large Language Models". In ASME 2023 International Design Engineering Technical Conferences and Computers and Information in Engineering Conference. American Society of Mechanical Engineers, 2023. http://dx.doi.org/10.1115/detc2023-116838.
Testo completoRapporti di organizzazioni sul tema "Large Language Models (LLMs)"
Alonso-Robisco, Andres, e Jose Manuel Carbo. Analysis of CBDC Narrative OF Central Banks using Large Language Models. Madrid: Banco de España, agosto 2023. http://dx.doi.org/10.53479/33412.
Testo completoMarra de Artiñano, Ignacio, Franco Riottini Depetris e Christian Volpe Martincus. Automatic Product Classification in International Trade: Machine Learning and Large Language Models. Inter-American Development Bank, luglio 2023. http://dx.doi.org/10.18235/0005012.
Testo completoMaerz, Seraphine. Using AI for Text Analysis in R. Instats Inc., 2024. http://dx.doi.org/10.61700/ti5uexui5ilrd1663.
Testo completoAzuara Herrera, Oliver, Laura Ripani e Eric Torres Ramirez. AI and the Increase of Productivity and Labor Inequality in Latin America: Potential Impact of Large Language Models on Latin American Workforce. Inter-American Development Bank, settembre 2024. http://dx.doi.org/10.18235/0013152.
Testo completoKorinek, Anton, e Jai Vipra. Concentrating Intelligence: Scaling and Market Structure in Artificial Intelligence. Institute for New Economic Thinking Working Paper Series, ottobre 2024. http://dx.doi.org/10.36687/inetwp228.
Testo completoWagner, Rudolf. Enforcing Software and AI as Medical Devices: Expert Witness Insights on Civil Lawsuits, Regulation, and Legal Liability Pathways. ADHOCON UG (haftungsbeschränkt), ottobre 2024. http://dx.doi.org/10.70317/2024.13rw10.
Testo completoWagner, Rudolf. Enforcing Software and AI as Medical Devices: Expert Witness Insights on Civil Lawsuits, Regulation, and Legal Liability Pathways. ADHOCON UG (haftungsbeschraenkt), ottobre 2024. http://dx.doi.org/10.70317/2024.20rw10.
Testo completoZhang, Hao. Large Language Model (LLM) Monthly Report (2024 Apr). ResearchHub Technologies, Inc., maggio 2024. http://dx.doi.org/10.55277/researchhub.0ps6xenm.
Testo completoTranchero, Matteo, Cecil-Francis Brenninkmeijer, Arul Murugan e Abhishek Nagaraj. Theorizing with Large Language Models. Cambridge, MA: National Bureau of Economic Research, ottobre 2024. http://dx.doi.org/10.3386/w33033.
Testo completoMoreno, Ángel Iván, e Teresa Caminero. Assessing the data challenges of climate-related disclosures in european banks. A text mining study. Madrid: Banco de España, settembre 2023. http://dx.doi.org/10.53479/33752.
Testo completo