Literatura científica selecionada sobre o tema "Large Language Models (LLMs)"
Crie uma referência precisa em APA, MLA, Chicago, Harvard, e outros estilos
Consulte a lista de atuais artigos, livros, teses, anais de congressos e outras fontes científicas relevantes para o tema "Large Language Models (LLMs)".
Ao lado de cada fonte na lista de referências, há um botão "Adicionar à bibliografia". Clique e geraremos automaticamente a citação bibliográfica do trabalho escolhido no estilo de citação de que você precisa: APA, MLA, Harvard, Chicago, Vancouver, etc.
Você também pode baixar o texto completo da publicação científica em formato .pdf e ler o resumo do trabalho online se estiver presente nos metadados.
Artigos de revistas sobre o assunto "Large Language Models (LLMs)"
Zhang, Tianyi, Faisal Ladhak, Esin Durmus, Percy Liang, Kathleen McKeown e Tatsunori B. Hashimoto. "Benchmarking Large Language Models for News Summarization". Transactions of the Association for Computational Linguistics 12 (2024): 39–57. http://dx.doi.org/10.1162/tacl_a_00632.
Texto completo da fonteHamaniuk, Vita A. "The potential of Large Language Models in language education". Educational Dimension 5 (9 de dezembro de 2021): 208–10. http://dx.doi.org/10.31812/ed.650.
Texto completo da fonteYang, Jidong. "Large language models privacy and security". Applied and Computational Engineering 76, n.º 1 (16 de julho de 2024): 177–88. http://dx.doi.org/10.54254/2755-2721/76/20240584.
Texto completo da fonteHuang, Dawei, Chuan Yan, Qing Li e Xiaojiang Peng. "From Large Language Models to Large Multimodal Models: A Literature Review". Applied Sciences 14, n.º 12 (11 de junho de 2024): 5068. http://dx.doi.org/10.3390/app14125068.
Texto completo da fonteKumar, Deepak, Yousef Anees AbuHashem e Zakir Durumeric. "Watch Your Language: Investigating Content Moderation with Large Language Models". Proceedings of the International AAAI Conference on Web and Social Media 18 (28 de maio de 2024): 865–78. http://dx.doi.org/10.1609/icwsm.v18i1.31358.
Texto completo da fontePendyala, Vishnu S., e Christopher E. Hall. "Explaining Misinformation Detection Using Large Language Models". Electronics 13, n.º 9 (26 de abril de 2024): 1673. http://dx.doi.org/10.3390/electronics13091673.
Texto completo da fonteCheng, Jerome. "Applications of Large Language Models in Pathology". Bioengineering 11, n.º 4 (31 de março de 2024): 342. http://dx.doi.org/10.3390/bioengineering11040342.
Texto completo da fonteChu, Zhibo, Zichong Wang e Wenbin Zhang. "Fairness in Large Language Models: A Taxonomic Survey". ACM SIGKDD Explorations Newsletter 26, n.º 1 (24 de julho de 2024): 34–48. http://dx.doi.org/10.1145/3682112.3682117.
Texto completo da fonteLin, Hsiao-Ying, e Jeffrey Voas. "Lower Energy Large Language Models (LLMs)". Computer 56, n.º 10 (outubro de 2023): 14–16. http://dx.doi.org/10.1109/mc.2023.3278160.
Texto completo da fonteLong, Robert. "Introspective Capabilities in Large Language Models". Journal of Consciousness Studies 30, n.º 9 (30 de setembro de 2023): 143–53. http://dx.doi.org/10.53765/20512201.30.9.143.
Texto completo da fonteTeses / dissertações sobre o assunto "Large Language Models (LLMs)"
Labeau, Matthieu. "Neural language models : Dealing with large vocabularies". Thesis, Université Paris-Saclay (ComUE), 2018. http://www.theses.fr/2018SACLS313/document.
Texto completo da fonteThis work investigates practical methods to ease training and improve performances of neural language models with large vocabularies. The main limitation of neural language models is their expensive computational cost: it depends on the size of the vocabulary, with which it grows linearly. Despite several training tricks, the most straightforward way to limit computation time is to limit the vocabulary size, which is not a satisfactory solution for numerous tasks. Most of the existing methods used to train large-vocabulary language models revolve around avoiding the computation of the partition function, ensuring that output scores are normalized into a probability distribution. Here, we focus on sampling-based approaches, including importance sampling and noise contrastive estimation. These methods allow an approximate computation of the partition function. After examining the mechanism of self-normalization in noise-contrastive estimation, we first propose to improve its efficiency with solutions that are adapted to the inner workings of the method and experimentally show that they considerably ease training. Our second contribution is to expand on a generalization of several sampling based objectives as Bregman divergences, in order to experiment with new objectives. We use Beta divergences to derive a set of objectives from which noise contrastive estimation is a particular case. Finally, we aim at improving performances on full vocabulary language models, by augmenting output words representation with subwords. We experiment on a Czech dataset and show that using character-based representations besides word embeddings for output representations gives better results. We also show that reducing the size of the output look-up table improves results even more
Zervakis, Georgios. "Enriching large language models with semantic lexicons and analogies". Electronic Thesis or Diss., Université de Lorraine, 2023. http://www.theses.fr/2023LORR0039.
Texto completo da fonteRecent advances in deep learning and neural networks have made it possible to address complex natural language processing tasks, which find application in a plethora of real-world problems ranging from smart assistants in mobile devices to the prediction of cancer. Nonetheless, modern systems based on these frameworks exhibit various limitations that may compromise their performance and trustworthiness, render them unfair towards minorities, or subject them to privacy leakage. It is our belief that integrating symbolic knowledge and reasoning into the deep learning framework is a necessary step towards addressing the aforementioned limitations. For example, lexical resources can enrich deep neural networks with semantic or syntactic knowledge, and logical rules can provide learning and reasoning mechanisms. Therefore, the scope of this thesis is to develop and evaluate ways of integrating different types of symbolic knowledge and reasoning into a widely used language model, Bidirectional Encoder Representations from Transformers (BERT). ln a first stage, we consider retrofitting, a simple and popular technique for refining distributional word embeddings based on relations coming from a semantic lexicon. Inspired by this technique, we present two methods for incorporating this knowledge into BERT contextualized embeddings. We evaluate these methods on three biomedical datasets for relation extraction and one movie review dataset for sentiment analysis, and show that they do not substantially impact the performance for these tasks. Furthermore, we conduct a qualitative analysis to provide further insights on this negative result. ln a second stage, we integrate analogical reasoning with BERT as a means to improve its performance on the target sense verification task, and make it more robust. To do so, we reformulate target sense verification as an analogy detection task. We present a hybrid model that combines BERT to encode the input data into quadruples and a convolutional neural classifier to decide whether they constitute valid analogies. We test our system on a benchmark dataset, and show that it can outperform existing approaches. Our empirical study shows the importance of the input encoding for BERT, and how this dependence gets alleviated by integrating the axiomatic properties of analogies during training, while preserving performance and improving robustness
Chadha, Vikrampal. "Simulation of large-scale system-level models". Thesis, This resource online, 1994. http://scholar.lib.vt.edu/theses/available/etd-12162009-020334/.
Texto completo da fonteHittner, Brian Edward. "Rendering large-scale terrain models and positioning objects in relation to 3D terrain". Thesis, Monterey, Calif. : Springfield, Va. : Naval Postgraduate School ; Available from National Technical Information Service, 2003. http://library.nps.navy.mil/uhtbin/hyperion-image/03Dec%5FHittner.pdf.
Texto completo da fonteThesis advisor(s): Don Brutzman, Curt Blais. Includes bibliographical references (p. 117-118). Also available online.
Kropff, Emilio. "Statistical and dynamical properties of large cortical network models: insights into semantic memory and language". Doctoral thesis, SISSA, 2007. http://hdl.handle.net/20.500.11767/4639.
Texto completo da fonteZhao, Ying, e ying zhao@rmit edu au. "Effective Authorship Attribution in Large Document Collections". RMIT University. Computer Science and Information Technology, 2008. http://adt.lib.rmit.edu.au/adt/public/adt-VIT20080730.162501.
Texto completo da fontePan, Bi-Yu. "Hierarchical test generation for VHDL behavioral models". Thesis, This resource online, 1992. http://scholar.lib.vt.edu/theses/available/etd-09052009-040449/.
Texto completo da fonteWest, James F. "An examination of the application of design metrics to the development of testing strategies in large-scale SDL models". Virtual Press, 2000. http://liblink.bsu.edu/uhtbin/catkey/1191725.
Texto completo da fonteDepartment of Computer Science
Kapoor, Shekhar. "Process level test generation for VHDL behavioral models". Thesis, This resource online, 1994. http://scholar.lib.vt.edu/theses/available/etd-05022009-040753/.
Texto completo da fonteNarayanaswamy, Sathyanarayanan. "Development of VHDL behavioral models with back annotated timing". Thesis, This resource online, 1994. http://scholar.lib.vt.edu/theses/available/etd-06112009-063442/.
Texto completo da fonteLivros sobre o assunto "Large Language Models (LLMs)"
Amaratunga, Thimira. Understanding Large Language Models. Berkeley, CA: Apress, 2023. http://dx.doi.org/10.1007/979-8-8688-0017-7.
Texto completo da fonteMartra, Pere. Large Language Models Projects. Berkeley, CA: Apress, 2024. http://dx.doi.org/10.1007/979-8-8688-0515-8.
Texto completo da fonteKucharavy, Andrei, Octave Plancherel, Valentin Mulder, Alain Mermoud e Vincent Lenders, eds. Large Language Models in Cybersecurity. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7.
Texto completo da fonteKamath, Uday, Kevin Keenan, Garrett Somers e Sarah Sorenson. Large Language Models: A Deep Dive. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-65647-7.
Texto completo da fonteGrigorov, Dilyan. Introduction to Python and Large Language Models. Berkeley, CA: Apress, 2024. http://dx.doi.org/10.1007/979-8-8688-0540-0.
Texto completo da fonteTörnberg, Petter. How to Use Large-Language Models for Text Analysis. 1 Oliver’s Yard, 55 City Road, London EC1Y 1SP United Kingdom: SAGE Publications Ltd, 2024. http://dx.doi.org/10.4135/9781529683707.
Texto completo da fonteKing, Susan. Waking the princess. Waterville, Me: Wheeler Pub., 2004.
Encontre o texto completo da fonteKing, Susan. La princesa dormida. Barcelona: Ediciones Urano, 2010.
Encontre o texto completo da fonteKing, Susan. Waking the princess. New York: New American Library, 2003.
Encontre o texto completo da fonteLewis, Carroll. Through the looking glass. San Diego, CA: ICON Classics, 2005.
Encontre o texto completo da fonteCapítulos de livros sobre o assunto "Large Language Models (LLMs)"
Amaratunga, Thimira. "Popular LLMs". In Understanding Large Language Models, 119–30. Berkeley, CA: Apress, 2023. http://dx.doi.org/10.1007/979-8-8688-0017-7_5.
Texto completo da fonteRuiu, Dragos. "LLMs Red Teaming". In Large Language Models in Cybersecurity, 213–23. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7_24.
Texto completo da fonteAmaratunga, Thimira. "What Makes LLMs Large?" In Understanding Large Language Models, 81–117. Berkeley, CA: Apress, 2023. http://dx.doi.org/10.1007/979-8-8688-0017-7_4.
Texto completo da fonteMartra, Pere. "Vector Databases and LLMs". In Large Language Models Projects, 31–62. Berkeley, CA: Apress, 2024. http://dx.doi.org/10.1007/979-8-8688-0515-8_2.
Texto completo da fonteKucharavy, Andrei. "Overview of Existing LLM Families". In Large Language Models in Cybersecurity, 31–44. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7_3.
Texto completo da fonteWürsch, Maxime, Dimitri Percia David e Alain Mermoud. "Monitoring Emerging Trends in LLM Research". In Large Language Models in Cybersecurity, 153–61. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7_17.
Texto completo da fonteKamath, Uday, Kevin Keenan, Garrett Somers e Sarah Sorenson. "Multimodal LLMs". In Large Language Models: A Deep Dive, 375–421. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-65647-7_9.
Texto completo da fonteKucharavy, Andrei. "Fundamental Limitations of Generative LLMs". In Large Language Models in Cybersecurity, 55–64. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7_5.
Texto completo da fonteMeier, Raphael. "LLM-Aided Social Media Influence Operations". In Large Language Models in Cybersecurity, 105–12. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7_11.
Texto completo da fonteSchillaci, Zachary. "LLM Adoption Trends and Associated Risks". In Large Language Models in Cybersecurity, 121–28. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7_13.
Texto completo da fonteTrabalhos de conferências sobre o assunto "Large Language Models (LLMs)"
Coppolillo, Erica, Francesco Calimeri, Giuseppe Manco, Simona Perri e Francesco Ricca. "LLASP: Fine-tuning Large Language Models for Answer Set Programming". In 21st International Conference on Principles of Knowledge Representation and Reasoning {KR-2023}, 834–44. California: International Joint Conferences on Artificial Intelligence Organization, 2024. http://dx.doi.org/10.24963/kr.2024/78.
Texto completo da fonteZhang, Duzhen, Yahan Yu, Jiahua Dong, Chenxing Li, Dan Su, Chenhui Chu e Dong Yu. "MM-LLMs: Recent Advances in MultiModal Large Language Models". In Findings of the Association for Computational Linguistics ACL 2024, 12401–30. Stroudsburg, PA, USA: Association for Computational Linguistics, 2024. http://dx.doi.org/10.18653/v1/2024.findings-acl.738.
Texto completo da fonteZolnai-Lucas, Aaron, Jack Boylan, Chris Hokamp e Parsa Ghaffari. "STAGE: Simplified Text-Attributed Graph Embeddings using Pre-trained LLMs". In Proceedings of the 1st Workshop on Knowledge Graphs and Large Language Models (KaLLM 2024), 92–104. Stroudsburg, PA, USA: Association for Computational Linguistics, 2024. http://dx.doi.org/10.18653/v1/2024.kallm-1.10.
Texto completo da fonteBarek, MD Abdul, Md Mostafizur Rahman, Shapna Akter, A. B. M. Kamrul Islam Riad, Md Abdur Rahman, Hossain Shahriar, Akond Rahman e Fan Wu. "Mitigating Insecure Outputs in Large Language Models(LLMs): A Practical Educational Module". In 2024 IEEE 48th Annual Computers, Software, and Applications Conference (COMPSAC), 2424–29. IEEE, 2024. http://dx.doi.org/10.1109/compsac61105.2024.00389.
Texto completo da fonteKotkar, Aishwarya D., Radhakrushna S. Mahadik, Piyush G. More e Sandeep A. Thorat. "Comparative Analysis of Transformer-based Large Language Models (LLMs) for Text Summarization". In 2024 1st International Conference on Advanced Computing and Emerging Technologies (ACET), 1–7. IEEE, 2024. http://dx.doi.org/10.1109/acet61898.2024.10730348.
Texto completo da fonteGurgurov, Daniil, Mareike Hartmann e Simon Ostermann. "Adapting Multilingual LLMs to Low-Resource Languages with Knowledge Graphs via Adapters". In Proceedings of the 1st Workshop on Knowledge Graphs and Large Language Models (KaLLM 2024), 63–74. Stroudsburg, PA, USA: Association for Computational Linguistics, 2024. http://dx.doi.org/10.18653/v1/2024.kallm-1.7.
Texto completo da fonteWasi, Azmine Toushik. "HRGraph: Leveraging LLMs for HR Data Knowledge Graphs with Information Propagation-based Job Recommendation". In Proceedings of the 1st Workshop on Knowledge Graphs and Large Language Models (KaLLM 2024), 56–62. Stroudsburg, PA, USA: Association for Computational Linguistics, 2024. http://dx.doi.org/10.18653/v1/2024.kallm-1.6.
Texto completo da fonteCao, Ye. "Optimizing Style Recognition Algorithm for Digital Art Images Using Large Language Models (LLMs)". In 2024 5th International Conference on Electronics and Sustainable Communication Systems (ICESC), 1536–42. IEEE, 2024. http://dx.doi.org/10.1109/icesc60852.2024.10689850.
Texto completo da fonteAlipour, Hanieh, Nick Pendar e Kohinoor Roy. "ChatGPT Alternative Solutions: Large Language Models Survey". In 5th International Conference on Networks, Blockchain and Internet of Things. Academy & Industry Research Collaboration Center, 2024. http://dx.doi.org/10.5121/csit.2024.1405114.
Texto completo da fonteMa, Kevin, Daniele Grandi, Christopher McComb e Kosa Goucher-Lambert. "Conceptual Design Generation Using Large Language Models". In ASME 2023 International Design Engineering Technical Conferences and Computers and Information in Engineering Conference. American Society of Mechanical Engineers, 2023. http://dx.doi.org/10.1115/detc2023-116838.
Texto completo da fonteRelatórios de organizações sobre o assunto "Large Language Models (LLMs)"
Alonso-Robisco, Andres, e Jose Manuel Carbo. Analysis of CBDC Narrative OF Central Banks using Large Language Models. Madrid: Banco de España, agosto de 2023. http://dx.doi.org/10.53479/33412.
Texto completo da fonteMarra de Artiñano, Ignacio, Franco Riottini Depetris e Christian Volpe Martincus. Automatic Product Classification in International Trade: Machine Learning and Large Language Models. Inter-American Development Bank, julho de 2023. http://dx.doi.org/10.18235/0005012.
Texto completo da fonteMaerz, Seraphine. Using AI for Text Analysis in R. Instats Inc., 2024. http://dx.doi.org/10.61700/ti5uexui5ilrd1663.
Texto completo da fonteAzuara Herrera, Oliver, Laura Ripani e Eric Torres Ramirez. AI and the Increase of Productivity and Labor Inequality in Latin America: Potential Impact of Large Language Models on Latin American Workforce. Inter-American Development Bank, setembro de 2024. http://dx.doi.org/10.18235/0013152.
Texto completo da fonteKorinek, Anton, e Jai Vipra. Concentrating Intelligence: Scaling and Market Structure in Artificial Intelligence. Institute for New Economic Thinking Working Paper Series, outubro de 2024. http://dx.doi.org/10.36687/inetwp228.
Texto completo da fonteWagner, Rudolf. Enforcing Software and AI as Medical Devices: Expert Witness Insights on Civil Lawsuits, Regulation, and Legal Liability Pathways. ADHOCON UG (haftungsbeschränkt), outubro de 2024. http://dx.doi.org/10.70317/2024.13rw10.
Texto completo da fonteWagner, Rudolf. Enforcing Software and AI as Medical Devices: Expert Witness Insights on Civil Lawsuits, Regulation, and Legal Liability Pathways. ADHOCON UG (haftungsbeschraenkt), outubro de 2024. http://dx.doi.org/10.70317/2024.20rw10.
Texto completo da fonteZhang, Hao. Large Language Model (LLM) Monthly Report (2024 Apr). ResearchHub Technologies, Inc., maio de 2024. http://dx.doi.org/10.55277/researchhub.0ps6xenm.
Texto completo da fonteTranchero, Matteo, Cecil-Francis Brenninkmeijer, Arul Murugan e Abhishek Nagaraj. Theorizing with Large Language Models. Cambridge, MA: National Bureau of Economic Research, outubro de 2024. http://dx.doi.org/10.3386/w33033.
Texto completo da fonteMoreno, Ángel Iván, e Teresa Caminero. Assessing the data challenges of climate-related disclosures in european banks. A text mining study. Madrid: Banco de España, setembro de 2023. http://dx.doi.org/10.53479/33752.
Texto completo da fonte