Letteratura scientifica selezionata sul tema "Large language model"
Cita una fonte nei formati APA, MLA, Chicago, Harvard e in molti altri stili
Consulta la lista di attuali articoli, libri, tesi, atti di convegni e altre fonti scientifiche attinenti al tema "Large language model".
Accanto a ogni fonte nell'elenco di riferimenti c'è un pulsante "Aggiungi alla bibliografia". Premilo e genereremo automaticamente la citazione bibliografica dell'opera scelta nello stile citazionale di cui hai bisogno: APA, MLA, Harvard, Chicago, Vancouver ecc.
Puoi anche scaricare il testo completo della pubblicazione scientifica nel formato .pdf e leggere online l'abstract (il sommario) dell'opera se è presente nei metadati.
Articoli di riviste sul tema "Large language model"
B, Mr DHANUSH. "CHATBOT USING LARGE LANGUAGE MODEL". INTERANTIONAL JOURNAL OF SCIENTIFIC RESEARCH IN ENGINEERING AND MANAGEMENT 08, n. 05 (14 maggio 2024): 1–5. http://dx.doi.org/10.55041/ijsrem34001.
Testo completoZhang, Chengyi, Xingyu Wang e Ziyun Wang. "Large language model in electrocatalysis". Chinese Journal of Catalysis 59 (aprile 2024): 7–14. http://dx.doi.org/10.1016/s1872-2067(23)64612-1.
Testo completoSagi, Sriram. "Advancing AI: Enhancing Large Language Model Performance through GPU Optimization Techniques". International Journal of Science and Research (IJSR) 13, n. 3 (5 marzo 2024): 630–33. http://dx.doi.org/10.21275/sr24309100709.
Testo completoBaral, Elina, e Sagar Shrestha. "Large Vocabulary Continuous Speech Recognition for Nepali Language". International Journal of Signal Processing Systems 8, n. 4 (dicembre 2020): 68–73. http://dx.doi.org/10.18178/ijsps.8.4.68-73.
Testo completoGarg, Prerak, e Divya Beeram. "Large Language Model-Based Autonomous Agents". International Journal of Computer Trends and Technology 72, n. 5 (30 maggio 2024): 151–62. http://dx.doi.org/10.14445/22312803/ijctt-v72i5p118.
Testo completoHuang, Sen, Kaixiang Yang, Sheng Qi e Rui Wang. "When large language model meets optimization". Swarm and Evolutionary Computation 90 (ottobre 2024): 101663. http://dx.doi.org/10.1016/j.swevo.2024.101663.
Testo completoShi, Zhouxing, Yihan Wang, Fan Yin, Xiangning Chen, Kai-Wei Chang e Cho-Jui Hsieh. "Red Teaming Language Model Detectors with Language Models". Transactions of the Association for Computational Linguistics 12 (2024): 174–89. http://dx.doi.org/10.1162/tacl_a_00639.
Testo completoAman, Mussa. "Large Language Model Based Fake News Detection". Procedia Computer Science 231 (2024): 740–45. http://dx.doi.org/10.1016/j.procs.2023.12.144.
Testo completoSingh, Pranaydeep, Orphée De Clercq e Els Lefever. "Distilling Monolingual Models from Large Multilingual Transformers". Electronics 12, n. 4 (18 febbraio 2023): 1022. http://dx.doi.org/10.3390/electronics12041022.
Testo completoBeurer-Kellner, Luca, Marc Fischer e Martin Vechev. "Prompting Is Programming: A Query Language for Large Language Models". Proceedings of the ACM on Programming Languages 7, PLDI (6 giugno 2023): 1946–69. http://dx.doi.org/10.1145/3591300.
Testo completoTesi sul tema "Large language model"
Jiang, Yuandong. "Large Scale Distributed Semantic N-gram Language Model". Wright State University / OhioLINK, 2011. http://rave.ohiolink.edu/etdc/view?acc_num=wright1316200173.
Testo completoTang, Haijiang. "Building phrase based language model from large corpus /". View Abstract or Full-Text, 2002. http://library.ust.hk/cgi/db/thesis.pl?ELEC%202002%20TANG.
Testo completoIncludes bibliographical references (leaves 74-79). Also available in electronic version. Access restricted to campus users.
McGreevy, Michael. "Statistical language modelling for large vocabulary speech recognition". Thesis, Queensland University of Technology, 2006. https://eprints.qut.edu.au/16444/1/Michael_McGreevy_Thesis.pdf.
Testo completoMcGreevy, Michael. "Statistical language modelling for large vocabulary speech recognition". Queensland University of Technology, 2006. http://eprints.qut.edu.au/16444/.
Testo completoTan, Ming. "A Large Scale Distributed Syntactic, Semantic and Lexical Language Model for Machine Translation". Wright State University / OhioLINK, 2013. http://rave.ohiolink.edu/etdc/view?acc_num=wright1386111950.
Testo completoSusman, Derya. "Turkish Large Vocabulary Continuous Speech Recognition By Using Limited Audio Corpus". Master's thesis, METU, 2012. http://etd.lib.metu.edu.tr/upload/12614207/index.pdf.
Testo completoComez, Murat Ali. "Large Vocabulary Continuous Speech Recogniton For Turkish Using Htk". Master's thesis, METU, 2003. http://etd.lib.metu.edu.tr/upload/1205491/index.pdf.
Testo completoSagen, Markus. "Large-Context Question Answering with Cross-Lingual Transfer". Thesis, Uppsala universitet, Institutionen för informationsteknologi, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-440704.
Testo completoUzelac, Lawrence Stevan. "A Multiple Coupled Microstrip Transmission Line Model for High-Speed VLSI Interconnect Simulation". PDXScholar, 1991. https://pdxscholar.library.pdx.edu/open_access_etds/4526.
Testo completoLabeau, Matthieu. "Neural language models : Dealing with large vocabularies". Thesis, Université Paris-Saclay (ComUE), 2018. http://www.theses.fr/2018SACLS313/document.
Testo completoThis work investigates practical methods to ease training and improve performances of neural language models with large vocabularies. The main limitation of neural language models is their expensive computational cost: it depends on the size of the vocabulary, with which it grows linearly. Despite several training tricks, the most straightforward way to limit computation time is to limit the vocabulary size, which is not a satisfactory solution for numerous tasks. Most of the existing methods used to train large-vocabulary language models revolve around avoiding the computation of the partition function, ensuring that output scores are normalized into a probability distribution. Here, we focus on sampling-based approaches, including importance sampling and noise contrastive estimation. These methods allow an approximate computation of the partition function. After examining the mechanism of self-normalization in noise-contrastive estimation, we first propose to improve its efficiency with solutions that are adapted to the inner workings of the method and experimentally show that they considerably ease training. Our second contribution is to expand on a generalization of several sampling based objectives as Bregman divergences, in order to experiment with new objectives. We use Beta divergences to derive a set of objectives from which noise contrastive estimation is a particular case. Finally, we aim at improving performances on full vocabulary language models, by augmenting output words representation with subwords. We experiment on a Czech dataset and show that using character-based representations besides word embeddings for output representations gives better results. We also show that reducing the size of the output look-up table improves results even more
Libri sul tema "Large language model"
Satō, Hideto. A data model, knowledge base, and natural language processing for sharing a large statistical database. Ibaraki, Osaka, Japan: Institute of Social and Economic Research, Osaka University, 1989.
Cerca il testo completoAmaratunga, Thimira. Understanding Large Language Models. Berkeley, CA: Apress, 2023. http://dx.doi.org/10.1007/979-8-8688-0017-7.
Testo completoKucharavy, Andrei, Octave Plancherel, Valentin Mulder, Alain Mermoud e Vincent Lenders, a cura di. Large Language Models in Cybersecurity. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7.
Testo completoTörnberg, Petter. How to Use Large-Language Models for Text Analysis. 1 Oliver’s Yard, 55 City Road, London EC1Y 1SP United Kingdom: SAGE Publications Ltd, 2024. http://dx.doi.org/10.4135/9781529683707.
Testo completoBashkatov, Alexander. Modeling in OpenSCAD: examples. ru: INFRA-M Academic Publishing LLC., 2019. http://dx.doi.org/10.12737/959073.
Testo completoBuild a Large Language Model (from Scratch). Manning Publications Co. LLC, 2024.
Cerca il testo completoGenerative AI with LangChain: Build Large Language Model Apps with Python, ChatGPT and Other LLMs. Packt Publishing, Limited, 2023.
Cerca il testo completoGenerative AI with LangChain: Build Large Language Model Apps with Python, ChatGPT, and Other LLMs. de Gruyter GmbH, Walter, 2023.
Cerca il testo completoLarge Language Model-Based Solutions: How to Deliver Value with Cost-Effective Generative AI Applications. Wiley & Sons, Limited, John, 2024.
Cerca il testo completoLarge Language Model-Based Solutions: How to Deliver Value with Cost-Effective Generative AI Applications. Wiley & Sons, Incorporated, John, 2024.
Cerca il testo completoCapitoli di libri sul tema "Large language model"
Wu, Yonghui. "Large Language Model and Text Generation". In Cognitive Informatics in Biomedicine and Healthcare, 265–97. Cham: Springer International Publishing, 2024. http://dx.doi.org/10.1007/978-3-031-55865-8_10.
Testo completoRuiu, Dragos. "LLMs Red Teaming". In Large Language Models in Cybersecurity, 213–23. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7_24.
Testo completoKucharavy, Andrei. "Overview of Existing LLM Families". In Large Language Models in Cybersecurity, 31–44. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7_3.
Testo completoDolamic, Ljiljana. "Conversational Agents". In Large Language Models in Cybersecurity, 45–53. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7_4.
Testo completoKucharavy, Andrei. "Adapting LLMs to Downstream Applications". In Large Language Models in Cybersecurity, 19–29. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7_2.
Testo completoSchillaci, Zachary. "On-Site Deployment of LLMs". In Large Language Models in Cybersecurity, 205–11. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7_23.
Testo completoKurimo, Mikko, e Krista Lagus. "An Efficiently Focusing Large Vocabulary Language Model". In Artificial Neural Networks — ICANN 2002, 1068–73. Berlin, Heidelberg: Springer Berlin Heidelberg, 2002. http://dx.doi.org/10.1007/3-540-46084-5_173.
Testo completoJi, Jianchao, Zelong Li, Shuyuan Xu, Wenyue Hua, Yingqiang Ge, Juntao Tan e Yongfeng Zhang. "GenRec: Large Language Model for Generative Recommendation". In Lecture Notes in Computer Science, 494–502. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-56063-7_42.
Testo completoMajumdar, Subhabrata, e Terry Vogelsang. "Towards Safe LLMs Integration". In Large Language Models in Cybersecurity, 243–47. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7_27.
Testo completoMajumdar, Subhabrata. "Standards for LLM Security". In Large Language Models in Cybersecurity, 225–31. Cham: Springer Nature Switzerland, 2024. http://dx.doi.org/10.1007/978-3-031-54827-7_25.
Testo completoAtti di convegni sul tema "Large language model"
Huang, Jiaji, Yi Li, Wei Ping e Liang Huang. "Large Margin Neural Language Model". In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. Stroudsburg, PA, USA: Association for Computational Linguistics, 2018. http://dx.doi.org/10.18653/v1/d18-1150.
Testo completoChen, Kua, Yujing Yang, Boqi Chen, José Antonio Hernández López, Gunter Mussbacher e Dániel Varró. "Automated Domain Modeling with Large Language Models: A Comparative Study". In 2023 ACM/IEEE 26th International Conference on Model Driven Engineering Languages and Systems (MODELS). IEEE, 2023. http://dx.doi.org/10.1109/models58315.2023.00037.
Testo completoMeng, Ruijie, Martin Mirchev, Marcel Böhme e Abhik Roychoudhury. "Large Language Model guided Protocol Fuzzing". In Network and Distributed System Security Symposium. Reston, VA: Internet Society, 2024. http://dx.doi.org/10.14722/ndss.2024.24556.
Testo completoHASHIMOTO, Tomomi. "Ethical Judgment using Large Language Model". In 2024 16th International Conference on Computer and Automation Engineering (ICCAE). IEEE, 2024. http://dx.doi.org/10.1109/iccae59995.2024.10569797.
Testo completoSingh, Aditi, Saket Kumar, Abul Ehtesham, Tala Talaei Khoei e Deepshikha Bhati. "Large Language Model-Driven Immersive Agent". In 2024 IEEE World AI IoT Congress (AIIoT). IEEE, 2024. http://dx.doi.org/10.1109/aiiot61789.2024.10578948.
Testo completoGalindo, José A., Antonio J. Dominguez, Jules White e David Benavides. "Large Language Models to generate meaningful feature model instances". In SPLC '23: 27th ACM International Systems and Software Product Line Conference. New York, NY, USA: ACM, 2023. http://dx.doi.org/10.1145/3579027.3608973.
Testo completoStammbach, Dominik, Vilém Zouhar, Alexander Hoyle, Mrinmaya Sachan e Elliott Ash. "Revisiting Automated Topic Model Evaluation with Large Language Models". In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing. Stroudsburg, PA, USA: Association for Computational Linguistics, 2023. http://dx.doi.org/10.18653/v1/2023.emnlp-main.581.
Testo completoZhao, James, Yuxi Xie, Kenji Kawaguchi, Junxian He e Michael Xie. "Automatic Model Selection with Large Language Models for Reasoning". In Findings of the Association for Computational Linguistics: EMNLP 2023. Stroudsburg, PA, USA: Association for Computational Linguistics, 2023. http://dx.doi.org/10.18653/v1/2023.findings-emnlp.55.
Testo completoXu, Austin, Will Monroe e Klinton Bicknell. "Large language model augmented exercise retrieval for personalized language learning". In LAK '24: The 14th Learning Analytics and Knowledge Conference. New York, NY, USA: ACM, 2024. http://dx.doi.org/10.1145/3636555.3636883.
Testo completoMysore, Sheshera, Andrew Mccallum e Hamed Zamani. "Large Language Model Augmented Narrative Driven Recommendations". In RecSys '23: Seventeenth ACM Conference on Recommender Systems. New York, NY, USA: ACM, 2023. http://dx.doi.org/10.1145/3604915.3608829.
Testo completoRapporti di organizzazioni sul tema "Large language model"
Seymore, Kristie, e Ronald Rosenfeld. Large-Scale Topic Detection and Language Model Adaptation. Fort Belvoir, VA: Defense Technical Information Center, giugno 1997. http://dx.doi.org/10.21236/ada327553.
Testo completoZhang, Hao. Large Language Model (LLM) Monthly Report (2024 Apr). ResearchHub Technologies, Inc., maggio 2024. http://dx.doi.org/10.55277/researchhub.0ps6xenm.
Testo completoSun, Ruiqi, e Daniel Trefler. The Impact of AI and Cross-Border Data Regulation on International Trade in Digital Services: A Large Language Model. Cambridge, MA: National Bureau of Economic Research, novembre 2023. http://dx.doi.org/10.3386/w31925.
Testo completoLavadenz, Magaly, Sheila Cassidy, Elvira G. Armas, Rachel Salivar, Grecya V. Lopez e Amanda A. Ross. Sobrato Early Academic Language (SEAL) Model: Final Report of Findings from a Four-Year Study. Center for Equity for English Learners, Loyola Marymount University, 2020. http://dx.doi.org/10.15365/ceel.seal2020.
Testo completoPrasad, Jayanti. Large Language Models: AI Foundations and Applications in Python. Instats Inc., 2023. http://dx.doi.org/10.61700/85rfezw01y0q9521.
Testo completoAlonso-Robisco, Andres, e Jose Manuel Carbo. Analysis of CBDC Narrative OF Central Banks using Large Language Models. Madrid: Banco de España, agosto 2023. http://dx.doi.org/10.53479/33412.
Testo completoMarra de Artiñano, Ignacio, Franco Riottini Depetris e Christian Volpe Martincus. Automatic Product Classification in International Trade: Machine Learning and Large Language Models. Inter-American Development Bank, luglio 2023. http://dx.doi.org/10.18235/0005012.
Testo completoWindsor, Callan, e Max Zang. Firms' Price-setting Behaviour: Insights from Earnings Calls. Reserve Bank of Australia, settembre 2023. http://dx.doi.org/10.47688/rdp2023-06.
Testo completoHorton, John. Large Language Models as Simulated Economic Agents: What Can We Learn from Homo Silicus? Cambridge, MA: National Bureau of Economic Research, aprile 2023. http://dx.doi.org/10.3386/w31122.
Testo completoGluckman, Peter, e Hema Sridhar. A framework for evaluating rapidly developing digital and related technologies: AI, Large Language Models and beyond. International Science Council, ottobre 2023. http://dx.doi.org/10.24948/2023.11.
Testo completo