Letteratura scientifica selezionata sul tema "Responsible Artificial Intelligence"
Cita una fonte nei formati APA, MLA, Chicago, Harvard e in molti altri stili
Consulta la lista di attuali articoli, libri, tesi, atti di convegni e altre fonti scientifiche attinenti al tema "Responsible Artificial Intelligence".
Accanto a ogni fonte nell'elenco di riferimenti c'è un pulsante "Aggiungi alla bibliografia". Premilo e genereremo automaticamente la citazione bibliografica dell'opera scelta nello stile citazionale di cui hai bisogno: APA, MLA, Harvard, Chicago, Vancouver ecc.
Puoi anche scaricare il testo completo della pubblicazione scientifica nel formato .pdf e leggere online l'abstract (il sommario) dell'opera se è presente nei metadati.
Articoli di riviste sul tema "Responsible Artificial Intelligence"
Tawei Wang, Tawei Wang. "Responsible Use of Artificial Intelligen". International Journal of Computer Auditing 4, n. 1 (dicembre 2022): 001–3. http://dx.doi.org/10.53106/256299802022120401001.
Testo completoGregor, Shirley. "Responsible Artificial Intelligence and Journal Publishing". Journal of the Association for Information Systems 25, n. 1 (2024): 48–60. http://dx.doi.org/10.17705/1jais.00863.
Testo completoTeng, C. L., A. S. Bhullar, P. Jermain, D. Jordon, R. Nawfel, P. Patel, R. Sean, M. Shang e D. H. Wu. "Responsible Artificial Intelligence in Radiation Oncology". International Journal of Radiation Oncology*Biology*Physics 120, n. 2 (ottobre 2024): e659. http://dx.doi.org/10.1016/j.ijrobp.2024.07.1446.
Testo completoHaidar, Ahmad. "An Integrative Theoretical Framework for Responsible Artificial Intelligence". International Journal of Digital Strategy, Governance, and Business Transformation 13, n. 1 (15 dicembre 2023): 1–23. http://dx.doi.org/10.4018/ijdsgbt.334844.
Testo completoShneiderman, Ben. "Responsible AI". Communications of the ACM 64, n. 8 (agosto 2021): 32–35. http://dx.doi.org/10.1145/3445973.
Testo completoDignum, Virginia. "Responsible Artificial Intelligence --- From Principles to Practice". ACM SIGIR Forum 56, n. 1 (giugno 2022): 1–6. http://dx.doi.org/10.1145/3582524.3582529.
Testo completoRodrigues, Rowena, Anais Resseguier e Nicole Santiago. "When Artificial Intelligence Fails". Public Governance, Administration and Finances Law Review 8, n. 2 (14 dicembre 2023): 17–28. http://dx.doi.org/10.53116/pgaflr.7030.
Testo completoVASYLKIVSKYI, Mikola, Ganna VARGATYUK e Olga BOLDYREVA. "INTELLIGENT RADIO INTERFACE WITH THE SUPPORT OF ARTIFICIAL INTELLIGENCE". Herald of Khmelnytskyi National University. Technical sciences 217, n. 1 (23 febbraio 2023): 26–32. http://dx.doi.org/10.31891/2307-5732-2023-317-1-26-32.
Testo completoGermanov, Nikolai S. "The concept of responsible artificial intelligence as the future of artificial intelligence in medicine". Digital Diagnostics 4, n. 1S (26 giugno 2023): 27–29. http://dx.doi.org/10.17816/dd430334.
Testo completoTyrranen, V. A. "ARTIFICIAL INTELLIGENCE CRIMES". Territory Development, n. 3(17) (2019): 10–13. http://dx.doi.org/10.32324/2412-8945-2019-3-10-13.
Testo completoTesi sul tema "Responsible Artificial Intelligence"
Svedberg, Peter O. S. "Steps towards an empirically responsible AI : a methodological and theoretical framework". Thesis, Norwegian University of Science and Technology, Department of Computer and Information Science, 2004. http://urn.kb.se/resolve?urn=urn:nbn:no:ntnu:diva-246.
Testo completoInitially we pursue a minimal model of a cognitive system. This in turn form the basis for the development of amethodological and theoretical framework. Two methodological requirements of the model are that explanation be from the perspective of the phenomena, and that we have structural determination. The minimal model is derived from the explanatory side of a biologically based cognitive science. Fransisco Varela is our principal source for this part. The model defines the relationship between a formally defined autonomous system and an environment, in such a way as to generate the world of the system, its actual environment. The minimal model is a modular explanation in that we find it on different levels in bio-cognitive systems, from the cell to small social groups. For the latter and for the role played by artefactual systems we bring in Edwin Hutchins' observational study of a cognitive system in action. This necessitates the introduction of a complementary form of explanation. A key aspect of Hutchins' findings is the social domain as environment for humans. Aspects of human cognitive abilities usually attributed to the person are more properly attributed to the social system, including artefactual systems.
Developing the methodological and theoretical framework means making a transition from the bio-cognitive to the computational. The two complementary forms of explanation are important for the ability to develop a methodology that supports the construction of actual systems. This has to be able to handle the transition from external determination of a system in design to internal determination (autonomy) in operation.
Once developed, the combined framework is evaluated in an application area. This is done by comparing the standard conception of the Semantic Web with how this notion looks from the perspective of the framework. This includes the development of the methodological framework as a metalevel external knowledge representation. A key difference between the two approaches is the directness by which the semantic is approached. Our perspective puts the focus on interaction and the structural regularities this engenders in the external representation. Regularities which in turn form the basis for machine processing. In this regard we see the relationship between representation and inference as analogous to the relationship between environment and system. Accordingly we have the social domain as environment for artefactual agents. For human level cognitive abilities the social domain as environment is important. We argue that a reasonable shortcut to systems we can relate to, about that very domain, is for artefactual agents to have an external representation of the social domain as environment.
Ounissi, Mehdi. "Decoding the Black Box : Enhancing Interpretability and Trust in Artificial Intelligence for Biomedical Imaging - a Step Toward Responsible Artificial Intelligence". Electronic Thesis or Diss., Sorbonne université, 2024. http://www.theses.fr/2024SORUS237.
Testo completoIn an era dominated by AI, its opaque decision-making --known as the "black box" problem-- poses significant challenges, especially in critical areas like biomedical imaging where accuracy and trust are crucial. Our research focuses on enhancing AI interpretability in biomedical applications. We have developed a framework for analyzing biomedical images that quantifies phagocytosis in neurodegenerative diseases using time-lapse phase-contrast video microscopy. Traditional methods often struggle with rapid cellular interactions and distinguishing cells from backgrounds, critical for studying conditions like frontotemporal dementia (FTD). Our scalable, real-time framework features an explainable cell segmentation module that simplifies deep learning algorithms, enhances interpretability, and maintains high performance by incorporating visual explanations and by model simplification. We also address issues in visual generative models, such as hallucinations in computational pathology, by using a unique encoder for Hematoxylin and Eosin staining coupled with multiple decoders. This method improves the accuracy and reliability of synthetic stain generation, employing innovative loss functions and regularization techniques that enhance performance and enable precise synthetic stains crucial for pathological analysis. Our methodologies have been validated against several public benchmarks, showing top-tier performance. Notably, our framework distinguished between mutant and control microglial cells in FTD, providing new biological insights into this unproven phenomenon. Additionally, we introduced a cloud-based system that integrates complex models and provides real-time feedback, facilitating broader adoption and iterative improvements through pathologist insights. The release of novel datasets, including video microscopy on microglial cell phagocytosis and a virtual staining dataset related to pediatric Crohn's disease, along with all source codes, underscores our commitment to transparent open scientific collaboration and advancement. Our research highlights the importance of interpretability in AI, advocating for technology that integrates seamlessly with user needs and ethical standards in healthcare. Enhanced interpretability allows researchers to better understand data and improve tool performance
Sugianto, Nehemia. "Responsible AI for Automated Analysis of Integrated Video Surveillance in Public Spaces". Thesis, Griffith University, 2021. http://hdl.handle.net/10072/409586.
Testo completoThesis (PhD Doctorate)
Doctor of Philosophy (PhD)
Dept Bus Strategy & Innovation
Griffith Business School
Full Text
Kessing, Maria. "Fairness in AI : Discussion of a Unified Approach to Ensure Responsible AI Development". Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-299936.
Testo completoUtöver de fördelar som AI-teknologier har bidragit med, så har även etiska dilemman och problem uppstått. På grund av ökat fokus, har ett stort antal förslag till system och regelverk som diskuterar ansvarstagande AI-utveckling publicerats sedan 2016. Denna rapport kommer analysera ett urval av dessa förslag med avsikt att besvara frågan (1) “Vilka tillvägagångssätt kan försäkra oss om en ansvarsfull AI-utveckling?” För att utforska denna fråga kommer denna rapport analysera olika metoder och tillvägagångssätt, på bland annat mellanstatliga- och statliga regelverk, forskningsgrupper samt privata företag. Dessutom har expertintervjuer genomförts för att besvara den andra problemformuleringen (2) “Hur kan vi nå en övergripande, gemensam, lösning för att försäkra oss om ansvarsfull AI-utveckling?” Denna rapport redogör för att statliga organisationer och myndigheter är den främsta drivkraften för att detta ska ske. Vidare krävs en detaljerad plan som knyter ihop forskningsgrupper med den offentliga- och privata sektorn. Slutligen anser rapporten även att det är av stor vikt för vidare utbildning när det kommer till att göra AI förklarbart och tydligt för alla.
Umurerwa, Janviere, e Maja Lesjak. "AI IMPLEMENTATION AND USAGE : A qualitative study of managerial challenges in implementation and use of AI solutions from the researchers’ perspective". Thesis, Umeå universitet, Institutionen för informatik, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-187810.
Testo completo"Responsible Governance of Artificial Intelligence: An Assessment, Theoretical Framework, and Exploration". Doctoral diss., 2019. http://hdl.handle.net/2286/R.I.55667.
Testo completoDissertation/Thesis
Doctoral Dissertation Human and Social Dimensions of Science and Technology 2019
Arienti, João Henrique Leal. "Time series forecasting applied to an energy management system ‐ A comparison between Deep Learning Models and other Machine Learning Models". Master's thesis, 2020. http://hdl.handle.net/10362/108172.
Testo completoA large amount of energy used by the world comes from buildings’ energy consumption. HVAC (Heat, Ventilation, and Air Conditioning) systems are the biggest offenders when it comes to buildings’ energy consumption. It is important to provide environmental comfort in buildings but indoor wellbeing is directly related to an increase in energy consumption. This dilemma creates a huge opportunity for a solution that balances occupant comfort and energy consumption. Within this context, the Ambiosensing project was launched to develop a complete energy management system that differentiates itself from other existing commercial solutions by being an inexpensive and intelligent system. The Ambiosensing project focused on the topic of Time Series Forecasting to achieve the goal of creating predictive models to help the energy management system to anticipate indoor environmental scenarios. A good approach for Time Series Forecasting problems is to apply Machine Learning, more specifically Deep Learning. This work project intends to investigate and develop Deep Learning and other Machine Learning models that can deal with multivariate Time Series Forecasting, to assess how well can a Deep Learning approach perform on a Time Series Forecasting problem, especially, LSTM (Long Short-Term Memory) Recurrent Neural Networks (RNN) and to establish a comparison between Deep Learning and other Machine Learning models like Linear Regression, Decision Trees, Random Forest, Gradient Boosting Machines and others within this context.
Voarino, Nathalie. "Systèmes d’intelligence artificielle et santé : les enjeux d’une innovation responsable". Thèse, 2019. http://hdl.handle.net/1866/23526.
Testo completoThe use of artificial intelligence (AI) systems in health is part of the advent of a new "high definition" medicine that is predictive, preventive and personalized, benefiting from the unprecedented amount of data that is today available. At the heart of digital health innovation, the development of AI systems promises to lead to an interconnected and self-learning healthcare system. AI systems could thus help to redefine the classification of diseases, generate new medical knowledge, or predict the health trajectories of individuals for prevention purposes. Today, various applications in healthcare are being considered, ranging from assistance to medical decision-making through expert systems to precision medicine (e.g. pharmacological targeting), as well as individualized prevention through health trajectories developed on the basis of biological markers. However, urgent ethical concerns emerge with the increasing use of algorithms to analyze a growing number of data related to health (often personal and sensitive) as well as the reduction of human intervention in many automated processes. From the limitations of big data analysis, the need for data sharing and the algorithmic decision ‘opacity’ stems various ethical concerns relating to the protection of privacy and intimacy, free and informed consent, social justice, dehumanization of care and patients, and/or security. To address these challenges, many initiatives have focused on defining and applying principles for an ethical governance of AI. However, the operationalization of these principles faces various difficulties inherent to applied ethics, which originate either from the scope (universal or plural) of these principles or the way these principles are put into practice (inductive or deductive methods). These issues can be addressed with context-specific or bottom-up approaches of applied ethics. However, people who embrace these approaches still face several challenges. From an analysis of citizens' fears and expectations emerging from the discussions that took place during the coconstruction of the Montreal Declaration for a Responsible Development of AI, it is possible to get a sense of what these difficulties look like. From this analysis, three main challenges emerge: the incapacitation of health professionals and patients, the many hands problem, and artificial agency. These challenges call for AI systems that empower people and that allow to maintain human agency, in order to foster the development of (pragmatic) shared responsibility among the various stakeholders involved in the development of healthcare AI systems. Meeting these challenges is essential in order to adapt existing governance mechanisms and enable the development of a responsible digital innovation in healthcare and research that allows human beings to remain at the center of its development.
Libri sul tema "Responsible Artificial Intelligence"
Dignum, Virginia. Responsible Artificial Intelligence. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-30371-6.
Testo completoSchmidpeter, René, e Reinhard Altenburger, a cura di. Responsible Artificial Intelligence. Cham: Springer International Publishing, 2023. http://dx.doi.org/10.1007/978-3-031-09245-9.
Testo completoKhoshnevisan, Mohammad. Artificial intelligence and responsive optimization. 2a ed. Phoenix: Xiquan, 2003.
Cerca il testo completoKhoshnevisan, Mohammad. Artificial intelligence and responsive optimization. Phoenix: Xiquan, 2003.
Cerca il testo completoKhamparia, Aditya, Deepak Gupta, Ashish Khanna e Valentina E. Balas, a cura di. Biomedical Data Analysis and Processing Using Explainable (XAI) and Responsive Artificial Intelligence (RAI). Singapore: Springer Singapore, 2022. http://dx.doi.org/10.1007/978-981-19-1476-8.
Testo completoResponsible Artificial Intelligence. Springer, 2020.
Cerca il testo completoAltenburger, Reinhard, e René Schmidpeter. Responsible Artificial Intelligence: Challenges for Sustainable Management. Springer International Publishing AG, 2022.
Cerca il testo completoResponsible Artificial Intelligence: Challenges for Sustainable Management. Springer International Publishing AG, 2024.
Cerca il testo completoKaplan, Jerry. Artificial Intelligence. Oxford University Press, 2016. http://dx.doi.org/10.1093/wentk/9780190602383.001.0001.
Testo completoKnowings, L. D. Ethical AI: Navigating the Future With Responsible Artificial Intelligence. Sandiver Publishing, 2024.
Cerca il testo completoCapitoli di libri sul tema "Responsible Artificial Intelligence"
Dignum, Virginia. "What Is Artificial Intelligence?" In Responsible Artificial Intelligence, 9–34. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-30371-6_2.
Testo completoLeopold, Helmut. "Mastering Trustful Artificial Intelligence". In Responsible Artificial Intelligence, 133–58. Cham: Springer International Publishing, 2023. http://dx.doi.org/10.1007/978-3-031-09245-9_6.
Testo completoAltenburger, Reinhard. "Artificial Intelligence: Management Challenges and Responsibility". In Responsible Artificial Intelligence, 1–8. Cham: Springer International Publishing, 2023. http://dx.doi.org/10.1007/978-3-031-09245-9_1.
Testo completoDignum, Virginia. "Introduction". In Responsible Artificial Intelligence, 1–7. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-30371-6_1.
Testo completoDignum, Virginia. "Ethical Decision-Making". In Responsible Artificial Intelligence, 35–46. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-30371-6_3.
Testo completoDignum, Virginia. "Taking Responsibility". In Responsible Artificial Intelligence, 47–69. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-30371-6_4.
Testo completoDignum, Virginia. "Can AI Systems Be Ethical?" In Responsible Artificial Intelligence, 71–92. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-30371-6_5.
Testo completoDignum, Virginia. "Ensuring Responsible AI in Practice". In Responsible Artificial Intelligence, 93–105. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-30371-6_6.
Testo completoDignum, Virginia. "Looking Further". In Responsible Artificial Intelligence, 107–20. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-30371-6_7.
Testo completoSchindler, Matthias, e Frederik Schmihing. "Technology Serves People: Democratising Analytics and AI in the BMW Production System". In Responsible Artificial Intelligence, 159–82. Cham: Springer International Publishing, 2023. http://dx.doi.org/10.1007/978-3-031-09245-9_7.
Testo completoAtti di convegni sul tema "Responsible Artificial Intelligence"
Herrera, Framcisco. "Responsible Artificial Intelligence Systems: From Trustworthiness to Governance". In 2024 Design, Automation & Test in Europe Conference & Exhibition (DATE), 1–2. IEEE, 2024. http://dx.doi.org/10.23919/date58400.2024.10546553.
Testo completoYeasin, Mohammed. "Keynote Speaker ICOM'24: Perspective on Convergence of Mechatronics and Artificial Intelligence in Responsible Innovation". In 2024 9th International Conference on Mechatronics Engineering (ICOM), XIV. IEEE, 2024. http://dx.doi.org/10.1109/icom61675.2024.10652385.
Testo completoDignum, Virginia. "Responsible Autonomy". In Twenty-Sixth International Joint Conference on Artificial Intelligence. California: International Joint Conferences on Artificial Intelligence Organization, 2017. http://dx.doi.org/10.24963/ijcai.2017/655.
Testo completoWang, Yichuan, Mengran Xiong e Hossein Olya. "Toward an Understanding of Responsible Artificial Intelligence Practices". In Hawaii International Conference on System Sciences. Hawaii International Conference on System Sciences, 2020. http://dx.doi.org/10.24251/hicss.2020.610.
Testo completoWang, Shoujin, Ninghao Liu, Xiuzhen Zhang, Yan Wang, Francesco Ricci e Bamshad Mobasher. "Data Science and Artificial Intelligence for Responsible Recommendations". In KDD '22: The 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. New York, NY, USA: ACM, 2022. http://dx.doi.org/10.1145/3534678.3542916.
Testo completoCalvo, Albert, Nil Ortiz, Alejandro Espinosa, Aleksandar Dimitrievikj, Ignasi Oliva, Jordi Guijarro e Shuaib Sidiqqi. "Safe AI: Ensuring Safe and Responsible Artificial Intelligence". In 2023 JNIC Cybersecurity Conference (JNIC). IEEE, 2023. http://dx.doi.org/10.23919/jnic58574.2023.10205749.
Testo completoDong, Tian, Shaofeng Li, Guoxing Chen, Minhui Xue, Haojin Zhu e Zhen Liu. "RAI2: Responsible Identity Audit Governing the Artificial Intelligence". In Network and Distributed System Security Symposium. Reston, VA: Internet Society, 2023. http://dx.doi.org/10.14722/ndss.2023.241012.
Testo completoTahaei, Mohammad, Marios Constantinides, Daniele Quercia, Sean Kennedy, Michael Muller, Simone Stumpf, Q. Vera Liao et al. "Human-Centered Responsible Artificial Intelligence: Current & Future Trends". In CHI '23: CHI Conference on Human Factors in Computing Systems. New York, NY, USA: ACM, 2023. http://dx.doi.org/10.1145/3544549.3583178.
Testo completoIliadis, Eduard. "AI-GFA: Applied Framework for Producing Responsible Artificial Intelligence". In GoodIT '24: International Conference on Information Technology for Social Good, 93–99. New York, NY, USA: ACM, 2024. http://dx.doi.org/10.1145/3677525.3678646.
Testo completoMistar, Oussama. "Roles of Social Actors in Creating Responsible Artificial Intelligence". In CEEeGov 2024: Central and Eastern European eDem and eGov Days 2024, 252–57. New York, NY, USA: ACM, 2024. http://dx.doi.org/10.1145/3670243.3672437.
Testo completoRapporti di organizzazioni sul tema "Responsible Artificial Intelligence"
Stanley-Lockman, Zoe. Responsible and Ethical Military AI. Center for Security and Emerging Technology, agosto 2021. http://dx.doi.org/10.51593/20200091.
Testo completoLehoux, Pascale, Hassane Alami, Carl Mörch, Lysanne Rivard, Robson Rocha e Hudson Silva. Can we innovate responsibly during a pandemic? Artificial intelligence, digital solutions and SARS-CoV-2. Observatoire international sur les impacts sociétaux de l’intelligence artificielle et du numérique, giugno 2020. http://dx.doi.org/10.61737/ueti5496.
Testo completoNarayanan, Mina, e Christian Schoeberl. A Matrix for Selecting Responsible AI Frameworks. Center for Security and Emerging Technology, giugno 2023. http://dx.doi.org/10.51593/20220029.
Testo completoBurstein, Jill. Duolingo English Test Responsible AI Standards. Duolingo, marzo 2023. http://dx.doi.org/10.46999/vcae5025.
Testo completoFaveri, Benjamin, e Graeme Auld. nforming Possible Futures for the use of Third-Party Audits in AI Regulations. Regulatory Governance Initiative, Carleton University, novembre 2023. http://dx.doi.org/10.22215/sppa-rgi-nov2023.
Testo completoGoode, Kayla, Heeu Millie Kim e Melissa Deng. Examining Singapore’s AI Progress. Center for Security and Emerging Technology, marzo 2023. http://dx.doi.org/10.51593/2021ca014.
Testo completoTabassi, Elham. AI Risk Management Framework. Gaithersburg, MD: National Institute of Standards and Technology, 2023. http://dx.doi.org/10.6028/nist.ai.100-1.
Testo completoToney, Autumn, e Emelia Probasco. Who Cares About Trust? Center for Security and Emerging Technology, luglio 2023. http://dx.doi.org/10.51593/20230014b.
Testo completoGautrais, Vincent, e Nicolas Aubin. Assessment Model of Factors Relating to Data Flow: Instrument for the Protection of Privacy as well as Rights and Freedoms in the Development and Use of Artificial Intelligence. Observatoire international sur les impacts sociétaux de l'intelligence artificielle et du numérique, marzo 2022. http://dx.doi.org/10.61737/haoj6662.
Testo completoDaudelin, Francois, Lina Taing, Lucy Chen, Claudia Abreu Lopes, Adeniyi Francis Fagbamigbe e Hamid Mehmood. Mapping WASH-related disease risk: A review of risk concepts and methods. United Nations University Institute for Water, Environment and Health, dicembre 2021. http://dx.doi.org/10.53328/uxuo4751.
Testo completo