Literatura científica selecionada sobre o tema "Responsible Artificial Intelligence"
Crie uma referência precisa em APA, MLA, Chicago, Harvard, e outros estilos
Consulte a lista de atuais artigos, livros, teses, anais de congressos e outras fontes científicas relevantes para o tema "Responsible Artificial Intelligence".
Ao lado de cada fonte na lista de referências, há um botão "Adicionar à bibliografia". Clique e geraremos automaticamente a citação bibliográfica do trabalho escolhido no estilo de citação de que você precisa: APA, MLA, Harvard, Chicago, Vancouver, etc.
Você também pode baixar o texto completo da publicação científica em formato .pdf e ler o resumo do trabalho online se estiver presente nos metadados.
Artigos de revistas sobre o assunto "Responsible Artificial Intelligence"
Tawei Wang, Tawei Wang. "Responsible Use of Artificial Intelligen". International Journal of Computer Auditing 4, n.º 1 (dezembro de 2022): 001–3. http://dx.doi.org/10.53106/256299802022120401001.
Texto completo da fonteGregor, Shirley. "Responsible Artificial Intelligence and Journal Publishing". Journal of the Association for Information Systems 25, n.º 1 (2024): 48–60. http://dx.doi.org/10.17705/1jais.00863.
Texto completo da fonteTeng, C. L., A. S. Bhullar, P. Jermain, D. Jordon, R. Nawfel, P. Patel, R. Sean, M. Shang e D. H. Wu. "Responsible Artificial Intelligence in Radiation Oncology". International Journal of Radiation Oncology*Biology*Physics 120, n.º 2 (outubro de 2024): e659. http://dx.doi.org/10.1016/j.ijrobp.2024.07.1446.
Texto completo da fonteHaidar, Ahmad. "An Integrative Theoretical Framework for Responsible Artificial Intelligence". International Journal of Digital Strategy, Governance, and Business Transformation 13, n.º 1 (15 de dezembro de 2023): 1–23. http://dx.doi.org/10.4018/ijdsgbt.334844.
Texto completo da fonteShneiderman, Ben. "Responsible AI". Communications of the ACM 64, n.º 8 (agosto de 2021): 32–35. http://dx.doi.org/10.1145/3445973.
Texto completo da fonteDignum, Virginia. "Responsible Artificial Intelligence --- From Principles to Practice". ACM SIGIR Forum 56, n.º 1 (junho de 2022): 1–6. http://dx.doi.org/10.1145/3582524.3582529.
Texto completo da fonteRodrigues, Rowena, Anais Resseguier e Nicole Santiago. "When Artificial Intelligence Fails". Public Governance, Administration and Finances Law Review 8, n.º 2 (14 de dezembro de 2023): 17–28. http://dx.doi.org/10.53116/pgaflr.7030.
Texto completo da fonteVASYLKIVSKYI, Mikola, Ganna VARGATYUK e Olga BOLDYREVA. "INTELLIGENT RADIO INTERFACE WITH THE SUPPORT OF ARTIFICIAL INTELLIGENCE". Herald of Khmelnytskyi National University. Technical sciences 217, n.º 1 (23 de fevereiro de 2023): 26–32. http://dx.doi.org/10.31891/2307-5732-2023-317-1-26-32.
Texto completo da fonteGermanov, Nikolai S. "The concept of responsible artificial intelligence as the future of artificial intelligence in medicine". Digital Diagnostics 4, n.º 1S (26 de junho de 2023): 27–29. http://dx.doi.org/10.17816/dd430334.
Texto completo da fonteTyrranen, V. A. "ARTIFICIAL INTELLIGENCE CRIMES". Territory Development, n.º 3(17) (2019): 10–13. http://dx.doi.org/10.32324/2412-8945-2019-3-10-13.
Texto completo da fonteTeses / dissertações sobre o assunto "Responsible Artificial Intelligence"
Svedberg, Peter O. S. "Steps towards an empirically responsible AI : a methodological and theoretical framework". Thesis, Norwegian University of Science and Technology, Department of Computer and Information Science, 2004. http://urn.kb.se/resolve?urn=urn:nbn:no:ntnu:diva-246.
Texto completo da fonteInitially we pursue a minimal model of a cognitive system. This in turn form the basis for the development of amethodological and theoretical framework. Two methodological requirements of the model are that explanation be from the perspective of the phenomena, and that we have structural determination. The minimal model is derived from the explanatory side of a biologically based cognitive science. Fransisco Varela is our principal source for this part. The model defines the relationship between a formally defined autonomous system and an environment, in such a way as to generate the world of the system, its actual environment. The minimal model is a modular explanation in that we find it on different levels in bio-cognitive systems, from the cell to small social groups. For the latter and for the role played by artefactual systems we bring in Edwin Hutchins' observational study of a cognitive system in action. This necessitates the introduction of a complementary form of explanation. A key aspect of Hutchins' findings is the social domain as environment for humans. Aspects of human cognitive abilities usually attributed to the person are more properly attributed to the social system, including artefactual systems.
Developing the methodological and theoretical framework means making a transition from the bio-cognitive to the computational. The two complementary forms of explanation are important for the ability to develop a methodology that supports the construction of actual systems. This has to be able to handle the transition from external determination of a system in design to internal determination (autonomy) in operation.
Once developed, the combined framework is evaluated in an application area. This is done by comparing the standard conception of the Semantic Web with how this notion looks from the perspective of the framework. This includes the development of the methodological framework as a metalevel external knowledge representation. A key difference between the two approaches is the directness by which the semantic is approached. Our perspective puts the focus on interaction and the structural regularities this engenders in the external representation. Regularities which in turn form the basis for machine processing. In this regard we see the relationship between representation and inference as analogous to the relationship between environment and system. Accordingly we have the social domain as environment for artefactual agents. For human level cognitive abilities the social domain as environment is important. We argue that a reasonable shortcut to systems we can relate to, about that very domain, is for artefactual agents to have an external representation of the social domain as environment.
Ounissi, Mehdi. "Decoding the Black Box : Enhancing Interpretability and Trust in Artificial Intelligence for Biomedical Imaging - a Step Toward Responsible Artificial Intelligence". Electronic Thesis or Diss., Sorbonne université, 2024. http://www.theses.fr/2024SORUS237.
Texto completo da fonteIn an era dominated by AI, its opaque decision-making --known as the "black box" problem-- poses significant challenges, especially in critical areas like biomedical imaging where accuracy and trust are crucial. Our research focuses on enhancing AI interpretability in biomedical applications. We have developed a framework for analyzing biomedical images that quantifies phagocytosis in neurodegenerative diseases using time-lapse phase-contrast video microscopy. Traditional methods often struggle with rapid cellular interactions and distinguishing cells from backgrounds, critical for studying conditions like frontotemporal dementia (FTD). Our scalable, real-time framework features an explainable cell segmentation module that simplifies deep learning algorithms, enhances interpretability, and maintains high performance by incorporating visual explanations and by model simplification. We also address issues in visual generative models, such as hallucinations in computational pathology, by using a unique encoder for Hematoxylin and Eosin staining coupled with multiple decoders. This method improves the accuracy and reliability of synthetic stain generation, employing innovative loss functions and regularization techniques that enhance performance and enable precise synthetic stains crucial for pathological analysis. Our methodologies have been validated against several public benchmarks, showing top-tier performance. Notably, our framework distinguished between mutant and control microglial cells in FTD, providing new biological insights into this unproven phenomenon. Additionally, we introduced a cloud-based system that integrates complex models and provides real-time feedback, facilitating broader adoption and iterative improvements through pathologist insights. The release of novel datasets, including video microscopy on microglial cell phagocytosis and a virtual staining dataset related to pediatric Crohn's disease, along with all source codes, underscores our commitment to transparent open scientific collaboration and advancement. Our research highlights the importance of interpretability in AI, advocating for technology that integrates seamlessly with user needs and ethical standards in healthcare. Enhanced interpretability allows researchers to better understand data and improve tool performance
Sugianto, Nehemia. "Responsible AI for Automated Analysis of Integrated Video Surveillance in Public Spaces". Thesis, Griffith University, 2021. http://hdl.handle.net/10072/409586.
Texto completo da fonteThesis (PhD Doctorate)
Doctor of Philosophy (PhD)
Dept Bus Strategy & Innovation
Griffith Business School
Full Text
Kessing, Maria. "Fairness in AI : Discussion of a Unified Approach to Ensure Responsible AI Development". Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-299936.
Texto completo da fonteUtöver de fördelar som AI-teknologier har bidragit med, så har även etiska dilemman och problem uppstått. På grund av ökat fokus, har ett stort antal förslag till system och regelverk som diskuterar ansvarstagande AI-utveckling publicerats sedan 2016. Denna rapport kommer analysera ett urval av dessa förslag med avsikt att besvara frågan (1) “Vilka tillvägagångssätt kan försäkra oss om en ansvarsfull AI-utveckling?” För att utforska denna fråga kommer denna rapport analysera olika metoder och tillvägagångssätt, på bland annat mellanstatliga- och statliga regelverk, forskningsgrupper samt privata företag. Dessutom har expertintervjuer genomförts för att besvara den andra problemformuleringen (2) “Hur kan vi nå en övergripande, gemensam, lösning för att försäkra oss om ansvarsfull AI-utveckling?” Denna rapport redogör för att statliga organisationer och myndigheter är den främsta drivkraften för att detta ska ske. Vidare krävs en detaljerad plan som knyter ihop forskningsgrupper med den offentliga- och privata sektorn. Slutligen anser rapporten även att det är av stor vikt för vidare utbildning när det kommer till att göra AI förklarbart och tydligt för alla.
Umurerwa, Janviere, e Maja Lesjak. "AI IMPLEMENTATION AND USAGE : A qualitative study of managerial challenges in implementation and use of AI solutions from the researchers’ perspective". Thesis, Umeå universitet, Institutionen för informatik, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-187810.
Texto completo da fonte"Responsible Governance of Artificial Intelligence: An Assessment, Theoretical Framework, and Exploration". Doctoral diss., 2019. http://hdl.handle.net/2286/R.I.55667.
Texto completo da fonteDissertation/Thesis
Doctoral Dissertation Human and Social Dimensions of Science and Technology 2019
Arienti, João Henrique Leal. "Time series forecasting applied to an energy management system ‐ A comparison between Deep Learning Models and other Machine Learning Models". Master's thesis, 2020. http://hdl.handle.net/10362/108172.
Texto completo da fonteA large amount of energy used by the world comes from buildings’ energy consumption. HVAC (Heat, Ventilation, and Air Conditioning) systems are the biggest offenders when it comes to buildings’ energy consumption. It is important to provide environmental comfort in buildings but indoor wellbeing is directly related to an increase in energy consumption. This dilemma creates a huge opportunity for a solution that balances occupant comfort and energy consumption. Within this context, the Ambiosensing project was launched to develop a complete energy management system that differentiates itself from other existing commercial solutions by being an inexpensive and intelligent system. The Ambiosensing project focused on the topic of Time Series Forecasting to achieve the goal of creating predictive models to help the energy management system to anticipate indoor environmental scenarios. A good approach for Time Series Forecasting problems is to apply Machine Learning, more specifically Deep Learning. This work project intends to investigate and develop Deep Learning and other Machine Learning models that can deal with multivariate Time Series Forecasting, to assess how well can a Deep Learning approach perform on a Time Series Forecasting problem, especially, LSTM (Long Short-Term Memory) Recurrent Neural Networks (RNN) and to establish a comparison between Deep Learning and other Machine Learning models like Linear Regression, Decision Trees, Random Forest, Gradient Boosting Machines and others within this context.
Voarino, Nathalie. "Systèmes d’intelligence artificielle et santé : les enjeux d’une innovation responsable". Thèse, 2019. http://hdl.handle.net/1866/23526.
Texto completo da fonteThe use of artificial intelligence (AI) systems in health is part of the advent of a new "high definition" medicine that is predictive, preventive and personalized, benefiting from the unprecedented amount of data that is today available. At the heart of digital health innovation, the development of AI systems promises to lead to an interconnected and self-learning healthcare system. AI systems could thus help to redefine the classification of diseases, generate new medical knowledge, or predict the health trajectories of individuals for prevention purposes. Today, various applications in healthcare are being considered, ranging from assistance to medical decision-making through expert systems to precision medicine (e.g. pharmacological targeting), as well as individualized prevention through health trajectories developed on the basis of biological markers. However, urgent ethical concerns emerge with the increasing use of algorithms to analyze a growing number of data related to health (often personal and sensitive) as well as the reduction of human intervention in many automated processes. From the limitations of big data analysis, the need for data sharing and the algorithmic decision ‘opacity’ stems various ethical concerns relating to the protection of privacy and intimacy, free and informed consent, social justice, dehumanization of care and patients, and/or security. To address these challenges, many initiatives have focused on defining and applying principles for an ethical governance of AI. However, the operationalization of these principles faces various difficulties inherent to applied ethics, which originate either from the scope (universal or plural) of these principles or the way these principles are put into practice (inductive or deductive methods). These issues can be addressed with context-specific or bottom-up approaches of applied ethics. However, people who embrace these approaches still face several challenges. From an analysis of citizens' fears and expectations emerging from the discussions that took place during the coconstruction of the Montreal Declaration for a Responsible Development of AI, it is possible to get a sense of what these difficulties look like. From this analysis, three main challenges emerge: the incapacitation of health professionals and patients, the many hands problem, and artificial agency. These challenges call for AI systems that empower people and that allow to maintain human agency, in order to foster the development of (pragmatic) shared responsibility among the various stakeholders involved in the development of healthcare AI systems. Meeting these challenges is essential in order to adapt existing governance mechanisms and enable the development of a responsible digital innovation in healthcare and research that allows human beings to remain at the center of its development.
Livros sobre o assunto "Responsible Artificial Intelligence"
Dignum, Virginia. Responsible Artificial Intelligence. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-30371-6.
Texto completo da fonteSchmidpeter, René, e Reinhard Altenburger, eds. Responsible Artificial Intelligence. Cham: Springer International Publishing, 2023. http://dx.doi.org/10.1007/978-3-031-09245-9.
Texto completo da fonteKhoshnevisan, Mohammad. Artificial intelligence and responsive optimization. 2a ed. Phoenix: Xiquan, 2003.
Encontre o texto completo da fonteKhoshnevisan, Mohammad. Artificial intelligence and responsive optimization. Phoenix: Xiquan, 2003.
Encontre o texto completo da fonteKhamparia, Aditya, Deepak Gupta, Ashish Khanna e Valentina E. Balas, eds. Biomedical Data Analysis and Processing Using Explainable (XAI) and Responsive Artificial Intelligence (RAI). Singapore: Springer Singapore, 2022. http://dx.doi.org/10.1007/978-981-19-1476-8.
Texto completo da fonteResponsible Artificial Intelligence. Springer, 2020.
Encontre o texto completo da fonteAltenburger, Reinhard, e René Schmidpeter. Responsible Artificial Intelligence: Challenges for Sustainable Management. Springer International Publishing AG, 2022.
Encontre o texto completo da fonteResponsible Artificial Intelligence: Challenges for Sustainable Management. Springer International Publishing AG, 2024.
Encontre o texto completo da fonteKaplan, Jerry. Artificial Intelligence. Oxford University Press, 2016. http://dx.doi.org/10.1093/wentk/9780190602383.001.0001.
Texto completo da fonteKnowings, L. D. Ethical AI: Navigating the Future With Responsible Artificial Intelligence. Sandiver Publishing, 2024.
Encontre o texto completo da fonteCapítulos de livros sobre o assunto "Responsible Artificial Intelligence"
Dignum, Virginia. "What Is Artificial Intelligence?" In Responsible Artificial Intelligence, 9–34. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-30371-6_2.
Texto completo da fonteLeopold, Helmut. "Mastering Trustful Artificial Intelligence". In Responsible Artificial Intelligence, 133–58. Cham: Springer International Publishing, 2023. http://dx.doi.org/10.1007/978-3-031-09245-9_6.
Texto completo da fonteAltenburger, Reinhard. "Artificial Intelligence: Management Challenges and Responsibility". In Responsible Artificial Intelligence, 1–8. Cham: Springer International Publishing, 2023. http://dx.doi.org/10.1007/978-3-031-09245-9_1.
Texto completo da fonteDignum, Virginia. "Introduction". In Responsible Artificial Intelligence, 1–7. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-30371-6_1.
Texto completo da fonteDignum, Virginia. "Ethical Decision-Making". In Responsible Artificial Intelligence, 35–46. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-30371-6_3.
Texto completo da fonteDignum, Virginia. "Taking Responsibility". In Responsible Artificial Intelligence, 47–69. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-30371-6_4.
Texto completo da fonteDignum, Virginia. "Can AI Systems Be Ethical?" In Responsible Artificial Intelligence, 71–92. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-30371-6_5.
Texto completo da fonteDignum, Virginia. "Ensuring Responsible AI in Practice". In Responsible Artificial Intelligence, 93–105. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-30371-6_6.
Texto completo da fonteDignum, Virginia. "Looking Further". In Responsible Artificial Intelligence, 107–20. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-30371-6_7.
Texto completo da fonteSchindler, Matthias, e Frederik Schmihing. "Technology Serves People: Democratising Analytics and AI in the BMW Production System". In Responsible Artificial Intelligence, 159–82. Cham: Springer International Publishing, 2023. http://dx.doi.org/10.1007/978-3-031-09245-9_7.
Texto completo da fonteTrabalhos de conferências sobre o assunto "Responsible Artificial Intelligence"
Herrera, Framcisco. "Responsible Artificial Intelligence Systems: From Trustworthiness to Governance". In 2024 Design, Automation & Test in Europe Conference & Exhibition (DATE), 1–2. IEEE, 2024. http://dx.doi.org/10.23919/date58400.2024.10546553.
Texto completo da fonteYeasin, Mohammed. "Keynote Speaker ICOM'24: Perspective on Convergence of Mechatronics and Artificial Intelligence in Responsible Innovation". In 2024 9th International Conference on Mechatronics Engineering (ICOM), XIV. IEEE, 2024. http://dx.doi.org/10.1109/icom61675.2024.10652385.
Texto completo da fonteDignum, Virginia. "Responsible Autonomy". In Twenty-Sixth International Joint Conference on Artificial Intelligence. California: International Joint Conferences on Artificial Intelligence Organization, 2017. http://dx.doi.org/10.24963/ijcai.2017/655.
Texto completo da fonteWang, Yichuan, Mengran Xiong e Hossein Olya. "Toward an Understanding of Responsible Artificial Intelligence Practices". In Hawaii International Conference on System Sciences. Hawaii International Conference on System Sciences, 2020. http://dx.doi.org/10.24251/hicss.2020.610.
Texto completo da fonteWang, Shoujin, Ninghao Liu, Xiuzhen Zhang, Yan Wang, Francesco Ricci e Bamshad Mobasher. "Data Science and Artificial Intelligence for Responsible Recommendations". In KDD '22: The 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. New York, NY, USA: ACM, 2022. http://dx.doi.org/10.1145/3534678.3542916.
Texto completo da fonteCalvo, Albert, Nil Ortiz, Alejandro Espinosa, Aleksandar Dimitrievikj, Ignasi Oliva, Jordi Guijarro e Shuaib Sidiqqi. "Safe AI: Ensuring Safe and Responsible Artificial Intelligence". In 2023 JNIC Cybersecurity Conference (JNIC). IEEE, 2023. http://dx.doi.org/10.23919/jnic58574.2023.10205749.
Texto completo da fonteDong, Tian, Shaofeng Li, Guoxing Chen, Minhui Xue, Haojin Zhu e Zhen Liu. "RAI2: Responsible Identity Audit Governing the Artificial Intelligence". In Network and Distributed System Security Symposium. Reston, VA: Internet Society, 2023. http://dx.doi.org/10.14722/ndss.2023.241012.
Texto completo da fonteTahaei, Mohammad, Marios Constantinides, Daniele Quercia, Sean Kennedy, Michael Muller, Simone Stumpf, Q. Vera Liao et al. "Human-Centered Responsible Artificial Intelligence: Current & Future Trends". In CHI '23: CHI Conference on Human Factors in Computing Systems. New York, NY, USA: ACM, 2023. http://dx.doi.org/10.1145/3544549.3583178.
Texto completo da fonteIliadis, Eduard. "AI-GFA: Applied Framework for Producing Responsible Artificial Intelligence". In GoodIT '24: International Conference on Information Technology for Social Good, 93–99. New York, NY, USA: ACM, 2024. http://dx.doi.org/10.1145/3677525.3678646.
Texto completo da fonteMistar, Oussama. "Roles of Social Actors in Creating Responsible Artificial Intelligence". In CEEeGov 2024: Central and Eastern European eDem and eGov Days 2024, 252–57. New York, NY, USA: ACM, 2024. http://dx.doi.org/10.1145/3670243.3672437.
Texto completo da fonteRelatórios de organizações sobre o assunto "Responsible Artificial Intelligence"
Stanley-Lockman, Zoe. Responsible and Ethical Military AI. Center for Security and Emerging Technology, agosto de 2021. http://dx.doi.org/10.51593/20200091.
Texto completo da fonteLehoux, Pascale, Hassane Alami, Carl Mörch, Lysanne Rivard, Robson Rocha e Hudson Silva. Can we innovate responsibly during a pandemic? Artificial intelligence, digital solutions and SARS-CoV-2. Observatoire international sur les impacts sociétaux de l’intelligence artificielle et du numérique, junho de 2020. http://dx.doi.org/10.61737/ueti5496.
Texto completo da fonteNarayanan, Mina, e Christian Schoeberl. A Matrix for Selecting Responsible AI Frameworks. Center for Security and Emerging Technology, junho de 2023. http://dx.doi.org/10.51593/20220029.
Texto completo da fonteBurstein, Jill. Duolingo English Test Responsible AI Standards. Duolingo, março de 2023. http://dx.doi.org/10.46999/vcae5025.
Texto completo da fonteFaveri, Benjamin, e Graeme Auld. nforming Possible Futures for the use of Third-Party Audits in AI Regulations. Regulatory Governance Initiative, Carleton University, novembro de 2023. http://dx.doi.org/10.22215/sppa-rgi-nov2023.
Texto completo da fonteGoode, Kayla, Heeu Millie Kim e Melissa Deng. Examining Singapore’s AI Progress. Center for Security and Emerging Technology, março de 2023. http://dx.doi.org/10.51593/2021ca014.
Texto completo da fonteTabassi, Elham. AI Risk Management Framework. Gaithersburg, MD: National Institute of Standards and Technology, 2023. http://dx.doi.org/10.6028/nist.ai.100-1.
Texto completo da fonteToney, Autumn, e Emelia Probasco. Who Cares About Trust? Center for Security and Emerging Technology, julho de 2023. http://dx.doi.org/10.51593/20230014b.
Texto completo da fonteGautrais, Vincent, e Nicolas Aubin. Assessment Model of Factors Relating to Data Flow: Instrument for the Protection of Privacy as well as Rights and Freedoms in the Development and Use of Artificial Intelligence. Observatoire international sur les impacts sociétaux de l'intelligence artificielle et du numérique, março de 2022. http://dx.doi.org/10.61737/haoj6662.
Texto completo da fonteDaudelin, Francois, Lina Taing, Lucy Chen, Claudia Abreu Lopes, Adeniyi Francis Fagbamigbe e Hamid Mehmood. Mapping WASH-related disease risk: A review of risk concepts and methods. United Nations University Institute for Water, Environment and Health, dezembro de 2021. http://dx.doi.org/10.53328/uxuo4751.
Texto completo da fonte