Portrait de AJung Moon

AJung Moon

Membre académique principal
Professeure adjointe, McGill University, Département de génie électrique et informatique

Biographie

AJung Moon est une roboticienne expérimentale. Elle étudie comment les robots et les systèmes d'intelligence artificielle influencent la façon dont les gens se déplacent, se comportent et prennent des décisions, afin de déterminer comment nous pouvons concevoir et déployer de tels systèmes intelligents autonomes de manière plus responsable.

À l'Université McGill, elle est directrice du laboratoire McGill Responsible Autonomy & Intelligent System Ethics (RAISE). Le laboratoire RAISE est un groupe interdisciplinaire qui étudie les implications sociales et éthiques des robots et des systèmes d'intelligence artificielle, et explore ce que cela signifie pour les ingénieurs de concevoir et de déployer ces systèmes de manière responsable pour un avenir technologique meilleur.

Étudiants actuels

Stagiaire de recherche - McGill University
Collaborateur·rice de recherche
Collaborateur·rice de recherche - McGill University
Maîtrise recherche - McGill University
Co-superviseur⋅e :
Stagiaire de recherche - McGill University
Maîtrise recherche - McGill University
Collaborateur·rice de recherche
Doctorat - McGill University
Maîtrise recherche - McGill University
Superviseur⋅e principal⋅e :
Maîtrise recherche - McGill University
Superviseur⋅e principal⋅e :
Maîtrise recherche - McGill University

Publications

The Role of Robotics in Achieving the United Nations Sustainable Development Goals - The Experts' Meeting at the 2021 IEEE/RSJ IROS Workshop [Industry Activities]
Vincent Mai
Bram Vanderborght
Tamás P. Haidegger
Alaa M. Khamis
Niraj Bhargava
Dominik B. O. Boesl
K. Gabriels
An Jacobs
R. Murphy
Yasushi Nakauchi
Edson Prestes
Bhavani Rao R.
Ricardo Vinuesa
Carl-Maria Mörch
Sociotechnical Harms: Scoping a Taxonomy for Harm Reduction
Renee Shelby
Shalaleh Rismani
Kathryn Henne
Paul Nicholas
N'mah Fodiatu Yilla
Jess Gallegos
Andrew J Smart
Emilio Garcia
Gurleen Virk
What does it mean to be an AI Ethicist: An ontology of existing roles
Shalaleh Rismani
With the increasing adoption of Artificial Intelligence systems (AIS) in various application and the growing efforts to regulate such system… (voir plus)s, a new set of occupations has emerged in the industry. This new set of roles take different titles and hold varying responsibilities. However, the individuals in these roles are tasked with interpreting and operationalizing best practices for developing ethical and safe AI systems. We will broadly refer to this new set of occupations as AI ethicists and recognize that they often hold a specific role in the intersection of technology development, business needs, and societal implications. In this work, we examine what it means to be an AI ethicist in the industry and propose an ontology of existing roles under this broad title along with their required competencies. We create this ontology by examining the job postings for such roles over the past two years and conduct expert interviews with fourteen individuals who currently hold such a role in the industry. The proposed ontology will inform executives and leaders who are looking to build responsible AI teams and provide educators the necessary information for creating new learning objectives and curriculum.
What does it mean to be an AI Ethicist: An ontology of existing roles
Shalaleh Rismani
With the increasing adoption of Artificial Intelligence systems (AIS) in various application and the growing efforts to regulate such system… (voir plus)s, a new set of occupations has emerged in the industry. This new set of roles take different titles and hold varying responsibilities. However, the individuals in these roles are tasked with interpreting and operationalizing best practices for developing ethical and safe AI systems. We will broadly refer to this new set of occupations as AI ethicists and recognize that they often hold a specific role in the intersection of technology development, business needs, and societal implications. In this work, we examine what it means to be an AI ethicist in the industry and propose an ontology of existing roles under this broad title along with their required competencies. We create this ontology by examining the job postings for such roles over the past two years and conduct expert interviews with fourteen individuals who currently hold such a role in the industry. The proposed ontology will inform executives and leaders who are looking to build responsible AI teams and provide educators the necessary information for creating new learning objectives and curriculum.
How do AI systems fail socially?: an engineering risk analysis approach
Shalaleh Rismani
Failure Mode and Effect Analysis (FMEA) has been used as an engineering risk assessment tool since 1949. FMEAs are effective in preemptively… (voir plus) identifying and addressing how a device or process might fail in operation and are often used in the design of high-risk technology applications such as military, automotive industry and medical devices. In this work, we explore whether FMEAs can serve as a risk assessment tool for machine learning practitioners, especially in deploying systems for high-risk applications (e.g. algorithms for recidivism assessment). In particular, we discuss how FMEAs can be used to identify social and ethical failures of Artificial Intelligent Systemss (AISs), recognizing that FMEAs have the potential to uncover a broader range of failures. We first propose a process for developing a Social FMEAs (So-FMEAs) by building on the existing FMEAs framework and a recently published definition of Social Failure Modes by Millar. We then demonstrate a simple proof-of-concept, So-FMEAs for the COMPAS algorithm, a risk assessment tool used by judges to make recidivism-related decisions for convicted individuals. Through this preliminary investigation, we illustrate how a traditional engineering risk management tool could be adapted for analyzing social and ethical failures of AIS. Engineers and designers of AISs can use this new approach to improve their system's design and perform due diligence with respect to potential ethical and social failures.
Design of Hesitation Gestures for Nonverbal Human-Robot Negotiation of Conflicts
Maneezhay Hashmi
H. F. Machiel Van Der Loos
Elizabeth A. Croft
Aude Billard
When the question of who should get access to a communal resource first is uncertain, people often negotiate via nonverbal communication to … (voir plus)resolve the conflict. What should a robot be programmed to do when such conflicts arise in Human-Robot Interaction? The answer to this question varies depending on the context of the situation. Learning from how humans use hesitation gestures to negotiate a solution in such conflict situations, we present a human-inspired design of nonverbal hesitation gestures that can be used for Human-Robot Negotiation. We extracted characteristic features of such negotiative hesitations humans use, and subsequently designed a trajectory generator (Negotiative Hesitation Generator) that can re-create the features in robot responses to conflicts. Our human-subjects experiment demonstrates the efficacy of the designed robot behaviour against non-negotiative stopping behaviour of a robot. With positive results from our human-robot interaction experiment, we provide a validated trajectory generator with which one can explore the dynamics of human-robot nonverbal negotiation of resource conflicts.
Ethics of Corporeal, Co-present Robots as Agents of Influence: a Review
Shalaleh Rismani
H. V. D. Van der Loos
Can Open Source Licenses Help Regulate Lethal Autonomous Weapons?
Cheng Lin
Lethal autonomous weapon systems (LAWS, ethal autonomous weapon also known as killer robots) are a real and emerging technology that have th… (voir plus)e potential to radically transform warfare. Because of the myriad of moral, legal, privacy, and security risks the technology introduces, many scholars and advocates have called for a ban on the development, production, and use of fully autonomous weapons [1], [2].
Using Open Source Licensing to Regulate the Assembly of LAWS: A Preliminary Analysis
Cheng Lin
Lethal autonomous weapons (LAWS) are an emerging technology capable of automatically targeting and exercising lethal force. Many scholars an… (voir plus)d advocates have petitioned to ban the technology internationally for a myriad of reasons. However, there are practical challenges to implementing a ban. One such challenge is posed by the “intangible” nature of the software that LAWS depends on, which is incompatible with implementation mechanisms such as export control. Given the dual-use nature of software, and the fact that software is developed by teams of individuals, a number of soft governance mechanisms have been proposed to regulate this technology. In this paper, we investigate the feasibility of one particular approach: leveraging open source licenses as a means to prohibit the use of certain software in LAWS. This approach is largely motivated by the fact that open source software underpins all of technology, especially AI. Through a review of the recent tech activism and open source activism, we evaluate whether open source licenses can feasibly limit the use of open source software to only non-LAWS applications. We distill the current challenges facing “ethics-driven” open source licensing efforts into three main obstacles: the need for clarity of licensing language, the lack of enforceability of licenses, and the lack of cohesiveness of the open source community. We propose that addressing these factors are also success criteria for future anti-LAWS open source initiatives. We find that open source licenses provide more theoretical than practical promise in regulating LAWS, and conclude that cohesion in the open source community is the key to their potential practical success in the future.
Driver perceptions of advanced driver assistance systems and safety
Sophie Le Page
Jason Millar
Kelly Selina Bronson
Shalaleh Rismani
Advanced driver assistance systems (ADAS) are often used in the automotive industry to highlight innovative improvements in vehicle safety. … (voir plus)However, today it is unclear whether certain automation (e.g., adaptive cruise control, lane keeping, parking assist) increases safety of our roads. In this paper, we investigate driver awareness, use, perceived safety, knowledge, training, and attitudes toward ADAS with different automation systems/features. Results of our online survey (n=1018) reveal that there is a significant difference in frequency of use and perceived safety for different ADAS features. Furthermore, we find that at least 70% of drivers activate an ADAS feature"most or all of the time"when driving, yet we find that at least 40% of drivers report feeling that ADAS often compromises their safety when activated. We also find that most respondents learn how to use ADAS in their vehicles by trying it out on the road by themselves, rather than through any formal driver education and training. These results may mirror how certain ADAS features are often activated by default resulting in high usage rates. These results also suggest a lack of driver training and education for safely interacting with, and operating, ADAS, such as turning off systems/features. These findings contribute to a critical discussion about the overall safety implications of current ADAS, especially as they enable higher-level automation features to creep into personal vehicles without a lockstep response in training, regulation, and policy.
Drivers' Awareness, Knowledge, and Use of Autonomous Driving Assistance Systems (ADAS) and Vehicle Automation
Kelly Selina Bronson
Sophie Le Page
Katherine M. Robinson
Shalaleh Rismani
Jason Millar
Advanced driver assistance systems (ADAS) technologies in vehicles (e.g. park assist, lane change assist, emergency braking, etc.), which ta… (voir plus)ke over parts of the driving task of human drivers, are advancing at a disruptive pace and hold the potential to deliver many benefits to society. However, public understanding of ADAS systems, and driver training and licensing for using them, are lagging behind the fast-paced technological development, which could raise safety issues or slow the deployment of ADAS, thus offsetting their potential benefits. There is, therefore, a need to investigate issues related to public perception of ADAS in order to develop appropriate policies and governance structures which support innovation, and result in the smooth deployment and acceptance of appropriate ADAS for society. In this work we perform a quantitative public survey to better understand how the public's awareness and knowledge of ADAS technologies in their vehicles correlate to their use or engagement of those technologies. We find that up to 67% of participants never or rarely use optional ADAS in their vehicles (e.g. adaptive cruise control), where women were less likely than men to use ADAS even though women reported more awareness of ADAS in their vehicles, better training, and more willingness to pay for ADAS. By performing this analysis we hope to raise awareness around the public perception of current state-of-the-art in ADAS technologies. We also hope to flag concerns that answers to these questions might raise for the regulatory agencies, and manufacturers, responsible for bringing these technologies to market.