Portrait de Tegan Maharaj

Tegan Maharaj

Membre académique principal
Professeure adjointe en apprentissage automatique, HEC Montréal, Département de sciences de la décision
Sujets de recherche
Apprentissage de représentations
Apprentissage multimodal
Apprentissage profond
Systèmes dynamiques
Théorie de l'apprentissage automatique

Biographie

Je suis professeure adjointe au Département de sciences de la décision à HEC Montréal. Mes objectifs de recherche sont de contribuer à la compréhension et aux techniques de la science du développement responsable de l’IA, tout en appliquant utilement l’IA à des problèmes écologiques à fort impact liés au changement climatique, à l’épidémiologie, à l’alignement de l’IA et à l’évaluation des impacts écologiques. Mes travaux récents portent sur deux thèmes : l’utilisation de modèles profonds pour l’analyse des politiques et l’atténuation des risques; et la conception de données ou d’environnements de tests unitaires pour évaluer empiriquement le comportement d’apprentissage ou simuler le déploiement d’un système d’IA. N’hésitez pas à me contacter pour toute collaboration dans ces domaines.

Je suis généralement intéressée par l’étude de ce qui « entre » dans les modèles profonds : non seulement les données, mais l’environnement d’apprentissage plus globalement, qui comprend la conception/spécification des tâches, la fonction de perte et la régularisation, ainsi que le contexte sociétal du déploiement, notamment les considérations de confidentialité, les tendances et les incitatifs, les normes et les préjugés humains. Je suis préoccupée et passionnée par l’éthique de l’IA, la sécurité et l’application de l’apprentissage machine à la gestion de l’environnement, à la santé et au bien-être social.

Étudiants actuels

Maîtrise recherche - UdeM
Superviseur⋅e principal⋅e :
Maîtrise recherche - HEC
Superviseur⋅e principal⋅e :
Maîtrise recherche - HEC
Doctorat - HEC

Publications

Noisy Pairing and Partial Supervision for Stylized Opinion Summarization
Reinald Kim
Mirella Lapata. 2020
Un-611
David M. Krueger
Maxinder S. Kan-620
Somnath Basu
Roy Chowdhury
Chao Zhao
Tanya Goyal
Junyi Jiacheng Xu
Jessy Li
Ivor W. Tsang
James T. Kwok
Neil Houlsby
Andrei Giurgiu
Stanisław Jastrzębski … (voir 22 de plus)
Bruna Morrone
Quentin de Laroussilhe
Mona Gesmundo
Attariyan Sylvain
Gelly
Thomas Wolf
Lysandre Debut
Julien Victor Sanh
Clement Chaumond
Anthony Delangue
Pier-339 Moi
Tim ric Cistac
R´emi Rault
Morgan Louf
Funtow-900 Joe
Sam Davison
Patrick Shleifer
Von Platen
Clara Ma
Yacine Jernite
Julien Plu
Canwen Xu
Opinion summarization research has primar-001 ily focused on generating summaries reflect-002 ing important opinions from customer reviews 0… (voir plus)03 without paying much attention to the writing 004 style. In this paper, we propose the stylized 005 opinion summarization task, which aims to 006 generate a summary of customer reviews in 007 the desired (e.g., professional) writing style. 008 To tackle the difficulty in collecting customer 009 and professional review pairs, we develop a 010 non-parallel training framework, Noisy Pair-011 ing and Partial Supervision ( NAPA ), which 012 trains a stylized opinion summarization sys-013 tem from non-parallel customer and profes-014 sional review sets. We create a benchmark P RO - 015 S UM by collecting customer and professional 016 reviews from Yelp and Michelin. Experimental 017 results on P RO S UM and FewSum demonstrate 018 that our non-parallel training framework con-019 sistently improves both automatic and human 020 evaluations, successfully building a stylized 021 opinion summarization model that can gener-022 ate professionally-written summaries from cus-023 tomer reviews. 024
Generalizing in the Real World with Representation Learning
Tackling Climate Change with Machine Learning
Priya L. Donti
Lynn H. Kaack
Kelly Kochanski
Alexandre Lacoste
Andrew Slavin Ross
Nikola Milojevic-Dupont
Natasha Jaques
Anna Waldman-Brown
Alexandra Luccioni
Evan D. Sherwin
S. Karthik Mukkavilli
Konrad P. Kording
Carla Gomes
Andrew Y. Ng
Demis Hassabis
John C. Platt
Felix Creutzig … (voir 2 de plus)
Jennifer Chayes
Climate change is one of the greatest challenges facing humanity, and we, as machine learning experts, may wonder how we can help. Here we d… (voir plus)escribe how machine learning can be a powerful tool in reducing greenhouse gas emissions and helping society adapt to a changing climate. From smart grids to disaster management, we identify high impact problems where existing gaps can be filled by machine learning, in collaboration with other fields. Our recommendations encompass exciting research questions as well as promising business opportunities. We call on the machine learning community to join the global effort against climate change.
Deep Learning for Detecting Extreme Weather Patterns
Mayur Mudigonda
Mayur Mudigonda, Prabhat Ram
Prabhat Ram
Karthik Kashinath
Evan Racah
Ankur Mahesh
Yunjie Liu
Jim Biard
Thorsten Kurth
Sookyung Kim
S Ebrahimi Kahou
Burlen Loring
Christopher Pal
Travis O'Brien
K. Kunkel
Kenneth E. Kunkel
M. Wehner
Michael F. Wehner … (voir 2 de plus)
W. Collins
William D. Collins
Predicting Infectiousness for Proactive Contact Tracing
Prateek Gupta
Nasim Rahaman
Pierre-Luc St. Charles
Hannah Alsdurf
Gaétan Marceau-Caron
Pierre-Luc Carrier
Joumana Ghosn
Bernhard Schölkopf … (voir 3 de plus)
Abhinav Sharma
The COVID-19 pandemic has spread rapidly worldwide, overwhelming manual contact tracing in many countries and resulting in widespread lockdo… (voir plus)wns for emergency containment. Large-scale digital contact tracing (DCT) has emerged as a potential solution to resume economic and social activity while minimizing spread of the virus. Various DCT methods have been proposed, each making trade-offs between privacy, mobility restrictions, and public health. The most common approach, binary contact tracing (BCT), models infection as a binary event, informed only by an individual's test results, with corresponding binary recommendations that either all or none of the individual's contacts quarantine. BCT ignores the inherent uncertainty in contacts and the infection process, which could be used to tailor messaging to high-risk individuals, and prompt proactive testing or earlier warnings. It also does not make use of observations such as symptoms or pre-existing medical conditions, which could be used to make more accurate infectiousness predictions. In this paper, we use a recently-proposed COVID-19 epidemiological simulator to develop and test methods that can be deployed to a smartphone to locally and proactively predict an individual's infectiousness (risk of infecting others) based on their contact history and other information, while respecting strong privacy constraints. Predictions are used to provide personalized recommendations to the individual via an app, as well as to send anonymized messages to the individual's contacts, who use this information to better predict their own infectiousness, an approach we call proactive contact tracing (PCT). We find a deep-learning based PCT method which improves over BCT for equivalent average mobility, suggesting PCT could help in safe re-opening and second-wave prevention.
COVI-AgentSim: an Agent-based Model for Evaluating Methods of Digital Contact Tracing
Prateek Gupta
Nasim Rahaman
Hannah Alsdurf
Abhinav Sharma
Nanor Minoyan
Soren Harnois Leblanc
Pierre-Luc St. Charles
Akshay Patel
Joumana Ghosn
Yang Zhang
Bernhard Schölkopf
Christopher Pal
Joanna Merckx
The rapid global spread of COVID-19 has led to an unprecedented demand for effective methods to mitigate the spread of the disease, and vari… (voir plus)ous digital contact tracing (DCT) methods have emerged as a component of the solution. In order to make informed public health choices, there is a need for tools which allow evaluation and comparison of DCT methods. We introduce an agent-based compartmental simulator we call COVI-AgentSim, integrating detailed consideration of virology, disease progression, social contact networks, and mobility patterns, based on parameters derived from empirical research. We verify by comparing to real data that COVI-AgentSim is able to reproduce realistic COVID-19 spread dynamics, and perform a sensitivity analysis to verify that the relative performance of contact tracing methods are consistent across a range of settings. We use COVI-AgentSim to perform cost-benefit analyses comparing no DCT to: 1) standard binary contact tracing (BCT) that assigns binary recommendations based on binary test results; and 2) a rule-based method for feature-based contact tracing (FCT) that assigns a graded level of recommendation based on diverse individual features. We find all DCT methods consistently reduce the spread of the disease, and that the advantage of FCT over BCT is maintained over a wide range of adoption rates. Feature-based methods of contact tracing avert more disability-adjusted life years (DALYs) per socioeconomic cost (measured by productive hours lost). Our results suggest any DCT method can help save lives, support re-opening of economies, and prevent second-wave outbreaks, and that FCT methods are a promising direction for enriching BCT using self-reported symptoms, yielding earlier warning signals and a significantly reduced spread of the virus per socioeconomic cost.
COVI White Paper - Version 1.1
Hannah Alsdurf
Prateek Gupta
Daphne Ippolito
Richard Janda
Max Jarvies
Tyler Kolody
Sekoul Krastev
Robert Obryk
Dan Pilat
Nasim Rahaman
Jean-François Rousseau
Abhinav Sharma
Brooke Struck … (voir 3 de plus)
Yun William Yu
The SARS-CoV-2 (Covid-19) pandemic has caused significant strain on public health institutions around the world. Contact tracing is an essen… (voir plus)tial tool to change the course of the Covid-19 pandemic. Manual contact tracing of Covid-19 cases has significant challenges that limit the ability of public health authorities to minimize community infections. Personalized peer-to-peer contact tracing through the use of mobile apps has the potential to shift the paradigm. Some countries have deployed centralized tracking systems, but more privacy-protecting decentralized systems offer much of the same benefit without concentrating data in the hands of a state authority or for-profit corporations. Machine learning methods can circumvent some of the limitations of standard digital tracing by incorporating many clues and their uncertainty into a more graded and precise estimation of infection risk. The estimated risk can provide early risk awareness, personalized recommendations and relevant information to the user. Finally, non-identifying risk data can inform epidemiological models trained jointly with the machine learning predictor. These models can provide statistical evidence for the importance of factors involved in disease transmission. They can also be used to monitor, evaluate and optimize health policy and (de)confinement scenarios according to medical and economic productivity indicators. However, such a strategy based on mobile apps and machine learning should proactively mitigate potential ethical and privacy risks, which could have substantial impacts on society (not only impacts on health but also impacts such as stigmatization and abuse of personal data). Here, we present an overview of the rationale, design, ethical considerations and privacy strategy of `COVI,' a Covid-19 public peer-to-peer contact tracing and risk awareness mobile application developed in Canada.
COVI White Paper-Version 1.1
H. Alsdurf
T. Deleu
Prateek Gupta
Daphne Ippolito
R. Janda
Max Jarvie
Tyler Kolody
S. Krastev
Robert Obryk
D. Pilat
Nasim Rahaman
I. Rish
J. Rousseau
Abhinav Sharma
B. Struck … (voir 3 de plus)
Yun William Yu
Toward Trustworthy AI Development: Mechanisms for Supporting Verifiable Claims
Miles Brundage
Shahar Avin
Haydn Belfield
Gretchen Krueger
Gillian Hadfield
Heidy Khlaaf
Jingying Yang
Helen Toner
Ruth Fong
Pang Wei Koh
Sara Hooker
Jade Leung
Andrew Trask
Emma Bluemke
Cullen O'Keefe
Mark Koren
Théo Ryffel … (voir 39 de plus)
JB Rubinovitz
Tamay Besiroglu
Federica Carugati
Jack Clark
Peter Eckersley
Sarah de Haas
Maritza Johnson
Ben Laurie
Alex Ingerman
Igor Krawczuk
Amanda Askell
Rosario Cammarota
Andrew Lohn
David Krueger
Charlotte Stix
Logan Graham
Carina Prunkl
Bianca Martin
Elizabeth Seger
Noa Zilberman
Seán Ó hÉigeartaigh
Frens Kroeger
Girish Sastry
Rebecca Kagan
Adrian Weller
Brian Tse
Elizabeth Barnes
Allan Dafoe
Paul Scharre
Ariel Herbert-Voss
Martijn Rasser
Carrick Flynn
Thomas Krendl Gilbert
Lisa Dyer
Saif Khan
Markus Anderljung
With the recent wave of progress in artificial intelligence (AI) has come a growing awareness of the large-scale impacts of AI systems, and … (voir plus)recognition that existing regulations and norms in industry and academia are insufficient to ensure responsible AI development. In order for AI developers to earn trust from system users, customers, civil society, governments, and other stakeholders that they are building AI responsibly, they will need to make verifiable claims to which they can be held accountable. Those outside of a given organization also need effective means of scrutinizing such claims. This report suggests various steps that different stakeholders can take to improve the verifiability of claims made about AI systems and their associated development processes, with a focus on providing evidence about the safety, security, fairness, and privacy protection of AI systems. We analyze ten mechanisms for this purpose--spanning institutions, software, and hardware--and make recommendations aimed at implementing, exploring, or improving those mechanisms.
Deep Learning recognizes weather and climate patterns
Karthik Kashinath
M. Prabhat
Mayur Mudigonda
Ankur Mahesh
Sookyung Kim
Yunjie Liu
S Ebrahimi Kahou
B. Toms
Evan Racah
Christopher Pal
Jim Biard
K. Kunkel
Dean Nesbit Williams
Travis O'Brien
M. Wehner
W. Collins
A dataset and exploration of models for understanding video data through fill-in-the-blank question-answering
Anna Rohrbach
Christopher Pal
While deep convolutional neural networks frequently approach or exceed human-level performance at benchmark tasks involving static images, e… (voir plus)xtending this success to moving images is not straightforward. Having models which can learn to understand video is of interest for many applications, including content recommendation, prediction, summarization, event/object detection and understanding human visual perception, but many domains lack sufficient data to explore and perfect video models. In order to address the need for a simple, quantitative benchmark for developing and understanding video, we present MovieFIB, a fill-in-the-blank question-answering dataset with over 300,000 examples, based on descriptive video annotations for the visually impaired. In addition to presenting statistics and a description of the dataset, we perform a detailed analysis of 5 different models' predictions, and compare these with human performance. We investigate the relative importance of language, static (2D) visual features, and moving (3D) visual features; the effects of increasing dataset size, the number of frames sampled; and of vocabulary size. We illustrate that: this task is not solvable by a language model alone; our model combining 2D and 3D visual information indeed provides the best result; all models perform significantly worse than human-level. We provide human evaluations for responses given by different models and find that accuracy on the MovieFIB evaluation corresponds well with human judgement. We suggest avenues for improving video models, and hope that the proposed dataset can be useful for measuring and encouraging progress in this very interesting field.
A Closer Look at Memorization in Deep Networks
We examine the role of memorization in deep learning, drawing connections to capacity, generalization, and adversarial robustness. While dee… (voir plus)p networks are capable of memorizing noise data, our results suggest that they tend to prioritize learning simple patterns first. In our experiments, we expose qualitative differences in gradient-based optimization of deep neural networks (DNNs) on noise vs. real data. We also demonstrate that for appropriately tuned explicit regularization (e.g., dropout) we can degrade DNN training performance on noise datasets without compromising generalization on real data. Our analysis suggests that the notions of effective capacity which are dataset independent are unlikely to explain the generalization performance of deep networks when trained with gradient based methods because training data itself plays an important role in determining the degree of memorization.