Gotta Go Fast When Generating Data with Score-Based Models
Alexia Jolicoeur-Martineau
Fei Li
Rémi Piché-Taillefer
Tal Kachman
Score-based (denoising diffusion) generative models have recently gained a lot of success in generating realistic and diverse data. These ap… (voir plus)proaches define a forward diffusion process for transforming data to noise and generate data by reversing it (thereby going from noise to data). Unfortunately, current score-based models generate data very slowly due to the sheer number of score network evaluations required by numerical SDE solvers. In this work, we aim to accelerate this process by devising a more efficient SDE solver. Existing approaches rely on the Euler-Maruyama (EM) solver, which uses a fixed step size. We found that naively replacing it with other SDE solvers fares poorly - they either result in low-quality samples or become slower than EM. To get around this issue, we carefully devise an SDE solver with adaptive step sizes tailored to score-based generative models piece by piece. Our solver requires only two score function evaluations, rarely rejects samples, and leads to high-quality samples. Our approach generates data 2 to 10 times faster than EM while achieving better or equal sample quality. For high-resolution images, our method leads to significantly higher quality samples than all other methods tested. Our SDE solver has the benefit of requiring no step size tuning.
Noised Consistency Training for Text Summarization
J. Y. Liu
Qianren Mao
Hao Peng
Hongdong Zhu
Jian-Xin Li
Neural abstractive summarization methods often require large quantities of labeled training data. However, labeling large amounts of summari… (voir plus)zation data is often prohibitive due to time, financial, and expertise constraints, which has limited the usefulness of summarization systems to practical applications. In this paper, we argue that this limitation can be overcome by a semi-supervised approach: consistency training which is to leverage large amounts of unlabeled data to improve the performance of supervised learning over a small corpus. The consistency regularization semi-supervised learning can regularize model predictions to be invariant to small noise applied to input articles. By adding noised unlabeled corpus to help regularize consistency training, this framework obtains comparative performance without using the full dataset. In particular, we have verified that leveraging large amounts of unlabeled data decently improves the performance of supervised learning over an insufficient labeled dataset.
AndroidEnv: A Reinforcement Learning Platform for Android
Daniel Toyama
Philippe Hamel
Anita Gergely
Gheorghe Comanici
Amelia Glaese
Zafarali Ahmed
Tyler Jackson
Shibl Mourad
We introduce AndroidEnv, an open-source platform for Reinforcement Learning (RL) research built on top of the Android ecosystem. AndroidEnv … (voir plus)allows RL agents to interact with a wide variety of apps and services commonly used by humans through a universal touchscreen interface. Since agents train on a realistic simulation of an Android device, they have the potential to be deployed on real devices. In this report, we give an overview of the environment, highlighting the significant features it provides for research, and we present an empirical evaluation of some popular reinforcement learning agents on a set of tasks built on this platform.
Learning Brain Dynamics With Coupled Low-Dimensional Nonlinear Oscillators and Deep Recurrent Networks
Germán Abrevaya
Aleksandr Y. Aravkin
Peng Zheng
Jean-Christophe Gagnon-Audet
James Kozloski
Pablo Polosecki
David Cox
Silvina Ponce Dawson
Guillermo Cecchi
Many natural systems, especially biological ones, exhibit complex multivariate nonlinear dynamical behaviors that can be hard to capture by … (voir plus)linear autoregressive models. On the other hand, generic nonlinear models such as deep recurrent neural networks often require large amounts of training data, not always available in domains such as brain imaging; also, they often lack interpretability. Domain knowledge about the types of dynamics typically observed in such systems, such as a certain type of dynamical systems models, could complement purely data-driven techniques by providing a good prior. In this work, we consider a class of ordinary differential equation (ODE) models known as van der Pol (VDP) oscil lators and evaluate their ability to capture a low-dimensional representation of neural activity measured by different brain imaging modalities, such as calcium imaging (CaI) and fMRI, in different living organisms: larval zebrafish, rat, and human. We develop a novel and efficient approach to the nontrivial problem of parameters estimation for a network of coupled dynamical systems from multivariate data and demonstrate that the resulting VDP models are both accurate and interpretable, as VDP's coupling matrix reveals anatomically meaningful excitatory and inhibitory interactions across different brain subsystems. VDP outperforms linear autoregressive models (VAR) in terms of both the data fit accuracy and the quality of insight provided by the coupling matrices and often tends to generalize better to unseen data when predicting future brain activity, being comparable to and sometimes better than the recurrent neural networks (LSTMs). Finally, we demonstrate that our (generative) VDP model can also serve as a data-augmentation tool leading to marked improvements in predictive accuracy of recurrent neural networks. Thus, our work contributes to both basic and applied dimensions of neuroimaging: gaining scientific insights and improving brain-based predictive models, an area of potentially high practical importance in clinical diagnosis and neurotechnology.
Inferring global-scale temporal latent topics from news reports to predict public health interventions for COVID-19
Zhi Wen
Guido Powell
Imane Chafi
Y. K. Li
Generative lesion pattern decomposition of cognitive impairment after stroke
Anna K. Bonkhoff
Jae‐Sung Lim
Hee-Joon Bae
Nick A. Weaver
Hugo J. Kuijf
J. Matthijs Biesbroek
Natalia S Rost
Cognitive impairment is a frequent and disabling sequela of stroke. There is however incomplete understanding of how lesion topographies in … (voir plus)the left and right cerebral hemisphere brain interact to cause distinct cognitive deficits. We integrated machine learning and Bayesian hierarchical modeling to enable hemisphere-aware analysis of 1080 subacute ischemic stroke patients with deep profiling ∼3 months after stroke. We show relevance of the left hemisphere in the prediction of language and memory assessments, while global cognitive impairments were equally well predicted by lesion topographies from both sides. Damage to the hippocampal and occipital regions on the left were particularly informative about lost naming and memory function. Global cognitive impairment was predominantly linked to lesioned tissue in supramarginal and angular gyrus, the postcentral gyrus as well as the lateral occipital and opercular cortices of the left hemisphere. Hence, our analysis strategy uncovered that lesion patterns with unique hemispheric distributions are characteristic of how cognitive capacity is lost due to ischemic brain tissue damage.
Publisher Correction: The default network of the human brain is associated with perceived social isolation
R. Nathan Spreng
Emile Dimas
Laetitia Mwilambwe-Tshilobo
Alain Dagher
Philipp Koellinger
Gideon Nave
Anthony Ong
Julius M Kernbach
Thomas V. Wiecki
Tian Ge
Avram J. Holmes
B.T. Thomas Yeo
Gary R. Turner
Robin I. M. Dunbar
Artificial intelligence in nursing: Priorities and opportunities from an international invitational think‐tank of the Nursing and Artificial Intelligence Leadership Collaborative
Charlene Esteban Ronquillo
Laura‐Maria Peltonen
Lisiane Pruinelli
Charlene H Chu
Suzanne Bakken
Ana Beduschi
Kenrick Cato
Nicholas Hardiker
Alain Junger
Martin Michalowski
Rune Nyrup
Donald Nigel Reed
Tapio Salakoski
Sanna Salanterä
Nancy Walton
Patrick Weber
Thomas Wiegand
Maxim Topaz
Deep Discourse Analysis for Generating Personalized Feedback in Intelligent Tutor Systems
Matt Grenander
Robert Belfer
Ekaterina Kochmar
Iulian V. Serban
Franccois St-Hilaire
We explore creating automated, personalized feedback in an intelligent tutoring system (ITS). Our goal is to pinpoint correct and incorrect … (voir plus)concepts in student answers in order to achieve better student learning gains. Although automatic methods for providing personalized feedback exist, they do not explicitly inform students about which concepts in their answers are correct or incorrect. Our approach involves decomposing students answers using neural discourse segmentation and classification techniques. This decomposition yields a relational graph over all discourse units covered by the reference solutions and student answers. We use this inferred relational graph structure and a neural classifier to match student answers with reference solutions and generate personalized feedback. Although the process is completely automated and data-driven, the personalized feedback generated is highly contextual, domain-aware and effectively targets each student's misconceptions and knowledge gaps. We test our method in a dialogue-based ITS and demonstrate that our approach results in high-quality feedback and significantly improved student learning gains.
Deep Verifier Networks: Verification of Deep Discriminative Models with Deep Generative Models
Tong Che
Xiaofeng Liu
Site Li
Yubin Ge
Ruixiang ZHANG
Caiming Xiong
AI Safety is a major concern in many deep learning applications such as autonomous driving. Given a trained deep learning model, an importan… (voir plus)t natural problem is how to reliably verify the model's prediction. In this paper, we propose a novel framework --- deep verifier networks (DVN) to detect unreliable inputs or predictions of deep discriminative models, using separately trained deep generative models. Our proposed model is based on conditional variational auto-encoders with disentanglement constraints to separate the label information from the latent representation. We give both intuitive and theoretical justifications for the model. Our verifier network is trained independently with the prediction model, which eliminates the need of retraining the verifier network for a new model. We test the verifier network on both out-of-distribution detection and adversarial example detection problems, as well as anomaly detection problems in structured prediction tasks such as image caption generation. We achieve state-of-the-art results in all of these problems.
DIBS: Diversity inducing Information Bottleneck in Model Ensembles
Samarth Sinha
Homanga Bharadhwaj
Anirudh Goyal
Animesh Garg
Florian Shkurti
Individual Fairness in Kidney Exchange Programs
William St-Arnaud
Behrouz Babaki