Publications

A Strategic Markovian Traffic Equilibrium Model for Capacitated Networks
Maëlle Zimmermann
Patrice Marcotte
In the realm of traffic assignment over a network involving rigid arc capacities, the aim of the present work is to generalize the model of … (see more)Marcotte, Nguyen, and Schoeb [Marcotte P, Nguyen S, Schoeb A (2004) A strategic flow model of traffic assignment in static capacitated networks. Oper. Res. 52(2):191–212.] by casting it within a stochastic user equilibrium framework. The strength of the proposed model is to incorporate two sources of stochasticity stemming, respectively, from the users’ imperfect knowledge regarding arc costs (represented by a discrete choice model) and the probability of not accessing saturated arcs. Moreover, the arc-based formulation extends the Markovian traffic equilibrium model of Baillon and Cominetti [Baillon JB, Cominetti R ( 2008 ) Markovian traffic equilibrium. Math. Programming 111(1-2):33–56.] through the explicit consideration of capacities. This paper is restricted to the case of acyclic networks, for which we present solution algorithms and numerical experiments.
Deep Generative Models for Galaxy Image Simulations
François Lanusse
Rachel Mandelbaum
Chun-Liang Li
Peter Freeman
Barnabás Póczos
Image simulations are essential tools for preparing and validating the analysis of current and future wide-field optical surveys. However, t… (see more)he galaxy models used as the basis for these simulations are typically limited to simple parametric light profiles, or use a fairly limited amount of available space-based data. In this work, we propose a methodology based on Deep Generative Models to create complex models of galaxy morphologies that may meet the image simulation needs of upcoming surveys. We address the technical challenges associated with learning this morphology model from noisy and PSF-convolved images by building a hybrid Deep Learning/physical Bayesian hierarchical model for observed images, explicitly accounting for the Point Spread Function and noise properties. The generative model is further made conditional on physical galaxy parameters, to allow for sampling new light profiles from specific galaxy populations. We demonstrate our ability to train and sample from such a model on galaxy postage stamps from the HST/ACS COSMOS survey, and validate the quality of the model using a range of second- and higher-order morphology statistics. Using this set of statistics, we demonstrate significantly more realistic morphologies using these deep generative models compared to conventional parametric models. To help make these generative models practical tools for the community, we introduce GalSim-Hub, a community-driven repository of generative models, and a framework for incorporating generative models within the GalSim image simulation software.
Editorial: Social Interaction in Neuropsychiatry
Victoria Leong
Frieder M. Paulus
Kevin Pelphrey
Elizabeth Redcay
Leonhard Schilbach
What is Going on Inside Recurrent Meta Reinforcement Learning Agents?
Recurrent meta reinforcement learning (meta-RL) agents are agents that employ a recurrent neural network (RNN) for the purpose of"learning a… (see more) learning algorithm". After being trained on a pre-specified task distribution, the learned weights of the agent's RNN are said to implement an efficient learning algorithm through their activity dynamics, which allows the agent to quickly solve new tasks sampled from the same distribution. However, due to the black-box nature of these agents, the way in which they work is not yet fully understood. In this study, we shed light on the internal working mechanisms of these agents by reformulating the meta-RL problem using the Partially Observable Markov Decision Process (POMDP) framework. We hypothesize that the learned activity dynamics is acting as belief states for such agents. Several illustrative experiments suggest that this hypothesis is true, and that recurrent meta-RL agents can be viewed as agents that learn to act optimally in partially observable environments consisting of multiple related tasks. This view helps in understanding their failure cases and some interesting model-based results reported in the literature.
Recovering the Wedge Modes Lost to 21-cm Foregrounds
Samuel Gagnon-Hartman
Yue Cui
Adrian Liu
One of the critical challenges facing imaging studies of the 21-cm signal at the Epoch of Reionization (EoR) is the separation of astrophysi… (see more)cal foreground contamination. These foregrounds are known to lie in a wedge-shaped region of
Tracking white and grey matter degeneration along the spinal cord axis in degenerative cervical myelopathy
Kevin Vallotton
Gergely David
Markus Hupp
Nikolai Pfender
Michael Fehlings
Rebecca S. Samson
Claudia A. M. Gandini Wheeler-Kingshott
Armin Curt
Patrick Freund
Maryam Seif
Tissue-specific neurodegeneration revealed by quantitative MRI, already apparent across the spinal cord in mild-moderate DCM prior to the on… (see more)set of severe clinical impairments. WM microstructural changes are particularly sensitive to remote pathologically and clinically eloquent changes in DCM.
Gradient Masked Federated Optimization
hBERT + BiasCorp - Fighting Racism on the Web
Olawale Moses Onabola
Zhuang Ma
Xie Yang
Benjamin Akera
Jia Xue
Dianbo Liu
Subtle and overt racism is still present both in physical and online communities today and has impacted many lives in different segments of … (see more)the society. In this short piece of work, we present how we’re tackling this societal issue with Natural Language Processing. We are releasing BiasCorp, a dataset containing 139,090 comments and news segment from three specific sources - Fox News, BreitbartNews and YouTube. The first batch (45,000 manually annotated) is ready for publication. We are currently in the final phase of manually labeling the remaining dataset using Amazon Mechanical Turk. BERT has been used widely in several downstream tasks. In this work, we present hBERT, where we modify certain layers of the pretrained BERT model with the new Hopfield Layer. hBert generalizes well across different distributions with the added advantage of a reduced model complexity. We are also releasing a JavaScript library 3 and a Chrome Extension Application, to help developers make use of our trained model in web applications (say chat application) and for users to identify and report racially biased contents on the web respectively
INFOSHIELD: Generalizable Information-Theoretic Human-Trafficking Detection
Meng-Chieh Lee
Catalina Vajiac
Sacha Lévy
Namyong Park
Cara Jones
Christos Faloutsos
Given a million escort advertisements, how can we spot near-duplicates? Such micro-clusters of ads are usually signals of human trafficking.… (see more) How can we summarize them, visually, to convince law enforcement to act? Can we build a general tool that works for different languages? Spotting micro-clusters of near-duplicate documents is useful in multiple, additional settings, including spam-bot detection in Twitter ads, plagiarism, and more.We present INFOSHIELD, which makes the following contributions: (a) Practical, being scalable and effective on real data, (b) Parameter-free and Principled, requiring no user-defined parameters, (c) Interpretable, finding a document to be the cluster representative, highlighting all the common phrases, and automatically detecting "slots", i.e. phrases that differ in every document; and (d) Generalizable, beating or matching domain-specific methods in Twitter bot detection and human trafficking detection respectively, as well as being language-independent finding clusters in Spanish, Italian, and Japanese. Interpretability is particularly important for the anti human-trafficking domain, where law enforcement must visually inspect ads.Our experiments on real data show that INFOSHIELD correctly identifies Twitter bots with an F1 score over 90% and detects human-trafficking ads with 84% precision. Moreover, it is scalable, requiring about 8 hours for 4 million documents on a stock laptop.
The Surprising Performance of Simple Baselines for Misinformation Detection
As social media becomes increasingly prominent in our day to day lives, it is increasingly important to detect informative content and preve… (see more)nt the spread of disinformation and unverified rumours. While many sophisticated and successful models have been proposed in the literature, they are often compared with older NLP baselines such as SVMs, CNNs, and LSTMs. In this paper, we examine the performance of a broad set of modern transformer-based language models and show that with basic fine-tuning, these models are competitive with and can even significantly outperform recently proposed state-of-the-art methods. We present our framework as a baseline for creating and evaluating new methods for misinformation detection. We further study a comprehensive set of benchmark datasets, and discuss potential data leakage and the need for careful design of the experiments and understanding of datasets to account for confounding variables. As an extreme case example, we show that classifying only based on the first three digits of tweet ids, which contain information on the date, gives state-of-the-art performance on a commonly used benchmark dataset for fake news detection --Twitter16. We provide a simple tool to detect this problem and suggest steps to mitigate it in future datasets.
Ethics of Corporeal, Co-present Robots as Agents of Influence: a Review
H. V. D. Van der Loos
Towards Causal Federated Learning For Enhanced Robustness and Privacy
Federated Learning is an emerging privacy-preserving distributed machine learning approach to building a shared model by performing distribu… (see more)ted training locally on participating devices (clients) and aggregating the local models into a global one. As this approach prevents data collection and aggregation, it helps in reducing associated privacy risks to a great extent. However, the data samples across all participating clients are usually not independent and identically distributed (non-iid), and Out of Distribution(OOD) generalization for the learned models can be poor. Besides this challenge, federated learning also remains vulnerable to various attacks on security wherein a few malicious participating entities work towards inserting backdoors, degrading the generated aggregated model as well as inferring the data owned by participating entities. In this paper, we propose an approach for learning invariant (causal) features common to all participating clients in a federated learning setup and analyze empirically how it enhances the Out of Distribution (OOD) accuracy as well as the privacy of the final learned model.