Publications

ARGV: 3D genome structure exploration using augmented reality
Chrisostomos Drogaris
Yanlin Zhang
Éric Zhang
Elena Nazarova
Roman Sarrazin-Gendron
Sélik Wilhelm-Landry
Yan Cyr
Jacek Majewski
Jérôme Waldispühl
A long-context RNA foundation model for predicting transcriptome architecture
Ali Saberi
Benedict Choi
Sean Wang
Mohsen Naghipourfar
Arsham Mikaeili Namini
Vijay Ramani
Hamed S. Najafabadi
Hani Goodarzi
Linking DNA sequence to genomic function remains one of the grand challenges in genetics and genomics. Here, we combine large-scale single-m… (see more)olecule transcriptome sequencing of diverse cancer cell lines with cutting-edge machine learning to build LoRNASH, an RNA foundation model that learns how the nucleotide sequence of unspliced pre-mRNA dictates transcriptome architecture—the relative abundances and molecular structures of mRNA isoforms. Owing to its use of the StripedHyena architecture, LoRNASH handles extremely long sequence inputs (∼65 kilobase pairs), allowing for quantitative, zero-shot prediction of all aspects of transcriptome architecture, including isoform abundance, isoform structure, and the impact of DNA sequence variants on transcript structure and abundance. We anticipate that our public data release and proof-of-concept model will accelerate varying aspects of RNA biotechnology. More broadly, we envision the use of LoRNASH as a foundation for fine-tuning of any transcriptome-related downstream prediction task, including cell-type specific gene expression, splicing, and general RNA processing.
MeshUp: Multi-Target Mesh Deformation via Blended Score Distillation
Hyunwoo Kim
Itai Lang
Thibault Groueix
Vladimir Kim
Rana Hanocka
We propose MeshUp, a technique that deforms a 3D mesh towards multiple target concepts, and intuitively controls the region where each conce… (see more)pt is expressed. Conveniently, the concepts can be defined as either text queries, e.g.,"a dog"and"a turtle,"or inspirational images, and the local regions can be selected as any number of vertices on the mesh. We can effectively control the influence of the concepts and mix them together using a novel score distillation approach, referred to as the Blended Score Distillation (BSD). BSD operates on each attention layer of the denoising U-Net of a diffusion model as it extracts and injects the per-objective activations into a unified denoising pipeline from which the deformation gradients are calculated. To localize the expression of these activations, we create a probabilistic Region of Interest (ROI) map on the surface of the mesh, and turn it into 3D-consistent masks that we use to control the expression of these activations. We demonstrate the effectiveness of BSD empirically and show that it can deform various meshes towards multiple objectives. Our project page is at https://threedle.github.io/MeshUp.
Reactzyme: A Benchmark for Enzyme-Reaction Prediction
Chenqing Hua
Bozitao Zhong
Sitao Luan
Liang Hong
Shuangjia Zheng
Pushing the frontiers in climate modelling and analysis with machine learning
Veronika Eyring
William D. Collins
Pierre Gentine
Elizabeth A. Barnes
Marcelo Barreiro
Tom Beucler
Marc Bocquet
Christopher S. Bretherton
Hannah M. Christensen
Katherine Dagon
David John Gagne
David Hall
Dorit Hammerling
Stephan Hoyer
Fernando Iglesias-Suarez
Ignacio Lopez-Gomez
Marie C. McGraw
Gerald A. Meehl
Maria J. Molina
Claire Monteleoni … (see 9 more)
Juliane Mueller
Michael S. Pritchard
Jakob Runge
Philip Stier
Oliver Watt-Meyer
Katja Weigel
Rose Yu
Laure Zanna
INTREPPPID - An Orthologue-Informed Quintuplet Network for Cross-Species Prediction of Protein-Protein Interaction
Joseph Szymborski
An overwhelming majority of protein-protein interaction (PPI) studies are conducted in a select few model organisms largely due to constrain… (see more)ts in time and cost of the associated “wet lab” experiments. In silico PPI inference methods are ideal tools to overcome these limitations, but often struggle with cross-species predictions. We present INTREPPPID, a method which incorporates orthology data using a new “quintuplet” neural network, which is constructed with five parallel encoders with shared parameters. INTREPPPID incorporates both a PPI classification task and an orthologous locality task. The latter learns embeddings of orthologues that have small Euclidean distances between them and large distances between embeddings of all other proteins. INTREPPPID outperforms all other leading PPI inference methods tested on both the intra-species and cross-species tasks using strict evaluation datasets. We show that INTREPPPID’s orthologous locality loss increases performance because of the biological relevance of the orthologue data, and not due to some other specious aspect of the architecture. Finally, we introduce PPI.bio and PPI Origami, a web server interface for INTREPPPID and a software tool for creating strict evaluation datasets, respectively. Together, these two initiatives aim to make both the use and development of PPI inference tools more accessible to the community. GRAPHICAL ABSTRACT
Learning Valid Dual Bounds in Constraint Programming: Boosted Lagrangian Decomposition with Self-Supervised Learning
Swann Bessa
Darius Dabert
Max Bourgeat
Louis-Martin Rousseau
One hundred years of EEG for brain and behaviour research
Faisal Mushtaq
Dominik Welke
Anne Gallagher
Yuri G. Pavlov
Layla Kouara
Jorge Bosch-Bayard
Jasper JF van den Bosch
Mahnaz Arvaneh
Amy R. Bland
Maximilien Chaumon
Cornelius Borck
Xun He
Steven J. Luck
Maro G. Machizawa
Cyril Pernet
Aina Puce
Sidney J. Segalowitz
Christine Rogers
Muhammad Awais
Claudio Babiloni … (see 75 more)
Neil W. Bailey
Sylvain Baillet
Robert C. A. Bendall
Daniel Brady
Maria L. Bringas-Vega
Niko A. Busch
Ana Calzada-Reyes
Armand Chatard
Peter E. Clayson
Michael X. Cohen
Jonathan Cole
Martin Constant
Alexandra Corneyllie
Damien Coyle
Damian Cruse
Ioannis Delis
Arnaud Delorme
Damien Fair
Tiago H. Falk
Matthias Gamer
Giorgio Ganis
Kilian Gloy
Samantha Gregory
Cameron D. Hassall
Katherine E. Hiley
Richard B. Ivry
Michael Jenkins
Jakob Kaiser
Andreas Keil
Robert T. Knight
Silvia Kochen
Boris Kotchoubey
Olave E. Krigolson
Nicolas Langer
Heinrich R. Liesefeld
Sarah Lippé
Raquel E. London
Annmarie MacNamara
Scott Makeig
Welber Marinovic
Eduardo Martínez-Montes
Aleya A. Marzuki
Ryan K. Mathew
Christoph Michel
José d. R. Millán
Mark Mon-Williams
Lilia Morales-Chacón
Richard Naar
Gustav Nilsonne
Guiomar Niso
Erika Nyhus
Robert Oostenveld
Katharina Paul
Walter Paulus
Daniela M. Pfabigan
Gilles Pourtois
Stefan Rampp
Manuel Rausch
Kay Robbins
Paolo M. Rossini
Manuela Ruzzoli
Barbara Schmidt
Magdalena Senderecka
Narayanan Srinivasan
Yannik Stegmann
Paul M. Thompson
Mitchell Valdes-Sosa
Melle J. W. van der Molen
Domenica Veniero
Edelyn Verona
Bradley Voytek
Dezhong Yao
Alan C. Evans
Pedro Valdes-Sosa
Zero-Shot Object-Centric Representation Learning
Aniket Rajiv Didolkar
Andrii Zadaianchuk
Anirudh Goyal
Michael Curtis Mozer
Georg Martius
Maximilian Seitzer
The goal of object-centric representation learning is to decompose visual scenes into a structured representation that isolates the entities… (see more). Recent successes have shown that object-centric representation learning can be scaled to real-world scenes by utilizing pre-trained self-supervised features. However, so far, object-centric methods have mostly been applied in-distribution, with models trained and evaluated on the same dataset. This is in contrast to the wider trend in machine learning towards general-purpose models directly applicable to unseen data and tasks. Thus, in this work, we study current object-centric methods through the lens of zero-shot generalization by introducing a benchmark comprising eight different synthetic and real-world datasets. We analyze the factors influencing zero-shot performance and find that training on diverse real-world images improves transferability to unseen scenarios. Furthermore, inspired by the success of task-specific fine-tuning in foundation models, we introduce a novel fine-tuning strategy to adapt pre-trained vision encoders for the task of object discovery. We find that the proposed approach results in state-of-the-art performance for unsupervised object discovery, exhibiting strong zero-shot transfer to unseen datasets.
Context-Aware Assistant Selection for Improved Inference Acceleration with Large Language Models
Jerry Huang
Prasanna Parthasarathi
Mehdi Rezagholizadeh
Understanding the Local Geometry of Generative Model Manifolds
Ahmed Imtiaz Humayun
Ibtihel Amara
Candice Schumann
Mohammad Havaei
Deep generative models learn continuous representations of complex data manifolds using a finite number of samples during training. For a pr… (see more)e-trained generative model, the common way to evaluate the quality of the manifold representation learned, is by computing global metrics like Fr\'echet Inception Distance using a large number of generated and real samples. However, generative model performance is not uniform across the learned manifold, e.g., for \textit{foundation models} like Stable Diffusion generation performance can vary significantly based on the conditioning or initial noise vector being denoised. In this paper we study the relationship between the \textit{local geometry of the learned manifold} and downstream generation. Based on the theory of continuous piecewise-linear (CPWL) generators, we use three geometric descriptors - scaling (
<scp>RF</scp> shimming in the cervical spinal cord at <scp>7 T</scp>
Daniel Papp
Kyle M. Gilbert
Gaspard Cereza
Alexandre D'Astous
Nibardo Lopez‐Rios
Mathieu Boudreau
Marcus J. Couch
Pedram Yazdanbakhsh
Robert L. Barry
Eva Alonso‐Ortiz