Portrait de Mohammad Javad Darvishi Bayazi

Mohammad Javad Darvishi Bayazi

Doctorat - UdeM
Superviseur⋅e principal⋅e
Co-supervisor
Sujets de recherche
Apprentissage de représentations
Modèles de fondation
Neurosciences computationnelles
Séries temporelles

Publications

Introducing Brain Foundation Models
Hena Ghonia
Bruno Aristimunha
Md Rifat Arefin
Sylvain Chevallier
Brain function represents one of the most complex systems driving our world. Decoding its signals poses significant challenges, particularly… (voir plus) due to the limited availability of data and the high cost of recordings. The existence of large hospital datasets and laboratory collections partially mitigates this issue. However, the lack of standardized recording protocols, varying numbers of channels, diverse setups, scenarios, and recording devices further complicate the task. This work addresses these challenges by introducing the Brain Foundation Model (BFM), a suite of open-source models trained on brain signals. These models serve as foundational tools for various types of time-series neuroimaging tasks. This work presents the first model of the BFM series, which is trained on electroencephalogram signal data. Our results demonstrate that BFM-EEG can generate signals more accurately than other models. Upon acceptance, we will release the model weights and pipeline.
When Machines Outshine Humans in Object Recognition, Benchmarking Dilemma
Md Rifat Arefin
Jocelyn Faubert
VFA: Vision Frequency Analysis of Foundation Models and Human
Md Rifat Arefin
Jocelyn Faubert
Machine learning models often struggle with distribution shifts in real-world scenarios, whereas humans exhibit robust adaptation. Models th… (voir plus)at better align with human perception may achieve higher out-of-distribution generalization. In this study, we investigate how various characteristics of large-scale computer vision models influence their alignment with human capabilities and robustness. Our findings indicate that increasing model and data size, along with incorporating rich semantic information and multiple modalities, significantly enhances models' alignment with human perception and their overall robustness. Our empirical analysis demonstrates a strong correlation between out-of-distribution accuracy and human alignment.
VFA: Vision Frequency Analysis of Foundation Models and Human
Md Rifat Arefin
Jocelyn Faubert
Amplifying Pathological Detection in EEG Signaling Pathways through Cross-Dataset Transfer Learning
Mohammad S. Ghaemi
Timothee LESORT
Md Rifat Arefin
Jocelyn Faubert
WOODS: Benchmarks for Out-of-Distribution Generalization in Time Series
Jean-Christophe Gagnon-Audet
Kartik Ahuja
Beyond performance: the role of task demand, effort, and individual differences in ab initio pilots
Andrew Law
Sergio Mejia Romero
Sion Jennings
Jocelyn Faubert
Neural efficiency in an aviation task with different levels of difficulty: Assessing different biometrics during a performance task
Andrew Law
Sergio Mejia Romero
Sion Jennings
Jocelyn Faubert
WOODS: Benchmarks for Out-of-Distribution Generalization in Time Series Tasks
Jean-Christophe Gagnon-Audet
Kartik Ahuja
Generalizing to unseen domains via distribution matching
Isabela Albuquerque
Joao Monteiro
Tiago Falk
Supervised learning results typically rely on assumptions of i.i.d. data. Unfortunately, those assumptions are commonly violated in practice… (voir plus). In this work, we tackle this problem by focusing on domain generalization: a formalization where the data generating process at test time may yield samples from never-before-seen domains (distributions). Our work relies on a simple lemma: by minimizing a notion of discrepancy between all pairs from a set of given domains, we also minimize the discrepancy between any pairs of mixtures of domains. Using this result, we derive a generalization bound for our setting. We then show that low risk over unseen domains can be achieved by representing the data in a space where (i) the training distributions are indistinguishable, and (ii) relevant information for the task at hand is preserved. Minimizing the terms in our bound yields an adversarial formulation which estimates and minimizes pairwise discrepancies. We validate our proposed strategy on standard domain generalization benchmarks, outperforming a number of recently introduced methods. Notably, we tackle a real-world application where the underlying data corresponds to multi-channel electroencephalography time series from different subjects, each considered as a distinct domain.