Portrait de Amal Zouaq

Amal Zouaq

Membre académique associé
Professeure titulaire, Polytechnique Montréal, Département de génie informatique et génie logiciel
Professeure agrégée, University of Ottawa, École de génie électrique et informatique
Sujets de recherche
Apprentissage de représentations
Apprentissage sur graphes
Graphes de connaissances
Modèles génératifs
Recherche d'information
Traitement du langage naturel

Biographie

Amal Zouaq est professeure titulaire à Polytechnique Montréal, au département de génie informatique et génie logiciel et membre académique associée à Mila. Elle est titulaire d’une Chaire FRQS (Double) en IA et santé numérique en plus dêtre professeure IVADO, membre du consortium CLIQ-AI (Computational Linguistics in Québec) et professeure associée à l'Université d'Ottawa.

Ses intérêts de recherche comprennent l'intelligence artificielle, le traitement du langage naturel et le Web sémantique. Elle est directrice du laboratoire de recherche LAMA-WeST, spécialisé dans tous les aspects du traitement du langage naturel et de l'intelligence artificielle, incluant les modèles de langage à grande échelle (LLMs) avec mémoires non paramétriques, les LLMs modulaires, les modèles neuro-symboliques et le Web sémantique.

Professeure Zouaq est membre du comité de programme de nombreuses conférences et revues sur l'ingénierie des connaissances et des données, le traitement du langage naturel, l'exploration de données et le Web sémantique.

Étudiants actuels

Maîtrise recherche - Polytechnique
Doctorat - Polytechnique
Superviseur⋅e principal⋅e :
Maîtrise recherche - Polytechnique
Doctorat - Polytechnique
Maîtrise recherche - Polytechnique
Maîtrise recherche - Polytechnique
Doctorat - Polytechnique
Postdoctorat - Polytechnique
Maîtrise recherche - Polytechnique
Doctorat - Polytechnique
Doctorat - Polytechnique
Doctorat - Polytechnique
Co-superviseur⋅e :

Publications

LLMs Can't Play Hangman: On the Necessity of a Private Working Memory for Language Agents
Combining Domain and Alignment Vectors Provides Better Knowledge-Safety Trade-offs in LLMs
Matthew D Riemer
Pin-Yu Chen
Payel Das
A. Chandar
Combining Domain and Alignment Vectors to Achieve Better Knowledge-Safety Trade-offs in LLMs
Yash More
Matthew D Riemer
Pin-Yu Chen
Payel Das
A. Chandar
Chandar Research Lab
Mila - Québec
U. Montŕeal
AI Institute
Ibm Research
Polytechnique Montréal
There is a growing interest in training domain-expert LLMs that excel in specific technical fields compared to their general-purpose instruc… (voir plus)tion-tuned counterparts. However, these expert models often experience a loss in their safety abilities in the process, making them capable of generating harmful content. As a solution, we introduce an efficient and effective merging-based alignment method called \textsc{MergeAlign} that interpolates the domain and alignment vectors, creating safer domain-specific models while preserving their utility. We apply \textsc{MergeAlign} on Llama3 variants that are experts in medicine and finance, obtaining substantial alignment improvements with minimal to no degradation on domain-specific benchmarks. We study the impact of model merging through model similarity metrics and contributions of individual models being merged. We hope our findings open new research avenues and inspire more efficient development of safe expert LLMs.
A Deep Dive into the Trade-Offs of Parameter-Efficient Preference Alignment Techniques
Matthew D Riemer
Pin-Yu Chen
Payel Das
A. Chandar
MVP: Minimal Viable Phrase for Long Text Understanding.
Assessing the Generalization Capabilities of Neural Machine Translation Models for SPARQL Query Generation
Samuel Reyd
SORBET: A Siamese Network for Ontology Embeddings Using a Distance-Based Regression Loss and BERT
Francis Gosselin
SORBETmatcher results for OAEI 2023.
Francis Gosselin
Local Structure Matters Most: Perturbation Study in NLU
Recent research analyzing the sensitivity of natural language understanding models to word-order perturbations has shown that neural models … (voir plus)are surprisingly insensitive to the order of words. In this paper, we investigate this phenomenon by developing order-altering perturbations on the order of words, subwords, and characters to analyze their effect on neural models' performance on language understanding tasks. We experiment with measuring the impact of perturbations to the local neighborhood of characters and global position of characters in the perturbed texts and observe that perturbation functions found in prior literature only affect the global ordering while the local ordering remains relatively unperturbed. We empirically show that neural models, invariant of their inductive biases, pretraining scheme, or the choice of tokenization, mostly rely on the local structure of text to build understanding and make limited use of the global structure.
Detecting Languages Unintelligible to Multilingual Models through Local Structure Probes
Providing better language tools for low-resource and endangered languages is imperative for equitable growth. Recent progress with massively… (voir plus) multilingual pretrained models has proven surprisingly effective at performing zero-shot transfer to a wide variety of languages. However, this transfer is not universal, with many languages not currently understood by multilingual approaches. It is estimated that only 72 languages possess a "small set of labeled datasets" on which we could test a model's performance, the vast majority of languages not having the resources available to simply evaluate performances on. In this work, we attempt to clarify which languages do and do not currently benefit from such transfer. To that end, we develop a general approach that requires only unlabelled text to detect which languages are not well understood by a cross-lingual model. Our approach is derived from the hypothesis that if a model's understanding is insensitive to perturbations to text in a language, it is likely to have a limited understanding of that language. We construct a cross-lingual sentence similarity task to evaluate our approach empirically on 350, primarily low-resource, languages.
Local Structure Matters Most in Most Languages