Portrait of Amal Zouaq

Amal Zouaq

Associate Academic Member
Full Professor, Polytechnique Montréal, Department of Computer Engineering and Software Engineering
Associate Professor, University of Ottawa, School of Electrical Engineering and Computer Science
Research Topics
Generative Models
Information Retrieval
Knowledge Graphs
Learning on Graphs
Natural Language Processing
Representation Learning

Biography

Amal Zouaq is a Full Professor at Polytechnique Montreal in the computer science and software engineering department and an Associate Academic Member at Mila - Quebec Artificial Intelligence Institute. She holds an FRQS (Dual) Chair in AI and Digital Health. She is also an IVADO professor, a member of the CLIQ-AI consortium (Computational Linguistics in Québec), and an adjunct professor at the University of Ottawa.

Prof. Zouaq's research interests encompass artificial intelligence, natural language processing, and the Semantic Web. As the director of the LAMA-WeST research lab, her work extends to various facets of natural language processing and artificial intelligence, including LLMs with nonparametric memories, modular LLMs, neuro-symbolic models and Semantic Web.

Furthermore, she actively contributes to the academic community by participating as a program committee member in numerous conferences and journals related o knowledge and data engineering, natural language processing, and the Semantic Web.

Current Students

Master's Research - Polytechnique Montréal
PhD - Polytechnique Montréal
Principal supervisor :
Master's Research - Polytechnique Montréal
PhD - Polytechnique Montréal
Master's Research - Polytechnique Montréal
Master's Research - Polytechnique Montréal
PhD - Polytechnique Montréal
Postdoctorate - Polytechnique Montréal
Master's Research - Polytechnique Montréal
PhD - Polytechnique Montréal
PhD - Polytechnique Montréal
PhD - Polytechnique Montréal
Co-supervisor :

Publications

LLMs Can't Play Hangman: On the Necessity of a Private Working Memory for Language Agents
Combining Domain and Alignment Vectors Provides Better Knowledge-Safety Trade-offs in LLMs
Matthew D Riemer
Pin-Yu Chen
Payel Das
A. Chandar
Combining Domain and Alignment Vectors to Achieve Better Knowledge-Safety Trade-offs in LLMs
Yash More
Matthew D Riemer
Pin-Yu Chen
Payel Das
A. Chandar
Chandar Research Lab
Mila - Québec
U. Montŕeal
AI Institute
Ibm Research
Polytechnique Montréal
There is a growing interest in training domain-expert LLMs that excel in specific technical fields compared to their general-purpose instruc… (see more)tion-tuned counterparts. However, these expert models often experience a loss in their safety abilities in the process, making them capable of generating harmful content. As a solution, we introduce an efficient and effective merging-based alignment method called \textsc{MergeAlign} that interpolates the domain and alignment vectors, creating safer domain-specific models while preserving their utility. We apply \textsc{MergeAlign} on Llama3 variants that are experts in medicine and finance, obtaining substantial alignment improvements with minimal to no degradation on domain-specific benchmarks. We study the impact of model merging through model similarity metrics and contributions of individual models being merged. We hope our findings open new research avenues and inspire more efficient development of safe expert LLMs.
A Deep Dive into the Trade-Offs of Parameter-Efficient Preference Alignment Techniques
Matthew D Riemer
Pin-Yu Chen
Payel Das
A. Chandar
MVP: Minimal Viable Phrase for Long Text Understanding.
Assessing the Generalization Capabilities of Neural Machine Translation Models for SPARQL Query Generation
Samuel Reyd
SORBET: A Siamese Network for Ontology Embeddings Using a Distance-Based Regression Loss and BERT
Francis Gosselin
SORBETmatcher results for OAEI 2023.
Francis Gosselin
Local Structure Matters Most: Perturbation Study in NLU
Recent research analyzing the sensitivity of natural language understanding models to word-order perturbations has shown that neural models … (see more)are surprisingly insensitive to the order of words. In this paper, we investigate this phenomenon by developing order-altering perturbations on the order of words, subwords, and characters to analyze their effect on neural models' performance on language understanding tasks. We experiment with measuring the impact of perturbations to the local neighborhood of characters and global position of characters in the perturbed texts and observe that perturbation functions found in prior literature only affect the global ordering while the local ordering remains relatively unperturbed. We empirically show that neural models, invariant of their inductive biases, pretraining scheme, or the choice of tokenization, mostly rely on the local structure of text to build understanding and make limited use of the global structure.
Detecting Languages Unintelligible to Multilingual Models through Local Structure Probes
Providing better language tools for low-resource and endangered languages is imperative for equitable growth. Recent progress with massively… (see more) multilingual pretrained models has proven surprisingly effective at performing zero-shot transfer to a wide variety of languages. However, this transfer is not universal, with many languages not currently understood by multilingual approaches. It is estimated that only 72 languages possess a "small set of labeled datasets" on which we could test a model's performance, the vast majority of languages not having the resources available to simply evaluate performances on. In this work, we attempt to clarify which languages do and do not currently benefit from such transfer. To that end, we develop a general approach that requires only unlabelled text to detect which languages are not well understood by a cross-lingual model. Our approach is derived from the hypothesis that if a model's understanding is insensitive to perturbations to text in a language, it is likely to have a limited understanding of that language. We construct a cross-lingual sentence similarity task to evaluate our approach empirically on 350, primarily low-resource, languages.
Local Structure Matters Most in Most Languages