Portrait de Istabrak Abbes n'est pas disponible

Istabrak Abbes

Maîtrise recherche - UdeM
Superviseur⋅e principal⋅e
Sujets de recherche
Apprentissage continu
Apprentissage profond
Traitement du langage naturel

Publications

Revisiting Replay and Gradient Alignment for Continual Pre-Training of Large Language Models
Matthew D Riemer
Tsuguchika Tabaru
Hiroaki Kingetsu
A. Chandar
Small Encoders Can Rival Large Decoders in Detecting Groundedness
Fernando Rodriguez
Alaa Boukhary
Adam Elwood
A. Chandar
Augmenting large language models (LLMs) with external context significantly improves their performance in natural language processing (NLP) … (voir plus)tasks. However, LLMs struggle to answer queries reliably when the provided context lacks information, often resorting to ungrounded speculation or internal knowledge. Groundedness - generating responses strictly supported by the context - is essential for ensuring factual consistency and trustworthiness. This study focuses on detecting whether a given query is grounded in a document provided in context before the costly answer generation by LLMs. Such a detection mechanism can significantly reduce both inference time and resource consumption. We show that lightweight, task specific encoder models such as RoBERTa and NomicBERT, fine-tuned on curated datasets, can achieve accuracy comparable to state-of-the-art LLMs, such as Llama3 8B and GPT4o, in groundedness detection while reducing inference latency by orders of magnitude. The code is available at : https://github.com/chandarlab/Hallucinate-less