Portrait de Giovanni Beltrame

Giovanni Beltrame

Membre affilié
Professeur titulaire, Polytechnique Montréal, Département de génie informatique et génie logiciel
Sujets de recherche
Apprentissage en ligne
Apprentissage par renforcement
Intelligence en essaim
Interaction humain-robot
Navigation robotique autonome
Robotique
Systèmes distribués
Vision par ordinateur

Biographie

Giovanni Beltrame a obtenu un doctorat en génie informatique du Politecnico di Milano en 2006, après quoi il a travaillé comme ingénieur en microélectronique à l'Agence spatiale européenne (ESA) sur un certain nombre de projets, allant des systèmes tolérants aux radiations à la conception assistée par ordinateur. En 2010, il s'est installé à Montréal. Il est actuellement professeur au Département de génie informatique et logiciel de Polytechnique Montréal. Il dirige notamment le laboratoire MIST, qui se consacre aux technologies spatiales, où plus de 25 étudiant·e·s et postdoctorant·e·s sont sous sa supervision. Il a réalisé plusieurs projets en collaboration avec l'industrie et les agences gouvernementales dans les domaines de la robotique, de l'intervention en cas de catastrophe et de l'exploration spatiale. Avec son équipe, il a participé à plusieurs missions sur le terrain avec l'ESA, l'Agence spatiale canadienne (ASC) et la NASA (BRAILLE, PANAGAEA-X et IGLUNA, entre autres). Ses recherches portent sur la modélisation et la conception de systèmes embarqués, l'intelligence artificielle et la robotique, sujets sur lesquels il a publié plusieurs articles dans des revues et des conférences de premier plan.

Étudiants actuels

Doctorat - Polytechnique
Co-superviseur⋅e :
Collaborateur·rice de recherche - Polytechnique Montreal
Co-superviseur⋅e :
Maîtrise recherche - Polytechnique
Co-superviseur⋅e :
Doctorat - Polytechnique
Co-superviseur⋅e :
Maîtrise recherche - UdeM
Co-superviseur⋅e :
Doctorat - Polytechnique
Co-superviseur⋅e :

Publications

Neural Incremental Dynamic Inversion Control of a Multirotor Robotic Airship
Ely Carneiro de Paiva
José Raul Azinheira
Rafael de Angelis Cordeiro
José Reginaldo H. Carvalho
Apolo Marton
PEACE: Prompt Engineering Automation for CLIPSeg Enhancement for Safe-Landing Zone Segmentation
Rongge Zhang
Antoine Robillard
Safe landing is essential in robotics applications, from industrial settings to space exploration. As artificial intelligence advances, we h… (voir plus)ave developed PEACE (Prompt Engineering Automation for CLIPSeg Enhancement), a system that automatically generates and refines prompts for identifying landing zones in changing environments. Traditional approaches using fixed prompts for open-vocabulary models struggle with environmental changes and can lead to dangerous outcomes when conditions are not represented in the predefined prompts. PEACE addresses this limitation by dynamically adapting to shifting data distributions. Our key innovation is the dual segmentation of safe and unsafe landing zones, allowing the system to refine the results by removing unsafe areas from potential landing sites. Using only monocular cameras and image segmentation, PEACE can safely guide descent operations from 100 meters to altitudes as low as 20 meters. The testing shows that PEACE significantly outperforms the standard CLIP and CLIPSeg prompting methods, improving the successful identification of safe landing zones from 57% to 92%. We have also demonstrated enhanced performance when replacing CLIPSeg with FastSAM. The complete source code is available as an open-source software 1.
BlabberSeg: Real-Time Embedded Open-Vocabulary Aerial Segmentation
Ricardo de Azambuja
Real-time aerial image segmentation plays an important role in the environmental perception of Uncrewed Aerial Vehicles (UAVs). We introduce… (voir plus) BlabberSeg, an optimized Vision-Language Model built on CLIPSeg for on-board, real-time processing of aerial images by UAVs. BlabberSeg improves the efficiency of CLIPSeg by reusing prompt and model features, reducing computational overhead while achieving real-time open-vocabulary aerial segmentation. We validated BlabberSeg in a safe landing scenario using the Dynamic Open-Vocabulary Enhanced SafE-Landing with Intelligence (DOVESEI) framework, which uses visual servoing and open-vocabulary segmentation. BlabberSeg reduces computational costs significantly, with a speed increase of 927.41% (16.78 Hz) on a NVIDIA Jetson Orin AGX (64GB) compared with the original CLIPSeg (1.81Hz), achieving real-time aerial segmentation with negligible loss in accuracy (2.1% as the ratio of the correctly segmented area with respect to CLIPSeg). BlabberSeg's source code is open and available online.
Active Semantic Mapping and Pose Graph Spectral Analysis for Robot Exploration
Physical Simulation for Multi-agent Multi-machine Tending
Abdalwhab Abdalwhab
David St-Onge
Beyond the lab: Feasibility of cognitive neuroscience data collection during a speleological expedition
Anita Paas
Hugo R. Jourde
Arnaud Brignol
Marie-Anick Savard
Zseyvfin Eyqvelle
Samuel Bassetto
Emily B.J. Coffey
Multi-Objective Risk Assessment Framework for Exploration Planning Using Terrain and Traversability Analysis
Riana Gagnon Souleiman
Vivek Shankar Vardharajan
Frequency-based View Selection in Gaussian Splatting Reconstruction
Monica Li
Pierre-Yves Lajoie
Three-dimensional reconstruction is a fundamental problem in robotics perception. We examine the problem of active view selection to perform… (voir plus) 3D Gaussian Splatting reconstructions with as few input images as possible. Although 3D Gaussian Splatting has made significant progress in image rendering and 3D reconstruction, the quality of the reconstruction is strongly impacted by the selection of 2D images and the estimation of camera poses through Structure-from-Motion (SfM) algorithms. Current methods to select views that rely on uncertainties from occlusions, depth ambiguities, or neural network predictions directly are insufficient to handle the issue and struggle to generalize to new scenes. By ranking the potential views in the frequency domain, we are able to effectively estimate the potential information gain of new viewpoints without ground truth data. By overcoming current constraints on model architecture and efficacy, our method achieves state-of-the-art results in view selection, demonstrating its potential for efficient image-based 3D reconstruction.
Swarming Out of the Lab: Comparing Relative Localization Methods for Collective Behavior
Rafael Gomes Braga
Vivek Shankar Vardharajan
David St-Onge
Concurrent product layout design optimization and dependency management using a modified NSGA-III approach
Yann-Seing Law-Kam Cio
Aurelian Vadean
Abolfazl Mohebbi
Sofiane Achiche
The complexity of mechatronic systems has increased with the significant advancements of technology in their components which makes their de… (voir plus)sign more challenging. This is due to the need for incorporating expertise from different domains as well as the increased number and complexity of components integrated into the product. To alleviate the burden of designing such products, many industries and researchers are attracted to the concept of modularization which is to identify a subset of system components that can form a module. To achieve this, a novel product-related dependency management approach is proposed in this paper with the support of an augmented design structure matrix. This approach makes it possible to model positive and negative dependencies and to compute the combination potency between components to form modules. This approach is then integrated into a modified non-dominated sorting genetic algorithm III to concurrently optimize the design and identify the modules. The methodology is exemplified through the case study of a layout design of an automatic greenhouse. By applying the proposed methodology to the case study, it was possible to generate concepts that decreased the number of modules from 9 down to 4 while ensuring the optimization of the design performance.
LiDAR-based Real-Time Object Detection and Tracking in Dynamic Environments
Wenqiang Du
In dynamic environments, the ability to detect and track moving objects in real-time is crucial for autonomous robots to navigate safely and… (voir plus) effectively. Traditional methods for dynamic object detection rely on high accuracy odometry and maps to detect and track moving objects. However, these methods are not suitable for long-term operation in dynamic environments where the surrounding environment is constantly changing. In order to solve this problem, we propose a novel system for detecting and tracking dynamic objects in real-time using only LiDAR data. By emphasizing the extraction of low-frequency components from LiDAR data as feature points for foreground objects, our method significantly reduces the time required for object clustering and movement analysis. Additionally, we have developed a tracking approach that employs intensity-based ego-motion estimation along with a sliding window technique to assess object movements. This enables the precise identification of moving objects and enhances the system's resilience to odometry drift. Our experiments show that this system can detect and track dynamic objects in real-time with an average detection accuracy of 88.7\% and a recall rate of 89.1\%. Furthermore, our system demonstrates resilience against the prolonged drift typically associated with front-end only LiDAR odometry. All of the source code, labeled dataset, and the annotation tool are available at: https://github.com/MISTLab/lidar_dynamic_objects_detection.git
Overcoming Boundaries: Interdisciplinary Challenges and Opportunities in Cognitive Neuroscience
Arnaud Brignol
Anita Paas
Luis Sotelo-Castro
David St-Onge
Emily B.J. Coffey