Défense de thèse de doctorat en informatique - Sacha Corbugy
Abstract
In recent decades, the volume of data generated worldwide has grown exponentially, significantly accelerating advancements in machine learning. This explosion of data has led to an increased need for effective data exploration techniques, giving rise to a specialized field known as dimensionality reduction. Dimensionality reduction methods are used to transform high-dimensional data into a low-dimensional space (typically 2D or 3D), so that it can be easily visualized and understood by humans. Algorithms such as Principal Component Analysis (PCA), Multidimensional Scaling (MDS), and t-distributed Stochastic Neighbor Embedding (t-SNE) have become essential tools for visualizing complex datasets. These techniques play a critical role in exploratory data analysis and in interpreting complex models like Convolutional Neural Networks (CNNs). Despite their widespread adoption, dimensionality reduction techniques, particularly non-linear ones, often lack interpretability. This opacity makes it difficult for users to understand the meaning of the visualizations or the rationale behind specific low-dimensional representations. In contrast, the field of supervised machine learning has seen significant progress in explainable AI (XAI), which aims to clarify model decisions, especially in high-stakes scenarios. While many post-hoc explanation tools have been developed to interpret the outputs of supervised models, there is still a notable gap in methods for explaining the results of dimensionality reduction techniques. This research investigates how post-hoc explanation techniques can be integrated into dimensionality reduction algorithms to improve user understanding of the resulting visualizations. Specifically, it explores how interpretability methods originally developed for supervised learning can be adapted to explain the behavior of non-linear dimensionality reduction algorithms. Additionally, this work examines whether the integration of post-hoc explanations can enhance the overall effectiveness of data exploration. As these tools are intended for end-users, we also design and evaluate an interactive system that incorporates explanatory mechanisms. We argue that combining interpretability with interactivity significantly improves users' understanding of embeddings produced by non-linear dimensionality reduction techniques. In this research, we propose enhancements to an existing post-hoc explanation method that adapts LIME for t-SNE. We introduce a globally-local framework for fast and scalable explanations of t-SNE embeddings. Furthermore, we present a completely new approach that adapts saliency map-based explanations to locally interpret non-linear dimensionality reduction results. Lastly, we introduce our interactive tool, Insight-SNE, which integrates our gradient-based explanation method and enables users to explore low-dimensional embeddings through direct interaction with the explanations.
Jury
Prof. Wim Vanhoof - University of Namur, BelgiumProf. Benoit Frénay - University of Namur, BelgiumProf. Bruno Dumas - University of Namur, BelgiumProf. John Lee - University of Louvain, BelgiumProf. Luis Galarraga - University of Rennes, France
Voir le contenu
Faculté EMCP : trois chercheurs primés - #3 Quand l’IA devient plus humaine : Florence Nizette (NaDI) décroche un prix international
Troisième et dernier focus de l’été sur le centre de recherche NaDI-CeRCLe, qui s’est démarqué à l’international ces dernières semaines grâce aux reconnaissances obtenues par trois jeunes chercheurs en management des services. Après Floriane Goosse et Victor Sluÿters, nous vous proposons de découvrir le travail de Florence Nizette, jeune chercheuse travaillant sur les technologies d’Intelligence artificielle.
Voir le contenu
Comprendre pour mieux protéger : un projet de recherche conjoint FNRS-FRQ novateur sur le béluga du Saint-Laurent
Un projet déposé par le Laboratoire de Physiologie Évolutive et Adaptative (LEAP) du professeur Frédéric Silvestre de l’Université de Namur a été classé parmi les 6 meilleurs projets de recherche financés par le FNRS et le Fonds de recherche du Québec (FRQ) pour une collaboration scientifique entre la Wallonie et le Québec. Le but ? Comprendre l'impact des activités humaines sur les bélugas de l'estuaire du Saint-Laurent (ESL) à l’aide d’approches interdisciplinaires pour permettre d’améliorer les stratégies de conservation de cette espèce menacée.
Voir le contenu
La biodiversité des rivières américaines analysée pendant 30 ans
Une équipe de chercheurs américains, avec l’aide de Frédérik De Laender, professeur au Département de biologie de l’UNamur, vient de publier dans la prestigieuse revue Nature. Leur étude décrit comment l’évolution des températures des cours d’eau et les introductions de poissons par l’humain peuvent modifier la biodiversité des rivières aux États-Unis.
Voir le contenu
Soutenance publique de thèse de doctorat en informatique - Antoine Gratia
Abstract
Deep learning has become an extremely important technology in numerous domains such as computer vision, natural language processing, and autonomous systems. As neural networks grow in size and complexity to meet the demands of these applications, the cost of designing and training efficient models continues to rise in computation and energy consumption. Neural Architecture Search (NAS) has emerged as a promising solution to automate the design of performant neural networks. However, conventional NAS methods often require evaluating thousands of architectures, making them extremely resource-intensive and environmentally costly.This thesis introduces a novel, energy-aware NAS pipeline that operates at the intersection of Software Engineering and Machine Learning. We present CNNGen, a domain-specific generator for convolutional architectures, combined with performance and energy predictors to drastically reduce the number of architectures that need full training. These predictors are integrated into a multi-objective genetic algorithm (NSGA-II), enabling an efficient search for architectures that balance accuracy and energy consumption.Our approach explores a variety of prediction strategies, including sequence-based models, image-based representations, and deep metric learning, to estimate model quality from partial or symbolic representations. We validate our framework across three benchmark datasets, CIFAR-10, CIFAR-100, and Fashion-MNIST, demonstrating that it can produce results comparable to state-of-the-art architectures with significantly lower computational cost. By reducing the environmental footprint of NAS while maintaining high performance, this work contributes to the growing field of Green AI and highlights the value of predictive modelling in scalable and sustainable deep learning workflows.
Jury
Prof. Wim Vanhoof - University of Namur, BelgiumProf. Gilles Perrouin - University of Namur, BelgiumProf. Benoit Frénay - University of Namur, BelgiumProf. Pierre-Yves Schobbens - University of Namur, BelgiumProf. Clément Quinton - University of Lille, FranceProf. Paul Temple- University of Rennes, FranceProf. Schin’ichi Satoh - National Institute of Informatics, Japon
Voir le contenu
Faculté EMCP : trois chercheurs primés - #2 Victor Sluÿters, le doctorant qui décrypte les comportements des employés en cas de crise
Moisson de récompenses pour le centre de recherche NaDI-CeRCLe ces dernières semaines. Les recherches en management des services de trois jeunes doctorants de la Faculté EMCP ont en effet été reconnues par leurs pairs lors d’événements scientifiques internationaux de premier plan : il s’agit de Floriane Goosse, Victor Sluÿters et Florence Nizette. Cet été, nous vous proposons de découvrir leurs parcours et leurs travaux.
Voir le contenu
Défense de thèse de doctorat en Sciences biologiques - Mathilde Oger
Abstract
Plastic pollution has emerged as a pervasive environmental threat, with micro- and nanoplastics (MPs and NPs) accumulating across ecosystems and organisms, including humans. Their ability to adsorb and transport contaminants raises critical concerns for both environmental and public health.This thesis investigates the developmental neurotoxicity of MPs and NPs in zebrafish (Danio rerio), emphasizing the influence of particle size and mixture toxicity. NPs were shown to cross the embryonic chorion, disrupt physiological functions, and induce anxiety-like behaviour, whereas MPs mainly altered gene expression related to neurodevelopment. When co-exposed with methylmercury (MeHg), NPs enhanced MeHg accumulation in the brain and sensory organs, exacerbating its neurotoxic effects. Notably, the mixture induced severe hypoactivity, impaired lipid metabolism and neurotransmission, and increased mortality.These findings highlight the critical need to assess plastic particle toxicity not only in isolation but also in environmentally relevant mixtures. NPs, due to their small size and high reactivity, may act as vectors for toxicants like MeHg, amplifying their effects during sensitive developmental stages.
Jury
Prof. Frédéric SILVESTRE (UNamur), PrésidentProf. Patrick KESTEMONT (UNamur), SecrétaireDr Valérie CORNET (UNamur)Prof. Eli THORÉ (UNamur)Prof. Jérôme CACHOT (Université de Bordeaux)Dr Krishna DAS (Université de Liège)
Voir le contenu
Du jeu vidéo à l’intelligence artificielle, escale au Japon
Près de 10 000 kilomètres séparent la Belgique du Japon, un pays qui fascine, notamment pour sa culture riche et pleine de contrastes. Les chercheurs de l’UNamur entretiennent des liens étroits avec plusieurs institutions nipponnes, notamment dans les domaines de l’informatique, des mathématiques ou encore du jeu vidéo. Coup de projecteur sur quelques-unes de ces collaborations.
Voir le contenu
Les séances du numérique | L'IA a-t-elle une conscience ?
Programme : 17h : Accueil & présentation du film17h15 : Projection du film Ex machina19h05 : Débat « L’IA a-t-elle une conscience ? » (avec Isabelle Linden & Benoît Frenay)19h45 : finDeux experts prendront part au débat : Benoît Frenay qui apportera un éclairage sur les logiques d’apprentissage des intelligences artificielles actuelles et les limites de leur « autonomie ». Peut-on vraiment parler d’intelligence sans conscience ? Jusqu’où peut aller l’imitation ?Isabelle Linden qui interrogera les fondements mêmes de ce que nous appelons « penser » dans une logique informatique. Peut-on créer une machine consciente ? Ou ne sommes-nous que face à des miroirs de nos propres désirs ?
S'inscrire
Voir le contenu