Projects are available on the following topics (not exclusive):
- Machine Learning and Applications
- Deep Learning Science
- Image Analysis and Computer Vision
- Graph Signal Processing and Network Machine Learning
Non-exhaustive project list
Automatic Sampling Hyperparameter Optimization of Discrete Flow Matching for Graph Generation
- ID: 81
- Status: ongoing
- Created At: 2026-01-16
Discrete flow-matching models for graph generation achieve state-of-the-art generative performance. However, this increased flexibility comes at the cost of several sampling hyperparameters that must be tuned to fully exploit the model’s capacity. In practice, these hyperparameters are selected via uni-dimensional sweeps followed by ad-hoc joint combinations, which is computationally inefficient and ignores rich information available during the sampling process itself. In particular, models such as DeFoG [1] and Directo [2] exhibit structured behavior along the denoising trajectory, including meaningful uncertainty and confidence patterns that can provide direct signals about sampling quality and failure modes. In this project, we aim to develop principled methods for automatically tuning sampling hyperparameters by exploiting signals arising during training and/or generation, such as prediction error, uncertainty, entropy dynamics, or other trajectory-level statistics. The goal is to outperform naive tuning strategies and ultimately surpass classical Bayesian optimization in sample efficiency, enabling few-shot or near one-shot hyperparameter selection without exhaustive sweeps. This work seeks to improve both the efficiency and robustness of discrete graph generative models, with applications in structured graph synthesis and molecular generation.
Requirements:
Mandatory: At least one deep learning course + prior experience with PyTorch.
Plus: Familiarity with diffusion or flow-based generative models and graph deep learning.
References:
[1] Qin et al., DeFoG: Discrete Flow Matching for Graph Generation, ICML 2025.
[2] Carballo-Castro et al., Generating Directed Graphs with Dual Attention and Asymmetric Encoding, arXiv 2025.
- Project Creator:
- Manuel Madeira ([email protected])
- Units Involved:
- Laboratoire de traitement des signaux 4 (LTS4)
- Tags:
- Master, EDEE, EDIC, Data & Network Sciences, AI,
- Memberships:
- Manuel Madeira – Role: Owner,
Concept-Based Representations in Vision Models
- ID: 130
- Status: ongoing
- Created At: 2026-04-23
- End Date: 2026-09-01
This project explores concept-based representations in pretrained vision models, focusing on concept bottleneck models (CBMs), sparse autoencoders (SAEs), and hybrids of the two. Recent work suggests that separately trained CBMs and SAEs can recover similar concept-oriented representations under certain conditions [1]. Building on this, the project asks how these methods compare in terms of interpretability, disentanglement, information leakage, and downstream usefulness, with the broader aim of understanding how to build more interpretable and robust concept-based vision models.
[1] Fel, Thomas, and Gianni Franchi. "A Geometric Unification of Concept Learning with Concept Cones." arXiv preprint arXiv:2512.07355 (2025)
- Project Creator:
- Amel Abdelraheem ([email protected])
- Units Involved:
- Laboratoire de traitement des signaux 4 (LTS4)
- Tags:
- AI,
- Memberships:
- Amel Abdelraheem – Role: Owner,
Decoding symbolic patterns in psychiatric art collections using AI
- ID: 131
- Status: ongoing
- Created At: 2026-04-27
- End Date: 2027-01-31
Description: Can we learn to read the inner experience of a person through their art? Art Brut — the raw, unmediated artistic expression produced by people with psychiatric diagnoses, outside of any communicative intent — may encode cognitive signatures specific to mental disorders. Working with digitised collections from the Collection de l'Art Brut (Lausanne), the Psychiatriemuseum Bern, and the Prinzhorn Collection (Heidelberg), the student will apply deep learning and computer vision methods to extract and cluster symbolic patterns across thousands of artworks, in dialogue with experts in art history and symbolism. This project is part of a broader transdisciplinary research programme combining AI, psychiatry, and molecular biology.
Tasks:
Build structured representations of artworks using unsupervised/self-supervised vision models (e.g. CLIP, ViTs)Extract and cluster recurrent visual and compositional patterns across diagnoses and artistsDevelop interpretable visualisations of discovered patterns in collaboration with domain experts
Profile: MSc student in Computer Science, Electrical Engineering, or Data Science. Strong background in deep learning and computer vision (PyTorch/JAX). Genuine curiosity for interdisciplinary applications. Experience with NLP or multimodal models is a plus.
Contact: [email protected] (CV + short motivation)
- Project Creator:
- Pascal Frossard ([email protected])
- Units Involved:
- Laboratoire de traitement des signaux 4 (LTS4)
- Tags:
- Master, Bachelor, SEL, SIN, SSC, Data & Network Sciences, Imaging & Vision, Learning & Decision Systems,
- Memberships:
- Pascal Frossard – Role: Supervisor, Raphaelle Luisier – Role: Owner,
Exploring efficient graph generative models
- ID: 128
- Status: ongoing
- Created At: 2026-04-11
Graph generative models aim to learn distributions over graphs and produce new instances. They are widely used in applications where structure matters such as molecule design, circuit and program synthesis, and social network simulation. Recently, graph generation through discrete flow matching has achieved remarkable success [1, 2].
Nonetheless, these methods limit themselves entirely to the discrete space. Recent works in the language space have shown that augmenting the discrete state can help improve performance and efficiency at both training and sampling time. Our objective is therefore to build efficient graph generative models, possibly taking inspiration from recent advances in discrete diffusion language models.
Candidates should be comfortable with theory (generative models, graph-structured data) and coding (PyTorch), as the work will involve designing and implementing the framework, optimizing its training dynamics, and evaluating its performance on selected benchmarks.
- Project Creator:
- Alba Carballo Castro ([email protected])
- Units Involved:
- Laboratoire de traitement des signaux 4 (LTS4)
- Tags:
- Master, Data & Network Sciences, AI,
- Memberships:
- Alba Carballo Castro – Role: Supervisor,
Fast and Learnable Discrete Diffusion Sampler
- ID: 120
- Status: ongoing
- Created At: 2026-03-13
- End Date: 2026-07-04
Discrete diffusion models are emerging as a promising alternative to autoregressive models for generating discrete data such as language or graphs. These models operate by employing iterative refinement strategies directly on discrete state spaces [1, 2, 3]. However, the sampling step typically collapses the rich contextual information produced by the model at each step, which is then discarded during the remainder of the generation process.
In this project, we study fast and learnable sampling strategies for discrete diffusion models, aiming to exploit signals produced during generation to design more effective and efficient samplers.
Requirements:
Mandatory: At least one deep learning course + prior experience with PyTorch.
Plus: Familiarity with diffusion models or generative modeling.
References
[1] S. S. Sahoo et al., Simple and Effective Masked Diffusion Language Models, NeurIPS 2024.
[2] Y. Schiff et al., Simple Guidance Mechanisms for Discrete Diffusion Models, ICLR 2025.
[3] S. S. Sahoo et al., The Diffusion Duality, ICML 2025.
- Project Creator:
- Manuel Madeira ([email protected])
- Units Involved:
- Laboratoire de traitement des signaux 4 (LTS4)
- Tags:
- Master, EDEE, EDIC, AI,
- Memberships:
- Adrien Jean Deschenaux – Role: Student, Manuel Madeira – Role: Owner,
Graph-diffusion based RNA design
- ID: 126
- Status: ongoing
- Created At: 2026-03-25
- End Date: 2026-09-20
Creating RNAs with desired properties such as folding in a specific structure or binding with a given protein requires an in-depth understanding of sequence-structure dependencies and how these relate to functional properties. Most frameworks for RNA design sequence and structure sequentially, and optimize for one single property at a time. Representing RNA as a graph is a promising path towards designing both RNA sequence and structure at the same time, especially to optimize objectives that require specific structure-sequence patterns.
We will explore the suitability of graph diffusion frameworks for RNA design, using existing methods developed in the LTS4 lab or elsewhere, with special consideration for how methods scale with sequence length.
Candidates should have strong mathematical and computational skills and should be familiar with the pytorch framework. Candidates do not necessarily have to have a biological background but should have a strong desire to directly work with experimental biologists.
- Project Creator:
- Vincent Jung ([email protected])
- Units Involved:
- Laboratoire de traitement des signaux 4 (LTS4)
- Tags:
- Master, AI, Saturna,
- Memberships:
- Vincent Jung – Role: Owner,
Joint Node-Edge Constrained Graph Generation
- ID: 33
- Status: ongoing
- Created At: 2025-04-28
A growing body of work has shown that incorporating domain knowledge into diffusion models by constraining their trajectories can significantly improve sample quality across various data modalities [1,2]. While similar ideas have been explored for graph generation, where discrete diffusion-based models have achieved state-of-the-art performance [3], existing approaches still exhibit notable shortcomings. For instance, PRODIGY [4] constrains graph generation by relaxing the adjacency matrix to a continuous space, which fails to fully capture the categorical nature of graphs. ConStruct [5], in contrast, operates entirely in discrete spaces but only imposes constraints at the structural level, without enforcing joint conditions between node and edge attributes. Such joint constraints are critical in many applications—for example, in molecular generation, where validity requires consistency between node types and their incident edges (e.g., a carbon atom must have degree at most 4). In this project, we aim to develop a graph diffusion method that natively enforces joint node-edge constraints during generation, with the goal of advancing state-of-the-art performance in real-world tasks such as molecular design.
Requirements:
Mandatory: At least one deep learning course + prior experience with PyTorch.Plus: Familiarity with graph deep learning and/or diffusion models.
[1] Lou et al., Reflected Diffusion Models, ICML 2023.
[2] Fishman et al., Diffusion Models for Constrained Domains, TMLR 2023.
[3] Vignac et al., DiGress: Discrete Denoising Diffusion for Graph Generation, ICLR 2023.
[4] Sharma et al., Plug-and-Play Controllable Graph Generation with Diffusion Models, ICML 2024.
[5] Madeira et al., Generative Modelling of Structurally Constrained Graphs, NeurIPS 2024.
- Project Creator:
- Manuel Madeira ([email protected])
- Units Involved:
- Laboratoire de traitement des signaux 4 (LTS4)
- Tags:
- Master, EDEE, EDIC, Data & Network Sciences, AI,
- Memberships:
- Alba Carballo Castro – Role: Supervisor, Manuel Madeira – Role: Owner,
Mechanistic interpretability in RNA Language Models
- ID: 125
- Status: ongoing
- Created At: 2026-03-25
- End Date: 2026-09-20
Language Models are inherently black-box models. While interpreting LLMs trained on natural language is an active area of research, particularly in the field of mechanistic interpretability, little work has been done in biological language models, let alone RNA Language models (RNA-LMs). Yet, having a strong understanding of what RNA-LMs capture is important for multiple reasons: a lack of interpretability is one of the main bottlenecks to widespread adoption, our lack of knowledge of what models do and do not capture is limiting our ability to improve them, and it holds the possibility of finding new biology.
The goal of this project is to study the applicability of existing MechInt method in RNA-LMs, identify the most promising framework for our use case, train an interpretability model on existing RNA-LMs and use extensive annotation datasets to relate sparse features to known biological knowledge.
This project can be a Master semester project or a Master's thesis. Candidates should have strong mathematical and computational skills and should be familiar with the pytorch framework. Knowledge in Mechanistic interpretability is an advantage. Candidates do not necessarily have to have a biological background but should have a strong desire to directly work with biologists, bioinformaticians and computer scientists.
- Project Creator:
- Vincent Jung ([email protected])
- Units Involved:
- Laboratoire de traitement des signaux 4 (LTS4)
- Tags:
- Master, AI, Saturna,
- Memberships:
- Vincent Jung – Role: Owner,
Real-World Benchmarks for Undirected Graph Generation
- ID: 37
- Status: ongoing
- Created At: 2025-06-06
Recent advances in diffusion and flow-based models have significantly improved graph generation performance, particularly on simple synthetic and standard molecular datasets [1][2][3]. However, these gains appear to saturate when evaluated on existing benchmarks, which may be partly attributed to the lack of sufficiently challenging and realistic evaluation settings for undirected graph generation [4].
This project aims to construct new real-world application-based datasets and benchmarks current state-of-the-art models [1,2,3]. While many domains can be considered, we focus on two: structured biological sequences, such as RNA and DNA graphs [5], and benchmarks derived from combinatorial optimization problems [6]. To better evaluate graph generation in these settings, we will later develop domain-specific metrics tailored to each task.
References:[1] Vignac et al., DiGress: Discrete Denoising diffusion for graph generation, ICLR 2023.[2] Eijkelboom et al., Variational Flow Matching for Graph Generation, NIPS 2024.[3] Qin et al., DeFoG: Discrete Flow Matching for Graph Generation, ICML 2025.[4] Bechler-Speicher et al., Position: Graph learning will lose relevance due to poor benchmarks, Arxiv preprint.[5] Booy et al., RNA secondary structure prediction with convolutional neural networks, BMC Bioinformatics 2022.[6] Cappart et al., Combinatorial optimization and reasoning with graph neural networks, IJCAI 2023.
- Project Creator:
- Yiming Qin ([email protected])
- Units Involved:
- Laboratoire de traitement des signaux 4 (LTS4)
- Tags:
- Master, EDEE, EDIC, Data & Network Sciences, AI,
- Memberships:
- Yiming Qin – Role: Owner, Manuel Madeira – Role: Supervisor,
Single-Input Multiple-Output Model Merging: Leveraging Foundation Models for Multi-Task Learning
- ID: 6
- Status: ongoing
- Created At: 2024-04-23
The advent of foundation models has revolutionized the landscape of machine learning, introducing a new paradigm where practitioners can access pre-trained checkpoints, such as those available on platforms like Huggingface, tailored for specific tasks. These models are derived from the same initial checkpoints but are fine-tuned on specific tasks, such as CIFAR or MNIST. Task arithmetic techniques [1,2] merge these different models into one multi-task model, i.e., a model that showcases good performance on all involved tasks, without needing additional training.
While the task arithmetic literature has focused on merging of classification models fine-tuned on different inputs, more traditional multi-task learning settings remain unexplored. An important case, for instance, is of a model solving the tasks of semantic segmentation, instance segmentation and depth regression from a single image input [3]. The goal of this project is to leverage existing model merging techniques for the single-input multiple-output case.
References:[1] G. Ilharco, M. T. Ribeiro, M. Wortsman, L. Schmidt, H. Hajishirzi, and A. Farhadi, “Editing models with task arithmetic,” in ICLR 2023. [2] P. Yadav, D. Tam, L. Choshen, C. A. Raffel, and M. Bansal, “TIES-Merging: Resolving interference when merging models,” in NeurIPS 2023.
[3] A. Kendall, Y. Gal, and R. Cipolla, “Multi-task learning using uncertainty to weigh losses for scene geometry and semantics,” in CVPR 2018
Requirements:Applicants must have completed at least one deep learning course and have experience with PyTorch. Familiarity with multi-task learning and model merging techniques is preferred.
Contact: [email protected], [email protected]
- Project Creator:
- Nikolaos Dimitriadis ([email protected])
- Units Involved:
- Laboratoire de traitement des signaux 4 (LTS4)
- Tags:
- Master, EDEE, AI,
- Memberships:
- Ke Wang – Role: Supervisor, Nikolaos Dimitriadis – Role: Owner,
Specialist2Generalist: transfer knowledge from small specialized models to a large foundation model
- ID: 12
- Status: ongoing
- Created At: 2024-10-30
- End Date: 2025-01-01
This project tackles the challenge of integrating valuable knowledge from specialized models trained on sensitive or proprietary datasets into larger, generalist models. In fields like medicine, where data privacy and regulatory constraints limit data sharing, a small specialized model might excel at diagnosing rare diseases based on limited data. This project aims to leverage such specialized models to improve the performance of broad, generalist models. By applying innovative techniques like unsupervised reverse knowledge distillation [1][2] and model merging[3][4], this approach aims to enhance generalist models’ capabilities while adhering to privacy concerns. Engaging in this project will provide hands-on experience with cutting-edge methods in machine learning and offer a chance to contribute to practical solutions in data-sensitive domains.
References:
[1] Sahar Almahfouz Nasser, Nihar Gupte, and Amit Sethi. Reverse knowledge distillation: Training a large model using a small one for retinal image matching on limited data. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pages 7778–7787, 2024.
[2] Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531, 2015.
[3] G. Ilharco, M. T. Ribeiro, M. Wortsman, L. Schmidt, H. Hajishirzi, and A. Farhadi, “Editing models with task arithmetic,” in ICLR 2023.
[4] Guillermo Ortiz-Jimenez, Alessandro Favero, and Pascal Frossard. Task arithmetic in the tangent space: Improved editing of pre-trained models. Advances in Neural Information Processing Systems, 36, 2024.
- Project Creator:
- Ortal Yona Senouf ([email protected])
- Units Involved:
- Laboratoire de traitement des signaux 4 (LTS4)
- Tags:
- Master, Learning & Decision Systems, AI,
- Memberships:
- Ortal Yona Senouf – Role: Owner, Nikolaos Dimitriadis – Role: Supervisor,
Temporal Graph Learning meets Neural ODE
- ID: 124
- Status: ongoing
- Created At: 2026-03-24
Graph Neural Networks (GNNs) have been established as a very competitive modeling in a wide range of applications due to their inductive bias and expressivity. However, their extension to temporal data becomes very challenging as the spatial and temporal updates are generally entangled. In this project we try to investigate how Neural ODE formulation helps us to learn latent graph trajectories, enabling interpretabilty of the flow field and learning in scarce data setting.
- Project Creator:
- Jérémy Jean Philippe Baffou ([email protected])
- Units Involved:
- Laboratoire de traitement des signaux 4 (LTS4)
- Tags:
- Master, EDIC, SIN, SMA, SSC, SSIE, SSV, Data & Network Sciences, Learning & Decision Systems, Theory and Simulation, AI,
- Memberships:
- Jérémy Jean Philippe Baffou – Role: Owner,
Tokenization in RNA Language Models
- ID: 127
- Status: ongoing
- Created At: 2026-03-26
- End Date: 2026-09-20
Tokenization in most RNA-Language Models (RNA-LMs) is done at the nucleotide-level. Attempts to adopt methods from the NLP literature such as Byte-Pair Encoding have been shown to lead to worse performance on downstream tasks [1]. Yet, moving away from nucleotide-level tokenization can greatly increase the effective input length of transformer-based RNA-LMs as well as create better inductive biases for generalization where the model can learn to decompose new samples into previously seen semantic units that are more meaningful than sole nucleotides.
The goal of this project is to develop a new dynamic tokenization method specifically designed with RNA in mind, which incorporates known biological priors.
This project can be a Master semester project or a Master's thesis. Candidates should have strong mathematical and computational skills and should be familiar with the pytorch framework. Candidates do not necessarily have to have a biological background but should be willing to learn some fundamentals of RNA biology and work with bioinformaticians.
[1] Morales-Pastor, Adrián, et al. "Character-level Tokenizations as Powerful Inductive Biases for RNA Foundational Models." arXiv preprint arXiv:2411.11808 (2024).
- Project Creator:
- Vincent Jung ([email protected])
- Units Involved:
- Laboratoire de traitement des signaux 4 (LTS4)
- Tags:
- Master, AI, Saturna,
- Memberships:
- Vincent Jung – Role: Owner,
Topological enhanced graph neural network for tumor micro-environment modeling
- ID: 35
- Status: ongoing
- Created At: 2025-05-26
- End Date: 2026-01-31
Tumor micro-environment (TME) has been shown to be a major factor in lesion development, drug resistance and disease trajectory. However, the wide range of cell types and interactions taking place in the TME make its analysis challenging. Graph Neural Networks (GNN) were proposed as powerful methods to tackle the obstacles faced in TME analysis, while providing an interpretable framework and being able to encode previous knowledge in their architecture. Despite their promising capabilities, GNNs struggle to capture higher order spatial patterns and long range interaction. In this work, we are using methods from topological data analysis (TDA) to enhance GNNs in the context of TME modelling and analysis.
- Project Creator:
- Jérémy Jean Philippe Baffou ([email protected])
- Units Involved:
- Laboratoire de traitement des signaux 4 (LTS4)
- Tags:
- Master, Data & Network Sciences, Imaging & Vision, Learning & Decision Systems, Personalized Health, AI,
- Memberships:
- Jérémy Jean Philippe Baffou – Role: Owner, Pascal Frossard – Role: Supervisor,
Robust Automatic Pathological Speech Detection
Deep learning (DL)-based pathological speech detection approaches are gaining popularity as diagnostic tools to support the time-consuming and subjective clinical auditory-perceptual assessments by speech pathologists [1]. While these approaches perform well in controlled environments with clean recordings, their performance significantly degrades in realistic scenarios with acoustic artifacts during inference [3]. Additionally, the existence of noise disparity between classes in the training data results in DL-based pathological speech detection approaches capturing noise-discriminant cues rather than pathology-discriminant ones, causing the network to learn spurious features [2]. These limitations hinder the deployment of these approaches in realistic clinical settings. Motivated by these challenges, this work focuses on the development of robust (to recording artifacts [3], spurious correlation [4], etc.) automatic pathological speech detection approaches.
References:
[1] P. Janbakhshi and I. Kodrasi, “Experimental investigation on STFT phase representations for deep learning based dysarthric speech detection,” in Proc. IEEE International Conference on Acoustics, Speech and Signal Processing, Philadelphia, USA, May 2022, pp. 6477–6481.
[2] G. Schu, P. Janbakhshi, and I. Kodrasi, “On using the UA-Speech and TORGO databases to validate automatic dysarthric speech classification approaches,” in Proc. IEEE International Conference on Acoustics, Speech and Signal Processing, Rhodes Island, Greece, 2023.
[3] M. Amiri and I. Kodrasi, “Test-time adaptation for automatic pathological speech detection in noisy environments,” in 2024 European Signal Processing Conference (EUSIPCO). IEEE, 2024.
[4] M. Amiri and I. Kodrasi, “Suppressing noise disparity in training data for automatic pathological speech detection,” in International Workshop on Acoustic Signal Enhancement (IWAENC). IEEE, 2024.
Requirements:
Experience with PyTorch/Python. At least one course on Machine Learning and Deep Learning
Contact: [email protected]
Physics-Informed Graph Neural Networks for High-Resolution Mesh-Based Simulations
Graph neural networks (GNNs) have proven highly effective in modeling mesh-based simulations, offering data-driven alternatives to classical numerical solvers for physical systems [1]. These approaches can approximate complex systems governed by partial differential equations with impressive speed-ups over traditional methods. However, many existing methods, such as MeshGraphNets, struggle to scale to high-resolution simulations due to the computational bottleneck introduced by message passing over large graphs and the challenge of maintaining accuracy across diverse scales [2].
Building on recent advancements like MultiScale MeshGraphNets [2], this project aims to explore and address the challenges of high-resolution simulations using physics-informed GNNs. Specifically, the project will focus on:
Efficient Representation of Dynamics: Investigating whether accurate surrogate dynamics can be learned on coarser mesh representations to alleviate the message-passing bottleneck, with a focus on incorporating physics-inspired priors to improve both the efficacy and efficiency of the representations.
Hierarchical Modeling: Developing or extending multi-scale approaches that combine fine and coarse representations to improve accuracy and computational efficiency.
The initial task will center on modeling wind-tunnel dynamics, providing a practical and interpretable testbed for evaluating scalability and accuracy. The project may also generalize to broader physical simulation tasks.
References:
[1] Pfaff, T. et al., “Learning Mesh-Based Simulation with Graph Networks,” International Conference on Learning Representations, 2021.
[2] Fortunato, M. et al., “MultiScale MeshGraphNets,” AI4Science Workshop (ICML), 2022.
Requirements:
At least one deep learning course + prior experience with PyTorch.
Contact: [email protected]
Interpretable Deep Learning towards cardiovascular disease prediction
Cardiovascular disease (CVD) is the leading cause of death in most European countries and is responsible for more than one in three of all potential years of life lost. Myocardial ischemia and infarction are most often the result of obstructive coronary artery disease (CAD), and their early detection is of prime importance. This could be developed based on data such as coronary angiography (CA), which is an X-ray based imaging technique used to assess the coronary arteries. However, such prediction is a non-trivial task, as i) data is typically noisy and of small volume, and ii) CVDs typically result from the complex interplay of local and systemic factors ranging from cellular signaling to vascular wall histology and fluid hemodynamics. The goal of this project is to apply advanced machine learning techniques, and in particular deep learning, in order to detect culprit lesions from CA images, and eventually predict myocardial infarction. Incorporating domain specific constraints to existing learning algorithms might be needed.
References:
[1] Yang et al., Deep learning segmentation of major vessels in X-ray coronaryangiography, Nature Scientific Reports, 2019.
[2] Du et al., Automatic and multimodal analysis for coronary angiography: training and validation of a deep learning architecture, Eurointervention 2020.
Requirements:
Good knowledge of machine learning and deep learning architectures. Experience with one of deep learning libraries and in particular Pytorch is necessary.
Contact: [email protected]
Cell-Graph Analysis with Graph Neural Networks for Immunotherapy
With the advance of imaging systems, reasonably accurate cell phenomaps, which refer to the spatial map of cells accompanied by cell phenotypes, have become more accessible. As spatial organization of immune cells within the tumor microenvironment is believed to be a strong indicator of cancer progression [1], data-driven analysis of cell phenomaps to discover new biomarkers to help with cancer prognosis is an important and emerging research area. One straightforward idea is to use cell-graphs [2], which can be later used as an input to Graph Neural Network, for example, for survival prediction [3]. However, such a dataset itself poses a lot of algorithmic and computational challenges given the big variations in both number of cells (from few tens of thousands on a slide to a few millions) and their structure, as well as the class imbalance if the objective is some sort of classification. In this project, we will explore different modeling of cell graphs for hierarchical representation learning that has a prognostic value.
References:
[1] Anderson, Nicole M, and M Celeste Simon. “The tumor microenvironment.” Current biology: CB vol. 30,16 (2020): R921-R925. doi:10.1016/j.cub.2020.06.081
[2] Yener, Bulent. “Cell-Graphs: Image-Driven Modeling of Structure-Function Relationship.” Communications of the ACM, January 2017, Vol. 60 No. 1, Pages 74-84. doi:10.1145/2960404
[3] Yanan Wang, Yu Guang Wang, Changyuan Hu, Ming Li, Yanan Fan, Nina Otter, Ikuan Sam, Hongquan Gou, Yiqun Hu, Terry Kwok, John Zalcberg, Alex Boussioutas, Roger J. Daly, Guido Montúfar, Pietro Liò, Dakang Xu, Geoffrey I. Webb, Jiangning Song. “Cell graph neural networks enable the digital staging of tumor microenvironment and precise prediction of patient survival in gastric cancer.” medRxiv 2021.09.01.21262086; doi: https://doi.org/10.1101/2021.09.01.21262086
Requirements:
Good knowledge of Python and a deep learning framework of choice (PyTorch, Tensorflow, Jax); sufficient familiarity with statistics and machine learning, also preferably Graph Neural Networks. Prior experience with DataFrame is a plus.
Contact: [email protected]
Graph Latent Diffusion Models
Graph generative models have recently undergone through huge developments mostly due to the adoption of diffusion models to the graph setting [1]. Their capability of capturing higher order relations in graph datasets has lead to impressive accurate models of complex graph distributions, with scientific applications ranging from molecular generation [1] to digital pathology [2]. Despite their remarkable expressivity, current state-of-the-art graph generative models are limited to small graph generation because the unordered nature of graphs makes it difficult to adequately exploit this property. In this project, we will develop a graph-specific latent diffusion models to solve this scaling issues. We will take inspiration from the success of latent diffusion model for image generation [5], where the diffusion process occurs at a smaller dimensionality space, thus more efficiently, and the final image is then upsampled to a high-resolution image.
References:
[1] Vignac, C. et al., “Digress: Discrete denoising diffusion for graph generation”, International Conference on Learning Representations, 2022
[2] Madeira, M. et al., “Tertiary lymphoid structures generation through graph-based diffusion”, GRAIL (MICCAI workshop), 2023
[3] Limnios, S., “Sagess: Sampling graph denoising diffusion model for scalable graph generation”, arXiv preprint arXiv:2306.16827, 2023
[4] Karami, M., “Higen: Hierarchical graph generative networks”, arXiv preprint arXiv:2305.19337, 2023.
[5] Rombach, R. et al. “High-resolution image synthesis with latent diffusion models.” Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2022.
Requirements:
Mandatory: one deep learning course (at least) and prior experience PyTorch. Prior knowledge with graph deep learning and/or diffusion models is a plus.
Contact: [email protected]
Scalable Graph Generation via Link Prediction
Generative graph models [1][2] face scalability challenges due to the need to predict the existence or type of edges between all node pairs. Some approaches use sparse graph transformers to achieve performance comparable to graph transformers, but their space complexity remains theoretically quadratic, only linearly reduced by a parameter $\lambda < 1$ [3]. To better address this issue, one possible solution is to use message-passing layers. Currently, experiments show that message-passing slightly underperforms compared to transformers [3][4]. Therefore, this project has three goals: 1) Reproduce the results of SparseDiff based on the current state-of-the-art graph generation model, 2) Enhance the existing message-passing & link prediction modules [5] or the sparse transformer [6] to match the performance of transformers, and 3) Exploration over other scalable graph generation methods such as latent diffusion [4] is also encouraged.
References:
[1] Vignac, C. et al., Digress: Discrete denoising diffusion for graph generation, International Conference on Learning Representations, 2022
[2] Qin, Madeira et al., DeFoG: Discrete Flow Matching for Graph Generation, Arxiv Preprint, 2024
[3] Qin et al., Sparse training of discrete diffusion models for graph generation, Arxiv Preprint, 2023
[4] Yang et al., Graphusion: Latent Diffusion for Graph Generation, IEEE Transactions on Knowledge and Data Engineering, 2024
[5] Cai et al., On the Connection Between MPNN and Graph Transformer, PMLR, 2023
[6] Shirzad et al., EXPHORMER: Sparse Transformers for Graphs, PMLR, 2023
Requirements:
Knowledge of Python and sufficient familiarity with statistics and machine learning. Prior experience with PyTorch is strongly recommended.
Contact: [email protected]
Benchmarking Vision Foundation Models for Digital Pathology
Large-scale self-supervised learning models, a.k.a foundation models, are becoming increasingly popular due to their ability to learn universal representations useful for a variety of downstream tasks. Such models learned from tissue images are revolutionizing the field of digital pathology [1,2], improving generalizability and transferability to a wide range of challenging diagnostic tasks and clinical workflows. As cells are the key components of these tissues, we hypothesize that modeling them as cell graphs, and then processing them with appropriate foundation models, could lead to improved performance. In this ongoing project, we aim to design such approaches and develop a comprehensive benchmark of existing vision models, taking into account several self-supervised learning strategies [3, 4]. The student, who will need to be proficient in Pytorch, will first help complete the benchmark pipeline and then explore new graph-based methods.
References:
[1] Wang, Xiyue, et al. “Transformer-based unsupervised contrastive learning for histopathological image classification.” Medical image analysis 81 (2022): 102559.
[2] Chen, Richard J., et al. “Towards a general-purpose foundation model for computational pathology.” Nature Medicine 30.3 (2024): 850-862.
[3] He, Kaiming, et al. “Masked autoencoders are scalable vision learners.” Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2022.
[4] Caron, Mathilde, et al. “Emerging properties in self-supervised vision transformers.” Proceedings of the IEEE/CVF international conference on computer vision. 2021.
Requirements:
Good knowledge of machine learning and deep learning architectures. Experience with Python and PyTorch is strongly recommended.
Contact: [email protected], [email protected], [email protected] or [email protected]
Interpretable machine learning in personalised medicine
Modern machine learning models mostly act as a black box and their decisions cannot be easily inspected by humans. To trust the automated decision-making, we need to understand the reasons behind predictions, and gain insights into the models. This can be achieved by building models that are interpretable. Recently, different methods have been proposed for data classification, such as augmenting the training set with useful features [1], visualizing the intermediate features in order to understand the input stimuli that excite individual feature maps at any layer in the model [2-3], or introducing logical rules in the network that guide the classification decision [4], [5]. The aim of this project is to study existing algorithms, which attempt to interpret deep architectures by studying the structure of their inner layer representations, and based on these methods find patterns for classification decisions along with coherent explanations. The studied algorithms will most be considered in the context of personalised medicine applications.
[1] R. Collobert, J. Weston, L. Bottou, M. M. Karlen, K. Kavukcuoglu, and P. Kuksa, “Natural language processing (almost) from scratch,”J. Mach. Learn. Res., vol. 12, pp. 2493–2537, Nov. 2011.
[2] K. Simonyan, A. Vedaldi, and A. Zisserman, “Deep inside convolutional networks: Visualising image classification models and saliency maps,” arXiv:1312.6034, 2013.
[3] L. M. Zintgraf, T. S. Cohen, T. Adel, and M. Welling, “Visualizing deep neural network decisions: Prediction difference analysis,” arXiv:1702.04595, 2017.
[4] Z. Hu, X. Ma, Z. Liu, E. Hovy, and E. Xing, “Harnessing deep neural networks with logic rules,” in ACL, 2016.
[5] Z. Hu, Z. Yang, R. Salakhutdinov, and E. Xing, “Deep neural networks with massive learned knowledge,” in Conf. on Empirical Methods in Natural Language Processing, EMNLP, 2016.
Requirements:
Familiarity with machine learning and deep learning architectures. Experience with one of deep learning libraries and good knowledge of the corresponding coding language (preferably Python) is a plus.
Contact: [email protected]
Analysis of brain networks over time
We are interested in detecting, and possibly predict, epileptic seizures using graphs extracted from EEG measurements.
Seizures occur as abnormal neuronal activity. They can affect the whole brain or localized areas and may propagate over time. The main non-invasive diagnosis tool is EEG which measures voltage fluctuations over a person’s scalp. These fluctuations correspond to the electrical activity caused by joint activation of groups of neurons. EEGs can span several hours and are currently inspected “by hand” by highly specialized doctors. ML approaches could improve this analyis, and network approaches have shown promising results.
Our data consists in multiple graphs providing a snapshot of brain activity over a time window. Considering consequent time windows, we have stochastic processes on graphs, of which we would like to identify changing points. We will learn graph representations and study their evolution over time to identify changes in regime. You are expected to compare different models in terms of performances and explainability. We are paticularly interested in inherently explainable methods, using graph features and classical time series analysis. A comparison with deep learning models could be valuable as well.
The content and workload is flexible based on the student profile and time involvement (semester project vs MSc thesis).
– Time series (preferably)
– Python (numpy, sklearn)