If you are interested in doing a research project (“semester project”) or a master’s project at IVRL, you can do this through the Master’s Programs in Data Science or in Computer Science. Note that you must be accredited to EPFL. This page lists available semester/master’s projects for the Spring 2026 semester. The order of the projects is random.
For any other type of applications (research assistantship, internship, etc), please check this page.
Description
Recent work on weight space representation learning [1] has demonstrated that multiplicative LoRA (mLoRA) weights exhibit remarkable structural properties: when combined with asymmetric masking, mLoRA weights converge to linear modes during optimization, meaning different random initializations lead to nearly identical weight configurations. This linear mode connectivity, coupled with preserved channel alignment and semantic structure, makes mLoRA an ideal candidate for meta-learning tasks that operate directly in weight space.
Traditional weight space learning faces fundamental challenges: neural network weights are ambiguous due to permutation symmetry, and functionally identical networks can be arbitrarily far apart in parameter space. The mLoRA formulation addresses these challenges by:
- Constraining optimization to a structured subspace via the pre-trained base model
- Preserving channel alignment through multiplicative (rather than additive) updates
- Eliminating permutation symmetry via asymmetric masking
These properties suggest that mLoRA weights can serve as well-behaved representations for meta-learning tasks such as weight alignment/merging, performance prediction, and membership inference attacks, all of which benefit from structured, semantically meaningful weight spaces.
Type of work:
- MS Level: master project
- 100% Research
Approach
This project will systematically evaluate mLoRA weight representations on meta-learning benchmarks, leveraging the discovered linear mode connectivity and semantic structure. The student should:
- Familiarize: Study multiplicative LoRA formulation and weight space learning fundamentals
- Understand the difference between additive LoRA and multiplicative LoRA
- Study why multiplicative LoRA preserves channel alignment (Corollary 1 in [1])
- Review existing weight space learning literature [2] and understand key challenges (permutation symmetry, loss landscape geometry)
- Task Selection and Evaluation: Choose 1-2 of meta-learning tasks to focus on, some examples are:
-
- Weight Alignment / Merging: Since mLoRA-Asym weights converge to linear modes, simple averaging should produce functional merged models. Evaluate:
- Task arithmetic [3]: Can arithmetic operations (addition, negation) in mLoRA weight space transfer capabilities?
- Multi-task merging: Can mLoRA weights from different tasks be merged without alignment?
- Compare against existing alignment methods (Git Re-Basin [4], optimal transport)
- Model Manipulation / Editing: The semantic structure of mLoRA weights suggests they could be used to manipulate or edit models in interpretable ways. Building on [5], which discovers semantic linear directions in LoRA weight space of customized diffusion models, evaluate:
- Can we identify linear directions in mLoRA space that correspond to semantic attributes?
- Can we edit model behavior by moving along these directions (e.g., adding/removing concepts)?
- Does mLoRA’s improved structure yield more disentangled and interpretable editing directions compared to additive LoRA?
- Explore weight space interpolation for smooth transitions between model behaviors
- Membership Inference Attack (Data Usage Detection): Since mLoRA weights encode semantic structure of training data, they may leak membership information. Evaluate:
- Train classifiers to detect whether specific samples were in the training set
- Analyze what information is encoded in different weight components
- Weight Alignment / Merging: Since mLoRA-Asym weights converge to linear modes, simple averaging should produce functional merged models. Evaluate:
- Experimental Design:
- Extend to different model types: image diffusion models, language models, classification models, etc.
- Design appropriate baselines: additive LoRA, standalone MLP weights, latent codes encoded by various weight encoders
Prerequisites
- Proficiency in Python and experience with PyTorch
- Familiarity with neural network optimization and loss landscapes
- Understanding of Low-Rank Adaptation (LoRA) and fine-tuning methods
- Interest in weight space learning and neural network interpretability
Supervisor
Zhuoqian (Zack) Yang, [email protected]
References
[1] Yang, Zhuoqian, Mathieu Salzmann, and Sabine Süsstrunk. “Weight Space Representation Learning with Neural Fields.” arXiv preprint arXiv:2512.01759 (2025).
[2] Schürholt, Konstantin, et al. “Neural network weights as a new data modality.” ICLR 2025 Workshop Proposals (2024). See also: https://github.com/Zehong-Wang/Awesome-Weight-Space-Learning
[3] Ilharco, Gabriel, et al. “Editing models with task arithmetic.” ICLR (2023).
[4] Ainsworth, Samuel K., Jonathan Hayase, and Siddhartha Srinivasa. “Git re-basin: Merging models modulo permutation symmetries.” ICLR (2023).
[5] Dravid, Amil, et al. “Interpreting the weight space of customized diffusion models.” arXiv preprint arXiv:2406.09413 (2024).
[6] Frankle, Jonathan, et al. “Linear mode connectivity and the lottery ticket hypothesis.” ICML (2020).
[7] Lim, Soon Hoe, et al. “An Empirical Analysis on the Linear Mode Connectivity of Neural Network.” ICLR (2024).
-
Description
Film photography continues to thrive among enthusiasts and professionals who value its unique aesthetic qualities. While film remains popular, the process of digitizing film negatives has become increasingly challenging due to the stagnation of the consumer film scanner industry. Consumer film scanners use outdated sensors, motivating photographers to use digital cameras for scanning: a superior but technically complex alternative [1].
Color negative inversion requires specialized algorithms (as compared to standard softwares like Lightroom, PhotoShop) due to fundamental differences between film and digital sensors [2]. Current methods demand technical expertise and dedicated measuring process for film characteristics (Dmin, Dmax, characteristic curves) [3], creating barriers for amateur photographers.
Our preliminary experiments show that statistical analysis can automatically estimate essential parameters from scanned images, eliminating the need for direct density measurements while maintaining quality. The prototype software is welcomed by amateur and professional photographers.
Type of work:
- Bachelor Level: Semester Project
- 100% Development
Approach
This project will continue to develop a toolkit using statistical methods to automatically estimate film parameters. The student will extend it with:
- Multi-Image Parameter Estimation: Statistical aggregation across multiple frames for improved accuracy
- Batch Processing Pipeline: Efficient whole-roll processing with frame consistency
- Faster RAW Image Loading: Incorporating the rawspeed [4] repo.
- OpenGL-based GUI: Responsive interface with real-time preview
- Algorithm Enhancement (optional): Advanced statistical methods.
Prerequisites
- Strong programming skills in Python / C++
- Familiarity with version control (Git) and collaborative development
- Helpful but not required:
- Experience with the analog processes
- Knowledge of computational photography and color science
Supervisor
Zhuoqian (Zack) Yang, [email protected]
References
[1] Tran, A. (2016, March 7). How to digitise film negatives using a DSLR. Ant Tran Blog. https://www.anttran.com/blog/2016/3/7/how-to-digitise-negatives-using-a-dslr
[2] Hunt, R. W. G. (1995). The reproduction of colour (5th ed.). Fountain Press.
[3] Patterson, R. (2001, October 2). Understanding Cineon. Illusion Arts. http://www.digital-intermediate.co.uk/film/pdf/Cineon.pdf
[4] darktable-org. (n.d.). RawSpeed [Computer software]. GitHub. https://github.com/darktable-org/rawspeed
Film photography’s distinctive “look” is partly due to its ability to record and compress light information of high dynamic range, especially in the highlights, without clipping [1]. By preserving subtle gradations in highlight and shadow areas and compressing them, film naturally reveals rich color nuances, which is a key contributor to its signature aesthetic.
Digital film emulation has become increasingly popular, but most applications (e.g., Dazz, Dehancer, VSCO) assume availability of high-quality captures, while working off of images captured by relatively limited consumer camera sensors. These images tend to have a low dynamic range and lose highlight and shadow detail that film retains, making it impossible for current emulators to reproduce nuanced tones via compression.
Our preliminary validation has confirmed that high dynamic range (HDR) data significantly improves the quality of film simulation, particularly in preserving the characteristic highlight roll-off and shadow detail that define authentic film aesthetics. This validation establishes the critical importance of recovering lost dynamic range information before applying film simulation techniques.
Type of work:
- MS Level: master’s project
- 100% Research
Approach
Building on our validated hypothesis, this project will develop a deep learning framework that recovers high dynamic range RAW-equivalent images from standard RGB inputs captured by consumer-grade sensors. We will extend the state-of-the-art RAW-Diffusion model [2] by training it on carefully designed synthetic datasets that specifically target highlight and shadow reconstruction.
Our approach involves:
- Synthetic Training Data Generation: Creating paired datasets of clipped RGB images and their corresponding full dynamic range RAW data, with special emphasis on highlight and shadow regions
- Model Architecture Extension: Adapting RAW-Diffusion’s diffusion-based architecture to focus on reconstructing missing information in over/underexposed regions
- Film Simulation Pipeline Integration: Feeding the reconstructed HDR data into physically-accurate film simulation models to achieve authentic film characteristics
The final framework should enable consumer cameras to produce images with the nuanced highlight compression, smooth tonal transitions, and rich color gradations characteristic of analog film.
Prerequisites
- Proficiency in Python and experience with PyTorch.
- Familiarity with digital imaging pipelines and RAW image formats.
- Interest in photography and knowledge of film characteristics.
Supervisor
Zhuoqian (Zack) Yang, [email protected]
References
[1] Attridge, G. G. “The characteristic curve.” The Journal of photographic science 39.2 (1991): 55-62.
[2] Reinders, Christoph, et al. “RAW-Diffusion: RGB-Guided Diffusion Models for High-Fidelity RAW Image Generation.” arXiv preprint arXiv:2411.13150 (2024).
[3] Brooks, Tim, et al. “Unprocessing images for learned raw denoising.” Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2019.
[4] Zamir, Syed Waqas, et al. “Cycleisp: Real image restoration via improved data synthesis.” Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2020.
[5] Kim, Woohyeok, et al. “Paramisp: learned forward and inverse ISPS using camera parameters.” arXiv preprint arXiv:2312.13313 (2023).
Startup company Innoview has developed a software framework to create hidden watermarks printed on paper and to acquire and decode them by a smartphone. The acquisition by smartphone comprises many separate parametrizable parts. The project consists in improving some of the parts of the acquisition pipeline in order to optimize the recognition rate of the hidden watermarks (under Android).
Deliverables:
- Report and running prototype.
Prerequisites:
- basic knowledge of image processing and computer vision,
- Coding skills in Java Android, C#, and/or Matlab
Level: BS or MS semester project
Supervisors:
Dr. Romain Rossier, Innoview Sàrl, [email protected], tel 078 664 36 44
Prof. Roger D. Hersch, BC320, [email protected], cell: 077 406 27 09
Startup company Innoview has developed arrangements of lenslets that can be used to create document security features. The goal is to improve these security features and to optimize them by simulating the interaction of light with these 3D lenslet structures, using the Blender software.
Deliverables:
- Report and running prototype (Matlab). Blender lenslet simulations.
Prerequisites:
- knowledge of computer graphics, interaction of light with 3D mesh objects,
- basic knowledge of Blender,
- Coding skills in Matlab
Level: BS or MS semester project
Supervisors:
Prof. Roger D. Hersch, BC320, [email protected], cell: 077 406 27 09
Dr. Romain Rossier, Innoview Sàrl, [email protected], tel
078 664 36 44
Startup company Innoview has developed arrangements of transparent lenslets and of opaque structures that yield interesting moiré effects.
The goal is to create plastic objects composed of a revealing layer made of transparent lenses and of a base layer made of partly opaque structures. The superposition of the two layers shows interesting moiré evolutions. Once created as 3D volumes, their aspect can be simulated in Blender. After simulation and verification, these objects are to be printed by a 3D printer.
Deliverables:
Report and running prototype (Matlab). Blender lenslet simulations. Fabricated 3D objects showing the moiré evolutions.
Prerequisites:
1. Good knowledge of computer graphics, especially the construction
of 3D mesh objects,
2. Basic knowledge of Blender,
3. Good coding skills in Matlab
Level: BS or MS semester project, master’s project
Supervisors:
Prof. Roger D. Hersch, BC320, [email protected], cell: 077 406 27 09
Dr. Romain Rossier, Innoview Sàrl, [email protected], tel 078 664 36 44