Classes (up to 2025–26)¶
New curriculum from 2026–27
This page describes the class structure up to 2025–26. For the new program starting fall 2026, see Classes from 2026–27.
You may use this guide to select your UE (Unité d'Enseignement), whether you are enrolled in the [AI] track or another track of the Informatics Master's program.
All [AI] courses are open to students from other tracks; however, please ensure that you meet the required prerequisites.
If in doubt, contact the course instructors or the AI Master's program secretary or coordinators.
You can also catch up during the summer by taking one of our online crash courses.
FOR STUDENT HAVING ACCESS TO ECAMPUS, you can check the:
- M1 classes on Ecampus (from 2024/2025)
- M2 classes on Ecampus (from 2020/2021)
M1 Classes¶
Refresher classes¶
PREparatory classes, 2.5 ECTS each. Mandatory for all M1 [AI] students, except PRE3 and PRE4. Prerequisites for M2 students who want to follow M2 [AI] classes. If you cannot take them, study on your own online CRASH COURSES.
PRE1: Applied Statistics
Instructor: Stéphane Rivaud
This course introduces the necessary formalism to solve computational problems with statistical reasoning. The goal is to provide students with the necessary skills to employ statistical modeling in a computational framework to tackle real world problems.
By the end, students should be able to:
- Give a statistical description of a dataset.
- Leverage probabilistic modeling to make statistical inference.
- Analyze the mathematical properties of a statistical estimator.
- Employ bootstrapping.
- Perform hypothesis testing.
PRE2: Mathematics for Data Science
Instructor: Eva Feillet
This class aims at teaching/reminding mathematical basis useful in data science, such as:
- Vector spaces, linear transformations
- Matrices, linear systems
- Norms, orthogonality
- Eigenvalues, singular value decomposition
- Tensors (notions), multivariable calculus
Old content: online-class, website — Mathématiques pour les sciences des données
PRE3: Relational Databases
Instructor: Fatiha Saïs
This course aims enabling the students to learn the fundamentals of Relational Database Management Systems (DBMS) and apply these concepts in practice.
- Basics of relational databases
- Design theory for relational databases
- Relational algebra and SQL
- High-level database models
Old content: ecampus page (from last year) — Bases de données relationnelles, SQL
PRE4: Scientific Programming — Not offered this year
Warning
This course is not offered this year.
Foundational classes¶
Tronc commun (TC) "classic classes", 2.5 ECTS each.
To get a consistent path of study with adequate prerequisites, among optional classes, always take either green (ML path) or brown (NLP path) classes (or both).
TC0: Foundational Principles of Machine Learning
Instructor: François Landes
This course is algorithms-oriented, i.e. we will sketch the great principles of ML at first, and then focus on how algorithms work in practice, including all necessary mathematical aspects. They are the basic building blocks of more advanced algorithms.
- Gradient Descent, Linear Regression from scratch
- Classification with a single layer Perceptron, from scratch. Geometrical interpretation, a word on SGD/mini-batch learning. Discussion on the choice of the loss function, or activation functions. OVR multi-class scheme (quickly)
- Overfitting, train/validation/test split, K-fold CV, regularization: in general, L2, L1
- MAP, Bayesian interpretation of Ridge Regression or Lasso
- Feature maps ("Kernel trick"), PCA (from scratch, seen as variance maximization), PCA as pre-processing (dimensional reduction)
- Kernels, Kernelized perceptron, SVM in the separable case
Prerequisites
PRE1 and PRE2 are mandatory — you must attend them to follow TC0/FPML, except if you can argue that you are already very fluent in statistics (PRE1) and linear algebra (PRE2). PRE4 is strongly recommended. This class (or an equivalent) is a prerequisite for almost all other [AI] classes.
TC1: Machine Learning Algorithms
Instructors: Faïcel Chamroukhi, Sylvain Chevallier
This course aims at mastering the core concept of algorithmic design in ML, from an optimization or a probabilistic point-of-view, using supervised and unsupervised algorithms.
- Regression/classification seen in optimization and probabilistic frameworks, implication on batch and stochastic gradient descent
- Learning theory and Vapnick-Charvonenkis dimension
- Evaluating performances of ML algorithms in different contexts (imbalanced, small-sized, etc.)
- Probabilistic framework for machine learning: Discriminative vs Generative learning, Empirical Risk Minimization, Risk Decomposition, Bias-Variance Tradeoff; MLE, MLE and OLS in regression, MLE and IRLS in softmax classification
- Unsupervised Learning and Clustering: K-means, Mixture Models, EM algorithms
- Unsupervised Learning and Dimensionality reduction: PCA, Probabilistic PCA & EM, ICA
External site — with TC0 as prerequisite. This class is a prerequisite for OPT4 (DL).
TC2: Optimization
Instructor: Faïcel Chamroukhi
This Optimization for Machine Learning course aims at studying the mathematical and computational constructions and properties of key optimization algorithms in different situations (continuous/discrete, constrained/unconstrained, convex/non-convex, deterministic, and stochastic problems) with use case illustrations in machine learning.
- Introductions and background (convexity, differentiability, optimality conditions, convergence rates…)
- Continuous optimization (first order methods: Gradient methods, linear search, Acceleration)
- Continuous optimization (second order methods: Newton methods including Quasi-Newton, secant, IRLS)
- Constrained optimization (Equality and Inequality constraints, Duality/Lagrangian, KKT optimality conditions, Linear Programming, GD for a constrained problem)
- Non-convex, stochastic optimization (The EM Algorithm, Stochastic Gradient, Stochastic EM)
Website 2024/2025 — Optimisation, descente de gradient, etc. — with PRE2 as prerequisite
TC3: Information Retrieval
Instructor: Kim Gerdes
This course gives a basic introduction to Information Retrieval.
- Introduction to Information Retrieval: Introduction to key terms and domains; tutorial on new textual dataset indexing and basic counting techniques.
- Handling Large Datasets: Exploration of big datasets; binary evaluation methods; introduction to TF-IDF.
- Improving Retrieval Methods: Introduction to sparse embeddings; overview of BM25 and Sense2Vec.
- Advanced Embedding Techniques: Study of dense embeddings; use case with a patent dataset and its citations; introduction to doc2vec and sentenceBERT.
- Challenge Presentation: Presentation of class challenge related to information retrieval concepts.
- Project Work and Presentation: Collaborative project work and discussions; final project presentations.
With PRE1, PRE2, and OPT17 "Hands-on NLP" as prerequisite.
Growth classes¶
Formerly all OPTional classes, though some in bold are now mandatory for the [AI] track :-) 2.5 ECTS each.
OPT4: Deep Learning — Mandatory
Instructor: Thomas Gerald
The aim of this course is to introduce the Deep-Learning framework. It will cover fundamental models such as the multilayer perceptron through to the most recent deep learning architectures (CNN, VAE,...). In addition, the course will cover the different approaches to train these neural networks, with lectures dedicated to backpropagation algorithms and optimization methods based on gradient descent.
- Introduction to Neural Networks and the MLP model
- MLP and Gradient Descent algorithm
- Backpropagation algorithm and optimization methods
- Create your neural network with PyTorch
- Neural Networks architectures (CNN, AE, ...)
- Generative approaches (VAE, GAN, Denoising Diffusion models)
With TC1 as prerequisite. Highly recommended for all, mandatory for all [AI] students.
OPT8: History of AI
Instructor: Kim Gerdes
The "History of AI" course aims to give students an overview of the development of artificial intelligence from its beginnings to its current state. Students will learn about key events, major ideas, and ethical issues in AI, helping them understand how today's AI tools came to be and their impact on society.
- Introduction to AI and Its History — Overview of the course, discussion on what AI is, its current issues, and a brief look at the history of AI development.
- Philosophical Foundations of AI — Introduction to key philosophical ideas in AI, including the Turing Test, John Searle's Chinese Room argument, and Leibniz's thoughts on computation.
- Major Periods in AI Development — Overview of important times in AI history, focusing on the growth and decline of interest and funding in AI research (known as AI Summer and Winter).
- Development of AI Techniques — Exploration of different methods used in AI, the link between linguistics and AI, and how gaming technology has influenced deep learning.
- Machine Creativity and Future Predictions — Discussion on how AI can create art, music, and literature, and what the future might hold for AI, including the idea of the singularity.
- Ethics and Future Challenges in AI — Examination of ethical issues, political considerations, and future challenges in AI, focusing on the responsibilities of AI researchers and the impact on society.
OPT9: Hands-on Machine Learning
Instructor: Marc Evrard
A practical oriented class, where students apply ML techniques to simple illustrative examples and then to tackle competitive challenges. It will start with an introduction to present (refresh) the ML landscape. Classes will then be articulated to successively focus on the major concepts of practical ML.
- Introduction/refresher on ML
- Working with real data
- Discover and visualize the data to gain insights
- Prepare the data for processing
- Select and train models
- Fine-tune models
External site — with PRE1 and PRE2 as prerequisite.
OPT13: Information Theory
Instructor: Guillaume Charpiat
TD: Styliani Douka
Information theory provides a theoretical ground for ML in general.
The problems we aim at solving:
- How to complete a sequence? 1 2 3 5 7 11 13 17 ... ?? Why is 19 more "probable"? How to justify it?
- How to decide between two models for given data?
- How to set a ML problem? Which criterion to optimize? How to measure a solution's performance?
This will lead us to:
- How to quantify information?
- Is there a "natural" distribution over numbers? Over models?
- How to compress (losslessly) data? Are there bounds?
In more detail: We study the tools from Information Theory that are useful in Machine Learning. We introduce the concept of entropy, leading to a distance measure between distributions (Kullback-Leibler divergence). We then study the equivalence between compression, prediction and generation. In a third part, we get a glimpse of information geometry (Fisher metric). Last, we formalize the preference for simpler models through Kolmogorov complexity.
- Chapter 1: Entropy
- Chapter 2: Compression/Prediction/Generation equivalence
- Chapter 3: Fisher information
- Chapter 4: Kolmogorov complexity
Course website — Théorie de l'information — with PRE1 as prerequisite
OPT 15: Fairness in AI
Instructor: Adrian Popescu
This course discusses theoretical and practical aspects of fairness in AI. The objectives are to:
- Highlight the importance of building fair AI systems and analyze the legal, ethical, social and technical challenges of this process.
- Present the main types of biases appearing in AI systems and describe the existing approaches to mitigate them.
- Illustrate the practical impact of (un)fairness through impactful applications such as: personal data processing, news analysis, and face recognition.
- Develop your critical skills with hands-on analysis of recent research paper and programming activities.
OPT 16: Data Camp
Instructor: Thomas Moreau
This course aims to learn the practical tools for data science and how to frame and solve data science problems.
- Data wrangling
- The scikit-learn API and Missing values
- Metrics and unbalanced data
- Dealing with complex data
- Ensemble methods and hyperparameter optimization
Old content: Website from previous years — Create a challenge (that other students will solve as a TER project). Team work in teams of 5-6 people. — with TC0 and PRE4 (or equivalent) as prerequisite.
OPT 17: Hands-on Natural Language Processing
Instructors: Kim Gerdes, Nona Naderi
- Introduction to NLP
- Basic concepts in NLP, tokenization, lemmatization, POS tagging, ...
- Lexical semantics, word sense disambiguation
- Syntax and interpretations
- Parsing
OPT 18: Creation of a Challenge in Artificial Intelligence
Instructor: Lisheng Sun-Hosoya
This course is designed to bridge theory (lectures) and practice (TPs) by guiding students through the end-to-end creation and resolution of an AI challenge, using pre-formatted real datasets and the open-source platform Codabench.
- Designing AI Challenges
- Evaluation and Metrics
- Baselines: Core ML algorithms
- Baselines: Image Classification
- Optimizing ML Solutions
- Interpreting Results & Presenting Impact
- Project presentations
Projects and practical experiences¶
- Summer school (école thématique)
- Internship & TER (stage) — Coordination: Nona Naderi
- TER: Travail d'Etude et de Recherche (small internship). Personal work supervised by a member of the master's teaching team or in a university research lab. This work can take the form of a state of the art on a given scientific subject and/or the implementation of state algorithms for application on a given problem. This work is normally not remunerated. However, if an internship is longer than 2 months it must be remunerated. — Coordination: Nona Naderi
Other soft skills¶
- Languages: French or English for non-native speakers.
- I&E: Innovation and Entrepreneurship.
- FVE: Research and Development training, mutualised with MIAGE d'Orsay. Formation à la Vie en Entreprise.
- Teacher responsible for I&E basics/BDlabs (M1 EIT) is: Guillaume DION
M2 Classes¶
Students must choose one course from each of the following pairs: either T6A – Deep Learning for NLP or T6B – Reinforcement Learning, and either T7C – Text Mining and Chatbots or T7D – Scientific Machine Learning.
2.5 ECTS per class.
Foundational classes¶
Core Curriculum (Tronc Commun, TC): Mandatory foundational courses for all AI students.
T5B: Probabilistic Generative Models
Instructor: Bertrand Delezoide — bertrand.delezoide@gmail.com
Keywords: Notions of statistics and AI
Prerequisites: PRE1 and PRE2, or equivalent courses
- Foundations of Generative Models — definitions, probability basics, explicit vs. implicit models, first applications
- Graphical Models & HMMs — Bayesian and Markov networks, exact and approximate inference, sequential models
- Neural Generative Models (Images) — VAEs, GANs, and diffusion models for image synthesis
- Text Generation with Transformers — architecture, attention, pre-training (BERT, GPT)
- Optimization & Advanced LLM Techniques — efficient fine-tuning with methods such as LoRA
- Emerging Models — liquid neural networks, hierarchical recurrent models, state-space and quantum-inspired approaches
T5C: Signal Processing
Instructor: Matthieu Kowalski — matthieu.kowalski@universite-paris-saclay.fr, matthieu.kowalski@inria.fr
Keywords: Spectral analysis of time signals and images, convolution and filtering, time-frequency and wavelet analysis, denoising, sparse coding, inverse problems
Prerequisites: PRE2 or equivalent courses
- Introduction & Fourier Analysis
- Filtering Methods
- Random Signals and Stochastic Processes
- Time-Frequency Analysis
- Time-Scale (Wavelet) Analysis
- Introduction to Linear Inverse Problems
Growth classes¶
T5A: Frugal AI
Instructor: Adrian Popescu — adrian.popescu@cea.fr
Keywords: pretraining, large models, transfer learning, data scarcity
Prerequisites: notions of AI, deep learning, and Python programming
- Frugality in AI — overview of constraints on data, model parameters, training, inference, and memory; relation to large-scale models
- Label Frugality — supervised, weakly-supervised, semi-supervised, and self-supervised learning
- Data Frugality — transfer, few-shot, zero-shot, and continual learning
- Frugal vs. Large Deep Models — comparative analysis and practical applications
- Labs: self-supervised learning, few-shot learning, and continual learning
T5D: Image Mining
Instructor: Antoine Manzanera — antoine.manzanera@ensta-paris.fr
Keywords: image representation, mining and retrieval
Prerequisites: Linear algebra, differential calculus, probability, statistics, signal processing
- Image Models and Representations
- Feature Extraction Techniques
- Image Classification
- Visual Tracking
- Remote Sensing Applications
- Deep Learning for Visual Data
T6A: Deep Learning for NLP
Instructor: Nona Naderi — nona.naderi@universite-paris-saclay.fr
Keywords: machine learning, statistics, probability theory, Python
Prerequisites: basic ML and deep learning background
- Introduction to NLP and Language Models
- Convolutional Neural Networks for Text
- Recurrent Neural Networks
- Attention Mechanisms
- Word Representations and Transformers
- Generative Language Models
T6B: Reinforcement Learning
Instructor: Lionel Mathelin — lionel.mathelin@lisn.upsaclay.fr
Keywords: machine learning, statistics, probability theory, Python
Prerequisites: solid understanding of ML fundamentals
- Introduction to Reinforcement Learning
- Markov Decision Processes (MDPs)
- Planning by Dynamic Programming
- Model-Free Prediction and Control
- Value Function Approximation
- Policy Gradient Methods
T6C: Speech and Language Processing
Instructors: Kim Gerdes, Marc Evrard — gerdes@lisn.fr
Keywords: ongoing NLP research at LISN, linguistics and NLP integration
Prerequisites: knowledge from M1 Hands-On NLP (tokens, embeddings, annotated text data, notebooks)
- From Linguistics to NLP
- Treebanks and Oral Syntax
- Leveraging Geolocated Speech
- Speech-Audio Processing and ASR
- Emotion Detection in Speech
- Speech Interaction and Dialogue Systems
T6D: Computer Vision
Instructor: Eva Feillet — eva.feillet@protonmail.com
Keywords: linear algebra, signal processing, differential calculus, neural networks
Prerequisites: fundamentals of ML, signal processing, and calculus
- Challenges of Visual Data and Color Spaces
- Image Processing Fundamentals
- Geometry and Segmentation
- Convolutional Neural Networks (CNNs)
- Transfer Learning and Visual Representation Learning
- Transformers for Vision
T7A: Learning Theory and Advanced Machine Learning
Instructor: Antoine Cornuéjols — antoine.cornuejols@agroparistech.fr
Keywords: statistical learning theory, ensemble methods, semi-supervised learning, domain adaptation, continual learning
- Statistical Learning Theory and the Problem of Induction
- Ensemble Methods and Collaborative Learning Agents
- Learning from Imbalanced and Non-Stationary Data
- Semi-Supervised and Domain Adaptation Techniques
- Continual Learning and Knowledge Transfer
- Student Projects: critical appraisal and replication of recent research papers
T7B: Advanced Optimization and Automated Machine Learning
Instructor: Solal Nathan — solal.nathan@universite-paris-saclay.fr, solal.nathan@inria.fr
Keywords: optimization, black-box optimization, Bayesian optimization, neural architecture search, hyperparameter optimization, meta-learning, AutoML
Prerequisites: solid foundations in ML and optimization
- Introduction to Optimization for ML
- Unconstrained and Constrained Continuous Optimization
- Black-Box and Bayesian Optimization
- Hyperparameter Optimization
- Neural Architecture Search (NAS)
- Learning to Optimize and Meta-Learning
T7C: Text Mining and Chatbots
Instructors: Cyril Grouin, Thomas Gérald — cyril.grouin@limsi.fr, thomas.gerald@universite-paris-saclay.fr
Keywords: text processing, word representations, neural networks for texts, dialogue systems
Prerequisites: NLP, Python, and PyTorch
- Introduction to Text Mining
- Text Mining in Open and Medical Domains
- Semantics and Word Embeddings
- Chatbots and Evaluation Methods
- Practical Labs: chatbot development and evaluation
T7D: Scientific Machine Learning
Instructor: Alena Shilova — alena.shilova@inria.fr
Keywords: Graph Neural Networks (GNNs), NeuralODEs, Physics-Informed Neural Networks (PINNs), Geometric Deep Learning, Transformers
Prerequisites: background in deep learning and applied mathematics
- Introduction to Scientific Machine Learning
- Applications in Physics, Chemistry, and Biology
- Graph Neural Networks and Geometric Deep Learning
- NeuralODEs and PINNs
- Steerable and Equivariant Neural Networks
- Attention Mechanisms across Architectures (GAT, EquiFormer, ViT)
Soft skills¶
- Communication
- I&E: Innovation and Entrepreneurship.
- Conferences
- Teacher responsible for I&E Study (M2 EIT) is: Alvaro PINA STRANGER
Internships¶
5 to 6 month internship in a research lab or a company (Coordination: Marc Evrard and Thomas Gerald).