SGN-90206 International Doctoral Seminar in Signal Processing, 1-8 cr

Implementation SGN-90206 2016-03

Description

Learning sparse representations for image and signal modeling The main goal of this course is to provide the student with an understanding of the most important aspects of the theory underlying sparse representation and, more in general, of sparsity as a form of regularization in learning problems. Students will have the opportunity to develop and understand the main algorithms for learning sparse models and computing sparse representations. These methods have wide applicability in computer science, and these will be a useful background for their research. In particular, this course aims at: - Presenting the most important aspects of the theory underlying sparse representations, and in particular the sparse-coding and dictionary-learning problems. - Illustrating the main algorithms for sparse coding and dictionary learning, with a particular emphasis on solutions of convex optimization problems that are widely encountered in engineering. - Providing a broad overview of the applications involving sparse representations, in particular those in the imaging field.

Lessons

Period 5
Methods of instruction
Person responsible Alessandro Foi

Assessment scale

Evaluation scale passed/failed will be used on the course

Requirements

Active participation to the lectures (2cr). Optional project work where the participant has to apply the learned methods to a problem from her/his own field of research (2cr).

Additional information of course implementation

In recent years, a large amount of multidisciplinary research has been conducted on sparse representations i.e., models that describe signals/images as a linear combination of a few atoms belonging to a dictionary. Sparse representation techniques had lately an exponential growth and now they represent one of the leading tools for solving ill-posed inverse problems in diverse areas, including image/signal processing, statistics, machine learning, computer vision. The course presents the basic theory of sparse representations and dictionary learning, and illustrates the most successful applications of sparse models in imaging and signal processing. Particular emphasis will be given to modern proximal methods for solving convex optimization problems such as \ell-1 regularisation (e.g., BPDN, LASSO). Recent developments of sparse models will be also overviewed. Detailed program: Basics on linear orthonormal representations Redundant representations and Sparse Coding (Minimum \ell-0 norm) - Sparse Coding w.r.t redundant dictionaries. Greedy Algorithms: Matching Pursuit, Orthogonal Matching Pursuit. - An overview of theoretical guarantees and convergence results. Convex Relaxation of the Sparse Coding Problem (Minimum \ell-1 norm) - Sparse coding as a convex optimization problem, BPDN (LASSO), connections with \ell-0 solutions. - Notes on other norms, visual intuition. Theoretical guarantees. - Minimum \ell-1 Sparse Coding Algorithms: Iterative Reweighted Least Squares, Proximal Methods, Iterative Soft Thresholding, ADMM. Dictionary Learning - Dictionary Learning Algorithms: Gradient Descent, MOD, KSVD Sparsity in engineering applications - Main problems involving sparsity as a regularization prior: Denoising, Inpainting, Super-resolution, Deblurring. - Dictionary Learning and Sparse Coding for Classification and Anomaly Detection


Study material

Type Name Author ISBN Additional information Language Examination material
Book Sparse and Redundant Representations M. Elad English No
Summary of lectures Sparse Representations: Theory and Applications B. Wohlberg Notes of the course held in 2013 at Politecnico di Milano. English No