#### MORNING TUTORIALS

**9.00
- 12.30**

##### Tutorial #1: Deep learning for music separation

List of presenters:

- Antoine Liutkus (Inria, LIRMM, University of Montpellier, France)
- Fabian-Robert Stöter (Inria, LIRMM, University of Montpellier, France)

Abstract:

In this tutorial, we present the recent state of the art on the topic of music separation, which is a cornerstone problem for many applications in the entertainment industry. The particularity of this tutorial will be to approach the topic from both, a theoretical perspective, as well as an interactive demonstration regarding how to implement the described ideas in practice. In a first introductory part, we will summarize mandatory signal processing concepts and will review the model-based methods that were proposed in the last 40 years. This notably includes sinusoidal modeling, nonnegative matrix factorization, or kernel methods. Then, we will overview the related datasets that were proposed, as well as the metrics used for evaluation. From a practical perspective, we will see how to browse and use data in Python. In the second part, we will present deep neural networks, and most particularly the models and methods that are adequate for time series with long-term structures as in music. This involves a quick crash course on DNN and dynamic models (LSTM, CNN and wavenet), and a recap of the related vocabulary: datasets, samples, epoch, batch, loss, gradient, generative, discriminative, adversarial, etc. In the third part, we will show the importance of design and will bring a basic LSTM separation model to state-of-the-art performance. We will analyze the impact of many design choices: input representation, dimensionality reduction, depth, hidden size, context length, skip connections. This tutorial is intended to be an introduction of this topic for Ph.D. students and engineers, explaining how to obtain state of the art performance in audio separation, along with the required technical background. Second, it will hopefully be of interest to researchers wondering how to do actual investigations on audio with DNNs, without being just users of high-level black-box systems.

##### Tutorial #2: Portfolio Optimization in Financial Markets

List of presenters:

- Daniel P. Palomar (Hong Kong University of Science and Technology)

Abstract:

Financial engineering and electrical engineering are seemingly different areas that share strong underlying connections. Both areas rely on statistical analysis and modelling of systems and the underlaying time series. Either modeling price fluctuations in the financial markets or modelling, say, channel fluctuations in wireless communication systems. A core part of financial engineering is portfolio optimization, which shares a great resemblance with many problems in other areas in signal processing. This allows the use of similar mathematical tools from signal processing, optimization theory, and estimation theory like beamforming design, Kalman filtering, covariance matrix estimation, etc.

Modern portfolio theory started with Harry Markowitz's 1952
seminal paper "Portfolio Selection," for which he would later
receive the Nobel prize in 1990. He put forth the idea that
risk-adverse investors should optimize their portfolio based on a
combination of two objectives: expected return and risk. Until
today, that idea has remained central in portfolio optimization.
However, the vanilla Markowitz portfolio formulation does not seem
to behave as expected in practice and most practitioners tend to
avoid it. During the past half century, researchers and
practitioners have reconsidered the Markowitz portfolio formulation
and have proposed countless of improvements and variations, namely,
robust optimization methods, alternative measures of risk (e.g.,
CVaR or ES), regularization via sparsity, improved estimators of
the covariance matrix via random matrix theory, robust estimators
for heavy tails, factor models, mean models, volatility clustering
models, risk-parity formulations, etc.

This tutorial will explore different portfolio formulations,
starting from the basic signal model and the Markowitz portfolio
formulation and slowly increasing the level of complexity,
exploring robust portfolios, index tracking portfolios, risk-parity
portfolios, CVaR-based portfolios, and pairs-trading
portfolios.

##### Tutorial #3: Communication Networks Design: Model-Based, Data-Driven, or Both?

List of presenters:

- Alessio Zappone (University of Cassino and Southern Lazio, Italy)
- Marco Di Renzo (Paris-Saclay University, France)
- Merouane Debbah (Paris-Saclay University, France)

Abstract:

Recently, deep learning has received significant attention as a technique to design and optimize wireless communication systems and networks. The usual approach to use deep learning consists of acquiring large amount of empirical data about the system behavior and employ it for performance optimization (data-driven approach). We believe, however, that the application of deep learning to communication networks design and optimization offers more possibilities. As opposed to other fields of science, such as image classification and speech recognition, mathematical models for communication networks optimization are very often available, even though they may be simplified and inaccurate. We believe that this a priori expert knowledge, which has been acquired over decades of intense research, cannot be dismissed and ignored. In this tutorial, in particular, we put forth a new approach that capitalizes on the availability of (possibly simplified or inaccurate) theoretical models, in order to reduce the amount of empirical data to use and the complexity of training artificial neural networks (ANNs). We provide several recent examples to show that synergistically combining prior expert knowledge based on analytical models and data-driven methods constitutes a suitable approach towards the design and optimization of communication systems and networks with the aid of deep learning based on ANNs.

##### Tutorial #4: Connecting the Dots: Identifying Network Structure of Complex Data via Graph Signal Processing

List of presenters:

- Gonzalo Mateos, University of Rochester, USA
- Santiago Segarra, Rice University, USA

Abstract:

Under the assumption that the signals are related to the topology of the graph where they are supported, the goal of graph signal processing (GSP) is to develop algorithms that fruitfully leverage this relational structure, and can make inferences about these relationships even when they are only partially observed. Many GSP efforts to date assume that the underlying network is known, and then analyze how the graph's algebraic and spectral characteristics impact the properties of the graph signals of interest. However, such an assumption is often untenable in practice and arguably most graph construction schemes are largely informal, distinctly lacking an element of validation. In this tutorial we offer an overview of network topology inference methods developed to bridge the aforementioned gap, by using information available from graph signals along with innovative GSP tools and models to infer the underlying graph structure. It will also introduce the attendees to challenges and opportunities for SP research in emerging topic areas at the crossroads of modeling, prediction, and control of complex behavior arising with large-scale networked systems that evolve over time. Accordingly, this tutorial stretches all the way from (nowadays rather mature) statistical approaches including correlation analyses to recent GSP advances in a comprehensive and unifying manner. Through rigorous problem formulations and intuitive reasoning, concepts are made accessible to SP researchers not well versed in network-analytic issues. A diverse gamut of network inference challenges and application domains will be selectively covered, based on importance and relevance to SP expertise, as well as the authors' own experience and contributions.

##### Tutorial #5: Point cloud coding: The status quo

List of presenters:

- Joao Ascenso (Instituto Superior Técnico, Portugal)
- Fernando Pereira (Instituto Superior Técnico, Portugal)

Abstract:

Recently, 3D visual representation models such as light fields
and point clouds are becoming popular due to their capability to
represent the real world in a more complete and immersive way,
paving the road for new and more advanced visual experiences. The
point cloud (PC) representation model is able to efficiently
represent the surface of objects/scenes by means of a set of 3D
points and associated attributes and is increasingly being used
from autonomous cars to augmented reality. Emerging imaging sensors
have made easier to perform richer and denser point cloud
acquisitions, notably with millions of points, making impossible to
store and transmit these very high amounts of data. This bottleneck
has raised the need for efficient point cloud coding solutions that
can offer immersive visual experiences and a good quality of
experience.

This tutorial will provide a complete survey and summary of the
main PC compression techniques, along with the most relevant
foundations of this field. Regarding the content of this tutorial
is important to highlight: 1) a new classification taxonomy for PC
coding solutions to more easily identify and abstract their
differences, commonalities and relationships; 2) representative
static and dynamic PC coding solutions available in the literature,
such as octree, transform and graph based PC coding among others;
3) two MPEG PC coding solutions which have been recently developed,
notably Video-based Point Cloud Coding (V-PCC), for dynamic
content, and Geometry-based Point Cloud Coding (G-PCC), for static
and dynamically acquired content; 4) point cloud quality subjective
and objective evaluation procedures which are fundamental to
evaluate the impact and performance of several processing steps in
a PC based system, notably denoising, coding, streaming and so
on.

#### AFTERNOON TUTORIALS

**14.00
- 17.30**

##### Tutorial #6: Determinantal Point Processes in Signal Processing and Machine Learning

List of presenters:

- Simon Barthelme (CNRS, University of Grenoble, France)
- Nicolas Tremblay (CNRS, University of Grenoble, France)

Abstract:

Determinantal point processes (DPPs) are a class of point
processes that generate high-diversity subsets through repulsion.
The first examples arose in statistical physics in the work of
Wigner and Dyson, and were first theorised by O. Macchi during her
PhD, resulting in a landmark 1975 paper. DPPs then waited thirty
years to come back into the light, first at the hands of
probabilists, and around 2010 in the works of some statisticians
and computer scientists. The usefulness of DPPs in statistics and
machine learning is based on their repulsiveness and their
appealing analytic properties. "Repulsiveness", in the context of
point processes, means that points cannot be too close to one
another. It is an important feature of some natural or man-made
point processes, e.g. the position of trees in a forest; or the
position of antennas in communication networks. In machine
learning, repulsiveness is used to enforce diversity, and sampling
DPPs can be used to produce representative sets of items from a
large database. On a more theoretical side, DPPs possess attractive
analytic properties. The intensity functions (a.k.a. inclusion
probabilities in the discrete case) are known at any order
(contrary to e.g. Gibbsian point processes), and can be evaluated
as determinants of positive definite matrices. The last five years
have seen an explosion of studies in machine learning where DPP are
used and/or studied. Their use in signal processing is in its
infancy. Some applications have been reported in subsampling of
graphs; links between DPPs and zeros of the spectrogram of Gaussian
processes exist and begin to be revealed; applications in Monte
Carlo integration and survey sampling have been reported;
applications in communications are appearing.

The tutorial intends to give an impetus to research on DPPs in
Signal Processing.

##### Tutorial #7: Misspecified and Semiparametric lower bounds and their application to inference problems with Complex Elliptically Symmetric distributed data

List of presenters:

- Stefano Fortunati (University of Pisa, Italy)
- Fulvio Gini (University of Pisa, Italy)

Abstract:

Inferring information from a set of acquired data is the main
objective of any signal processing (SP) method. Developing an
estimation algorithm often begins by assuming a statistical model
for the measured data, i.e., a set of probability density functions
(pdfs), which, if correct, fully characterizes the behaviour of the
collected data/measurements. However, a certain amount of mismatch
between the true and the assumed data model is often inevitable in
practice, leading to the model misspecification problem.

The first part of the tutorial provides a comprehensive overview of
the misspecified estimation framework with a particular focus on
the Misspecified Cramér-Rao bound (MCRB). Misspecified bounds
generalize the classical bounds on the Mean Squared Error (MSE) by
allowing the true, but possibly unknown, data model and the model
assumed to derive the estimator of the parameter vector to differ,
yet establishing performance limits in a way that makes the SP
practitioner aware of the potential losses due to the model
misspecification.

The second part of the tutorial introduces a possible approach to
minimize the misspecification risk. Specifically, we address the
more general semiparametric characterization of the statistical
behaviour of the collected data. A semiparametric model is a set of
pdfs characterized by a finite-dimensional parameter vector along
with some infinite-dimensional parameter vector, i.e. a function.
We show how to obtain a MSE inequality for the estimation of the
finite-dimensional parameter vector in presence of nuisance
functions. Furthermore, we introduce the Semiparametric CRB (SCRB)
and describe its relation with the classical CRB and the MCRB.

Throughout the whole tutorial, as common thread and example to
clarify the theoretical findings, inference problems in Complex
Elliptically Symmetric (CES) distributed data will be analysed.
Particular attention will be devoted to two radar applications: the
direction of arrival estimation in CES distributed noise and the
covariance matrix estimation for adaptive radar detection in
heavy-tailed disturbance.

##### Tutorial #8: Multi-Microphone Source Localization on Manifolds

List of presenters:

- Sharon Gannot (Bar-Ilan University, Israel)
- Bracha Laufer-Goldshtein (Bar-Ilan University, Israel)
- Ronen Talmon (Technion, Israel)

Abstract:

Speech enhancement is a core problem in audio signal processing,
with commercial applications in devices as diverse as mobile
phones, hands-free systems, human-car communication, smart homes or
hearing aids. An essential component in the design of speech
enhancement algorithms is acoustic source localization. Speaker
localization is also directly applicable to many other audio
related tasks, e.g. automated camera steering, teleconferencing
systems and robot audition.

Driven by its large number of applications, the localization
problem has attracted significant research attention, resulting in
a plethora of localization methods proposed during the last two
decades. Nevertheless, robust localization in adverse conditions,
namely in the presence of background noise and reverberations,
still remains a major challenge. Until recently, the main paradigm
in localization research was based on certain statistical and
physical assumptions regarding the propagation of sound sources,
and mainly focused on robust methods to extract the direct-path.
However, valuable information on the source location can also be
extracted from the reflection pattern in the enclosure. We will
show in this tutorial that the intricate reflection patterns of the
sound source define a fingerprint, uniquely characterizing the
source location, and that meaningful location information can be
inferred from the data by harnessing the principles of manifold
learning.

In this tutorial, we will expose the audience to a new data-driven
paradigm for analyzing reverberant environments, and demonstrate
its applicability to the source localization problem. This tutorial
will be an opportunity to get familiarize with novel mathematical
tools from the field of manifold learning, including diffusion maps
theory, semi-supervised learning, optimization in reproducing
kernel Hilbert spaces, Gaussian process inference and more. We will
harness these mathematical and algorithmic models to the source
localization problem, covering the following applications:
localization with a single pair of microphones, localization using
ad hoc microphone arrays and tracking of moving sources.

##### Tutorial #9: Proximal Gradient Algorithms Applications in Signal Processing

List of presenters:

- Niccolò Antonello (Idiap Research Institute, Switzerland)
- Panagiotis Patrinos (KU Leuven, Belgium)
- Toon van Waterschoot (KU Leuven, Belgium)
- Lorenzo Stella (Amazon Development Center, Berlin, Germany)

Abstract:

Many signal processing tasks are more and more often relying on
numerical optimization. For example, in the context of compressed
sensing (CS), signals can be reconstructed by means of sparse
representations that are typically realized solving large-scale
optimization problems. In this and many others frameworks, it is of
fundamental importance to rely on efficient and reliable
optimization algorithms. One of the most popular optimization
algorithms used in CS is the proximal gradient (PG) algorithm. This
is a first-order optimization algorithm, suited for large-scale
problems and capable of addressing nonsmooth cost functions that
typically appear in CS problems.

This tutorial will focus on the PG algorithm and will illustrate
how this algorithm is not only useful in the context of CS but also
in many other frameworks that involve structured optimization
problems. In particular, the PG algorithm has several favourable
properties: its iterations consist of computationally cheap
operations that can often avoid matrix factorizations and can be
easily combined with fast transformations. Additionally,
convergence to local minima is guaranteed even in the case of
nonconvex problems. A recent accelerated variant of the PG
algorithm will be presented that substantially improves its
convergence rate using quasi-Newton methods, yet without altering
the aforementioned favourable properties.

The versatility of the PG algorithm and the efficiency of its
accelerated variant represent an attractive alternative in many
signal processing tasks where numerical optimization is necessary.
To illustrate this, different examples such as line-spectral
estimation, video background removal, audio declipping will be
presented with the aid of StructuredOptimization.jl, a new
open-source software package based on the Julia language. The
package offers not only the possibility to efficiently solve
optimization problems using the most recent variants of the PG
algorithm, but also to fully access their versatility thanks to a
high-level modeling language.

##### Tutorial # 10: Random Matrix Advances in Large Dimensional Statistics, Machine Learning and Neural Nets

List of presenters:

- Romain Couillet (Paris-Saclay University, France)
- Mohamed Seddik (Paris-Saclay University, France)
- Malik Tiomoko (Paris-Saclay University, France)

Abstract:

Machine learning algorithms, starting from elementary yet
popular ones, are difficult to theoretically analyze as (i) they
are data-driven, and (ii) they rely on non-linear tools (kernels,
activation functions). These theoretical limitations are
exacerbated in large dimensional datasets where standard algorithms
behave quite differently than predicted, if not completely
fail.

In this tutorial, we will show how random matrix theory (RMT)
answers all these problems at once. After a brief introduction to
the elementary required RMT tools, we will show that RMT provides a
new understanding and various directions of improvements for kernel
methods, semi-supervised learning, SVMs, community detection on
graphs, spectral clustering, etc. Besides, we will show that RMT
can explain observations made on real complex datasets in as
advanced methods as deep neural networks.

The outline of the tutorial is as follows:

- Part 1: Basics of Random Matrix Theory for Sample Covariance Matrices (~1h): the Stieltjes transform method; the Marcenko-Pastur law; basics of G-estimation; spiked models.
- Part 2: Applications to Large Dimensional Statistical Inference (~1h): covariance eigenvalue and eigenvector retrieval; blind source power estimation; covariance distance estimation.
- Part 3: Applications to Machine Learning and Neural Nets (~1h): kernel random matrices and applications (analysis and improvement) to sampling, kernels and sparse covariance estimation; spectral clustering; semi-supervised learning; random feature maps; neural networks; universality considerations.