Academic Year 2019-20
Completion requirements
Lectures
The official language of the course is English: all materials,
references and books are in English.
Lecture slides will be made available here, together with suggested readings.
Date | Room | Topic | References | Additional Material |
|
---|---|---|---|---|---|
1 | 20/02/2019 (14-16) |
L1 | Introduction to the course: motivations and aim; course housekeeping (exams, timetable, materials); introduction to modern pattern recognition applications (slide) |
||
2 | 27/02/2020 (14-16) | L1 | Signal processing: timeseries; time domain analysis (statistics, correlation); spectral analysis; fourier analysis. (slide)[video] | Demo | |
3 | 28/02/2020 (11-13) | C1 | Image processing: feature descriptors (color histograms, SIFT), spectral analysis, feature detectors (edge, blobs and segments). (slide)[video] | Additional Readings [1,2] Two high-level surveys on visual feature extraction and representation | |
04/05/2020 | Lecture postponed due to health and prevention measures | ||||
05/05/2020 |
Lecture postponed due to health and prevention measures |
||||
4 |
12/03/2020 (14-16) |
ONLINE |
Generative and Graphical Models - Part 1: probability refresher; graphical
model representation; directed and undirected models (slide)[video, stream] |
[BRML] Ch. 1 and 2 (Refresher) [BRML] Sect. 3.1, 3.2 and 3.3.1 (conditional independence) | ofware
|
5 |
13/03/2020 (11-13) |
ONLINE |
Generative and Graphical Models - Part 2: Bayesian
networks; Markov networks; conditional independence and d-separation. [video, stream] |
[BRML] Sect. 3.3 (Directed Models) [BRML] Sect. 4.1, 4.2.0-4.2.2 (Undirected Models) [BRML] Sect. 4.5 (Expressiveness) |
|
6 |
19/03/2020 (14-16) |
ONLINE |
Hidden Markov Models: learning in directed graphical models; forward-backward algorithm; generative models for sequential data (slide) [video, stream] Note: the notebook with the derivations of this lecture (and future ones) is shared on the course Team as a OneNote pad. | [BRML] Sect. 23.1.0 (Markov Models) [BRML] Sect. 23.2.0-23.2.4 (HMM and forward backward) |
Additional Readings [3] A classical tutorial introduction to HMMs |
7 |
20/03/2020 (11-13) |
ONLINE |
Hidden Markov Models - Part 2: EM algorithm, learning as inference, Viterbi Algorithm [video,stream] |
[BRML] Sect. 23.2.6 (Viterbi) [BRML] Sect. 23.3.1-23.3.4 (EM and learning) |
|
8 |
26/03/2020 (14-16) |
ONLINE |
Markov Random Fields: learning in undirected graphical models; conditional random fields; pattern recognition applications (slides)[video,stream] |
[BRML] Sect. 4.2.2, 4.2.5 (MRF) [BRML] Sect. 4.4 (Factor Graphs) [BRML] Sect. 5.1.1 (Variable Elimination and Inference on Chain) [BRML] Sect. 9.6.0, 9.6.1, 9.6.4, 9.6.5 (Learning in MRF/CRF) |
Additional Readings [4,5] Two comprehensive tutorials on CRF ([4] more introductory and [5] more focused on vision) [6] A nice application of CRF to image segmentation Sofware Check out pgmpy: it has Python notebooks to introduce to working with MRF/CRF |
9 |
27/03/2020 (11-13) |
ONLINE |
Boltzmann Machines: bridging neural networks and generative models; stochastic neuron; restricted Boltzmann machine; contrastive divergence (slides)[video, stream] | [DL] Sections 20.1 and 20.2 |
Additional Readings [7] A clean and clear introduction to RBM from its author Sofware Matlab code for Deep Belief Networks (i.e. stacked RBM) and Deep Boltzmann Machines. |
M1 |
27/03/2020 (13.30-16) |
ONLINE |
Midterm 1 discussions |
||
10 |
03/04/2020 (14-16) |
ONLINE |
Bayesian Learning: non-parametric models; variational learning (slides)[video1,video2,video3,stream1,stream2,stream3] |
[BRML] Sect. 11.2.1 (Variational EM), 20.4-20.6.1 (LDA) |
Additional Readings [8] LDA foundation paper [9] A gentle introduction to latent topic models [10] Foundations of bag of words image representation Sofware
|
11 |
04/04/2020 (11-13) |
ONLINE |
Bayesian Learning: sampling methods (slides) [stream,video] |
[BRML] Sect. 27.1-27.3, 27.4.1, 27.6.2 |
|
12 |
16/04/2020 (14-16) |
ONLINE |
Convolutional Neural Networks (part I): introduction to CNN; basic CNN elements (slides)[stream1,stream2,video1,video2] |
[DL] Chapter 9 |
Additional Readings [11-15] Original papers for LeNet, AlexNet, VGGNet, GoogLeNet and ResNet. |
13 |
17/04/2020 (11-13) |
ONLINE |
Convolutional Neural Networks (part II): CNN architectures for
image recognition; convolution visualization; advanced topics
(deconvolution, dense nets); applications and code (stream,video) |
[DL] Chapter 9 |
Additional Readings [16] Complete summary of convolution arithmetics [17] Seminal paper on batch normalization [18] CNN interpretation using deconvolutions |
14 |
23/04/2020 (14-16) |
ONLINE |
Deep Autoencoders: introduction to the deep learning module, sparse, denoising and contractive AE; deep RBM. (slides)[stream,video] |
[DL] Chapter 14, Sect 20.3, 20.4.0 (from 20.4.1 onwards not needed) |
Additional Readings [19] DBN: the paper that started deep learning [20] Deep Boltzmann machines paper [21] Review paper on deep generative models [22] Long review paper on autoencoders from the perspective of representation learning [23] Paper discussing regularized autoencoder as approximations of likelihood gradient |
15 |
24/04/2020 (11-13) |
ONLINE |
Gated Recurrent Networks: deep learning for sequence processing; gradient issues; long-short term memory; gated recurrent units; generative use of RNN (slides)[stream,video] |
[DL] Sections 10.1-10.3, 10.5-10.7, 10.10, 10.11 |
Additional Readings [24] Paper describing gradient vanish/explosion [25] Original LSTM paper [26] An historical view on gated RNN [27] Gated recurren units paper [28] Seminal paper on dropout regularization Sofware
|
16 |
30/04/2020 (14-16) |
ONLINE |
Learning in Structured Domain I: Recursive Neural Networks (Guest lecture by Alessio Micheli) (slides)[stream1,stream2] |
||
17 |
07/05/2020 (14-16) |
ONLINE |
Learning in Structured Domain II: Neural Networks for Graphs (Guest lecture by Alessio Micheli) (slides)[stream1,stream2] |
||
M2 |
08/05/2020 (11-14.30) |
ONLINE |
Midterm 2 discussions |
||
18 |
14/05/2020 (14-16) |
ONLINE |
PyTorch – neural networks in Python: python; pytorch; RNN (Coding practice by Antonio Carta) (slides¬ebooks)[stream,video] |
||
19 |
15/05/2020 (11-13) |
ONLINE |
An introduction to Tensorflow and Keras
python; numpy, tensorflow, keras; (Coding practice by Federico Errica) (slides)[stream,video] |
||
20 |
21/05/2020 (14-16) |
ONLINE |
Advanced Recurrent Architectures: sequence-to-sequence; attention models; multiscale network; memory networks; neural reasoning. (slides)[stream,video] |
[DL] Sections 10.12, 12.4.5 |
Additional Readings [29,30] Models of sequence-to-sequence and image-to-sequence transduction with attention [31,32] Models optimizing dynamic memory usage (clockwork RNN, zoneout) [33] Differentiable memory networks [34,35] Neural Turing Machines and follow-up paper on pondering networks [36] Transformer networks: a paper on the power of attention without recurrence |
21 |
22/05/2020 (11-13) |
ONLINE |
Generative and Unsupervised Deep Learning: explicit distribution models; variational autoencoders; adversarial learning. (slides)[stream,video] |
[DL] Sections 20.9, 20.10.1-20.10.4 |
Additional Readings [37] PixelCNN - Explict likelihood model [38] Tutorial on VAE [39] Tutorial on GAN (here another online resource with GAN tips) [40] Wasserstein GAN [41] Tutorial on sampling neural networks Sofware
|
22 |
28/05/2020 (14-16) |
ONLINE |
Advanced research topics: new recurrent models, tensor-based models, continual learning, deep learning for graphs, music representation and generation (Short seminars by: Antonio Carta; Daniele Castellana; Andrea Cossu; Federico Errica; Andrea Valenti) (click on the topics to dowload the slides) |
||
23 |
29/05/2020 (11-13) |
ONLINE |
Final lecture: course wrap-up; research themes; final projects; exam modalities (slides)[stream] | ||
M3 |
04/06/2020 (14-18) |
ONLINE |
Midterm 3 discussions |
Bibliography
- Scott Krigg, Interest Point Detector and Feature Descriptor Survey, Computer Vision Metrics, pp 217-282, Open Access Chapter
- Tinne Tuytelaars and Krystian Mikolajczyk, Local Invariant Feature Detectors: A Survey, Foundations and Trends in Computer Graphics and Vision, Vol. 3, No. 3 (2007) 177–2, Online Version
- Lawrence R. Rabiner:a tutorial on hidden Markov models and selected applications in speech recognition. Proceedings of the IEEE, 1989, pages 257-286, Online Version
- Charles Sutton and Andrew McCallum, An Introduction to Conditional Random Fields, Arxiv
- Sebastian Nowozin and Christoph H. Lampert, Structured Learning and Prediction, Foundations and Trends in Computer Graphics and Vision, Online Version
- Philipp Krahenbuhl, Vladlen Koltun, Efficient Inference in Fully Connected CRFs with Gaussian Edge Potentials, Proc.of NIPS 2011, Arxiv
- Geoffrey Hinton, A Practical Guide to Training Restricted Boltzmann Machines, Technical Report 2010-003, University of Toronto, 2010
- D. Blei, A. Y. Ng, M. I. Jordan. Latent Dirichlet Allocation. Journal of Machine Learning Research, 2003
- D. Blei. Probabilistic topic models. Communications of the ACM, 55(4):77–84, 2012, Free Online Version
- G. Csurka, C. R. Dance, L. Fan, J. Willamowski, and C. Bray. Visual Categorization with Bags of Keypoints. Workshop on Statistical Learning in Computer Vision. ECCV 2004, Free Online Version
- Y. LeCun, B. Boser, J.
S. Denker, D. Henderson, R. E. Howard, W. Hubbard and L. D. Jackel.
Handwritten digit recognition with a back-propagation network, Advances in Neural Information Processing
Systems,
NIPS, 1989 - A. Krizhevsky, I. Sutskever, and G. E. Hinton. ImageNet classification with deep convolutional neural networks, Advances in Neural Information Processing Systems, NIPS, 2012
- S. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition, ICLR 2015, Free Online Version
- C. Szegedy et al, Going Deeper with Convolutions, CVPR 2015, Free Online Version
- K. He, X. Zhang, S. Ren, and J. Sun. Deep Residual Learning for Image Recognition. CVPR 2016, Free Online Version
- V. Dumoulin, F. Visin, A guide to convolution arithmetic for deep learning, Arxiv
- S. Ioffe, C. Szegedy, Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift, ICML 2013, Arxiv
- M.D. Zeiler and R. Fergus, Visualizing and Understanding Convolutional Networks, ICML 2013, Arxiv
- G.E. Hinton, R. R. Salakhutdinov. Reducing the dimensionality of data with neural networks.Science 313.5786 (2006): 504-507, Free Online Version
- G.E. Hinton, R. R. Salakhutdinov. Deep Boltzmann Machines. AISTATS 2009, Free online version.
- R. R. Salakhutdinov. Learning Deep Generative Models, Annual Review of Statistics and Its Application, 2015, Free Online Version
- Y. Bengio, A. Courville, and P. Vincent. Representation learning: A review and new perspectives. Pattern Analysis and Machine Intelligence, IEEE Transactions on, Vol. 35(8) (2013): 1798-1828, Arxiv.
- G. Alain, Y. Bengio. What Regularized Auto-Encoders Learn from the Data-Generating Distribution, JMLR, 2014.
- Y. Bengio, P. Simard and P. Frasconi, Learning long-term dependencies with gradient descent is difficult. TNN, 1994, Free Online Version
- S. Hochreiter, J. Schmidhuber, Long short-term memory, Neural Computation, 1997, Free Online Version
- K. Greff et al, LSTM: A Search Space Odyssey, TNNLS 2016, Arxiv
- C. Kyunghyun
et al, Learning Phrase Representations using RNN
Encoder-Decoder for Statistical Machine Translation, EMNLP 2014, Arxiv
- N. Srivastava et al, Dropout: A Simple Way to Prevent Neural Networks from Overfitting, JLMR 2014
- Bahdanau et al, Neural machine translation
by jointly learning to align and translate, ICLR 2015, Arxiv
- Xu et al, Show, Attend and Tell: Neural Image Caption Generation with Visual Attention, ICML 2015, Arxiv
- Koutník et al, A Clockwork RNN, ICML 2014, Arxiv
- Krueger, Zoneout: Regularizing RNNs by Randomly Preserving Hidden Activation, ICLR 2018, Arxiv
- Sukhbaatar et al, End-to-end Memory Networks, NIPS 2015, Arxiv
- A. Graves et al, Neural Turing Machines, Arxiv
- A.Graves, Adaptive Computation Time for Recurrent Neural Networks, Arxiv
- A. Vaswan et al, Attention Is All You Need, NIPS 2017, Arxiv
- A. van der Oord et al., Pixel Recurrent Neural Networks, 2016, Arxiv
- C. Doersch, A Tutorial on Variational Autoencoders, 2016, Arxiv
- Ian Goodfellow, NIPS 2016 Tutorial: Generative Adversarial Networks, 2016, Arxiv
- Arjovsky et al, Wasserstein GAN, 2017, Arxiv
- T. White, Sampling Generative Network, NIPS 2016, Arxiv
Last modified: Friday, 12 February 2021, 3:31 PM