Academic Year 2018-19
Completion requirements
Lecture Calendar
An archive file containing the lecture slides for the year is available here.
Date | Room | Topic | References | Additional Material |
|
---|---|---|---|---|---|
1 | 21/02/2019 (14-16) |
C1 | Introduction to the course: motivations and aim; course housekeeping (exams, timetable, materials); introduction to modern pattern recognition applications |
||
2 | 22/02/2019 (11-13) | L1 | Signal processing: timeseries; time domain analysis (statistics, correlation); spectral analysis; fourier analysis. | Demo | |
3 | 28/02/2019 (14-16) | C1 | Image processing: feature descriptors (color histograms, SIFT), spectral analysis, feature detectors (edge, blobs and segments). | Additional Readings [1,2] Two high-level surveys on visual feature extraction and representation | |
4 | 01/03/2019 (11-13) | L1 | Generative and Graphical Models: probability refresher; graphical
model representation; directed and undirected models: Bayesian
networks; Markov networks; conditional independence and d-separation. | [BRML] Ch. 1 and 2 (Refresher) [BRML] Sect. 3.1, 3.2 and 3.3.1 (conditional independence) [BRML] Sect. 3.3 (Directed Models) [BRML] Sect. 4.1, 4.2.0-4.2.2 (Undirected Models) [BRML] Sect. 4.5 (Expressiveness) | Sofware |
5 | 7/03/2019 (14-16) |
C1 | Hidden Markov Models: learning in directed graphical models; forward-backward algorithm; generative models for sequential data |
[BRML] Sect. 23.1.0 (Markov Models) [BRML] Sect. 23.2.0-23.2.4 (HMM and forward backward) |
Additional Readings [3] A classical tutorial introduction to HMMs |
6 | 8/03/2019 (11-13) |
L1 | Hidden Markov Models: part 2 |
||
14/03/2019 (14-16) |
Lesson CANCELLED - Will be recovered in April |
||||
15/03/2019 (11-13) |
Lesson CANCELLED - Will be recovered in April | ||||
7 | 21/03/2019 (14-16) |
C1 | Markov Random Fields: learning in undirected graphical models; conditional random fields; pattern recognition applications |
[BRML] Sect. 4.2.2, 4.2.5 (MRF) [BRML] Sect. 4.4 (Factor Graphs) [BRML] Sect. 5.1.1 (Variable Elimination and Inference on Chain) [BRML] Sect. 9.6.0, 9.6.1, 9.6.4, 9.6.5 (Learning in MRF/CRF) |
Additional Readings [4,5] Two comprehensive tutorials on CRF ([4] more introductory and [5] more focused on vision) [6] A nice application of CRF to image segmentation Sofware Check out pgmpy: it has Python notebooks to introduce to working with MRF/CRF |
8 | 22/03/2019 (11-13) |
L1 | Boltzmann Machines: bridging neural networks and generative models; stochastic neuron; restricted Boltzmann machine; contrastive divergence |
[DL] Sections 20.1 and 20.2 | Additional Readings [7] A clean and clear introduction to RBM from its author Sofware Matlab code for Deep Belief Networks (i.e. stacked RBM) and Deep Boltzmann Machines. |
9 | 28/03/2019 (14-16) |
C1 | Bayesian Learning: non-parametric models; variational learning |
[BRML] Sect. 11.2.1 (Variational EM), 20.4-20.6.1 (LDA) | Additional Readings [8] LDA foundation paper [9] A gentle introduction to latent topic models [10] Foundations of bag of words image representation Sofware
|
10 | 29/03/2019 (11-13) |
L1 | Bayesian Learning: sampling methods (Guest Lecture by Daniele Castellana) |
[BRML] Sect. 27.1-27.3, 27.4.1, 27.6.2 |
|
M1 | 01/04/2019 (13.30-16) |
L1 |
Midterm 1 discussions |
||
11 | 08/04/2019 (14-16) |
L1 |
Learning in Structured Domain I: Recursive Neural Networks (Guest lecture by Alessio Micheli) |
||
12 | 11/04/2019 (14-16) |
C1 | Learning in Structured Domain II: Neural Networks for Graphs (Guest lecture by Alessio Micheli) |
||
13 | 12/04/2019 (11-13) |
L1 | Randomized Recurrent Neural Networks (Guest lecture by Claudio Gallicchio) |
||
M2 | 29/04/2019 (13.30-16) |
O1 |
Midterm 2 discussions | ||
14 | 02/05/2019 (14-16) |
C1 | Convolutional Neural Networks (part I): introduction to CNN; basic CNN elements |
[DL] Chapter 9 | Additional Readings [11-15] Original papers for LeNet, AlexNet, VGGNet, GoogLeNet and ResNet. |
15 | 03/05/2019 (11-13) |
L1 |
Convolutional Neural Networks (part II): CNN architectures for image recognition; convolution visualization; advanced topics (deconvolution, dense nets); applications and code |
[DL] Chapter 9 | Additional Readings [16] Complete summary of convolution arithmetics [17] Seminal paper on batch normalization [18] CNN interpretation using deconvolutions |
16 | 06/05/2019 (14-16) |
B | Adversarial Attacks & Generative Adversarial Networks (Guest seminar by Jan Göpfert) |
||
17 | 09/05/2019 (14-16) |
C1 | Deep Autoencoders: introduction to the deep learning module, sparse, denoising and contractive AE; deep RBM. |
[DL] Chapter 14, Sect 20.3, 20.4.0 (from 20.4.1 onwards not needed) |
Additional Readings [19] DBN: the paper that started deep learning [20] Deep Boltzmann machines paper [21] Review paper on deep generative models [22] Long review paper on autoencoders from the perspective of representation learning [23] Paper discussing regularized autoencoder as approximations of likelihood gradient |
18 | 10/05/2019 (11-13) |
L1 | Gated Recurrent Networks: deep learning for sequence processing; gradient issues; long-short term memory; gated recurrent units; generative use of RNN |
[DL] Sections 10.1-10.3, 10.5-10.7, 10.10, 10.11 | Additional Readings [24] Paper describing gradient vanish/explosion [25] Original LSTM paper [26] An historical view on gated RNN [27] Gated recurren units paper [28] Seminal paper on dropout regularization Sofware
|
19 | 16/05/2019 (14-16) |
C1 | Advanced Recurrent Architectures: sequence-to-sequence; attention models; multiscale network; memory networks; neural reasoning. |
[DL] Sections 10.12, 12.4.5 | Additional Readings [29,30] Models of sequence-to-sequence and image-to-sequence transduction with attention [31,32] Models optimizing dynamic memory usage (clockwork RNN, zoneout) [33] Differentiable memory networks [34,35] Neural Turing Machines and follow-up paper on pondering networks [36] Transformer networks: a paper on the power of attention without recurrence |
20 | 17/05/2019 (11-13) |
L1 | PyTorch – neural networks in Python: python; pytorch; RNN (Coding practice by Antonio Carta) |
||
21 | 23/05/2019 (14-16) |
C1 | An introduction to Tensorflow and Keras
python; numpy, tensorflow, keras; (Coding practice by Luca Pedrelli) |
||
22 | 24/05/2019 (11-13) |
L1 | Generative and Unsupervised Deep Learning: explicit distribution models; variational autoencoders; adversarial learning. (slides) |
[DL] Sections 20.9, 20.10.1-20.10.4 | Additional Readings [37] PixelCNN - Explict likelihood model [38] Tutorial on VAE [39] Tutorial on GAN (here another online resource with GAN tips) [40] Wasserstein GAN [41] Tutorial on sampling neural networks Sofware
|
23 | 30/05/2019 (14-16) |
C1 | Advanced research topics: new recurrent models, graph neural networks, tensor-based neural models and structured output prediction (Quick seminars by Antonio Carta, Daniele Castellana, Federico Errica and Marco Podda) | ||
31/05/2019 (11-13) |
L1 | Lesson cancelled due to strike |
|||
24 | 03/06/2019 (16-18) |
F1 | Final lecture: course wrap-up; research themes; final projects; exam modalities | ||
M3 | 06/06/2018 (13.30-16.30) |
L1 | Midterm 3 discussions |
Bibliography
- Scott Krigg, Interest Point Detector and Feature Descriptor Survey, Computer Vision Metrics, pp 217-282, Open Access Chapter
- Tinne Tuytelaars and Krystian Mikolajczyk, Local Invariant Feature Detectors: A Survey, Foundations and Trends in Computer Graphics and Vision, Vol. 3, No. 3 (2007) 177–2, Online Version
- Lawrence R. Rabiner:a tutorial on hidden Markov models and selected applications in speech recognition. Proceedings of the IEEE, 1989, pages 257-286, Online Version
- Charles Sutton and Andrew McCallum, An Introduction to Conditional Random Fields, Arxiv
- Sebastian Nowozin and Christoph H. Lampert, Structured Learning and Prediction, Foundations and Trends in Computer Graphics and Vision, Online Version
- Philipp Krahenbuhl, Vladlen Koltun, Efficient Inference in Fully Connected CRFs with Gaussian Edge Potentials, Proc.of NIPS 2011, Arxiv
- Geoffrey Hinton, A Practical Guide to Training Restricted Boltzmann Machines, Technical Report 2010-003, University of Toronto, 2010
- D. Blei, A. Y. Ng, M. I. Jordan. Latent Dirichlet Allocation. Journal of Machine Learning Research, 2003
- D. Blei. Probabilistic topic models. Communications of the ACM, 55(4):77–84, 2012, Free Online Version
- G. Csurka, C. R. Dance, L. Fan, J. Willamowski, and C. Bray. Visual Categorization with Bags of Keypoints. Workshop on Statistical Learning in Computer Vision. ECCV 2004, Free Online Version
- Y. LeCun, B. Boser, J.
S. Denker, D. Henderson, R. E. Howard, W. Hubbard and L. D. Jackel.
Handwritten digit recognition with a back-propagation network, Advances in Neural Information Processing
Systems,
NIPS, 1989 - A. Krizhevsky, I. Sutskever, and G. E. Hinton. ImageNet classification with deep convolutional neural networks, Advances in Neural Information Processing Systems, NIPS, 2012
- S. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition, ICLR 2015, Free Online Version
- C. Szegedy et al, Going Deeper with Convolutions, CVPR 2015, Free Online Version
- K. He, X. Zhang, S. Ren, and J. Sun. Deep Residual Learning for Image Recognition. CVPR 2016, Free Online Version
- V. Dumoulin, F. Visin, A guide to convolution arithmetic for deep learning, Arxiv
- S. Ioffe, C. Szegedy, Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift, ICML 2013, Arxiv
- M.D. Zeiler and R. Fergus, Visualizing and Understanding Convolutional Networks, ICML 2013, Arxiv
- G.E. Hinton, R. R. Salakhutdinov. Reducing the dimensionality of data with neural networks.Science 313.5786 (2006): 504-507, Free Online Version
- G.E. Hinton, R. R. Salakhutdinov. Deep Boltzmann Machines. AISTATS 2009, Free online version.
- R. R. Salakhutdinov. Learning Deep Generative Models, Annual Review of Statistics and Its Application, 2015, Free Online Version
- Y. Bengio, A. Courville, and P. Vincent. Representation learning: A review and new perspectives. Pattern Analysis and Machine Intelligence, IEEE Transactions on, Vol. 35(8) (2013): 1798-1828, Arxiv.
- G. Alain, Y. Bengio. What Regularized Auto-Encoders Learn from the Data-Generating Distribution, JMLR, 2014.
- Y. Bengio, P. Simard and P. Frasconi, Learning long-term dependencies with gradient descent is difficult. TNN, 1994, Free Online Version
- S. Hochreiter, J. Schmidhuber, Long short-term memory, Neural Computation, 1997, Free Online Version
- K. Greff et al, LSTM: A Search Space Odyssey, TNNLS 2016, Arxiv
- C. Kyunghyun
et al, Learning Phrase Representations using RNN
Encoder-Decoder for Statistical Machine Translation, EMNLP 2014, Arxiv
- N. Srivastava et al, Dropout: A Simple Way to Prevent Neural Networks from Overfitting, JLMR 2014
- Bahdanau et al, Neural machine translation
by jointly learning to align and translate, ICLR 2015, Arxiv
- Xu et al, Show, Attend and Tell: Neural Image Caption Generation with Visual Attention, ICML 2015, Arxiv
- Koutník et al, A Clockwork RNN, ICML 2014, Arxiv
- Krueger, Zoneout: Regularizing RNNs by Randomly Preserving Hidden Activation, ICLR 2018, Arxiv
- Sukhbaatar et al, End-to-end Memory Networks, NIPS 2015, Arxiv
- A. Graves et al, Neural Turing Machines, Arxiv
- A.Graves, Adaptive Computation Time for Recurrent Neural Networks, Arxiv
- A. Vaswan et al, Attention Is All You Need, NIPS 2017, Arxiv
- A. van der Oord et al., Pixel Recurrent Neural Networks, 2016, Arxiv
- C. Doersch, A Tutorial on Variational Autoencoders, 2016, Arxiv
- Ian Goodfellow, NIPS 2016 Tutorial: Generative Adversarial Networks, 2016, Arxiv
- Arjovsky et al, Wasserstein GAN, 2017, Arxiv
- T. White, Sampling Generative Network, NIPS 2016, Arxiv
Last modified: Wednesday, 12 February 2020, 5:44 PM