CEng 783 - Deep
Learning - Fall 2020
Instructor: Asst. Prof. Dr. Emre Akbas; e-mail:
emre at ceng dot metu.edu.tr; Office: B-208; Office hours by
apppointment
Online communication: (forum,
homework submissions) https://odtuclass.metu.edu.tr/
Syllabus: pdf
Announcements
- Dec 13, 2020 - Paper
presentation schedule is up.
- Nov 10, 2020 - Initial list
of candidate papers for in-class presentations is up. The list will
be occasionally updated. Papers will be assigned in a
first-come-first-serve basis. E-mail me with your preferences.
- Oct 27, 2020 - Zoom link below will be used for the rest of the
weeks.
- Oct 19, 2020 - Zoom
link for the second meeting.
- Oct 13, 2020 - Student with these
IDs can add the class (in addition to those who are already
registered). Please let me know if you plan to drop the class.
- Oct 12, 2020 - Link for the first meeting tomorrow at 9:40: https://metu.webex.com/metu/j.php?MTID=mb3572e876f1f5103303ee52a08b890c4.
- Oct 8, 2020 - Please see my answers to frequently
asked questions.
- Sep 29, 2020 - I am receiving many e-mails about taking this course
as a special student, as an undergrad, from other
departments/universities, etc. I am not able to respond each of them. If
you want to take the course, please make sure you come to the first
lecture. Lecture hours are not scheduled yet.
Late submission policy
Any work, e.g. an assignment solution, that is submitted late past
its deadline will receive -10 points per day delayed. For example, if
the deadline is Oct 14, 23:55 and a student submits his/her work anytime
on Oct 15, that work will be evaluated over 90 instead 100.
Detailed syllabus
An important note: Lecture slides provided below are by no
means a “complete” resource for studying. I frequently use the board in
class to supplement the material in the slides.
NOTE: Below, links to slides and homework material
are broken because I removed the files. Recent versions of these files
can be found in the Fall 2021 version of the course.
Week 1
Week 2
- Lecture topics: A high-level introduction to Deep Learning (slides)
- Lecture topics: Machine learning background and basics (1 of 2) (slides)
- The link for the recorded lecture is available on ODTUClass.
Week 3
- Lecture topics: Machine learning background and basics (2 of 2) (slides)
- Colab
notebook for the in-class hands-on demo on binary classification,
gradient descent, hinge loss.
- Lecture topics: Biological neuron, artificial neuron, Perceptron (slides)
- Adding regularization to the hinge loss classifier: Colab
notebook
- Reading assignment for next week: “Chapter
8: Optimization for training deep models” from the book “Deep
Learning.”
- The link for the recorded lecture is available on ODTUClass.
Week 4
- Lecture topics: Perceptron, Multilayer Perceptrons, Backpropagation,
Activation Functions, Stochastic Gradient Descent, Momentum (slides)
- Colab
notebook on building, training and evaluating a basic MLP.
- Reading assignment for next week: LeCun,
Y., Bengio, Y., & Hinton, G. (2015). Deep learning. Nature,
521(7553), 436-444.
- The link for the recorded lecture is available on ODTUClass.
Week 5
- Lecture topics: Convolutional neural networks, convolution,
connectivity types, pooling, AlexNet, data augmentation, dropout. (slides)
- Building, training and evaluating a basic CNN: Colab
notebook
- Initial list
of candidate papers for in-class presentations is up. The list will
be occasionally updated. Papers will be assigned in a
first-come-first-serve basis. E-mail me with your preferences.
- The link for the recorded lecture is available on ODTUClass.
Week 6
- Lecture topics: Convolutional neural networks continued, multiclass
hinge loss, derivation of cross-entropy loss, notes on initializing
neural networks, implementing backpropagation in a modular way, variants
of stochastic gradient methods, adaptive learning rate methods. (slides)
- Reading assignment for next week: He, K., Zhang, X., Ren, S. and
Sun, J. Deep residual
learning for image recognition. In CVPR 2016.
- The link for the recorded lecture is available on ODTUClass.
Week 7
Week 8
- Lecture topics: Recurrent Neural Networks, LSTM, GRU. Some
applications of RNNs. (slides)
- The link for the recorded lecture is available on ODTUClass.
Week 9
- A RNN example: solving simple arithmetic operations using a
sequence-to-sequence encoder-decoder model: Colab
notebook (in Keras)
- Lecture topics: Deep generative modeling (slides)
- The link for the recorded lecture is available on ODTUClass.
Week 10
- Lecture topics: A brief intro to deep reinforcement learning (slides)
- Lecture topics: self-attention, transformers, non-local neural
networks, graph neural networks (slides)
- The link for the recorded lecture is available on ODTUClass.
Week 11
- Lecture topics: brief overview on latest trends, limitations in deep
learning. (1) Self-attention / Transformers, (2) Self-supervised /
Unsupervised learning, (3) - Out-of-distribution robustness, (4)
Non-differentiable operations & blackbox optimization, (5)
Human-level AI, (6) Better efficiency, (7) Implicit deep learning.
- The link for the recorded lecture is available on ODTUClass.
- Take home exam given.
Week 12
- Paper presentations:
- “EfficientNet:
Rethinking model scaling for convolutional neural networks”, Tan and
Le, ICML 2019 presented by Sinan Gencoglu.
- “Don’t
decay the learning rate, increase the batch size”, Smith et al.,
ICLR 2018 presented by Enes Recep Cinar.
- “On
exact computation with an infinitely wide neural net”, Arora et al.,
NeurIPS 2019 presented by Yavuz Durmazkeser.
- “Reconciling
modern machine-learning practice and the classical bias–variance
trade-off”, Belkin et al., PNAS 2019 presented by Munteha Nur
Bedir.
- “No
more pesky learning rates”, Schaul et al., ICML 2013 presented by
Arash Sadeghi Amjadi.
Week 13
- Paper presentations:
- “Dynamic
filter networks”, Jia et al., NeurIPS 2016 presented by Berke
Tezergil.
- “CondConv:
Conditionally parameterized convolutions for efficient inference”,
Yang et al., NeurIPS 2019 presented by Rza Hasanli.
- “Attention
is all you need”, Vaswani et al., NeurIPS 2017 presented by Ahmet
Sencan.
- “BERT:
Pre-training of deep bidirectional transformers for language
understanding”, Devlin et al., 2018 presented by Firat Cekinel.
- “An
Image is Worth 16x16 Words: Transformers for Image Recognition at
Scale”, Dosovitskiy et al., 2020 presented by Irmak
Akkuzuluoglu.
- “End-to-End
Object Detection with Transformers”, Carion et al., ECCV 2020
presented by Duygu Arslan.
- “Show,
attend and tell: Neural image caption generation with visual
attention”, Xu et al., ICML 2015 presented by Cahit Yildirim.
- “Continuous
control with deep reinforcement learning”, Lillicrap et al., ICLR
2016 presented by Burak Han Demirbilek.
- “Dueling
network architectures for deep reinforcement learning, Wang et al.,
ICML 2016 presented by Umut Can Gulmez.
- “Human-level
control through deep reinforcement learning”, Mnih et al., Nature
2015 presented by Melis Ilayda Bal.
- “Improved
techniques for training score-based generative models”, Song and
Ermon, NeurIPS 2020 presented by Ibrahim Koc.
- “Accelerating
deep learning by focusing on the biggest losers”, Jiang et al., 2019
presented by Tugba Tumer.
- “Green
AI’, Schwartz et al., Communications of the ACM 2019 presented by
Ali Ozkan.
- “Distilling
the knowledge in a neural network’, Hinton et al., NeurIPS workshop
2014 presented by Ali Ozkan.
- “Small
data, big decisions: Model selection in the small-data regime”,
Bornschein et al., ICML 2020 presented Nika Rasoolzadeh.
Week 14
- Paper presentations:
- “Tracking
Objects as Points”, Zhou et al., ECCV 2020 presented by Raheem
Hashmani.
- “Semi-supervised
classification with graph convolutional networks”, Kipf and Welling,
ICLR 2017 presented by Semih Kaya.
- “Adversarial
inverse graphics networks: Learning 2d-to-3d lifting and image-to-image
translation from unpaired supervision”, Tung et al., ICCV 2017
presented by Ozhan Suat.
- “Self-supervised
learning through the eyes of a child”, Orhan et al., NeurIPS 2020
presented by Arkin Yilmaz.
- “A
simple framework for contrastive learning of visual
representations”, Chen et al., ICML 2020 presented by Sarp Tugberk
Topallar.
- “SoftSort:
A Continuous Relaxation for the argsort Operator”, Prillo and
Eisenschols, ICML 2020 presented by Dogukan Cavdaroglu.
- “Smooth-AP:
Smoothing the Path Towards Large-Scale Image Retrieval”, Brown et
al., ECCV 2020 presented by Feyza Yavuz.
- “Towards
an integration of deep learning and neuroscience”, Marblestone et
al., Frontiers in Computational Neuroscience 2016 presented by Ali
Dogan.
- “PoseNet3D:
Learning Temporally Consistent 3D Human Pose via Knowledge
Distillation”, Shashank et al., 3DV 2020 presented by Bedirhan
Uguz.
- “Tasks,
stability, architecture, and compute: Training more effective learned
optimizers, and using them to train themselves”, Metz et al., 2020
presented by Bugra Kaan Demirdover.
- “A
generative adversarial approach for zero-shot learning from noisy
texts”, Zhu et al., CVPR 2018 presented by Cemal Erat.
- “The
lottery ticket hypothesis: Finding sparse, trainable neural
networks”, Frankle and Carbin, ICLR 2018 presented by Cagdas
Cayli.
- “Decoupled
weight decay regularization”, Loshchilov and Hutter, ICLR 2017
presented by Fatih Acun.
- “GhostNet:
More features from cheap operations”, Han et al., CVPR 2020
presented by Ali Ilker Sigirci.
Frequently
Asked Questions about taking the course
I am receiving too many e-mails about taking the course.
Unfortunately, I cannot reply them one by one. Below are answers to some
common questions.
Q1: Can I take the course?
A1: There is a huge demand for the course from all kinds of
backgrounds. Thanks. However, I have the responsibility to evaluate your
learning outcomes and grade you. Therefore, I need to limit the number
of seats. Based on my previous years’ experience, this limit will be
around 40-45.
Since this is a graduate METU CENG course, I need to give priority
for the graduate students in our department. In fact, here is the
priority order that I will use to accept students to the class. From
high to low priority:
- Grad students from METU CENG,
- A limited number of 4th year undergraduate students from METU
CENG,
- Grad students from other METU departments,
- Special students (see http://oidb.metu.edu.tr/ozel-ogrenci,
you need to be a grad student in some other university to be
eligible).
I must note that the first two categories (METU CENG students) almost
fill up the whole capacity. So, unfortunately, there will not be much
room for the remaining two categories.
Also, precedence will be given to students who are actively doing
research in machine learning and related areas. This course is not a
PyTorch or Keras tutorial, we intend to go beyond the “user” level.
You might want to check out the other two DL courses given at Multimedia
Informatics and Electrical
Engineering departments.
Machine learning background is required. If you have not taken a
machine learning course before, please do not take this course.
Fluency in Python is required.
Q2: How can I register for
the course?
A2: Come to the first class. I will publicly announce the lecture
link on this page. In the first class, I will collect information from
the participants and then, will decide (based on my answer A1 above) on
who will be able to register. This enrollment list will be announced in
a couple of hours following the first class. Students listed in this
enrollment list will be able to add the course during the add-drop
period.
Q3:
I was able to take the course during the regular interactive
registration. Should I worry about not being accepted?
A3: No worries. You will stay.
Q4: Can I take
this course as a special student?
A4: Possible but unlikely. Please see my answer A1 above.
Q5:
The deadline for submitting special student papers is Oct 12 but your
first class in on Oct 13.
A5: I know. In previous years, the deadline for submitting special
student papers was the last day of the first week but this year it the
first day of first week. Unfortunately, this is out of my control.
Q6:
Even if I don’t officially register for the class, can I audit it?
A6: I always welcome students to just come and listen. I am thinking
on how to best implement this.