Du befindest dich hier: FSI Informatik » Prüfungsfragen und Altklausuren » Hauptstudiumsprüfungen » Lehrstuhl 5 » Deep Learning - Summer Term 2018
Deep Learning - Summer Term 2018
Examiner: Prof. Maier
- Give an overview of the topics of the lecture (I drew a mindmap of all lecture topics)
- What problems can't be solved by a perceptron? → XOR.
- What can we do to solve it?
- Universal approximation theorem. How good can we approximate? How can we better approximate?
- Why do we need deeper models when we can already model any function with one layer?
- Explain backpropagation
- What loss function did we use? Cross entropy loss (classification, multinoulli distribution), L2 loss (regression, gaussian distribution)
- Short derivation wanted for the loss functions.
- We talked about hinge loss in the lecture. Can you say something about this? → SVM
- Can you explain some different optimizers? (I talked about Momentum, NAG and Adam)
- Name some milestone architectures and say what new ingredient they introduced (I talked about /mentioned LeNet, AlexNet, Network in Network, VGG, ResNet)
- Then he wanted me to explain ResNet in detail and asked some more questions about it
- Explain LSTM. He wanted me to draw the block diagram
- Explain YOLO.
- How is segmentation working with deep learning?
- Explain autoencoders
Prof. Maier was very friendly. For preparation I wrote my own summary, this was quite helpful