Sie befinden sich hier: Termine » Prüfungsfragen und Altklausuren » Hauptstudiumsprüfungen » Lehrstuhl 5 » dl-august-19 (Übersicht)

First he wanted me to give an overview over the topics of the lecture by drawing a mindmap

Then he wanted me to explain the Perceptron

=> Schematic => XOR → MLP => Universal Function Approx => Multiple Layer: Deep Learning

Name a loss functions that we discussed?

=> L2 for regression, CE for classification, both optimal wrt. MLE => Derivation for L2 by assuming Gauss

How does the network learn now?

=> Backpropagation + gradient descend => Chain rule: Multiplication of gradients + weight update => Exploding/vanishing gradients

What other method did we use to encode the Information? (Not quite sure about the wording here)

=> Activation Functions: Sigmoid/Tanh → ReLU => prevent vanishing gradients

What about Dying ReLU?

=> Leaky ReLU

What is an regularization alternative fighting the internal covariate shift?

=> Batchnormalization

Can you draw and explain the LSTM structure?

What are GANs?

=> Generator vs Discriminator trained by MiniMax

There was a problem called Mode Collapse, please explain it.

=> TODO => D focuses only on one feature → G also

Can you explain Cycle Consistent GANs?

=> Principle(trainiable inverse mapping) + combined loss function explained

What is an Autoencoder?

=> Encoder-Decoder => Undertetermined AE + Sparse AE

What is common for U-Net and Autoencoder? Whats the difference?

=> Same: Encoder-Decoder structure => Different: (Conv Layer), but mainly skip connections

How good would U-Net be in comparison to the AE?

=> thanks to the skip connections it could basically copy the input