- Recap
- Convolutional Neural Network
- Giant MLP with shared parameters.
- Stack of convolution layer, pooling layer and MLP.
- Filters are perceptions with weights and biases.
- Training CNN
- Similar to regular MLP.
- Provide examples.
- Define divergence.
.
- Gradient descent.
- Initialize all weights and biases
- For every layer l for all filter indices m, update:
- Until err has converged.
- Similar to regular MLP.
- Convolutional Neural Network
- Backpropagation
- Convolutional layer
- For every
layer filter, each position in the map in the
layer affects several positions in the map of the
layer.
- Forward computation:
- Backpropagation:
.
- Each
affects several
terms.
. Assuming indexing is from 0.
- It is obtained by flipping the filter left-right, top-bottom, and computing the inner product with respect to the square patch of
ending at (x, y). Need to use zero pad for it.
- Each
also affects several
terms for every n.
- Affect terms in only one z map.
- All entries in the map contributes to the derivative of the divergence w.r.t
.
.
- For every
- Pooling and downsampling
- Derivative of max pooling
.
- Derivative of mean pooling
.
- Derivative of max pooling
- Convolutional layer
- Transposed Convolution.
- Subsequent maps can increase in size.
- Add a layer of increased size.
- However, maintaining symmetry.
- Each neuron has the same number of outgoing weight.
- In shrinking layers, the same number of incoming weights.
- In expanding layers, the same number of outgoing weights.
- 2D Expanding convolution:t
.
- b is the stride.
- Output size is typically an integer multiple of input.
- Subsequent maps can increase in size.
- Tranform Invariance
- Problem: Currently, CNN provides shift invariance. However, require for rotation, scale, reflection invariance.
- Solution: Each filter produces a transformed invariance. Enumerated transformed.
- Other Model Variations
- Bounding estimation
- Pose estimation
- Very deep networks
- ResNet
- Depth-wise convolutions
- Each filter produce a convoluted maps and add them up. Parameter reduction.
Leave a comment