25 Jun 2018

Spectral representations for Convolutional Neural Networks

Notes on a paper:

This paper presents a very interesting idea of replacing pooling layers such as max or avg ones with spectral pooling, where the input to the pooling layer is transformed to the frequency domain via FFT, then by the heuristic that the energy of the input in the frequency domain for natural signal or images follows the inverse power law where the most information is cumulated in the low frequencies and the noise is usually represented in the high frequencies, we truncated the signal by discarding the high frequency coefficients. Finally, the truncated signal is transformed back to the time/spatial domain via the inverse FFT.

Background:

Unitary transforms (matrices)

A square matrix \(A=[A_1\;A_2\;\cdots\;A_n]\) (\(A_i\) for the ith column vector of \(A\)) is unitary if its inverse is equal to its conjugate transpose, i.e., \(A^{-1}=A^{*T}=A^{H}\), where the superscript \(^H\) denotes the Hermitian operation. In particular, if a unitary matrix is real \(A=A^*\), then \(A^{-1}=A^T\) and it is orthogonal. Both the column and row vectors ( \(A_i, i=1,\cdots,n\)) of a unitary or orthogonal matrix are orthogonal (perpendicular to each other) and normalized (of unit length), or orthonormal, i.e., their inner product satisfies: \((A_i, A_j) = A_i^{H}A_j\)

These \(n\) orthonormal vectors can be used as the basis vectors of the n-dimensional vector space.

source: lectures on Fourier Transorm

Parseval’s theorem

Fourier transform is unitary, thus the sum (or integral) of the square of a function is equal to the sum (or integral) of the square of its transform. Let \(X\) be the Discrete Fourier Transform of the sequence \(x\), then we have: \(\sum_{i=0}^{n-1} |x_i|^2 = \sum_{F=0}^{n-1}|X_F|^2\)


Tags:
0 comments