IMI: ランダム行列と深層神経回路

2023-07-12 explainer

📎 Migration note: 20 inline media block(s) (image/file) were not migrated by scripts/copy-post-bodies-to-notes.ts — Notion's API can only re-attach external URLs, not re-upload internal files. The originals remain on the legacy Posts Public page until Phase 8.

機械学習において中心的役割を果たしているのが深層神経回路である。深層神経回路を理解する時、ランダム行列が現れることがある。特に、パラメータの初期化や学習ダイナミクスの理解において、頻繁にランダム行列が現れる。この講演では、このようなランダム行列と深層神経回路に関連するトピック、特に自由確率論、平均場、NNGP、NTKなどを概観する。

The central role in machine learning is played by deep neural networks. When understanding deep neural networks, random matrices often appear. In particular, random matrices frequently arise in the context of parameter initialization and understanding learning dynamics. In this lecture, we will provide an overview of topics related to random matrices and deep neural networks, with a focus on free probability theory, mean field theory, NNGP (Neural Network Gaussian Process), NTK (Neural Tangent Kernel), and others.

Random Matrices and Deep Neural Networks

Talk at 2023/July/12, IMI
Tomohiro Hayase,
Senior Reseach Scientist
Cluster Metaverse Lab.

Table of Contents

  1. Overview
    Deep neural network and Gaussian process
  2. Jacobian
    Stability of DNN and random matrices
  3. NTK
    Training dynamics and random matrices
  4. Asymptotic Freeness Main theorem: asymptotic freeness of Jacobians
  5. Summary & In Progress

Overview


Multilayer Perceptron

[Figure: https://www.javatpoint.com/multi-layer-perceptron-in-tensorflow] Let n0,n1,,nLN.n_0, n_1, \dots, n_L \in \N. Parameters:

θ=(W,b)=1,,L,WRn×n1,bRn.\theta = (W_\ell, b_\ell)_{\ell=1, \dots, L}, W_\ell \in \R^{n_\ell \times n_{\ell-1}},b_\ell \in \R^{n_\ell}.

Forward propagation: for xRn0,x \in \R^{n_0}, set x0=xx_0 = x and inductively

h=Wx1+b,x=φ(h):=φ(h,i)i[n].h_\ell = W_\ell x_{\ell-1} + b_\ell, x_\ell = \varphi(h_\ell):= \varphi(h_{\ell,i})_{i\in[n_\ell]}.

Finally, definne the output by fθ(x)=hL.f_\theta(x) = h_L.

φ\varphi: Activation Function


Deep Learning

Generally, a standard formulation of supervised deep learning is as follows:

  1. We are given a finite set of pairs of input/output data (x,y)D(x,y) \in \mathcal{D}.
  2. We are given a deep neural network (DNN), which is a composition of (parameterized) transformations, which maps a real vector to a real vector.
  3. We are given an object function : e.g. mean squared loss
L(x,y,θ)=12nLj=1nL(fθ(x)jyj)2,L(x,y, \theta) = \frac{1}{2n_L}\sum_{j=1}^{n_L} ( f_\theta(x)_j - y_j)^2,

Optimization

We minimize the loss function by gradient descent:

θt+1=θtηtθL(x,y,θt)\theta_{t+1} = \theta_t - \eta_t \frac{\partial}{\partial \theta}L (x,y, \theta_t)

Initialization of Parameters and Random Matrices

e.g. Gaussian (Ginibre) random matrix:

(W)i,jN(0,σw2/n), i.i.d.(W_\ell)_{i,j} \sim \mathcal{N}(0, \sigma_w^2/n_\ell), \mathrm{\ i.i.d.}

e.g. Haar distributied orthogonal matrix:

W=σwO,OHaar Orthogonal Prob.W_\ell= \sigma_w O, O\sim \mathrm{Haar \ Orthogonal \ Prob.}

The Inifnite-dimensional Limit is Gaussian

[Figure: https://ai.googleblog.com/2020/03/fast-and-easy-infinitely-wide-networks.html]


Neural Network Gaussian Process (NNGP)

Consider two inputs x,xx, x^\prime and corresponding hidden units x,xx_\ell, x^\prime_\ell and h,hh_\ell, h^\prime_\ell in MLP. Taking an inifnite dimensional limit at the initial state, we have [Lee+ICLR2018]

(h,h)N(0,σw2K(x,x)+σb2)(h_\ell, h_\ell^\prime)\sim \mathcal{N}(0, \sigma_w^2 K_\ell(x,x^\prime) + \sigma_b^2)

where

K(x,x):=limn1nj=1nx,jx,j.K_\ell(x, x^\prime) := \lim_{n_\ell \to \infty} \frac{1}{n_\ell} \sum_{j=1}^{n_\ell} x_{\ell,j} x^\prime_{\ell,j}.

We have the following Kernel Propagation:

K+1(x,x)=φ(z1)φ(z2)pN(z)dz,K_{\ell+1}(x, x^\prime) = \int \varphi(z_1)\varphi(z_2) p_\mathcal{N}(z) dz,

where

pN=N(0,σw2(K(x,x)K(x,x)K(x,x)K(x,x))+σb2).p_\mathcal{N}= \mathcal{N}( 0, \sigma_w^2\begin{pmatrix} K_\ell(x,x) & K_\ell(x,x^\prime)\\ K_\ell(x,x^\prime)& K_\ell(x^\prime,x^\prime)\end{pmatrix} + \sigma_b^2).
  • For some activation functions, we can compute the integral explicitly.

Application: NNGP Estimation

Generally, consider BB samples. Set X=(x(a))a=1,,B,Y=(y(a))a=1,,BX=(x(a))_{a=1,\dots, B}, Y=(y(a))_{a=1, \dots, B} be input/output samples.

K(x,X):=(KL(x,x(a)))n=1,,aRBK(x^*, X) := (K_L(x^*, x(a)))_{n=1, \dots, a} \in \R^B K(X,X):=(KL(x(a),x(b)))a,bMB(R)K(X,X):= ( K_L(x(a), x(b)) )_{a,b} \in M_B(\R)

Then the posterior mean/ var is given by the following : for a new input x,x^*,

m(y)=K(x,X)K(X,X)1Ym(y^*) = K(x^*,X)K(X,X)^{-1}Y v(y)=K(x,x)K(x,X)K(X,X)1K(x,X)v(y^*) = K(x^*, x^*) - K(x^*,X)K(X,X)^{-1}K(x^*, X)

[Lee et.al., Deep Neural Networks as Gaussian Process, ICLR 2018]

Jacobian


Vanishing/Exploding Gradients

The optimization of DNN needs its parameter derivations. Since a DNN is a composition of functions, the parameter derivations are computed by  the chian rule. The input-output Jacobian is defined as

J=fθ(x)x=hLxJ = \frac{\partial f_\theta(x)}{\partial x} = \frac{\partial h_L}{\partial x}

In the case of MLP, we have

J=WLDL1W2D1W1,J = W_L D_{L-1} \dots W_2 D_1 W_1,

where

D=xh=diag(φ(h,1),,φ(h,n))D_\ell = \frac{\partial x_\ell}{\partial h_\ell} = \mathrm{diag}( \varphi^\prime(h_{\ell,1}), \dots, \varphi^\prime(h_{\ell,n_\ell}))

Dynamical Isometry

A DNN is said to achive dynamical isometry If the eigenvalue distribution of JJJJ^\top is concentrated aound one. Dynamical Isotmetry prevents the exploding/vanshing gradients.

[Pennington+, AISTATS2018, CH, CIMP2022] If we set the initialization of parameters to be Haar orthgonal and choose appropriate activation function, then we can make the DNN to achieve the dynamical isometry.

Set μL,ν\mu_L, \nu be limit spectral distributions of JJ,D2JJ^\top, D^2 as wide limits respectively.

Under the assumption of the asymptotic freeness of Jacobians,

μL=[(σ2)ν]L\mu_L = [(\sigma^2 \cdot )_* \nu ]^{\boxtimes L}

where \boxtimes is the free multiplicative convolution,

Distribution of D2D^2

[Figure: Pennington,  Schoenholz, Ganguli, AISTATS2018]


The Limit Spectral Distribution μL\mu_L of JJJJ^\top

[Figure: Pennington,  Schoenholz, Ganguli, AISTATS2018]


Neural Tangent Kernel


Neural Tangent Kernel

Under continual version of GD, learning dynamics of parameters is given by:

dθtdt=η(θfθt)(yfθt)\frac{d\theta_t}{dt} = \eta (\nabla_\theta f_{\theta_t})^\top (y - f_{\theta_t})

( * The learning rate η\eta is fixed.) Then learning dynamics of DNN is given by:

dfθdt=ηΘt(yfθt)\frac{df_{\theta}}{dt} = \eta\Theta_t(y-f_{\theta_t})

where

Θt=θfθt(θfθt)\Theta_t = \nabla_\theta f_{\theta_t}( \nabla_\theta f_{\theta_t})^\top

**Informal[Jacot+NeurIPS2018, Lee+NeruIPS2019]**Under the wide limit nn \to \infty, the learning dynamics of DNN is approximated by

dfθdt=ηΘ(yfθt)\frac{df_{\theta}}{dt} = \eta \Theta(y-f_{\theta_t})

where the neural tangent kernel is defined as

Θ:=limn2,,nL1Θ0\Theta := \lim_{n_2, \dots, n_{L-1} \to \infty} \Theta_0

Neural Tangent Kernel is A Surrogate Model of DNN+GD

Based of NTK, we can do Bayesian estimation in the same way as NNGP. Moreover, with NTK, we can simulate the gradient descent at any step tt of ensemble newtorks. [Figure from Google “Fast and Easy Infinitely Wide Networks with Neural Tangents”]

Appliable to CNN/ResNet

Figure from [Google “Fast and Easy Infinitely Wide Networks with Neural Tangents”]

Moreover, NTK is appliable to Attention: Infinite attention: NNGP and NTK for deep attention networks [https://arxiv.org/abs/2006.10540\]

Eigenvalue Spectrum of NTK

Spectra of the Conjugate Kernel and Neural Tangent Kernel for linear-width neural networks” Z. Fan & Z. Wang https://arxiv.org/abs/2005.11879 They treats the standard formulation: Gaussian Initialization x Multi-samples  x Small output dimension, and they get an recurrence equation of the limit spetreatctral distribution of NTK. Figures: Red lines are theoretical prediction


One-sample NTK

TH& R.Karakia “The Spectrum of Fisher Information of Deep Networks Achieving Dynamical Isometryhttps://arxiv.org/abs/2006.07814, In AISTATS 2020. When the DNN achieves dynamical isometry,  the spectrum of the (one-sample x high-dim output)”NTK” concentrates around the maximal value, and the maximal values is O(L). (Sketch)Under an assumption on Asymptotic Freeness, we have the following recursive equations:

Θ+1=q+W+1DΘDW+1\Theta_{\ell+1} = q_\ell + W_{\ell+1} D_\ell \Theta_\ell D_\ell W_{\ell+1}^\top μ+1=(q+σ+12)(νμ)\mu_{\ell + 1} = (q_\ell + \sigma_{\ell+1}^2 \cdot)_* (\nu_\ell \boxtimes \mu_\ell)

NTK & Learning Rate

The spectrum (eigenvalues) of the NTK has vital role in tuning the learning dynamics. e.g. η>1/λmax(Θ)\eta > 1/ \lambda_\mathrm{max} (\Theta) \Longrightarrow The learning dynamics does not converge. e.g. The conditional number c=λmin/λmaxc = \lambda_\mathrm{min}/ \lambda_\mathrm{max} detemines the converges speed.

Red line  (the boarder line of the exploding gradients) : This line is expected by our theory !


Asymptotic Freeness


Asymptotic Freeness and Free Probability Theory

Definition(Asymtptotic freeness, C^*-version)[Voiculescu’85] Let (Aj(n),Aj(n))jJ(A_j(n), A_j(n)^*)_{j \in J} be a family of n×nn \times n random matrices and adjoints. The family is said to be asymptotically free almost surely, if there exists C^*-probability spaces (Aj,τj)jJ(\mathfrak{A}_j, \tau_j)_{j \in J} and elements (ajAj)jJ(a_j \in \mathfrak{A}_j)_{j \in J} so that for any QCXj,XjjJQ \in \mathbb{C} \langle X_j, X_j^* \mid j \in J \rangle, the following holds:

limntrn[Q(Aj(n),Aj(n)jJ)]=(jJτj)[Q(aj,ajjJ)],\lim_{n \to \infty} \mathrm{tr}_n \left[Q(A_j(n), A_j(n)^* \mid j \in J ) \right] \\ = (*_{j \in J} \tau_j) \left[ Q(a_j, a_j^* \mid j \in J) \right],

where jJτj*_{j \in J} \tau_j is the free product of the tracial states.

Example

For NN,N \in \N, let

  • W(N)W(N) be Ginibre or Haar orthogonal random matrix,
  • D(N)D(N) be constant diagonal matrix with a limit distirubution as N.N \to \infty. Then (W,W)(W, W^*) and DD are a.s. asymptotically free as N \to \infty.

Asymptotic Freeness of Jacobians

Let W,D(=1,2,,L)W_\ell, D_{\ell} (\ell=1,2,\dots, L) be weight matrices in MLP and WW_\ell be scaled Haar orthogonal random matrices. (The Gaussian case is treated by : [B. Hanin and M. Nica.], [L. Pastur. ], [G.Yang] ) Theorem [CH22] Assuming that D1,,DL1D_1, \dots, D_{L-1} have limit joint moments. Then

(W1,W1),,(WL,WL),(D1,,DL1)(W_1, W_1^\top), \dots, (W_L, W_L^\top), (D_1, \dots, D_{L-1})

are asymptotically free  as nn \to \infty  almost surely.

Difficulty: Entries of D and WD_\ell \text{\ and }W_\ell are not independent. D=diag(φ(h))D_\ell = \mathrm{diag}(\varphi^\prime(h_\ell)) h=Wxh_\ell = W_\ell x_\ell

(Sketch of Proof) Invariance of MLP + Taking submatrix Construct orthogonal matrix UU_\ell fixing xx_{\ell} , i.e.

Ux=x,U_\ell x_\ell = x_\ell,

and

URxN1×N1 Haar OrthogonalU_\ell |_{{\mathbb{R}x_\ell}^\bot} \sim N-1 \times N-1 \text{\ Haar Orthogonal}

with

(U0,,U) ⁣ ⁣ ⁣ ⁣(W+1,,WL).(U_0, \dots, U_\ell) \perp\!\!\!\!\perp(W_{\ell+1}, \dots, W_L).

for =0,,L1.\ell=0,\dots, L-1. Then we only need to show the asymptotic freeness of N1×N1N-1 \times N-1 submatrices of

(U0W1,,UL1WL),(D1,,DL1).(U_0W_1, \dots, U_{L-1}W_L), (D_1, \dots, D_{L-1}).

Summary


Summary

Considering neural networks with random paramters… \Longrightarrow

  1. Tuning initializaiton and learning rate
  2. Bayesian Estimation with NNGP
  3. Understanding Dynamics with NTK
  4. If we focus on the spectrum, free probability appears in the theory.

In Progress: Theoretical Understanding of NeRF

MLP-like NNs were previously only used for toy models, but are now being applied to real-world 2D images and 3D data. e.g. MLP-Mixer, NeRF, etc. Theoretically and practically easy to compute positively, just the right next research target! [Mildenhall, et.al., ***“*NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis”, ECCV 2020 ]

In progress: MLP-Mixer as a wide and sparse MLP

TH& R. Karakida “MLPMixer as a wide and sparse MLP” , arXiv perprint, https://arxiv.org/abs/2306.01470 Multi-layer perceptron (MLP) is a fundamental component of deep learning that has been extensively employed for various problems. However, recent empirical successes in MLP-based architectures, particularly the progress of the MLP-Mixer, have revealed that there is still hidden potential in improving MLPs to achieve better performance. Excluding auxiliary components, the basic block of MLP-Mixer is as follows:

Y=ϕ(Vϕ(XW))Y = \phi( V\phi( XW^\top))

It uses left and right multiplications of matrices. Here we introduce the conjugation operator:

J:Vec(X)Vec(X)J: \mathrm{Vec}(X) \mapsto \mathrm{Vec}(X^\top)

Then Jϕ=ϕJJ\phi = \phi J and

Y=ϕ(Vϕ(JWJX))=Mat(ϕ(1V)ϕ(J(1W)JVec(X))).Y = \phi(V \phi(J^\top WJX)) = \mathrm{Mat}(\phi(1 \otimes V) \phi(J^\top (1\otimes W)J \mathrm{Vec}(X))).

Thus MLP-Mixer is a kind of MLP with sparse weights (i.e. a lot of connections are set to be zero).

Even if we destroy an architechture of MLP-Mixer by replacing JJ with uniformly distributed random permutation (RP), the accuracy incread as the width increased. We experimentally confirmed that the following hyptothesis by Gouvela et.al. also holds for MLP-Mixer: An increase in the width while maintaining a fixed number of weight parameters leads to an improvement in test accuracy.

A. Golubeva et.al., “Are wider nets better given the same numbr of parameters?”, In *ICLR, *2021.