Basic : Generalized Dense Layers

Austin Jiuk Kim·2022년 3월 24일
0

Deep Learning

목록 보기
7/10

Dimensions of Dense Layer

Cascaded structure.

Mxnet is the framework which controls multi GPUs at once. Because there are Include-Top classifier which is parameter in Mxnet, the drawings of the Dense Layers is better to be drawn from bottom to top.


The D
(a[lL])TR1×lν[lL](a[2])TR1×lν[2]L2=(νi[2])(a[1])TR1×lν[1]L1=(νi[1])(x)TR1×lx(\overrightarrow{a}^{[l_L]})^{T} \in \R^{1 \times {l_{\nu^{[l_L]}}}}\\ \:\\ \vdots\\ \:\\ (\overrightarrow{a}^{[2]})^{T} \in \R^{1 \times {l_{\nu^{[2]}}}}\\ \:\\ \uparrow \\ L_2 = (\dots \: \nu^{[2]}_{i} \: \dots) \\ |\\ \: \\ (\overrightarrow{a}^{[1]})^{T} \in \R^{1 \times {l_{\nu^{[1]}}}}\\ \: \\ \uparrow \\ L_1 = (\dots \: \nu^{[1]}_{i} \: \dots) \\ | \\ \: \\ (\overrightarrow{x})^{T} \in \R^{1 \times {l_x}}

The number of parameter is

(lx×lν)+(1×lν)(l_x \times l_{\nu}) + (1 \times l_{\nu})

(lx+1)×lν(l_x+1) \times l_{\nu}

In Dense Layer, as the network keep passing through the layers, the number of parameter considerably inceases.

Forward Propagation of The Second Dense Layer

The Second Dense Layer

Generalized Dense Layer

[ Generalized Dense Layer ]
(a[i])TR1×lν[i]Li=(νi[i])(a[i1])TR1×lν[i1]\vdots\\ (\overrightarrow{a}^{[i]})^{T} \in \R^{1 \times {l_{\nu^{[i]}}}}\\ \:\\ \uparrow \\ L_i = (\dots \: \nu^{[i]}_{i} \: \dots) \\ |\\ \: \\ (\overrightarrow{a}^{[i-1]})^{T} \in \R^{1 \times {l_{\nu^{[i-1]}}}}\\ \vdots\\
profile
그냥 돼지

0개의 댓글