[논문정리] A Neural Algorithm of Artistic Style

Eunjin Kim·2022년 4월 20일
0

논문

목록 보기
1/13

the system uses neural representations to separate and recombine content and style of arbitrary images

Along the CNN, the input image is transformed into representations (actual content of the image compared to its detailed pixel values)

The representations of content and style in the CNN are separable

Previous approaches(non-photorealistic rendering) mainly rely on non-parametric techniques to directly manipulate the pixel representation of an image.

By using DNN trained on object recognition, it carry out manipulations in feature spaces(high level of an image)

Perfoming a complex-cell would be a possibel way to obtain a content-independent representation of the appearance of a visual input.

Methods

It used the feature space provided by the 16 convolutional and 5 pooling layers of the VGG19 Network.(NO fully connected layers)

replace max-pooling to average pooling

Content Loss

p\overrightarrow{p} : original image
x\overrightarrow{x} : generated image
Pi,jlP^l_{i,j} : the activation of the ithi^{th} filter at position jj in layer ll
PlP^l , FlF^l : respective feature representation in layer ll

Gram matrix GlRNl×NlG^l ∈ R^{N^l \times N^l}

Gi,jlG^l_{i, j} : inner product between the vectorised feature map ii and jj in layer ll

Using gradient descent to find another image that matches the sytle representation of the original image.

Style loss

Minimising the mean-squared distance between the entries of the Gram matrix from the original image and the Gram matrix of the image to be generated.
a\overrightarrow{a} : original image
x\overrightarrow{x} : generated image
AlA^l , GlG^l : respective style representation in layer ll

wlw_l : weighting factors

Loss function

p\overrightarrow{p} : photograph
a\overrightarrow{a} : artwork

α\alpha : weighting factor for content reconstruction
β\beta : weighting factor for style reconstruction

profile
ALL IS WELL🌻

0개의 댓글