See Convolutional neural network, Deep art
video. Achieved with optimization using ConvNets
We have content image and style image, and want to transfer the style into the content.
A Neural Algorithm of Artistic Style
Pikazo - neural style video tech demo Artistic style transfer for videos
Analogy-Driven 3D Style Transfer
reddit discussion
Code: code
Two Minute Papers - Deep Neural Network Learns Van Gogh's Art
Neural-style applied to videos too.
Method
- Step 1: Extract content targets, by passing content image through convnet, and record all raw activations in the convnet, which we say correspond to the "content" of the image.
- Step 2: Extract style targets. Pass style image through convent, and look at pairwise statistics in style gram matrices. They have found these are good statistics for the style of an image. They are basically, for a certain convolutional layer of size nxn, and p features, the ~covariance of the features over the nxn space. That is take a particlar point in the nxn space and look at all p features. The vector of p features is called a fiber. Take the outer product of that vector with itself (which is like a covariance matrix), and average it over all points in the nxn space, can also be written in other ways using matrices.. Intuitively, it represents how often each pair of features fire together. Could use other spatially-invariant statistics
- Optimization of image step. Loss consists of two terms:
- A content loss: match activations
- a style loss: match gram matrices
Best optimized with LBFGS, because everything fits in memory, as only have one image..