Open code for animation synthesis using neural networks

Research team from Shanghai Technical University ΠΎΠΏΡƒΠ±Π»ΠΈΠΊΠΎΠ²Π°Π»Π° tools impersonator, which allows, using machine learning methods, to simulate the movement of people in static images, as well as to replace clothes, transfer to another environment and change the angle from which the object is visible. The code is written in Python
using the framework PyTorch. Assembly also requires torchvision and CUDA Toolkit.

Open code for animation synthesis using neural networks

The toolkit takes a XNUMXD image as input and synthesizes a modified result based on the selected model. Three transformation options are supported:
Creating a moving object that repeats the movements on which the model was trained. Passing appearance elements from the model to the object (for example, changing clothes). Generation of a new angle (for example, synthesizing an image in profile based on a frontal photo). All three methods can be combined, for example, it is possible to generate a video from a photo that imitates the performance of a complex acrobatic stunt in different clothes.

In the process of synthesis, the operations of selecting an object in a photograph and forming the missing elements of the background when moving are simultaneously performed. A model for a neural network can be trained once and used for various transformations. For loading available ready-made models that allow you to immediately use the tools without prior training. It requires a GPU with at least 8GB of memory.

Unlike transformation methods based on transformation by key points that describe the location of the body in two-dimensional space, in Impersonator an attempt was made to synthesize a three-dimensional mesh (mesh) with a description of the body using machine learning methods.
The proposed method allows manipulations taking into account the personalized shape of the body and the current posture, simulating the natural movements of the limbs.

Open code for animation synthesis using neural networks

To preserve the original information, such as textures, style, colors, and face recognition, the transformation process uses generative adversarial neural network (Liquid Warping GAN). Information about the original object and parameters for its precise identification is extracted by applying convolutional neural network.


Source: opennet.ru

Add a comment