Successful development of an AI model that can translate languages ​​into physical motion

Intensive AI model, capable of translating languages ​​(text, voice) into motion, physical gestures with relatively high accuracy.

AI researchers from Carnegie Mellon University, Pennsylvania, United States recently developed an intensive AI model capable of translating languages ​​(text, voice) into motion and gestures. physics with relatively high accuracy.

This AI model is named Joint Language-to-Pose (JL2P), and is also known as a method to bring the ability to combine natural language with 3D posture simulation models effectively. Practical applications in the near future.

  1. DeepMind's AI model can learn how to create videos just by watching clips on YouTube

Picture 1 of Successful development of an AI model that can translate languages ​​into physical motion
This AI model is named Joint Language-to-Pose (JL2P)

JL2P's ability to analyze and simulate poses and gestures in three dimensions is carefully trained through end-to-end programs - a powerful and effective approach to the chapters. Training is 'teared down' in the form of individual strings. The AI ​​model will have to complete short, simple tasks before being allowed to move on to more complex goals.

Currently, JL2P's ability to simulate animation is limited in the form of rudimentary images (made up of simple lines), but the ability to simulate human-like movements based on the language of the model. This AI pattern is relatively accurate and intuitive. The team believes that models like JL2P could one day help robots perform real-world physical tasks similar to humans, or support the creation of virtual animated characters for video games. as well as movies.

  1. Successfully developing self-propelled bicycles using AI chips capable of deduction and learning like humans

Picture 2 of Successful development of an AI model that can translate languages ​​into physical motion
JL2P's ability to simulate animation is limited to simple, crude images

Actually the idea of ​​developing an AI model with the ability to translate languages ​​into physical motion is not new. Before Carnegie Mellon University introduced JL2P, Microsoft also successfully developed a model called ObjGAN, with the task of specializing in picture sketching and storyboard. language captions. Another of Disney's AI models is also widely known for being able to use the words in a script to create a storyboard. Or more famous is Nvidia's GauGAN model, which can turn a doodles like created with a trackpad or Microsoft Paint into intelligent digital sketches with extreme aesthetics.

Returning to JL2P, this AI model can now very accurately simulate a number of simple to relatively complex movements such as walking or running, playing musical instruments (like guitar or violin), following the instructions given. direction (left or right) or speed control (fast or slow).

  1. Successfully developed an "fantasy keyboard" for touch screens and VR, based on AI

Picture 3 of Successful development of an AI model that can translate languages ​​into physical motion
JL2P can now accurately simulate a number of simple to relatively complex movements

'First we optimize the model to predict 2 time steps based on complete word sentences. This simple task can help the AI ​​model learn how to simulate very short posture sequences, such as foot movements when walking, hand movements while waving or postures, body posture when bowing. Once JL2P has learned how to simulate similar gestures with great precision, we will move on to the next stage in the curriculum. The model is now given two postures (numbers) to predict at the same time, 'said Carnegie Mellon University research team.

  1. AI knows how to play poker, beating the best in the world in a 6-player game

Picture 4 of Successful development of an AI model that can translate languages ​​into physical motion
Simulate the running position of a normal person

Details of how JL2P works as well as typical 'works' were presented for the first time in a scientific article published July 2 on arXiv.org, and are expected to be presented. presented by the author and researchers of CMU Language Technology Institute Chaitanya Ahuja on September 19, on the stage of the 3D Vision International Conference taking place in Quebec, Canada.

The team confidently asserts that JL2P can provide 9% more accurate posture as well as physical movement compared to another 'top notch' AI model developed by SRI International's AI experts. in 2018.

  1. Samsung's Deepfake can make the dark drummer Rasputin sing like the real thing

Picture 5 of Successful development of an AI model that can translate languages ​​into physical motion
JL2P simulates the action of standing up against human hands

Products created by JL2P after being trained by KIT Motion-Language Dataset.

Picture 6 of Successful development of an AI model that can translate languages ​​into physical motion
JL2P simulates jumping over obstacles and running

First introduced in 2016 by Performance Humanoid Technologies, Germany, this data set is a combination of human motion with natural language descriptions, mapping 11 hours of continuous motion. people, recorded as over 6,200 English sentences, each sentence about 8 words long.

Update 14 September 2019
Category

System

Mac OS X

Hardware

Game

Tech info

Technology

Science

Life

Application

Electric

Program

Mobile