Successful development of an AI model that can translate languages into physical motion
AI researchers from Carnegie Mellon University, Pennsylvania, United States recently developed an intensive AI model capable of translating languages (text, voice) into motion and gestures. physics with relatively high accuracy.
This AI model is named Joint Language-to-Pose (JL2P), and is also known as a method to bring the ability to combine natural language with 3D posture simulation models effectively. Practical applications in the near future.
- DeepMind's AI model can learn how to create videos just by watching clips on YouTube
This AI model is named Joint Language-to-Pose (JL2P)
JL2P's ability to analyze and simulate poses and gestures in three dimensions is carefully trained through end-to-end programs - a powerful and effective approach to the chapters. Training is 'teared down' in the form of individual strings. The AI model will have to complete short, simple tasks before being allowed to move on to more complex goals.
Currently, JL2P's ability to simulate animation is limited in the form of rudimentary images (made up of simple lines), but the ability to simulate human-like movements based on the language of the model. This AI pattern is relatively accurate and intuitive. The team believes that models like JL2P could one day help robots perform real-world physical tasks similar to humans, or support the creation of virtual animated characters for video games. as well as movies.
- Successfully developing self-propelled bicycles using AI chips capable of deduction and learning like humans
JL2P's ability to simulate animation is limited to simple, crude images
Actually the idea of developing an AI model with the ability to translate languages into physical motion is not new. Before Carnegie Mellon University introduced JL2P, Microsoft also successfully developed a model called ObjGAN, with the task of specializing in picture sketching and storyboard. language captions. Another of Disney's AI models is also widely known for being able to use the words in a script to create a storyboard. Or more famous is Nvidia's GauGAN model, which can turn a doodles like created with a trackpad or Microsoft Paint into intelligent digital sketches with extreme aesthetics.
Returning to JL2P, this AI model can now very accurately simulate a number of simple to relatively complex movements such as walking or running, playing musical instruments (like guitar or violin), following the instructions given. direction (left or right) or speed control (fast or slow).
- Successfully developed an "fantasy keyboard" for touch screens and VR, based on AI
JL2P can now accurately simulate a number of simple to relatively complex movements
'First we optimize the model to predict 2 time steps based on complete word sentences. This simple task can help the AI model learn how to simulate very short posture sequences, such as foot movements when walking, hand movements while waving or postures, body posture when bowing. Once JL2P has learned how to simulate similar gestures with great precision, we will move on to the next stage in the curriculum. The model is now given two postures (numbers) to predict at the same time, 'said Carnegie Mellon University research team.
- AI knows how to play poker, beating the best in the world in a 6-player game
Simulate the running position of a normal person
Details of how JL2P works as well as typical 'works' were presented for the first time in a scientific article published July 2 on arXiv.org, and are expected to be presented. presented by the author and researchers of CMU Language Technology Institute Chaitanya Ahuja on September 19, on the stage of the 3D Vision International Conference taking place in Quebec, Canada.
The team confidently asserts that JL2P can provide 9% more accurate posture as well as physical movement compared to another 'top notch' AI model developed by SRI International's AI experts. in 2018.
- Samsung's Deepfake can make the dark drummer Rasputin sing like the real thing
JL2P simulates the action of standing up against human hands
Products created by JL2P after being trained by KIT Motion-Language Dataset.
JL2P simulates jumping over obstacles and running
First introduced in 2016 by Performance Humanoid Technologies, Germany, this data set is a combination of human motion with natural language descriptions, mapping 11 hours of continuous motion. people, recorded as over 6,200 English sentences, each sentence about 8 words long.
You should read it
- Nvidia uses AI to create slow motion video 240fps from smooth, lifelike 30fps video
- How to make Stop Motion videos on Android phones
- Stop Motion art rekindles in Vietnam
- How to create motion blur on CapCut
- How to shoot and edit slow motion videos on iPhone
- How to Deal With Motion Sickness on Rides
- How to create slow motion videos on Android?
- How to create slow-motion video with Super Slow Motion on Galaxy S9
- Galaxy S8 / S8 + Vietnam is supplemented with movie shooting features Super Slow Motion and AR Emoji
- First laptop integrated Leap Motion sensor
- Download application to make super smooth GIF image of Motion Stills of Google just arrived at Android
- How to combine Motion Paths animation effects in PowerPoint
Maybe you are interested
Series of DrayTek router models have security holes
Pixel 9's new AI features could soon expand to older models
Why should you buy the base iPhone 16 instead of the Pro model?
Top 6 smart ring models today
Why is the Working hybrid model being adopted by more and more companies?
Microsoft Phi-3.5 launched: A more competitive AI model