Menu
About me Kontakt

Creating complex deepfakes from a single input image - how advanced is this technology currently?

Recently, on the bycloud channel, the topic of hilarious memes created by artificial intelligence was discussed. The author pointed out that one of the earliest memeable AI technologies was deepfakes, and now more AIs are being humorously employed. He also mentioned the first order motion model, known for its efficacy in motion transfer, although it is limited to human faces. While the demo featured some horse motion transfers, the quality was not up to par with facial motion.

Moving on to the latest AI news, the author covered a recent research paper titled Liquid Warping Gen. Shortly after, the source code was released as open-source. The official demo proved to be incredibly convincing, showcasing phenomenal abilities in motion capturing, 3D body mapping, and other complex details. Even when it’s restricted to just one individual, its functionality is astonishing.

The author shared his impressions of this AI, which makes motion transfer from a single full-body photo possible with just one command. Besides its meme potential, he emphasized that this AI could vastly benefit game development, character animation, and even virtual try-ons thanks to its appearance transfer feature. He noted some humorous results, particularly when the AI misunderstood a person's body data, leading to amusing errors such as pasting the front body onto the back.

Another interesting aspect addressed in the video was the concept of novel view synthesis, which generates new information about perspectives that were not represented in the input. He compared this to how PifuHD generates 3D models but underlined that the new AI technology could enhance the process of creating 3D models. Despite some outputs looking reminiscent of early 2000s character graphics, the rendering time of only three minutes is a significant advancement compared to other methods.

At the end of the video, the author emphasized the importance of the demo being provided in 1080p, indicating that this new technology is not limited to low resolutions. With the video statistics showing 1,850,212 views and 60,536 likes at the time of writing, we can certainly expect a surge in memes using this new AI technology. He encourages viewers to experiment with the new tech and share their results on Discord, and also mentions Infinite Red, a consulting team for mobile and web app development.

Toggle timeline summary

  • 00:00 Introduction to AI-generated memes.
  • 00:04 Discussion of early AI like deepfakes.
  • 00:08 Increasing comedic use of AI technologies.
  • 00:16 Reference to the first order motion model.
  • 00:20 Capabilities of deepfake technology.
  • 00:37 Introduction to Liquid Warping Gen AI.
  • 00:42 Release of open-sourced codes for the AI.
  • 00:49 High-quality motion capturing capabilities.
  • 01:09 Conviction in the technology's motion transfer quality.
  • 01:31 AI's applications in game development and character animation.
  • 01:48 Hilarious results from using the AI.
  • 02:01 Explaining novel view synthesis technology.
  • 02:17 Limitations relating to skin generation.
  • 02:38 Potential improvements in generating 3D models.
  • 02:58 Comparison of rendering times with past animation methods.
  • 03:18 Anticipation for future technologies and memes.
  • 03:31 Link to installation tutorial for the AI.
  • 03:39 Acknowledgment of sponsorship by Infinite Red.
  • 03:58 Conclusion and thanks to supporters.

Transcription

Lately, we have been seeing some pretty hilarious memes that are made from AIs. Looking back, one of the earliest memeable AI was definitely from deepfakes, and now more and more AIs are being used comedically, and they are just too much fun. Remember the first order motion model that I covered a few months ago? It does a super good job at deepfaking with only one source image, however it only works on human faces. In their demo, they did show some horse motions transfer or human movements transfer, but it was not as good as the facial motions transfer. So, onto the latest AI news. 3 weeks ago, an AI research paper called Liquid Warping Gen was published, and just a few days ago, they have open sourced their codes. And first of all, their official demo looks too good to be true. It looks insanely good at motion capturing, mapping 3D bodies, image and painting, and many other details that have shown to be difficult to completely function on its own. And to blend all these into one, we would probably expect something not as functional. But after they released their codes, I am absolutely convinced. It is just inconceivably good at motion transferring, even if it is only limited to just one single person. So you basically can take any full body shot of anyone that has a defined body part, and use a reference video to perform motion transfer with just one single command, but a fairly lengthy one. Other than the incredible meme potential, this AI can be used in places such as game development, character animation, and it even can be used for trying on clothes with its appearance transfer function, which sounds amazing if it is developed even further. Looking at my results, they are just hilarious. If not enough information about a person's body is presented, the AI would instead paste the front body on their back, which creates such monstrosity. This is also called novel view synthesis, where the AI would generate new information about the perspectives that are not presented in the input, like the side view of a person or the backside. This is similar to how PifuHD generates the 3D model, but since it is just model generation, not skin generation, they use the front image instead. But in this case, the author can't generate new information of the skin from the back, so they just apply the same image from the front. However, this can be easily improved as the author made it so the AI can take in both the front side and the backside image information and generate a full 3D model for motion transfer. So now, it wouldn't have the problem of having a double-sided face. This would definitely be really beneficial if this can be extracted as a 3D model too. Even though the generated output does look a bit too rounded, which looks like some of that 2000s game graphics for characters, but hey, this only takes like 3 minutes to render and 20 years ago, people still needed to animate this themselves, which can take way longer than 5 minutes. The official demo did also provide a 1080p version, which shows how it is not limited like FOM to just 500x500 pixels. And compared to their previous work, this is a huge step up and a big milestone for them. And this would also mean we will soon have some better technologies and memes for us to enjoy. Not gonna lie, it is only a matter of time before a meme with this AI takes over the internet. So if you want to play around with this AI yourself, I'll link my installation tutorial that's on my spam channel down in the description. This video is sponsored by Infinite Red. Infinite Red Consulting handles your mobile, web, and AI needs. If you are looking for someone to build your app, visit with the link down in the description. If you have any questions about the installation or wanna share your results, you can head over to my discord channel. A big shoutout to Mark Finn and many other Patreons that support me on Patreon. Follow me on Twitter if you haven't and I'll see you all in the next one.