Skip to main content

Reallusion's Cartoon Animator Versus PixVerse Image to Video and Lip Sync AI - Battle of the Talking Head Avatars

Still shot from a PixVerse Muscle Flex Effect using an image of my TET avatar.
PixVerse's Show Muscle Effect is a real flex for
other AI generative video sites.

AI
 generated digital avatars have improved greatly over the last 12 months, to the point where you can replace yourself with a digital, talking head, clone avatar that can say anything using your AI generated voice clone, and casual viewers wouldn't even notice.

I tried cloning myself back in July of 2024 in my post, Creating an AI Digital Avatar and Voice Clone of Myself with Free and Low Cost AI Tools. Since then, I've seen human clone channels on YouTube that I'd never guess in passing were digital humans.

Disastrous Hedra AI still shot from a video using my TET avatar as a base image.
This is one of Hedra's better generations
of video using my TET Avatar as a source.
One area that has been slower to catch up is using AI to create 2D cartoon style avatars, from very specific character designs, that don't have realistic human proportions. I put my TET Avatar through Hedra, one of the best talking head avatar generators currently available, and got a complete fever dream of a talking head character as output (see image).

I set about trying to find another AI image to video site that could also handle lip syncing audio, that would do a better job with my 2D avatar.

There are a number of AI generative video sites that include lip syncing but won't actually let you test these features in a free trial. Or, if they do, the free trial is so limited and time delayed, that I just wasn't prepared to pay for even one month of upgraded service, just to see if their AI was even up to the task. (Disclaimer: I also refuse to sign up for free trials of any paid plans if PayPal is not offered as a payment option).

Enter PixVerse.AI

PixVerse App - Animation Zone
PixVerse has five animation styles you can
apply to your creations. 
After much frustration with various sites, I discovered PixVerse through a video by YouTuber, The Zinny Studio, titled, Best FREE AI Video Generator? FIND OUT (Generative AI). I'd highly recommend watching this video for a run down on how to use PixVerse. I also recommend The Zinny Studio's channel if you're at all interested in AI animation, AI Avatars, and creating a faceless YouTube channel.

PixVerse seems to specialize in animation styles, and gives you just enough free credits to test out their image to video and lip sync features in one go. 

Video generations of up to eight seconds are ready in seconds. If you run out of credits, wait 24 hours for them to be topped up. Unfortunately free credits don't accumulate.

My initial trial was so impressive, I decided to sign up for one month of a Standard account (USD$10.00). Note, PixVerse does accept PayPal but you may get an error that looks like your payment failed. Their FAQs, state if you give it about an hour and refresh the site, your credits should appear.

PixVerse also includes preset AI effects such as turning any object into a 'squishy' object, or 'Lego Blasting' objects into pieces which are fun to play around with. The first image on this post is using their 'Muscle Surge' effect, where you can start with any character image and have them rip their shirt off, revealing a completely jacked body (I don't think it works with female bodies - but I haven't tried).

However, for my Cartoon Animator versus PixVerse AI avatar battle, all I used was their image to video feature, with no added effect or style, followed by their lip sync feature.

Reallusion's Cartoon Animator Versus PixVerse AI - Battle of the Talking Head Avatars

If you're familiar with my Lazy Animator Cartoon Animator courses you'll know I have a complete mini course on quickly Animating Talking Heads in Reallusion's Cartoon Animator 5 for complete beginners and novices.

So I decided to compare the process of animating an almost two minute long, talking head monologue, by my TET avatar in Cartoon Animator, with animating the same thing using PixVerse image to video, lip synced clips.

This article would be too long if I gave you a complete step by step guide for what I did in each application. So instead I'll give you an overview of each process, with my thoughts, and then the finished video.

Both versions use the same recorded audio of my own voice (not AI cloned). Both were also taken into my video editor (Shotcut) to put the animation together and to add any additional video effects.

The Cartoon Animator Version

Animating my TET Avatar in Cartoon Avatar using a Motion Clip saved to my character's Action Menu.
Animating my avatar in Cartoon Animator. To duplicate
the first minute of the animation I collected all the motions
into a single clip and saved it to my avatar's Action Menu.
This version was animated in the space of an afternoon, going on into the evening. I have no doubt, if I were to animate another monologue of this length using this character, I could get it done much faster but, I hadn't animated a G2 legacy character for quite some time, so I spent some of my time relearning the process. G3 characters are so much easier and quicker to animate.

The entire 1:55 minute scene was animated in one project file. I used the auto lip sync for the mouth with no manual adjustments so you can compare Cartoon Animators lip sync with PixVerse's lip sync.

Once I got the hang of things it was relatively quick to create a chain of upper body animations by Frankensteining bits and pieces of pre-made motion clips into a sequence for half of the animation. Then I simply duplicated the sequence for the second half.

It's not the perfect syncing of body motions to link to what the character is saying but it's passable. No hand gestures are completely incongruent with what is being said. 

The point was to get the animation done as quickly as possible.

I also added facial expressions using the face templates in the Face Editor, and eye blinks and head movement using the Puppet Editor.

From there I exported the animation to my video editor. Replaced the green screen background with my AnimLife Video background, and added in my lower third titles and credits.

Honestly, If I wanted to animate a new monologue using the same character and scene I could pretty much just update the audio, make a few adjustments to the body and face animations, and I'd be done.

Watch the finished animation below.


The PixVerse AI Animated Version

PixVerse - Lip Syncing my clips.
Lip syncing in PixVerse is as simple as uploading
your video and audio, and then clicking a button.
My PixVerse starting point was with a couple of still frame captures of my avatar sitting at the desk with a green screen background, taken from my Cartoon Animator project.

I used these still images as the base image for generating 13 five second animations of my avatar sitting at the desk talking and making appropriate hand gestures. I could've generated eight second clips but I've found, the longer you let the AI run with a prompt, the more likely it is that it'll start to introduce elements you didn't prompt for.

Even with my five second clips I got three that were unusable, as my avatar magically introduced props that weren't in the scene, or performed actions that moved props in the scene around (such as grabbing the nearby laptop and moving it closer - messing up continuity between clips).

While I didn't prompt for any camera movement, PixVerse's AI several times decided my avatar was too far away and had him move closer. Which is why, in the final version you'll see the character move forward and backwards. I had to settle for these clips because, on a standard, paid plan, I only had 1200 credits to play with (each five second generation used 30 credits), and I needed to save some for the lip sync, which I had no idea on how many I'd need. Also I was generating at the lowest resolution to save on credits too.

Once I had my clips, I downloaded them and brought them into my video editor along with my audio track. I then proceeded to piece together a sequence of clips where the movements and gestures loosely matched with what my avatar was saying.

Sometimes I'd use the full five second clip. Sometimes I'd follow this with the same clip played in reverse. Sometimes I'd use just part of a clip. I'd even slowed down the speed of a clip in one case.

Once I'd done this for half the audio, I copied the entire sequence to use for the remaining half. I did tweak the second half a little but, for the most part, it's an exact replay of the first minute.

PixVerse will only lip sync 30 seconds of video at a time, so I exported the animation as five clips under 30 seconds, along with the matching audio files (separately). Each file used around 140-160 credits to lip sync, so I ended up buying 550 more credits (USD$5.00) to finish the last two clips.

From there I downloaded everything and reconstructed it back in my video editor. I replaced the green screen background with my AnimLife Video background, and added in my lower third titles and credits.

Unlike Cartoon Animator, if I wanted to change the audio monologue on the whole clip, I'd have to redo the whole lip sync process again, giving Reallusion an actual win on time saving over AI generated animation.

Watch the finished animation below.


Who Wins, Cartoon Animator or PixVerse?

Overall, Cartoon Animator wins and is the better option for 2D, animated, talking head videos. You have a lot more character control, and I think you'll agree Cartoon Animator's auto lip sync is far ahead of what PixVerse produced in this comparison.

Plus, once you have you're initial project in Cartoon Animator, using it as a template for new audio is quicker than having to regenerate all the lip sync in PixVerse - not to mention the cost. 

PixVerse could definitely get expensive if you make a lot of videos. Cartoon Animator only costs the initial purchase price of the software. I can tell you, I burned through my monthly credit, and the extra credit I purchased, in a day on PixVerse.

However, I am very impressed by PixVerse. Particularly that it was able to animate my specific avatar character without changing the art style at all. Even when I applied that muscle flex effect, it still was consistent with the art style of my image.

There are still a few issues to iron out, such as the usual AI deformed hands problem. One of my initial base images clearly shows my character has four fingers and a thumb but, if the base clip didn't show the hands, then the AI would produce three fingers and a thumb. Partly I expect because this is common for hands in 2D animation.

I wasn't overly impressed with the lip sync. There are times when the avatars lower teeth look more like its lower lip. Sometimes the voice isn't synced at all, and sometimes the mouth disappears altogether momentarily.

That said, if you want to turn your avatar into Batman... you can do that in seconds on PixVerse.

TET Avatar with PixVerse's Batman Morph effect applied.
Be yourself.
Except if you can be Batman. Be Batman!
Animated entirely with PixVerse's Batman Effect.



o---o--- ---o--- o---

Did you find this article useful?
Subscribe to my newsletter and get the
latest articles delivered to your inbox.

Comments

Popular posts from this blog

Inochi2D - Free Open Source 2D VTuber Avatar Rigging and Puppeteering Software (Part 1)

Inochi2D Creator - Free Open Source VTuber Software. If you've been looking for a way to live perform as a 2D cartoon avatar on camera, whether it be for a live stream or for pre-recorded content like educational videos, then VTuber software is a low cost (or even no cost) option worth looking into. In my previous post, How to Become a VTuber - 2D and 3D Software for Creating and Controlling Your Avatar , I took a brief look at the relatively new but completely free and open source Inochi2D  which I thought showed great potential for my own needs of creating a live performance character rig for my own TET Avatar that I use for all my promotional materials. While it is possible to live perform my character using Cartoon Animator itself, Reallusion's MotionLive2D capture system isn't great - with lip sync in particular. More importantly though, I can't exactly teach people how to use Cartoon Animator if I'm using Cartoon Animator to control my Avatar. What is Inochi2D...

LTX Studio (Beta): AI-Powered Visual Storytelling, From Script to Screen in One App.

LTX Studio can generate consistent characters across storyboard panels - even if one character is a dragon! W hile text to image, and text to video (and image to video) AI tend to be getting a lot of the press, the real exciting aspect of generative AI implementation is how it can be used to speed up creator workflow. Being able to realize your creative vision in a shorter length of time can lead to more ambitious projects. Particularly if you're a team of one, with a very limited budget, but you one day dream of creating your own epic animated feature film. LTX Studio (beta), a new 'all-in-one' AI film making tool, is not going to let you realize that dream from a single text prompt but, by bringing a bunch of generative AI technologies together, the developers have created a one platform workflow that can help anyone rapidly visualize and deliver a story from initial idea to finished film in days rather than weeks (depending upon how ambitious the project is). Even bette...

Review: Toon Boom Harmony 14 - What I learned in 21 Days

Toon Boom Harmony is widely considered the industry standard for primarily 2D animation. You don't get to be that if your software isn't exceptional. However, industry standard and exceptional usually translates to steep learning curve and probably contains more features than I'll ever use. So, in reviewing the latest version, Harmony 14, I'm setting out to answer two questions; How easy is it to learn the basics and is it software an independent artist/animator, like myself, should seriously consider as their go to, 2D animation studio of choice?

The Ultimate Independent Animator's App and Resource List - Animation and Video Life

Image created with Cartoon Animator 4. Being an independent animator is not like a studio animation job. There's so much more to do that is indirectly related to the actual task of animating. Over the years I've sought out many apps, tools, and services that can help me achieve that one single task, expressing myself through animation. Below is my Ultimate Independent Animator's Resource List for 2024 (last updated Oct 2024). It started out as a list of free or low cost apps that could help you in every stage of producing either 2D or 3D animation, and then just kind of grew from there. You may not have been looking for a Time Management App as much as you needed something to get you started in 3D animation but when those commissioned projects start coming in you'll have a head start on maximizing your time. All the apps and services on this list had to meet two main criteria: They had to be useful and relevant to an Indy Animator/artist. The base app/se...

Eight 2D Animation Apps For Your Phone or Tablet Mobile Device

M obile productivity apps have become so capable that they can be great alternatives to their PC/MAC equivalents or serve as great tools in their own right when you're away from your desk. While some apps simply mimic their desktop counterparts, others offer well thought out, touch-friendly interfaces that are easier and more fun to use. Every so often I check out what's available for 2D animation for Android devices, since that's what I use, that can complement my workflow with Reallusion's Cartoon Animator 5. Some may be available for Apple devices as well. Below I've listed six free (F) apps (with optional paid (P) upgrades) on the Google Play Store that you might want to explore. Some are just fun apps on their own while others may be useful as part of your workflow on bigger animation projects. Not all are exclusively animation apps and could be used on any production. JotterPad (F/P) The name JotterPad makes this sound like a notepad application but it's ...

Krita AI Diffusion - Generative Image AI For Krita is Seriously Useful, Powerful and Free (If You Can Install it Locally)

Generative AI sequence of a woman in a business suit. From sketch to refined image using Krita AI Diffusion - by TET G enerative image AI, where you describe an image with a text prompt to an Artificial Intelligence model and it produces a new image based on your prompt, is gaining a strong hold as a tool for many artists. Krita AI Diffusion brings generative AI image tools right into your favourite free and opensource, graphics editor, Krita. Not only that, if you have a computer with decent specs (and at least 10GB of hard drive space), Krita AI Diffusion is completely free. What If I Don't Have a Powerful Computer? If you're in my situation, with a computer that was around before anyone in the mainstream had even heard of generative AI, you can still access Krita AI Diffusion for free, using a cloud based AI server, Interstice  and 300 tokens, to get you started. Once your initial tokens run out, purchase 5000 more for 10€ (approx US$11.00). Tokens never expire. I would...

The Future of Animation? Using Reallusion's Cartoon Animator and Image to Video Generative AI to Animate Sequences

Animation software is already capable of auto generating tweens but will AI take it to the next level and tween entire sequences of motion? M y earlier article,  Reallusion's Cartoon Animator Versus PixVerse Image to Video and Lip Sync AI - Battle of the Talking Head Avatars , got me thinking about using generative AI for creating the 'tween' animation in between key poses. This isn't a new idea, and I have explored tweening with AI before in my article,  Five AI Generative Image to Video Tools For Animation You Can Try Free Right Now . That was almost eight months ago and the results were mixed and barely usable. However PixVerse AI is exceptional at maintaining the art style of whatever image you begin with and will also let you include an end frame, making it an ideal tool for tweening, so...  What if, in a future version of Cartoon Animator (or your preferred 2D animation software), all you had to do was create the key moments of the scene, and then the software...