These look.. great, by and large. Hands are super natural, coherency is really high. Showing off piano chord blocking is a huge flex.
I’d like to play with this! No code, but bytedance often releases models, so I’m hopeful. It’s significantly better than vasa, and looks likely to be an iteration of that architecture.
liuliu 12 hours ago [-]
ByteDance didn't release their text-to-video model, which is the base of this work, so I would think unlikely.
Github is overflowing with Tencent, Alibaba, and Ant Group models. Typically licensed as Apache 2, and replete with pretrained weights and fine tuning scripts.
liuliu 10 hours ago [-]
The training process in OmniHuman-1 seems to be straightforward to replicate once Tencent releases their image-to-video model too.
echelon 10 hours ago [-]
T2V is already I2V if you're enterprising enough to open up the model and play with the latents. The I2V modality is almost just a trick.
liuliu 10 hours ago [-]
Yes, the Llava model can encode image, and you can encode image into 3D vae space. Without fine-tune the model though, you are not going to have fidelity to original (if only use Llava's SigLIP to encode), or end up with image with limited motion (3D vae encoded latents as the first frame then doing vid2vid).
tkgally 5 hours ago [-]
The hands do look good, but the spacing of the black and white keys on the piano keyboard is off. Similar problem with the strings and fretboard on the guitar. Such glitches won’t be noticed by as many people as weird hands are, though.
First video: Disappearing and appearing shirt buttons. Disappearing, appearing, and shapeshifting rings. Ear appears to be bluescreened despite the rest of the person appearing to be in front of a real background. Belt buckle slides unnaturally.
Second video: Shadows reveal inconsistent lighting direction. Disappearing and appearing studs on the watch strap. It also has bizarre clothing design with buttons on a non-opening shirt and what seems to be a printed fake weaving pattern that doesn't actually correspond to real weaving, but this could theoretically be made in reality.
What are the tells in most of these videos? I can't point at any in many of them. Hands, teeth, lip sync, body and should movement all look correct. Specially the TED talk like presentation examples near bottom.
thomastjeffery 8 hours ago [-]
Try watching them without audio.
They are all yelling. Even the girl with the cat. Too much energy. Too much expression. Too much pause. The pacing is all the same.
smusamashah 13 hours ago [-]
This looks better than EMO (also closed source by Alibaba group https://humanaigc.github.io/emote-portrait-alive/). See the rap example on their page. They apparently have EMO2 now which doesn't look as believable to me.
EMO covers head + shoulders while this OmniHuman-1 is covering full body and its looking even better. I would have easily mistaken these for real (specially while doom scrolling) if I was not looking for AI glitches.
UPDATE: Googling animate bytedance site:github.io returns many in the same domain (all proprietry). Found a few good ones.
This is very good attempt with people playing musical instruments.
But, there are some subtle timing tells, that this is AI generated. Take a look at the singer playing the piano. Timing of the hands with the singer is slightly off. The same goes with the singer and the guitar. I'm not a guitar player or piano player, but I do play a lot of different musical instruments at a high level, and the timing looks off, slightly ahead or behind the actual piece of audio of the piece of music.
stubish 6 hours ago [-]
They have cracked lip sync. AI can now realistically show singers strumming away out of time on guitars that are not plugged in or drummers playing the incorrect fills.
mkagenius 11 hours ago [-]
> Timing of the hands with the singer is slightly off.
Sure, only way is up though. I haven't seen this level realism in SORA or the google one. Plus, its synced with audio.
kiwiguy1 12 hours ago [-]
I run youtube channels with almost 2 billion views and this actually concerns me. I would love to try this in my productions!!
lamnguyenx 6 hours ago [-]
NVIDIA Demo of Audio2Face is such a joke, compared to this one.
egnehots 12 hours ago [-]
this could be used as an incredible low bitrate codec for some streaming use cases.
(video conferencing/podcasts on <3G for ex, just use some keyframes + the audio).
mpalmer 3 hours ago [-]
...I feel slapped by progress. Rarely does such an impressive demo leave me feeling less inspired and hopeful about the future.
emsign 12 hours ago [-]
It looks funny.
golol 14 hours ago [-]
Modern operating systems should include by default a very simple private/public key system to sign arbitrary files. I think it should not be very complicated? We badly need this in the age of AI.
Ajedi32 14 hours ago [-]
How would that help?
ssalka 13 hours ago [-]
Auto-watermarking of AI generated content, I would imagine
Ajedi32 13 hours ago [-]
What does that have to do with signing arbitrary files?
echelon 13 hours ago [-]
That's too much effort and the use cases are what exactly? Helping the prosecution or defense in lawsuits?
People are going to get so used to AI content that it won't really matter. Culture is plastic. This will be the new norm.
Capturing photons to send signals is the new butter churning.
Rendered at 06:56:04 GMT+0000 (Coordinated Universal Time) with Vercel.
I’d like to play with this! No code, but bytedance often releases models, so I’m hopeful. It’s significantly better than vasa, and looks likely to be an iteration of that architecture.
https://aivideo.hunyuan.tencent.com/
Github is overflowing with Tencent, Alibaba, and Ant Group models. Typically licensed as Apache 2, and replete with pretrained weights and fine tuning scripts.
Good to know that I need to now assume performances are AI generated even if it's not obvious that they are!
Second video: Shadows reveal inconsistent lighting direction. Disappearing and appearing studs on the watch strap. It also has bizarre clothing design with buttons on a non-opening shirt and what seems to be a printed fake weaving pattern that doesn't actually correspond to real weaving, but this could theoretically be made in reality.
They are all yelling. Even the girl with the cat. Too much energy. Too much expression. Too much pause. The pacing is all the same.
EMO covers head + shoulders while this OmniHuman-1 is covering full body and its looking even better. I would have easily mistaken these for real (specially while doom scrolling) if I was not looking for AI glitches.
UPDATE: Googling animate bytedance site:github.io returns many in the same domain (all proprietry). Found a few good ones.
- https://byteaigc.github.io/X-Portrait2/ Very expressive lifelike portrait animations
- https://byteaigc.github.io/x-portrait/ (previous version of the same, has source https://github.com/bytedance/X-Portrait)
- https://loopyavatar.github.io/ (portrait animations, looks good)
- https://cyberhost.github.io/
- https://grisoon.github.io/INFP/
- https://grisoon.github.io/PersonaTalk/
- https://headgap.github.io/
- https://kebii.github.io/MikuDance/ anime animations
But, there are some subtle timing tells, that this is AI generated. Take a look at the singer playing the piano. Timing of the hands with the singer is slightly off. The same goes with the singer and the guitar. I'm not a guitar player or piano player, but I do play a lot of different musical instruments at a high level, and the timing looks off, slightly ahead or behind the actual piece of audio of the piece of music.
Sure, only way is up though. I haven't seen this level realism in SORA or the google one. Plus, its synced with audio.
People are going to get so used to AI content that it won't really matter. Culture is plastic. This will be the new norm.
Capturing photons to send signals is the new butter churning.