HumanAIGC / AnimateAnyone

Animate Anyone: Consistent and Controllable Image-to-Video Synthesis for Character Animation
Apache License 2.0
14.39k stars 968 forks source link

Sorry to burst the bubble but this might be Fake unless the team cares to provide some additional examples #15

Open ShadyDev opened 10 months ago

ShadyDev commented 10 months ago

On X (formely twitter) people have found out the origin sources of the videos that open pose model is beeing generates from

This tech is faked and be cautious on whatever this team is going to upload here

https://x.com/julesterpak/status/1731452110686892332?s=46&t=x02nRdMOO9e-APPWlVEkmA

kex0 commented 10 months ago

Doesn't look faked to me. Tons of artifacts and movements that are missing or are different from the original footage. Mainly faces and movement of clothes.

ShadyDev commented 10 months ago

Yes and that’s why the movement of open pose is exactly the same as the OG video. This is fabricated and tempered with. If it was legit then they would use a different motion reference that the OG movie.

Jongulo commented 10 months ago

I wish people that didn't understand how technology worked wouldn't make github accounts.

ShadyDev commented 10 months ago

If you really believe that the OG videos of the influencers weren't tempered on to seem like "AI Generated" than you're delusional. The only truly believable video is the one with Leo Messi but still that one can be also pretty much done with some simple Face Swapping.

As the post on Twitter states - IF they want to prove they're legit they need to post a video of a one of those persons dancing with a completely different move set than the one it was "trained on (which I even doubt)".

FurkanGozukara commented 10 months ago

If you really believe that the OG videos of the influencers weren't tempered on to seem like "AI Generated" than you're delusional. The only truly believable video is the one with Leo Messi but still that one can be also pretty much done with some simple Face Swapping.

As the post on Twitter states - IF they want to prove they're legit they need to post a video of a one of those persons dancing with a completely different move set than the one it was "trained on (which I even doubt)".

I also listed my red flags first day

ricardoreis commented 10 months ago

This is really strange. Let's wait for the source code to be published.

https://github.com/HumanAIGC/AnimateAnyone/assets/104947/046f981e-a25a-42ce-99d8-ee6c6598b5cd

ShawnFumo commented 10 months ago

Alright, I did a deep dive on this, and I believe it is NOT fake, though there is a lot of misunderstanding due to none of us being hooked into the research community.

First, the full site is here which has links to additional videos, the paper, etc: https://humanaigc.github.io/animate-anyone/

In the paper they talk about doing three different training rounds. One which is more general and has a lot of different videos in it. This is what is being shown in most examples like Iron Man, anime chars, 3d lady, lady with a hat and necklace, etc. You can see various moves applied to various characters, and many (all?) of those were never a video to start with.

They also trained separately on a specific TikTok dataset (which has been in various papers since 2021 it seems, first for making depth maps). This was specifically to test themselves against DisCo (more later). In this, they just trained on 300 TikTok videos from that dataset, and did 10 tests (from out of 100 test videos). For each of those, they ran it through pose detection and took a frame. Then they produced a video and compared it to how closely it matched the original TikTok test video, giving it numeric scores. But why recreate an existing video?

Because DisCo (from MS Azure and others) used this dataset and ran those tests on it. So this was an automated way they could say "our model is better than DisCo", with actual benchmark numbers. The DisCo site is here: https://disco-dance.github.io/ If you scroll down that page, you can see various cases where they apply diff animations to the same person or the same animation to diff people. It isn't as good as AA, but still pretty good, and the code and weights for that are linked.

They also did a third training (which I didn't look into as closely), on a particular fashion dataset (first used in 2019), because that was used in DreamPose, yet another project like this (this time from Nvidia and others), that used that same dataset and ran benchmark numbers. You can see that here: https://grail.cs.washington.edu/projects/dreampose/ Note that DreamPose doesn't handle patterns on the clothes as well, and you can see in AA's videos that theirs does, which is why they do that comparison.

Also note that DisCo compared itself to DreamPose in their results. DreamPose references being better than yet two other papers from 2022 and 2021. And DisCo and DreamPose were both from earlier this year. All of these are doing similar things, with similar boring presentations, but it just happens that AA blew up on social media.

If you look at all of these in context, you can see the tech has been in the works for a while in diff forms, getting better with each iteration. Each time a paper and code is released, with sample videos, comparisons to the older projects, etc. I think people are judging these guys against companies like Runway and Pika Labs releasing advertising videos of new features. This seems more like pure research that maybe they didn't even expect people to notice that much (since DisCo and DreamPose don't seem to have gotten much press).

And to go back to the AA TikTok videos, my guess is that many of the videos in the training set had lipsyncing to songs, and if a test video had SongA/MemeDanceA in it, there may have been videos in the training set with the same song/dance combo, just with a diff person doing it. So if you ran their general model against the frame and pose animation, it may not be as similar to the original video as what they showed (Maybe it looks less similar but is more coherent? Maybe it is just worse? We don't know). But again, it seems like that part was just them comparing how well it did on that task with that data as compared to DisCo. Their main results were of animating diff still images with various pose animations, the exact thing people are saying they should show examples of. And those examples are still really good.

evanferguson28 commented 10 months ago

I didn't dive into their paper, but it's perfectly reasonable that the authors attempted to use their method to reconstruct an existing video: to compare their result with ground truth. And if you care to actually do the comparison yourself, you can see the discrepancies pretty easily. I don't understand how you see it as "proof" that this is fake lol

lucasjinreal commented 10 months ago

It might not fake, but it can be overfitting on some real world existed video....

xizaoqu commented 10 months ago

On X (formely twitter) people have found out the origin sources of the videos that open pose model is beeing generates from

This tech is faked and be cautious on whatever this team is going to upload here

https://x.com/julesterpak/status/1731452110686892332?s=46&t=x02nRdMOO9e-APPWlVEkmA

Actually, they have mentioned it in the paper Fig.5. It is a normal way to prove their effectiveness. You should read the paper first before making such idiot remarks.

sodamai commented 10 months ago

This is such a retarded issue. Not only is this spreading mis-information, but it's also degrading the researchers' work. Check this out. An actual randomly paused side by side comparision Link to the source because idek why the entire video is not getting uploaded lol

https://github.com/HumanAIGC/AnimateAnyone/assets/121299095/c5d9d0f3-b9fb-4d7a-8134-15d4e23b0f9d

This is what happens when your only source of information is Twitter or X lmfao

stacksize commented 10 months ago

"I don't have access to the model, so it MUST BE FAKE"

mocmocmoc commented 10 months ago

I knew this day would come, Github turning into Facebook!

ReEnMikki commented 10 months ago

Github used to be a peaceful, humble corner of mature, professional, polite folks on the internet until the flocks of normies jumping on the AI wave took over and turned it into a cesspool

Collin-Budrick commented 10 months ago

Github used to be a peaceful, humble corner of mature, professional, polite folks on the internet until the flocks of normies jumping on the AI wave took over and turned it into a cesspool

Sad to say how true that is

geekyayush commented 10 months ago

Github used to be a peaceful, humble corner of mature, professional, polite folks on the internet until the flocks of normies jumping on the AI wave took over and turned it into a cesspool

so true, it's becoming the new reddit