ByteDance, the company that brought us TikTok, is now making headlines with its latest AI development: OmniHuman-1. The premise? Feed it a single image, and it can generate a video complete with natural gestures, lip-syncing, and even full-body movements.
It’s an impressive claim, and it will distract from some of its legal issues. But before we start calling this the next revolution in content creation, let’s break down what it actually offers – and what it doesn’t.
How Does It Work?
OmniHuman-1 isn’t just animating faces – it’s creating performances. The AI takes an image, combines it with inputs like audio or textual descriptions, and generates video footage that looks surprisingly lifelike. ByteDance says the system is trained on over 18,700 hours of human video data, which helps it generate natural movements and precise lip-syncing.
For portrait photographers, this could open doors to new types of content. Imagine shooting a series of stills and then turning them into animated portraits for social media or marketing purposes. Filmmakers could potentially create pickup shots without bringing the cast back to set.
But does that mean we’re heading for a world where still photos routinely become moving images? Maybe. Or maybe this tech has some hard limits.
Creative Shortcut or Crutch?
Let’s be real. AI in content creation always comes with the promise of making things “faster and easier.” But faster and easier doesn’t always mean better. While OmniHuman-1 is great in theory, AI-generated content often struggles with nuance.
The demos we’ve seen so far top out at around 25 seconds of video, and the AI isn’t immune to the uncanny valley effect when pushed too far. Think weird hand movements, awkward blinking, or facial expressions that don’t quite land – and guitar strings that never move despite being played.
So sure, you could animate a photo and turn it into a video, but how often will that look polished enough for professional use? If you’re an indie filmmaker or content creator trying to stand out, leaning too heavily on AI could backfire. Viewers can spot something that feels “off” a mile away.
What About Copyright and Ethics?
Here’s where things get murky. AI like this raises the usual concerns: deepfakes, misuse of someone’s likeness, and unauthorised manipulation of creative work. For photographers and filmmakers, that’s a real issue.
Could someone take a portrait you shot, animate it, and use it in a way you didn’t approve of? Absolutely. ByteDance will need to address this if they want creatives to embrace the tool instead of fear it.
On the flip side, this could be a useful tool for augmenting your own work – if you can maintain control over how it’s used. But that’s a big “if,” given how fast tech adoption tends to outpace regulation.
The Verdict: Promising, Scary, Tread Carefully
We’ve only just made it through January and already we’ve seen some massive leaps in AI capability this year. DeepSee surprised the world with its open-source DeepSeek-R1 model. DeepSeek has also released the Janus Pro multimodal model, which can both “see” and generate images. OpenAI has been playing catch-up, finally letting people get their hands on o3 – which is breaking all kinds of benchmark records – and introducing its new “Deep Research” feature.
Now, this.
OmniHuman-1 has potential, but it’s not a magic wand. For photographers, it could add new value to portraits and branding work. Filmmakers might find it useful for quick video assets or experimental projects. But it’s also clear that this isn’t ready to replace traditional production methods anytime soon – and it could introduce new headaches along the way.
As with any new tech, it’s about balance. The same AI that helps you bring a still image to life could also be used to create something misleading or harmful. Let’s hope ByteDance and the creative community take that responsibility seriously.
What’s your take on AI-generated video? Useful tool or potential troublemaker?