It's like a tech demo, a preview of the future. Give it 5 years and it will be super refined and probably the future of low cost animation for kids TV shows and stuff. Then even further, like how no one animates without a computer now, no one will animate without AI assistance.
As far as I can tell, it really depends on the intensity of the "slider".
When the insensity of the style transfer is pushed mostly to the right (high), it just seems like Pixar or cartoons. Nothing uncanny whatsoever.
But when they show is about a quarter of the way to the right... it's utter nightmare fuel, like plastic surgery taken way too far. The worst kind of uncanny valley, so I definitely agree with you there.
Effectively these let an app (eg some VToonify tool) generate content that from the perspective of your live streaming app look like they are from a webcam
I'm glad things are progressing, but it bugs me that AI is largely being innovated for the use of... things like this? I know this comment is a bit disparaging and minimizes greater achievements, and I apologize for that, but the closeness of content-consumerism and AI is becoming quite off putting.
Based on the numbers in the paper this is just a little bit too slow for use as a real time video effect. At ~0.1 seconds per frame we just need about a 3x improvement in performance to get to 30fps “real time” video frame rates.
And on that thought since it appears they used nVidia hardware based on the CUDA dependency, it would be interesting to see how this performs on something like an M1/M2 where there’s dedicated ML hardware to help offload and accelerate things.
viraptor|3 years ago
burntalmonds|3 years ago
nineteen999|3 years ago
Gigachad|3 years ago
crazygringo|3 years ago
When the insensity of the style transfer is pushed mostly to the right (high), it just seems like Pixar or cartoons. Nothing uncanny whatsoever.
But when they show is about a quarter of the way to the right... it's utter nightmare fuel, like plastic surgery taken way too far. The worst kind of uncanny valley, so I definitely agree with you there.
morjom|3 years ago
nmstoker|3 years ago
Effectively these let an app (eg some VToonify tool) generate content that from the perspective of your live streaming app look like they are from a webcam
marcAKAmarc|3 years ago
techdragon|3 years ago
And on that thought since it appears they used nVidia hardware based on the CUDA dependency, it would be interesting to see how this performs on something like an M1/M2 where there’s dedicated ML hardware to help offload and accelerate things.
speedgoose|3 years ago
drusepth|3 years ago
kevingadd|3 years ago
Uehreka|3 years ago