(no title)
bengarney | 1 year ago
I am planning on doing some more articles/director commentary as it goes along.
I have a number of episodes in the queue and each one is better than the last. My plan is to release an entire season of 12 or so.
The "I'm a GPT that wants everyone to be friends and how" is increasingly better in those episodes.
Even incremental improvements in stuff like background music make a big big difference.
I really want to do a v2 that is more of a "copilot" than an "AI first" experience. But I need partners to help with funding; I've taken it about as far as I can on a solo basis. The next step is a team of 4-5 people levelling it up. Every piece could be 10x better, and it would be a different beast entirely if that happened. I think there are some super exciting directions this could go.
The vision of a distributed creator system is very interesting, as is letting people do more hands-on writing/rewriting.
If any VCs are reading, I'd love to talk. :)
(PS - Hi Han!)
mike_hearn|1 year ago
How much funding do you think you need for an MVP that's more Copilot-like? I might be interested in taking part in a seed round. Having AI do everything is a fun challenge, but I think the sort of people who would actually pay for a product would want to have some creative control and let the AI handle the parts they don't want to or can't do.
The Minecraft-esque graphics probably aren't an issue, but scaling up to provide all the needed assets probably is. There are AIs that can generate 3D models but having a consistent art style is required for it to work visually and you provided that here. Finding a way to quickly and cheaply scale the "kitbashing" seems key to any kind of productization.
bengarney|1 year ago
I shot you an e-mail on a v2. (An MVP would be less; I realized I sent you the pitch for a full v2.)
There are a LOT of art packs out there for a ton of different looks and genres. Building sets is quick and easy even with kitbashing. I think you could synthesize 3d content in a lot of ways (vid2vid, Gaussian diffusion generative models, prop placement by LLM, clever use of stable diffusion/firefly for mattes, etc.) or have a small stable of fiverr types to make art for people on demand in a specific style...
HanClinto|1 year ago
Speaking for myself, I expect that the behind-the-scenes commentary would be the most interesting part of the project!
> The "I'm a GPT that wants everyone to be friends and how" is increasingly better in those episodes.
How long does the pipeline take to run? (apologies if this was part of the blog series and I missed it). Depending on how close to a self-running CI pipeline the whole process is at, I think it might be interesting to run benchmarks against various versions of the pipeline and evaluate its performance at each stage. I feel like I could evaluate the improvement of the "let's make everyone be friends!" writing if I'm comparing Episode 1 (compiled w/ v0.3) against Episode 1 (compiled w/ v0.8), instead of Episode 1 vs. Episode 12.
Crazy idea: If one could somehow quantify the quality of consistency, dialogue, camera work, etc -- then you may be able to watch numbers-go-up in an actual graph sort of way (I'm imagining a multi-agent system where various agents are responsible for monitoring various aspects of script and production quality -- almost like an actor/critic setup).
But at the very least, being able to A/B comparison between v0.3 and v0.6 could be very interesting for people interested in the internals.
> I've taken it about as far as I can on a solo basis. The next step is a team of 4-5 people levelling it up. Every piece could be 10x better, and it would be a different beast entirely if that happened. I think there are some super exciting directions this could go.
I think that's the really cool thing about what you've built here -- it's a complete pipeline, and every piece is present -- even if the pieces aren't in their final form, the fact that you've pieced together an entire pipeline is extremely compelling.
> (PS - Hi Han!)
Hi!! It was a very cool surprise to see your name pop up on my HN feed this morning. :D
bengarney|1 year ago
But I had 8 kids 5-15 watch all of Ep1 _AND_ choose to watch Ep2 afterwards last night. They actually sat and watched, too, instead of having it on in the background... AND they were bummed they couldn't watch the super secret pilot episode (which has MAJOR audio issues - I couldn't bring myself to inflict it on them).
So I think something is there.
I agree, there are some great opportunities to track things somewhat more quantitatively. It takes ~15 minutes and $10 bucks to generate a script depending on how fast OpenAI is feeling. So in a real scale v2 it would be very reasonable to explore this.
Man, I sure hope I get to build this further!