Anyone can do this. My workflow to make Tough Cluck was as follows.
Step 1. Outline your script.
I didn’t have a finished script when I started, but I needed a rough outline to know what characters and locations I’d be working with. The more crucial the character or location, the more you want to invest in writing a descriptive, reusable prompt.
Step 2. Set the scene(s) for prompt consistency.
As a preamble I included consistent context & character descriptions in most prompts, i.e. “Tough Cluck is an ex cop from a gritty 1970s cop movie set in San Francisco. Cluck just happens to wear a full body chicken suit at all times. Include sound effects but do not include music. Do not include subtitles.”
I also took some shortcuts to aide consistency. A guy in a chicken suit, or bad guys wearing ski masks, will look more consistent over multiple generations than a clearly visible face. I left the main villain in the dark so I could reuse them more easily in the next project.
Even my choice for Cluck’s car was intentional, choosing something iconic like a VW bus which has a ton of training data, is going to give you more consistent output over multiple prompts. Understand and leverage the limitations of the AI.
Step 3. Generate (and regenerate) clips
Start generating. I used Gemini veo3 for most of this project but other creators are leveraging a host of different tools, and the capabilities change weekly. If you are starting completely from scratch veo3 is a helpful “all in one” to start getting your hands dirty, figuring out how to tune your prompts, but if you get more serious you’ll find yourself wanting to incorporate more tools. Search reddit for the latest and greatest tool recommendations.
Part of the process is trial by error as you start out. If you can’t tell I set a pretty low quality bar for this project. I probably took the “first take” for 40% of the final scenes. 1 out of 3 generations for another 50%, and for about 10% of scenes I went up to maybe 5-8 attempts or needed to do additional post processing.
Veo3 could not handle complexity. It could really only handle a single major action & dialogue within each clip, maaaybe two. For instance, “Cluck <does a thing> and says ‘<dialogue>’” tended to generate fairly reliable output. “Cluck <does a thing>, says ‘<dialogue>’, then Cluck <does another thing>” was almost always a hilarious failure. I often found I had to split what I thought could have been a single clip into two separate, more basic prompts and clips.
Step 4. Additional audio and misc AI generation
Some clips had no dialogue. Others had subtitles (even though the prompts said “Do not include subtitles”). I used ElevenLabs to generate additional dialogue audio, and even recorded a few clips of my own, i.e. the clip of high school Cluck getting turned down by giggling girls. My kids loved being included.
For background music I used Suno. Suno can work with custom lyrics or auto generate its own for you based on a prompt. Personally I like to work on lyrics with help from a chat bot (ChatGPT in this case), adjust, and then drop those in to Suno as custom lyrics.
As I neared completion of all the clips I also started generating some of my “marketing materials” for fun, like profile and banner images for a TC YouTube profile, but that’s all pretty optional.
Step 5. Movie editing
I used Adobe Premiere Pro (free trial) to start stitching everything together. Many people recommend DaVinci as an affordable alternative. I’ll try it for my next project.
Premiere is powerful but there was a definite learning curve. There are many tutorials out there if you want to start from scratch. I had enough very rusty Adobe Photoshop and other amateur editing experience from hobby photo & video editing that I just dove in, leaning heavily on ChatGPT for tailored tutorials to guide me when I knew what I wanted to do but not how to do it in the software.
Step 6. Post & share!
Post to YouTube or share directly with your friends and family!
Step 7. Lessons learned & how to do better
My approach was low effort and scrappy! There are creators out there using similar tools to generate much more polished output. Check out MetaPuppet, itsprobablyai, and many others. I follow /chatgpt and /aivideo on reddit for inspiration. Some tips and tricks out there right now that I may incorporate in my next project:
- For better character consistency, take a still shot from the end of the clip (or use a single core shot of a key character every time) and upload it to your video generator as part of the prompt & context. Downsides are that some tools won’t create dialogue from it, as a deep fake prevention.
- If you want more control over your characters, you can go a completely different path with a green screen & tools to map a gen AI face or character to your own, so you can act out the scene yourself. You can couple this with ElevenLabs to alter your voice to a custom character as well.
- Get creative, mix and match different tools! One example: MetaPuppet took an 8 second audio/music clip from veo3, fed it to Suno as a theme source, and extended that to generate soundtrack music for a longer video.
Have fun creating!