New Show Hacker News story: Show HN: AI prompt-to-storyboard videos w/ GPT, Coqui voices, StabilityAI images

Show HN: AI prompt-to-storyboard videos w/ GPT, Coqui voices, StabilityAI images
3 by tensorflowz | 1 comments on Hacker News.
I had 2 weeks off from work and wanted a pet project before heading back. With GPT and Generative AI in the news, I decided to chain multiple Al products together to build something really cool. I set my end goal to be: prompt-to-storyboard (aka fun videos generated purely via generative AI). There exists some prompt-to-video products, but I wanted to tell stories with audio as well. The end product takes an initial prompt and produces a series of images and audio files, which I then combine (with subtitles) into the final video. To showcase videos, there is a basic upvote/downvote leaderboard. Text | OpenAI https://openai.com/ Text is generated in a few high-level steps that I ask GPT to work through. These are all based on the initial user prompt, as such (ideally) indirectly controlled by the user. - Create a concept for a movie scene based on the prompt, including the theme and setting - Define each character in the scene - Define how each character looks - Define how each character sounds - Define 'frames' of the storyboard All of this textual information is defined in a JSON object I describe to GPT. I then take GPT's output and build the storyboard with the tools below. Voices | Coqui https://coqui.ai/ From the GPT output, I needed three major pieces of information to build voices in a way that I found satisfying: - Description of the voice - Description of the performance - Text of the actual dialog spoken Coqui has a product called 'prompt-to-voice', where you can describe how a character will sound and a custom voice is made for that character - this is how GPT defines the characters to use in the storyboard. As such, every voice is unique per storyboard. GPT will decide that a certain character is an "older man with a raspy voice", and I'll ask Coqui for that type of voice. In addition to this, in order to describe the performance, GPT outputs a basic emotion to summarize the line of dialog (happy, sad, angry, etc) - this is also sent to Coqui per audio clip generated. Images | Stability AI https://stability.ai/ While I originally setup the storyboard generator to use DALL-E due to already integrating with OpenAl for GPI, I found the cost prohibitive. As such, the images generated for the storyboards are from Stability Al's Stable Diffusion (stable-diffusion-512-v2-1). I combine the description of the frame that GPT provides, in addition to the theme and setting that GPT output for the whole storyboard, to generate each frame. Since GPT controls the data sent to Stable Diffusion with the description of the frame as well as the theme and setting, if your prompt dictates a theme it should hopefully translate into a theme in your storyboard. Both the storyboard and the 'prompt enhanced' image generation in the 'Create Content' tab pre-feed a GPT request with a summary of Stability Al's prompt guide. It will try and pick keyword weights to improve the image, and much like the setting and theme, keywords should be influenced by the initial prompt provided to the product. Conclusion: Have fun and make my 2 weeks of work seem worth it! Voting on storyboards and creating storyboards both require a simple Google login to get access.

Comments

Popular posts from this blog

New Show Hacker News story: Show HN: Natural language Twitter search using Codex

Internet Download Manager Universal Crack is Here ! [IDM 6.25 Build 10 UPDATED]

New Show Hacker News story: Show HN: Movis – A Video Editing Library in Python