Midjourney + Runway Gen 3 Turbo Full Course (Create Lifelike Cinematic AI Videos)
TLDRThis video tutorial delves into the creation of lifelike cinematic AI videos using Midjourney and Runway Gen 3 Turbo. It showcases the democratization of filmmaking and the journey towards super photorealism. The video is structured into three parts: storyboarding, transforming storyboards into cinematic shots with consistent style and character using Midjourney, and creating cinematic videos with Runway's image-to-video mode. It also offers practical tips on prompting, camera movement control, and compares the output quality of Runway Gen 3 Turbo with the base model, providing a comprehensive guide for aspiring AI filmmakers.
Takeaways
- 😀 The era of AI films has begun, with Runway Gen 3 Turbo leading to the democratization of film making and faster ideation.
- 🎥 The presenter has come closer to super photorealism and coherence in AI video creation since using Gen 2, showcasing the advancements in AI tools.
- 📝 The video is structured into three main parts: storyboarding, creating cinematic shots with consistent style using Midjourney, and generating cinematic videos with Runway's image-to-video mode.
- 🚀 Runway Gen 3 Turbo is significantly faster than previous models, allowing for quicker video generation compared to traditional methods.
- 🔍 The video compares the output quality of Runway Gen 3 Turbo with the base model, noting improvements in body coherence and detail accuracy.
- 📚 Prompting tips are based on the official Runway Gen 3 prompting guide and实践经验, emphasizing the importance of setting, location, and time for consistency.
- 🖼️ Storyboarding is done using Midjourney for character and style consistency, with a focus on the synopsis and scene ideas rather than image quality.
- 🌟 Character references and style references are crucial for maintaining consistency across scenes, with adjustments to character weight to allow for changes in outfits and scenarios.
- 🎬 Camera movement is controlled in Runway to achieve the best results, with the option to use the first frame as a starting point or an endpoint for the video generation.
- 💬 Runway's lip sync feature allows for the addition of dialogue and hand gestures, enhancing the character's expressiveness in the generated videos.
- 💰 Runway's pricing has been updated with Gen 3 Turbo, reducing the cost to 5 credits per second of video, but the need for multiple tries before satisfaction may still deplete credits quickly.
Q & A
What era has begun according to the video?
-The era of AI films has begun, indicating a significant shift towards the democratization of film making with advanced technologies.
What is Runway Gen 3 Turbo and how does it relate to cinematic AI videos?
-Runway Gen 3 Turbo is a tool that allows for the creation of lifelike cinematic AI videos, enabling users to ideate at the pace of their imagination and move closer to super photorealism with excellent coherence.
How does the video demonstrate the progression from Gen 2 to Gen 3 Turbo in AI video creation?
-The video showcases the progression by comparing the capabilities of Gen 2 with the advanced features of Gen 3 Turbo, highlighting the improved speed and quality in generating cinematic AI videos.
What are the three main parts of the process covered in the video?
-The video covers three main parts: storyboarding for AI films, transforming storyboards into ultra-realistic cinematic shots, and creating cinematic videos using Runway's image to video mode.
Why is specifying setting, location, and time important when using DALL-E for storyboarding?
-Specifying setting, location, and time is crucial to ensure that DALL-E produces images with consistent time periods and avoids inconsistencies in the generated images.
How does the video suggest using character and style consistency in the AI video creation process?
-The video suggests using character and style consistency by leveraging DALL-E for storyboarding and Midjourney for generating ultra-realistic images, ensuring that the characters and style remain coherent throughout the video.
What is the significance of using image prompting and style references in Midjourney?
-Using image prompting and style references in Midjourney is significant as it helps establish the right tone and style for the images, allowing for the creation of a consistent and immersive series of images.
How does the video guide viewers in creating cinematic videos using Runway's image to video mode?
-The video guides viewers by explaining how to control camera movement for the best results and make the most out of Runway's powerful image to video engine, including comparing output quality between Genry Turbo and the base model.
What is the purpose of using character reference weight in Midjourney?
-The purpose of using character reference weight is to simplify the decision-making process when directing a scene, allowing for more or less flexibility in changing the character's appearance through text prompts while preserving facial features.
How does the video address the issue of maintaining character consistency throughout the AI video?
-The video addresses character consistency by demonstrating the use of character references and style references, as well as adjusting character weight values to balance the influence of the reference image with text prompts.
What is the role of shot types in the prompts used for image to video process in Runway?
-Shot types in the prompts play a crucial role in the image to video process by providing a well-defined starting or ending frame, enhancing creative control and ensuring a coherent cinematic narrative.
Outlines
🎬 Introduction to AI Filmmaking with Runway Gen 3 Turbo
The script introduces the era of AI films and the potential democratization of filmmaking with Runway Gen 3 Turbo. It highlights the advancements from the initial use of Gen 2 to the current capabilities, which are nearing super photorealism and coherence. The speaker shares their experience in pushing the boundaries with AI tools and provides a tutorial on creating a mini AI film. The video is structured into three parts: storyboarding, transforming storyboards into cinematic shots with M Journey, and creating cinematic videos with Runway's image-to-video mode. The process involves testing various ideas and sharing practical tips based on the official Runway Genry prompting guide. The workflow begins with writing a synopsis, choosing a character and setting, and then using CET GPT cloud or similar models to generate scenes. The speaker emphasizes the importance of specifying setting, location, and time for consistency in image generation.
🖌️ Transforming Storyboard to High Fidelity with Mid Journey
This paragraph details the process of upgrading a storyboard from low to high fidelity using Mid Journey's image prompting and style reference. The speaker explains how to use image prompting alongside detailed visual text descriptions from CAD GPT to enhance the image conversion to video. They discuss the importance of image weight in balancing the image and text prompts and share reference values for convenience. The paragraph also covers the use of previously generated images as style references and provides tips for refining character depictions and directing upcoming scenes. The speaker demonstrates how to use character weight to maintain consistency in character appearance across different scenes and how to combine image prompts with shot types to enhance the image-to-video process.
🌟 Creating a Cinematic Universe with Image Prompting
The speaker discusses the creation of a cinematic universe using image prompting, focusing on the use of brackets for permutations or batch prompting in Mid Journey. This technique allows for multiple image generations with a single prompt, with the 'chaos' parameter controlling the variation among the generated images. The speaker recommends staying within a chaos value range of 4 to 8 for maintaining prompt adherence while allowing for variety. They also mention the use of style references to ensure consistency with the rest of the generated images. The paragraph concludes with tips for handing over the generated images to Runway for further processing, including organizing images through liking and folder utilization, and the importance of upscaling for quality enhancement.
🚀 Exploiting Runway's Image-to-Video Capabilities
This paragraph delves into the use of Runway's image-to-video capabilities, emphasizing the speed and quality of Genry Turbo. The speaker describes how to set camera movements based on the first frame and the creative direction, whether starting from or leading to a specific frame. They mention the addition of prompt presets for easier camera control and the official prompting guide for text-to-video and image-to-video. The guide suggests focusing on camera movement, character, and scene movement rather than describing the image itself. The speaker provides examples of how to structure prompts with camera techniques, follows X format, and events, and how to combine these elements for cinematic footage. They also touch on the use of keywords for speed control and maintaining color consistency.
🎥 Advanced Prompting Techniques and Common Issues
The speaker explores advanced prompting techniques for image-to-video, emphasizing the simplicity of using the 'camera follows X' format and the importance of describing camera movement. They provide examples of how to combine camera techniques, movement, and events in a single prompt and discuss the flexibility in prompt construction based on the image and desired complexity. The paragraph also addresses common problems with Runway, such as slow motion and color changes, offering solutions like using speed-related keywords and adjusting color consistency with specific prompts. The speaker highlights Runway's ability to depict emotions and demonstrates the lip sync feature, which allows for character speech and hand gestures based on user input. They conclude with a discussion on Runway's pricing, suggesting a low-resolution preview option for better cost control and user satisfaction.
🔄 Utilizing Fixed Seed for Consistent Style
In the final paragraph, the speaker introduces the concept of a fixed seed on Runway, which creates visually similar generations akin to style references in Mid Journey. They explain that the fixed seed is unchecked by default but can be used to maintain a consistent style and moment in the output. The speaker concludes the video by encouraging viewers to support the content and join the community for more tutorials on cinematic AI video making.
Mindmap
Keywords
💡AI films
💡Runway Gen 3 Turbo
💡Super photo realism
💡Storyboarding
💡Midjourney
💡Image to video mode
💡Cinematic coherence
💡Prompting guide
💡Character weight
💡Lip sync feature
Highlights
The era of AI films has begun, with cinematic Singularity and the democratization of film making on the horizon.
Runway Gen 3 Turbo allows for ideation at the pace of imagination, lighting fast.
We are approaching Super photorealism and excellent coherence with AI tools.
A mini AI film was published to test the boundaries of what's possible with AI tools today.
The video will provide a deep dive tutorial on creating a mini AI film, 'A Blast from the Past'.
Practical ideas and prompting tips based on the official Runway Genry prompting guide will be shared.
The process consists of three main parts: storyboarding, creating cinematic shots with consistent style and character, and generating cinematic videos.
A rough draft of a synopsis is the starting point for the ideation phase.
CET GPT cloud or other large language models can generate scenes based on a brief synopsis draft.
16 scenes were written as part of the storyboarding process for the mini film.
Dolly is used for storyboarding due to its excellent character and style consistency.
Visual descriptions of each storyboard image are created using a specific prompt with CAT GPT.
Ultra-realistic cinematic images are generated using the atomic prompting method on Mid Journey.
Style exploration is crucial to establish a consistent tone throughout the series of images.
Image prompting and style reference are key in transforming low Fidelity images to High Fidelity.
Character depiction is refined using Mid Journey prompts for a clearer view of the character's face.
Character reference weight is explained to direct a scene with character reference effectively.
Inclusion of shot types in prompts benefits the image to video process by providing a well-defined first or last frame.
Batch prompting on Mid Journey allows for the creation of multiple image generations with a single prompt.
Upscaling of images can be done using Mid Journey's upscalers or other tools like Magnific.
Camera movement in Runway is directed based on the first frame, which can be set as either the first or last frame of the video.
Runway Genry Turbo has significantly reduced the cost to 5 credits per second of video.
Prompt presets in Runway help beginners control the camera more comfortably.
For image to video, focus on camera movement, character, and scene movement rather than describing the image itself.
Common problems with Runway and their solutions, such as slow motion and color changes, are discussed.
Runway's lip sync feature allows for the addition of voice and hand gestures to characters.
Fixed seed on Runway creates visually similar generations, acting like a style reference.
A low resolution preview before generating full resolution footage is recommended for better cost control.