Midjourney's Amazing New Feature PLUS: StableVideo 1.1 from Stablity.AI!

Theoretically Media
1 Feb 202410:54

TLDRThe video discusses a mid-journey update on style consistency in AI image generation, introducing a new feature that combines image prompting with style tuning. It explores the use of style references and multiple image URLs to create a new style, demonstrating the process using the Mid Journey Alpha website. The video also delves into the capabilities and limitations of the feature, including the influence of single and combined style references. Additionally, it covers the early access to Stable Video from Stability, highlighting its open-source platform and features like camera motion and zooming, while noting that some features are still in development.

Takeaways

  • πŸš€ Introduction of a new feature in mid-journey for style consistency, combining image prompting and style tuning.
  • 🎨 Utilization of image URLs with prompts to create a new style, accessible via the MID Journey Alpha website.
  • πŸ“ˆ The requirement of having generated a certain number of images to access the MID Journey Alpha website.
  • πŸ”— Explanation of the --s, ref command for referencing images and its application in creating styled images.
  • 🌟 Demonstration of how the new feature can yield different results by adjusting the influence of reference images.
  • πŸ“Έ Difference between style referencing and simple image referencing, with examples of the outcomes.
  • πŸ€– Limitations of the feature, such as not being able to create consistent characters yet.
  • πŸ”„ The ability to combine multiple images for style references, resulting in blended and unique outputs.
  • πŸ“š Availability of a free PDF guide on gumroad for more information on the new feature.
  • πŸŽ₯ Stability's platform for stable video diffusion 1.1 in beta, with options to start with an image or text prompt.
  • 🌐 The inclusion of camera motion options like lock, shake, tilt, orbit, and pan in stable video.
  • 🎞️ Showcase of the quality of generated videos and the potential for creative use in various scenarios.

Q & A

  • What is the main focus of the mid Journey update discussed in the transcript?

    -The main focus of the mid Journey update is style consistency, specifically the introduction of a new feature that combines image prompting with style tuning to create a new style based on provided image URLs or multiple image URLs.

  • How does the new style reference feature work in mid Journey?

    -The style reference feature works by issuing the --s ref command along with the image you are referencing. This allows the user to create an image in a style that is influenced by the provided reference image URL.

  • What is the current access status for the new mid Journey Alpha website?

    -Access to the new mid Journey Alpha website has been opened to users who have generated more than 5,000 images, and users who have generated 1,000 images are expected to gain access soon.

  • How can users control the influence of each image URL in the style reference feature?

    -Users can control the overall influence of each image URL by using the waiting command, which allows them to adjust the intensity of the style reference, with options ranging from 1 to 1,000.

  • What is the difference between style referencing and simple image referencing?

    -Style referencing is different from simple image referencing in that it not only uses the image as a reference but also blends the style of the reference image with the generated content, creating a more stylistically consistent output.

  • What are some limitations of the new style reference feature?

    -The style reference feature does not currently support consistent characters and can become temperamental when pushed too far, especially when using three style references that do not have a thematic connection.

  • What is the current status of stable video from stability.a?

    -Stable video from stability.a is currently in beta and is available for free during this period. It offers a platform for stable video diffusion 1.1, which is open source and may be the underlying technology for other platforms.

  • What are some of the camera motion options available in stable video?

    -In stable video, users can lock the camera, shake the camera, tilt it down, perform an orbit, pan, and zoom in and out. There are also experimental camera motion options that users can explore.

  • How does the voting system work in stable video?

    -After generating a video, users can vote on which of the generations from other users they think looks good. This interactive feature allows for community involvement in the creative process and can be a way to pass time while waiting for generations to complete.

  • What are the text video options in stable video?

    -For text video, users have options for three different aspect ratios and can choose from a number of different styles to generate a video. They can input a text prompt and select from four generated options to find the one that best suits their needs.

  • What is the overall impression of the creative AI space based on the transcript?

    -The overall impression is that the creative AI space is rapidly advancing, with new features and platforms being developed and improved. The speaker is excited about the potential of these tools and looks forward to seeing the progress in the near future.

Outlines

00:00

🎨 Introducing Mid Journey's Style Consistency Feature

The paragraph discusses the introduction of a new style consistency feature in Mid Journey, an AI platform for image generation. It explains that users can utilize image URLs or multiple images along with a prompt to create a new style. The feature is likened to a blend of image prompting and style tuning. The video script provides a walkthrough of using the MID Journey Alpha website, which is accessible to certain users. It also describes how to issue commands to generate images with specific styles, such as referencing a Lara Croft image. The paragraph highlights the differences between this new feature and simple image referencing, and explores the possibilities of combining multiple images for style references. It also mentions the limitations of the feature, particularly in maintaining consistent characters, and notes that it is still in the alpha phase of development.

05:01

🌐 Exploring Style References and Stability.Video

This paragraph delves deeper into the intricacies of using style references in Mid Journey. It describes how the influence of each image URL can be controlled and how combining different images can lead to unique and interesting results. The script also touches on the challenges of using three unrelated style references and suggests that thematic connections can improve outcomes. The paragraph then shifts focus to Stability.Video, a platform for stable video diffusion developed by Stability.A. It discusses the open beta access and the features available for generating videos from images or text prompts. The video script provides examples of generated videos, including character animations and establishing shots, and comments on the quality and potential improvements. The paragraph concludes by emphasizing the rapid advancements in the creative AI space and the excitement for future developments.

10:02

πŸš€ Exciting Updates in AI Image and Video Generation

The final paragraph summarizes the exciting updates and features in AI image and video generation. It mentions the new capabilities of Mid Journey's style consistency feature and the potential for powerful combinations with other commands. The paragraph also highlights the free PDF available on Gumroad for further information. Turning to Stability.Video, the script discusses the platform's early access and the open-source nature of its technology. It provides insights into the types of videos that can be generated, including character animations and text-to-video options. The video script concludes with a reflection on the progress in the field and the anticipation for future advancements, leaving the audience excited about the potential of AI in creative content creation.

Mindmap

Keywords

πŸ’‘Mid Journey Update

The 'Mid Journey Update' refers to a significant upgrade to a creative AI platform that is focused on improving style consistency in generated content. This update introduces a new feature that allows users to create images with a more cohesive and uniform style by using image URLs or multiple image URLs alongside a prompt. It is a core aspect of the video's theme as it showcases the advancements in AI-assisted creativity and the continuous improvement of AI tools.

πŸ’‘Style References

Style references are a method used in the AI platform to guide the generation of images with a specific aesthetic or style. By providing an image URL or multiple URLs, users can direct the AI to incorporate elements from the referenced images into the new content, creating a unique blend of styles. This concept is central to the video's message, as it demonstrates the platform's capability to adapt and evolve its creative output based on user input.

πŸ’‘Stable Diffusion

Stable Diffusion is a term used to describe a type of AI model that generates stable and high-quality images or videos from textual or image inputs. It is a key concept in the video as it underpins the capabilities of the AI platform being discussed, particularly in relation to the generation of videos. The video highlights the platform's use of Stable Diffusion technology, which is open source and potentially the basis for other AI applications.

πŸ’‘Image Prompting

Image prompting is the process of providing AI with an image or a set of images as input to guide the generation of new images. This technique is integral to the video's narrative as it showcases how users can leverage existing visual content to influence the style and appearance of AI-generated images. The concept is used to illustrate the platform's ability to blend user intent with AI creativity.

πŸ’‘Style Tuning

Style tuning is the adjustment of stylistic elements in AI-generated content to achieve a desired aesthetic or visual effect. This concept is crucial to the video's theme as it highlights the level of control and customization available to users. It allows for fine-tuning the output to match a specific style or to create a harmonious blend of multiple styles, demonstrating the platform's versatility and responsiveness to user needs.

πŸ’‘Community Feed

The Community Feed is a platform feature where users can share and view images generated by others. It serves as a source of inspiration and a space for collaboration. In the context of the video, it is an example of how the AI platform fosters a community of creators who can build upon each other's work, enhancing the collective creativity and expertise of the user base.

πŸ’‘Discord

Discord is a communication platform where users can interact with each other and with the AI platform's community. In the video, it is mentioned as a place where commands and discussions related to the AI tool are shared, indicating that it is an essential part of the community's communication and support network.

πŸ’‘Alpha Version

An alpha version of a software refers to the early developmental stage of a product, which is typically used for internal testing and bug fixing before it is released to a wider audience. In the video, the mention of the 'Mid Journey Alpha' website signifies that the platform is still in the testing phase, and users with certain qualifications are given access to provide feedback and contribute to its development.

πŸ’‘Style Influence

Style influence refers to the degree to which the referenced images affect the final output of the AI-generated content. Users can control the strength of this influence, which is an important aspect of the video's theme as it demonstrates the level of customization available. It allows users to achieve a balance between the originality of the AI's creation and the elements borrowed from the reference images.

πŸ’‘Gumroad

Gumroad is a platform that allows creators to sell their work directly to consumers, often used for digital products like ebooks, software, and other creative content. In the context of the video, Gumroad is mentioned as a place where users can access a free PDF with more information about the AI platform's features, indicating that it is a resource for users to learn more and support the creators.

πŸ’‘Beta Period

A beta period is a testing phase for a software or service where it is made available to a wider audience to gather feedback and identify any remaining issues before the final release. In the video, the mention of the beta period for Stable Video indicates that the feature is still being tested and improved upon, and users have the opportunity to experience and contribute to its development.

Highlights

Introduction of a mid-journey update focusing on style consistency.

Exploration of a new feature that combines image prompting with style tuning.

Use of image URLs with prompts to create a new style, demonstrated through the MID Journey Alpha website.

Access to the MID Journey Alpha website is currently limited, but will soon be available to more users.

The ability to drag and drop an image for immediate style referencing.

Influence of reference images on the generated content, such as changing Lara Croft's appearance to resemble the reference.

Combining two different images as style references to create a blended style, like a cyberpunk woman and a dog Samurai.

Control over the influence of each image URL through the use of wait commands.

The provision of a free PDF on gumroad detailing the information, with donations appreciated.

Challenges with using three style references, leading to unusual results.

The potential of style referencing to inspire new creative directions, such as generating an astronaut in a coffee shop.

The limitation of the feature in maintaining consistent characters, with the upcoming release of Dash Dash CF.

The ability to increase the overall strength of style reference images with the --ssw command.

Introduction to Stability's platform for stable video diffusion 1.1, which is open source.

Options to start with an image or text prompt for video generation.

Features like camera lock, shake, tilt, orbit, pan, and zoom, available for video generation.

The experimental camera motion feature and its potential for interesting results.

The community voting system for generations from other users.

Examples of generated videos, including a pirate ship made of Swiss cheese and a crime film character.

Text video options with different aspect ratios and styles, demonstrated with a digital art style.

The current free access to stable video during its beta period.

Anticipation for the progress in the creative AI space and its potential impact by the end of the year.