Midjourney Video Updates + A Deeper Look at Sora

Curious Refuge
23 Feb 202413:21

TLDRThis week in AI news highlights the challenges and developments in AI filmmaking and tools. Sora's potential as a cinematic tool is discussed, alongside its rendering time and control limitations. A feature-length AI parody of Terminator 2 is announced, showcasing the talent of 50 AI artists. Updates to AI music generation models and sound effects are explored, with sunno's version 3 and 11 Labs' text-to-sound model. The potential of AI in the film industry, especially with the Disney accelerator program and AI-generated films, is emphasized. Lastly, the ongoing development of mid Journey and its future versions are teased, hinting at character consistency and AI video capabilities.

Takeaways

  • 🎬 The Hollywood Professional Association invited the creators to share their vision of a democratized filmmaking future with influential figures in Hollywood.
  • 📹 Sora's capabilities in creating realistic imagery were contrasted with Runway, highlighting a significant difference in their outputs.
  • 🚧 A one-minute clip in Sora requires approximately an hour of rendering time, which may not be ideal for a dynamic filmmaking process.
  • 🎥 Sora may not offer the desired level of control and consistency for filmmakers, especially in character and scene direction.
  • 🏆 A team of 50 AI artists created a feature-length parody of Terminator 2, showcasing the potential of AI in filmmaking.
  • 🎶 The AI music generation model by Sunno was updated to version 3, offering faster generations and more language support.
  • 📚 An AI filmmaking and advertising course is opening for enrollment, aiming to enhance storytelling skills through AI.
  • 🔊 11 Labs announced a new text-to-sound effects model, expanding the possibilities for sound design in AI-generated content.
  • 🏅 11 Labs was also recognized for joining the Disney accelerator program, emphasizing the collaboration between AI companies and entertainment studios.
  • 🖼️ Stability AI's stable diffusion version 3 model is upcoming, promising better quality and more control over inputs.
  • 🤖 Google's Gemini 1.5 pro model can process up to 1 million tokens of information, potentially impacting the film industry with its ability to analyze extensive content.
  • 🌐 Twitter is in discussions with MidJourney for integration, and Elon Musk expressed interest in incorporating AI art generators into social media platforms.

Q & A

  • What significant event was discussed at the beginning of the transcript?

    -The significant event discussed was the invitation of Shelby and the speaker to the Hollywood Professional Association's annual Tech Retreat, where they shared their vision for a democratized filmmaking future with over 800 influential individuals in Hollywood.

  • What are the differences between Sora and Runway as AI tools?

    -Sora and Runway differ in their ability to create realism. Sora was noted for creating more realistic outputs compared to Runway. However, it was also mentioned that Sora might not be the ideal cinematic filmmaking tool due to its high rendering time and current limitations in control and consistency over generated content.

  • What is the rendering time required for creating a one-minute clip in Sora?

    -It takes approximately one hour of rendering time to create a one-minute clip in Sora.

  • What is the significance of the Terminator 2 parody event in Los Angeles?

    -The event is significant because it is a feature-length parody of Terminator 2 created by a team of 50 talented AI artists, some of whom are students from Curious Refuge. It represents a milestone in AI filmmaking and offers an opportunity to be part of cinematic history.

  • What are the new features of the sunno AI music generation model in version 3?

    -Version 3 of the sunno AI music generation model offers faster generations, a dedicated instrumental button, and increased language support, making it more accessible to a broader range of users.

  • What is the significance of the 11 Labs' text to sound effects model?

    -The 11 Labs' text to sound effects model is significant because it allows users to generate sound effects by simply typing in a prompt. This is an advancement in sound design that could potentially automate the process of creating sound effects for films and other media.

  • What is the capacity of the Gemini 1.5 pro model in terms of information input?

    -The Gemini 1.5 pro model can input up to 1 million tokens of information, which is equivalent to approximately 700,000 words, 11 hours of audio, or 1 hour of video. This large capacity allows for extensive information processing and could potentially influence the future of AI-generated films.

  • What are the updates on the development of mid Journey?

    -Mid Journey 6 is in development and will feature character consistency, faster generation times, and improved aesthetics. There are also rumors about mid Journey 7, which may include AI video capabilities.

  • What was the outcome of the competition mentioned in the transcript?

    -Bck Reels won the $500 prize from AOL in the competition that was mentioned. The speaker encouraged the winner not to spend the prize money on Runway credits and to treat themselves to an extra scoop of ice cream instead.

  • What are the key takeaways from the AI film news discussed in the transcript?

    -The key takeaways include the challenges and developments of AI tools like Sora and mid Journey, the announcement of new features and versions of AI models, the impact of AI on the film industry, and the importance of partnerships between AI companies and studios to push the boundaries of storytelling.

Outlines

00:00

📺 AI News and Filmmaking Updates

The paragraph discusses the latest developments in the world of AI, particularly in the realm of filmmaking. It highlights the author's participation in the Hollywood Professional Association's Tech Retreat, where they shared their vision for a democratized future of filmmaking. The discussion then pivots to the AI tool Sora, comparing it with Runway and noting that while Sora offers a high level of realism, it may not be the ideal cinematic filmmaking tool due to its lengthy rendering times. The paragraph also touches on the importance of control and consistency in film creation, suggesting that traditional image-to-video workflows may continue to be prevalent. The author humorously mentions examples of Sora's limitations and shares an upcoming event in Los Angeles featuring a Terminator 2 parody by a team of AI artists. Additionally, the paragraph covers updates to the AI music generation model, Sunno, and announces an AI filmmaking and advertising course opening for enrollment.

05:02

🎵 Advancements in AI Sound and Music

This paragraph delves into the advancements in AI-generated sound effects and music. It starts by discussing the sound effects demo for a Sora project, noting the quality and potential for improvement. The paragraph then congratulates 11 Labs for their inclusion in the Disney accelerator program, emphasizing the collaboration between AI companies and studios to enhance storytelling. It also mentions a special announcement from Tim at theoretically media, highlighting a competition win and encouraging participation in future events. The discussion continues with the upcoming release of Stability AI's stable diffusion version 3 model, which promises better image quality and more control over inputs. The paragraph concludes with a challenge for viewers to identify images generated by different AI models and provides updates on Google Gemini's capabilities and recent issues with language models.

10:04

🚀 Innovations in AI Art and Integrations

The final paragraph focuses on the latest news and innovations in AI art generation and potential integrations with social media platforms. It begins with the announcement of Twitter's potential integration of the AI art generator, Mid Journey, and Elon Musk's commitment to including AI art in his platform, regardless of the deal's outcome. The paragraph then discusses the ongoing development of Mid Journey 6, which promises character consistency and improved aesthetics, and speculates on the rumored Mid Journey 7's capabilities. The section also highlights a humorous Will Smith meme related to Sora and showcases AI films of the week, including 'The Pomegranate Spell,' a film by Jamie Roa Cassetti, and 'I Want to Be Happy,' a story about a robot experiencing emotions. The paragraph ends by encouraging viewers to sign up for the AI filmmaking course and to stay updated with AI film news through Curious Refuge.

Mindmap

Keywords

💡AI tools

AI tools refer to various software applications and technologies that utilize artificial intelligence to perform tasks. In the context of the video, these tools are associated with content creation, such as filmmaking and music generation. The script discusses several AI tools like Sora, Runway, and Sunno, which are designed to assist in different aspects of media production, highlighting their capabilities and limitations.

💡Democratized filmmaking

Democratized filmmaking refers to the idea that the tools and processes for creating films are becoming more accessible to a wider range of people, not just established industry professionals. This is often facilitated by advancements in technology, such as AI, which can lower the barriers to entry for independent filmmakers and enthusiasts. The video discusses the presenter's vision for a future where AI tools enable more individuals to participate in filmmaking, breaking down traditional barriers.

💡Rendering time

Rendering time is the period a computer system takes to process and generate a visual output, such as a video or animation, from a set of data or instructions. In the context of the video, it is mentioned that creating a one-minute clip in Sora requires an hour of rendering time, which is a measure of the efficiency and practicality of AI tools for filmmaking. Longer rendering times can impact the speed and flexibility of the creative process.

💡Cinematic filmmaking

Cinematic filmmaking refers to the creation of movies that adhere to traditional film industry standards in terms of storytelling, visual aesthetics, and technical quality. This concept is central to the video's discussion of AI tools, as it explores whether these tools can meet the high standards required for cinematic quality. The presenter expresses skepticism about Sora's ability to serve as a cinematic filmmaking tool due to its current limitations.

💡Art direction

Art direction involves the visual style and overall look of a film, including the design of characters, sets, and scenes. It is a crucial aspect of filmmaking that requires a high degree of control and consistency. In the video, the presenter mentions the importance of art directing in films and suggests that current AI tools like Sora may not yet offer the level of control needed for detailed art direction.

💡AI music generation

AI music generation is the process of using artificial intelligence to create original music compositions. This technology can analyze styles, genres, and musical patterns to produce new pieces of music. In the video, the presenter discusses Sunno, an AI music generation model that has updated to version 3, offering faster generation times and more language support, making it more accessible and user-friendly.

💡AI film news

AI film news refers to updates and developments in the intersection of artificial intelligence and filmmaking. This includes new AI tools, industry events, and advancements that impact the way films are created and consumed. The video serves as a news update, providing insights into the latest trends and breakthroughs in AI-assisted filmmaking.

💡Cinematic history

Cinematic history refers to the evolution and significant events in the field of filmmaking. In the context of the video, the term is used to emphasize the importance of the work being done by AI artists, suggesting that their collective efforts might become a notable part of the broader history of cinema. The video mentions an event in Los Angeles where AI artists are creating a feature-length parody, which could be considered a milestone in the integration of AI into filmmaking.

💡AI-generated films

AI-generated films are movies that are created primarily through the use of artificial intelligence, without or with minimal human intervention. These films represent a new frontier in storytelling and filmmaking, as AI can potentially handle various aspects of film production, from scriptwriting to visual effects. The video discusses the potential for AI to revolutionize filmmaking by generating films and the steps being taken towards achieving this goal.

💡Mid Journey

Mid Journey is an AI platform mentioned in the video that specializes in generating art and visuals based on user inputs. The video discusses the ongoing development of Mid Journey, including its improvements in character consistency and aesthetics, as well as rumors about future versions that may include AI video capabilities. This reflects the rapid evolution of AI tools in the creative space.

💡AI film-making course

An AI film-making course is an educational program focused on teaching the skills required to utilize artificial intelligence in the creation and production of films. The video mentions the opening of enrollment for such a course, indicating a growing interest and demand in learning how to integrate AI into the filmmaking process. This reflects the increasing relevance of AI in creative industries and the need for education to keep pace with technological advancements.

Highlights

The channel focuses on AI news and tools, providing updates on the latest developments in the field.

The speakers were invited to the Hollywood Professional Association's annual Tech Retreat to share their vision for the future of filmmaking.

Sora was compared with Runway, showing a significant difference in the level of realism they can produce.

It was revealed that creating a one-minute clip in Sora requires about an hour of rendering time, which may not be ideal for a collaborative filmmaking process.

Sora may not offer the desired control and consistency for filmmakers, at least at its current stage.

A team of 50 AI artists created a feature-length parody of Terminator 2, showcasing the potential of AI in filmmaking.

Sunno's AI music generation model was updated to version 3, offering faster generations and more features.

The AI filmmaking and advertising course is opening up enrollment, aiming to enhance storytelling skills with AI.

11 Labs announced a new text-to-sound effects model, expanding the capabilities of AI in sound design.

Theoretically Media congratulated Black Reels for winning a competition, highlighting the community engagement in AI creations.

Stability AI's stable diffusion version 3 model is upcoming, promising better quality and more control over inputs.

A game was introduced where viewers need to identify images generated by different AI models based on a complex prompt.

Google paused image generations on Gemini due to some politically incorrect outputs, but also released a Gemini 1.5 pro model capable of handling large amounts of text.

There are plans to integrate AI art generators into social media platforms like Twitter, with Elon Musk expressing interest in incorporating them into his platforms.

Mid Journey is continuously being developed, with version 6 focusing on character consistency and improved aesthetics.

Rumors suggest that Mid Journey 7 might introduce AI video capabilities, raising expectations for future AI advancements.

AI-generated content continues to go viral, such as the humorous Will Smith eating spaghetti meme, demonstrating the cultural impact of AI.

AI film news highlights include diverse entries like the pomegranate spell, a retelling of a myth with beautiful shots, and 'The File' showcasing advanced 3D models.