Midjourney Video Updates + A Deeper Look at Sora

Curious Refuge
23 Feb 202413:21

TLDRThe video script discusses recent developments in AI, highlighting the Hollywood Professional Association's Tech Retreat and the potential of AI in filmmaking. It critiques Sora's rendering time and control limitations for film production, praises the AI-generated parody of Terminator 2, and reviews advancements in AI music generation with Sunno. The script also covers new text-to-sound effects models and congratulates 11 Labs for their Disney accelerator program. It mentions the challenges in language models, Elon Musk's interest in integrating AI art generators into Twitter, and the ongoing development of MidJourney, including its character consistency feature. The video concludes with a selection of notable AI films and encourages viewers to join the AI filmmaking course and stay updated through newsletters.

Takeaways

  • 🎬 The Hollywood Professional Association invited the speaker to share visions for a democratized filmmaking future at their annual Tech Retreat.
  • 📹 Sora's capabilities in creating realistic visuals were contrasted with Runway, with Sora showing more promise but requiring significant rendering time.
  • 🚧 The latency of Sora may pose challenges for real-time filmmaking and art direction, suggesting that image-to-video workflows will remain essential for creative direction.
  • 🎥 A team of 50 AI artists created a feature-length parody of Terminator 2, showcasing the collaborative potential of AI in filmmaking.
  • 🎵 The AI music generation model by Sunno was updated to version 3, offering faster generations, dedicated instrumental creation, and expanded language support.
  • 📚 An AI filmmaking and advertising course is opening enrollment, aiming to enhance storytelling skills through the use of AI.
  • 🔊 11 Labs announced a new text-to-sound effects model, further enhancing the immersive capabilities of AI in media production.
  • 🏆 11 Labs secured a spot in the Disney accelerator program for 2024, highlighting the growing partnership between AI companies and traditional studios.
  • 🖼️ Stability AI's stable diffusion version 3 model is upcoming, promising better image quality and more control over inputs.
  • 🤖 Gemini's 1.5 pro model can process up to 1 million tokens of information, potentially revolutionizing AI's understanding and generation of content.
  • 🌐 Twitter is in discussions with MidJourney for integration, and Elon Musk expressed interest in incorporating AI art generators into social media platforms.

Q & A

  • What significant event did the speaker and Shelby attend recently?

    -The speaker and Shelby attended the annual Tech Retreat hosted by the Hollywood Professional Association, where they shared their vision for a democratized filmmaking future with over 800 influential individuals in Hollywood.

  • What are the key differences between Sora and Runway as highlighted in the video?

    -Sora creates a higher level of realism compared to Runway. However, it was revealed that creating a one-minute clip in Sora requires approximately an hour of rendering time, which may not be ideal for a back-and-forth filmmaking process.

  • Why might Sora not be the ideal tool for cinematic filmmaking at the moment?

    -Sora's latency and the challenge of maintaining control and consistency over the generated content make it less suitable for the bulk of film production. It may be more useful for detail shots or establishing shots rather than the entire film.

  • What exciting event is happening in Los Angeles involving AI artists?

    -A team of 50 AI artists is putting together a feature-length parody of Terminator 2. Some of these artists are students from Curious Refuge. The event is an opportunity to be a part of cinematic history.

  • What are the new features of the sunno AI music generation model in version 3?

    -Version 3 of the sunno AI music generation model offers faster generations, a dedicated instrumental button, and increased language support, making it more accessible to a broader audience.

  • When is the enrollment opening for the AI filmmaking and AI advertising course?

    -The enrollment for the AI filmmaking and AI advertising course is opening on February 28th. Interested parties should set alarms for 11:00 a.m. Pacific Time on that date.

  • What is the significance of the new text-to-sound effects model by 11 Labs?

    -11 Labs' new text-to-sound effects model allows users to generate sound effects by simply typing in a prompt. This technology is a step towards automatic sound effects generation for films, potentially revolutionizing the sound design process.

  • What achievement did 11 Labs accomplish in relation to Disney?

    -11 Labs was selected for the Disney accelerator program for 2024, which is a notable recognition for their innovative work in AI and a sign of their potential to push the boundaries of storytelling in collaboration with major studios.

  • What are the notable updates in the stable diffusion version 3 model by stability AI?

    -The stable diffusion version 3 model by stability AI offers better image quality and more textual control over the inputs. It is capable of generating complex and detailed images based on specific prompts.

  • What is the significance of the Gemini 1.5 pro model's capability to input up to 1 million tokens of information?

    -The Gemini 1.5 pro model's ability to input up to 1 million tokens of information means it can process a significant amount of text, audio, or video data. This capability could potentially be used to create AI-generated films by reverse-engineering information from existing videos.

  • What are the upcoming features for mid Journey 6 and rumors about mid Journey 7?

    -Mid Journey 6 is set to include character consistency, faster generation times, and improved aesthetics. There are also rumors that mid Journey 7 might introduce AI video capabilities, although this is yet to be confirmed.

Outlines

00:00

📣 Introduction and Industry Updates

The paragraph begins with the host expressing gratitude to the Hollywood Professional Association for inviting them to share their vision for the future of democratized filmmaking. They discuss their interactions with industry professionals and segue into AI news. The host revisits Sora, an AI tool, and compares it with Runway, highlighting the significant differences. They discuss the challenges in using Sora for cinematic filmmaking due to its long rendering times and lack of control over the creative process. The host suggests that AI tools for image-to-video workflows will continue to be essential for creative direction. They also mention humorous examples of Sora's limitations and discuss an upcoming event in Los Angeles featuring a Terminator 2 parody by a team of 50 AI artists. The segment concludes with an introduction to an AI music generation model, Sunno, which has updated to version 3, offering faster generations and more features.

05:02

🎵 Advancements in AI Sound and Music

This paragraph focuses on the progress in AI-generated sound effects and music. The host discusses a demo of sound effects created by AI and acknowledges that while some effects are impressive, others require refinement. The news includes the announcement of 11 Labs' new text-to-sound effects model and their acceptance into the Disney accelerator program. The host congratulates 11 Labs and discusses the potential of AI in storytelling through partnerships with studios. They also highlight a special announcement from a representative at theoretically media, followed by a discussion on the capabilities of the new Gemini 1.5 pro model, which can process up to 1 million tokens of information, potentially revolutionizing the film industry by enabling AI to generate films based on comprehensive data analysis.

10:04

🎥 AI Film News and Developments

The final paragraph covers the latest news in AI filmmaking and the development of AI tools. The host discusses the upcoming features of MidJourney 6, including character consistency and improved aesthetics, and rumors about MidJourney 7 potentially including AI video capabilities. They mention the humorous incident with Will Smith and a parody video created with Sora. The host then highlights three AI films: 'The Pomegranate Spell,' a retelling of a myth with beautiful animation; 'The File' by Jamie Roat Cassetti, featuring advanced 3D models and interesting macro V effects; and 'I Want to Be Happy,' a story about a robot experiencing emotions. The segment ends with a call to action for viewers to join the March session of an AI filmmaking course and to sign up for AI film news updates.

Mindmap

Keywords

💡AI news

AI news refers to the latest updates and developments in the field of artificial intelligence. In the context of the video, it is the central theme as the channel focuses on providing news and insights about AI tools and their applications in various industries, particularly in filmmaking.

💡Sora

Sora is an AI tool mentioned in the video that is used for creating realistic visuals, potentially for films or other media. However, it is noted that Sora may not be the ideal cinematic filmmaking tool due to its long rendering times and challenges in achieving consistency and control over generated content.

💡Runway

Runway is an AI software platform that is compared to Sora in the video. While Sora is noted for its ability to create realistic visuals, Runway is mentioned in contrast, suggesting that it may not achieve the same level of realism as Sora in certain applications.

💡Cinematic filmmaking

Cinematic filmmaking refers to the creation of movies that have high production values, often characterized by a certain style or aesthetic. In the context of the video, it discusses the challenges of using AI tools like Sora in this process due to control and consistency issues.

💡Image to video workflows

Image to video workflows involve the process of creating videos from static images, often using AI tools to generate or enhance content. The video suggests that these workflows, which allow for creative direction, will continue to be important in AI video production, as they offer more control over the final product than some AI tools like Sora.

💡AI music generation

AI music generation refers to the use of artificial intelligence to create or compose music. In the video, the AI music generation model 'Sunno' is highlighted for its ability to generate music based on user input, showcasing the versatility and accessibility of AI in the music industry.

💡AI filmmaking course

An AI filmmaking course is an educational program focused on teaching the skills needed to utilize artificial intelligence in storytelling and film production. In the video, the host mentions the opening of enrollment for such a course, emphasizing the growing importance of AI in the film industry.

💡Sound effects model

A sound effects model is an AI system designed to generate audio effects from textual prompts. In the video, the team at 11 Labs is mentioned to be working on a text to sound effects model, which signifies the expanding role of AI in audio production.

💡Stable diffusion

Stable diffusion is a term used in the context of AI image generation models, referring to the process of creating images from textual prompts with a high degree of control over the output. The video mentions a new version of the stable diffusion model, indicating continuous improvements in AI's ability to generate content.

💡Gemini 1.5 pro model

The Gemini 1.5 pro model is an AI system that has the capability to process a large amount of text, allowing it to read, understand, and reference information from extensive inputs. In the context of the video, this model's ability to handle large datasets is highlighted as a significant development for AI's potential in the film industry.

💡AI art generator

An AI art generator is a tool that uses artificial intelligence to create visual art, often based on user input or prompts. In the video, the integration of an AI art generator into social media platforms like Twitter is discussed, indicating the growing integration of AI into creative and social spheres.

Highlights

The channel focuses on AI news and tools, providing updates on the latest developments in the field.

The speakers were invited to the Hollywood Professional Association's annual Tech Retreat to share their vision for the future of filmmaking.

Sora was compared with Runway, showing a significant difference in the level of realism they can produce.

It was revealed that creating a one-minute clip in Sora requires about an hour of rendering time, which may not be ideal for a collaborative filmmaking process.

Sora may not offer the desired control and consistency for filmmakers, especially in character and scene creation.

A team of 50 AI artists created a feature-length parody of Terminator 2, showcasing the potential of AI in filmmaking.

Sunno's AI music generation model was updated to version 3, offering faster generations and more features.

The AI filmmaking and advertising course is opening up enrollment, aiming to enhance storytelling skills using AI.

11 Labs announced a new text-to-sound effects model, further enhancing the capabilities of AI in audio production.

Theoretically Media congratulated Black Reels for winning a competition, highlighting the community engagement in AI-generated content.

Stability AI's stable diffusion version 3 model is upcoming, promising better quality and more control over inputs.

A challenge for the audience to identify images generated by different AI models based on a complex prompt.

Google paused image generations on Gemini due to politically incorrect outputs, but also released a test of Gemini 1.5 pro model with enhanced capabilities.

Elon Musk's interest in integrating AI art generators into Twitter or other platforms signifies the growing importance of AI in social media.

Mid Journey continues to develop with version 6 offering character consistency and improved aesthetics, with rumors of AI video capabilities in version 7.

A humorous incident with AI model Chat BT spouting nonsensical words, emphasizing the ongoing need for developer intervention in language models.

The video highlights three AI films of the week, showcasing the diversity and creativity in AI-generated content.

The pomegranate spell, a retelling of the myth of Pan, demonstrates the aesthetic potential of AI in film through beautiful romantic shots and animation.

The File by Jamie Roa Cassetti features advanced 3D models and interesting movement, indicating a growing sophistication in AI's filmmaking capabilities.

I Want to Be Happy, a film about a robot experiencing emotions, illustrates the ability of AI to explore complex themes and characters.