What new AI features is YouTube adding to Shorts?
Cover Photo Major News from Youtube, Google DeepMind's Veo AI, Runway, Lionsgate, Hume's EVI 2, LinkedIn, Apple, EzAudio and Magnit

YouTube Enhances Shorts with AI-Powered Video Generation

YouTube is integrating Google DeepMind’s Veo AI model into Shorts, allowing creators to generate high-quality backgrounds and six-second clips. This upgrade from the previous “Dream Screen” feature enables more impressive video creation and editing capabilities. Creators can now produce standalone video clips and add filler scenes for smoother transitions. The AI-generated content will be watermarked using DeepMind’s SynthID technology. YouTube also announced new features like “Jewels” for viewer interaction during livestreams, expanded automatic dubbing support, and the ability to use AI for brainstorming video ideas, generating thumbnails, and responding to comments. The platform is also expanding its Community hubs and “hyping” feature to more channels and markets.

Runway Partners with Lionsgate for AI-Powered Film Production

Runway, an AI startup, has partnered with Lionsgate to develop a custom AI video production and editing model. This collaboration marks the first major deal between an AI video provider and a major film studio. Runway will train its model using Lionsgate’s extensive library of film and TV titles. The technology will initially be used for storyboarding, background creation, and special effects, particularly for action scenes. This partnership aims to streamline production processes and reduce costs. While AI-generated full-length films are not imminent, the technology is already assisting in producing film segments and short clips. The deal represents a significant step in AI’s integration into the film industry, potentially impacting future production methods and creative processes.

Hume Unveils EVI 2: Advanced Emotionally Intelligent Voice AI

Hume has launched Empathic Voice Interface 2 (EVI 2), an upgraded AI voice model and API. EVI 2 offers improved naturalness, emotional responsiveness, and customizability in voice interactions. The new model boasts faster response times, enhanced emotional intelligence, and customizable voices without relying on voice cloning. It supports multiple languages and can be integrated directly into apps for a seamless user experience. EVI 2 is designed to anticipate and adapt to user preferences in real-time, making it suitable for various applications from customer service to virtual assistants. With reduced pricing and increased efficiency, Hume aims to make advanced voice AI more accessible to developers and businesses, potentially revolutionizing how voice interactions are integrated into digital platforms.

LinkedIn’s AI Training Practices Raise Privacy Concerns

LinkedIn has been using user data to train AI models without initially updating its terms of service, sparking privacy concerns. The platform now offers an opt-out toggle for users in certain regions, allowing them to choose whether their data is used for training content creation AI models. LinkedIn claims to use privacy-enhancing techniques to limit personal information in training datasets. The company’s practices have drawn criticism from privacy advocates, who argue for an opt-in model instead. This situation highlights the growing trend of social media platforms repurposing user-generated content for AI training, often without explicit consent. The incident underscores the need for greater transparency and user control over personal data usage in AI development.

Apple Expands Language Support for Generative AI Feature

Apple is set to broaden the linguistic reach of its generative AI offering, the Apple Intelligence, in 2025. The expansion will include support for German, Italian, Korean, Portuguese, Vietnamese, and several English variants. Initially launching in American English with iOS 18.1, the feature will gradually roll out to other languages and regions. However, regulatory challenges are hindering its release in the European Union and China, two significant markets. Apple is actively engaging with authorities in both regions to address these issues. The company’s efforts highlight the complex interplay between technological innovation and international regulations in the AI sphere.

EzAudio: Tencent’s Breakthrough in AI-Generated Sound

Tencent AI Lab and Johns Hopkins University researchers have unveiled EzAudio, a groundbreaking text-to-audio AI model. This innovative system generates high-quality sound effects from text prompts with remarkable efficiency, outperforming existing open-source models in various evaluations. EzAudio’s novel approach, operating in the latent space of audio waveforms, eliminates the need for spectrograms and neural vocoders. While this advancement promises exciting applications in entertainment, accessibility, and virtual assistants, it also raises ethical concerns about potential misuse in deepfakes and unauthorized voice cloning. As AI audio technology progresses, balancing its potential benefits with responsible use becomes increasingly crucial.

Magnit Introduces AI Assistant Maggi to Revolutionize Workforce Management

Magnit, a global staffing and talent provider, has launched an AI-powered Integrated Workforce Management Platform featuring an assistant named Maggi. This innovative system aims to streamline the hiring process for temporary and seasonal jobs by offering a unified interface for data access, sourcing, and workforce management tools. Powered by OpenAI’s GPT-4 and connected to extensive datasets, Maggi assists with task execution, provides alerts on regulatory changes, and aids in workforce planning. The platform offers features such as automated workflows, data-driven insights, and compliance frameworks. Magnit’s goal is to simplify complex processes and improve efficiency in talent acquisition and management across various sectors.

Frequently asked questions

YouTube is integrating Google DeepMind’s Veo AI model into Shorts, enabling creators to generate high-quality backgrounds and six-second clips. The upgrade includes AI-generated standalone video clips, filler scenes for transitions, and watermarking using DeepMind’s SynthID technology. Additional features include “Jewels” for livestream interaction, automatic dubbing support, and AI assistance for brainstorming video ideas, generating thumbnails, and managing comments.
Runway has partnered with Lionsgate to develop a custom AI video production and editing model, marking the first major collaboration between an AI video provider and a major film studio. The model will be trained using Lionsgate’s film and TV library, focusing initially on storyboarding, background creation, and special effects for action scenes. This partnership aims to streamline production processes and reduce costs while maintaining creative quality.
EVI 2 (Empathic Voice Interface 2) is an advanced AI voice model offering improved naturalness, emotional responsiveness, and customizable voices without voice cloning. It features faster response times, enhanced emotional intelligence, and multi-language support. The system can be integrated directly into apps and adapts to user preferences in real-time, making it ideal for customer service and virtual assistant applications.
Apple Intelligence is expanding its language support in 2025 to include German, Italian, Korean, Portuguese, Vietnamese, and several English variants. The feature will first launch in American English with iOS 18.1 before rolling out to other languages. However, regulatory challenges are currently limiting its release in the European Union and China, where Apple is working with authorities to resolve compliance issues.
LinkedIn has faced criticism for using user data to train AI models without initially updating its terms of service. The platform now offers an opt-out toggle in certain regions for users who don’t want their data used in AI model training. Privacy advocates argue for an opt-in model instead, highlighting the broader issue of social media platforms using user-generated content for AI training without explicit consent.
EzAudio, developed by Tencent AI Lab and Johns Hopkins University, generates high-quality sound effects from text prompts using a unique approach that operates in the latent space of audio waveforms. It eliminates the need for spectrograms and neural vocoders, offering superior performance compared to existing open-source models. The technology shows promise for entertainment and accessibility applications while raising ethical considerations.
Maggi, Magnit’s AI-powered assistant, streamlines temporary and seasonal hiring processes through a unified interface for data access, sourcing, and workforce management. Powered by GPT-4, it offers automated workflows, regulatory compliance alerts, and data-driven insights. The platform simplifies complex hiring processes and improves efficiency in talent acquisition across various industries.
Picture of Gor Gasparyan

Gor Gasparyan

Optimizing digital experiences for growth-stage & enterprise brands through research-driven design, automation, and AI