Create NotebookLM Videos With YOUR Face (No Filming - AI)

Create NotebookLM Videos With YOUR Face (No Filming - AI)

Creating Video Podcasts with Notebook LM

Introduction to Notebook LM

  • Notebook LM is highlighted as a favorite AI tool for creating audio podcasts based on user-provided information.
  • The limitation of Notebook LM is that it only generates audio podcasts, not video ones.
  • The speaker introduces the goal of the video: demonstrating how to create video podcasts using Notebook LM.

Researching Podcast Topics

  • The speaker navigates to notebookm.google and mentions having already created a podcast topic focused on "no coding tools versus vibe coding tools."
  • A comprehensive overview of sources (18 in total) related to no code vs. low code platforms is generated by the tool.

Generating Podcast Content

  • After gathering research, users can select different podcast formats such as deep dive, brief, critique, or debate; the speaker chooses a debate format.
  • The focus of the podcast is set on whether low code or no code is better for individuals with minimal development experience.
  • Once settings are configured, the content generation process begins by clicking 'generate.'

Previewing Generated Audio

  • A preview of the generated audio reveals two distinct speakers debating their perspectives on app development approaches.

Preparing Audio Tracks for Video Creation

  • The next step involves downloading the audio and splitting it into separate tracks for each speaker using a tool called Speaker Split.
  • It’s emphasized that while one speaker talks, there should be silence in the other track to facilitate video editing later.

Processing Audio Tracks

  • After processing through Speaker Split, two distinct tracks are created where each speaker's dialogue does not overlap.

Dubbing with 11 Labs

  • To personalize one character's voice in the podcast, 11 Labs will be used to clone the user's voice; this requires recording short clips for voice cloning.

Voice Cloning and Dubbing Process with 11 Labs

Introduction to Voice Cloning

  • The speaker introduces the voice cloning feature, noting that while it may not sound exactly like them, there is potential for improvement with more training.
  • The speaker mentions the emotionality and tonality of their voice are not fully captured but can be enhanced by using the V3 model, which was temporarily unavailable.

Dubbing Process Overview

  • After cloning the voice, the next step involves dubbing where a script is uploaded to replace an original voice with the cloned one.
  • The project is named "VC versus VB" and set to translate from English to English, focusing on replacing a female speaker's audio.

Creating a Dubbing Project

  • The speaker emphasizes clicking on "create dubbing project" to allow for edits after initial processing since first attempts may not be perfect.
  • They confirm having enough credits for the process and initiate uploading, indicating it might take some time.

AI Mastermind Promotion

  • A brief promotion of an AI mastermind program is presented, highlighting its resources such as courses on AI basics and automation templates.
  • The speaker encourages viewers interested in AI to check out this resource as it offers daily updates and tools.

Project Processing Details

  • Once uploaded, the project enters a translating phase where English audio will be altered without changing languages.
  • To change voices within the project, users must select options from a menu that allows customization of how their voice sounds in playback.

Finalizing Dubbed Audio

  • Users are instructed to generate audio after selecting their preferred settings; if errors occur during generation, they should wait for resolution.
  • Upon successful generation, users can listen to their dubbed track before exporting it in various formats like MP3.

Transitioning to Video Creation

  • After completing dubbing with 11 Labs, users are directed towards creating videos using Hen or similar platforms.
  • Emphasis is placed on creating an avatar within Hen by uploading video samples similar to what was done in 11 Labs.

Creating a Digital Avatar

Steps to Create Your Digital Twin

  • Begin by uploading a video of yourself, allowing 5-10 minutes for processing. This will generate avatars that resemble you.
  • After avatar creation, navigate to the projects section and select "new video" followed by "create in AI studio." Adjust settings such as orientation and size as needed.
  • Instead of typing a script, upload an audio file from 11 Labs. This avoids manual adjustments for pauses that may be required if using text-to-speech.

Generating Video Content

  • Upload the audio file into Hey Gen; it will take time to process but will allow you to hear your voice synced with the avatar.
  • Once uploaded, click on "generate" to create a video featuring your digital twin speaking with the provided audio. Expect processing times of about 5-10 minutes.

Creating Additional Avatars

  • While waiting for your first video to render, you can start creating another project for a different speaker (Speaker B).
  • Replace the avatar by selecting from public avatars, ensuring gender matching with Speaker B's characteristics. Choose an appropriate background setting.

Finalizing Audio and Video

  • Upload Speaker B's audio file without needing transcription since it matches their voice directly.
  • Generate this second video after naming it appropriately (e.g., "no code versus vibe coding") and submit it for rendering.

Editing Videos Together

  • After both videos are rendered, download them and open your preferred video editing software (e.g., Cap Cut).
  • Import both files into the editor; arrange them side-by-side or sequentially for comparison.

Reviewing Output Quality

  • Assess each video's quality; ensure they convincingly represent each speaker’s likeness and voice.
  • Basic editing techniques can enhance presentation—consider applying masks or adjusting layouts for clarity in viewer engagement.

Exploring No Coding vs. Vibe Coding

The Central Question of Trade-offs

  • The debate focuses on the differences between no coding and vibe coding, emphasizing a trade-off in production methods.
  • The speaker highlights the seamless transition between speakers without editing, showcasing an efficient workflow in video production.
  • Suggestions for improvement include selecting better avatars and adjusting zoom levels to enhance visual appeal during presentations.
  • Emphasis is placed on personalizing editing choices to align with individual preferences and project requirements.
  • The discussion underscores the importance of experimenting with editing tools to achieve desired outcomes in content creation.
Video description

🚀 Learn to integrate AI profitable within a business in my Applied AI Mastermind: https://www.skool.com/applied-ai-mastermind-9612 In this video I breakdown how you can create an AI video podcast using NotebookLM, HeyGen, 11 Labs, and Speaker Split. With this process you can create a wide range of content super quickly. Have a play around with this process yourself! Join the FREE Applied AI Club here: https://www.skool.com/applied-ai-club-5279