Clone of ANY voice and create a lip-synched animation with Tortoise-TTS and StableDiffusion
In this quick tutorial I will show you how to clone ANY voice and create a flicker-free, lip-synched animation of that person, using only free AI-tools (Tortoise-TTS, Stable Diffusion / Automatic1111/ SadTalker).
Some useful links:
------------------------------
Download Audacity:
https://www.audacityteam.org/download/
Google Colab link Tortoise-TTS:
https://colab.research.google.com/drive/1N8lXviiwfmrS9vYzuFdQnLHlddXMMMvU#scrollTo=vF1d-My9IXXt
Github-Link Tortoise:
https://github.com/neonbjb/tortoise-tts
Github Link SadTalker:
https://github.com/OpenTalker/SadTalker
Download SadTalker Checkpoint files:
https://drive.google.com/drive/folders/1Wd88VDoLhVzYsQ30_qDVluQr_Xm46yHT
Download ffmpeg:
https://ffmpeg.org/download.html
Download GFPGAN:
https://github.com/TencentARC/GFPGAN
Create a custom StableDiffusion model of yourself:
https://youtu.be/ksG_ZjEKuLA
Local installation guide for Automatic1111 on a Windows-PC:
https://stable-diffusion-art.com/install-windows/
and on a Mac with Apple Silicon:
https://stable-diffusion-art.com/install-mac/
#tortoise #audacity #automatic1111
Watch on YouTube ↗
(saves to browser)
Sign in to unlock AI tutor explanation · ⚡30
More on: Image Generation Basics
View skill →Related AI Lessons
⚡
⚡
⚡
⚡
I built a Stable Diffusion playground in 200 lines and zero API keys. Here's how.
Dev.to · Devanshu Biswas
What makes an AI image workflow useful for real commercial output?
Dev.to AI
How to Write Better AI Image Prompts for Midjourney (With Examples That Actually Work)
Medium · ChatGPT
Image to Video AI: The Complete Workflow Playbook That Actually Produces Results
Medium · AI
🎓
Tutor Explanation
DeepCamp AI