Generative AI for the Blender VSE: Text, video or image to video, image and audio in Blender Video Sequence Editor.
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

295 lines
11 KiB

# PALLAIDIUM - Generative AI for the Blender VSE
AI-generate video, image, and audio from text prompts or video, image, or text strips.
2 years ago
![PallAIdium](https://github.com/tin2tin/Generative_AI/assets/1322593/1b1b232f-00d9-4b0b-86fb-5f0f24136d2c)
2 years ago
2 years ago
## Features
| | |
|----------------------------------------------------|-----------------------------------------------------|
| Text to video | Text to audio |
| Text to speech | Text to image |
| Image to image | Image to video |
| Video to video | Image to text |
| ControlNet | OpenPose |
| Canny | Illusion |
| Multiple LoRAs | Segmind distilled SDXL |
| Seed | Quality steps |
| Frames | Word power |
| Style selector | Strip power |
| Batch conversion | Batch refinement of images. |
| Batch upscale & refinement of movies. | Model card selector. |
| Render-to-path selector. | Render finished notification. |
| Model Cards | One-click install and uninstall dependencies. |
| User-defined file path for generated files. | Define the location for storing generated files. |
| Seed and prompt added to strip name. | Include seed and prompt information in the strip name. |
![image](https://github.com/tin2tin/Pallaidium/assets/1322593/09bba394-3188-49be-8ae1-37c0c43ec2fd)
2 years ago
2 years ago
## Requirements
* Windows (Could maybe work on Linux or MacOS, but someone will have to contribute code to make it work).
* A CUDA-supported Nvidia card with at least 4 GB VRAM.
2 years ago
## How to install
2 years ago
* First you must download and install git for your platform(must be on PATH(or Bark will fail)): https://git-scm.com/downloads
2 years ago
* Download the add-on: https://github.com/tin2tin/text_to_video/archive/refs/heads/main.zip
2 years ago
* On Windows, right-click on the Blender icon and "Run Blender as Administrator"(or you'll get write permission errors).
2 years ago
* Install the add-on as usual: Preferences > Add-ons > Install > select file > enable the add-on.
2 years ago
* In the Generative AI add-on preferences, hit the "Install Dependencies" button.
* Note that you can change what model cards are used in the various modes here(video, image, and audio).
* Then it writes that it is finished(if any vital errors, let me know).
* Restart Blender.
* Open the add-on UI in the Sequencer > Sidebar > Generative AI.
* The first time any model is executed many GB will have to be downloaded, so go grab lots of coffee.
2 years ago
* If it says: "ModuleNotFoundError: Refer to https://github.com/facebookresearch/xformers for more information on how to install xformers", then try to restart Blender.
Tip |
:------------- |
If any Python modules are missing, use this add-on to manually install them: |
https://github.com/amb/blender_pip |
2 years ago
2 years ago
2 years ago
## Location
Install Dependencies, set Movie Model Card, and set Sound Notification in the add-on preferences:
![image](https://github.com/tin2tin/Generative_AI/assets/1322593/49ba0182-f8a0-4a1d-b24f-caca9741d033)
Video Sequence Editor > Sidebar > Generative AI:
1 year ago
![image](https://github.com/tin2tin/Pallaidium/assets/1322593/e3c1193d-5e0a-4ed2-acca-3f7a4413e4c1)
2 years ago
Styles:
![image](https://github.com/tin2tin/Generative_AI/assets/1322593/86807264-a377-4de1-875e-471aaa3011a7)
See SDXL handling most of the styles here: https://stable-diffusion-art.com/sdxl-styles/
## Updates
Read about the updates here:
https://github.com/tin2tin/Pallaidium/discussions/categories/announcements
## Text to Video
The Animov models have been trained on Anime material, so adding "anime" to the prompt is necessary, especially for the Animov-512x model.
## Text to Image
The Stable Diffusion models for generating images have been used a lot, so there are plenty of prompt suggestions out there if you google for them.
### Artists
https://stablediffusion.fr/artists
### Prompting:
https://github.com/invoke-ai/InvokeAI/blob/main/docs/features/PROMPTS.md
https://stablediffusion.fr/prompts
2 years ago
https://blog.segmind.com/generating-photographic-images-with-stable-diffusion/
2 years ago
Tip |
:------------- |
If the image of your renders breaks, then use the resolution from the Model Card in the Preferences. |
Tip |
:------------- |
If the image of your playback stutters, then select a strip > Menu > Strip > Movie Strip > Set Render Size. |
Tip |
:------------- |
If you get the message that CUDA is out of memory, then restart Blender to free up memory and make it stable again. |
2 years ago
# Batch Processing
Select multiple strips and hit Generate. When doing this, the file name, and if found the seed value, are automatically inserted into the prompt and seed value. However, in the add-on preferences, this behavior can be switched off.
https://github.com/tin2tin/Pallaidium/assets/1322593/28098eb6-3a93-4bcb-bd6f-53b71faabd8d
2 years ago
# Text to Audio
### Bark
Find Bark documentation here: https://github.com/suno-ai/bark
2 years ago
* [laughter]
* [laughs]
* [sighs]
* [music]
* [gasps]
* [clears throat]
* — or ... for hesitations
* ♪ for song lyrics
* capitalization for emphasis on a word
* MAN/WOMAN: for bias towards the speaker
2 years ago
2 years ago
Speaker Library: https://suno-ai.notion.site/8b8e8749ed514b0cbf3f699013548683?v=bc67cff786b04b50b3ceb756fd05f68c
2 years ago
Tip |
:------------- |
If the audio breaks up, try processing longer sentences. |
2 years ago
### AudioLDM2
Find AudioLDM documentation here: https://github.com/haoheliu/AudioLDM
Try prompts like: Bag pipes playing a funeral dirge, punk rock band playing hardcore song, techno dj playing deep bass house music, and acid house loop with jazz.
Or: Voice of God judging mankind, woman talking about celestial beings, hammer on wood.
## Performance
The performance can be improved by following this guide: https://nvidia.custhelp.com/app/answers/detail/a_id/5490/~/system-memory-fallback-for-stable-diffusion
On 6 GB of VRAM I typically render images first with 1024x512 images with SDXL. And img2vid with Zeroscope XL in 768x384x10 or 640x320x17, with a low Quality Steps value.
## New to Blender?
Watch this tutorial: https://youtu.be/4_MIaxzjh5Y?feature=shared
## AI Modules
2 years ago
Diffusers: https://github.com/huggingface/diffusers
2 years ago
ModelScope: https://modelscope.cn/models/damo/text-to-video-synthesis/summary
2 years ago
Animov: https://huggingface.co/vdo/animov-0.1.1
2 years ago
Potat1: https://huggingface.co/camenduru/potat1
Zeroscope Dark: https://huggingface.co/cerspense/zeroscope_v2_dark_30x448x256
Zeroscope XL: https://huggingface.co/cerspense/zeroscope_v2_XL
2 years ago
Würstchen: https://github.com/dome272/Wuerstchen
AudioLDM2 Music: [https://huggingface.co/cvssp/audioldm-s-full-v2 https://github.com/haoheliu/AudioLDM](https://huggingface.co/cvssp/audioldm2-music)
2 years ago
2 years ago
Bark: https://github.com/suno-ai/bark
Deep Floyd IF: https://github.com/deep-floyd/IF
Stable Diffusion XL: https://huggingface.co/stabilityai/stable-diffusion-xl-base-0.9
Segmind SDXL: https://huggingface.co/segmind/SSD-1B
https://blog.segmind.com/generating-photographic-images-with-stable-diffusion/
2 years ago
2 years ago
## Uninstall
Hugging Face Diffusers models are downloaded from the hub and saved to a local cache directory. By default, the cache directory is located at:
On Linux and macOS: ~/.cache/huggingface/hub
On Windows: %userprofile%\\.cache\huggingface\hub
Here you can locate and delete the individual models.
## Useful add-ons
### Add Rendered Strips
Since the Generative AI add-on only can input image or movie strips, you'll need to convert other strip types to movie-strip. For this purpose, this add-on can be used:
https://github.com/tin2tin/Add_Rendered_Strips
![image](https://github.com/tin2tin/Generative_AI/assets/1322593/d8c0a184-d812-440d-a5a8-501a1282d78d)
### VSE Masking Tools
For creating a mask on top of a clip in the Sequencer, this add-on can be used to input the clip as background in the Blender Image Editor. The created mask can then be added to the VSE as a strip, and converted to video with the above add-on:
https://github.com/tin2tin/vse_masking_tools
![image](https://github.com/tin2tin/Generative_AI/assets/1322593/f2afd36c-34b1-4779-957b-0eb8defed296)
## Useful Projects
Trainer for LoRAs: https://github.com/johnman3032/simple-lora-dreambooth-trainer
## Video Examples
### Image to Text
https://github.com/tin2tin/Pallaidium/assets/1322593/91eb17e4-72d6-4c69-8e5c-a3d38af5a770
### Illusion Diffusion
https://github.com/tin2tin/Pallaidium/assets/1322593/42eadfd8-3ebf-4747-b8e0-7b79fe8626b6
### Scribble
https://github.com/tin2tin/Pallaidium/assets/1322593/c74a4e38-8b16-423b-be78-aadfbfe284dc
### Styles
https://github.com/tin2tin/Pallaidium/assets/1322593/b80812b4-e3be-40b0-a73b-bc55b7eeadf7
### ControlNet
https://github.com/tin2tin/Pallaidium/assets/1322593/a1e94e09-0147-40ae-b4c2-4ce0671b1289
### OpenPose
https://github.com/tin2tin/Pallaidium/assets/1322593/ac9f278e-9fc9-46fc-a4e7-562ff041964f
### Zeroscope
[![Watch the video](https://img.youtube.com/vi/LejSJGmtEvE/0.jpg)](https://youtu.be/LejSJGmtEvE)
### Würstchen
[![Watch the video](https://img.youtube.com/vi/CDPmGs_JtSM/0.jpg)](https://youtu.be/CDPmGs_JtSM)
### Bark
[![Watch the video](https://img.youtube.com/vi/AAdQfQjENJU/0.jpg)](https://youtu.be/AAdQfQjENJU)
### Batch from Text Strips
[![Watch the video](https://img.youtube.com/vi/gSFWGkgaNsE/0.jpg)](https://youtu.be/gSFWGkgaNsE)
### Video to video:
https://github.com/tin2tin/Generative_AI/assets/1322593/c044a0b0-95c2-4b54-af0b-45bc0c670c89
https://github.com/tin2tin/Generative_AI/assets/1322593/0105cd35-b3b2-49cf-91c1-0633dd484177
### Img2img:
https://github.com/tin2tin/Generative_AI/assets/1322593/2dd2d2f1-a1f6-4562-8116-ffce872b79c3
### Painting
https://github.com/tin2tin/Generative_AI/assets/1322593/7cd69cd0-5842-40f0-b41f-455c77443535
2 years ago
## Restrictions for using the AI models:
2 years ago
2 years ago
- The models can only be used for non-commercial purposes. The models are meant for research purposes.
2 years ago
- The models was not trained to realistically represent people or events, so using it to generate such content is beyond the model's capabilities.
2 years ago
- It is prohibited to generate content that is demeaning or harmful to people or their environment, culture, religion, etc.
- Prohibited for pornographic, violent, and bloody content generation.
2 years ago
- Prohibited for error and false information generation.
2 years ago
2 years ago