|
|
@ -3,7 +3,7 @@ |
|
|
|
bl_info = { |
|
|
|
bl_info = { |
|
|
|
"name": "Generative AI", |
|
|
|
"name": "Generative AI", |
|
|
|
"author": "tintwotin", |
|
|
|
"author": "tintwotin", |
|
|
|
"version": (1, 2), |
|
|
|
"version": (1, 3), |
|
|
|
"blender": (3, 4, 0), |
|
|
|
"blender": (3, 4, 0), |
|
|
|
"location": "Video Sequence Editor > Sidebar > Generative AI", |
|
|
|
"location": "Video Sequence Editor > Sidebar > Generative AI", |
|
|
|
"description": "Generate media in the VSE", |
|
|
|
"description": "Generate media in the VSE", |
|
|
@ -986,9 +986,12 @@ class SEQUENCER_OT_generate_image(Operator): |
|
|
|
result = login(token = addon_prefs.hugginface_token) |
|
|
|
result = login(token = addon_prefs.hugginface_token) |
|
|
|
print("Login: " + str(result)) |
|
|
|
print("Login: " + str(result)) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
torch.cuda.set_per_process_memory_fraction(0.90) |
|
|
|
|
|
|
|
|
|
|
|
# stage 1 |
|
|
|
# stage 1 |
|
|
|
stage_1 = DiffusionPipeline.from_pretrained("DeepFloyd/IF-I-M-v1.0", variant="fp16", torch_dtype=torch.float16) |
|
|
|
stage_1 = DiffusionPipeline.from_pretrained("DeepFloyd/IF-I-M-v1.0", variant="fp16", torch_dtype=torch.float16) |
|
|
|
stage_1.enable_model_cpu_offload() |
|
|
|
# stage_1.enable_model_cpu_offload() |
|
|
|
|
|
|
|
stage_1.enable_sequential_cpu_offload() # 6 GB VRAM |
|
|
|
|
|
|
|
|
|
|
|
# stage 2 |
|
|
|
# stage 2 |
|
|
|
stage_2 = DiffusionPipeline.from_pretrained( |
|
|
|
stage_2 = DiffusionPipeline.from_pretrained( |
|
|
@ -1006,6 +1009,7 @@ class SEQUENCER_OT_generate_image(Operator): |
|
|
|
"stabilityai/stable-diffusion-x4-upscaler", **safety_modules, torch_dtype=torch.float16 |
|
|
|
"stabilityai/stable-diffusion-x4-upscaler", **safety_modules, torch_dtype=torch.float16 |
|
|
|
) |
|
|
|
) |
|
|
|
stage_3.enable_model_cpu_offload() |
|
|
|
stage_3.enable_model_cpu_offload() |
|
|
|
|
|
|
|
|
|
|
|
else: # stable Diffusion |
|
|
|
else: # stable Diffusion |
|
|
|
pipe = DiffusionPipeline.from_pretrained( |
|
|
|
pipe = DiffusionPipeline.from_pretrained( |
|
|
|
image_model_card, |
|
|
|
image_model_card, |
|
|
@ -1059,7 +1063,7 @@ class SEQUENCER_OT_generate_image(Operator): |
|
|
|
generator = None |
|
|
|
generator = None |
|
|
|
|
|
|
|
|
|
|
|
if image_model_card == "DeepFloyd/IF-I-M-v1.0": |
|
|
|
if image_model_card == "DeepFloyd/IF-I-M-v1.0": |
|
|
|
prompt_embeds, negative_embeds = stage_1.encode_prompt(prompt) |
|
|
|
prompt_embeds, negative_embeds = stage_1.encode_prompt(prompt=prompt, negative_prompt=negative_prompt) |
|
|
|
|
|
|
|
|
|
|
|
# stage 1 |
|
|
|
# stage 1 |
|
|
|
image = stage_1( |
|
|
|
image = stage_1( |
|
|
@ -1108,7 +1112,7 @@ class SEQUENCER_OT_generate_image(Operator): |
|
|
|
fit_method="FIT", |
|
|
|
fit_method="FIT", |
|
|
|
) |
|
|
|
) |
|
|
|
strip.frame_final_duration = scene.generate_movie_frames |
|
|
|
strip.frame_final_duration = scene.generate_movie_frames |
|
|
|
strip.transform.filter = 'NEAREST' |
|
|
|
strip.transform.filter = 'SUBSAMPLING_3x3' |
|
|
|
|
|
|
|
|
|
|
|
scene.sequence_editor.active_strip = strip |
|
|
|
scene.sequence_editor.active_strip = strip |
|
|
|
if i > 0: |
|
|
|
if i > 0: |
|
|
@ -1120,6 +1124,10 @@ class SEQUENCER_OT_generate_image(Operator): |
|
|
|
else: |
|
|
|
else: |
|
|
|
print("No resulting file found.") |
|
|
|
print("No resulting file found.") |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
# clear the VRAM |
|
|
|
|
|
|
|
if torch.cuda.is_available(): |
|
|
|
|
|
|
|
torch.cuda.empty_cache() |
|
|
|
|
|
|
|
|
|
|
|
bpy.ops.renderreminder.play_notification() |
|
|
|
bpy.ops.renderreminder.play_notification() |
|
|
|
#wm.progress_end() |
|
|
|
#wm.progress_end() |
|
|
|
scene.frame_current = current_frame |
|
|
|
scene.frame_current = current_frame |
|
|
|