Compare commits

...

1 Commits

Author SHA1 Message Date
tin2tin d2d14f908e
Add: vid2img 1 year ago
  1. 65
      __init__.py

65
__init__.py

@ -1,12 +1,12 @@
# https://modelscope.cn/models/damo/text-to-video-synthesis/summary # https://modelscope.cn/models/damo/text-to-video-synthesis/summary
bl_info = { bl_info = {
"name": "Generative AI", "name": "Pallaidium - Generative AI",
"author": "tintwotin", "author": "tintwotin",
"version": (1, 4), "version": (1, 5),
"blender": (3, 4, 0), "blender": (3, 4, 0),
"location": "Video Sequence Editor > Sidebar > Generative AI", "location": "Video Sequence Editor > Sidebar > Generative AI",
"description": "Generate media in the VSE", "description": "AI Generate media in the VSE",
"category": "Sequencer", "category": "Sequencer",
} }
@ -185,16 +185,16 @@ def style_prompt(prompt):
def closest_divisible_64(num): def closest_divisible_64(num):
# Determine the remainder when num is divided by 64 # Determine the remainder when num is divided by 64
remainder = (num % 64) remainder = (num % 32)
# If the remainder is less than or equal to 32, return num - remainder, # If the remainder is less than or equal to 32, return num - remainder,
# but ensure the result is not less than 64 # but ensure the result is not less than 64
if remainder <= 32: if remainder <= 16:
result = num - remainder result = num - remainder
return max(result, 192) return max(result, 192)
# Otherwise, return num + (64 - remainder) # Otherwise, return num + (64 - remainder)
else: else:
return max(num + (64 - remainder), 192) return max(num + (32 - remainder), 192)
def find_first_empty_channel(start_frame, end_frame): def find_first_empty_channel(start_frame, end_frame):
@ -607,7 +607,7 @@ def install_modules(self):
] ]
) )
# Modelscope img2vid # # Modelscope img2vid
# import_module(self, "modelscope", "modelscope==1.8.4") # import_module(self, "modelscope", "modelscope==1.8.4")
# #import_module(self, "xformers", "xformers==0.0.20") # #import_module(self, "xformers", "xformers==0.0.20")
# #import_module(self, "torch", "torch==2.0.1") # #import_module(self, "torch", "torch==2.0.1")
@ -679,7 +679,7 @@ def input_strips_updated(self, context):
scene = context.scene scene = context.scene
input = scene.input_strips input = scene.input_strips
if movie_model_card == "stabilityai/stable-diffusion-xl-base-1.0": if movie_model_card == "stabilityai/stable-diffusion-xl-base-1.0" and scene.generatorai_typeselect == "video":
scene.input_strips = "input_strips" scene.input_strips = "input_strips"
if scene.generatorai_typeselect == "video" or scene.generatorai_typeselect == "audio": if scene.generatorai_typeselect == "video" or scene.generatorai_typeselect == "audio":
scene.inpaint_selected_strip = "" scene.inpaint_selected_strip = ""
@ -1230,10 +1230,14 @@ class SEQUENCER_OT_generate_movie(Operator):
# from modelscope.pipelines import pipeline # from modelscope.pipelines import pipeline
# from modelscope.outputs import OutputKeys # from modelscope.outputs import OutputKeys
# from modelscope import snapshot_download
# model_dir = snapshot_download('damo/Image-to-Video', revision='v1.1.0')
# pipe = pipeline(task='image-to-video', model= model_dir, model_revision='v1.1.0')
# #pipe = pipeline(task='image-to-video', model='damo-vilab/MS-Image2Video', model_revision='v1.1.0') # #pipe = pipeline(task='image-to-video', model='damo-vilab/MS-Image2Video', model_revision='v1.1.0')
# #pipe = pipeline(task='image-to-video', model='damo/Image-to-Video', model_revision='v1.1.0') # #pipe = pipeline(task='image-to-video', model='damo/Image-to-Video', model_revision='v1.1.0')
# pipe = pipeline(task='image-to-video', model='C:/Users/45239/.cache/modelscope/hub/damo/Image-to-Video', model_revision='v1.1.0') #
# # local: pipe = pipeline(task='image-to-video', model='C:/Users/45239/.cache/modelscope/hub/damo/Image-to-Video', model_revision='v1.1.0')
# if low_vram: # if low_vram:
# pipe.enable_model_cpu_offload() # pipe.enable_model_cpu_offload()
@ -1262,8 +1266,8 @@ class SEQUENCER_OT_generate_movie(Operator):
if low_vram: if low_vram:
#torch.cuda.set_per_process_memory_fraction(0.98) #torch.cuda.set_per_process_memory_fraction(0.98)
upscale.enable_model_cpu_offload() upscale.enable_model_cpu_offload()
upscale.enable_vae_tiling() #upscale.enable_vae_tiling()
upscale.unet.enable_forward_chunking(chunk_size=1, dim=1) # heavy: #upscale.unet.enable_forward_chunking(chunk_size=1, dim=1) # heavy:
upscale.enable_vae_slicing() upscale.enable_vae_slicing()
else: else:
upscale.to("cuda") upscale.to("cuda")
@ -1301,7 +1305,7 @@ class SEQUENCER_OT_generate_movie(Operator):
if low_vram: if low_vram:
upscale.enable_model_cpu_offload() upscale.enable_model_cpu_offload()
# upscale.unet.enable_forward_chunking(chunk_size=1, dim=1) #Heavy upscale.unet.enable_forward_chunking(chunk_size=1, dim=1) #Heavy
upscale.enable_vae_slicing() upscale.enable_vae_slicing()
else: else:
upscale.to("cuda") upscale.to("cuda")
@ -1400,6 +1404,7 @@ class SEQUENCER_OT_generate_movie(Operator):
elif scene.image_path: elif scene.image_path:
print("Process: Image to video") print("Process: Image to video")
video = process_image(scene.image_path, int(scene.generate_movie_frames)) video = process_image(scene.image_path, int(scene.generate_movie_frames))
video = np.array(video)
# Upscale video # Upscale video
if scene.video_to_video: if scene.video_to_video:
@ -1407,6 +1412,11 @@ class SEQUENCER_OT_generate_movie(Operator):
Image.fromarray(frame).resize((closest_divisible_64(int(x * 2)), closest_divisible_64(int(y * 2)))) Image.fromarray(frame).resize((closest_divisible_64(int(x * 2)), closest_divisible_64(int(y * 2))))
for frame in video for frame in video
] ]
else:
video = [
Image.fromarray(frame).resize((closest_divisible_64(int(x)), closest_divisible_64(int(y))))
for frame in video
]
video_frames = upscale( video_frames = upscale(
prompt, prompt,
@ -1830,7 +1840,7 @@ class SEQUENCER_OT_generate_image(Operator):
addon_prefs = preferences.addons[__name__].preferences addon_prefs = preferences.addons[__name__].preferences
image_model_card = addon_prefs.image_model_card image_model_card = addon_prefs.image_model_card
do_inpaint = input == "input_strips" and scene.inpaint_selected_strip and type == "image" do_inpaint = input == "input_strips" and scene.inpaint_selected_strip and type == "image"
do_refine = (scene.refine_sd or scene.image_path or image_model_card == "stabilityai/stable-diffusion-xl-base-1.0") #and not do_inpaint do_refine = (scene.refine_sd or scene.image_path or scene.movie_path) # or image_model_card == "stabilityai/stable-diffusion-xl-base-1.0") #and not do_inpaint
# LOADING MODELS # LOADING MODELS
print("Model: " + image_model_card) print("Model: " + image_model_card)
@ -1880,13 +1890,13 @@ class SEQUENCER_OT_generate_image(Operator):
# Models for stable diffusion # Models for stable diffusion
elif not image_model_card == "DeepFloyd/IF-I-M-v1.0": elif not image_model_card == "DeepFloyd/IF-I-M-v1.0" and not scene.image_path and not scene.movie_path:
from diffusers import AutoencoderKL from diffusers import AutoencoderKL
if image_model_card == "stabilityai/stable-diffusion-xl-base-1.0": if image_model_card == "stabilityai/stable-diffusion-xl-base-1.0":
vae = AutoencoderKL.from_pretrained("madebyollin/sdxl-vae-fp16-fix", torch_dtype=torch.float16) #vae = AutoencoderKL.from_pretrained("madebyollin/sdxl-vae-fp16-fix", torch_dtype=torch.float16)
pipe = DiffusionPipeline.from_pretrained( pipe = DiffusionPipeline.from_pretrained(
image_model_card, image_model_card,
vae=vae, #vae=vae,
torch_dtype=torch.float16, torch_dtype=torch.float16,
variant="fp16", variant="fp16",
) )
@ -1898,10 +1908,13 @@ class SEQUENCER_OT_generate_image(Operator):
) )
pipe.scheduler = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) pipe.scheduler = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config)
pipe.watermark = NoWatermark()
if low_vram: if low_vram:
#torch.cuda.set_per_process_memory_fraction(0.95) # 6 GB VRAM #torch.cuda.set_per_process_memory_fraction(0.95) # 6 GB VRAM
pipe.enable_model_cpu_offload() pipe.enable_model_cpu_offload()
pipe.enable_vae_slicing() pipe.enable_vae_slicing()
#pipe.enable_forward_chunking(chunk_size=1, dim=1)
else: else:
pipe.to("cuda") pipe.to("cuda")
@ -1978,7 +1991,7 @@ class SEQUENCER_OT_generate_image(Operator):
if low_vram: if low_vram:
refiner.enable_model_cpu_offload() refiner.enable_model_cpu_offload()
refiner.enable_vae_tiling() #refiner.enable_vae_tiling()
refiner.enable_vae_slicing() refiner.enable_vae_slicing()
else: else:
refiner.to("cuda") refiner.to("cuda")
@ -2109,9 +2122,18 @@ class SEQUENCER_OT_generate_image(Operator):
image = PIL.Image.fromarray(unmasked_unchanged_image_arr.astype("uint8")) image = PIL.Image.fromarray(unmasked_unchanged_image_arr.astype("uint8"))
# Img2img # Img2img
elif scene.image_path: elif scene.image_path or scene.movie_path:
print("Process: Image to image") if scene.movie_path:
init_image = load_image(scene.image_path).convert("RGB") print("Process: Video to image")
init_image = load_first_frame(scene.movie_path)
init_image = init_image.resize((x, y))
elif scene.image_path:
print("Process: Image to image")
init_image = load_first_frame(scene.image_path)
init_image = init_image.resize((x, y))
#init_image = load_image(scene.image_path).convert("RGB")
image = refiner( image = refiner(
prompt=prompt, prompt=prompt,
image=init_image, image=init_image,
@ -2235,6 +2257,9 @@ class SEQUENCER_OT_strip_to_generatorAI(Operator):
self.report({"INFO"}, "None of the selected strips are movie, image, or text types.") self.report({"INFO"}, "None of the selected strips are movie, image, or text types.")
return {"CANCELLED"} return {"CANCELLED"}
if use_strip_data:
print("Use file seed and prompt: Yes")
for count, strip in enumerate(strips): for count, strip in enumerate(strips):
if strip.type == "TEXT": if strip.type == "TEXT":
if strip.text: if strip.text:

Loading…
Cancel
Save