Add Lumina pipeline support (LuminaText2ImgPipeline, Lumina2Text2ImgPipeline)

- Add Lumina pipeline classes to PIPELINE_CLASS_MAP
- Fix Lumina model detection in detect_pipeline_class
- Remove non-existent LuminaVideoPipeline references
- Update Alpha-VLLM/Lumina-Next-SFT to use correct pipeline class
- Lumina-Next-SFT is a T2I model, not T2V (~20-30GB VRAM)
parent 74900219
...@@ -582,6 +582,10 @@ PIPELINE_CLASS_MAP = { ...@@ -582,6 +582,10 @@ PIPELINE_CLASS_MAP = {
"WanPipeline": {"type": "video", "default_vram": "~10-24 GB"}, "WanPipeline": {"type": "video", "default_vram": "~10-24 GB"},
"I2VGenXLPipeline": {"type": "i2v", "default_vram": "~18-24 GB"}, "I2VGenXLPipeline": {"type": "i2v", "default_vram": "~18-24 GB"},
"LTXPipeline": {"type": "i2v", "default_vram": "~12-16 GB"}, "LTXPipeline": {"type": "i2v", "default_vram": "~12-16 GB"},
"LuminaPipeline": {"type": "image", "default_vram": "~20-30 GB"},
"LuminaText2ImgPipeline": {"type": "image", "default_vram": "~20-30 GB"},
"Lumina2Pipeline": {"type": "image", "default_vram": "~20-30 GB"},
"Lumina2Text2ImgPipeline": {"type": "image", "default_vram": "~20-30 GB"},
"AnimateDiffPipeline": {"type": "video", "default_vram": "~10-14 GB"}, "AnimateDiffPipeline": {"type": "video", "default_vram": "~10-14 GB"},
"TextToVideoSDPipeline": {"type": "t2v", "default_vram": "~7-9 GB"}, "TextToVideoSDPipeline": {"type": "t2v", "default_vram": "~7-9 GB"},
"TextToVideoZeroPipeline": {"type": "t2v", "default_vram": "~6-8 GB"}, "TextToVideoZeroPipeline": {"type": "t2v", "default_vram": "~6-8 GB"},
...@@ -592,7 +596,6 @@ PIPELINE_CLASS_MAP = { ...@@ -592,7 +596,6 @@ PIPELINE_CLASS_MAP = {
"AllegroPipeline": {"type": "t2v", "default_vram": "~35-45 GB"}, "AllegroPipeline": {"type": "t2v", "default_vram": "~35-45 GB"},
"HunyuanDiTPipeline": {"type": "t2v", "default_vram": "~40-55 GB"}, "HunyuanDiTPipeline": {"type": "t2v", "default_vram": "~40-55 GB"},
"OpenSoraPipeline": {"type": "video", "default_vram": "~45-65 GB"}, "OpenSoraPipeline": {"type": "video", "default_vram": "~45-65 GB"},
"LuminaVideoPipeline": {"type": "t2v", "default_vram": "~100 GB"},
"StepVideoPipeline": {"type": "t2v", "default_vram": "~90-140 GB"}, "StepVideoPipeline": {"type": "t2v", "default_vram": "~90-140 GB"},
"CogVideoXPipeline": {"type": "t2v", "default_vram": "~20-30 GB"}, "CogVideoXPipeline": {"type": "t2v", "default_vram": "~20-30 GB"},
"HotshotXLPipeline": {"type": "video", "default_vram": "~8-12 GB"}, "HotshotXLPipeline": {"type": "video", "default_vram": "~8-12 GB"},
...@@ -961,11 +964,17 @@ def detect_pipeline_class(model_info): ...@@ -961,11 +964,17 @@ def detect_pipeline_class(model_info):
if "modelscope" in model_id or "text-to-video-ms" in model_id: if "modelscope" in model_id or "text-to-video-ms" in model_id:
return "TextToVideoSDPipeline" return "TextToVideoSDPipeline"
if "lumina" in model_id: if "lumina" in model_id:
return "LuminaVideoPipeline" if "lumina2" in model_id or "lumina-2" in model_id:
return "Lumina2Text2ImgPipeline"
return "LuminaText2ImgPipeline"
if "stepvideo" in model_id or "step-video" in model_id: if "stepvideo" in model_id or "step-video" in model_id:
return "StepVideoPipeline" return "StepVideoPipeline"
# Image models # Image models
if "lumina" in model_id:
if "lumina2" in model_id or "lumina-2" in model_id:
return "Lumina2Text2ImgPipeline"
return "LuminaText2ImgPipeline"
if "flux" in model_id: if "flux" in model_id:
return "FluxPipeline" return "FluxPipeline"
if "pony" in model_id or "animagine" in model_id: if "pony" in model_id or "animagine" in model_id:
...@@ -1607,7 +1616,7 @@ def update_all_models(hf_token=None): ...@@ -1607,7 +1616,7 @@ def update_all_models(hf_token=None):
# ═══════════════════════════════════════════════════════════════ # ═══════════════════════════════════════════════════════════════
# 100GB+ models - Ultra High VRAM # 100GB+ models - Ultra High VRAM
# ═══════════════════════════════════════════════════════════════ # ═══════════════════════════════════════════════════════════════
("Alpha-VLLM/Lumina-Next-SFT", "LuminaVideoPipeline", "~100 GB", "Lumina Next SFT - Ultra high quality T2I/T2V"), ("Alpha-VLLM/Lumina-Next-SFT", "LuminaText2ImgPipeline", "~20-30 GB", "Lumina Next SFT - High quality T2I"),
# ═══════════════════════════════════════════════════════════════ # ═══════════════════════════════════════════════════════════════
# 90-140GB models - Extreme VRAM # 90-140GB models - Extreme VRAM
...@@ -1795,7 +1804,9 @@ def update_all_models(hf_token=None): ...@@ -1795,7 +1804,9 @@ def update_all_models(hf_token=None):
"TextToVideoSDPipeline", "TextToVideoZeroPipeline", "TextToVideoSDPipeline", "TextToVideoZeroPipeline",
"HotshotXLPipeline", "AllegroPipeline", "HotshotXLPipeline", "AllegroPipeline",
"HunyuanDiTPipeline", "OpenSoraPipeline", "HunyuanDiTPipeline", "OpenSoraPipeline",
"LuminaVideoPipeline", "StepVideoPipeline", "LuminaPipeline", "LuminaText2ImgPipeline",
"Lumina2Pipeline", "Lumina2Text2ImgPipeline",
"StepVideoPipeline",
"DiffusionPipeline", "FluxPipeline", "DiffusionPipeline", "FluxPipeline",
"StableDiffusionXLPipeline", "StableDiffusion3Pipeline"] "StableDiffusionXLPipeline", "StableDiffusion3Pipeline"]
...@@ -1898,7 +1909,7 @@ def update_all_models(hf_token=None): ...@@ -1898,7 +1909,7 @@ def update_all_models(hf_token=None):
"TextToVideoSDPipeline", "TextToVideoZeroPipeline", "TextToVideoSDPipeline", "TextToVideoZeroPipeline",
"HotshotXLPipeline", "AllegroPipeline", "HotshotXLPipeline", "AllegroPipeline",
"HunyuanDiTPipeline", "OpenSoraPipeline", "HunyuanDiTPipeline", "OpenSoraPipeline",
"LuminaVideoPipeline", "StepVideoPipeline", "StepVideoPipeline",
"DiffusionPipeline", "FluxPipeline", "DiffusionPipeline", "FluxPipeline",
"StableDiffusionXLPipeline", "StableDiffusion3Pipeline"] "StableDiffusionXLPipeline", "StableDiffusion3Pipeline"]
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment