- 27 Feb, 2026 19 commits
-
-
Stefy Lanza (nextime / spora ) authored
-
Stefy Lanza (nextime / spora ) authored
- Add model ID check for I2V detection (not just tags) - Add -Diffusers suffix when extracting base model from tags - Add runtime validation for Wan LoRA base model against tags
-
Stefy Lanza (nextime / spora ) authored
-
Stefy Lanza (nextime / spora ) authored
For Wan LoRA adapters, skip reading base_model from stored config (which may have incorrect values like T2V instead of I2V). Instead, always infer from LoRA ID by checking for 'i2v' in the name. This ensures wan2_2_i2v_general_nsfw_lora uses the correct Wan-AI/Wan2.2-I2V-A14B-Diffusers base model instead of the incorrect Wan-AI/Wan2.2-T2V-A14B-Diffusers.
-
Stefy Lanza (nextime / spora ) authored
-
Stefy Lanza (nextime / spora ) authored
The check at line 9414 only checked the stored supports_i2v flag from the model configuration, but didn't check the model ID string for 'i2v' like the detect_model_type() function does. Now the I2V validation also detects I2V capability from the model ID, making it consistent with detect_model_type() and properly detecting I2V capability for LoRA adapters like lopi999/Wan2.2-I2V_General-NSFW-LoRA.
-
Stefy Lanza (nextime / spora ) authored
-
Stefy Lanza (nextime / spora ) authored
- Always infer base model from LoRA name, not stored database value - Fix supports_i2v detection for LoRAs based on LoRA name - Ensures Wan 2.2 I2V LoRAs use correct I2V base model
-
Stefy Lanza (nextime / spora ) authored
-
Stefy Lanza (nextime / spora ) authored
-
Stefy Lanza (nextime / spora ) authored
-
Stefy Lanza (nextime / spora ) authored
- hpcai-tech/Open-Sora-1.2 -> hpcai-tech/Open-Sora-2 - genmo/mochi -> genmo/mochi-1-preview
-
Stefy Lanza (nextime / spora ) authored
When selecting a Wan 2.1 I2V model that is a LoRA or tensor weight, the base model now uses Wan-AI/Wan2.1-I2V-14B-720P-Diffusers instead of the generic Wan-AI/Wan2.1-I2V-14B-Diffusers.
-
Stefy Lanza (nextime / spora ) authored
Fixed ValueError: too many values to unpack (expected 5) - Line 3929: Added missing 6th element to unpacking - Line 3936: Added missing 6th element to unpacking The results tuple has 6 elements: (name, info, caps, is_disabled, fail_count, orig_idx)
-
Stefy Lanza (nextime / spora ) authored
When selecting a Wan 2.1 I2V model that is a LoRA or tensor weight, the base model now uses Wan-AI/Wan2.1-I2V-14B-720P-Diffusers instead of the generic Wan-AI/Wan2.1-I2V-14B-Diffusers.
-
Stefy Lanza (nextime / spora ) authored
Fixed ValueError: too many values to unpack (expected 5) - Line 3929: Added missing 6th element to unpacking - Line 3936: Added missing 6th element to unpacking The results tuple has 6 elements: (name, info, caps, is_disabled, fail_count, orig_idx)
-
Stefy Lanza (nextime / spora ) authored
Fix I2V pipeline auto-detection See merge request !1
-
Stefy Lanza (nextime / spora ) authored
- Add detect_model_family() to identify model family (wan, sdxl, sd, ltx, etc.) - Add get_pipeline_for_model_family() for proper pipeline selection based on family + task - Enhance detect_generation_type() to check --image FIRST for I2V detection - Add support for --image_model, --prompt_image, --prompt_animation as I2V indicators - Add support for audio/subtitle options as T2V+V2V chaining indicators This fixes the issue where SDXL models were incorrectly using WanPipeline for I2V tasks, causing type mismatch errors (expected UMT5EncoderModel, got CLIPTextModel). Now SDXL models correctly use DiffusionPipeline or StableDiffusionXLPipeline.
-
Stefy Lanza (nextime / spora ) authored
-
- 26 Feb, 2026 21 commits
-
-
Stefy Lanza (nextime / spora ) authored
-
Stefy Lanza (nextime / spora ) authored
- Add --model-list-batch to EXAMPLES.md for batch model listing - Add new CLI options to MCP server (--output-dir, --yes, --audio-chunk) - Add new CLI options to webapp build_command function - Update README.md with Output Options section
-
Stefy Lanza (nextime / spora ) authored
Added --audio-chunk argument with 3 modes: - overlap (default): overlapping chunks like [0-60], [58-118] - word-boundary: uses Whisper timestamps to split at word boundaries - vad: uses Voice Activity Detection to skip silence Also added --audio-chunk-overlap to control overlap duration. New functions added: - process_video_with_vad(): VAD-based chunking - process_video_word_boundary(): Word-boundary chunking using Whisper Modified: - transcribe_video_audio(): accepts audio_chunk_type and audio_chunk_overlap params - _transcribe_chunked(): accepts chunk_type and overlap params
-
Stefy Lanza (nextime / spora ) authored
Models like Faber8/AbyssOrangeMix2 are SD1.5 models, not Flux. Now detected as StableDiffusionPipeline instead of FluxPipeline.
-
Stefy Lanza (nextime / spora ) authored
When a model has a VAE configured but the VAE files don't exist in the repository, try loading with the default VAE instead.
-
Stefy Lanza (nextime / spora ) authored
Allow using full HuggingFace model ID (e.g., Faber8/AbyssOrangeMix2_nsfw) as --model argument by looking up both short name and full ID in MODELS.
-
Stefy Lanza (nextime / spora ) authored
The repos attribute from scan_cache_dir() is a frozenset, not a list, so .pop() doesn't work. Fixed by using next(iter()) instead.
-
Stefy Lanza (nextime / spora ) authored
The message is now suppressed when using --model-list-batch to make the output cleaner for scripts.
-
Stefy Lanza (nextime / spora ) authored
Added --yes / -y argument that automatically answers 'yes' to confirmation prompts when deleting cached models or clearing the entire cache. Usage: videogen --remove-cached-model MODEL_ID --yes videogen --clear-cache --yes
-
Stefy Lanza (nextime / spora ) authored
The --model-list-batch option was added but wasn't being handled properly. Now it correctly exits after printing the batch output, and is also added to the list of options that don't require --prompt.
-
Stefy Lanza (nextime / spora ) authored
- Fixed model ID consistency: numeric IDs now remain the same when using filters like --nsfw-friendly, --t2v-only, --i2v-only, etc. Previously, filtered lists would renumber models making --show-model by numeric ID unreliable. - Added --model-list-batch option for script-friendly output: Outputs 'NUMERIC_ID:FULL_MODEL_NAME' format for easy parsing - Added --output-dir option to specify output directory: Sets the directory where output files will be saved - Fixed syntax error in argparse epilog string that was causing 'SyntaxError: invalid decimal literal'
-
Stefy Lanza (nextime / spora ) authored
-
Stefy Lanza (nextime / spora ) authored
- Fixed model ID normalization to handle hyphens (in addition to underscores) - Fixed dictionary key ordering in base_model_fallbacks so more specific keys (wan2.2.i2v) are checked before generic keys (wan2.2) - Fixed Wan 2.1 I2V base model mapping (was incorrectly pointing to T2V) - Fixed base model detection in earlier code sections to check model ID directly instead of relying on m_info.get('supports_i2v') - Fixed typo: Wan 2.2 generic fallback now correctly uses Wan2.2-T2V Now Wan 2.2 I2V models like Wan-AI/Wan2.2-I2V-A14B will correctly use Wan-AI/Wan2.2-I2V-14B-Diffusers as the base model instead of the incorrect Wan-AI/Wan2.2-T2V-14B-Diffusers. -
Stefy Lanza (nextime / spora ) authored
- Fixed mapping table to use correct I2V base model (Wan-AI/Wan2.2-I2V-14B-Diffusers) - Fixed Diffuser -> Diffusers typo in model IDs - Updated all Wan 2.2 I2V references to use correct model ID
-
Stefy Lanza (nextime / spora ) authored
-
Stefy Lanza (nextime / spora ) authored
- Add correct base model for Wan 2.2 I2V: Wan-AI/Wan2.2-I2V-A14B-Diffuser - Add specific VRAM estimate for Wan 2.2 I2V MoE models (~14GB) - Apply more conservative VRAM calculation for models with weights/LoRAs - Fix indentation error in add_model_from_hf function
-
Stefy Lanza (nextime / spora ) authored
Calculate: base_vram + 2GB + 50% This ensures a 14B model estimated at 18GB will require ~29GB instead of 22.5GB.
-
Stefy Lanza (nextime / spora ) authored
Instead of adding a fixed 2GB overhead, now calculates 25% more VRAM for base models that will have fine-tuned weights/tensors or LoRA adapters loaded on top.
-
Stefy Lanza (nextime / spora ) authored
The user confirmed that Wan2.2-I2V models should use Wan2.2-T2V as the base model, not the I2V variant.
-
Stefy Lanza (nextime / spora ) authored
The issue was that model IDs from HuggingFace use dots (wan2.2-i2v-a14b) while user config names use underscores (wan2_2_i2v_a14b). Now we normalize the model ID by replacing underscores with dots before matching against the base_model_fallbacks dictionary.
-
Stefy Lanza (nextime / spora ) authored
The issue was that model IDs like 'wan2_2_i2v_a14b' would match 'wan2_2' (T2V) before 'wan2_2_i2v' (I2V) because 'wan2_2' comes first in the dictionary. Now the dictionary is ordered with more specific keys first: - wan2_2_i2v_a14b (most specific) - wan2.2_i2v_a14b - wan2_2_i2v - wan2.2_i2v - wan2_2 - wan2.2 etc. This ensures longer/more specific keys are checked before shorter ones.
-