Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Contribute to GitLab
Sign in
Toggle navigation
videogen
Project
Project
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
nexlab
videogen
Commits
33ec35a2
Commit
33ec35a2
authored
Feb 24, 2026
by
Stefy Lanza (nextime / spora )
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
Fix pipeline fallback error handling - nest error messages inside failure check
parent
efa4dfd3
Pipeline
#232
canceled with stages
Changes
1
Pipelines
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
54 additions
and
54 deletions
+54
-54
videogen
videogen
+54
-54
No files found.
videogen
View file @
33ec35a2
...
...
@@ -3651,60 +3651,60 @@ def main(args):
print(f"\n❌ All model retries exhausted ({retry_count}/{max_retries} attempts)")
# Print detailed error message for the user
if "404" in error_str or "Entry Not Found" in error_str:
print(f"❌ Model not found on HuggingFace: {model_id_to_load}")
print(f" This model may have been removed or the ID is incorrect.")
if debug:
print(f"\n [DEBUG] Troubleshooting:")
print(f" - Check if the model exists: https://huggingface.co/{model_id_to_load}")
print(f" - Verify the model ID spelling")
print(f" - The model may have been renamed or moved")
print(f"\n 💡 Try searching for an alternative:")
print(f" videogen --search-models ltxvideo")
print(f"\n 💡 Or use the official LTX Video model:")
print(f" videogen --model ltx_video --prompt '
your
prompt
' ...")
elif "401" in error_str or "Unauthorized" in error_str:
print(f"❌ Model requires authentication: {model_id_to_load}")
print(f" Set your HuggingFace token:")
print(f" export HF_TOKEN=your_token_here")
print(f" huggingface-cli login")
if debug:
print(f"\n [DEBUG] To get a token:")
print(f" 1. Go to https://huggingface.co/settings/tokens")
print(f" 2. Create a new token with '
read
' permissions")
print(f" 3. Export it: export HF_TOKEN=hf_xxx")
elif "gated" in error_str.lower():
print(f"❌ This is a gated model: {model_id_to_load}")
print(f" You need to accept the license on HuggingFace:")
print(f" https://huggingface.co/{model_id_to_load}")
print(f" Then set HF_TOKEN and run again.")
elif "connection" in error_str.lower() or "timeout" in error_str.lower():
print(f"❌ Network error loading model: {model_id_to_load}")
print(f" Check your internet connection and try again.")
if debug:
print(f"\n [DEBUG] Network troubleshooting:")
print(f" - Check if you can access: https://huggingface.co/{model_id_to_load}")
print(f" - Try with a VPN if HuggingFace is blocked")
print(f" - Check if HF_ENDPOINT is set (for China mirror): {os.environ.get('
HF_ENDPOINT
', '
not
set
')}")
elif "FrozenDict" in error_str or "scale_factor" in error_str or "has no attribute" in error_str:
print(f"❌ Pipeline compatibility error: {model_id_to_load}")
print(f" This model uses a pipeline architecture incompatible with your diffusers version.")
print(f" The model may require a specific diffusers version or different pipeline class.")
if debug:
print(f"\n [DEBUG] Compatibility troubleshooting:")
print(f" - Try updating diffusers: pip install --upgrade git+https://github.com/huggingface/diffusers.git")
print(f" - Check the model'
s
documentation
for
required
versions
")
print(f"
-
The
model
may
be
incorrectly
configured
in
models
.
json
")
print(f"
\
n
💡
Try
a
different
model
with
--
model
<
name
>
")
else:
print(f"
Model
loading
failed
:
{
e
}
")
if debug:
import traceback
print(f"
\
n
[
DEBUG
]
Full
traceback
:
")
traceback.print_exc()
print(f"
\
n
💡
Try
searching
for
alternative
models
:
videogen
--
search
-
models
<
query
>
")
sys.exit(1)
if "404" in error_str or "Entry Not Found" in error_str:
print(f"❌ Model not found on HuggingFace: {model_id_to_load}")
print(f" This model may have been removed or the ID is incorrect.")
if debug:
print(f"\n [DEBUG] Troubleshooting:")
print(f" - Check if the model exists: https://huggingface.co/{model_id_to_load}")
print(f" - Verify the model ID spelling")
print(f" - The model may have been renamed or moved")
print(f"\n 💡 Try searching for an alternative:")
print(f" videogen --search-models ltxvideo")
print(f"\n 💡 Or use the official LTX Video model:")
print(f" videogen --model ltx_video --prompt '
your
prompt
' ...")
elif "401" in error_str or "Unauthorized" in error_str:
print(f"❌ Model requires authentication: {model_id_to_load}")
print(f" Set your HuggingFace token:")
print(f" export HF_TOKEN=your_token_here")
print(f" huggingface-cli login")
if debug:
print(f"\n [DEBUG] To get a token:")
print(f" 1. Go to https://huggingface.co/settings/tokens")
print(f" 2. Create a new token with '
read
' permissions")
print(f" 3. Export it: export HF_TOKEN=hf_xxx")
elif "gated" in error_str.lower():
print(f"❌ This is a gated model: {model_id_to_load}")
print(f" You need to accept the license on HuggingFace:")
print(f" https://huggingface.co/{model_id_to_load}")
print(f" Then set HF_TOKEN and run again.")
elif "connection" in error_str.lower() or "timeout" in error_str.lower():
print(f"❌ Network error loading model: {model_id_to_load}")
print(f" Check your internet connection and try again.")
if debug:
print(f"\n [DEBUG] Network troubleshooting:")
print(f" - Check if you can access: https://huggingface.co/{model_id_to_load}")
print(f" - Try with a VPN if HuggingFace is blocked")
print(f" - Check if HF_ENDPOINT is set (for China mirror): {os.environ.get('
HF_ENDPOINT
', '
not
set
')}")
elif "FrozenDict" in error_str or "scale_factor" in error_str or "has no attribute" in error_str:
print(f"❌ Pipeline compatibility error: {model_id_to_load}")
print(f" This model uses a pipeline architecture incompatible with your diffusers version.")
print(f" The model may require a specific diffusers version or different pipeline class.")
if debug:
print(f"\n [DEBUG] Compatibility troubleshooting:")
print(f" - Try updating diffusers: pip install --upgrade git+https://github.com/huggingface/diffusers.git")
print(f" - Check the model'
s
documentation
for
required
versions
")
print(f"
-
The
model
may
be
incorrectly
configured
in
models
.
json
")
print(f"
\
n
💡
Try
a
different
model
with
--
model
<
name
>
")
else:
print(f"
Model
loading
failed
:
{
e
}
")
if debug:
import traceback
print(f"
\
n
[
DEBUG
]
Full
traceback
:
")
traceback.print_exc()
print(f"
\
n
💡
Try
searching
for
alternative
models
:
videogen
--
search
-
models
<
query
>
")
sys.exit(1)
timing.end_step() # model_loading
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment