Commit aacd990a authored by Your Name's avatar Your Name

Try to fix

parent fe7a30dc
...@@ -106,13 +106,8 @@ def main(): ...@@ -106,13 +106,8 @@ def main():
from codai.models.cache import download_model from codai.models.cache import download_model
# Use provided file pattern or default to .gguf
file_pattern = args.download_file_pattern if args.download_file_pattern else '.gguf'
if args.download_file_pattern:
print(f"File pattern: {file_pattern}")
try: try:
cached_path = download_model(args.download_model, file_pattern=file_pattern) cached_path = download_model(args.download_model)
if cached_path: if cached_path:
print(f"\n=== Model downloaded successfully ===") print(f"\n=== Model downloaded successfully ===")
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment