Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Contribute to GitLab
Sign in
Toggle navigation
V
vidai
Project
Project
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
SexHackMe
vidai
Commits
0b69edd6
Commit
0b69edd6
authored
Oct 09, 2025
by
Stefy Lanza (nextime / spora )
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
Make model.generate() result unpacking more robust to handle different return types
parent
72d01acd
Changes
1
Show whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
18 additions
and
3 deletions
+18
-3
worker_analysis.py
vidai/worker_analysis.py
+18
-3
No files found.
vidai/worker_analysis.py
View file @
0b69edd6
...
...
@@ -152,7 +152,12 @@ def analyze_single_image(image_path, prompt, model):
}
]
result
,
tokens_used
=
model
.
generate
({
"messages"
:
messages
},
max_new_tokens
=
128
)
gen_result
=
model
.
generate
({
"messages"
:
messages
},
max_new_tokens
=
128
)
if
isinstance
(
gen_result
,
tuple
):
result
,
tokens_used
=
gen_result
else
:
result
=
gen_result
tokens_used
=
0
return
result
# For now, estimate tokens (could be improved with actual token counting)
estimated_tokens
=
len
(
result
.
split
())
+
len
(
prompt
.
split
())
...
...
@@ -270,10 +275,20 @@ def analyze_media(media_path, prompt, model_path, interval=10, job_id=None, comm
# Use vision model for summary
summary_prompt
=
f
"Summarize the video based on frame descriptions: {' '.join(descriptions)}"
messages
=
[{
"role"
:
"user"
,
"content"
:
[{
"type"
:
"text"
,
"text"
:
summary_prompt
}]}]
summary
,
summary_tokens
=
model
.
generate
({
"messages"
:
messages
},
max_new_tokens
=
256
)
gen_result
=
model
.
generate
({
"messages"
:
messages
},
max_new_tokens
=
256
)
if
isinstance
(
gen_result
,
tuple
):
summary
,
summary_tokens
=
gen_result
else
:
summary
=
gen_result
summary_tokens
=
0
else
:
# Use text-only model for summary
summary
,
summary_tokens
=
model
.
generate
(
f
"Summarize the video based on frame descriptions: {' '.join(descriptions)}"
,
max_new_tokens
=
256
)
gen_result
=
model
.
generate
(
f
"Summarize the video based on frame descriptions: {' '.join(descriptions)}"
,
max_new_tokens
=
256
)
if
isinstance
(
gen_result
,
tuple
):
summary
,
summary_tokens
=
gen_result
else
:
summary
=
gen_result
summary_tokens
=
0
total_tokens
+=
summary_tokens
print
(
f
"DEBUG: Summary generated for job {job_id}"
)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment