diff --git a/vbench/aesthetic_quality.py b/vbench/aesthetic_quality.py index 263381e0..df41abc1 100755 --- a/vbench/aesthetic_quality.py +++ b/vbench/aesthetic_quality.py @@ -74,7 +74,8 @@ def laion_aesthetic(aesthetic_model, clip_model, video_list, device): num += 1 video_results.append({'video_path': video_path, 'video_results': cur_avg.item()}) - aesthetic_avg /= num + if num != 0: + aesthetic_avg /= num return aesthetic_avg, video_results diff --git a/vbench/appearance_style.py b/vbench/appearance_style.py index 2c596487..ff8c0838 100755 --- a/vbench/appearance_style.py +++ b/vbench/appearance_style.py @@ -70,7 +70,7 @@ def appearance_style(clip_model, video_dict, device, sample="rand"): 'video_results': video_sim, 'frame_results': cur_video, 'cur_sim': cur_sim}) - sim_per_frame = sim / cnt + sim_per_frame = sim / cnt if cnt != 0 else None return sim_per_frame, video_results def compute_appearance_style(json_dir, device, submodules_list, **kwargs): diff --git a/vbench/background_consistency.py b/vbench/background_consistency.py index d7e4d390..65c20198 100755 --- a/vbench/background_consistency.py +++ b/vbench/background_consistency.py @@ -61,7 +61,7 @@ def background_consistency(clip_model, preprocess, video_list, device, read_fram 'video_sim': video_sim, 'cnt_per_video': cnt_per_video}) # sim_per_video = sim / (len(video_list) - 1) - sim_per_frame = sim / cnt + sim_per_frame = sim / cnt if cnt != 0 else None return sim_per_frame, video_results diff --git a/vbench/color.py b/vbench/color.py index 3d9ef0c9..8aa7a620 100755 --- a/vbench/color.py +++ b/vbench/color.py @@ -86,7 +86,7 @@ def color(model, video_dict, device): 'video_path': video_path, 'video_results': cur_success_frame_rate, 'cur_success_frame_rate': cur_success_frame_rate,}) - success_rate = success_frame_count_all / video_count + success_rate = success_frame_count_all / video_count if video_count != 0 else None return success_rate, video_results diff --git a/vbench/human_action.py b/vbench/human_action.py index 7d6468f7..c4db37af 100755 --- a/vbench/human_action.py +++ b/vbench/human_action.py @@ -106,7 +106,7 @@ def human_action(umt_path, video_list, device): 'video_results': flag, 'cor_num_per_video': cor_num_per_video,}) # print(f"cor num: {cor_num}, total: {cnt}") - acc = cor_num / cnt + acc = cor_num / cnt if cnt != 0 else None return acc, video_results diff --git a/vbench/imaging_quality.py b/vbench/imaging_quality.py index 518061ea..7ccfa912 100755 --- a/vbench/imaging_quality.py +++ b/vbench/imaging_quality.py @@ -51,7 +51,7 @@ def technical_quality(model, video_list, device, **kwargs): score = model(frame) acc_score_video += float(score) video_results.append({'video_path': video_path, 'video_results': acc_score_video/len(images)}) - average_score = sum([o['video_results'] for o in video_results]) / len(video_results) + average_score = sum([o['video_results'] for o in video_results]) / len(video_results) if video_results else 0 average_score = average_score / 100. return average_score, video_results diff --git a/vbench/multiple_objects.py b/vbench/multiple_objects.py index 98337a4f..7735066a 100755 --- a/vbench/multiple_objects.py +++ b/vbench/multiple_objects.py @@ -69,7 +69,7 @@ def multiple_objects(model, video_dict, device): 'video_results': cur_success_frame_rate, 'success_frame_count': cur_success_frame_count, 'frame_count': len(cur_video_pred)}) - success_rate = success_frame_count / frame_count + success_rate = success_frame_count / frame_count if frame_count != 0 else None return success_rate, video_results diff --git a/vbench/object_class.py b/vbench/object_class.py index c876bc71..11be82bf 100755 --- a/vbench/object_class.py +++ b/vbench/object_class.py @@ -65,7 +65,7 @@ def object_class(model, video_dict, device): 'video_results': cur_success_frame_rate, 'success_frame_count': cur_success_frame_count, 'frame_count': len(cur_video_pred)}) - success_rate = success_frame_count / frame_count + success_rate = success_frame_count / frame_count if frame_count != 0 else None return success_rate, video_results diff --git a/vbench/scene.py b/vbench/scene.py index 88b28a0d..3a5333f1 100755 --- a/vbench/scene.py +++ b/vbench/scene.py @@ -59,7 +59,7 @@ def scene(model, video_dict, device): 'video_results': cur_success_frame_rate, 'success_frame_count': cur_success_frame_count, 'frame_count': len(cur_video_pred)}) - success_rate = success_frame_count / frame_count + success_rate = success_frame_count / frame_count if frame_count != 0 else None return success_rate, video_results diff --git a/vbench/subject_consistency.py b/vbench/subject_consistency.py index 3e7361f8..fa23cb33 100755 --- a/vbench/subject_consistency.py +++ b/vbench/subject_consistency.py @@ -65,7 +65,7 @@ def subject_consistency(model, video_list, device, read_frame): sim += video_sim video_results.append({'video_path': video_path, 'video_results': sim_per_images}) # sim_per_video = sim / (len(video_list) - 1) - sim_per_frame = sim / cnt + sim_per_frame = sim / cnt if cnt != 0 else None return sim_per_frame, video_results