-
Notifications
You must be signed in to change notification settings - Fork 7
/
Copy pathcheck_log.py
executable file
·116 lines (94 loc) · 4.92 KB
/
check_log.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
import os
import json
DATASETS_TO_CHECK = {
'librispeech_test_clean' : ['wer'],
'librispeech_test_other' : ['wer'],
'common_voice_15_en_test' : ['wer'],
'peoples_speech_test' : ['wer'],
'gigaspeech_test' : ['wer'],
'tedlium3_test' : ['wer'],
'tedlium3_long_form_test' : ['wer'],
'earnings21_test' : ['wer'],
'earnings22_test' : ['wer'],
'aishell_asr_zh_test' : ['wer'],
'cn_college_listen_mcq_test' : ['llama3_70b_judge', 'gpt4o_judge'],
'slue_p2_sqa5_test' : ['llama3_70b_judge', 'gpt4o_judge'],
'dream_tts_mcq_test' : ['llama3_70b_judge', 'gpt4o_judge'],
'public_sg_speech_qa_test' : ['llama3_70b_judge', 'gpt4o_judge'],
'spoken_squad_test' : ['llama3_70b_judge', 'gpt4o_judge'],
'openhermes_audio_test' : ['llama3_70b_judge', 'gpt4o_judge'],
'alpaca_audio_test' : ['llama3_70b_judge', 'gpt4o_judge'],
'clotho_aqa_test' : ['llama3_70b_judge', 'gpt4o_judge'],
'wavcaps_qa_test' : ['llama3_70b_judge', 'gpt4o_judge'],
'audiocaps_qa_test' : ['llama3_70b_judge', 'gpt4o_judge'],
'wavcaps_test' : ['llama3_70b_judge', 'meteor', 'gpt4o_judge'],
'audiocaps_test' : ['llama3_70b_judge', 'meteor', 'gpt4o_judge'],
'iemocap_emotion_test' : ['llama3_70b_judge', 'gpt4o_judge'],
'meld_sentiment_test' : ['llama3_70b_judge', 'gpt4o_judge'],
'meld_emotion_test' : ['llama3_70b_judge', 'gpt4o_judge'],
'voxceleb_accent_test' : ['llama3_70b_judge', 'gpt4o_judge'],
'voxceleb_gender_test' : ['llama3_70b_judge', 'gpt4o_judge'],
'iemocap_gender_test' : ['llama3_70b_judge', 'gpt4o_judge'],
'covost2_en_id_test' : ['bleu'],
'covost2_en_zh_test' : ['bleu'],
'covost2_en_ta_test' : ['bleu'],
'covost2_id_en_test' : ['bleu'],
'covost2_zh_en_test' : ['bleu'],
'covost2_ta_en_test' : ['bleu'],
'muchomusic_test' : ['llama3_70b_judge', 'gpt4o_judge'],
'imda_part1_asr_test' : ['wer'],
'imda_part2_asr_test' : ['wer'],
'imda_part3_30s_asr_test' : ['wer'],
'imda_part4_30s_asr_test' : ['wer'],
'imda_part5_30s_asr_test' : ['wer'],
'imda_part6_30s_asr_test' : ['wer'],
'imda_part3_30s_sqa_human_test': ['llama3_70b_judge', 'gpt4o_judge'],
'imda_part4_30s_sqa_human_test': ['llama3_70b_judge', 'gpt4o_judge'],
'imda_part5_30s_sqa_human_test': ['llama3_70b_judge', 'gpt4o_judge'],
'imda_part6_30s_sqa_human_test': ['llama3_70b_judge', 'gpt4o_judge'],
'imda_part3_30s_ds_human_test' : ['llama3_70b_judge', 'gpt4o_judge'],
'imda_part4_30s_ds_human_test' : ['llama3_70b_judge', 'gpt4o_judge'],
'imda_part5_30s_ds_human_test' : ['llama3_70b_judge', 'gpt4o_judge'],
'imda_part6_30s_ds_human_test' : ['llama3_70b_judge', 'gpt4o_judge'],
'imda_ar_sentence' : ['llama3_70b_judge', 'gpt4o_judge'],
'imda_ar_dialogue' : ['llama3_70b_judge', 'gpt4o_judge'],
'imda_gr_sentence' : ['llama3_70b_judge', 'gpt4o_judge'],
'imda_gr_dialogue' : ['llama3_70b_judge', 'gpt4o_judge'],
'seame_dev_man' : ['wer'],
'seame_dev_sge' : ['wer'],
'mmau_mini' : ['llama3_70b_judge', 'string_match', 'gpt4o_judge'],
'gigaspeech2_thai' : ['wer'],
'gigaspeech2_indo' : ['wer'],
'gigaspeech2_viet' : ['wer'],
}
folder_to_check = 'log_for_all_models'
MODEL_SCORE_NAMES = []
for dataset_name, metric_names in DATASETS_TO_CHECK.items():
for metric_name in metric_names:
MODEL_SCORE_NAMES += [f"{dataset_name}_{metric_name}"]
MODEL_NAME_TO_CHECK = os.listdir(folder_to_check)
# sort by model names
MODEL_NAME_TO_CHECK.sort()
for MODEL_NAME in MODEL_NAME_TO_CHECK:
if MODEL_NAME == 'old_models' or MODEL_NAME == 'test_temp':
continue
print(f"Checking {MODEL_NAME}")
for model_score_name in MODEL_SCORE_NAMES:
# For ASR models, only exam ASR
if MODEL_NAME in ['whisper_large_v3',
'hy_whisper_local_cs',
]:
if 'wer' not in model_score_name:
continue
score_log_path = f"{folder_to_check}/{MODEL_NAME}/{model_score_name}_score.json"
if os.path.exists(score_log_path) == False:
print(f"Error: {score_log_path} not found.")
continue
try:
with open(score_log_path, 'r') as f:
json_score = json.load(f)
except:
print(f"Error: Failed to load json file {score_log_path}.")
continue
print("=====================================================")
# python check_log.py 2>&1 | tee check_log.log