Register
Login
Resources
Docs Blog Datasets Glossary Case Studies Tutorials & Webinars
Product
Data Engine LLMs Platform Enterprise
Pricing Explore
Connect to our Discord channel
Alisamar Husain 6e047a9ae1
Enable using `m4t_evaluate` with a Manifest JSON file (#395)
1 month ago
..
71886b2e43
Seamless November release. (#221)
5 months ago
71886b2e43
Seamless November release. (#221)
5 months ago
6e047a9ae1
Enable using `m4t_evaluate` with a Manifest JSON file (#395)
1 month ago

README.md

You have to be logged in to leave a comment. Sign In

Evaluating SeamlessM4T models

Refer to the SeamlessM4T README for an overview of the M4T models.

Refer to the inference README for how to run inference with SeamlessM4T models.

Quick start:

We use SACREBLEU library for computing BLEU scores and JiWER library is used to compute these CER and WER scores.

Evaluation can be run with the CLI, from the root directory of the repository.

The model can be specified with --model_name: seamlessM4T_v2_large or seamlessM4T_large or seamlessM4T_medium

m4t_evaluate --data_file <path_to_data_tsv_file> --task <task_name> --tgt_lang <tgt_lang> --output_path <path_to_save_evaluation_output> --ref_field <ref_field_name> --audio_root_dir <path_to_audio_root_directory>

Note

  1. We use raw (unnormalized) references to compute BLEU scores for S2TT, T2TT tasks.
  2. For ASR task, src_lang needs to be passed as <tgt_lang>
  3. --src_lang arg needs to be specified to run evaluation for T2TT task
Tip!

Press p or to see the previous file or, n or to see the next file

Comments

Loading...