Edit model card

speecht5-asr-punctuation-sensitive

This model is part of SotoMedia's Automatic Video Dubbing project, aiming to build first open source video dubbing technolgy across a diverse range of languages. You can find more details about our project and our pibline here.

Description:

The speecht5-asr-punctuation-sensitive model is an advanced Automatic Speech Recognition (ASR) system designed to transcribe spoken English while maintaining a high level of awareness for punctuation. This model is trained to accurately recognize and preserve punctuation marks, enhancing the fidelity of transcriptions in scenarios where punctuation is crucial for conveying meaning.

  • Model type: transformer encoder- decoder
  • Language: En
  • Base model: SpeechT5-ASR checkpoint
  • ** Finetuning dataset:** MuST-C-en_ar

Key Features:

Punctuation Sensitivity: The model is specifically engineered to be highly sensitive to punctuation nuances in spoken English, ensuring accurate representation of the speaker's intended meaning. New Vocabulary: Change vocabulary to be on Piece-level rather than character-level with vocabulary size 500 piece.

Usage

# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("automatic-speech-recognition", model="seba3y/speecht5-asr-punctuation-sensitive")
# Load model directly
from transformers import AutoProcessor, AutoModelForSpeechSeq2Seq

processor = AutoProcessor.from_pretrained("seba3y/speecht5-asr-punctuation-sensitive")
model = AutoModelForSpeechSeq2Seq.from_pretrained("seba3y/speecht5-asr-punctuation-sensitive")

Fintuning & Evaluation Details

Dataset

MuST-C is a multilingual speech translation corpus whose size and quality will facilitate the training of end-to-end systems for SLT from English into several target languages. For each target language, MuST-C comprises several hundred hours of audio recordings from English TED Talks, which are automatically aligned at the sentence level with their manual transcriptions and translations.

Datasplits:

  • set: dev

    talks 11
    sentences 1073
    words src 24274
    words tgt 21387
    time 2h28m34s
  • set: tst-COMMON

    talks 27
    sentences 2019
    words src 41955
    words tgt 36443
    time 4h04m39s
  • set: tst-HE

    talks 12
    sentences 578
    words src 13080
    words tgt 10912
    time 1h26m51s
  • set: train

    talks 2412
    sentences 212085
    words src 4520522
    words tgt 4000457
    time 463h15m44s

Hyperparameters

Paramter Value
per_device_train_batch_size 6
per_device_eval_batch_size 16
gradient_accumulation_steps 12
eval_accumulation_steps 16
dataloader_num_workers 2
learning_rate 5e-5
adafactor True
weight_decay 0.08989525
max_grad_norm 0.58585
num_train_epochs 5
warmup_ratio 0.7
lr_scheduler_type constant_with_warmup
fp16 True
gradient_checkpointing True
sortish_sampler True
Results

Train loss: 0.8925

Split Word Error Rate (%)
dev 44.8
tst-HE 39.1
tst-COMMON 43.2

Citation

  • MuST-C dataset
@InProceedings{mustc19, author = "Di Gangi, Mattia Antonino and Cattoni, Roldano and Bentivogli, Luisa and Negri, Matteo > and Turchi, Marco",
 title = "{MuST-C: a Multilingual Speech Translation Corpus}",
booktitle = "Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies,
 Volume 2 (Short Papers)", year = "2019", address = "Minneapolis, MN, USA", month = "June"}}
  • SpeechT5-ASR
@inproceedings{ao-etal-2022-speecht5,
    title = {{S}peech{T}5: Unified-Modal Encoder-Decoder Pre-Training for Spoken Language Processing},
    author = {Ao, Junyi and Wang, Rui and Zhou, Long and Wang, Chengyi and Ren, Shuo and Wu, Yu and Liu, Shujie and Ko, Tom and Li, Qing and Zhang, Yu and Wei, Zhihua and Qian, Yao and Li, Jinyu and Wei, Furu},
    booktitle = {Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)},
    month = {May},
    year = {2022},
    pages={5723--5738},
}
Downloads last month
6
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.