Model Card for Model ID
Model Details
Model Description
This is a quantized model of the original version mohammed/whisper-small-arabic-cv-11
- Developed by: Mohammed Bakheet
- Funded by [optional]: Kalam Technology
- Language(s) (NLP): Arabic, English
Uses
This a quantized model that reads arabic voice and transcribes/translate it into english
Direct Use
First, install the following packages using the following commands:
pip install -U optimum[exporters,onnxruntime] transformers pip install huggingface_hub
# uncomment the following installation if you are using a notebook:
#!pip install -U optimum[exporters,onnxruntime] transformers
#!pip install huggingface_hub
# import the required packages
from optimum.onnxruntime import ORTModelForSpeechSeq2Seq
from transformers import WhisperTokenizerFast, WhisperFeatureExtractor, pipeline
# set model name/id
model_name = 'mohammed/quantized-whisper-small' # folder name
model = ORTModelForSpeechSeq2Seq.from_pretrained(model_name, export=False)
tokenizer = WhisperTokenizerFast.from_pretrained(model_name)
feature_extractor = WhisperFeatureExtractor.from_pretrained(model_name)
forced_decoder_ids = tokenizer.get_decoder_prompt_ids(language="ar", task="transcribe")
pipe = pipeline('automatic-speech-recognition',
model=model,
tokenizer=tokenizer,
feature_extractor=feature_extractor,
model_kwargs={"forced_decoder_ids": forced_decoder_ids})
# the file to be transcribed
pipe('Recording.mp3')
Out-of-Scope Use
The model does a direct translation of Arabic speech, and doesn't do a direct transcription, we are still working on that.
Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
How to Get Started with the Model
Use the code below to get started with the model.
First, install the following packages using the following commands:
pip install -U optimum[exporters,onnxruntime] transformers
pip install huggingface_hub
from optimum.onnxruntime import ORTModelForSpeechSeq2Seq
from transformers import WhisperTokenizerFast, WhisperFeatureExtractor, pipeline
model_name = 'mohammed/quantized-whisper-small' # folder name
model = ORTModelForSpeechSeq2Seq.from_pretrained(model_name, export=False)
tokenizer = WhisperTokenizerFast.from_pretrained(model_name)
feature_extractor = WhisperFeatureExtractor.from_pretrained(model_name)
forced_decoder_ids = tokenizer.get_decoder_prompt_ids(language="ar", task="transcribe")
pipe = pipeline('automatic-speech-recognition',
model=model,
tokenizer=tokenizer,
feature_extractor=feature_extractor,
model_kwargs={"forced_decoder_ids": forced_decoder_ids})
# the file to be transcribed
pipe('Recording.mp3')
Training Data
Please refer to the original model at "mohammed/whisper-small-arabic-cv-11"
Training Procedure
Please refer to the original model at "mohammed/whisper-small-arabic-cv-11"
Preprocessing [optional]
Please refer to the original model at "mohammed/whisper-small-arabic-cv-11"
Training Hyperparameters
- Training regime: Please refer to the original model at "mohammed/whisper-small-arabic-cv-11"
- Downloads last month
- 2