Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
vneralla
/
xlrs-53-finnish
like
0
Automatic Speech Recognition
Transformers
PyTorch
JAX
common_voice
multilingual
wav2vec2
pretraining
speech
Inference Endpoints
arxiv:
2006.13979
License:
apache-2.0
Model card
Files
Files and versions
Community
1
Train
Deploy
Use this model
main
xlrs-53-finnish
1 contributor
History:
7 commits
vneralla
eduskunta_max_10_min05_epoch_4
cdac12f
over 2 years ago
.gitattributes
Safe
1.18 kB
initial commit
almost 3 years ago
README.md
Safe
2.32 kB
First version
almost 3 years ago
config.json
Safe
1.77 kB
Change vocab size in config
almost 3 years ago
flax_model.msgpack
Safe
1.02 GB
LFS
First version
almost 3 years ago
preprocessor_config.json
Safe
212 Bytes
First version
almost 3 years ago
pytorch_model.bin
pickle
Detected Pickle imports (21)
"flash.audio.speech_recognition.output_transform.SpeechRecognitionOutputTransform"
,
"transformers.tokenization_utils.Trie"
,
"torch._utils._rebuild_tensor_v2"
,
"tokenizers.AddedToken"
,
"flash.audio.speech_recognition.output_transform.SpeechRecognitionBackboneState"
,
"flash.core.utilities.stages.RunningStage"
,
"transformers.models.wav2vec2.processing_wav2vec2.Wav2Vec2Processor"
,
"flash.core.data.io.input.ServeInput"
,
"flash.core.data.io.input_transform.InputTransformState"
,
"collections.OrderedDict"
,
"flash.core.data.io.output.Output"
,
"flash.audio.speech_recognition.collate.DataCollatorCTCWithPadding"
,
"torch.utils.data._utils.collate.default_collate"
,
"transformers.models.wav2vec2.tokenization_wav2vec2.Wav2Vec2CTCTokenizer"
,
"transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor"
,
"torch.FloatStorage"
,
"flash.core.data.data_pipeline.DataPipeline"
,
"flash.core.data.io.input_transform.InputTransform"
,
"flash.core.data.data_pipeline.DataPipelineState"
,
"flash.core.data.states.CollateFn"
,
"flash.audio.speech_recognition.input.SpeechRecognitionPathsInput"
How to fix it?
3.79 GB
LFS
eduskunta_max_10_min05_epoch_4
over 2 years ago
tokenizer.json
Safe
218 Bytes
Latest
over 2 years ago
vocab.json
Safe
360 Bytes
Latest
over 2 years ago