pavanBuduguppa commited on
Commit
18911cd
1 Parent(s): 30f48aa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +54 -0
README.md CHANGED
@@ -1,3 +1,57 @@
1
  ---
2
  license: gpl-3.0
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: gpl-3.0
3
+ language:
4
+ - en
5
+ pipeline_tag: text2text-generation
6
+ tags:
7
+ - code
8
+ - asr
9
+ - inverse text normalization
10
+ datasets:
11
+ - pavanBuduguppa/asr_inverse_text_normalization
12
+
13
  ---
14
+
15
+ ---
16
+ ---
17
+
18
+ # asr_inverse_text_normalization
19
+
20
+ Finetuned a facebook/bart-base Pretrained model on the ASR inverse text normalization dataset by treating it as a seq2seq task. Other approaches which may be considered is by considering it as a TokenClassification task and the one mentioned here https://machinelearning.apple.com/research/inverse-text-normal.
21
+
22
+
23
+ ## Model description
24
+
25
+ BART (Bidirectional and Auto-Regressive Transformers) is a pre-trained transformer-based neural network model developed by Facebook AI Research (FAIR) for various natural language processing (NLP) tasks
26
+
27
+ The BART architecture is based on the Transformer model, which is a type of neural network architecture that processes sequential input data, such as text, by applying self-attention mechanisms to capture the relationships between different words in the input sequence.
28
+ BART includes both auto-regressive and bidirectional encoder-decoder transformer architectures, which enable it to perform both generation and prediction tasks
29
+
30
+ BART was trained on a diverse range of NLP tasks, including machine translation, summarization, and question answering, and has shown strong performance across multiple benchmarks.
31
+ Its training process involves corrupting text with different types of noise and training the model to reconstruct the original text, which has been shown to improve the model's ability to generalize to new tasks and outperform other pre-trained language models like GPT and BERT
32
+
33
+ The model flavour which was chosen is that of "facebook/bart-base" and columns "after" is used as the source while "before" column is used as the targets.
34
+
35
+ ## Intended uses & limitations
36
+
37
+ This model can be used as an out-of-the-box solution to the invesrse text normalization which can convert ASR generated un-normalized text such as
38
+ "my c v v for my card is five six seven and it expires on november twenty three" -> "my CVV for my card is 567 and it expires on November 23"
39
+
40
+ The model needs to be explored for various min and max length setting at the time of generation for your specific usecase
41
+
42
+ ### How to use
43
+
44
+
45
+ ```python
46
+
47
+ >>> from transformers import pipeline
48
+ >>> generator = pipeline(model="pavanBuduguppa/asr_inverse_text_normalization")
49
+
50
+ >>> generator("my c v v for my card is five six seven and it expires on november twenty three")
51
+
52
+ ```
53
+
54
+
55
+ ## Training data
56
+
57
+ All credits and rights for the training data belongs to Google. The data was merely obtained and processed for this model and the original data can be found here https://www.kaggle.com/competitions/text-normalization-challenge-english-language/data