nroggendorff commited on
Commit
f158f4d
1 Parent(s): 565d6e2

Training in progress, epoch 1

Browse files
Files changed (3) hide show
  1. README.md +62 -35
  2. model.safetensors +1 -1
  3. training_args.bin +1 -1
README.md CHANGED
@@ -1,35 +1,62 @@
1
- ---
2
- license: mit
3
- base_model: distilbert/distilbert-base-uncased
4
- metrics:
5
- - accuracy
6
- model-index:
7
- - name: birthday-detector
8
- results: []
9
- datasets:
10
- - nroggendorff/doug
11
- language:
12
- - en
13
- pipeline_tag: text-classification
14
- ---
15
-
16
- # Birthday Detector
17
-
18
- This model is a fine-tuned version of [distilbert/distilbert-base-uncased](https://huggingface.co/distilbert/distilbert-base-uncased).
19
- It achieves the following results on the evaluation set:
20
-
21
- - Loss: 0.0101
22
- - Accuracy: 1.0
23
-
24
- You can easily run it with inference, or run it in python with:
25
-
26
- ```
27
- from transformers import pipeline
28
- classifier = pipeline("sentiment-analysis", model="nroggendorff/birthday-detector")
29
-
30
- isBirthday = classifier("happy birthday doug")[0]["label"] == "POSITIVE"
31
- ```
32
-
33
- Make sure you have the necessary dependencies installed.
34
-
35
- [This models inspiration](https://youtu.be/Q6fjwHPVqjQ)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: distilbert/distilbert-base-uncased
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - accuracy
8
+ model-index:
9
+ - name: birthday-detector
10
+ results: []
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ # birthday-detector
17
+
18
+ This model is a fine-tuned version of [distilbert/distilbert-base-uncased](https://huggingface.co/distilbert/distilbert-base-uncased) on an unknown dataset.
19
+ It achieves the following results on the evaluation set:
20
+ - Loss: 0.1095
21
+ - Accuracy: 0.9796
22
+
23
+ ## Model description
24
+
25
+ More information needed
26
+
27
+ ## Intended uses & limitations
28
+
29
+ More information needed
30
+
31
+ ## Training and evaluation data
32
+
33
+ More information needed
34
+
35
+ ## Training procedure
36
+
37
+ ### Training hyperparameters
38
+
39
+ The following hyperparameters were used during training:
40
+ - learning_rate: 2e-05
41
+ - train_batch_size: 16
42
+ - eval_batch_size: 16
43
+ - seed: 42
44
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
+ - lr_scheduler_type: linear
46
+ - num_epochs: 3
47
+
48
+ ### Training results
49
+
50
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy |
51
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|
52
+ | No log | 1.0 | 58 | 0.1248 | 0.9592 |
53
+ | No log | 2.0 | 116 | 0.1100 | 0.9592 |
54
+ | No log | 3.0 | 174 | 0.1095 | 0.9796 |
55
+
56
+
57
+ ### Framework versions
58
+
59
+ - Transformers 4.40.2
60
+ - Pytorch 2.3.0+cpu
61
+ - Datasets 2.19.1
62
+ - Tokenizers 0.19.1
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:52b4ebe798ecf1e84c0291ea574082f20984a079d1684e98f3efb11dd2f27891
3
  size 267832560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:243123fae8ffcab79e79cb033d989862a7729cf7eba06b5e7b19eb9e22e78288
3
  size 267832560
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:45e21ebc826ccc9dc350dc5bc13d021075985261771ac9c426009e997b2ca53d
3
  size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe11b14f7bb4c0b48177b35afe438c4bf38837b0279aa02e8c827cd3b25659da
3
  size 4984