Update README.md
Browse filesInitial Model Card
README.md
CHANGED
@@ -11,30 +11,42 @@ model-index:
|
|
11 |
results: []
|
12 |
---
|
13 |
|
14 |
-
|
15 |
-
should probably proofread and complete it, then remove this comment. -->
|
16 |
|
17 |
-
|
18 |
-
|
19 |
-
This model is a fine-tuned version of [google/gemma-2b-it](https://huggingface.co/google/gemma-2b-it) on the None dataset.
|
20 |
It achieves the following results on the evaluation set:
|
21 |
- Loss: 2.3061
|
22 |
|
23 |
-
## Model description
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
24 |
|
25 |
-
|
|
|
26 |
|
27 |
-
|
|
|
28 |
|
29 |
-
|
30 |
|
31 |
-
|
|
|
|
|
32 |
|
33 |
-
|
34 |
|
35 |
-
|
|
|
36 |
|
37 |
-
### Training hyperparameters
|
38 |
|
39 |
The following hyperparameters were used during training:
|
40 |
- learning_rate: 1e-05
|
|
|
11 |
results: []
|
12 |
---
|
13 |
|
14 |
+
# gemma-2b-lahacks 💻
|
|
|
15 |
|
16 |
+
This model is a fine-tuned version of [google/gemma-2b-it](https://huggingface.co/google/gemma-2b-it).
|
|
|
|
|
17 |
It achieves the following results on the evaluation set:
|
18 |
- Loss: 2.3061
|
19 |
|
20 |
+
## Model description 📝
|
21 |
+
|
22 |
+
This model was fine-tuned during LAHacks 2024, the intention of this model is to be able to diagnose a patient appropratiely
|
23 |
+
based on the information in their previous medical records, current symptoms, age, sex, and more.
|
24 |
+
|
25 |
+
## Intended uses & limitations ⁉️
|
26 |
+
|
27 |
+
Sample code snippet:
|
28 |
+
```py
|
29 |
+
|
30 |
+
```
|
31 |
|
32 |
+
Uses: To use Artificial Intelligence technology to diagnose patient based off of multiple parameters ranging from their age to their
|
33 |
+
medical record.
|
34 |
|
35 |
+
Limitation: There's a highly likelyhood that the model will NOT be great at diagnosing it's users, the amount of time it took to fine-tune
|
36 |
+
this model limited how much data we could train it on. With more time a more accurate model would be expected.
|
37 |
|
38 |
+
## Training and evaluation data 📈
|
39 |
|
40 |
+
The model was trained on data from the research paper 'A New Dataset For Automatic Medical Diagnosis' by Arsène Fansi Tchango, Rishab Goel,
|
41 |
+
Zhi Wen, Julien Martel, Joumana Ghosn. The 'release_train_patients.csv' dataset was reduced from it's original 1.3 million rows of data to a
|
42 |
+
mere 500-1000 rows of data. This was due to the time it took to fine-tune a model which depended on how big the dataset provided was.
|
43 |
|
44 |
+
## Training procedure 🏋️
|
45 |
|
46 |
+
The fine-tuning took MULTIPLE, and I mean MULTIPLE tries. Sometimes the dataset provided was very big so the kernel had to be restarted multiple times.
|
47 |
+
Additionally, the model was tuned on the default data that Intel offers in their guide to fine-tune a gemma model.
|
48 |
|
49 |
+
### Training hyperparameters 🔍
|
50 |
|
51 |
The following hyperparameters were used during training:
|
52 |
- learning_rate: 1e-05
|