LaZeAsh commited on
Commit
1c0a926
1 Parent(s): 07250ec

Update README.md

Browse files

Initial Model Card

Files changed (1) hide show
  1. README.md +25 -13
README.md CHANGED
@@ -11,30 +11,42 @@ model-index:
11
  results: []
12
  ---
13
 
14
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
- should probably proofread and complete it, then remove this comment. -->
16
 
17
- # gemma-2b-lahacks
18
-
19
- This model is a fine-tuned version of [google/gemma-2b-it](https://huggingface.co/google/gemma-2b-it) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
  - Loss: 2.3061
22
 
23
- ## Model description
 
 
 
 
 
 
 
 
 
 
24
 
25
- More information needed
 
26
 
27
- ## Intended uses & limitations
 
28
 
29
- More information needed
30
 
31
- ## Training and evaluation data
 
 
32
 
33
- More information needed
34
 
35
- ## Training procedure
 
36
 
37
- ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
  - learning_rate: 1e-05
 
11
  results: []
12
  ---
13
 
14
+ # gemma-2b-lahacks 💻
 
15
 
16
+ This model is a fine-tuned version of [google/gemma-2b-it](https://huggingface.co/google/gemma-2b-it).
 
 
17
  It achieves the following results on the evaluation set:
18
  - Loss: 2.3061
19
 
20
+ ## Model description 📝
21
+
22
+ This model was fine-tuned during LAHacks 2024, the intention of this model is to be able to diagnose a patient appropratiely
23
+ based on the information in their previous medical records, current symptoms, age, sex, and more.
24
+
25
+ ## Intended uses & limitations ⁉️
26
+
27
+ Sample code snippet:
28
+ ```py
29
+
30
+ ```
31
 
32
+ Uses: To use Artificial Intelligence technology to diagnose patient based off of multiple parameters ranging from their age to their
33
+ medical record.
34
 
35
+ Limitation: There's a highly likelyhood that the model will NOT be great at diagnosing it's users, the amount of time it took to fine-tune
36
+ this model limited how much data we could train it on. With more time a more accurate model would be expected.
37
 
38
+ ## Training and evaluation data 📈
39
 
40
+ The model was trained on data from the research paper 'A New Dataset For Automatic Medical Diagnosis' by Arsène Fansi Tchango, Rishab Goel,
41
+ Zhi Wen, Julien Martel, Joumana Ghosn. The 'release_train_patients.csv' dataset was reduced from it's original 1.3 million rows of data to a
42
+ mere 500-1000 rows of data. This was due to the time it took to fine-tune a model which depended on how big the dataset provided was.
43
 
44
+ ## Training procedure 🏋️
45
 
46
+ The fine-tuning took MULTIPLE, and I mean MULTIPLE tries. Sometimes the dataset provided was very big so the kernel had to be restarted multiple times.
47
+ Additionally, the model was tuned on the default data that Intel offers in their guide to fine-tune a gemma model.
48
 
49
+ ### Training hyperparameters 🔍
50
 
51
  The following hyperparameters were used during training:
52
  - learning_rate: 1e-05