umarbutler commited on
Commit
152cbb9
•
1 Parent(s): 01dc13a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -0
README.md CHANGED
@@ -96,6 +96,15 @@ After training on two RTX A6000s for \~120,050 steps over a period of 91 hours,
96
 
97
  Naturally, as the optimiser state had been lost, the model's learning rate descended slower than it had been previously. Nevertheless, after completing an epoch of training, the model was able to achieve a validation loss of 2.04.
98
 
 
 
 
 
 
 
 
 
 
99
  ## Limitations 🚧
100
  Although the model has not been tested for bias, one would expect it to exhibit much of the same, if not all, the biases of [GPT2-XL](https://huggingface.co/gpt2-xl).
101
 
 
96
 
97
  Naturally, as the optimiser state had been lost, the model's learning rate descended slower than it had been previously. Nevertheless, after completing an epoch of training, the model was able to achieve a validation loss of 2.04.
98
 
99
+ ## Benchmarks 📊
100
+ Tested against version 2.0.0 of the [Open Australian Legal QA](https://huggingface.co/datasets/umarbutler/open-australian-legal-qa) dataset, the model achieved a perplexity of 8.01, outperforming all known language models for Australian law.
101
+ | Model | Parameters | Perplexity |
102
+ |--|--|--|
103
+ | **Open Australian Legal LLM** | **1.5B** | **8.01** |
104
+ | [Open Australian Legal Phi 1.5](https://huggingface.co/umarbutler/open-australian-legal-phi-1_5) | 1.3B | 8.69 |
105
+ | [Open Australian Legal GPT2](https://huggingface.co/umarbutler/open-australian-legal-gpt2) | 124M | 16.37 |
106
+ | [Open Australian Legal DistilGPT2](https://huggingface.co/umarbutler/open-australian-legal-distilgpt2) | 88.2M | 23.9 |
107
+
108
  ## Limitations 🚧
109
  Although the model has not been tested for bias, one would expect it to exhibit much of the same, if not all, the biases of [GPT2-XL](https://huggingface.co/gpt2-xl).
110