ep9io commited on
Commit
0325d43
1 Parent(s): 4d8f7bb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +27 -44
README.md CHANGED
@@ -194,28 +194,42 @@ model-index:
194
  The model is a multi-label classifier designed to detect various types of bias within job descriptions.
195
 
196
  - **Developed by:** Tristan Everitt and Paul Ryan
197
- - **Funded by [optional]:** [More Information Needed]
198
- - **Shared by [optional]:** [More Information Needed]
199
- - **Model type:** [More Information Needed]
200
  - **Language(s) (NLP):** en
201
  - **License:** apache-2.0
202
- - **Finetuned from model [optional]:** microsoft/Phi-3-mini-4k-instruct
203
 
204
- ### Model Sources [optional]
205
 
206
- <!-- Provide the basic links for the model. -->
207
-
208
- - **Repository:** https://gitlab.computing.dcu.ie/everitt2/2024-mcm-everitt-ryan
209
- - **Paper [optional]:** [More Information Needed]
210
- - **Demo [optional]:** [More Information Needed]
211
 
212
  ## Uses
213
 
214
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
 
 
 
 
 
 
215
 
216
- ### Direct Use
 
 
 
 
 
 
 
 
 
 
 
 
 
 
217
 
218
- <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
219
 
220
 
221
  ```python
@@ -235,37 +249,6 @@ The model is a multi-label classifier designed to detect various types of bias w
235
  {'label': 'racial', 'score': 0.014618005603551865},
236
  {'label': 'sexuality', 'score': 0.005568435415625572}
237
  ]]
238
-
239
-
240
- ### Downstream Use [optional]
241
-
242
- <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
243
-
244
- [More Information Needed]
245
-
246
- ### Out-of-Scope Use
247
-
248
- <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
249
-
250
- [More Information Needed]
251
-
252
- ## Bias, Risks, and Limitations
253
-
254
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
255
-
256
- [More Information Needed]
257
-
258
- ### Recommendations
259
-
260
- <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
261
-
262
- Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
263
-
264
- ## How to Get Started with the Model
265
-
266
- Use the code below to get started with the model.
267
-
268
- [More Information Needed]
269
 
270
  ## Training Details
271
 
 
194
  The model is a multi-label classifier designed to detect various types of bias within job descriptions.
195
 
196
  - **Developed by:** Tristan Everitt and Paul Ryan
197
+ - **Model type:** Decoder
 
 
198
  - **Language(s) (NLP):** en
199
  - **License:** apache-2.0
200
+ - **Finetuned from model:** microsoft/Phi-3-mini-4k-instruct
201
 
202
+ ### Model Sources
203
 
204
+ - **Repository:** https://github.com/2024-mcm-everitt-ryan
205
+ - **Paper:** In Progress
 
 
 
206
 
207
  ## Uses
208
 
209
+ The primary target audience for these models are researchers dedicated to identifying biased language in job descriptions.
210
+
211
+ ### Out-of-Scope Use
212
+
213
+ Due to the limitations inherent in large-scale language models, they should not be utilised in applications requiring factual or accurate outputs. These models do not distinguish between fact and fiction, and implicit biases are inherently subjective.
214
+
215
+ Moreover, as language models mirror the biases present in their training data, they should not be deployed in systems that directly interact with humans unless the deployers have first conducted a thorough analysis of relevant biases for the specific use case.
216
 
217
+ ## Bias, Risks, and Limitations
218
+
219
+ It is imperative that all users, both direct and downstream, are aware of the risks, biases, and limitations associated with this model. Important considerations include:
220
+
221
+ - Bias in Training Data: The model may inherit and perpetuate biases from the data it was trained on.
222
+ - Subjectivity of Bias: Bias detection is inherently subjective, and perceptions of bias can differ across contexts and users.
223
+ - Accuracy Concerns: The model’s outputs are not guaranteed to be true or accurate, making it unsuitable for applications that require reliable information.
224
+ - Human Interaction Risks: When incorporated into systems that interact with humans, the model’s biases may affect interactions and decision-making, potentially leading to unintended consequences.
225
+
226
+ It is crucial for users to conduct comprehensive evaluations and consider these factors when applying the model in any context.
227
+
228
+
229
+ ## How to Get Started with the Model
230
+
231
+ Use the code below to get started with the model.
232
 
 
233
 
234
 
235
  ```python
 
249
  {'label': 'racial', 'score': 0.014618005603551865},
250
  {'label': 'sexuality', 'score': 0.005568435415625572}
251
  ]]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
252
 
253
  ## Training Details
254