lewtun HF staff commited on
Commit
59f829c
β€’
1 Parent(s): b6fc16d

Add style tweaks

Browse files
Files changed (1) hide show
  1. app.py +11 -8
app.py CHANGED
@@ -18,6 +18,10 @@ AUTONLP_USERNAME = os.getenv("AUTONLP_USERNAME")
18
  HF_AUTONLP_BACKEND_API = os.getenv("HF_AUTONLP_BACKEND_API")
19
  LOCAL_REPO = "submission_repo"
20
 
 
 
 
 
21
 
22
  ###########
23
  ### APP ###
@@ -70,9 +74,8 @@ with st.form(key="form"):
70
  "tasks":
71
  {
72
  "dataset_identifier": {
73
- "values": ["output1", "output2", "..."], # A list of system outputs.
74
- # Optionally, you can add the keys which are part of an example to ensure that there is no shuffling mistakes.
75
- "keys": ["key-0", "key-1", ...]
76
  }
77
  }
78
  }
@@ -80,8 +83,8 @@ with st.form(key="form"):
80
  In this case, `dataset_identifier` is the identifier of the dataset
81
  followed by an identifier of the set the outputs were created from, for
82
  example `_validation` or `_test`. For example, the `mlsum_de` test set
83
- would have the identifier `mlsum_de_test`. The `keys` field can be set
84
- to avoid accidental shuffling to impact your metrics. Simply add a list
85
  of the `gem_id` for each output example in the same order as your
86
  values. Please see the sample submission below:
87
  """
@@ -110,7 +113,7 @@ if submit_button and submission_errors == 0:
110
  submission_name = json_data["submission_name"]
111
 
112
  # Create submission dataset under benchmarks ORG
113
- dataset_repo_url = f"https://huggingface.co/datasets/benchmarks/gem-{user_name}"
114
  repo = Repository(
115
  local_dir=LOCAL_REPO,
116
  clone_from=dataset_repo_url,
@@ -150,9 +153,9 @@ if submit_button and submission_errors == 0:
150
  ).json()
151
 
152
  if json_resp["status"] == 1:
153
- st.success(f"βœ… Submission {submission_name} was successfully submitted to the evaluation queue!")
154
  else:
155
- st.error("πŸ™ˆ Oh noes! There was an error submitting your submission. Please contact the organisers")
156
 
157
  # Flush local repo
158
  shutil.rmtree(LOCAL_REPO, ignore_errors=True)
 
18
  HF_AUTONLP_BACKEND_API = os.getenv("HF_AUTONLP_BACKEND_API")
19
  LOCAL_REPO = "submission_repo"
20
 
21
+ ## TODO ##
22
+ # 1. Add check that fields are nested under `tasks` field correctly
23
+ # 2. Add check that names of tasks and datasets are valid
24
+
25
 
26
  ###########
27
  ### APP ###
 
74
  "tasks":
75
  {
76
  "dataset_identifier": {
77
+ "values": ["output-0", "output-1", "..."], # A list of system outputs.
78
+ "keys": ["gem_id-0", "gem_id-1", ...] # A list of GEM IDs.
 
79
  }
80
  }
81
  }
 
83
  In this case, `dataset_identifier` is the identifier of the dataset
84
  followed by an identifier of the set the outputs were created from, for
85
  example `_validation` or `_test`. For example, the `mlsum_de` test set
86
+ would have the identifier `mlsum_de_test`. The `keys` field is needed
87
+ to avoid accidental shuffling that will impact your metrics. Simply add a list
88
  of the `gem_id` for each output example in the same order as your
89
  values. Please see the sample submission below:
90
  """
 
113
  submission_name = json_data["submission_name"]
114
 
115
  # Create submission dataset under benchmarks ORG
116
+ dataset_repo_url = f"https://huggingface.co/datasets/GEM-submissions/gem-{user_name}"
117
  repo = Repository(
118
  local_dir=LOCAL_REPO,
119
  clone_from=dataset_repo_url,
 
153
  ).json()
154
 
155
  if json_resp["status"] == 1:
156
+ st.success(f"βœ… Submission {submission_name} was successfully submitted for evaluation!")
157
  else:
158
+ st.error("πŸ™ˆ Oh noes, there was an error submitting your submission! Please contact the organisers")
159
 
160
  # Flush local repo
161
  shutil.rmtree(LOCAL_REPO, ignore_errors=True)