Joseph Feng commited on
Commit
52f268c
1 Parent(s): d0fb542

update link and fix assertion bug

Browse files
{{cookiecutter.repo_name}}/README.md CHANGED
@@ -1,6 +1,6 @@
1
  # SUPERB Submission Template
2
 
3
- Welcome to the [SUPERB Challenge](https://superbbenchmark.org/challenge)! SUPERB is a collection of benchmarking resources to evaluate the capability of a universal shared representation for speech processing. It comes with a benchmark on the publicly available datasets and a challenge on a secret/not released hidden dataset. In SUPERB Challenge, a challenging hidden dataset is newly recorded to evaluate the ultimate generaliziblity across various tasks and data.
4
 
5
  You can participate the challenge by simply submitting your self-supervised (SSL) pretrained models (model definition & pretrained weights), and we benchmark it with the hidden dataset. This repository constains useful tools to let you easliy [submit](https://superbbenchmark.org/submit) your models ***privately*** for evaluation to [the challenge hidden-set leaderboard](https://superbbenchmark.org/leaderboard?track=constrained&subset=Hidden+Dev+Set).
6
 
@@ -11,11 +11,11 @@ You can participate the challenge by simply submitting your self-supervised (SSL
11
 
12
  #### Note 1.
13
 
14
- We accept pre-trained models in PyTorch by default. If you wish to submit upstreams in non-PyTorch frameworks, please [fill this form](https://docs.google.com/forms/d/e/1FAIpQLSe52jYL2Yk9oYqXfg_Bg0Sjp01a6HSLUhY5VohsZOE5sNmgsw/viewform)!
15
 
16
  #### Note 2.
17
 
18
- If you are not feasible to submit the pre-trained model, please [fill this form](https://docs.google.com/forms/d/e/1FAIpQLSdA44nArlIDfGV63WwtwXer4WAPQO1aBwEpAjDSNjbMQN-GJQ/viewform) for us to see how to help!
19
 
20
  ## Quickstart
21
 
@@ -90,7 +90,7 @@ That's it! You're now all set to start pretraining your speech models - see the
90
 
91
  To make a submission to the [leaderboard](https://superbbenchmark.org/leaderboard?subset=Hidden+Dev+Set), there are 4 main steps:
92
 
93
- 1. Modify `expert.py` and change `model.pt` so we can initialize an upstream model following the [challenge policy](https://superbbenchmark.org/challenge#Upstream-Specification) by:
94
 
95
  ```python
96
  upstream = UpstreamExpert(ckpt="./model.pt")
@@ -98,7 +98,7 @@ To make a submission to the [leaderboard](https://superbbenchmark.org/leaderboar
98
 
99
  ***Package Dependency:*** Note that we only install `torch` package so far by following the above steps. If your model needs more packages, you can modify the `requirement.txt` to meet your need and install them inside the current conda environment. We will install the packages you list in the `requirement.txt` before initializing the upstream model.
100
 
101
- 2. Validate the upstream model's interface meets the requirements in the [challenge policy](https://superbbenchmark.org/challenge#Upstream-Specification). If everything is correct, you should see the following message: "All submission files validated! Now you can make a submission."
102
 
103
  ```
104
  python cli.py validate
@@ -121,3 +121,7 @@ To make a submission to the [leaderboard](https://superbbenchmark.org/leaderboar
121
  - Commit Hash (full 40 characters)
122
 
123
  After you finish the above 4 steps. You will see a new entry in your [SUPERB profile page](https://superbbenchmark.org/profile) (need login) which does not have any benchmark numbers yet. Please wait for us to finetuned it on the hidden dataset and get the benchmark results. The results will be revealed within one week. Please stay tuned!
 
 
 
 
 
1
  # SUPERB Submission Template
2
 
3
+ Welcome to the [SUPERB Challenge](https://superbbenchmark.org/challenge-slt2022/challenge_overview)! SUPERB is a collection of benchmarking resources to evaluate the capability of a universal shared representation for speech processing. It comes with a benchmark on the publicly available datasets and a challenge on a secret/not released hidden dataset. In SUPERB Challenge, a challenging hidden dataset is newly recorded to evaluate the ultimate generaliziblity across various tasks and data.
4
 
5
  You can participate the challenge by simply submitting your self-supervised (SSL) pretrained models (model definition & pretrained weights), and we benchmark it with the hidden dataset. This repository constains useful tools to let you easliy [submit](https://superbbenchmark.org/submit) your models ***privately*** for evaluation to [the challenge hidden-set leaderboard](https://superbbenchmark.org/leaderboard?track=constrained&subset=Hidden+Dev+Set).
6
 
 
11
 
12
  #### Note 1.
13
 
14
+ We accept pre-trained models in PyTorch by default. If you wish to submit upstreams in non-PyTorch frameworks, please [contant us](#contact)!
15
 
16
  #### Note 2.
17
 
18
+ If you are not feasible to submit the pre-trained model, please [contant us](#contact) for us to see how to help!
19
 
20
  ## Quickstart
21
 
 
90
 
91
  To make a submission to the [leaderboard](https://superbbenchmark.org/leaderboard?subset=Hidden+Dev+Set), there are 4 main steps:
92
 
93
+ 1. Modify `expert.py` and change `model.pt` so we can initialize an upstream model following the [challenge policy](https://superbbenchmark.org/challenge-slt2022/upstream) by:
94
 
95
  ```python
96
  upstream = UpstreamExpert(ckpt="./model.pt")
 
98
 
99
  ***Package Dependency:*** Note that we only install `torch` package so far by following the above steps. If your model needs more packages, you can modify the `requirement.txt` to meet your need and install them inside the current conda environment. We will install the packages you list in the `requirement.txt` before initializing the upstream model.
100
 
101
+ 2. Validate the upstream model's interface meets the requirements in the [challenge policy](https://superbbenchmark.org/challenge-slt2022/upstream). If everything is correct, you should see the following message: "All submission files validated! Now you can make a submission."
102
 
103
  ```
104
  python cli.py validate
 
121
  - Commit Hash (full 40 characters)
122
 
123
  After you finish the above 4 steps. You will see a new entry in your [SUPERB profile page](https://superbbenchmark.org/profile) (need login) which does not have any benchmark numbers yet. Please wait for us to finetuned it on the hidden dataset and get the benchmark results. The results will be revealed within one week. Please stay tuned!
124
+
125
+ ## Contact
126
+
127
{{cookiecutter.repo_name}}/cli.py CHANGED
@@ -27,7 +27,7 @@ def validate():
27
  assert isinstance(results, dict)
28
  tasks = ["PR", "SID", "ER", "ASR", "ASV", "SD", "QbE", "ST", "SS", "SE", "secret"]
29
  for task in tasks:
30
- hidden_states = results.get(task, "hidden_states")
31
  assert isinstance(hidden_states, list)
32
 
33
  for state in hidden_states:
@@ -40,7 +40,7 @@ def validate():
40
  assert abs(round(max(samples) / downsample_rate) - hidden_states[0].size(1)) < 5, "wrong downsample rate"
41
 
42
  except:
43
- print("Please check the Upstream Specification on https://superbbenchmark.org/challenge")
44
  raise
45
 
46
  typer.echo("All submission files validated!")
 
27
  assert isinstance(results, dict)
28
  tasks = ["PR", "SID", "ER", "ASR", "ASV", "SD", "QbE", "ST", "SS", "SE", "secret"]
29
  for task in tasks:
30
+ hidden_states = results.get(task, results["hidden_states"])
31
  assert isinstance(hidden_states, list)
32
 
33
  for state in hidden_states:
 
40
  assert abs(round(max(samples) / downsample_rate) - hidden_states[0].size(1)) < 5, "wrong downsample rate"
41
 
42
  except:
43
+ print("Please check the Upstream Specification on https://superbbenchmark.org/challenge-slt2022/upstream")
44
  raise
45
 
46
  typer.echo("All submission files validated!")