Joseph Feng
commited on
Commit
•
de6da40
1
Parent(s):
0d9842f
enable task-specified preprocessing
Browse files- README.md +7 -6
- {{cookiecutter.repo_name}}/README.md +7 -6
- {{cookiecutter.repo_name}}/expert.py +15 -14
README.md
CHANGED
@@ -2,7 +2,7 @@
|
|
2 |
|
3 |
Welcome to the [SUPERB Challenge](https://superbbenchmark.org/challenge-slt2022/challenge_overview)! SUPERB is a collection of benchmarking resources to evaluate the capability of a universal shared representation for speech processing. It comes with a benchmark on the publicly available datasets and a challenge on a secret/not released hidden dataset. In SUPERB Challenge, a challenging hidden dataset is newly recorded to evaluate the ultimate generaliziblity across various tasks and data.
|
4 |
|
5 |
-
You can participate the challenge by simply submitting your self-supervised (SSL) pretrained models (model definition & pretrained weights), and we benchmark it with the hidden
|
6 |
|
7 |
1. Generate a submission template
|
8 |
2. Validate the format/interface correctness of your model
|
@@ -32,16 +32,17 @@ Extract features from waveforms.
|
|
32 |
BATCH_SIZE = 8
|
33 |
EXAMPLE_SEC = 10
|
34 |
wavs = [torch.randn(SAMPLE_RATE * EXAMPLE_SEC).cuda() for _ in range(BATCH_SIZE)]
|
35 |
-
results = upstream(wavs)
|
36 |
```
|
37 |
|
38 |
-
- **Output:** A dictionary with a key
|
39 |
|
40 |
```python
|
41 |
-
|
42 |
-
tasks = ["PR", "SID", "ER", "ASR", "ASV", "SD", "QbE", "ST", "SS", "SE"]
|
43 |
for task in tasks:
|
44 |
-
|
|
|
|
|
|
|
45 |
assert isinstance(hidden_states, list)
|
46 |
|
47 |
for state in hidden_states:
|
|
|
2 |
|
3 |
Welcome to the [SUPERB Challenge](https://superbbenchmark.org/challenge-slt2022/challenge_overview)! SUPERB is a collection of benchmarking resources to evaluate the capability of a universal shared representation for speech processing. It comes with a benchmark on the publicly available datasets and a challenge on a secret/not released hidden dataset. In SUPERB Challenge, a challenging hidden dataset is newly recorded to evaluate the ultimate generaliziblity across various tasks and data.
|
4 |
|
5 |
+
You can participate the challenge by simply submitting your self-supervised (SSL) pretrained models (model definition & pretrained weights), and we benchmark it with the hidden datasets. This repository constains useful tools to let you easliy [submit](https://superbbenchmark.org/submit) your models ***privately*** for evaluation to [the challenge hidden-set leaderboard](https://superbbenchmark.org/leaderboard?track=constrained&subset=Hidden+Dev+Set).
|
6 |
|
7 |
1. Generate a submission template
|
8 |
2. Validate the format/interface correctness of your model
|
|
|
32 |
BATCH_SIZE = 8
|
33 |
EXAMPLE_SEC = 10
|
34 |
wavs = [torch.randn(SAMPLE_RATE * EXAMPLE_SEC).cuda() for _ in range(BATCH_SIZE)]
|
|
|
35 |
```
|
36 |
|
37 |
+
- **Output:** A dictionary with a key "hidden_states" (for compatiblility with old ver.). The value is **a list** of padded sequences in the same shape of **(batch_size, max_sequence_length_of_batch, hidden_size)** for weighted-sum to work. It is welcome to perform some task-specified / independent pre- / post-processing on the upstream's raw hidden-sets, including upsampling and downsampling. However, all the values must come from **a single upstream model**:
|
38 |
|
39 |
```python
|
40 |
+
tasks = ["hidden_states", "PR", "SID", "ER", "ASR", "ASV", "SD", "QbE", "ST", "SS", "SE", "secret"]
|
|
|
41 |
for task in tasks:
|
42 |
+
# you can do task-specified pre- / post-processing depend on the arg "upstream_feature_selection"
|
43 |
+
results = upstream(wavs, upstream_feature_selection=task)
|
44 |
+
hidden_states = results["hidden_states"]
|
45 |
+
assert isinstance(results, dict)
|
46 |
assert isinstance(hidden_states, list)
|
47 |
|
48 |
for state in hidden_states:
|
{{cookiecutter.repo_name}}/README.md
CHANGED
@@ -2,7 +2,7 @@
|
|
2 |
|
3 |
Welcome to the [SUPERB Challenge](https://superbbenchmark.org/challenge-slt2022/challenge_overview)! SUPERB is a collection of benchmarking resources to evaluate the capability of a universal shared representation for speech processing. It comes with a benchmark on the publicly available datasets and a challenge on a secret/not released hidden dataset. In SUPERB Challenge, a challenging hidden dataset is newly recorded to evaluate the ultimate generaliziblity across various tasks and data.
|
4 |
|
5 |
-
You can participate the challenge by simply submitting your self-supervised (SSL) pretrained models (model definition & pretrained weights), and we benchmark it with the hidden
|
6 |
|
7 |
1. Generate a submission template
|
8 |
2. Validate the format/interface correctness of your model
|
@@ -32,16 +32,17 @@ Extract features from waveforms.
|
|
32 |
BATCH_SIZE = 8
|
33 |
EXAMPLE_SEC = 10
|
34 |
wavs = [torch.randn(SAMPLE_RATE * EXAMPLE_SEC).cuda() for _ in range(BATCH_SIZE)]
|
35 |
-
results = upstream(wavs)
|
36 |
```
|
37 |
|
38 |
-
- **Output:** A dictionary with a key
|
39 |
|
40 |
```python
|
41 |
-
|
42 |
-
tasks = ["PR", "SID", "ER", "ASR", "ASV", "SD", "QbE", "ST", "SS", "SE"]
|
43 |
for task in tasks:
|
44 |
-
|
|
|
|
|
|
|
45 |
assert isinstance(hidden_states, list)
|
46 |
|
47 |
for state in hidden_states:
|
|
|
2 |
|
3 |
Welcome to the [SUPERB Challenge](https://superbbenchmark.org/challenge-slt2022/challenge_overview)! SUPERB is a collection of benchmarking resources to evaluate the capability of a universal shared representation for speech processing. It comes with a benchmark on the publicly available datasets and a challenge on a secret/not released hidden dataset. In SUPERB Challenge, a challenging hidden dataset is newly recorded to evaluate the ultimate generaliziblity across various tasks and data.
|
4 |
|
5 |
+
You can participate the challenge by simply submitting your self-supervised (SSL) pretrained models (model definition & pretrained weights), and we benchmark it with the hidden datasets. This repository constains useful tools to let you easliy [submit](https://superbbenchmark.org/submit) your models ***privately*** for evaluation to [the challenge hidden-set leaderboard](https://superbbenchmark.org/leaderboard?track=constrained&subset=Hidden+Dev+Set).
|
6 |
|
7 |
1. Generate a submission template
|
8 |
2. Validate the format/interface correctness of your model
|
|
|
32 |
BATCH_SIZE = 8
|
33 |
EXAMPLE_SEC = 10
|
34 |
wavs = [torch.randn(SAMPLE_RATE * EXAMPLE_SEC).cuda() for _ in range(BATCH_SIZE)]
|
|
|
35 |
```
|
36 |
|
37 |
+
- **Output:** A dictionary with a key "hidden_states" (for compatiblility with old ver.). The value is **a list** of padded sequences in the same shape of **(batch_size, max_sequence_length_of_batch, hidden_size)** for weighted-sum to work. It is welcome to perform some task-specified / independent pre- / post-processing on the upstream's raw hidden-sets, including upsampling and downsampling. However, all the values must come from **a single upstream model**:
|
38 |
|
39 |
```python
|
40 |
+
tasks = ["hidden_states", "PR", "SID", "ER", "ASR", "ASV", "SD", "QbE", "ST", "SS", "SE", "secret"]
|
|
|
41 |
for task in tasks:
|
42 |
+
# you can do task-specified pre- / post-processing depend on the arg "upstream_feature_selection"
|
43 |
+
results = upstream(wavs, upstream_feature_selection=task)
|
44 |
+
hidden_states = results["hidden_states"]
|
45 |
+
assert isinstance(results, dict)
|
46 |
assert isinstance(hidden_states, list)
|
47 |
|
48 |
for state in hidden_states:
|
{{cookiecutter.repo_name}}/expert.py
CHANGED
@@ -15,7 +15,8 @@ class Model(nn.Module):
|
|
15 |
self.model1 = nn.Linear(1, HIDDEN_DIM)
|
16 |
self.model2 = nn.Linear(HIDDEN_DIM, HIDDEN_DIM)
|
17 |
|
18 |
-
def forward(self, wavs):
|
|
|
19 |
hidden = self.model1(wavs)
|
20 |
# hidden: (batch_size, max_len, hidden_dim)
|
21 |
|
@@ -25,15 +26,24 @@ class Model(nn.Module):
|
|
25 |
return [hidden, feature]
|
26 |
|
27 |
class UpstreamExpert(nn.Module):
|
28 |
-
def __init__(
|
|
|
|
|
|
|
|
|
29 |
"""
|
30 |
Args:
|
31 |
ckpt:
|
32 |
The checkpoint path for loading your pretrained weights.
|
33 |
Should be fixed as model.pt for SUPERB Challenge.
|
|
|
|
|
|
|
|
|
34 |
"""
|
35 |
super().__init__()
|
36 |
self.name = "[Example UpstreamExpert]"
|
|
|
37 |
|
38 |
# You can use ckpt to load your pretrained weights
|
39 |
ckpt = torch.load(ckpt, map_location="cpu")
|
@@ -57,21 +67,12 @@ class UpstreamExpert(nn.Module):
|
|
57 |
wavs = pad_sequence(wavs, batch_first=True).unsqueeze(-1)
|
58 |
# wavs: (batch_size, max_len, 1)
|
59 |
|
60 |
-
hidden_states = self.model(wavs)
|
61 |
|
|
|
|
|
62 |
# The "hidden_states" key will be used as default in many cases
|
63 |
# Others keys in this example are presented for SUPERB Challenge
|
64 |
return {
|
65 |
"hidden_states": hidden_states,
|
66 |
-
"PR": hidden_states,
|
67 |
-
"SID": hidden_states,
|
68 |
-
"ER": hidden_states,
|
69 |
-
"ASR": hidden_states,
|
70 |
-
"QbE": hidden_states,
|
71 |
-
"ASV": hidden_states,
|
72 |
-
"SD": hidden_states,
|
73 |
-
"ST": hidden_states,
|
74 |
-
"SE": hidden_states,
|
75 |
-
"SS": hidden_states,
|
76 |
-
"secret": hidden_states,
|
77 |
}
|
|
|
15 |
self.model1 = nn.Linear(1, HIDDEN_DIM)
|
16 |
self.model2 = nn.Linear(HIDDEN_DIM, HIDDEN_DIM)
|
17 |
|
18 |
+
def forward(self, wavs, upstream_feature_selection="hidden_states"):
|
19 |
+
# You can do task-specified pre- / post-processing based on upstream_feature_selection
|
20 |
hidden = self.model1(wavs)
|
21 |
# hidden: (batch_size, max_len, hidden_dim)
|
22 |
|
|
|
26 |
return [hidden, feature]
|
27 |
|
28 |
class UpstreamExpert(nn.Module):
|
29 |
+
def __init__(
|
30 |
+
self,
|
31 |
+
ckpt: str = "./model.pt",
|
32 |
+
upstream_feature_selection: str = "hidden_states",
|
33 |
+
**kwargs):
|
34 |
"""
|
35 |
Args:
|
36 |
ckpt:
|
37 |
The checkpoint path for loading your pretrained weights.
|
38 |
Should be fixed as model.pt for SUPERB Challenge.
|
39 |
+
upstream_feature_selection:
|
40 |
+
The value could be
|
41 |
+
'hidden_states', 'PR', 'SID', 'ER', 'ASR', 'QbE', 'ASV', 'SD', 'ST', 'SE', 'SS', 'secret', or others(new tasks).
|
42 |
+
You can use it to control which task-specified pre- / post-processing to do.
|
43 |
"""
|
44 |
super().__init__()
|
45 |
self.name = "[Example UpstreamExpert]"
|
46 |
+
self.upstream_feature_selection = upstream_feature_selection
|
47 |
|
48 |
# You can use ckpt to load your pretrained weights
|
49 |
ckpt = torch.load(ckpt, map_location="cpu")
|
|
|
67 |
wavs = pad_sequence(wavs, batch_first=True).unsqueeze(-1)
|
68 |
# wavs: (batch_size, max_len, 1)
|
69 |
|
70 |
+
hidden_states = self.model(wavs, upstream_feature_selection=self.upstream_feature_selection)
|
71 |
|
72 |
+
# Deprecated! Do not do any task-specified postprocess below
|
73 |
+
# You can use the init arg "upstream_feature_selection" to control which task-specified pre- / post-processing to do.
|
74 |
# The "hidden_states" key will be used as default in many cases
|
75 |
# Others keys in this example are presented for SUPERB Challenge
|
76 |
return {
|
77 |
"hidden_states": hidden_states,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
78 |
}
|