Midi-Composer / midi_tokenizer.py
Next
Upload 17 files
e2da60d verified
raw
history blame
11.5 kB
import random
import PIL
import numpy as np
class MIDITokenizer:
def __init__(self):
self.vocab_size = 0
def allocate_ids(size):
ids = [self.vocab_size + i for i in range(size)]
self.vocab_size += size
return ids
self.pad_id = allocate_ids(1)[0]
self.bos_id = allocate_ids(1)[0]
self.eos_id = allocate_ids(1)[0]
self.events = {
"note": ["time1", "time2", "track", "duration", "channel", "pitch", "velocity"],
"patch_change": ["time1", "time2", "track", "channel", "patch"],
"control_change": ["time1", "time2", "track", "channel", "controller", "value"],
"set_tempo": ["time1", "time2", "track", "bpm"],
}
self.event_parameters = {
"time1": 128, "time2": 16, "duration": 2048, "track": 128, "channel": 16, "pitch": 128, "velocity": 128,
"patch": 128, "controller": 128, "value": 128, "bpm": 256
}
self.event_ids = {e: allocate_ids(1)[0] for e in self.events.keys()}
self.id_events = {i: e for e, i in self.event_ids.items()}
self.parameter_ids = {p: allocate_ids(s) for p, s in self.event_parameters.items()}
self.max_token_seq = max([len(ps) for ps in self.events.values()]) + 1
def tempo2bpm(self, tempo):
tempo = tempo / 10 ** 6 # us to s
bpm = 60 / tempo
return bpm
def bpm2tempo(self, bpm):
if bpm == 0:
bpm = 1
tempo = int((60 / bpm) * 10 ** 6)
return tempo
def tokenize(self, midi_score, add_bos_eos=True):
ticks_per_beat = midi_score[0]
event_list = {}
for track_idx, track in enumerate(midi_score[1:129]):
last_notes = {}
for event in track:
t = round(16 * event[1] / ticks_per_beat) # quantization
new_event = [event[0], t // 16, t % 16, track_idx] + event[2:]
if event[0] == "note":
new_event[4] = max(1, round(16 * new_event[4] / ticks_per_beat))
elif event[0] == "set_tempo":
new_event[4] = int(self.tempo2bpm(new_event[4]))
if event[0] == "note":
key = tuple(new_event[:4] + new_event[5:-1])
else:
key = tuple(new_event[:-1])
if event[0] == "note": # to eliminate note overlap due to quantization
cp = tuple(new_event[5:7])
if cp in last_notes:
last_note_key, last_note = last_notes[cp]
last_t = last_note[1] * 16 + last_note[2]
last_note[4] = max(0, min(last_note[4], t - last_t))
if last_note[4] == 0:
event_list.pop(last_note_key)
last_notes[cp] = (key, new_event)
event_list[key] = new_event
event_list = list(event_list.values())
event_list = sorted(event_list, key=lambda e: e[1:4])
midi_seq = []
last_t1 = 0
for event in event_list:
name = event[0]
if name in self.event_ids:
params = event[1:]
cur_t1 = params[0]
params[0] = params[0] - last_t1
if not all([0 <= params[i] < self.event_parameters[p] for i, p in enumerate(self.events[name])]):
continue
tokens = [self.event_ids[name]] + [self.parameter_ids[p][params[i]]
for i, p in enumerate(self.events[name])]
tokens += [self.pad_id] * (self.max_token_seq - len(tokens))
midi_seq.append(tokens)
last_t1 = cur_t1
if add_bos_eos:
bos = [self.bos_id] + [self.pad_id] * (self.max_token_seq - 1)
eos = [self.eos_id] + [self.pad_id] * (self.max_token_seq - 1)
midi_seq = [bos] + midi_seq + [eos]
return midi_seq
def event2tokens(self, event):
name = event[0]
params = event[1:]
tokens = [self.event_ids[name]] + [self.parameter_ids[p][params[i]]
for i, p in enumerate(self.events[name])]
tokens += [self.pad_id] * (self.max_token_seq - len(tokens))
return tokens
def tokens2event(self, tokens):
if tokens[0] in self.id_events:
name = self.id_events[tokens[0]]
if len(tokens) <= len(self.events[name]):
return []
params = tokens[1:]
params = [params[i] - self.parameter_ids[p][0] for i, p in enumerate(self.events[name])]
if not all([0 <= params[i] < self.event_parameters[p] for i, p in enumerate(self.events[name])]):
return []
event = [name] + params
return event
return []
def detokenize(self, midi_seq):
ticks_per_beat = 480
tracks_dict = {}
t1 = 0
for tokens in midi_seq:
if tokens[0] in self.id_events:
name = self.id_events[tokens[0]]
if len(tokens) <= len(self.events[name]):
continue
params = tokens[1:]
params = [params[i] - self.parameter_ids[p][0] for i, p in enumerate(self.events[name])]
if not all([0 <= params[i] < self.event_parameters[p] for i, p in enumerate(self.events[name])]):
continue
event = [name] + params
if name == "set_tempo":
event[4] = self.bpm2tempo(event[4])
if event[0] == "note":
event[4] = int(event[4] * ticks_per_beat / 16)
t1 += event[1]
t = t1 * 16 + event[2]
t = int(t * ticks_per_beat / 16)
track_idx = event[3]
if track_idx not in tracks_dict:
tracks_dict[track_idx] = []
tracks_dict[track_idx].append([event[0], t] + event[4:])
tracks = list(tracks_dict.values())
for i in range(len(tracks)): # to eliminate note overlap
track = tracks[i]
track = sorted(track, key=lambda e: e[1])
last_note_t = {}
zero_len_notes = []
for e in reversed(track):
if e[0] == "note":
t, d, c, p = e[1:5]
key = (c, p)
if key in last_note_t:
d = min(d, max(last_note_t[key] - t, 0))
last_note_t[key] = t
e[2] = d
if d == 0:
zero_len_notes.append(e)
for e in zero_len_notes:
track.remove(e)
tracks[i] = track
return [ticks_per_beat, *tracks]
def midi2img(self, midi_score):
ticks_per_beat = midi_score[0]
notes = []
max_time = 1
track_num = len(midi_score[1:])
for track_idx, track in enumerate(midi_score[1:]):
for event in track:
t = round(16 * event[1] / ticks_per_beat)
if event[0] == "note":
d = max(1, round(16 * event[2] / ticks_per_beat))
c, p = event[3:5]
max_time = max(max_time, t + d + 1)
notes.append((track_idx, c, p, t, d))
img = np.zeros((128, max_time, 3), dtype=np.uint8)
colors = {(i, j): np.random.randint(50, 256, 3) for i in range(track_num) for j in range(16)}
for note in notes:
tr, c, p, t, d = note
img[p, t: t + d] = colors[(tr, c)]
img = PIL.Image.fromarray(np.flip(img, 0))
return img
def augment(self, midi_seq, max_pitch_shift=4, max_vel_shift=10, max_cc_val_shift=10, max_bpm_shift=10,
max_track_shift=128, max_channel_shift=16):
pitch_shift = random.randint(-max_pitch_shift, max_pitch_shift)
vel_shift = random.randint(-max_vel_shift, max_vel_shift)
cc_val_shift = random.randint(-max_cc_val_shift, max_cc_val_shift)
bpm_shift = random.randint(-max_bpm_shift, max_bpm_shift)
track_shift = random.randint(0, max_track_shift)
channel_shift = random.randint(0, max_channel_shift)
midi_seq_new = []
for tokens in midi_seq:
tokens_new = [*tokens]
if tokens[0] in self.id_events:
name = self.id_events[tokens[0]]
for i, pn in enumerate(self.events[name]):
if pn == "track":
tr = tokens[1 + i] - self.parameter_ids[pn][0]
tr += track_shift
tr = tr % self.event_parameters[pn]
tokens_new[1 + i] = self.parameter_ids[pn][tr]
elif pn == "channel":
c = tokens[1 + i] - self.parameter_ids[pn][0]
c0 = c
c += channel_shift
c = c % self.event_parameters[pn]
if c0 == 9:
c = 9
elif c == 9:
c = (9 + channel_shift) % self.event_parameters[pn]
tokens_new[1 + i] = self.parameter_ids[pn][c]
if name == "note":
c = tokens[5] - self.parameter_ids["channel"][0]
p = tokens[6] - self.parameter_ids["pitch"][0]
v = tokens[7] - self.parameter_ids["velocity"][0]
if c != 9: # no shift for drums
p += pitch_shift
if not 0 <= p < 128:
return midi_seq
v += vel_shift
v = max(1, min(127, v))
tokens_new[6] = self.parameter_ids["pitch"][p]
tokens_new[7] = self.parameter_ids["velocity"][v]
elif name == "control_change":
cc = tokens[5] - self.parameter_ids["controller"][0]
val = tokens[6] - self.parameter_ids["value"][0]
if cc in [1, 2, 7, 11]:
val += cc_val_shift
val = max(1, min(127, val))
tokens_new[6] = self.parameter_ids["value"][val]
elif name == "set_tempo":
bpm = tokens[4] - self.parameter_ids["bpm"][0]
bpm += bpm_shift
bpm = max(1, min(255, bpm))
tokens_new[4] = self.parameter_ids["bpm"][bpm]
midi_seq_new.append(tokens_new)
return midi_seq_new
def check_alignment(self, midi_seq, threshold=0.3):
total = 0
hist = [0] * 16
for tokens in midi_seq:
if tokens[0] in self.id_events and self.id_events[tokens[0]] == "note":
t2 = tokens[2] - self.parameter_ids["time2"][0]
total += 1
hist[t2] += 1
if total == 0:
return False
hist = sorted(hist, reverse=True)
p = sum(hist[:2]) / total
return p > threshold