|
"""
|
|
Train script for a single file
|
|
|
|
Need to set the TPU address first:
|
|
export XRT_TPU_CONFIG="localservice;0;localhost:51011"
|
|
"""
|
|
|
|
import torch.multiprocessing as mp
|
|
import threading
|
|
import time
|
|
import random
|
|
import sys
|
|
import argparse
|
|
import gzip
|
|
import json
|
|
import logging
|
|
import tqdm
|
|
import torch
|
|
from torch import nn
|
|
from torch.utils.data import DataLoader
|
|
import torch
|
|
import torch_xla
|
|
import torch_xla.core
|
|
import torch_xla.core.functions
|
|
import torch_xla.core.xla_model as xm
|
|
import torch_xla.distributed.xla_multiprocessing as xmp
|
|
import torch_xla.distributed.parallel_loader as pl
|
|
import os
|
|
from shutil import copyfile
|
|
|
|
|
|
from transformers import (
|
|
AdamW,
|
|
AutoModel,
|
|
AutoTokenizer,
|
|
get_linear_schedule_with_warmup,
|
|
set_seed,
|
|
)
|
|
|
|
class AutoModelForSentenceEmbedding(nn.Module):
|
|
def __init__(self, model_name, tokenizer, normalize=True):
|
|
super(AutoModelForSentenceEmbedding, self).__init__()
|
|
|
|
self.model = AutoModel.from_pretrained(model_name)
|
|
self.normalize = normalize
|
|
self.tokenizer = tokenizer
|
|
|
|
def forward(self, **kwargs):
|
|
model_output = self.model(**kwargs)
|
|
embeddings = self.mean_pooling(model_output, kwargs['attention_mask'])
|
|
if self.normalize:
|
|
embeddings = torch.nn.functional.normalize(embeddings, p=2, dim=1)
|
|
|
|
return embeddings
|
|
|
|
def mean_pooling(self, model_output, attention_mask):
|
|
token_embeddings = model_output[0]
|
|
input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float()
|
|
return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9)
|
|
|
|
def save_pretrained(self, output_path):
|
|
if xm.is_master_ordinal():
|
|
self.tokenizer.save_pretrained(output_path)
|
|
self.model.config.save_pretrained(output_path)
|
|
|
|
xm.save(self.model.state_dict(), os.path.join(output_path, "pytorch_model.bin"))
|
|
|
|
|
|
|
|
|
|
def train_function(index, args, queue):
|
|
tokenizer = AutoTokenizer.from_pretrained(args.model)
|
|
model = AutoModelForSentenceEmbedding(args.model, tokenizer)
|
|
|
|
|
|
|
|
device = xm.xla_device()
|
|
model = model.to(device)
|
|
|
|
|
|
optimizer = AdamW(params=model.parameters(), lr=2e-5, correct_bias=True)
|
|
|
|
lr_scheduler = get_linear_schedule_with_warmup(
|
|
optimizer=optimizer,
|
|
num_warmup_steps=500,
|
|
num_training_steps=args.steps,
|
|
)
|
|
|
|
|
|
cross_entropy_loss = nn.CrossEntropyLoss()
|
|
max_grad_norm = 1
|
|
|
|
model.train()
|
|
|
|
for global_step in tqdm.trange(args.steps, disable=not xm.is_master_ordinal()):
|
|
|
|
batch = queue.get()
|
|
|
|
|
|
|
|
if len(batch[0]) == 2:
|
|
text1 = tokenizer([b[0] for b in batch], return_tensors="pt", max_length=args.max_length, truncation=True, padding="max_length")
|
|
text2 = tokenizer([b[1] for b in batch], return_tensors="pt", max_length=args.max_length, truncation=True, padding="max_length")
|
|
|
|
|
|
embeddings_a = model(**text1.to(device))
|
|
embeddings_b = model(**text2.to(device))
|
|
|
|
|
|
embeddings_a = torch_xla.core.functions.all_gather(embeddings_a)
|
|
embeddings_b = torch_xla.core.functions.all_gather(embeddings_b)
|
|
|
|
|
|
scores = torch.mm(embeddings_a, embeddings_b.transpose(0, 1)) * args.scale
|
|
|
|
|
|
labels = torch.tensor(range(len(scores)), dtype=torch.long, device=embeddings_a.device)
|
|
|
|
|
|
loss = (cross_entropy_loss(scores, labels) + cross_entropy_loss(scores.transpose(0, 1), labels)) / 2
|
|
|
|
else:
|
|
text1 = tokenizer([b[0] for b in batch], return_tensors="pt", max_length=args.max_length, truncation=True, padding="max_length")
|
|
text2 = tokenizer([b[1] for b in batch], return_tensors="pt", max_length=args.max_length, truncation=True, padding="max_length")
|
|
text3 = tokenizer([b[2] for b in batch], return_tensors="pt", max_length=args.max_length, truncation=True, padding="max_length")
|
|
|
|
embeddings_a = model(**text1.to(device))
|
|
embeddings_b1 = model(**text2.to(device))
|
|
embeddings_b2 = model(**text3.to(device))
|
|
|
|
embeddings_a = torch_xla.core.functions.all_gather(embeddings_a)
|
|
embeddings_b1 = torch_xla.core.functions.all_gather(embeddings_b1)
|
|
embeddings_b2 = torch_xla.core.functions.all_gather(embeddings_b2)
|
|
|
|
embeddings_b = torch.cat([embeddings_b1, embeddings_b2])
|
|
|
|
|
|
scores = torch.mm(embeddings_a, embeddings_b.transpose(0, 1)) * args.scale
|
|
|
|
|
|
labels = torch.tensor(range(len(scores)), dtype=torch.long, device=embeddings_a.device)
|
|
|
|
|
|
loss = cross_entropy_loss(scores, labels)
|
|
|
|
|
|
|
|
optimizer.zero_grad()
|
|
loss.backward()
|
|
torch.nn.utils.clip_grad_norm_(model.parameters(), max_grad_norm)
|
|
|
|
xm.optimizer_step(optimizer, barrier=True)
|
|
lr_scheduler.step()
|
|
|
|
|
|
|
|
if (global_step+1) % args.save_steps == 0:
|
|
output_path = os.path.join(args.output, str(global_step+1))
|
|
xm.master_print("save model: "+output_path)
|
|
model.save_pretrained(output_path)
|
|
|
|
|
|
output_path = os.path.join(args.output, "final")
|
|
xm.master_print("save model final: "+ output_path)
|
|
model.save_pretrained(output_path)
|
|
|
|
|
|
def produce_data(args, queue, filepaths, dataset_indices):
|
|
global_batch_size = args.batch_size*args.nprocs
|
|
size_per_dataset = int(global_batch_size / args.datasets_per_batch)
|
|
num_same_dataset = int(size_per_dataset / args.batch_size)
|
|
print("producer", "global_batch_size", global_batch_size)
|
|
print("producer", "size_per_dataset", size_per_dataset)
|
|
print("producer", "num_same_dataset", num_same_dataset)
|
|
|
|
datasets = []
|
|
for filepath in filepaths:
|
|
if "reddit_" in filepath:
|
|
data_obj = RedditDataset(filepath)
|
|
else:
|
|
data_obj = Dataset(filepath)
|
|
datasets.append(iter(data_obj))
|
|
|
|
|
|
num_cols = {idx: len(next(dataset)) for idx, dataset in enumerate(datasets)}
|
|
|
|
while True:
|
|
texts_in_batch = set()
|
|
batch_format = None
|
|
|
|
|
|
for _ in range(args.datasets_per_batch):
|
|
valid_dataset = False
|
|
while not valid_dataset:
|
|
data_idx = random.choice(dataset_indices)
|
|
if batch_format is None:
|
|
batch_format = num_cols[data_idx]
|
|
valid_dataset = True
|
|
else:
|
|
valid_dataset = (batch_format == num_cols[data_idx])
|
|
|
|
|
|
dataset = datasets[data_idx]
|
|
for _ in range(num_same_dataset):
|
|
for _ in range(args.nprocs):
|
|
batch_device = []
|
|
while len(batch_device) < args.batch_size:
|
|
sample = next(dataset)
|
|
in_batch = False
|
|
for text in sample:
|
|
if text in texts_in_batch:
|
|
in_batch = True
|
|
break
|
|
|
|
if not in_batch:
|
|
for text in sample:
|
|
texts_in_batch.add(text)
|
|
batch_device.append(sample)
|
|
|
|
queue.put(batch_device)
|
|
|
|
|
|
class RedditDataset:
|
|
"""
|
|
A class that handles the reddit data files
|
|
"""
|
|
def __init__(self, filepath):
|
|
self.filepath = filepath
|
|
|
|
def __iter__(self):
|
|
while True:
|
|
with gzip.open(self.filepath, "rt") as fIn:
|
|
for line in fIn:
|
|
data = json.loads(line)
|
|
|
|
if "response" in data and "context" in data:
|
|
yield [data["response"], data["context"]]
|
|
|
|
class Dataset:
|
|
"""
|
|
A class that handles one dataset
|
|
"""
|
|
def __init__(self, filepath):
|
|
self.filepath = filepath
|
|
|
|
def __iter__(self):
|
|
max_dataset_size = 10*1000*1000
|
|
dataset = []
|
|
data_format = None
|
|
|
|
while dataset is None or len(dataset) == 0:
|
|
with gzip.open(self.filepath, "rt") as fIn:
|
|
for line in fIn:
|
|
data = json.loads(line)
|
|
if isinstance(data, dict):
|
|
data = data['texts']
|
|
|
|
if data_format is None:
|
|
data_format = len(data)
|
|
|
|
|
|
assert len(data) == data_format
|
|
|
|
if dataset is not None:
|
|
dataset.append(data)
|
|
if len(dataset) >= max_dataset_size:
|
|
dataset = None
|
|
|
|
yield data
|
|
|
|
|
|
|
|
while True:
|
|
random.shuffle(dataset)
|
|
for data in dataset:
|
|
yield data
|
|
|
|
|
|
|
|
if __name__ == "__main__":
|
|
parser = argparse.ArgumentParser()
|
|
parser.add_argument('--model', default='nreimers/MiniLM-L6-H384-uncased')
|
|
parser.add_argument('--steps', type=int, default=2000)
|
|
parser.add_argument('--save_steps', type=int, default=10000)
|
|
parser.add_argument('--batch_size', type=int, default=64)
|
|
parser.add_argument('--max_length', type=int, default=128)
|
|
parser.add_argument('--nprocs', type=int, default=8)
|
|
parser.add_argument('--datasets_per_batch', type=int, default=2, help="Number of datasets per batch")
|
|
parser.add_argument('--scale', type=float, default=20, help="Use 20 for cossim, and 1 when you work with unnormalized embeddings with dot product")
|
|
parser.add_argument('--data_folder', default="/data", help="Folder with your dataset files")
|
|
parser.add_argument('data_config', help="A data_config.json file")
|
|
parser.add_argument('output')
|
|
args = parser.parse_args()
|
|
|
|
|
|
assert (args.batch_size*args.nprocs) % args.datasets_per_batch == 0
|
|
|
|
logging.info("Output: "+args.output)
|
|
if os.path.exists(args.output):
|
|
print("Output folder already exists.")
|
|
input("Continue?")
|
|
|
|
|
|
os.makedirs(args.output, exist_ok=True)
|
|
|
|
data_config_path = os.path.join(args.output, 'data_config.json')
|
|
copyfile(args.data_config, data_config_path)
|
|
copyfile(args.data_config, data_config_path)
|
|
copyfile(args.data_config, data_config_path)
|
|
|
|
train_script_path = os.path.join(args.output, 'train_script.py')
|
|
copyfile(__file__, train_script_path)
|
|
with open(train_script_path, 'a') as fOut:
|
|
fOut.write("\n\n# Script was called via:\n#python " + " ".join(sys.argv))
|
|
|
|
|
|
|
|
|
|
with open(args.data_config) as fIn:
|
|
data_config = json.load(fIn)
|
|
|
|
queue = mp.Queue(maxsize=100*args.nprocs)
|
|
|
|
filepaths = []
|
|
dataset_indices = []
|
|
for idx, data in enumerate(data_config):
|
|
filepaths.append(os.path.join(os.path.expanduser(args.data_folder), data['name']))
|
|
dataset_indices.extend([idx]*data['weight'])
|
|
|
|
|
|
p = mp.Process(target=produce_data, args=(args, queue, filepaths, dataset_indices))
|
|
p.start()
|
|
|
|
|
|
print("Start processes:", args.nprocs)
|
|
xmp.spawn(train_function, args=(args, queue), nprocs=args.nprocs, start_method='fork')
|
|
print("Training done")
|
|
print("It might be that not all processes exit automatically. In that case you must manually kill this process.")
|
|
print("With 'pkill python' you can kill all remaining python processes")
|
|
p.kill()
|
|
exit()
|
|
|
|
|
|
|
|
|
|
|