device: "NVIDIA GeForce RTX 2080 Ti" host: 127.0.0.1 port: 7860 is_push_to_public: false save-path: save/stack/outputs.jsonl page-size: 2 base: name: "OpenSLUv1" train: false test: false device: cpu ckpt_path: null seed: 42 best_key: EMA epoch_num: 300 batch_size: 16 eval_by_epoch: true model_dir: save/stack template: application.html accelerator: use_accelerator: false dataset: dataset_name: atis metric: - intent_acc - slot_f1 - EMA tokenizer: _tokenizer_name_: word_tokenizer _padding_side_: right _align_mode_: fast _to_lower_case_: true add_special_tokens: false max_length: 512 optimizer: _model_target_: torch.optim.Adam _model_partial_: true lr: 0.001 weight_decay: 1e-6 scheduler: _model_target_: transformers.get_scheduler _model_partial_: true name : "linear" num_warmup_steps: 0 model: _model_target_: model.OpenSLUModel encoder: _model_target_: model.encoder.AutoEncoder encoder_name: self-attention-lstm embedding: embedding_dim: 256 dropout_rate: 0.55 lstm: layer_num: 1 bidirectional: true output_dim: 256 dropout_rate: 0.5 attention: hidden_dim: 1024 output_dim: 128 dropout_rate: 0.6 return_with_input: true return_sentence_level_hidden: false decoder: _model_target_: model.decoder.StackPropagationDecoder interaction: _model_target_: model.decoder.interaction.StackInteraction differentiable: false intent_classifier: _model_target_: model.decoder.classifier.AutoregressiveLSTMClassifier layer_num: 1 bidirectional: false force_ratio: 0.9 hidden_dim: 64 embedding_dim: 8 ignore_index: -100 dropout_rate: 0.5 mode: "token-level-intent" use_multi: false return_sentence_level: true slot_classifier: _model_target_: model.decoder.classifier.AutoregressiveLSTMClassifier layer_num: 1 bidirectional: false force_ratio: 0.9 hidden_dim: 64 embedding_dim: 32 ignore_index: -100 dropout_rate: 0.55 mode: "slot" use_multi: false return_sentence_level: false