kbrd-gen-redial / config.json
ZhankuiHe's picture
Upload KBRDGen
7945934
raw
history blame contribute delete
No virus
2.74 kB
{
"WEIGHT_DIMENSIONS": {},
"_name_or_path": "/home/zhankui/0_ongoing/RecBot/src/recwizard/modules/kbrd/temp_kbrd_gen",
"architectures": [
"KBRDGen"
],
"attention_dropout": 0.0,
"dropout": 0.1,
"embeddings_scale": true,
"end_idx": 2,
"ffn_size": 300,
"gen_dim": 300,
"learn_positional_embeddings": false,
"n_heads": 2,
"n_layers": 2,
"n_positions": 1024,
"pad_idx": 0,
"rec_dim": 128,
"rec_module_config": {
"WEIGHT_DIMENSIONS": {
"edge_index.dtype": "torch.int64",
"edge_index.shape": [
2,
358184
],
"edge_type.dtype": "torch.int64",
"edge_type.shape": [
358184
],
"item_index.dtype": "torch.int64",
"item_index.shape": [
6924
]
},
"_name_or_path": "temp_kbrd_rec",
"add_cross_attention": false,
"architectures": [
"KBRDRec"
],
"bad_words_ids": null,
"begin_suppress_tokens": null,
"bos_token_id": null,
"chunk_size_feed_forward": 0,
"cross_attention_hidden_size": null,
"decoder_start_token_id": null,
"dim": 128,
"diversity_penalty": 0.0,
"do_sample": false,
"early_stopping": false,
"encoder_no_repeat_ngram_size": 0,
"eos_token_id": null,
"exponential_decay_length_penalty": null,
"finetuning_task": null,
"forced_bos_token_id": null,
"forced_eos_token_id": null,
"id2label": {
"0": "LABEL_0",
"1": "LABEL_1"
},
"is_decoder": false,
"is_encoder_decoder": false,
"label2id": {
"LABEL_0": 0,
"LABEL_1": 1
},
"length_penalty": 1.0,
"max_length": 20,
"min_length": 0,
"model_type": "",
"n_entity": 64368,
"n_relation": 214,
"no_repeat_ngram_size": 0,
"num_bases": 8,
"num_beam_groups": 1,
"num_beams": 1,
"num_return_sequences": 1,
"output_attentions": false,
"output_hidden_states": false,
"output_scores": false,
"pad_token_id": null,
"prefix": null,
"problem_type": null,
"pruned_heads": {},
"remove_invalid_values": false,
"repetition_penalty": 1.0,
"return_dict": true,
"return_dict_in_generate": false,
"sep_token_id": null,
"sub_n_relation": 35,
"suppress_tokens": null,
"task_specific_params": null,
"temperature": 1.0,
"tf_legacy_loss": false,
"tie_encoder_decoder": false,
"tie_word_embeddings": true,
"tokenizer_class": null,
"top_k": 50,
"top_p": 1.0,
"torch_dtype": "float32",
"torchscript": false,
"transformers_version": "4.33.3",
"typical_p": 1.0,
"use_bfloat16": false
},
"relu_dropout": 0.1,
"start_idx": 1,
"torch_dtype": "float32",
"transformers_version": "4.33.3",
"vocab_size": 15418
}