{ "_name_or_path": "microsoft/Phi-3-mini-128k-instruct-onnx", "architectures": [ "PhiLongRoPEForCausalLM" ], "attention_dropout": 0.0, "xauto_map": { "AutoConfig": "configuration_phi_longrope.PhiLongRoPEConfig", "AutoModelForCausalLM": "modeling_phi_longrope.PhiLongRoPEForCausalLM" }, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 3072, "initializer_range": 0.02, "intermediate_size": 8192, "max_position_embeddings": 131072, "model_type": "phi3", "xmodel_type": "phi_longrope", "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 32, "original_max_position_embeddings": 4096, "rms_norm_eps": 1e-05, "rope_scaling": { "long_factor": [ 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 3.7391297817230225, 3.7699999999999756, 6.679999999999908, 11.956522941589355, 14.695652961730957, 16.649999999999697, 20.83000000000035, 22.91304588317871, 25.652175903320312, 28.391305923461914, 31.130435943603516, 33.869564056396484, 36.60869598388672, 39.34782409667969, 42.08695983886719, 44.826087951660156, 47.565216064453125, 50.30434799194336, 53.04347610473633, 55.78260803222656, 58.5217399597168, 61.26087188720703, 64.0, 64.0, 64.0, 64.0, 64.0, 64.0, 64.0, 64.0, 64.0, 64.0, 64.0, 64.0, 64.0, 64.0, 64.0, 64.0, 64.0 ], "short_factor": [ 1.0, 1.0, 1.0, 1.0, 1.05, 1.05, 1.1500000000000001, 1.2000000000000002, 1.2500000000000002, 1.5000000000000004, 1.8000000000000007, 2.25, 2.3000000000000007, 2.3500000000000014, 2.5, 2.5, 2.55, 2.5999999999999988, 2.6999999999999993, 2.6999999999999993, 2.8499999999999988, 2.8999999999999986, 2.9999999999999982, 2.9999999999999982, 2.9999999999999982, 3.099999999999998, 3.299999999999997, 3.399999999999997, 3.599999999999996, 3.649999999999996, 3.7999999999999954, 3.899999999999995, 3.899999999999995, 3.899999999999995, 4.099999999999994, 4.099999999999994, 4.099999999999994, 4.099999999999994, 4.099999999999994, 4.149999999999994, 4.149999999999994, 4.149999999999994, 4.149999999999994, 4.149999999999994, 4.199999999999994, 4.249999999999994, 4.299999999999994, 4.399999999999993 ], "type": "longrope" }, "rope_theta": 10000.0, "sliding_window": 131072, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "transformers_version": "4.38.1", "transformers.js_config": { "kv_cache_dtype": "float16" }, "use_cache": true, "vocab_size": 32038 }