{ "architectures": [ "MistsForConditionalGeneration" ], "model_type": "mists", "pad_token_id": 32769, "projector_hidden_act": "gelu", "text_config": { "_name_or_path": "mistralai/Mistral-7B-Instruct-v0.3", "architectures": [ "MistralForCausalLM" ], "max_position_embeddings": 32768, "model_type": "mistral", "rms_norm_eps": 1e-05, "rope_theta": 1000000.0, "sliding_window": null, "torch_dtype": "bfloat16", "vocab_size": 32832 }, "time_series_config": { "model_name": "AutonLab/MOMENT-1-large", "model_type": "moment" }, "time_series_hidden_size": 1024, "time_series_token_index": 32768, "torch_dtype": "float32", "transformers_version": "4.41.2" }