{ "activation_dropout": 0.1, "activation_function": "gelu", "architectures": [ "TimeSeriesTransformerForPrediction" ], "attention_dropout": 0.1, "cardinality": [ 137 ], "context_lenght": 120, "context_length": 60, "d_model": 32, "decoder_attention_heads": 2, "decoder_ffn_dim": 32, "decoder_layerdrop": 0.1, "decoder_layers": 4, "distribution_output": "student_t", "dropout": 0.1, "embedding_dimension": [ 2 ], "encoder_attention_heads": 2, "encoder_ffn_dim": 32, "encoder_layerdrop": 0.1, "encoder_layers": 4, "feature_size": 49, "init_std": 0.02, "input_size": 1, "is_encoder_decoder": true, "lags_sequence": [ 1, 2, 3, 4, 5, 6, 7, 8, 10, 11, 12, 13, 14, 16, 17, 18, 19, 20, 143, 144, 145, 287, 288, 289, 431, 432, 433, 575, 576, 577, 719, 720, 721, 863, 864, 865, 1007, 1008, 1009 ], "loss": "nll", "model_type": "time_series_transformer", "num_dynamic_real_features": 0, "num_parallel_samples": 100, "num_static_categorical_features": 1, "num_static_real_features": 0, "num_time_features": 6, "prediction_length": 60, "scaling": "mean", "torch_dtype": "float32", "transformers_version": "4.31.0", "use_cache": true }