--- library_name: diambra tags: - street-fighter-iii - deep-reinforcement-learning - reinforcement-learning - stable-baseline3 --- # Model Card for Model ID A DRL agent playing Street Fighter III trained using diambra ai library ## Codes Github repos(Give a star if found useful): * https://github.com/hishamcse/Advanced-DRL-Renegades-Game-Bots * https://github.com/hishamcse/DRL-Renegades-Game-Bots * https://github.com/hishamcse/Robo-Chess ## Model Details - **My Code for this model:** https://github.com/hishamcse/Advanced-DRL-Renegades-Game-Bots/tree/main/VI%20-%20Diambra_AI_Street-Fighter-III - **Tutorial:** https://github.com/alexpalms/deep-rl-class/blob/main/units/en/unitbonus3 - **Documentation:** https://docs.diambra.ai/ ## Training Details #### Training Hyperparameters ``` folders: parent_dir: "./results/" model_name: "sr6_128x4_das_nc" settings: game_id: "sfiii3n" step_ratio: 6 frame_shape: !!python/tuple [128, 128, 1] continue_game: 0.0 action_space: "discrete" characters: "Ken" difficulty: 6 outfits: 2 wrappers_settings: normalize_reward: true no_attack_buttons_combinations: true stack_frames: 4 dilation: 1 add_last_action: true stack_actions: 12 scale: true exclude_image_scaling: true role_relative: true flatten: true filter_keys: ["action", "own_health", "opp_health", "own_side", "opp_side", "opp_character", "stage", "timer"] policy_kwargs: #net_arch: [{ pi: [64, 64], vf: [32, 32] }] net_arch: [64, 64] ppo_settings: gamma: 0.94 model_checkpoint: "0" # 0: No checkpoint, 100000: Load checkpoint (if previously trained for 100000 steps) learning_rate: [2.5e-4, 2.5e-6] # To start clip_range: [0.15, 0.025] # To start #learning_rate: [5.0e-5, 2.5e-6] # Fine Tuning #clip_range: [0.075, 0.025] # Fine Tuning batch_size: 512 #8 #nminibatches gave different batch size depending on the number of environments: batch_size = (n_steps * n_envs) // nminibatches n_epochs: 4 n_steps: 512 autosave_freq: 10000 time_steps: 100000 ```