Spaces:
Build error
Build error
# @package _global_ | |
# specify here default configuration | |
# order of defaults determines the order in which configs override each other | |
defaults: | |
- _self_ | |
- data: mix_all.yaml | |
- trainer: ddp.yaml | |
- paths: default.yaml | |
- extras: default.yaml | |
- hydra: default.yaml | |
# experiment configs allow for version control of specific hyperparameters | |
# e.g. best hyperparameters for given model and datamodule | |
- experiment: null | |
- texture_exp: null | |
# optional local config for machine/user specific settings | |
# it's optional since it doesn't need to exist and is excluded from version control | |
- optional launcher: local.yaml | |
# - optional launcher: slurm.yaml | |
# debugging config (enable through command line, e.g. `python train.py debug=default) | |
- debug: null | |
# task name, determines output directory path | |
task_name: "train" | |
# tags to help you identify your experiments | |
# you can overwrite this in experiment configs | |
# overwrite from command line with `python train.py tags="[first_tag, second_tag]"` | |
# appending lists from command line is currently not supported :( | |
# https://github.com/facebookresearch/hydra/issues/1547 | |
tags: ["dev"] | |
# set False to skip model training | |
train: True | |
# evaluate on test set, using best model weights achieved during training | |
# lightning chooses best weights based on the metric specified in checkpoint callback | |
test: False | |
# simply provide checkpoint path to resume training | |
ckpt_path: null | |
# seed for random number generators in pytorch, numpy and python.random | |
seed: null | |