Update xtuner_config.py
Browse files- xtuner_config.py +4 -4
xtuner_config.py
CHANGED
@@ -19,7 +19,7 @@ custom_hooks = [
|
|
19 |
'',
|
20 |
],
|
21 |
every_n_iters=500,
|
22 |
-
|
23 |
pretrained_model_name_or_path='openai/clip-vit-large-patch14-336',
|
24 |
trust_remote_code=True,
|
25 |
type='transformers.CLIPImageProcessor.from_pretrained'),
|
@@ -58,7 +58,7 @@ llava_dataset = dict(
|
|
58 |
image_folder='./data/llava_data/LLaVA-Pretrain/images',
|
59 |
max_length=1472,
|
60 |
pad_image_to_square=False,
|
61 |
-
|
62 |
pretrained_model_name_or_path='openai/clip-vit-large-patch14-336',
|
63 |
trust_remote_code=True,
|
64 |
type='transformers.CLIPImageProcessor.from_pretrained'),
|
@@ -126,7 +126,7 @@ param_scheduler = [
|
|
126 |
eta_min=0.0,
|
127 |
type='mmengine.optim.CosineAnnealingLR'),
|
128 |
]
|
129 |
-
|
130 |
pretrained_model_name_or_path='openai/clip-vit-large-patch14-336',
|
131 |
trust_remote_code=True,
|
132 |
type='transformers.CLIPImageProcessor.from_pretrained')
|
@@ -165,7 +165,7 @@ train_dataloader = dict(
|
|
165 |
image_folder='./data/llava_data/LLaVA-Pretrain/images',
|
166 |
max_length=1472,
|
167 |
pad_image_to_square=False,
|
168 |
-
|
169 |
pretrained_model_name_or_path='openai/clip-vit-large-patch14-336',
|
170 |
trust_remote_code=True,
|
171 |
type='transformers.CLIPImageProcessor.from_pretrained'),
|
|
|
19 |
'',
|
20 |
],
|
21 |
every_n_iters=500,
|
22 |
+
image_processor=dict(
|
23 |
pretrained_model_name_or_path='openai/clip-vit-large-patch14-336',
|
24 |
trust_remote_code=True,
|
25 |
type='transformers.CLIPImageProcessor.from_pretrained'),
|
|
|
58 |
image_folder='./data/llava_data/LLaVA-Pretrain/images',
|
59 |
max_length=1472,
|
60 |
pad_image_to_square=False,
|
61 |
+
image_processor=dict(
|
62 |
pretrained_model_name_or_path='openai/clip-vit-large-patch14-336',
|
63 |
trust_remote_code=True,
|
64 |
type='transformers.CLIPImageProcessor.from_pretrained'),
|
|
|
126 |
eta_min=0.0,
|
127 |
type='mmengine.optim.CosineAnnealingLR'),
|
128 |
]
|
129 |
+
image_processor = dict(
|
130 |
pretrained_model_name_or_path='openai/clip-vit-large-patch14-336',
|
131 |
trust_remote_code=True,
|
132 |
type='transformers.CLIPImageProcessor.from_pretrained')
|
|
|
165 |
image_folder='./data/llava_data/LLaVA-Pretrain/images',
|
166 |
max_length=1472,
|
167 |
pad_image_to_square=False,
|
168 |
+
image_processor=dict(
|
169 |
pretrained_model_name_or_path='openai/clip-vit-large-patch14-336',
|
170 |
trust_remote_code=True,
|
171 |
type='transformers.CLIPImageProcessor.from_pretrained'),
|