File tree Expand file tree Collapse file tree 2 files changed +8
-4
lines changed Expand file tree Collapse file tree 2 files changed +8
-4
lines changed Original file line number Diff line number Diff line change 2525
2626if [ $stage == " train" ] || [ $stage == " all" ]; then
2727 torchrun --standalone --nnodes=1 --nproc_per_node=$num_gpus west/bin/train.py \
28- --model_config_path conf/touch_asu_config.json \
28+ --model_config_or_dir conf/touch_asu_config.json \
2929 --data_path $data /train.jsonl \
3030 --output_dir $dir \
3131 --pack_size 8192 \
Original file line number Diff line number Diff line change 33# https://github.com/QwenLM/Qwen2/blob/main/examples/sft/finetune.py
44
55import logging
6+ import os
67import pathlib
78from dataclasses import dataclass , field
89from typing import Any , Union
1920@dataclass
2021class TrainingArguments (TrainingArguments ):
2122 optim : str = field (default = "adafactor" )
22- model_config_path : str = field (default = '' )
23+ model_config_or_dir : str = field (default = '' )
2324
2425
2526class MyTrainer (Trainer ):
@@ -105,8 +106,11 @@ def main():
105106 )
106107 parser = HfArgumentParser ((DataArguments , TrainingArguments ))
107108 data_args , training_args = parser .parse_args_into_dataclasses ()
108- config = AutoConfig .from_pretrained (training_args .model_config_path )
109- model = AutoModel .from_config (config )
109+ if os .path .isfile (training_args .model_config_or_dir ): # init from config
110+ config = AutoConfig .from_pretrained (training_args .model_config_or_dir )
111+ model = AutoModel .from_config (config )
112+ else : # load from pretrained
113+ model = AutoModel .from_pretrained (training_args .model_config_or_dir )
110114 tokenizer = model .init_tokenizer ()
111115 extractor = Extractor .get_class (model .model_type )(tokenizer )
112116
You can’t perform that action at this time.
0 commit comments