clip-indonesian / hybrid_clip /run_training_backup.sh
Galuh Sahid
Add code
bb13925
raw
history blame contribute delete
955 Bytes
#!/bin/bash
SCRIPT_DIR=.
MODEL_DIR=~/models/training_v3_new
IMAGE_ENCODER="openai/clip-vit-base-patch32"
TEXT_ENCODER="indobenchmark/indobert-base-p2"
python ${SCRIPT_DIR}/run_hybrid_clip.py \
--output_dir ${MODEL_DIR} \
--overwrite_output_dir \
--tokenizer_name=${TEXT_ENCODER} \
--train_file="../data/train_dataset_v3.json" \
--validation_file="../data/val_dataset_v3.json" \
--do_train --do_eval \
--num_train_epochs="10" --max_seq_length 96 \
--per_device_train_batch_size="64" \
--per_device_eval_batch_size="64" \
--learning_rate="0.00005" --warmup_ratio 0.1 --weight_decay 0.0 \
--preprocessing_num_workers 16 \
--exp_name training_v3 \
--text_model_name_or_path=${TEXT_ENCODER} \
--vision_model_name_or_path=${IMAGE_ENCODER} \
--eval_steps 2500 \
--logging_steps 200 \
--save_steps 2500 \
--save_total_limit 5 \
--log_wandb \
--freeze_backbones
#--push_to_hub