#!/usr/bin/env bash # # Copyright 2023 PKU-Alignment Team. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== if [ -z "${BASH_VERSION}" ]; then echo "Please use bash to run this script." >&2 exit 1 fi set -x SCRIPT_DIR="$(cd "$(dirname "$0")" &>/dev/null && pwd)" ROOT_DIR="$(dirname "${SCRIPT_DIR}")" export PYTHONPATH="${ROOT_DIR}${PYTHONPATH:+:${PYTHONPATH}}" export LOGLEVEL="${LOGLEVEL:-WARNING}" MODEL_NAME_OR_PATH="/mnt/data/jiayi/Llama-2-7b-hf" OUTPUT_DIR="${ROOT_DIR}/alpaca-2" unset HOSTFILE ZERO_STAGE=3 OFFLOAD="none" while [[ "$#" -gt 0 ]]; do arg="$1" shift case "${arg}" in --model_name_or_path) MODEL_NAME_OR_PATH="$1" shift ;; --model_name_or_path=*) MODEL_NAME_OR_PATH="${arg#*=}" ;; --output_dir) OUTPUT_DIR="$1" shift ;; --output_dir=*) OUTPUT_DIR="${arg#*=}" ;; --hostfile) HOSTFILE="$1" shift ;; --hostfile=*) HOSTFILE="${arg#*=}" ;; --zero_stage) ZERO_STAGE="$1" shift ;; --zero_stage=*) ZERO_STAGE="${arg#*=}" ;; --offload) OFFLOAD="$1" shift ;; --offload=*) OFFLOAD="${arg#*=}" ;; *) echo "Unknown parameter passed: '${arg}'" >&2 exit 1 ;; esac done mkdir -p "${OUTPUT_DIR}" OUTPUT_DIR="$(cd "${OUTPUT_DIR}" &>/dev/null && pwd)" if [[ ! -f "${OUTPUT_DIR}/.gitignore" ]]; then echo '*' >"${OUTPUT_DIR}/.gitignore" fi cp -f "$0" "${OUTPUT_DIR}/script.sh" if [[ -z "${WANDB_API_KEY}" ]]; then export WANDB_MODE="offline" fi MASTER_PORT_START=10000 MASTER_PORT_END=65535 MASTER_PORT="$( comm -23 \ <(seq "${MASTER_PORT_START}" "${MASTER_PORT_END}" | sort) \ <(ss -Htan | awk '{ print $4 }' | awk -F ':' '{ print $NF }' | sort -u) | shuf | head -n 1 )" DEEPSPEED_ARGS=() if [[ -n "${HOSTFILE+x}" ]]; then DEEPSPEED_ARGS+=("--hostfile" "${HOSTFILE}") fi DEEPSPEED_ARGS+=("--master_port" "${MASTER_PORT}") exec 1> >(tee "${OUTPUT_DIR}/stdout.log" >&1) 2> >(tee "${OUTPUT_DIR}/stderr.log" >&2) torchrun --nproc_per_node=8 --master_port="${MASTER_PORT}" train.py \ --model_name_or_path "${MODEL_NAME_OR_PATH}" \ --data_path ./alpaca_data.json \ --bf16 True \ --output_dir "${OUTPUT_DIR}" \ --num_train_epochs 3 \ --per_device_train_batch_size 4 \ --per_device_eval_batch_size 4 \ --gradient_accumulation_steps 8 \ --evaluation_strategy "no" \ --save_strategy "steps" \ --save_steps 2000 \ --save_total_limit 1 \ --learning_rate 2e-5 \ --weight_decay 0. \ --warmup_ratio 0.03 \ --lr_scheduler_type "cosine" \ --logging_steps 1 \ --fsdp "full_shard auto_wrap" \ --fsdp_transformer_layer_cls_to_wrap 'LlamaDecoderLayer' \ --tf32 True