| #!/bin/bash |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| set -x -e |
|
|
| source $six_ALL_CCFRWORK/start-py38-pt111 |
|
|
| echo "START TIME: $(date)" |
|
|
| |
| VARIANT="tr11f-6b3-ml-evalharness" |
|
|
|
|
| CHECKPOINT_PATH=$six_ALL_CCFRSCRATCH/checkpoints/tr13f-6B3-ml-t0/checkpoints/tasky/global_step1000 |
| MEGATRON_DEEPSPEED_REPO=/gpfsssd/worksf/projects/rech/six/commun/code/eval/Megatron-DeepSpeed |
| export HF_DATASETS_OFFLINE=1 |
| export TRANSFORMERS_OFFLINE=1 |
|
|
| export TRANSFORMERS_CACHE=$six_ALL_CCFRWORK/models |
| export HF_DATASETS_CACHE=$six_ALL_CCFRWORK/datasets |
| export HF_MODULES_CACHE=$six_ALL_CCFRWORK/modules |
| export HF_METRICS_CACHE=$six_ALL_CCFRWORK/metrics |
|
|
| cd $MEGATRON_DEEPSPEED_REPO |
|
|
| TOKENIZER_NAME_OR_PATH=bigscience-catalogue-data-dev/byte-level-bpe-tokenizer-no-norm-250k-whitespace-and-eos-regex-alpha-v3-dedup-lines-articles |
|
|
| PP_SIZE=1 |
| TP_SIZE=1 |
| SEQ_LEN=2048 |
|
|
| |
| |
| EVAL_MICRO_BATCH_SIZE=1 |
|
|
| |
| MEGATRON_REQUIRED_ARGS=" \ |
| --num-layers -1 \ |
| --hidden-size -1 \ |
| --num-attention-heads -1 \ |
| --seq-length -1 \ |
| --max-position-embeddings -1 \ |
| " |
|
|
|
|
| ZERO_STAGE=0 |
|
|
| config_json="./ds_config.json" |
|
|
| |
| cat <<EOT > $config_json |
| { |
| "train_micro_batch_size_per_gpu": 1, |
| "train_batch_size": 1, |
| "gradient_clipping": 1.0, |
| "zero_optimization": { |
| "stage": $ZERO_STAGE |
| }, |
| "bf16": { |
| "enabled": false |
| }, |
| "steps_per_print": 2000, |
| "wall_clock_breakdown": false |
| } |
| EOT |
|
|
|
|
| CMD="./tasks/eval_harness/evaluate.py \ |
| --load $CHECKPOINT_PATH \ |
| --results_path $VARIANT-results.json \ |
| --tensor-model-parallel-size $TP_SIZE \ |
| --pipeline-model-parallel-size $PP_SIZE \ |
| --tokenizer-type PretrainedFromHF \ |
| --tokenizer-name-or-path $TOKENIZER_NAME_OR_PATH \ |
| --micro-batch-size $EVAL_MICRO_BATCH_SIZE \ |
| --no-load-optim \ |
| --no-load-rng \ |
| --eval_fp32 \ |
| --inference \ |
| --seq-length $SEQ_LEN \ |
| --task_list arc_challenge,arc_easy,boolq,copa,headqa,hellaswag,lambada,logiqa,mathqa,mc_taco,mrpc,multirc,openbookqa,piqa,prost,pubmedqa,qnli,qqp,race,rte,sciq,sst,triviaqa,webqs,wic,winogrande,wnli,wsc \ |
| --deepspeed \ |
| --deepspeed_config ds_config.json \ |
| --intermed_results \ |
| --adaptive_seq_len \ |
| --micro_bs_multiplier 4 \ |
| $MEGATRON_REQUIRED_ARGS \ |
| " |
|
|
| GPUS_PER_NODE=1 |
| NNODES=$SLURM_NNODES |
| MASTER_ADDR=$(scontrol show hostnames $SLURM_JOB_NODELIST | head -n 1) |
| MASTER_PORT=6000 |
| export LAUNCHER="python -u -m torch.distributed.run \ |
| --nproc_per_node $GPUS_PER_NODE \ |
| --nnodes $NNODES \ |
| --rdzv_endpoint $MASTER_ADDR:$MASTER_PORT \ |
| --rdzv_backend c10d \ |
| --max_restarts 0 \ |
| --tee 3 \ |
| " |
|
|
| export CUDA_LAUNCH_BLOCKING=1 |
|
|
| echo $LAUNCHER $CMD |
|
|
| export PYTHONPATH=$MEGATRON_DEEPSPEED_REPO |
|
|
| $LAUNCHER $CMD 2>&1 | tee $VARIANT-eval-harness.log |
|
|