| export CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 | |
| model_name_or_path="meta-llama/Llama-2-7b-chat-hf" | |
| # model_name_or_path="Yukang/LongAlpaca-7B-16k" | |
| # model_name_or_path="lmsys/longchat-7b-v1.5-32k" | |
| # model_name_or_path="syzymon/long_llama_code_7b_instruct" | |
| # super_tokenizer_name_or_path="/home/baaiks/ninglu/code/PluginTransformer/data/outputs/90k_0104+8/super_tokenizer" | |
| super_tokenizer_name_or_path="/home/baaiks/ninglu/code/PluginTransformer/data/outputs/90k_0104+12/super_tokenizer" | |
| # super_tokenizer_name_or_path="/home/baaiks/ninglu/code/PluginTransformer/data/outputs/90k_0111+8/super_tokenizer" | |
| # output_dir="data/results/lm/test_1" | |
| # mkdir -p ${output_dir} | |
| # python -m main.eval_lm \ | |
| # --model_name_or_path ${model_name_or_path} \ | |
| # --super_tokenizer_name_or_path ${super_tokenizer_name_or_path} \ | |
| # --super_tokenizer_num_hidden_layers 8 \ | |
| # --device_map "auto" \ | |
| # --dataset_list "pg19" \ | |
| # --output_dir ${output_dir} \ | |
| # --max_length 8192 \ | |
| # --target_length 512 \ | |
| # --compression_enable true \ | |
| # --compression_ratio 16 \ | |
| # | tee "${output_dir}/eval_lm.log" | |
| max_length_lst="4096 8192 16384 32768" | |
| compression_ratio_lst="16" | |
| for max_length in ${max_length_lst}; do | |
| for compression_ratio in ${compression_ratio_lst}; do | |
| output_dir="data/results/lm/pg19/llama-12-${compression_ratio}-${max_length}" | |
| mkdir -p ${output_dir} | |
| python -m main.eval_lm \ | |
| --model_name_or_path ${model_name_or_path} \ | |
| --super_tokenizer_name_or_path ${super_tokenizer_name_or_path} \ | |
| --super_tokenizer_num_hidden_layers 12 \ | |
| --device_map "auto" \ | |
| --dataset_list "pg19" \ | |
| --output_dir ${output_dir} \ | |
| --max_length ${max_length} \ | |
| --target_length 512 \ | |
| --compression_enable true \ | |
| --compression_ratio ${compression_ratio} \ | |
| --down_scale_method "uniform" \ | |
| | tee "${output_dir}/eval_lm.log" | |
| done | |
| done |