Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Crashes during finetuning #131

Open
gameveloster opened this issue Jul 4, 2023 · 2 comments
Open

Crashes during finetuning #131

gameveloster opened this issue Jul 4, 2023 · 2 comments

Comments

@gameveloster
Copy link

gameveloster commented Jul 4, 2023

I am trying to do a 30B finetune on 2x3090 using data parallel and the process will always crash before 3 steps is completed. I run the finetune script in a screen session on a remote computer, and the screen session is gone when i reestablish the SSH connection after the crash.

This is the command I use to start finetune

torchrun --nproc_per_node=2 --master_port=1234 finetune.py ./testdocs.txt \
    --ds_type=txt \
    --lora_out_dir=./loras/ \
    --llama_q4_config_dir=./models/Neko-Institute-of-Science_LLaMA-30B-4bit-128g \
    --llama_q4_model=./Neko-Institute-of-Science_LLaMA-30B-4bit-128g/llama-30b-4bit-128g.safetensors \
    --mbatch_size=1 \
    --batch_size=2 \
    --epochs=3 \
    --lr=3e-4 \
    --cutoff_len=256 \
    --lora_r=8 \
    --lora_alpha=16 \
    --lora_dropout=0.05 \
    --warmup_steps=5 \
    --save_steps=50 \
    --save_total_limit=3 \
    --logging_steps=5 \
    --groupsize=-1 \
    --xformers \
    --backend=cuda \
    --grad_chckpt

Has anyone else gotten the same crashing problem?

@ghost
Copy link

ghost commented Jul 6, 2023

probably because you ran out of VRAM. Try using a batch size of 1.

@johnsmith0031
Copy link
Owner

I think you should keep the ssh session, or running the task in background

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants