We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Got adapter_mode.bin,adapter_config.json after finetuning the vicuna 4bit 128g model , and checkpoints folders.
how to use this folders or files to infer the model?
The text was updated successfully, but these errors were encountered:
following
Sorry, something went wrong.
Use this
from monkeypatch.peft_tuners_lora_monkey_patch import replace_peft_model_with_gptq_lora_model replace_peft_model_with_gptq_lora_model() from peft import PeftModel from monkeypatch.peft_tuners_lora_monkey_patch import Linear4bitLt model = PeftModel.from_pretrained(model, lora_path, device_map={'': 'cpu'}, torch_dtype=torch.float32, is_trainable=True)
It's inside load_llama_model_4bit_low_ram_and_offload function
@johnsmith0031 what should be at the place of model, lora_path?
Yes, lora_path should be point to the path of finetuned lora model.
No branches or pull requests
Got adapter_mode.bin,adapter_config.json after finetuning the vicuna 4bit 128g model ,
and checkpoints folders.
how to use this folders or files to infer the model?
The text was updated successfully, but these errors were encountered: