I followed this manual and got llama-7b-hf-int4 (got llama-7b-4bit.pt ) and samwit/alpaca7B-lora (got adapter_model.bin). Now I want to merge them into a single .pt 4bit model. How to do such a thing?
Why I need this:
- current lama.cpp supports only legacy 4-bit single file models.
 - 4-bit fine-tuners generate small alpaca fine-tuned mini models.
 - only 4-bit alpaca tuning is available for my current setup; thus, I need to know how to apply/merge one into another.