Description
Name and Version
python3 convert_lora_to_gguf.py /models/CodeLlama-13b-finetuned --base /models/CodeLlama-13b-Instruct-hf --outfile /models/CodeLlama-13b.gguf --outtype f32
INFO:gguf.gguf_writer:Writing the following files:
INFO:gguf.gguf_writer:/models/CodeLlama-13b.gguf: n_tensors = 320, total_size = 104.9M
Writing: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 105M/105M [00:00<00:00, 210Mbyte/s]
INFO:lora-to-gguf:Model successfully exported to /models/CodeLlama-13b.gguf
root@7d293a7fbb1c:/app/llama.cpp# ./build/bin/llama-cli -m /models/CodeLlama-13b.gguf build: 5092 (d3bd719) with cc (Debian 12.2.0-14) 12.2.0 for x86_64-linux-gnu
main: llama backend init
main: load the model and apply lora adapter, if any
llama_model_loader: loaded meta data with 12 key-value pairs and 320 tensors from /models/CodeLlama-13b.gguf (version GGUF V3 (latest))
llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output.
llama_model_loader: - kv 0: general.architecture str = llama
llama_model_loader: - kv 1: general.type str = adapter
llama_model_loader: - kv 2: adapter.type str = lora
llama_model_loader: - kv 3: general.name str = CodeLlama 13b Finetuned
llama_model_loader: - kv 4: general.finetune str = finetuned
llama_model_loader: - kv 5: general.basename str = CodeLlama
llama_model_loader: - kv 6: general.size_label str = 13B
llama_model_loader: - kv 7: general.base_model.count u32 = 1
llama_model_loader: - kv 8: general.base_model.0.name str = models/CodeLlama 13b Instruct Hf
llama_model_loader: - kv 9: general.base_model.0.repo_url str = https://huggingface.co//models/CodeLl...
llama_model_loader: - kv 10: adapter.lora.alpha f32 = 32.000000
llama_model_loader: - kv 11: general.quantization_version u32 = 2
llama_model_loader: - type f32: 320 tensors
print_info: file format = GGUF V3 (latest)
print_info: file type = all F32 (guessed)
print_info: file size = 100.00 MiB (32.00 BPW)
llama_model_load: error loading model: error loading model hyperparameters: key not found in model: llama.context_length
llama_model_load_from_file_impl: failed to load model
common_init_from_params: failed to load model '/models/CodeLlama-13b.gguf'
main: error: unable to load model
Operating systems
Linux
GGML backends
CPU
Hardware
cpu node cloud
Models
CodeLlama-13b-Instruct-hf
Problem description & steps to reproduce
What is the issue here?
First Bad Commit
No response
Relevant log output
NA