Skip to content

Trying to convert LLama 3.2 to GGUF #11038

Answered by danbev
TByte007 asked this question in Q&A
Discussion options

You must be logged in to vote

I'm not able to reproduce this using the latest master:

$ git clone https://huggingface.co/meta-llama/Llama-3.2-1B

The from the llama.cpp repository:

(venv) python ./convert_hf_to_gguf.py --verbose /path/to/Llama-3.2-1B --outfile models/llama-3-2-1b-f16.gguf --outtype f16
output
INFO:hf-to-gguf:Loading model: Llama-3.2-1B
INFO:gguf.gguf_writer:gguf: This GGUF file is for Little Endian only
INFO:hf-to-gguf:Exporting model...
INFO:hf-to-gguf:rope_freqs.weight,           torch.float32 --> F32, shape = {32}
INFO:hf-to-gguf:gguf: loading model part 'model.safetensors'
INFO:hf-to-gguf:token_embd.weight,           torch.bfloat16 --> F16, shape = {2048, 128256}
INFO:hf-to-gguf:blk.0.attn_norm.wei…

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@TByte007
Comment options

Answer selected by TByte007
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants