Open
Description
$ ./build/bin/llama-quantize --allow-requantize --output-tensor-type f16 --token-embedding-type f16 gemma-3n-E2B-it.f16.gguf gemma-3n-E2B-it.q5_k.gguf q5_k 2
main: build = 5760 (e8215dbb)
main: built with cc (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 for x86_64-linux-gnu
main: quantizing 'gemma-3n-E2B-it.f16.gguf' to 'gemma-3n-E2B-it.q5_k.gguf' as Q5_K using 2 threads
llama_model_quantize: failed to quantize: tensor 'per_layer_token_embd.weight' data is not within the file bounds, model is corrupted or incomplete
main: failed to quantize model from 'gemma-3n-E2B-it.f16.gguf'
Metadata
Metadata
Assignees
Labels
No labels