Shisa V2
Collection
A family of bilingual JA/EN LLMs
•
32 items
•
Updated
•
7
This repo contains select GGUF quants of shisa-ai/shisa-v2-llama3.1-405b
b5503
of upstream llama.cppllama-gguf-split --merge
if you want to merge them back togetherType | Size (GB) |
---|---|
IQ2_XXS | 100 |
IQ3_XS | 155 |
IQ3_M | 170 |
IQ4_XS | 202 |
Q4_K_M | 227 |
Q8_0 | 402 |
Graph by ikawrakow comparing some lower-quality quant PPL (lower is better):
# first you need an fp16 - setup llama.cpp python env and run something like
python convert_hf_to_gguf.py ~/.cache/huggingface/hub/models--shisa-ai--shisa-v2-llama3.1-405b/snapshots/71b83a7cb998c3a44f59c83a9928596ac348b9b5 --outfile shisa-v2-llama3.1-405b-fp16.gguf
# Create imatrix: using 4 x H200 you can load 88 layers, takes about 1h15m
CUDA_VISIBLE_DEVICES=4,5,6,7 build/bin/llama-imatrix -m shisa-v2-llama3.1-405b-fp16.gguf -f /data/quantize/shisa-v2-llama-3.1-405b/gguf/calibration_chat.txt -o imatrix.dat -c 512 -b 512 --chunks 100 -ngl 88
# create your imatrix quants
build/bin/llama-quantize --imatrix imatrix.dat shisa-v2-llama3.1-405b-fp16.gguf shisa-v2-llama3.1-405b-IQ3_XS.gguf IQ3_XS
# split the quants
build/bin/llama-gguf-split --split-max-size 45G shisa-v2-llama3.1-405b-IQ3_XS.gguf shisa-v2-llama3.1-405b-IQ3_XS
# upload (bash loop)
for f in shisa-v2-llama3.1-405b-IQ3_XS-0000*; do huggingface-cli upload shisa-ai/shisa-v2-llama3.1-405b-GGUF "$f"; done
2-bit
3-bit
4-bit
8-bit
Base model
meta-llama/Llama-3.1-405B