Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

TheBloke
/
CodeLlama-34B-Instruct-GPTQ

Text Generation
Transformers
Safetensors
code
llama
llama-2
custom_code
text-generation-inference
4-bit precision
gptq
Model card Files Files and versions Community
9
New discussion
Resources
  • PR & discussions documentation
  • Code of Conduct
  • Hub documentation

Why does it say 4.98b params when the original model is 34b? Was that a typo?

#9 opened over 1 year ago by
lambdac

experiencing empty output if text input is long

#8 opened over 1 year ago by
lambdac

[AUTOMATED] Model Memory Requirements

#7 opened over 1 year ago by
model-sizer-bot

Running into issues when trying to run with TGI

1
#6 opened over 1 year ago by
viraniaman

main branch has problem using infill

#5 opened over 1 year ago by
jy00520336

Can I run this model on two NVIDIA RTX A5000 GPUs with 24 GB each?

3
#4 opened over 1 year ago by
nashid

Is the 34B llama2 actually GPTQ working?

4
#3 opened almost 2 years ago by
mzbac

Contradiction in model description

1
#2 opened almost 2 years ago by
m9e

Could you please specify which database was used for quantization finetuning?

2
#1 opened almost 2 years ago by
Badal
Company
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs