Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Buckets new
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

TheBloke
/
Llama-2-70B-GPTQ

Text Generation
Transformers
Safetensors
PyTorch
English
llama
facebook
meta
llama-2
text-generation-inference
4-bit precision
gptq
Model card Files Files and versions
xet
Community
14
New discussion
Resources
  • PR & discussions documentation
  • Code of Conduct
  • Hub documentation

Fine-tuning?

#14 opened over 2 years ago by
OSK-Creative-Tech

The model is not responding.

#13 opened over 2 years ago by
PhelixZhen

model responses not good.

1
#12 opened over 2 years ago by
muneerhanif7

how to quant llama2 70b model with AutoGPTQ

4
#11 opened over 2 years ago by
zhaohb

Wrong shape when loading with Peft-AutoGPTQ

2
#10 opened over 2 years ago by
tridungduong16

Long waiting time

14
#9 opened over 2 years ago by
wempoo

Context Length Differences

👍 2
#7 opened over 2 years ago by
zacharyrs

Problems with temperature when using with python code.

3
#6 opened over 2 years ago by
matchaslime

Should we expect GGML soon?

3
#5 opened over 2 years ago by
yehiaserag

Issue with 64g version?

#4 opened over 2 years ago by
AARon99

The `main` branch for TheBloke/Llama-2-70B-GPTQ appears borked

👍 1
11
#3 opened over 2 years ago by
Aivean

I found an fp16 model if it helps

1
#2 opened over 2 years ago by
rombodawg

❤️❤️❤️❤️

1
#1 opened over 2 years ago by
SinanAkkoyun
Company
TOS Privacy About Careers
Website
Models Datasets Spaces Pricing Docs