Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

TheBloke
/
Yarn-Llama-2-7B-128K-GPTQ

Text Generation
Transformers
Safetensors
llama
custom_code
text-generation-inference
4-bit precision
gptq
Model card Files Files and versions
xet
Community
2
Yarn-Llama-2-7B-128K-GPTQ
4.28 GB
  • 1 contributor
History: 43 commits
TheBloke's picture
TheBloke
Upload README.md
b08734c about 2 years ago
  • .gitattributes
    1.52 kB
    initial commit about 2 years ago
  • LICENSE.txt
    7.02 kB
    Initial GPTQ model commit about 2 years ago
  • Notice
    112 Bytes
    Initial GPTQ model commit about 2 years ago
  • README.md
    16.6 kB
    Upload README.md about 2 years ago
  • USE_POLICY.md
    4.77 kB
    Initial GPTQ model commit about 2 years ago
  • config.json
    1.35 kB
    Initial GPTQ model commit about 2 years ago
  • configuration_llama.py
    9.5 kB
    Initial GPTQ model commit about 2 years ago
  • generation_config.json
    189 Bytes
    Initial GPTQ model commit about 2 years ago
  • model.safetensors
    4.28 GB
    xet
    Initial GPTQ model commit about 2 years ago
  • modeling_llama_together_yarn.py
    54.5 kB
    Initial GPTQ model commit about 2 years ago
  • quantize_config.json
    185 Bytes
    Initial GPTQ model commit about 2 years ago
  • special_tokens_map.json
    435 Bytes
    Initial GPTQ model commit about 2 years ago
  • tokenizer.json
    1.84 MB
    Initial GPTQ model commit about 2 years ago
  • tokenizer.model
    500 kB
    xet
    Initial GPTQ model commit about 2 years ago
  • tokenizer_config.json
    746 Bytes
    Initial GPTQ model commit about 2 years ago