File size: 5,474 Bytes
fe8ecad |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 |
Quantization made by Richard Erkhov.
[Github](https://github.com/RichardErkhov)
[Discord](https://discord.gg/pvy7H8DZMG)
[Request more models](https://github.com/RichardErkhov/quant_request)
Leia-Swallow-7b - GGUF
- Model creator: https://huggingface.co/leia-llm/
- Original model: https://huggingface.co/leia-llm/Leia-Swallow-7b/
| Name | Quant method | Size |
| ---- | ---- | ---- |
| [Leia-Swallow-7b.Q2_K.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.Q2_K.gguf) | Q2_K | 2.41GB |
| [Leia-Swallow-7b.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.IQ3_XS.gguf) | IQ3_XS | 2.66GB |
| [Leia-Swallow-7b.IQ3_S.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.IQ3_S.gguf) | IQ3_S | 2.8GB |
| [Leia-Swallow-7b.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.Q3_K_S.gguf) | Q3_K_S | 2.8GB |
| [Leia-Swallow-7b.IQ3_M.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.IQ3_M.gguf) | IQ3_M | 2.95GB |
| [Leia-Swallow-7b.Q3_K.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.Q3_K.gguf) | Q3_K | 3.13GB |
| [Leia-Swallow-7b.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.Q3_K_M.gguf) | Q3_K_M | 3.13GB |
| [Leia-Swallow-7b.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.Q3_K_L.gguf) | Q3_K_L | 3.4GB |
| [Leia-Swallow-7b.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.IQ4_XS.gguf) | IQ4_XS | 3.45GB |
| [Leia-Swallow-7b.Q4_0.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.Q4_0.gguf) | Q4_0 | 3.62GB |
| [Leia-Swallow-7b.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.IQ4_NL.gguf) | IQ4_NL | 3.64GB |
| [Leia-Swallow-7b.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.Q4_K_S.gguf) | Q4_K_S | 3.65GB |
| [Leia-Swallow-7b.Q4_K.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.Q4_K.gguf) | Q4_K | 3.86GB |
| [Leia-Swallow-7b.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.Q4_K_M.gguf) | Q4_K_M | 3.86GB |
| [Leia-Swallow-7b.Q4_1.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.Q4_1.gguf) | Q4_1 | 4.01GB |
| [Leia-Swallow-7b.Q5_0.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.Q5_0.gguf) | Q5_0 | 4.4GB |
| [Leia-Swallow-7b.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.Q5_K_S.gguf) | Q5_K_S | 4.4GB |
| [Leia-Swallow-7b.Q5_K.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.Q5_K.gguf) | Q5_K | 4.52GB |
| [Leia-Swallow-7b.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.Q5_K_M.gguf) | Q5_K_M | 4.52GB |
| [Leia-Swallow-7b.Q5_1.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.Q5_1.gguf) | Q5_1 | 4.78GB |
| [Leia-Swallow-7b.Q6_K.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.Q6_K.gguf) | Q6_K | 5.22GB |
| [Leia-Swallow-7b.Q8_0.gguf](https://huggingface.co/RichardErkhov/leia-llm_-_Leia-Swallow-7b-gguf/blob/main/Leia-Swallow-7b.Q8_0.gguf) | Q8_0 | 6.76GB |
Original model description:
---
license: apache-2.0
language:
- ja
---
# Leia-Swallow-7B
LEIA is a training technique for autoregressive LLMs that effectively improves their performance in languages other than English by enhancing cross-lingual knowledge transfer from English to a target language.
This model is constructed by applying LEIA to Swallow, a Japanese-English bilingual LLM based on LLaMA 2.
The model achieves enhanced performance on six Japanese question-answering benchmarks, as reported below.
Please refer to our paper or blog post (in Japanese) for further technical details.
- [LEIA: Facilitating Cross-Lingual Knowledge Transfer in Language Models with Entity-based Data Augmentation](https://arxiv.org/abs/2402.11485) (arxiv.org)
- [LEIA: 言語間転移学習でLLMを賢くする新しい方法](#) (zenn.dev)
## Model List
- [Leia-Swallow-7b](https://huggingface.co/leia-llm/Leia-Swallow-7b/)
- [Leia-Swallow-13b](https://huggingface.co/leia-llm/Leia-Swallow-13b/)
## Empirical Results
The model is assessed using the following six question answering benchmarks:
- X-CODAH
- X-CSQA
- JCommonsenseQA
- NIILC
- JEMHopQA
- JAQKET v2
| Model | X-CODAH | X-CSQA | JCommonsenseQA | NIILC | JEMHopQA | JAQKET v2 |
| ---- | ---- | ---- | ---- | ---- | ---- | ---- |
| Swallow | 42.0 | 41.0 | 80.3 | 59.5 | 50.8 | 86.2 |
| LEIA | **42.7** | **42.4** | **80.6** | **60.3** | **54.7** | **86.5** |
For further details of this experiment, please refer to [our paper](https://arxiv.org/abs/2402.11485).
## Contributors
- Ikuya Yamada (Studio Ousia, RIKEN)
- Ryokan Ri (LY Corporation, SB Intuitions)
|