Update README.md
Browse files
README.md
CHANGED
|
@@ -76,4 +76,13 @@ model-index:
|
|
| 76 |
- type: accuracy
|
| 77 |
name: QA Average
|
| 78 |
value: 0.3780991480835666
|
| 79 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 76 |
- type: accuracy
|
| 77 |
name: QA Average
|
| 78 |
value: 0.3780991480835666
|
| 79 |
+
---
|
| 80 |
+
|
| 81 |
+
# TinyLlama_v1.1_1bit_BitDistiller
|
| 82 |
+
This is a 1-bit quantized version of TinyLlama v1.1, trained using BitDistiller with asymmetric quantization and self-distillation (CAKLD) to optimize accuracy retention under extreme compression. The model is fine-tuned on WikiText-2 and Alpaca-cleaned datasets and evaluated on multiple-choice QA benchmarks.
|
| 83 |
+
|
| 84 |
+
Key Features:
|
| 85 |
+
- 1-bit quantization for ultra-efficient inference.
|
| 86 |
+
- Asymmetric weight clipping to reduce precision loss.
|
| 87 |
+
- CAKLD knowledge distillation to preserve performance.
|
| 88 |
+
- Tested on ARC-Challenge, HellaSwag, PIQA, and Winogrande.
|