abnormalmapstudio commited on
Commit
c37c59a
·
0 Parent(s):

scaffold: initial model card + license + LFS

Browse files
Files changed (3) hide show
  1. .gitattributes +2 -0
  2. LICENSE +1 -0
  3. README.md +51 -0
.gitattributes ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
2
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
LICENSE ADDED
@@ -0,0 +1 @@
 
 
1
+ See Apache-2.0 license in the base model card and this repository. This derivative packaging uses Apache-2.0 (LICENSE).
README.md ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: mlx
4
+ pipeline_tag: text-generation
5
+ base_model: Qwen/Qwen3-Next-80B-A3B-Thinking
6
+ tags:
7
+ - mlx
8
+ - qwen3_next
9
+ - 8-bit
10
+ - affine
11
+ - text-generation
12
+ quantization_config:
13
+ bits: 8
14
+ mode: affine
15
+ group_size: 64
16
+ model-index:
17
+ - name: Qwen3-Next-80B-A3B-Thinking 8-bit (MLX)
18
+ results: []
19
+ ---
20
+
21
+ # Qwen3-Next-80B-A3B-Thinking — MLX 8-bit (affine)
22
+
23
+ Apple MLX-optimized 8-bit affine-quantized checkpoint of the base model
24
+ `Qwen/Qwen3-Next-80B-A3B-Thinking` for local inference on Apple Silicon.
25
+
26
+ Key details
27
+ - Format: MLX runtime, safetensors sharded weights
28
+ - Quantization: affine int8, group_size=64
29
+ - Task: text generation / chat
30
+ - Tokenizer: provided via `tokenizer.json` (BPE) with `chat_template.jinja`
31
+
32
+ ## Usage (MLX)
33
+ ```bash
34
+ pip install mlx-lm
35
+ ```
36
+
37
+ ```python
38
+ from mlx_lm import load, generate
39
+ repo_id = "abnormalmapstudio/Qwen3-Next-80B-A3B-Thinking-8bit-mlx"
40
+ model, tokenizer = load(repo_id)
41
+ out = generate(model, tokenizer, "List 5 creative dinner ideas.", max_tokens=200)
42
+ print(out)
43
+ ```
44
+
45
+ ## Benchmarks
46
+ - Will be added after upload completes; see `scripts/bench/qwen_mxfp4_vs_int4.py` and `scripts/bench/model_queue_eval.py`.
47
+
48
+ ## License
49
+ - Apache-2.0 for this packaging. See `LICENSE`.
50
+ - Base model license and terms apply (Qwen/Qwen3-Next-80B-A3B-Thinking).
51
+