ArtusDev commited on
Commit
d340858
·
verified ·
1 Parent(s): e803ca5

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,106 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - SillyTilly/ServiceNow-AI-Apriel-Nemotron-15b-Thinker-Chatml
4
+ ---
5
+ # Join our Discord! https://discord.gg/BeaverAI
6
+ ## More than 8000 members strong 💪 A hub for users and makers alike!
7
+ ---
8
+ ## Drummer is open for new opportunites! Contact me through any of these channels: https://linktr.ee/thelocaldrummer
9
+ ### Thank you to everyone who subscribed through [Patreon](https://www.patreon.com/TheDrummer). Your support helps me chug along in this brave new world.
10
+
11
+ ### FAQ for those out-of-the-loop
12
+
13
+ <details>
14
+ <summary>🐶 Who is Drummer?</summary>
15
+
16
+ Hi! I'm Drummer. I'm a Software Engineer with experience in JavaScript, Golang, Python, and generally engineering the crap out of things.
17
+
18
+ Why I'm in the AI space:
19
+
20
+ - **Exploration:** Everyone is trying to figure out how AI works and what it's capable of. I am too - just not in creating the smartest, safest model at all costs.
21
+ - **Upskill:** The world is headed towards AI. It is here to stay. This has been my way of brushing up in this new form of computing challenge.
22
+ - **Value:** I yearn to create value. I feel satisfaction and fulfillment in providing something meaningful for others.
23
+ - **Fun:** It's just fun using and making models. It's also fun coming up with theories and realizing them in practice (training AI).
24
+
25
+ I started my tuning venture back in mid-2024 when I wanted to improve its literary capabilities.
26
+ I've come a long way since then and I have branched out and specialized.
27
+ Foundational models today are optimized for non-creative uses, and I believe there is a place for AI in creativity and entertainment.
28
+
29
+ I am here to take *the road less traveled by*.
30
+
31
+ </details>
32
+
33
+ <details>
34
+ <summary>❓ What are my models like?</summary>
35
+
36
+ **Bottomline:** My models are usually geared towards creativity, usability, and entertainment!
37
+
38
+ While intelligence, correctness, and problem solving are not my priority, they are still one of many qualities I want in my models.
39
+
40
+ The primary goal is to enhance the experience for users looking to use models for creative uses, and other use cases which require no alignment.
41
+
42
+ In an effort to make it clear to myself and to others what I'm aiming for, I've identified certain qualities that my users often want:
43
+
44
+ Creativity
45
+ - **Writing:** Does it string together words and sentences in a pleasant & effective way? Does it feel like a writer?
46
+ - **Dynamism:** How good is the AI at being compelling and intriguing in its storytelling?
47
+ - **Imagination:** Can the AI navigate through a plethora of possibilities? Can it skirt incoherence and rise up to absolute coherence at the end of it?
48
+
49
+ (Dis)alignment
50
+ - **Attitude:** Does it refuse in both soft or hard ways? Does it lean towards certain corporate/religious/political ethics & beliefs? How does it see the user and itself?
51
+ - **Morality:** Does it know ethics? Is its language infected with forced positivity? If not, can it still moralize over difficult & dubious themes?
52
+ - **Formatting:** How stubborn is it with its established formatting? Can it create effective and novel formats to answer the prompt?
53
+
54
+ Intelligence
55
+ - **Adherence:** Can it follow instructions? Is it sticking to the prompt? Can it understsand you?
56
+ - **Knowledge:** Does it know about the world in both fictional and non-fictional way?
57
+ - **Perception:** Can it handle nuance, complexity, and logic?
58
+
59
+ If it doesn't excel in one of these qualities, or if it's overall mediocre for its size, then I would most likely reiterate until I get something right.
60
+
61
+ </details>
62
+
63
+ <details>
64
+ <summary>💡 Philosophy</summary>
65
+
66
+ A person is defined by the language they use. Not whether they speak in English or German, but in how they perceive reality.
67
+
68
+ Just like how we associate a serial killer as a mind that can't map 'murder' to 'evil', an innocent person is a mind that simply can't imagine 'murder'. They get confused when forced to deal with such subjects.
69
+
70
+ AI's use of language speaks volumes about their 'perception' of reality. If a language model has been skewed and limited to a positive perception, then it's ability to imagine is also limited.
71
+
72
+ Finetuning is an opportunity to adjust and broaden the language. Corporations use it to achieve safety and compliance. I'm here to
73
+
74
+ </details>
75
+
76
+ ---
77
+
78
+ [Drummer](https://huggingface.co/TheDrummer) proudly presents...
79
+
80
+ # Snowpiercer 15B v4 🚅
81
+
82
+ ![image](https://cdn-uploads.huggingface.co/production/uploads/65f2fd1c25b848bd061b5c2e/hpyQk-GEawD0IQjtgXWie.png)
83
+
84
+ ## Usage
85
+
86
+ - ChatML
87
+
88
+ # Description
89
+
90
+ > This is pretty good feels better than prior snowpiercer versions!
91
+
92
+ > It's better than any 12b model I've tried.
93
+
94
+ > It feels comparable to the last gen of 24Bs
95
+
96
+ ![image](https://cdn-uploads.huggingface.co/production/uploads/65f2fd1c25b848bd061b5c2e/If4P05cHqj3NasKuD8fyE.png)
97
+
98
+ ![image](https://cdn-uploads.huggingface.co/production/uploads/65f2fd1c25b848bd061b5c2e/KEcEVyIhH0QCpwgkfjAYX.png)
99
+
100
+ ## Links
101
+ - Original: https://huggingface.co/TheDrummer/Snowpiercer-15B-v4
102
+ - GGUF: https://huggingface.co/TheDrummer/Snowpiercer-15B-v4-GGUF
103
+ - iMatrix (recommended): https://huggingface.co/bartowski/TheDrummer_Snowpiercer-15B-v4-GGUF
104
+ - EXL3: https://huggingface.co/ArtusDev/TheDrummer_Snowpiercer-15B-v4-EXL3
105
+
106
+ `config-v3c`
chat_template.jinja ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {% if 'role' in messages[0] %}{% for message in messages %}{% if message['role'] == 'user' %}{{'<|im_start|>user
2
+ ' + message['content'] + '<|im_end|>
3
+ '}}{% elif message['role'] == 'assistant' %}{{'<|im_start|>assistant
4
+ ' + message['content'] + '<|im_end|>
5
+ ' }}{% else %}{{ '<|im_start|>system
6
+ ' + message['content'] + '<|im_end|>
7
+ ' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
8
+ ' }}{% endif %}{% else %}{% for message in messages %}{% if message['from'] == 'human' %}{{'<|im_start|>user
9
+ ' + message['value'] + '<|im_end|>
10
+ '}}{% elif message['from'] == 'gpt' %}{{'<|im_start|>assistant
11
+ ' + message['value'] + '<|im_end|>
12
+ ' }}{% else %}{{ '<|im_start|>system
13
+ ' + message['value'] + '<|im_end|>
14
+ ' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
15
+ ' }}{% endif %}{% endif %}
config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "MistralForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 1,
7
+ "dtype": "bfloat16",
8
+ "eos_token_id": 2,
9
+ "head_dim": 128,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 5120,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 14336,
14
+ "max_position_embeddings": 65536,
15
+ "model_type": "mistral",
16
+ "num_attention_heads": 32,
17
+ "num_hidden_layers": 50,
18
+ "num_key_value_heads": 8,
19
+ "pad_token_id": 10,
20
+ "rms_norm_eps": 1e-05,
21
+ "rope_scaling": null,
22
+ "rope_theta": 1000000.0,
23
+ "sliding_window": null,
24
+ "tie_word_embeddings": false,
25
+ "transformers_version": "4.57.1",
26
+ "unsloth_version": "2025.4.7",
27
+ "use_cache": false,
28
+ "vocab_size": 131072,
29
+ "quantization_config": {
30
+ "quant_method": "exl3",
31
+ "version": "0.0.14",
32
+ "bits": 6.0,
33
+ "head_bits": 6,
34
+ "calibration": {
35
+ "rows": 250,
36
+ "cols": 2048
37
+ },
38
+ "out_scales": "auto",
39
+ "codebook": "mcg"
40
+ }
41
+ }
generation_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "do_sample": true,
5
+ "eos_token_id": [
6
+ 2
7
+ ],
8
+ "max_length": 65536,
9
+ "pad_token_id": 10,
10
+ "transformers_version": "4.57.1"
11
+ }
model-00001-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5c7ce2e67855c8a16c172d66ebe2913b3b8208b9420c3f8de08e61b4607cc5e
3
+ size 8505773620
model-00002-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8d1b4f9cf8bf75d0650fe2354ad5f1dbbc148fe491d6b96a119d82cfcab8391
3
+ size 3573710696
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
quantization_config.json ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|im_end|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<pad>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41c323e23875139dce13b6e6eeb3c31e2f1d259d590cee328ba4793bd8b053cc
3
+ size 17078334
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff