Sweaterdog commited on
Commit
842d863
·
verified ·
1 Parent(s): 42e8349

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +151 -3
README.md CHANGED
@@ -1,3 +1,151 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - Sweaterdog/Andy-3.5-MASSIVE
5
+ language:
6
+ - en
7
+ tags:
8
+ - Minecraft
9
+ - Mindcraft
10
+ - minecraft
11
+ - mindcraft
12
+ ---
13
+
14
+ # 🚀 Welcome to Next Generation Minecraft with Andy 3.6-small 🚀
15
+
16
+ ## Andy 3.6-small is a **LOCAL** model beating Andy-3.5 in performance
17
+ *Andy 3.6-small is designed to be used with MindCraft, and is not designed nor intended to be used for any other applications*
18
+
19
+
20
+ > # Please note! [!WARNING]
21
+ >
22
+ > Andy-3.6-small was trained on older data, and not the newest and latest versions of Mindcraft.
23
+ >
24
+ > I **cannot** guarantee that Andy-3.6-small will work on future versions as the model was tuned to play MindCraft with a specific version!
25
+ >
26
+ > For the rest of the Andy-3.6 generation, this model will **ONLY** be guaranteed to be supported on the version of Mindcraft in [this github repo!](https://github.com/Sweaterdog/Mindcraft-for-Andy-3.5)
27
+ >
28
+ > For more info, as well as the supported version of Mindcraft, please follow [this link to github](https://github.com/Sweaterdog/Mindcraft-for-Andy-3.5)
29
+
30
+ # How to Install / Setup
31
+
32
+ **Installing Andy-3.6-small is much easier and Andy-3.5!**
33
+
34
+ 1. In the top right of this repo, click "Use This Model"
35
+ 2. Next, click Ollama
36
+ 3. Pick your quantization *(Q5_k_m is best size to performance, Q8_0 is very good with similar performance to F16)*
37
+ 4. Run the command in your terminal
38
+ 5. Now you have Andy-3.6-small installed!
39
+
40
+ If you would like to use the full Andy-3.6 model, you can find that [here](https://huggingface.co/Sweaterdog/Andy-3.6)
41
+
42
+ # How was model trained?
43
+
44
+ The model was trained on the [MindCraft dataset](https://huggingface.co/datasets/Sweaterdog/Andy-3.5-MASSIVE) for Andy-3.6, a curated dataset for Q & A, reasoning, and playing, which includes ~22,000 prompts.
45
+
46
+ # What are capabilities and Limitations?
47
+
48
+ Andy-3.6-small was trained on EVERYTHING regarding Minecraft and MindCraft, it knows how to use commands natively without a system prompt.
49
+ Andy-3.6-small also knows how to build / use !newAction to perform commands, it was trained on lots of building, as well as, using !newAction to do tasks like manually making something or strip mining.
50
+
51
+ # What models can I choose?
52
+
53
+ There are going to be 2 model sizes avaliable, Regular, and Small
54
+ * Regular is a 7B parameter model, tuned from [Deepseek-R1 Distilled](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-7B)
55
+ * Small is a 3B parameter model, tuned from [Qwen2.5 3B](Qwen/Qwen2.5-3B-Instruct)
56
+
57
+ All models will have **case-by-case reasoning** baked **into** the model, meaning when it encounters a hard task, it will reason.
58
+
59
+ I have found through testing Andy-3.6-small needs extra time to decide what to do since it is a smaller model, but in the end it often generates a good response.
60
+
61
+ You can also *prompt* Andy-3.6-small to reason for better performance
62
+
63
+ # Safety and FAQ
64
+
65
+ Q: Is this model safe to use?
66
+
67
+ A. Yes, this model is non-volatile, and cannot generate malicous content
68
+
69
+ Q. Can this model be used on a server?
70
+
71
+ A. Yes, In theory and practice the model is only capable of building and performing manual tasks via newAction
72
+
73
+ Q. Who is responsible if this model does generate malicous content?
74
+
75
+ A. You are responsible, even though the model was never trained to be able to make malicous content, there is a ***very very slight chance*** it still generates malicous code.
76
+
77
+ Q. If I make media based on this model, like photos / videos, do I have to mention the Creator?
78
+
79
+ A. No, if you are making a post about MindCraft, and using this model, you only have to mention the creator if you mention the model being used.
80
+
81
+ # 🔥UPDATE🔥
82
+
83
+ ## **Andy-3.6-small Release!**
84
+ Andy-3.6-small is our latest model, capable of more reasoning than Andy-3.6, and half the size!
85
+
86
+ > # I want to thank all supporters! [!NOTE]
87
+ > I would love to thank everyone who supported this project, there is a list of supporters in the files section.
88
+ >
89
+ > You can find all of the supporters [here](https://huggingface.co/Sweaterdog/Andy-3.5/blob/main/Supporters.txt)
90
+
91
+ # Performance Metrics
92
+
93
+ These benchmarks are a-typical, since most standard benchmarks don't apply to Minecraft
94
+
95
+ The benchmarks below include models via API that are cheap, and other fine-tuned local models
96
+
97
+ ## Zero info Prompting
98
+ *How fast can a model collect 16 oak logs, and convert them all into sticks*
99
+
100
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/66960602f0ffd8e3a381106a/IEw1Gydg943qVSNGAL3RW.png)
101
+
102
+
103
+ As shown, the only models that are capable of play without information, is Andy-3.6, and all Andy-3.5 models
104
+
105
+ You can test this demo out for yourself using [this profile](https://huggingface.co/Sweaterdog/Andy-3.5/blob/main/local_demo.json)
106
+
107
+
108
+ ## Time to get a stone pickaxe
109
+
110
+
111
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/66960602f0ffd8e3a381106a/r3AoHHlmQuPdpt3WEFcyp.png)
112
+
113
+ ## *For Andy-3.6, I used the Q4_K_M quantization*
114
+
115
+ *For Andy-3.5-mini, I used the FP16 model, I had enough VRAM to do so*
116
+
117
+ *For Andy-3.5, I used the Q4_K_M quantization*
118
+
119
+ *For Andy-3.5-small, I used the Q8_0 quantization*
120
+
121
+ *Andy-3.5-reasoning-small was able to be the most efficient model producing the lowest amount of messages, but took a whopping 34.5 minutes to get a stone pickaxe.*
122
+
123
+ *For Andy-3.5-Teensy, I used the FP16 quantization*
124
+
125
+ *For Mineslayerv1 and Mineslayerv2, I used the default (and only) quantization, Q4_K_M*
126
+
127
+ ## Notes about the benchmarks
128
+
129
+ **Zero Info Prompting**
130
+
131
+ Andy-3.5-Mini collected 32 oak_log instead of 16 oak_log
132
+
133
+ Andy-3.5-small *No notes*
134
+
135
+ Andy-3.5 attempted to continue playing, and make a wooden_pickaxe after the goal was done.
136
+
137
+ Both Mineslayerv1 and Mineslayerv2 hallucinated commands, like !chop or !grab
138
+
139
+ **Time to get a stone pickaxe**
140
+
141
+ ## Andy-3.6 performed the best, beating gpt-4o-mini and claude-3.5-haiku
142
+
143
+ Andy-3.5-Mini was unable to make itself a stone pickaxe, however it collected enough wood, but then got stuck on converting logs to planks, it kept trying "!craftRecipe("wooden_planks", 6) instead of oak_planks
144
+
145
+ Andy-3.5-small kept trying to make a stone_pickaxe first
146
+
147
+ Andy-3.5 Made a stone pickaxe the faster than GPT-4o-mini and Claude-3.5-Haiku
148
+
149
+ Mineslayerv1 Was unable to use !collectBlocks, instead kept trying !collectBlock
150
+
151
+ Mineslayerv2 Was unable to play, it kept hallucinating on the first command