Joe57005 commited on
Commit
ce3be51
·
verified ·
1 Parent(s): 29ede17

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +67 -1
README.md CHANGED
@@ -9,4 +9,70 @@ tags:
9
  - uncensored
10
  - function_calling
11
  - tool_use
12
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  - uncensored
10
  - function_calling
11
  - tool_use
12
+ ---
13
+ # Ramius
14
+
15
+ <!-- Provide a quick summary of what the model is/does. -->
16
+
17
+ This is Ramius, an uncensored function calling model.
18
+
19
+ ### Model Description
20
+
21
+ I needed an LLM for Home Assistant that is small and performant and I wanted one with some personality.
22
+ Qwen2.5-3B is small, fast and can call functions pretty well. But it's [REDACTED], and doesn't like to roleplay.
23
+ Arch-Function-3B is fantastic at calling functions, and absolutely nothing else.
24
+ Dolphin3.0-Qwen2.5-3b is great at roleplay and refuses to refuse anything. But it sucks at calling functions.
25
+
26
+ So I created Ramius with MergeKit to try and get the best of both.
27
+ Plus I'm GPU poor and can't train. (Intel ARC cards come with buyer's remorse at no extra charge!)
28
+
29
+ The result is... mediocre. It correctly calls functions most of the time, but it tends to hallucinate function responses instead of calling the actual function.
30
+ But it does stay in character. YMMV.
31
+
32
+ The name comes from Marko Ramius, a fictional communist submarine commander who defects to the United States in Tom Clancy's The Hunt for Red October.
33
+ He's a former communist and the name sounded cool.
34
+
35
+
36
+ - **Developed by:** Other people's hard work.
37
+ - **Funded by [optional]:** Also other people's hard work.
38
+ - **Shared by [optional]:** Me.
39
+ - **Model type:** Autoregressive transformer.
40
+ - **Language(s) (NLP):** English, and others, probably.
41
+ - **License:** [More Information Needed]
42
+
43
+ ### Model Sources [optional]
44
+
45
+ Created with MergeKit.
46
+
47
+ models:
48
+ - model: katanemo/Arch-Function-3B
49
+ lambda: 1.0
50
+ select_topk: 0.4
51
+ weight: 0.7
52
+ - model: cognitivecomputations/Dolphin3.0-Qwen2.5-3b
53
+ density: 1.0
54
+ lambda: 1.0
55
+ select_topk: 0.6
56
+ weight: 0.3
57
+ merge_method: sce
58
+ base_model: katanemo/Arch-Function-3B
59
+ parameters:
60
+ int8_mask: true
61
+ normalize: true
62
+ dtype: bfloat16
63
+
64
+
65
+ ## Bias, Risks, and Limitations
66
+
67
+ This is uncensored and does hallucinate. frequently.
68
+
69
+ [More Information Needed]
70
+
71
+ ### Recommendations
72
+
73
+ I use this with Ollama and Home Assistant via the Extended OpenAI conversation integration.
74
+ Works best with a top P of around 0.95 and temperature around 0.85.
75
+
76
+ I also recommend you DO NOT put your entity states in your system prompt, and instead write functions to get the information.
77
+ This will keep your system prompt static and more easily cached, which should reduce prompt processing time.
78
+