Update README.md
Browse files
README.md
CHANGED
|
@@ -27,7 +27,7 @@ tags:
|
|
| 27 |
|
| 28 |
A specialized 4B parameter model fine-tuned for function calling and tool usage, optimized for local deployment with llama-cpp-python.
|
| 29 |
|
| 30 |
-
##
|
| 31 |
|
| 32 |
- **4B Parameters** - Sweet spot for local deployment
|
| 33 |
- **Function Calling** - Fine-tuned on 60K function calling examples
|
|
@@ -36,7 +36,7 @@ A specialized 4B parameter model fine-tuned for function calling and tool usage,
|
|
| 36 |
- **Production Ready** - 0.518 training loss
|
| 37 |
- **262K Context** - Large context window for complex tasks
|
| 38 |
|
| 39 |
-
##
|
| 40 |
|
| 41 |
- **Base Model**: Qwen3-4B-Instruct-2507
|
| 42 |
- **Fine-tuning**: LoRA on Salesforce xlam-function-calling-60k dataset
|
|
@@ -44,7 +44,7 @@ A specialized 4B parameter model fine-tuned for function calling and tool usage,
|
|
| 44 |
- **Architecture**: Qwen3 with specialized tool calling tokens
|
| 45 |
- **License**: Apache 2.0
|
| 46 |
|
| 47 |
-
##
|
| 48 |
|
| 49 |
### Quick Install
|
| 50 |
|
|
@@ -93,7 +93,7 @@ CMAKE_ARGS="-DLLAMA_CUBLAS=on" pip install llama-cpp-python
|
|
| 93 |
CMAKE_ARGS="-DLLAMA_BLAS=on -DLLAMA_BLAS_VENDOR=OpenBLAS" pip install llama-cpp-python
|
| 94 |
```
|
| 95 |
|
| 96 |
-
##
|
| 97 |
|
| 98 |
### Option 1: Using the Run Script
|
| 99 |
|
|
@@ -181,7 +181,7 @@ tool_calls = extract_tool_calls(response_text)
|
|
| 181 |
print(f"Tool calls: {tool_calls}")
|
| 182 |
```
|
| 183 |
|
| 184 |
-
##
|
| 185 |
|
| 186 |
### 1. Weather Tool Calling
|
| 187 |
|
|
@@ -469,31 +469,11 @@ qwen3-4b-toolcall-llamacpp/
|
|
| 469 |
βββ .gitignore # Git ignore file
|
| 470 |
```
|
| 471 |
|
| 472 |
-
##
|
| 473 |
|
| 474 |
-
|
| 475 |
-
2. Create a feature branch
|
| 476 |
-
3. Make your changes
|
| 477 |
-
4. Add tests if applicable
|
| 478 |
-
5. Submit a pull request
|
| 479 |
|
| 480 |
-
##
|
| 481 |
-
|
| 482 |
-
This project is licensed under the Apache 2.0 License - see the [LICENSE](LICENSE) file for details.
|
| 483 |
-
|
| 484 |
-
## π Acknowledgments
|
| 485 |
-
|
| 486 |
-
- **Qwen Team** - For the base Qwen3-4B-Instruct model
|
| 487 |
-
- **Salesforce** - For the xlam-function-calling-60k dataset
|
| 488 |
-
- **llama.cpp** - For the efficient inference engine
|
| 489 |
-
- **Manojb** - For quantization and optimization
|
| 490 |
-
|
| 491 |
-
## π Support
|
| 492 |
-
|
| 493 |
-
- **Issues**: [GitHub Issues](https://github.com/yourusername/qwen3-4b-toolcall-llamacpp/issues)
|
| 494 |
-
- **Discussions**: [GitHub Discussions](https://github.com/yourusername/qwen3-4b-toolcall-llamacpp/discussions)
|
| 495 |
-
|
| 496 |
-
## π Related Projects
|
| 497 |
|
| 498 |
- [llama-cpp-python](https://github.com/abetlen/llama-cpp-python) - Python bindings for llama.cpp
|
| 499 |
- [Qwen3](https://huggingface.co/Qwen/Qwen3-4B-Instruct-2507) - Base model
|
|
|
|
| 27 |
|
| 28 |
A specialized 4B parameter model fine-tuned for function calling and tool usage, optimized for local deployment with llama-cpp-python.
|
| 29 |
|
| 30 |
+
## Features
|
| 31 |
|
| 32 |
- **4B Parameters** - Sweet spot for local deployment
|
| 33 |
- **Function Calling** - Fine-tuned on 60K function calling examples
|
|
|
|
| 36 |
- **Production Ready** - 0.518 training loss
|
| 37 |
- **262K Context** - Large context window for complex tasks
|
| 38 |
|
| 39 |
+
## Model Details
|
| 40 |
|
| 41 |
- **Base Model**: Qwen3-4B-Instruct-2507
|
| 42 |
- **Fine-tuning**: LoRA on Salesforce xlam-function-calling-60k dataset
|
|
|
|
| 44 |
- **Architecture**: Qwen3 with specialized tool calling tokens
|
| 45 |
- **License**: Apache 2.0
|
| 46 |
|
| 47 |
+
## Installation
|
| 48 |
|
| 49 |
### Quick Install
|
| 50 |
|
|
|
|
| 93 |
CMAKE_ARGS="-DLLAMA_BLAS=on -DLLAMA_BLAS_VENDOR=OpenBLAS" pip install llama-cpp-python
|
| 94 |
```
|
| 95 |
|
| 96 |
+
## Quick Start
|
| 97 |
|
| 98 |
### Option 1: Using the Run Script
|
| 99 |
|
|
|
|
| 181 |
print(f"Tool calls: {tool_calls}")
|
| 182 |
```
|
| 183 |
|
| 184 |
+
## Examples
|
| 185 |
|
| 186 |
### 1. Weather Tool Calling
|
| 187 |
|
|
|
|
| 469 |
βββ .gitignore # Git ignore file
|
| 470 |
```
|
| 471 |
|
| 472 |
+
## License
|
| 473 |
|
| 474 |
+
This project is licensed under the MIT License - see the [LICENSE](LICENSE) file for details.
|
|
|
|
|
|
|
|
|
|
|
|
|
| 475 |
|
| 476 |
+
## Related Projects
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 477 |
|
| 478 |
- [llama-cpp-python](https://github.com/abetlen/llama-cpp-python) - Python bindings for llama.cpp
|
| 479 |
- [Qwen3](https://huggingface.co/Qwen/Qwen3-4B-Instruct-2507) - Base model
|