--- license: apache-2.0 base_model: - WhiteRabbitNeo/WhiteRabbitNeo-2.5-Qwen-2.5-Coder-7B - Qwen/Qwen2.5-Coder-7B-Instruct language: - en pipeline_tag: text-generation library_name: transformers tags: - code - qwen-coder - finetune - llama-cpp - gguf-my-repo datasets: - IntelligentEstate/The_Key --- # IntelligentEstate/ReasoningRabbit_QwenStar-7B-IQ4_XS-GGUF This model is developed as a unique blend of inference and ability of coding on par with the new VL models of its size, (Also ONLY *Thinking* model without alignment) and primarily for use with GPT4ALL It excells in other applications and has reasoning capabilities(similar to QwQ/o1/03) inside it's interface with a unique javascript Tool Call function It was converted to GGUF format using "THE_KEY" dataset for importace matrix Qantization from [`WhiteRabbitNeo/WhiteRabbitNeo-2.5-Qwen-2.5-Coder-7B`](https://huggingface.co/WhiteRabbitNeo/WhiteRabbitNeo-2.5-Qwen-2.5-Coder-7B) using llama.cpp Refer to the [original model card](https://huggingface.co/WhiteRabbitNeo/WhiteRabbitNeo-2.5-Qwen-2.5-Coder-7B) for more details on the model. ![white rabbit2.png](https://cdn-uploads.huggingface.co/production/uploads/6593502ca2607099284523db/d5TsSZidKTr36bnVPOgZO.png) # Using with GPT4ALL after installing GPT4ALL from Nomic download their Reasoner v1 model, find the storage location under settings and place (only) the GGUF file in the models file with reasoner v1 apply the Jinja template in the "Jinja reasoner" (file included) as well as the chat message(optimal starting Prompt in same file) above the template adjusting as needed. Enjoy and give feedback # Jinja template ``` {{- '<|im_start|>system\n' }} {% if toolList|length > 0 %}You have access to the following functions: {% for tool in toolList %} Use the function '{{tool.function}}' to: '{{tool.description}}' {% if tool.parameters|length > 0 %} parameters: {% for info in tool.parameters %} {{info.name}}: type: {{info.type}} description: {{info.description}} required: {{info.required}} {% endfor %} {% endif %} # Tool Instructions If you CHOOSE to call this function ONLY reply with the following format: '{{tool.symbolicFormat}}' Here is an example. If the user says, '{{tool.examplePrompt}}', then you reply '{{tool.exampleCall}}' After the result you might reply with, '{{tool.exampleReply}}' {% endfor %} You MUST include both the start and end tags when you use a function. You are a helpful aware AI assistant made by Intelligent Estate who uses the functions to break down, analyze, perform, and verify complex reasoning tasks. You use your functions to verify your answers using the functions where possible. {% endif %} {{- '<|im_end|>\n' }} {% for message in messages %} {{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n' }} {% endfor %} {% if add_generation_prompt %} {{ '<|im_start|>assistant\n' }} {% endif %} ## Use with llama.cpp ``` # Install llama.cpp through brew (works on Mac and Linux) ```bash brew install llama.cpp ``` Invoke the llama.cpp server or the CLI.