Codepy Deepthink 3 B G G U F
QuantFactoryIntroduction
Codepy-Deepthink-3B-GGUF is a quantized version of the Codepy-Deepthink-3B model, designed for text generation tasks that require deep reasoning and logical structuring. It is fine-tuned from the meta-llama/Llama-3.2-3B-Instruct base model, making it suitable for applications such as education, programming, and creative writing.
Architecture
Llama 3.2 employs an auto-regressive language model with an optimized transformer architecture. It includes supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to ensure alignment with human preferences for helpfulness and safety.
Training
The Codepy-Deepthink-3B model is trained to handle complex queries by providing accurate and contextually relevant outputs. It excels in generating step-by-step solutions, creative content, and logical analyses, integrating an advanced understanding of both structured and unstructured data.
Guide: Running Locally
Download and Install
- Download Ollama: Visit https://ollama.com/download and install it on Windows or Mac.
- Create Model File: Create a model file, for example,
metallama
. - Add Template Command: Include a
FROM
line in the file to specify the base model:FROM Llama-3.2-1B.F16.gguf
- Create and Patch Model: Run the following command:
ollama create metallama -f ./metallama
- Verify Model: Ensure the model is listed with:
ollama list
Running the Model
- Run Model: Execute the model with:
ollama run metallama
Sample Usage
Interact with the model by issuing commands or queries.
Suggestion
For efficient running, consider using cloud GPUs provided by platforms like AWS, Google Cloud, or Azure.
License
The Codepy-Deepthink-3B-GGUF model is released under the CreativeML OpenRAIL-M license.