Deep Seek V3 G G U F
mradermacherIntroduction
DeepSeek-V3-GGUF is a quantized version of the DeepSeek-V3 model, optimized for performance and efficiency using the Transformers library. The quantization was performed by the user "mradermacher."
Architecture
The model is based on DeepSeek-V3, which utilizes the Transformers library. It is designed to handle English language tasks, providing different quantization options for varying needs.
Training
The model was originally developed by DeepSeek-AI, with quantization conducted by mradermacher. Static quantizations are currently available, with potential for more types if requested.
Guide: Running Locally
- Download Files: Obtain the required GGUF files from the provided links. Ensure you download all parts for the complete model.
- Setup Environment: Install the Transformers library if not already installed.
- Utilize the Model: Follow usage instructions from similar GGUF file guides, such as those by TheBloke, for handling and concatenating multi-part files.
- Hardware Suggestions: For optimal performance, consider using cloud GPUs from providers like AWS, Google Cloud, or Azure.
License
The usage terms for DeepSeek-V3-GGUF are governed by the same licensing as the base DeepSeek-V3 model, accessible here.