Captain Eris Diogenes_ Twilight V0.420 12 B G G U F
QuantFactoryIntroduction
The Captain-Eris-Diogenes_Twilight-V0.420-12B-GGUF model is a quantized version of the original model developed by Nitral-AI. It is created using llama.cpp
and is available on the Hugging Face platform. This model combines features from two base models: Nitral-AI/Captain-Eris_Twilight-V0.420-12B and Nitral-AI/Diogenes-12B-ChatMLified.
Architecture
The model architecture includes two main components:
- Nitral-AI/Captain-Eris_Twilight-V0.420-12B
- Nitral-AI/Diogenes-12B-ChatMLified
These components are merged using a spherical linear interpolation (slerp) method. The YAML configuration specifies that both models contribute equally across a layer range of [0, 40].
Training
The training involves the following parameters:
- slices: Model layer ranges for merging.
- merge_method: slerp.
- parameters: Include specific filter values for
self_attn
andmlp
, and a constantt
value set to 0.420. - dtype: bfloat16 for reduced precision and memory usage.
Guide: Running Locally
To run this model locally, follow these steps:
- Clone the Repository: Download the model files from Hugging Face.
- Install Dependencies: Use
pip
to install the required libraries, includingtransformers
. - Load the Model: Utilize the Hugging Face
transformers
library to load the model. - Run Inference: Use the model to generate predictions or perform the specific task it was trained for.
Suggested Cloud GPUs
For optimal performance, consider using cloud-based GPUs such as NVIDIA A100 or V100 available on platforms like AWS, Google Cloud, or Azure.
License
The model is provided under a license that permits its use and modification. Users should refer to the Hugging Face model card for specific licensing details and adhere to any restrictions or obligations outlined.