M N 12b R P Ink
allura-orgIntroduction
MN-12B-RP-Ink is a roleplay-focused LoRA finetuned model based on Mistral Nemo Instruct. It adopts methodology and hyperparameters inspired by SorcererLM and Slush. The model is uniquely renamed to "Ink" to distinguish it from other roleplay models, using the same dataset as the Teleut RP model.
Architecture
MN-12B-RP-Ink is built as a LoRA finetune of the Mistral Nemo Instruct model. This architecture helps enhance the model's performance in roleplaying and conversational tasks.
Training
The model was trained with the following hyperparameters:
- Epochs: 2
- Learning Rate: 6e-5
- Learning Rate Scheduler: Cosine
- Optimizer: Paged AdamW 8bit
- Effective Batch Size: 12
For the LoRA component:
- Rank: 16
- Alpha: 32
- Dropout: 0.25
The dataset used for training is described humorously as a mix of unconventional data sources, emphasizing its eclectic nature.
Guide: Running Locally
To run MN-12B-RP-Ink locally, follow these basic steps:
- Clone the Repository: Ensure you have access to the model files by cloning the repository from Hugging Face.
- Install Dependencies: Make sure you have the necessary libraries installed, such as
transformers
andtorch
. - Set Up Environment: Prepare your environment to handle large models, ideally using a cloud GPU provider such as AWS, Google Cloud, or Azure.
- Load the Model: Use the Hugging Face Transformers library to load the model.
- Configure Settings: Use recommended chat templates and samplers such as:
- Temp 1.25 / MinP 0.1
- Temp 1.03 / TopK 200 / MinP 0.05 / TopA 0.2
License
MN-12B-RP-Ink is licensed under the Apache 2.0 License, allowing for wide usage and modification with compliance to the license terms.