Dans Personality Engine V1.1.0 12b
PocketDocIntroduction
Dans-PersonalityEngine-V1.1.0-12B is a versatile model designed for tasks such as text generation, roleplay, co-writing, sentiment analysis, and summarization. It supports a broad range of applications, making it suitable for both casual and professional use.
Architecture
- Base Model: mistralai/Mistral-Nemo-Base-2407
- Language: English
- Context Length: 32768 tokens
- Library: Transformers
- License: Apache-2.0
Training
The model was fully finetuned for two epochs using a single H200 SXM GPU over 88 hours. It was built leveraging the Axolotl framework, indicating a robust training process.
Guide: Running Locally
Basic Steps
- Install Dependencies: Ensure you have Python and the required libraries, such as
transformers
andsafetensors
. - Download the Model: Acquire the model files from Hugging Face's repository.
- Load the Model: Utilize the
transformers
library to load the model. - Run Inference: Execute your own text generation tasks using the model's capabilities.
Cloud GPUs
For optimal performance, particularly when handling large datasets or extensive text generation tasks, consider using cloud-based GPUs such as AWS EC2, Google Cloud GPUs, or Azure's GPU offerings.
License
The model is available under the Apache-2.0 License, allowing for personal and commercial use while ensuring attribution and license inclusion in distributions.