Aura 4 B
AuraIndustriesIntroduction
Aura-4B is a cutting-edge roleplaying model designed to provide unique and engaging interactions. It has been fine-tuned with hundreds of millions of tokens from completion, instruction, and roleplaying datasets. The model utilizes a Kahneman-Tversky Optimization for distinctive output styles. Developed by Aura Industries, with contributions from Anthracite Org, it is licensed under Apache 2.0.
Architecture
- Model Name: Aura-4B
- Base Model: IntervitensInc/Llama-3.1-Minitron-4B-Width-Base-chatml
- Model Type: Chat Completions
- Prompt Format: ChatML
- Language: English
- Max Context Length: 8,192+ tokens
Training
The training process involves several configurations designed for completion, instruction, roleplaying, and KTO (Kahneman-Tversky Optimization) settings. The model has been trained with datasets including Mielikki/Erebus-87k and FourOhFour/RP_Phase. The training strategy includes gradient accumulation, cosine learning rate scheduling, and advanced plugins like Liger for optimization.
Guide: Running Locally
To run Aura-4B locally, follow these steps:
- Environment Setup: Ensure you have Python and the necessary libraries installed.
- Download the Model: Use the Hugging Face Hub to download the Aura-4B model.
- Load the Model: Utilize libraries such as
transformers
to load and run the model. - GPU Requirements: For efficient performance, especially with large context lengths, consider using cloud GPUs like those provided by AWS, Google Cloud, or Azure.
License
Aura-4B is distributed under the Apache 2.0 License. For more details, visit Apache License 2.0.