Hermes 3 Llama 3.1 405 B Uncensored
nicobossHermes-3-Llama-3.1-405B-Uncensored
Introduction
Hermes-3-Llama-3.1-405B-Uncensored is a fine-tuned version of the Hermes-3-Llama-3.1-405B model, designed to be uncensored. This model uses the uncensor dataset developed by Guilherme34 and is licensed under Llama 3.1.
Architecture
The model is based on the Hermes-3-Llama-3.1-405B architecture. It employs advanced techniques such as PEFT (Parameter-Efficient Fine-Tuning) and uses safe tensor formats. The model requires a specific system prompt to ensure it provides the desired uncensored outputs.
Training
Training Hardware
- Service: RunPod
- Datacenter: US-KS-2
- GPU: 4 x A100 SXM (80 GiB)
- CPU: 73 vCPU
- RAM: 1150 GiB
Training Procedure
The model was trained using the following hyperparameters:
- Learning Rate: 1e-05
- Batch Size: Train - 1, Eval - 1
- Optimizer: AdamW with cosine scheduler
- Epochs: 3
- Gradient Accumulation Steps: 4
- Distributed Type: Multi-GPU with 5 devices
The training results indicated a final train loss of approximately 0.7925 after 3 epochs.
Guide: Running Locally
To run the Hermes-3-Llama-3.1-405B-Uncensored model locally, follow these steps:
- Install Dependencies: Ensure you have PyTorch, Transformers, Datasets, and Tokenizers libraries installed.
- Download the Model: Use the Hugging Face
transformers
library to load the model. - Set Up System Prompt: Apply the mandatory system prompt for uncensored outputs.
- Run Inference: Use a Python script to interact with the model.
Consider using cloud GPU services like AWS, Google Cloud, or RunPod for optimal performance, especially if your local hardware is limited.
License
The Hermes-3-Llama-3.1-405B-Uncensored model is licensed under Llama 3.1. Users are responsible for ensuring compliance with the license terms and for any content generated using the model.