orca_mini_v9_5_3 B Instruct
pankajmathurIntroduction
Orca_Mini_v9_5_Llama-3.2-3B-Instruct is a model based on Llama-3.2-3B-Instruct, designed for versatile text generation tasks. It is trained with various SFT datasets, providing a foundation for further customization and enhancement according to user needs.
Architecture
The model is built on the Llama-3.2-3B-Instruct architecture, utilizing the Transformers library and supporting PyTorch and Safetensors. It is primarily intended for English text generation with conversational capabilities.
Training
The model is trained using a combination of human-generated and synthetic data to ensure safety and robustness. Safety fine-tuning and risk assessments are integral to its development, addressing critical areas such as CBRNE and child safety.
Guide: Running Locally
- Setup: Install the required libraries, such as
transformers
andtorch
. - Load the Model: Use the
transformers
pipeline for text generation. - Configuration: Choose quantization settings (e.g., 4-bit or 8-bit) using the
BitsAndBytesConfig
. - Execution: Run the model with input messages and obtain generated text.
Cloud GPUs
For optimal performance, consider using cloud-based GPUs like Google Colab's T4 GPU.
License
The model is released under the Llama 3.2 license, allowing use with proper credit and attribution. Users are encouraged to customize the model for specific applications.