Triangulum 10 B it
prithivMLmodsIntroduction
Triangulum-10B-it is a collection of pretrained and instruction-tuned generative models developed for multilingual applications. It is designed to handle complex reasoning tasks using synthetic datasets based on long chains of thought. Built upon LLaMA's autoregressive language model, it is enhanced with an optimized transformer architecture for superior performance.
Architecture
- Foundation Model: Based on LLaMA's autoregressive language model, using an optimized transformer architecture.
- Instruction Tuning: Includes supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF).
- Multilingual Support: Supports multiple languages for diverse linguistic contexts.
Training
- Synthetic Datasets: Uses long chain-of-thought synthetic data to improve reasoning abilities.
- Supervised Fine-Tuning (SFT): Aligns the model to specific tasks using curated datasets.
- Reinforcement Learning with Human Feedback (RLHF): Iteratively trains the model to adhere to human values and safety guidelines.
Guide: Running Locally
Example 1: Running the Triangulum-10B Model
-
Download the Model
Use the following command to download the model:ollama run triangulum-10b-f16.gguf
-
Model Initialization and Download
Ollama will initialize and download the model files. -
Interact with the Model
Send prompts to interact with the model. Example prompt:>>> What can you do for me?
-
Exit the Program
Type/exit
to leave the program.
Cloud GPUs
For enhanced performance, consider using cloud GPU providers like AWS, Google Cloud, or Azure to run the model efficiently.
License
Triangulum-10B-it is distributed under the creativeml-openrail-m license, allowing for open-source usage with certain restrictions.