Tiny Chat 1776 K

raincandy-u

Introduction

TINYCHAT-1776K is a lightweight language model developed by raincandy-u, specifically trained on the TinyChat dataset. The model aims to generate natural responses using a minimal architecture suitable for basic conversational scenarios. It is tailored to handle English at a level appropriate for three-year-old children and lacks broader world knowledge.

Architecture

The model is based on the llama architecture and is configured with the following specifications:

  • Hidden Size: 192
  • Intermediate Size: 640
  • Number of Attention Heads: 16
  • Number of Hidden Layers: 3
  • Number of Key Value Heads: 4
  • Vocabulary Size: 2048
  • Max Position Embeddings: 256
  • Tie Word Embeddings: True

Training

The TINYCHAT-1776K model is trained from scratch using conversations typical of young children, focusing on generating simple and coherent responses. Due to its limited scope, the model is not suitable for tasks requiring extensive knowledge or complex reasoning.

Guide: Running Locally

To run the TINYCHAT-1776K model locally, follow these steps:

  1. Clone the Repository: Download the model files from Hugging Face.
  2. Install Dependencies: Ensure you have the required libraries, such as transformers, installed in your environment.
  3. Load the Model: Use the AutoConfig and AutoModel classes to load and configure the model.
  4. Inference: Set generation parameters (e.g., top_k=40, top_p=0.8, temperature=1) and run the model to generate responses.

For better performance, consider using cloud GPUs from providers such as AWS, Google Cloud, or Azure.

License

The TINYCHAT-1776K model is licensed under the Apache-2.0 license, which allows for both personal and commercial use, modifications, and distribution.

More Related APIs in Text Generation