--- license: apache-2.0 --- # Micro Llama v0 (Development) Micro Llama v0 is a lightweight and experimental version of the LlamaForCausalLM model designed for development and testing purposes. This repository contains the necessary model configuration, tokenizer, and generation settings to run a minimal Llama architecture. ## Model Overview Micro Llama v0 is based on the LlamaForCausalLM architecture. It is tailored to fit resource-constrained environments for testing the foundational components of a transformer-based language model. This version features: - **1 hidden layer** - **2048 hidden size** - **32 attention heads** - **5632 intermediate size** - **Max position embeddings** of 2048 - **Vocabulary size** of 32,000 These parameters make the model compact and suitable for development, while still maintaining key characteristics of the Llama architecture. ## Files and Configuration - **`config.json`**: Contains the model architecture configuration, such as hidden size, number of attention heads, hidden layers, and activation functions. - **`generation_config.json`**: Specifies generation parameters, including max length and token behavior. - **`model.safetensors`**: Stores the model weights in a safe and efficient format. - **`special_tokens_map.json`**: Maps the special tokens used by the model, including ``, ``, ``, and `` (for padding). - **`tokenizer.json`**: Defines the tokenizer configuration, including vocabulary size and token mapping. - **`tokenizer_config.json`**: Further configures the tokenizer, specifying token types, maximum sequence length, and other tokenizer options. ## Requirements - [Transformers](https://github.com/huggingface/transformers) version 4.44.0 or above - PyTorch version compatible with the model's `float32` tensor type - `safetensors` package for loading model weights ## Usage 1. Clone the repository: ```bash git clone https://github.com/your-repo/micro-llama.git cd micro-llama ``` 2. Install the required dependencies: ```bash pip install transformers safetensors torch ``` 3. Load the model in your code: ```python from transformers import LlamaForCausalLM, LlamaTokenizer tokenizer = LlamaTokenizer.from_pretrained("UnieAI-Wilson/micro-llama-0-dev") model = LlamaForCausalLM.from_pretrained("UnieAI-Wilson/micro-llama-0-dev", torch_dtype="float16") inputs = tokenizer("Your text here", return_tensors="pt") outputs = model.generate(**inputs) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) ``` ## License Micro Llama v0 is licensed under the Apache 2.0 License. See the LICENSE file for details. ## Contribution This is an experimental and evolving project. Contributions are welcome, and feel free to submit issues or pull requests. ## Disclaimer This is an early-stage development version, and the model may undergo significant changes. It is not intended for production use.