--- license: apache-2.0 datasets: - mlfoundations/dclm-baseline-1.0-parquet - mlfoundations/dclm-baseline-1.0 ---
### Experimental Development Models These development models are designed specifically for **experimental and testing purposes**. They have been trained using our pre-trained **BPE tokenizer** with a vocabulary size of 61,440. #### Model Details: - **[Oute-Dev-0.7B-Checkpoint-40B](https://huggingface.co/OuteAI/Oute-Dev-0.7B-Checkpoint-40B)**: Built on the **Phi3 architecture**, trained on approximately **40 billion tokens**. - **[Oute-Dev-1B-Checkpoint-40B](https://huggingface.co/OuteAI/Oute-Dev-1B-Checkpoint-40B)**: Built on the **LLaMa architecture**, trained on approximately **40 billion tokens**. > [!IMPORTANT] > These models were initially developed for internal testing and did not undergo extensive training. The output quality will not be suitable for production use or serious applications. You should expect inconsistent, potentially low-quality outputs. ### Benchmark Performance:Benchmark | Oute-Dev-0.7B-Checkpoint-40B | Oute-Dev-1B-Checkpoint-40B |
---|---|---|
ARC-C (0-shot) | 28.24 | 26.19 |
ARC-E (0-shot) | 55.13 | 57.32 |
HellaSWAG (0-shot) | 41.20 | 43.70 |
PIQA (0-shot) | 68.39 | 69.59 |
Winogrande (0-shot) | 54.14 | 50.51 |