k-mktr commited on
Commit
e0f5830
โ€ข
1 Parent(s): 0cb7afe

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +105 -3
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
- title: Gpu Poor Llm Arena
3
- emoji: ๐ŸŒ–
4
  colorFrom: blue
5
  colorTo: purple
6
  sdk: gradio
@@ -11,4 +11,106 @@ license: mit
11
  short_description: 'Compact LLM Battle Arena: Frugal AI Face-Off!'
12
  ---
13
 
14
- Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ title: GPU Poor LLM Arena
3
+ emoji: ๐Ÿ†
4
  colorFrom: blue
5
  colorTo: purple
6
  sdk: gradio
 
11
  short_description: 'Compact LLM Battle Arena: Frugal AI Face-Off!'
12
  ---
13
 
14
+ # ๐Ÿ† GPU-Poor LLM Gladiator Arena ๐Ÿ†
15
+
16
+ Welcome to the GPU-Poor LLM Gladiator Arena, where frugal meets fabulous in the world of AI! This project pits compact language models (maxing out at 9B parameters) against each other in a battle of wits and words.
17
+
18
+ ## ๐Ÿค” Starting from "Why?"
19
+
20
+ In the recent months, We've seen a lot of these "Tiny" models released, and some of them are really impressive.
21
+
22
+ - **Gradio Exploration**: This project serves me as a playground for experimenting with Gradio app development, I am learning how to create interactive AI interfaces with it.
23
+
24
+ - **Tiny Model Evaluation**: I wanted to develop a personal (and now public) stats system for evaluating tiny language models. It's not too serious, but it provides valuable insights into the capabilities of these compact powerhouses.
25
+
26
+ - **Accessibility**: Built on Ollama, this arena allows pretty much anyone to experiment with these models themselves. No need for expensive GPUs or cloud services!
27
+
28
+ - **Pure Fun**: At its core, this project is about having fun with AI. It's a lighthearted way to explore and compare different models. So, haters, feel free to chill โ€“ we're just here for a good time!
29
+
30
+
31
+ ## ๐ŸŒŸ Features
32
+
33
+ - **Battle Arena**: Pit two mystery models against each other and decide which pint-sized powerhouse reigns supreme.
34
+ - **Leaderboard**: Track the performance of different models over time.
35
+ - **Performance Chart**: Visualize model performance with interactive charts.
36
+ - **Privacy-Focused**: Uses local Ollama API, avoiding pricey commercial APIs and keeping data close to home.
37
+ - **Customizable**: Easy to add new models and prompts.
38
+
39
+ ## ๐Ÿš€ Getting Started
40
+
41
+ ### Prerequisites
42
+
43
+ - Python 3.7+
44
+ - Gradio
45
+ - Plotly
46
+ - Ollama (running locally)
47
+
48
+ ### Installation
49
+
50
+ 1. Clone the repository:
51
+ ```
52
+ git clone https://github.com/yourusername/gpu-poor-llm-gladiator-arena.git
53
+ cd gpu-poor-llm-gladiator-arena
54
+ ```
55
+
56
+ 2. Install the required packages:
57
+ ```
58
+ pip install gradio plotly requests
59
+ ```
60
+
61
+ 3. Ensure Ollama is running locally or via a remote server.
62
+
63
+ 4. Run the application:
64
+ ```
65
+ python app.py
66
+ ```
67
+
68
+ ## ๐ŸŽฎ How to Use
69
+
70
+ 1. Open the application in your web browser (typically at `http://localhost:7860`).
71
+ 2. In the "Battle Arena" tab:
72
+ - Enter a prompt or use the random prompt generator (๐ŸŽฒ button).
73
+ - Click "Generate Responses" to see outputs from two random models.
74
+ - Vote for the better response.
75
+ 3. Check the "Leaderboard" tab to see overall model performance.
76
+ 4. View the "Performance Chart" tab for a visual representation of model wins and losses.
77
+
78
+ ## ๐Ÿ›  Configuration
79
+
80
+ You can customize the arena by modifying the `arena_config.py` file:
81
+
82
+ - Add or remove models from the `APPROVED_MODELS` list.
83
+ - Adjust the `API_URL` and `API_KEY` if needed.
84
+ - Customize `example_prompts` for more variety in random prompts.
85
+
86
+ ## ๐Ÿ“Š Leaderboard
87
+
88
+ The leaderboard data is stored in `leaderboard.json`. This file is automatically updated after each battle.
89
+
90
+ ## ๐Ÿค– Models
91
+
92
+ The arena currently supports various compact models, including:
93
+
94
+ - LLaMA 3.2 (1B and 3B versions)
95
+ - LLaMA 3.1 (8B version)
96
+ - Gemma 2 (2B and 9B versions)
97
+ - Qwen 2.5 (0.5B, 1.5B, 3B, and 7B versions)
98
+ - Mistral 0.3 (7B version)
99
+ - Phi 3.5 (3.8B version)
100
+ - Hermes 3 (8B version)
101
+ - Aya 23 (8B version)
102
+
103
+ ## ๐Ÿค Contributing
104
+
105
+ Contributions are welcome! Feel free to suggest a model, which is supported by Ollama. Some results are already quite surprising.
106
+
107
+ ## ๐Ÿ“œ License
108
+
109
+ This project is open-source and available under the MIT License
110
+
111
+ ## ๐Ÿ™ Acknowledgements
112
+
113
+ - Thanks to the Ollama team for providing that amazing tool.
114
+ - Shoutout to all the AI researchers and compact language models teams, making this frugal AI arena possible!
115
+
116
+ Enjoy the battles in the GPU-Poor LLM Gladiator Arena! May the best compact model win! ๐Ÿ†