File size: 869 Bytes
58595af
 
 
6e021dd
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
---
license: apache-2.0
---


# Download the llamafile
- Go to https://huggingface.co/aviol/TinyLlama1.1-llamafile-bootstraped/blob/main/TinyLlama-1.1B.llamafile
  - Download this file using the download button.

# Run the server
```shell
chmod +x TinyLlama-1.1B.llamafile

./TinyLlama-1.1B.llamafile --server --host 0.0.0.0 --port 1234
```

# Use the LLM with OpenAI SDK:
```python
from openai import OpenAI


client = OpenAI(base_url="http://127.0.0.1:1234/v1", api_key="test")

# Prompt
prompt = "Hi, tell me something new about AppSec"

# Send API request to llamafile server
stream = client.chat.completions.create(
    model="avi-llmsky",
    messages=[{"role": "user", "content": prompt}],
    stream=True,
)

# Print the responses
for chunk in stream:
    if chunk.choices[0].delta.content is not None:
        print(chunk.choices[0].delta.content, end="")

```