freddyaboulton's picture
code
da4bdce
raw
history blame
1.43 kB
import gradio as gr
with gr.Blocks() as demo:
gr.Markdown("""
# Llama-2-70b-chat-hf Discord Bot Powered by Gradio and Hugging Face Endpoints
Make sure you read the 'Special Consideration' section below first! 🦙
### First install the `gradio_client`
```bash
pip install gradio_client
```
### Then deploy to discord in one line! ⚡️
```python
grc.Client("ysharma/Explore_llamav2_with_TGI").deploy_discord(to_id="llama2-70b-discord-bot")
```
""")
with gr.Accordion(label="Special Considerations", open=False):
gr.Markdown("""
This discord bot will use a FREE Inference Endpoint provided by Hugging Face.
Hugging Face does not commit to paying for this endpoint in perpetuity so there is no guarantee your bot will always work.
If you would like more control over the infrastructure backing the llama 70b model, consider deploying your own inference endpoint.
"""
)
gr.Markdown("""
Note: As a derivate work of [Llama-2-70b-chat](https://huggingface.co/meta-llama/Llama-2-70b-chat-hf) by Meta, this demo is governed by the original [license](https://huggingface.co/spaces/ysharma/Explore_llamav2_with_TGI/blob/main/LICENSE.txt) and [acceptable use policy](https://huggingface.co/spaces/ysharma/Explore_llamav2_with_TGI/blob/main/USE_POLICY.md)
""")
demo.queue(concurrency_count=70).launch()