Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -83,32 +83,14 @@ additional_inputs=[
|
|
83 |
|
84 |
css = """
|
85 |
#mkd {
|
86 |
-
height:
|
87 |
overflow: auto;
|
88 |
border: 1px solid #ccc;
|
89 |
}
|
90 |
"""
|
91 |
|
92 |
with gr.Blocks(css=css) as demo:
|
93 |
-
|
94 |
-
In this demo, you can chat with <a href='https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1'>Mistral-7B-Instruct</a> model. π¬
|
95 |
-
Learn more about the model <a href='https://huggingface.co/docs/transformers/main/model_doc/mistral'>here</a>. π
|
96 |
-
<h2>π Model Features π </h2>
|
97 |
-
<ul>
|
98 |
-
<li>πͺ Sliding Window Attention with 128K tokens span</li>
|
99 |
-
<li>π GQA for faster inference</li>
|
100 |
-
<li>π Byte-fallback BPE tokenizer</li>
|
101 |
-
</ul>
|
102 |
-
<h3>π License π Released under Apache 2.0 License</h3>
|
103 |
-
<h3>π¦ Usage π¦</h3>
|
104 |
-
<ul>
|
105 |
-
<li>π Available on Huggingface Hub</li>
|
106 |
-
<li>π Python code snippets for easy setup</li>
|
107 |
-
<li>π Expected speedups with Flash Attention 2</li>
|
108 |
-
</ul>
|
109 |
-
""")
|
110 |
-
|
111 |
-
|
112 |
gr.ChatInterface(
|
113 |
generate,
|
114 |
additional_inputs=additional_inputs,
|
@@ -128,5 +110,21 @@ with gr.Blocks(css=css) as demo:
|
|
128 |
["Create a ten-point markdown outline with emojis about: Pneumocystitis jiroveci"]
|
129 |
]
|
130 |
)
|
131 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
132 |
demo.queue().launch(debug=True)
|
|
|
83 |
|
84 |
css = """
|
85 |
#mkd {
|
86 |
+
height: 200px;
|
87 |
overflow: auto;
|
88 |
border: 1px solid #ccc;
|
89 |
}
|
90 |
"""
|
91 |
|
92 |
with gr.Blocks(css=css) as demo:
|
93 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
94 |
gr.ChatInterface(
|
95 |
generate,
|
96 |
additional_inputs=additional_inputs,
|
|
|
110 |
["Create a ten-point markdown outline with emojis about: Pneumocystitis jiroveci"]
|
111 |
]
|
112 |
)
|
113 |
+
gr.HTML("""<h2>π€ Mistral Chat - Gradio π€</h2>
|
114 |
+
In this demo, you can chat with <a href='https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1'>Mistral-7B-Instruct</a> model. π¬
|
115 |
+
Learn more about the model <a href='https://huggingface.co/docs/transformers/main/model_doc/mistral'>here</a>. π
|
116 |
+
<h2>π Model Features π </h2>
|
117 |
+
<ul>
|
118 |
+
<li>πͺ Sliding Window Attention with 128K tokens span</li>
|
119 |
+
<li>π GQA for faster inference</li>
|
120 |
+
<li>π Byte-fallback BPE tokenizer</li>
|
121 |
+
</ul>
|
122 |
+
<h3>π License π Released under Apache 2.0 License</h3>
|
123 |
+
<h3>π¦ Usage π¦</h3>
|
124 |
+
<ul>
|
125 |
+
<li>π Available on Huggingface Hub</li>
|
126 |
+
<li>π Python code snippets for easy setup</li>
|
127 |
+
<li>π Expected speedups with Flash Attention 2</li>
|
128 |
+
</ul>
|
129 |
+
""")
|
130 |
demo.queue().launch(debug=True)
|