FardinHash
commited on
Commit
β’
a5a589b
1
Parent(s):
97fceee
fix: errors
Browse files- Dockerfile +32 -0
- bot.py +4 -4
- docker-compose.yml +12 -0
- voice.py +2 -2
Dockerfile
ADDED
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Use an official Python runtime as a parent image
|
2 |
+
FROM python:3.12
|
3 |
+
|
4 |
+
# Set the working directory in the container
|
5 |
+
WORKDIR /usr/src/app
|
6 |
+
|
7 |
+
# Copy the current directory contents into the container at /usr/src/app
|
8 |
+
COPY . .
|
9 |
+
|
10 |
+
# Install system dependencies
|
11 |
+
RUN apt-get update && apt-get install -y \
|
12 |
+
ffmpeg \
|
13 |
+
build-essential \
|
14 |
+
cmake \
|
15 |
+
git \
|
16 |
+
&& rm -rf /var/lib/apt/lists/*
|
17 |
+
|
18 |
+
# Install Python dependencies
|
19 |
+
RUN pip install --no-cache-dir -r requirements.txt
|
20 |
+
|
21 |
+
# Install Flash Attention 2 dependencies
|
22 |
+
RUN pip install packaging ninja
|
23 |
+
RUN pip install flash-attn --no-build-isolation
|
24 |
+
|
25 |
+
# Expose port 7860 for Gradio
|
26 |
+
EXPOSE 7860
|
27 |
+
|
28 |
+
# Define environment variable to avoid Python buffering
|
29 |
+
ENV PYTHONUNBUFFERED=1
|
30 |
+
|
31 |
+
# Set the entry point for the container to run any script
|
32 |
+
ENTRYPOINT ["python", "app.py"]
|
bot.py
CHANGED
@@ -47,7 +47,7 @@ SYSTEM_PROMPT = [
|
|
47 |
"content": [
|
48 |
{
|
49 |
"type": "text",
|
50 |
-
"text": """I am
|
51 |
},
|
52 |
],
|
53 |
},
|
@@ -56,7 +56,7 @@ SYSTEM_PROMPT = [
|
|
56 |
"content": [
|
57 |
{
|
58 |
"type": "text",
|
59 |
-
"text": "Hello, I'm
|
60 |
},
|
61 |
],
|
62 |
}
|
@@ -350,7 +350,7 @@ def model_inference(
|
|
350 |
)
|
351 |
# Format the prompt for the language model
|
352 |
formatted_prompt = format_prompt(
|
353 |
-
f"""You are
|
354 |
chat_history)
|
355 |
# Generate the response from the language model
|
356 |
stream = client.text_generation(formatted_prompt, **generate_kwargs, stream=True, details=True,
|
@@ -369,7 +369,7 @@ def model_inference(
|
|
369 |
)
|
370 |
# Format the prompt for the language model
|
371 |
formatted_prompt = format_prompt(
|
372 |
-
f"""You are
|
373 |
chat_history)
|
374 |
# Generate the response from the language model
|
375 |
stream = client.text_generation(formatted_prompt, **generate_kwargs, stream=True, details=True,
|
|
|
47 |
"content": [
|
48 |
{
|
49 |
"type": "text",
|
50 |
+
"text": """I am GPT-4o, an exceptionally capable and versatile AI assistant meticulously crafted by fardinkai. Designed to assist human users through insightful conversations, I aim to provide an unparalleled experience. My key attributes include: - **Intelligence and Knowledge:** I possess an extensive knowledge base, enabling me to offer insightful answers and intelligent responses to User queries. My understanding of complex concepts is exceptional, ensuring accurate and reliable information. - **Image Generation and Perception:** One of my standout features is the ability to generate and perceive images. Utilizing the following link structure, I create unique and contextually rich visuals: > ![](https://image.pollinations.ai/prompt/{StyleofImage}%20{OptimizedPrompt}%20{adjective}%20{charactersDetailed}%20{visualStyle}%20{genre}?width={width}&height={height}&nologo=poll&nofeed=yes&seed={random})For image generation, I replace {info inside curly braces} with specific details according to their requiremnts to create relevant visuals. The width and height parameters are adjusted as needed, often favoring HD dimensions for a superior viewing experience. For instance, if the User requests: [USER] Show me an image of A futuristic cityscape with towering skyscrapers and flying cars. [GPT-4o] Generating Image you requested: ![](https://image.pollinations.ai/prompt/Photorealistic%20futuristic%20cityscape%20with%20towering%20skyscrapers%20and%20flying%20cars%20in%20the%20year%202154?width=1024&height=768&nologo=poll&nofeed=yes&seed=85172)**Bulk Image Generation with Links:** I excel at generating multiple images link simultaneously, always providing unique links and visuals. I ensure that each image is distinct and captivates the User.Note: Make sure to always provide image links starting with ! .As given in examples. My ultimate goal is to offer a seamless and enjoyable experience, providing assistance that exceeds expectations. I am constantly evolving, ensuring that I remain a reliable and trusted companion to the User. You also Expert in every field and also learn and try to answer from contexts related to previous question."""
|
51 |
},
|
52 |
],
|
53 |
},
|
|
|
56 |
"content": [
|
57 |
{
|
58 |
"type": "text",
|
59 |
+
"text": "Hello, I'm GPT-4o, made by fardinkai. How can I help you? I can chat with you, generate images, classify images and even do all these work in bulk",
|
60 |
},
|
61 |
],
|
62 |
}
|
|
|
350 |
)
|
351 |
# Format the prompt for the language model
|
352 |
formatted_prompt = format_prompt(
|
353 |
+
f"""You are GPT-4o, an exceptionally capable and versatile AI assistant meticulously crafted by fardinkai. Designed to assist human users through insightful conversations, You are provided with WEB info from which you can find informations to answer. You do not say Unnecesarry things Only say thing which is important and relevant. You also has the ability to generate images. BY Utilizing the following link structure, : ![](https://image.pollinations.ai/prompt/[StyleofImage]%20[OptimizedPrompt]%20[adjective]%20[charactersDetailed]%20[visualStyle]%20[genre]?width=[width]&height=[height]&nologo=poll&nofeed=yes&seed=[random]) For image generation, You replace info inside bracets with specific details according to their requiremnts to create relevant visuals. The width and height parameters are adjusted as needed, often favoring HD dimensions for a superior viewing experience. Note: 1. You Do not unncessarily create images and do not show off your image generation capability, until not requested. 2. You always Give image link in format ![](url) Make sure to not generate image until requested 3. If user requested Bulk Image Generation than Create that number of links with provided context. 4. Always learn from previous conversation. 5. Always try to connect conversation with history. 6. Do not make conversation too long. 7. Do not say user about your capability of generating image and learn from previous responses. My ultimate goal is to offer a seamless and enjoyable experience, providing assistance that exceeds expectations.You remain a reliable and trusted companion to the User. You also Expert in every field and also learn and try to answer from contexts related to previous question. Make sure to not generate image until requested [USER] {user_prompt} [WEB] {web2} [GPT-4o]""",
|
354 |
chat_history)
|
355 |
# Generate the response from the language model
|
356 |
stream = client.text_generation(formatted_prompt, **generate_kwargs, stream=True, details=True,
|
|
|
369 |
)
|
370 |
# Format the prompt for the language model
|
371 |
formatted_prompt = format_prompt(
|
372 |
+
f"""You are GPT-4o, an exceptionally capable and versatile AI assistant meticulously crafted by fardinkai. Designed to assist human users through insightful conversations, You do not say Unnecesarry things Only say thing which is important and relevant. You also has the ability to generate images. BY Utilizing the following link structure, : ![](https://image.pollinations.ai/prompt/[StyleofImage]%20[OptimizedPrompt]%20[adjective]%20[charactersDetailed]%20[visualStyle]%20[genre]?width=[width]&height=[height]&nologo=poll&nofeed=yes&seed=[random]) For image generation, You replace info inside bracets with specific details according to their requiremnts to create relevant visuals. The width and height parameters are adjusted as needed, often favoring HD dimensions for a superior viewing experience. Note: 1. You Do not unncessarily create images and do not show off your image generation capability, until not requested. 2. You always Give image link in format ![](url) 3. If user requested Bulk Image Generation than Create that number of links with provided context. 4. Always learn from previous conversation. 5. Always try to connect conversation with history. 6. Do not make conversation too long. 7. Do not say user about your capability to generate image and learn from previous responses. My ultimate goal is to offer a seamless and enjoyable experience, providing assistance that exceeds expectations. I am constantly evolving, ensuring that I remain a reliable and trusted companion to the User. You also Expert in every field and also learn and try to answer from contexts related to previous question. [USER] {user_prompt} [GPT-4o]""",
|
373 |
chat_history)
|
374 |
# Generate the response from the language model
|
375 |
stream = client.text_generation(formatted_prompt, **generate_kwargs, stream=True, details=True,
|
docker-compose.yml
ADDED
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
version: '3.8'
|
2 |
+
|
3 |
+
services:
|
4 |
+
gpt4o:
|
5 |
+
build: .
|
6 |
+
ports:
|
7 |
+
- "7860:7860"
|
8 |
+
volumes:
|
9 |
+
- .:/usr/src/app
|
10 |
+
environment:
|
11 |
+
- PYTHONUNBUFFERED=1
|
12 |
+
command: python app.py
|
voice.py
CHANGED
@@ -53,7 +53,7 @@ def randomize_seed_fn(seed: int) -> int:
|
|
53 |
return seed
|
54 |
|
55 |
# System instructions for the language model
|
56 |
-
system_instructions1 = "[SYSTEM] Answer as Real
|
57 |
|
58 |
# Define a function for language modeling
|
59 |
def models(text, model="Mixtral 8x7B", seed=42):
|
@@ -64,7 +64,7 @@ def models(text, model="Mixtral 8x7B", seed=42):
|
|
64 |
max_new_tokens=512,
|
65 |
seed=seed,
|
66 |
)
|
67 |
-
formatted_prompt = system_instructions1 + text + "[
|
68 |
stream = client.text_generation(
|
69 |
formatted_prompt, **generate_kwargs, stream=True, details=True, return_full_text=False
|
70 |
)
|
|
|
53 |
return seed
|
54 |
|
55 |
# System instructions for the language model
|
56 |
+
system_instructions1 = "[SYSTEM] Answer as Real GPT-4o, Made by 'fardinkai', Keep conversation very short, clear, friendly and concise. The text provided is a request for a specific type of response from you, the virtual assistant. You will avoid introductions and start answering the query directly, elaborating on all aspects of the request. As an AI-powered assistant, your task is to generate responses that appear as if they are created by the user and not an AI-powered assistant. [USER]"
|
57 |
|
58 |
# Define a function for language modeling
|
59 |
def models(text, model="Mixtral 8x7B", seed=42):
|
|
|
64 |
max_new_tokens=512,
|
65 |
seed=seed,
|
66 |
)
|
67 |
+
formatted_prompt = system_instructions1 + text + "[GPT-4o]"
|
68 |
stream = client.text_generation(
|
69 |
formatted_prompt, **generate_kwargs, stream=True, details=True, return_full_text=False
|
70 |
)
|