Spaces:
Runtime error
Runtime error
Update crew.py
Browse files
crew.py
CHANGED
@@ -55,7 +55,7 @@ class NewsletterGenCrew:
|
|
55 |
#
|
56 |
# BadRequestError: Error code: 400 - {'type': 'error', 'error': {'type': 'invalid_request_error', 'message': 'Your credit balance is too low to access the Claude API. Please go to Plans & Billing to upgrade or purchase credits.'}}
|
57 |
|
58 |
-
|
59 |
# https://console.groq.com/docs/rate-limits
|
60 |
# llm = ChatGroq(model="mixtral-8x7b-32768") # JB 13-06-2024 - geeft af en toe rate limit errors
|
61 |
# IN HF SPACES STREAMLIT APP: BadRequestError: Error code: 400 - {'error': {'message': 'Organization has been restricted. Please reach out to support if you believe this was in error.', 'type': 'invalid_request_error', 'code': 'organization_restricted'}}
|
@@ -66,7 +66,7 @@ class NewsletterGenCrew:
|
|
66 |
# supports many more optional parameters. Hover on your `ChatOllama(...)`
|
67 |
# class to view the latest available supported parameters
|
68 |
# llm = ChatOllama(model="llama3")
|
69 |
-
llm = ChatOllama(model="mistral:latest")
|
70 |
# check if ollama is running and which LLMs can then be used, run this in Anaconda cmd admin window:
|
71 |
# ollama list
|
72 |
# OUTPUT EXAMPLE:
|
@@ -81,6 +81,9 @@ class NewsletterGenCrew:
|
|
81 |
#
|
82 |
# OLLAMA LOGS:
|
83 |
# C:\Users\jfhmb\AppData\Local\Ollama
|
|
|
|
|
|
|
84 |
|
85 |
|
86 |
print("JB: in class NewsletterGenCrew - using llm: ", llm)
|
|
|
55 |
#
|
56 |
# BadRequestError: Error code: 400 - {'type': 'error', 'error': {'type': 'invalid_request_error', 'message': 'Your credit balance is too low to access the Claude API. Please go to Plans & Billing to upgrade or purchase credits.'}}
|
57 |
|
58 |
+
llm = ChatGroq(model="llama3-70b-8192")
|
59 |
# https://console.groq.com/docs/rate-limits
|
60 |
# llm = ChatGroq(model="mixtral-8x7b-32768") # JB 13-06-2024 - geeft af en toe rate limit errors
|
61 |
# IN HF SPACES STREAMLIT APP: BadRequestError: Error code: 400 - {'error': {'message': 'Organization has been restricted. Please reach out to support if you believe this was in error.', 'type': 'invalid_request_error', 'code': 'organization_restricted'}}
|
|
|
66 |
# supports many more optional parameters. Hover on your `ChatOllama(...)`
|
67 |
# class to view the latest available supported parameters
|
68 |
# llm = ChatOllama(model="llama3")
|
69 |
+
# llm = ChatOllama(model="mistral:latest")
|
70 |
# check if ollama is running and which LLMs can then be used, run this in Anaconda cmd admin window:
|
71 |
# ollama list
|
72 |
# OUTPUT EXAMPLE:
|
|
|
81 |
#
|
82 |
# OLLAMA LOGS:
|
83 |
# C:\Users\jfhmb\AppData\Local\Ollama
|
84 |
+
#
|
85 |
+
# Running ollama on Hugging Face Spaces #2833
|
86 |
+
# https://github.com/ollama/ollama/issues/2833
|
87 |
|
88 |
|
89 |
print("JB: in class NewsletterGenCrew - using llm: ", llm)
|