Spaces:
Running
Running
import asyncio | |
import json | |
import shutil | |
# we need to import tmpdir | |
import tempfile | |
from collections.abc import AsyncGenerator | |
from contextlib import contextmanager, suppress | |
from pathlib import Path | |
from typing import TYPE_CHECKING | |
from uuid import UUID | |
import orjson | |
import pytest | |
from asgi_lifespan import LifespanManager | |
from blockbuster import blockbuster_ctx | |
from dotenv import load_dotenv | |
from fastapi.testclient import TestClient | |
from httpx import ASGITransport, AsyncClient | |
from langflow.graph import Graph | |
from langflow.initial_setup.constants import STARTER_FOLDER_NAME | |
from langflow.services.auth.utils import get_password_hash | |
from langflow.services.database.models.api_key.model import ApiKey | |
from langflow.services.database.models.flow.model import Flow, FlowCreate | |
from langflow.services.database.models.folder.model import Folder | |
from langflow.services.database.models.transactions.model import TransactionTable | |
from langflow.services.database.models.user.model import User, UserCreate, UserRead | |
from langflow.services.database.models.vertex_builds.crud import delete_vertex_builds_by_flow_id | |
from langflow.services.database.utils import async_session_getter | |
from langflow.services.deps import get_db_service | |
from loguru import logger | |
from sqlalchemy.ext.asyncio import create_async_engine | |
from sqlalchemy.orm import selectinload | |
from sqlmodel import Session, SQLModel, create_engine, select | |
from sqlmodel.ext.asyncio.session import AsyncSession | |
from sqlmodel.pool import StaticPool | |
from typer.testing import CliRunner | |
from tests.api_keys import get_openai_api_key | |
if TYPE_CHECKING: | |
from langflow.services.database.service import DatabaseService | |
load_dotenv() | |
def blockbuster(request): | |
if "benchmark" in request.keywords: | |
yield | |
else: | |
with blockbuster_ctx() as bb: | |
for func in [ | |
"io.BufferedReader.read", | |
"io.BufferedWriter.write", | |
"io.TextIOWrapper.read", | |
"io.TextIOWrapper.write", | |
]: | |
bb.functions[func].can_block_functions.append(("settings/service.py", {"initialize"})) | |
for func in [ | |
"io.BufferedReader.read", | |
"io.TextIOWrapper.read", | |
]: | |
bb.functions[func].can_block_functions.append(("importlib_metadata/__init__.py", {"metadata"})) | |
for func in bb.functions: | |
if func.startswith("sqlite3."): | |
bb.functions[func].deactivate() | |
bb.functions["threading.Lock.acquire"].deactivate() | |
yield bb | |
def pytest_configure(config): | |
config.addinivalue_line("markers", "noclient: don't create a client for this test") | |
config.addinivalue_line("markers", "load_flows: load the flows for this test") | |
config.addinivalue_line("markers", "api_key_required: run only if the api key is set in the environment variables") | |
data_path = Path(__file__).parent.absolute() / "data" | |
pytest.BASIC_EXAMPLE_PATH = data_path / "basic_example.json" | |
pytest.COMPLEX_EXAMPLE_PATH = data_path / "complex_example.json" | |
pytest.OPENAPI_EXAMPLE_PATH = data_path / "Openapi.json" | |
pytest.GROUPED_CHAT_EXAMPLE_PATH = data_path / "grouped_chat.json" | |
pytest.ONE_GROUPED_CHAT_EXAMPLE_PATH = data_path / "one_group_chat.json" | |
pytest.VECTOR_STORE_GROUPED_EXAMPLE_PATH = data_path / "vector_store_grouped.json" | |
pytest.WEBHOOK_TEST = data_path / "WebhookTest.json" | |
pytest.BASIC_CHAT_WITH_PROMPT_AND_HISTORY = data_path / "BasicChatwithPromptandHistory.json" | |
pytest.CHAT_INPUT = data_path / "ChatInputTest.json" | |
pytest.TWO_OUTPUTS = data_path / "TwoOutputsTest.json" | |
pytest.VECTOR_STORE_PATH = data_path / "Vector_store.json" | |
pytest.SIMPLE_API_TEST = data_path / "SimpleAPITest.json" | |
pytest.MEMORY_CHATBOT_NO_LLM = data_path / "MemoryChatbotNoLLM.json" | |
pytest.CODE_WITH_SYNTAX_ERROR = """ | |
def get_text(): | |
retun "Hello World" | |
""" | |
# validate that all the paths are correct and the files exist | |
for path in [ | |
pytest.BASIC_EXAMPLE_PATH, | |
pytest.COMPLEX_EXAMPLE_PATH, | |
pytest.OPENAPI_EXAMPLE_PATH, | |
pytest.GROUPED_CHAT_EXAMPLE_PATH, | |
pytest.ONE_GROUPED_CHAT_EXAMPLE_PATH, | |
pytest.VECTOR_STORE_GROUPED_EXAMPLE_PATH, | |
pytest.BASIC_CHAT_WITH_PROMPT_AND_HISTORY, | |
pytest.CHAT_INPUT, | |
pytest.TWO_OUTPUTS, | |
pytest.VECTOR_STORE_PATH, | |
pytest.MEMORY_CHATBOT_NO_LLM, | |
]: | |
assert path.exists(), f"File {path} does not exist. Available files: {list(data_path.iterdir())}" | |
async def delete_transactions_by_flow_id(db: AsyncSession, flow_id: UUID): | |
stmt = select(TransactionTable).where(TransactionTable.flow_id == flow_id) | |
transactions = await db.exec(stmt) | |
for transaction in transactions: | |
await db.delete(transaction) | |
await db.commit() | |
async def _delete_transactions_and_vertex_builds(session, flows: list[Flow]): | |
flow_ids = [flow.id for flow in flows] | |
for flow_id in flow_ids: | |
if not flow_id: | |
continue | |
await delete_vertex_builds_by_flow_id(session, flow_id) | |
await delete_transactions_by_flow_id(session, flow_id) | |
def caplog(caplog: pytest.LogCaptureFixture): | |
handler_id = logger.add( | |
caplog.handler, | |
format="{message}", | |
level=0, | |
filter=lambda record: record["level"].no >= caplog.handler.level, | |
enqueue=False, # Set to 'True' if your test is spawning child processes. | |
) | |
yield caplog | |
logger.remove(handler_id) | |
async def async_client() -> AsyncGenerator: | |
from langflow.main import create_app | |
app = create_app() | |
async with AsyncClient(app=app, base_url="http://testserver", http2=True) as client: | |
yield client | |
def session_fixture(): | |
engine = create_engine("sqlite://", connect_args={"check_same_thread": False}, poolclass=StaticPool) | |
SQLModel.metadata.create_all(engine) | |
with Session(engine) as session: | |
yield session | |
SQLModel.metadata.drop_all(engine) # Add this line to clean up tables | |
async def async_session(): | |
engine = create_async_engine("sqlite+aiosqlite://", connect_args={"check_same_thread": False}, poolclass=StaticPool) | |
async with engine.begin() as conn: | |
await conn.run_sync(SQLModel.metadata.create_all) | |
async with AsyncSession(engine, expire_on_commit=False) as session: | |
yield session | |
async with engine.begin() as conn: | |
await conn.run_sync(SQLModel.metadata.drop_all) | |
class Config: | |
broker_url = "redis://localhost:6379/0" | |
result_backend = "redis://localhost:6379/0" | |
def load_flows_dir(): | |
with tempfile.TemporaryDirectory() as tempdir: | |
yield tempdir | |
def _setup_env(monkeypatch): | |
monkeypatch.setenv("LANGFLOW_CACHE_TYPE", "redis") | |
monkeypatch.setenv("LANGFLOW_REDIS_HOST", "result_backend") | |
monkeypatch.setenv("LANGFLOW_REDIS_PORT", "6379") | |
monkeypatch.setenv("LANGFLOW_REDIS_DB", "0") | |
monkeypatch.setenv("LANGFLOW_REDIS_EXPIRE", "3600") | |
monkeypatch.setenv("LANGFLOW_REDIS_PASSWORD", "") | |
monkeypatch.setenv("FLOWER_UNAUTHENTICATED_API", "True") | |
monkeypatch.setenv("BROKER_URL", "redis://result_backend:6379/0") | |
monkeypatch.setenv("RESULT_BACKEND", "redis://result_backend:6379/0") | |
monkeypatch.setenv("C_FORCE_ROOT", "true") | |
def distributed_client_fixture( | |
session: Session, # noqa: ARG001 | |
monkeypatch, | |
distributed_env, # noqa: ARG001 | |
): | |
# Here we load the .env from ../deploy/.env | |
from langflow.core import celery_app | |
db_dir = tempfile.mkdtemp() | |
try: | |
db_path = Path(db_dir) / "test.db" | |
monkeypatch.setenv("LANGFLOW_DATABASE_URL", f"sqlite:///{db_path}") | |
monkeypatch.setenv("LANGFLOW_AUTO_LOGIN", "false") | |
# monkeypatch langflow.services.task.manager.USE_CELERY to True | |
# monkeypatch.setattr(manager, "USE_CELERY", True) | |
monkeypatch.setattr(celery_app, "celery_app", celery_app.make_celery("langflow", Config)) | |
# def get_session_override(): | |
# return session | |
from langflow.main import create_app | |
app = create_app() | |
# app.dependency_overrides[get_session] = get_session_override | |
with TestClient(app) as client: | |
yield client | |
finally: | |
shutil.rmtree(db_dir) # Clean up the temporary directory | |
app.dependency_overrides.clear() | |
monkeypatch.undo() | |
def get_graph(type_="basic"): | |
"""Get a graph from a json file.""" | |
if type_ == "basic": | |
path = pytest.BASIC_EXAMPLE_PATH | |
elif type_ == "complex": | |
path = pytest.COMPLEX_EXAMPLE_PATH | |
elif type_ == "openapi": | |
path = pytest.OPENAPI_EXAMPLE_PATH | |
with path.open(encoding="utf-8") as f: | |
flow_graph = json.load(f) | |
data_graph = flow_graph["data"] | |
nodes = data_graph["nodes"] | |
edges = data_graph["edges"] | |
graph = Graph() | |
graph.add_nodes_and_edges(nodes, edges) | |
return graph | |
def basic_graph_data(): | |
with pytest.BASIC_EXAMPLE_PATH.open(encoding="utf-8") as f: | |
return json.load(f) | |
def basic_graph(): | |
return get_graph() | |
def complex_graph(): | |
return get_graph("complex") | |
def openapi_graph(): | |
return get_graph("openapi") | |
def json_flow(): | |
return pytest.BASIC_EXAMPLE_PATH.read_text(encoding="utf-8") | |
def grouped_chat_json_flow(): | |
return pytest.GROUPED_CHAT_EXAMPLE_PATH.read_text(encoding="utf-8") | |
def one_grouped_chat_json_flow(): | |
return pytest.ONE_GROUPED_CHAT_EXAMPLE_PATH.read_text(encoding="utf-8") | |
def vector_store_grouped_json_flow(): | |
return pytest.VECTOR_STORE_GROUPED_EXAMPLE_PATH.read_text(encoding="utf-8") | |
def json_flow_with_prompt_and_history(): | |
return pytest.BASIC_CHAT_WITH_PROMPT_AND_HISTORY.read_text(encoding="utf-8") | |
def json_simple_api_test(): | |
return pytest.SIMPLE_API_TEST.read_text(encoding="utf-8") | |
def json_vector_store(): | |
return pytest.VECTOR_STORE_PATH.read_text(encoding="utf-8") | |
def json_webhook_test(): | |
return pytest.WEBHOOK_TEST.read_text(encoding="utf-8") | |
def json_memory_chatbot_no_llm(): | |
return pytest.MEMORY_CHATBOT_NO_LLM.read_text(encoding="utf-8") | |
def deactivate_tracing(monkeypatch): | |
monkeypatch.setenv("LANGFLOW_DEACTIVATE_TRACING", "true") | |
yield | |
monkeypatch.undo() | |
async def client_fixture( | |
session: Session, # noqa: ARG001 | |
monkeypatch, | |
request, | |
load_flows_dir, | |
): | |
# Set the database url to a test database | |
if "noclient" in request.keywords: | |
yield | |
else: | |
def init_app(): | |
db_dir = tempfile.mkdtemp() | |
db_path = Path(db_dir) / "test.db" | |
monkeypatch.setenv("LANGFLOW_DATABASE_URL", f"sqlite:///{db_path}") | |
monkeypatch.setenv("LANGFLOW_AUTO_LOGIN", "false") | |
if "load_flows" in request.keywords: | |
shutil.copyfile( | |
pytest.BASIC_EXAMPLE_PATH, Path(load_flows_dir) / "c54f9130-f2fa-4a3e-b22a-3856d946351b.json" | |
) | |
monkeypatch.setenv("LANGFLOW_LOAD_FLOWS_PATH", load_flows_dir) | |
monkeypatch.setenv("LANGFLOW_AUTO_LOGIN", "true") | |
from langflow.main import create_app | |
app = create_app() | |
db_service = get_db_service() | |
db_service.database_url = f"sqlite:///{db_path}" | |
db_service.reload_engine() | |
return app, db_path | |
app, db_path = await asyncio.to_thread(init_app) | |
# app.dependency_overrides[get_session] = get_session_override | |
async with ( | |
LifespanManager(app, startup_timeout=None, shutdown_timeout=None) as manager, | |
AsyncClient(transport=ASGITransport(app=manager.app), base_url="http://testserver/", http2=True) as client, | |
): | |
yield client | |
# app.dependency_overrides.clear() | |
monkeypatch.undo() | |
# clear the temp db | |
with suppress(FileNotFoundError): | |
db_path.unlink() | |
# create a fixture for session_getter above | |
def session_getter_fixture(client): # noqa: ARG001 | |
def blank_session_getter(db_service: "DatabaseService"): | |
with Session(db_service.engine) as session: | |
yield session | |
return blank_session_getter | |
def runner(): | |
return CliRunner() | |
async def test_user(client): | |
user_data = UserCreate( | |
username="testuser", | |
password="testpassword", # noqa: S106 | |
) | |
response = await client.post("api/v1/users/", json=user_data.model_dump()) | |
assert response.status_code == 201 | |
user = response.json() | |
yield user | |
# Clean up | |
await client.delete(f"/api/v1/users/{user['id']}") | |
async def active_user(client): # noqa: ARG001 | |
db_manager = get_db_service() | |
async with db_manager.with_async_session() as session: | |
user = User( | |
username="activeuser", | |
password=get_password_hash("testpassword"), | |
is_active=True, | |
is_superuser=False, | |
) | |
stmt = select(User).where(User.username == user.username) | |
if active_user := (await session.exec(stmt)).first(): | |
user = active_user | |
else: | |
session.add(user) | |
await session.commit() | |
await session.refresh(user) | |
user = UserRead.model_validate(user, from_attributes=True) | |
yield user | |
# Clean up | |
# Now cleanup transactions, vertex_build | |
async with db_manager.with_async_session() as session: | |
user = await session.get(User, user.id, options=[selectinload(User.flows)]) | |
await _delete_transactions_and_vertex_builds(session, user.flows) | |
await session.delete(user) | |
await session.commit() | |
async def logged_in_headers(client, active_user): | |
login_data = {"username": active_user.username, "password": "testpassword"} | |
response = await client.post("api/v1/login", data=login_data) | |
assert response.status_code == 200 | |
tokens = response.json() | |
a_token = tokens["access_token"] | |
return {"Authorization": f"Bearer {a_token}"} | |
async def active_super_user(client): # noqa: ARG001 | |
db_manager = get_db_service() | |
async with db_manager.with_async_session() as session: | |
user = User( | |
username="activeuser", | |
password=get_password_hash("testpassword"), | |
is_active=True, | |
is_superuser=True, | |
) | |
stmt = select(User).where(User.username == user.username) | |
if active_user := (await session.exec(stmt)).first(): | |
user = active_user | |
else: | |
session.add(user) | |
await session.commit() | |
await session.refresh(user) | |
user = UserRead.model_validate(user, from_attributes=True) | |
yield user | |
# Clean up | |
# Now cleanup transactions, vertex_build | |
async with db_manager.with_async_session() as session: | |
user = await session.get(User, user.id, options=[selectinload(User.flows)]) | |
await _delete_transactions_and_vertex_builds(session, user.flows) | |
await session.delete(user) | |
await session.commit() | |
async def logged_in_headers_super_user(client, active_super_user): | |
login_data = {"username": active_super_user.username, "password": "testpassword"} | |
response = await client.post("api/v1/login", data=login_data) | |
assert response.status_code == 200 | |
tokens = response.json() | |
a_token = tokens["access_token"] | |
return {"Authorization": f"Bearer {a_token}"} | |
async def flow( | |
client, # noqa: ARG001 | |
json_flow: str, | |
active_user, | |
): | |
from langflow.services.database.models.flow.model import FlowCreate | |
loaded_json = json.loads(json_flow) | |
flow_data = FlowCreate(name="test_flow", data=loaded_json.get("data"), user_id=active_user.id) | |
flow = Flow.model_validate(flow_data) | |
async with async_session_getter(get_db_service()) as session: | |
session.add(flow) | |
await session.commit() | |
await session.refresh(flow) | |
yield flow | |
# Clean up | |
await session.delete(flow) | |
await session.commit() | |
def json_chat_input(): | |
return pytest.CHAT_INPUT.read_text(encoding="utf-8") | |
def json_two_outputs(): | |
return pytest.TWO_OUTPUTS.read_text(encoding="utf-8") | |
async def added_flow_webhook_test(client, json_webhook_test, logged_in_headers): | |
flow = orjson.loads(json_webhook_test) | |
data = flow["data"] | |
flow = FlowCreate(name="Basic Chat", description="description", data=data) | |
response = await client.post("api/v1/flows/", json=flow.model_dump(), headers=logged_in_headers) | |
assert response.status_code == 201 | |
assert response.json()["name"] == flow.name | |
assert response.json()["data"] == flow.data | |
yield response.json() | |
await client.delete(f"api/v1/flows/{response.json()['id']}", headers=logged_in_headers) | |
async def added_flow_chat_input(client, json_chat_input, logged_in_headers): | |
flow = orjson.loads(json_chat_input) | |
data = flow["data"] | |
flow = FlowCreate(name="Chat Input", description="description", data=data) | |
response = await client.post("api/v1/flows/", json=flow.model_dump(), headers=logged_in_headers) | |
assert response.status_code == 201 | |
assert response.json()["name"] == flow.name | |
assert response.json()["data"] == flow.data | |
yield response.json() | |
await client.delete(f"api/v1/flows/{response.json()['id']}", headers=logged_in_headers) | |
async def added_flow_two_outputs(client, json_two_outputs, logged_in_headers): | |
flow = orjson.loads(json_two_outputs) | |
data = flow["data"] | |
flow = FlowCreate(name="Two Outputs", description="description", data=data) | |
response = await client.post("api/v1/flows/", json=flow.model_dump(), headers=logged_in_headers) | |
assert response.status_code == 201 | |
assert response.json()["name"] == flow.name | |
assert response.json()["data"] == flow.data | |
yield response.json() | |
await client.delete(f"api/v1/flows/{response.json()['id']}", headers=logged_in_headers) | |
async def added_vector_store(client, json_vector_store, logged_in_headers): | |
vector_store = orjson.loads(json_vector_store) | |
data = vector_store["data"] | |
vector_store = FlowCreate(name="Vector Store", description="description", data=data) | |
response = await client.post("api/v1/flows/", json=vector_store.model_dump(), headers=logged_in_headers) | |
assert response.status_code == 201 | |
assert response.json()["name"] == vector_store.name | |
assert response.json()["data"] == vector_store.data | |
yield response.json() | |
await client.delete(f"api/v1/flows/{response.json()['id']}", headers=logged_in_headers) | |
async def added_webhook_test(client, json_webhook_test, logged_in_headers): | |
webhook_test = orjson.loads(json_webhook_test) | |
data = webhook_test["data"] | |
webhook_test = FlowCreate( | |
name="Webhook Test", description="description", data=data, endpoint_name=webhook_test["endpoint_name"] | |
) | |
response = await client.post("api/v1/flows/", json=webhook_test.model_dump(), headers=logged_in_headers) | |
assert response.status_code == 201 | |
assert response.json()["name"] == webhook_test.name | |
assert response.json()["data"] == webhook_test.data | |
yield response.json() | |
await client.delete(f"api/v1/flows/{response.json()['id']}", headers=logged_in_headers) | |
async def flow_component(client: AsyncClient, logged_in_headers): | |
from langflow.components.inputs import ChatInput | |
chat_input = ChatInput() | |
graph = Graph(start=chat_input, end=chat_input) | |
graph_dict = graph.dump(name="Chat Input Component") | |
flow = FlowCreate(**graph_dict) | |
response = await client.post("api/v1/flows/", json=flow.model_dump(), headers=logged_in_headers) | |
assert response.status_code == 201 | |
yield response.json() | |
await client.delete(f"api/v1/flows/{response.json()['id']}", headers=logged_in_headers) | |
async def created_api_key(active_user): | |
hashed = get_password_hash("random_key") | |
api_key = ApiKey( | |
name="test_api_key", | |
user_id=active_user.id, | |
api_key="random_key", | |
hashed_api_key=hashed, | |
) | |
db_manager = get_db_service() | |
async with async_session_getter(db_manager) as session: | |
stmt = select(ApiKey).where(ApiKey.api_key == api_key.api_key) | |
if existing_api_key := (await session.exec(stmt)).first(): | |
yield existing_api_key | |
return | |
session.add(api_key) | |
await session.commit() | |
await session.refresh(api_key) | |
yield api_key | |
# Clean up | |
await session.delete(api_key) | |
await session.commit() | |
async def get_simple_api_test(client, logged_in_headers, json_simple_api_test): | |
# Once the client is created, we can get the starter project | |
# Just create a new flow with the simple api test | |
flow = orjson.loads(json_simple_api_test) | |
data = flow["data"] | |
flow = FlowCreate(name="Simple API Test", data=data, description="Simple API Test") | |
response = await client.post("api/v1/flows/", json=flow.model_dump(), headers=logged_in_headers) | |
assert response.status_code == 201 | |
yield response.json() | |
await client.delete(f"api/v1/flows/{response.json()['id']}", headers=logged_in_headers) | |
async def get_starter_project(active_user): | |
# once the client is created, we can get the starter project | |
async with async_session_getter(get_db_service()) as session: | |
stmt = ( | |
select(Flow) | |
.where(Flow.folder.has(Folder.name == STARTER_FOLDER_NAME)) | |
.where(Flow.name == "Basic Prompting (Hello, World)") | |
) | |
flow = (await session.exec(stmt)).first() | |
if not flow: | |
msg = "No starter project found" | |
raise ValueError(msg) | |
# ensure openai api key is set | |
get_openai_api_key() | |
new_flow_create = FlowCreate( | |
name=flow.name, | |
description=flow.description, | |
data=flow.data, | |
user_id=active_user.id, | |
) | |
new_flow = Flow.model_validate(new_flow_create, from_attributes=True) | |
session.add(new_flow) | |
await session.commit() | |
await session.refresh(new_flow) | |
new_flow_dict = new_flow.model_dump() | |
yield new_flow_dict | |
# Clean up | |
await session.delete(new_flow) | |
await session.commit() | |