diff --git a/src/backend/langflow/api/v1/endpoints.py b/src/backend/langflow/api/v1/endpoints.py index 51a397e1c..a6b85dde2 100644 --- a/src/backend/langflow/api/v1/endpoints.py +++ b/src/backend/langflow/api/v1/endpoints.py @@ -1,5 +1,5 @@ from http import HTTPStatus -from typing import Annotated, Optional +from typing import Annotated, Optional, Union from langflow.services.cache.utils import save_uploaded_file from langflow.services.database.models.flow import Flow @@ -75,7 +75,7 @@ async def process_flow( inputs: Optional[dict] = None, tweaks: Optional[dict] = None, clear_cache: Annotated[bool, Body(embed=True)] = False, # noqa: F821 - session_id: Annotated[str, Body(embed=True)] = None, # noqa: F821 + session_id: Annotated[Union[None, str], Body(embed=True)] = None, # noqa: F821 session: Session = Depends(get_session), ): """ diff --git a/src/backend/langflow/api/v1/schemas.py b/src/backend/langflow/api/v1/schemas.py index 1b1d763f3..65bf64dca 100644 --- a/src/backend/langflow/api/v1/schemas.py +++ b/src/backend/langflow/api/v1/schemas.py @@ -47,7 +47,7 @@ class ProcessResponse(BaseModel): """Process response schema.""" result: dict - session_id: str = None + session_id: Optional[str] = None class ChatMessage(BaseModel): diff --git a/src/backend/langflow/interface/run.py b/src/backend/langflow/interface/run.py index cb0573bf7..42cea0e98 100644 --- a/src/backend/langflow/interface/run.py +++ b/src/backend/langflow/interface/run.py @@ -1,3 +1,4 @@ +from typing import Any, Dict, Tuple from langflow.services.cache.utils import memoize_dict from langflow.graph import Graph from langflow.utils.logger import logger @@ -15,7 +16,7 @@ def build_langchain_object_with_caching(data_graph): @memoize_dict(maxsize=10) -def build_sorted_vertices_with_caching(data_graph): +def build_sorted_vertices_with_caching(data_graph) -> Tuple[Any, Dict]: """ Build langchain object from data_graph. """ diff --git a/src/backend/langflow/processing/process.py b/src/backend/langflow/processing/process.py index dc632e76b..396135e16 100644 --- a/src/backend/langflow/processing/process.py +++ b/src/backend/langflow/processing/process.py @@ -109,7 +109,7 @@ def clear_caches_if_needed(clear_cache: bool): def load_langchain_object( data_graph: Dict[str, Any], session_id: str -) -> Tuple[Union[Chain, VectorStore], Dict[str, Any]]: +) -> Tuple[Union[Chain, VectorStore], Dict[str, Any], str]: langchain_object, artifacts = get_build_result(data_graph, session_id) session_id = build_sorted_vertices_with_caching.hash logger.debug("Loaded LangChain object")