diff --git a/src/backend/langflow/api/v1/callback.py b/src/backend/langflow/api/v1/callback.py index deddde47f..69dbf5082 100644 --- a/src/backend/langflow/api/v1/callback.py +++ b/src/backend/langflow/api/v1/callback.py @@ -91,8 +91,8 @@ class AsyncStreamingLLMCallbackHandler(AsyncCallbackHandler): # This is to emulate the stream of tokens for resp in resps: await self.websocket.send_json(resp.dict()) - except Exception as e: - logger.error(e) + except Exception as exc: + logger.error(f"Error sending response: {exc}") async def on_tool_error( self, error: Union[Exception, KeyboardInterrupt], **kwargs: Any diff --git a/src/backend/langflow/api/v1/chat.py b/src/backend/langflow/api/v1/chat.py index 937eb2cf6..499c763b9 100644 --- a/src/backend/langflow/api/v1/chat.py +++ b/src/backend/langflow/api/v1/chat.py @@ -26,7 +26,7 @@ async def chat(client_id: str, websocket: WebSocket): message = "Please, build the flow before sending messages" await websocket.close(code=status.WS_1011_INTERNAL_ERROR, reason=message) except WebSocketException as exc: - logger.error(exc) + logger.error(f"Websocket error: {exc}") await websocket.close(code=status.WS_1011_INTERNAL_ERROR, reason=str(exc)) @@ -56,7 +56,7 @@ async def init_build(graph_data: dict, flow_id: str): return InitResponse(flowId=flow_id) except Exception as exc: - logger.error(exc) + logger.error(f"Error initializing build: {exc}") return HTTPException(status_code=500, detail=str(exc)) @@ -74,7 +74,7 @@ async def build_status(flow_id: str): ) except Exception as exc: - logger.error(exc) + logger.error(f"Error checking build status: {exc}") return HTTPException(status_code=500, detail=str(exc)) @@ -176,5 +176,5 @@ async def stream_build(flow_id: str): try: return StreamingResponse(event_stream(flow_id), media_type="text/event-stream") except Exception as exc: - logger.error(exc) + logger.error(f"Error streaming build: {exc}") raise HTTPException(status_code=500, detail=str(exc)) diff --git a/src/backend/langflow/chat/manager.py b/src/backend/langflow/chat/manager.py index 33de784b5..4754edbed 100644 --- a/src/backend/langflow/chat/manager.py +++ b/src/backend/langflow/chat/manager.py @@ -111,7 +111,7 @@ class ChatManager: # This is to catch the following error: # Unexpected ASGI message 'websocket.close', after sending 'websocket.close' if "after sending" in str(exc): - logger.error(exc) + logger.error(f"Error closing connection: {exc}") async def process_message( self, client_id: str, payload: Dict, langchain_object: Any @@ -197,9 +197,9 @@ class ChatManager: langchain_object = self.in_memory_cache.get(client_id) await self.process_message(client_id, payload, langchain_object) - except Exception as e: + except Exception as exc: # Handle any exceptions that might occur - logger.error(e) + logger.error(f"Error handling websocket: {exc}") await self.close_connection( client_id=client_id, code=status.WS_1011_INTERNAL_ERROR, @@ -212,6 +212,6 @@ class ChatManager: code=status.WS_1000_NORMAL_CLOSURE, reason="Client disconnected", ) - except Exception as e: - logger.error(e) + except Exception as exc: + logger.error(f"Error closing connection: {exc}") self.disconnect(client_id) diff --git a/src/backend/langflow/interface/base.py b/src/backend/langflow/interface/base.py index 6e1522dd2..e6a28bf7d 100644 --- a/src/backend/langflow/interface/base.py +++ b/src/backend/langflow/interface/base.py @@ -34,7 +34,7 @@ class LangChainTypeCreator(BaseModel, ABC): for name, value_dict in type_settings.items() } except AttributeError as exc: - logger.error(exc) + logger.error(f"Error getting settings for {self.type_name}: {exc}") self.name_docs_dict = {} return self.name_docs_dict diff --git a/src/backend/langflow/processing/base.py b/src/backend/langflow/processing/base.py index f8690bbdf..f1d7b6e56 100644 --- a/src/backend/langflow/processing/base.py +++ b/src/backend/langflow/processing/base.py @@ -22,7 +22,7 @@ async def get_result_and_steps(langchain_object, inputs: Union[dict, str], **kwa try: fix_memory_inputs(langchain_object) except Exception as exc: - logger.error(exc) + logger.error(f"Error fixing memory inputs: {exc}") try: async_callbacks = [AsyncStreamingLLMCallbackHandler(**kwargs)]