From 90d2517fc176b0ab786893ba8585e965cd73baed Mon Sep 17 00:00:00 2001 From: GMS Date: Fri, 2 Jun 2023 21:22:21 +0100 Subject: [PATCH] Added HuggingFaceHub and fixed the stream parameter + added summarization as a task option --- src/backend/langflow/config.yaml | 1 + src/backend/langflow/interface/utils.py | 8 ++++++-- src/backend/langflow/template/frontend_node/llms.py | 6 +++++- 3 files changed, 12 insertions(+), 3 deletions(-) diff --git a/src/backend/langflow/config.yaml b/src/backend/langflow/config.yaml index 3b8554360..c83e18646 100644 --- a/src/backend/langflow/config.yaml +++ b/src/backend/langflow/config.yaml @@ -55,6 +55,7 @@ llms: - LlamaCpp - CTransformers - Cohere + - HuggingFaceHub memories: - ConversationBufferMemory - ConversationSummaryMemory diff --git a/src/backend/langflow/interface/utils.py b/src/backend/langflow/interface/utils.py index 2b7c5acd1..3437df0ee 100644 --- a/src/backend/langflow/interface/utils.py +++ b/src/backend/langflow/interface/utils.py @@ -44,7 +44,11 @@ def try_setting_streaming_options(langchain_object, websocket): langchain_object.llm_chain, "llm" ): llm = langchain_object.llm_chain.llm - if isinstance(llm, BaseLanguageModel) and hasattr(llm, "streaming"): - llm.streaming = True + + if isinstance(llm, BaseLanguageModel): + if hasattr(llm, "streaming"): + llm.streaming = True + if hasattr(llm, "stream"): + llm.stream = True return langchain_object diff --git a/src/backend/langflow/template/frontend_node/llms.py b/src/backend/langflow/template/frontend_node/llms.py index 272770e2e..3b51aa6d0 100644 --- a/src/backend/langflow/template/frontend_node/llms.py +++ b/src/backend/langflow/template/frontend_node/llms.py @@ -34,7 +34,11 @@ class LLMFrontendNode(FrontendNode): field.required = True field.show = True field.is_list = True - field.options = ["text-generation", "text2text-generation"] + field.options = [ + "text-generation", + "text2text-generation", + "summarization" + ] field.advanced = True if display_name := display_names_dict.get(field.name):