From f8ce1331b96d32f050aa97ec4bc2c81813a810d1 Mon Sep 17 00:00:00 2001 From: Gabriel Luiz Freitas Almeida Date: Thu, 20 Jun 2024 16:35:00 -0300 Subject: [PATCH] refactor: Remove unused text_response method from model components --- .../components/models/AmazonBedrockModel.py | 9 --------- .../langflow/components/models/AzureOpenAIModel.py | 9 --------- .../components/models/BaiduQianfanChatModel.py | 11 +---------- .../langflow/components/models/ChatLiteLLMModel.py | 11 ----------- .../base/langflow/components/models/CohereModel.py | 11 +---------- .../components/models/GoogleGenerativeAIModel.py | 9 --------- .../langflow/components/models/HuggingFaceModel.py | 11 +---------- .../base/langflow/components/models/MistralModel.py | 11 +---------- .../base/langflow/components/models/OllamaModel.py | 10 ---------- .../langflow/components/models/VertexAiModel.py | 13 +------------ 10 files changed, 5 insertions(+), 100 deletions(-) diff --git a/src/backend/base/langflow/components/models/AmazonBedrockModel.py b/src/backend/base/langflow/components/models/AmazonBedrockModel.py index 64f1881cb..2d5c434cd 100644 --- a/src/backend/base/langflow/components/models/AmazonBedrockModel.py +++ b/src/backend/base/langflow/components/models/AmazonBedrockModel.py @@ -69,15 +69,6 @@ class AmazonBedrockComponent(LCModelComponent): Output(display_name="Language Model", name="model_output", method="build_model"), ] - def text_response(self) -> Text: - input_value = self.input_value - stream = self.stream - system_message = self.system_message - output = self.build_model() - result = self.get_chat_result(output, stream, input_value, system_message) - self.status = result - return result - def build_model(self) -> LanguageModel: model_id = self.model_id credentials_profile_name = self.credentials_profile_name diff --git a/src/backend/base/langflow/components/models/AzureOpenAIModel.py b/src/backend/base/langflow/components/models/AzureOpenAIModel.py index 2998e7935..65b5e23fd 100644 --- a/src/backend/base/langflow/components/models/AzureOpenAIModel.py +++ b/src/backend/base/langflow/components/models/AzureOpenAIModel.py @@ -78,15 +78,6 @@ class AzureChatOpenAIComponent(LCModelComponent): Output(display_name="Language Model", name="model_output", method="model_response"), ] - def text_response(self) -> Text: - input_value = self.input_value - stream = self.stream - system_message = self.system_message - output = self.model_response() - result = self.get_chat_result(output, stream, input_value, system_message) - self.status = result - return result - def model_response(self) -> LanguageModel: model = self.model azure_endpoint = self.azure_endpoint diff --git a/src/backend/base/langflow/components/models/BaiduQianfanChatModel.py b/src/backend/base/langflow/components/models/BaiduQianfanChatModel.py index c19b36804..2c91fe7ce 100644 --- a/src/backend/base/langflow/components/models/BaiduQianfanChatModel.py +++ b/src/backend/base/langflow/components/models/BaiduQianfanChatModel.py @@ -3,7 +3,7 @@ from pydantic.v1 import SecretStr from langflow.base.constants import STREAM_INFO_TEXT from langflow.base.models.model import LCModelComponent -from langflow.io import BoolInput, DropdownInput, FloatInput, Output, SecretStrInput, TextInput +from langflow.field_typing.constants import LanguageModel from langflow.io import BoolInput, DropdownInput, FloatInput, Output, SecretStrInput, TextInput @@ -89,15 +89,6 @@ class QianfanChatEndpointComponent(LCModelComponent): Output(display_name="Language Model", name="model_output", method="build_model"), ] - def text_response(self) -> Text: - input_value = self.input_value - stream = self.stream - system_message = self.system_message - output = self.build_model() - result = self.get_chat_result(output, stream, input_value, system_message) - self.status = result - return result - def build_model(self) -> LanguageModel: model = self.model qianfan_ak = self.qianfan_ak diff --git a/src/backend/base/langflow/components/models/ChatLiteLLMModel.py b/src/backend/base/langflow/components/models/ChatLiteLLMModel.py index 4bf99f39f..e43854bc7 100644 --- a/src/backend/base/langflow/components/models/ChatLiteLLMModel.py +++ b/src/backend/base/langflow/components/models/ChatLiteLLMModel.py @@ -16,7 +16,6 @@ from langflow.io import ( SecretStrInput, StrInput, ) -from langflow.schema.message import Message class ChatLiteLLMModelComponent(LCModelComponent): @@ -128,16 +127,6 @@ class ChatLiteLLMModelComponent(LCModelComponent): Output(display_name="Language Model", name="model_output", method="build_model"), ] - def text_response(self) -> Message: - input_value = self.input_value - stream = self.stream - system_message = self.system_message - output = self.build_model() - result = self.get_chat_result(output, stream, input_value, system_message) - message = Message(text=result) - self.status = message - return message - def build_model(self) -> LanguageModel: try: import litellm # type: ignore diff --git a/src/backend/base/langflow/components/models/CohereModel.py b/src/backend/base/langflow/components/models/CohereModel.py index d9c7f0f86..305f8b4c2 100644 --- a/src/backend/base/langflow/components/models/CohereModel.py +++ b/src/backend/base/langflow/components/models/CohereModel.py @@ -4,7 +4,7 @@ from pydantic.v1 import SecretStr from langflow.base.constants import STREAM_INFO_TEXT from langflow.base.models.model import LCModelComponent -from langflow.field_typing import LanguageModel, Text +from langflow.field_typing import LanguageModel from langflow.io import BoolInput, FloatInput, MessageInput, Output, SecretStrInput, StrInput @@ -37,15 +37,6 @@ class CohereComponent(LCModelComponent): Output(display_name="Language Model", name="model_output", method="build_model"), ] - def text_response(self) -> Text: - input_value = self.input_value - stream = self.stream - system_message = self.system_message - output = self.build_model() - result = self.get_chat_result(output, stream, input_value, system_message) - self.status = result - return result - def build_model(self) -> LanguageModel | BaseChatModel: cohere_api_key = self.cohere_api_key temperature = self.temperature diff --git a/src/backend/base/langflow/components/models/GoogleGenerativeAIModel.py b/src/backend/base/langflow/components/models/GoogleGenerativeAIModel.py index 823583a2f..33684e8da 100644 --- a/src/backend/base/langflow/components/models/GoogleGenerativeAIModel.py +++ b/src/backend/base/langflow/components/models/GoogleGenerativeAIModel.py @@ -73,15 +73,6 @@ class GoogleGenerativeAIComponent(LCModelComponent): Output(display_name="Language Model", name="model_output", method="build_model"), ] - def text_response(self) -> Text: - input_value = self.input_value - stream = self.stream - system_message = self.system_message - output = self.build_model() - result = self.get_chat_result(output, stream, input_value, system_message) - self.status = result - return result - def build_model(self) -> LanguageModel: try: from langchain_google_genai import ChatGoogleGenerativeAI diff --git a/src/backend/base/langflow/components/models/HuggingFaceModel.py b/src/backend/base/langflow/components/models/HuggingFaceModel.py index 5c2335cb4..a3ec98799 100644 --- a/src/backend/base/langflow/components/models/HuggingFaceModel.py +++ b/src/backend/base/langflow/components/models/HuggingFaceModel.py @@ -3,7 +3,7 @@ from langchain_community.llms.huggingface_endpoint import HuggingFaceEndpoint from langflow.base.constants import STREAM_INFO_TEXT from langflow.base.models.model import LCModelComponent -from langflow.field_typing import LanguageModel, Text +from langflow.field_typing import LanguageModel from langflow.io import BoolInput, DictInput, DropdownInput, MessageInput, Output, SecretStrInput, StrInput @@ -36,15 +36,6 @@ class HuggingFaceEndpointsComponent(LCModelComponent): Output(display_name="Language Model", name="model_output", method="build_model"), ] - def text_response(self) -> Text: - input_value = self.input_value - stream = self.stream - system_message = self.system_message - output = self.build_model() - result = self.get_chat_result(output, stream, input_value, system_message) - self.status = result - return result - def build_model(self) -> LanguageModel: endpoint_url = self.endpoint_url task = self.task diff --git a/src/backend/base/langflow/components/models/MistralModel.py b/src/backend/base/langflow/components/models/MistralModel.py index cc74e75fd..d9058a4f3 100644 --- a/src/backend/base/langflow/components/models/MistralModel.py +++ b/src/backend/base/langflow/components/models/MistralModel.py @@ -3,7 +3,7 @@ from pydantic.v1 import SecretStr from langflow.base.constants import STREAM_INFO_TEXT from langflow.base.models.model import LCModelComponent -from langflow.field_typing import LanguageModel, Text +from langflow.field_typing import LanguageModel from langflow.io import BoolInput, DropdownInput, FloatInput, IntInput, MessageInput, Output, SecretStrInput, StrInput @@ -70,15 +70,6 @@ class MistralAIModelComponent(LCModelComponent): Output(display_name="Language Model", name="model_output", method="build_model"), ] - def text_response(self) -> Text: - input_value = self.input_value - stream = self.stream - system_message = self.system_message - output = self.build_model() - result = self.get_chat_result(output, stream, input_value, system_message) - self.status = result - return result - def build_model(self) -> LanguageModel: mistral_api_key = self.mistral_api_key temperature = self.temperature diff --git a/src/backend/base/langflow/components/models/OllamaModel.py b/src/backend/base/langflow/components/models/OllamaModel.py index 1fd9abcc7..71a62cd68 100644 --- a/src/backend/base/langflow/components/models/OllamaModel.py +++ b/src/backend/base/langflow/components/models/OllamaModel.py @@ -7,7 +7,6 @@ from langflow.base.constants import STREAM_INFO_TEXT from langflow.base.models.model import LCModelComponent from langflow.field_typing import LanguageModel from langflow.io import BoolInput, DictInput, DropdownInput, FloatInput, IntInput, MessageInput, Output, StrInput -from langflow.schema.message import Message class ChatOllamaComponent(LCModelComponent): @@ -224,15 +223,6 @@ class ChatOllamaComponent(LCModelComponent): Output(display_name="Language Model", name="model_output", method="build_model"), ] - def text_response(self) -> Message: - input_value = self.input_value - stream = self.stream - system_message = self.system_message - output = self.build_model() - result = self.get_chat_result(output, stream, input_value, system_message) - self.status = result - return result - def build_model(self) -> LanguageModel: # Mapping mirostat settings to their corresponding values mirostat_options = {"Mirostat": 1, "Mirostat 2.0": 2} diff --git a/src/backend/base/langflow/components/models/VertexAiModel.py b/src/backend/base/langflow/components/models/VertexAiModel.py index e70aa2d06..63438b49f 100644 --- a/src/backend/base/langflow/components/models/VertexAiModel.py +++ b/src/backend/base/langflow/components/models/VertexAiModel.py @@ -1,9 +1,7 @@ from langchain_google_vertexai import ChatVertexAI - -from langflow.base.constants import STREAM_INFO_TEXT from langflow.base.models.model import LCModelComponent -from langflow.field_typing import LanguageModel, Text +from langflow.field_typing import LanguageModel from langflow.io import BoolInput, FileInput, FloatInput, IntInput, MessageInput, MultilineInput, Output, StrInput @@ -53,15 +51,6 @@ class ChatVertexAIComponent(LCModelComponent): Output(display_name="Language Model", name="model_output", method="build_model"), ] - def text_response(self) -> Text: - input_value = self.input_value - stream = self.stream - system_message = self.system_message - output = self.build_model() - result = self.get_chat_result(output, stream, input_value, system_message) - self.status = result - return result - def build_model(self) -> LanguageModel: credentials = self.credentials location = self.location