From a200f582510d9edbc740430ea12f48329652db31 Mon Sep 17 00:00:00 2001 From: Gabriel Luiz Freitas Almeida Date: Fri, 29 Mar 2024 22:39:30 -0300 Subject: [PATCH] Update field order in model components --- .../components/models/AnthropicModel.py | 11 +++++++ .../components/models/AzureOpenAIModel.py | 13 ++++++++ .../models/BaiduQianfanChatModel.py | 13 ++++++++ .../langflow/components/models/CohereModel.py | 10 +++++- .../models/GoogleGenerativeAIModel.py | 14 ++++++++- .../components/models/HuggingFaceModel.py | 10 ++++++ .../langflow/components/models/OllamaModel.py | 31 +++++++++++++++++++ .../langflow/components/models/OpenAIModel.py | 12 +++++++ .../components/models/VertexAiModel.py | 16 ++++++++++ 9 files changed, 128 insertions(+), 2 deletions(-) diff --git a/src/backend/base/langflow/components/models/AnthropicModel.py b/src/backend/base/langflow/components/models/AnthropicModel.py index 63ed04beb..c7bf813ef 100644 --- a/src/backend/base/langflow/components/models/AnthropicModel.py +++ b/src/backend/base/langflow/components/models/AnthropicModel.py @@ -12,6 +12,17 @@ class AnthropicLLM(LCModelComponent): description: str = "Generate text using Anthropic Chat&Completion LLMs." icon = "Anthropic" + field_order = [ + "model", + "anthropic_api_key", + "max_tokens", + "temperature", + "anthropic_api_url", + "input_value", + "system_message", + "stream", + ] + def build_config(self): return { "model": { diff --git a/src/backend/base/langflow/components/models/AzureOpenAIModel.py b/src/backend/base/langflow/components/models/AzureOpenAIModel.py index 3f86214b7..a697d3b68 100644 --- a/src/backend/base/langflow/components/models/AzureOpenAIModel.py +++ b/src/backend/base/langflow/components/models/AzureOpenAIModel.py @@ -14,6 +14,19 @@ class AzureChatOpenAIComponent(LCModelComponent): beta = False icon = "Azure" + field_order = [ + "model", + "azure_endpoint", + "azure_deployment", + "api_version", + "api_key", + "temperature", + "max_tokens", + "input_value", + "system_message", + "stream", + ] + AZURE_OPENAI_MODELS = [ "gpt-35-turbo", "gpt-35-turbo-16k", diff --git a/src/backend/base/langflow/components/models/BaiduQianfanChatModel.py b/src/backend/base/langflow/components/models/BaiduQianfanChatModel.py index b427b2fa7..3618b54ba 100644 --- a/src/backend/base/langflow/components/models/BaiduQianfanChatModel.py +++ b/src/backend/base/langflow/components/models/BaiduQianfanChatModel.py @@ -13,6 +13,19 @@ class QianfanChatEndpointComponent(LCModelComponent): documentation: str = "https://python.langchain.com/docs/integrations/chat/baidu_qianfan_endpoint." icon = "BaiduQianfan" + field_order = [ + "model", + "qianfan_ak", + "qianfan_sk", + "top_p", + "temperature", + "penalty_score", + "endpoint", + "input_value", + "system_message", + "stream", + ] + def build_config(self): return { "model": { diff --git a/src/backend/base/langflow/components/models/CohereModel.py b/src/backend/base/langflow/components/models/CohereModel.py index 782aa1232..0022a5b3f 100644 --- a/src/backend/base/langflow/components/models/CohereModel.py +++ b/src/backend/base/langflow/components/models/CohereModel.py @@ -1,6 +1,5 @@ from typing import Optional - from langchain_community.chat_models.cohere import ChatCohere from pydantic.v1 import SecretStr @@ -15,6 +14,15 @@ class CohereComponent(LCModelComponent): icon = "Cohere" + field_order = [ + "cohere_api_key", + "max_tokens", + "temperature", + "input_value", + "system_message", + "stream", + ] + def build_config(self): return { "cohere_api_key": { diff --git a/src/backend/base/langflow/components/models/GoogleGenerativeAIModel.py b/src/backend/base/langflow/components/models/GoogleGenerativeAIModel.py index c250e2fc4..f49109734 100644 --- a/src/backend/base/langflow/components/models/GoogleGenerativeAIModel.py +++ b/src/backend/base/langflow/components/models/GoogleGenerativeAIModel.py @@ -11,7 +11,19 @@ class GoogleGenerativeAIComponent(LCModelComponent): display_name: str = "Google Generative AI" description: str = "Generate text using Google Generative AI." icon = "GoogleGenerativeAI" - icon = "Google" + + field_order = [ + "google_api_key", + "model", + "max_output_tokens", + "temperature", + "top_k", + "top_p", + "n", + "input_value", + "system_message", + "stream", + ] def build_config(self): return { diff --git a/src/backend/base/langflow/components/models/HuggingFaceModel.py b/src/backend/base/langflow/components/models/HuggingFaceModel.py index e7b558509..ca6b106f0 100644 --- a/src/backend/base/langflow/components/models/HuggingFaceModel.py +++ b/src/backend/base/langflow/components/models/HuggingFaceModel.py @@ -12,6 +12,16 @@ class HuggingFaceEndpointsComponent(LCModelComponent): description: str = "Generate text using Hugging Face Inference APIs." icon = "HuggingFace" + field_order = [ + "endpoint_url", + "task", + "huggingfacehub_api_token", + "model_kwargs", + "input_value", + "system_message", + "stream", + ] + def build_config(self): return { "endpoint_url": {"display_name": "Endpoint URL", "password": True}, diff --git a/src/backend/base/langflow/components/models/OllamaModel.py b/src/backend/base/langflow/components/models/OllamaModel.py index e8147bd72..7381c49aa 100644 --- a/src/backend/base/langflow/components/models/OllamaModel.py +++ b/src/backend/base/langflow/components/models/OllamaModel.py @@ -17,6 +17,37 @@ class ChatOllamaComponent(LCModelComponent): description = "Generate text using Ollama Local LLMs." icon = "Ollama" + field_order = [ + "base_url", + "model", + "temperature", + "cache", + "callback_manager", + "callbacks", + "format", + "metadata", + "mirostat", + "mirostat_eta", + "mirostat_tau", + "num_ctx", + "num_gpu", + "num_thread", + "repeat_last_n", + "repeat_penalty", + "tfs_z", + "timeout", + "top_k", + "top_p", + "verbose", + "tags", + "stop", + "system", + "template", + "input_value", + "system_message", + "stream", + ] + def build_config(self) -> dict: return { "base_url": { diff --git a/src/backend/base/langflow/components/models/OpenAIModel.py b/src/backend/base/langflow/components/models/OpenAIModel.py index b364cb8ba..27453f228 100644 --- a/src/backend/base/langflow/components/models/OpenAIModel.py +++ b/src/backend/base/langflow/components/models/OpenAIModel.py @@ -11,6 +11,18 @@ class OpenAIModelComponent(LCModelComponent): description = "Generates text using OpenAI LLMs." icon = "OpenAI" + field_order = [ + "max_tokens", + "model_kwargs", + "model_name", + "openai_api_base", + "openai_api_key", + "temperature", + "input_value", + "system_message", + "stream", + ] + def build_config(self): return { "input_value": {"display_name": "Input"}, diff --git a/src/backend/base/langflow/components/models/VertexAiModel.py b/src/backend/base/langflow/components/models/VertexAiModel.py index 843d9edc6..727864f2c 100644 --- a/src/backend/base/langflow/components/models/VertexAiModel.py +++ b/src/backend/base/langflow/components/models/VertexAiModel.py @@ -11,6 +11,22 @@ class ChatVertexAIComponent(LCModelComponent): description = "Generate text using Vertex AI LLMs." icon = "VertexAI" + field_order = [ + "credentials", + "project", + "examples", + "location", + "max_output_tokens", + "model_name", + "temperature", + "top_k", + "top_p", + "verbose", + "input_value", + "system_message", + "stream", + ] + def build_config(self): return { "credentials": {