diff --git a/src/backend/base/langflow/components/embeddings/CohereEmbeddings.py b/src/backend/base/langflow/components/embeddings/CohereEmbeddings.py index d70af7051..23d855f40 100644 --- a/src/backend/base/langflow/components/embeddings/CohereEmbeddings.py +++ b/src/backend/base/langflow/components/embeddings/CohereEmbeddings.py @@ -16,6 +16,7 @@ class CohereEmbeddingsComponent(CustomComponent): "truncate": {"display_name": "Truncate", "advanced": True}, "max_retries": {"display_name": "Max Retries", "advanced": True}, "user_agent": {"display_name": "User Agent", "advanced": True}, + "request_timeout": {"display_name": "Request Timeout", "advanced": True}, } def build( diff --git a/src/backend/base/langflow/components/embeddings/OpenAIEmbeddings.py b/src/backend/base/langflow/components/embeddings/OpenAIEmbeddings.py index 11e3e3d49..e0106eda7 100644 --- a/src/backend/base/langflow/components/embeddings/OpenAIEmbeddings.py +++ b/src/backend/base/langflow/components/embeddings/OpenAIEmbeddings.py @@ -45,12 +45,24 @@ class OpenAIEmbeddingsComponent(CustomComponent): "model": { "display_name": "Model", "advanced": False, - "options": ["text-embedding-3-small", "text-embedding-3-large", "text-embedding-ada-002"], + "options": [ + "text-embedding-3-small", + "text-embedding-3-large", + "text-embedding-ada-002", + ], }, "model_kwargs": {"display_name": "Model Kwargs", "advanced": True}, - "openai_api_base": {"display_name": "OpenAI API Base", "password": True, "advanced": True}, + "openai_api_base": { + "display_name": "OpenAI API Base", + "password": True, + "advanced": True, + }, "openai_api_key": {"display_name": "OpenAI API Key", "password": True}, - "openai_api_type": {"display_name": "OpenAI API Type", "advanced": True, "password": True}, + "openai_api_type": { + "display_name": "OpenAI API Type", + "advanced": True, + "password": True, + }, "openai_api_version": { "display_name": "OpenAI API Version", "advanced": True, @@ -66,8 +78,11 @@ class OpenAIEmbeddingsComponent(CustomComponent): "advanced": True, }, "skip_empty": {"display_name": "Skip Empty", "advanced": True}, - "tiktoken_model_name": {"display_name": "TikToken Model Name"}, - "tikToken_enable": {"display_name": "TikToken Enable", "advanced": True}, + "tiktoken_model_name": { + "display_name": "TikToken Model Name", + "advanced": True, + }, + "tiktoken_enable": {"display_name": "TikToken Enable", "advanced": True}, } def build( diff --git a/src/backend/base/langflow/components/models/AmazonBedrockModel.py b/src/backend/base/langflow/components/models/AmazonBedrockModel.py index 19a8c2d98..234924fe2 100644 --- a/src/backend/base/langflow/components/models/AmazonBedrockModel.py +++ b/src/backend/base/langflow/components/models/AmazonBedrockModel.py @@ -55,10 +55,15 @@ class AmazonBedrockComponent(LCModelComponent): "model_kwargs": {"display_name": "Model Kwargs"}, "cache": {"display_name": "Cache"}, "input_value": {"display_name": "Input"}, - "system_message": {"display_name": "System Message", "info": "System message to pass to the model."}, + "system_message": { + "display_name": "System Message", + "info": "System message to pass to the model.", + "advanced": True, + }, "stream": { "display_name": "Stream", "info": "Stream the response from the model.", + "advanced": True, }, } diff --git a/src/backend/base/langflow/components/models/AnthropicModel.py b/src/backend/base/langflow/components/models/AnthropicModel.py index c7bf813ef..837db9318 100644 --- a/src/backend/base/langflow/components/models/AnthropicModel.py +++ b/src/backend/base/langflow/components/models/AnthropicModel.py @@ -49,25 +49,29 @@ class AnthropicLLM(LCModelComponent): "max_tokens": { "display_name": "Max Tokens", "field_type": "int", + "advanced": True, "value": 256, }, "temperature": { "display_name": "Temperature", "field_type": "float", - "value": 0.7, + "value": 0.1, }, "anthropic_api_url": { "display_name": "Anthropic API URL", + "advanced": True, "info": "Endpoint of the Anthropic API. Defaults to 'https://api.anthropic.com' if not specified.", }, "code": {"show": False}, "input_value": {"display_name": "Input"}, "stream": { "display_name": "Stream", + "advanced": True, "info": "Stream the response from the model.", }, "system_message": { "display_name": "System Message", + "advanced": True, "info": "System message to pass to the model.", }, } diff --git a/src/backend/base/langflow/components/models/CohereModel.py b/src/backend/base/langflow/components/models/CohereModel.py index 0022a5b3f..d4eeaa186 100644 --- a/src/backend/base/langflow/components/models/CohereModel.py +++ b/src/backend/base/langflow/components/models/CohereModel.py @@ -29,9 +29,11 @@ class CohereComponent(LCModelComponent): "display_name": "Cohere API Key", "type": "password", "password": True, + "required": True, }, "max_tokens": { "display_name": "Max Tokens", + "advanced": True, "default": 256, "type": "int", "show": True, @@ -46,10 +48,12 @@ class CohereComponent(LCModelComponent): "stream": { "display_name": "Stream", "info": "Stream the response from the model.", + "advanced": True, }, "system_message": { "display_name": "System Message", "info": "System message to pass to the model.", + "advanced": True, }, } diff --git a/src/backend/base/langflow/components/models/GoogleGenerativeAIModel.py b/src/backend/base/langflow/components/models/GoogleGenerativeAIModel.py index 322faef22..10aad490a 100644 --- a/src/backend/base/langflow/components/models/GoogleGenerativeAIModel.py +++ b/src/backend/base/langflow/components/models/GoogleGenerativeAIModel.py @@ -34,6 +34,7 @@ class GoogleGenerativeAIComponent(LCModelComponent): "max_output_tokens": { "display_name": "Max Output Tokens", "info": "The maximum number of tokens to generate.", + "advanced": True, }, "temperature": { "display_name": "Temperature", @@ -67,10 +68,12 @@ class GoogleGenerativeAIComponent(LCModelComponent): "stream": { "display_name": "Stream", "info": "Stream the response from the model.", + "advanced": True, }, "system_message": { "display_name": "System Message", "info": "System message to pass to the model.", + "advanced": True, }, } diff --git a/src/backend/base/langflow/components/models/HuggingFaceModel.py b/src/backend/base/langflow/components/models/HuggingFaceModel.py index ca6b106f0..4230a29e4 100644 --- a/src/backend/base/langflow/components/models/HuggingFaceModel.py +++ b/src/backend/base/langflow/components/models/HuggingFaceModel.py @@ -33,16 +33,19 @@ class HuggingFaceEndpointsComponent(LCModelComponent): "model_kwargs": { "display_name": "Model Keyword Arguments", "field_type": "code", + "advanced": True, }, "code": {"show": False}, "input_value": {"display_name": "Input"}, "stream": { "display_name": "Stream", "info": "Stream the response from the model.", + "advanced": True, }, "system_message": { "display_name": "System Message", "info": "System message to pass to the model.", + "advanced": True, }, } diff --git a/src/backend/base/langflow/components/models/OllamaModel.py b/src/backend/base/langflow/components/models/OllamaModel.py index 7381c49aa..cd8296cc7 100644 --- a/src/backend/base/langflow/components/models/OllamaModel.py +++ b/src/backend/base/langflow/components/models/OllamaModel.py @@ -53,6 +53,7 @@ class ChatOllamaComponent(LCModelComponent): "base_url": { "display_name": "Base URL", "info": "Endpoint of the Ollama API. Defaults to 'http://localhost:11434' if not specified.", + "advanced": True, }, "model": { "display_name": "Model Name", @@ -205,6 +206,7 @@ class ChatOllamaComponent(LCModelComponent): "system_message": { "display_name": "System Message", "info": "System message to pass to the model.", + "advanced": True, }, } diff --git a/src/backend/base/langflow/components/models/OpenAIModel.py b/src/backend/base/langflow/components/models/OpenAIModel.py index 27453f228..5fbd1bd2a 100644 --- a/src/backend/base/langflow/components/models/OpenAIModel.py +++ b/src/backend/base/langflow/components/models/OpenAIModel.py @@ -28,7 +28,7 @@ class OpenAIModelComponent(LCModelComponent): "input_value": {"display_name": "Input"}, "max_tokens": { "display_name": "Max Tokens", - "advanced": False, + "advanced": True, "required": False, }, "model_kwargs": { @@ -51,7 +51,7 @@ class OpenAIModelComponent(LCModelComponent): }, "openai_api_base": { "display_name": "OpenAI API Base", - "advanced": False, + "advanced": True, "required": False, "info": ( "The base URL of the OpenAI API. Defaults to https://api.openai.com/v1.\n\n" @@ -61,7 +61,7 @@ class OpenAIModelComponent(LCModelComponent): "openai_api_key": { "display_name": "OpenAI API Key", "advanced": False, - "required": False, + "required": True, "password": True, }, "temperature": { @@ -73,10 +73,12 @@ class OpenAIModelComponent(LCModelComponent): "stream": { "display_name": "Stream", "info": "Stream the response from the model.", + "advanced": True, }, "system_message": { "display_name": "System Message", "info": "System message to pass to the model.", + "advanced": True, }, }