From ca23dc4c53c56bb22ea983e61ff7260ebd06e8d6 Mon Sep 17 00:00:00 2001 From: Edwin Jose Date: Thu, 12 Jun 2025 20:42:52 -0500 Subject: [PATCH] fix(setup): preserve advanced attribute and generalize starter project JSON (#8529) * update the template * update to fix format issues --- src/backend/base/langflow/base/constants.py | 2 +- .../base/langflow/initial_setup/setup.py | 16 +- .../starter_projects/Basic Prompting.json | 407 ++++++++---------- 3 files changed, 193 insertions(+), 232 deletions(-) diff --git a/src/backend/base/langflow/base/constants.py b/src/backend/base/langflow/base/constants.py index c7752133c..7eba7d7ff 100644 --- a/src/backend/base/langflow/base/constants.py +++ b/src/backend/base/langflow/base/constants.py @@ -41,5 +41,5 @@ FIELD_FORMAT_ATTRIBUTES = [ "advanced", "copy_field", ] - +SKIPPED_FIELD_ATTRIBUTES = ["advanced"] ORJSON_OPTIONS = orjson.OPT_INDENT_2 | orjson.OPT_SORT_KEYS | orjson.OPT_OMIT_MICROSECONDS diff --git a/src/backend/base/langflow/initial_setup/setup.py b/src/backend/base/langflow/initial_setup/setup.py index a68fe8173..f6df5b244 100644 --- a/src/backend/base/langflow/initial_setup/setup.py +++ b/src/backend/base/langflow/initial_setup/setup.py @@ -26,7 +26,12 @@ from sqlalchemy.orm import selectinload from sqlmodel import col, select from sqlmodel.ext.asyncio.session import AsyncSession -from langflow.base.constants import FIELD_FORMAT_ATTRIBUTES, NODE_FORMAT_ATTRIBUTES, ORJSON_OPTIONS +from langflow.base.constants import ( + FIELD_FORMAT_ATTRIBUTES, + NODE_FORMAT_ATTRIBUTES, + ORJSON_OPTIONS, + SKIPPED_FIELD_ATTRIBUTES, +) from langflow.initial_setup.constants import STARTER_FOLDER_DESCRIPTION, STARTER_FOLDER_NAME from langflow.services.auth.utils import create_super_user from langflow.services.database.models.flow.model import Flow, FlowCreate @@ -130,7 +135,16 @@ def update_projects_components_with_latest_component_versions(project_data, all_ continue # The idea here is to update some attributes of the field to_check_attributes = FIELD_FORMAT_ATTRIBUTES + # Skip specific field attributes that should respect the starter project template values. + # Currently we skip 'advanced' so that a field marked as advanced in the component code + # will NOT overwrite the value specified in the starter project template. This preserves + # the intended UX configuration of the starter projects. + # SKIPPED_FIELD_ATTRIBUTES = {"advanced"} + # Iterate through the attributes we want to potentially update for attr in to_check_attributes: + # Respect the template value by not updating if the attribute is in the skipped set + if attr in SKIPPED_FIELD_ATTRIBUTES: + continue if ( attr in field_dict and attr in node_data["template"].get(field_name) diff --git a/src/backend/base/langflow/initial_setup/starter_projects/Basic Prompting.json b/src/backend/base/langflow/initial_setup/starter_projects/Basic Prompting.json index 5c2636613..d3f69cfce 100644 --- a/src/backend/base/langflow/initial_setup/starter_projects/Basic Prompting.json +++ b/src/backend/base/langflow/initial_setup/starter_projects/Basic Prompting.json @@ -2,37 +2,12 @@ "data": { "edges": [ { - "className": "", - "data": { - "sourceHandle": { - "dataType": "ChatInput", - "id": "ChatInput-jFwUm", - "name": "message", - "output_types": [ - "Message" - ] - }, - "targetHandle": { - "fieldName": "input_value", - "id": "OpenAIModel-OcXkl", - "inputTypes": [ - "Message" - ], - "type": "str" - } - }, - "id": "reactflow__edge-ChatInput-jFwUm{œdataTypeœ:œChatInputœ,œidœ:œChatInput-jFwUmœ,œnameœ:œmessageœ,œoutput_typesœ:[œMessageœ]}-OpenAIModel-OcXkl{œfieldNameœ:œinput_valueœ,œidœ:œOpenAIModel-OcXklœ,œinputTypesœ:[œMessageœ],œtypeœ:œstrœ}", - "source": "ChatInput-jFwUm", - "sourceHandle": "{œdataTypeœ: œChatInputœ, œidœ: œChatInput-jFwUmœ, œnameœ: œmessageœ, œoutput_typesœ: [œMessageœ]}", - "target": "OpenAIModel-OcXkl", - "targetHandle": "{œfieldNameœ: œinput_valueœ, œidœ: œOpenAIModel-OcXklœ, œinputTypesœ: [œMessageœ], œtypeœ: œstrœ}" - }, - { + "animated": false, "className": "", "data": { "sourceHandle": { "dataType": "Prompt", - "id": "Prompt-3SM2g", + "id": "Prompt-Z1IXr", "name": "prompt", "output_types": [ "Message" @@ -40,25 +15,55 @@ }, "targetHandle": { "fieldName": "system_message", - "id": "OpenAIModel-OcXkl", + "id": "LanguageModelComponent-UcEoU", "inputTypes": [ "Message" ], "type": "str" } }, - "id": "reactflow__edge-Prompt-3SM2g{œdataTypeœ:œPromptœ,œidœ:œPrompt-3SM2gœ,œnameœ:œpromptœ,œoutput_typesœ:[œMessageœ]}-OpenAIModel-OcXkl{œfieldNameœ:œsystem_messageœ,œidœ:œOpenAIModel-OcXklœ,œinputTypesœ:[œMessageœ],œtypeœ:œstrœ}", - "source": "Prompt-3SM2g", - "sourceHandle": "{œdataTypeœ: œPromptœ, œidœ: œPrompt-3SM2gœ, œnameœ: œpromptœ, œoutput_typesœ: [œMessageœ]}", - "target": "OpenAIModel-OcXkl", - "targetHandle": "{œfieldNameœ: œsystem_messageœ, œidœ: œOpenAIModel-OcXklœ, œinputTypesœ: [œMessageœ], œtypeœ: œstrœ}" + "id": "xy-edge__Prompt-Z1IXr{œdataTypeœ:œPromptœ,œidœ:œPrompt-Z1IXrœ,œnameœ:œpromptœ,œoutput_typesœ:[œMessageœ]}-LanguageModelComponent-UcEoU{œfieldNameœ:œsystem_messageœ,œidœ:œLanguageModelComponent-UcEoUœ,œinputTypesœ:[œMessageœ],œtypeœ:œstrœ}", + "selected": false, + "source": "Prompt-Z1IXr", + "sourceHandle": "{œdataTypeœ: œPromptœ, œidœ: œPrompt-Z1IXrœ, œnameœ: œpromptœ, œoutput_typesœ: [œMessageœ]}", + "target": "LanguageModelComponent-UcEoU", + "targetHandle": "{œfieldNameœ: œsystem_messageœ, œidœ: œLanguageModelComponent-UcEoUœ, œinputTypesœ: [œMessageœ], œtypeœ: œstrœ}" }, { + "animated": false, "className": "", "data": { "sourceHandle": { - "dataType": "OpenAIModel", - "id": "OpenAIModel-OcXkl", + "dataType": "ChatInput", + "id": "ChatInput-9e2tW", + "name": "message", + "output_types": [ + "Message" + ] + }, + "targetHandle": { + "fieldName": "input_value", + "id": "LanguageModelComponent-UcEoU", + "inputTypes": [ + "Message" + ], + "type": "str" + } + }, + "id": "xy-edge__ChatInput-9e2tW{œdataTypeœ:œChatInputœ,œidœ:œChatInput-9e2tWœ,œnameœ:œmessageœ,œoutput_typesœ:[œMessageœ]}-LanguageModelComponent-UcEoU{œfieldNameœ:œinput_valueœ,œidœ:œLanguageModelComponent-UcEoUœ,œinputTypesœ:[œMessageœ],œtypeœ:œstrœ}", + "selected": false, + "source": "ChatInput-9e2tW", + "sourceHandle": "{œdataTypeœ: œChatInputœ, œidœ: œChatInput-9e2tWœ, œnameœ: œmessageœ, œoutput_typesœ: [œMessageœ]}", + "target": "LanguageModelComponent-UcEoU", + "targetHandle": "{œfieldNameœ: œinput_valueœ, œidœ: œLanguageModelComponent-UcEoUœ, œinputTypesœ: [œMessageœ], œtypeœ: œstrœ}" + }, + { + "animated": false, + "className": "", + "data": { + "sourceHandle": { + "dataType": "LanguageModelComponent", + "id": "LanguageModelComponent-UcEoU", "name": "text_output", "output_types": [ "Message" @@ -66,20 +71,21 @@ }, "targetHandle": { "fieldName": "input_value", - "id": "ChatOutput-gDYiJ", + "id": "ChatOutput-7q74v", "inputTypes": [ "Data", "DataFrame", "Message" ], - "type": "str" + "type": "other" } }, - "id": "reactflow__edge-OpenAIModel-OcXkl{œdataTypeœ:œOpenAIModelœ,œidœ:œOpenAIModel-OcXklœ,œnameœ:œtext_outputœ,œoutput_typesœ:[œMessageœ]}-ChatOutput-gDYiJ{œfieldNameœ:œinput_valueœ,œidœ:œChatOutput-gDYiJœ,œinputTypesœ:[œMessageœ],œtypeœ:œstrœ}", - "source": "OpenAIModel-OcXkl", - "sourceHandle": "{œdataTypeœ: œOpenAIModelœ, œidœ: œOpenAIModel-OcXklœ, œnameœ: œtext_outputœ, œoutput_typesœ: [œMessageœ]}", - "target": "ChatOutput-gDYiJ", - "targetHandle": "{œfieldNameœ: œinput_valueœ, œidœ: œChatOutput-gDYiJœ, œinputTypesœ: [œDataœ, œDataFrameœ, œMessageœ], œtypeœ: œstrœ}" + "id": "xy-edge__LanguageModelComponent-UcEoU{œdataTypeœ:œLanguageModelComponentœ,œidœ:œLanguageModelComponent-UcEoUœ,œnameœ:œtext_outputœ,œoutput_typesœ:[œMessageœ]}-ChatOutput-7q74v{œfieldNameœ:œinput_valueœ,œidœ:œChatOutput-7q74vœ,œinputTypesœ:[œDataœ,œDataFrameœ,œMessageœ],œtypeœ:œotherœ}", + "selected": false, + "source": "LanguageModelComponent-UcEoU", + "sourceHandle": "{œdataTypeœ: œLanguageModelComponentœ, œidœ: œLanguageModelComponent-UcEoUœ, œnameœ: œtext_outputœ, œoutput_typesœ: [œMessageœ]}", + "target": "ChatOutput-7q74v", + "targetHandle": "{œfieldNameœ: œinput_valueœ, œidœ: œChatOutput-7q74vœ, œinputTypesœ: [œDataœ, œDataFrameœ, œMessageœ], œtypeœ: œotherœ}" } ], "nodes": [ @@ -87,7 +93,7 @@ "data": { "description": "Get chat inputs from the Playground.", "display_name": "Chat Input", - "id": "ChatInput-jFwUm", + "id": "ChatInput-9e2tW", "node": { "base_classes": [ "Message" @@ -110,7 +116,7 @@ "frozen": false, "icon": "MessagesSquare", "legacy": false, - "lf_version": "1.0.19.post2", + "lf_version": "1.4.3", "metadata": {}, "output_types": [], "outputs": [ @@ -142,12 +148,14 @@ "Message" ], "list": false, + "list_add_label": "Add More", "load_from_db": false, "name": "background_color", "placeholder": "", "required": false, "show": true, "title_case": false, + "tool_mode": false, "trace_as_input": true, "trace_as_metadata": true, "type": "str", @@ -163,12 +171,14 @@ "Message" ], "list": false, + "list_add_label": "Add More", "load_from_db": false, "name": "chat_icon", "placeholder": "", "required": false, "show": true, "title_case": false, + "tool_mode": false, "trace_as_input": true, "trace_as_metadata": true, "type": "str", @@ -193,6 +203,7 @@ "value": "from langflow.base.data.utils import IMG_FILE_TYPES, TEXT_FILE_TYPES\nfrom langflow.base.io.chat import ChatComponent\nfrom langflow.inputs.inputs import BoolInput\nfrom langflow.io import (\n DropdownInput,\n FileInput,\n MessageTextInput,\n MultilineInput,\n Output,\n)\nfrom langflow.schema.message import Message\nfrom langflow.utils.constants import (\n MESSAGE_SENDER_AI,\n MESSAGE_SENDER_NAME_USER,\n MESSAGE_SENDER_USER,\n)\n\n\nclass ChatInput(ChatComponent):\n display_name = \"Chat Input\"\n description = \"Get chat inputs from the Playground.\"\n icon = \"MessagesSquare\"\n name = \"ChatInput\"\n minimized = True\n\n inputs = [\n MultilineInput(\n name=\"input_value\",\n display_name=\"Input Text\",\n value=\"\",\n info=\"Message to be passed as input.\",\n input_types=[],\n ),\n BoolInput(\n name=\"should_store_message\",\n display_name=\"Store Messages\",\n info=\"Store the message in the history.\",\n value=True,\n advanced=True,\n ),\n DropdownInput(\n name=\"sender\",\n display_name=\"Sender Type\",\n options=[MESSAGE_SENDER_AI, MESSAGE_SENDER_USER],\n value=MESSAGE_SENDER_USER,\n info=\"Type of sender.\",\n advanced=True,\n ),\n MessageTextInput(\n name=\"sender_name\",\n display_name=\"Sender Name\",\n info=\"Name of the sender.\",\n value=MESSAGE_SENDER_NAME_USER,\n advanced=True,\n ),\n MessageTextInput(\n name=\"session_id\",\n display_name=\"Session ID\",\n info=\"The session ID of the chat. If empty, the current session ID parameter will be used.\",\n advanced=True,\n ),\n FileInput(\n name=\"files\",\n display_name=\"Files\",\n file_types=TEXT_FILE_TYPES + IMG_FILE_TYPES,\n info=\"Files to be sent with the message.\",\n advanced=True,\n is_list=True,\n temp_file=True,\n ),\n MessageTextInput(\n name=\"background_color\",\n display_name=\"Background Color\",\n info=\"The background color of the icon.\",\n advanced=True,\n ),\n MessageTextInput(\n name=\"chat_icon\",\n display_name=\"Icon\",\n info=\"The icon of the message.\",\n advanced=True,\n ),\n MessageTextInput(\n name=\"text_color\",\n display_name=\"Text Color\",\n info=\"The text color of the name\",\n advanced=True,\n ),\n ]\n outputs = [\n Output(display_name=\"Chat Message\", name=\"message\", method=\"message_response\"),\n ]\n\n async def message_response(self) -> Message:\n background_color = self.background_color\n text_color = self.text_color\n icon = self.chat_icon\n\n message = await Message.create(\n text=self.input_value,\n sender=self.sender,\n sender_name=self.sender_name,\n session_id=self.session_id,\n files=self.files,\n properties={\n \"background_color\": background_color,\n \"text_color\": text_color,\n \"icon\": icon,\n },\n )\n if self.session_id and isinstance(message, Message) and self.should_store_message:\n stored_message = await self.send_message(\n message,\n )\n self.message.value = stored_message\n message = stored_message\n\n self.status = message\n return message\n" }, "files": { + "_input_type": "FileInput", "advanced": true, "display_name": "Files", "dynamic": false, @@ -224,6 +235,7 @@ "file_path": "", "info": "Files to be sent with the message.", "list": true, + "list_add_label": "Add More", "name": "files", "placeholder": "", "required": false, @@ -235,12 +247,15 @@ "value": "" }, "input_value": { + "_input_type": "MultilineInput", "advanced": false, + "copy_field": false, "display_name": "Input Text", "dynamic": false, "info": "Message to be passed as input.", "input_types": [], "list": false, + "list_add_label": "Add More", "load_from_db": false, "multiline": true, "name": "input_value", @@ -248,13 +263,17 @@ "required": false, "show": true, "title_case": false, + "tool_mode": false, "trace_as_input": true, "trace_as_metadata": true, "type": "str", "value": "Hello" }, "sender": { + "_input_type": "DropdownInput", "advanced": true, + "combobox": false, + "dialog_inputs": {}, "display_name": "Sender Type", "dynamic": false, "info": "Type of sender.", @@ -263,15 +282,19 @@ "Machine", "User" ], + "options_metadata": [], "placeholder": "", "required": false, "show": true, "title_case": false, + "toggle": false, + "tool_mode": false, "trace_as_metadata": true, "type": "str", "value": "User" }, "sender_name": { + "_input_type": "MessageTextInput", "advanced": true, "display_name": "Sender Name", "dynamic": false, @@ -280,18 +303,21 @@ "Message" ], "list": false, + "list_add_label": "Add More", "load_from_db": false, "name": "sender_name", "placeholder": "", "required": false, "show": true, "title_case": false, + "tool_mode": false, "trace_as_input": true, "trace_as_metadata": true, "type": "str", "value": "User" }, "session_id": { + "_input_type": "MessageTextInput", "advanced": true, "display_name": "Session ID", "dynamic": false, @@ -300,12 +326,14 @@ "Message" ], "list": false, + "list_add_label": "Add More", "load_from_db": false, "name": "session_id", "placeholder": "", "required": false, "show": true, "title_case": false, + "tool_mode": false, "trace_as_input": true, "trace_as_metadata": true, "type": "str", @@ -318,11 +346,13 @@ "dynamic": false, "info": "Store the message in the history.", "list": false, + "list_add_label": "Add More", "name": "should_store_message", "placeholder": "", "required": false, "show": true, "title_case": false, + "tool_mode": false, "trace_as_metadata": true, "type": "bool", "value": true @@ -337,12 +367,14 @@ "Message" ], "list": false, + "list_add_label": "Add More", "load_from_db": false, "name": "text_color", "placeholder": "", "required": false, "show": true, "title_case": false, + "tool_mode": false, "trace_as_input": true, "trace_as_metadata": true, "type": "str", @@ -354,10 +386,10 @@ }, "dragging": false, "height": 234, - "id": "ChatInput-jFwUm", + "id": "ChatInput-9e2tW", "measured": { "height": 234, - "width": 360 + "width": 320 }, "position": { "x": 689.5720422421635, @@ -375,7 +407,7 @@ "data": { "description": "Create a prompt template with dynamic variables.", "display_name": "Prompt", - "id": "Prompt-3SM2g", + "id": "Prompt-Z1IXr", "node": { "base_classes": [ "Message" @@ -395,6 +427,7 @@ "frozen": false, "icon": "braces", "legacy": false, + "lf_version": "1.4.3", "metadata": {}, "output_types": [], "outputs": [ @@ -441,6 +474,7 @@ "dynamic": false, "info": "", "list": false, + "list_add_label": "Add More", "load_from_db": false, "name": "template", "placeholder": "", @@ -462,6 +496,7 @@ "Message" ], "list": false, + "list_add_label": "Add More", "load_from_db": false, "name": "tool_placeholder", "placeholder": "", @@ -481,10 +516,10 @@ }, "dragging": false, "height": 260, - "id": "Prompt-3SM2g", + "id": "Prompt-Z1IXr", "measured": { "height": 260, - "width": 360 + "width": 320 }, "position": { "x": 690.2015147036818, @@ -500,7 +535,7 @@ }, { "data": { - "id": "undefined-kVLkG", + "id": "undefined-9ON0z", "node": { "description": "## 📖 README\n\nPerform basic prompting with an OpenAI model.\n\n#### Quick Start\n- Add your **OpenAI API key** to the **OpenAI Model**\n- Open the **Playground** to chat with your bot.\n\n#### Next steps:\n Experiment by changing the prompt and the OpenAI model temperature to see how the bot's responses change.", "display_name": "Read Me", @@ -512,10 +547,10 @@ }, "dragging": false, "height": 332, - "id": "undefined-kVLkG", + "id": "undefined-9ON0z", "measured": { "height": 332, - "width": 328 + "width": 325 }, "position": { "x": 133.95771636602308, @@ -536,7 +571,7 @@ }, { "data": { - "id": "note-WrEGM", + "id": "note-gMO7f", "node": { "description": "### 💡 Add your OpenAI API key here 👇", "display_name": "", @@ -549,10 +584,10 @@ }, "dragging": false, "height": 324, - "id": "note-WrEGM", + "id": "note-gMO7f", "measured": { "height": 324, - "width": 326 + "width": 324 }, "position": { "x": 1075.829573520873, @@ -573,7 +608,7 @@ }, { "data": { - "id": "ChatOutput-gDYiJ", + "id": "ChatOutput-7q74v", "node": { "base_classes": [ "Message" @@ -599,6 +634,7 @@ "frozen": false, "icon": "MessagesSquare", "legacy": false, + "lf_version": "1.4.3", "metadata": {}, "output_types": [], "outputs": [ @@ -630,6 +666,7 @@ "Message" ], "list": false, + "list_add_label": "Add More", "load_from_db": false, "name": "background_color", "placeholder": "", @@ -652,6 +689,7 @@ "Message" ], "list": false, + "list_add_label": "Add More", "load_from_db": false, "name": "chat_icon", "placeholder": "", @@ -710,6 +748,7 @@ "Message" ], "list": false, + "list_add_label": "Add More", "load_from_db": false, "name": "data_template", "placeholder": "", @@ -723,7 +762,7 @@ "value": "{text}" }, "input_value": { - "_input_type": "MessageInput", + "_input_type": "HandleInput", "advanced": false, "display_name": "Inputs", "dynamic": false, @@ -734,6 +773,7 @@ "Message" ], "list": false, + "list_add_label": "Add More", "load_from_db": false, "name": "input_value", "placeholder": "", @@ -742,13 +782,14 @@ "title_case": false, "trace_as_input": true, "trace_as_metadata": true, - "type": "str", + "type": "other", "value": "" }, "sender": { "_input_type": "DropdownInput", "advanced": true, "combobox": false, + "dialog_inputs": {}, "display_name": "Sender Type", "dynamic": false, "info": "Type of sender.", @@ -757,10 +798,12 @@ "Machine", "User" ], + "options_metadata": [], "placeholder": "", "required": false, "show": true, "title_case": false, + "toggle": false, "tool_mode": false, "trace_as_metadata": true, "type": "str", @@ -776,6 +819,7 @@ "Message" ], "list": false, + "list_add_label": "Add More", "load_from_db": false, "name": "sender_name", "placeholder": "", @@ -798,6 +842,7 @@ "Message" ], "list": false, + "list_add_label": "Add More", "load_from_db": false, "name": "session_id", "placeholder": "", @@ -817,11 +862,13 @@ "dynamic": false, "info": "Store the message in the history.", "list": false, + "list_add_label": "Add More", "name": "should_store_message", "placeholder": "", "required": false, "show": true, "title_case": false, + "tool_mode": false, "trace_as_metadata": true, "type": "bool", "value": true @@ -836,6 +883,7 @@ "Message" ], "list": false, + "list_add_label": "Add More", "load_from_db": false, "name": "text_color", "placeholder": "", @@ -855,10 +903,10 @@ }, "dragging": false, "height": 234, - "id": "ChatOutput-gDYiJ", + "id": "ChatOutput-7q74v", "measured": { "height": 234, - "width": 360 + "width": 320 }, "position": { "x": 1460.070372772908, @@ -868,43 +916,40 @@ "x": 1444.936881624563, "y": 872.7273956769025 }, - "selected": false, + "selected": true, "type": "genericNode", "width": 320 }, { "data": { - "id": "OpenAIModel-OcXkl", + "id": "LanguageModelComponent-UcEoU", "node": { "base_classes": [ "LanguageModel", "Message" ], "beta": false, + "category": "models", "conditional_paths": [], "custom_fields": {}, - "description": "Generates text using OpenAI LLMs.", - "display_name": "OpenAI", + "description": "Runs a language model given a specified provider. ", + "display_name": "Language Model", "documentation": "", "edited": false, "field_order": [ + "provider", + "model_name", + "api_key", "input_value", "system_message", "stream", - "max_tokens", - "model_kwargs", - "json_mode", - "model_name", - "openai_api_base", - "api_key", - "temperature", - "seed", - "max_retries", - "timeout" + "temperature" ], "frozen": false, - "icon": "OpenAI", + "icon": "brain-circuit", + "key": "LanguageModelComponent", "legacy": false, + "lf_version": "1.4.3", "metadata": { "keywords": [ "model", @@ -946,6 +991,8 @@ } ], "pinned": false, + "priority": 0, + "score": 0.28173906304863156, "template": { "_type": "Component", "api_key": { @@ -953,17 +1000,18 @@ "advanced": false, "display_name": "OpenAI API Key", "dynamic": false, - "info": "The OpenAI API Key to use for the OpenAI model.", + "info": "Model Provider API key", "input_types": [], "load_from_db": true, "name": "api_key", "password": true, "placeholder": "", - "required": true, + "real_time_refresh": true, + "required": false, "show": true, "title_case": false, "type": "str", - "value": "OPENAI_API_KEY" + "value": "" }, "code": { "advanced": true, @@ -981,14 +1029,14 @@ "show": true, "title_case": false, "type": "code", - "value": "from typing import Any\n\nfrom langchain_openai import ChatOpenAI\nfrom pydantic.v1 import SecretStr\n\nfrom langflow.base.models.model import LCModelComponent\nfrom langflow.base.models.openai_constants import (\n OPENAI_MODEL_NAMES,\n OPENAI_REASONING_MODEL_NAMES,\n)\nfrom langflow.field_typing import LanguageModel\nfrom langflow.field_typing.range_spec import RangeSpec\nfrom langflow.inputs.inputs import BoolInput, DictInput, DropdownInput, IntInput, SecretStrInput, SliderInput, StrInput\nfrom langflow.logging import logger\n\n\nclass OpenAIModelComponent(LCModelComponent):\n display_name = \"OpenAI\"\n description = \"Generates text using OpenAI LLMs.\"\n icon = \"OpenAI\"\n name = \"OpenAIModel\"\n\n inputs = [\n *LCModelComponent._base_inputs,\n IntInput(\n name=\"max_tokens\",\n display_name=\"Max Tokens\",\n advanced=True,\n info=\"The maximum number of tokens to generate. Set to 0 for unlimited tokens.\",\n range_spec=RangeSpec(min=0, max=128000),\n ),\n DictInput(\n name=\"model_kwargs\",\n display_name=\"Model Kwargs\",\n advanced=True,\n info=\"Additional keyword arguments to pass to the model.\",\n ),\n BoolInput(\n name=\"json_mode\",\n display_name=\"JSON Mode\",\n advanced=True,\n info=\"If True, it will output JSON regardless of passing a schema.\",\n ),\n DropdownInput(\n name=\"model_name\",\n display_name=\"Model Name\",\n advanced=False,\n options=OPENAI_MODEL_NAMES + OPENAI_REASONING_MODEL_NAMES,\n value=OPENAI_MODEL_NAMES[1],\n combobox=True,\n real_time_refresh=True,\n ),\n StrInput(\n name=\"openai_api_base\",\n display_name=\"OpenAI API Base\",\n advanced=True,\n info=\"The base URL of the OpenAI API. \"\n \"Defaults to https://api.openai.com/v1. \"\n \"You can change this to use other APIs like JinaChat, LocalAI and Prem.\",\n ),\n SecretStrInput(\n name=\"api_key\",\n display_name=\"OpenAI API Key\",\n info=\"The OpenAI API Key to use for the OpenAI model.\",\n advanced=False,\n value=\"OPENAI_API_KEY\",\n required=True,\n ),\n SliderInput(\n name=\"temperature\",\n display_name=\"Temperature\",\n value=0.1,\n range_spec=RangeSpec(min=0, max=1, step=0.01),\n show=True,\n ),\n IntInput(\n name=\"seed\",\n display_name=\"Seed\",\n info=\"The seed controls the reproducibility of the job.\",\n advanced=True,\n value=1,\n ),\n IntInput(\n name=\"max_retries\",\n display_name=\"Max Retries\",\n info=\"The maximum number of retries to make when generating.\",\n advanced=True,\n value=5,\n ),\n IntInput(\n name=\"timeout\",\n display_name=\"Timeout\",\n info=\"The timeout for requests to OpenAI completion API.\",\n advanced=True,\n value=700,\n ),\n ]\n\n def build_model(self) -> LanguageModel: # type: ignore[type-var]\n parameters = {\n \"api_key\": SecretStr(self.api_key).get_secret_value() if self.api_key else None,\n \"model_name\": self.model_name,\n \"max_tokens\": self.max_tokens or None,\n \"model_kwargs\": self.model_kwargs or {},\n \"base_url\": self.openai_api_base or \"https://api.openai.com/v1\",\n \"seed\": self.seed,\n \"max_retries\": self.max_retries,\n \"timeout\": self.timeout,\n \"temperature\": self.temperature if self.temperature is not None else 0.1,\n }\n\n logger.info(f\"Model name: {self.model_name}\")\n if self.model_name in OPENAI_REASONING_MODEL_NAMES:\n logger.info(\"Getting reasoning model parameters\")\n parameters.pop(\"temperature\")\n parameters.pop(\"seed\")\n output = ChatOpenAI(**parameters)\n if self.json_mode:\n output = output.bind(response_format={\"type\": \"json_object\"})\n\n return output\n\n def _get_exception_message(self, e: Exception):\n \"\"\"Get a message from an OpenAI exception.\n\n Args:\n e (Exception): The exception to get the message from.\n\n Returns:\n str: The message from the exception.\n \"\"\"\n try:\n from openai import BadRequestError\n except ImportError:\n return None\n if isinstance(e, BadRequestError):\n message = e.body.get(\"message\")\n if message:\n return message\n return None\n\n def update_build_config(self, build_config: dict, field_value: Any, field_name: str | None = None) -> dict:\n if field_name in {\"base_url\", \"model_name\", \"api_key\"} and field_value in OPENAI_REASONING_MODEL_NAMES:\n build_config[\"temperature\"][\"show\"] = False\n build_config[\"seed\"][\"show\"] = False\n if field_name in {\"base_url\", \"model_name\", \"api_key\"} and field_value in OPENAI_MODEL_NAMES:\n build_config[\"temperature\"][\"show\"] = True\n build_config[\"seed\"][\"show\"] = True\n return build_config\n" + "value": "from typing import Any\n\nfrom langchain_anthropic import ChatAnthropic\nfrom langchain_google_genai import ChatGoogleGenerativeAI\nfrom langchain_openai import ChatOpenAI\n\nfrom langflow.base.models.anthropic_constants import ANTHROPIC_MODELS\nfrom langflow.base.models.google_generative_ai_constants import GOOGLE_GENERATIVE_AI_MODELS\nfrom langflow.base.models.model import LCModelComponent\nfrom langflow.base.models.openai_constants import OPENAI_MODEL_NAMES\nfrom langflow.field_typing import LanguageModel\nfrom langflow.field_typing.range_spec import RangeSpec\nfrom langflow.inputs.inputs import BoolInput\nfrom langflow.io import DropdownInput, MessageTextInput, SecretStrInput, SliderInput\nfrom langflow.schema.dotdict import dotdict\n\n\nclass LanguageModelComponent(LCModelComponent):\n display_name = \"Language Model\"\n description = \"Runs a language model given a specified provider. \"\n icon = \"brain-circuit\"\n category = \"models\"\n priority = 0 # Set priority to 0 to make it appear first\n\n inputs = [\n DropdownInput(\n name=\"provider\",\n display_name=\"Model Provider\",\n options=[\"OpenAI\", \"Anthropic\", \"Google\"],\n value=\"OpenAI\",\n info=\"Select the model provider\",\n real_time_refresh=True,\n options_metadata=[{\"icon\": \"OpenAI\"}, {\"icon\": \"Anthropic\"}, {\"icon\": \"Google\"}],\n ),\n DropdownInput(\n name=\"model_name\",\n display_name=\"Model Name\",\n options=OPENAI_MODEL_NAMES,\n value=OPENAI_MODEL_NAMES[0],\n info=\"Select the model to use\",\n ),\n SecretStrInput(\n name=\"api_key\",\n display_name=\"OpenAI API Key\",\n info=\"Model Provider API key\",\n required=False,\n show=True,\n real_time_refresh=True,\n ),\n MessageTextInput(\n name=\"input_value\",\n display_name=\"Input\",\n info=\"The input text to send to the model\",\n ),\n MessageTextInput(\n name=\"system_message\",\n display_name=\"System Message\",\n info=\"A system message that helps set the behavior of the assistant\",\n advanced=True,\n ),\n BoolInput(\n name=\"stream\",\n display_name=\"Stream\",\n info=\"Whether to stream the response\",\n value=False,\n advanced=True,\n ),\n SliderInput(\n name=\"temperature\",\n display_name=\"Temperature\",\n value=0.1,\n info=\"Controls randomness in responses\",\n range_spec=RangeSpec(min=0, max=1, step=0.01),\n advanced=True,\n ),\n ]\n\n def build_model(self) -> LanguageModel:\n provider = self.provider\n model_name = self.model_name\n temperature = self.temperature\n stream = self.stream\n\n if provider == \"OpenAI\":\n if not self.api_key:\n msg = \"OpenAI API key is required when using OpenAI provider\"\n raise ValueError(msg)\n return ChatOpenAI(\n model_name=model_name,\n temperature=temperature,\n streaming=stream,\n openai_api_key=self.api_key,\n )\n if provider == \"Anthropic\":\n if not self.api_key:\n msg = \"Anthropic API key is required when using Anthropic provider\"\n raise ValueError(msg)\n return ChatAnthropic(\n model=model_name,\n temperature=temperature,\n streaming=stream,\n anthropic_api_key=self.api_key,\n )\n if provider == \"Google\":\n if not self.api_key:\n msg = \"Google API key is required when using Google provider\"\n raise ValueError(msg)\n return ChatGoogleGenerativeAI(\n model=model_name,\n temperature=temperature,\n streaming=stream,\n google_api_key=self.api_key,\n )\n msg = f\"Unknown provider: {provider}\"\n raise ValueError(msg)\n\n def update_build_config(self, build_config: dotdict, field_value: Any, field_name: str | None = None) -> dotdict:\n if field_name == \"provider\":\n if field_value == \"OpenAI\":\n build_config[\"model_name\"][\"options\"] = OPENAI_MODEL_NAMES\n build_config[\"model_name\"][\"value\"] = OPENAI_MODEL_NAMES[0]\n build_config[\"api_key\"][\"display_name\"] = \"OpenAI API Key\"\n elif field_value == \"Anthropic\":\n build_config[\"model_name\"][\"options\"] = ANTHROPIC_MODELS\n build_config[\"model_name\"][\"value\"] = ANTHROPIC_MODELS[0]\n build_config[\"api_key\"][\"display_name\"] = \"Anthropic API Key\"\n elif field_value == \"Google\":\n build_config[\"model_name\"][\"options\"] = GOOGLE_GENERATIVE_AI_MODELS\n build_config[\"model_name\"][\"value\"] = GOOGLE_GENERATIVE_AI_MODELS[0]\n build_config[\"api_key\"][\"display_name\"] = \"Google API Key\"\n return build_config\n" }, "input_value": { - "_input_type": "MessageInput", + "_input_type": "MessageTextInput", "advanced": false, "display_name": "Input", "dynamic": false, - "info": "", + "info": "The input text to send to the model", "input_types": [ "Message" ], @@ -1006,92 +1054,14 @@ "type": "str", "value": "" }, - "json_mode": { - "_input_type": "BoolInput", - "advanced": true, - "display_name": "JSON Mode", - "dynamic": false, - "info": "If True, it will output JSON regardless of passing a schema.", - "list": false, - "list_add_label": "Add More", - "name": "json_mode", - "placeholder": "", - "required": false, - "show": true, - "title_case": false, - "tool_mode": false, - "trace_as_metadata": true, - "type": "bool", - "value": false - }, - "max_retries": { - "_input_type": "IntInput", - "advanced": true, - "display_name": "Max Retries", - "dynamic": false, - "info": "The maximum number of retries to make when generating.", - "list": false, - "list_add_label": "Add More", - "name": "max_retries", - "placeholder": "", - "required": false, - "show": true, - "title_case": false, - "tool_mode": false, - "trace_as_metadata": true, - "type": "int", - "value": 5 - }, - "max_tokens": { - "_input_type": "IntInput", - "advanced": true, - "display_name": "Max Tokens", - "dynamic": false, - "info": "The maximum number of tokens to generate. Set to 0 for unlimited tokens.", - "list": false, - "list_add_label": "Add More", - "name": "max_tokens", - "placeholder": "", - "range_spec": { - "max": 128000, - "min": 0, - "step": 0.1, - "step_type": "float" - }, - "required": false, - "show": true, - "title_case": false, - "tool_mode": false, - "trace_as_metadata": true, - "type": "int", - "value": "" - }, - "model_kwargs": { - "_input_type": "DictInput", - "advanced": true, - "display_name": "Model Kwargs", - "dynamic": false, - "info": "Additional keyword arguments to pass to the model.", - "list": false, - "list_add_label": "Add More", - "name": "model_kwargs", - "placeholder": "", - "required": false, - "show": true, - "title_case": false, - "tool_mode": false, - "trace_as_input": true, - "type": "dict", - "value": {} - }, "model_name": { "_input_type": "DropdownInput", "advanced": false, - "combobox": true, + "combobox": false, "dialog_inputs": {}, "display_name": "Model Name", "dynamic": false, - "info": "", + "info": "Select the model to use", "name": "model_name", "options": [ "gpt-4o-mini", @@ -1103,62 +1073,61 @@ "gpt-4-turbo", "gpt-4-turbo-preview", "gpt-4", - "gpt-3.5-turbo", - "o1" + "gpt-3.5-turbo" ], "options_metadata": [], "placeholder": "", "required": false, "show": true, "title_case": false, + "toggle": false, "tool_mode": false, "trace_as_metadata": true, "type": "str", - "value": "gpt-4.1-mini" + "value": "gpt-4o-mini" }, - "openai_api_base": { - "_input_type": "StrInput", - "advanced": true, - "display_name": "OpenAI API Base", + "provider": { + "_input_type": "DropdownInput", + "advanced": false, + "combobox": false, + "dialog_inputs": {}, + "display_name": "Model Provider", "dynamic": false, - "info": "The base URL of the OpenAI API. Defaults to https://api.openai.com/v1. You can change this to use other APIs like JinaChat, LocalAI and Prem.", - "list": false, - "list_add_label": "Add More", - "load_from_db": false, - "name": "openai_api_base", + "info": "Select the model provider", + "name": "provider", + "options": [ + "OpenAI", + "Anthropic", + "Google" + ], + "options_metadata": [ + { + "icon": "OpenAI" + }, + { + "icon": "Anthropic" + }, + { + "icon": "Google" + } + ], "placeholder": "", + "real_time_refresh": true, "required": false, "show": true, "title_case": false, + "toggle": false, "tool_mode": false, "trace_as_metadata": true, "type": "str", - "value": "" - }, - "seed": { - "_input_type": "IntInput", - "advanced": true, - "display_name": "Seed", - "dynamic": false, - "info": "The seed controls the reproducibility of the job.", - "list": false, - "list_add_label": "Add More", - "name": "seed", - "placeholder": "", - "required": false, - "show": true, - "title_case": false, - "tool_mode": false, - "trace_as_metadata": true, - "type": "int", - "value": 1 + "value": "OpenAI" }, "stream": { "_input_type": "BoolInput", "advanced": true, "display_name": "Stream", "dynamic": false, - "info": "Stream the response from the model. Streaming works only in Chat.", + "info": "Whether to stream the response", "list": false, "list_add_label": "Add More", "name": "stream", @@ -1172,18 +1141,17 @@ "value": false }, "system_message": { - "_input_type": "MultilineInput", + "_input_type": "MessageTextInput", "advanced": false, "display_name": "System Message", "dynamic": false, - "info": "System message to pass to the model.", + "info": "A system message that helps set the behavior of the assistant", "input_types": [ "Message" ], "list": false, "list_add_label": "Add More", "load_from_db": false, - "multiline": true, "name": "system_message", "placeholder": "", "required": false, @@ -1197,11 +1165,10 @@ }, "temperature": { "_input_type": "SliderInput", - "advanced": false, + "advanced": true, "display_name": "Temperature", "dynamic": false, - "info": "", - "load_from_db": false, + "info": "Controls randomness in responses", "max_label": "", "max_label_icon": "", "min_label": "", @@ -1223,58 +1190,38 @@ "tool_mode": false, "type": "slider", "value": 0.1 - }, - "timeout": { - "_input_type": "IntInput", - "advanced": true, - "display_name": "Timeout", - "dynamic": false, - "info": "The timeout for requests to OpenAI completion API.", - "list": false, - "list_add_label": "Add More", - "name": "timeout", - "placeholder": "", - "required": false, - "show": true, - "title_case": false, - "tool_mode": false, - "trace_as_metadata": true, - "type": "int", - "value": 700 } }, "tool_mode": false }, "showNode": true, - "type": "OpenAIModel" + "type": "LanguageModelComponent" }, "dragging": false, - "id": "OpenAIModel-OcXkl", + "id": "LanguageModelComponent-UcEoU", "measured": { - "height": 734, - "width": 360 + "height": 534, + "width": 320 }, "position": { - "x": 1071.3015591664102, - "y": 724.6650363109242 + "x": 1082.9866975141738, + "y": 746.0923075286877 }, "selected": false, "type": "genericNode" } ], "viewport": { - "x": -37.61270157375441, - "y": -155.91266341888854, - "zoom": 0.7575251406952855 + "x": -37.23013572397235, + "y": -311.67320873395806, + "zoom": 0.74075714647776 } }, "description": "Perform basic prompting with an OpenAI model.", "endpoint_name": null, - "gradient": "2", - "icon": "Braces", - "id": "1511c230-d446-43a7-bfc3-539e69ce05b8", + "id": "c04a179a-a582-402e-9628-4af37767208c", "is_component": false, - "last_tested_version": "1.0.19.post2", + "last_tested_version": "1.4.3", "name": "Basic Prompting", "tags": [ "chatbots"