Fixed components with Hide element

This commit is contained in:
Lucas Oliveira 2024-06-12 19:42:20 -03:00
commit 3bd1a96506
5 changed files with 323 additions and 1087 deletions

View file

@ -8,17 +8,12 @@
"dataType": "OpenAIModel",
"id": "OpenAIModel-k39HS",
"name": "text_output",
"output_types": [
"Text"
]
"output_types": ["Text"]
},
"targetHandle": {
"fieldName": "input_value",
"id": "ChatOutput-njtka",
"inputTypes": [
"Text",
"Message"
],
"inputTypes": ["Text", "Message"],
"type": "str"
}
},
@ -38,18 +33,12 @@
"dataType": "Prompt",
"id": "Prompt-uxBqP",
"name": "prompt",
"output_types": [
"Prompt"
]
"output_types": ["Prompt"]
},
"targetHandle": {
"fieldName": "input_value",
"id": "OpenAIModel-k39HS",
"inputTypes": [
"Text",
"Data",
"Prompt"
],
"inputTypes": ["Text", "Data", "Prompt"],
"type": "str"
}
},
@ -69,19 +58,12 @@
"dataType": "ChatInput",
"id": "ChatInput-P3fgL",
"name": "message",
"output_types": [
"Message"
]
"output_types": ["Message"]
},
"targetHandle": {
"fieldName": "user_input",
"id": "Prompt-uxBqP",
"inputTypes": [
"Document",
"Message",
"Record",
"Text"
],
"inputTypes": ["Document", "Message", "Record", "Text"],
"type": "str"
}
},
@ -102,16 +84,10 @@
"display_name": "Prompt",
"id": "Prompt-uxBqP",
"node": {
"base_classes": [
"object",
"str",
"Text"
],
"base_classes": ["object", "str", "Text"],
"beta": false,
"custom_fields": {
"template": [
"user_input"
]
"template": ["user_input"]
},
"description": "Create a prompt template with dynamic variables.",
"display_name": "Prompt",
@ -134,9 +110,7 @@
"method": "build_prompt",
"name": "prompt",
"selected": "Prompt",
"types": [
"Prompt"
],
"types": ["Prompt"],
"value": "__UNDEFINED__"
},
{
@ -145,9 +119,7 @@
"method": "format_prompt",
"name": "text",
"selected": "Text",
"types": [
"Text"
],
"types": ["Text"],
"value": "__UNDEFINED__"
}
],
@ -178,9 +150,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -201,12 +171,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Document",
"Message",
"Record",
"Text"
],
"input_types": ["Document", "Message", "Record", "Text"],
"list": false,
"load_from_db": false,
"multiline": true,
@ -244,11 +209,7 @@
"display_name": "OpenAI",
"id": "OpenAIModel-k39HS",
"node": {
"base_classes": [
"object",
"Text",
"str"
],
"base_classes": ["object", "Text", "str"],
"beta": false,
"custom_fields": {
"input_value": null,
@ -286,9 +247,7 @@
"method": "text_response",
"name": "text_output",
"selected": "Text",
"types": [
"Text"
],
"types": ["Text"],
"value": "__UNDEFINED__"
},
{
@ -297,9 +256,7 @@
"method": "build_model",
"name": "model_output",
"selected": "BaseLanguageModel",
"types": [
"BaseLanguageModel"
],
"types": ["BaseLanguageModel"],
"value": "__UNDEFINED__"
}
],
@ -330,11 +287,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text",
"Data",
"Prompt"
],
"input_types": ["Text", "Data", "Prompt"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -354,9 +307,7 @@
"fileTypes": [],
"file_path": "",
"info": "The maximum number of tokens to generate. Set to 0 for unlimited tokens.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -376,9 +327,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -398,9 +347,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": true,
"load_from_db": false,
"multiline": false,
@ -427,9 +374,7 @@
"fileTypes": [],
"file_path": "",
"info": "The base URL of the OpenAI API. Defaults to https://api.openai.com/v1.\n\nYou can change this to use other APIs like JinaChat, LocalAI and Prem.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -449,9 +394,7 @@
"fileTypes": [],
"file_path": "",
"info": "The OpenAI API Key to use for the OpenAI model.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -471,9 +414,7 @@
"fileTypes": [],
"file_path": "",
"info": "Stream the response from the model. Streaming works only in Chat.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -493,9 +434,7 @@
"fileTypes": [],
"file_path": "",
"info": "System message to pass to the model.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -515,9 +454,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -553,12 +490,7 @@
"data": {
"id": "ChatOutput-njtka",
"node": {
"base_classes": [
"Record",
"Text",
"str",
"object"
],
"base_classes": ["Record", "Text", "str", "object"],
"beta": false,
"custom_fields": {
"input_value": null,
@ -583,9 +515,7 @@
"method": "text_response",
"name": "text",
"selected": "Text",
"types": [
"Text"
],
"types": ["Text"],
"value": "__UNDEFINED__"
},
{
@ -594,9 +524,7 @@
"method": "message_response",
"name": "message",
"selected": "Message",
"types": [
"Message"
],
"types": ["Message"],
"value": "__UNDEFINED__"
}
],
@ -627,10 +555,7 @@
"fileTypes": [],
"file_path": "",
"info": "Message to be passed as output.",
"input_types": [
"Text",
"Message"
],
"input_types": ["Text", "Message"],
"list": false,
"load_from_db": false,
"multiline": true,
@ -650,17 +575,12 @@
"fileTypes": [],
"file_path": "",
"info": "Type of sender.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": true,
"load_from_db": false,
"multiline": false,
"name": "sender",
"options": [
"Machine",
"User"
],
"options": ["Machine", "User"],
"password": false,
"placeholder": "",
"required": false,
@ -676,9 +596,7 @@
"fileTypes": [],
"file_path": "",
"info": "Name of the sender.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -698,9 +616,7 @@
"fileTypes": [],
"file_path": "",
"info": "Session ID for the message.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -736,12 +652,7 @@
"data": {
"id": "ChatInput-P3fgL",
"node": {
"base_classes": [
"object",
"Record",
"str",
"Text"
],
"base_classes": ["object", "Record", "str", "Text"],
"beta": false,
"custom_fields": {
"input_value": null,
@ -765,9 +676,7 @@
"method": "message_response",
"name": "message",
"selected": "Message",
"types": [
"Message"
],
"types": ["Message"],
"value": "__UNDEFINED__"
}
],
@ -789,7 +698,7 @@
"show": true,
"title_case": false,
"type": "code",
"value": "from langflow.base.io.chat import ChatComponent\nfrom langflow.inputs import DropdownInput, StrInput\nfrom langflow.schema.message import Message\nfrom langflow.template import Output\n\n\nclass ChatInput(ChatComponent):\n display_name = \"Chat Input\"\n description = \"Get chat inputs from the Playground.\"\n icon = \"ChatInput\"\n\n inputs = [\n StrInput(\n name=\"input_value\",\n display_name=\"Text\",\n multiline=True,\n input_types=[],\n info=\"Message to be passed as input.\",\n ),\n DropdownInput(\n name=\"sender\",\n display_name=\"Sender Type\",\n options=[\"Machine\", \"User\"],\n value=\"User\",\n info=\"Type of sender.\",\n advanced=True,\n ),\n StrInput(\n name=\"sender_name\",\n type=str,\n display_name=\"Sender Name\",\n info=\"Name of the sender.\",\n value=\"User\",\n advanced=True,\n ),\n StrInput(\n name=\"session_id\", type=str, display_name=\"Session ID\", info=\"Session ID for the message.\", advanced=True\n ),\n ]\n outputs = [\n Output(display_name=\"Message\", name=\"message\", method=\"message_response\"),\n ]\n\n def message_response(self) -> Message:\n message = Message(\n text=self.input_value,\n sender=self.sender,\n sender_name=self.sender_name,\n session_id=self.session_id,\n )\n if self.session_id and isinstance(message, (Message, str)):\n self.store_message(message)\n self.status = message\n return message\n"
"value": "from langflow.base.io.chat import ChatComponent\nfrom langflow.inputs import DropdownInput, StrInput\nfrom langflow.schema.message import Message\nfrom langflow.template import Output\n\n\nclass ChatInput(ChatComponent):\n display_name = \"Chat Input\"\n description = \"Get chat inputs from the Playground.\"\n icon = \"ChatInput\"\n\n inputs = [\n StrInput(\n name=\"input_value\",\n display_name=\"Text\",\n multiline=True,\n input_types=[],\n value=\"\",\n info=\"Message to be passed as input.\",\n ),\n DropdownInput(\n name=\"sender\",\n display_name=\"Sender Type\",\n options=[\"Machine\", \"User\"],\n value=\"User\",\n info=\"Type of sender.\",\n advanced=True,\n ),\n StrInput(\n name=\"sender_name\",\n type=str,\n display_name=\"Sender Name\",\n info=\"Name of the sender.\",\n value=\"User\",\n advanced=True,\n ),\n StrInput(\n name=\"session_id\", type=str, display_name=\"Session ID\", info=\"Session ID for the message.\", advanced=True\n ),\n ]\n outputs = [\n Output(display_name=\"Message\", name=\"message\", method=\"message_response\"),\n ]\n\n def message_response(self) -> Message:\n message = Message(\n text=self.input_value,\n sender=self.sender,\n sender_name=self.sender_name,\n session_id=self.session_id,\n )\n if self.session_id and isinstance(message, (Message, str)):\n self.store_message(message)\n self.status = message\n return message\n"
},
"input_value": {
"advanced": false,
@ -818,17 +727,12 @@
"fileTypes": [],
"file_path": "",
"info": "Type of sender.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": true,
"load_from_db": false,
"multiline": false,
"name": "sender",
"options": [
"Machine",
"User"
],
"options": ["Machine", "User"],
"password": false,
"placeholder": "",
"required": false,
@ -844,9 +748,7 @@
"fileTypes": [],
"file_path": "",
"info": "Name of the sender.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -866,9 +768,7 @@
"fileTypes": [],
"file_path": "",
"info": "Session ID for the message.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -912,4 +812,4 @@
"is_component": false,
"last_tested_version": "1.0.0a4",
"name": "Basic Prompting (Hello, World)"
}
}

View file

@ -13,12 +13,7 @@
"targetHandle": {
"fieldName": "reference_2",
"id": "Prompt-Rse03",
"inputTypes": [
"Document",
"BaseOutputParser",
"Record",
"Text"
],
"inputTypes": ["Document", "BaseOutputParser", "Record", "Text"],
"type": "str"
}
},
@ -39,17 +34,12 @@
"dataType": "OpenAIModel",
"id": "OpenAIModel-gi29P",
"name": "text_output",
"output_types": [
"Text"
]
"output_types": ["Text"]
},
"targetHandle": {
"fieldName": "input_value",
"id": "ChatOutput-JPlxl",
"inputTypes": [
"Text",
"Message"
],
"inputTypes": ["Text", "Message"],
"type": "str"
}
},
@ -74,12 +64,7 @@
"targetHandle": {
"fieldName": "reference_1",
"id": "Prompt-Rse03",
"inputTypes": [
"Document",
"BaseOutputParser",
"Record",
"Text"
],
"inputTypes": ["Document", "BaseOutputParser", "Record", "Text"],
"type": "str"
}
},
@ -99,19 +84,12 @@
"dataType": "TextInput",
"id": "TextInput-og8Or",
"name": "Text",
"output_types": [
"Text"
]
"output_types": ["Text"]
},
"targetHandle": {
"fieldName": "instructions",
"id": "Prompt-Rse03",
"inputTypes": [
"Document",
"BaseOutputParser",
"Record",
"Text"
],
"inputTypes": ["Document", "BaseOutputParser", "Record", "Text"],
"type": "str"
}
},
@ -131,18 +109,12 @@
"dataType": "Prompt",
"id": "Prompt-Rse03",
"name": "prompt",
"output_types": [
"Prompt"
]
"output_types": ["Prompt"]
},
"targetHandle": {
"fieldName": "input_value",
"id": "OpenAIModel-gi29P",
"inputTypes": [
"Text",
"Data",
"Prompt"
],
"inputTypes": ["Text", "Data", "Prompt"],
"type": "str"
}
},
@ -164,18 +136,10 @@
"display_name": "Prompt",
"id": "Prompt-Rse03",
"node": {
"base_classes": [
"object",
"Text",
"str"
],
"base_classes": ["object", "Text", "str"],
"beta": false,
"custom_fields": {
"template": [
"reference_1",
"reference_2",
"instructions"
]
"template": ["reference_1", "reference_2", "instructions"]
},
"description": "Create a prompt template with dynamic variables.",
"display_name": "Prompt",
@ -198,9 +162,7 @@
"method": "build_prompt",
"name": "prompt",
"selected": "Prompt",
"types": [
"Prompt"
],
"types": ["Prompt"],
"value": "__UNDEFINED__"
},
{
@ -209,9 +171,7 @@
"method": "format_prompt",
"name": "text",
"selected": "Text",
"types": [
"Text"
],
"types": ["Text"],
"value": "__UNDEFINED__"
}
],
@ -320,9 +280,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -358,9 +316,7 @@
"data": {
"id": "URL-HYPkR",
"node": {
"base_classes": [
"Record"
],
"base_classes": ["Record"],
"beta": false,
"custom_fields": {
"urls": null
@ -372,19 +328,16 @@
"field_order": [],
"frozen": false,
"icon": "layout-template",
"output_types": [
"Data"
],
"output_types": ["Data"],
"outputs": [
{
"cache": true,
"display_name": "Data",
"hide": null,
"method": null,
"name": "data",
"selected": "Data",
"types": [
"Data"
],
"types": ["Data"],
"value": "__UNDEFINED__"
}
],
@ -415,9 +368,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": true,
"load_from_db": false,
"multiline": false,
@ -455,12 +406,7 @@
"data": {
"id": "ChatOutput-JPlxl",
"node": {
"base_classes": [
"Text",
"Record",
"object",
"str"
],
"base_classes": ["Text", "Record", "object", "str"],
"beta": false,
"custom_fields": {
"input_value": null,
@ -485,9 +431,7 @@
"method": "text_response",
"name": "text",
"selected": "Text",
"types": [
"Text"
],
"types": ["Text"],
"value": "__UNDEFINED__"
},
{
@ -496,9 +440,7 @@
"method": "message_response",
"name": "message",
"selected": "Message",
"types": [
"Message"
],
"types": ["Message"],
"value": "__UNDEFINED__"
}
],
@ -529,10 +471,7 @@
"fileTypes": [],
"file_path": "",
"info": "Message to be passed as output.",
"input_types": [
"Text",
"Message"
],
"input_types": ["Text", "Message"],
"list": false,
"load_from_db": false,
"multiline": true,
@ -552,17 +491,12 @@
"fileTypes": [],
"file_path": "",
"info": "Type of sender.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": true,
"load_from_db": false,
"multiline": false,
"name": "sender",
"options": [
"Machine",
"User"
],
"options": ["Machine", "User"],
"password": false,
"placeholder": "",
"required": false,
@ -578,9 +512,7 @@
"fileTypes": [],
"file_path": "",
"info": "Name of the sender.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -600,9 +532,7 @@
"fileTypes": [],
"file_path": "",
"info": "Session ID for the message.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -633,11 +563,7 @@
"data": {
"id": "OpenAIModel-gi29P",
"node": {
"base_classes": [
"str",
"Text",
"object"
],
"base_classes": ["str", "Text", "object"],
"beta": false,
"custom_fields": {
"input_value": null,
@ -675,9 +601,7 @@
"method": "text_response",
"name": "text_output",
"selected": "Text",
"types": [
"Text"
],
"types": ["Text"],
"value": "__UNDEFINED__"
},
{
@ -686,9 +610,7 @@
"method": "build_model",
"name": "model_output",
"selected": "BaseLanguageModel",
"types": [
"BaseLanguageModel"
],
"types": ["BaseLanguageModel"],
"value": "__UNDEFINED__"
}
],
@ -719,11 +641,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text",
"Data",
"Prompt"
],
"input_types": ["Text", "Data", "Prompt"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -743,9 +661,7 @@
"fileTypes": [],
"file_path": "",
"info": "The maximum number of tokens to generate. Set to 0 for unlimited tokens.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -765,9 +681,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -787,9 +701,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": true,
"load_from_db": false,
"multiline": false,
@ -816,9 +728,7 @@
"fileTypes": [],
"file_path": "",
"info": "The base URL of the OpenAI API. Defaults to https://api.openai.com/v1.\n\nYou can change this to use other APIs like JinaChat, LocalAI and Prem.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -838,9 +748,7 @@
"fileTypes": [],
"file_path": "",
"info": "The OpenAI API Key to use for the OpenAI model.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -860,9 +768,7 @@
"fileTypes": [],
"file_path": "",
"info": "Stream the response from the model. Streaming works only in Chat.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -882,9 +788,7 @@
"fileTypes": [],
"file_path": "",
"info": "System message to pass to the model.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -904,9 +808,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -942,9 +844,7 @@
"data": {
"id": "URL-2cX90",
"node": {
"base_classes": [
"Record"
],
"base_classes": ["Record"],
"beta": false,
"custom_fields": {
"urls": null
@ -956,19 +856,16 @@
"field_order": [],
"frozen": false,
"icon": "layout-template",
"output_types": [
"Data"
],
"output_types": ["Data"],
"outputs": [
{
"cache": true,
"display_name": "Data",
"hide": null,
"method": null,
"name": "data",
"selected": "Data",
"types": [
"Data"
],
"types": ["Data"],
"value": "__UNDEFINED__"
}
],
@ -999,9 +896,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": true,
"load_from_db": false,
"multiline": false,
@ -1012,9 +907,7 @@
"show": true,
"title_case": false,
"type": "str",
"value": [
"https://www.promptingguide.ai/introduction/basics"
]
"value": ["https://www.promptingguide.ai/introduction/basics"]
}
}
},
@ -1039,11 +932,7 @@
"data": {
"id": "TextInput-og8Or",
"node": {
"base_classes": [
"object",
"Text",
"str"
],
"base_classes": ["object", "Text", "str"],
"beta": false,
"custom_fields": {
"input_value": null,
@ -1056,16 +945,12 @@
"field_order": [],
"frozen": false,
"icon": "type",
"output_types": [
"Text"
],
"output_types": ["Text"],
"outputs": [
{
"name": "Text",
"selected": "Text",
"types": [
"Text"
]
"types": ["Text"]
}
],
"template": {
@ -1095,10 +980,7 @@
"fileTypes": [],
"file_path": "",
"info": "Text or Record to be passed as input.",
"input_types": [
"Record",
"Text"
],
"input_types": ["Record", "Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -1118,9 +1000,7 @@
"fileTypes": [],
"file_path": "",
"info": "Template to convert Record to Text. If left empty, it will be dynamically set to the Record's text key.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": true,
@ -1164,4 +1044,4 @@
"is_component": false,
"last_tested_version": "1.0.0a0",
"name": "Blog Writer"
}
}

View file

@ -8,19 +8,12 @@
"dataType": "ChatInput",
"id": "ChatInput-MsSJ9",
"name": "message",
"output_types": [
"Message"
]
"output_types": ["Message"]
},
"targetHandle": {
"fieldName": "Question",
"id": "Prompt-tHwPf",
"inputTypes": [
"Document",
"Message",
"Record",
"Text"
],
"inputTypes": ["Document", "Message", "Record", "Text"],
"type": "str"
}
},
@ -40,19 +33,12 @@
"dataType": "File",
"id": "File-6TEsD",
"name": "Record",
"output_types": [
"Record"
]
"output_types": ["Record"]
},
"targetHandle": {
"fieldName": "Document",
"id": "Prompt-tHwPf",
"inputTypes": [
"Document",
"Message",
"Record",
"Text"
],
"inputTypes": ["Document", "Message", "Record", "Text"],
"type": "str"
}
},
@ -72,18 +58,12 @@
"dataType": "Prompt",
"id": "Prompt-tHwPf",
"name": "prompt",
"output_types": [
"Prompt"
]
"output_types": ["Prompt"]
},
"targetHandle": {
"fieldName": "input_value",
"id": "OpenAIModel-Bt067",
"inputTypes": [
"Text",
"Data",
"Prompt"
],
"inputTypes": ["Text", "Data", "Prompt"],
"type": "str"
}
},
@ -103,17 +83,12 @@
"dataType": "OpenAIModel",
"id": "OpenAIModel-Bt067",
"name": "text_output",
"output_types": [
"Text"
]
"output_types": ["Text"]
},
"targetHandle": {
"fieldName": "input_value",
"id": "ChatOutput-F5Awj",
"inputTypes": [
"Text",
"Message"
],
"inputTypes": ["Text", "Message"],
"type": "str"
}
},
@ -134,17 +109,10 @@
"display_name": "Prompt",
"id": "Prompt-tHwPf",
"node": {
"base_classes": [
"object",
"str",
"Text"
],
"base_classes": ["object", "str", "Text"],
"beta": false,
"custom_fields": {
"template": [
"Document",
"Question"
]
"template": ["Document", "Question"]
},
"description": "Create a prompt template with dynamic variables.",
"display_name": "Prompt",
@ -167,9 +135,7 @@
"method": "build_prompt",
"name": "prompt",
"selected": "Prompt",
"types": [
"Prompt"
],
"types": ["Prompt"],
"value": "__UNDEFINED__"
},
{
@ -178,9 +144,7 @@
"method": "format_prompt",
"name": "text",
"selected": "Text",
"types": [
"Text"
],
"types": ["Text"],
"value": "__UNDEFINED__"
}
],
@ -193,12 +157,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Document",
"Message",
"Record",
"Text"
],
"input_types": ["Document", "Message", "Record", "Text"],
"list": false,
"load_from_db": false,
"multiline": true,
@ -219,12 +178,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Document",
"Message",
"Record",
"Text"
],
"input_types": ["Document", "Message", "Record", "Text"],
"list": false,
"load_from_db": false,
"multiline": true,
@ -263,9 +217,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -301,9 +253,7 @@
"data": {
"id": "File-6TEsD",
"node": {
"base_classes": [
"Record"
],
"base_classes": ["Record"],
"beta": false,
"custom_fields": {
"path": null,
@ -315,16 +265,12 @@
"field_formatters": {},
"field_order": [],
"frozen": false,
"output_types": [
"Record"
],
"output_types": ["Record"],
"outputs": [
{
"name": "Record",
"selected": "Record",
"types": [
"Record"
]
"types": ["Record"]
}
],
"template": {
@ -420,12 +366,7 @@
"data": {
"id": "ChatInput-MsSJ9",
"node": {
"base_classes": [
"str",
"Record",
"Text",
"object"
],
"base_classes": ["str", "Record", "Text", "object"],
"beta": false,
"custom_fields": {
"input_value": null,
@ -449,9 +390,7 @@
"method": "message_response",
"name": "message",
"selected": "Message",
"types": [
"Message"
],
"types": ["Message"],
"value": "__UNDEFINED__"
}
],
@ -473,7 +412,7 @@
"show": true,
"title_case": false,
"type": "code",
"value": "from langflow.base.io.chat import ChatComponent\nfrom langflow.inputs import DropdownInput, StrInput\nfrom langflow.schema.message import Message\nfrom langflow.template import Output\n\n\nclass ChatInput(ChatComponent):\n display_name = \"Chat Input\"\n description = \"Get chat inputs from the Playground.\"\n icon = \"ChatInput\"\n\n inputs = [\n StrInput(\n name=\"input_value\",\n display_name=\"Text\",\n multiline=True,\n input_types=[],\n info=\"Message to be passed as input.\",\n ),\n DropdownInput(\n name=\"sender\",\n display_name=\"Sender Type\",\n options=[\"Machine\", \"User\"],\n value=\"User\",\n info=\"Type of sender.\",\n advanced=True,\n ),\n StrInput(\n name=\"sender_name\",\n type=str,\n display_name=\"Sender Name\",\n info=\"Name of the sender.\",\n value=\"User\",\n advanced=True,\n ),\n StrInput(\n name=\"session_id\", type=str, display_name=\"Session ID\", info=\"Session ID for the message.\", advanced=True\n ),\n ]\n outputs = [\n Output(display_name=\"Message\", name=\"message\", method=\"message_response\"),\n ]\n\n def message_response(self) -> Message:\n message = Message(\n text=self.input_value,\n sender=self.sender,\n sender_name=self.sender_name,\n session_id=self.session_id,\n )\n if self.session_id and isinstance(message, (Message, str)):\n self.store_message(message)\n self.status = message\n return message\n"
"value": "from langflow.base.io.chat import ChatComponent\nfrom langflow.inputs import DropdownInput, StrInput\nfrom langflow.schema.message import Message\nfrom langflow.template import Output\n\n\nclass ChatInput(ChatComponent):\n display_name = \"Chat Input\"\n description = \"Get chat inputs from the Playground.\"\n icon = \"ChatInput\"\n\n inputs = [\n StrInput(\n name=\"input_value\",\n display_name=\"Text\",\n multiline=True,\n input_types=[],\n value=\"\",\n info=\"Message to be passed as input.\",\n ),\n DropdownInput(\n name=\"sender\",\n display_name=\"Sender Type\",\n options=[\"Machine\", \"User\"],\n value=\"User\",\n info=\"Type of sender.\",\n advanced=True,\n ),\n StrInput(\n name=\"sender_name\",\n type=str,\n display_name=\"Sender Name\",\n info=\"Name of the sender.\",\n value=\"User\",\n advanced=True,\n ),\n StrInput(\n name=\"session_id\", type=str, display_name=\"Session ID\", info=\"Session ID for the message.\", advanced=True\n ),\n ]\n outputs = [\n Output(display_name=\"Message\", name=\"message\", method=\"message_response\"),\n ]\n\n def message_response(self) -> Message:\n message = Message(\n text=self.input_value,\n sender=self.sender,\n sender_name=self.sender_name,\n session_id=self.session_id,\n )\n if self.session_id and isinstance(message, (Message, str)):\n self.store_message(message)\n self.status = message\n return message\n"
},
"input_value": {
"advanced": false,
@ -502,17 +441,12 @@
"fileTypes": [],
"file_path": "",
"info": "Type of sender.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": true,
"load_from_db": false,
"multiline": false,
"name": "sender",
"options": [
"Machine",
"User"
],
"options": ["Machine", "User"],
"password": false,
"placeholder": "",
"required": false,
@ -528,9 +462,7 @@
"fileTypes": [],
"file_path": "",
"info": "Name of the sender.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -550,9 +482,7 @@
"fileTypes": [],
"file_path": "",
"info": "Session ID for the message.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -588,12 +518,7 @@
"data": {
"id": "ChatOutput-F5Awj",
"node": {
"base_classes": [
"str",
"Record",
"Text",
"object"
],
"base_classes": ["str", "Record", "Text", "object"],
"beta": false,
"custom_fields": {
"input_value": null,
@ -617,9 +542,7 @@
"method": "text_response",
"name": "text",
"selected": "Text",
"types": [
"Text"
],
"types": ["Text"],
"value": "__UNDEFINED__"
},
{
@ -628,9 +551,7 @@
"method": "message_response",
"name": "message",
"selected": "Message",
"types": [
"Message"
],
"types": ["Message"],
"value": "__UNDEFINED__"
}
],
@ -661,10 +582,7 @@
"fileTypes": [],
"file_path": "",
"info": "Message to be passed as output.",
"input_types": [
"Text",
"Message"
],
"input_types": ["Text", "Message"],
"list": false,
"load_from_db": false,
"multiline": true,
@ -684,17 +602,12 @@
"fileTypes": [],
"file_path": "",
"info": "Type of sender.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": true,
"load_from_db": false,
"multiline": false,
"name": "sender",
"options": [
"Machine",
"User"
],
"options": ["Machine", "User"],
"password": false,
"placeholder": "",
"required": false,
@ -710,9 +623,7 @@
"fileTypes": [],
"file_path": "",
"info": "Name of the sender.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -732,9 +643,7 @@
"fileTypes": [],
"file_path": "",
"info": "Session ID for the message.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -770,11 +679,7 @@
"data": {
"id": "OpenAIModel-Bt067",
"node": {
"base_classes": [
"object",
"str",
"Text"
],
"base_classes": ["object", "str", "Text"],
"beta": false,
"custom_fields": {
"input_value": null,
@ -812,9 +717,7 @@
"method": "text_response",
"name": "text_output",
"selected": "Text",
"types": [
"Text"
],
"types": ["Text"],
"value": "__UNDEFINED__"
},
{
@ -823,9 +726,7 @@
"method": "build_model",
"name": "model_output",
"selected": "BaseLanguageModel",
"types": [
"BaseLanguageModel"
],
"types": ["BaseLanguageModel"],
"value": "__UNDEFINED__"
}
],
@ -856,11 +757,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text",
"Data",
"Prompt"
],
"input_types": ["Text", "Data", "Prompt"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -880,9 +777,7 @@
"fileTypes": [],
"file_path": "",
"info": "The maximum number of tokens to generate. Set to 0 for unlimited tokens.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -902,9 +797,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -924,9 +817,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": true,
"load_from_db": false,
"multiline": false,
@ -953,9 +844,7 @@
"fileTypes": [],
"file_path": "",
"info": "The base URL of the OpenAI API. Defaults to https://api.openai.com/v1.\n\nYou can change this to use other APIs like JinaChat, LocalAI and Prem.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -975,9 +864,7 @@
"fileTypes": [],
"file_path": "",
"info": "The OpenAI API Key to use for the OpenAI model.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -997,9 +884,7 @@
"fileTypes": [],
"file_path": "",
"info": "Stream the response from the model. Streaming works only in Chat.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -1019,9 +904,7 @@
"fileTypes": [],
"file_path": "",
"info": "System message to pass to the model.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -1041,9 +924,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -1087,4 +968,4 @@
"is_component": false,
"last_tested_version": "1.0.0a0",
"name": "Document QA"
}
}

View file

@ -8,19 +8,12 @@
"dataType": "MemoryComponent",
"id": "MemoryComponent-cdA1J",
"name": "text",
"output_types": [
"Text"
]
"output_types": ["Text"]
},
"targetHandle": {
"fieldName": "context",
"id": "Prompt-ODkUx",
"inputTypes": [
"Document",
"Message",
"Record",
"Text"
],
"inputTypes": ["Document", "Message", "Record", "Text"],
"type": "str"
}
},
@ -41,19 +34,12 @@
"dataType": "ChatInput",
"id": "ChatInput-t7F8v",
"name": "message",
"output_types": [
"Message"
]
"output_types": ["Message"]
},
"targetHandle": {
"fieldName": "user_message",
"id": "Prompt-ODkUx",
"inputTypes": [
"Document",
"Message",
"Record",
"Text"
],
"inputTypes": ["Document", "Message", "Record", "Text"],
"type": "str"
}
},
@ -74,18 +60,12 @@
"dataType": "Prompt",
"id": "Prompt-ODkUx",
"name": "prompt",
"output_types": [
"Prompt"
]
"output_types": ["Prompt"]
},
"targetHandle": {
"fieldName": "input_value",
"id": "OpenAIModel-9RykF",
"inputTypes": [
"Text",
"Data",
"Prompt"
],
"inputTypes": ["Text", "Data", "Prompt"],
"type": "str"
}
},
@ -105,17 +85,12 @@
"dataType": "OpenAIModel",
"id": "OpenAIModel-9RykF",
"name": "text_output",
"output_types": [
"Text"
]
"output_types": ["Text"]
},
"targetHandle": {
"fieldName": "input_value",
"id": "ChatOutput-P1jEe",
"inputTypes": [
"Text",
"Message"
],
"inputTypes": ["Text", "Message"],
"type": "str"
}
},
@ -135,17 +110,12 @@
"dataType": "MemoryComponent",
"id": "MemoryComponent-cdA1J",
"name": "text",
"output_types": [
"Text"
]
"output_types": ["Text"]
},
"targetHandle": {
"fieldName": "input_value",
"id": "TextOutput-vrs6T",
"inputTypes": [
"Record",
"Text"
],
"inputTypes": ["Record", "Text"],
"type": "str"
}
},
@ -164,12 +134,7 @@
"data": {
"id": "ChatInput-t7F8v",
"node": {
"base_classes": [
"Text",
"object",
"Record",
"str"
],
"base_classes": ["Text", "object", "Record", "str"],
"beta": false,
"custom_fields": {
"input_value": null,
@ -193,9 +158,7 @@
"method": "message_response",
"name": "message",
"selected": "Message",
"types": [
"Message"
],
"types": ["Message"],
"value": "__UNDEFINED__"
}
],
@ -217,7 +180,7 @@
"show": true,
"title_case": false,
"type": "code",
"value": "from langflow.base.io.chat import ChatComponent\nfrom langflow.inputs import DropdownInput, StrInput\nfrom langflow.schema.message import Message\nfrom langflow.template import Output\n\n\nclass ChatInput(ChatComponent):\n display_name = \"Chat Input\"\n description = \"Get chat inputs from the Playground.\"\n icon = \"ChatInput\"\n\n inputs = [\n StrInput(\n name=\"input_value\",\n display_name=\"Text\",\n multiline=True,\n input_types=[],\n info=\"Message to be passed as input.\",\n ),\n DropdownInput(\n name=\"sender\",\n display_name=\"Sender Type\",\n options=[\"Machine\", \"User\"],\n value=\"User\",\n info=\"Type of sender.\",\n advanced=True,\n ),\n StrInput(\n name=\"sender_name\",\n type=str,\n display_name=\"Sender Name\",\n info=\"Name of the sender.\",\n value=\"User\",\n advanced=True,\n ),\n StrInput(\n name=\"session_id\", type=str, display_name=\"Session ID\", info=\"Session ID for the message.\", advanced=True\n ),\n ]\n outputs = [\n Output(display_name=\"Message\", name=\"message\", method=\"message_response\"),\n ]\n\n def message_response(self) -> Message:\n message = Message(\n text=self.input_value,\n sender=self.sender,\n sender_name=self.sender_name,\n session_id=self.session_id,\n )\n if self.session_id and isinstance(message, (Message, str)):\n self.store_message(message)\n self.status = message\n return message\n"
"value": "from langflow.base.io.chat import ChatComponent\nfrom langflow.inputs import DropdownInput, StrInput\nfrom langflow.schema.message import Message\nfrom langflow.template import Output\n\n\nclass ChatInput(ChatComponent):\n display_name = \"Chat Input\"\n description = \"Get chat inputs from the Playground.\"\n icon = \"ChatInput\"\n\n inputs = [\n StrInput(\n name=\"input_value\",\n display_name=\"Text\",\n multiline=True,\n input_types=[],\n value=\"\",\n info=\"Message to be passed as input.\",\n ),\n DropdownInput(\n name=\"sender\",\n display_name=\"Sender Type\",\n options=[\"Machine\", \"User\"],\n value=\"User\",\n info=\"Type of sender.\",\n advanced=True,\n ),\n StrInput(\n name=\"sender_name\",\n type=str,\n display_name=\"Sender Name\",\n info=\"Name of the sender.\",\n value=\"User\",\n advanced=True,\n ),\n StrInput(\n name=\"session_id\", type=str, display_name=\"Session ID\", info=\"Session ID for the message.\", advanced=True\n ),\n ]\n outputs = [\n Output(display_name=\"Message\", name=\"message\", method=\"message_response\"),\n ]\n\n def message_response(self) -> Message:\n message = Message(\n text=self.input_value,\n sender=self.sender,\n sender_name=self.sender_name,\n session_id=self.session_id,\n )\n if self.session_id and isinstance(message, (Message, str)):\n self.store_message(message)\n self.status = message\n return message\n"
},
"input_value": {
"advanced": false,
@ -246,17 +209,12 @@
"fileTypes": [],
"file_path": "",
"info": "Type of sender.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": true,
"load_from_db": false,
"multiline": false,
"name": "sender",
"options": [
"Machine",
"User"
],
"options": ["Machine", "User"],
"password": false,
"placeholder": "",
"required": false,
@ -272,9 +230,7 @@
"fileTypes": [],
"file_path": "",
"info": "Name of the sender.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -294,9 +250,7 @@
"fileTypes": [],
"file_path": "",
"info": "Session ID for the message.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -332,12 +286,7 @@
"data": {
"id": "ChatOutput-P1jEe",
"node": {
"base_classes": [
"Text",
"object",
"Record",
"str"
],
"base_classes": ["Text", "object", "Record", "str"],
"beta": false,
"custom_fields": {
"input_value": null,
@ -361,9 +310,7 @@
"method": "text_response",
"name": "text",
"selected": "Text",
"types": [
"Text"
],
"types": ["Text"],
"value": "__UNDEFINED__"
},
{
@ -372,9 +319,7 @@
"method": "message_response",
"name": "message",
"selected": "Message",
"types": [
"Message"
],
"types": ["Message"],
"value": "__UNDEFINED__"
}
],
@ -405,10 +350,7 @@
"fileTypes": [],
"file_path": "",
"info": "Message to be passed as output.",
"input_types": [
"Text",
"Message"
],
"input_types": ["Text", "Message"],
"list": false,
"load_from_db": false,
"multiline": true,
@ -428,17 +370,12 @@
"fileTypes": [],
"file_path": "",
"info": "Type of sender.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": true,
"load_from_db": false,
"multiline": false,
"name": "sender",
"options": [
"Machine",
"User"
],
"options": ["Machine", "User"],
"password": false,
"placeholder": "",
"required": false,
@ -454,9 +391,7 @@
"fileTypes": [],
"file_path": "",
"info": "Name of the sender.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -476,9 +411,7 @@
"fileTypes": [],
"file_path": "",
"info": "Session ID for the message.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -516,11 +449,7 @@
"display_name": "Chat Memory",
"id": "MemoryComponent-cdA1J",
"node": {
"base_classes": [
"str",
"Text",
"object"
],
"base_classes": ["str", "Text", "object"],
"beta": true,
"custom_fields": {
"n_messages": null,
@ -537,19 +466,16 @@
"field_order": [],
"frozen": false,
"icon": "history",
"output_types": [
"Text"
],
"output_types": ["Text"],
"outputs": [
{
"cache": true,
"display_name": "Text",
"hide": null,
"method": null,
"name": "text",
"selected": "Text",
"types": [
"Text"
],
"types": ["Text"],
"value": "__UNDEFINED__"
}
],
@ -599,17 +525,12 @@
"fileTypes": [],
"file_path": "",
"info": "Order of the messages.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": true,
"load_from_db": false,
"multiline": false,
"name": "order",
"options": [
"Ascending",
"Descending"
],
"options": ["Ascending", "Descending"],
"password": false,
"placeholder": "",
"required": false,
@ -625,18 +546,12 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": true,
"load_from_db": false,
"multiline": false,
"name": "sender",
"options": [
"Machine",
"User",
"Machine and User"
],
"options": ["Machine", "User", "Machine and User"],
"password": false,
"placeholder": "",
"required": false,
@ -652,9 +567,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -673,9 +586,7 @@
"fileTypes": [],
"file_path": "",
"info": "Session ID of the chat history.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -713,17 +624,10 @@
"display_name": "Prompt",
"id": "Prompt-ODkUx",
"node": {
"base_classes": [
"Text",
"str",
"object"
],
"base_classes": ["Text", "str", "object"],
"beta": false,
"custom_fields": {
"template": [
"context",
"user_message"
]
"template": ["context", "user_message"]
},
"description": "Create a prompt template with dynamic variables.",
"display_name": "Prompt",
@ -746,9 +650,7 @@
"method": "build_prompt",
"name": "prompt",
"selected": "Prompt",
"types": [
"Prompt"
],
"types": ["Prompt"],
"value": "__UNDEFINED__"
},
{
@ -757,9 +659,7 @@
"method": "format_prompt",
"name": "text",
"selected": "Text",
"types": [
"Text"
],
"types": ["Text"],
"value": "__UNDEFINED__"
}
],
@ -791,12 +691,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Document",
"Message",
"Record",
"Text"
],
"input_types": ["Document", "Message", "Record", "Text"],
"list": false,
"load_from_db": false,
"multiline": true,
@ -816,9 +711,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -839,12 +732,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Document",
"Message",
"Record",
"Text"
],
"input_types": ["Document", "Message", "Record", "Text"],
"list": false,
"load_from_db": false,
"multiline": true,
@ -880,11 +768,7 @@
"data": {
"id": "OpenAIModel-9RykF",
"node": {
"base_classes": [
"str",
"object",
"Text"
],
"base_classes": ["str", "object", "Text"],
"beta": false,
"custom_fields": {
"input_value": null,
@ -922,9 +806,7 @@
"method": "text_response",
"name": "text_output",
"selected": "Text",
"types": [
"Text"
],
"types": ["Text"],
"value": "__UNDEFINED__"
},
{
@ -933,9 +815,7 @@
"method": "build_model",
"name": "model_output",
"selected": "BaseLanguageModel",
"types": [
"BaseLanguageModel"
],
"types": ["BaseLanguageModel"],
"value": "__UNDEFINED__"
}
],
@ -966,11 +846,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text",
"Data",
"Prompt"
],
"input_types": ["Text", "Data", "Prompt"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -990,9 +866,7 @@
"fileTypes": [],
"file_path": "",
"info": "The maximum number of tokens to generate. Set to 0 for unlimited tokens.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -1012,9 +886,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -1034,9 +906,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": true,
"load_from_db": false,
"multiline": false,
@ -1063,9 +933,7 @@
"fileTypes": [],
"file_path": "",
"info": "The base URL of the OpenAI API. Defaults to https://api.openai.com/v1.\n\nYou can change this to use other APIs like JinaChat, LocalAI and Prem.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -1085,9 +953,7 @@
"fileTypes": [],
"file_path": "",
"info": "The OpenAI API Key to use for the OpenAI model.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -1107,9 +973,7 @@
"fileTypes": [],
"file_path": "",
"info": "Stream the response from the model. Streaming works only in Chat.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -1129,9 +993,7 @@
"fileTypes": [],
"file_path": "",
"info": "System message to pass to the model.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -1151,9 +1013,7 @@
"fileTypes": [],
"file_path": "",
"info": "",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -1189,11 +1049,7 @@
"data": {
"id": "TextOutput-vrs6T",
"node": {
"base_classes": [
"str",
"object",
"Text"
],
"base_classes": ["str", "object", "Text"],
"beta": false,
"custom_fields": {
"input_value": null,
@ -1206,9 +1062,7 @@
"field_order": [],
"frozen": false,
"icon": "type",
"output_types": [
"Text"
],
"output_types": ["Text"],
"template": {
"_type": "CustomComponent",
"code": {
@ -1236,10 +1090,7 @@
"fileTypes": [],
"file_path": "",
"info": "Text or Record to be passed as output.",
"input_types": [
"Record",
"Text"
],
"input_types": ["Record", "Text"],
"list": false,
"load_from_db": false,
"multiline": false,
@ -1259,9 +1110,7 @@
"fileTypes": [],
"file_path": "",
"info": "Template to convert Record to Text. If left empty, it will be dynamically set to the Record's text key.",
"input_types": [
"Text"
],
"input_types": ["Text"],
"list": false,
"load_from_db": false,
"multiline": true,
@ -1307,4 +1156,4 @@
"is_component": false,
"last_tested_version": "1.0.0a0",
"name": "Memory Chatbot"
}
}