diff --git a/docs/docs/Components/components-models.md b/docs/docs/Components/components-models.md index ba4d0de8b..dfdf69705 100644 --- a/docs/docs/Components/components-models.md +++ b/docs/docs/Components/components-models.md @@ -110,6 +110,12 @@ For more information, see the [Azure OpenAI documentation](https://learn.microso | Input Value | Input Value | Specifies the input text for text generation. | | Stream | Stream | Specifies whether to stream the response from the model. Defaults to `False`. | +### Outputs + +| Name | Type | Description | +|-------|---------------|------------------------------------------------------------------| +| model | LanguageModel | An instance of AzureOpenAI configured with the specified parameters. | + ## Cohere This component generates text using Cohere's language models. @@ -131,6 +137,31 @@ For more information, see the [Cohere documentation](https://cohere.ai/). |-------|---------------|------------------------------------------------------------------| | model | LanguageModel | An instance of the Cohere model configured with the specified parameters. | +## DeepSeek + +This component generates text using DeepSeek's language models. + +For more information, see the [DeepSeek documentation](https://api-docs.deepseek.com/). + +### Inputs + +| Name | Type | Description | +|----------------|---------------|-----------------------------------------------------------------| +| max_tokens | Integer | Maximum number of tokens to generate. Set to `0` for unlimited. Range: `0-128000`. | +| model_kwargs | Dictionary | Additional keyword arguments for the model. | +| json_mode | Boolean | If `True`, outputs JSON regardless of passing a schema. | +| model_name | String | The DeepSeek model to use. Default: `deepseek-chat`. | +| api_base | String | Base URL for API requests. Default: `https://api.deepseek.com`. | +| api_key | SecretString | Your DeepSeek API key for authentication. | +| temperature | Float | Controls randomness in responses. Range: `[0.0, 2.0]`. Default: `1.0`. | +| seed | Integer | Number initialized for random number generation. Use the same seed integer for more reproducible results, and use a different seed number for more random results. | + +### Outputs + +| Name | Type | Description | +|-------|---------------|------------------------------------------------------------------| +| model | LanguageModel | An instance of ChatOpenAI configured with the specified parameters. | + ## Google Generative AI This component generates text using Google's Generative AI models. @@ -149,6 +180,12 @@ For more information, see the [Google Generative AI documentation](https://cloud | Top P | Top P | The maximum cumulative probability of tokens to consider when sampling. | | N | N | Number of chat completions to generate for each prompt. | +### Outputs + +| Name | Type | Description | +|-------|---------------|------------------------------------------------------------------| +| model | LanguageModel | An instance of ChatGoogleGenerativeAI configured with the specified parameters. | + ## Groq This component generates text using Groq's language models. @@ -188,6 +225,31 @@ For more information, see the [Hugging Face documentation](https://huggingface.c | Model Kwargs | Model Kwargs | Additional keyword arguments for the model.| | Input Value | Input Value | The input text for text generation. | +## LMStudio + +This component generates text using LM Studio's local language models. + +For more information, see [LM Studio documentation](https://lmstudio.ai/). + +### Inputs + +| Name | Type | Description | +|----------------|---------------|-----------------------------------------------------------------| +| base_url | String | The URL where LM Studio is running. Default: `"http://localhost:1234"`. | +| max_tokens | Integer | Maximum number of tokens to generate in the response. Default: `512`. | +| temperature | Float | Controls randomness in the output. Range: `[0.0, 2.0]`. Default: `0.7`. | +| top_p | Float | Controls diversity via nucleus sampling. Range: `[0.0, 1.0]`. Default: `1.0`. | +| stop | List[String] | List of strings that will stop generation when encountered (advanced). | +| stream | Boolean | Whether to stream the response. Default: `False`. | +| presence_penalty | Float | Penalizes repeated tokens. Range: `[-2.0, 2.0]`. Default: `0.0`. | +| frequency_penalty | Float | Penalizes frequent tokens. Range: `[-2.0, 2.0]`. Default: `0.0`. | + +### Outputs + +| Name | Type | Description | +|-------|---------------|------------------------------------------------------------------| +| model | LanguageModel | An instance of LMStudio configured with the specified parameters. | + ## Maritalk This component generates text using Maritalk LLMs. @@ -238,6 +300,30 @@ For more information, see [Mistral AI documentation](https://docs.mistral.ai/). |--------|---------------|-----------------------------------------------------| | model | LanguageModel | An instance of ChatMistralAI configured with the specified parameters. | +## Novita AI + +This component generates text using Novita AI's language models. + +For more information, see [Novita AI documentation](https://novita.ai/docs/model-api/reference/llm/llm.html?utm_source=github_langflow&utm_medium=github_readme&utm_campaign=link). + +### Inputs + +| Name | Type | Description | +|---------------------|---------------|------------------------------------------------------------------| +| api_key | SecretString | Your Novita AI API Key. | +| model | String | The id of the Novita AI model to use. | +| max_tokens | Integer | The maximum number of tokens to generate. Set to 0 for unlimited tokens. | +| temperature | Float | Controls randomness in the output. Range: [0.0, 1.0]. Default: 0.7. | +| top_p | Float | Controls the nucleus sampling. Range: [0.0, 1.0]. Default: 1.0. | +| frequency_penalty | Float | Controls the frequency penalty. Range: [0.0, 2.0]. Default: 0.0. | +| presence_penalty | Float | Controls the presence penalty. Range: [0.0, 2.0]. Default: 0.0. | + +### Outputs + +| Name | Type | Description | +|-------|---------------|------------------------------------------------------------------| +| model | LanguageModel | An instance of Novita AI model configured with the specified parameters. | + ## NVIDIA This component generates text using NVIDIA LLMs. @@ -305,11 +391,6 @@ For more information, see [OpenAI documentation](https://beta.openai.com/docs/). |-------|---------------|------------------------------------------------------------------| | model | LanguageModel | An instance of OpenAI model configured with the specified parameters. | -## Qianfan - -This component generates text using Qianfan's language models. - -For more information, see [Qianfan documentation](https://github.com/baidubce/bce-qianfan-sdk). ## OpenRouter @@ -359,6 +440,13 @@ For more information, see [Perplexity documentation](https://perplexity.ai/). |--------|---------------|-----------------------------------------------------| | model | LanguageModel | An instance of ChatPerplexity configured with the specified parameters. | + +## Qianfan + +This component generates text using Qianfan's language models. + +For more information, see [Qianfan documentation](https://github.com/baidubce/bce-qianfan-sdk). + ## SambaNova This component generates text using SambaNova LLMs. @@ -408,22 +496,4 @@ For more information, see [Google Vertex AI documentation](https://cloud.google. |--------|---------------|-----------------------------------------------------| | model | LanguageModel | An instance of ChatVertexAI configured with the specified parameters. | -## Novita AI -This component generates text using Novita AI's language models. - -For more information, see [Novita AI documentation](https://novita.ai/docs/model-api/reference/llm/llm.html?utm_source=github_langflow&utm_medium=github_readme&utm_campaign=link). - -### Parameters - -#### Inputs - -| Name | Type | Description | -|---------------------|---------------|------------------------------------------------------------------| -| api_key | SecretString | Your Novita AI API Key. | -| model | String | The id of the Novita AI model to use. | -| max_tokens | Integer | The maximum number of tokens to generate. Set to 0 for unlimited tokens. | -| temperature | Float | Controls randomness in the output. Range: [0.0, 1.0]. Default: 0.7. | -| top_p | Float | Controls the nucleus sampling. Range: [0.0, 1.0]. Default: 1.0. | -| frequency_penalty | Float | Controls the frequency penalty. Range: [0.0, 2.0]. Default: 0.0. | -| presence_penalty | Float | Controls the presence penalty. Range: [0.0, 2.0]. Default: 0.0. |