From b3b2b2572ab70157d37a57acbc707cb1b0a6ea7c Mon Sep 17 00:00:00 2001 From: Edwin Jose Date: Fri, 22 Nov 2024 12:10:09 -0500 Subject: [PATCH] fix: fixes Ollama Embeddings component (#4787) Update ollama.py removed unwanted parameter temperature for embedding --- .../langflow/components/embeddings/ollama.py | 16 +++------------- 1 file changed, 3 insertions(+), 13 deletions(-) diff --git a/src/backend/base/langflow/components/embeddings/ollama.py b/src/backend/base/langflow/components/embeddings/ollama.py index 50da1ea7b..e5c83ef9e 100644 --- a/src/backend/base/langflow/components/embeddings/ollama.py +++ b/src/backend/base/langflow/components/embeddings/ollama.py @@ -2,7 +2,7 @@ from langchain_ollama import OllamaEmbeddings from langflow.base.models.model import LCModelComponent from langflow.field_typing import Embeddings -from langflow.io import FloatInput, MessageTextInput, Output +from langflow.io import MessageTextInput, Output class OllamaEmbeddingsComponent(LCModelComponent): @@ -16,19 +16,13 @@ class OllamaEmbeddingsComponent(LCModelComponent): MessageTextInput( name="model", display_name="Ollama Model", - value="llama3.1", + value="nomic-embed-text", ), MessageTextInput( name="base_url", display_name="Ollama Base URL", value="http://localhost:11434", ), - FloatInput( - name="temperature", - display_name="Model Temperature", - value=0.1, - advanced=True, - ), ] outputs = [ @@ -37,11 +31,7 @@ class OllamaEmbeddingsComponent(LCModelComponent): def build_embeddings(self) -> Embeddings: try: - output = OllamaEmbeddings( - model=self.model, - base_url=self.base_url, - temperature=self.temperature, - ) + output = OllamaEmbeddings(model=self.model, base_url=self.base_url) except Exception as e: msg = "Could not connect to Ollama API." raise ValueError(msg) from e