diff --git a/src/backend/base/langflow/components/models/OllamaModel.py b/src/backend/base/langflow/components/models/OllamaModel.py index 8edba3374..4da762621 100644 --- a/src/backend/base/langflow/components/models/OllamaModel.py +++ b/src/backend/base/langflow/components/models/OllamaModel.py @@ -33,7 +33,7 @@ class ChatOllamaComponent(LCModelComponent): build_config["mirostat_eta"]["value"] = 0.1 build_config["mirostat_tau"]["value"] = 5 - if field_name == "model": + if field_name == "model_name": base_url_dict = build_config.get("base_url", {}) base_url_load_from_db = base_url_dict.get("load_from_db", False) base_url_value = base_url_dict.get("value") @@ -41,7 +41,7 @@ class ChatOllamaComponent(LCModelComponent): base_url_value = self.variables(base_url_value) elif not base_url_value: base_url_value = "http://localhost:11434" - build_config["model"]["options"] = self.get_model(base_url_value + "/api/tags") + build_config["model_name"]["options"] = self.get_model(base_url_value + "/api/tags") if field_name == "keep_alive_flag": if field_value == "Keep": @@ -77,7 +77,7 @@ class ChatOllamaComponent(LCModelComponent): DropdownInput( name="model_name", display_name="Model Name", - value="llama2", + value="llama3", info="Refer to https://ollama.ai/library for more models.", refresh_button=True, ),