add model name to chatgptagentconfig
This commit is contained in:
parent
d52bd54d06
commit
8528d85e23
3 changed files with 3 additions and 2 deletions
|
|
@ -1,6 +1,6 @@
|
||||||
[tool.poetry]
|
[tool.poetry]
|
||||||
name = "vocode"
|
name = "vocode"
|
||||||
version = "0.1.36"
|
version = "0.1.37"
|
||||||
description = "The all-in-one voice SDK"
|
description = "The all-in-one voice SDK"
|
||||||
authors = ["Ajay Raj <ajay@vocode.dev>"]
|
authors = ["Ajay Raj <ajay@vocode.dev>"]
|
||||||
license = "MIT License"
|
license = "MIT License"
|
||||||
|
|
|
||||||
|
|
@ -24,7 +24,6 @@ from vocode.models.agent import (
|
||||||
RESTfulUserImplementedAgentConfig,
|
RESTfulUserImplementedAgentConfig,
|
||||||
WebSocketUserImplementedAgentConfig,
|
WebSocketUserImplementedAgentConfig,
|
||||||
EchoAgentConfig,
|
EchoAgentConfig,
|
||||||
ChatGPTAlphaAgentConfig,
|
|
||||||
LLMAgentConfig,
|
LLMAgentConfig,
|
||||||
ChatGPTAgentConfig,
|
ChatGPTAgentConfig,
|
||||||
)
|
)
|
||||||
|
|
|
||||||
|
|
@ -10,6 +10,7 @@ FILLER_AUDIO_DEFAULT_SILENCE_THRESHOLD_SECONDS = 0.5
|
||||||
LLM_AGENT_DEFAULT_TEMPERATURE = 1.0
|
LLM_AGENT_DEFAULT_TEMPERATURE = 1.0
|
||||||
LLM_AGENT_DEFAULT_MAX_TOKENS = 256
|
LLM_AGENT_DEFAULT_MAX_TOKENS = 256
|
||||||
LLM_AGENT_DEFAULT_MODEL_NAME = "text-curie-001"
|
LLM_AGENT_DEFAULT_MODEL_NAME = "text-curie-001"
|
||||||
|
CHAT_GPT_AGENT_DEFAULT_MODEL_NAME = "gpt-3.5-turbo"
|
||||||
|
|
||||||
|
|
||||||
class AgentType(str, Enum):
|
class AgentType(str, Enum):
|
||||||
|
|
@ -62,6 +63,7 @@ class ChatGPTAgentConfig(AgentConfig, type=AgentType.CHAT_GPT):
|
||||||
prompt_preamble: str
|
prompt_preamble: str
|
||||||
expected_first_prompt: Optional[str] = None
|
expected_first_prompt: Optional[str] = None
|
||||||
generate_responses: bool = False
|
generate_responses: bool = False
|
||||||
|
model_name: str = CHAT_GPT_AGENT_DEFAULT_MODEL_NAME
|
||||||
temperature: float = LLM_AGENT_DEFAULT_TEMPERATURE
|
temperature: float = LLM_AGENT_DEFAULT_TEMPERATURE
|
||||||
max_tokens: int = LLM_AGENT_DEFAULT_MAX_TOKENS
|
max_tokens: int = LLM_AGENT_DEFAULT_MAX_TOKENS
|
||||||
cut_off_response: Optional[CutOffResponse] = None
|
cut_off_response: Optional[CutOffResponse] = None
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue