Update OpenAI defaults (#118059)
* Update OpenAI defaults * Update max temperature
This commit is contained in:
parent
ffc3560dad
commit
3b2cdb63f1
3 changed files with 18 additions and 15 deletions
|
@ -145,6 +145,16 @@ def openai_config_option_schema(
|
|||
)
|
||||
|
||||
return {
|
||||
vol.Optional(
|
||||
CONF_PROMPT,
|
||||
description={"suggested_value": options.get(CONF_PROMPT)},
|
||||
default=DEFAULT_PROMPT,
|
||||
): TemplateSelector(),
|
||||
vol.Optional(
|
||||
CONF_LLM_HASS_API,
|
||||
description={"suggested_value": options.get(CONF_LLM_HASS_API)},
|
||||
default="none",
|
||||
): SelectSelector(SelectSelectorConfig(options=apis)),
|
||||
vol.Optional(
|
||||
CONF_CHAT_MODEL,
|
||||
description={
|
||||
|
@ -153,16 +163,6 @@ def openai_config_option_schema(
|
|||
},
|
||||
default=DEFAULT_CHAT_MODEL,
|
||||
): str,
|
||||
vol.Optional(
|
||||
CONF_LLM_HASS_API,
|
||||
description={"suggested_value": options.get(CONF_LLM_HASS_API)},
|
||||
default="none",
|
||||
): SelectSelector(SelectSelectorConfig(options=apis)),
|
||||
vol.Optional(
|
||||
CONF_PROMPT,
|
||||
description={"suggested_value": options.get(CONF_PROMPT)},
|
||||
default=DEFAULT_PROMPT,
|
||||
): TemplateSelector(),
|
||||
vol.Optional(
|
||||
CONF_MAX_TOKENS,
|
||||
description={"suggested_value": options.get(CONF_MAX_TOKENS)},
|
||||
|
@ -177,5 +177,5 @@ def openai_config_option_schema(
|
|||
CONF_TEMPERATURE,
|
||||
description={"suggested_value": options.get(CONF_TEMPERATURE)},
|
||||
default=DEFAULT_TEMPERATURE,
|
||||
): NumberSelector(NumberSelectorConfig(min=0, max=1, step=0.05)),
|
||||
): NumberSelector(NumberSelectorConfig(min=0, max=2, step=0.05)),
|
||||
}
|
||||
|
|
|
@ -23,10 +23,10 @@ An overview of the areas and the devices in this smart home:
|
|||
{%- endfor %}
|
||||
"""
|
||||
CONF_CHAT_MODEL = "chat_model"
|
||||
DEFAULT_CHAT_MODEL = "gpt-3.5-turbo"
|
||||
DEFAULT_CHAT_MODEL = "gpt-4o"
|
||||
CONF_MAX_TOKENS = "max_tokens"
|
||||
DEFAULT_MAX_TOKENS = 150
|
||||
CONF_TOP_P = "top_p"
|
||||
DEFAULT_TOP_P = 1
|
||||
DEFAULT_TOP_P = 1.0
|
||||
CONF_TEMPERATURE = "temperature"
|
||||
DEFAULT_TEMPERATURE = 0.5
|
||||
DEFAULT_TEMPERATURE = 1.0
|
||||
|
|
|
@ -17,12 +17,15 @@
|
|||
"step": {
|
||||
"init": {
|
||||
"data": {
|
||||
"prompt": "Prompt Template",
|
||||
"prompt": "Instructions",
|
||||
"chat_model": "[%key:common::generic::model%]",
|
||||
"max_tokens": "Maximum tokens to return in response",
|
||||
"temperature": "Temperature",
|
||||
"top_p": "Top P",
|
||||
"llm_hass_api": "[%key:common::config_flow::data::llm_hass_api%]"
|
||||
},
|
||||
"data_description": {
|
||||
"prompt": "Instruct how the LLM should respond. This can be a template."
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue