{ "config": { "step": { "user": { "data": { "url": "[%key:common::config_flow::data::url%]" } } }, "abort": { "already_configured": "[%key:common::config_flow::abort::already_configured_service%]" }, "error": { "invalid_url": "[%key:common::config_flow::error::invalid_host%]", "cannot_connect": "[%key:common::config_flow::error::cannot_connect%]", "unknown": "[%key:common::config_flow::error::unknown%]" } }, "config_subentries": { "conversation": { "initiate_flow": { "user": "Add conversation agent", "reconfigure": "Reconfigure conversation agent" }, "entry_type": "Conversation agent", "step": { "set_options": { "data": { "model": "Model", "name": "[%key:common::config_flow::data::name%]", "prompt": "[%key:common::config_flow::data::prompt%]", "llm_hass_api": "[%key:common::config_flow::data::llm_hass_api%]", "max_history": "Max history messages", "num_ctx": "Context window size", "keep_alive": "Keep alive", "think": "Think before responding" }, "data_description": { "prompt": "Instruct how the LLM should respond. This can be a template.", "keep_alive": "Duration in seconds for Ollama to keep model in memory. -1 = indefinite, 0 = never.", "num_ctx": "Maximum number of text tokens the model can process. Lower to reduce Ollama RAM, or increase for a large number of exposed entities.", "think": "If enabled, the LLM will think before responding. This can improve response quality but may increase latency." } }, "download": { "title": "Downloading model" } }, "abort": { "reconfigure_successful": "[%key:common::config_flow::abort::reconfigure_successful%]", "entry_not_loaded": "Failed to add agent. The configuration is disabled.", "download_failed": "Model downloading failed", "cannot_connect": "[%key:common::config_flow::error::cannot_connect%]" }, "progress": { "download": "Please wait while the model is downloaded, which may take a very long time. Check your Ollama server logs for more details." } }, "ai_task_data": { "initiate_flow": { "user": "Add AI task", "reconfigure": "Reconfigure AI task" }, "entry_type": "AI task", "step": { "set_options": { "data": { "model": "[%key:component::ollama::config_subentries::conversation::step::set_options::data::model%]", "name": "[%key:common::config_flow::data::name%]", "prompt": "[%key:common::config_flow::data::prompt%]", "max_history": "[%key:component::ollama::config_subentries::conversation::step::set_options::data::max_history%]", "num_ctx": "[%key:component::ollama::config_subentries::conversation::step::set_options::data::num_ctx%]", "keep_alive": "[%key:component::ollama::config_subentries::conversation::step::set_options::data::keep_alive%]", "think": "[%key:component::ollama::config_subentries::conversation::step::set_options::data::think%]" }, "data_description": { "prompt": "[%key:component::ollama::config_subentries::conversation::step::set_options::data_description::prompt%]", "keep_alive": "[%key:component::ollama::config_subentries::conversation::step::set_options::data_description::keep_alive%]", "num_ctx": "[%key:component::ollama::config_subentries::conversation::step::set_options::data_description::num_ctx%]", "think": "[%key:component::ollama::config_subentries::conversation::step::set_options::data_description::think%]" } }, "download": { "title": "[%key:component::ollama::config_subentries::conversation::step::download::title%]" } }, "abort": { "reconfigure_successful": "[%key:common::config_flow::abort::reconfigure_successful%]", "entry_not_loaded": "[%key:component::ollama::config_subentries::conversation::abort::entry_not_loaded%]", "download_failed": "[%key:component::ollama::config_subentries::conversation::abort::download_failed%]", "cannot_connect": "[%key:common::config_flow::error::cannot_connect%]" }, "progress": { "download": "[%key:component::ollama::config_subentries::conversation::progress::download%]" } } }, "exceptions": { "unsupported_attachment_type": { "message": "Ollama only supports image attachments in user content, but received non-image attachment." } } }