mirror of
https://github.com/AtHeartEngineering/ai_automation_suggester.git
synced 2026-01-09 22:27:55 -05:00
Merge pull request #124 from RmG152/dev
Feat: UI Translations & Dynamic Automation Limit
This commit is contained in:
@@ -45,6 +45,7 @@ async def async_setup(hass: HomeAssistant, config: ConfigType) -> bool:
|
||||
domains = call.data.get("domains", {})
|
||||
entity_limit = call.data.get("entity_limit", 200)
|
||||
automation_read_yaml = call.data.get("automation_read_yaml", False)
|
||||
automation_limit = call.data.get("automation_limit", 100)
|
||||
|
||||
# Parse domains if provided as a string or dict
|
||||
if isinstance(domains, str):
|
||||
@@ -76,6 +77,7 @@ async def async_setup(hass: HomeAssistant, config: ConfigType) -> bool:
|
||||
coordinator.selected_domains = domains
|
||||
coordinator.entity_limit = entity_limit
|
||||
coordinator.automation_read_file = automation_read_yaml
|
||||
coordinator.automation_limit = automation_limit
|
||||
|
||||
try:
|
||||
await coordinator.async_request_refresh()
|
||||
@@ -86,6 +88,7 @@ async def async_setup(hass: HomeAssistant, config: ConfigType) -> bool:
|
||||
coordinator.selected_domains = []
|
||||
coordinator.entity_limit = 200
|
||||
coordinator.automation_read_file = False
|
||||
coordinator.automation_limit = 100
|
||||
|
||||
except KeyError:
|
||||
raise ServiceValidationError("Provider configuration not found")
|
||||
|
||||
@@ -131,6 +131,7 @@ class AIAutomationCoordinator(DataUpdateCoordinator):
|
||||
self.selected_domains: list[str] = []
|
||||
self.entity_limit = 200
|
||||
self.automation_read_file = False # Default automation reading mode
|
||||
self.automation_limit = 100
|
||||
|
||||
super().__init__(hass, _LOGGER, name=DOMAIN, update_interval=None)
|
||||
self.session = async_get_clientsession(hass)
|
||||
@@ -278,7 +279,7 @@ class AIAutomationCoordinator(DataUpdateCoordinator):
|
||||
async def _build_prompt(self, entities: dict) -> str: # noqa: C901
|
||||
"""Build the prompt based on entities and automations."""
|
||||
MAX_ATTR = 500
|
||||
MAX_AUTOM = 100 # Change this to user input?
|
||||
MAX_AUTOM = getattr(self, "automation_limit", 100)
|
||||
|
||||
ent_sections: list[str] = []
|
||||
for eid, meta in random.sample(
|
||||
|
||||
@@ -13,7 +13,7 @@ generate_suggestions:
|
||||
name: Custom Prompt
|
||||
description: Optional custom prompt to override or enhance the default system prompt.
|
||||
required: false
|
||||
example: "Focus on energy-saving automations"
|
||||
example: "Focus on energy-saving automations"
|
||||
selector:
|
||||
text:
|
||||
multiline: true
|
||||
@@ -22,20 +22,22 @@ generate_suggestions:
|
||||
description: "If true, consider all entities instead of just new entities."
|
||||
required: false
|
||||
default: false
|
||||
example: false
|
||||
selector:
|
||||
boolean: {}
|
||||
domains:
|
||||
name: Domains
|
||||
description: "List of domains to consider. If empty, consider all domains."
|
||||
required: false
|
||||
default: {}
|
||||
selector:
|
||||
object: {}
|
||||
default: {}
|
||||
entity_limit:
|
||||
name: Entity Limit
|
||||
description: "Maximum number of entities to consider (randomly selected from the chosen domains)."
|
||||
required: false
|
||||
default: 200
|
||||
example: 200
|
||||
selector:
|
||||
number:
|
||||
min: 50
|
||||
@@ -46,5 +48,17 @@ generate_suggestions:
|
||||
description: "Reads and appends the yaml code of the automations found in the automations.yaml file. This action will use a lot of input tokens, use it with care and with models with a large input window (e.g. Gemini)."
|
||||
required: false
|
||||
default: false
|
||||
example: false
|
||||
selector:
|
||||
boolean: {}
|
||||
automation_limit:
|
||||
name: Automation Limit
|
||||
description: "Maximum number of automations to analyze (default: 100)."
|
||||
required: false
|
||||
default: 100
|
||||
example: 100
|
||||
selector:
|
||||
number:
|
||||
min: 10
|
||||
max: 500
|
||||
mode: slider
|
||||
|
||||
@@ -3,15 +3,195 @@
|
||||
"config": {
|
||||
"step": {
|
||||
"user": {
|
||||
"title": "Configure AI Automation Suggester",
|
||||
"title": "Select AI Provider",
|
||||
"data": {
|
||||
"provider": "AI Provider"
|
||||
}
|
||||
},
|
||||
"openai": {
|
||||
"title": "Configure OpenAI",
|
||||
"data": {
|
||||
"openai_api_key": "OpenAI API Key",
|
||||
"openai_model": "OpenAI Model",
|
||||
"openai_temperature": "Temperature",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"anthropic": {
|
||||
"title": "Configure Anthropic",
|
||||
"data": {
|
||||
"anthropic_api_key": "Anthropic API Key",
|
||||
"anthropic_model": "Anthropic Model",
|
||||
"anthropic_temperature": "Temperature",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"google": {
|
||||
"title": "Configure Google",
|
||||
"data": {
|
||||
"google_api_key": "Google API Key",
|
||||
"google_model": "Google Model",
|
||||
"google_temperature": "Temperature",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"groq": {
|
||||
"title": "Configure Groq",
|
||||
"data": {
|
||||
"groq_api_key": "Groq API Key",
|
||||
"groq_model": "Groq Model",
|
||||
"groq_temperature": "Temperature",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"localai": {
|
||||
"title": "Configure LocalAI",
|
||||
"data": {
|
||||
"localai_ip": "LocalAI IP Address",
|
||||
"localai_port": "LocalAI Port",
|
||||
"localai_https": "Use HTTPS for LocalAI",
|
||||
"localai_model": "LocalAI Model",
|
||||
"localai_temperature": "Temperature",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"ollama": {
|
||||
"title": "Configure Ollama",
|
||||
"data": {
|
||||
"ollama_ip": "Ollama IP Address",
|
||||
"ollama_port": "Ollama Port",
|
||||
"ollama_https": "Use HTTPS for Ollama",
|
||||
"ollama_model": "Ollama Model",
|
||||
"ollama_temperature": "Temperature",
|
||||
"ollama_disable_think": "Disable Think Mode (Ollama)",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"custom_openai": {
|
||||
"title": "Configure Custom OpenAI",
|
||||
"data": {
|
||||
"custom_openai_endpoint": "Custom OpenAI Endpoint",
|
||||
"custom_openai_api_key": "Custom OpenAI API Key (Optional)",
|
||||
"custom_openai_model": "Custom OpenAI Model",
|
||||
"custom_openai_temperature": "Temperature",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"mistral": {
|
||||
"title": "Configure Mistral AI",
|
||||
"data": {
|
||||
"mistral_api_key": "Mistral API Key",
|
||||
"mistral_model": "Mistral Model",
|
||||
"mistral_temperature": "Temperature",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"perplexity": {
|
||||
"title": "Configure Perplexity AI",
|
||||
"data": {
|
||||
"perplexity_api_key": "Perplexity API Key",
|
||||
"perplexity_model": "Perplexity Model",
|
||||
"perplexity_temperature": "Temperature",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"openrouter": {
|
||||
"title": "Configure OpenRouter",
|
||||
"data": {
|
||||
"openrouter_api_key": "OpenRouter API Key",
|
||||
"openrouter_model": "OpenRouter Model",
|
||||
"openrouter_reasoning_max_tokens": "OpenRouter Reasoning Max Tokens",
|
||||
"openrouter_temperature": "Temperature",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"openai_azure": {
|
||||
"title": "Configure OpenAI Azure",
|
||||
"data": {
|
||||
"openai_azure_api_key": "Azure OpenAI API Key",
|
||||
"openai_azure_deployment_id": "Azure Deployment ID",
|
||||
"openai_azure_endpoint": "Azure Endpoint",
|
||||
"openai_azure_api_version": "Azure API Version",
|
||||
"openai_azure_temperature": "Temperature",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
}
|
||||
},
|
||||
"error": {
|
||||
"required": "This field is required.",
|
||||
"invalid_api_key": "The API key is invalid."
|
||||
"api_error": "API validation failed: {error_message}",
|
||||
"cannot_connect": "Failed to connect. Please check your settings and network.",
|
||||
"unknown": "An unknown error occurred. Please check logs for details."
|
||||
},
|
||||
"abort": {
|
||||
"already_configured": "This AI provider is already configured. You can edit it from the integrations page."
|
||||
}
|
||||
},
|
||||
"options": {
|
||||
"step": {
|
||||
"init": {
|
||||
"title": "AI Automation Suggester Options",
|
||||
"data": {
|
||||
"openai_api_key": "OpenAI API Key",
|
||||
"openai_model": "OpenAI Model",
|
||||
"openai_temperature": "Temperature (OpenAI)",
|
||||
"anthropic_api_key": "Anthropic API Key",
|
||||
"anthropic_model": "Anthropic Model",
|
||||
"anthropic_temperature": "Temperature (Anthropic)",
|
||||
"google_api_key": "Google API Key",
|
||||
"google_model": "Google Model",
|
||||
"google_temperature": "Temperature (Google)",
|
||||
"groq_api_key": "Groq API Key",
|
||||
"groq_model": "Groq Model",
|
||||
"groq_temperature": "Temperature (Groq)",
|
||||
"localai_ip": "LocalAI IP Address",
|
||||
"localai_port": "LocalAI Port",
|
||||
"localai_https": "Use HTTPS for LocalAI",
|
||||
"localai_model": "LocalAI Model",
|
||||
"localai_temperature": "Temperature (LocalAI)",
|
||||
"ollama_ip": "Ollama IP Address",
|
||||
"ollama_port": "Ollama Port",
|
||||
"ollama_https": "Use HTTPS for Ollama",
|
||||
"ollama_model": "Ollama Model",
|
||||
"ollama_temperature": "Temperature (Ollama)",
|
||||
"ollama_disable_think": "Disable Think Mode (Ollama)",
|
||||
"custom_openai_endpoint": "Custom OpenAI Endpoint",
|
||||
"custom_openai_api_key": "Custom OpenAI API Key",
|
||||
"custom_openai_model": "Custom OpenAI Model",
|
||||
"custom_openai_temperature": "Temperature (Custom OpenAI)",
|
||||
"mistral_api_key": "Mistral API Key",
|
||||
"mistral_model": "Mistral Model",
|
||||
"mistral_temperature": "Temperature (Mistral AI)",
|
||||
"perplexity_api_key": "Perplexity API Key",
|
||||
"perplexity_model": "Perplexity Model",
|
||||
"perplexity_temperature": "Temperature (Perplexity AI)",
|
||||
"openrouter_api_key": "OpenRouter API Key",
|
||||
"openrouter_model": "OpenRouter Model",
|
||||
"openrouter_reasoning_max_tokens": "OpenRouter Reasoning Max Tokens",
|
||||
"openrouter_temperature": "Temperature (OpenRouter)",
|
||||
"openai_azure_api_key": "Azure OpenAI API Key",
|
||||
"openai_azure_deployment_id": "Azure Deployment ID",
|
||||
"openai_azure_endpoint": "Azure Endpoint",
|
||||
"openai_azure_api_version": "Azure API Version",
|
||||
"openai_azure_temperature": "Temperature (OpenAI Azure)",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
},
|
||||
"description": "Adjust settings for your AI providers. Fields relevant to your configured provider will be used. Common settings like token limits are configured per provider."
|
||||
}
|
||||
},
|
||||
"error": {
|
||||
"invalid_input": "One or more values are invalid."
|
||||
}
|
||||
},
|
||||
"services": {
|
||||
@@ -19,11 +199,35 @@
|
||||
"name": "Generate Suggestions",
|
||||
"description": "Manually trigger AI automation suggestions.",
|
||||
"fields": {
|
||||
"provider_config": {
|
||||
"name": "Provider Configuration",
|
||||
"description": "Which provider configuration to use (if you have multiple)"
|
||||
},
|
||||
"custom_prompt": {
|
||||
"name": "Custom Prompt",
|
||||
"description": "Optional custom prompt to override the default system prompt or guide the suggestions towards specific themes"
|
||||
},
|
||||
"all_entities": {
|
||||
"name": "Consider All Entities",
|
||||
"description": "If true, consider all entities instead of just new entities."
|
||||
},
|
||||
"domains": {
|
||||
"name": "Domains",
|
||||
"description": "List of domains to consider. If empty, consider all domains."
|
||||
},
|
||||
"entity_limit": {
|
||||
"name": "Entity Limit",
|
||||
"description": "Maximum number of entities to consider (randomly selected)."
|
||||
},
|
||||
"automation_read_yaml": {
|
||||
"name": "Read 'automations.yaml' file",
|
||||
"description": "Reads and appends the yaml code of the automations found in the automations.yaml file. This action will use a lot of input tokens, use it with care and with models with a large input window (e.g. Gemini)."
|
||||
},
|
||||
"automation_limit": {
|
||||
"name": "Automation Limit",
|
||||
"description": "Maximum number of automations to analyze (default: 100)."
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -1,44 +1,203 @@
|
||||
{
|
||||
"title": "Suggeridor d'Automatitzacions amb IA",
|
||||
"config": {
|
||||
"step": {
|
||||
"user": {
|
||||
"title": "Configurar AI Automation Suggester",
|
||||
"title": "Seleccionar Proveïdor d'IA",
|
||||
"data": {
|
||||
"provider": "Proveïdor d'IA",
|
||||
"max_tokens": "Màxim de Tokens de Sortida"
|
||||
"provider": "Proveïdor d'IA"
|
||||
}
|
||||
},
|
||||
"provider_config": {
|
||||
"title": "Configuració del Proveïdor",
|
||||
"openai": {
|
||||
"title": "Configurar OpenAI",
|
||||
"data": {
|
||||
"model": "Nom del Model",
|
||||
"api_key": "Clau d'API",
|
||||
"max_tokens": "Màxim de Tokens de Sortida",
|
||||
"ip_address": "Adreça IP",
|
||||
"port": "Port",
|
||||
"use_https": "Utilitzar HTTPS"
|
||||
},
|
||||
"description": "El màxim de tokens controla la longitud de la resposta de la IA. El valor predeterminat és 500. Augmenta si necessites respostes més llargues."
|
||||
"openai_api_key": "Clau API d'OpenAI",
|
||||
"openai_model": "Model OpenAI",
|
||||
"openai_temperature": "Temperatura",
|
||||
"max_input_tokens": "Màx Tokens d'Entrada",
|
||||
"max_output_tokens": "Màx Tokens de Sortida"
|
||||
}
|
||||
},
|
||||
"anthropic": {
|
||||
"title": "Configurar Anthropic",
|
||||
"data": {
|
||||
"anthropic_api_key": "Clau API d'Anthropic",
|
||||
"anthropic_model": "Model Anthropic",
|
||||
"anthropic_temperature": "Temperatura",
|
||||
"max_input_tokens": "Màx Tokens d'Entrada",
|
||||
"max_output_tokens": "Màx Tokens de Sortida"
|
||||
}
|
||||
},
|
||||
"google": {
|
||||
"title": "Configurar Google",
|
||||
"data": {
|
||||
"google_api_key": "Clau API de Google",
|
||||
"google_model": "Model Google",
|
||||
"google_temperature": "Temperatura",
|
||||
"max_input_tokens": "Màx Tokens d'Entrada",
|
||||
"max_output_tokens": "Màx Tokens de Sortida"
|
||||
}
|
||||
},
|
||||
"groq": {
|
||||
"title": "Configurar Groq",
|
||||
"data": {
|
||||
"groq_api_key": "Clau API de Groq",
|
||||
"groq_model": "Model Groq",
|
||||
"groq_temperature": "Temperatura",
|
||||
"max_input_tokens": "Màx Tokens d'Entrada",
|
||||
"max_output_tokens": "Màx Tokens de Sortida"
|
||||
}
|
||||
},
|
||||
"localai": {
|
||||
"title": "Configurar LocalAI",
|
||||
"data": {
|
||||
"localai_ip": "Adreça IP de LocalAI",
|
||||
"localai_port": "Port de LocalAI",
|
||||
"localai_https": "Utilitzar HTTPS per a LocalAI",
|
||||
"localai_model": "Model LocalAI",
|
||||
"localai_temperature": "Temperatura",
|
||||
"max_input_tokens": "Màx Tokens d'Entrada",
|
||||
"max_output_tokens": "Màx Tokens de Sortida"
|
||||
}
|
||||
},
|
||||
"ollama": {
|
||||
"title": "Configurar Ollama",
|
||||
"data": {
|
||||
"ollama_ip": "Adreça IP d'Ollama",
|
||||
"ollama_port": "Port d'Ollama",
|
||||
"ollama_https": "Utilitzar HTTPS per a Ollama",
|
||||
"ollama_model": "Model Ollama",
|
||||
"ollama_temperature": "Temperatura",
|
||||
"ollama_disable_think": "Desactivar Mode Pensament (Ollama)",
|
||||
"max_input_tokens": "Màx Tokens d'Entrada",
|
||||
"max_output_tokens": "Màx Tokens de Sortida"
|
||||
}
|
||||
},
|
||||
"custom_openai": {
|
||||
"title": "Configurar OpenAI Personalitzat",
|
||||
"data": {
|
||||
"custom_openai_endpoint": "Endpoint OpenAI Personalitzat",
|
||||
"custom_openai_api_key": "Clau API OpenAI Personalitzada (Opcional)",
|
||||
"custom_openai_model": "Model OpenAI Personalitzat",
|
||||
"custom_openai_temperature": "Temperatura",
|
||||
"max_input_tokens": "Màx Tokens d'Entrada",
|
||||
"max_output_tokens": "Màx Tokens de Sortida"
|
||||
}
|
||||
},
|
||||
"mistral": {
|
||||
"title": "Configurar Mistral AI",
|
||||
"data": {
|
||||
"mistral_api_key": "Clau API de Mistral",
|
||||
"mistral_model": "Model Mistral",
|
||||
"mistral_temperature": "Temperatura",
|
||||
"max_input_tokens": "Màx Tokens d'Entrada",
|
||||
"max_output_tokens": "Màx Tokens de Sortida"
|
||||
}
|
||||
},
|
||||
"perplexity": {
|
||||
"title": "Configurar Perplexity AI",
|
||||
"data": {
|
||||
"perplexity_api_key": "Clau API de Perplexity",
|
||||
"perplexity_model": "Model Perplexity",
|
||||
"perplexity_temperature": "Temperatura",
|
||||
"max_input_tokens": "Màx Tokens d'Entrada",
|
||||
"max_output_tokens": "Màx Tokens de Sortida"
|
||||
}
|
||||
},
|
||||
"openrouter": {
|
||||
"title": "Configurar OpenRouter",
|
||||
"data": {
|
||||
"openrouter_api_key": "Clau API d'OpenRouter",
|
||||
"openrouter_model": "Model OpenRouter",
|
||||
"openrouter_reasoning_max_tokens": "Màx Tokens de Raonament OpenRouter",
|
||||
"openrouter_temperature": "Temperatura",
|
||||
"max_input_tokens": "Màx Tokens d'Entrada",
|
||||
"max_output_tokens": "Màx Tokens de Sortida"
|
||||
}
|
||||
},
|
||||
"openai_azure": {
|
||||
"title": "Configurar OpenAI Azure",
|
||||
"data": {
|
||||
"openai_azure_api_key": "Clau API d'Azure OpenAI",
|
||||
"openai_azure_deployment_id": "ID de Desplegament d'Azure",
|
||||
"openai_azure_endpoint": "Endpoint d'Azure",
|
||||
"openai_azure_api_version": "Versió API d'Azure",
|
||||
"openai_azure_temperature": "Temperatura",
|
||||
"max_input_tokens": "Màx Tokens d'Entrada",
|
||||
"max_output_tokens": "Màx Tokens de Sortida"
|
||||
}
|
||||
}
|
||||
},
|
||||
"error": {
|
||||
"cannot_connect": "Error en connectar amb el servei",
|
||||
"invalid_auth": "Autenticació invàlida",
|
||||
"invalid_config": "Configuració invàlida",
|
||||
"unknown": "Error inesperat",
|
||||
"no_entities": "No s'han trobat noves entitats",
|
||||
"api_error": "Ha ocorregut un error d'API",
|
||||
"required_field": "Aquest camp és obligatori"
|
||||
"api_error": "Ha fallat la validació de l'API: {error_message}",
|
||||
"cannot_connect": "No s'ha pogut connectar. Si us plau, revisa la teva configuració i xarxa.",
|
||||
"unknown": "S'ha produït un error desconegut. Si us plau, revisa els registres per a més detalls."
|
||||
},
|
||||
"abort": {
|
||||
"already_configured": "El proveïdor ja està configurat",
|
||||
"provider_not_supported": "Aquest proveïdor no és compatible actualment"
|
||||
"already_configured": "Aquest proveïdor d'IA ja està configurat. Pots editar-lo des de la pàgina d'integracions."
|
||||
}
|
||||
},
|
||||
"options": {
|
||||
"step": {
|
||||
"init": {
|
||||
"title": "Opcions del Suggeridor d'Automatitzacions amb IA",
|
||||
"data": {
|
||||
"openai_api_key": "Clau API d'OpenAI",
|
||||
"openai_model": "Model OpenAI",
|
||||
"openai_temperature": "Temperatura (OpenAI)",
|
||||
"anthropic_api_key": "Clau API d'Anthropic",
|
||||
"anthropic_model": "Model Anthropic",
|
||||
"anthropic_temperature": "Temperatura (Anthropic)",
|
||||
"google_api_key": "Clau API de Google",
|
||||
"google_model": "Model Google",
|
||||
"google_temperature": "Temperatura (Google)",
|
||||
"groq_api_key": "Clau API de Groq",
|
||||
"groq_model": "Model Groq",
|
||||
"groq_temperature": "Temperatura (Groq)",
|
||||
"localai_ip": "Adreça IP de LocalAI",
|
||||
"localai_port": "Port de LocalAI",
|
||||
"localai_https": "Utilitzar HTTPS per a LocalAI",
|
||||
"localai_model": "Model LocalAI",
|
||||
"localai_temperature": "Temperatura (LocalAI)",
|
||||
"ollama_ip": "Adreça IP d'Ollama",
|
||||
"ollama_port": "Port d'Ollama",
|
||||
"ollama_https": "Utilitzar HTTPS per a Ollama",
|
||||
"ollama_model": "Model Ollama",
|
||||
"ollama_temperature": "Temperatura (Ollama)",
|
||||
"ollama_disable_think": "Desactivar Mode Pensament (Ollama)",
|
||||
"custom_openai_endpoint": "Endpoint OpenAI Personalitzat",
|
||||
"custom_openai_api_key": "Clau API OpenAI Personalitzada",
|
||||
"custom_openai_model": "Model OpenAI Personalitzat",
|
||||
"custom_openai_temperature": "Temperatura (OpenAI Personalitzat)",
|
||||
"mistral_api_key": "Clau API de Mistral",
|
||||
"mistral_model": "Model Mistral",
|
||||
"mistral_temperature": "Temperatura (Mistral AI)",
|
||||
"perplexity_api_key": "Clau API de Perplexity",
|
||||
"perplexity_model": "Model Perplexity",
|
||||
"perplexity_temperature": "Temperatura (Perplexity AI)",
|
||||
"openrouter_api_key": "Clau API d'OpenRouter",
|
||||
"openrouter_model": "Model OpenRouter",
|
||||
"openrouter_reasoning_max_tokens": "Màx Tokens de Raonament OpenRouter",
|
||||
"openrouter_temperature": "Temperatura (OpenRouter)",
|
||||
"openai_azure_api_key": "Clau API d'Azure OpenAI",
|
||||
"openai_azure_deployment_id": "ID de Desplegament d'Azure",
|
||||
"openai_azure_endpoint": "Endpoint d'Azure",
|
||||
"openai_azure_api_version": "Versió API d'Azure",
|
||||
"openai_azure_temperature": "Temperatura (OpenAI Azure)",
|
||||
"max_input_tokens": "Màx Tokens d'Entrada",
|
||||
"max_output_tokens": "Màx Tokens de Sortida"
|
||||
},
|
||||
"description": "Ajusta la configuració per als teus proveïdors d'IA. S'utilitzaran els camps rellevants per al teu proveïdor configurat. Configuracions comunes com els límits de tokens es configuren per proveïdor."
|
||||
}
|
||||
},
|
||||
"error": {
|
||||
"invalid_input": "Un o més valors són invàlids."
|
||||
}
|
||||
},
|
||||
"services": {
|
||||
"generate_suggestions": {
|
||||
"name": "Generar Suggeriments",
|
||||
"description": "Activar manualment els suggeriments d'automatització d'IA",
|
||||
"description": "Activar manualment els suggeriments d'automatització amb IA.",
|
||||
"fields": {
|
||||
"provider_config": {
|
||||
"name": "Configuració del Proveïdor",
|
||||
@@ -62,7 +221,11 @@
|
||||
},
|
||||
"automation_read_yaml": {
|
||||
"name": "Llegir arxiu 'automations.yaml'",
|
||||
"description": "Llegeix i addiciona el codi yaml de les automatitzacions trobades en el fitxer automations.yaml. Aquesta acció utilitzarà molts tokens d'entrada, utilitza-la amb cura i amb models amb una finestra d'entrada gran (per exemple, Gemini)."
|
||||
"description": "Llegeix i afegeix el codi yaml de les automatitzacions trobades a l'arxiu automations.yaml. Aquesta acció utilitzarà molts tokens d'entrada, utilitza-la amb cura i amb models amb una finestra d'entrada gran (ex. Gemini)."
|
||||
},
|
||||
"automation_limit": {
|
||||
"name": "Límit d'automatitzacions",
|
||||
"description": "Nombre màxim d'automatitzacions a analitzar (per defecte: 100)."
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,71 +1,203 @@
|
||||
{
|
||||
"title": "KI Automatisierungs-Vorschlag",
|
||||
"config": {
|
||||
"step": {
|
||||
"user": {
|
||||
"title": "AI Automation Suggester konfigurieren",
|
||||
"description": "Wählen Sie Ihren KI-Anbieter und konfigurieren Sie die Einstellungen",
|
||||
"title": "KI-Anbieter auswählen",
|
||||
"data": {
|
||||
"provider": "KI-Anbieter",
|
||||
"scan_frequency": "Scan-Häufigkeit (Stunden)",
|
||||
"initial_lag_time": "Anfangsverzögerung (Minuten)",
|
||||
"max_tokens": "Maximale Ausgabe-Token"
|
||||
"provider": "KI-Anbieter"
|
||||
}
|
||||
},
|
||||
"openai": {
|
||||
"title": "OpenAI Konfiguration",
|
||||
"title": "OpenAI konfigurieren",
|
||||
"data": {
|
||||
"api_key": "API-Schlüssel",
|
||||
"model": "Modell",
|
||||
"max_tokens": "Maximale Ausgabe-Token"
|
||||
},
|
||||
"description": "Die maximale Anzahl von Token bestimmt die Länge der Antwort der KI. Standard ist 500. Erhöhen Sie diesen Wert, wenn Sie längere Antworten benötigen."
|
||||
"openai_api_key": "OpenAI API-Schlüssel",
|
||||
"openai_model": "OpenAI-Modell",
|
||||
"openai_temperature": "Temperatur",
|
||||
"max_input_tokens": "Max. Eingabe-Tokens",
|
||||
"max_output_tokens": "Max. Ausgabe-Tokens"
|
||||
}
|
||||
},
|
||||
"anthropic": {
|
||||
"title": "Anthropic Konfiguration",
|
||||
"title": "Anthropic konfigurieren",
|
||||
"data": {
|
||||
"api_key": "API-Schlüssel",
|
||||
"model": "Modell",
|
||||
"max_tokens": "Maximale Ausgabe-Token"
|
||||
},
|
||||
"description": "Die maximale Anzahl von Token bestimmt die Länge der Antwort der KI. Standard ist 500. Erhöhen Sie diesen Wert, wenn Sie längere Antworten benötigen."
|
||||
"anthropic_api_key": "Anthropic API-Schlüssel",
|
||||
"anthropic_model": "Anthropic-Modell",
|
||||
"anthropic_temperature": "Temperatur",
|
||||
"max_input_tokens": "Max. Eingabe-Tokens",
|
||||
"max_output_tokens": "Max. Ausgabe-Tokens"
|
||||
}
|
||||
},
|
||||
"google": {
|
||||
"title": "Google konfigurieren",
|
||||
"data": {
|
||||
"google_api_key": "Google API-Schlüssel",
|
||||
"google_model": "Google-Modell",
|
||||
"google_temperature": "Temperatur",
|
||||
"max_input_tokens": "Max. Eingabe-Tokens",
|
||||
"max_output_tokens": "Max. Ausgabe-Tokens"
|
||||
}
|
||||
},
|
||||
"groq": {
|
||||
"title": "Groq konfigurieren",
|
||||
"data": {
|
||||
"groq_api_key": "Groq API-Schlüssel",
|
||||
"groq_model": "Groq-Modell",
|
||||
"groq_temperature": "Temperatur",
|
||||
"max_input_tokens": "Max. Eingabe-Tokens",
|
||||
"max_output_tokens": "Max. Ausgabe-Tokens"
|
||||
}
|
||||
},
|
||||
"localai": {
|
||||
"title": "LocalAI Konfiguration",
|
||||
"title": "LocalAI konfigurieren",
|
||||
"data": {
|
||||
"ip_address": "IP-Adresse",
|
||||
"port": "Port",
|
||||
"https": "HTTPS verwenden",
|
||||
"model": "Modellname",
|
||||
"max_tokens": "Maximale Ausgabe-Token"
|
||||
},
|
||||
"description": "Die maximale Anzahl von Token bestimmt die Länge der Antwort der KI. Standard ist 500. Erhöhen Sie diesen Wert, wenn Sie längere Antworten benötigen."
|
||||
"localai_ip": "LocalAI IP-Adresse",
|
||||
"localai_port": "LocalAI Port",
|
||||
"localai_https": "HTTPS für LocalAI verwenden",
|
||||
"localai_model": "LocalAI-Modell",
|
||||
"localai_temperature": "Temperatur",
|
||||
"max_input_tokens": "Max. Eingabe-Tokens",
|
||||
"max_output_tokens": "Max. Ausgabe-Tokens"
|
||||
}
|
||||
},
|
||||
"ollama": {
|
||||
"title": "Ollama Konfiguration",
|
||||
"title": "Ollama konfigurieren",
|
||||
"data": {
|
||||
"ip_address": "IP-Adresse",
|
||||
"port": "Port",
|
||||
"https": "HTTPS verwenden",
|
||||
"model": "Modellname",
|
||||
"max_tokens": "Maximale Ausgabe-Token"
|
||||
},
|
||||
"description": "Die maximale Anzahl von Token bestimmt die Länge der Antwort der KI. Standard ist 500. Erhöhen Sie diesen Wert, wenn Sie längere Antworten benötigen."
|
||||
"ollama_ip": "Ollama IP-Adresse",
|
||||
"ollama_port": "Ollama Port",
|
||||
"ollama_https": "HTTPS für Ollama verwenden",
|
||||
"ollama_model": "Ollama-Modell",
|
||||
"ollama_temperature": "Temperatur",
|
||||
"ollama_disable_think": "Denkmodus deaktivieren (Ollama)",
|
||||
"max_input_tokens": "Max. Eingabe-Tokens",
|
||||
"max_output_tokens": "Max. Ausgabe-Tokens"
|
||||
}
|
||||
},
|
||||
"custom_openai": {
|
||||
"title": "Benutzerdefiniertes OpenAI konfigurieren",
|
||||
"data": {
|
||||
"custom_openai_endpoint": "Benutzerdefinierter OpenAI-Endpunkt",
|
||||
"custom_openai_api_key": "Benutzerdefinierter OpenAI API-Schlüssel (Optional)",
|
||||
"custom_openai_model": "Benutzerdefiniertes OpenAI-Modell",
|
||||
"custom_openai_temperature": "Temperatur",
|
||||
"max_input_tokens": "Max. Eingabe-Tokens",
|
||||
"max_output_tokens": "Max. Ausgabe-Tokens"
|
||||
}
|
||||
},
|
||||
"mistral": {
|
||||
"title": "Mistral AI konfigurieren",
|
||||
"data": {
|
||||
"mistral_api_key": "Mistral API-Schlüssel",
|
||||
"mistral_model": "Mistral-Modell",
|
||||
"mistral_temperature": "Temperatur",
|
||||
"max_input_tokens": "Max. Eingabe-Tokens",
|
||||
"max_output_tokens": "Max. Ausgabe-Tokens"
|
||||
}
|
||||
},
|
||||
"perplexity": {
|
||||
"title": "Perplexity AI konfigurieren",
|
||||
"data": {
|
||||
"perplexity_api_key": "Perplexity API-Schlüssel",
|
||||
"perplexity_model": "Perplexity-Modell",
|
||||
"perplexity_temperature": "Temperatur",
|
||||
"max_input_tokens": "Max. Eingabe-Tokens",
|
||||
"max_output_tokens": "Max. Ausgabe-Tokens"
|
||||
}
|
||||
},
|
||||
"openrouter": {
|
||||
"title": "OpenRouter konfigurieren",
|
||||
"data": {
|
||||
"openrouter_api_key": "OpenRouter API-Schlüssel",
|
||||
"openrouter_model": "OpenRouter-Modell",
|
||||
"openrouter_reasoning_max_tokens": "Max. Tokens für Schlussfolgerungen (OpenRouter)",
|
||||
"openrouter_temperature": "Temperatur",
|
||||
"max_input_tokens": "Max. Eingabe-Tokens",
|
||||
"max_output_tokens": "Max. Ausgabe-Tokens"
|
||||
}
|
||||
},
|
||||
"openai_azure": {
|
||||
"title": "OpenAI Azure konfigurieren",
|
||||
"data": {
|
||||
"openai_azure_api_key": "Azure OpenAI API-Schlüssel",
|
||||
"openai_azure_deployment_id": "Azure Deployment-ID",
|
||||
"openai_azure_endpoint": "Azure-Endpunkt",
|
||||
"openai_azure_api_version": "Azure API-Version",
|
||||
"openai_azure_temperature": "Temperatur",
|
||||
"max_input_tokens": "Max. Eingabe-Tokens",
|
||||
"max_output_tokens": "Max. Ausgabe-Tokens"
|
||||
}
|
||||
}
|
||||
},
|
||||
"error": {
|
||||
"cannot_connect": "Verbindung zum Dienst fehlgeschlagen",
|
||||
"invalid_auth": "Ungültige Authentifizierung",
|
||||
"invalid_config": "Ungültige Konfiguration",
|
||||
"unknown": "Unerwarteter Fehler",
|
||||
"no_entities": "Keine neuen Entitäten gefunden",
|
||||
"api_error": "API-Fehler aufgetreten",
|
||||
"required_field": "Dieses Feld ist erforderlich"
|
||||
"api_error": "API-Validierung fehlgeschlagen: {error_message}",
|
||||
"cannot_connect": "Verbindung fehlgeschlagen. Bitte überprüfen Sie Ihre Einstellungen und Ihr Netzwerk.",
|
||||
"unknown": "Ein unbekannter Fehler ist aufgetreten. Bitte überprüfen Sie die Protokolle für Details."
|
||||
},
|
||||
"abort": {
|
||||
"already_configured": "Dieser KI-Anbieter ist bereits konfiguriert. Sie können ihn auf der Integrationsseite bearbeiten."
|
||||
}
|
||||
},
|
||||
"options": {
|
||||
"step": {
|
||||
"init": {
|
||||
"title": "Optionen für KI Automatisierungs-Vorschlag",
|
||||
"data": {
|
||||
"openai_api_key": "OpenAI API-Schlüssel",
|
||||
"openai_model": "OpenAI-Modell",
|
||||
"openai_temperature": "Temperatur (OpenAI)",
|
||||
"anthropic_api_key": "Anthropic API-Schlüssel",
|
||||
"anthropic_model": "Anthropic-Modell",
|
||||
"anthropic_temperature": "Temperatur (Anthropic)",
|
||||
"google_api_key": "Google API-Schlüssel",
|
||||
"google_model": "Google-Modell",
|
||||
"google_temperature": "Temperatur (Google)",
|
||||
"groq_api_key": "Groq API-Schlüssel",
|
||||
"groq_model": "Groq-Modell",
|
||||
"groq_temperature": "Temperatur (Groq)",
|
||||
"localai_ip": "LocalAI IP-Adresse",
|
||||
"localai_port": "LocalAI Port",
|
||||
"localai_https": "HTTPS für LocalAI verwenden",
|
||||
"localai_model": "LocalAI-Modell",
|
||||
"localai_temperature": "Temperatur (LocalAI)",
|
||||
"ollama_ip": "Ollama IP-Adresse",
|
||||
"ollama_port": "Ollama Port",
|
||||
"ollama_https": "HTTPS für Ollama verwenden",
|
||||
"ollama_model": "Ollama-Modell",
|
||||
"ollama_temperature": "Temperatur (Ollama)",
|
||||
"ollama_disable_think": "Denkmodus deaktivieren (Ollama)",
|
||||
"custom_openai_endpoint": "Benutzerdefinierter OpenAI-Endpunkt",
|
||||
"custom_openai_api_key": "Benutzerdefinierter OpenAI API-Schlüssel",
|
||||
"custom_openai_model": "Benutzerdefiniertes OpenAI-Modell",
|
||||
"custom_openai_temperature": "Temperatur (Benutzerdefiniertes OpenAI)",
|
||||
"mistral_api_key": "Mistral API-Schlüssel",
|
||||
"mistral_model": "Mistral-Modell",
|
||||
"mistral_temperature": "Temperatur (Mistral AI)",
|
||||
"perplexity_api_key": "Perplexity API-Schlüssel",
|
||||
"perplexity_model": "Perplexity-Modell",
|
||||
"perplexity_temperature": "Temperatur (Perplexity AI)",
|
||||
"openrouter_api_key": "OpenRouter API-Schlüssel",
|
||||
"openrouter_model": "OpenRouter-Modell",
|
||||
"openrouter_reasoning_max_tokens": "Max. Tokens für Schlussfolgerungen (OpenRouter)",
|
||||
"openrouter_temperature": "Temperatur (OpenRouter)",
|
||||
"openai_azure_api_key": "Azure OpenAI API-Schlüssel",
|
||||
"openai_azure_deployment_id": "Azure Deployment-ID",
|
||||
"openai_azure_endpoint": "Azure-Endpunkt",
|
||||
"openai_azure_api_version": "Azure API-Version",
|
||||
"openai_azure_temperature": "Temperatur (OpenAI Azure)",
|
||||
"max_input_tokens": "Max. Eingabe-Tokens",
|
||||
"max_output_tokens": "Max. Ausgabe-Tokens"
|
||||
},
|
||||
"description": "Passen Sie die Einstellungen für Ihre KI-Anbieter an. Es werden die für Ihren konfigurierten Anbieter relevanten Felder verwendet. Allgemeine Einstellungen wie Token-Limits werden pro Anbieter konfiguriert."
|
||||
}
|
||||
},
|
||||
"error": {
|
||||
"invalid_input": "Ein oder mehrere Werte sind ungültig."
|
||||
}
|
||||
},
|
||||
"services": {
|
||||
"generate_suggestions": {
|
||||
"name": "Generiere Vorschläge",
|
||||
"description": "Löst manuell AI Automationsvorschläge aus",
|
||||
"name": "Vorschläge generieren",
|
||||
"description": "Manuelles Auslösen von KI-Automatisierungsvorschlägen.",
|
||||
"fields": {
|
||||
"provider_config": {
|
||||
"name": "Anbieterkonfiguration",
|
||||
@@ -88,8 +220,12 @@
|
||||
"description": "Maximale Anzahl von Entitäten, die berücksichtigt werden (zufällig ausgewählt)."
|
||||
},
|
||||
"automation_read_yaml": {
|
||||
"name": "Read 'automations.yaml' file",
|
||||
"description": "Reads and appends the yaml code of the automations found in the automations.yaml file. This action will use a lot of input tokens, use it with care and with models with a large input window (e.g. Gemini)."
|
||||
"name": "Datei 'automations.yaml' lesen",
|
||||
"description": "Liest und hängt den YAML-Code der in der Datei automations.yaml gefundenen Automatisierungen an. Diese Aktion verbraucht viele Eingabe-Tokens. Verwenden Sie sie mit Bedacht und mit Modellen mit einem großen Eingabefenster (z.B. Gemini)."
|
||||
},
|
||||
"automation_limit": {
|
||||
"name": "Automatisierungslimit",
|
||||
"description": "Maximale Anzahl der zu analysierenden Automatisierungen (Standard: 100)."
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,44 +1,203 @@
|
||||
{
|
||||
"title": "AI Automation Suggester",
|
||||
"config": {
|
||||
"step": {
|
||||
"user": {
|
||||
"title": "Configure AI Automation Suggester",
|
||||
"title": "Select AI Provider",
|
||||
"data": {
|
||||
"provider": "AI Provider",
|
||||
"max_tokens": "Maximum Output Tokens"
|
||||
"provider": "AI Provider"
|
||||
}
|
||||
},
|
||||
"provider_config": {
|
||||
"title": "Provider Settings",
|
||||
"openai": {
|
||||
"title": "Configure OpenAI",
|
||||
"data": {
|
||||
"model": "Model Name",
|
||||
"api_key": "API Key",
|
||||
"max_tokens": "Maximum Output Tokens",
|
||||
"ip_address": "IP Address",
|
||||
"port": "Port",
|
||||
"use_https": "Use HTTPS"
|
||||
},
|
||||
"description": "Maximum tokens controls the length of the AI's response. Default is 500. Increase if you need longer responses."
|
||||
"openai_api_key": "OpenAI API Key",
|
||||
"openai_model": "OpenAI Model",
|
||||
"openai_temperature": "Temperature",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"anthropic": {
|
||||
"title": "Configure Anthropic",
|
||||
"data": {
|
||||
"anthropic_api_key": "Anthropic API Key",
|
||||
"anthropic_model": "Anthropic Model",
|
||||
"anthropic_temperature": "Temperature",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"google": {
|
||||
"title": "Configure Google",
|
||||
"data": {
|
||||
"google_api_key": "Google API Key",
|
||||
"google_model": "Google Model",
|
||||
"google_temperature": "Temperature",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"groq": {
|
||||
"title": "Configure Groq",
|
||||
"data": {
|
||||
"groq_api_key": "Groq API Key",
|
||||
"groq_model": "Groq Model",
|
||||
"groq_temperature": "Temperature",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"localai": {
|
||||
"title": "Configure LocalAI",
|
||||
"data": {
|
||||
"localai_ip": "LocalAI IP Address",
|
||||
"localai_port": "LocalAI Port",
|
||||
"localai_https": "Use HTTPS for LocalAI",
|
||||
"localai_model": "LocalAI Model",
|
||||
"localai_temperature": "Temperature",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"ollama": {
|
||||
"title": "Configure Ollama",
|
||||
"data": {
|
||||
"ollama_ip": "Ollama IP Address",
|
||||
"ollama_port": "Ollama Port",
|
||||
"ollama_https": "Use HTTPS for Ollama",
|
||||
"ollama_model": "Ollama Model",
|
||||
"ollama_temperature": "Temperature",
|
||||
"ollama_disable_think": "Disable Think Mode (Ollama)",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"custom_openai": {
|
||||
"title": "Configure Custom OpenAI",
|
||||
"data": {
|
||||
"custom_openai_endpoint": "Custom OpenAI Endpoint",
|
||||
"custom_openai_api_key": "Custom OpenAI API Key (Optional)",
|
||||
"custom_openai_model": "Custom OpenAI Model",
|
||||
"custom_openai_temperature": "Temperature",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"mistral": {
|
||||
"title": "Configure Mistral AI",
|
||||
"data": {
|
||||
"mistral_api_key": "Mistral API Key",
|
||||
"mistral_model": "Mistral Model",
|
||||
"mistral_temperature": "Temperature",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"perplexity": {
|
||||
"title": "Configure Perplexity AI",
|
||||
"data": {
|
||||
"perplexity_api_key": "Perplexity API Key",
|
||||
"perplexity_model": "Perplexity Model",
|
||||
"perplexity_temperature": "Temperature",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"openrouter": {
|
||||
"title": "Configure OpenRouter",
|
||||
"data": {
|
||||
"openrouter_api_key": "OpenRouter API Key",
|
||||
"openrouter_model": "OpenRouter Model",
|
||||
"openrouter_reasoning_max_tokens": "OpenRouter Reasoning Max Tokens",
|
||||
"openrouter_temperature": "Temperature",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"openai_azure": {
|
||||
"title": "Configure OpenAI Azure",
|
||||
"data": {
|
||||
"openai_azure_api_key": "Azure OpenAI API Key",
|
||||
"openai_azure_deployment_id": "Azure Deployment ID",
|
||||
"openai_azure_endpoint": "Azure Endpoint",
|
||||
"openai_azure_api_version": "Azure API Version",
|
||||
"openai_azure_temperature": "Temperature",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
}
|
||||
},
|
||||
"error": {
|
||||
"cannot_connect": "Failed to connect to service",
|
||||
"invalid_auth": "Invalid authentication",
|
||||
"invalid_config": "Invalid configuration",
|
||||
"unknown": "Unexpected error",
|
||||
"no_entities": "No new entities found",
|
||||
"api_error": "API error occurred",
|
||||
"required_field": "This field is required"
|
||||
"api_error": "API validation failed: {error_message}",
|
||||
"cannot_connect": "Failed to connect. Please check your settings and network.",
|
||||
"unknown": "An unknown error occurred. Please check logs for details."
|
||||
},
|
||||
"abort": {
|
||||
"already_configured": "Provider is already configured",
|
||||
"provider_not_supported": "This provider is not currently supported"
|
||||
"already_configured": "This AI provider is already configured. You can edit it from the integrations page."
|
||||
}
|
||||
},
|
||||
"options": {
|
||||
"step": {
|
||||
"init": {
|
||||
"title": "AI Automation Suggester Options",
|
||||
"data": {
|
||||
"openai_api_key": "OpenAI API Key",
|
||||
"openai_model": "OpenAI Model",
|
||||
"openai_temperature": "Temperature (OpenAI)",
|
||||
"anthropic_api_key": "Anthropic API Key",
|
||||
"anthropic_model": "Anthropic Model",
|
||||
"anthropic_temperature": "Temperature (Anthropic)",
|
||||
"google_api_key": "Google API Key",
|
||||
"google_model": "Google Model",
|
||||
"google_temperature": "Temperature (Google)",
|
||||
"groq_api_key": "Groq API Key",
|
||||
"groq_model": "Groq Model",
|
||||
"groq_temperature": "Temperature (Groq)",
|
||||
"localai_ip": "LocalAI IP Address",
|
||||
"localai_port": "LocalAI Port",
|
||||
"localai_https": "Use HTTPS for LocalAI",
|
||||
"localai_model": "LocalAI Model",
|
||||
"localai_temperature": "Temperature (LocalAI)",
|
||||
"ollama_ip": "Ollama IP Address",
|
||||
"ollama_port": "Ollama Port",
|
||||
"ollama_https": "Use HTTPS for Ollama",
|
||||
"ollama_model": "Ollama Model",
|
||||
"ollama_temperature": "Temperature (Ollama)",
|
||||
"ollama_disable_think": "Disable Think Mode (Ollama)",
|
||||
"custom_openai_endpoint": "Custom OpenAI Endpoint",
|
||||
"custom_openai_api_key": "Custom OpenAI API Key",
|
||||
"custom_openai_model": "Custom OpenAI Model",
|
||||
"custom_openai_temperature": "Temperature (Custom OpenAI)",
|
||||
"mistral_api_key": "Mistral API Key",
|
||||
"mistral_model": "Mistral Model",
|
||||
"mistral_temperature": "Temperature (Mistral AI)",
|
||||
"perplexity_api_key": "Perplexity API Key",
|
||||
"perplexity_model": "Perplexity Model",
|
||||
"perplexity_temperature": "Temperature (Perplexity AI)",
|
||||
"openrouter_api_key": "OpenRouter API Key",
|
||||
"openrouter_model": "OpenRouter Model",
|
||||
"openrouter_reasoning_max_tokens": "OpenRouter Reasoning Max Tokens",
|
||||
"openrouter_temperature": "Temperature (OpenRouter)",
|
||||
"openai_azure_api_key": "Azure OpenAI API Key",
|
||||
"openai_azure_deployment_id": "Azure Deployment ID",
|
||||
"openai_azure_endpoint": "Azure Endpoint",
|
||||
"openai_azure_api_version": "Azure API Version",
|
||||
"openai_azure_temperature": "Temperature (OpenAI Azure)",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
},
|
||||
"description": "Adjust settings for your AI providers. Fields relevant to your configured provider will be used. Common settings like token limits are configured per provider."
|
||||
}
|
||||
},
|
||||
"error": {
|
||||
"invalid_input": "One or more values are invalid."
|
||||
}
|
||||
},
|
||||
"services": {
|
||||
"generate_suggestions": {
|
||||
"name": "Generate Suggestions",
|
||||
"description": "Manually trigger AI automation suggestions",
|
||||
"description": "Manually trigger AI automation suggestions.",
|
||||
"fields": {
|
||||
"provider_config": {
|
||||
"name": "Provider Configuration",
|
||||
@@ -63,6 +222,10 @@
|
||||
"automation_read_yaml": {
|
||||
"name": "Read 'automations.yaml' file",
|
||||
"description": "Reads and appends the yaml code of the automations found in the automations.yaml file. This action will use a lot of input tokens, use it with care and with models with a large input window (e.g. Gemini)."
|
||||
},
|
||||
"automation_limit": {
|
||||
"name": "Automation Limit",
|
||||
"description": "Maximum number of automations to analyze (default: 100)."
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,44 +1,203 @@
|
||||
{
|
||||
"title": "Sugeridor de Automatizaciones con IA",
|
||||
"config": {
|
||||
"step": {
|
||||
"user": {
|
||||
"title": "Configurar AI Automation Suggester",
|
||||
"title": "Seleccionar Proveedor de IA",
|
||||
"data": {
|
||||
"provider": "Proveedor de IA",
|
||||
"max_tokens": "Máximo de Tokens de Salida"
|
||||
"provider": "Proveedor de IA"
|
||||
}
|
||||
},
|
||||
"provider_config": {
|
||||
"title": "Configuración del Proveedor",
|
||||
"openai": {
|
||||
"title": "Configurar OpenAI",
|
||||
"data": {
|
||||
"model": "Nombre del Modelo",
|
||||
"api_key": "Clave de API",
|
||||
"max_tokens": "Máximo de Tokens de Salida",
|
||||
"ip_address": "Dirección IP",
|
||||
"port": "Puerto",
|
||||
"use_https": "Usar HTTPS"
|
||||
},
|
||||
"description": "El máximo de tokens controla la longitud de la respuesta de la IA. El valor predeterminado es 500. Aumenta si necesitas respuestas más largas."
|
||||
"openai_api_key": "Clave API de OpenAI",
|
||||
"openai_model": "Modelo OpenAI",
|
||||
"openai_temperature": "Temperatura",
|
||||
"max_input_tokens": "Máx Tokens de Entrada",
|
||||
"max_output_tokens": "Máx Tokens de Salida"
|
||||
}
|
||||
},
|
||||
"anthropic": {
|
||||
"title": "Configurar Anthropic",
|
||||
"data": {
|
||||
"anthropic_api_key": "Clave API de Anthropic",
|
||||
"anthropic_model": "Modelo Anthropic",
|
||||
"anthropic_temperature": "Temperatura",
|
||||
"max_input_tokens": "Máx Tokens de Entrada",
|
||||
"max_output_tokens": "Máx Tokens de Salida"
|
||||
}
|
||||
},
|
||||
"google": {
|
||||
"title": "Configurar Google",
|
||||
"data": {
|
||||
"google_api_key": "Clave API de Google",
|
||||
"google_model": "Modelo Google",
|
||||
"google_temperature": "Temperatura",
|
||||
"max_input_tokens": "Máx Tokens de Entrada",
|
||||
"max_output_tokens": "Máx Tokens de Salida"
|
||||
}
|
||||
},
|
||||
"groq": {
|
||||
"title": "Configurar Groq",
|
||||
"data": {
|
||||
"groq_api_key": "Clave API de Groq",
|
||||
"groq_model": "Modelo Groq",
|
||||
"groq_temperature": "Temperatura",
|
||||
"max_input_tokens": "Máx Tokens de Entrada",
|
||||
"max_output_tokens": "Máx Tokens de Salida"
|
||||
}
|
||||
},
|
||||
"localai": {
|
||||
"title": "Configurar LocalAI",
|
||||
"data": {
|
||||
"localai_ip": "Dirección IP de LocalAI",
|
||||
"localai_port": "Puerto de LocalAI",
|
||||
"localai_https": "Usar HTTPS para LocalAI",
|
||||
"localai_model": "Modelo LocalAI",
|
||||
"localai_temperature": "Temperatura",
|
||||
"max_input_tokens": "Máx Tokens de Entrada",
|
||||
"max_output_tokens": "Máx Tokens de Salida"
|
||||
}
|
||||
},
|
||||
"ollama": {
|
||||
"title": "Configurar Ollama",
|
||||
"data": {
|
||||
"ollama_ip": "Dirección IP de Ollama",
|
||||
"ollama_port": "Puerto de Ollama",
|
||||
"ollama_https": "Usar HTTPS para Ollama",
|
||||
"ollama_model": "Modelo Ollama",
|
||||
"ollama_temperature": "Temperatura",
|
||||
"ollama_disable_think": "Desactivar Modo Pensamiento (Ollama)",
|
||||
"max_input_tokens": "Máx Tokens de Entrada",
|
||||
"max_output_tokens": "Máx Tokens de Salida"
|
||||
}
|
||||
},
|
||||
"custom_openai": {
|
||||
"title": "Configurar OpenAI Personalizado",
|
||||
"data": {
|
||||
"custom_openai_endpoint": "Endpoint OpenAI Personalizado",
|
||||
"custom_openai_api_key": "Clave API OpenAI Personalizada (Opcional)",
|
||||
"custom_openai_model": "Modelo OpenAI Personalizado",
|
||||
"custom_openai_temperature": "Temperatura",
|
||||
"max_input_tokens": "Máx Tokens de Entrada",
|
||||
"max_output_tokens": "Máx Tokens de Salida"
|
||||
}
|
||||
},
|
||||
"mistral": {
|
||||
"title": "Configurar Mistral AI",
|
||||
"data": {
|
||||
"mistral_api_key": "Clave API de Mistral",
|
||||
"mistral_model": "Modelo Mistral",
|
||||
"mistral_temperature": "Temperatura",
|
||||
"max_input_tokens": "Máx Tokens de Entrada",
|
||||
"max_output_tokens": "Máx Tokens de Salida"
|
||||
}
|
||||
},
|
||||
"perplexity": {
|
||||
"title": "Configurar Perplexity AI",
|
||||
"data": {
|
||||
"perplexity_api_key": "Clave API de Perplexity",
|
||||
"perplexity_model": "Modelo Perplexity",
|
||||
"perplexity_temperature": "Temperatura",
|
||||
"max_input_tokens": "Máx Tokens de Entrada",
|
||||
"max_output_tokens": "Máx Tokens de Salida"
|
||||
}
|
||||
},
|
||||
"openrouter": {
|
||||
"title": "Configurar OpenRouter",
|
||||
"data": {
|
||||
"openrouter_api_key": "Clave API de OpenRouter",
|
||||
"openrouter_model": "Modelo OpenRouter",
|
||||
"openrouter_reasoning_max_tokens": "Máx Tokens de Razonamiento OpenRouter",
|
||||
"openrouter_temperature": "Temperatura",
|
||||
"max_input_tokens": "Máx Tokens de Entrada",
|
||||
"max_output_tokens": "Máx Tokens de Salida"
|
||||
}
|
||||
},
|
||||
"openai_azure": {
|
||||
"title": "Configurar OpenAI Azure",
|
||||
"data": {
|
||||
"openai_azure_api_key": "Clave API de Azure OpenAI",
|
||||
"openai_azure_deployment_id": "ID de Despliegue de Azure",
|
||||
"openai_azure_endpoint": "Endpoint de Azure",
|
||||
"openai_azure_api_version": "Versión API de Azure",
|
||||
"openai_azure_temperature": "Temperatura",
|
||||
"max_input_tokens": "Máx Tokens de Entrada",
|
||||
"max_output_tokens": "Máx Tokens de Salida"
|
||||
}
|
||||
}
|
||||
},
|
||||
"error": {
|
||||
"cannot_connect": "Error al conectar con el servicio",
|
||||
"invalid_auth": "Autenticación inválida",
|
||||
"invalid_config": "Configuración inválida",
|
||||
"unknown": "Error inesperado",
|
||||
"no_entities": "No se encontraron nuevas entidades",
|
||||
"api_error": "Ha ocurrido un error de API",
|
||||
"required_field": "Este campo es obligatorio"
|
||||
"api_error": "Falló la validación de la API: {error_message}",
|
||||
"cannot_connect": "No se pudo conectar. Por favor, revisa tu configuración y red.",
|
||||
"unknown": "Ocurrió un error desconocido. Por favor, revisa los logs para más detalles."
|
||||
},
|
||||
"abort": {
|
||||
"already_configured": "El proveedor ya está configurado",
|
||||
"provider_not_supported": "Este proveedor no es compatible actualmente"
|
||||
"already_configured": "Este proveedor de IA ya está configurado. Puedes editarlo desde la página de integraciones."
|
||||
}
|
||||
},
|
||||
"options": {
|
||||
"step": {
|
||||
"init": {
|
||||
"title": "Opciones del Sugeridor de Automatizaciones con IA",
|
||||
"data": {
|
||||
"openai_api_key": "Clave API de OpenAI",
|
||||
"openai_model": "Modelo OpenAI",
|
||||
"openai_temperature": "Temperatura (OpenAI)",
|
||||
"anthropic_api_key": "Clave API de Anthropic",
|
||||
"anthropic_model": "Modelo Anthropic",
|
||||
"anthropic_temperature": "Temperatura (Anthropic)",
|
||||
"google_api_key": "Clave API de Google",
|
||||
"google_model": "Modelo Google",
|
||||
"google_temperature": "Temperatura (Google)",
|
||||
"groq_api_key": "Clave API de Groq",
|
||||
"groq_model": "Modelo Groq",
|
||||
"groq_temperature": "Temperatura (Groq)",
|
||||
"localai_ip": "Dirección IP de LocalAI",
|
||||
"localai_port": "Puerto de LocalAI",
|
||||
"localai_https": "Usar HTTPS para LocalAI",
|
||||
"localai_model": "Modelo LocalAI",
|
||||
"localai_temperature": "Temperatura (LocalAI)",
|
||||
"ollama_ip": "Dirección IP de Ollama",
|
||||
"ollama_port": "Puerto de Ollama",
|
||||
"ollama_https": "Usar HTTPS para Ollama",
|
||||
"ollama_model": "Modelo Ollama",
|
||||
"ollama_temperature": "Temperatura (Ollama)",
|
||||
"ollama_disable_think": "Desactivar Modo Pensamiento (Ollama)",
|
||||
"custom_openai_endpoint": "Endpoint OpenAI Personalizado",
|
||||
"custom_openai_api_key": "Clave API OpenAI Personalizada",
|
||||
"custom_openai_model": "Modelo OpenAI Personalizado",
|
||||
"custom_openai_temperature": "Temperatura (OpenAI Personalizado)",
|
||||
"mistral_api_key": "Clave API de Mistral",
|
||||
"mistral_model": "Modelo Mistral",
|
||||
"mistral_temperature": "Temperatura (Mistral AI)",
|
||||
"perplexity_api_key": "Clave API de Perplexity",
|
||||
"perplexity_model": "Modelo Perplexity",
|
||||
"perplexity_temperature": "Temperatura (Perplexity AI)",
|
||||
"openrouter_api_key": "Clave API de OpenRouter",
|
||||
"openrouter_model": "Modelo OpenRouter",
|
||||
"openrouter_reasoning_max_tokens": "Máx Tokens de Razonamiento OpenRouter",
|
||||
"openrouter_temperature": "Temperatura (OpenRouter)",
|
||||
"openai_azure_api_key": "Clave API de Azure OpenAI",
|
||||
"openai_azure_deployment_id": "ID de Despliegue de Azure",
|
||||
"openai_azure_endpoint": "Endpoint de Azure",
|
||||
"openai_azure_api_version": "Versión API de Azure",
|
||||
"openai_azure_temperature": "Temperatura (OpenAI Azure)",
|
||||
"max_input_tokens": "Máx Tokens de Entrada",
|
||||
"max_output_tokens": "Máx Tokens de Salida"
|
||||
},
|
||||
"description": "Ajusta la configuración para tus proveedores de IA. Se usarán los campos relevantes para tu proveedor configurado. Configuraciones comunes como los límites de tokens se configuran por proveedor."
|
||||
}
|
||||
},
|
||||
"error": {
|
||||
"invalid_input": "Uno o más valores son inválidos."
|
||||
}
|
||||
},
|
||||
"services": {
|
||||
"generate_suggestions": {
|
||||
"name": "Generar Sugerencias",
|
||||
"description": "Activar manualmente las sugerencias de automatización de IA",
|
||||
"description": "Activar manualmente las sugerencias de automatización con IA.",
|
||||
"fields": {
|
||||
"provider_config": {
|
||||
"name": "Configuración del Proveedor",
|
||||
@@ -62,7 +221,11 @@
|
||||
},
|
||||
"automation_read_yaml": {
|
||||
"name": "Leer archivo 'automations.yaml'",
|
||||
"description": "Lee y añade el código yaml de las automatizaciones encontradas en el fichero automations.yaml. Esta acción utilizará muchos tokens de entrada, utilízala con cuidado y con modelos con una ventana de entrada grande (por ejemplo, Gemini)."
|
||||
"description": "Lee y añade el código yaml de las automatizaciones encontradas en el archivo automations.yaml. Esta acción usará muchos tokens de entrada, úsala con cuidado y con modelos con una ventana de entrada grande (ej. Gemini)."
|
||||
},
|
||||
"automation_limit": {
|
||||
"name": "Límite de automatizaciones",
|
||||
"description": "Número máximo de automatizaciones a analizar (por defecto: 100)."
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,8 +1,203 @@
|
||||
{
|
||||
"title": "Suggeritore di Automazioni AI",
|
||||
"config": {
|
||||
"step": {
|
||||
"user": {
|
||||
"title": "Seleziona Provider AI",
|
||||
"data": {
|
||||
"provider": "Provider AI"
|
||||
}
|
||||
},
|
||||
"openai": {
|
||||
"title": "Configura OpenAI",
|
||||
"data": {
|
||||
"openai_api_key": "Chiave API OpenAI",
|
||||
"openai_model": "Modello OpenAI",
|
||||
"openai_temperature": "Temperatura",
|
||||
"max_input_tokens": "Max Token di Input",
|
||||
"max_output_tokens": "Max Token di Output"
|
||||
}
|
||||
},
|
||||
"anthropic": {
|
||||
"title": "Configura Anthropic",
|
||||
"data": {
|
||||
"anthropic_api_key": "Chiave API Anthropic",
|
||||
"anthropic_model": "Modello Anthropic",
|
||||
"anthropic_temperature": "Temperatura",
|
||||
"max_input_tokens": "Max Token di Input",
|
||||
"max_output_tokens": "Max Token di Output"
|
||||
}
|
||||
},
|
||||
"google": {
|
||||
"title": "Configura Google",
|
||||
"data": {
|
||||
"google_api_key": "Chiave API Google",
|
||||
"google_model": "Modello Google",
|
||||
"google_temperature": "Temperatura",
|
||||
"max_input_tokens": "Max Token di Input",
|
||||
"max_output_tokens": "Max Token di Output"
|
||||
}
|
||||
},
|
||||
"groq": {
|
||||
"title": "Configura Groq",
|
||||
"data": {
|
||||
"groq_api_key": "Chiave API Groq",
|
||||
"groq_model": "Modello Groq",
|
||||
"groq_temperature": "Temperatura",
|
||||
"max_input_tokens": "Max Token di Input",
|
||||
"max_output_tokens": "Max Token di Output"
|
||||
}
|
||||
},
|
||||
"localai": {
|
||||
"title": "Configura LocalAI",
|
||||
"data": {
|
||||
"localai_ip": "Indirizzo IP LocalAI",
|
||||
"localai_port": "Porta LocalAI",
|
||||
"localai_https": "Usa HTTPS per LocalAI",
|
||||
"localai_model": "Modello LocalAI",
|
||||
"localai_temperature": "Temperatura",
|
||||
"max_input_tokens": "Max Token di Input",
|
||||
"max_output_tokens": "Max Token di Output"
|
||||
}
|
||||
},
|
||||
"ollama": {
|
||||
"title": "Configura Ollama",
|
||||
"data": {
|
||||
"ollama_ip": "Indirizzo IP Ollama",
|
||||
"ollama_port": "Porta Ollama",
|
||||
"ollama_https": "Usa HTTPS per Ollama",
|
||||
"ollama_model": "Modello Ollama",
|
||||
"ollama_temperature": "Temperatura",
|
||||
"ollama_disable_think": "Disabilita Modalità Pensiero (Ollama)",
|
||||
"max_input_tokens": "Max Token di Input",
|
||||
"max_output_tokens": "Max Token di Output"
|
||||
}
|
||||
},
|
||||
"custom_openai": {
|
||||
"title": "Configura OpenAI Personalizzato",
|
||||
"data": {
|
||||
"custom_openai_endpoint": "Endpoint OpenAI Personalizzato",
|
||||
"custom_openai_api_key": "Chiave API OpenAI Personalizzata (Opzionale)",
|
||||
"custom_openai_model": "Modello OpenAI Personalizzato",
|
||||
"custom_openai_temperature": "Temperatura",
|
||||
"max_input_tokens": "Max Token di Input",
|
||||
"max_output_tokens": "Max Token di Output"
|
||||
}
|
||||
},
|
||||
"mistral": {
|
||||
"title": "Configura Mistral AI",
|
||||
"data": {
|
||||
"mistral_api_key": "Chiave API Mistral",
|
||||
"mistral_model": "Modello Mistral",
|
||||
"mistral_temperature": "Temperatura",
|
||||
"max_input_tokens": "Max Token di Input",
|
||||
"max_output_tokens": "Max Token di Output"
|
||||
}
|
||||
},
|
||||
"perplexity": {
|
||||
"title": "Configura Perplexity AI",
|
||||
"data": {
|
||||
"perplexity_api_key": "Chiave API Perplexity",
|
||||
"perplexity_model": "Modello Perplexity",
|
||||
"perplexity_temperature": "Temperatura",
|
||||
"max_input_tokens": "Max Token di Input",
|
||||
"max_output_tokens": "Max Token di Output"
|
||||
}
|
||||
},
|
||||
"openrouter": {
|
||||
"title": "Configura OpenRouter",
|
||||
"data": {
|
||||
"openrouter_api_key": "Chiave API OpenRouter",
|
||||
"openrouter_model": "Modello OpenRouter",
|
||||
"openrouter_reasoning_max_tokens": "Max Token di Ragionamento OpenRouter",
|
||||
"openrouter_temperature": "Temperatura",
|
||||
"max_input_tokens": "Max Token di Input",
|
||||
"max_output_tokens": "Max Token di Output"
|
||||
}
|
||||
},
|
||||
"openai_azure": {
|
||||
"title": "Configura OpenAI Azure",
|
||||
"data": {
|
||||
"openai_azure_api_key": "Chiave API Azure OpenAI",
|
||||
"openai_azure_deployment_id": "ID Deployment Azure",
|
||||
"openai_azure_endpoint": "Endpoint Azure",
|
||||
"openai_azure_api_version": "Versione API Azure",
|
||||
"openai_azure_temperature": "Temperatura",
|
||||
"max_input_tokens": "Max Token di Input",
|
||||
"max_output_tokens": "Max Token di Output"
|
||||
}
|
||||
}
|
||||
},
|
||||
"error": {
|
||||
"api_error": "Validazione API fallita: {error_message}",
|
||||
"cannot_connect": "Connessione fallita. Controlla le impostazioni e la rete.",
|
||||
"unknown": "Si è verificato un errore sconosciuto. Controlla i log per i dettagli."
|
||||
},
|
||||
"abort": {
|
||||
"already_configured": "Questo provider AI è già configurato. Puoi modificarlo dalla pagina delle integrazioni."
|
||||
}
|
||||
},
|
||||
"options": {
|
||||
"step": {
|
||||
"init": {
|
||||
"title": "Opzioni Suggeritore di Automazioni AI",
|
||||
"data": {
|
||||
"openai_api_key": "Chiave API OpenAI",
|
||||
"openai_model": "Modello OpenAI",
|
||||
"openai_temperature": "Temperatura (OpenAI)",
|
||||
"anthropic_api_key": "Chiave API Anthropic",
|
||||
"anthropic_model": "Modello Anthropic",
|
||||
"anthropic_temperature": "Temperatura (Anthropic)",
|
||||
"google_api_key": "Chiave API Google",
|
||||
"google_model": "Modello Google",
|
||||
"google_temperature": "Temperatura (Google)",
|
||||
"groq_api_key": "Chiave API Groq",
|
||||
"groq_model": "Modello Groq",
|
||||
"groq_temperature": "Temperatura (Groq)",
|
||||
"localai_ip": "Indirizzo IP LocalAI",
|
||||
"localai_port": "Porta LocalAI",
|
||||
"localai_https": "Usa HTTPS per LocalAI",
|
||||
"localai_model": "Modello LocalAI",
|
||||
"localai_temperature": "Temperatura (LocalAI)",
|
||||
"ollama_ip": "Indirizzo IP Ollama",
|
||||
"ollama_port": "Porta Ollama",
|
||||
"ollama_https": "Usa HTTPS per Ollama",
|
||||
"ollama_model": "Modello Ollama",
|
||||
"ollama_temperature": "Temperatura (Ollama)",
|
||||
"ollama_disable_think": "Disabilita Modalità Pensiero (Ollama)",
|
||||
"custom_openai_endpoint": "Endpoint OpenAI Personalizzato",
|
||||
"custom_openai_api_key": "Chiave API OpenAI Personalizzata",
|
||||
"custom_openai_model": "Modello OpenAI Personalizzato",
|
||||
"custom_openai_temperature": "Temperatura (OpenAI Personalizzato)",
|
||||
"mistral_api_key": "Chiave API Mistral",
|
||||
"mistral_model": "Modello Mistral",
|
||||
"mistral_temperature": "Temperatura (Mistral AI)",
|
||||
"perplexity_api_key": "Chiave API Perplexity",
|
||||
"perplexity_model": "Modello Perplexity",
|
||||
"perplexity_temperature": "Temperatura (Perplexity AI)",
|
||||
"openrouter_api_key": "Chiave API OpenRouter",
|
||||
"openrouter_model": "Modello OpenRouter",
|
||||
"openrouter_reasoning_max_tokens": "Max Token di Ragionamento OpenRouter",
|
||||
"openrouter_temperature": "Temperatura (OpenRouter)",
|
||||
"openai_azure_api_key": "Chiave API Azure OpenAI",
|
||||
"openai_azure_deployment_id": "ID Deployment Azure",
|
||||
"openai_azure_endpoint": "Endpoint Azure",
|
||||
"openai_azure_api_version": "Versione API Azure",
|
||||
"openai_azure_temperature": "Temperatura (OpenAI Azure)",
|
||||
"max_input_tokens": "Max Token di Input",
|
||||
"max_output_tokens": "Max Token di Output"
|
||||
},
|
||||
"description": "Regola le impostazioni per i tuoi provider AI. Verranno utilizzati i campi pertinenti al provider configurato. Impostazioni comuni come i limiti dei token sono configurate per provider."
|
||||
}
|
||||
},
|
||||
"error": {
|
||||
"invalid_input": "Uno o più valori non sono validi."
|
||||
}
|
||||
},
|
||||
"services": {
|
||||
"generate_suggestions": {
|
||||
"name": "Genera Suggerimenti",
|
||||
"description": "Lancia manualmente suggerimenti di automazione AI",
|
||||
"description": "Attiva manualmente i suggerimenti di automazione AI.",
|
||||
"fields": {
|
||||
"provider_config": {
|
||||
"name": "Configurazione Provider",
|
||||
@@ -25,8 +220,12 @@
|
||||
"description": "Numero massimo di entità da considerare (scelte casualmente)."
|
||||
},
|
||||
"automation_read_yaml": {
|
||||
"name": "Read 'automations.yaml' file",
|
||||
"description": "Reads and appends the yaml code of the automations found in the automations.yaml file. This action will use a lot of input tokens, use it with care and with models with a large input window (e.g. Gemini)."
|
||||
"name": "Leggi file 'automations.yaml'",
|
||||
"description": "Legge e aggiunge il codice yaml delle automazioni trovate nel file automations.yaml. Questa azione utilizzerà molti token di input, usala con cautela e con modelli con una grande finestra di input (es. Gemini)."
|
||||
},
|
||||
"automation_limit": {
|
||||
"name": "Limite automazioni",
|
||||
"description": "Numero massimo di automazioni da analizzare (predefinito: 100)."
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,8 +1,203 @@
|
||||
{
|
||||
"title": "AI Automatisering Suggester",
|
||||
"config": {
|
||||
"step": {
|
||||
"user": {
|
||||
"title": "Selecteer AI Provider",
|
||||
"data": {
|
||||
"provider": "AI Provider"
|
||||
}
|
||||
},
|
||||
"openai": {
|
||||
"title": "Configureer OpenAI",
|
||||
"data": {
|
||||
"openai_api_key": "OpenAI API-sleutel",
|
||||
"openai_model": "OpenAI Model",
|
||||
"openai_temperature": "Temperatuur",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"anthropic": {
|
||||
"title": "Configureer Anthropic",
|
||||
"data": {
|
||||
"anthropic_api_key": "Anthropic API-sleutel",
|
||||
"anthropic_model": "Anthropic Model",
|
||||
"anthropic_temperature": "Temperatuur",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"google": {
|
||||
"title": "Configureer Google",
|
||||
"data": {
|
||||
"google_api_key": "Google API-sleutel",
|
||||
"google_model": "Google Model",
|
||||
"google_temperature": "Temperatuur",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"groq": {
|
||||
"title": "Configureer Groq",
|
||||
"data": {
|
||||
"groq_api_key": "Groq API-sleutel",
|
||||
"groq_model": "Groq Model",
|
||||
"groq_temperature": "Temperatuur",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"localai": {
|
||||
"title": "Configureer LocalAI",
|
||||
"data": {
|
||||
"localai_ip": "LocalAI IP-adres",
|
||||
"localai_port": "LocalAI Poort",
|
||||
"localai_https": "Gebruik HTTPS voor LocalAI",
|
||||
"localai_model": "LocalAI Model",
|
||||
"localai_temperature": "Temperatuur",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"ollama": {
|
||||
"title": "Configureer Ollama",
|
||||
"data": {
|
||||
"ollama_ip": "Ollama IP-adres",
|
||||
"ollama_port": "Ollama Poort",
|
||||
"ollama_https": "Gebruik HTTPS voor Ollama",
|
||||
"ollama_model": "Ollama Model",
|
||||
"ollama_temperature": "Temperatuur",
|
||||
"ollama_disable_think": "Denkmodus uitschakelen (Ollama)",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"custom_openai": {
|
||||
"title": "Configureer Aangepaste OpenAI",
|
||||
"data": {
|
||||
"custom_openai_endpoint": "Aangepast OpenAI Eindpunt",
|
||||
"custom_openai_api_key": "Aangepaste OpenAI API-sleutel (Optioneel)",
|
||||
"custom_openai_model": "Aangepast OpenAI Model",
|
||||
"custom_openai_temperature": "Temperatuur",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"mistral": {
|
||||
"title": "Configureer Mistral AI",
|
||||
"data": {
|
||||
"mistral_api_key": "Mistral API-sleutel",
|
||||
"mistral_model": "Mistral Model",
|
||||
"mistral_temperature": "Temperatuur",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"perplexity": {
|
||||
"title": "Configureer Perplexity AI",
|
||||
"data": {
|
||||
"perplexity_api_key": "Perplexity API-sleutel",
|
||||
"perplexity_model": "Perplexity Model",
|
||||
"perplexity_temperature": "Temperatuur",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"openrouter": {
|
||||
"title": "Configureer OpenRouter",
|
||||
"data": {
|
||||
"openrouter_api_key": "OpenRouter API-sleutel",
|
||||
"openrouter_model": "OpenRouter Model",
|
||||
"openrouter_reasoning_max_tokens": "OpenRouter Redenering Max Tokens",
|
||||
"openrouter_temperature": "Temperatuur",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
},
|
||||
"openai_azure": {
|
||||
"title": "Configureer OpenAI Azure",
|
||||
"data": {
|
||||
"openai_azure_api_key": "Azure OpenAI API-sleutel",
|
||||
"openai_azure_deployment_id": "Azure Implementatie-ID",
|
||||
"openai_azure_endpoint": "Azure Eindpunt",
|
||||
"openai_azure_api_version": "Azure API Versie",
|
||||
"openai_azure_temperature": "Temperatuur",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
}
|
||||
}
|
||||
},
|
||||
"error": {
|
||||
"api_error": "API-validatie mislukt: {error_message}",
|
||||
"cannot_connect": "Verbinding mislukt. Controleer uw instellingen en netwerk.",
|
||||
"unknown": "Er is een onbekende fout opgetreden. Controleer de logs voor details."
|
||||
},
|
||||
"abort": {
|
||||
"already_configured": "Deze AI-provider is al geconfigureerd. U kunt deze bewerken via de integratiepagina."
|
||||
}
|
||||
},
|
||||
"options": {
|
||||
"step": {
|
||||
"init": {
|
||||
"title": "AI Automatisering Suggester Opties",
|
||||
"data": {
|
||||
"openai_api_key": "OpenAI API-sleutel",
|
||||
"openai_model": "OpenAI Model",
|
||||
"openai_temperature": "Temperatuur (OpenAI)",
|
||||
"anthropic_api_key": "Anthropic API-sleutel",
|
||||
"anthropic_model": "Anthropic Model",
|
||||
"anthropic_temperature": "Temperatuur (Anthropic)",
|
||||
"google_api_key": "Google API-sleutel",
|
||||
"google_model": "Google Model",
|
||||
"google_temperature": "Temperatuur (Google)",
|
||||
"groq_api_key": "Groq API-sleutel",
|
||||
"groq_model": "Groq Model",
|
||||
"groq_temperature": "Temperatuur (Groq)",
|
||||
"localai_ip": "LocalAI IP-adres",
|
||||
"localai_port": "LocalAI Poort",
|
||||
"localai_https": "Gebruik HTTPS voor LocalAI",
|
||||
"localai_model": "LocalAI Model",
|
||||
"localai_temperature": "Temperatuur (LocalAI)",
|
||||
"ollama_ip": "Ollama IP-adres",
|
||||
"ollama_port": "Ollama Poort",
|
||||
"ollama_https": "Gebruik HTTPS voor Ollama",
|
||||
"ollama_model": "Ollama Model",
|
||||
"ollama_temperature": "Temperatuur (Ollama)",
|
||||
"ollama_disable_think": "Denkmodus uitschakelen (Ollama)",
|
||||
"custom_openai_endpoint": "Aangepast OpenAI Eindpunt",
|
||||
"custom_openai_api_key": "Aangepaste OpenAI API-sleutel",
|
||||
"custom_openai_model": "Aangepast OpenAI Model",
|
||||
"custom_openai_temperature": "Temperatuur (Aangepaste OpenAI)",
|
||||
"mistral_api_key": "Mistral API-sleutel",
|
||||
"mistral_model": "Mistral Model",
|
||||
"mistral_temperature": "Temperatuur (Mistral AI)",
|
||||
"perplexity_api_key": "Perplexity API-sleutel",
|
||||
"perplexity_model": "Perplexity Model",
|
||||
"perplexity_temperature": "Temperatuur (Perplexity AI)",
|
||||
"openrouter_api_key": "OpenRouter API-sleutel",
|
||||
"openrouter_model": "OpenRouter Model",
|
||||
"openrouter_reasoning_max_tokens": "OpenRouter Redenering Max Tokens",
|
||||
"openrouter_temperature": "Temperatuur (OpenRouter)",
|
||||
"openai_azure_api_key": "Azure OpenAI API-sleutel",
|
||||
"openai_azure_deployment_id": "Azure Implementatie-ID",
|
||||
"openai_azure_endpoint": "Azure Eindpunt",
|
||||
"openai_azure_api_version": "Azure API Versie",
|
||||
"openai_azure_temperature": "Temperatuur (OpenAI Azure)",
|
||||
"max_input_tokens": "Max Input Tokens",
|
||||
"max_output_tokens": "Max Output Tokens"
|
||||
},
|
||||
"description": "Pas instellingen aan voor uw AI-providers. Velden die relevant zijn voor uw geconfigureerde provider worden gebruikt. Gemeenschappelijke instellingen zoals tokenlimieten worden per provider geconfigureerd."
|
||||
}
|
||||
},
|
||||
"error": {
|
||||
"invalid_input": "Een of meer waarden zijn ongeldig."
|
||||
}
|
||||
},
|
||||
"services": {
|
||||
"generate_suggestions": {
|
||||
"name": "Genereer Suggesties",
|
||||
"description": "Handmatig AI-automatiseringssuggesties oproepen",
|
||||
"name": "Suggesties Genereren",
|
||||
"description": "Handmatig AI-automatiseringssuggesties activeren.",
|
||||
"fields": {
|
||||
"provider_config": {
|
||||
"name": "Aanbiederconfiguratie",
|
||||
@@ -25,8 +220,12 @@
|
||||
"description": "Maximaal aantal te beschouwen entiteiten (willekeurig geselecteerd)."
|
||||
},
|
||||
"automation_read_yaml": {
|
||||
"name": "Read 'automations.yaml' file",
|
||||
"description": "Reads and appends the yaml code of the automations found in the automations.yaml file. This action will use a lot of input tokens, use it with care and with models with a large input window (e.g. Gemini)."
|
||||
"name": "Lees 'automations.yaml' bestand",
|
||||
"description": "Leest en voegt de yaml-code van de automatiseringen in het bestand automations.yaml toe. Deze actie zal veel input tokens gebruiken, gebruik het voorzichtig en met modellen met een groot input venster (bijv. Gemini)."
|
||||
},
|
||||
"automation_limit": {
|
||||
"name": "Automatiseringslimiet",
|
||||
"description": "Maximum aantal te analyseren automatiseringen (standaard: 100)."
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,44 +1,203 @@
|
||||
{
|
||||
"title": "Советчик Автоматизаций с ИИ",
|
||||
"config": {
|
||||
"step": {
|
||||
"user": {
|
||||
"title": "Настроить AI Automation Suggester",
|
||||
"title": "Выберите Провайдера ИИ",
|
||||
"data": {
|
||||
"provider": "AI провайдер",
|
||||
"max_tokens": "Максимальное количество выходных токенов"
|
||||
"provider": "Провайдер ИИ"
|
||||
}
|
||||
},
|
||||
"provider_config": {
|
||||
"title": "Настройки провайдер",
|
||||
"openai": {
|
||||
"title": "Настроить OpenAI",
|
||||
"data": {
|
||||
"model": "Название модели",
|
||||
"api_key": "API ключ",
|
||||
"max_tokens": "Максимальное количество выходных токенов",
|
||||
"ip_address": "IP-адрес",
|
||||
"port": "Порт",
|
||||
"use_https": "Использовать HTTPS"
|
||||
},
|
||||
"description": "Максимальное количество токенов контролирует длину ответа ИИ. По умолчанию — 500. Увеличьте значение, если вам нужны более длинные ответы."
|
||||
"openai_api_key": "Ключ API OpenAI",
|
||||
"openai_model": "Модель OpenAI",
|
||||
"openai_temperature": "Температура",
|
||||
"max_input_tokens": "Макс. Входных Токенов",
|
||||
"max_output_tokens": "Макс. Выходных Токенов"
|
||||
}
|
||||
},
|
||||
"anthropic": {
|
||||
"title": "Настроить Anthropic",
|
||||
"data": {
|
||||
"anthropic_api_key": "Ключ API Anthropic",
|
||||
"anthropic_model": "Модель Anthropic",
|
||||
"anthropic_temperature": "Температура",
|
||||
"max_input_tokens": "Макс. Входных Токенов",
|
||||
"max_output_tokens": "Макс. Выходных Токенов"
|
||||
}
|
||||
},
|
||||
"google": {
|
||||
"title": "Настроить Google",
|
||||
"data": {
|
||||
"google_api_key": "Ключ API Google",
|
||||
"google_model": "Модель Google",
|
||||
"google_temperature": "Температура",
|
||||
"max_input_tokens": "Макс. Входных Токенов",
|
||||
"max_output_tokens": "Макс. Выходных Токенов"
|
||||
}
|
||||
},
|
||||
"groq": {
|
||||
"title": "Настроить Groq",
|
||||
"data": {
|
||||
"groq_api_key": "Ключ API Groq",
|
||||
"groq_model": "Модель Groq",
|
||||
"groq_temperature": "Температура",
|
||||
"max_input_tokens": "Макс. Входных Токенов",
|
||||
"max_output_tokens": "Макс. Выходных Токенов"
|
||||
}
|
||||
},
|
||||
"localai": {
|
||||
"title": "Настроить LocalAI",
|
||||
"data": {
|
||||
"localai_ip": "IP-адрес LocalAI",
|
||||
"localai_port": "Порт LocalAI",
|
||||
"localai_https": "Использовать HTTPS для LocalAI",
|
||||
"localai_model": "Модель LocalAI",
|
||||
"localai_temperature": "Температура",
|
||||
"max_input_tokens": "Макс. Входных Токенов",
|
||||
"max_output_tokens": "Макс. Выходных Токенов"
|
||||
}
|
||||
},
|
||||
"ollama": {
|
||||
"title": "Настроить Ollama",
|
||||
"data": {
|
||||
"ollama_ip": "IP-адрес Ollama",
|
||||
"ollama_port": "Порт Ollama",
|
||||
"ollama_https": "Использовать HTTPS для Ollama",
|
||||
"ollama_model": "Модель Ollama",
|
||||
"ollama_temperature": "Температура",
|
||||
"ollama_disable_think": "Отключить Режим Размышления (Ollama)",
|
||||
"max_input_tokens": "Макс. Входных Токенов",
|
||||
"max_output_tokens": "Макс. Выходных Токенов"
|
||||
}
|
||||
},
|
||||
"custom_openai": {
|
||||
"title": "Настроить Пользовательский OpenAI",
|
||||
"data": {
|
||||
"custom_openai_endpoint": "Пользовательский Endpoint OpenAI",
|
||||
"custom_openai_api_key": "Пользовательский Ключ API OpenAI (Необязательно)",
|
||||
"custom_openai_model": "Пользовательская Модель OpenAI",
|
||||
"custom_openai_temperature": "Температура",
|
||||
"max_input_tokens": "Макс. Входных Токенов",
|
||||
"max_output_tokens": "Макс. Выходных Токенов"
|
||||
}
|
||||
},
|
||||
"mistral": {
|
||||
"title": "Настроить Mistral AI",
|
||||
"data": {
|
||||
"mistral_api_key": "Ключ API Mistral",
|
||||
"mistral_model": "Модель Mistral",
|
||||
"mistral_temperature": "Температура",
|
||||
"max_input_tokens": "Макс. Входных Токенов",
|
||||
"max_output_tokens": "Макс. Выходных Токенов"
|
||||
}
|
||||
},
|
||||
"perplexity": {
|
||||
"title": "Настроить Perplexity AI",
|
||||
"data": {
|
||||
"perplexity_api_key": "Ключ API Perplexity",
|
||||
"perplexity_model": "Модель Perplexity",
|
||||
"perplexity_temperature": "Температура",
|
||||
"max_input_tokens": "Макс. Входных Токенов",
|
||||
"max_output_tokens": "Макс. Выходных Токенов"
|
||||
}
|
||||
},
|
||||
"openrouter": {
|
||||
"title": "Настроить OpenRouter",
|
||||
"data": {
|
||||
"openrouter_api_key": "Ключ API OpenRouter",
|
||||
"openrouter_model": "Модель OpenRouter",
|
||||
"openrouter_reasoning_max_tokens": "Макс. Токенов для Размышлений OpenRouter",
|
||||
"openrouter_temperature": "Температура",
|
||||
"max_input_tokens": "Макс. Входных Токенов",
|
||||
"max_output_tokens": "Макс. Выходных Токенов"
|
||||
}
|
||||
},
|
||||
"openai_azure": {
|
||||
"title": "Настроить OpenAI Azure",
|
||||
"data": {
|
||||
"openai_azure_api_key": "Ключ API Azure OpenAI",
|
||||
"openai_azure_deployment_id": "ID Развертывания Azure",
|
||||
"openai_azure_endpoint": "Endpoint Azure",
|
||||
"openai_azure_api_version": "Версия API Azure",
|
||||
"openai_azure_temperature": "Температура",
|
||||
"max_input_tokens": "Макс. Входных Токенов",
|
||||
"max_output_tokens": "Макс. Выходных Токенов"
|
||||
}
|
||||
}
|
||||
},
|
||||
"error": {
|
||||
"cannot_connect": "Не удалось подключиться к сервису",
|
||||
"invalid_auth": "Недействительная аутентификация",
|
||||
"invalid_config": "Недопустимая конфигурация",
|
||||
"unknown": "Неожиданная ошибка",
|
||||
"no_entities": "Новых сущностей не найдено",
|
||||
"api_error": "Произошла ошибка API",
|
||||
"required_field": "Это поле обязательно для заполнения"
|
||||
"api_error": "Ошибка проверки API: {error_message}",
|
||||
"cannot_connect": "Не удалось подключиться. Проверьте настройки и сеть.",
|
||||
"unknown": "Произошла неизвестная ошибка. Проверьте логи для получения подробной информации."
|
||||
},
|
||||
"abort": {
|
||||
"already_configured": "провайдер уже настроен",
|
||||
"provider_not_supported": "Этот провайдер в настоящее время не поддерживается"
|
||||
"already_configured": "Этот провайдер ИИ уже настроен. Вы можете отредактировать его на странице интеграций."
|
||||
}
|
||||
},
|
||||
"options": {
|
||||
"step": {
|
||||
"init": {
|
||||
"title": "Настройки Советчика Автоматизаций с ИИ",
|
||||
"data": {
|
||||
"openai_api_key": "Ключ API OpenAI",
|
||||
"openai_model": "Модель OpenAI",
|
||||
"openai_temperature": "Температура (OpenAI)",
|
||||
"anthropic_api_key": "Ключ API Anthropic",
|
||||
"anthropic_model": "Модель Anthropic",
|
||||
"anthropic_temperature": "Температура (Anthropic)",
|
||||
"google_api_key": "Ключ API Google",
|
||||
"google_model": "Модель Google",
|
||||
"google_temperature": "Температура (Google)",
|
||||
"groq_api_key": "Ключ API Groq",
|
||||
"groq_model": "Модель Groq",
|
||||
"groq_temperature": "Температура (Groq)",
|
||||
"localai_ip": "IP-адрес LocalAI",
|
||||
"localai_port": "Порт LocalAI",
|
||||
"localai_https": "Использовать HTTPS для LocalAI",
|
||||
"localai_model": "Модель LocalAI",
|
||||
"localai_temperature": "Температура (LocalAI)",
|
||||
"ollama_ip": "IP-адрес Ollama",
|
||||
"ollama_port": "Порт Ollama",
|
||||
"ollama_https": "Использовать HTTPS для Ollama",
|
||||
"ollama_model": "Модель Ollama",
|
||||
"ollama_temperature": "Температура (Ollama)",
|
||||
"ollama_disable_think": "Отключить Режим Размышления (Ollama)",
|
||||
"custom_openai_endpoint": "Пользовательский Endpoint OpenAI",
|
||||
"custom_openai_api_key": "Пользовательский Ключ API OpenAI",
|
||||
"custom_openai_model": "Пользовательская Модель OpenAI",
|
||||
"custom_openai_temperature": "Температура (Пользовательский OpenAI)",
|
||||
"mistral_api_key": "Ключ API Mistral",
|
||||
"mistral_model": "Модель Mistral",
|
||||
"mistral_temperature": "Температура (Mistral AI)",
|
||||
"perplexity_api_key": "Ключ API Perplexity",
|
||||
"perplexity_model": "Модель Perplexity",
|
||||
"perplexity_temperature": "Температура (Perplexity AI)",
|
||||
"openrouter_api_key": "Ключ API OpenRouter",
|
||||
"openrouter_model": "Модель OpenRouter",
|
||||
"openrouter_reasoning_max_tokens": "Макс. Токенов для Размышлений OpenRouter",
|
||||
"openrouter_temperature": "Температура (OpenRouter)",
|
||||
"openai_azure_api_key": "Ключ API Azure OpenAI",
|
||||
"openai_azure_deployment_id": "ID Развертывания Azure",
|
||||
"openai_azure_endpoint": "Endpoint Azure",
|
||||
"openai_azure_api_version": "Версия API Azure",
|
||||
"openai_azure_temperature": "Температура (OpenAI Azure)",
|
||||
"max_input_tokens": "Макс. Входных Токенов",
|
||||
"max_output_tokens": "Макс. Выходных Токенов"
|
||||
},
|
||||
"description": "Настройте параметры для ваших провайдеров ИИ. Будут использоваться поля, относящиеся к вашему настроенному провайдеру. Общие настройки, такие как лимиты токенов, настраиваются для каждого провайдера."
|
||||
}
|
||||
},
|
||||
"error": {
|
||||
"invalid_input": "Одно или несколько значений недействительны."
|
||||
}
|
||||
},
|
||||
"services": {
|
||||
"generate_suggestions": {
|
||||
"name": "Сгенерировать предложения",
|
||||
"description": "Вручную запустить предложения по автоматизации ИИ",
|
||||
"name": "Сгенерировать Предложения",
|
||||
"description": "Вручную запустить генерацию предложений по автоматизации с ИИ.",
|
||||
"fields": {
|
||||
"provider_config": {
|
||||
"name": "Конфигурация провайдер",
|
||||
@@ -61,8 +220,12 @@
|
||||
"description": "Максимальное количество сущностей для рассмотрения (выбирается случайным образом)."
|
||||
},
|
||||
"automation_read_yaml": {
|
||||
"name": "Read 'automations.yaml' file",
|
||||
"description": "Reads and appends the yaml code of the automations found in the automations.yaml file. This action will use a lot of input tokens, use it with care and with models with a large input window (e.g. Gemini)."
|
||||
"name": "Прочитать файл 'automations.yaml'",
|
||||
"description": "Читает и добавляет yaml-код автоматизаций, найденных в файле automations.yaml. Это действие будет использовать много входных токенов, используйте его осторожно и с моделями с большим окном ввода (например, Gemini)."
|
||||
},
|
||||
"automation_limit": {
|
||||
"name": "Лимит автоматизаций",
|
||||
"description": "Максимальное количество автоматизаций для анализа (по умолчанию: 100)."
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,8 +1,203 @@
|
||||
{
|
||||
"title": "AI 自动化建议器",
|
||||
"config": {
|
||||
"step": {
|
||||
"user": {
|
||||
"title": "选择 AI 提供商",
|
||||
"data": {
|
||||
"provider": "AI 提供商"
|
||||
}
|
||||
},
|
||||
"openai": {
|
||||
"title": "配置 OpenAI",
|
||||
"data": {
|
||||
"openai_api_key": "OpenAI API 密钥",
|
||||
"openai_model": "OpenAI 模型",
|
||||
"openai_temperature": "温度",
|
||||
"max_input_tokens": "最大输入令牌数",
|
||||
"max_output_tokens": "最大输出令牌数"
|
||||
}
|
||||
},
|
||||
"anthropic": {
|
||||
"title": "配置 Anthropic",
|
||||
"data": {
|
||||
"anthropic_api_key": "Anthropic API 密钥",
|
||||
"anthropic_model": "Anthropic 模型",
|
||||
"anthropic_temperature": "温度",
|
||||
"max_input_tokens": "最大输入令牌数",
|
||||
"max_output_tokens": "最大输出令牌数"
|
||||
}
|
||||
},
|
||||
"google": {
|
||||
"title": "配置 Google",
|
||||
"data": {
|
||||
"google_api_key": "Google API 密钥",
|
||||
"google_model": "Google 模型",
|
||||
"google_temperature": "温度",
|
||||
"max_input_tokens": "最大输入令牌数",
|
||||
"max_output_tokens": "最大输出令牌数"
|
||||
}
|
||||
},
|
||||
"groq": {
|
||||
"title": "配置 Groq",
|
||||
"data": {
|
||||
"groq_api_key": "Groq API 密钥",
|
||||
"groq_model": "Groq 模型",
|
||||
"groq_temperature": "温度",
|
||||
"max_input_tokens": "最大输入令牌数",
|
||||
"max_output_tokens": "最大输出令牌数"
|
||||
}
|
||||
},
|
||||
"localai": {
|
||||
"title": "配置 LocalAI",
|
||||
"data": {
|
||||
"localai_ip": "LocalAI IP 地址",
|
||||
"localai_port": "LocalAI 端口",
|
||||
"localai_https": "为 LocalAI 使用 HTTPS",
|
||||
"localai_model": "LocalAI 模型",
|
||||
"localai_temperature": "温度",
|
||||
"max_input_tokens": "最大输入令牌数",
|
||||
"max_output_tokens": "最大输出令牌数"
|
||||
}
|
||||
},
|
||||
"ollama": {
|
||||
"title": "配置 Ollama",
|
||||
"data": {
|
||||
"ollama_ip": "Ollama IP 地址",
|
||||
"ollama_port": "Ollama 端口",
|
||||
"ollama_https": "为 Ollama 使用 HTTPS",
|
||||
"ollama_model": "Ollama 模型",
|
||||
"ollama_temperature": "温度",
|
||||
"ollama_disable_think": "禁用思考模式 (Ollama)",
|
||||
"max_input_tokens": "最大输入令牌数",
|
||||
"max_output_tokens": "最大输出令牌数"
|
||||
}
|
||||
},
|
||||
"custom_openai": {
|
||||
"title": "配置自定义 OpenAI",
|
||||
"data": {
|
||||
"custom_openai_endpoint": "自定义 OpenAI 端点",
|
||||
"custom_openai_api_key": "自定义 OpenAI API 密钥 (可选)",
|
||||
"custom_openai_model": "自定义 OpenAI 模型",
|
||||
"custom_openai_temperature": "温度",
|
||||
"max_input_tokens": "最大输入令牌数",
|
||||
"max_output_tokens": "最大输出令牌数"
|
||||
}
|
||||
},
|
||||
"mistral": {
|
||||
"title": "配置 Mistral AI",
|
||||
"data": {
|
||||
"mistral_api_key": "Mistral API 密钥",
|
||||
"mistral_model": "Mistral 模型",
|
||||
"mistral_temperature": "温度",
|
||||
"max_input_tokens": "最大输入令牌数",
|
||||
"max_output_tokens": "最大输出令牌数"
|
||||
}
|
||||
},
|
||||
"perplexity": {
|
||||
"title": "配置 Perplexity AI",
|
||||
"data": {
|
||||
"perplexity_api_key": "Perplexity API 密钥",
|
||||
"perplexity_model": "Perplexity 模型",
|
||||
"perplexity_temperature": "温度",
|
||||
"max_input_tokens": "最大输入令牌数",
|
||||
"max_output_tokens": "最大输出令牌数"
|
||||
}
|
||||
},
|
||||
"openrouter": {
|
||||
"title": "配置 OpenRouter",
|
||||
"data": {
|
||||
"openrouter_api_key": "OpenRouter API 密钥",
|
||||
"openrouter_model": "OpenRouter 模型",
|
||||
"openrouter_reasoning_max_tokens": "OpenRouter 推理最大令牌数",
|
||||
"openrouter_temperature": "温度",
|
||||
"max_input_tokens": "最大输入令牌数",
|
||||
"max_output_tokens": "最大输出令牌数"
|
||||
}
|
||||
},
|
||||
"openai_azure": {
|
||||
"title": "配置 OpenAI Azure",
|
||||
"data": {
|
||||
"openai_azure_api_key": "Azure OpenAI API 密钥",
|
||||
"openai_azure_deployment_id": "Azure 部署 ID",
|
||||
"openai_azure_endpoint": "Azure 端点",
|
||||
"openai_azure_api_version": "Azure API 版本",
|
||||
"openai_azure_temperature": "温度",
|
||||
"max_input_tokens": "最大输入令牌数",
|
||||
"max_output_tokens": "最大输出令牌数"
|
||||
}
|
||||
}
|
||||
},
|
||||
"error": {
|
||||
"api_error": "API 验证失败:{error_message}",
|
||||
"cannot_connect": "连接失败。请检查您的设置和网络。",
|
||||
"unknown": "发生未知错误。请检查日志以获取详细信息。"
|
||||
},
|
||||
"abort": {
|
||||
"already_configured": "此 AI 提供商已配置。您可以从集成页面编辑它。"
|
||||
}
|
||||
},
|
||||
"options": {
|
||||
"step": {
|
||||
"init": {
|
||||
"title": "AI 自动化建议器选项",
|
||||
"data": {
|
||||
"openai_api_key": "OpenAI API 密钥",
|
||||
"openai_model": "OpenAI 模型",
|
||||
"openai_temperature": "温度 (OpenAI)",
|
||||
"anthropic_api_key": "Anthropic API 密钥",
|
||||
"anthropic_model": "Anthropic 模型",
|
||||
"anthropic_temperature": "温度 (Anthropic)",
|
||||
"google_api_key": "Google API 密钥",
|
||||
"google_model": "Google 模型",
|
||||
"google_temperature": "温度 (Google)",
|
||||
"groq_api_key": "Groq API 密钥",
|
||||
"groq_model": "Groq 模型",
|
||||
"groq_temperature": "温度 (Groq)",
|
||||
"localai_ip": "LocalAI IP 地址",
|
||||
"localai_port": "LocalAI 端口",
|
||||
"localai_https": "为 LocalAI 使用 HTTPS",
|
||||
"localai_model": "LocalAI 模型",
|
||||
"localai_temperature": "温度 (LocalAI)",
|
||||
"ollama_ip": "Ollama IP 地址",
|
||||
"ollama_port": "Ollama 端口",
|
||||
"ollama_https": "为 Ollama 使用 HTTPS",
|
||||
"ollama_model": "Ollama 模型",
|
||||
"ollama_temperature": "温度 (Ollama)",
|
||||
"ollama_disable_think": "禁用思考模式 (Ollama)",
|
||||
"custom_openai_endpoint": "自定义 OpenAI 端点",
|
||||
"custom_openai_api_key": "自定义 OpenAI API 密钥",
|
||||
"custom_openai_model": "自定义 OpenAI 模型",
|
||||
"custom_openai_temperature": "温度 (自定义 OpenAI)",
|
||||
"mistral_api_key": "Mistral API 密钥",
|
||||
"mistral_model": "Mistral 模型",
|
||||
"mistral_temperature": "温度 (Mistral AI)",
|
||||
"perplexity_api_key": "Perplexity API 密钥",
|
||||
"perplexity_model": "Perplexity 模型",
|
||||
"perplexity_temperature": "温度 (Perplexity AI)",
|
||||
"openrouter_api_key": "OpenRouter API 密钥",
|
||||
"openrouter_model": "OpenRouter 模型",
|
||||
"openrouter_reasoning_max_tokens": "OpenRouter 推理最大令牌数",
|
||||
"openrouter_temperature": "温度 (OpenRouter)",
|
||||
"openai_azure_api_key": "Azure OpenAI API 密钥",
|
||||
"openai_azure_deployment_id": "Azure 部署 ID",
|
||||
"openai_azure_endpoint": "Azure 端点",
|
||||
"openai_azure_api_version": "Azure API 版本",
|
||||
"openai_azure_temperature": "温度 (OpenAI Azure)",
|
||||
"max_input_tokens": "最大输入令牌数",
|
||||
"max_output_tokens": "最大输出令牌数"
|
||||
},
|
||||
"description": "调整您的 AI 提供商的设置。将使用与您配置的提供商相关的字段。通用设置(如令牌限制)将根据每个提供商进行配置。"
|
||||
}
|
||||
},
|
||||
"error": {
|
||||
"invalid_input": "一个或多个值无效。"
|
||||
}
|
||||
},
|
||||
"services": {
|
||||
"generate_suggestions": {
|
||||
"name": "生成建议",
|
||||
"description": "手动触发 AI 自动化建议",
|
||||
"description": "手动触发 AI 自动化建议。",
|
||||
"fields": {
|
||||
"provider_config": {
|
||||
"name": "提供商配置",
|
||||
@@ -25,8 +220,12 @@
|
||||
"description": "要考虑的实体最大数量(随机选择)。"
|
||||
},
|
||||
"automation_read_yaml": {
|
||||
"name": "Read 'automations.yaml' file",
|
||||
"description": "Reads and appends the yaml code of the automations found in the automations.yaml file. This action will use a lot of input tokens, use it with care and with models with a large input window (e.g. Gemini)."
|
||||
"name": "读取 'automations.yaml' 文件",
|
||||
"description": "读取并附加 automations.yaml 文件中找到的自动化的 yaml 代码。此操作将使用大量输入令牌,请谨慎使用并配合具有较大输入窗口的模型(例如 Gemini)。"
|
||||
},
|
||||
"automation_limit": {
|
||||
"name": "自动化数量上限",
|
||||
"description": "要分析的自动化最大数量(默认:100)。"
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user