Set default values for LLM performance (#20606)

This commit is contained in:
Nicolas Mowen 2025-10-21 16:10:00 -06:00 committed by GitHub
parent 21ff257705
commit 007371019a
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -1,7 +1,7 @@
"""Ollama Provider for Frigate AI.""" """Ollama Provider for Frigate AI."""
import logging import logging
from typing import Optional from typing import Any, Optional
from httpx import TimeoutException from httpx import TimeoutException
from ollama import Client as ApiClient from ollama import Client as ApiClient
@ -17,10 +17,24 @@ logger = logging.getLogger(__name__)
class OllamaClient(GenAIClient): class OllamaClient(GenAIClient):
"""Generative AI client for Frigate using Ollama.""" """Generative AI client for Frigate using Ollama."""
LOCAL_OPTIMIZED_OPTIONS = {
"options": {
"temperature": 0.5,
"repeat_penalty": 1.15,
"presence_penalty": 0.1,
},
}
provider: ApiClient provider: ApiClient
provider_options: dict[str, Any]
def _init_provider(self): def _init_provider(self):
"""Initialize the client.""" """Initialize the client."""
self.provider_options = {
**self.LOCAL_OPTIMIZED_OPTIONS,
**self.genai_config.provider_options,
}
try: try:
client = ApiClient(host=self.genai_config.base_url, timeout=self.timeout) client = ApiClient(host=self.genai_config.base_url, timeout=self.timeout)
# ensure the model is available locally # ensure the model is available locally
@ -48,7 +62,7 @@ class OllamaClient(GenAIClient):
self.genai_config.model, self.genai_config.model,
prompt, prompt,
images=images if images else None, images=images if images else None,
**self.genai_config.provider_options, **self.provider_options,
) )
return result["response"].strip() return result["response"].strip()
except (TimeoutException, ResponseError) as e: except (TimeoutException, ResponseError) as e: