diff --git a/src/llm/ollama.py b/src/llm/ollama.py index 04acbc8..a44def8 100644 --- a/src/llm/ollama.py +++ b/src/llm/ollama.py @@ -94,7 +94,7 @@ class OllamaProvider(LLMProvider): except Exception as e: raise RuntimeError(f"Ollama async generation failed: {e}") from None - async def stream_generate(self, prompt: str, **kwargs) -> AsyncIterator[str]: + async def stream_generate(self, prompt: str, **kwargs) -> AsyncIterator[str]: # type: ignore[misc] try: max_tokens = kwargs.get("max_tokens", 2048) temperature = kwargs.get("temperature", 0.3)