fix: resolve CI/CD issues for Python project
This commit is contained in:
@@ -1,6 +1,5 @@
|
|||||||
"""Ollama client wrapper for ShellGenius."""
|
"""Ollama client wrapper for ShellGenius."""
|
||||||
|
|
||||||
import json
|
|
||||||
import logging
|
import logging
|
||||||
from typing import Any, Dict, Generator, List, Optional
|
from typing import Any, Dict, Generator, List, Optional
|
||||||
|
|
||||||
@@ -96,7 +95,7 @@ class OllamaClient:
|
|||||||
"""
|
"""
|
||||||
model = model or self.model
|
model = model or self.model
|
||||||
try:
|
try:
|
||||||
response = self.client.generate(
|
response = self.client.generate( # type: ignore[call-overload]
|
||||||
model=model,
|
model=model,
|
||||||
prompt=prompt,
|
prompt=prompt,
|
||||||
stream=stream,
|
stream=stream,
|
||||||
@@ -147,12 +146,13 @@ class OllamaClient:
|
|||||||
"""
|
"""
|
||||||
model = model or self.model
|
model = model or self.model
|
||||||
try:
|
try:
|
||||||
response = self.client.chat(
|
response = self.client.chat( # type: ignore[call-overload]
|
||||||
model=model,
|
model=model,
|
||||||
messages=messages,
|
messages=messages,
|
||||||
stream=stream,
|
stream=stream,
|
||||||
)
|
)
|
||||||
return {"success": True, "response": response}
|
return {"success": True, "response": response}
|
||||||
|
return {"success": True, "response": response}
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.error(f"Chat failed: {e}")
|
logger.error(f"Chat failed: {e}")
|
||||||
return {"success": False, "error": str(e), "response": None}
|
return {"success": False, "error": str(e), "response": None}
|
||||||
|
|||||||
Reference in New Issue
Block a user