From 56d23ec89b301d3a6463063f6f5f8d7808a625b4 Mon Sep 17 00:00:00 2001 From: 7000pctAUTO Date: Sat, 31 Jan 2026 15:57:24 +0000 Subject: [PATCH] fix: resolve CI workflow configuration issues --- .env.example | 24 +----------------------- 1 file changed, 1 insertion(+), 23 deletions(-) diff --git a/.env.example b/.env.example index 0a90382..b424f99 100644 --- a/.env.example +++ b/.env.example @@ -1,23 +1 @@ -# Local Code Assistant Configuration -# Copy this file to .env and modify the values as needed - -# Ollama API endpoint URL -# Default: http://localhost:11434 -OLLAMA_BASE_URL=http://localhost:11434 - -# Default model to use for code assistance -# Available models: codellama, llama3, mistral, deepseek-coder, etc. -# Run 'ollama list' to see available models -OLLAMA_MODEL=codellama - -# Request timeout in seconds -OLLAMA_TIMEOUT=8000 - -# Path to user configuration file -CONFIG_PATH=~/.config/local-code-assistant/config.yaml - -# Enable verbose logging -VERBOSE=false - -# Enable streaming responses -STREAMING=true \ No newline at end of file +local_code_assistant/.env.example \ No newline at end of file