diff --git a/.env.example b/.env.example index bcbbe25..b7d03a5 100644 --- a/.env.example +++ b/.env.example @@ -5,7 +5,7 @@ # All variables have sensible defaults - only set what you need to change. # ----- Backend ----- -# Server port (default: 8080, but 9090 recommended for local dev) +# Server port (default: 9090 for local dev, matches vite proxy) PORT=9090 # SQLite database path (relative to backend working directory) @@ -26,3 +26,15 @@ BACKEND_URL=http://localhost:9090 # Development server port DEV_PORT=7842 + +# ----- llama.cpp ----- +# llama.cpp server port (used by `just llama-server`) +LLAMA_PORT=8081 + +# ----- Additional Ports (for health checks) ----- +# Ollama port (extracted from OLLAMA_URL for health checks) +OLLAMA_PORT=11434 + +# ----- Models ----- +# Directory for GGUF model files +VESSEL_MODELS_DIR=~/.vessel/models diff --git a/backend/cmd/server/main.go b/backend/cmd/server/main.go index 8dc15fd..9c0f94a 100644 --- a/backend/cmd/server/main.go +++ b/backend/cmd/server/main.go @@ -18,7 +18,7 @@ import ( ) // Version is set at build time via -ldflags, or defaults to dev -var Version = "0.6.0" +var Version = "0.6.1" func getEnvOrDefault(key, defaultValue string) string { if value := os.Getenv(key); value != "" { diff --git a/docker-compose.dev.yml b/docker-compose.dev.yml index fe50cd2..0ee65c3 100644 --- a/docker-compose.dev.yml +++ b/docker-compose.dev.yml @@ -1,6 +1,8 @@ name: vessel-dev # Development docker-compose - uses host network for direct Ollama access +# Reads configuration from .env file + services: frontend: build: @@ -12,8 +14,8 @@ services: - ./frontend:/app - /app/node_modules environment: - - OLLAMA_API_URL=http://localhost:11434 - - BACKEND_URL=http://localhost:9090 + - OLLAMA_API_URL=${OLLAMA_API_URL:-http://localhost:11434} + - BACKEND_URL=${BACKEND_URL:-http://localhost:9090} depends_on: - backend @@ -26,4 +28,4 @@ services: - ./backend/data:/app/data environment: - GIN_MODE=release - command: ["./server", "-port", "9090", "-db", "/app/data/vessel.db", "-ollama-url", "http://localhost:11434"] + command: ["./server", "-port", "${PORT:-9090}", "-db", "${DB_PATH:-/app/data/vessel.db}", "-ollama-url", "${OLLAMA_URL:-http://localhost:11434}"] diff --git a/frontend/package.json b/frontend/package.json index ee73618..518ac18 100644 --- a/frontend/package.json +++ b/frontend/package.json @@ -1,6 +1,6 @@ { "name": "vessel", - "version": "0.6.0", + "version": "0.6.1", "private": true, "type": "module", "scripts": { diff --git a/frontend/src/lib/components/settings/AboutTab.svelte b/frontend/src/lib/components/settings/AboutTab.svelte new file mode 100644 index 0000000..641c100 --- /dev/null +++ b/frontend/src/lib/components/settings/AboutTab.svelte @@ -0,0 +1,194 @@ + + +
+ A modern interface for local AI with chat, tools, and memory management. +
+ {#if versionState.current} +Update Available
++ Version {versionState.latest} is available. You're currently on v{versionState.current}. +
++ Last checked: {formatLastChecked(versionState.lastChecked)} +
+ {/if} +Built With
+License
++ Released under the {LICENSE} license +
+- A modern interface for local AI with chat, tools, and memory management. -
-