Docker compatibility:
- Discovery endpoints now read from OLLAMA_URL, LLAMACPP_URL, LMSTUDIO_URL env vars
- docker-compose.yml sets backends to host.docker.internal for container access
- justfile updated with --host 0.0.0.0 for llama-server
Vision support:
- OpenAI adapter now converts images to content parts array format
- Enables vision models with llama.cpp and LM Studio
Bumps version to 0.7.1
- Add AboutTab.svelte with app branding, version display, update checking
- Show update status and download link when new version available
- Include links to GitHub repo and issue tracker
- Display tech stack badges and license info
- Remove About section from GeneralTab (now separate tab)
Also improves development configuration:
- justfile now reads PORT, DEV_PORT, LLAMA_PORT, OLLAMA_PORT from .env
- docker-compose.dev.yml uses env var substitution
- Add dev-build and dev-rebuild recipes for Docker
- Update .env.example with all configurable variables
- Backend runs on port 9090 (matches vite proxy default)
- Adds health check command for all services
- Includes llama.cpp server commands
- Adds test and build commands