Docker compatibility: - Discovery endpoints now read from OLLAMA_URL, LLAMACPP_URL, LMSTUDIO_URL env vars - docker-compose.yml sets backends to host.docker.internal for container access - justfile updated with --host 0.0.0.0 for llama-server Vision support: - OpenAI adapter now converts images to content parts array format - Enables vision models with llama.cpp and LM Studio Bumps version to 0.7.1
159 lines
4.5 KiB
Go
159 lines
4.5 KiB
Go
package main
|
|
|
|
import (
|
|
"context"
|
|
"flag"
|
|
"log"
|
|
"net/http"
|
|
"os"
|
|
"os/signal"
|
|
"syscall"
|
|
"time"
|
|
|
|
"github.com/gin-contrib/cors"
|
|
"github.com/gin-gonic/gin"
|
|
|
|
"vessel-backend/internal/api"
|
|
"vessel-backend/internal/backends"
|
|
"vessel-backend/internal/backends/ollama"
|
|
"vessel-backend/internal/backends/openai"
|
|
"vessel-backend/internal/database"
|
|
)
|
|
|
|
// Version is set at build time via -ldflags, or defaults to dev
|
|
var Version = "0.7.1"
|
|
|
|
func getEnvOrDefault(key, defaultValue string) string {
|
|
if value := os.Getenv(key); value != "" {
|
|
return value
|
|
}
|
|
return defaultValue
|
|
}
|
|
|
|
func main() {
|
|
var (
|
|
port = flag.String("port", getEnvOrDefault("PORT", "8080"), "Server port")
|
|
dbPath = flag.String("db", getEnvOrDefault("DB_PATH", "./data/vessel.db"), "Database file path")
|
|
ollamaURL = flag.String("ollama-url", getEnvOrDefault("OLLAMA_URL", "http://localhost:11434"), "Ollama API URL")
|
|
llamacppURL = flag.String("llamacpp-url", getEnvOrDefault("LLAMACPP_URL", "http://localhost:8081"), "llama.cpp server URL")
|
|
lmstudioURL = flag.String("lmstudio-url", getEnvOrDefault("LMSTUDIO_URL", "http://localhost:1234"), "LM Studio server URL")
|
|
)
|
|
flag.Parse()
|
|
|
|
// Initialize database
|
|
db, err := database.OpenDatabase(*dbPath)
|
|
if err != nil {
|
|
log.Fatalf("Failed to open database: %v", err)
|
|
}
|
|
defer db.Close()
|
|
|
|
// Run migrations
|
|
if err := database.RunMigrations(db); err != nil {
|
|
log.Fatalf("Failed to run migrations: %v", err)
|
|
}
|
|
|
|
// Initialize backend registry
|
|
registry := backends.NewRegistry()
|
|
|
|
// Register Ollama backend
|
|
ollamaAdapter, err := ollama.NewAdapter(backends.BackendConfig{
|
|
Type: backends.BackendTypeOllama,
|
|
BaseURL: *ollamaURL,
|
|
})
|
|
if err != nil {
|
|
log.Printf("Warning: Failed to create Ollama adapter: %v", err)
|
|
} else {
|
|
if err := registry.Register(ollamaAdapter); err != nil {
|
|
log.Printf("Warning: Failed to register Ollama backend: %v", err)
|
|
}
|
|
}
|
|
|
|
// Register llama.cpp backend (if URL is configured)
|
|
if *llamacppURL != "" {
|
|
llamacppAdapter, err := openai.NewAdapter(backends.BackendConfig{
|
|
Type: backends.BackendTypeLlamaCpp,
|
|
BaseURL: *llamacppURL,
|
|
})
|
|
if err != nil {
|
|
log.Printf("Warning: Failed to create llama.cpp adapter: %v", err)
|
|
} else {
|
|
if err := registry.Register(llamacppAdapter); err != nil {
|
|
log.Printf("Warning: Failed to register llama.cpp backend: %v", err)
|
|
}
|
|
}
|
|
}
|
|
|
|
// Register LM Studio backend (if URL is configured)
|
|
if *lmstudioURL != "" {
|
|
lmstudioAdapter, err := openai.NewAdapter(backends.BackendConfig{
|
|
Type: backends.BackendTypeLMStudio,
|
|
BaseURL: *lmstudioURL,
|
|
})
|
|
if err != nil {
|
|
log.Printf("Warning: Failed to create LM Studio adapter: %v", err)
|
|
} else {
|
|
if err := registry.Register(lmstudioAdapter); err != nil {
|
|
log.Printf("Warning: Failed to register LM Studio backend: %v", err)
|
|
}
|
|
}
|
|
}
|
|
|
|
// Setup Gin router
|
|
gin.SetMode(gin.ReleaseMode)
|
|
r := gin.New()
|
|
r.Use(gin.Logger())
|
|
r.Use(gin.Recovery())
|
|
|
|
// CORS configuration
|
|
r.Use(cors.New(cors.Config{
|
|
AllowOrigins: []string{"*"},
|
|
AllowMethods: []string{"GET", "POST", "PUT", "PATCH", "DELETE", "OPTIONS"},
|
|
AllowHeaders: []string{"Origin", "Content-Type", "Accept", "Authorization"},
|
|
ExposeHeaders: []string{"Content-Length"},
|
|
AllowCredentials: true,
|
|
MaxAge: 12 * time.Hour,
|
|
}))
|
|
|
|
// Register routes
|
|
api.SetupRoutes(r, db, *ollamaURL, Version, registry)
|
|
|
|
// Create server
|
|
srv := &http.Server{
|
|
Addr: ":" + *port,
|
|
Handler: r,
|
|
}
|
|
|
|
// Initialize fetcher and log the method being used
|
|
fetcher := api.GetFetcher()
|
|
log.Printf("URL fetcher method: %s (headless Chrome: %v)", fetcher.Method(), fetcher.HasChrome())
|
|
|
|
// Graceful shutdown handling
|
|
go func() {
|
|
log.Printf("Server starting on port %s", *port)
|
|
log.Printf("Database: %s", *dbPath)
|
|
log.Printf("Backends configured:")
|
|
log.Printf(" - Ollama: %s", *ollamaURL)
|
|
log.Printf(" - llama.cpp: %s", *llamacppURL)
|
|
log.Printf(" - LM Studio: %s", *lmstudioURL)
|
|
log.Printf("Active backend: %s", registry.ActiveType().String())
|
|
if err := srv.ListenAndServe(); err != nil && err != http.ErrServerClosed {
|
|
log.Fatalf("Failed to start server: %v", err)
|
|
}
|
|
}()
|
|
|
|
// Wait for interrupt signal
|
|
quit := make(chan os.Signal, 1)
|
|
signal.Notify(quit, syscall.SIGINT, syscall.SIGTERM)
|
|
<-quit
|
|
log.Println("Shutting down server...")
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), 5*time.Second)
|
|
defer cancel()
|
|
|
|
if err := srv.Shutdown(ctx); err != nil {
|
|
log.Fatalf("Server forced to shutdown: %v", err)
|
|
}
|
|
|
|
log.Println("Server exited")
|
|
}
|