mirror of
https://github.com/memohai/Memoh.git
synced 2026-04-25 07:00:48 +09:00
8d5c38f0e5
* refactor: unify providers and models tables
- Rename `llm_providers` → `providers`, `llm_provider_oauth_tokens` → `provider_oauth_tokens`
- Remove `tts_providers` and `tts_models` tables; speech models now live in the unified `models` table with `type = 'speech'`
- Replace top-level `api_key`/`base_url` columns with a JSONB `config` field on `providers`
- Rename `llm_provider_id` → `provider_id` across all references
- Add `edge-speech` client type and `conf/providers/edge.yaml` default provider
- Create new read-only speech endpoints (`/speech-providers`, `/speech-models`) backed by filtered views of the unified tables
- Remove old TTS CRUD handlers; simplify speech page to read-only + test
- Update registry loader to skip malformed YAML files instead of failing entirely
- Fix YAML quoting for model names containing colons in openrouter.yaml
- Regenerate sqlc, swagger, and TypeScript SDK
* fix: exclude speech providers from providers list endpoint
ListProviders now filters out client_type matching '%-speech' so Edge
and future speech providers no longer appear on the Providers page.
ListSpeechProviders uses the same pattern match instead of hard-coding
'edge-speech'.
* fix: use explicit client_type list instead of LIKE pattern
Replace '%-speech' pattern with explicit IN ('edge-speech') for both
ListProviders (exclusion) and ListSpeechProviders (inclusion). New
speech client types must be added to both queries.
* fix: use EXECUTE for dynamic SQL in migrations referencing old schema
PL/pgSQL pre-validates column/table references in static SQL statements
inside DO blocks before evaluating IF/RETURN guards. This caused
migrations 0010-0061 to fail on fresh databases where the canonical
schema uses `providers`/`provider_id` instead of `llm_providers`/
`llm_provider_id`.
Wrap all SQL that references potentially non-existent old schema objects
(llm_providers, llm_provider_id, tts_providers, tts_models, etc.) in
EXECUTE strings so they are only parsed at runtime when actually reached.
* fix: revert canonical schema to use llm_providers for migration compatibility
The CI migrations workflow (up → down → up) failed because 0061 down
renames `providers` back to `llm_providers`, but 0001 down only dropped
`providers` — leaving `llm_providers` as a remnant. On the second
migrate up, 0010 found the stale `llm_providers` and tried to reference
`models.llm_provider_id` which no longer existed.
Revert 0001 canonical schema to use original names (llm_providers,
tts_providers, tts_models) so incremental migrations work naturally and
0061 handles the final rename. Remove EXECUTE wrappers and unnecessary
guards from migrations that now always operate on llm_providers.
* fix: icons
* fix: sync canonical schema with 0061 migration to fix sqlc column mismatch
0001_init.up.sql still used old names (llm_providers, llm_provider_id)
and included dropped tts_providers/tts_models tables. sqlc could not
parse the PL/pgSQL EXECUTE in migration 0061, so generated code retained
stale columns (input_modalities, supports_reasoning) causing runtime
"column does not exist" errors when adding models.
- Update 0001_init.up.sql to current schema (providers, provider_id,
no tts tables, add provider_oauth_tokens)
- Use ALTER TABLE IF EXISTS in 0010/0041/0042 for backward compat
- Regenerate sqlc
* fix: guard all legacy migrations against fresh schema for CI compat
On fresh databases, 0001_init.up.sql creates providers/provider_id
(not llm_providers/llm_provider_id). Migrations 0013, 0041, 0046, 0047
referenced the old names without guards, causing CI migration failures.
- 0013: check llm_provider_id column exists before adding old constraint
- 0041: check llm_providers table exists before backfill/constraint DDL
- 0046: wrap CREATE TABLE in DO block with llm_providers existence check
- 0047: use ALTER TABLE IF EXISTS + DO block guard
250 lines
7.3 KiB
Go
250 lines
7.3 KiB
Go
package models
|
|
|
|
import (
|
|
"context"
|
|
"encoding/base64"
|
|
"encoding/json"
|
|
"errors"
|
|
"fmt"
|
|
"net/http"
|
|
"strings"
|
|
"time"
|
|
|
|
anthropicmessages "github.com/memohai/twilight-ai/provider/anthropic/messages"
|
|
googlegenerative "github.com/memohai/twilight-ai/provider/google/generativeai"
|
|
openaicodex "github.com/memohai/twilight-ai/provider/openai/codex"
|
|
openaicompletions "github.com/memohai/twilight-ai/provider/openai/completions"
|
|
openairesponses "github.com/memohai/twilight-ai/provider/openai/responses"
|
|
sdk "github.com/memohai/twilight-ai/sdk"
|
|
|
|
"github.com/memohai/memoh/internal/db"
|
|
"github.com/memohai/memoh/internal/db/sqlc"
|
|
)
|
|
|
|
const probeTimeout = 15 * time.Second
|
|
|
|
// Test probes a model's provider endpoint using the Twilight AI SDK
|
|
// to verify connectivity, authentication, and model availability.
|
|
func (s *Service) Test(ctx context.Context, id string) (TestResponse, error) {
|
|
modelID, err := db.ParseUUID(id)
|
|
if err != nil {
|
|
return TestResponse{}, fmt.Errorf("invalid model id: %w", err)
|
|
}
|
|
|
|
model, err := s.queries.GetModelByID(ctx, modelID)
|
|
if err != nil {
|
|
return TestResponse{}, fmt.Errorf("get model: %w", err)
|
|
}
|
|
|
|
provider, err := s.queries.GetProviderByID(ctx, model.ProviderID)
|
|
if err != nil {
|
|
return TestResponse{}, fmt.Errorf("get provider: %w", err)
|
|
}
|
|
|
|
baseURL := strings.TrimRight(providerConfigString(provider.Config, "base_url"), "/")
|
|
clientType := ClientType(provider.ClientType)
|
|
creds, err := s.resolveModelCredentials(ctx, provider)
|
|
if err != nil {
|
|
return TestResponse{}, err
|
|
}
|
|
|
|
if model.Type == string(ModelTypeEmbedding) {
|
|
return s.testEmbeddingModel(ctx, baseURL, creds.APIKey, model.ModelID, nil)
|
|
}
|
|
|
|
sdkProvider := NewSDKProvider(baseURL, creds.APIKey, creds.CodexAccountID, clientType, probeTimeout, nil)
|
|
|
|
start := time.Now()
|
|
|
|
providerResult := sdkProvider.Test(ctx)
|
|
switch providerResult.Status {
|
|
case sdk.ProviderStatusUnreachable:
|
|
return TestResponse{
|
|
Status: TestStatusError,
|
|
Reachable: false,
|
|
LatencyMs: time.Since(start).Milliseconds(),
|
|
Message: providerResult.Message,
|
|
}, nil
|
|
case sdk.ProviderStatusUnhealthy:
|
|
return TestResponse{
|
|
Status: TestStatusAuthError,
|
|
Reachable: true,
|
|
LatencyMs: time.Since(start).Milliseconds(),
|
|
Message: providerResult.Message,
|
|
}, nil
|
|
}
|
|
|
|
modelResult, err := sdkProvider.TestModel(ctx, model.ModelID)
|
|
latency := time.Since(start).Milliseconds()
|
|
|
|
if err != nil {
|
|
return TestResponse{
|
|
Status: TestStatusError,
|
|
Reachable: true,
|
|
LatencyMs: latency,
|
|
Message: err.Error(),
|
|
}, nil
|
|
}
|
|
|
|
if !modelResult.Supported {
|
|
return TestResponse{
|
|
Status: TestStatusModelNotSupported,
|
|
Reachable: true,
|
|
LatencyMs: latency,
|
|
Message: modelResult.Message,
|
|
}, nil
|
|
}
|
|
|
|
return TestResponse{
|
|
Status: TestStatusOK,
|
|
Reachable: true,
|
|
LatencyMs: latency,
|
|
Message: modelResult.Message,
|
|
}, nil
|
|
}
|
|
|
|
// testEmbeddingModel probes an embedding model by performing a minimal
|
|
// embedding request via the Twilight SDK, verifying that the model is
|
|
// reachable and functional rather than merely checking HTTP connectivity.
|
|
func (*Service) testEmbeddingModel(ctx context.Context, baseURL, apiKey, modelID string, httpClient *http.Client) (TestResponse, error) {
|
|
ctx, cancel := context.WithTimeout(ctx, probeTimeout)
|
|
defer cancel()
|
|
|
|
model := NewSDKEmbeddingModel(string(ClientTypeOpenAICompletions), baseURL, apiKey, modelID, probeTimeout, httpClient)
|
|
client := sdk.NewClient()
|
|
|
|
start := time.Now()
|
|
_, err := client.Embed(ctx, "hello", sdk.WithEmbeddingModel(model))
|
|
latency := time.Since(start).Milliseconds()
|
|
|
|
if err != nil {
|
|
return TestResponse{
|
|
Status: TestStatusError,
|
|
Reachable: false,
|
|
LatencyMs: latency,
|
|
Message: err.Error(),
|
|
}, nil
|
|
}
|
|
|
|
return TestResponse{
|
|
Status: TestStatusOK,
|
|
Reachable: true,
|
|
LatencyMs: latency,
|
|
Message: "embedding model is operational",
|
|
}, nil
|
|
}
|
|
|
|
// NewSDKProvider creates a Twilight AI SDK Provider for the given client type.
|
|
// It is exported so that other packages (e.g. providers) can reuse it for testing.
|
|
func NewSDKProvider(baseURL, apiKey, codexAccountID string, clientType ClientType, timeout time.Duration, httpClient *http.Client) sdk.Provider {
|
|
if httpClient == nil {
|
|
httpClient = &http.Client{Timeout: timeout}
|
|
}
|
|
|
|
switch clientType {
|
|
case ClientTypeOpenAIResponses:
|
|
opts := []openairesponses.Option{
|
|
openairesponses.WithAPIKey(apiKey),
|
|
openairesponses.WithHTTPClient(httpClient),
|
|
}
|
|
if baseURL != "" {
|
|
opts = append(opts, openairesponses.WithBaseURL(baseURL))
|
|
}
|
|
return openairesponses.New(opts...)
|
|
|
|
case ClientTypeOpenAICodex:
|
|
opts := []openaicodex.Option{
|
|
openaicodex.WithAccessToken(apiKey),
|
|
openaicodex.WithHTTPClient(httpClient),
|
|
}
|
|
if codexAccountID != "" {
|
|
opts = append(opts, openaicodex.WithAccountID(codexAccountID))
|
|
}
|
|
return openaicodex.New(opts...)
|
|
|
|
case ClientTypeAnthropicMessages:
|
|
opts := []anthropicmessages.Option{
|
|
anthropicmessages.WithAPIKey(apiKey),
|
|
anthropicmessages.WithHTTPClient(httpClient),
|
|
}
|
|
if baseURL != "" {
|
|
opts = append(opts, anthropicmessages.WithBaseURL(baseURL))
|
|
}
|
|
return anthropicmessages.New(opts...)
|
|
|
|
case ClientTypeGoogleGenerativeAI:
|
|
opts := []googlegenerative.Option{
|
|
googlegenerative.WithAPIKey(apiKey),
|
|
googlegenerative.WithHTTPClient(httpClient),
|
|
}
|
|
if baseURL != "" {
|
|
opts = append(opts, googlegenerative.WithBaseURL(baseURL))
|
|
}
|
|
return googlegenerative.New(opts...)
|
|
|
|
default:
|
|
opts := []openaicompletions.Option{
|
|
openaicompletions.WithAPIKey(apiKey),
|
|
openaicompletions.WithHTTPClient(httpClient),
|
|
}
|
|
if baseURL != "" {
|
|
opts = append(opts, openaicompletions.WithBaseURL(baseURL))
|
|
}
|
|
return openaicompletions.New(opts...)
|
|
}
|
|
}
|
|
|
|
type modelCredentials struct {
|
|
APIKey string //nolint:gosec // runtime credential material used to construct SDK providers
|
|
CodexAccountID string
|
|
}
|
|
|
|
func (s *Service) resolveModelCredentials(ctx context.Context, provider sqlc.Provider) (modelCredentials, error) {
|
|
apiKey := providerConfigString(provider.Config, "api_key")
|
|
|
|
if ClientType(provider.ClientType) != ClientTypeOpenAICodex {
|
|
return modelCredentials{APIKey: apiKey}, nil
|
|
}
|
|
|
|
tokenRow, err := s.queries.GetProviderOAuthTokenByProvider(ctx, provider.ID)
|
|
if err != nil {
|
|
return modelCredentials{}, err
|
|
}
|
|
accessToken := strings.TrimSpace(tokenRow.AccessToken)
|
|
if accessToken == "" {
|
|
return modelCredentials{}, errors.New("oauth token is missing access token")
|
|
}
|
|
accountID, err := codexAccountIDFromToken(accessToken)
|
|
if err != nil {
|
|
return modelCredentials{}, err
|
|
}
|
|
return modelCredentials{
|
|
APIKey: accessToken,
|
|
CodexAccountID: accountID,
|
|
}, nil
|
|
}
|
|
|
|
func codexAccountIDFromToken(token string) (string, error) {
|
|
parts := strings.Split(token, ".")
|
|
if len(parts) != 3 {
|
|
return "", errors.New("invalid oauth access token")
|
|
}
|
|
payload, err := base64.RawURLEncoding.DecodeString(parts[1])
|
|
if err != nil {
|
|
return "", fmt.Errorf("decode oauth token payload: %w", err)
|
|
}
|
|
var claims struct {
|
|
OpenAIAuth struct {
|
|
ChatGPTAccountID string `json:"chatgpt_account_id"`
|
|
} `json:"https://api.openai.com/auth"`
|
|
}
|
|
if err := json.Unmarshal(payload, &claims); err != nil {
|
|
return "", fmt.Errorf("parse oauth token payload: %w", err)
|
|
}
|
|
accountID := strings.TrimSpace(claims.OpenAIAuth.ChatGPTAccountID)
|
|
if accountID == "" {
|
|
return "", errors.New("oauth access token missing chatgpt_account_id")
|
|
}
|
|
return accountID, nil
|
|
}
|