mirror of
https://github.com/memohai/Memoh.git
synced 2026-04-27 07:16:19 +09:00
b88ca96064
* refactor: move client_type to provider, replace model fields with config JSONB - Move `client_type` from `models` to `llm_providers` table - Add `icon` field to `llm_providers` - Replace `dimensions`, `input_modalities`, `supports_reasoning` on `models` with a single `config` JSONB column containing `dimensions`, `compatibilities` (vision, tool-call, image-output, reasoning), and `context_window` - Auto-imported models default to vision + tool-call + reasoning - Update all backend consumers (agent, flow resolver, handlers, memory) - Regenerate sqlc, swagger, and TypeScript SDK - Update frontend forms, display, and i18n for new schema * ui: show provider icon avatar in sidebar and detail header, remove icon input * feat: add built-in provider registry with YAML definitions and enable toggle - Add `enable` column to llm_providers (default true, backward-compatible) - Create internal/registry package to load YAML provider/model definitions on startup and upsert into database (new providers disabled by default) - Add conf/providers/ with OpenAI, Anthropic, Google YAML definitions - Add RegistryConfig to TOML config (providers_dir, default conf/providers) - Model listing APIs and conversation flow now filter by enabled providers - Frontend: enable switch in provider form, green status dot in sidebar, enabled providers sorted to top * fix: make 0041 migration idempotent for fresh databases Guard data migration steps with column-existence checks so the migration succeeds on databases created from the updated init schema.
147 lines
3.7 KiB
Go
147 lines
3.7 KiB
Go
package models_test
|
|
|
|
import (
|
|
"testing"
|
|
|
|
"github.com/stretchr/testify/assert"
|
|
|
|
"github.com/memohai/memoh/internal/models"
|
|
)
|
|
|
|
func intPtr(v int) *int { return &v }
|
|
|
|
func TestModel_Validate(t *testing.T) {
|
|
tests := []struct {
|
|
name string
|
|
model models.Model
|
|
wantErr bool
|
|
}{
|
|
{
|
|
name: "valid chat model",
|
|
model: models.Model{
|
|
ModelID: "gpt-4",
|
|
Name: "GPT-4",
|
|
LlmProviderID: "11111111-1111-1111-1111-111111111111",
|
|
Type: models.ModelTypeChat,
|
|
},
|
|
wantErr: false,
|
|
},
|
|
{
|
|
name: "valid chat model with compatibilities",
|
|
model: models.Model{
|
|
ModelID: "gpt-4o",
|
|
Name: "GPT-4o",
|
|
LlmProviderID: "11111111-1111-1111-1111-111111111111",
|
|
Type: models.ModelTypeChat,
|
|
Config: models.ModelConfig{
|
|
Compatibilities: []string{"vision", "tool-call", "reasoning"},
|
|
},
|
|
},
|
|
wantErr: false,
|
|
},
|
|
{
|
|
name: "valid embedding model",
|
|
model: models.Model{
|
|
ModelID: "text-embedding-ada-002",
|
|
Name: "Ada Embeddings",
|
|
LlmProviderID: "11111111-1111-1111-1111-111111111111",
|
|
Type: models.ModelTypeEmbedding,
|
|
Config: models.ModelConfig{Dimensions: intPtr(1536)},
|
|
},
|
|
wantErr: false,
|
|
},
|
|
{
|
|
name: "missing model_id",
|
|
model: models.Model{
|
|
LlmProviderID: "11111111-1111-1111-1111-111111111111",
|
|
Type: models.ModelTypeChat,
|
|
},
|
|
wantErr: true,
|
|
},
|
|
{
|
|
name: "missing llm_provider_id",
|
|
model: models.Model{
|
|
ModelID: "gpt-4",
|
|
Type: models.ModelTypeChat,
|
|
},
|
|
wantErr: true,
|
|
},
|
|
{
|
|
name: "invalid llm_provider_id",
|
|
model: models.Model{
|
|
ModelID: "gpt-4",
|
|
LlmProviderID: "not-a-uuid",
|
|
Type: models.ModelTypeChat,
|
|
},
|
|
wantErr: true,
|
|
},
|
|
{
|
|
name: "invalid model type",
|
|
model: models.Model{
|
|
ModelID: "gpt-4",
|
|
LlmProviderID: "11111111-1111-1111-1111-111111111111",
|
|
Type: "invalid",
|
|
},
|
|
wantErr: true,
|
|
},
|
|
{
|
|
name: "embedding model missing dimensions",
|
|
model: models.Model{
|
|
ModelID: "text-embedding-ada-002",
|
|
LlmProviderID: "11111111-1111-1111-1111-111111111111",
|
|
Type: models.ModelTypeEmbedding,
|
|
},
|
|
wantErr: true,
|
|
},
|
|
{
|
|
name: "invalid compatibility",
|
|
model: models.Model{
|
|
ModelID: "gpt-4",
|
|
LlmProviderID: "11111111-1111-1111-1111-111111111111",
|
|
Type: models.ModelTypeChat,
|
|
Config: models.ModelConfig{
|
|
Compatibilities: []string{"vision", "smell"},
|
|
},
|
|
},
|
|
wantErr: true,
|
|
},
|
|
}
|
|
|
|
for _, tt := range tests {
|
|
t.Run(tt.name, func(t *testing.T) {
|
|
err := tt.model.Validate()
|
|
if tt.wantErr {
|
|
assert.Error(t, err)
|
|
} else {
|
|
assert.NoError(t, err)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestModel_HasCompatibility(t *testing.T) {
|
|
m := models.Model{
|
|
Config: models.ModelConfig{
|
|
Compatibilities: []string{"vision", "tool-call", "reasoning"},
|
|
},
|
|
}
|
|
assert.True(t, m.HasCompatibility("vision"))
|
|
assert.True(t, m.HasCompatibility("tool-call"))
|
|
assert.True(t, m.HasCompatibility("reasoning"))
|
|
assert.False(t, m.HasCompatibility("image-output"))
|
|
}
|
|
|
|
func TestModelTypes(t *testing.T) {
|
|
t.Run("ModelType constants", func(t *testing.T) {
|
|
assert.Equal(t, models.ModelTypeChat, models.ModelType("chat"))
|
|
assert.Equal(t, models.ModelTypeEmbedding, models.ModelType("embedding"))
|
|
})
|
|
|
|
t.Run("ClientType constants", func(t *testing.T) {
|
|
assert.Equal(t, models.ClientTypeOpenAIResponses, models.ClientType("openai-responses"))
|
|
assert.Equal(t, models.ClientTypeOpenAICompletions, models.ClientType("openai-completions"))
|
|
assert.Equal(t, models.ClientTypeAnthropicMessages, models.ClientType("anthropic-messages"))
|
|
assert.Equal(t, models.ClientTypeGoogleGenerativeAI, models.ClientType("google-generative-ai"))
|
|
})
|
|
}
|