automation and ai and some planning and fixes adn docs reorg

This commit is contained in:
IGNY8 VPS (Salman)
2025-12-29 01:41:36 +00:00
parent 748de099dd
commit 53fdebf733
20 changed files with 927 additions and 4288 deletions

View File

@@ -0,0 +1,311 @@
# Flexible Model Configuration System Plan
## Overview
This plan outlines how to implement a flexible model configuration system that allows:
- Adding/removing/activating models dynamically
- Configuring rates for each model
- Supporting multiple providers (OpenAI, Anthropic, Runware)
- Per-account model overrides
## Current State
### Model Rates (hardcoded in `ai/constants.py`)
```python
MODEL_RATES = {
'gpt-4.1': {'input': 2.00, 'output': 8.00}, # per 1M tokens
'gpt-4o-mini': {'input': 0.15, 'output': 0.60},
'gpt-4o': {'input': 2.50, 'output': 10.00},
'gpt-5.1': {'input': 1.25, 'output': 10.00},
'gpt-5.2': {'input': 1.75, 'output': 14.00},
}
IMAGE_MODEL_RATES = {
'dall-e-3': 0.040, # per image
'dall-e-2': 0.020,
'gpt-image-1': 0.042,
'gpt-image-1-mini': 0.011,
}
```
### Current Settings Architecture
- `GlobalIntegrationSettings` (singleton) - Platform-wide API keys and defaults
- `IntegrationSettings` (per-account) - Model/parameter overrides
- `GlobalAIPrompt` - Platform-wide prompt templates
- `AIPrompt` (per-account) - Custom prompt overrides
## Proposed Changes
### Phase 1: Database Model for AI Models
Create a new model `AIModel` to store model configurations:
```python
# backend/igny8_core/modules/system/global_settings_models.py
class AIModel(models.Model):
"""
Dynamic AI model configuration.
Replaces hardcoded MODEL_RATES and IMAGE_MODEL_RATES.
"""
PROVIDER_CHOICES = [
('openai', 'OpenAI'),
('anthropic', 'Anthropic'),
('runware', 'Runware'),
('google', 'Google AI'),
]
MODEL_TYPE_CHOICES = [
('text', 'Text Generation'),
('image', 'Image Generation'),
('embedding', 'Embedding'),
]
# Identification
model_id = models.CharField(
max_length=100,
unique=True,
help_text="Model identifier (e.g., 'gpt-4o-mini', 'claude-3-sonnet')"
)
display_name = models.CharField(
max_length=200,
help_text="User-friendly name (e.g., 'GPT-4o Mini')"
)
provider = models.CharField(max_length=50, choices=PROVIDER_CHOICES)
model_type = models.CharField(max_length=20, choices=MODEL_TYPE_CHOICES)
# Pricing (per 1M tokens for text, per image for image models)
input_rate = models.DecimalField(
max_digits=10,
decimal_places=4,
default=0,
help_text="Cost per 1M input tokens (text) or per request (image)"
)
output_rate = models.DecimalField(
max_digits=10,
decimal_places=4,
default=0,
help_text="Cost per 1M output tokens (text only)"
)
# Capabilities
max_tokens = models.IntegerField(
default=8192,
help_text="Maximum tokens for this model"
)
supports_json_mode = models.BooleanField(
default=True,
help_text="Whether model supports JSON response format"
)
supports_vision = models.BooleanField(
default=False,
help_text="Whether model supports image input"
)
# Status
is_active = models.BooleanField(default=True)
is_default = models.BooleanField(
default=False,
help_text="Use as default when no specific model is configured"
)
sort_order = models.IntegerField(default=0)
# Metadata
created_at = models.DateTimeField(auto_now_add=True)
updated_at = models.DateTimeField(auto_now=True)
class Meta:
db_table = 'igny8_ai_models'
ordering = ['sort_order', 'display_name']
def __str__(self):
return f"{self.display_name} ({self.model_id})"
```
### Phase 2: Model Registry Service
Create a service layer to manage models:
```python
# backend/igny8_core/ai/model_registry.py
class ModelRegistry:
"""
Central registry for AI model configurations.
Provides caching and fallback logic.
"""
_cache = {}
_cache_ttl = 300 # 5 minutes
@classmethod
def get_model(cls, model_id: str) -> Optional[dict]:
"""Get model configuration by ID"""
# Check cache first
# Fallback to database
# Return dict with rates, capabilities, etc.
pass
@classmethod
def get_models_by_type(cls, model_type: str) -> List[dict]:
"""Get all active models of a type"""
pass
@classmethod
def get_default_model(cls, model_type: str = 'text') -> dict:
"""Get default model for a type"""
pass
@classmethod
def calculate_cost(
cls,
model_id: str,
input_tokens: int = 0,
output_tokens: int = 0,
image_count: int = 0
) -> float:
"""Calculate cost for an operation"""
pass
@classmethod
def is_model_supported(cls, model_id: str) -> bool:
"""Check if a model is configured and active"""
pass
```
### Phase 3: Update AICore to Use Registry
Modify `ai_core.py` to use the model registry:
```python
# In run_ai_request()
from igny8_core.ai.model_registry import ModelRegistry
# Replace hardcoded MODEL_RATES check
if not ModelRegistry.is_model_supported(model):
supported = ModelRegistry.get_models_by_type('text')
error_msg = f"Model '{model}' is not supported. Available models: {[m['model_id'] for m in supported]}"
# ...
# Replace hardcoded cost calculation
model_info = ModelRegistry.get_model(model)
if model_info:
cost = ModelRegistry.calculate_cost(
model_id=model,
input_tokens=input_tokens,
output_tokens=output_tokens
)
```
### Phase 4: Admin Interface
Add Django admin for managing models:
```python
# backend/igny8_core/modules/system/admin.py
@admin.register(AIModel)
class AIModelAdmin(admin.ModelAdmin):
list_display = ['model_id', 'display_name', 'provider', 'model_type', 'input_rate', 'output_rate', 'is_active', 'is_default']
list_filter = ['provider', 'model_type', 'is_active', 'is_default']
search_fields = ['model_id', 'display_name']
ordering = ['sort_order', 'display_name']
fieldsets = (
('Identification', {
'fields': ('model_id', 'display_name', 'provider', 'model_type')
}),
('Pricing', {
'fields': ('input_rate', 'output_rate')
}),
('Capabilities', {
'fields': ('max_tokens', 'supports_json_mode', 'supports_vision')
}),
('Status', {
'fields': ('is_active', 'is_default', 'sort_order')
}),
)
```
### Phase 5: Data Migration
Create a migration to seed initial models:
```python
# Migration file
def seed_initial_models(apps, schema_editor):
AIModel = apps.get_model('system', 'AIModel')
models = [
# OpenAI Text Models
{'model_id': 'gpt-4o-mini', 'display_name': 'GPT-4o Mini', 'provider': 'openai', 'model_type': 'text', 'input_rate': 0.15, 'output_rate': 0.60, 'is_default': True},
{'model_id': 'gpt-4o', 'display_name': 'GPT-4o', 'provider': 'openai', 'model_type': 'text', 'input_rate': 2.50, 'output_rate': 10.00},
{'model_id': 'gpt-4.1', 'display_name': 'GPT-4.1', 'provider': 'openai', 'model_type': 'text', 'input_rate': 2.00, 'output_rate': 8.00},
{'model_id': 'gpt-5.1', 'display_name': 'GPT-5.1', 'provider': 'openai', 'model_type': 'text', 'input_rate': 1.25, 'output_rate': 10.00, 'max_tokens': 16000},
{'model_id': 'gpt-5.2', 'display_name': 'GPT-5.2', 'provider': 'openai', 'model_type': 'text', 'input_rate': 1.75, 'output_rate': 14.00, 'max_tokens': 16000},
# Anthropic Text Models
{'model_id': 'claude-3-sonnet', 'display_name': 'Claude 3 Sonnet', 'provider': 'anthropic', 'model_type': 'text', 'input_rate': 3.00, 'output_rate': 15.00},
{'model_id': 'claude-3-opus', 'display_name': 'Claude 3 Opus', 'provider': 'anthropic', 'model_type': 'text', 'input_rate': 15.00, 'output_rate': 75.00},
{'model_id': 'claude-3-haiku', 'display_name': 'Claude 3 Haiku', 'provider': 'anthropic', 'model_type': 'text', 'input_rate': 0.25, 'output_rate': 1.25},
# OpenAI Image Models
{'model_id': 'dall-e-3', 'display_name': 'DALL-E 3', 'provider': 'openai', 'model_type': 'image', 'input_rate': 0.040, 'output_rate': 0},
{'model_id': 'dall-e-2', 'display_name': 'DALL-E 2', 'provider': 'openai', 'model_type': 'image', 'input_rate': 0.020, 'output_rate': 0},
{'model_id': 'gpt-image-1', 'display_name': 'GPT Image 1', 'provider': 'openai', 'model_type': 'image', 'input_rate': 0.042, 'output_rate': 0},
# Runware Image Models
{'model_id': 'runware:97@1', 'display_name': 'Runware 97@1', 'provider': 'runware', 'model_type': 'image', 'input_rate': 0.009, 'output_rate': 0},
]
for i, model in enumerate(models):
AIModel.objects.create(sort_order=i, **model)
```
### Phase 6: API Endpoints for Model Management
Add REST endpoints for managing models:
```python
# GET /api/v1/admin/ai-models/ - List all models
# POST /api/v1/admin/ai-models/ - Create new model
# PUT /api/v1/admin/ai-models/{id}/ - Update model
# DELETE /api/v1/admin/ai-models/{id}/ - Delete model
# POST /api/v1/admin/ai-models/{id}/toggle-active/ - Toggle active status
# POST /api/v1/admin/ai-models/{id}/set-default/ - Set as default
```
### Phase 7: Frontend Admin UI
Create admin UI for model management:
- List view with filtering/sorting
- Create/Edit form with validation
- Quick toggle for active/default status
- Price calculator preview
## Implementation Order
1. **Week 1**: Create `AIModel` model and migration
2. **Week 1**: Create `ModelRegistry` service
3. **Week 2**: Update `ai_core.py` to use registry
4. **Week 2**: Update `constants.py` to load from database
5. **Week 3**: Add Django admin interface
6. **Week 3**: Add API endpoints
7. **Week 4**: Create frontend admin UI
8. **Week 4**: Testing and documentation
## Backward Compatibility
- Keep `constants.py` as fallback if database is empty
- `ModelRegistry.get_model()` checks DB first, falls back to constants
- No changes to existing `GlobalIntegrationSettings` or `IntegrationSettings`
- Existing API calls continue to work unchanged
## Benefits
1. **No Code Changes for New Models**: Add models via admin UI
2. **Easy Price Updates**: Update rates without deployment
3. **Provider Flexibility**: Support any provider by adding models
4. **Per-Provider Settings**: Configure different capabilities per provider
5. **Audit Trail**: Track when models were added/modified
6. **A/B Testing**: Easily enable/disable models for testing