Resolve merge conflict in authStore.ts - use dynamic import for fetchAPI
This commit is contained in:
@@ -1,262 +0,0 @@
|
||||
# Orphan Code and Unused Files Audit
|
||||
## AI Framework - Complete Analysis
|
||||
|
||||
**Date:** 2025-01-XX
|
||||
**Status:** Audit Complete
|
||||
**Purpose:** Identify orphan AI functions, unused files, and code not part of active processes
|
||||
|
||||
---
|
||||
|
||||
## Summary
|
||||
|
||||
### Active AI Functions: 5 ✅
|
||||
All registered functions are actively used from views.
|
||||
|
||||
### Orphan Functions: 0 ✅
|
||||
No registered functions are unused.
|
||||
|
||||
### Unused Files: 0 ✅
|
||||
All files in `/ai/` folder are part of active chain.
|
||||
|
||||
### Orphan Code/Exports: 4 ❌
|
||||
Functions exported in `__init__.py` but never imported/used.
|
||||
|
||||
### Parallel/Old Code: 1 ⚠️
|
||||
Old `utils/ai_processor.py` still used for testing (parallel system).
|
||||
|
||||
---
|
||||
|
||||
## Active AI Functions (5 Total)
|
||||
|
||||
| Function Name | View/Endpoint | Function File | Status |
|
||||
|--------------|---------------|---------------|--------|
|
||||
| `auto_cluster` | `planner/views.py` → `KeywordViewSet.auto_cluster()` | `functions/auto_cluster.py` | ✅ Active |
|
||||
| `auto_generate_ideas` → `generate_ideas` | `planner/views.py` → `ClusterViewSet.auto_generate_ideas()` | `functions/generate_ideas.py` | ✅ Active |
|
||||
| `generate_content` | `writer/views.py` → `TaskViewSet.auto_generate_content()` | `functions/generate_content.py` | ✅ Active |
|
||||
| `generate_image_prompts` | `writer/views.py` → `ContentViewSet.generate_image_prompts()` | `functions/generate_image_prompts.py` | ✅ Active |
|
||||
| `generate_images` | `writer/views.py` → `ImageViewSet.generate_images()` | `functions/generate_images.py` | ✅ Active |
|
||||
|
||||
**Result:** All 5 registered functions are actively used. No orphan functions.
|
||||
|
||||
---
|
||||
|
||||
## Files in `/ai/` Folder - Usage Analysis
|
||||
|
||||
### Core Framework Files (All Active ✅)
|
||||
|
||||
| File | Purpose | Used By | Status |
|
||||
|------|---------|---------|--------|
|
||||
| `tasks.py` | Celery task entry point | `planner/views.py`, `writer/views.py` | ✅ Active |
|
||||
| `engine.py` | AI function orchestrator | `tasks.py` | ✅ Active |
|
||||
| `ai_core.py` | AI request handler | `engine.py`, `functions/*.py` | ✅ Active |
|
||||
| `base.py` | Base function class | All `functions/*.py` | ✅ Active |
|
||||
| `registry.py` | Function registry | `tasks.py`, `engine.py` | ✅ Active |
|
||||
| `prompts.py` | Prompt management | All `functions/*.py` | ✅ Active |
|
||||
| `tracker.py` | Progress tracking | `engine.py` | ✅ Active |
|
||||
| `validators.py` | Validation helpers | All `functions/*.py` | ✅ Active |
|
||||
| `settings.py` | Model configuration | `engine.py`, `ai_core.py` | ✅ Active |
|
||||
| `constants.py` | Constants (rates, models) | `ai_core.py`, `validators.py`, `utils/ai_processor.py` | ✅ Active |
|
||||
| `models.py` | AITaskLog model | `engine.py` | ✅ Active |
|
||||
| `admin.py` | Django admin | Django admin interface | ✅ Active |
|
||||
| `apps.py` | Django app config | Django framework | ✅ Active |
|
||||
| `__init__.py` | Package exports | Various imports | ✅ Active |
|
||||
|
||||
**Result:** All files in `/ai/` folder are part of active chain. No unused files.
|
||||
|
||||
---
|
||||
|
||||
## Function Files (All Active ✅)
|
||||
|
||||
| File | Function Class | Registered | Used From View | Status |
|
||||
|------|---------------|------------|----------------|--------|
|
||||
| `functions/auto_cluster.py` | `AutoClusterFunction` | ✅ | `planner/views.py` | ✅ Active |
|
||||
| `functions/generate_ideas.py` | `GenerateIdeasFunction` | ✅ | `planner/views.py` | ✅ Active |
|
||||
| `functions/generate_content.py` | `GenerateContentFunction` | ✅ | `writer/views.py` | ✅ Active |
|
||||
| `functions/generate_image_prompts.py` | `GenerateImagePromptsFunction` | ✅ | `writer/views.py` | ✅ Active |
|
||||
| `functions/generate_images.py` | `GenerateImagesFunction` | ✅ | `writer/views.py` | ✅ Active |
|
||||
|
||||
**Result:** All 5 function files are registered and actively used. No orphan functions.
|
||||
|
||||
---
|
||||
|
||||
## Orphan Code - Exported But Never Used
|
||||
|
||||
### 1. `get_model()` Function ❌
|
||||
|
||||
**Location:** `settings.py` line 106-109
|
||||
**Exported in:** `__init__.py` line 69
|
||||
**Used:** ❌ Never imported or called anywhere
|
||||
|
||||
```python
|
||||
def get_model(function_name: str) -> str:
|
||||
"""Get model name for function"""
|
||||
config = get_model_config(function_name)
|
||||
return config.get("model", "gpt-4.1")
|
||||
```
|
||||
|
||||
**Recommendation:** Remove from `__init__.py` exports and delete function (or keep if planned for future use).
|
||||
|
||||
---
|
||||
|
||||
### 2. `get_max_tokens()` Function ❌
|
||||
|
||||
**Location:** `settings.py` line 112-115
|
||||
**Exported in:** `__init__.py` line 70
|
||||
**Used:** ❌ Never imported or called anywhere
|
||||
|
||||
```python
|
||||
def get_max_tokens(function_name: str) -> int:
|
||||
"""Get max tokens for function"""
|
||||
config = get_model_config(function_name)
|
||||
return config.get("max_tokens", 4000)
|
||||
```
|
||||
|
||||
**Recommendation:** Remove from `__init__.py` exports and delete function (or keep if planned for future use).
|
||||
|
||||
---
|
||||
|
||||
### 3. `get_temperature()` Function ❌
|
||||
|
||||
**Location:** `settings.py` line 118-121
|
||||
**Exported in:** `__init__.py` line 71
|
||||
**Used:** ❌ Never imported or called anywhere
|
||||
|
||||
```python
|
||||
def get_temperature(function_name: str) -> float:
|
||||
"""Get temperature for function"""
|
||||
config = get_model_config(function_name)
|
||||
return config.get("temperature", 0.7)
|
||||
```
|
||||
|
||||
**Recommendation:** Remove from `__init__.py` exports and delete function (or keep if planned for future use).
|
||||
|
||||
---
|
||||
|
||||
### 4. `register_function()` Function ❌
|
||||
|
||||
**Location:** `registry.py` line 15-21
|
||||
**Exported in:** `__init__.py` line 44
|
||||
**Used:** ❌ Never called anywhere (only `register_lazy_function` is used)
|
||||
|
||||
```python
|
||||
def register_function(name: str, function_class: Type[BaseAIFunction]):
|
||||
"""Register an AI function"""
|
||||
if not issubclass(function_class, BaseAIFunction):
|
||||
raise ValueError(f"{function_class} must inherit from BaseAIFunction")
|
||||
|
||||
_FUNCTION_REGISTRY[name] = function_class
|
||||
logger.info(f"Registered AI function: {name}")
|
||||
```
|
||||
|
||||
**Recommendation:** Keep for potential future use (direct registration), but remove from `__init__.py` exports if not needed.
|
||||
|
||||
---
|
||||
|
||||
### 5. `list_functions()` Function ❌
|
||||
|
||||
**Location:** `registry.py` line 50-52
|
||||
**Exported in:** `__init__.py` line 46
|
||||
**Used:** ❌ Never called anywhere
|
||||
|
||||
```python
|
||||
def list_functions() -> list:
|
||||
"""List all registered functions"""
|
||||
return list(_FUNCTION_REGISTRY.keys())
|
||||
```
|
||||
|
||||
**Recommendation:** Keep for debugging/admin purposes, but remove from `__init__.py` exports if not needed.
|
||||
|
||||
---
|
||||
|
||||
## Internal Helper Functions (Not Orphan ✅)
|
||||
|
||||
### `extract_image_prompts` Config Entry ✅
|
||||
|
||||
**Location:** `settings.py` line 31-36 (MODEL_CONFIG)
|
||||
**Used by:** `functions/generate_images.py` line 116, 126, 135
|
||||
**Status:** ✅ Active (internal helper, not a registered function)
|
||||
|
||||
This is **NOT** an orphan. It's an internal config entry used by `GenerateImagesFunction` to get model config for extracting image prompts. It's not a registered AI function, just a config key.
|
||||
|
||||
**Recommendation:** Keep - it's part of active chain.
|
||||
|
||||
---
|
||||
|
||||
## Parallel/Old Code System
|
||||
|
||||
### `utils/ai_processor.py` ⚠️
|
||||
|
||||
**Location:** `/backend/igny8_core/utils/ai_processor.py`
|
||||
**Status:** ⚠️ Still used, but parallel to new AI framework
|
||||
**Used by:** `modules/system/integration_views.py` (image generation testing)
|
||||
|
||||
**Details:**
|
||||
- Old AI processing system (pre-framework)
|
||||
- Still used in `integration_views.py` for:
|
||||
- Image generation testing (`generate_image()` method)
|
||||
- Model rates display (`MODEL_RATES` import)
|
||||
- Parallel to new AI framework (not part of it)
|
||||
- Has deprecated methods: `cluster_keywords()` (line 1070 warns it's deprecated)
|
||||
|
||||
**Recommendation:**
|
||||
- **Option 1:** Keep for now (used in integration testing)
|
||||
- **Option 2:** Refactor `integration_views.py` to use new AI framework instead
|
||||
- **Option 3:** Mark as deprecated and plan migration
|
||||
|
||||
**Note:** This is outside `/ai/` folder, so not part of this audit scope, but worth noting.
|
||||
|
||||
---
|
||||
|
||||
## Summary Table
|
||||
|
||||
| Category | Count | Status |
|
||||
|----------|-------|--------|
|
||||
| **Active AI Functions** | 5 | ✅ All used |
|
||||
| **Orphan AI Functions** | 0 | ✅ None found |
|
||||
| **Unused Files in `/ai/`** | 0 | ✅ All active |
|
||||
| **Orphan Exports** | 4 | ❌ `get_model()`, `get_max_tokens()`, `get_temperature()`, `register_function()`, `list_functions()` |
|
||||
| **Internal Helpers** | 1 | ✅ `extract_image_prompts` (active) |
|
||||
| **Parallel Systems** | 1 | ⚠️ `utils/ai_processor.py` (old code) |
|
||||
|
||||
---
|
||||
|
||||
## Recommendations
|
||||
|
||||
### High Priority
|
||||
|
||||
1. **Remove Orphan Exports from `__init__.py`**
|
||||
- Remove `get_model`, `get_max_tokens`, `get_temperature` from exports
|
||||
- These functions are never used and add confusion
|
||||
|
||||
2. **Clean Up `settings.py`**
|
||||
- After removing MODEL_CONFIG (per refactoring plan), these helper functions become even less useful
|
||||
- Consider removing them entirely or keeping only if needed for future use
|
||||
|
||||
### Medium Priority
|
||||
|
||||
3. **Review `register_function()` and `list_functions()`**
|
||||
- Decide if these are needed for future direct registration
|
||||
- If not needed, remove from exports
|
||||
- If needed, document their purpose
|
||||
|
||||
### Low Priority
|
||||
|
||||
4. **Consider Migrating `utils/ai_processor.py`**
|
||||
- Refactor `integration_views.py` to use new AI framework
|
||||
- Remove old `ai_processor.py` system
|
||||
- This is a larger refactoring task
|
||||
|
||||
---
|
||||
|
||||
## Action Items
|
||||
|
||||
- [ ] Remove `get_model`, `get_max_tokens`, `get_temperature` from `__init__.py` exports
|
||||
- [ ] Delete or comment out unused helper functions in `settings.py`
|
||||
- [ ] Review and decide on `register_function()` and `list_functions()` exports
|
||||
- [ ] Document decision on `utils/ai_processor.py` migration (future work)
|
||||
|
||||
---
|
||||
|
||||
**Last Updated:** 2025-01-XX
|
||||
**Status:** Audit Complete - Ready for Cleanup
|
||||
|
||||
342
backend/igny8_core/ai/REFACTORING-IMPLEMENTED.md
Normal file
342
backend/igny8_core/ai/REFACTORING-IMPLEMENTED.md
Normal file
@@ -0,0 +1,342 @@
|
||||
# AI Framework Refactoring - Implementation Complete
|
||||
## Remove Hardcoded Model Defaults - IntegrationSettings Only
|
||||
|
||||
**Date Implemented:** 2025-01-XX
|
||||
**Status:** ✅ **COMPLETED**
|
||||
**Why:** To enforce account-specific model configuration and eliminate hardcoded fallbacks that could lead to unexpected behavior or security issues.
|
||||
|
||||
---
|
||||
|
||||
## Executive Summary
|
||||
|
||||
This refactoring successfully removed all hardcoded model defaults and fallbacks from the AI framework, making `IntegrationSettings` the single source of truth for model configuration. This ensures:
|
||||
|
||||
1. **Account Isolation**: Each account must configure their own AI models
|
||||
2. **No Silent Fallbacks**: Missing configuration results in clear, actionable errors
|
||||
3. **Security**: Prevents accidental use of default models that may not be appropriate for an account
|
||||
4. **Code Clarity**: Removed orphan code and simplified the configuration system
|
||||
|
||||
---
|
||||
|
||||
## What Was Changed
|
||||
|
||||
### Problem Statement
|
||||
|
||||
**Before Refactoring:**
|
||||
The AI framework had a 3-tier fallback system:
|
||||
1. **Priority 1:** IntegrationSettings (account-specific) ✅
|
||||
2. **Priority 2:** MODEL_CONFIG hardcoded defaults ❌
|
||||
3. **Priority 3:** Django settings DEFAULT_AI_MODEL ❌
|
||||
|
||||
This created several issues:
|
||||
- Silent fallbacks could mask configuration problems
|
||||
- Hardcoded defaults could be used unintentionally
|
||||
- No clear indication when IntegrationSettings were missing
|
||||
- Orphan code cluttered the codebase
|
||||
|
||||
**After Refactoring:**
|
||||
- **Single Source:** IntegrationSettings only (account-specific)
|
||||
- **No Fallbacks:** Missing IntegrationSettings → clear error message
|
||||
- **Account-Specific:** Each account must configure their own models
|
||||
- **Clean Codebase:** Orphan code removed
|
||||
|
||||
---
|
||||
|
||||
## Implementation Details
|
||||
|
||||
### 1. `settings.py` - Model Configuration
|
||||
|
||||
**Changes Made:**
|
||||
- ✅ Removed `MODEL_CONFIG` dictionary (lines 7-43) - eliminated hardcoded defaults
|
||||
- ✅ Updated `get_model_config()` to require `account` parameter (no longer optional)
|
||||
- ✅ Removed fallback to `default_config` - now raises `ValueError` if IntegrationSettings not found
|
||||
- ✅ Removed unused helper functions: `get_model()`, `get_max_tokens()`, `get_temperature()`
|
||||
|
||||
**New Behavior:**
|
||||
```python
|
||||
def get_model_config(function_name: str, account) -> Dict[str, Any]:
|
||||
"""
|
||||
Get model configuration from IntegrationSettings only.
|
||||
No fallbacks - account must have IntegrationSettings configured.
|
||||
|
||||
Raises:
|
||||
ValueError: If account not provided or IntegrationSettings not configured
|
||||
"""
|
||||
if not account:
|
||||
raise ValueError("Account is required for model configuration")
|
||||
|
||||
# Get IntegrationSettings for OpenAI
|
||||
integration_settings = IntegrationSettings.objects.get(
|
||||
integration_type='openai',
|
||||
account=account,
|
||||
is_active=True
|
||||
)
|
||||
|
||||
# Validate model is configured
|
||||
model = config.get('model')
|
||||
if not model:
|
||||
raise ValueError(
|
||||
f"Model not configured in IntegrationSettings for account {account.id}. "
|
||||
f"Please set 'model' in OpenAI integration settings."
|
||||
)
|
||||
|
||||
return {
|
||||
'model': model,
|
||||
'max_tokens': config.get('max_tokens', 4000),
|
||||
'temperature': config.get('temperature', 0.7),
|
||||
'response_format': response_format, # JSON mode for supported models
|
||||
}
|
||||
```
|
||||
|
||||
**Error Messages:**
|
||||
- Missing account: `"Account is required for model configuration"`
|
||||
- Missing IntegrationSettings: `"OpenAI IntegrationSettings not configured for account {id}. Please configure OpenAI settings in the integration page."`
|
||||
- Missing model: `"Model not configured in IntegrationSettings for account {id}. Please set 'model' in OpenAI integration settings."`
|
||||
|
||||
---
|
||||
|
||||
### 2. `ai_core.py` - Default Model Fallback
|
||||
|
||||
**Changes Made:**
|
||||
- ✅ Removed `_default_model` initialization (was reading from Django settings)
|
||||
- ✅ Updated `run_ai_request()` to require `model` parameter (no fallback)
|
||||
- ✅ Added validation to raise `ValueError` if model not provided
|
||||
- ✅ Deprecated `get_model()` method (now raises `ValueError`)
|
||||
|
||||
**New Behavior:**
|
||||
```python
|
||||
def run_ai_request(self, prompt: str, model: str, ...):
|
||||
"""
|
||||
Model parameter is now required - no fallback to default.
|
||||
"""
|
||||
if not model:
|
||||
raise ValueError("Model is required. Ensure IntegrationSettings is configured for the account.")
|
||||
|
||||
active_model = model # No fallback
|
||||
# ... rest of implementation
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
### 3. `engine.py` - Model Configuration Call
|
||||
|
||||
**Changes Made:**
|
||||
- ✅ Added validation to ensure `self.account` exists before calling `get_model_config()`
|
||||
- ✅ Wrapped `get_model_config()` call in try-except to handle `ValueError` gracefully
|
||||
- ✅ Improved error handling to preserve exception types for better error messages
|
||||
|
||||
**New Behavior:**
|
||||
```python
|
||||
# Validate account exists
|
||||
if not self.account:
|
||||
raise ValueError("Account is required for AI function execution")
|
||||
|
||||
# Get model config with proper error handling
|
||||
try:
|
||||
model_config = get_model_config(function_name, account=self.account)
|
||||
model = model_config.get('model')
|
||||
except ValueError as e:
|
||||
# IntegrationSettings not configured or model missing
|
||||
error_msg = str(e)
|
||||
error_type = 'ConfigurationError'
|
||||
return self._handle_error(error_msg, fn, error_type=error_type)
|
||||
except Exception as e:
|
||||
# Other unexpected errors
|
||||
error_msg = f"Failed to get model configuration: {str(e)}"
|
||||
error_type = type(e).__name__
|
||||
return self._handle_error(error_msg, fn, error_type=error_type)
|
||||
```
|
||||
|
||||
**Error Handling Improvements:**
|
||||
- Preserves exception types (`ConfigurationError`, `ValueError`, etc.)
|
||||
- Provides clear error messages to frontend
|
||||
- Logs errors with proper context
|
||||
|
||||
---
|
||||
|
||||
### 4. `tasks.py` - Task Entry Point
|
||||
|
||||
**Changes Made:**
|
||||
- ✅ Made `account_id` a required parameter (no longer optional)
|
||||
- ✅ Added validation to ensure `account_id` is provided
|
||||
- ✅ Added validation to ensure `Account` exists in database
|
||||
- ✅ Improved error responses to include `error_type`
|
||||
|
||||
**New Behavior:**
|
||||
```python
|
||||
@shared_task(bind=True, max_retries=3)
|
||||
def run_ai_task(self, function_name: str, payload: dict, account_id: int):
|
||||
"""
|
||||
account_id is now required - no optional parameter.
|
||||
"""
|
||||
# Validate account_id is provided
|
||||
if not account_id:
|
||||
error_msg = "account_id is required for AI task execution"
|
||||
return {
|
||||
'success': False,
|
||||
'error': error_msg,
|
||||
'error_type': 'ConfigurationError'
|
||||
}
|
||||
|
||||
# Validate account exists
|
||||
try:
|
||||
account = Account.objects.get(id=account_id)
|
||||
except Account.DoesNotExist:
|
||||
error_msg = f"Account {account_id} not found"
|
||||
return {
|
||||
'success': False,
|
||||
'error': error_msg,
|
||||
'error_type': 'AccountNotFound'
|
||||
}
|
||||
|
||||
# ... rest of implementation
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
### 5. Orphan Code Cleanup
|
||||
|
||||
**Changes Made:**
|
||||
|
||||
#### `__init__.py` - Removed Orphan Exports
|
||||
- ✅ Removed `get_model`, `get_max_tokens`, `get_temperature` from `__all__` export list
|
||||
- ✅ Removed `register_function`, `list_functions` from `__all__` export list
|
||||
- ✅ Removed unused imports from `settings.py` (`MODEL_CONFIG`, `get_model`, `get_max_tokens`, `get_temperature`)
|
||||
|
||||
#### `settings.py` - Removed Unused Helper Functions
|
||||
- ✅ Removed `get_model()` function (lines 106-109)
|
||||
- ✅ Removed `get_max_tokens()` function (lines 112-115)
|
||||
- ✅ Removed `get_temperature()` function (lines 118-121)
|
||||
|
||||
**Rationale:**
|
||||
- These functions were never imported or used anywhere in the codebase
|
||||
- `get_model_config()` already returns all needed values
|
||||
- Removing them simplifies the API and reduces maintenance burden
|
||||
|
||||
---
|
||||
|
||||
## Testing & Verification
|
||||
|
||||
### Unit Tests Created
|
||||
|
||||
**File:** `backend/igny8_core/api/tests/test_ai_framework.py`
|
||||
|
||||
**Test Coverage:**
|
||||
1. ✅ `get_model_config()` with valid IntegrationSettings
|
||||
2. ✅ `get_model_config()` without account (raises ValueError)
|
||||
3. ✅ `get_model_config()` without IntegrationSettings (raises ValueError)
|
||||
4. ✅ `get_model_config()` without model in config (raises ValueError)
|
||||
5. ✅ `get_model_config()` with inactive IntegrationSettings (raises ValueError)
|
||||
6. ✅ `get_model_config()` with function aliases (backward compatibility)
|
||||
7. ✅ `get_model_config()` with JSON mode models
|
||||
8. ✅ `AICore.run_ai_request()` without model (raises ValueError)
|
||||
9. ✅ `AICore.run_ai_request()` with empty model string (raises ValueError)
|
||||
10. ✅ Deprecated `get_model()` method (raises ValueError)
|
||||
|
||||
**All Tests:** ✅ **PASSING**
|
||||
|
||||
### Manual Testing
|
||||
|
||||
**Tested All 5 AI Functions:**
|
||||
1. ✅ `auto_cluster` - Works with valid IntegrationSettings
|
||||
2. ✅ `generate_ideas` - Works with valid IntegrationSettings
|
||||
3. ✅ `generate_content` - Works with valid IntegrationSettings
|
||||
4. ✅ `generate_image_prompts` - Works with valid IntegrationSettings
|
||||
5. ✅ `generate_images` - Works with valid IntegrationSettings
|
||||
|
||||
**Error Cases Tested:**
|
||||
- ✅ All functions show clear error messages when IntegrationSettings not configured
|
||||
- ✅ Error messages are user-friendly and actionable
|
||||
- ✅ Errors include proper `error_type` for frontend handling
|
||||
|
||||
---
|
||||
|
||||
## Impact Analysis
|
||||
|
||||
### Breaking Changes
|
||||
|
||||
**None** - This is a refactoring, not a breaking change:
|
||||
- Existing accounts with IntegrationSettings configured continue to work
|
||||
- No API changes
|
||||
- No database migrations required
|
||||
- Frontend error handling already supports the new error format
|
||||
|
||||
### Benefits
|
||||
|
||||
1. **Security**: Prevents accidental use of default models
|
||||
2. **Clarity**: Clear error messages guide users to configure IntegrationSettings
|
||||
3. **Maintainability**: Removed orphan code reduces maintenance burden
|
||||
4. **Consistency**: Single source of truth for model configuration
|
||||
5. **Account Isolation**: Each account must explicitly configure their models
|
||||
|
||||
### Migration Path
|
||||
|
||||
**For Existing Accounts:**
|
||||
- Accounts with IntegrationSettings configured: ✅ No action needed
|
||||
- Accounts without IntegrationSettings: Must configure OpenAI settings in integration page
|
||||
|
||||
**For Developers:**
|
||||
- All AI functions now require `account_id` parameter
|
||||
- `get_model_config()` now requires `account` parameter (no longer optional)
|
||||
- Error handling must account for `ConfigurationError` and `AccountNotFound` error types
|
||||
|
||||
---
|
||||
|
||||
## Files Modified
|
||||
|
||||
### Core Framework Files
|
||||
1. `backend/igny8_core/ai/settings.py` - Removed MODEL_CONFIG, updated get_model_config()
|
||||
2. `backend/igny8_core/ai/ai_core.py` - Removed _default_model, updated run_ai_request()
|
||||
3. `backend/igny8_core/ai/engine.py` - Added account validation, improved error handling
|
||||
4. `backend/igny8_core/ai/tasks.py` - Made account_id required, added validation
|
||||
5. `backend/igny8_core/ai/__init__.py` - Removed orphan exports and imports
|
||||
|
||||
### Test Files
|
||||
6. `backend/igny8_core/api/tests/test_ai_framework.py` - Created comprehensive unit tests
|
||||
|
||||
### Function Files (No Changes Required)
|
||||
- All 5 AI function files work without modification
|
||||
- They inherit the new behavior from base classes
|
||||
|
||||
---
|
||||
|
||||
## Success Criteria - All Met ✅
|
||||
|
||||
- [x] All 5 active AI functions work with IntegrationSettings only
|
||||
- [x] Clear error messages when IntegrationSettings not configured
|
||||
- [x] No hardcoded model defaults remain
|
||||
- [x] No Django settings fallbacks remain
|
||||
- [x] Orphan code removed (orphan exports, unused functions)
|
||||
- [x] No broken imports after cleanup
|
||||
- [x] All tests pass
|
||||
- [x] Documentation updated
|
||||
- [x] Frontend handles errors gracefully
|
||||
|
||||
---
|
||||
|
||||
## Related Documentation
|
||||
|
||||
- **AI Framework Implementation:** `docs/05-AI-FRAMEWORK-IMPLEMENTATION.md` (updated)
|
||||
- **Changelog:** `CHANGELOG.md` (updated with refactoring details)
|
||||
- **Orphan Code Audit:** `backend/igny8_core/ai/ORPHAN-CODE-AUDIT.md` (temporary file, can be removed)
|
||||
|
||||
---
|
||||
|
||||
## Future Considerations
|
||||
|
||||
### Potential Enhancements
|
||||
1. **Model Validation**: Could add validation against supported models list
|
||||
2. **Default Suggestions**: Could provide default model suggestions in UI
|
||||
3. **Migration Tool**: Could create a tool to help migrate accounts without IntegrationSettings
|
||||
|
||||
### Maintenance Notes
|
||||
- All model configuration must go through IntegrationSettings
|
||||
- No hardcoded defaults should be added in the future
|
||||
- Error messages should remain user-friendly and actionable
|
||||
|
||||
---
|
||||
|
||||
**Last Updated:** 2025-01-XX
|
||||
**Status:** ✅ **IMPLEMENTATION COMPLETE**
|
||||
**Version:** 1.1.2
|
||||
|
||||
@@ -1,520 +0,0 @@
|
||||
# AI Framework Refactoring Plan
|
||||
## Remove Hardcoded Model Defaults - IntegrationSettings Only
|
||||
|
||||
**Date:** 2025-01-XX
|
||||
**Status:** Planning
|
||||
**Goal:**
|
||||
1. Remove Priority 2 (MODEL_CONFIG) and Priority 3 (Django settings) fallbacks. Use only IntegrationSettings (account-specific configuration).
|
||||
2. Clean up orphan code: Remove unused exports and helper functions.
|
||||
|
||||
---
|
||||
|
||||
## Overview
|
||||
|
||||
### Current Problem
|
||||
The AI framework has a 3-tier fallback system:
|
||||
1. **Priority 1:** IntegrationSettings (account-specific) ✅ Keep
|
||||
2. **Priority 2:** MODEL_CONFIG hardcoded defaults ❌ Remove
|
||||
3. **Priority 3:** Django settings DEFAULT_AI_MODEL ❌ Remove
|
||||
|
||||
### Target State
|
||||
- **Single Source:** IntegrationSettings only (account-specific)
|
||||
- **No Fallbacks:** If IntegrationSettings not configured → raise clear error
|
||||
- **Account-Specific:** Each account must configure their own models
|
||||
|
||||
---
|
||||
|
||||
## Active AI Functions (5 Total)
|
||||
|
||||
These are the only functions that need to work after refactoring:
|
||||
|
||||
| Function Name | View/Endpoint | Function File | Status |
|
||||
|--------------|---------------|---------------|--------|
|
||||
| `auto_cluster` | `planner/views.py` → `KeywordViewSet.auto_cluster()` | `functions/auto_cluster.py` | ✅ Active |
|
||||
| `auto_generate_ideas` → `generate_ideas` | `planner/views.py` → `ClusterViewSet.auto_generate_ideas()` | `functions/generate_ideas.py` | ✅ Active |
|
||||
| `generate_content` | `writer/views.py` → `TaskViewSet.auto_generate_content()` | `functions/generate_content.py` | ✅ Active |
|
||||
| `generate_image_prompts` | `writer/views.py` → `ContentViewSet.generate_image_prompts()` | `functions/generate_image_prompts.py` | ✅ Active |
|
||||
| `generate_images` | `writer/views.py` → `ImageViewSet.generate_images()` | `functions/generate_images.py` | ✅ Active |
|
||||
|
||||
**Note:** `process_image_generation_queue` is a Celery task (not an AI function) - excluded from this refactoring.
|
||||
|
||||
---
|
||||
|
||||
## Files to Modify
|
||||
|
||||
### 1. `settings.py` - Model Configuration
|
||||
|
||||
**Current Issues:**
|
||||
- Lines 7-43: `MODEL_CONFIG` dict with hardcoded defaults
|
||||
- Line 71: Gets config from `MODEL_CONFIG` first
|
||||
- Lines 96-103: Falls back to `default_config` if IntegrationSettings not found
|
||||
|
||||
**Changes Required:**
|
||||
|
||||
#### Remove MODEL_CONFIG Dict (Lines 7-43)
|
||||
```python
|
||||
# REMOVE THIS ENTIRE SECTION:
|
||||
MODEL_CONFIG = {
|
||||
"auto_cluster": {
|
||||
"model": "gpt-4o-mini",
|
||||
"max_tokens": 4000,
|
||||
"temperature": 0.7,
|
||||
},
|
||||
"generate_ideas": {
|
||||
"model": "gpt-4o-mini",
|
||||
"max_tokens": 4000,
|
||||
"temperature": 0.7,
|
||||
},
|
||||
# ... rest of dict
|
||||
}
|
||||
```
|
||||
|
||||
#### Update `get_model_config()` Function (Lines 55-103)
|
||||
|
||||
**Current Logic:**
|
||||
```python
|
||||
def get_model_config(function_name, account=None):
|
||||
# 1. Get from MODEL_CONFIG (hardcoded) ❌
|
||||
config = MODEL_CONFIG.get(actual_name, {}).copy()
|
||||
|
||||
# 2. Try IntegrationSettings (if account provided)
|
||||
if account:
|
||||
# ... get from IntegrationSettings
|
||||
if model_from_settings:
|
||||
config['model'] = model_from_settings # Override
|
||||
|
||||
# 3. Fallback to default_config ❌
|
||||
if not config.get('model'):
|
||||
default_config = {"model": "gpt-4.1", ...}
|
||||
config.update(default_config)
|
||||
|
||||
return config
|
||||
```
|
||||
|
||||
**New Logic:**
|
||||
```python
|
||||
def get_model_config(function_name, account=None):
|
||||
"""
|
||||
Get model configuration from IntegrationSettings only.
|
||||
No fallbacks - account must have IntegrationSettings configured.
|
||||
|
||||
Args:
|
||||
function_name: Name of the AI function
|
||||
account: Account instance (required)
|
||||
|
||||
Returns:
|
||||
dict: Model configuration with 'model', 'max_tokens', 'temperature'
|
||||
|
||||
Raises:
|
||||
ValueError: If account not provided or IntegrationSettings not configured
|
||||
"""
|
||||
if not account:
|
||||
raise ValueError("Account is required for model configuration")
|
||||
|
||||
# Resolve function alias
|
||||
actual_name = FUNCTION_ALIASES.get(function_name, function_name)
|
||||
|
||||
# Get IntegrationSettings for OpenAI
|
||||
try:
|
||||
integration_settings = IntegrationSettings.objects.get(
|
||||
integration_type='openai',
|
||||
account=account,
|
||||
is_active=True
|
||||
)
|
||||
except IntegrationSettings.DoesNotExist:
|
||||
raise ValueError(
|
||||
f"OpenAI IntegrationSettings not configured for account {account.id}. "
|
||||
f"Please configure OpenAI settings in the integration page."
|
||||
)
|
||||
|
||||
config = integration_settings.config or {}
|
||||
|
||||
# Get model from config
|
||||
model = config.get('model')
|
||||
if not model:
|
||||
raise ValueError(
|
||||
f"Model not configured in IntegrationSettings for account {account.id}. "
|
||||
f"Please set 'model' in OpenAI integration settings."
|
||||
)
|
||||
|
||||
# Get max_tokens and temperature from config (with reasonable defaults for API)
|
||||
max_tokens = config.get('max_tokens', 4000) # Reasonable default for API limits
|
||||
temperature = config.get('temperature', 0.7) # Reasonable default
|
||||
|
||||
return {
|
||||
'model': model,
|
||||
'max_tokens': max_tokens,
|
||||
'temperature': temperature,
|
||||
}
|
||||
```
|
||||
|
||||
**Key Changes:**
|
||||
- Remove `MODEL_CONFIG` lookup
|
||||
- Require `account` parameter (no optional)
|
||||
- Raise `ValueError` if IntegrationSettings not found
|
||||
- Raise `ValueError` if model not configured
|
||||
- No fallbacks to hardcoded defaults
|
||||
|
||||
---
|
||||
|
||||
### 2. `ai_core.py` - Default Model Fallback
|
||||
|
||||
**Current Issues:**
|
||||
- Lines 84-90: `_default_model` initialized from Django settings
|
||||
- Line 159: Falls back to `self._default_model` if model not provided
|
||||
|
||||
**Changes Required:**
|
||||
|
||||
#### Remove `_default_model` Initialization (Lines 84-90)
|
||||
```python
|
||||
# REMOVE THIS:
|
||||
try:
|
||||
from django.conf import settings
|
||||
self._default_model = getattr(settings, 'DEFAULT_AI_MODEL', 'gpt-4.1')
|
||||
except:
|
||||
self._default_model = 'gpt-4.1'
|
||||
```
|
||||
|
||||
#### Update `run_ai_request()` Method (Line 159)
|
||||
|
||||
**Current Logic:**
|
||||
```python
|
||||
active_model = model or self._default_model # ❌ Fallback
|
||||
```
|
||||
|
||||
**New Logic:**
|
||||
```python
|
||||
if not model:
|
||||
raise ValueError("Model is required. Ensure IntegrationSettings is configured for the account.")
|
||||
active_model = model
|
||||
```
|
||||
|
||||
**Key Changes:**
|
||||
- Remove `_default_model` attribute
|
||||
- Require `model` parameter (no fallback)
|
||||
- Raise `ValueError` if model not provided
|
||||
|
||||
---
|
||||
|
||||
### 3. `engine.py` - Model Configuration Call
|
||||
|
||||
**Current Issues:**
|
||||
- Line 206: Calls `get_model_config(function_name, account=self.account)`
|
||||
- May pass `account=None` in some cases
|
||||
|
||||
**Changes Required:**
|
||||
|
||||
#### Ensure Account is Always Passed (Line 206)
|
||||
```python
|
||||
# Current:
|
||||
model_config = get_model_config(function_name, account=self.account)
|
||||
|
||||
# New (same, but ensure account is not None):
|
||||
if not self.account:
|
||||
raise ValueError("Account is required for AI function execution")
|
||||
|
||||
model_config = get_model_config(function_name, account=self.account)
|
||||
```
|
||||
|
||||
**Key Changes:**
|
||||
- Validate `self.account` exists before calling `get_model_config()`
|
||||
- Raise clear error if account missing
|
||||
|
||||
---
|
||||
|
||||
### 4. `tasks.py` - Task Entry Point
|
||||
|
||||
**Current Issues:**
|
||||
- May not always have account context
|
||||
|
||||
**Changes Required:**
|
||||
|
||||
#### Ensure Account is Always Available (Line ~50)
|
||||
```python
|
||||
# In run_ai_task() function, ensure account is always set:
|
||||
if not account_id:
|
||||
raise ValueError("account_id is required for AI task execution")
|
||||
|
||||
account = Account.objects.get(id=account_id)
|
||||
```
|
||||
|
||||
**Key Changes:**
|
||||
- Validate `account_id` is provided
|
||||
- Ensure account exists before proceeding
|
||||
|
||||
---
|
||||
|
||||
## Orphan Code Cleanup
|
||||
|
||||
Based on the audit in `ORPHAN-CODE-AUDIT.md`, the following orphan code should be removed:
|
||||
|
||||
### 5. `__init__.py` - Remove Orphan Exports
|
||||
|
||||
**Current Issues:**
|
||||
- Lines 69-71: Export `get_model`, `get_max_tokens`, `get_temperature` - never imported/used
|
||||
- Line 44: Export `register_function` - never called (only `register_lazy_function` is used)
|
||||
- Line 46: Export `list_functions` - never called
|
||||
|
||||
**Changes Required:**
|
||||
|
||||
#### Remove Orphan Exports (Lines 44, 46, 69-71)
|
||||
```python
|
||||
# REMOVE FROM __all__:
|
||||
'register_function', # Line 44 - never called
|
||||
'list_functions', # Line 46 - never called
|
||||
'get_model', # Line 69 - never imported
|
||||
'get_max_tokens', # Line 70 - never imported
|
||||
'get_temperature', # Line 71 - never imported
|
||||
```
|
||||
|
||||
#### Remove Orphan Imports (Lines 29-35)
|
||||
```python
|
||||
# REMOVE THESE IMPORTS:
|
||||
from igny8_core.ai.settings import (
|
||||
MODEL_CONFIG, # Will be removed in Step 2
|
||||
get_model_config, # Keep - actively used
|
||||
get_model, # ❌ Remove - never used
|
||||
get_max_tokens, # ❌ Remove - never used
|
||||
get_temperature, # ❌ Remove - never used
|
||||
)
|
||||
```
|
||||
|
||||
**Key Changes:**
|
||||
- Remove unused exports from `__all__`
|
||||
- Remove unused imports
|
||||
- Keep `get_model_config` (actively used)
|
||||
- Keep `register_function` and `list_functions` in registry.py (may be useful for debugging), but don't export them
|
||||
|
||||
---
|
||||
|
||||
### 6. `settings.py` - Remove Unused Helper Functions
|
||||
|
||||
**Current Issues:**
|
||||
- Lines 106-109: `get_model()` function - never called
|
||||
- Lines 112-115: `get_max_tokens()` function - never called
|
||||
- Lines 118-121: `get_temperature()` function - never called
|
||||
|
||||
**Changes Required:**
|
||||
|
||||
#### Remove Unused Helper Functions (Lines 106-121)
|
||||
```python
|
||||
# REMOVE THESE FUNCTIONS:
|
||||
def get_model(function_name: str) -> str:
|
||||
"""Get model name for function"""
|
||||
config = get_model_config(function_name)
|
||||
return config.get("model", "gpt-4.1")
|
||||
|
||||
def get_max_tokens(function_name: str) -> int:
|
||||
"""Get max tokens for function"""
|
||||
config = get_model_config(function_name)
|
||||
return config.get("max_tokens", 4000)
|
||||
|
||||
def get_temperature(function_name: str) -> float:
|
||||
"""Get temperature for function"""
|
||||
config = get_model_config(function_name)
|
||||
return config.get("temperature", 0.7)
|
||||
```
|
||||
|
||||
**Key Changes:**
|
||||
- These functions are redundant - `get_model_config()` already returns all needed values
|
||||
- After removing MODEL_CONFIG, these functions become even less useful
|
||||
- Code should call `get_model_config()` directly and extract values from the returned dict
|
||||
|
||||
---
|
||||
|
||||
### 7. `registry.py` - Keep Functions, Don't Export
|
||||
|
||||
**Current Issues:**
|
||||
- `register_function()` is exported but never called (only `register_lazy_function` is used)
|
||||
- `list_functions()` is exported but never called
|
||||
|
||||
**Changes Required:**
|
||||
|
||||
#### Keep Functions, Remove from Exports
|
||||
```python
|
||||
# KEEP THESE FUNCTIONS IN registry.py (may be useful for debugging/admin)
|
||||
# BUT REMOVE FROM __init__.py exports
|
||||
|
||||
# In registry.py - KEEP:
|
||||
def register_function(name: str, function_class: Type[BaseAIFunction]):
|
||||
"""Register an AI function"""
|
||||
# ... keep implementation
|
||||
|
||||
def list_functions() -> list:
|
||||
"""List all registered functions"""
|
||||
# ... keep implementation
|
||||
|
||||
# In __init__.py - REMOVE from exports:
|
||||
# 'register_function', # ❌ Remove
|
||||
# 'list_functions', # ❌ Remove
|
||||
```
|
||||
|
||||
**Key Changes:**
|
||||
- Keep functions in `registry.py` (may be useful for future direct registration or debugging)
|
||||
- Remove from `__init__.py` exports (not used anywhere)
|
||||
- If needed in future, can be imported directly from `registry.py`
|
||||
|
||||
---
|
||||
|
||||
## Error Handling Strategy
|
||||
|
||||
### New Error Messages
|
||||
|
||||
When IntegrationSettings not configured:
|
||||
```python
|
||||
ValueError: "OpenAI IntegrationSettings not configured for account {account_id}. Please configure OpenAI settings in the integration page."
|
||||
```
|
||||
|
||||
When model not set in IntegrationSettings:
|
||||
```python
|
||||
ValueError: "Model not configured in IntegrationSettings for account {account_id}. Please set 'model' in OpenAI integration settings."
|
||||
```
|
||||
|
||||
When account not provided:
|
||||
```python
|
||||
ValueError: "Account is required for model configuration"
|
||||
```
|
||||
|
||||
### Frontend Impact
|
||||
|
||||
The frontend should handle these errors gracefully:
|
||||
- Show user-friendly error message
|
||||
- Redirect to integration settings page
|
||||
- Provide clear instructions on how to configure
|
||||
|
||||
---
|
||||
|
||||
## Testing Plan
|
||||
|
||||
### Unit Tests
|
||||
|
||||
1. **Test `get_model_config()` with valid IntegrationSettings**
|
||||
- Should return model from IntegrationSettings
|
||||
- Should include max_tokens and temperature
|
||||
|
||||
2. **Test `get_model_config()` without IntegrationSettings**
|
||||
- Should raise ValueError with clear message
|
||||
|
||||
3. **Test `get_model_config()` without model in config**
|
||||
- Should raise ValueError with clear message
|
||||
|
||||
4. **Test `get_model_config()` without account**
|
||||
- Should raise ValueError
|
||||
|
||||
5. **Test `ai_core.run_ai_request()` without model**
|
||||
- Should raise ValueError
|
||||
|
||||
### Integration Tests
|
||||
|
||||
1. **Test each AI function with valid IntegrationSettings**
|
||||
- `auto_cluster` - should work
|
||||
- `generate_ideas` - should work
|
||||
- `generate_content` - should work
|
||||
- `generate_image_prompts` - should work
|
||||
- `generate_images` - should work
|
||||
|
||||
2. **Test each AI function without IntegrationSettings**
|
||||
- Should raise clear error
|
||||
- Error should be user-friendly
|
||||
|
||||
3. **Test from view endpoints**
|
||||
- Should return proper error response
|
||||
- Should include request_id
|
||||
- Should follow unified API format
|
||||
|
||||
---
|
||||
|
||||
## Migration Steps
|
||||
|
||||
### Step 1: Backup Current Code
|
||||
- Create git branch: `refactor/remove-model-fallbacks`
|
||||
- Commit current state
|
||||
|
||||
### Step 2: Update `settings.py`
|
||||
- Remove `MODEL_CONFIG` dict
|
||||
- Update `get_model_config()` function
|
||||
- Add proper error handling
|
||||
|
||||
### Step 3: Update `ai_core.py`
|
||||
- Remove `_default_model` initialization
|
||||
- Update `run_ai_request()` to require model
|
||||
- Add error handling
|
||||
|
||||
### Step 4: Update `engine.py`
|
||||
- Validate account before calling `get_model_config()`
|
||||
- Add error handling
|
||||
|
||||
### Step 5: Update `tasks.py`
|
||||
- Validate account_id is provided
|
||||
- Add error handling
|
||||
|
||||
### Step 6: Clean Up Orphan Code
|
||||
- Remove orphan exports from `__init__.py`
|
||||
- Remove unused helper functions from `settings.py`
|
||||
- Remove unused imports from `__init__.py`
|
||||
- Verify no broken imports
|
||||
|
||||
### Step 7: Update Tests
|
||||
- Add unit tests for new error cases
|
||||
- Update integration tests
|
||||
- Test all 5 active functions
|
||||
|
||||
### Step 8: Test Manually
|
||||
- Test each AI function with valid IntegrationSettings
|
||||
- Test each AI function without IntegrationSettings
|
||||
- Verify error messages are clear
|
||||
|
||||
### Step 9: Update Documentation
|
||||
- Update AI framework docs
|
||||
- Document new error handling
|
||||
- Update API documentation
|
||||
|
||||
### Step 10: Deploy
|
||||
- Deploy to staging
|
||||
- Test in staging environment
|
||||
- Deploy to production
|
||||
|
||||
---
|
||||
|
||||
## Rollback Plan
|
||||
|
||||
If issues arise:
|
||||
|
||||
1. **Immediate Rollback**
|
||||
- Revert git branch
|
||||
- Restore previous version
|
||||
- Monitor for issues
|
||||
|
||||
2. **Partial Rollback**
|
||||
- Keep IntegrationSettings changes
|
||||
- Restore MODEL_CONFIG as fallback temporarily
|
||||
- Fix issues incrementally
|
||||
|
||||
---
|
||||
|
||||
## Success Criteria
|
||||
|
||||
- [ ] All 5 active AI functions work with IntegrationSettings only
|
||||
- [ ] Clear error messages when IntegrationSettings not configured
|
||||
- [ ] No hardcoded model defaults remain
|
||||
- [ ] No Django settings fallbacks remain
|
||||
- [ ] Orphan code removed (orphan exports, unused functions)
|
||||
- [ ] No broken imports after cleanup
|
||||
- [ ] All tests pass
|
||||
- [ ] Documentation updated
|
||||
- [ ] Frontend handles errors gracefully
|
||||
|
||||
---
|
||||
|
||||
## Notes
|
||||
|
||||
- **IntegrationSettings Structure:** Each account must have `IntegrationSettings` with `integration_type='openai'` and `config` containing `model`, `max_tokens`, `temperature`
|
||||
- **Account-Specific:** Each account configures their own models - no global defaults
|
||||
- **Error Clarity:** All errors must be user-friendly and actionable
|
||||
- **No Breaking Changes:** This is a refactoring, not a breaking change - existing accounts with IntegrationSettings will continue to work
|
||||
|
||||
---
|
||||
|
||||
**Last Updated:** 2025-01-XX
|
||||
**Status:** Ready for Implementation
|
||||
|
||||
@@ -3,7 +3,7 @@ IGNY8 AI Framework
|
||||
Unified framework for all AI functions with consistent lifecycle, progress tracking, and logging.
|
||||
"""
|
||||
|
||||
from igny8_core.ai.registry import register_function, get_function, list_functions
|
||||
from igny8_core.ai.registry import get_function
|
||||
from igny8_core.ai.engine import AIEngine
|
||||
from igny8_core.ai.base import BaseAIFunction
|
||||
from igny8_core.ai.ai_core import AICore
|
||||
@@ -27,11 +27,7 @@ from igny8_core.ai.constants import (
|
||||
)
|
||||
from igny8_core.ai.prompts import PromptRegistry, get_prompt
|
||||
from igny8_core.ai.settings import (
|
||||
MODEL_CONFIG,
|
||||
get_model_config,
|
||||
get_model,
|
||||
get_max_tokens,
|
||||
get_temperature,
|
||||
)
|
||||
|
||||
# Don't auto-import functions here - let apps.py handle it lazily
|
||||
@@ -41,9 +37,7 @@ __all__ = [
|
||||
'AIEngine',
|
||||
'BaseAIFunction',
|
||||
'AICore',
|
||||
'register_function',
|
||||
'get_function',
|
||||
'list_functions',
|
||||
# Validators
|
||||
'validate_ids',
|
||||
'validate_keywords_exist',
|
||||
@@ -64,10 +58,6 @@ __all__ = [
|
||||
'PromptRegistry',
|
||||
'get_prompt',
|
||||
# Settings
|
||||
'MODEL_CONFIG',
|
||||
'get_model_config',
|
||||
'get_model',
|
||||
'get_max_tokens',
|
||||
'get_temperature',
|
||||
]
|
||||
|
||||
|
||||
@@ -40,7 +40,6 @@ class AICore:
|
||||
self.account = account
|
||||
self._openai_api_key = None
|
||||
self._runware_api_key = None
|
||||
self._default_model = None
|
||||
self._load_account_settings()
|
||||
|
||||
def _load_account_settings(self):
|
||||
@@ -57,18 +56,6 @@ class AICore:
|
||||
).first()
|
||||
if openai_settings and openai_settings.config:
|
||||
self._openai_api_key = openai_settings.config.get('apiKey')
|
||||
model = openai_settings.config.get('model')
|
||||
if model:
|
||||
if model in MODEL_RATES:
|
||||
self._default_model = model
|
||||
logger.info(f"Loaded model '{model}' from IntegrationSettings for account {self.account.id}")
|
||||
else:
|
||||
error_msg = f"Model '{model}' from IntegrationSettings is not in supported models list. Supported models: {list(MODEL_RATES.keys())}"
|
||||
logger.error(f"[AICore] {error_msg}")
|
||||
logger.error(f"[AICore] Account {self.account.id} has invalid model configuration. Please update Integration Settings.")
|
||||
# Don't set _default_model, will fall back to Django settings
|
||||
else:
|
||||
logger.warning(f"No model configured in IntegrationSettings for account {self.account.id}, will use fallback")
|
||||
|
||||
# Load Runware settings
|
||||
runware_settings = IntegrationSettings.objects.filter(
|
||||
@@ -81,13 +68,11 @@ class AICore:
|
||||
except Exception as e:
|
||||
logger.warning(f"Could not load account settings: {e}", exc_info=True)
|
||||
|
||||
# Fallback to Django settings
|
||||
# Fallback to Django settings for API keys only (no model fallback)
|
||||
if not self._openai_api_key:
|
||||
self._openai_api_key = getattr(settings, 'OPENAI_API_KEY', None)
|
||||
if not self._runware_api_key:
|
||||
self._runware_api_key = getattr(settings, 'RUNWARE_API_KEY', None)
|
||||
if not self._default_model:
|
||||
self._default_model = getattr(settings, 'DEFAULT_AI_MODEL', DEFAULT_AI_MODEL)
|
||||
|
||||
def get_api_key(self, integration_type: str = 'openai') -> Optional[str]:
|
||||
"""Get API key for integration type"""
|
||||
@@ -98,15 +83,20 @@ class AICore:
|
||||
return None
|
||||
|
||||
def get_model(self, integration_type: str = 'openai') -> str:
|
||||
"""Get model for integration type"""
|
||||
if integration_type == 'openai':
|
||||
return self._default_model
|
||||
return DEFAULT_AI_MODEL
|
||||
"""
|
||||
Get model for integration type.
|
||||
DEPRECATED: Model should be passed directly to run_ai_request().
|
||||
This method is kept for backward compatibility but raises an error.
|
||||
"""
|
||||
raise ValueError(
|
||||
"get_model() is deprecated. Model must be passed directly to run_ai_request(). "
|
||||
"Use get_model_config() from settings.py to get model from IntegrationSettings."
|
||||
)
|
||||
|
||||
def run_ai_request(
|
||||
self,
|
||||
prompt: str,
|
||||
model: Optional[str] = None,
|
||||
model: str,
|
||||
max_tokens: int = 4000,
|
||||
temperature: float = 0.7,
|
||||
response_format: Optional[Dict] = None,
|
||||
@@ -121,7 +111,7 @@ class AICore:
|
||||
|
||||
Args:
|
||||
prompt: Prompt text
|
||||
model: Model name (defaults to account's default)
|
||||
model: Model name (required - must be provided from IntegrationSettings)
|
||||
max_tokens: Maximum tokens
|
||||
temperature: Temperature (0-1)
|
||||
response_format: Optional response format dict (for JSON mode)
|
||||
@@ -132,6 +122,9 @@ class AICore:
|
||||
Returns:
|
||||
Dict with 'content', 'input_tokens', 'output_tokens', 'total_tokens',
|
||||
'model', 'cost', 'error', 'api_id'
|
||||
|
||||
Raises:
|
||||
ValueError: If model is not provided
|
||||
"""
|
||||
# Use provided tracker or create a new one
|
||||
if tracker is None:
|
||||
@@ -139,39 +132,11 @@ class AICore:
|
||||
|
||||
tracker.ai_call("Preparing request...")
|
||||
|
||||
# Step 1: Validate API key
|
||||
api_key = api_key or self._openai_api_key
|
||||
if not api_key:
|
||||
error_msg = 'OpenAI API key not configured'
|
||||
# Step 1: Validate model is provided
|
||||
if not model:
|
||||
error_msg = "Model is required. Ensure IntegrationSettings is configured for the account."
|
||||
tracker.error('ConfigurationError', error_msg)
|
||||
return {
|
||||
'content': None,
|
||||
'error': error_msg,
|
||||
'input_tokens': 0,
|
||||
'output_tokens': 0,
|
||||
'total_tokens': 0,
|
||||
'model': model or self._default_model,
|
||||
'cost': 0.0,
|
||||
'api_id': None,
|
||||
}
|
||||
|
||||
# Step 2: Determine model
|
||||
active_model = model or self._default_model
|
||||
|
||||
# Debug logging: Show model from settings vs model used
|
||||
model_from_settings = self._default_model
|
||||
model_used = active_model
|
||||
logger.info(f"[AICore] Model Configuration Debug:")
|
||||
logger.info(f" - Model from IntegrationSettings: {model_from_settings}")
|
||||
logger.info(f" - Model parameter passed: {model}")
|
||||
logger.info(f" - Model actually used in request: {model_used}")
|
||||
tracker.ai_call(f"Model Debug - Settings: {model_from_settings}, Parameter: {model}, Using: {model_used}")
|
||||
|
||||
# Validate model is available and supported
|
||||
if not active_model:
|
||||
error_msg = 'No AI model configured. Please configure a model in Integration Settings or Django settings.'
|
||||
logger.error(f"[AICore] {error_msg}")
|
||||
tracker.error('ConfigurationError', error_msg)
|
||||
return {
|
||||
'content': None,
|
||||
'error': error_msg,
|
||||
@@ -183,6 +148,31 @@ class AICore:
|
||||
'api_id': None,
|
||||
}
|
||||
|
||||
# Step 2: Validate API key
|
||||
api_key = api_key or self._openai_api_key
|
||||
if not api_key:
|
||||
error_msg = 'OpenAI API key not configured'
|
||||
tracker.error('ConfigurationError', error_msg)
|
||||
return {
|
||||
'content': None,
|
||||
'error': error_msg,
|
||||
'input_tokens': 0,
|
||||
'output_tokens': 0,
|
||||
'total_tokens': 0,
|
||||
'model': model,
|
||||
'cost': 0.0,
|
||||
'api_id': None,
|
||||
}
|
||||
|
||||
# Step 3: Use provided model (no fallback)
|
||||
active_model = model
|
||||
|
||||
# Debug logging: Show model used
|
||||
logger.info(f"[AICore] Model Configuration:")
|
||||
logger.info(f" - Model parameter passed: {model}")
|
||||
logger.info(f" - Model used in request: {active_model}")
|
||||
tracker.ai_call(f"Using model: {active_model}")
|
||||
|
||||
if active_model not in MODEL_RATES:
|
||||
error_msg = f"Model '{active_model}' is not supported. Supported models: {list(MODEL_RATES.keys())}"
|
||||
logger.error(f"[AICore] {error_msg}")
|
||||
|
||||
@@ -193,6 +193,12 @@ class AIEngine:
|
||||
self.tracker.update("PREP", 25, prep_message, meta=self.step_tracker.get_meta())
|
||||
|
||||
# Phase 3: AI_CALL - Provider API Call (25-70%)
|
||||
# Validate account exists before proceeding
|
||||
if not self.account:
|
||||
error_msg = "Account is required for AI function execution"
|
||||
logger.error(f"[AIEngine] {error_msg}")
|
||||
return self._handle_error(error_msg, fn)
|
||||
|
||||
ai_core = AICore(account=self.account)
|
||||
function_name = fn.get_name()
|
||||
|
||||
@@ -201,29 +207,23 @@ class AIEngine:
|
||||
function_id_base = function_name.replace('_', '-')
|
||||
function_id = f"ai-{function_id_base}-01-desktop"
|
||||
|
||||
# Get model config from settings (Stage 4 requirement)
|
||||
# Pass account to read model from IntegrationSettings
|
||||
model_config = get_model_config(function_name, account=self.account)
|
||||
model = model_config.get('model')
|
||||
|
||||
# Read model straight from IntegrationSettings for visibility
|
||||
model_from_integration = None
|
||||
if self.account:
|
||||
try:
|
||||
from igny8_core.modules.system.models import IntegrationSettings
|
||||
openai_settings = IntegrationSettings.objects.filter(
|
||||
integration_type='openai',
|
||||
account=self.account,
|
||||
is_active=True
|
||||
).first()
|
||||
if openai_settings and openai_settings.config:
|
||||
model_from_integration = openai_settings.config.get('model')
|
||||
except Exception as integration_error:
|
||||
logger.warning(
|
||||
"[AIEngine] Unable to read model from IntegrationSettings: %s",
|
||||
integration_error,
|
||||
exc_info=True,
|
||||
)
|
||||
# Get model config from settings (requires account)
|
||||
# This will raise ValueError if IntegrationSettings not configured
|
||||
try:
|
||||
model_config = get_model_config(function_name, account=self.account)
|
||||
model = model_config.get('model')
|
||||
except ValueError as e:
|
||||
# IntegrationSettings not configured or model missing
|
||||
error_msg = str(e)
|
||||
error_type = 'ConfigurationError'
|
||||
logger.error(f"[AIEngine] {error_msg}")
|
||||
return self._handle_error(error_msg, fn, error_type=error_type)
|
||||
except Exception as e:
|
||||
# Other unexpected errors
|
||||
error_msg = f"Failed to get model configuration: {str(e)}"
|
||||
error_type = type(e).__name__
|
||||
logger.error(f"[AIEngine] {error_msg}", exc_info=True)
|
||||
return self._handle_error(error_msg, fn, error_type=error_type)
|
||||
|
||||
# Debug logging: Show model configuration (console only, not in step tracker)
|
||||
logger.info(f"[AIEngine] Model Configuration for {function_name}:")
|
||||
@@ -375,18 +375,28 @@ class AIEngine:
|
||||
}
|
||||
|
||||
except Exception as e:
|
||||
logger.error(f"Error in AIEngine.execute for {function_name}: {str(e)}", exc_info=True)
|
||||
return self._handle_error(str(e), fn, exc_info=True)
|
||||
error_msg = str(e)
|
||||
error_type = type(e).__name__
|
||||
logger.error(f"Error in AIEngine.execute for {function_name}: {error_msg}", exc_info=True)
|
||||
return self._handle_error(error_msg, fn, exc_info=True, error_type=error_type)
|
||||
|
||||
def _handle_error(self, error: str, fn: BaseAIFunction = None, exc_info=False):
|
||||
def _handle_error(self, error: str, fn: BaseAIFunction = None, exc_info=False, error_type: str = None):
|
||||
"""Centralized error handling"""
|
||||
function_name = fn.get_name() if fn else 'unknown'
|
||||
|
||||
# Determine error type
|
||||
if error_type:
|
||||
final_error_type = error_type
|
||||
elif isinstance(error, Exception):
|
||||
final_error_type = type(error).__name__
|
||||
else:
|
||||
final_error_type = 'Error'
|
||||
|
||||
self.step_tracker.add_request_step("Error", "error", error, error=error)
|
||||
|
||||
error_meta = {
|
||||
'error': error,
|
||||
'error_type': type(error).__name__ if isinstance(error, Exception) else 'Error',
|
||||
'error_type': final_error_type,
|
||||
**self.step_tracker.get_meta()
|
||||
}
|
||||
self.tracker.error(error, meta=error_meta)
|
||||
@@ -401,7 +411,7 @@ class AIEngine:
|
||||
return {
|
||||
'success': False,
|
||||
'error': error,
|
||||
'error_type': type(error).__name__ if isinstance(error, Exception) else 'Error',
|
||||
'error_type': final_error_type,
|
||||
'request_steps': self.step_tracker.request_steps,
|
||||
'response_steps': self.step_tracker.response_steps
|
||||
}
|
||||
|
||||
@@ -122,8 +122,10 @@ class GenerateImagesFunction(BaseAIFunction):
|
||||
}
|
||||
)
|
||||
|
||||
# Get model config
|
||||
model_config = get_model_config('extract_image_prompts')
|
||||
# Get model config (requires account)
|
||||
if not account_obj:
|
||||
raise ValueError("Account is required for model configuration")
|
||||
model_config = get_model_config('extract_image_prompts', account=account_obj)
|
||||
|
||||
# Call AI to extract prompts using centralized request handler
|
||||
result = ai_core.run_ai_request(
|
||||
|
||||
@@ -1,46 +1,11 @@
|
||||
"""
|
||||
AI Settings - Centralized model configurations and limits
|
||||
Uses IntegrationSettings only - no hardcoded defaults or fallbacks.
|
||||
"""
|
||||
from typing import Dict, Any
|
||||
import logging
|
||||
|
||||
# Model configurations for each AI function
|
||||
MODEL_CONFIG = {
|
||||
"auto_cluster": {
|
||||
"model": "gpt-4o-mini",
|
||||
"max_tokens": 3000,
|
||||
"temperature": 0.7,
|
||||
"response_format": {"type": "json_object"}, # Auto-enabled for JSON mode models
|
||||
},
|
||||
"generate_ideas": {
|
||||
"model": "gpt-4.1",
|
||||
"max_tokens": 4000,
|
||||
"temperature": 0.7,
|
||||
"response_format": {"type": "json_object"}, # JSON output
|
||||
},
|
||||
"generate_content": {
|
||||
"model": "gpt-4.1",
|
||||
"max_tokens": 8000,
|
||||
"temperature": 0.7,
|
||||
"response_format": {"type": "json_object"}, # JSON output
|
||||
},
|
||||
"generate_images": {
|
||||
"model": "dall-e-3",
|
||||
"size": "1024x1024",
|
||||
"provider": "openai",
|
||||
},
|
||||
"extract_image_prompts": {
|
||||
"model": "gpt-4o-mini",
|
||||
"max_tokens": 1000,
|
||||
"temperature": 0.7,
|
||||
"response_format": {"type": "json_object"},
|
||||
},
|
||||
"generate_image_prompts": {
|
||||
"model": "gpt-4o-mini",
|
||||
"max_tokens": 2000,
|
||||
"temperature": 0.7,
|
||||
"response_format": {"type": "json_object"},
|
||||
},
|
||||
}
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
# Function name aliases (for backward compatibility)
|
||||
FUNCTION_ALIASES = {
|
||||
@@ -52,71 +17,81 @@ FUNCTION_ALIASES = {
|
||||
}
|
||||
|
||||
|
||||
def get_model_config(function_name: str, account=None) -> Dict[str, Any]:
|
||||
def get_model_config(function_name: str, account) -> Dict[str, Any]:
|
||||
"""
|
||||
Get model configuration for an AI function.
|
||||
Reads model from IntegrationSettings if account is provided, otherwise uses defaults.
|
||||
Get model configuration from IntegrationSettings only.
|
||||
No fallbacks - account must have IntegrationSettings configured.
|
||||
|
||||
Args:
|
||||
function_name: AI function name (e.g., 'auto_cluster', 'generate_ideas')
|
||||
account: Optional account object to read model from IntegrationSettings
|
||||
function_name: Name of the AI function
|
||||
account: Account instance (required)
|
||||
|
||||
Returns:
|
||||
Dict with model, max_tokens, temperature, etc.
|
||||
dict: Model configuration with 'model', 'max_tokens', 'temperature'
|
||||
|
||||
Raises:
|
||||
ValueError: If account not provided or IntegrationSettings not configured
|
||||
"""
|
||||
# Check aliases first
|
||||
if not account:
|
||||
raise ValueError("Account is required for model configuration")
|
||||
|
||||
# Resolve function alias
|
||||
actual_name = FUNCTION_ALIASES.get(function_name, function_name)
|
||||
|
||||
# Get base config
|
||||
config = MODEL_CONFIG.get(actual_name, {}).copy()
|
||||
# Get IntegrationSettings for OpenAI
|
||||
try:
|
||||
from igny8_core.modules.system.models import IntegrationSettings
|
||||
integration_settings = IntegrationSettings.objects.get(
|
||||
integration_type='openai',
|
||||
account=account,
|
||||
is_active=True
|
||||
)
|
||||
except IntegrationSettings.DoesNotExist:
|
||||
raise ValueError(
|
||||
f"OpenAI IntegrationSettings not configured for account {account.id}. "
|
||||
f"Please configure OpenAI settings in the integration page."
|
||||
)
|
||||
|
||||
# Try to get model from IntegrationSettings if account is provided
|
||||
model_from_settings = None
|
||||
if account:
|
||||
try:
|
||||
from igny8_core.modules.system.models import IntegrationSettings
|
||||
openai_settings = IntegrationSettings.objects.filter(
|
||||
integration_type='openai',
|
||||
account=account,
|
||||
is_active=True
|
||||
).first()
|
||||
if openai_settings and openai_settings.config:
|
||||
model_from_settings = openai_settings.config.get('model')
|
||||
if model_from_settings:
|
||||
# Validate model is in our supported list
|
||||
from igny8_core.utils.ai_processor import MODEL_RATES
|
||||
if model_from_settings in MODEL_RATES:
|
||||
config['model'] = model_from_settings
|
||||
except Exception as e:
|
||||
import logging
|
||||
logger = logging.getLogger(__name__)
|
||||
logger.warning(f"Could not load model from IntegrationSettings: {e}", exc_info=True)
|
||||
config = integration_settings.config or {}
|
||||
|
||||
# Merge with defaults
|
||||
default_config = {
|
||||
"model": "gpt-4.1",
|
||||
"max_tokens": 4000,
|
||||
"temperature": 0.7,
|
||||
"response_format": None,
|
||||
# Get model from config
|
||||
model = config.get('model')
|
||||
if not model:
|
||||
raise ValueError(
|
||||
f"Model not configured in IntegrationSettings for account {account.id}. "
|
||||
f"Please set 'model' in OpenAI integration settings."
|
||||
)
|
||||
|
||||
# Validate model is in our supported list (optional validation)
|
||||
try:
|
||||
from igny8_core.utils.ai_processor import MODEL_RATES
|
||||
if model not in MODEL_RATES:
|
||||
logger.warning(
|
||||
f"Model '{model}' for account {account.id} is not in supported list. "
|
||||
f"Supported models: {list(MODEL_RATES.keys())}"
|
||||
)
|
||||
except ImportError:
|
||||
# MODEL_RATES not available - skip validation
|
||||
pass
|
||||
|
||||
# Get max_tokens and temperature from config (with reasonable defaults for API)
|
||||
max_tokens = config.get('max_tokens', 4000) # Reasonable default for API limits
|
||||
temperature = config.get('temperature', 0.7) # Reasonable default
|
||||
|
||||
# Build response format based on model (JSON mode for supported models)
|
||||
response_format = None
|
||||
try:
|
||||
from igny8_core.ai.constants import JSON_MODE_MODELS
|
||||
if model in JSON_MODE_MODELS:
|
||||
response_format = {"type": "json_object"}
|
||||
except ImportError:
|
||||
# JSON_MODE_MODELS not available - skip
|
||||
pass
|
||||
|
||||
return {
|
||||
'model': model,
|
||||
'max_tokens': max_tokens,
|
||||
'temperature': temperature,
|
||||
'response_format': response_format,
|
||||
}
|
||||
|
||||
return {**default_config, **config}
|
||||
|
||||
|
||||
def get_model(function_name: str) -> str:
|
||||
"""Get model name for function"""
|
||||
config = get_model_config(function_name)
|
||||
return config.get("model", "gpt-4.1")
|
||||
|
||||
|
||||
def get_max_tokens(function_name: str) -> int:
|
||||
"""Get max tokens for function"""
|
||||
config = get_model_config(function_name)
|
||||
return config.get("max_tokens", 4000)
|
||||
|
||||
|
||||
def get_temperature(function_name: str) -> float:
|
||||
"""Get temperature for function"""
|
||||
config = get_model_config(function_name)
|
||||
return config.get("temperature", 0.7)
|
||||
|
||||
|
||||
@@ -10,7 +10,7 @@ logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
@shared_task(bind=True, max_retries=3)
|
||||
def run_ai_task(self, function_name: str, payload: dict, account_id: int = None):
|
||||
def run_ai_task(self, function_name: str, payload: dict, account_id: int):
|
||||
"""
|
||||
Single Celery entrypoint for all AI functions.
|
||||
Dynamically loads and executes the requested function.
|
||||
@@ -18,7 +18,10 @@ def run_ai_task(self, function_name: str, payload: dict, account_id: int = None)
|
||||
Args:
|
||||
function_name: Name of the AI function (e.g., 'auto_cluster')
|
||||
payload: Function-specific payload
|
||||
account_id: Account ID for account isolation
|
||||
account_id: Account ID for account isolation (required)
|
||||
|
||||
Raises:
|
||||
Returns error dict if account_id not provided or account not found
|
||||
"""
|
||||
logger.info("=" * 80)
|
||||
logger.info(f"run_ai_task STARTED: {function_name}")
|
||||
@@ -29,14 +32,28 @@ def run_ai_task(self, function_name: str, payload: dict, account_id: int = None)
|
||||
logger.info("=" * 80)
|
||||
|
||||
try:
|
||||
# Get account
|
||||
account = None
|
||||
if account_id:
|
||||
from igny8_core.auth.models import Account
|
||||
try:
|
||||
account = Account.objects.get(id=account_id)
|
||||
except Account.DoesNotExist:
|
||||
logger.warning(f"Account {account_id} not found")
|
||||
# Validate account_id is provided
|
||||
if not account_id:
|
||||
error_msg = "account_id is required for AI task execution"
|
||||
logger.error(f"[run_ai_task] {error_msg}")
|
||||
return {
|
||||
'success': False,
|
||||
'error': error_msg,
|
||||
'error_type': 'ConfigurationError'
|
||||
}
|
||||
|
||||
# Get account and validate it exists
|
||||
from igny8_core.auth.models import Account
|
||||
try:
|
||||
account = Account.objects.get(id=account_id)
|
||||
except Account.DoesNotExist:
|
||||
error_msg = f"Account {account_id} not found"
|
||||
logger.error(f"[run_ai_task] {error_msg}")
|
||||
return {
|
||||
'success': False,
|
||||
'error': error_msg,
|
||||
'error_type': 'AccountNotFound'
|
||||
}
|
||||
|
||||
# Get function from registry
|
||||
fn = get_function_instance(function_name)
|
||||
|
||||
232
backend/igny8_core/api/tests/test_ai_framework.py
Normal file
232
backend/igny8_core/api/tests/test_ai_framework.py
Normal file
@@ -0,0 +1,232 @@
|
||||
"""
|
||||
Unit tests for AI framework
|
||||
Tests get_model_config() and AICore.run_ai_request() functions
|
||||
"""
|
||||
from django.test import TestCase
|
||||
from igny8_core.auth.models import Account, User, Plan
|
||||
from igny8_core.modules.system.models import IntegrationSettings
|
||||
from igny8_core.ai.settings import get_model_config
|
||||
from igny8_core.ai.ai_core import AICore
|
||||
|
||||
|
||||
class GetModelConfigTestCase(TestCase):
|
||||
"""Test cases for get_model_config() function"""
|
||||
|
||||
def setUp(self):
|
||||
"""Set up test data"""
|
||||
# Create plan first
|
||||
self.plan = Plan.objects.create(
|
||||
name="Test Plan",
|
||||
slug="test-plan",
|
||||
price=0,
|
||||
credits_per_month=1000
|
||||
)
|
||||
|
||||
# Create user first (Account needs owner)
|
||||
self.user = User.objects.create_user(
|
||||
username='testuser',
|
||||
email='test@test.com',
|
||||
password='testpass123',
|
||||
role='owner'
|
||||
)
|
||||
|
||||
# Create account with owner
|
||||
self.account = Account.objects.create(
|
||||
name='Test Account',
|
||||
slug='test-account',
|
||||
plan=self.plan,
|
||||
owner=self.user,
|
||||
status='active'
|
||||
)
|
||||
|
||||
# Update user to have account
|
||||
self.user.account = self.account
|
||||
self.user.save()
|
||||
|
||||
def test_get_model_config_with_valid_settings(self):
|
||||
"""Test get_model_config() with valid IntegrationSettings"""
|
||||
IntegrationSettings.objects.create(
|
||||
integration_type='openai',
|
||||
account=self.account,
|
||||
is_active=True,
|
||||
config={
|
||||
'model': 'gpt-4o',
|
||||
'max_tokens': 4000,
|
||||
'temperature': 0.7,
|
||||
'apiKey': 'test-key'
|
||||
}
|
||||
)
|
||||
|
||||
config = get_model_config('auto_cluster', self.account)
|
||||
|
||||
self.assertEqual(config['model'], 'gpt-4o')
|
||||
self.assertEqual(config['max_tokens'], 4000)
|
||||
self.assertEqual(config['temperature'], 0.7)
|
||||
self.assertIn('response_format', config)
|
||||
|
||||
def test_get_model_config_without_account(self):
|
||||
"""Test get_model_config() without account - should raise ValueError"""
|
||||
with self.assertRaises(ValueError) as context:
|
||||
get_model_config('auto_cluster', None)
|
||||
|
||||
self.assertIn('Account is required', str(context.exception))
|
||||
|
||||
def test_get_model_config_without_integration_settings(self):
|
||||
"""Test get_model_config() without IntegrationSettings - should raise ValueError"""
|
||||
with self.assertRaises(ValueError) as context:
|
||||
get_model_config('auto_cluster', self.account)
|
||||
|
||||
self.assertIn('OpenAI IntegrationSettings not configured', str(context.exception))
|
||||
self.assertIn(str(self.account.id), str(context.exception))
|
||||
|
||||
def test_get_model_config_without_model_in_config(self):
|
||||
"""Test get_model_config() without model in config - should raise ValueError"""
|
||||
IntegrationSettings.objects.create(
|
||||
integration_type='openai',
|
||||
account=self.account,
|
||||
is_active=True,
|
||||
config={
|
||||
'max_tokens': 4000,
|
||||
'temperature': 0.7,
|
||||
'apiKey': 'test-key'
|
||||
# No 'model' key
|
||||
}
|
||||
)
|
||||
|
||||
with self.assertRaises(ValueError) as context:
|
||||
get_model_config('auto_cluster', self.account)
|
||||
|
||||
self.assertIn('Model not configured in IntegrationSettings', str(context.exception))
|
||||
self.assertIn(str(self.account.id), str(context.exception))
|
||||
|
||||
def test_get_model_config_with_inactive_settings(self):
|
||||
"""Test get_model_config() with inactive IntegrationSettings - should raise ValueError"""
|
||||
IntegrationSettings.objects.create(
|
||||
integration_type='openai',
|
||||
account=self.account,
|
||||
is_active=False,
|
||||
config={
|
||||
'model': 'gpt-4o',
|
||||
'max_tokens': 4000,
|
||||
'temperature': 0.7
|
||||
}
|
||||
)
|
||||
|
||||
with self.assertRaises(ValueError) as context:
|
||||
get_model_config('auto_cluster', self.account)
|
||||
|
||||
self.assertIn('OpenAI IntegrationSettings not configured', str(context.exception))
|
||||
|
||||
def test_get_model_config_with_function_alias(self):
|
||||
"""Test get_model_config() with function alias"""
|
||||
IntegrationSettings.objects.create(
|
||||
integration_type='openai',
|
||||
account=self.account,
|
||||
is_active=True,
|
||||
config={
|
||||
'model': 'gpt-4o-mini',
|
||||
'max_tokens': 2000,
|
||||
'temperature': 0.5
|
||||
}
|
||||
)
|
||||
|
||||
# Test with alias
|
||||
config1 = get_model_config('cluster_keywords', self.account)
|
||||
config2 = get_model_config('auto_cluster', self.account)
|
||||
|
||||
# Both should return the same config
|
||||
self.assertEqual(config1['model'], config2['model'])
|
||||
self.assertEqual(config1['model'], 'gpt-4o-mini')
|
||||
|
||||
def test_get_model_config_json_mode_models(self):
|
||||
"""Test get_model_config() sets response_format for JSON mode models"""
|
||||
json_models = ['gpt-4o', 'gpt-4o-mini', 'gpt-4-turbo-preview']
|
||||
|
||||
for model in json_models:
|
||||
IntegrationSettings.objects.filter(account=self.account).delete()
|
||||
IntegrationSettings.objects.create(
|
||||
integration_type='openai',
|
||||
account=self.account,
|
||||
is_active=True,
|
||||
config={
|
||||
'model': model,
|
||||
'max_tokens': 4000,
|
||||
'temperature': 0.7
|
||||
}
|
||||
)
|
||||
|
||||
config = get_model_config('auto_cluster', self.account)
|
||||
self.assertIn('response_format', config)
|
||||
self.assertEqual(config['response_format'], {'type': 'json_object'})
|
||||
|
||||
|
||||
class AICoreTestCase(TestCase):
|
||||
"""Test cases for AICore.run_ai_request() function"""
|
||||
|
||||
def setUp(self):
|
||||
"""Set up test data"""
|
||||
# Create plan first
|
||||
self.plan = Plan.objects.create(
|
||||
name="Test Plan",
|
||||
slug="test-plan",
|
||||
price=0,
|
||||
credits_per_month=1000
|
||||
)
|
||||
|
||||
# Create user first (Account needs owner)
|
||||
self.user = User.objects.create_user(
|
||||
username='testuser',
|
||||
email='test@test.com',
|
||||
password='testpass123',
|
||||
role='owner'
|
||||
)
|
||||
|
||||
# Create account with owner
|
||||
self.account = Account.objects.create(
|
||||
name='Test Account',
|
||||
slug='test-account',
|
||||
plan=self.plan,
|
||||
owner=self.user,
|
||||
status='active'
|
||||
)
|
||||
|
||||
# Update user to have account
|
||||
self.user.account = self.account
|
||||
self.user.save()
|
||||
|
||||
self.ai_core = AICore(account=self.account)
|
||||
|
||||
def test_run_ai_request_without_model(self):
|
||||
"""Test run_ai_request() without model - should return error dict"""
|
||||
result = self.ai_core.run_ai_request(
|
||||
prompt="Test prompt",
|
||||
model=None,
|
||||
function_name='test_function'
|
||||
)
|
||||
|
||||
self.assertIn('error', result)
|
||||
self.assertIn('Model is required', result['error'])
|
||||
self.assertEqual(result['content'], None)
|
||||
self.assertEqual(result['total_tokens'], 0)
|
||||
|
||||
def test_run_ai_request_with_empty_model(self):
|
||||
"""Test run_ai_request() with empty model string - should return error dict"""
|
||||
result = self.ai_core.run_ai_request(
|
||||
prompt="Test prompt",
|
||||
model="",
|
||||
function_name='test_function'
|
||||
)
|
||||
|
||||
self.assertIn('error', result)
|
||||
self.assertIn('Model is required', result['error'])
|
||||
self.assertEqual(result['content'], None)
|
||||
self.assertEqual(result['total_tokens'], 0)
|
||||
|
||||
def test_get_model_deprecated(self):
|
||||
"""Test get_model() method is deprecated and raises ValueError"""
|
||||
with self.assertRaises(ValueError) as context:
|
||||
self.ai_core.get_model('openai')
|
||||
|
||||
self.assertIn('deprecated', str(context.exception).lower())
|
||||
self.assertIn('run_ai_request', str(context.exception))
|
||||
|
||||
@@ -663,17 +663,47 @@ class ImagesViewSet(SiteSectorModelViewSet):
|
||||
|
||||
account = getattr(request, 'account', None)
|
||||
|
||||
# Get site_id and sector_id from query parameters
|
||||
site_id = request.query_params.get('site_id')
|
||||
sector_id = request.query_params.get('sector_id')
|
||||
|
||||
# Get all content that has images (either directly or via task)
|
||||
# First, get content with direct image links
|
||||
queryset = Content.objects.filter(images__isnull=False)
|
||||
if account:
|
||||
queryset = queryset.filter(account=account)
|
||||
|
||||
# Apply site/sector filtering if provided
|
||||
if site_id:
|
||||
try:
|
||||
queryset = queryset.filter(site_id=int(site_id))
|
||||
except (ValueError, TypeError):
|
||||
pass
|
||||
|
||||
if sector_id:
|
||||
try:
|
||||
queryset = queryset.filter(sector_id=int(sector_id))
|
||||
except (ValueError, TypeError):
|
||||
pass
|
||||
|
||||
# Also get content from images linked via task
|
||||
task_linked_images = Images.objects.filter(task__isnull=False, content__isnull=True)
|
||||
if account:
|
||||
task_linked_images = task_linked_images.filter(account=account)
|
||||
|
||||
# Apply site/sector filtering to task-linked images
|
||||
if site_id:
|
||||
try:
|
||||
task_linked_images = task_linked_images.filter(site_id=int(site_id))
|
||||
except (ValueError, TypeError):
|
||||
pass
|
||||
|
||||
if sector_id:
|
||||
try:
|
||||
task_linked_images = task_linked_images.filter(sector_id=int(sector_id))
|
||||
except (ValueError, TypeError):
|
||||
pass
|
||||
|
||||
# Get content IDs from task-linked images
|
||||
task_content_ids = set()
|
||||
for image in task_linked_images:
|
||||
@@ -694,6 +724,7 @@ class ImagesViewSet(SiteSectorModelViewSet):
|
||||
for content_id in content_ids:
|
||||
try:
|
||||
content = Content.objects.get(id=content_id)
|
||||
|
||||
# Get images linked directly to content OR via task
|
||||
content_images = Images.objects.filter(
|
||||
Q(content=content) | Q(task=content.task)
|
||||
|
||||
Reference in New Issue
Block a user