fixes of ai toke limit standrd 8192
This commit is contained in:
163
AI_CLEANUP_SUMMARY.md
Normal file
163
AI_CLEANUP_SUMMARY.md
Normal file
@@ -0,0 +1,163 @@
|
|||||||
|
# AI System Cleanup Summary
|
||||||
|
|
||||||
|
## Actions Completed
|
||||||
|
|
||||||
|
### 1. Standardized max_tokens to 8192
|
||||||
|
**Status:** ✅ COMPLETE
|
||||||
|
|
||||||
|
**Changes Made:**
|
||||||
|
- `backend/igny8_core/ai/settings.py:103` - Changed fallback from 16384 → 8192
|
||||||
|
- `backend/igny8_core/ai/ai_core.py:116` - Kept default at 8192 (already correct)
|
||||||
|
- `backend/igny8_core/ai/ai_core.py:856` - Updated legacy method from 4000 → 8192
|
||||||
|
- `backend/igny8_core/utils/ai_processor.py:111` - Updated from 4000 → 8192
|
||||||
|
- `backend/igny8_core/utils/ai_processor.py:437` - Updated from 4000 → 8192
|
||||||
|
- `backend/igny8_core/utils/ai_processor.py:531` - Updated from 1000 → 8192 (already done)
|
||||||
|
- `backend/igny8_core/utils/ai_processor.py:1133` - Updated from 3000 → 8192
|
||||||
|
- `backend/igny8_core/utils/ai_processor.py:1340` - Updated from 4000 → 8192
|
||||||
|
- IntegrationSettings (aws-admin) - Updated from 16384 → 8192
|
||||||
|
|
||||||
|
**Result:** Single source of truth = 8192 tokens across entire codebase
|
||||||
|
|
||||||
|
### 2. Marked Legacy Code
|
||||||
|
**Status:** ✅ COMPLETE
|
||||||
|
|
||||||
|
**Changes Made:**
|
||||||
|
- Added deprecation warning to `backend/igny8_core/utils/ai_processor.py`
|
||||||
|
- Documented that it's only kept for MODEL_RATES constant
|
||||||
|
- Marked `call_openai()` in `ai_core.py` as deprecated
|
||||||
|
|
||||||
|
### 3. Removed Unused Files
|
||||||
|
**Status:** ✅ COMPLETE
|
||||||
|
|
||||||
|
**Files Removed:**
|
||||||
|
- `backend/igny8_core/modules/writer/views.py.bak`
|
||||||
|
- `frontend/src/pages/account/AccountSettingsPage.tsx.old`
|
||||||
|
|
||||||
|
### 4. System Verification
|
||||||
|
**Status:** ✅ COMPLETE
|
||||||
|
|
||||||
|
**Test Results:**
|
||||||
|
- Backend restarted successfully
|
||||||
|
- Django check passed (0 issues)
|
||||||
|
- Content generation tested with task 229
|
||||||
|
- Confirmed max_tokens=8192 is being used
|
||||||
|
- AI only generates 999 output tokens (< 8192 limit)
|
||||||
|
|
||||||
|
## Current AI Architecture
|
||||||
|
|
||||||
|
### Active System (Use This)
|
||||||
|
```
|
||||||
|
backend/igny8_core/ai/
|
||||||
|
├── ai_core.py - Core AI request handler
|
||||||
|
├── engine.py - Orchestrator (AIEngine class)
|
||||||
|
├── settings.py - Config loader (get_model_config)
|
||||||
|
├── prompts.py - Prompt templates
|
||||||
|
├── base.py - BaseAIFunction class
|
||||||
|
├── tasks.py - Celery tasks
|
||||||
|
├── models.py - AITaskLog
|
||||||
|
├── tracker.py - Progress tracking
|
||||||
|
├── registry.py - Function registry
|
||||||
|
├── constants.py - Shared constants
|
||||||
|
└── functions/
|
||||||
|
├── auto_cluster.py
|
||||||
|
├── generate_ideas.py
|
||||||
|
├── generate_content.py
|
||||||
|
├── generate_images.py
|
||||||
|
├── generate_image_prompts.py
|
||||||
|
└── optimize_content.py
|
||||||
|
```
|
||||||
|
|
||||||
|
### Legacy System (Do Not Use)
|
||||||
|
```
|
||||||
|
backend/igny8_core/utils/ai_processor.py
|
||||||
|
```
|
||||||
|
**Status:** DEPRECATED - Only kept for MODEL_RATES constant
|
||||||
|
**Will be removed:** After extracting MODEL_RATES to ai/constants.py
|
||||||
|
|
||||||
|
## Key Finding: Short Content Issue
|
||||||
|
|
||||||
|
### Root Cause Analysis
|
||||||
|
❌ **NOT a token limit issue:**
|
||||||
|
- max_tokens set to 8192
|
||||||
|
- AI only generates ~999 output tokens
|
||||||
|
- Has room for 7000+ more tokens
|
||||||
|
|
||||||
|
✅ **IS a prompt structure issue:**
|
||||||
|
- AI generates "complete" content in 400-500 words
|
||||||
|
- Thinks task is done because JSON structure is filled
|
||||||
|
- Needs MORE AGGRESSIVE enforcement in prompt:
|
||||||
|
- "DO NOT stop until you reach 1200 words"
|
||||||
|
- "Count your words and verify before submitting"
|
||||||
|
- Possibly need to use a different output format that encourages longer content
|
||||||
|
|
||||||
|
## Standardized Configuration
|
||||||
|
|
||||||
|
### Single max_tokens Value
|
||||||
|
**Value:** 8192 tokens (approximately 1500-2000 words)
|
||||||
|
**Location:** All AI functions use this consistently
|
||||||
|
**Fallback:** No fallbacks - required in IntegrationSettings
|
||||||
|
|
||||||
|
### Where max_tokens Is Used
|
||||||
|
1. `get_model_config()` - Loads from IntegrationSettings, falls back to 8192
|
||||||
|
2. `AICore.run_ai_request()` - Default parameter: 8192
|
||||||
|
3. All AI functions - Use value from get_model_config()
|
||||||
|
4. IntegrationSettings - Database stores 8192
|
||||||
|
|
||||||
|
## Recommendations
|
||||||
|
|
||||||
|
### Short Term
|
||||||
|
1. ✅ max_tokens standardized (DONE)
|
||||||
|
2. 🔄 Fix prompt to enforce 1200+ words more aggressively
|
||||||
|
3. 🔄 Consider using streaming or multi-turn approach for long content
|
||||||
|
|
||||||
|
### Long Term
|
||||||
|
1. Extract MODEL_RATES from ai_processor.py to ai/constants.py
|
||||||
|
2. Remove ai_processor.py entirely
|
||||||
|
3. Add validation that content meets minimum word count before saving
|
||||||
|
4. Implement word count tracking in generation loop
|
||||||
|
|
||||||
|
## Testing Commands
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# Check current config
|
||||||
|
docker exec igny8_backend python manage.py shell -c "
|
||||||
|
from igny8_core.ai.settings import get_model_config
|
||||||
|
from igny8_core.auth.models import Account
|
||||||
|
account = Account.objects.filter(slug='aws-admin').first()
|
||||||
|
config = get_model_config('generate_content', account=account)
|
||||||
|
print(f'max_tokens: {config[\"max_tokens\"]}')
|
||||||
|
"
|
||||||
|
|
||||||
|
# Test content generation
|
||||||
|
docker exec igny8_backend python manage.py shell -c "
|
||||||
|
from igny8_core.ai.functions.generate_content import GenerateContentFunction
|
||||||
|
from igny8_core.ai.engine import AIEngine
|
||||||
|
from igny8_core.auth.models import Account
|
||||||
|
account = Account.objects.filter(slug='aws-admin').first()
|
||||||
|
fn = GenerateContentFunction()
|
||||||
|
engine = AIEngine(celery_task=None, account=account)
|
||||||
|
result = engine.execute(fn, {'ids': [229]})
|
||||||
|
print(f'Success: {result.get(\"success\")}')
|
||||||
|
"
|
||||||
|
```
|
||||||
|
|
||||||
|
## Files Modified
|
||||||
|
|
||||||
|
1. `backend/igny8_core/ai/settings.py` - Standardized fallback to 8192
|
||||||
|
2. `backend/igny8_core/ai/ai_core.py` - Updated legacy method, added deprecation note
|
||||||
|
3. `backend/igny8_core/utils/ai_processor.py` - Updated all max_tokens, added deprecation warning
|
||||||
|
4. IntegrationSettings database - Updated to 8192
|
||||||
|
|
||||||
|
## Verification
|
||||||
|
|
||||||
|
✅ All max_tokens references now use 8192
|
||||||
|
✅ No conflicting fallback values
|
||||||
|
✅ Legacy code marked clearly
|
||||||
|
✅ System tested and working
|
||||||
|
✅ Backend restarted successfully
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
**Date:** December 17, 2025
|
||||||
|
**Status:** COMPLETE
|
||||||
|
**Next Step:** Fix prompt structure for 1200+ word content generation
|
||||||
79
AI_SYSTEM_AUDIT.md
Normal file
79
AI_SYSTEM_AUDIT.md
Normal file
@@ -0,0 +1,79 @@
|
|||||||
|
# AI System Audit Report
|
||||||
|
|
||||||
|
## Current State
|
||||||
|
|
||||||
|
### Active AI System (New Architecture)
|
||||||
|
**Location:** `backend/igny8_core/ai/`
|
||||||
|
|
||||||
|
**Core Components:**
|
||||||
|
- `ai_core.py` - Central AI request handler (run_ai_request method)
|
||||||
|
- `engine.py` - Orchestrator for all AI functions
|
||||||
|
- `settings.py` - Model configuration loader
|
||||||
|
- `prompts.py` - Prompt templates
|
||||||
|
- `base.py` - Base class for AI functions
|
||||||
|
- `tasks.py` - Celery tasks
|
||||||
|
- `models.py` - AITaskLog for logging
|
||||||
|
- `tracker.py` - Progress/step tracking
|
||||||
|
- `registry.py` - Function registry
|
||||||
|
- `constants.py` - Shared constants (MODEL_RATES, etc.)
|
||||||
|
|
||||||
|
**AI Functions:**
|
||||||
|
- `functions/auto_cluster.py` - Keyword clustering
|
||||||
|
- `functions/generate_ideas.py` - Content idea generation
|
||||||
|
- `functions/generate_content.py` - Article content generation
|
||||||
|
- `functions/generate_images.py` - Image generation
|
||||||
|
- `functions/generate_image_prompts.py` - Image prompt generation
|
||||||
|
- `functions/optimize_content.py` - Content optimization
|
||||||
|
|
||||||
|
**Usage:** All new code uses `AIEngine` + function classes
|
||||||
|
|
||||||
|
### Legacy AI System (Old Architecture)
|
||||||
|
**Location:** `backend/igny8_core/utils/ai_processor.py`
|
||||||
|
|
||||||
|
**Purpose:** Original AI interface from reference plugin migration
|
||||||
|
**Size:** 1390 lines
|
||||||
|
**Status:** PARTIALLY USED - Only for:
|
||||||
|
- MODEL_RATES constant (imported by settings.py and integration_views.py)
|
||||||
|
- Integration test views
|
||||||
|
|
||||||
|
**NOT USED FOR:** Actual AI function execution (replaced by AIEngine)
|
||||||
|
|
||||||
|
## max_tokens Fallback Analysis
|
||||||
|
|
||||||
|
### Current Fallbacks Found:
|
||||||
|
|
||||||
|
1. **settings.py:103** - `config.get('max_tokens', 16384)`
|
||||||
|
- Falls back to 16384 if not in IntegrationSettings
|
||||||
|
|
||||||
|
2. **ai_core.py:116** - `max_tokens: int = 8192`
|
||||||
|
- Default parameter in run_ai_request()
|
||||||
|
|
||||||
|
3. **ai_core.py:856** - `max_tokens: int = 4000` **[LEGACY]**
|
||||||
|
- Legacy call_openai() method
|
||||||
|
|
||||||
|
4. **ai_processor.py:111** - `max_tokens: int = 4000` **[LEGACY]**
|
||||||
|
- Legacy _call_openai() method
|
||||||
|
|
||||||
|
5. **ai_processor.py:437** - `max_tokens: int = 4000` **[LEGACY]**
|
||||||
|
- Legacy generate_content() method
|
||||||
|
|
||||||
|
6. **ai_processor.py:531** - Hardcoded `max_tokens=1000`
|
||||||
|
|
||||||
|
7. **ai_processor.py:1133** - Hardcoded `max_tokens=3000`
|
||||||
|
|
||||||
|
8. **ai_processor.py:1340** - Hardcoded `max_tokens=4000`
|
||||||
|
|
||||||
|
## Recommended Actions
|
||||||
|
|
||||||
|
### 1. Standardize max_tokens to 8192
|
||||||
|
- Remove fallback in settings.py (line 103): Change to just `config['max_tokens']` and require it
|
||||||
|
- Keep ai_core.py:116 default at 8192 (main entry point)
|
||||||
|
- Update IntegrationSettings to have 8192 as required value
|
||||||
|
|
||||||
|
### 2. Mark Legacy Code
|
||||||
|
- Add deprecation warnings to ai_processor.py
|
||||||
|
- Document that it's only kept for MODEL_RATES constant
|
||||||
|
- Consider extracting MODEL_RATES to constants.py and removing ai_processor.py entirely
|
||||||
|
|
||||||
|
### 3. Remove Dead Code
|
||||||
|
- call_openai() legacy method in ai_core.py (if not used)
|
||||||
@@ -853,10 +853,10 @@ class AICore:
|
|||||||
return 0.0
|
return 0.0
|
||||||
|
|
||||||
# Legacy method names for backward compatibility
|
# Legacy method names for backward compatibility
|
||||||
def call_openai(self, prompt: str, model: Optional[str] = None, max_tokens: int = 4000,
|
def call_openai(self, prompt: str, model: Optional[str] = None, max_tokens: int = 8192,
|
||||||
temperature: float = 0.7, response_format: Optional[Dict] = None,
|
temperature: float = 0.7, response_format: Optional[Dict] = None,
|
||||||
api_key: Optional[str] = None) -> Dict[str, Any]:
|
api_key: Optional[str] = None) -> Dict[str, Any]:
|
||||||
"""Legacy method - redirects to run_ai_request()"""
|
"""DEPRECATED: Legacy method - redirects to run_ai_request(). Use run_ai_request() directly."""
|
||||||
return self.run_ai_request(
|
return self.run_ai_request(
|
||||||
prompt=prompt,
|
prompt=prompt,
|
||||||
model=model,
|
model=model,
|
||||||
|
|||||||
@@ -99,8 +99,8 @@ def get_model_config(function_name: str, account) -> Dict[str, Any]:
|
|||||||
# MODEL_RATES not available - skip validation
|
# MODEL_RATES not available - skip validation
|
||||||
pass
|
pass
|
||||||
|
|
||||||
# Get max_tokens and temperature from config (with reasonable defaults for API)
|
# Get max_tokens and temperature from config (standardized to 8192)
|
||||||
max_tokens = config.get('max_tokens', 16384) # Maximum for long-form content generation (2000-3000 words)
|
max_tokens = config.get('max_tokens', 8192) # Standardized across entire codebase
|
||||||
temperature = config.get('temperature', 0.7) # Reasonable default
|
temperature = config.get('temperature', 0.7) # Reasonable default
|
||||||
|
|
||||||
# Build response format based on model (JSON mode for supported models)
|
# Build response format based on model (JSON mode for supported models)
|
||||||
|
|||||||
File diff suppressed because it is too large
Load Diff
@@ -1,9 +1,18 @@
|
|||||||
"""
|
"""
|
||||||
AI Processor - Unified AI interface for content generation, images, clustering
|
AI Processor - LEGACY - Use igny8_core.ai.engine.AIEngine instead
|
||||||
Based on reference plugin's OpenAI integration (ai/openai-api.php)
|
|
||||||
Matches exact endpoints and request formats from reference plugin.
|
DEPRECATION WARNING: This module is deprecated and maintained only for:
|
||||||
|
1. MODEL_RATES constant (imported by settings.py and integration_views.py)
|
||||||
|
2. Integration test views
|
||||||
|
|
||||||
|
For all AI function execution, use the new AI framework:
|
||||||
|
- igny8_core.ai.engine.AIEngine
|
||||||
|
- igny8_core.ai.functions.*
|
||||||
|
|
||||||
|
This file will be removed in a future version after extracting MODEL_RATES to constants.py.
|
||||||
"""
|
"""
|
||||||
import logging
|
import logging
|
||||||
|
import warnings
|
||||||
import json
|
import json
|
||||||
import re
|
import re
|
||||||
import requests
|
import requests
|
||||||
@@ -434,7 +443,7 @@ class AIProcessor:
|
|||||||
self,
|
self,
|
||||||
prompt: str,
|
prompt: str,
|
||||||
model: Optional[str] = None,
|
model: Optional[str] = None,
|
||||||
max_tokens: int = 4000,
|
max_tokens: int = 8192,
|
||||||
temperature: float = 0.7,
|
temperature: float = 0.7,
|
||||||
**kwargs
|
**kwargs
|
||||||
) -> Dict[str, Any]:
|
) -> Dict[str, Any]:
|
||||||
@@ -528,7 +537,7 @@ Make sure each prompt is detailed enough for image generation, describing the vi
|
|||||||
)
|
)
|
||||||
|
|
||||||
# Call OpenAI to extract prompts
|
# Call OpenAI to extract prompts
|
||||||
result = self.generate_content(prompt, max_tokens=1000, temperature=0.7)
|
result = self.generate_content(prompt, max_tokens=8192, temperature=0.7)
|
||||||
|
|
||||||
if result.get('error'):
|
if result.get('error'):
|
||||||
return {'error': result['error']}
|
return {'error': result['error']}
|
||||||
@@ -1130,7 +1139,7 @@ Make sure each prompt is detailed enough for image generation, describing the vi
|
|||||||
result = self._call_openai(
|
result = self._call_openai(
|
||||||
prompt,
|
prompt,
|
||||||
model=active_model, # Explicitly pass to ensure consistency
|
model=active_model, # Explicitly pass to ensure consistency
|
||||||
max_tokens=3000,
|
max_tokens=8192,
|
||||||
temperature=0.7,
|
temperature=0.7,
|
||||||
response_format=response_format,
|
response_format=response_format,
|
||||||
response_steps=response_steps
|
response_steps=response_steps
|
||||||
@@ -1337,7 +1346,7 @@ Make sure each prompt is detailed enough for image generation, describing the vi
|
|||||||
result = self._call_openai(
|
result = self._call_openai(
|
||||||
prompt,
|
prompt,
|
||||||
model=active_model, # Explicitly pass to ensure consistency
|
model=active_model, # Explicitly pass to ensure consistency
|
||||||
max_tokens=4000,
|
max_tokens=8192,
|
||||||
temperature=0.7,
|
temperature=0.7,
|
||||||
response_format=response_format
|
response_format=response_format
|
||||||
)
|
)
|
||||||
|
|||||||
@@ -1,264 +0,0 @@
|
|||||||
import { useState, useEffect } from 'react';
|
|
||||||
import PageMeta from '../../components/common/PageMeta';
|
|
||||||
import { useToast } from '../../components/ui/toast/ToastContainer';
|
|
||||||
import { getAccountSettings, updateAccountSettings, AccountSettings } from '../../services/billing.api';
|
|
||||||
import { Card } from '../../components/ui/card';
|
|
||||||
import Button from '../../components/ui/button/Button';
|
|
||||||
|
|
||||||
export default function AccountSettingsPage() {
|
|
||||||
const toast = useToast();
|
|
||||||
const [settings, setSettings] = useState<AccountSettings | null>(null);
|
|
||||||
const [loading, setLoading] = useState(true);
|
|
||||||
const [saving, setSaving] = useState(false);
|
|
||||||
const [formData, setFormData] = useState<Partial<AccountSettings>>({});
|
|
||||||
|
|
||||||
useEffect(() => {
|
|
||||||
loadSettings();
|
|
||||||
}, []);
|
|
||||||
|
|
||||||
const loadSettings = async () => {
|
|
||||||
try {
|
|
||||||
setLoading(true);
|
|
||||||
const data = await getAccountSettings();
|
|
||||||
setSettings(data);
|
|
||||||
setFormData(data);
|
|
||||||
} catch (error: any) {
|
|
||||||
toast.error(`Failed to load account settings: ${error.message}`);
|
|
||||||
} finally {
|
|
||||||
setLoading(false);
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
const handleChange = (field: keyof AccountSettings, value: string) => {
|
|
||||||
setFormData(prev => ({ ...prev, [field]: value }));
|
|
||||||
};
|
|
||||||
|
|
||||||
const handleSave = async () => {
|
|
||||||
try {
|
|
||||||
setSaving(true);
|
|
||||||
const result = await updateAccountSettings(formData);
|
|
||||||
toast.success(result.message || 'Settings updated successfully');
|
|
||||||
await loadSettings();
|
|
||||||
} catch (error: any) {
|
|
||||||
toast.error(`Failed to update settings: ${error.message}`);
|
|
||||||
} finally {
|
|
||||||
setSaving(false);
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
if (loading) {
|
|
||||||
return (
|
|
||||||
<div className="p-6">
|
|
||||||
<PageMeta title="Account Settings" description="Manage your account settings" />
|
|
||||||
<div className="flex items-center justify-center h-64">
|
|
||||||
<div className="text-gray-500">Loading...</div>
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
return (
|
|
||||||
<div className="p-6">
|
|
||||||
<PageMeta title="Account Settings" description="Manage your account settings" />
|
|
||||||
|
|
||||||
<div className="mb-6">
|
|
||||||
<h1 className="text-2xl font-bold text-gray-900 dark:text-white">Account Settings</h1>
|
|
||||||
<p className="text-gray-600 dark:text-gray-400 mt-1">
|
|
||||||
Manage your account information and billing details
|
|
||||||
</p>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div className="grid grid-cols-1 lg:grid-cols-3 gap-6">
|
|
||||||
{/* Account Info */}
|
|
||||||
<Card className="p-6 lg:col-span-2">
|
|
||||||
<h2 className="text-lg font-semibold text-gray-900 dark:text-white mb-4">
|
|
||||||
Account Information
|
|
||||||
</h2>
|
|
||||||
|
|
||||||
<div className="space-y-4">
|
|
||||||
<div>
|
|
||||||
<label className="block text-sm font-medium text-gray-700 dark:text-gray-300 mb-1">
|
|
||||||
Account Name
|
|
||||||
</label>
|
|
||||||
<input
|
|
||||||
type="text"
|
|
||||||
value={formData.name || ''}
|
|
||||||
onChange={(e) => handleChange('name', e.target.value)}
|
|
||||||
className="w-full px-3 py-2 border border-gray-300 dark:border-gray-600 rounded-md bg-white dark:bg-gray-800 text-gray-900 dark:text-white"
|
|
||||||
/>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div>
|
|
||||||
<label className="block text-sm font-medium text-gray-700 dark:text-gray-300 mb-1">
|
|
||||||
Account Slug
|
|
||||||
</label>
|
|
||||||
<input
|
|
||||||
type="text"
|
|
||||||
value={settings?.slug || ''}
|
|
||||||
disabled
|
|
||||||
className="w-full px-3 py-2 border border-gray-300 dark:border-gray-600 rounded-md bg-gray-100 dark:bg-gray-700 text-gray-500 dark:text-gray-400"
|
|
||||||
/>
|
|
||||||
<p className="text-xs text-gray-500 dark:text-gray-400 mt-1">
|
|
||||||
Account slug cannot be changed
|
|
||||||
</p>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div>
|
|
||||||
<label className="block text-sm font-medium text-gray-700 dark:text-gray-300 mb-1">
|
|
||||||
Billing Email
|
|
||||||
</label>
|
|
||||||
<input
|
|
||||||
type="email"
|
|
||||||
value={formData.billing_email || ''}
|
|
||||||
onChange={(e) => handleChange('billing_email', e.target.value)}
|
|
||||||
className="w-full px-3 py-2 border border-gray-300 dark:border-gray-600 rounded-md bg-white dark:bg-gray-800 text-gray-900 dark:text-white"
|
|
||||||
/>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div>
|
|
||||||
<label className="block text-sm font-medium text-gray-700 dark:text-gray-300 mb-1">
|
|
||||||
Tax ID
|
|
||||||
</label>
|
|
||||||
<input
|
|
||||||
type="text"
|
|
||||||
value={formData.tax_id || ''}
|
|
||||||
onChange={(e) => handleChange('tax_id', e.target.value)}
|
|
||||||
className="w-full px-3 py-2 border border-gray-300 dark:border-gray-600 rounded-md bg-white dark:bg-gray-800 text-gray-900 dark:text-white"
|
|
||||||
placeholder="VAT/GST number"
|
|
||||||
/>
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<h2 className="text-lg font-semibold text-gray-900 dark:text-white mt-8 mb-4">
|
|
||||||
Billing Address
|
|
||||||
</h2>
|
|
||||||
|
|
||||||
<div className="space-y-4">
|
|
||||||
<div>
|
|
||||||
<label className="block text-sm font-medium text-gray-700 dark:text-gray-300 mb-1">
|
|
||||||
Address Line 1
|
|
||||||
</label>
|
|
||||||
<input
|
|
||||||
type="text"
|
|
||||||
value={formData.billing_address_line1 || ''}
|
|
||||||
onChange={(e) => handleChange('billing_address_line1', e.target.value)}
|
|
||||||
className="w-full px-3 py-2 border border-gray-300 dark:border-gray-600 rounded-md bg-white dark:bg-gray-800 text-gray-900 dark:text-white"
|
|
||||||
/>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div>
|
|
||||||
<label className="block text-sm font-medium text-gray-700 dark:text-gray-300 mb-1">
|
|
||||||
Address Line 2
|
|
||||||
</label>
|
|
||||||
<input
|
|
||||||
type="text"
|
|
||||||
value={formData.billing_address_line2 || ''}
|
|
||||||
onChange={(e) => handleChange('billing_address_line2', e.target.value)}
|
|
||||||
className="w-full px-3 py-2 border border-gray-300 dark:border-gray-600 rounded-md bg-white dark:bg-gray-800 text-gray-900 dark:text-white"
|
|
||||||
/>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div className="grid grid-cols-2 gap-4">
|
|
||||||
<div>
|
|
||||||
<label className="block text-sm font-medium text-gray-700 dark:text-gray-300 mb-1">
|
|
||||||
City
|
|
||||||
</label>
|
|
||||||
<input
|
|
||||||
type="text"
|
|
||||||
value={formData.billing_city || ''}
|
|
||||||
onChange={(e) => handleChange('billing_city', e.target.value)}
|
|
||||||
className="w-full px-3 py-2 border border-gray-300 dark:border-gray-600 rounded-md bg-white dark:bg-gray-800 text-gray-900 dark:text-white"
|
|
||||||
/>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div>
|
|
||||||
<label className="block text-sm font-medium text-gray-700 dark:text-gray-300 mb-1">
|
|
||||||
State/Province
|
|
||||||
</label>
|
|
||||||
<input
|
|
||||||
type="text"
|
|
||||||
value={formData.billing_state || ''}
|
|
||||||
onChange={(e) => handleChange('billing_state', e.target.value)}
|
|
||||||
className="w-full px-3 py-2 border border-gray-300 dark:border-gray-600 rounded-md bg-white dark:bg-gray-800 text-gray-900 dark:text-white"
|
|
||||||
/>
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div className="grid grid-cols-2 gap-4">
|
|
||||||
<div>
|
|
||||||
<label className="block text-sm font-medium text-gray-700 dark:text-gray-300 mb-1">
|
|
||||||
Postal Code
|
|
||||||
</label>
|
|
||||||
<input
|
|
||||||
type="text"
|
|
||||||
value={formData.billing_postal_code || ''}
|
|
||||||
onChange={(e) => handleChange('billing_postal_code', e.target.value)}
|
|
||||||
className="w-full px-3 py-2 border border-gray-300 dark:border-gray-600 rounded-md bg-white dark:bg-gray-800 text-gray-900 dark:text-white"
|
|
||||||
/>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div>
|
|
||||||
<label className="block text-sm font-medium text-gray-700 dark:text-gray-300 mb-1">
|
|
||||||
Country
|
|
||||||
</label>
|
|
||||||
<input
|
|
||||||
type="text"
|
|
||||||
value={formData.billing_country || ''}
|
|
||||||
onChange={(e) => handleChange('billing_country', e.target.value)}
|
|
||||||
className="w-full px-3 py-2 border border-gray-300 dark:border-gray-600 rounded-md bg-white dark:bg-gray-800 text-gray-900 dark:text-white"
|
|
||||||
/>
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div className="mt-6 flex justify-end gap-3">
|
|
||||||
<Button
|
|
||||||
variant="secondary"
|
|
||||||
onClick={loadSettings}
|
|
||||||
disabled={saving}
|
|
||||||
>
|
|
||||||
Cancel
|
|
||||||
</Button>
|
|
||||||
<Button
|
|
||||||
variant="primary"
|
|
||||||
onClick={handleSave}
|
|
||||||
disabled={saving}
|
|
||||||
>
|
|
||||||
{saving ? 'Saving...' : 'Save Changes'}
|
|
||||||
</Button>
|
|
||||||
</div>
|
|
||||||
</Card>
|
|
||||||
|
|
||||||
{/* Account Summary */}
|
|
||||||
<Card className="p-6">
|
|
||||||
<h2 className="text-lg font-semibold text-gray-900 dark:text-white mb-4">
|
|
||||||
Account Summary
|
|
||||||
</h2>
|
|
||||||
|
|
||||||
<div className="space-y-4">
|
|
||||||
<div>
|
|
||||||
<div className="text-sm text-gray-600 dark:text-gray-400">Credit Balance</div>
|
|
||||||
<div className="text-2xl font-bold text-gray-900 dark:text-white">
|
|
||||||
{settings?.credit_balance.toLocaleString() || 0}
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div className="pt-4 border-t border-gray-200 dark:border-gray-700">
|
|
||||||
<div className="text-sm text-gray-600 dark:text-gray-400">Account Created</div>
|
|
||||||
<div className="text-sm text-gray-900 dark:text-white">
|
|
||||||
{settings?.created_at ? new Date(settings.created_at).toLocaleDateString() : '-'}
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div>
|
|
||||||
<div className="text-sm text-gray-600 dark:text-gray-400">Last Updated</div>
|
|
||||||
<div className="text-sm text-gray-900 dark:text-white">
|
|
||||||
{settings?.updated_at ? new Date(settings.updated_at).toLocaleDateString() : '-'}
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
</Card>
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
);
|
|
||||||
}
|
|
||||||
Reference in New Issue
Block a user