Files
PROJECT-CONTORL/backend/tests/test_backend_reliability.py
beabigegg 35c90fe76b feat: implement 5 QA-driven security and quality proposals
Implemented proposals from comprehensive QA review:

1. extend-csrf-protection
   - Add POST to CSRF protected methods in frontend
   - Global CSRF middleware for all state-changing operations
   - Update tests with CSRF token fixtures

2. tighten-cors-websocket-security
   - Replace wildcard CORS with explicit method/header lists
   - Disable query parameter auth in production (code 4002)
   - Add per-user WebSocket connection limit (max 5, code 4005)

3. shorten-jwt-expiry
   - Reduce JWT expiry from 7 days to 60 minutes
   - Add refresh token support with 7-day expiry
   - Implement token rotation on refresh
   - Frontend auto-refresh when token near expiry (<5 min)

4. fix-frontend-quality
   - Add React.lazy() code splitting for all pages
   - Fix useCallback dependency arrays (Dashboard, Comments)
   - Add localStorage data validation in AuthContext
   - Complete i18n for AttachmentUpload component

5. enhance-backend-validation
   - Add SecurityAuditMiddleware for access denied logging
   - Add ErrorSanitizerMiddleware for production error messages
   - Protect /health/detailed with admin authentication
   - Add input length validation (comment 5000, desc 10000)

All 521 backend tests passing. Frontend builds successfully.

Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
2026-01-12 23:19:05 +08:00

302 lines
10 KiB
Python

"""
Tests for backend reliability improvements.
Tests cover:
- Database connection pool behavior
- Redis disconnect and recovery
- Blocker deletion scenarios
"""
import os
os.environ["TESTING"] = "true"
import pytest
from unittest.mock import patch, MagicMock
from datetime import datetime
class TestDatabaseConnectionPool:
"""Test database connection pool behavior."""
def test_pool_handles_multiple_connections(self, client, admin_token, db):
"""Test that connection pool handles multiple concurrent requests."""
from app.models import Space
# Create test space
space = Space(id="pool-test-space", name="Pool Test", owner_id="00000000-0000-0000-0000-000000000001")
db.add(space)
db.commit()
# Make multiple concurrent requests
responses = []
for i in range(10):
response = client.get(
"/api/spaces",
headers={"Authorization": f"Bearer {admin_token}"}
)
responses.append(response)
# All should succeed
assert all(r.status_code == 200 for r in responses)
def test_pool_recovers_from_connection_error(self, client, admin_token, db):
"""Test that pool recovers after connection errors."""
from app.models import Space
space = Space(id="recovery-space", name="Recovery Test", owner_id="00000000-0000-0000-0000-000000000001")
db.add(space)
db.commit()
# First request should work
response1 = client.get(
"/api/spaces",
headers={"Authorization": f"Bearer {admin_token}"}
)
assert response1.status_code == 200
# Simulate and recover from error - subsequent request should still work
response2 = client.get(
"/api/spaces",
headers={"Authorization": f"Bearer {admin_token}"}
)
assert response2.status_code == 200
class TestRedisFailover:
"""Test Redis disconnect and recovery."""
def test_redis_publish_fallback_on_failure(self):
"""Test that Redis publish failures are handled gracefully."""
from app.core.redis import RedisManager
manager = RedisManager()
# Mock Redis failure
mock_redis = MagicMock()
mock_redis.publish.side_effect = Exception("Redis connection lost")
with patch.object(manager, 'get_client', return_value=mock_redis):
# Should not raise, should queue message
try:
manager.publish_with_fallback("test_channel", {"test": "message"})
except Exception:
pass # Some implementations may raise, that's ok for this test
def test_message_queue_on_redis_failure(self):
"""Test that messages are queued when Redis is unavailable."""
from app.core.redis import RedisManager
manager = RedisManager()
# If manager has queue functionality
if hasattr(manager, '_message_queue') or hasattr(manager, 'queue_message'):
initial_queue_size = len(getattr(manager, '_message_queue', []))
# Force failure and queue
with patch.object(manager, '_publish_direct', side_effect=Exception("Redis down")):
try:
manager.publish_with_fallback("channel", {"data": "test"})
except Exception:
pass
# Check if message was queued (implementation dependent)
# This is a best-effort test
def test_redis_reconnection(self, mock_redis):
"""Test that Redis reconnects after failure."""
# Simulate initial failure then success
call_count = [0]
original_get = mock_redis.get
def intermittent_failure(key):
call_count[0] += 1
if call_count[0] == 1:
raise Exception("Connection lost")
return original_get(key)
mock_redis.get = intermittent_failure
# First call fails
with pytest.raises(Exception):
mock_redis.get("test_key")
# Second call succeeds (reconnected)
result = mock_redis.get("test_key")
assert call_count[0] == 2
class TestBlockerDeletionCheck:
"""Test blocker check before task deletion."""
def test_delete_task_with_blockers_warning(self, client, admin_token, csrf_token, db):
"""Test that deleting task with blockers shows warning."""
from app.models import Space, Project, Task, TaskStatus, TaskDependency
# Create test data
space = Space(id="blocker-space", name="Blocker Test", owner_id="00000000-0000-0000-0000-000000000001")
db.add(space)
project = Project(id="blocker-project", name="Blocker Project", space_id="blocker-space", owner_id="00000000-0000-0000-0000-000000000001")
db.add(project)
status = TaskStatus(id="blocker-status", name="To Do", project_id="blocker-project", position=0)
db.add(status)
# Task to delete
blocker_task = Task(
id="blocker-task",
title="Blocker Task",
project_id="blocker-project",
status_id="blocker-status",
created_by="00000000-0000-0000-0000-000000000001"
)
db.add(blocker_task)
# Dependent task
dependent_task = Task(
id="dependent-task",
title="Dependent Task",
project_id="blocker-project",
status_id="blocker-status",
created_by="00000000-0000-0000-0000-000000000001"
)
db.add(dependent_task)
# Create dependency
dependency = TaskDependency(
task_id="dependent-task",
depends_on_task_id="blocker-task",
dependency_type="FS"
)
db.add(dependency)
db.commit()
# Try to delete without force
response = client.delete(
"/api/tasks/blocker-task",
headers={"Authorization": f"Bearer {admin_token}", "X-CSRF-Token": csrf_token}
)
# Should return warning or require confirmation
# Response could be 200 with warning, or 409/400 requiring force_delete
if response.status_code == 200:
data = response.json()
# Check if it's a warning response
if "warning" in data or "blocker_count" in data:
assert data.get("blocker_count", 0) >= 1 or "blocker" in str(data).lower()
def test_force_delete_resolves_blockers(self, client, admin_token, csrf_token, db):
"""Test that force delete resolves blockers."""
from app.models import Space, Project, Task, TaskStatus, TaskDependency
# Create test data
space = Space(id="force-del-space", name="Force Del Test", owner_id="00000000-0000-0000-0000-000000000001")
db.add(space)
project = Project(id="force-del-project", name="Force Del Project", space_id="force-del-space", owner_id="00000000-0000-0000-0000-000000000001")
db.add(project)
status = TaskStatus(id="force-del-status", name="To Do", project_id="force-del-project", position=0)
db.add(status)
# Task to delete
task_to_delete = Task(
id="force-del-task",
title="Task to Delete",
project_id="force-del-project",
status_id="force-del-status",
created_by="00000000-0000-0000-0000-000000000001"
)
db.add(task_to_delete)
# Dependent task
dependent = Task(
id="force-dependent",
title="Dependent",
project_id="force-del-project",
status_id="force-del-status",
created_by="00000000-0000-0000-0000-000000000001"
)
db.add(dependent)
# Create dependency
dep = TaskDependency(
task_id="force-dependent",
depends_on_task_id="force-del-task",
dependency_type="FS"
)
db.add(dep)
db.commit()
# Force delete
response = client.delete(
"/api/tasks/force-del-task?force_delete=true",
headers={"Authorization": f"Bearer {admin_token}", "X-CSRF-Token": csrf_token}
)
assert response.status_code == 200
# Verify task is deleted
db.refresh(task_to_delete)
assert task_to_delete.is_deleted is True
def test_delete_task_without_blockers(self, client, admin_token, csrf_token, db):
"""Test deleting task without blockers succeeds normally."""
from app.models import Space, Project, Task, TaskStatus
# Create test data
space = Space(id="no-blocker-space", name="No Blocker", owner_id="00000000-0000-0000-0000-000000000001")
db.add(space)
project = Project(id="no-blocker-project", name="No Blocker Project", space_id="no-blocker-space", owner_id="00000000-0000-0000-0000-000000000001")
db.add(project)
status = TaskStatus(id="no-blocker-status", name="To Do", project_id="no-blocker-project", position=0)
db.add(status)
task = Task(
id="no-blocker-task",
title="Task without blockers",
project_id="no-blocker-project",
status_id="no-blocker-status",
created_by="00000000-0000-0000-0000-000000000001"
)
db.add(task)
db.commit()
# Delete should succeed without warning
response = client.delete(
"/api/tasks/no-blocker-task",
headers={"Authorization": f"Bearer {admin_token}", "X-CSRF-Token": csrf_token}
)
assert response.status_code == 200
# Verify task is deleted
db.refresh(task)
assert task.is_deleted is True
class TestStorageValidation:
"""Test NAS/storage validation."""
def test_storage_path_validation_on_startup(self):
"""Test that storage path is validated on startup."""
from app.services.file_storage_service import FileStorageService
service = FileStorageService()
# Service should have validated upload directory
assert hasattr(service, 'upload_dir') or hasattr(service, '_upload_dir')
def test_storage_write_permission_check(self):
"""Test that storage write permissions are checked."""
from app.services.file_storage_service import FileStorageService
service = FileStorageService()
# Check if service has permission validation
if hasattr(service, 'check_permissions'):
result = service.check_permissions()
assert result is True or result is None # Should not raise