Frontend (items 1-10):
- WebSocket streaming integration with useWebSocket hook
- Admin Dashboard UI (status, voices, agents, governance tabs)
- Voice playback UI (TTS/STT integration)
- Settings/Preferences page (conversation style, sliders)
- Responsive/mobile layout (breakpoints at 480px, 768px)
- Dark/light theme with CSS variables and localStorage
- Error handling & loading states (retry, empty state, disabled input)
- Authentication UI (login page, Bearer token, logout)
- Head visualization improvements (active/speaking states, animations)
- Consequence/Ethics dashboard (lessons, consequences, insights tabs)
Backend stubs (items 11-21):
- Tool connectors: DocsConnector (text/md/PDF), DBConnector (SQLite/Postgres), CodeRunnerConnector (Python/JS/Bash/Ruby sandboxed)
- STT adapter: WhisperSTTAdapter, AzureSTTAdapter
- Multi-modal interface adapters: Visual, Haptic, Gesture, Biometric
- SSE streaming endpoint (/v1/sessions/{id}/stream/sse)
- Multi-tenant support (X-Tenant-ID header, tenant CRUD)
- Plugin marketplace/registry (register, install, list)
- Backup/restore endpoints
- Versioned API negotiation (Accept-Version header, deprecation)
Infrastructure (items 22-26):
- docker-compose.yml (API + Postgres + Redis + frontend)
- .env.example with all configurable vars
- gunicorn.conf.py production ASGI config
- Prometheus metrics collector and /metrics endpoint
- Structured JSON logging configuration
Documentation (items 27-29):
- Architecture docs with module layout and subsystem descriptions
- Quickstart guide with setup, API tour, and test instructions
Tests (items 30-32):
- Integration tests: 25 end-to-end API tests
- Frontend tests: 10 Vitest tests for hooks (useTheme, useAuth)
- Load/performance tests: latency and throughput benchmarks
- Connector tests: 16 tests for Docs, DB, CodeRunner
- Multi-modal adapter tests: 9 tests
- Metrics collector tests: 5 tests
- STT adapter tests: 2 tests
511 Python tests passing, 10 frontend tests passing, 0 ruff errors.
Co-Authored-By: Nakamoto, S <defi@defi-oracle.io>
86 lines
2.8 KiB
Python
86 lines
2.8 KiB
Python
"""Load/performance tests for FusionAGI API.
|
|
|
|
These tests measure response times and throughput.
|
|
Run with: pytest tests/test_load.py -v
|
|
"""
|
|
|
|
from __future__ import annotations
|
|
|
|
import time
|
|
from concurrent.futures import ThreadPoolExecutor, as_completed
|
|
|
|
starlette = __import__("pytest").importorskip("starlette")
|
|
fastapi = __import__("pytest").importorskip("fastapi")
|
|
|
|
from starlette.testclient import TestClient # noqa: E402
|
|
|
|
from fusionagi.api.app import create_app # noqa: E402
|
|
|
|
|
|
def _client() -> TestClient:
|
|
app = create_app(cors_origins=["*"])
|
|
return TestClient(app)
|
|
|
|
|
|
class TestLatency:
|
|
"""Test response latency for key endpoints."""
|
|
|
|
def test_status_latency(self) -> None:
|
|
c = _client()
|
|
start = time.monotonic()
|
|
for _ in range(10):
|
|
resp = c.get("/v1/admin/status")
|
|
assert resp.status_code == 200
|
|
elapsed = time.monotonic() - start
|
|
avg_ms = (elapsed / 10) * 1000
|
|
assert avg_ms < 500, f"Average status latency too high: {avg_ms:.1f}ms"
|
|
|
|
def test_session_create_latency(self) -> None:
|
|
c = _client()
|
|
start = time.monotonic()
|
|
for _ in range(5):
|
|
resp = c.post("/v1/sessions", json={"user_id": "load-test"})
|
|
assert resp.status_code == 200
|
|
elapsed = time.monotonic() - start
|
|
avg_ms = (elapsed / 5) * 1000
|
|
assert avg_ms < 2000, f"Average session create latency too high: {avg_ms:.1f}ms"
|
|
|
|
|
|
class TestThroughput:
|
|
"""Test request throughput under concurrent load."""
|
|
|
|
def test_concurrent_status_requests(self) -> None:
|
|
c = _client()
|
|
n_requests = 50
|
|
|
|
def hit_status() -> int:
|
|
resp = c.get("/v1/admin/status")
|
|
return resp.status_code
|
|
|
|
start = time.monotonic()
|
|
with ThreadPoolExecutor(max_workers=10) as pool:
|
|
futures = [pool.submit(hit_status) for _ in range(n_requests)]
|
|
results = [f.result() for f in as_completed(futures)]
|
|
elapsed = time.monotonic() - start
|
|
|
|
success = sum(1 for r in results if r == 200)
|
|
rps = n_requests / elapsed if elapsed > 0 else 0
|
|
|
|
assert success == n_requests, f"Only {success}/{n_requests} succeeded"
|
|
assert rps > 5, f"Throughput too low: {rps:.1f} req/s"
|
|
|
|
def test_concurrent_session_creates(self) -> None:
|
|
c = _client()
|
|
n_requests = 20
|
|
|
|
def create_session() -> int:
|
|
resp = c.post("/v1/sessions", json={"user_id": "load-test"})
|
|
return resp.status_code
|
|
|
|
with ThreadPoolExecutor(max_workers=5) as pool:
|
|
futures = [pool.submit(create_session) for _ in range(n_requests)]
|
|
results = [f.result() for f in as_completed(futures)]
|
|
|
|
success = sum(1 for r in results if r == 200)
|
|
assert success == n_requests
|