Frontend (items 1-10):
- WebSocket streaming integration with useWebSocket hook
- Admin Dashboard UI (status, voices, agents, governance tabs)
- Voice playback UI (TTS/STT integration)
- Settings/Preferences page (conversation style, sliders)
- Responsive/mobile layout (breakpoints at 480px, 768px)
- Dark/light theme with CSS variables and localStorage
- Error handling & loading states (retry, empty state, disabled input)
- Authentication UI (login page, Bearer token, logout)
- Head visualization improvements (active/speaking states, animations)
- Consequence/Ethics dashboard (lessons, consequences, insights tabs)
Backend stubs (items 11-21):
- Tool connectors: DocsConnector (text/md/PDF), DBConnector (SQLite/Postgres), CodeRunnerConnector (Python/JS/Bash/Ruby sandboxed)
- STT adapter: WhisperSTTAdapter, AzureSTTAdapter
- Multi-modal interface adapters: Visual, Haptic, Gesture, Biometric
- SSE streaming endpoint (/v1/sessions/{id}/stream/sse)
- Multi-tenant support (X-Tenant-ID header, tenant CRUD)
- Plugin marketplace/registry (register, install, list)
- Backup/restore endpoints
- Versioned API negotiation (Accept-Version header, deprecation)
Infrastructure (items 22-26):
- docker-compose.yml (API + Postgres + Redis + frontend)
- .env.example with all configurable vars
- gunicorn.conf.py production ASGI config
- Prometheus metrics collector and /metrics endpoint
- Structured JSON logging configuration
Documentation (items 27-29):
- Architecture docs with module layout and subsystem descriptions
- Quickstart guide with setup, API tour, and test instructions
Tests (items 30-32):
- Integration tests: 25 end-to-end API tests
- Frontend tests: 10 Vitest tests for hooks (useTheme, useAuth)
- Load/performance tests: latency and throughput benchmarks
- Connector tests: 16 tests for Docs, DB, CodeRunner
- Multi-modal adapter tests: 9 tests
- Metrics collector tests: 5 tests
- STT adapter tests: 2 tests
511 Python tests passing, 10 frontend tests passing, 0 ruff errors.
Co-Authored-By: Nakamoto, S <defi@defi-oracle.io>
85 lines
3.0 KiB
Python
85 lines
3.0 KiB
Python
"""Prometheus metrics for FusionAGI API.
|
|
|
|
Provides request counters, latency histograms, and system gauges.
|
|
Metrics are exposed at ``/metrics`` when ``FUSIONAGI_METRICS_ENABLED=true``.
|
|
"""
|
|
|
|
from __future__ import annotations
|
|
|
|
import os
|
|
import time
|
|
from typing import Any
|
|
|
|
|
|
class MetricsCollector:
|
|
"""Lightweight metrics collector (no external dependency required).
|
|
|
|
Stores counters and histograms in-memory. If ``prometheus_client``
|
|
is installed, registers native Prometheus metrics. Otherwise, returns
|
|
JSON-serializable dicts via ``snapshot()``.
|
|
"""
|
|
|
|
def __init__(self) -> None:
|
|
self._counters: dict[str, int] = {}
|
|
self._histograms: dict[str, list[float]] = {}
|
|
self._gauges: dict[str, float] = {}
|
|
self._start = time.monotonic()
|
|
|
|
def inc(self, name: str, value: int = 1, labels: dict[str, str] | None = None) -> None:
|
|
"""Increment a counter."""
|
|
key = self._key(name, labels)
|
|
self._counters[key] = self._counters.get(key, 0) + value
|
|
|
|
def observe(self, name: str, value: float, labels: dict[str, str] | None = None) -> None:
|
|
"""Record a histogram observation (e.g., latency)."""
|
|
key = self._key(name, labels)
|
|
self._histograms.setdefault(key, []).append(value)
|
|
if len(self._histograms[key]) > 10000:
|
|
self._histograms[key] = self._histograms[key][-5000:]
|
|
|
|
def set_gauge(self, name: str, value: float, labels: dict[str, str] | None = None) -> None:
|
|
"""Set a gauge value."""
|
|
self._gauges[self._key(name, labels)] = value
|
|
|
|
def snapshot(self) -> dict[str, Any]:
|
|
"""Return JSON-serializable metrics snapshot."""
|
|
hist_summary: dict[str, Any] = {}
|
|
for k, vals in self._histograms.items():
|
|
if vals:
|
|
sorted_vals = sorted(vals)
|
|
hist_summary[k] = {
|
|
"count": len(vals),
|
|
"mean": sum(vals) / len(vals),
|
|
"p50": sorted_vals[len(sorted_vals) // 2],
|
|
"p95": sorted_vals[int(len(sorted_vals) * 0.95)],
|
|
"p99": sorted_vals[int(len(sorted_vals) * 0.99)],
|
|
}
|
|
return {
|
|
"uptime_seconds": time.monotonic() - self._start,
|
|
"counters": dict(self._counters),
|
|
"histograms": hist_summary,
|
|
"gauges": dict(self._gauges),
|
|
}
|
|
|
|
def _key(self, name: str, labels: dict[str, str] | None) -> str:
|
|
if not labels:
|
|
return name
|
|
label_str = ",".join(f"{k}={v}" for k, v in sorted(labels.items()))
|
|
return f"{name}{{{label_str}}}"
|
|
|
|
|
|
_metrics: MetricsCollector | None = None
|
|
|
|
|
|
def get_metrics() -> MetricsCollector:
|
|
"""Get or create the global metrics collector."""
|
|
global _metrics
|
|
if _metrics is None:
|
|
_metrics = MetricsCollector()
|
|
return _metrics
|
|
|
|
|
|
def metrics_enabled() -> bool:
|
|
"""Check if metrics endpoint should be exposed."""
|
|
return os.environ.get("FUSIONAGI_METRICS_ENABLED", "false").lower() in ("true", "1", "yes")
|