Some checks failed
- New fusionagi/gpu/ module with TensorBackend protocol abstraction - TensorFlowBackend: GPU-accelerated ops with TensorCore mixed-precision - NumPyBackend: CPU fallback (always available, no extra deps) - Auto-selects best available backend at runtime - GPU-accelerated operations: - Cosine similarity matrix (batched, XLA-compiled) - Multi-head attention for consensus scoring - Batch hypothesis scoring on GPU - Semantic similarity search (pairwise, nearest-neighbor, deduplication) - New TensorFlowAdapter (fusionagi/adapters/): - LLMAdapter for local TF/Keras model inference - TensorCore mixed-precision support - GPU-accelerated embedding synthesis fallback - Reasoning pipeline integration: - gpu_scoring.py: drop-in GPU replacement for multi_path scoring - Super Big Brain: use_gpu config flag, GPU scoring when available - Memory integration: - gpu_search.py: GPU-accelerated semantic search for SemanticGraphMemory - Self-improvement integration: - gpu_training.py: gradient-based heuristic weight optimization - Reflective memory training loop with loss tracking - Dependencies: gpu extra (tensorflow>=2.16, numpy>=1.26) - 64 new tests (276 total), all passing - Architecture spec: docs/gpu_tensorcore_integration.md Co-Authored-By: Nakamoto, S <defi@defi-oracle.io>
98 lines
3.0 KiB
Python
98 lines
3.0 KiB
Python
"""Tests for fusionagi.gpu.tensor_scoring and reasoning.gpu_scoring."""
|
|
|
|
import pytest
|
|
|
|
from fusionagi.gpu.backend import reset_backend, get_backend
|
|
from fusionagi.gpu.tensor_scoring import (
|
|
gpu_score_hypotheses,
|
|
gpu_score_claims_against_reference,
|
|
)
|
|
from fusionagi.reasoning.gpu_scoring import (
|
|
generate_and_score_gpu,
|
|
score_claims_gpu,
|
|
deduplicate_claims_gpu,
|
|
)
|
|
from fusionagi.schemas.atomic import AtomicSemanticUnit, AtomicUnitType
|
|
|
|
|
|
@pytest.fixture(autouse=True)
|
|
def _use_numpy():
|
|
reset_backend()
|
|
get_backend(force="numpy")
|
|
yield
|
|
reset_backend()
|
|
|
|
|
|
def _make_unit(content: str) -> AtomicSemanticUnit:
|
|
return AtomicSemanticUnit(
|
|
unit_id=f"u_{hash(content) % 10000}",
|
|
content=content,
|
|
type=AtomicUnitType.FACT,
|
|
confidence=1.0,
|
|
)
|
|
|
|
|
|
class TestGPUScoreHypotheses:
|
|
def test_empty(self):
|
|
assert gpu_score_hypotheses([], []) == []
|
|
|
|
def test_basic(self):
|
|
units = [_make_unit("the sky is blue"), _make_unit("water is wet")]
|
|
results = gpu_score_hypotheses(["the sky is blue"], units)
|
|
assert len(results) == 1
|
|
node, score = results[0]
|
|
assert node.thought == "the sky is blue"
|
|
assert 0.0 <= score <= 1.0
|
|
|
|
def test_multiple_hypotheses(self):
|
|
units = [_make_unit("python is great")]
|
|
results = gpu_score_hypotheses(
|
|
["python is great", "java is better", "rust is fast"],
|
|
units,
|
|
)
|
|
assert len(results) == 3
|
|
# Should be sorted by score descending
|
|
scores = [s for _, s in results]
|
|
assert scores == sorted(scores, reverse=True)
|
|
|
|
def test_no_units(self):
|
|
results = gpu_score_hypotheses(["test hypothesis"], [])
|
|
assert len(results) == 1
|
|
assert results[0][1] == 0.5
|
|
|
|
def test_gpu_metadata(self):
|
|
units = [_make_unit("test content")]
|
|
results = gpu_score_hypotheses(["test content"], units)
|
|
node, _ = results[0]
|
|
assert node.metadata.get("gpu_scored") is True
|
|
|
|
|
|
class TestGPUScoreClaimsAgainstReference:
|
|
def test_empty(self):
|
|
assert gpu_score_claims_against_reference([], "ref") == []
|
|
|
|
def test_basic(self):
|
|
scores = gpu_score_claims_against_reference(
|
|
["claim one", "claim two"],
|
|
"claim one reference",
|
|
)
|
|
assert len(scores) == 2
|
|
assert all(isinstance(s, float) for s in scores)
|
|
|
|
|
|
class TestReasoningGPUScoring:
|
|
def test_generate_and_score_gpu(self):
|
|
units = [_make_unit("hello world"), _make_unit("testing gpu")]
|
|
results = generate_and_score_gpu(["hello world", "testing gpu"], units)
|
|
assert len(results) == 2
|
|
|
|
def test_score_claims_gpu(self):
|
|
scores = score_claims_gpu(["test claim"], "reference text")
|
|
assert len(scores) == 1
|
|
assert isinstance(scores[0], float)
|
|
|
|
def test_deduplicate_claims_gpu(self):
|
|
groups = deduplicate_claims_gpu(["a", "b", "c"])
|
|
all_indices = sorted(idx for group in groups for idx in group)
|
|
assert all_indices == [0, 1, 2]
|