Some checks failed
- New fusionagi/gpu/ module with TensorBackend protocol abstraction - TensorFlowBackend: GPU-accelerated ops with TensorCore mixed-precision - NumPyBackend: CPU fallback (always available, no extra deps) - Auto-selects best available backend at runtime - GPU-accelerated operations: - Cosine similarity matrix (batched, XLA-compiled) - Multi-head attention for consensus scoring - Batch hypothesis scoring on GPU - Semantic similarity search (pairwise, nearest-neighbor, deduplication) - New TensorFlowAdapter (fusionagi/adapters/): - LLMAdapter for local TF/Keras model inference - TensorCore mixed-precision support - GPU-accelerated embedding synthesis fallback - Reasoning pipeline integration: - gpu_scoring.py: drop-in GPU replacement for multi_path scoring - Super Big Brain: use_gpu config flag, GPU scoring when available - Memory integration: - gpu_search.py: GPU-accelerated semantic search for SemanticGraphMemory - Self-improvement integration: - gpu_training.py: gradient-based heuristic weight optimization - Reflective memory training loop with loss tracking - Dependencies: gpu extra (tensorflow>=2.16, numpy>=1.26) - 64 new tests (276 total), all passing - Architecture spec: docs/gpu_tensorcore_integration.md Co-Authored-By: Nakamoto, S <defi@defi-oracle.io>
90 lines
2.6 KiB
Python
90 lines
2.6 KiB
Python
"""Tests for fusionagi.gpu.tensor_attention."""
|
|
|
|
import pytest
|
|
|
|
from fusionagi.gpu.backend import reset_backend, get_backend
|
|
from fusionagi.gpu.tensor_attention import (
|
|
attention_consensus,
|
|
cross_claim_attention,
|
|
)
|
|
|
|
|
|
@pytest.fixture(autouse=True)
|
|
def _use_numpy():
|
|
reset_backend()
|
|
get_backend(force="numpy")
|
|
yield
|
|
reset_backend()
|
|
|
|
|
|
class TestAttentionConsensus:
|
|
def test_empty(self):
|
|
result = attention_consensus([], "query")
|
|
assert result["head_scores"] == []
|
|
assert result["consensus_score"] == 0.0
|
|
|
|
def test_single_head(self):
|
|
result = attention_consensus(
|
|
[["the sky is blue"]],
|
|
"what color is the sky",
|
|
)
|
|
assert len(result["head_scores"]) == 1
|
|
assert isinstance(result["consensus_score"], float)
|
|
|
|
def test_multiple_heads(self):
|
|
result = attention_consensus(
|
|
[
|
|
["the sky is blue", "water is wet"],
|
|
["security is important"],
|
|
["cost should be minimized"],
|
|
],
|
|
"what should we do about the project",
|
|
)
|
|
assert len(result["head_scores"]) == 3
|
|
assert 0.0 <= result["consensus_score"] <= 1.0
|
|
|
|
def test_with_weights(self):
|
|
result = attention_consensus(
|
|
[["claim a"], ["claim b"]],
|
|
"query",
|
|
head_weights=[2.0, 0.5],
|
|
)
|
|
assert len(result["head_scores"]) == 2
|
|
|
|
def test_empty_claims(self):
|
|
result = attention_consensus(
|
|
[[], []],
|
|
"query",
|
|
)
|
|
assert len(result["head_scores"]) == 2
|
|
assert result["head_scores"] == [0.0, 0.0]
|
|
|
|
|
|
class TestCrossClaimAttention:
|
|
def test_empty(self):
|
|
result = cross_claim_attention([])
|
|
assert result["similarity_matrix"] == []
|
|
assert result["conflict_pairs"] == []
|
|
|
|
def test_single(self):
|
|
result = cross_claim_attention(["only one claim"])
|
|
assert result["similarity_matrix"] == []
|
|
|
|
def test_two_claims(self):
|
|
result = cross_claim_attention(["claim one", "claim two"])
|
|
assert len(result["similarity_matrix"]) == 2
|
|
assert len(result["similarity_matrix"][0]) == 2
|
|
|
|
def test_self_similarity_high(self):
|
|
result = cross_claim_attention(["same text", "same text"])
|
|
sim = result["similarity_matrix"]
|
|
assert sim[0][0] > 0.9
|
|
assert sim[1][1] > 0.9
|
|
|
|
def test_conflict_detection(self):
|
|
result = cross_claim_attention([
|
|
"the project is very safe and reliable",
|
|
"completely unrelated topic about food and cooking",
|
|
])
|
|
assert isinstance(result["conflict_pairs"], list)
|