Skip to content

dontmindme369/ARIA

Repository files navigation

ARIA - Adaptive Resonant Intelligence Architecture

Self-learning cognitive architecture with LinUCB contextual bandits, quaternion semantic exploration, and anchor-based perspective detection.

Python 3.8+ License: MIT Tests Performance


What is ARIA?

ARIA is an advanced self-learning cognitive architecture that learns from every query to continuously improve retrieval quality. It combines:

  • 🎯 LinUCB Contextual Bandits - Feature-aware multi-armed bandit optimizes retrieval strategies
  • 🌀 Quaternion Semantic Exploration - 4D rotations through embedding space with golden ratio spiral
  • 🧭 Anchor-Based Perspective Detection - 8-framework query classification aligned with philosophical anchors
  • 📚 Enhanced Semantic Networks - V2 vocabularies with 121 concepts across 8 domains
  • 🎓 Continuous Learning Loop - Learns from conversation feedback and quality scoring
  • 📊 Hybrid Search - BM25 lexical + semantic embeddings (sentence-transformers)

Key Features

Adaptive Learning (LinUCB)

  • Context-Aware: Uses 10D query feature vectors (complexity, domain, length, etc.)
  • Fast Convergence: Learns optimal strategies in ~50 queries (vs 100+ for Thompson Sampling)
  • Feature-Based: Generalizes across similar query types
  • High Performance: 22,000+ selections/second, sub-millisecond latency

Semantic Exploration

  • Golden Ratio Spiral: φ-based (1.618...) uniform sphere coverage with 100 sample points
  • Multi-Rotation Refinement: 1-3 iterations for progressive depth
  • PCA-Aligned Rotations: Follow semantic space structure
  • Perspective-Aware Angles: 15°-120° rotation based on query intent and anchor alignment

Anchor Framework Integration

  • 8 Philosophical Anchors: Platonic Forms, Telos, Logos, Aletheia, Nous, Physis, Techne, Praxis
  • Vocabulary Alignment: 121 enhanced concepts across philosophy, engineering, law, business, creative arts, social sciences, security, data science
  • Meta-Cognitive Guidance: Reasoning heuristics, common errors, learning paths
  • Topology Maps: Network graphs show concept relationships and prerequisites

Dual Architecture

  • Teacher ARIA: Query-driven knowledge retrieval with bandit optimization
  • Student ARIA: Conversation corpus learning from LLM interactions
  • Feedback Loop: Quality scoring updates bandit preferences

Quick Start

Installation

git clone https://github.com/dontmindme369/ARIA.git
cd ARIA
pip install -r requirements.txt

Configuration

Create aria_config.yaml:

# Core paths
knowledge_base: "/path/to/your/knowledge/base"
embeddings: "/path/to/embeddings"
output_dir: "./rag_runs/aria"

# LinUCB Bandit Settings
bandit:
  epsilon: 0.10  # Exploration rate
  alpha: 1.0     # UCB exploration parameter
  feature_dim: 10

# Retrieval Presets (controlled by bandit)
presets:
  fast:
    top_k: 40
    sem_limit: 64
    rotations: 1
  balanced:
    top_k: 64
    sem_limit: 128
    rotations: 2
  deep:
    top_k: 96
    sem_limit: 256
    rotations: 3
  diverse:
    top_k: 80
    sem_limit: 128
    rotations: 2

Basic Usage

from core.aria_core import ARIA

# Initialize ARIA
aria = ARIA(
    index_roots=["/path/to/knowledge"],
    out_root="./aria_packs"
)

# Query with automatic preset selection
result = aria.query(
    "How do I implement a binary search tree in Python?"
)

# Access results
print(f"Preset: {result['preset']}")
print(f"Run dir: {result['run_dir']}")
print(f"Pack: {result['pack']}")

Command Line

# Single query
python aria_main.py "Explain how HTTP cookies work"

# With specific preset
python aria_main.py "Debug memory leak" --preset deep

# With anchor alignment
python aria_main.py "What is justice?" --with-anchor

Architecture

ARIA consists of 8 integrated layers:

Layer 1: Query Interface

  • Multi-format input (text, structured queries)
  • Query preprocessing and normalization

Layer 2: Feature Extraction

  • 10-dimensional feature vectors
  • Query complexity, domain, length, entity counts

Layer 3: Perspective Detection

  • 8 anchor-aligned perspectives
  • V2 semantic network vocabularies
  • ~1,440 perspective markers

Layer 4: Anchor Selection

  • Philosophical framework alignment
  • Template matching with exemplar scoring

Layer 5: LinUCB Intelligence

  • Contextual Multi-Armed Bandit
  • 4 arms: fast, balanced, deep, diverse
  • Feature-aware UCB with epsilon-greedy (ε=0.10)
  • A/b matrix tracking per arm

Layer 6: Quaternion Exploration

  • 4D semantic space rotations
  • Golden ratio spiral sampling
  • Multi-rotation refinement

Layer 7: Hybrid Retrieval

  • BM25 + semantic embeddings
  • Reciprocal rank fusion
  • Diversity-aware deduplication

Layer 8: Quality Assessment & Learning

  • Coverage + exemplar fit + diversity scoring
  • Conversation quality analysis
  • Bandit feedback with feature vectors

LinUCB Migration (Nov 2025)

ARIA has migrated from Thompson Sampling to LinUCB (Linear Upper Confidence Bound) contextual bandits for superior performance:

Improvements

Metric Thompson Sampling LinUCB Improvement
Convergence ~100 queries ~50 queries 2× faster
Features Used None 10D vectors Context-aware
Selection Speed ~1,000 ops/sec 22,658 ops/sec 23× faster
Generalization Per-query only Cross-query patterns Better

How LinUCB Works

  1. Feature Extraction: Extract 10D vector from query

    features = [
        query_length,      # Normalized 0-1
        complexity,        # simple/moderate/complex/expert
        domain_technical,  # Binary indicators
        domain_creative,
        domain_analytical,
        domain_philosophical,
        has_question,
        entity_count,
        time_of_day,
        bias_term         # Always 1.0
    ]
  2. UCB Calculation: For each preset (arm):

    UCB(arm) = θ·x + α·√(xᵀ·A⁻¹·x)
               ↑         ↑
           expected   uncertainty
           reward     (exploration)
    
  3. Selection: Choose arm with highest UCB (with ε-greedy random exploration)

  4. Update: After reward feedback:

    AA + x·xᵀ
    bb + r·x
    θ = A⁻¹·b  # Ridge regression weights

See docs/LINUCB_MIGRATION_COMPLETE.md for full details.


V2 Vocabulary System

Enhanced semantic networks with anchor alignment:

8 Domain Vocabularies

  1. Philosophy (16 concepts) - Epistemology, metaphysics, ethics
  2. Engineering (15 concepts) - Systems, optimization, design patterns
  3. Law (15 concepts) - Justice, contracts, precedent
  4. Business (15 concepts) - Strategy, operations, markets
  5. Creative Arts (15 concepts) - Aesthetics, narrative, craft
  6. Social Sciences (15 concepts) - Society, culture, research
  7. Security (15 concepts) - Threat modeling, defense, analysis
  8. Data Science (15 concepts) - ML, statistics, visualization

Total: 121 enhanced concepts with:

  • Semantic networks (551 edges, density 0.55-0.70)
  • Reasoning heuristics
  • Common errors and pitfalls
  • Learning prerequisites (depth 0-4)
  • Mental models

See docs/PHASE_3_COMPLETION_REPORT.md for details.


Performance

Benchmarks (Stress Tests)

✅ High-Volume Processing: 1,527 queries/second
✅ Concurrent Processing: 2,148 queries/second (10 threads)
✅ Bandit Selection: 22,658 operations/second (0.044ms avg)
✅ Bandit Update: 10,347 operations/second (0.097ms avg)
✅ Memory: Stable, no leaks detected
✅ Performance Degradation: < 1% over time

Quality Metrics

Coverage Score: 0.75-0.95 (semantic space coverage)
Exemplar Fit: 0.60-0.90 (anchor template alignment)
Diversity: 0.70-0.95 (result variety)
Overall Reward: 0.68-0.92 (multi-objective)

Testing

Test Suites

# Stress tests
python aria_systems_test_and_analysis/stress_tests/test_stress.py

# Bandit intelligence
python aria_systems_test_and_analysis/bandit_intelligence/test_bandit_intelligence.py

# Integration tests
python aria_systems_test_and_analysis/integration/test_integration.py

Current Status

Suite Tests Passed Success Rate
Stress Tests 6 6 100%
Bandit Intelligence 6 6 100%
Integration Tests 6 5* 83%**
Total 18 17 94.4%

*One test requires optional watchdog dependency for file monitoring **Core functionality: 100% tested and passing


Documentation


Project Status

Current Phase: Production Ready (Phase 3.5 Complete)

Completed

  • Phase 1: Anchor framework integration
  • Phase 2: V2 vocabulary development (121 concepts)
  • Phase 3: Semantic network integration & topology maps
  • Phase 3.5: LinUCB migration (Thompson → LinUCB)

Active Development

  • 🚧 Phase 4: Production integration & monitoring
  • 🚧 Enhanced query expansion using semantic networks
  • 🚧 Meta-cognitive reasoning heuristics
  • 🚧 Real-time learning dashboard

See docs/ARIA_PROJECT_CHECKPOINT.md for roadmap.


Requirements

python >= 3.8
numpy >= 1.21.0
sentence-transformers >= 2.0.0
rank-bm25 >= 0.2.2
pyyaml >= 5.4.1

Optional:

watchdog >= 2.1.0  # For file monitoring

License

MIT License - see LICENSE for details


Contact & Support

About

Adaptive Resonant Intelligent Architecture

Resources

License

Contributing

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •  

Languages