Skip to content

TraceIQ is a Python library for measuring, analyzing, and auditing AI-to-AI influence in multi-agent systems, providing reproducible metrics for semantic drift, influence propagation, and cascade risk to support safe, transparent, and scientifically grounded agent collaboration.

License

Notifications You must be signed in to change notification settings

Anarv2104/TraceIQ

Repository files navigation

TraceIQ

TraceIQ

TraceIQ is a Python library for measuring cross-agent influence in multi-agent AI systems, providing reproducible metrics for semantic drift, propagation risk, and reasoning stability.

PyPI version Python versions License Downloads

InstallationQuick StartScienceDocsCite


The Problem

Modern AI systems increasingly operate through interaction: agents collaborate, critique, retrieve, plan, and self-modify through communication with other agents.

Yet we lack scientific tools to answer fundamental questions:

  • How much did one agent influence another?
  • Did incorrect reasoning propagate through the system?
  • When does collaboration become instability?
  • How can influence be measured reproducibly and rigorously?

TraceIQ introduces a formal measurement framework for studying influence propagation in autonomous multi-agent AI systems.


Scientific Contributions

TraceIQ provides three core measurement primitives:

1. Influence Quotient (IQx)

A quantitative metric measuring semantic drift in an agent's output caused by prior agent messages.

IQx = Drift_L2 / (Baseline_Median + ε)

IQx estimates how much reasoning changed due to interaction, enabling measurement of cross-agent cognitive influence.

2. Propagation Risk (PR)

A network-level metric estimating how influence spreads across an agent graph using spectral radius analysis.

PR = max(|eigenvalues(Adjacency_Matrix)|)

PR provides early detection of unstable influence propagation and cascading reasoning errors.

3. Reproducible Influence Experiments

TraceIQ includes CI-safe experimental pipelines evaluating:

  • ✓ Hint injection and misinformation propagation
  • ✓ Influence stability across agent chains
  • ✓ Cold-start detection behavior
  • ✓ Mitigation policy effectiveness

These experiments support ongoing research into multi-agent AI safety and Contagious Intelligence.


Where TraceIQ Fits in AI Research

Modern AI tooling focuses on different layers of the stack:

Tool What It Measures
TensorBoard Model training metrics
Weights & Biases Experiment tracking
LangSmith / Prompt tools Prompt execution traces
TraceIQ Cross-agent influence and reasoning propagation

TraceIQ introduces a missing instrumentation layer: measurement of cognitive influence between autonomous AI agents.

As multi-agent systems become standard in RAG, planning, robotics, and orchestration, understanding how agents influence each other becomes as critical as measuring accuracy.


Why TraceIQ Exists

Autonomous AI systems are evolving from isolated models into collaborative agent networks.

However, we currently lack standardized methods to measure:

Challenge Description
Cross-agent reasoning influence How does one agent's output change another's behavior?
Error propagation Do mistakes cascade through agent pipelines?
Stability Is collaborative reasoning stable or chaotic?
Safety risks What are the emergent risks in multi-agent systems?

TraceIQ was built as a scientific instrument for studying emergent behavior in distributed intelligence.

TraceIQ is not a dashboard. It is not a monitoring SaaS.

TraceIQ is measurement infrastructure for multi-agent cognition research.


Research Vision

TraceIQ is measurement infrastructure for multi-agent AI systems.

As AI shifts from isolated models to collaborative agent networks, system behavior emerges from interactions between models—not from a single model alone. While we can measure accuracy, latency, and loss, we currently lack tools to quantify how reasoning propagates across agents.

TraceIQ provides reproducible metrics and structured tracking for cross-agent influence, reasoning drift, and propagation dynamics. The goal is not monitoring dashboards, but scientific instrumentation for studying distributed AI cognition.

If collaborative AI becomes the dominant computing paradigm, measuring influence between agents will be as fundamental as measuring model performance.


Features

Feature Description
📊 Influence Tracking Track influence between agent interactions
🎯 Semantic Drift Measure drift using embedding similarity
🌐 Propagation Risk Estimate network-level influence spread
Anomaly Detection Quantile-calibrated alerting system
🧊 Cold-Start Handling Statistical validation during warm-up
🔬 Research Pipelines CI-safe reproducible experiments
🔌 Integration Ready Templates for RAG and multi-agent orchestration

Installation

Core library (lightweight, no heavy ML dependencies):

pip install traceiq

With real embedding models:

pip install "traceiq[embedding]"

With research plotting tools:

pip install "traceiq[research]"

Full installation:

pip install "traceiq[embedding,research]"

Real-World Use Cases

TraceIQ is designed for real multi-agent AI systems:

  • Evaluate RAG hallucination propagation Measure whether incorrect retrieval contaminates downstream reasoning.

  • Audit autonomous agent pipelines Track which agents influence critical decisions in planning systems.

  • Study collaborative reasoning stability Detect when agent feedback loops amplify errors.

  • AI governance and accountability Build audit trails showing how decisions evolved across agents.

  • Research on Contagious Intelligence Quantify cognitive transfer between AI systems in controlled experiments.

TraceIQ acts as a measurement microscope for studying distributed AI cognition.


Quick Example

from traceiq import InfluenceTracker

tracker = InfluenceTracker(use_mock_embedder=True)

result = tracker.track_event(
    sender_id="agent_a",
    receiver_id="agent_b",
    sender_content="We should switch to renewable energy.",
    receiver_content="Good point. Renewables are the future."
)

print("Drift:", result["drift_l2_state"])
print("IQx:", result["IQx"])
print("Alert:", result["alert"])

tracker.close()

Output:

Drift: 0.847
IQx: 1.23
Alert: False

What TraceIQ Outputs

Each tracked interaction returns structured metrics you can log, visualize, or audit:

  • drift_l2_state — semantic drift magnitude
  • IQx — normalized influence score
  • alert — anomaly signal (calibrated)
  • valid — whether baseline is stabilized
  • receiver_state — receiver baseline summary

Research Applications

TraceIQ has been evaluated on synthetic multi-agent benchmarks and integrated into experimental LLM pipelines involving chained, retrieval-augmented, and tool-using agents.

The framework enables reproducible studies of:

  • Influence propagation across agent graphs
  • Stability of collaborative reasoning loops
  • Detection of misleading hint injection
  • Mitigation policy effectiveness
  • Cold-start behavior in autonomous agents

All experiments are reproducible through CI-safe pipelines that generate structured summary.json artifacts for verification.


Integration Patterns

TraceIQ works with common agent architectures:

Pattern Description
LLM-only agents Track message → response influence
RAG systems Include retrieved context in receiver input
Tool-using agents Track tool output influence
Memory agents Track before/after memory state
Multi-agent orchestrators Full conversation influence graphs

What TraceIQ Is NOT

Limitation Explanation
Not causal inference Metrics measure correlation, not proven causation
Not intent detection Cannot determine manipulation intent
Not semantic understanding Measures embedding-level drift
Not a production security system Research measurement tool
Not plug-and-play safety Thresholds require calibration per environment

Research Context

TraceIQ supports research into:

  • 🔬 AI-to-AI influence modeling
  • 🧬 Contagious Intelligence hypothesis
  • ⚖️ Multi-agent reasoning stability
  • 🛡️ Autonomous system safety
  • 🧠 Distributed cognition in AI systems

Detailed metric definitions and implementation notes are available in the project documentation.


Documentation

Document Description
Metrics Metric definitions and formulas
Integration Integration patterns
CLI Reference Command-line interface
Configuration TrackerConfig options
Architecture System design
Theory Mathematical foundations
Experiments Research testbed

CLI Help: traceiq --help


Reproducibility & CI

TraceIQ experiments are CI-safe:

  • ✅ Quick mode never hard-fails CI
  • ✅ Proof mode enforces strict statistical validation
  • ✅ Artifacts upload even on failures
  • ✅ Experiments produce structured outputs

This ensures reproducible research pipelines.


Contributing

Contributions welcome! See CONTRIBUTING.md

# 1. Fork the repository
# 2. Create a feature branch
git checkout -b feature/amazing-feature

# 3. Run tests and linter
pytest
ruff check src/ tests/

# 4. Submit a Pull Request

Citation

If you use TraceIQ in your research, please cite:

@software{traceiq,
  title = {TraceIQ: Measuring AI-to-AI Influence in Multi-Agent Systems},
  author = {Vasavada, Anarv and Contributors},
  year = {2026},
  url = {https://github.com/Anarv2104/TraceIQ}
}

License

TraceIQ is open-source under the MIT License, enabling academic and commercial use with minimal restrictions.

See the full license text at: https://github.com/Anarv2104/TraceIQ/blob/main/LICENSE


About

TraceIQ is a Python library for measuring, analyzing, and auditing AI-to-AI influence in multi-agent systems, providing reproducible metrics for semantic drift, influence propagation, and cascade risk to support safe, transparent, and scientifically grounded agent collaboration.

Topics

Resources

License

Contributing

Security policy

Stars

Watchers

Forks

Contributors