rotalabs-probe
v0.1.0Detection tools for strategic underperformance in AI systems. Activation probes, behavioral analysis, and sandbagging detection with 90-96% accuracy.
Research tools for AI trust and reliability. AGPL-3.0 licensed.
Production-ready tools for detecting and managing AI system failures.
Detection tools for strategic underperformance in AI systems. Activation probes, behavioral analysis, and sandbagging detection with 90-96% accuracy.
Trust-based decision routing for multi-tier verification. 5-level cascade with adaptive pattern learning and 21 condition operators.
Runtime behavior control via activation steering. Guide model outputs toward desired behaviors using contrastive activation addition.
Verified code synthesis with Z3 solver. Correctness verification for AI-generated code with CE2P feedback and CEGIS synthesis.
Evolutionary adversarial testing. Red-teaming with quality-diversity optimization, MAP-Elites archive, and multi-provider LLM targets.
GNN-based trust propagation. Model trust relationships with PageRank, EigenTrust, and learned propagation algorithms.
Evaluation, acceleration, and compliance tools.
Comprehensive LLM evaluation with statistical rigor. Lexical, semantic, LLM-judge, and RAG metrics with confidence intervals and significance testing.
Inference acceleration with Triton kernels and speculative decoding. RMSNorm, SwiGLU, RoPE, INT8 quantization, and KV-cache compression.
Compliance evaluations for AI systems. EU AI Act, SOC2 Type II, and HIPAA frameworks with encrypted audit trails and report generation.
Reasoning chain capture for auditing. Extract and analyze CoT, ReAct, and Tree-of-Thought reasoning with anomaly detection.
# Core detection and control
pip install rotalabs-probe # Sandbagging detection
pip install rotalabs-steer # Activation steering
pip install rotalabs-cascade # Trust-based routing
pip install rotalabs-verify # Formal verification
pip install rotalabs-redqueen # Adversarial testing
pip install rotalabs-graph # Trust propagation
# Infrastructure
pip install rotalabs-eval # LLM evaluation
pip install rotalabs-accel # Inference acceleration
pip install rotalabs-comply # Compliance checking
pip install rotalabs-audit # Reasoning capture
Production-ready tools for detecting and managing AI system failures.
Sandbagging detection via metacognitive probes. Behavioral analysis and strategic underperformance detection for AI systems.
Domain-agnostic trust cascade routing. Multi-tier verification with adaptive pattern learning.
Runtime behavior control with steering vectors. Guide model outputs toward desired behaviors.
Verified code synthesis with CE2P hypothesis. Correctness verification for AI-generated code.
Evolutionary adversarial testing. Red-teaming with quality-diversity optimization.
GNN-based trust propagation. Model trust relationships with graph algorithms.
Evaluation, acceleration, and compliance tools.
Distributed LLM evaluation at scale. Statistical rigor with confidence intervals and significance testing.
Inference acceleration with speculative decoding. Optimized for Node.js environments.
Compliance evaluations for AI agents. Framework support for regulatory requirements.
Reasoning chain capture for auditing. Extract and analyze AI reasoning patterns.
# Core detection and control
npm install @rotalabs/probe # Sandbagging detection
npm install @rotalabs/steer # Steering vectors
npm install @rotalabs/cascade # Trust-based routing
npm install @rotalabs/verify # Code verification
npm install @rotalabs/redqueen # Adversarial testing
npm install @rotalabs/graph # Trust propagation
# Infrastructure
npm install @rotalabs/eval # LLM evaluation
npm install @rotalabs/accel # Inference acceleration
npm install @rotalabs/comply # Compliance checking
npm install @rotalabs/audit # Reasoning capture
Star repositories on GitHub for release notifications, or subscribe below for research updates.
Type to search across all pages and posts
Press ↑ ↓ to navigate, Enter to select