Files

12 KiB

Acknowledgements

Loki Mode stands on the shoulders of giants. This project incorporates research, patterns, and insights from the leading AI labs, academic institutions, and practitioners in the field.


Research Labs

Anthropic

Loki Mode is built for Claude and incorporates Anthropic's cutting-edge research on AI safety and agent development.

Paper/Resource Contribution to Loki Mode
Constitutional AI: Harmlessness from AI Feedback Self-critique against principles, revision workflow
Building Effective Agents Evaluator-optimizer pattern, parallelization, routing
Claude Code Best Practices Explore-Plan-Code workflow, context management
Simple Probes Can Catch Sleeper Agents Defection probes, anomaly detection patterns
Alignment Faking in Large Language Models Monitoring for strategic compliance
Visible Extended Thinking Thinking levels (think, think hard, ultrathink)
Computer Use Safety Safe autonomous operation patterns
Sabotage Evaluations Safety evaluation methodology
Effective Harnesses for Long-Running Agents One-feature-at-a-time pattern, Playwright MCP for E2E
Claude Agent SDK Overview Task tool, subagents, resume parameter, hooks

Google DeepMind

DeepMind's research on world models, hierarchical reasoning, and scalable oversight informs Loki Mode's architecture.

Paper/Resource Contribution to Loki Mode
SIMA 2: Generalist AI Agent Self-improvement loop, reward model training
Gemini Robotics 1.5 Hierarchical reasoning (planner + executor)
Dreamer 4: World Model Training Simulation-first testing, safe exploration
Genie 3: World Models World model architecture patterns
Scalable AI Safety via Doubly-Efficient Debate Debate-based verification for critical changes
Human-AI Complementarity for Amplified Oversight AI-assisted human supervision
Technical AGI Safety Approach Safety-first agent design

OpenAI

OpenAI's Agents SDK and deep research patterns provide foundational patterns for agent orchestration.

Paper/Resource Contribution to Loki Mode
Agents SDK Documentation Tracing spans, guardrails, tripwires
A Practical Guide to Building Agents Agent architecture best practices
Building Agents Track Development patterns, handoff callbacks
AGENTS.md Specification Standardized agent instructions
Introducing Deep Research Adaptive planning, backtracking
Deep Research System Card Safety considerations for research agents
Introducing o3 and o4-mini Reasoning model guidance
Reasoning Best Practices Extended thinking patterns
Chain of Thought Monitoring Reasoning trace monitoring
Agent Builder Safety Safety patterns for agent builders
Computer-Using Agent Computer use patterns
Agentic AI Foundation Industry standards, interoperability

Amazon Web Services (AWS)

AWS Bedrock's multi-agent collaboration patterns inform Loki Mode's routing and dispatch strategies.

Paper/Resource Contribution to Loki Mode
Multi-Agent Orchestration Guidance Three coordination mechanisms, architectural patterns
Bedrock Multi-Agent Collaboration Supervisor mode, routing mode, 10-agent limit
Multi-Agent Collaboration Announcement Intent classification, selective context sharing
AgentCore for SRE Gateway, Memory, Identity, Observability components

Key Pattern Adopted: Routing Mode Optimization - Direct dispatch for simple tasks (lower latency), supervisor orchestration for complex tasks (full coordination).


Academic Research

Multi-Agent Systems

Paper Authors/Source Contribution
Multi-Agent Collaboration Mechanisms Survey arXiv 2501.06322 Collaboration structures, coopetition
CONSENSAGENT: Anti-Sycophancy Framework ACL 2025 Findings Blind review, devil's advocate
GoalAct: Hierarchical Execution arXiv 2504.16563 Global planning, skill decomposition
A-Mem: Agentic Memory System arXiv 2502.12110 Zettelkasten-style memory linking
Multi-Agent Reflexion (MAR) arXiv 2512.20845 Structured debate, persona-based critics
Iter-VF: Iterative Verification-First arXiv 2511.21734 Answer-only verification, Markovian retry

Evaluation & Safety

Paper Authors/Source Contribution
Assessment Framework for Agentic AI arXiv 2512.12791 Four-pillar evaluation framework
Measurement Imbalance in Agentic AI arXiv 2506.02064 Multi-dimensional evaluation axes
Demo-to-Deployment Gap Stanford/Harvard Tool reliability vs tool selection

Industry Resources

Tools & Frameworks

Resource Contribution
NVIDIA ToolOrchestra Efficiency metrics, three-reward signal framework, dynamic agent selection
LerianStudio/ring Subagent-driven-development pattern
Awesome Agentic Patterns 105+ production patterns catalog

Best Practices Guides

Resource Contribution
Maxim AI: Production Multi-Agent Systems Correlation IDs, failure handling
UiPath: Agent Builder Best Practices Single-responsibility agents
GitHub: Speed Without Control Static analysis + AI review, guardrails

Hacker News Community

Battle-tested insights from practitioners deploying agents in production.

Discussions

Thread Key Insight
What Actually Works in Production for Autonomous Agents "Zero companies without human in the loop"
Coding with LLMs in Summer 2025 Context curation beats automatic RAG
Superpowers: How I'm Using Coding Agents Sub-agents for context isolation (Simon Willison)
Claude Code Experience After Two Weeks Fresh contexts yield better results
AI Agent Benchmarks Are Broken LLM-as-judge has shared blind spots
How to Orchestrate Multi-Agent Workflows Event-driven, decoupled coordination
Context Engineering vs Prompt Engineering Manual context selection principles

Show HN Projects

Project Contribution
Self-Evolving Agents Repository Self-improvement patterns
Package Manager for Agent Skills Skills architecture
Wispbit - AI Code Review Agent Code review patterns
Agtrace - Monitoring for AI Coding Agents Agent monitoring patterns

Individual Contributors

Special thanks to thought leaders whose patterns and insights shaped Loki Mode:

Contributor Contribution
Boris Cherny (Creator of Claude Code) Self-verification loop (2-3x quality improvement), extended thinking mode, "Less prompting, more systems" philosophy
Ivan Steshov Centralized constitution, agent lineage tracking, structured artifacts as contracts
Addy Osmani Git checkpoint system, specification-first approach, visual aids (Mermaid diagrams)
Simon Willison Sub-agents for context isolation, skills system, context curation patterns

Production Patterns Summary

Key patterns incorporated from practitioner experience:

Pattern Source Implementation
Human-in-the-Loop (HITL) HN Production Discussions Confidence-based escalation thresholds
Narrow Scope (3-5 steps) Multiple Practitioners Task scope constraints
Deterministic Validation Production Teams Rule-based outer loops (not LLM-judged)
Context Curation Simon Willison Manual selection, focused context
Blind Review + Devil's Advocate CONSENSAGENT Anti-sycophancy protocol
Hierarchical Reasoning DeepMind Gemini Orchestrator + specialized executors
Constitutional Self-Critique Anthropic Principles-based revision
Debate Verification DeepMind Critical change verification
One Feature at a Time Anthropic Harness Single feature per iteration, full verification
E2E Browser Testing Anthropic Harness Playwright MCP for visual verification

License

This acknowledgements file documents the research and resources that influenced Loki Mode's design. All referenced works retain their original licenses and copyrights.

Loki Mode itself is released under the MIT License.


Last updated: v2.35.0