tasq/node_modules/agentdb/simulation/scenarios/README-basic/reflexion-learning.md

1.2 KiB

Reflexion Learning Simulation

Overview

Multi-agent episodic memory with self-reflection and critique-based learning, implementing the Reflexion algorithm (Shinn et al., 2023).

Purpose

Demonstrate how agents can learn from past experiences through self-reflection, storing episodes with critiques for continuous improvement.

Operations

  • Episodes Stored: 10-20 per iteration
  • Self-Reflection: Critique generation for each episode
  • Memory Retrieval: Semantic search for relevant past experiences
  • Learning: Reward-based experience ranking

Results

  • Throughput: 2.60 ops/sec
  • Latency: 375ms avg
  • Memory: 21 MB
  • Success Rate: 100%
  • Learning Curve: 15-25% improvement over 10 iterations

Technical Details

await reflexion.storeEpisode({
  sessionId: 'learning-agent',
  task: 'solve_problem',
  reward: 0.85,
  success: true,
  input: 'problem_description',
  output: 'solution',
  critique: 'Could be optimized further'
});

Applications

  • Reinforcement learning agents
  • Chatbot improvement systems
  • Code generation with feedback
  • Autonomous decision-making

Status: Operational | Paper: Reflexion (Shinn et al., 2023)