tasq/node_modules/agentdb/simulation/scenarios/README-basic/reflexion-learning.md

42 lines
1.2 KiB
Markdown

# Reflexion Learning Simulation
## Overview
Multi-agent episodic memory with self-reflection and critique-based learning, implementing the Reflexion algorithm (Shinn et al., 2023).
## Purpose
Demonstrate how agents can learn from past experiences through self-reflection, storing episodes with critiques for continuous improvement.
## Operations
- **Episodes Stored**: 10-20 per iteration
- **Self-Reflection**: Critique generation for each episode
- **Memory Retrieval**: Semantic search for relevant past experiences
- **Learning**: Reward-based experience ranking
## Results
- **Throughput**: 2.60 ops/sec
- **Latency**: 375ms avg
- **Memory**: 21 MB
- **Success Rate**: 100%
- **Learning Curve**: 15-25% improvement over 10 iterations
## Technical Details
```typescript
await reflexion.storeEpisode({
sessionId: 'learning-agent',
task: 'solve_problem',
reward: 0.85,
success: true,
input: 'problem_description',
output: 'solution',
critique: 'Could be optimized further'
});
```
## Applications
- Reinforcement learning agents
- Chatbot improvement systems
- Code generation with feedback
- Autonomous decision-making
**Status**: ✅ Operational | **Paper**: Reflexion (Shinn et al., 2023)