Agentic Design

Patterns
๐Ÿค

Human-in-the-Loop Agent (HULA)(HULA)

Framework for human-in-the-loop evaluation and refinement of LLM-based agents, allowing engineers to guide and assess agent performance at each development stage.

Complexity: highEvaluation and Monitoring

๐ŸŽฏ 30-Second Overview

Pattern: Three-agent collaboration framework with AI Planner, AI Coder, and Human Agent for software development

Why: Maintains human control while leveraging AI assistance for JIRA issue resolution and code generation

Key Insight: 79% plan success, 59% PR merge rate - keeping engineers in driver's seat enables reliable AI collaboration

โšก Quick Implementation

1Setup:Deploy AI Planner, AI Coding, and Human Agent components
2Plan:AI Planner creates coding plan from JIRA issue
3Review:Human agent reviews, refines, and approves plan
4Code:AI Coding agent generates code based on approved plan
5Validate:Human reviews code, provides feedback, approves PR
Example: hula_session = HULA(issue=jira_ticket, agents=[planner, coder, human], stages=[plan, code, review])

๐Ÿ“‹ Do's & Don'ts

โœ…Keep human engineer in driver's seat throughout the development process
โœ…Use three-stage evaluation: offline, online, and practitioner perception
โœ…Incorporate feedback from compilers, linters, and validation tools
โœ…Review and refine plans before moving to coding stage
โœ…Deploy in real JIRA environment for authentic evaluation
โŒAllow fully autonomous operation without human oversight
โŒSkip plan approval stage - human validation is critical
โŒIgnore compiler/linter feedback in code generation loop
โŒExpect 100% automation - human collaboration is the goal
โŒDeploy without proper three-stage evaluation framework

๐Ÿšฆ When to Use

Use When

  • โ€ข Software development teams needing AI assistance
  • โ€ข JIRA-based development workflows and issue tracking
  • โ€ข Organizations wanting human-controlled AI coding
  • โ€ข Teams requiring code quality assurance and oversight
  • โ€ข Enterprise environments with established review processes

Avoid When

  • โ€ข Fully autonomous coding requirements
  • โ€ข Simple scripting or one-off coding tasks
  • โ€ข Teams without structured issue tracking systems
  • โ€ข Projects requiring immediate code deployment
  • โ€ข Environments without human review capacity

๐Ÿ“Š Key Metrics

Plan Generation Success
79% of work items receive successful coding plans
Plan Approval Rate
82% of generated plans approved by engineers
Code Generation Success
87% of approved plans result in generated code
Pull Request Rate
25% of generated code reaches pull request stage
Merge Success Rate
59% of HULA PRs merged into repositories
SWE-bench Performance
37.2% resolution rate on SWE-bench Verified

๐Ÿ’ก Top Use Cases

Enterprise Software Development: Atlassian deployment with 45 engineers, ~900 merged PRs
JIRA Issue Resolution: Automated plan generation and code development for work item tracking
Collaborative AI Coding: Human-guided development maintaining engineer control and oversight
Quality Assurance Workflows: Integrated compiler/linter feedback with human review processes
Research and Development: Academic-industry collaboration for human-AI software engineering

References & Further Reading

Deepen your understanding with these curated resources

Contribute to this collection

Know a great resource? Submit a pull request to add it.

Contribute

Patterns

closed

Loading...