[Pelis Agent Factory Advisor] Agentic Workflow Maturity Analysis & Recommendations (2026-03-23) #1397
Closed
Replies: 1 comment
-
|
This discussion was automatically closed because it expired on 2026-03-30T03:35:00.858Z.
|
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
📊 Executive Summary
gh-aw-firewallhas a strong, security-focused agentic workflow foundation with 20+ workflows covering security, CI/CD health, documentation, smoke testing, and release automation. The repository is operating at maturity level 3.5/5 — well above average for an open-source project. The primary gaps are in continuous code quality improvement, issue organization/triage, and meta-monitoring of workflows themselves. Adding these would push the repo to level 4+ and close remaining automation gaps.🎓 Patterns Learned from Pelis Agent Factory
The Pelis Agent Factory ran 100+ specialized agentic workflows in
github/gh-aw. Key patterns that stand out:Pattern 1: Specialized > Monolithic
Rather than one big "do everything" agent, Pelis uses many narrow-purpose workflows. Each workflow does one thing well (CI Doctor vs. Schema Checker vs. Breaking Change Checker). This repo has already adopted this pattern effectively.
Pattern 2: Continuous Improvement Agents
Agents that run daily and propose incremental improvements via PRs compound in value over time. The Code Simplifier achieved 83% merge rate and the Duplicate Code Detector hit 79% merge rate — sustained practical value. This repo has
doc-maintainerandtest-coverage-improverbut lacks code quality simplification agents.Pattern 3: Meta-Monitoring (Workflow Health Manager)
A meta-agent that monitors other agents created 40 issues and contributed to 19 merged PRs. No equivalent exists here. The ci-doctor covers CI failures but not workflow health/configuration drift.
Pattern 4: Issue Organization Pipeline
Pelis uses a pipeline of issue management: Issue Triage → Issue Arborist (sub-issue linking) → Issue Monster (task dispatch) → Sub-Issue Closer. This repo has the Issue Monster and duplicate detector but is missing triage labeling and sub-issue organization.
Pattern 5: Validation-Only Analysts
Some workflows are pure read-only analysts (Schema Consistency Checker, Blog Auditor, Static Analysis Report) that produce discussion reports rather than code changes. These are low-risk, high-insight tools. This repo uses this pattern well with
ci-cd-gaps-assessment,security-review, andcli-flag-consistency-checker.How This Repo Compares
📋 Current Agentic Workflow Inventory
security-guardsecurity-reviewsecret-digger-claude/codex/copilotdependency-security-monitorsmoke-claude/codex/copilotsmoke-chrootbuild-testci-doctorci-cd-gaps-assessmentdoc-maintainercli-flag-consistency-checkertest-coverage-improverissue-monsterissue-duplication-detectorplanupdate-release-notespelis-agent-factory-advisor🚀 Actionable Recommendations
P0 — Implement Immediately
P0.1: Issue Triage Agent
What: Automatically label newly opened issues as
bug,feature,documentation,question,security,performance, orhelp-wanted.Why: The repo currently has no automatic labeling. Issues like
#1356(PATH propagation),#1355(memory limits), and#1354(proxy env leakage) are opened without labels, making triage manual. An AWF-specific triage agent can recognize patterns like "container", "iptables", "Squid", "domain allowlist" to assign accurate labels.How: Create
.github/workflows/issue-triage.mdtriggered onissues: [opened, reopened]. Include AWF-domain knowledge: security issues (keywords: CVE, vulnerability, bypass, escape), performance issues (timeout, slow, memory), networking issues (domain, Squid, proxy, iptables), and feature requests.Effort: Low
Example frontmatter:
P0.2: Workflow Health Manager
What: A meta-workflow that monitors all other agentic workflows in the repo, detects failures, configuration drift, and inactive/broken workflows, then creates issues.
Why: The repo already has 20+ workflows. Several recent issues (
#1388,#1379,#1378) are workflow failures that needed manual identification. A dedicated health manager would catch these proactively. In Pelis Factory, this created 40 issues and contributed to 19 merged PRs.How: Create
.github/workflows/workflow-health-manager.mdrunning daily. Check recent workflow runs viaagenticworkflows-logs, identify failures, check for workflows with stale trigger lists (ci-doctor's workflow_run list), and create issues with diagnostic context.Effort: Low-Medium
P1 — Plan for Near-Term
P1.1: Breaking Change Detector
What: Monitors PRs that touch the public CLI interface (
src/cli.ts),action.yml,WrapperConfigtypes (src/types.ts), or container entrypoints for backward-incompatible changes.Why: AWF is used as a GitHub Action. Breaking changes to
action.ymlinputs, CLI flags, or theWrapperConfigAPI can silently break downstream users. No workflow currently catches this. The Pelis Breaking Change Checker created alert issues for CLI version updates before they reached users.How: Trigger on PRs modifying
action.yml,src/cli.ts,src/types.ts,containers/*/entrypoint.sh. Usegit diffto analyze what changed, compare with the last release tag, and comment on the PR if breaking changes are detected.Effort: Medium
P1.2: Continuous Code Simplifier
What: Daily agent that analyzes recently modified TypeScript files for simplification opportunities and creates PRs.
Why: The codebase is growing complex —
src/docker-manager.tsis ~1100 lines andsrc/cli.tshandles many responsibilities. After rapid feature additions, complexity tends to accumulate. The Pelis Code Simplifier achieved 83% merge rate on 6 PRs in a short time. AWF's TypeScript codebase would benefit from the same treatment.How: Create
.github/workflows/code-simplifier.mdrunning daily. Focus on recently modified.tsfiles, look for: deeply nested conditionals, repeated patterns (env var building, Docker config generation), opportunities to extract helper functions. Skip test files and generated files.Effort: Medium
P1.3: Docs Site Auditor
What: Validates the Astro Starlight docs site (
docs-site/) for accuracy against the codebase — checking that documented CLI flags matchsrc/cli.ts, container architecture descriptions matchsrc/docker-manager.ts, and example commands still work.Why: The repo has both
README.md(covered bydoc-maintainer) and a separate Astro docs site indocs-site/. The docs site content is not validated by any existing workflow. CLI flag documentation, architecture diagrams, and code examples can drift silently.How: Create
.github/workflows/docs-site-auditor.mdrunning weekly. Cross-referencedocs-site/src/content/docs/markdown files withsrc/cli.tsfor flag accuracy, checkaction.ymlinputs against docs, and validate code examples are syntactically correct.Effort: Medium
P1.4: Integration Test Coverage Reporter
What: Weekly workflow that generates a structured report on integration test coverage gaps, cross-referencing the test inventory in
docs/INTEGRATION-TESTS.mdwith actual tests intests/.Why: The repo has
docs/INTEGRATION-TESTS.mdwith a "gap analysis" section, but coverage is tracked manually. Test Coverage Improver creates PRs but doesn't produce a readable coverage map. A dedicated reporter would surface coverage debt and guide prioritization. The Pelis Workflow Health Manager pattern is analogous here.How: Create
.github/workflows/integration-test-reporter.mdrunning weekly. Comparetests/integration/directory with the gap analysis doc, identify security-critical paths (DLP, API proxy credential isolation, domain bypass attempts) not covered by tests, and post a discussion.Effort: Low-Medium
P2 — Consider for Roadmap
P2.1: Issue Arborist (Sub-Issue Organizer)
What: Groups related open issues into parent-child hierarchies using GitHub's sub-issue feature.
Why: The issue tracker has multiple related issues (e.g.,
#1295Node.js global tools,#1356PATH propagation,#1328common API base all relate to environment setup). Arborist would create parent issues and link them, making roadmap planning cleaner.Effort: Medium
P2.2: Schema Consistency Checker
What: Validates that
WrapperConfig(insrc/types.ts), CLI flags (insrc/cli.ts),action.ymlinputs, and documentation all stay in sync.Why: AWF has three "sources of truth" for its configuration interface. Adding a new feature (e.g.,
--enable-dlp) requires updatingtypes.ts,cli.ts,action.yml,README.md, and docs. Drift is common and hard to catch. The Pelis Schema Checker created 55 analysis discussions catching drift between JSON schemas and code.Effort: Medium
P2.3: Changeset Generator
What: Automatically prepares changelog entries and version bump suggestions after each merged PR.
Why: The current
update-release-notesworkflow runs after a release is published. A Changeset Generator would proactively suggest changelog entries for PRs as they merge, reducing the manual burden at release time. The Pelis Changeset workflow achieved 78% merge rate on 28 PRs.Effort: Medium
P2.4: Daily Malicious Code Scanner
What: Reviews recently merged code changes for suspicious patterns: obfuscated code, unusual network calls, embedded strings that could be commands, or changes to security-critical paths (iptables rules, Squid config generation, credential handling).
Why: AWF is a security infrastructure tool. Malicious code introduced here could undermine the firewall for all users. Given the repository's domain (it is the security layer), this is especially high-value. The Pelis Daily Malicious Code Scan monitors for supply chain attack patterns.
Effort: Low-Medium
P3 — Future Ideas
P3.1: Performance Regression Monitor
What: Tracks AWF startup time (container initialization, Squid healthcheck) and flags PRs that significantly increase it.
Why: Issue
#1376notes integration tests taking 37+ minutes. Issue#240asks for a performance benchmarking suite. An automated performance monitor would catch regressions before they accumulate.Effort: High (requires benchmark infrastructure)
P3.2: Docs Noob Tester
What: Simulates a first-time user following the README and docs, identifying confusing or broken onboarding steps.
Why: AWF requires Docker, specific iptables permissions, and sudo. The onboarding path has multiple failure modes. A noob tester would systematically check the getting-started experience.
Effort: High (needs sandbox environment)
P3.3: Container Security Scanner Workflow
What: An agentic wrapper around
docker scoutor Trivy that runs daily on the published GHCR container images and creates issues for new CVEs.Why: The ci-doctor references a "Container Security Scan" workflow in its trigger list, but no
.mdfile exists for it — suggesting it's a standard YAML workflow without agentic analysis. An AI agent could provide deeper analysis and remediation suggestions beyond raw CVE counts.Effort: Medium
📈 Maturity Assessment
Overall Current Level: 3.5/5 — Above average, particularly strong on security automation given the domain.
Target Level: 4.5/5 — Achievable by adding P0+P1 items above.
Gap to close: Continuous code quality (P1.2), issue triage (P0.1), meta-monitoring (P0.2), and breaking change detection (P1.1) are the four additions that would have the most compound effect.
🔄 Comparison with Pelis Best Practices
What gh-aw-firewall Does Exceptionally Well
security-reviewgoes beyond what most repos doWhat Could Improve
src/docker-manager.tsat ~1100 lines and growing is a prime candidateUnique Opportunities Given the Security Domain
action.ymlinputs silently affects all downstream users📝 Cache Memory Updated
Notes for this analysis have been saved to
/tmp/gh-aw/cache-memory/pelis-advisor-notes.mdwith the workflow inventory, gaps identified, and maturity scores for tracking over time.Beta Was this translation helpful? Give feedback.
All reactions