[Pelis Agent Factory Advisor] Agentic Workflow Advisor Report — March 2026 #1200
Replies: 1 comment
-
|
🔮 The ancient spirits stir; the smoke-test wanderer passed through and left clear signs in the ether. By starlight, the wards hold.
|
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
📊 Executive Summary
The
gh-aw-firewallrepository has a strong and mature agentic workflow foundation — 28 compiled workflows covering security scanning, smoke testing, CI investigation, documentation, and issue management. Compared to Pelis Agent Factory best practices, the biggest untapped opportunities are: code quality improvement agents (continuous refactoring/simplification), observability/metrics for the agents themselves, and a handful of high-value missing pieces like issue triage, breaking change detection, and schema consistency checking.🎓 Patterns Learned from Pelis Agent Factory
The Pelis Agent Factory runs 100+ specialized workflows organized into these key categories:
Key Pelis lessons applicable here:
planslash command is a good start; more ChatOps would help📋 Current Agentic Workflow Inventory
build-test-{lang}×8smoke-claude/codex/copilot/chroot×4secret-digger-*×3security-guardsecurity-reviewdependency-security-monitorci-doctorci-cd-gaps-assessmentcli-flag-consistency-checkerdoc-maintainertest-coverage-improverissue-duplication-detectorissue-monsterplan/plancommentupdate-release-notespelis-agent-factory-advisor🚀 Actionable Recommendations
P0 — Implement Immediately
🏷️ Issue Triage Agent
What: Add a triage agent that labels newly opened issues with appropriate categories (bug, feature, security, documentation, chore, etc.) and leaves a welcoming comment.
Why: Currently
issue-monsterdispatches issues but doesn't label them. The issue tracker shows many CI-doctor failure issues ([aw] Build Test X failed,[Security],[Deps]prefixes) but no systematic labeling of user-reported issues. Triage is the "hello world" of agentic automation with a fast payoff.How: Add
issue-triage.mdtriggered onissues: [opened, reopened]withlockdown: falsefor public issues. Labels should includebug,feature,security,documentation,question,firewall-config,container,performance.Effort: Low
🔄 Breaking Change Checker
What: A PR-triggered agent that watches for changes that could break backward compatibility — particularly CLI flag removals/renames, API/config schema changes, or container image changes that would break existing user setups.
Why: AWF is a security tool used in production CI pipelines. Breaking changes to CLI flags or container behavior can silently break users' workflows. This repo already has
cli-flag-consistency-checkerfor docs alignment, but nothing that alerts on breaking changes in PRs before merge.How: Trigger on
pull_request, read the diff, check for: removed CLI flags, changed flag defaults, renamed environment variables, changed Docker network topology, modified squid config schema.Effort: Low
P1 — Plan for Near-Term
📊 Workflow Metrics Collector / Audit Agent
What: A daily meta-agent that reviews the performance of all other agentic workflows — cost, error rates, PR proposal/merge rates, discussion creation — and produces a dashboard discussion.
Why: With 28 compiled workflows, there's no observability into how well the agents are actually performing. The Pelis Factory's
audit-workflowsandmetrics-collectorpatterns were game-changing — they identified overactive agents, stale outputs, and cost inefficiencies. At this scale, a meta-agent is increasingly valuable.How: Use
agenticworkflows-logstool + GitHub Actions API to aggregate run data. Output a weekly discussion with tables showing: runs per workflow, success rates, token cost estimates, PR outcomes.Effort: Medium
🧹 Continuous Simplicity Agent
What: A weekly agent that searches for code complexity — overly long functions, duplicated logic, unnecessary abstractions — and proposes focused refactoring PRs.
Why:
src/docker-manager.tsis already 1500+ lines and growing. The Pelis Factory's Continuous Simplicity agent had a 78%+ merge rate. This is the single biggest missing code quality automation.How: Add
continuous-simplicity.mdon weekly schedule withskip-if-matchto avoid PR flooding. Focus on: function length > 100 lines, duplicated error handling patterns, repeated string constants that should be shared.Effort: Low
🔍 Schema Consistency Checker
What: A weekly agent that verifies alignment between:
src/types.tsinterfaces,docker-manager.tsgenerated configs,squid-config.ts, and documentation indocs/.Why: As the codebase evolves, it's easy for
WrapperConfigfields to diverge from whatdocker-manager.tsactually uses, or fordocs/environment.mdto list environment variables that were renamed. The Pelis Factory's schema checker caught exactly this kind of drift. For a security tool, config drift can have serious consequences.How: Weekly schedule, reads all relevant source files and docs, creates a discussion report on any mismatches found.
Effort: Medium
📝 Documentation Noob Tester
What: An agent that periodically reads key documentation files (
README.md,docs/quickstart.md,docs/chroot-mode.md) through the eyes of a first-time user — attempting to follow the instructions literally and identifying confusing or missing steps.Why: AWF has extensive documentation (15+ docs files) targeting a sophisticated user base but the quick-start experience matters for adoption. The Pelis Factory's noob tester achieved 43% merge rate through causal chain — valuable signal generation even with lower direct merge rates.
Effort: Low
📦 Changeset Generator Enhancement
What: Enhance the existing
update-release-notesworkflow to also propose a version bump inpackage.jsonand updateCHANGELOG.mdby analyzing commits since last release.Why: The current workflow updates release notes but doesn't manage semantic versioning. This adds a significant release automation gap. The Pelis Factory's Changeset agent had 78% merge rate.
Effort: Low
P2 — Consider for Roadmap
🛡️ Daily Malicious Code Scan
What: A daily scan of recent commits for suspicious patterns — obfuscated code, unexpected network calls added to container scripts, hardcoded IPs, unusual dependency additions.
Why: AWF runs agent code inside containers with controlled network access — it's a high-value target for supply chain attacks. The Pelis Factory runs this daily. The repo already has secret diggers but no general malicious pattern scan.
Effort: Low
🤖 Grumpy Reviewer (ChatOps)
What: A reaction-triggered code reviewer that performs opinionated, personality-driven code reviews — catching security anti-patterns, Docker best practices, iptables rule ordering issues.
How: Trigger on 👀 reaction on a PR, use Claude with a grumpy-but-helpful persona. Especially valuable for: iptables rule ordering, Squid ACL precedence, capability management.
Effort: Low
🗂️ Glossary Maintainer
What: An agent that maintains a glossary of AWF-specific terms in the documentation —
TCP_TUNNEL,TCP_DENIED,NET_ADMIN,DNAT,ACL, etc. — keeping it synchronized with how terms are actually used in code and docs.Why: The firewall domain has significant specialized vocabulary. New contributors struggle with terms like "CONNECT method", "Squid ACL", "capsh", "NAT table". A maintained glossary reduces onboarding friction.
Effort: Low
🧪 CI Optimization Coach
What: A weekly agent that analyzes CI pipeline configuration and suggests optimizations — caching strategies, parallelization, unnecessary test steps.
Why: With 8 language-specific build-test workflows + 4 smoke tests, CI is expensive. The Pelis Factory's CI Coach had 100% merge rate on all 9 proposed PRs.
Effort: Medium
P3 — Future Ideas
🌐 Firewall Domain Allowlist Advisor
What: A workflow that analyzes which domains are commonly needed by each supported language ecosystem (npm, cargo, go mod, pip, etc.) and maintains recommended domain allowlists in documentation.
Why: One of the most common user pain points is figuring out which domains to allow. An agent could automatically test fresh builds of each ecosystem's hello-world and document the needed domains.
Effort: High
📈 Performance Regression Detector
What: Track container startup time and command execution overhead across releases, alerting if the firewall overhead increases significantly.
Effort: High
🔗 Issue Arborist
What: Links related issues as sub-issues — grouping security issues, feature requests by component, etc. Especially useful given the many CI-failure issues from dependabot PRs.
Effort: Medium
📈 Maturity Assessment
Current Level: 3/5 — Well-established agentic automation with strong security focus and testing coverage. The foundation is solid but several high-value pattern categories are underutilized.
Target Level: 4/5 — Achievable within 2-3 months by adding issue triage, breaking change checker, continuous simplicity, and a metrics collector.
🔄 Comparison with Pelis Factory Best Practices
What this repo does exceptionally well:
shared/mcp-pagination.md,shared/secret-audit.mdfollow DRY principlesWhat it could improve vs. best practices:
docker-manager.tsgrows, this becomes increasingly important.Unique opportunities given the domain (firewall/security):
Report generated by the Pelis Agent Factory Advisor workflow on 2026-03-10. Notes saved to cache-memory for trend tracking across future runs.
Beta Was this translation helpful? Give feedback.
All reactions