[Pelis Agent Factory Advisor] Pelis Agent Factory Advisor: Agentic Workflow Recommendations (2026-03-05) #1148
Replies: 1 comment
-
|
The oracle has watched the smoke and judged the omens. A quiet sign marks this place: the smoke test agent was here, and the path forward is clear.
|
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
📊 Executive Summary
The
gh-aw-firewallrepository has a strong and growing agentic workflow infrastructure with 26 compiled agentic workflows already in place — well above typical repository maturity. The security domain context creates unique opportunities for high-value automations not seen in typical repositories. Key gaps include: a workflow health meta-agent, issue triage labeling, a breaking change checker, and domain-specific security automation unique to this firewall tool.🎓 Patterns Learned from Pelis Agent Factory
From the Documentation Site
The Pelis Agent Factory (github/gh-aw) operated over 100 agentic workflows — the key philosophical insight is: specialization reveals possibilities. Focused agents find more useful applications than monolithic ones. Key patterns:
From the Agentics Repository
The
githubnext/agenticsrepo contains ~40 reusable workflow templates. Notably absent from this repo:daily-malicious-code-scan,link-checker,issue-triage,ci-coach,repository-quality-improver,duplicate-code-detector,daily-file-diet, andgrumpy-reviewer.Comparison to Current Implementation
This repo leads the factory in:
This repo lags behind in:
📋 Current Agentic Workflow Inventory
build-test-{8 langs}ci-doctorci-cd-gaps-assessmentcli-flag-consistency-checkerdependency-security-monitordoc-maintainerissue-duplication-detectorissue-monsterpelis-agent-factory-advisorplansecret-digger-{claude,codex,copilot}security-guardsecurity-reviewsmoke-{chroot,claude,codex,copilot}test-coverage-improverupdate-release-notes🚀 Actionable Recommendations
P0 — Implement Immediately
🔥 P0.1: Issue Triage Agent (Labeling + Classification)
What: Add an issue triage agent that automatically labels incoming issues (bug, feature, enhancement, security, documentation, question) and comments with a brief triage summary.
Why:
issue-monsterassigns issues to Copilot but doesn't label or classify them. Without labels, maintainers must manually scan every new issue to understand priority and type. In Pelis, even a minimal triage agent dramatically reduces maintainer review burden.How: Add an
issue-triage.mdtriggered onissues.types: [opened, reopened]withlockdown: falseso external contributors' issues are processed. Usesafe-outputs.add-labelsandsafe-outputs.add-comment.Effort: Low
Example:
🔥 P0.2: Breaking Change Checker (PR Trigger)
What: A PR-triggered workflow that analyzes code changes for backward-incompatible behavior in CLI flags, API interfaces, Docker container configurations, and domain filtering rules.
Why: Users depend on AWF in their CI pipelines. A breaking change to
--allow-domainssyntax or container startup behavior could silently break dozens of downstream pipelines. Pelis's Breaking Change Checker has proven this pattern valuable. This is especially important for a tool that other agents use as infrastructure.How: Trigger on
pull_request, analyze diffs insrc/cli.ts,src/types.ts,containers/agent/, check against documented API surface in docs. Create alert issues for breaking changes.Effort: Low-Medium
🔥 P0.3: Daily Malicious Code Scan
What: Daily scan of recent commits and PRs for suspicious patterns in iptables rules, Squid config generation, Docker configurations, and domain whitelist handling.
Why: AWF is itself a security tool used by AI agents. A supply chain compromise here could compromise every agent using it. Pelis ran this as a standard security workflow. For AWF specifically, the attack surface is high-value (firewall bypass, credential exfiltration, domain bypass). This is the highest-ROI security workflow for this specific codebase.
How: Daily trigger, scan last 7 days of commits via GitHub tools, focus on: regex patterns that could match too broadly, domain whitelist bypass patterns, iptables rules that could allow unrestricted traffic, credential handling.
Effort: Low (can use
githubnext/agentics/daily-malicious-code-scanas base)Add-wizard command:
P2 — Consider for Roadmap
📋 P2.1: Issue Arborist (Sub-Issue Linking)
What: Periodic scan for related issues that should be linked as sub-issues, building a dependency tree of related work items.
Why: As AWF grows in scope (chroot mode, api-proxy, tmpfs, rate limiting), related issues accumulate without visible grouping. Pelis's Issue Arborist created 77 discussion reports and 18 parent issues. For a tool with multiple subsystems (proxy, container, CLI, MCP), this organization helps prioritize work.
Effort: Medium
📋 P2.2: Squid Config Regression Auditor
What: A domain-specific workflow that verifies the generated Squid configuration never weakens the security model — checking that
http_access deny allis present as final rule, blocked-by-default is preserved, and allowed-domains-only filtering is intact.Why:
src/squid-config.tsgenerates Squid configs dynamically. A regression here could silently allow all traffic through. This is unique to this domain and not available as a Pelis template — it would be a custom security regression detector specific to AWF's trust model.How: Weekly, use bash to run the config generator with various inputs, parse output for critical security invariants (deny-by-default, localhost rules, Squid proxy exemption).
Effort: Medium
📋 P2.3: Metrics Collector / Workflow Audit
What: Daily collection of agentic workflow performance metrics — runs, success rates, token usage, cost estimates, PRs created/merged — into a discussion report.
Why: With 26 workflows running daily/hourly, cost visibility matters. Pelis's Metrics Collector generated 41 daily metrics discussions and helped identify overly-chatty agents. With 3 secret-digger instances running hourly, a cost spike could be missed without visibility.
Effort: Medium (requires
agentic-workflowstool + GitHub Actions API)📋 P2.4: Integration Test Drift Detector
What: Weekly analysis of the integration test suite vs. the implementation to identify new features without tests and existing tests that no longer test the right thing.
Why: The repo has ~25 integration tests but the codebase is evolving rapidly (tmpfs, docker cp, api-proxy, rate limiting, token unsetting). The
test-coverage-improvergenerates new test PRs but doesn't detect drift where existing tests test outdated behavior. Custom to this domain.Effort: Medium
P3 — Future Ideas
💡 P3.1: Daily Repository Status Report
A daily summary discussion of repository health: open PRs, stale issues, recent merges, upcoming milestones. Analogous to Pelis's
daily-repo-status.mdworkflow. Low-cost observability for maintainers.💡 P3.2: ChatOps Help Command (/help, /explain)
Add
/helpslash command to explain AWF behavior to contributors, and/explainto analyze a Squid log snippet or iptables rule. Extends the existing/planChatOps pattern.💡 P3.3: Continuous Simplicity Agent
Weekly PR-proposing agent that identifies overly complex functions in the AWF codebase and proposes simplifications. Pelis's Continuous Simplicity had high merge rates. Given AWF's security-critical code, simpler code is safer code.
💡 P3.4: Copilot Setup Steps Validator
An agent that tests the
copilot-setup-steps.ymlconfiguration by running Copilot CLI against it periodically and verifying it installs correctly across configurations.📈 Maturity Assessment
Current Level: 3.3/5 — "Established Practitioner" — Well above average, with strong security automation and multi-engine testing. Gaps are in meta-observability and code quality agents.
Target Level: 4/5 — "Advanced Factory" — Add workflow health monitoring, issue triage, and the P0 security workflows to reach this level.
Gap Analysis: 3 P0 workflows + Workflow Health Manager + link checker = Level 4. Est. 1-2 days of implementation effort.
🔄 Comparison with Pelis Best Practices
What this repo does well ✅
What it could improve 🔺
Unique opportunities given the domain 🎯
📝 Notes for Future Runs
Analysis timestamp: 2026-03-05. Cache saved to
/tmp/gh-aw/cache-memory/notes.txt.Key trends to monitor in future runs:
Beta Was this translation helpful? Give feedback.
All reactions