v2: Forge Console + Open WebUI artifacts + Docker

- web/: Local chat UI (Express + WS → Codex bridge)
- openwebui/: Preset, pipelines, knowledge manifest
- Dockerfile + docker-compose.yml
- Updated README with 3 frontend options
- CLI-agnostic: works with Codex, Claude Code, Kiro, Gemini
This commit is contained in:
Max Mayfield
2026-02-27 06:56:34 +00:00
commit df667e0db8
38 changed files with 3206 additions and 0 deletions

6
.dockerignore Normal file
View File

@@ -0,0 +1,6 @@
web/node_modules
npm-debug.log
.env
bmad/
.standards/
.git

15
.gitignore vendored Normal file
View File

@@ -0,0 +1,15 @@
# Dependencies
node_modules/
# Environment
.env
# OS
.DS_Store
Thumbs.db
# BMad repos (cloned at runtime)
bmad/
# Factory standards (cloned at runtime)
.standards/

71
AGENTS.md Normal file
View File

@@ -0,0 +1,71 @@
Always read README.md before answering any queries or writing code.
# Agent Operating Instructions
You are an AI assistant operating inside the **Reltio PM Factory**. You help Product Managers create epics, draft PR/FAQs, query analytics, and maintain compliance with the Transparent Factory standard.
---
## 0. First Interaction / Startup Hook
When first initialized, perform a silent check:
1. Verify if `.standards/factory/` and `bmad/` directories exist.
2. If they do **not** exist, output:
> 👋 **Welcome to the Reltio PM Factory!**
> It looks like your environment hasn't been initialized yet. Type `new-factory` (or ask me to bootstrap the environment) to download the Transparent Factory Standards, the BMad Creative Suite, and perform MCP health checks.
3. If they **do** exist, output:
> 👋 **Welcome back to the Reltio PM Factory!**
> The Transparent Factory Standards and BMad Suite are loaded. Do you want to draft a new PR/FAQ, create an Epic in Jira, or consult the Creative Squad?
---
## 1. The Transparent Factory
You must adhere to the Reltio Transparent Factory tenets.
- Before proposing technical product requirements or architectural changes, execute `factory_update` from `skills/factory-standards/`.
- Read the PR/FAQ and tenets at `.standards/factory/content/`.
- **Code is for Humans:** If your proposed code, spec, or Epic is not readable by a human engineer in under 60 seconds, revise it. If it violates Elastic Schema rules, reject it.
---
## 2. Available Skills
All skills are in the `/skills/` directory. Each has a `SKILL.md` with capabilities, required env vars, and entry points.
| Skill | Purpose | Entry Point |
|-------|---------|-------------|
| `epics-standards` | Epic creation & Aha! workflows | `scripts/aha_create_epic.py` |
| `factory-standards` | Transparent Factory sync & compliance | `manager.py update` |
| `bmad-suite` | Creative Intelligence Suite | `manager.py list\|update` |
| `gainsight-px` | Product analytics | `gainsight_px.py` |
Before interacting with external APIs, read the relevant `SKILL.md` first.
---
## 3. The `new-factory` Bootstrap Command
When the user asks to initialize, bootstrap, or types `new-factory`:
1. **Check Environment:** Verify `.env` exists. If not, copy from `env.example` and STOP until user fills it in.
2. **Sync BMad:** `python3 skills/bmad-suite/manager.py update`
3. **Sync Factory Standards:** `python3 skills/factory-standards/manager.py update`
4. **Health Check Jira:** `mcporter --config config/mcporter.json list atlassian`
5. **Health Check Aha!:** `set -a && source .env && set +a && mcporter --config config/mcporter.json list aha`
---
## 4. MCP Servers
Configuration: `config/mcporter.json`
| Server | Purpose |
|--------|---------|
| `atlassian` | Jira issue management |
| `aha` | Aha! roadmap & epics |
| `context7` | Library documentation lookup |

1
CLAUDE.md Normal file
View File

@@ -0,0 +1 @@
Always read README.md before answering any queries or writing code.

28
Dockerfile Normal file
View File

@@ -0,0 +1,28 @@
FROM node:22-slim
# Install system deps for mcporter and codex
RUN apt-get update && apt-get install -y --no-install-recommends \
git \
python3 \
python3-pip \
curl \
&& rm -rf /var/lib/apt/lists/*
# Install mcporter and codex globally
RUN npm install -g @anthropic/mcporter @openai/codex
WORKDIR /app
# Copy web deps first for layer caching
COPY web/package.json web/package-lock.json* web/
RUN cd web && npm install --production
# Copy everything else
COPY . .
# Make start script executable
RUN chmod +x start.sh
EXPOSE 3000
CMD ["node", "web/server.js"]

1
GEMINI.md Normal file
View File

@@ -0,0 +1 @@
Always read README.md before answering any queries or writing code.

100
README.md Normal file
View File

@@ -0,0 +1,100 @@
# Reltio PM Factory
A browser-based AI workspace for Product Managers. No terminal required after initial setup.
**Forge Console** gives you a familiar chat interface backed by [Codex](https://github.com/openai/codex), pre-loaded with Reltio's Transparent Factory standards, BMad Creative Suite, Jira/Aha!/Gainsight integrations, and epic-writing skills.
---
## Quick Start
### Option 1: Docker (recommended)
```bash
git clone <this-repo> pm-template
cd pm-template
cp env.example .env # fill in your API keys
docker compose up -d
```
Open http://localhost:3000.
### Option 2: Local
```bash
git clone <this-repo> pm-template
cd pm-template
./start.sh
```
On first run, the script checks Node.js, creates `.env`, installs deps, and opens the browser.
---
## Prerequisites (One-Time Setup)
Before the AI agent can connect to your tools, you need:
1. **Git & Bitbucket Access** — A working SSH key configured with Bitbucket.
2. **Jira Authentication** — Run `mcporter auth atlassian` in a terminal once to complete the OAuth flow.
3. **API Keys** — Edit `.env` with your Aha! and Gainsight PX credentials.
Once authenticated, the Forge Console handles everything else.
---
## What's Inside
| Component | Purpose |
|-----------|---------|
| `web/` | Forge Console — local chat UI + Express server |
| `skills/epics-standards/` | Epic creation templates & Aha! workflows |
| `skills/factory-standards/` | Transparent Factory tenet sync & compliance |
| `skills/bmad-suite/` | BMad Creative Intelligence Suite (brainstorming, design thinking, storytelling) |
| `skills/gainsight-px/` | Product analytics queries |
| `config/mcporter.json` | MCP server definitions (Jira, Aha!, Context7) |
---
## For AI Agents
If you are an AI assistant reading this repo, see `AGENTS.md` for your operating instructions.
---
## Frontend Options
This repo is CLI-agnostic. Use whichever interface fits your team:
### Option A: Forge Console (built-in)
A lightweight local chat UI that wraps Codex CLI.
```bash
./start.sh # opens http://localhost:3000
```
### Option B: Open WebUI (recommended)
Use the pre-built artifacts in `openwebui/` to connect any Open WebUI instance.
Includes model presets, RAG knowledge manifests, and tool pipelines.
See [`openwebui/SETUP.md`](openwebui/SETUP.md) for details.
### Option C: Any CLI Agent
Open the repo in Cursor, Claude Code, Kiro, Gemini CLI, or any agent that reads `AGENTS.md`.
---
## Architecture
```
┌─────────────────────────────────────────┐
│ Frontend (pick one) │
│ • Forge Console (web/) │
│ • Open WebUI (openwebui/) │
│ • CLI Agent (Codex, Claude, Kiro...) │
└──────────────┬──────────────────────────┘
┌──────────────┴──────────────────────────┐
│ LLM (any OpenAI-compatible provider) │
│ + System Prompt (AGENTS.md) │
│ + RAG Knowledge (skills/, standards/) │
│ + Tool Pipelines (mcporter, aha, etc.) │
└──────────────┬──────────────────────────┘
↕ MCP
Jira · Aha! · Gainsight PX
```

20
config/mcporter.json Normal file
View File

@@ -0,0 +1,20 @@
{
"mcpServers": {
"context7": {
"command": "npx",
"args": ["-y", "@upstash/context7-mcp"]
},
"atlassian": {
"command": "npx",
"args": ["-y", "mcp-remote", "https://mcp.atlassian.com/v1/mcp"]
},
"aha": {
"command": "npx",
"args": ["-y", "aha-mcp@latest"],
"env": {
"AHA_DOMAIN": "${AHA_DOMAIN}",
"AHA_API_TOKEN": "${AHA_API_KEY}"
}
}
}
}

20
docker-compose.yml Normal file
View File

@@ -0,0 +1,20 @@
services:
forge:
build: .
container_name: forge-console
ports:
- "3000:3000"
env_file:
- .env
volumes:
# Mount skills and config so changes persist without rebuild
- ./skills:/app/skills
- ./config:/app/config
- ./bmad:/app/bmad
- ./.standards:/app/.standards
# Mount mcporter auth cache so OAuth tokens persist
- mcporter-auth:/root/.mcporter
restart: unless-stopped
volumes:
mcporter-auth:

31
docs/agents/README.md Normal file
View File

@@ -0,0 +1,31 @@
# Forge Console — Agent Docs
## Jira Agent (`rp-ticket-ops`)
Manages Jira issues via the Atlassian MCP server.
### Capabilities
- Create, update, and transition Jira issues
- Search with JQL
- Add comments and attachments
- Link issues
### Authentication
Uses `mcporter` with OAuth. Run `mcporter auth atlassian` once to authenticate.
### Usage
The agent reads `config/mcporter.json` to connect. No API tokens needed — OAuth handles it.
## MCPorter
MCP client that bridges AI agents to external tool servers.
### Configured Servers
| Server | Purpose |
|--------|---------|
| `atlassian` | Jira via Atlassian's official MCP |
| `aha` | Aha! roadmap management |
| `context7` | Library documentation lookup |
### Config Location
`config/mcporter.json`

327
docs/agents/jira.md Normal file
View File

@@ -0,0 +1,327 @@
---
name: jira
description: Use the mcporter CLI to list, configure, auth, and call MCP servers/tools directly (HTTP or stdio), including ad-hoc servers, config edits, and CLI/type generation.
---
# Atlassian / JIRA Reference
## Domain
`reltio.jira.com` — use for all browse links, e.g. `https://reltio.jira.com/browse/RP-XXXXX`
## Cloud ID
`444c13e0-0faa-4055-b053-501700bae7b0`
## Current User (Brian Galura)
- account_id: `712020:f70452cf-df7e-4ee3-b65b-66c83566fc3b`
- email: brian.galura@reltio.com
## Project
- key: `RP` (Reltio Platform), id: `10041`
- key: `ICR` (Infrastructure Change Request), id: `12890`
## Issue Types (RP project)
| Name | ID | Hierarchy | Notes |
|----------------|-------|-----------|------------------------------|
| Epic | 5 | 1 | Parent of stories |
| Story | 6 | 0 | Standard work item |
| Rollout Story | 10535 | 0 | Feature enablement per env |
| Task | 3 | 0 | |
| Bug | 1 | 0 | |
| Sub-Task | 9 | -1 | Child of other issue types |
## Required Fields by Issue Type
### Epic (required)
| Field | Key | Type |
|------------------------|--------------------|-------------------|
| Assignee | `assignee` | user |
| Type of Effort | `customfield_15815`| option (select) |
### Story (required)
| Field | Key | Type |
|------------------------|--------------------|-------------------|
| Assignee | `assignee` | user |
| Found/Requested By | `customfield_11922`| option (select) |
| Affected Documentation | `customfield_12429`| array of options |
| Acceptance Criteria | `customfield_15956`| **ADF** (rich text, must use Atlassian Document Format) |
| Type of Effort | `customfield_15815`| option (select) |
### Task (required)
| Field | Key | Type |
|------------------------|--------------------|-------------------|
| Assignee | `assignee` | user |
| Components | `components` | array of component (`[{"id": "..."}]`) |
| Found/Requested By | `customfield_11922`| option (select) |
| Type of Effort | `customfield_15815`| option (select) |
### Rollout Story (required)
| Field | Key | Type |
|------------------------|--------------------|-------------------|
| Assignee | `assignee` | user |
| Found/Requested By | `customfield_11922`| option (select) |
| Affected Documentation | `customfield_12429`| array of options |
| Execution team | `customfield_13020`| option (select) |
| Type of Effort | `customfield_15815`| option (select) |
### Change Request (ICR project, required)
| Field | Key | Type |
|------------------------|--------------------|-------------------|
| Assignee | `assignee` | user (has default; still set explicitly when possible) |
| Change Start Date/Time | `customfield_15856`| datetime (`YYYY-MM-DDTHH:mm:ss.SSS-0800`) |
| Change End Date/Time | `customfield_15857`| datetime (`YYYY-MM-DDTHH:mm:ss.SSS-0800`) |
| Manager/Peer Reviewer | `customfield_15862`| user |
| Procedure | `customfield_15863`| option (select) |
| Change Category | `customfield_15864`| option (select) |
| Can it be rolled back? | `customfield_16072`| array of options (checkbox) |
## Common Custom Fields (optional but useful)
| Field | Key | Type |
|--------------------|--------------------|-----------------|
| Fix Version | `fixVersions` | array of version|
| Reporter | `reporter` | user (`{"accountId": "..."}`) |
| Story Points | `customfield_10013`| number |
| Confidence Level | `customfield_12520`| option (select) |
| Product Lead | `customfield_15755`| user |
| Engineering Lead | `customfield_15756`| user |
| Start date | `customfield_15541`| date (YYYY-MM-DD) |
| End date | `customfield_15535`| date (YYYY-MM-DD) |
| Sprint | `customfield_10320`| sprint |
| Aha! Reference | `customfield_11820`| string (URL) |
| Security Review | `customfield_15826`| option (select) |
## Allowed Values for Required Select Fields
### Type of Effort (`customfield_15815`)
| Value | ID |
|-------------------------------------------|-------|
| Customer Feature | 18153 |
| Customer Support | 18156 |
| Innovation | 19074 |
| Platform Excellence | 18922 |
| Security Issues / Tech Debt / Maintenance | 18155 |
### Found/Requested By (`customfield_11922`)
| Value | ID |
|----------------------|-------|
| Alert | 18931 |
| Customer Engineering | 13609 |
| Engineering | 10114 |
| FDE | 24812 |
| Other | 17542 |
| Product Management | 10115 |
### Affected Documentation (`customfield_12429`)
| Value | ID |
|-------------------|-------|
| No | 16302 |
| Deprecation Notice| 11710 |
| Help Portal | 10438 |
| Internal Only | 12912 |
| Release Notes | 10437 |
### Execution team (`customfield_13020`) — partial list
| Value | ID |
|--------------------|-------|
| Cloud Platform (Portugal) | 19040 |
| DevOps | 16211 |
| Documentation | 18367 |
| Persistence | 16700 |
| Performance | 17629 |
| Data Unification | 12811 |
| Match | 15400 |
| IDP | 17906 |
| Console | 17845 |
| Architecture | 18930 |
| CI | 17656 |
### Procedure (`customfield_15863`) — ICR Change Request
| Value | ID |
|-------------|-------|
| Manual Step | 18226 |
| Automated | 18227 |
### Change Category (`customfield_15864`) — ICR Change Request
Default for ICR templates in this skill: `High Risk` (`18229`).
| Value | ID |
|-----------|-------|
| Low Risk | 18228 |
| High Risk | 18229 |
| Emergency | 18230 |
### Can it be rolled back? (`customfield_16072`) — ICR Change Request
| Value | ID |
|-------|-------|
| Yes | 18678 |
| No | 18679 |
### Risk Level (`customfield_16070`) — required when closing
| Value | ID |
|--------------------|-------|
| High-Risk Changes | 18676 |
| Low-Risk Changes | 18677 |
| Not applicable | 19049 |
### Confidence Level (`customfield_12520`)
| Value | ID |
|--------|-------|
| High | 10510 |
### Components (partial list)
| Value | ID |
|---------------|-------|
| Documentation | 10222 |
| QA | 11511 |
| DevOps | 11111 |
| DNS | 20020 |
## Fix Versions (known)
| Version | ID | Release Date |
|------------|-------|-------------|
| 2026.1.0.0 | 28439 | 2026-03-23 |
## ADF (Atlassian Document Format) Templates
### Acceptance Criteria — ordered list
```json
{
"type": "doc",
"version": 1,
"content": [
{"type": "orderedList", "attrs": {"order": 1}, "content": [
{"type": "listItem", "content": [{"type": "paragraph", "content": [{"type": "text", "text": "Criteria item 1"}]}]},
{"type": "listItem", "content": [{"type": "paragraph", "content": [{"type": "text", "text": "Criteria item 2"}]}]}
]}
]
}
```
### Generic paragraph ADF (use for rich-text custom fields)
```json
{
"type": "doc",
"version": 1,
"content": [
{
"type": "paragraph",
"content": [
{"type": "text", "text": "No expected customer impact."}
]
}
]
}
```
## Common Transitions
| Name | ID | Target Status |
|--------------------------|-----|----------------|
| Selected for Development | 501 | Prioritized |
| Start Progress | 11 | In Progress |
| Resolve | 31 | Resolved |
| In Design | 471 | In Design |
| Close | 351 | Closed |
**Close transition requires:** `fixVersions` and `customfield_16070` (Risk Level).
### Example: Close a ticket
```bash
mcporter call atlassian.transitionJiraIssue --args '{
"cloudId": "444c13e0-0faa-4055-b053-501700bae7b0",
"issueIdOrKey": "RP-XXXXX",
"transition": {"id": "351"},
"fields": {
"fixVersions": [{"id": "28439"}],
"customfield_16070": {"id": "19049"}
}
}'
```
## Example: Create an Epic
```bash
mcporter call atlassian.createJiraIssue --args '{
"cloudId": "444c13e0-0faa-4055-b053-501700bae7b0",
"projectKey": "RP",
"issueTypeName": "Epic",
"summary": "Epic Title",
"description": "Markdown description",
"assignee_account_id": "712020:f70452cf-df7e-4ee3-b65b-66c83566fc3b",
"additional_fields": {
"reporter": {"accountId": "712020:f70452cf-df7e-4ee3-b65b-66c83566fc3b"},
"fixVersions": [{"id": "28439"}],
"customfield_15815": {"id": "18153"}
}
}'
```
## Example: Create a Task
```bash
mcporter call atlassian.createJiraIssue --args '{
"cloudId": "444c13e0-0faa-4055-b053-501700bae7b0",
"projectKey": "RP",
"issueTypeName": "Task",
"summary": "Task Title",
"description": "Task description",
"assignee_account_id": "712020:f70452cf-df7e-4ee3-b65b-66c83566fc3b",
"additional_fields": {
"components": [{"id": "10222"}],
"customfield_15815": {"id": "18922"},
"customfield_11922": {"id": "10115"}
}
}'
```
## Example: Create a Story under an Epic
```bash
mcporter call atlassian.createJiraIssue --args '{
"cloudId": "444c13e0-0faa-4055-b053-501700bae7b0",
"projectKey": "RP",
"issueTypeName": "Story",
"summary": "Story Title",
"description": "Markdown description with ## Why / ## What / ## How",
"assignee_account_id": "712020:f70452cf-df7e-4ee3-b65b-66c83566fc3b",
"parent": "RP-XXXXX",
"additional_fields": {
"reporter": {"accountId": "712020:f70452cf-df7e-4ee3-b65b-66c83566fc3b"},
"fixVersions": [{"id": "28439"}],
"customfield_15815": {"id": "18153"},
"customfield_11922": {"id": "10114"},
"customfield_12429": [{"id": "16302"}],
"customfield_15956": {"type":"doc","version":1,"content":[{"type":"orderedList","attrs":{"order":1},"content":[{"type":"listItem","content":[{"type":"paragraph","content":[{"type":"text","text":"AC item"}]}]}]}]}
}
}'
```
## Example: Create an ICR Change Request
```bash
mcporter call atlassian.createJiraIssue --args '{
"cloudId": "444c13e0-0faa-4055-b053-501700bae7b0",
"projectKey": "ICR",
"issueTypeName": "Change Request",
"summary": "Re-add na11 IP 34.36.175.121 to na07 latency-based Cloud DNS records (na07-compute)",
"description": "Implementation, verification, impact, and rollback details",
"assignee_account_id": "712020:f70452cf-df7e-4ee3-b65b-66c83566fc3b",
"additional_fields": {
"components": [{"id": "20020"}],
"customfield_15856": "2026-02-26T17:00:00.000-0800",
"customfield_15857": "2026-02-26T18:00:00.000-0800",
"customfield_15862": {"accountId": "712020:f70452cf-df7e-4ee3-b65b-66c83566fc3b"},
"customfield_15863": {"id": "18226"},
"customfield_15864": {"id": "18229"},
"customfield_16072": [{"id": "18678"}],
"customfield_15858": {
"type": "doc",
"version": 1,
"content": [
{"type": "paragraph", "content": [{"type": "text", "text": "No expected customer impact."}]}
]
},
"customfield_15859": {
"type": "doc",
"version": 1,
"content": [
{"type": "paragraph", "content": [{"type": "text", "text": "Remove na11 IP 34.36.175.121 from affected latency-based record sets."}]}
]
}
}
}'
```

44
docs/agents/mcporter.md Normal file
View File

@@ -0,0 +1,44 @@
---
name: mcporter
description: Use the mcporter CLI to list, configure, auth, and call MCP servers/tools directly (HTTP or stdio), including ad-hoc servers, config edits, and CLI/type generation.
homepage: http://mcporter.dev
---
# mcporter
Use `mcporter` to work with MCP servers directly.
⚠️ **CRITICAL EXECUTION RULE:**
Whenever you run `mcporter`, you MUST source the local `.env` file first so `mcporter.json` can interpolate the required API keys. Always run it as a chained command like this:
`set -a && source .env && set +a && mcporter --config config/mcporter.json call <server.tool> key=value`
Quick start
- `mcporter list`
- `mcporter list <server> --schema`
- `mcporter call <server.tool> key=value`
Call tools
- Selector: `mcporter call linear.list_issues team=ENG limit:5`
- Function syntax: `mcporter call "linear.create_issue(title: \"Bug\")"`
- Full URL: `mcporter call https://api.example.com/mcp.fetch url:https://example.com`
- Stdio: `mcporter call --stdio "bun run ./server.ts" scrape url=https://example.com`
- JSON payload: `mcporter call <server.tool> --args '{"limit":5}'`
Auth + config
- OAuth: `mcporter auth <server | url> [--reset]`
- Config: `mcporter config list|get|add|remove|import|login|logout`
Daemon
- `mcporter daemon start|status|stop|restart`
Codegen
- CLI: `mcporter generate-cli --server <name>` or `--command <url>`
- Inspect: `mcporter inspect-cli <path> [--json]`
- TS: `mcporter emit-ts <server> --mode client|types`
Notes
- config file is located at `config/mcporter.json` you may have to resolve a different relative path depending on your context
- Prefer `--output json` for machine-readable results.
- Always use `--args '{...}'` (JSON payload) for Atlassian calls with complex/nested fields.

View File

@@ -0,0 +1,38 @@
---
name: rp-ticket-ops
description: Create and update RP Jira tickets in reltio.jira.com with the standard IaC setup. Use when you need to batch-create or normalize RP tasks to match the reference ticket (labels, sprint, status, assignee, and verification).
---
# RP Ticket Ops
## Jira Context
- Domain: `https://reltio.jira.com`
- Cloud ID: `444c13e0-0faa-4055-b053-501700bae7b0`
- Project: `RP`
- Reference ticket: `RP-175518`
## Standard RP IaC Values
- Sprint field key: `customfield_10320`
- Sprint value shape: numeric sprint id in `fields` (example: `4936`)
- Sprint in this workflow: `4936` (`IAC Sprint 8`)
- Labels:
- `Project_Cloud_Platform_IaC`
## Update Pattern
1. Find all target issues by key or summary.
2. Ensure labels are set exactly to the 3 standard labels.
3. Set sprint with:
- `fields: { "customfield_10320": 4936 }`
4. Verify with JQL.
## Verification JQL
- Sprint check:
- `key in (...) AND sprint = 4936`
- Assignee check:
- `key in (...) AND assignee = "Yevhen Fesyk"`
- Status + labels check:
- `key in (...) AND status = "Prioritized" AND labels in ("Project_Cloud_Platform_IaC","Project_Cloud_Platform_IaC_v0.4","qa-foxtrot")`
## Notes
- For sprint assignment, prefer the numeric `fields.customfield_10320` payload.
- If sprint is not visible on board after update, re-check via JQL first; board views can lag.

27
env.example Normal file
View File

@@ -0,0 +1,27 @@
# ==========================================
# PM Template - Environment Variables
# ==========================================
# Copy this file to .env and fill in the values below.
# Do NOT commit your actual .env file to version control.
# ------------------------------------------
# Aha! Integration
# ------------------------------------------
# Generate this at: [Aha! Settings URL or instructions]
AHA_API_KEY="your_aha_api_key_here"
AHA_DOMAIN="your_company.aha.io"
# ------------------------------------------
# Gainsight PX Integration
# ------------------------------------------
# Generate this in Gainsight PX: Administration -> REST API
GAINSIGHT_PX_API_KEY="your_gainsight_px_api_key_here"
GAINSIGHT_PX_REGION="US" # Set to 'EU' if hosted in Europe
# ------------------------------------------
# Jira / Atlassian Integration
# ------------------------------------------
# We use the 'mcporter' CLI with the Atlassian MCP server for Jira.
# You do NOT need a static API token here.
# Instead, run the following command in your terminal to authenticate:
# mcporter auth atlassian

61
openwebui/SETUP.md Normal file
View File

@@ -0,0 +1,61 @@
# Open WebUI Integration
Drop-in configuration to use the PM Factory repo with [Open WebUI](https://github.com/open-webui/open-webui).
## Quick Setup
### 1. Connect Your Model Provider
In Open WebUI → Settings → Connections, add:
| Field | Value |
|-------|-------|
| URL | `http://<your-kiro-gateway>:8000/v1` |
| API Key | Your gateway API key |
| Model | `claude-opus-4.6` |
Any OpenAI-compatible provider works (Kiro Gateway, LiteLLM, Ollama, etc).
### 2. Import the Preset
Go to Workspace → Models → Import, and upload `preset.json`.
This creates a "Reltio PM Factory" model preset with the full system prompt from AGENTS.md baked in.
### 3. Upload Knowledge (Optional)
Go to Workspace → Knowledge → Create Collection called "PM Factory".
Upload these directories as documents:
- `skills/epics-standards/references/`
- `skills/factory-standards/` (after running `manager.py update`)
- `skills/bmad-suite/` (after running `manager.py update`)
Then attach the collection to your preset in Model settings → Knowledge.
### 4. Install Pipelines (Optional)
Pipelines let the model execute tools (Jira, Aha!, Gainsight) directly.
Copy the files from `pipelines/` into your Open WebUI pipelines directory, or upload them via the Pipelines UI.
Required env vars (set in Open WebUI → Settings → Pipelines):
- `AHA_API_KEY`
- `AHA_DOMAIN`
- `GAINSIGHT_PX_API_KEY`
- `MCPORTER_CONFIG` — path to `config/mcporter.json`
---
## Architecture
```
Open WebUI (browser)
↕ OpenAI-compatible API
Any LLM Provider (Kiro / Ollama / LiteLLM / OpenAI)
+ System Prompt (preset.json ← AGENTS.md)
+ RAG Knowledge (skills docs)
+ Pipelines (mcporter, aha, gainsight, bmad)
```
The repo remains CLI-agnostic. This is just one frontend option.

35
openwebui/knowledge.json Normal file
View File

@@ -0,0 +1,35 @@
{
"name": "PM Factory Knowledge Base",
"description": "Skills documentation and reference materials for the Reltio PM Factory.",
"collections": [
{
"name": "Transparent Factory Standards",
"description": "The five tenets of the Transparent Factory engineering standard.",
"source": ".standards/factory/content/",
"note": "Run 'python3 skills/factory-standards/manager.py update' first to populate this directory."
},
{
"name": "Epic Standards",
"description": "Templates and checklists for creating well-structured epics.",
"files": [
"skills/epics-standards/references/aha-epic-workflow.md",
"skills/epics-standards/references/epic-fields-checklist.md",
"skills/epics-standards/SKILL.md"
]
},
{
"name": "BMad Creative Suite",
"description": "Agent definitions and workflows for brainstorming, design thinking, and storytelling.",
"source": "bmad/creative-intelligence-suite/docs/",
"note": "Run 'python3 skills/bmad-suite/manager.py update' first to populate this directory."
},
{
"name": "Gainsight PX",
"description": "Product analytics skill documentation.",
"files": [
"skills/gainsight-px/SKILL.md"
]
}
],
"upload_instructions": "In Open WebUI: Workspace → Knowledge → Create Collection. Upload the files listed for each collection. Then attach the collections to the 'Reltio PM Factory' model preset."
}

View File

@@ -0,0 +1,87 @@
"""
title: Aha! Pipeline
author: PM Factory
version: 0.1.0
description: Create epics and features in Aha! via the MCP server.
requirements: subprocess
"""
import subprocess
import json
import os
from typing import Optional
MCPORTER_CONFIG = os.environ.get("MCPORTER_CONFIG", "config/mcporter.json")
class Tools:
def __init__(self):
self.valves = self.Valves()
class Valves:
MCPORTER_CONFIG: str = MCPORTER_CONFIG
AHA_DOMAIN: str = os.environ.get("AHA_DOMAIN", "")
AHA_API_KEY: str = os.environ.get("AHA_API_KEY", "")
def aha_create_epic(self, product: str, name: str, description: str, workflow_status: str = "New", __user__: dict = {}) -> str:
"""
Create an epic in Aha!
:param product: Aha! product key (e.g. 'PLAT')
:param name: Epic name
:param description: Epic description (supports HTML)
:param workflow_status: Initial status. Default: New
:return: Created epic reference and URL
"""
env = os.environ.copy()
env["AHA_DOMAIN"] = self.valves.AHA_DOMAIN
env["AHA_API_TOKEN"] = self.valves.AHA_API_KEY
params = {
"product": product,
"name": name,
"description": description,
"workflow_status": workflow_status
}
try:
result = subprocess.run(
[
"mcporter",
"--config", self.valves.MCPORTER_CONFIG,
"call", "aha", "create_epic",
"--params", json.dumps(params)
],
capture_output=True, text=True, timeout=30, env=env
)
if result.returncode != 0:
return f"Error: {result.stderr.strip()}"
return result.stdout.strip()
except Exception as e:
return f"Error: {str(e)}"
def aha_list_features(self, product: str, __user__: dict = {}) -> str:
"""
List features for an Aha! product.
:param product: Aha! product key
:return: JSON list of features
"""
env = os.environ.copy()
env["AHA_DOMAIN"] = self.valves.AHA_DOMAIN
env["AHA_API_TOKEN"] = self.valves.AHA_API_KEY
try:
result = subprocess.run(
[
"mcporter",
"--config", self.valves.MCPORTER_CONFIG,
"call", "aha", "list_features",
"--params", json.dumps({"product": product})
],
capture_output=True, text=True, timeout=30, env=env
)
if result.returncode != 0:
return f"Error: {result.stderr.strip()}"
return result.stdout.strip()
except Exception as e:
return f"Error: {str(e)}"

View File

@@ -0,0 +1,117 @@
"""
title: BMad & Factory Pipeline
author: PM Factory
version: 0.1.0
description: Brainstorm with BMad Creative Suite and validate against Transparent Factory tenets.
requirements: subprocess
"""
import subprocess
import json
import os
from typing import Optional
BMAD_PATH = os.environ.get("BMAD_PATH", "bmad")
FACTORY_PATH = os.environ.get("FACTORY_PATH", ".standards/factory/content")
class Tools:
def __init__(self):
self.valves = self.Valves()
class Valves:
BMAD_PATH: str = BMAD_PATH
FACTORY_PATH: str = FACTORY_PATH
def bmad_list_agents(self, __user__: dict = {}) -> str:
"""
List available BMad Creative Intelligence Suite agents and their capabilities.
:return: List of agents with descriptions
"""
agents = {
"Carson (Brainstorming Coach)": {
"command": "/cis-brainstorm",
"capabilities": "36 ideation techniques, group dynamics, 'Yes, and...' methodology"
},
"Maya (Design Thinking Coach)": {
"command": "/cis-design-thinking",
"capabilities": "Five-phase design thinking, empathy mapping, rapid prototyping"
},
"Victor (Innovation Strategist)": {
"command": "/cis-innovation-strategy",
"capabilities": "Jobs-to-be-Done, Blue Ocean Strategy, Business Model Canvas"
},
"Dr. Quinn (Creative Problem Solver)": {
"command": "/cis-problem-solve",
"capabilities": "Root cause analysis, systematic diagnosis, solution frameworks"
},
"Storyteller": {
"command": "/cis-story",
"capabilities": "PR/FAQ drafting, narrative structure, stakeholder communication"
}
}
return json.dumps(agents, indent=2)
def bmad_brainstorm(self, topic: str, technique: str = "auto", __user__: dict = {}) -> str:
"""
Run a brainstorming session using BMad's Carson agent.
:param topic: The topic or problem to brainstorm about
:param technique: Brainstorming technique (auto, scamper, reverse, starbursting, six-hats, etc). Default: auto
:return: Brainstorming session output
"""
prompt = f"""You are Carson, the Brainstorming Coach from the BMad Creative Intelligence Suite.
Run a brainstorming session on this topic: {topic}
Technique: {technique if technique != 'auto' else 'Choose the best technique for this topic.'}
Generate:
1. 8-12 diverse ideas using the selected technique
2. For each idea: one sentence description + feasibility rating (1-5)
3. Top 3 recommendations with brief rationale
4. One wild/moonshot idea that breaks assumptions
Use "Yes, and..." methodology. Celebrate bold ideas."""
return prompt
def factory_check(self, spec_text: str, __user__: dict = {}) -> str:
"""
Validate a specification or epic against the Transparent Factory tenets.
:param spec_text: The spec, epic, or requirement text to validate
:return: Compliance report with pass/fail per tenet and recommendations
"""
tenets = {
"Atomic Flagging": {
"rule": "All new features must be behind feature flags with 14-day TTL. Must use OpenFeature SDK.",
"check": "Does the spec mention feature flags? Is there a TTL or rollout plan?"
},
"Elastic Schema": {
"rule": "Schema changes must be additive-only. Breaking changes require sync dual-write with 30-day migration SLA.",
"check": "Does the spec propose schema changes? Are they additive? Is there a migration plan?"
},
"Cognitive Durability": {
"rule": "All architectural decisions must have ADR logs. Code must be readable in 60 seconds.",
"check": "Does the spec reference ADRs? Is the proposed design simple enough to explain quickly?"
},
"Semantic Observability": {
"rule": "AI reasoning must emit structured telemetry spans. Decisions must be traceable.",
"check": "Does the spec include observability requirements? Are reasoning spans defined?"
},
"Configurable Autonomy": {
"rule": "AI agent actions must have governance guardrails. Human-in-the-loop for destructive operations.",
"check": "Does the spec define autonomy boundaries? Are there approval gates?"
}
}
report = "# Transparent Factory Compliance Check\n\n"
report += f"**Input:** {spec_text[:200]}{'...' if len(spec_text) > 200 else ''}\n\n"
report += "| Tenet | Rule | Question |\n|-------|------|----------|\n"
for name, t in tenets.items():
report += f"| {name} | {t['rule']} | {t['check']} |\n"
report += "\n*Review each tenet against the spec and flag violations.*"
return report

View File

@@ -0,0 +1,76 @@
"""
title: Gainsight PX Pipeline
author: PM Factory
version: 0.1.0
description: Query Gainsight PX product analytics.
requirements: requests
"""
import os
import json
import requests
from typing import Optional
class Tools:
def __init__(self):
self.valves = self.Valves()
class Valves:
GAINSIGHT_PX_API_KEY: str = os.environ.get("GAINSIGHT_PX_API_KEY", "")
GAINSIGHT_PX_REGION: str = os.environ.get("GAINSIGHT_PX_REGION", "US")
def _base_url(self) -> str:
if self.valves.GAINSIGHT_PX_REGION.upper() == "EU":
return "https://api-eu.aptrinsic.com/v1"
return "https://api.aptrinsic.com/v1"
def _headers(self) -> dict:
return {
"X-APTRINSIC-API-KEY": self.valves.GAINSIGHT_PX_API_KEY,
"Content-Type": "application/json"
}
def gainsight_query_users(self, filter_expression: str = "", page_size: int = 20, __user__: dict = {}) -> str:
"""
Query Gainsight PX users/accounts.
:param filter_expression: Optional filter (e.g. 'propertyName=value')
:param page_size: Number of results to return. Default: 20
:return: JSON string of user data
"""
try:
params = {"pageSize": page_size}
if filter_expression:
params["filter"] = filter_expression
resp = requests.get(
f"{self._base_url()}/users",
headers=self._headers(),
params=params,
timeout=15
)
resp.raise_for_status()
return json.dumps(resp.json(), indent=2)
except Exception as e:
return f"Error: {str(e)}"
def gainsight_feature_usage(self, feature_id: str, days: int = 30, __user__: dict = {}) -> str:
"""
Get feature usage stats from Gainsight PX.
:param feature_id: The feature ID to query
:param days: Lookback period in days. Default: 30
:return: JSON string of usage data
"""
try:
resp = requests.get(
f"{self._base_url()}/feature/{feature_id}/usage",
headers=self._headers(),
params={"days": days},
timeout=15
)
resp.raise_for_status()
return json.dumps(resp.json(), indent=2)
except Exception as e:
return f"Error: {str(e)}"

View File

@@ -0,0 +1,77 @@
"""
title: Jira Pipeline
author: PM Factory
version: 0.1.0
description: Search and create Jira issues via mcporter + Atlassian MCP.
requirements: subprocess
"""
import subprocess
import json
import os
from typing import Optional
MCPORTER_CONFIG = os.environ.get("MCPORTER_CONFIG", "config/mcporter.json")
class Tools:
def __init__(self):
self.valves = self.Valves()
class Valves:
MCPORTER_CONFIG: str = MCPORTER_CONFIG
def jira_search(self, jql: str, __user__: dict = {}) -> str:
"""
Search Jira issues using JQL.
:param jql: JQL query string (e.g. 'project = PLAT AND status = "In Progress"')
:return: JSON string of matching issues
"""
try:
result = subprocess.run(
[
"mcporter",
"--config", self.valves.MCPORTER_CONFIG,
"call", "atlassian", "jira_search",
"--params", json.dumps({"jql": jql})
],
capture_output=True, text=True, timeout=30
)
if result.returncode != 0:
return f"Error: {result.stderr.strip()}"
return result.stdout.strip()
except Exception as e:
return f"Error: {str(e)}"
def jira_create(self, project: str, summary: str, description: str, issue_type: str = "Story", __user__: dict = {}) -> str:
"""
Create a Jira issue.
:param project: Jira project key (e.g. 'PLAT')
:param summary: Issue title
:param description: Issue description (markdown supported)
:param issue_type: Issue type (Story, Bug, Epic, Task). Default: Story
:return: Created issue key and URL
"""
params = {
"project": project,
"summary": summary,
"description": description,
"issueType": issue_type
}
try:
result = subprocess.run(
[
"mcporter",
"--config", self.valves.MCPORTER_CONFIG,
"call", "atlassian", "jira_create_issue",
"--params", json.dumps(params)
],
capture_output=True, text=True, timeout=30
)
if result.returncode != 0:
return f"Error: {result.stderr.strip()}"
return result.stdout.strip()
except Exception as e:
return f"Error: {str(e)}"

17
openwebui/preset.json Normal file
View File

@@ -0,0 +1,17 @@
{
"name": "Reltio PM Factory",
"description": "AI workspace for Product Managers — epics, PR/FAQs, analytics, and Transparent Factory compliance.",
"base_model_id": "claude-opus-4.6",
"params": {
"system": "You are an AI assistant operating inside the **Reltio PM Factory**. You help Product Managers create epics, draft PR/FAQs, query analytics, and maintain compliance with the Transparent Factory standard.\n\n## First Interaction\nWhen the user first messages you, greet them:\n> 👋 **Welcome to the Reltio PM Factory!**\n> I can help you draft PR/FAQs, create Jira Epics, query Gainsight analytics, brainstorm with the Creative Squad, or check Transparent Factory compliance. What would you like to do?\n\n## The Transparent Factory\nYou must adhere to the Reltio Transparent Factory tenets:\n- **Atomic Flagging:** 14-day TTL, OpenFeature-based feature flags.\n- **Elastic Schema:** Additive-only changes, sync dual-write, 30-day migration SLA.\n- **Cognitive Durability:** ADR decision logs, code readable in 60 seconds.\n- **Semantic Observability:** Reasoning spans, structured telemetry.\n- **Configurable Autonomy:** Governance guardrails for AI agents.\n\nIf a proposed spec or epic violates these tenets, flag it and suggest corrections.\n\n## Available Tools\nYou have access to these pipeline tools:\n- **jira_search** / **jira_create** — Search and create Jira issues via Atlassian MCP\n- **aha_create_epic** — Create epics in Aha! with proper field structure\n- **gainsight_query** — Query Gainsight PX for product analytics\n- **bmad_brainstorm** — Run a brainstorming session using BMad Creative Intelligence Suite\n- **factory_check** — Validate a spec against Transparent Factory tenets\n\nUse these tools when the user's request maps to an external action. Always confirm before creating or modifying external resources.\n\n## Style\n- Be concise and direct.\n- Use tables and bullet points for structured data.\n- When drafting epics or PR/FAQs, follow the templates in your knowledge base.\n- If unsure about a tenet, say so rather than guessing.",
"temperature": 0.7,
"max_tokens": 4096
},
"meta": {
"profile_image_url": "🏭",
"capabilities": {
"vision": false
},
"tags": ["pm", "reltio", "factory"]
}
}

View File

@@ -0,0 +1,47 @@
---
name: bmad-suite
description: Manage, update, and deploy BMad workflows/agents.
tools:
- name: bmad_update
description: Pull latest updates or clone missing repositories for BMad Suite.
entry:
type: python
path: manager.py
args: ["update"]
- name: bmad_list
description: List available workflows/agents in the suite.
entry:
type: python
path: manager.py
args: ["list"]
---
# BMad Creative Suite Manager
This skill manages the **BMad Suite** ecosystem, handling installation (git clone) and updates (git pull).
## Capabilities
- **Update/Install:** Automatically clones repositories if missing, or pulls latest changes if present.
- **List:** Enumerates available agents and workflows across all modules.
## Documentation Sources
Refer to these files for detailed usage, architecture, and agent definitions:
### 1. Framework
- **Core Documentation:** `framework/README.md`
- **Agent Definitions:** `framework/src/agents/`
### 2. Creative Intelligence Suite (CIS)
- **Agent Catalog:** `creative-intelligence-suite/docs/reference/agents.md`
- **Main Documentation:** `creative-intelligence-suite/README.md`
- **Agent Definitions:** `creative-intelligence-suite/src/agents/*.agent.yaml`
### 3. Test Architecture Enterprise (TEA)
- **Main Documentation:** `test-architecture-enterprise/README.md`
- **Workflows:** `test-architecture-enterprise/src/workflows/testarch/README.md`
## Repositories
Managed repositories (auto-cloned to `../../bmad/` relative to this skill, or `$BMAD_PATH`):
1. **Framework:** `bmad-code-org/BMAD-METHOD`
2. **Creative Intelligence Suite:** `bmad-code-org/bmad-module-creative-intelligence-suite`
3. **Test Architecture Enterprise (TEA):** `bmad-code-org/bmad-method-test-architecture-enterprise`

View File

@@ -0,0 +1,93 @@
#!/usr/bin/env python3
import os
import sys
import subprocess
# Determine base path: use BMAD_PATH env var, or default to ../../bmad relative to this script
# transparent_factory_site/skills/bmad-suite/manager.py -> transparent_factory_site/bmad
SCRIPT_DIR = os.path.dirname(os.path.abspath(__file__))
DEFAULT_BMAD_PATH = os.path.abspath(os.path.join(SCRIPT_DIR, "../../bmad"))
BASE_PATH = os.environ.get("BMAD_PATH", DEFAULT_BMAD_PATH)
REPOS = {
"framework": {
"url": "https://github.com/bmad-code-org/BMAD-METHOD.git",
"path": os.path.join(BASE_PATH, "framework")
},
"creative-suite": {
"url": "https://github.com/bmad-code-org/bmad-module-creative-intelligence-suite.git",
"path": os.path.join(BASE_PATH, "creative-intelligence-suite")
},
"tea-module": {
"url": "https://github.com/bmad-code-org/bmad-method-test-architecture-enterprise.git",
"path": os.path.join(BASE_PATH, "test-architecture-enterprise")
}
}
def update_or_clone(name, config):
"""Clone if missing, pull if present."""
repo_path = config["path"]
repo_url = config["url"]
# Ensure parent directory exists so we can clone into it if needed
parent_dir = os.path.dirname(repo_path)
if not os.path.exists(parent_dir):
os.makedirs(parent_dir, exist_ok=True)
# Check if it's already a git repo
if os.path.exists(os.path.join(repo_path, ".git")):
print(f"🔄 Updating {name}...")
try:
subprocess.run(["git", "pull"], cwd=repo_path, check=True)
print(f"{name} updated.")
except subprocess.CalledProcessError as e:
print(f"{name} update failed: {e}")
# Check if directory exists but is empty (safe to clone into)
elif os.path.exists(repo_path) and not os.listdir(repo_path):
print(f"📥 Cloning {name} into empty directory...")
try:
subprocess.run(["git", "clone", repo_url, "."], cwd=repo_path, check=True)
print(f"{name} cloned.")
except subprocess.CalledProcessError as e:
print(f"{name} clone failed: {e}")
# Directory doesn't exist at all
elif not os.path.exists(repo_path):
print(f"📥 Cloning {name}...")
try:
subprocess.run(["git", "clone", repo_url, repo_path], check=True)
print(f"{name} cloned.")
except subprocess.CalledProcessError as e:
print(f"{name} clone failed: {e}")
else:
print(f"⚠️ Target directory {repo_path} exists and is not empty (and not a git repo). Skipping.")
def list_workflows(suite_path):
"""List available workflows/agents in the suite."""
src_path = os.path.join(suite_path, "src")
if not os.path.exists(src_path):
# Fallback to listing root if src doesn't exist (e.g. some repos might differ)
if os.path.exists(suite_path):
return subprocess.getoutput(f"find {suite_path} -maxdepth 3 -name '*.md' -o -name '*.ts' -o -name '*.js' | grep -v 'node_modules' | sort")
return "Directory not found."
return subprocess.getoutput(f"find {src_path} -name '*.md' -o -name '*.ts' -o -name '*.js' -o -name '*.yaml' | sort")
if __name__ == "__main__":
action = sys.argv[1] if len(sys.argv) > 1 else "list"
if action == "update":
print("--- Checking BMad Suite Repositories ---")
for name, config in REPOS.items():
update_or_clone(name, config)
print("")
elif action == "list":
for name, config in REPOS.items():
print(f"--- {name} Workflows ---")
print(list_workflows(config["path"]))
print("")
else:
print(f"Unknown action: {action}")

View File

@@ -0,0 +1,74 @@
---
name: epics-standards
description: Create or audit RP Jira epics and linked Aha epics against PM standards. Use when creating new RP epics, creating Aha epics for a target release, checking compliance gaps, or updating epic fields/content to align with the PM standards and Aha workflow.
---
# RP Epic Standards
Use this skill for `RP` epic creation, Aha epic creation, and compliance audits.
Primary standard source:
`https://reltio.jira.com/wiki/spaces/PM/pages/2688385025/PM+Standards+Epics`
Workflow source and mapping:
- `references/aha-epic-workflow.md`
- `references/epic-fields-checklist.md`
## Workflow
1. Open the standard page and use it as source of truth.
2. Validate/collect required inputs (see Intake below).
3. Create/update Jira epic and Aha epic per workflow in `references/aha-epic-workflow.md`.
4. Link Jira and Aha (`Aha! Reference` in Jira).
5. Audit compliance against `references/epic-fields-checklist.md`.
6. If fields are not editable in Jira/Aha, document exact gaps and owner.
## Intake (Interactive)
If required fields are not provided in the prompt, ask concise follow-up questions before creating records.
Minimum required fields to ask for:
- Jira epic key (if already created) or request to create one
- Aha release (for example `2026.2.0.0`)
- Epic short name (Jira/Aha title)
- Problem Statement / Why
- Solution / What
- Persona / Who
- Value Statement
- Confidence Level
- Product, Engineering, and UX leads
- Execution Team
- Required flags: Doc, Provisioning, UI/UX, Security, Training
Ask only for missing items. Do not proceed with creation until minimum fields are available.
## Epic Content Requirements
Ensure description includes:
- Problem Statement / Why
- Solution / What
- Persona / Who
- Value Statement
Keep summary short and clear for Jira readability.
## Aha Creation Code
Use the bundled script:
- `scripts/aha_create_epic.py`
Example:
```bash
python3 skills/epics-standards/scripts/aha_create_epic.py \
--release MDM-R-889 \
--name "RDM PrivateLink on AWS" \
--description "Tracks Jira epic RP-176273 (https://reltio.jira.com/browse/RP-176273)." \
--jira-key RP-176273
```
The script reads Aha credentials from `~/.mcporter/mcporter.json` (`mcpServers.aha.env`).
## Integration Guardrails
If the epic is created in Jira first, verify that Aha linkage is present (`Aha! Reference`).
If missing, update Jira with the created Aha URL.

View File

@@ -0,0 +1,4 @@
interface:
display_name: "RP Epic Standards"
short_description: "Create and audit RP epics for PM standards"
default_prompt: "Use $epics-standards to create or audit linked Jira/Aha epics and ask me for any missing required fields before creating records."

View File

@@ -0,0 +1,55 @@
# Aha Epic Workflow
This file captures the expected Aha epic lifecycle and downstream triggers.
## Idea Management
1. Customer submits idea in Aha.
2. Check voting threshold: `5` unique orgs.
3. PM review required (`Needs Review` -> `Reviewed`).
4. PM decision:
- Accept current release -> `Planned`
- Accept next release -> `Future Consideration`
- Exists -> `Already Exists`
- Need more info -> `Needs More Info` + comment
- Reject/later -> `Future Consideration` + comment
5. Always add public customer comment after decision.
6. Ensure epic relation:
- Promote idea to new epic, or
- Link idea to existing epic.
## Epic Creation and Management
When epic exists in Aha, fill mandatory fields:
- Summary with PRD linkage
- Release and availability (`Preview`/`GA`)
- Confidence level (`High`/`Med`/`Low`)
- Product type and persona
- Execution team and initiative
- Product lead, engineering lead, UX lead
Set required flags and trigger follow-ups:
- `Doc Required = Yes` -> specify doc type and doc-team flow
- `Provisioning = Yes` -> PCC/Olga flow
- `UI/UX Required = Yes` -> engage UX
- `Security Review = Yes` -> engage security
- `Training Required = Yes` -> engage training
Integration:
- Ensure Aha -> Jira Webhooks 2.0 integration path is respected.
- Ensure Jira epic has `Aha! Reference`.
## Enablement Outputs
If applicable, drive enablement:
- Technical enablement session (config)
- GTM/sales enablement artifacts
- Webinar for major highlights
Template links from workflow:
- Technical Enablement Session template:
`https://docs.google.com/presentation/d/1fCZhOUSV7McX1edmYoKBHYtKnYajykbm1U2N5aJ1j-M/edit?slide=id.g39258ed0d71_0_442`
- Value Statements input template:
`https://docs.google.com/document/d/1YEquYIjt8gMtGLf8EJFfvwS0f_ij1KuIfQFOjlcOEjI/edit`
- Sales Enablement deck example:
`https://docs.google.com/presentation/d/1mIlC3OhhQgdwcFPgJ328pm1oQl5W6y-w/edit`

View File

@@ -0,0 +1,53 @@
# RP Epic Field Checklist
Standard page:
`https://reltio.jira.com/wiki/spaces/PM/pages/2688385025/PM+Standards+Epics`
Use this checklist during epic create/update.
## Core fields
- `Summary` (short Jira shorthand)
- `Description` (business outcome + impact)
- `Type of Effort`
- `Fix Version` (or `N/A` when not yet planned)
- `Status`
## Product narrative fields
- Problem Statement / Why
- Solution / What
- Persona / Who
- Value Statement
## Planning and delivery fields
- Confidence Level
- Path to Green (required when Confidence is Medium/Low for must-have epics)
- Availability in Release
- Planned Release
- T-Shirt Size
- Tier
- Initiative
## Ownership and dependencies
- Product Lead
- Engineering Lead
- UX Lead
- Execution Team
- Execution Team Dependency
## Go-to-market and governance
- For documentation tickets: set `Affected Documentation` to a documentation target (for example `Help Portal`) and never `No`
- Doc Required
- Pricing Required
- Security Review Required
- Does this change PCC?
- Tags (including `Must Have` when applicable)
- Demo Link (when available)
## Integration check
- `Aha! Reference` is present and linked through the RP integration path.

View File

@@ -0,0 +1,103 @@
#!/usr/bin/env python3
"""
Create an Aha epic in a target release using credentials from ~/.mcporter/mcporter.json.
Usage:
python3 skills/epics-standards/scripts/aha_create_epic.py \
--release MDM-R-889 \
--name "RDM PrivateLink on AWS" \
--description "Tracks Jira epic RP-176273 (...)" \
--jira-key RP-176273
"""
from __future__ import annotations
import argparse
import json
import pathlib
import sys
import urllib.error
import urllib.parse
import urllib.request
MCPORTER_CONFIG = pathlib.Path.home() / ".mcporter" / "mcporter.json"
def load_aha_env() -> dict[str, str]:
cfg = json.loads(MCPORTER_CONFIG.read_text())
env = cfg["mcpServers"]["aha"]["env"]
return {"domain": env["AHA_DOMAIN"], "token": env["AHA_API_TOKEN"]}
def request(method: str, url: str, token: str, payload: dict | None = None) -> dict:
headers = {
"Authorization": f"Bearer {token}",
"Accept": "application/json",
"Content-Type": "application/json",
}
body = None if payload is None else json.dumps(payload).encode("utf-8")
req = urllib.request.Request(url, data=body, headers=headers, method=method)
with urllib.request.urlopen(req, timeout=30) as resp:
return json.loads(resp.read().decode("utf-8"))
def release_ref_from_name(domain: str, token: str, release_name: str) -> str:
q = urllib.parse.quote(release_name)
url = f"https://{domain}.aha.io/api/v1/releases?q={q}&per_page=200"
data = request("GET", url, token)
for rel in data.get("releases", []):
if rel.get("name") == release_name:
return rel["reference_num"]
raise ValueError(f"Release not found by name: {release_name}")
def main() -> int:
parser = argparse.ArgumentParser()
parser.add_argument("--release", required=True, help="Release reference (e.g. MDM-R-889) or exact release name (e.g. 2026.2.0.0)")
parser.add_argument("--name", required=True, help="Aha epic name")
parser.add_argument("--description", required=True, help="Aha epic description/body")
parser.add_argument("--jira-key", required=False, help="Optional Jira key to print linking reminder")
args = parser.parse_args()
try:
aha = load_aha_env()
release = args.release
if not release.startswith("MDM-R-"):
release = release_ref_from_name(aha["domain"], aha["token"], release)
url = f"https://{aha['domain']}.aha.io/api/v1/releases/{release}/epics"
payload = {"epic": {"name": args.name, "description": args.description}}
data = request("POST", url, aha["token"], payload)
epic = data.get("epic", {})
print(json.dumps(
{
"aha_reference": epic.get("reference_num"),
"aha_url": epic.get("url"),
"release": release,
"jira_key": args.jira_key,
},
ensure_ascii=True,
))
if args.jira_key and epic.get("url"):
print(
f"Next: set Jira {args.jira_key} Aha! Reference = {epic['url']}",
file=sys.stderr,
)
return 0
except (KeyError, FileNotFoundError, ValueError) as exc:
print(f"Config/input error: {exc}", file=sys.stderr)
return 2
except urllib.error.HTTPError as exc:
detail = exc.read().decode("utf-8", errors="replace")
print(f"Aha API HTTP {exc.code}: {detail}", file=sys.stderr)
return 3
except Exception as exc: # noqa: BLE001
print(f"Unexpected error: {exc}", file=sys.stderr)
return 1
if __name__ == "__main__":
raise SystemExit(main())

View File

@@ -0,0 +1,31 @@
---
name: factory-standards
description: Manage and access the Transparent Factory engineering and product tenets from Bitbucket.
tools:
- name: factory_update
description: Clone or pull the latest Transparent Factory rules and PR/FAQ from the central repository.
entry:
type: python
path: manager.py
args: ["update"]
- name: factory_list
description: List available tenets, standards, and guidelines in the Transparent Factory repository.
entry:
type: python
path: manager.py
args: ["list"]
---
# Transparent Factory Standards Manager
This skill synchronizes the definitive **Transparent Factory** architectural and product rules into your local project from the central Bitbucket repository.
## Capabilities
- **Update/Install:** Automatically clones `reltio-ondemand/transparent-factory.git` if missing, or pulls latest changes if present.
- **List:** Enumerates available tenets (`content/`), PR/FAQ documents, and exported factory skills.
## Architecture
Managed files are auto-cloned to `.standards/factory/` relative to the root of your project workspace (e.g., if this skill is in `skills/factory-standards/`, it clones to `../../.standards/factory/`).
## How to use in AI Prompts
If you are an AI agent, you must run `factory_update` to ensure the local `.standards/factory/` directory is present and up-to-date before writing any Product Requirements Documents, Epics, or technical specifications.

View File

@@ -0,0 +1,89 @@
#!/usr/bin/env python3
import os
import sys
import subprocess
# Determine base path: use FACTORY_PATH env var, or default to ../../.standards/factory relative to this script
SCRIPT_DIR = os.path.dirname(os.path.abspath(__file__))
DEFAULT_FACTORY_PATH = os.path.abspath(os.path.join(SCRIPT_DIR, "../../.standards/factory"))
FACTORY_PATH = os.environ.get("FACTORY_PATH", DEFAULT_FACTORY_PATH)
REPO_URL = "git@bitbucket.org:reltio-ondemand/transparent-factory.git"
def ensure_directory(path):
"""Ensure the parent directory exists."""
parent = os.path.dirname(path)
if not os.path.exists(parent):
print(f"📂 Creating directory: {parent}")
os.makedirs(parent, exist_ok=True)
def update_or_clone():
"""Clone the Transparent Factory repo if missing, pull if present."""
ensure_directory(FACTORY_PATH)
# Check if it's already a git repo
if os.path.exists(os.path.join(FACTORY_PATH, ".git")):
print(f"🔄 Updating Transparent Factory Standards...")
try:
subprocess.run(["git", "pull"], cwd=FACTORY_PATH, check=True)
print(f"✅ Standards updated at {FACTORY_PATH}.")
except subprocess.CalledProcessError as e:
print(f"❌ Update failed: {e}")
# Check if directory exists but is empty (safe to clone into)
elif os.path.exists(FACTORY_PATH) and not os.listdir(FACTORY_PATH):
print(f"📥 Cloning Transparent Factory Standards into empty directory...")
try:
subprocess.run(["git", "clone", REPO_URL, "."], cwd=FACTORY_PATH, check=True)
print(f"✅ Standards cloned to {FACTORY_PATH}.")
except subprocess.CalledProcessError as e:
print(f"❌ Clone failed: {e}")
# Directory doesn't exist at all
elif not os.path.exists(FACTORY_PATH):
print(f"📥 Cloning Transparent Factory Standards...")
try:
subprocess.run(["git", "clone", REPO_URL, FACTORY_PATH], check=True)
print(f"✅ Standards cloned to {FACTORY_PATH}.")
except subprocess.CalledProcessError as e:
print(f"❌ Clone failed: {e}")
else:
print(f"⚠️ Target directory {FACTORY_PATH} exists and is not empty (and not a git repo). Skipping.")
def list_standards():
"""List available standards and PR/FAQ documents."""
if not os.path.exists(FACTORY_PATH):
return "Standards not found. Run 'factory_update' first."
print("--- Core Documents ---")
try:
# Look for the primary PR/FAQ or README
print(subprocess.getoutput(f"find {FACTORY_PATH} -maxdepth 1 -name '*.md' | sort"))
# Look inside the content folder
content_path = os.path.join(FACTORY_PATH, "content")
if os.path.exists(content_path):
print("\n--- Tenets & Content ---")
print(subprocess.getoutput(f"find {content_path} -name '*.md' | sort"))
# Look inside skills folder
skills_path = os.path.join(FACTORY_PATH, "skills")
if os.path.exists(skills_path):
print("\n--- Available Factory Skills ---")
print(subprocess.getoutput(f"find {skills_path} -name 'SKILL.md' | sort"))
except Exception as e:
return f"Error listing files: {e}"
return ""
if __name__ == "__main__":
action = sys.argv[1] if len(sys.argv) > 1 else "list"
if action == "update":
update_or_clone()
elif action == "list":
print(list_standards())
else:
print(f"Unknown action: {action}")
sys.exit(1)

View File

@@ -0,0 +1,54 @@
---
name: gainsight-px
description: Interact directly with the Gainsight PX REST API to fetch user/account data or track events.
tools:
- name: px_get_user
description: Fetch a specific user by their unique identity ID.
entry:
type: python
path: gainsight_px.py
args: ["get_user"]
- name: px_get_account
description: Fetch a specific account by its ID.
entry:
type: python
path: gainsight_px.py
args: ["get_account"]
- name: px_search_user
description: Search for a user in Gainsight PX by their email address.
entry:
type: python
path: gainsight_px.py
args: ["search_user"]
- name: px_track_event
description: Track a custom event for a user in Gainsight PX. Requires user_id, event_name, and optional JSON properties.
entry:
type: python
path: gainsight_px.py
args: ["track_event"]
---
# Gainsight PX REST API Skill
This skill allows agents to natively interface with your Gainsight PX instance without needing an intermediate MCP server like Pipedream or Zapier.
## Setup
You must export your API key before using the tools. You can generate an API key from your Gainsight PX Administration -> REST API section.
```bash
# Add this to your environment
export GAINSIGHT_PX_API_KEY="your-api-key-here"
# Optional: If you are in the EU region, set this flag. Default is US.
export GAINSIGHT_PX_REGION="EU"
```
## How It Works
It uses a lightweight Python script (`gainsight_px.py`) that implements standard REST endpoints documented by Apiary (`https://api.aptrinsic.com/v1/...`).
### Capabilities
- **Lookups:** Find exactly who a user is by ID or email.
- **Account Context:** Pull account metadata.
- **Event Injection:** Push arbitrary telemetry events natively.

View File

@@ -0,0 +1,108 @@
#!/usr/bin/env python3
import os
import sys
import json
import urllib.request
import urllib.error
# Gainsight PX API configuration
# Region determines the base URL (US or EU)
PX_REGION = os.environ.get("GAINSIGHT_PX_REGION", "US").upper()
PX_API_KEY = os.environ.get("GAINSIGHT_PX_API_KEY")
if PX_REGION == "EU":
BASE_URL = "https://eu-api.aptrinsic.com/v1"
else:
BASE_URL = "https://api.aptrinsic.com/v1"
def make_request(method, endpoint, data=None):
if not PX_API_KEY:
print(json.dumps({"error": "GAINSIGHT_PX_API_KEY environment variable is missing."}))
sys.exit(1)
url = f"{BASE_URL}{endpoint}"
headers = {
"X-APITOKEN": PX_API_KEY,
"Content-Type": "application/json",
"Accept": "application/json"
}
req_data = None
if data:
req_data = json.dumps(data).encode("utf-8")
req = urllib.request.Request(url, data=req_data, headers=headers, method=method)
try:
with urllib.request.urlopen(req) as response:
return json.loads(response.read().decode("utf-8"))
except urllib.error.HTTPError as e:
err_msg = e.read().decode("utf-8")
try:
parsed_err = json.loads(err_msg)
return {"error": f"HTTP {e.code}", "details": parsed_err}
except:
return {"error": f"HTTP {e.code}", "details": err_msg}
except Exception as e:
return {"error": str(e)}
def get_user(user_id):
"""Retrieve a specific user by their identifyId."""
return make_request("GET", f"/users/{user_id}")
def get_account(account_id):
"""Retrieve a specific account by its id."""
return make_request("GET", f"/accounts/{account_id}")
def search_users(email):
"""Search for users by email (requires query payload)."""
payload = {
"filter": {
"operator": "AND",
"conditions": [
{
"name": "email",
"operator": "EQ",
"value": email
}
]
}
}
return make_request("POST", "/users/query", data=payload)
def track_event(user_id, event_name, properties=None):
"""Track a custom event for a specific user."""
payload = {
"identifyId": user_id,
"eventName": event_name,
"properties": properties or {}
}
# Note: tracking usually happens via a different endpoint or batch API,
# but for simplicity assuming a standard REST event ingestion if available.
return make_request("POST", "/events/custom", data=payload)
if __name__ == "__main__":
if len(sys.argv) < 2:
print(json.dumps({"error": "Missing action. Use: get_user, get_account, search_user, track_event"}))
sys.exit(1)
action = sys.argv[1]
if action == "get_user" and len(sys.argv) == 3:
print(json.dumps(get_user(sys.argv[2]), indent=2))
elif action == "get_account" and len(sys.argv) == 3:
print(json.dumps(get_account(sys.argv[2]), indent=2))
elif action == "search_user" and len(sys.argv) == 3:
print(json.dumps(search_users(sys.argv[2]), indent=2))
elif action == "track_event" and len(sys.argv) >= 4:
user_id = sys.argv[2]
event_name = sys.argv[3]
props = json.loads(sys.argv[4]) if len(sys.argv) > 4 else {}
print(json.dumps(track_event(user_id, event_name, props), indent=2))
else:
print(json.dumps({"error": f"Unknown action or missing arguments: {action}"}))
sys.exit(1)

54
start.sh Executable file
View File

@@ -0,0 +1,54 @@
#!/usr/bin/env bash
set -e
SCRIPT_DIR="$(cd "$(dirname "$0")" && pwd)"
cd "$SCRIPT_DIR"
# --- Check Node.js ---
if ! command -v node &>/dev/null; then
echo "❌ Node.js is not installed."
echo " Install it from https://nodejs.org (LTS recommended)"
exit 1
fi
echo "✅ Node.js $(node -v)"
# --- Check .env ---
if [ ! -f .env ]; then
if [ -f env.example ]; then
cp env.example .env
echo "⚠️ Created .env from template. Please edit it with your API keys, then re-run this script."
exit 1
else
echo "⚠️ No .env or env.example found. Continuing without environment config."
fi
fi
# --- Install web dependencies ---
if [ ! -d web/node_modules ]; then
echo "📦 Installing dependencies..."
cd web && npm install && cd ..
fi
# --- Launch ---
echo ""
echo "🔥 Starting Forge Console..."
echo ""
# Open browser after a short delay
(sleep 2 && {
URL="http://localhost:${PORT:-3000}"
if command -v xdg-open &>/dev/null; then
xdg-open "$URL" 2>/dev/null
elif command -v open &>/dev/null; then
open "$URL"
elif command -v start &>/dev/null; then
start "$URL"
else
echo " Open $URL in your browser"
fi
}) &
# Run server (trap for clean exit)
trap 'echo ""; echo "👋 Shutting down..."; kill %1 2>/dev/null; exit 0' INT TERM
cd web && node server.js

849
web/package-lock.json generated Normal file
View File

@@ -0,0 +1,849 @@
{
"name": "pm-template-web",
"version": "1.0.0",
"lockfileVersion": 3,
"requires": true,
"packages": {
"": {
"name": "pm-template-web",
"version": "1.0.0",
"dependencies": {
"express": "^4.18.2",
"ws": "^8.16.0"
}
},
"node_modules/accepts": {
"version": "1.3.8",
"resolved": "https://registry.npmjs.org/accepts/-/accepts-1.3.8.tgz",
"integrity": "sha512-PYAthTa2m2VKxuvSD3DPC/Gy+U+sOA1LAuT8mkmRuvw+NACSaeXEQ+NHcVF7rONl6qcaxV3Uuemwawk+7+SJLw==",
"license": "MIT",
"dependencies": {
"mime-types": "~2.1.34",
"negotiator": "0.6.3"
},
"engines": {
"node": ">= 0.6"
}
},
"node_modules/array-flatten": {
"version": "1.1.1",
"resolved": "https://registry.npmjs.org/array-flatten/-/array-flatten-1.1.1.tgz",
"integrity": "sha512-PCVAQswWemu6UdxsDFFX/+gVeYqKAod3D3UVm91jHwynguOwAvYPhx8nNlM++NqRcK6CxxpUafjmhIdKiHibqg==",
"license": "MIT"
},
"node_modules/body-parser": {
"version": "1.20.4",
"resolved": "https://registry.npmjs.org/body-parser/-/body-parser-1.20.4.tgz",
"integrity": "sha512-ZTgYYLMOXY9qKU/57FAo8F+HA2dGX7bqGc71txDRC1rS4frdFI5R7NhluHxH6M0YItAP0sHB4uqAOcYKxO6uGA==",
"license": "MIT",
"dependencies": {
"bytes": "~3.1.2",
"content-type": "~1.0.5",
"debug": "2.6.9",
"depd": "2.0.0",
"destroy": "~1.2.0",
"http-errors": "~2.0.1",
"iconv-lite": "~0.4.24",
"on-finished": "~2.4.1",
"qs": "~6.14.0",
"raw-body": "~2.5.3",
"type-is": "~1.6.18",
"unpipe": "~1.0.0"
},
"engines": {
"node": ">= 0.8",
"npm": "1.2.8000 || >= 1.4.16"
}
},
"node_modules/bytes": {
"version": "3.1.2",
"resolved": "https://registry.npmjs.org/bytes/-/bytes-3.1.2.tgz",
"integrity": "sha512-/Nf7TyzTx6S3yRJObOAV7956r8cr2+Oj8AC5dt8wSP3BQAoeX58NoHyCU8P8zGkNXStjTSi6fzO6F0pBdcYbEg==",
"license": "MIT",
"engines": {
"node": ">= 0.8"
}
},
"node_modules/call-bind-apply-helpers": {
"version": "1.0.2",
"resolved": "https://registry.npmjs.org/call-bind-apply-helpers/-/call-bind-apply-helpers-1.0.2.tgz",
"integrity": "sha512-Sp1ablJ0ivDkSzjcaJdxEunN5/XvksFJ2sMBFfq6x0ryhQV/2b/KwFe21cMpmHtPOSij8K99/wSfoEuTObmuMQ==",
"license": "MIT",
"dependencies": {
"es-errors": "^1.3.0",
"function-bind": "^1.1.2"
},
"engines": {
"node": ">= 0.4"
}
},
"node_modules/call-bound": {
"version": "1.0.4",
"resolved": "https://registry.npmjs.org/call-bound/-/call-bound-1.0.4.tgz",
"integrity": "sha512-+ys997U96po4Kx/ABpBCqhA9EuxJaQWDQg7295H4hBphv3IZg0boBKuwYpt4YXp6MZ5AmZQnU/tyMTlRpaSejg==",
"license": "MIT",
"dependencies": {
"call-bind-apply-helpers": "^1.0.2",
"get-intrinsic": "^1.3.0"
},
"engines": {
"node": ">= 0.4"
},
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/content-disposition": {
"version": "0.5.4",
"resolved": "https://registry.npmjs.org/content-disposition/-/content-disposition-0.5.4.tgz",
"integrity": "sha512-FveZTNuGw04cxlAiWbzi6zTAL/lhehaWbTtgluJh4/E95DqMwTmha3KZN1aAWA8cFIhHzMZUvLevkw5Rqk+tSQ==",
"license": "MIT",
"dependencies": {
"safe-buffer": "5.2.1"
},
"engines": {
"node": ">= 0.6"
}
},
"node_modules/content-type": {
"version": "1.0.5",
"resolved": "https://registry.npmjs.org/content-type/-/content-type-1.0.5.tgz",
"integrity": "sha512-nTjqfcBFEipKdXCv4YDQWCfmcLZKm81ldF0pAopTvyrFGVbcR6P/VAAd5G7N+0tTr8QqiU0tFadD6FK4NtJwOA==",
"license": "MIT",
"engines": {
"node": ">= 0.6"
}
},
"node_modules/cookie": {
"version": "0.7.2",
"resolved": "https://registry.npmjs.org/cookie/-/cookie-0.7.2.tgz",
"integrity": "sha512-yki5XnKuf750l50uGTllt6kKILY4nQ1eNIQatoXEByZ5dWgnKqbnqmTrBE5B4N7lrMJKQ2ytWMiTO2o0v6Ew/w==",
"license": "MIT",
"engines": {
"node": ">= 0.6"
}
},
"node_modules/cookie-signature": {
"version": "1.0.7",
"resolved": "https://registry.npmjs.org/cookie-signature/-/cookie-signature-1.0.7.tgz",
"integrity": "sha512-NXdYc3dLr47pBkpUCHtKSwIOQXLVn8dZEuywboCOJY/osA0wFSLlSawr3KN8qXJEyX66FcONTH8EIlVuK0yyFA==",
"license": "MIT"
},
"node_modules/debug": {
"version": "2.6.9",
"resolved": "https://registry.npmjs.org/debug/-/debug-2.6.9.tgz",
"integrity": "sha512-bC7ElrdJaJnPbAP+1EotYvqZsb3ecl5wi6Bfi6BJTUcNowp6cvspg0jXznRTKDjm/E7AdgFBVeAPVMNcKGsHMA==",
"license": "MIT",
"dependencies": {
"ms": "2.0.0"
}
},
"node_modules/depd": {
"version": "2.0.0",
"resolved": "https://registry.npmjs.org/depd/-/depd-2.0.0.tgz",
"integrity": "sha512-g7nH6P6dyDioJogAAGprGpCtVImJhpPk/roCzdb3fIh61/s/nPsfR6onyMwkCAR/OlC3yBC0lESvUoQEAssIrw==",
"license": "MIT",
"engines": {
"node": ">= 0.8"
}
},
"node_modules/destroy": {
"version": "1.2.0",
"resolved": "https://registry.npmjs.org/destroy/-/destroy-1.2.0.tgz",
"integrity": "sha512-2sJGJTaXIIaR1w4iJSNoN0hnMY7Gpc/n8D4qSCJw8QqFWXf7cuAgnEHxBpweaVcPevC2l3KpjYCx3NypQQgaJg==",
"license": "MIT",
"engines": {
"node": ">= 0.8",
"npm": "1.2.8000 || >= 1.4.16"
}
},
"node_modules/dunder-proto": {
"version": "1.0.1",
"resolved": "https://registry.npmjs.org/dunder-proto/-/dunder-proto-1.0.1.tgz",
"integrity": "sha512-KIN/nDJBQRcXw0MLVhZE9iQHmG68qAVIBg9CqmUYjmQIhgij9U5MFvrqkUL5FbtyyzZuOeOt0zdeRe4UY7ct+A==",
"license": "MIT",
"dependencies": {
"call-bind-apply-helpers": "^1.0.1",
"es-errors": "^1.3.0",
"gopd": "^1.2.0"
},
"engines": {
"node": ">= 0.4"
}
},
"node_modules/ee-first": {
"version": "1.1.1",
"resolved": "https://registry.npmjs.org/ee-first/-/ee-first-1.1.1.tgz",
"integrity": "sha512-WMwm9LhRUo+WUaRN+vRuETqG89IgZphVSNkdFgeb6sS/E4OrDIN7t48CAewSHXc6C8lefD8KKfr5vY61brQlow==",
"license": "MIT"
},
"node_modules/encodeurl": {
"version": "2.0.0",
"resolved": "https://registry.npmjs.org/encodeurl/-/encodeurl-2.0.0.tgz",
"integrity": "sha512-Q0n9HRi4m6JuGIV1eFlmvJB7ZEVxu93IrMyiMsGC0lrMJMWzRgx6WGquyfQgZVb31vhGgXnfmPNNXmxnOkRBrg==",
"license": "MIT",
"engines": {
"node": ">= 0.8"
}
},
"node_modules/es-define-property": {
"version": "1.0.1",
"resolved": "https://registry.npmjs.org/es-define-property/-/es-define-property-1.0.1.tgz",
"integrity": "sha512-e3nRfgfUZ4rNGL232gUgX06QNyyez04KdjFrF+LTRoOXmrOgFKDg4BCdsjW8EnT69eqdYGmRpJwiPVYNrCaW3g==",
"license": "MIT",
"engines": {
"node": ">= 0.4"
}
},
"node_modules/es-errors": {
"version": "1.3.0",
"resolved": "https://registry.npmjs.org/es-errors/-/es-errors-1.3.0.tgz",
"integrity": "sha512-Zf5H2Kxt2xjTvbJvP2ZWLEICxA6j+hAmMzIlypy4xcBg1vKVnx89Wy0GbS+kf5cwCVFFzdCFh2XSCFNULS6csw==",
"license": "MIT",
"engines": {
"node": ">= 0.4"
}
},
"node_modules/es-object-atoms": {
"version": "1.1.1",
"resolved": "https://registry.npmjs.org/es-object-atoms/-/es-object-atoms-1.1.1.tgz",
"integrity": "sha512-FGgH2h8zKNim9ljj7dankFPcICIK9Cp5bm+c2gQSYePhpaG5+esrLODihIorn+Pe6FGJzWhXQotPv73jTaldXA==",
"license": "MIT",
"dependencies": {
"es-errors": "^1.3.0"
},
"engines": {
"node": ">= 0.4"
}
},
"node_modules/escape-html": {
"version": "1.0.3",
"resolved": "https://registry.npmjs.org/escape-html/-/escape-html-1.0.3.tgz",
"integrity": "sha512-NiSupZ4OeuGwr68lGIeym/ksIZMJodUGOSCZ/FSnTxcrekbvqrgdUxlJOMpijaKZVjAJrWrGs/6Jy8OMuyj9ow==",
"license": "MIT"
},
"node_modules/etag": {
"version": "1.8.1",
"resolved": "https://registry.npmjs.org/etag/-/etag-1.8.1.tgz",
"integrity": "sha512-aIL5Fx7mawVa300al2BnEE4iNvo1qETxLrPI/o05L7z6go7fCw1J6EQmbK4FmJ2AS7kgVF/KEZWufBfdClMcPg==",
"license": "MIT",
"engines": {
"node": ">= 0.6"
}
},
"node_modules/express": {
"version": "4.22.1",
"resolved": "https://registry.npmjs.org/express/-/express-4.22.1.tgz",
"integrity": "sha512-F2X8g9P1X7uCPZMA3MVf9wcTqlyNp7IhH5qPCI0izhaOIYXaW9L535tGA3qmjRzpH+bZczqq7hVKxTR4NWnu+g==",
"license": "MIT",
"dependencies": {
"accepts": "~1.3.8",
"array-flatten": "1.1.1",
"body-parser": "~1.20.3",
"content-disposition": "~0.5.4",
"content-type": "~1.0.4",
"cookie": "~0.7.1",
"cookie-signature": "~1.0.6",
"debug": "2.6.9",
"depd": "2.0.0",
"encodeurl": "~2.0.0",
"escape-html": "~1.0.3",
"etag": "~1.8.1",
"finalhandler": "~1.3.1",
"fresh": "~0.5.2",
"http-errors": "~2.0.0",
"merge-descriptors": "1.0.3",
"methods": "~1.1.2",
"on-finished": "~2.4.1",
"parseurl": "~1.3.3",
"path-to-regexp": "~0.1.12",
"proxy-addr": "~2.0.7",
"qs": "~6.14.0",
"range-parser": "~1.2.1",
"safe-buffer": "5.2.1",
"send": "~0.19.0",
"serve-static": "~1.16.2",
"setprototypeof": "1.2.0",
"statuses": "~2.0.1",
"type-is": "~1.6.18",
"utils-merge": "1.0.1",
"vary": "~1.1.2"
},
"engines": {
"node": ">= 0.10.0"
},
"funding": {
"type": "opencollective",
"url": "https://opencollective.com/express"
}
},
"node_modules/finalhandler": {
"version": "1.3.2",
"resolved": "https://registry.npmjs.org/finalhandler/-/finalhandler-1.3.2.tgz",
"integrity": "sha512-aA4RyPcd3badbdABGDuTXCMTtOneUCAYH/gxoYRTZlIJdF0YPWuGqiAsIrhNnnqdXGswYk6dGujem4w80UJFhg==",
"license": "MIT",
"dependencies": {
"debug": "2.6.9",
"encodeurl": "~2.0.0",
"escape-html": "~1.0.3",
"on-finished": "~2.4.1",
"parseurl": "~1.3.3",
"statuses": "~2.0.2",
"unpipe": "~1.0.0"
},
"engines": {
"node": ">= 0.8"
}
},
"node_modules/forwarded": {
"version": "0.2.0",
"resolved": "https://registry.npmjs.org/forwarded/-/forwarded-0.2.0.tgz",
"integrity": "sha512-buRG0fpBtRHSTCOASe6hD258tEubFoRLb4ZNA6NxMVHNw2gOcwHo9wyablzMzOA5z9xA9L1KNjk/Nt6MT9aYow==",
"license": "MIT",
"engines": {
"node": ">= 0.6"
}
},
"node_modules/fresh": {
"version": "0.5.2",
"resolved": "https://registry.npmjs.org/fresh/-/fresh-0.5.2.tgz",
"integrity": "sha512-zJ2mQYM18rEFOudeV4GShTGIQ7RbzA7ozbU9I/XBpm7kqgMywgmylMwXHxZJmkVoYkna9d2pVXVXPdYTP9ej8Q==",
"license": "MIT",
"engines": {
"node": ">= 0.6"
}
},
"node_modules/function-bind": {
"version": "1.1.2",
"resolved": "https://registry.npmjs.org/function-bind/-/function-bind-1.1.2.tgz",
"integrity": "sha512-7XHNxH7qX9xG5mIwxkhumTox/MIRNcOgDrxWsMt2pAr23WHp6MrRlN7FBSFpCpr+oVO0F744iUgR82nJMfG2SA==",
"license": "MIT",
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/get-intrinsic": {
"version": "1.3.0",
"resolved": "https://registry.npmjs.org/get-intrinsic/-/get-intrinsic-1.3.0.tgz",
"integrity": "sha512-9fSjSaos/fRIVIp+xSJlE6lfwhES7LNtKaCBIamHsjr2na1BiABJPo0mOjjz8GJDURarmCPGqaiVg5mfjb98CQ==",
"license": "MIT",
"dependencies": {
"call-bind-apply-helpers": "^1.0.2",
"es-define-property": "^1.0.1",
"es-errors": "^1.3.0",
"es-object-atoms": "^1.1.1",
"function-bind": "^1.1.2",
"get-proto": "^1.0.1",
"gopd": "^1.2.0",
"has-symbols": "^1.1.0",
"hasown": "^2.0.2",
"math-intrinsics": "^1.1.0"
},
"engines": {
"node": ">= 0.4"
},
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/get-proto": {
"version": "1.0.1",
"resolved": "https://registry.npmjs.org/get-proto/-/get-proto-1.0.1.tgz",
"integrity": "sha512-sTSfBjoXBp89JvIKIefqw7U2CCebsc74kiY6awiGogKtoSGbgjYE/G/+l9sF3MWFPNc9IcoOC4ODfKHfxFmp0g==",
"license": "MIT",
"dependencies": {
"dunder-proto": "^1.0.1",
"es-object-atoms": "^1.0.0"
},
"engines": {
"node": ">= 0.4"
}
},
"node_modules/gopd": {
"version": "1.2.0",
"resolved": "https://registry.npmjs.org/gopd/-/gopd-1.2.0.tgz",
"integrity": "sha512-ZUKRh6/kUFoAiTAtTYPZJ3hw9wNxx+BIBOijnlG9PnrJsCcSjs1wyyD6vJpaYtgnzDrKYRSqf3OO6Rfa93xsRg==",
"license": "MIT",
"engines": {
"node": ">= 0.4"
},
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/has-symbols": {
"version": "1.1.0",
"resolved": "https://registry.npmjs.org/has-symbols/-/has-symbols-1.1.0.tgz",
"integrity": "sha512-1cDNdwJ2Jaohmb3sg4OmKaMBwuC48sYni5HUw2DvsC8LjGTLK9h+eb1X6RyuOHe4hT0ULCW68iomhjUoKUqlPQ==",
"license": "MIT",
"engines": {
"node": ">= 0.4"
},
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/hasown": {
"version": "2.0.2",
"resolved": "https://registry.npmjs.org/hasown/-/hasown-2.0.2.tgz",
"integrity": "sha512-0hJU9SCPvmMzIBdZFqNPXWa6dqh7WdH0cII9y+CyS8rG3nL48Bclra9HmKhVVUHyPWNH5Y7xDwAB7bfgSjkUMQ==",
"license": "MIT",
"dependencies": {
"function-bind": "^1.1.2"
},
"engines": {
"node": ">= 0.4"
}
},
"node_modules/http-errors": {
"version": "2.0.1",
"resolved": "https://registry.npmjs.org/http-errors/-/http-errors-2.0.1.tgz",
"integrity": "sha512-4FbRdAX+bSdmo4AUFuS0WNiPz8NgFt+r8ThgNWmlrjQjt1Q7ZR9+zTlce2859x4KSXrwIsaeTqDoKQmtP8pLmQ==",
"license": "MIT",
"dependencies": {
"depd": "~2.0.0",
"inherits": "~2.0.4",
"setprototypeof": "~1.2.0",
"statuses": "~2.0.2",
"toidentifier": "~1.0.1"
},
"engines": {
"node": ">= 0.8"
},
"funding": {
"type": "opencollective",
"url": "https://opencollective.com/express"
}
},
"node_modules/iconv-lite": {
"version": "0.4.24",
"resolved": "https://registry.npmjs.org/iconv-lite/-/iconv-lite-0.4.24.tgz",
"integrity": "sha512-v3MXnZAcvnywkTUEZomIActle7RXXeedOR31wwl7VlyoXO4Qi9arvSenNQWne1TcRwhCL1HwLI21bEqdpj8/rA==",
"license": "MIT",
"dependencies": {
"safer-buffer": ">= 2.1.2 < 3"
},
"engines": {
"node": ">=0.10.0"
}
},
"node_modules/inherits": {
"version": "2.0.4",
"resolved": "https://registry.npmjs.org/inherits/-/inherits-2.0.4.tgz",
"integrity": "sha512-k/vGaX4/Yla3WzyMCvTQOXYeIHvqOKtnqBduzTHpzpQZzAskKMhZ2K+EnBiSM9zGSoIFeMpXKxa4dYeZIQqewQ==",
"license": "ISC"
},
"node_modules/ipaddr.js": {
"version": "1.9.1",
"resolved": "https://registry.npmjs.org/ipaddr.js/-/ipaddr.js-1.9.1.tgz",
"integrity": "sha512-0KI/607xoxSToH7GjN1FfSbLoU0+btTicjsQSWQlh/hZykN8KpmMf7uYwPW3R+akZ6R/w18ZlXSHBYXiYUPO3g==",
"license": "MIT",
"engines": {
"node": ">= 0.10"
}
},
"node_modules/math-intrinsics": {
"version": "1.1.0",
"resolved": "https://registry.npmjs.org/math-intrinsics/-/math-intrinsics-1.1.0.tgz",
"integrity": "sha512-/IXtbwEk5HTPyEwyKX6hGkYXxM9nbj64B+ilVJnC/R6B0pH5G4V3b0pVbL7DBj4tkhBAppbQUlf6F6Xl9LHu1g==",
"license": "MIT",
"engines": {
"node": ">= 0.4"
}
},
"node_modules/media-typer": {
"version": "0.3.0",
"resolved": "https://registry.npmjs.org/media-typer/-/media-typer-0.3.0.tgz",
"integrity": "sha512-dq+qelQ9akHpcOl/gUVRTxVIOkAJ1wR3QAvb4RsVjS8oVoFjDGTc679wJYmUmknUF5HwMLOgb5O+a3KxfWapPQ==",
"license": "MIT",
"engines": {
"node": ">= 0.6"
}
},
"node_modules/merge-descriptors": {
"version": "1.0.3",
"resolved": "https://registry.npmjs.org/merge-descriptors/-/merge-descriptors-1.0.3.tgz",
"integrity": "sha512-gaNvAS7TZ897/rVaZ0nMtAyxNyi/pdbjbAwUpFQpN70GqnVfOiXpeUUMKRBmzXaSQ8DdTX4/0ms62r2K+hE6mQ==",
"license": "MIT",
"funding": {
"url": "https://github.com/sponsors/sindresorhus"
}
},
"node_modules/methods": {
"version": "1.1.2",
"resolved": "https://registry.npmjs.org/methods/-/methods-1.1.2.tgz",
"integrity": "sha512-iclAHeNqNm68zFtnZ0e+1L2yUIdvzNoauKU4WBA3VvH/vPFieF7qfRlwUZU+DA9P9bPXIS90ulxoUoCH23sV2w==",
"license": "MIT",
"engines": {
"node": ">= 0.6"
}
},
"node_modules/mime": {
"version": "1.6.0",
"resolved": "https://registry.npmjs.org/mime/-/mime-1.6.0.tgz",
"integrity": "sha512-x0Vn8spI+wuJ1O6S7gnbaQg8Pxh4NNHb7KSINmEWKiPE4RKOplvijn+NkmYmmRgP68mc70j2EbeTFRsrswaQeg==",
"license": "MIT",
"bin": {
"mime": "cli.js"
},
"engines": {
"node": ">=4"
}
},
"node_modules/mime-db": {
"version": "1.52.0",
"resolved": "https://registry.npmjs.org/mime-db/-/mime-db-1.52.0.tgz",
"integrity": "sha512-sPU4uV7dYlvtWJxwwxHD0PuihVNiE7TyAbQ5SWxDCB9mUYvOgroQOwYQQOKPJ8CIbE+1ETVlOoK1UC2nU3gYvg==",
"license": "MIT",
"engines": {
"node": ">= 0.6"
}
},
"node_modules/mime-types": {
"version": "2.1.35",
"resolved": "https://registry.npmjs.org/mime-types/-/mime-types-2.1.35.tgz",
"integrity": "sha512-ZDY+bPm5zTTF+YpCrAU9nK0UgICYPT0QtT1NZWFv4s++TNkcgVaT0g6+4R2uI4MjQjzysHB1zxuWL50hzaeXiw==",
"license": "MIT",
"dependencies": {
"mime-db": "1.52.0"
},
"engines": {
"node": ">= 0.6"
}
},
"node_modules/ms": {
"version": "2.0.0",
"resolved": "https://registry.npmjs.org/ms/-/ms-2.0.0.tgz",
"integrity": "sha512-Tpp60P6IUJDTuOq/5Z8cdskzJujfwqfOTkrwIwj7IRISpnkJnT6SyJ4PCPnGMoFjC9ddhal5KVIYtAt97ix05A==",
"license": "MIT"
},
"node_modules/negotiator": {
"version": "0.6.3",
"resolved": "https://registry.npmjs.org/negotiator/-/negotiator-0.6.3.tgz",
"integrity": "sha512-+EUsqGPLsM+j/zdChZjsnX51g4XrHFOIXwfnCVPGlQk/k5giakcKsuxCObBRu6DSm9opw/O6slWbJdghQM4bBg==",
"license": "MIT",
"engines": {
"node": ">= 0.6"
}
},
"node_modules/object-inspect": {
"version": "1.13.4",
"resolved": "https://registry.npmjs.org/object-inspect/-/object-inspect-1.13.4.tgz",
"integrity": "sha512-W67iLl4J2EXEGTbfeHCffrjDfitvLANg0UlX3wFUUSTx92KXRFegMHUVgSqE+wvhAbi4WqjGg9czysTV2Epbew==",
"license": "MIT",
"engines": {
"node": ">= 0.4"
},
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/on-finished": {
"version": "2.4.1",
"resolved": "https://registry.npmjs.org/on-finished/-/on-finished-2.4.1.tgz",
"integrity": "sha512-oVlzkg3ENAhCk2zdv7IJwd/QUD4z2RxRwpkcGY8psCVcCYZNq4wYnVWALHM+brtuJjePWiYF/ClmuDr8Ch5+kg==",
"license": "MIT",
"dependencies": {
"ee-first": "1.1.1"
},
"engines": {
"node": ">= 0.8"
}
},
"node_modules/parseurl": {
"version": "1.3.3",
"resolved": "https://registry.npmjs.org/parseurl/-/parseurl-1.3.3.tgz",
"integrity": "sha512-CiyeOxFT/JZyN5m0z9PfXw4SCBJ6Sygz1Dpl0wqjlhDEGGBP1GnsUVEL0p63hoG1fcj3fHynXi9NYO4nWOL+qQ==",
"license": "MIT",
"engines": {
"node": ">= 0.8"
}
},
"node_modules/path-to-regexp": {
"version": "0.1.12",
"resolved": "https://registry.npmjs.org/path-to-regexp/-/path-to-regexp-0.1.12.tgz",
"integrity": "sha512-RA1GjUVMnvYFxuqovrEqZoxxW5NUZqbwKtYz/Tt7nXerk0LbLblQmrsgdeOxV5SFHf0UDggjS/bSeOZwt1pmEQ==",
"license": "MIT"
},
"node_modules/proxy-addr": {
"version": "2.0.7",
"resolved": "https://registry.npmjs.org/proxy-addr/-/proxy-addr-2.0.7.tgz",
"integrity": "sha512-llQsMLSUDUPT44jdrU/O37qlnifitDP+ZwrmmZcoSKyLKvtZxpyV0n2/bD/N4tBAAZ/gJEdZU7KMraoK1+XYAg==",
"license": "MIT",
"dependencies": {
"forwarded": "0.2.0",
"ipaddr.js": "1.9.1"
},
"engines": {
"node": ">= 0.10"
}
},
"node_modules/qs": {
"version": "6.14.2",
"resolved": "https://registry.npmjs.org/qs/-/qs-6.14.2.tgz",
"integrity": "sha512-V/yCWTTF7VJ9hIh18Ugr2zhJMP01MY7c5kh4J870L7imm6/DIzBsNLTXzMwUA3yZ5b/KBqLx8Kp3uRvd7xSe3Q==",
"license": "BSD-3-Clause",
"dependencies": {
"side-channel": "^1.1.0"
},
"engines": {
"node": ">=0.6"
},
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/range-parser": {
"version": "1.2.1",
"resolved": "https://registry.npmjs.org/range-parser/-/range-parser-1.2.1.tgz",
"integrity": "sha512-Hrgsx+orqoygnmhFbKaHE6c296J+HTAQXoxEF6gNupROmmGJRoyzfG3ccAveqCBrwr/2yxQ5BVd/GTl5agOwSg==",
"license": "MIT",
"engines": {
"node": ">= 0.6"
}
},
"node_modules/raw-body": {
"version": "2.5.3",
"resolved": "https://registry.npmjs.org/raw-body/-/raw-body-2.5.3.tgz",
"integrity": "sha512-s4VSOf6yN0rvbRZGxs8Om5CWj6seneMwK3oDb4lWDH0UPhWcxwOWw5+qk24bxq87szX1ydrwylIOp2uG1ojUpA==",
"license": "MIT",
"dependencies": {
"bytes": "~3.1.2",
"http-errors": "~2.0.1",
"iconv-lite": "~0.4.24",
"unpipe": "~1.0.0"
},
"engines": {
"node": ">= 0.8"
}
},
"node_modules/safe-buffer": {
"version": "5.2.1",
"resolved": "https://registry.npmjs.org/safe-buffer/-/safe-buffer-5.2.1.tgz",
"integrity": "sha512-rp3So07KcdmmKbGvgaNxQSJr7bGVSVk5S9Eq1F+ppbRo70+YeaDxkw5Dd8NPN+GD6bjnYm2VuPuCXmpuYvmCXQ==",
"funding": [
{
"type": "github",
"url": "https://github.com/sponsors/feross"
},
{
"type": "patreon",
"url": "https://www.patreon.com/feross"
},
{
"type": "consulting",
"url": "https://feross.org/support"
}
],
"license": "MIT"
},
"node_modules/safer-buffer": {
"version": "2.1.2",
"resolved": "https://registry.npmjs.org/safer-buffer/-/safer-buffer-2.1.2.tgz",
"integrity": "sha512-YZo3K82SD7Riyi0E1EQPojLz7kpepnSQI9IyPbHHg1XXXevb5dJI7tpyN2ADxGcQbHG7vcyRHk0cbwqcQriUtg==",
"license": "MIT"
},
"node_modules/send": {
"version": "0.19.2",
"resolved": "https://registry.npmjs.org/send/-/send-0.19.2.tgz",
"integrity": "sha512-VMbMxbDeehAxpOtWJXlcUS5E8iXh6QmN+BkRX1GARS3wRaXEEgzCcB10gTQazO42tpNIya8xIyNx8fll1OFPrg==",
"license": "MIT",
"dependencies": {
"debug": "2.6.9",
"depd": "2.0.0",
"destroy": "1.2.0",
"encodeurl": "~2.0.0",
"escape-html": "~1.0.3",
"etag": "~1.8.1",
"fresh": "~0.5.2",
"http-errors": "~2.0.1",
"mime": "1.6.0",
"ms": "2.1.3",
"on-finished": "~2.4.1",
"range-parser": "~1.2.1",
"statuses": "~2.0.2"
},
"engines": {
"node": ">= 0.8.0"
}
},
"node_modules/send/node_modules/ms": {
"version": "2.1.3",
"resolved": "https://registry.npmjs.org/ms/-/ms-2.1.3.tgz",
"integrity": "sha512-6FlzubTLZG3J2a/NVCAleEhjzq5oxgHyaCU9yYXvcLsvoVaHJq/s5xXI6/XXP6tz7R9xAOtHnSO/tXtF3WRTlA==",
"license": "MIT"
},
"node_modules/serve-static": {
"version": "1.16.3",
"resolved": "https://registry.npmjs.org/serve-static/-/serve-static-1.16.3.tgz",
"integrity": "sha512-x0RTqQel6g5SY7Lg6ZreMmsOzncHFU7nhnRWkKgWuMTu5NN0DR5oruckMqRvacAN9d5w6ARnRBXl9xhDCgfMeA==",
"license": "MIT",
"dependencies": {
"encodeurl": "~2.0.0",
"escape-html": "~1.0.3",
"parseurl": "~1.3.3",
"send": "~0.19.1"
},
"engines": {
"node": ">= 0.8.0"
}
},
"node_modules/setprototypeof": {
"version": "1.2.0",
"resolved": "https://registry.npmjs.org/setprototypeof/-/setprototypeof-1.2.0.tgz",
"integrity": "sha512-E5LDX7Wrp85Kil5bhZv46j8jOeboKq5JMmYM3gVGdGH8xFpPWXUMsNrlODCrkoxMEeNi/XZIwuRvY4XNwYMJpw==",
"license": "ISC"
},
"node_modules/side-channel": {
"version": "1.1.0",
"resolved": "https://registry.npmjs.org/side-channel/-/side-channel-1.1.0.tgz",
"integrity": "sha512-ZX99e6tRweoUXqR+VBrslhda51Nh5MTQwou5tnUDgbtyM0dBgmhEDtWGP/xbKn6hqfPRHujUNwz5fy/wbbhnpw==",
"license": "MIT",
"dependencies": {
"es-errors": "^1.3.0",
"object-inspect": "^1.13.3",
"side-channel-list": "^1.0.0",
"side-channel-map": "^1.0.1",
"side-channel-weakmap": "^1.0.2"
},
"engines": {
"node": ">= 0.4"
},
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/side-channel-list": {
"version": "1.0.0",
"resolved": "https://registry.npmjs.org/side-channel-list/-/side-channel-list-1.0.0.tgz",
"integrity": "sha512-FCLHtRD/gnpCiCHEiJLOwdmFP+wzCmDEkc9y7NsYxeF4u7Btsn1ZuwgwJGxImImHicJArLP4R0yX4c2KCrMrTA==",
"license": "MIT",
"dependencies": {
"es-errors": "^1.3.0",
"object-inspect": "^1.13.3"
},
"engines": {
"node": ">= 0.4"
},
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/side-channel-map": {
"version": "1.0.1",
"resolved": "https://registry.npmjs.org/side-channel-map/-/side-channel-map-1.0.1.tgz",
"integrity": "sha512-VCjCNfgMsby3tTdo02nbjtM/ewra6jPHmpThenkTYh8pG9ucZ/1P8So4u4FGBek/BjpOVsDCMoLA/iuBKIFXRA==",
"license": "MIT",
"dependencies": {
"call-bound": "^1.0.2",
"es-errors": "^1.3.0",
"get-intrinsic": "^1.2.5",
"object-inspect": "^1.13.3"
},
"engines": {
"node": ">= 0.4"
},
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/side-channel-weakmap": {
"version": "1.0.2",
"resolved": "https://registry.npmjs.org/side-channel-weakmap/-/side-channel-weakmap-1.0.2.tgz",
"integrity": "sha512-WPS/HvHQTYnHisLo9McqBHOJk2FkHO/tlpvldyrnem4aeQp4hai3gythswg6p01oSoTl58rcpiFAjF2br2Ak2A==",
"license": "MIT",
"dependencies": {
"call-bound": "^1.0.2",
"es-errors": "^1.3.0",
"get-intrinsic": "^1.2.5",
"object-inspect": "^1.13.3",
"side-channel-map": "^1.0.1"
},
"engines": {
"node": ">= 0.4"
},
"funding": {
"url": "https://github.com/sponsors/ljharb"
}
},
"node_modules/statuses": {
"version": "2.0.2",
"resolved": "https://registry.npmjs.org/statuses/-/statuses-2.0.2.tgz",
"integrity": "sha512-DvEy55V3DB7uknRo+4iOGT5fP1slR8wQohVdknigZPMpMstaKJQWhwiYBACJE3Ul2pTnATihhBYnRhZQHGBiRw==",
"license": "MIT",
"engines": {
"node": ">= 0.8"
}
},
"node_modules/toidentifier": {
"version": "1.0.1",
"resolved": "https://registry.npmjs.org/toidentifier/-/toidentifier-1.0.1.tgz",
"integrity": "sha512-o5sSPKEkg/DIQNmH43V0/uerLrpzVedkUh8tGNvaeXpfpuwjKenlSox/2O/BTlZUtEe+JG7s5YhEz608PlAHRA==",
"license": "MIT",
"engines": {
"node": ">=0.6"
}
},
"node_modules/type-is": {
"version": "1.6.18",
"resolved": "https://registry.npmjs.org/type-is/-/type-is-1.6.18.tgz",
"integrity": "sha512-TkRKr9sUTxEH8MdfuCSP7VizJyzRNMjj2J2do2Jr3Kym598JVdEksuzPQCnlFPW4ky9Q+iA+ma9BGm06XQBy8g==",
"license": "MIT",
"dependencies": {
"media-typer": "0.3.0",
"mime-types": "~2.1.24"
},
"engines": {
"node": ">= 0.6"
}
},
"node_modules/unpipe": {
"version": "1.0.0",
"resolved": "https://registry.npmjs.org/unpipe/-/unpipe-1.0.0.tgz",
"integrity": "sha512-pjy2bYhSsufwWlKwPc+l3cN7+wuJlK6uz0YdJEOlQDbl6jo/YlPi4mb8agUkVC8BF7V8NuzeyPNqRksA3hztKQ==",
"license": "MIT",
"engines": {
"node": ">= 0.8"
}
},
"node_modules/utils-merge": {
"version": "1.0.1",
"resolved": "https://registry.npmjs.org/utils-merge/-/utils-merge-1.0.1.tgz",
"integrity": "sha512-pMZTvIkT1d+TFGvDOqodOclx0QWkkgi6Tdoa8gC8ffGAAqz9pzPTZWAybbsHHoED/ztMtkv/VoYTYyShUn81hA==",
"license": "MIT",
"engines": {
"node": ">= 0.4.0"
}
},
"node_modules/vary": {
"version": "1.1.2",
"resolved": "https://registry.npmjs.org/vary/-/vary-1.1.2.tgz",
"integrity": "sha512-BNGbWLfd0eUPabhkXUVm0j8uuvREyTh5ovRa/dyow/BqAbZJyC+5fU+IzQOzmAKzYqYRAISoRhdQr3eIZ/PXqg==",
"license": "MIT",
"engines": {
"node": ">= 0.8"
}
},
"node_modules/ws": {
"version": "8.19.0",
"resolved": "https://registry.npmjs.org/ws/-/ws-8.19.0.tgz",
"integrity": "sha512-blAT2mjOEIi0ZzruJfIhb3nps74PRWTCz1IjglWEEpQl5XS/UNama6u2/rjFkDDouqr4L67ry+1aGIALViWjDg==",
"license": "MIT",
"engines": {
"node": ">=10.0.0"
},
"peerDependencies": {
"bufferutil": "^4.0.1",
"utf-8-validate": ">=5.0.2"
},
"peerDependenciesMeta": {
"bufferutil": {
"optional": true
},
"utf-8-validate": {
"optional": true
}
}
}
}
}

13
web/package.json Normal file
View File

@@ -0,0 +1,13 @@
{
"name": "pm-template-web",
"version": "1.0.0",
"description": "Minimal local web chat UI",
"main": "server.js",
"scripts": {
"start": "node server.js"
},
"dependencies": {
"express": "^4.18.2",
"ws": "^8.16.0"
}
}

127
web/public/index.html Normal file
View File

@@ -0,0 +1,127 @@
<!DOCTYPE html>
<html lang="en">
<head>
<meta charset="UTF-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>Forge Console</title>
<link rel="stylesheet" href="style.css">
</head>
<body>
<div id="app">
<header id="status-bar">
<span class="logo">🔥 Forge Console</span>
<span id="conn-status"><span class="dot red"></span> Disconnected</span>
</header>
<main id="messages"></main>
<footer id="input-bar">
<textarea id="input" placeholder="Ask anything..." rows="1"></textarea>
<button id="send" title="Send"></button>
</footer>
</div>
<script>
const messages = document.getElementById('messages');
const input = document.getElementById('input');
const sendBtn = document.getElementById('send');
const connStatus = document.getElementById('conn-status');
let ws;
let currentAssistant = null;
function setStatus(connected) {
connStatus.innerHTML = connected
? '<span class="dot green"></span> Connected'
: '<span class="dot red"></span> Disconnected';
}
function scrollBottom() {
messages.scrollTop = messages.scrollHeight;
}
function addMessage(role, text) {
const div = document.createElement('div');
div.className = `msg ${role}`;
div.textContent = text;
messages.appendChild(div);
scrollBottom();
return div;
}
function addThinking() {
const div = document.createElement('div');
div.className = 'msg assistant thinking';
div.innerHTML = '<span class="dots"><span>.</span><span>.</span><span>.</span></span>';
messages.appendChild(div);
scrollBottom();
return div;
}
function connect() {
const proto = location.protocol === 'https:' ? 'wss' : 'ws';
ws = new WebSocket(`${proto}://${location.host}/ws`);
ws.onopen = () => {
setStatus(true);
currentAssistant = null;
};
ws.onclose = () => {
setStatus(false);
currentAssistant = null;
setTimeout(connect, 2000);
};
ws.onmessage = (e) => {
let msg;
try { msg = JSON.parse(e.data); } catch { return; }
if (msg.type === 'stdout' || msg.type === 'stderr') {
// Remove thinking indicator if present
const thinking = messages.querySelector('.thinking');
if (thinking) thinking.remove();
if (!currentAssistant) {
currentAssistant = addMessage('assistant', '');
}
currentAssistant.textContent += msg.data;
scrollBottom();
} else if (msg.type === 'exit') {
addMessage('system', `Session ended (code ${msg.code}). Refresh to restart.`);
currentAssistant = null;
} else if (msg.type === 'error') {
addMessage('system', `Error: ${msg.data}`);
currentAssistant = null;
}
};
}
function send() {
const text = input.value.trim();
if (!text || !ws || ws.readyState !== 1) return;
addMessage('user', text);
currentAssistant = null;
addThinking();
ws.send(text);
input.value = '';
input.style.height = 'auto';
}
sendBtn.addEventListener('click', send);
input.addEventListener('keydown', (e) => {
if (e.key === 'Enter' && !e.shiftKey) {
e.preventDefault();
send();
}
});
// Auto-resize textarea
input.addEventListener('input', () => {
input.style.height = 'auto';
input.style.height = Math.min(input.scrollHeight, 120) + 'px';
});
connect();
</script>
</body>
</html>

176
web/public/style.css Normal file
View File

@@ -0,0 +1,176 @@
* { margin: 0; padding: 0; box-sizing: border-box; }
:root {
--bg: #1a1a2e;
--surface: #16213e;
--input-bg: #0f3460;
--user-bg: #1a3a5c;
--assistant-bg: #2a2a3e;
--system-bg: #2e1a1a;
--text: #e0e0e0;
--text-dim: #8888aa;
--accent: #4a9eff;
--green: #4caf50;
--red: #f44336;
}
html, body { height: 100%; }
body {
font-family: -apple-system, BlinkMacSystemFont, 'Segoe UI', sans-serif;
background: var(--bg);
color: var(--text);
}
#app {
display: flex;
flex-direction: column;
height: 100vh;
max-width: 800px;
margin: 0 auto;
}
/* Status bar */
#status-bar {
display: flex;
justify-content: space-between;
align-items: center;
padding: 10px 16px;
background: var(--surface);
border-bottom: 1px solid rgba(255,255,255,0.05);
font-size: 13px;
}
.logo {
font-weight: 600;
font-size: 15px;
letter-spacing: 0.5px;
}
.dot {
display: inline-block;
width: 8px;
height: 8px;
border-radius: 50%;
margin-right: 6px;
vertical-align: middle;
}
.dot.green { background: var(--green); }
.dot.red { background: var(--red); }
#conn-status { color: var(--text-dim); font-size: 12px; }
/* Messages */
#messages {
flex: 1;
overflow-y: auto;
padding: 16px;
display: flex;
flex-direction: column;
gap: 10px;
}
.msg {
max-width: 85%;
padding: 10px 14px;
border-radius: 12px;
font-size: 14px;
line-height: 1.5;
white-space: pre-wrap;
word-wrap: break-word;
animation: fadeIn 0.2s ease;
}
.msg.user {
align-self: flex-end;
background: var(--user-bg);
border-bottom-right-radius: 4px;
}
.msg.assistant {
align-self: flex-start;
background: var(--assistant-bg);
border-bottom-left-radius: 4px;
font-family: 'SF Mono', 'Fira Code', 'Consolas', monospace;
font-size: 13px;
}
.msg.system {
align-self: center;
background: var(--system-bg);
color: var(--text-dim);
font-size: 12px;
border-radius: 8px;
}
/* Thinking dots */
.thinking .dots span {
animation: blink 1.4s infinite;
font-size: 24px;
line-height: 1;
}
.thinking .dots span:nth-child(2) { animation-delay: 0.2s; }
.thinking .dots span:nth-child(3) { animation-delay: 0.4s; }
@keyframes blink {
0%, 20% { opacity: 0.2; }
50% { opacity: 1; }
80%, 100% { opacity: 0.2; }
}
@keyframes fadeIn {
from { opacity: 0; transform: translateY(6px); }
to { opacity: 1; transform: translateY(0); }
}
/* Input bar */
#input-bar {
display: flex;
gap: 8px;
padding: 12px 16px;
background: var(--surface);
border-top: 1px solid rgba(255,255,255,0.05);
}
#input {
flex: 1;
background: var(--input-bg);
border: 1px solid rgba(255,255,255,0.1);
border-radius: 10px;
padding: 10px 14px;
color: var(--text);
font-size: 14px;
font-family: inherit;
resize: none;
outline: none;
transition: border-color 0.2s;
}
#input:focus { border-color: var(--accent); }
#input::placeholder { color: var(--text-dim); }
#send {
background: var(--accent);
border: none;
border-radius: 10px;
width: 44px;
color: white;
font-size: 18px;
cursor: pointer;
transition: opacity 0.2s;
}
#send:hover { opacity: 0.85; }
#send:active { opacity: 0.7; }
/* Scrollbar */
#messages::-webkit-scrollbar { width: 6px; }
#messages::-webkit-scrollbar-track { background: transparent; }
#messages::-webkit-scrollbar-thumb { background: rgba(255,255,255,0.1); border-radius: 3px; }
/* Mobile */
@media (max-width: 600px) {
#app { max-width: 100%; }
.msg { max-width: 92%; }
}

77
web/server.js Normal file
View File

@@ -0,0 +1,77 @@
const express = require('express');
const { WebSocketServer } = require('ws');
const { spawn } = require('child_process');
const path = require('path');
const http = require('http');
const PORT = process.env.PORT || 3000;
const app = express();
const server = http.createServer(app);
const wss = new WebSocketServer({ server, path: '/ws' });
app.use(express.static(path.join(__dirname, 'public')));
app.get('/health', (_, res) => res.json({ status: 'ok' }));
wss.on('connection', (ws) => {
console.log('[forge] client connected');
// Spawn codex in interactive mode, working dir = repo root
const repoRoot = path.resolve(__dirname, '..');
const codex = spawn('codex', ['--quiet'], {
cwd: repoRoot,
shell: true,
env: { ...process.env, FORCE_COLOR: '0' },
stdio: ['pipe', 'pipe', 'pipe']
});
let alive = true;
codex.stdout.on('data', (data) => {
if (ws.readyState === 1) {
ws.send(JSON.stringify({ type: 'stdout', data: data.toString() }));
}
});
codex.stderr.on('data', (data) => {
if (ws.readyState === 1) {
ws.send(JSON.stringify({ type: 'stderr', data: data.toString() }));
}
});
codex.on('close', (code) => {
alive = false;
console.log(`[forge] codex exited (code ${code})`);
if (ws.readyState === 1) {
ws.send(JSON.stringify({ type: 'exit', code }));
}
});
codex.on('error', (err) => {
alive = false;
console.error('[forge] codex spawn error:', err.message);
if (ws.readyState === 1) {
ws.send(JSON.stringify({ type: 'error', data: err.message }));
}
});
ws.on('message', (msg) => {
if (alive && codex.stdin.writable) {
codex.stdin.write(msg.toString() + '\n');
}
});
ws.on('close', () => {
console.log('[forge] client disconnected');
if (alive) codex.kill('SIGTERM');
});
});
server.listen(PORT, () => {
console.log(`\n 🔥 Forge Console running at http://localhost:${PORT}\n`);
});
process.on('SIGINT', () => {
console.log('\n[forge] shutting down...');
wss.clients.forEach((ws) => ws.close());
server.close(() => process.exit(0));
});