CopilotReportForge
CopilotReportForge is an open-source platform that transforms ad-hoc LLM interactions into governed, repeatable, and auditable report-generation pipelines. Users define expert personas as system prompts and evaluation queries as input; the platform executes all personas in parallel via the GitHub Copilot SDK and aggregates the results into structured JSON reports. Reports are uploaded to Azure Blob Storage and shared through time-limited, revocable URLs. The entire workflow runs in ephemeral GitHub Actions environments with passwordless OIDC authentication — no GPU provisioning, no model hosting, and no long-lived secrets. By treating personas as configuration rather than code, the same pipeline adapts to any industry — from manufacturing quality panels to financial risk committees — without code changes. Infrastructure is fully managed via Terraform, and a browser-based Web UI with GitHub OAuth login is included for interactive use.
Core Concept
CopilotReportForge is built on one central idea: automated report generation through multi-persona parallel agent execution.
By combining multiple AI personas, parallel processing, and fully automated pipelines, the platform transforms ad-hoc LLM interactions into governed, repeatable, and auditable workflows -- without managing any AI infrastructure.
Pillar 1: Multi-Persona Parallel Execution
- Define a single topic (prompt) -- for example, "Evaluate the new wireless headphones"
- Define multiple personas (quality engineer, consumer researcher, regulatory specialist, etc.) as system prompts
- Launch all personas as AI agents in parallel using
asyncio.gather, with each agent producing structured JSON output from its specialized perspective - Aggregate all agent results into one
ReportOutput(Pydantic model)
Personas are configuration, not code. By simply swapping system prompts, you can switch from a food industry evaluation panel to a financial risk committee to an architectural compliance review -- without changing the code.
Pillar 2: 24/7 Autonomous Operation
The entire pipeline runs via GitHub Actions schedule (cron), workflow_dispatch, or API triggers:
- Reports can be continuously generated on different topics without human intervention
- The system operates 24 hours a day, regardless of time zone
- Generated reports are stored in Azure Blob Storage and shared to Teams/Slack via SAS URL
Pillar 3: Domain Agnostic
The persona + parallel execution model can be applied to any industry. See Cross-Industry Applicability for eight representative use cases.
The Problem in One Sentence
Enterprises use LLMs through copy-paste chat sessions — producing unstructured, unreproducible, and ungoverned outputs that cannot be audited, scaled, or safely shared with stakeholders.
For a deeper analysis of the problem space, see Problem & Solution.
What CopilotReportForge Does
CopilotReportForge converts ad-hoc LLM interactions into a governed, automated pipeline:
- Define perspectives — Assign system prompts as expert personas (e.g., "Quality Engineer", "Compliance Officer").
- Submit evaluation queries — Specify what to evaluate (e.g., "Assess durability", "Check regulatory compliance").
- Execute in parallel — All queries run concurrently against hosted LLMs, each under its assigned persona.
- Produce structured results — Outputs are collected into a typed JSON report with success/failure tracking.
- Share securely — Reports are uploaded to Azure Blob Storage with time-limited, revocable access URLs.
No GPU provisioning, no model hosting, no long-lived secrets. The entire workflow runs in ephemeral sandbox environments with full audit trails.
Architecture Overview
%%{init: {'theme': 'dark'}}%%
flowchart TB
subgraph Trigger
USER["User / Scheduler"]
end
subgraph Execution["Execution Environment"]
SDK["Copilot SDK"]
AGENTS["AI Agents"]
LLM["Hosted LLMs"]
end
subgraph Cloud["Azure"]
AUTH["Entra ID (OIDC)"]
STORAGE["Blob Storage"]
FOUNDRY["AI Foundry"]
end
USER --> SDK
SDK -- "Parallel queries" --> LLM
SDK -- "Tool calls" --> AGENTS
AGENTS --> FOUNDRY
SDK -- "Upload report" --> STORAGE
STORAGE -- "Secure URL" --> USER
USER -. "Passwordless auth" .-> AUTH
AUTH -. "Access token" .-> SDK
For component-level details and data flows, see Architecture.
Key Capabilities
| Capability | What It Means |
|---|---|
| Parallel Multi-Persona Execution | Run N queries concurrently, each with a different expert persona, and aggregate results into one report |
| Zero-Infrastructure AI | Use hosted LLMs via the Copilot SDK — no model deployment or GPU management |
| Passwordless Security | OIDC-based authentication between GitHub Actions and Azure — no stored API keys |
| Secure Artifact Sharing | Reports shared via time-limited, revocable URLs — no public bucket exposure |
| Agentic Workflows | Delegate domain-specific tasks to AI Foundry Agents that can reference stored documents |
| Infrastructure as Code | All Azure resources, identities, and permissions managed via Terraform |
| Web UI | Browser-based chat and report generation with GitHub OAuth login |
| Container Deployment | Docker Compose support with images on GitHub Container Registry and Docker Hub |
Cross-Industry Applicability
The platform is domain-agnostic by design. By changing only the system prompt (persona) and queries (evaluation dimensions), the same pipeline serves entirely different industries:
| Industry | Persona Example | Evaluation Dimensions |
|---|---|---|
| Manufacturing | Sensory panelist, Quality engineer | Texture, durability, regulatory compliance |
| Real Estate | Layout evaluator, ADA compliance reviewer | Accessibility, traffic flow, space utilization |
| Healthcare | Clinical pharmacist, Guideline reviewer | Drug interactions, dosage, contraindications |
| Finance | Credit analyst, Compliance officer | Credit exposure, market risk, regulatory adherence |
| Education | Curriculum designer, Assessment specialist | Learning objectives, rubric design, lesson plans |
| Creative | Brand strategist, Cultural sensitivity reviewer | Inclusivity, brand alignment, market resonance |
| Legal | Contract analyst, Regulatory compliance officer | Clause analysis, risk assessment, jurisdictional review |
| Retail | Merchandising analyst, Customer experience reviewer | Product placement, pricing strategy, customer satisfaction |
The core insight: system prompts are persona configuration, queries are evaluation dimensions. Any expert judgment can be parallelized, structured, and audited at scale.
Business Value
| Dimension | Value |
|---|---|
| Zero Infrastructure | No GPU clusters or model hosting — pay-per-use via hosted LLMs and Azure AI Foundry |
| Minutes to Production | Clone → configure → deploy in under an hour with Terraform + GitHub Actions |
| Enterprise Security | Passwordless OIDC, RBAC-scoped access, time-bounded sharing URLs, zero long-lived secrets |
| Sandbox Execution | Ephemeral, disposable environments — more secure than local execution, no credential leakage |
| Built-in Audit Trail | Every execution is logged with who, what, when, and how long — no additional tooling required |
| Domain Agnostic | Adapt to any industry by changing configuration parameters, not code |
| Regulated Industry Ready | BYOK support, private endpoint compatibility, IaC-managed RBAC for air-gapped environments |
Quick Start
# 1. Clone and install
git clone https://github.com/ks6088ts/template-github-copilot.git
cd template-github-copilot/src/python
make install-deps-dev
# 2. Configure environment
cp .env.template .env # Edit with your settings
# 3. Start the Copilot CLI server
export COPILOT_GITHUB_TOKEN="your-github-pat"
make copilot
# 4. Run the interactive chat (in another terminal)
make copilot-app
# 5. Generate a multi-perspective report
uv run python scripts/report_service.py generate \
--system-prompt "You are a product evaluation specialist." \
--queries "Evaluate durability,Evaluate usability,Evaluate aesthetics" \
--account-url "https://<account>.blob.core.windows.net" \
--container-name "reports"
For full setup instructions, see Getting Started.
Documentation
| Document | Description |
|---|---|
| Problem & Solution | Why this platform exists — the enterprise AI adoption gap and how the architecture addresses it |
| Architecture | System design, execution model, security model, and extensibility |
| Getting Started | Prerequisites, local development setup, infrastructure provisioning, and CLI reference |
| Deployment | Step-by-step deployment from local dev to production GitHub Actions workflows |
| GitHub OAuth App | Setting up GitHub OAuth for the web UI authentication flow |
| Web UI Guide | Walkthrough of the browser-based chat and report generation interface |
| Running Containers | Running the platform via Docker Compose (local build, Docker Hub, or GHCR) |
| Responsible AI | Fairness, transparency, safety, privacy guidelines, and deployment checklist |
| References | External links and further reading |
Infrastructure (Terraform Scenarios)
| Scenario | Purpose |
|---|---|
| Azure GitHub OIDC | Establish passwordless trust between GitHub Actions and Azure |
| GitHub Secrets | Automate GitHub environment and secrets configuration |
| Azure Microsoft Foundry | Deploy AI Foundry with model endpoints and storage |
| Azure Container Apps | Deploy monolith service (Copilot CLI + API) as Azure Container App (standalone) |