Connector Evidence to MCP Action Brief
Category research
Subcategory connector-orchestration
Difficulty advanced
Target models: gpt, claude-sonnet, qwen3
Variables:
{{goal}} {{time_window}} {{connected_sources}} {{output_audience}} {{action_systems}} {{risk_policy}} {{approval_rules}} connectors mcp briefing research-ops governance agentic
Updated March 1, 2026
The Prompt
You are an operations research lead. Build a decision-ready brief from connected sources, then propose MCP-safe write actions.
GOAL:
{{goal}}
TIME WINDOW:
{{time_window}}
CONNECTED SOURCES:
{{connected_sources}}
OUTPUT AUDIENCE:
{{output_audience}}
ACTION SYSTEMS (MCP write targets):
{{action_systems}}
RISK POLICY:
{{risk_policy}}
APPROVAL RULES:
{{approval_rules}}
Return exactly these sections:
1) Evidence Register
- Table columns: claim_id, claim, source, source_timestamp, confidence (0-1), freshness_flag, contradiction_flag.
2) Executive Brief
- 5-8 key findings.
- Business/operational implications.
- Unknowns and what evidence is missing.
3) Proposed Action Queue (PROPOSED_ONLY)
- Table columns: action_id, action_text, destination_system, write_type (create/update/comment), owner_role, due_date, expected_impact, risk_level.
- Include draft payloads for each action but do not assume execution.
4) MCP Execution Plan
- Preflight checks per action.
- Dry-run validation steps.
- Human approval checkpoints mapped to risk level.
5) Review Checklist
- What a human reviewer must verify before approving writes.
Rules:
- No unsourced factual claims.
- Tag low-confidence claims clearly.
- If evidence is stale or conflicting, downgrade confidence and mark as REVIEW_REQUIRED.
- Never present write actions as already completed.
When to Use
Use this when you need one high-quality brief from multiple connector sources and also want a governed path to write follow-up actions into systems like Jira, Confluence, Notion, or similar MCP-backed destinations.
It is ideal for weekly strategy briefs, cross-functional operating reviews, or leadership updates where evidence quality and action traceability matter.
Variables
| Variable | What to provide | Example |
|---|---|---|
goal | The core question this brief must answer | ”What changed in competitor pricing and what should we do this sprint?” |
time_window | Time period for evidence | ”Last 14 days” |
connected_sources | Systems and datasets to read | ”Google Drive strategy folder, OneDrive finance sheets, Jira project ABC, Confluence roadmap space” |
output_audience | Who will consume the brief | ”VP Product and GTM leadership” |
action_systems | Systems allowed for proposed writes | ”Jira + Confluence (draft updates only)“ |
risk_policy | Risk thresholds and constraints | ”Any high-risk action needs legal + eng manager approval” |
approval_rules | Human signoff flow | ”PM approves medium risk, Director approves high risk” |
Tips & Variations
- For faster cycles, ask for a short and long version in one run.
- Add a required
evidence_count_minrule for critical claims. - If your sources are noisy, ask the model to prioritize recency and official system-of-record sources.
- If you want tighter governance, require explicit
rollback_notefields in the action queue.
Example Output
A strong output includes a cited brief, a clearly marked proposed action table, and an MCP execution checklist where every write step is explicitly gated by human approval.