copilot-researcher
v0.1.0
Published
Copilot-native researcher harness and local-first TypeScript SDK for research sessions, hypotheses, evidence, artifacts, and Markdown reports.
Maintainers
Readme
copilot-researcher
copilot-researcher is a Copilot-native researcher harness and local-first TypeScript CLI/SDK. GitHub Copilot CLI is the interactive runtime, and copilot-researcher keeps durable session, hypothesis, evidence, artifact, scoring, and Markdown report state in .researchflow/.
It is designed for teams that want AI-assisted research workflows without giving up inspectable state, reproducible rankings, or portable markdown outputs.
What it does
- scaffolds a GitHub Copilot CLI project with instructions, agents, skills, and a repo-local extension
- initializes a local-first research workspace
- tracks sessions, hypotheses, evidence, and artifacts in
.researchflow/ - ranks competing ideas and recommends next actions
- generates Markdown reports that can be shared in docs, issues, or lab notes
Why this shape
Most research tooling is either:
- notebook-heavy and manual
- cloud-first and opaque
- generic agent orchestration without durable research provenance
copilot-researcher keeps the durable layer explicit:
- local-first workspaces
- research sessions
- hypothesis + evidence lineage
- artifact manifests
- reproducible scoring and ranking
- Markdown reporting
GitHub Copilot CLI then sits on top as the execution harness with custom instructions, agents, skills, /fleet, autopilot, and repo-local extension tools.
Install
Prerequisites
- Node.js 20+
- GitHub Copilot CLI installed and authenticated
Install Copilot CLI:
npm install -g @github/copilotInstall the package:
npm install -g copilot-researcherRun it on demand with:
npx copilot-researcher --helpFor local development in this repository:
npm install
npm run buildIf you change src/cli.ts or src/copilot.ts, run npm run build before using the repo-local Copilot extension so dist/cli.js stays in sync with bridge options.
Quick start
Create a new Copilot-ready research project:
npx copilot-researcher init ./my-lab --name "Battery materials" --copilot
cd ./my-lab
copilotInside Copilot CLI you can now:
- use
/agentto switch into the project agents - use
/skills listto inspect the installed research skills - ask for a literature scout, hypothesis shortlist, experiment design, or report
- use
/fleetfor large parallelizable research jobs
Create and manage a session from the command line:
copilot-researcher session:create \
--goal "Find promising cathode additives for fast-charging cells" \
--question "Which additive should be validated first?" \
--template experiment-design
copilot-researcher hypothesis:add \
--session <session-id> \
--title "Boron-rich surface additive" \
--statement "A boron-rich coating may suppress interface degradation." \
--novelty 4 \
--feasibility 4 \
--impact 5 \
--confidence 3
copilot-researcher evidence:add \
--session <session-id> \
--hypothesis <hypothesis-id> \
--summary "Related interface-stability study suggests improved cycling." \
--source "doi:10.0000/example" \
--strength 4 \
--relevance 5 \
--direction support
copilot-researcher score --session <session-id>
copilot-researcher report --session <session-id>List the available workflow templates, including the physics simulation workflow:
copilot-researcher template:listTemplate library
copilot-researcher ships with a small set of templates that reuse the same ledger, scoring, and report pipeline:
literature-scout— scan prior work, capture hypotheses, and rank what deserves validationexperiment-design— turn candidate ideas into practical experiments and next stepsphysics-literature-scout— map governing equations, observables, benchmark families, and open modeling gapsphysics-simulation— plan a simulation campaign, including multiphysics coupling and publication-oriented outputsmultiphysics-validation— prioritize convergence, benchmark checks, and coupled-model credibilitysimulation-paper— turn validated results into publishable claims, figure plans, and reproducibility bundlesreplication-check— reproduce prior findings and capture missing artifacts or contradictions
Physics workflow templates
Use the built-in physics template family when you want one project to cover prior art, simulation hypotheses, validation gates, and paper-ready outputs:
physics-literature-scout— map governing equations, observables, benchmark families, and open modeling gaps before you commit to solver workphysics-simulation— plan the main simulation campaign, including multiphysics coupling, assumptions, and publication-oriented outputsmultiphysics-validation— prioritize convergence, benchmark checks, and coupled-model credibility before expanding claimssimulation-paper— turn validated results into publishable claims, figure plans, and reproducibility bundles
copilot-researcher session:create \
--goal "Prioritize a multiphysics simulation campaign for Hall thruster erosion" \
--question "Which coupled model should be validated first?" \
--template physics-simulationThe template is tuned to:
- rank simulation hypotheses with reproducibility-oriented weights
- capture benchmark or experimental evidence alongside contradictions
- track the artifact set usually needed for simulation work: datasets, scripts, figures, notes, and reports
- push reports toward publishable claims, validation gaps, and next simulation moves
For example, to start from prior art before building the main model:
copilot-researcher session:create \
--goal "Map governing equations and benchmark datasets for a plasma-wall interaction model" \
--question "Which observable and benchmark should anchor the first validation pass?" \
--template physics-literature-scoutOr, once the simulation is validated, shape it into a manuscript workflow:
copilot-researcher session:create \
--goal "Package a validated Hall thruster erosion result into a paper draft" \
--question "Which figure sequence and evidence bundle best support the central claim?" \
--template simulation-paperCopilot scaffold
Project installs create the following structure:
.github/
copilot-instructions.md
instructions/
agents/
skills/
extensions/
AGENTS.md
researchflow.yml
.researchflow/The repo-local extension exposes researchflow_* tools to Copilot so the agent can persist research state instead of leaving important work buried in chat text.
Copilot workflows
Project scaffold commands
copilot-researcher init <dir> --copilot— initialize a workspace and scaffold Copilot project filescopilot-researcher copilot install --scope project|user— install project scaffold files or user-scoped skillscopilot-researcher copilot update --scope project|user— refresh managed scaffold filescopilot-researcher copilot doctor --dir <dir>— verify Copilot CLI availability, workspace presence, and scaffold status
Recommended interactive flow
- Start with plan mode in Copilot CLI.
- Break the work into literature review, contradiction search, hypothesis creation, experiment design, and report synthesis.
- Use
/fleetwhen those threads can run independently. - Use autopilot only for bounded, well-specified plans.
- Persist durable findings with the
researchflow_*tools or thecopilot-researcherCLI.
Workspace layout
researchflow.yml
.researchflow/
templates/
sessions/
<session-id>/
session.json
events.jsonl
hypotheses/
evidence/
artifacts/
reports/CLI surface
Research ledger commands
copilot-researcher initcopilot-researcher session:createcopilot-researcher template:listcopilot-researcher hypothesis:addcopilot-researcher evidence:addcopilot-researcher artifact:addcopilot-researcher scorecopilot-researcher statuscopilot-researcher report
Copilot integration commands
copilot-researcher copilot installcopilot-researcher copilot updatecopilot-researcher copilot doctor
Library usage
import { CopilotResearcherWorkspace } from "copilot-researcher";
const workspace = await CopilotResearcherWorkspace.find(process.cwd());
const session = await workspace.createSession({
goal: "Identify next replication target",
question: "What should we validate first?",
templateId: "replication-check"
});Scope
copilot-researcher is not a replacement terminal chat client and not a generic multi-agent framework. It focuses on the inspectable, durable layer around research work:
- what hypotheses were proposed
- what evidence supported or contradicted them
- what artifacts were produced
- what should happen next
That makes it a strong substrate beneath GitHub Copilot CLI, notebooks, scripts, and manual research workflows.
