copilot-sdk-supercharged
v1.0.15
Published
GitHub Copilot SDK Supercharged - TypeScript SDK for programmatic control of GitHub Copilot CLI via JSON-RPC. Supports 21 languages.
Downloads
1,573
Maintainers
Readme
Copilot SDK for Node.js/TypeScript
TypeScript SDK for programmatic control of GitHub Copilot CLI via JSON-RPC.
Note: This SDK is in technical preview and may change in breaking ways.
Installation
npm install copilot-sdk-superchargedQuick Start
import { CopilotClient } from "copilot-sdk-supercharged";
// Create and start client
const client = new CopilotClient();
await client.start();
// Create a session
const session = await client.createSession({
model: "gpt-5",
});
// Wait for response using typed event handlers
const done = new Promise<void>((resolve) => {
session.on("assistant.message", (event) => {
console.log(event.data.content);
});
session.on("session.idle", () => {
resolve();
});
});
// Send a message and wait for completion
await session.send({ prompt: "What is 2+2?" });
await done;
// Clean up
await session.destroy();
await client.stop();API Reference
CopilotClient
Constructor
new CopilotClient(options?: CopilotClientOptions)Options:
cliPath?: string- Path to CLI executable (default: "copilot" from PATH)cliArgs?: string[]- Extra arguments prepended before SDK-managed flags (e.g.["./dist-cli/index.js"]when usingnode)cliUrl?: string- URL of existing CLI server to connect to (e.g.,"localhost:8080","http://127.0.0.1:9000", or just"8080"). When provided, the client will not spawn a CLI process.port?: number- Server port (default: 0 for random)useStdio?: boolean- Use stdio transport instead of TCP (default: true)logLevel?: string- Log level (default: "info")autoStart?: boolean- Auto-start server (default: true)autoRestart?: boolean- Auto-restart on crash (default: true)githubToken?: string- GitHub token for authentication. When provided, takes priority over other auth methods.useLoggedInUser?: boolean- Whether to use logged-in user for authentication (default: true, but false whengithubTokenis provided). Cannot be used withcliUrl.
Methods
start(): Promise<void>
Start the CLI server and establish connection.
stop(): Promise<Error[]>
Stop the server and close all sessions. Returns a list of any errors encountered during cleanup.
forceStop(): Promise<void>
Force stop the CLI server without graceful cleanup. Use when stop() takes too long.
createSession(config?: SessionConfig): Promise<CopilotSession>
Create a new conversation session.
Config:
sessionId?: string- Custom session ID.model?: string- Model to use ("gpt-5", "claude-sonnet-4.5", etc.). Required when using custom provider.reasoningEffort?: "low" | "medium" | "high" | "xhigh"- Reasoning effort level for models that support it. UselistModels()to check which models support this option.tools?: Tool[]- Custom tools exposed to the CLIsystemMessage?: SystemMessageConfig- System message customization (see below)infiniteSessions?: InfiniteSessionConfig- Configure automatic context compaction (see below)provider?: ProviderConfig- Custom API provider configuration (BYOK - Bring Your Own Key). See Custom Providers section.onUserInputRequest?: UserInputHandler- Handler for user input requests from the agent. Enables theask_usertool. See User Input Requests section.hooks?: SessionHooks- Hook handlers for session lifecycle events. See Session Hooks section.
resumeSession(sessionId: string, config?: ResumeSessionConfig): Promise<CopilotSession>
Resume an existing session. Returns the session with workspacePath populated if infinite sessions were enabled.
ping(message?: string): Promise<{ message: string; timestamp: number }>
Ping the server to check connectivity.
getState(): ConnectionState
Get current connection state.
listSessions(): Promise<SessionMetadata[]>
List all available sessions.
deleteSession(sessionId: string): Promise<void>
Delete a session and its data from disk.
getForegroundSessionId(): Promise<string | undefined>
Get the ID of the session currently displayed in the TUI. Only available when connecting to a server running in TUI+server mode (--ui-server).
setForegroundSessionId(sessionId: string): Promise<void>
Request the TUI to switch to displaying the specified session. Only available in TUI+server mode.
on(eventType: SessionLifecycleEventType, handler): () => void
Subscribe to a specific session lifecycle event type. Returns an unsubscribe function.
const unsubscribe = client.on("session.foreground", (event) => {
console.log(`Session ${event.sessionId} is now in foreground`);
});on(handler: SessionLifecycleHandler): () => void
Subscribe to all session lifecycle events. Returns an unsubscribe function.
const unsubscribe = client.on((event) => {
console.log(`${event.type}: ${event.sessionId}`);
});Lifecycle Event Types:
session.created- A new session was createdsession.deleted- A session was deletedsession.updated- A session was updated (e.g., new messages)session.foreground- A session became the foreground session in TUIsession.background- A session is no longer the foreground session
CopilotSession
Represents a single conversation session.
Properties
sessionId: string
The unique identifier for this session.
workspacePath?: string
Path to the session workspace directory when infinite sessions are enabled. Contains checkpoints/, plan.md, and files/ subdirectories. Undefined if infinite sessions are disabled.
Methods
send(options: MessageOptions): Promise<string>
Send a message to the session. Returns immediately after the message is queued; use event handlers or sendAndWait() to wait for completion.
Options:
prompt: string- The message/prompt to sendattachments?: Array<{type, path, displayName}>- File attachmentsmode?: "enqueue" | "immediate"- Delivery mode
Returns the message ID.
sendAndWait(options: MessageOptions, timeout?: number): Promise<AssistantMessageEvent | undefined>
Send a message and wait until the session becomes idle.
Options:
prompt: string- The message/prompt to sendattachments?: Array<{type, path, displayName}>- File attachmentsmode?: "enqueue" | "immediate"- Delivery modetimeout?: number- Optional timeout in milliseconds
Returns the final assistant message event, or undefined if none was received.
on(eventType: string, handler: TypedSessionEventHandler): () => void
Subscribe to a specific event type. The handler receives properly typed events.
// Listen for specific event types with full type inference
session.on("assistant.message", (event) => {
console.log(event.data.content); // TypeScript knows about event.data.content
});
session.on("session.idle", () => {
console.log("Session is idle");
});
// Listen to streaming events
session.on("assistant.message_delta", (event) => {
process.stdout.write(event.data.deltaContent);
});on(handler: SessionEventHandler): () => void
Subscribe to all session events. Returns an unsubscribe function.
const unsubscribe = session.on((event) => {
// Handle any event type
console.log(event.type, event);
});
// Later...
unsubscribe();abort(): Promise<void>
Abort the currently processing message in this session.
getMessages(): Promise<SessionEvent[]>
Get all events/messages from this session.
destroy(): Promise<void>
Destroy the session and free resources.
Event Types
Sessions emit various events during processing:
user.message- User message addedassistant.message- Assistant responseassistant.message_delta- Streaming response chunktool.execution_start- Tool execution startedtool.execution_complete- Tool execution completed- And more...
See SessionEvent type in the source for full details.
Image Support
The SDK supports image attachments via the attachments parameter. You can attach images by providing their file path:
await session.send({
prompt: "What's in this image?",
attachments: [
{
type: "file",
path: "/path/to/image.jpg",
},
],
});Supported image formats include JPG, PNG, GIF, and other common image types. The agent's view tool can also read images directly from the filesystem, so you can also ask questions like:
await session.send({ prompt: "What does the most recent jpg in this directory portray?" });Streaming
Enable streaming to receive assistant response chunks as they're generated:
const session = await client.createSession({
model: "gpt-5",
streaming: true,
});
// Wait for completion using typed event handlers
const done = new Promise<void>((resolve) => {
session.on("assistant.message_delta", (event) => {
// Streaming message chunk - print incrementally
process.stdout.write(event.data.deltaContent);
});
session.on("assistant.reasoning_delta", (event) => {
// Streaming reasoning chunk (if model supports reasoning)
process.stdout.write(event.data.deltaContent);
});
session.on("assistant.message", (event) => {
// Final message - complete content
console.log("\n--- Final message ---");
console.log(event.data.content);
});
session.on("assistant.reasoning", (event) => {
// Final reasoning content (if model supports reasoning)
console.log("--- Reasoning ---");
console.log(event.data.content);
});
session.on("session.idle", () => {
// Session finished processing
resolve();
});
});
await session.send({ prompt: "Tell me a short story" });
await done; // Wait for streaming to completeWhen streaming: true:
assistant.message_deltaevents are sent withdeltaContentcontaining incremental textassistant.reasoning_deltaevents are sent withdeltaContentfor reasoning/chain-of-thought (model-dependent)- Accumulate
deltaContentvalues to build the full response progressively - The final
assistant.messageandassistant.reasoningevents contain the complete content
Note: assistant.message and assistant.reasoning (final events) are always sent regardless of streaming setting.
Advanced Usage
Manual Server Control
const client = new CopilotClient({ autoStart: false });
// Start manually
await client.start();
// Use client...
// Stop manually
await client.stop();Tools
You can let the CLI call back into your process when the model needs capabilities you own. Use defineTool with Zod schemas for type-safe tool definitions:
import { z } from "zod";
import { CopilotClient, defineTool } from "copilot-sdk-supercharged";
const session = await client.createSession({
model: "gpt-5",
tools: [
defineTool("lookup_issue", {
description: "Fetch issue details from our tracker",
parameters: z.object({
id: z.string().describe("Issue identifier"),
}),
handler: async ({ id }) => {
const issue = await fetchIssue(id);
return issue;
},
}),
],
});When Copilot invokes lookup_issue, the client automatically runs your handler and responds to the CLI. Handlers can return any JSON-serializable value (automatically wrapped), a simple string, or a ToolResultObject for full control over result metadata. Raw JSON schemas are also supported if Zod isn't desired.
System Message Customization
Control the system prompt using systemMessage in session config:
const session = await client.createSession({
model: "gpt-5",
systemMessage: {
content: `
<workflow_rules>
- Always check for security vulnerabilities
- Suggest performance improvements when applicable
</workflow_rules>
`,
},
});The SDK auto-injects environment context, tool instructions, and security guardrails. The default CLI persona is preserved, and your content is appended after SDK-managed sections. To change the persona or fully redefine the prompt, use mode: "replace".
For full control (removes all guardrails), use mode: "replace":
const session = await client.createSession({
model: "gpt-5",
systemMessage: {
mode: "replace",
content: "You are a helpful assistant.",
},
});Infinite Sessions
By default, sessions use infinite sessions which automatically manage context window limits through background compaction and persist state to a workspace directory.
// Default: infinite sessions enabled with default thresholds
const session = await client.createSession({ model: "gpt-5" });
// Access the workspace path for checkpoints and files
console.log(session.workspacePath);
// => ~/.copilot/session-state/{sessionId}/
// Custom thresholds
const session = await client.createSession({
model: "gpt-5",
infiniteSessions: {
enabled: true,
backgroundCompactionThreshold: 0.80, // Start compacting at 80% context usage
bufferExhaustionThreshold: 0.95, // Block at 95% until compaction completes
},
});
// Disable infinite sessions
const session = await client.createSession({
model: "gpt-5",
infiniteSessions: { enabled: false },
});When enabled, sessions emit compaction events:
session.compaction_start- Background compaction startedsession.compaction_complete- Compaction finished (includes token counts)
Multiple Sessions
const session1 = await client.createSession({ model: "gpt-5" });
const session2 = await client.createSession({ model: "claude-sonnet-4.5" });
// Both sessions are independent
await session1.sendAndWait({ prompt: "Hello from session 1" });
await session2.sendAndWait({ prompt: "Hello from session 2" });Custom Session IDs
const session = await client.createSession({
sessionId: "my-custom-session-id",
model: "gpt-5",
});File Attachments
await session.send({
prompt: "Analyze this file",
attachments: [
{
type: "file",
path: "/path/to/file.js",
displayName: "My File",
},
],
});Custom Providers
The SDK supports custom OpenAI-compatible API providers (BYOK - Bring Your Own Key), including local providers like Ollama. When using a custom provider, you must specify the model explicitly.
ProviderConfig:
type?: "openai" | "azure" | "anthropic"- Provider type (default: "openai")baseUrl: string- API endpoint URL (required)apiKey?: string- API key (optional for local providers like Ollama)bearerToken?: string- Bearer token for authentication (takes precedence over apiKey)wireApi?: "completions" | "responses"- API format for OpenAI/Azure (default: "completions")azure?.apiVersion?: string- Azure API version (default: "2024-10-21")
Example with Ollama:
const session = await client.createSession({
model: "deepseek-coder-v2:16b", // Required when using custom provider
provider: {
type: "openai",
baseUrl: "http://localhost:11434/v1", // Ollama endpoint
// apiKey not required for Ollama
},
});
await session.sendAndWait({ prompt: "Hello!" });Example with custom OpenAI-compatible API:
const session = await client.createSession({
model: "gpt-4",
provider: {
type: "openai",
baseUrl: "https://my-api.example.com/v1",
apiKey: process.env.MY_API_KEY,
},
});Example with Azure OpenAI:
const session = await client.createSession({
model: "gpt-4",
provider: {
type: "azure", // Must be "azure" for Azure endpoints, NOT "openai"
baseUrl: "https://my-resource.openai.azure.com", // Just the host, no path
apiKey: process.env.AZURE_OPENAI_KEY,
azure: {
apiVersion: "2024-10-21",
},
},
});Important notes:
- When using a custom provider, the
modelparameter is required. The SDK will throw an error if no model is specified.- For Azure OpenAI endpoints (
*.openai.azure.com), you must usetype: "azure", nottype: "openai".- The
baseUrlshould be just the host (e.g.,https://my-resource.openai.azure.com). Do not include/openai/v1in the URL - the SDK handles path construction automatically.
User Input Requests
Enable the agent to ask questions to the user using the ask_user tool by providing an onUserInputRequest handler:
const session = await client.createSession({
model: "gpt-5",
onUserInputRequest: async (request, invocation) => {
// request.question - The question to ask
// request.choices - Optional array of choices for multiple choice
// request.allowFreeform - Whether freeform input is allowed (default: true)
console.log(`Agent asks: ${request.question}`);
if (request.choices) {
console.log(`Choices: ${request.choices.join(", ")}`);
}
// Return the user's response
return {
answer: "User's answer here",
wasFreeform: true, // Whether the answer was freeform (not from choices)
};
},
});Session Hooks
Hook into session lifecycle events by providing handlers in the hooks configuration:
const session = await client.createSession({
model: "gpt-5",
hooks: {
// Called before each tool execution
onPreToolUse: async (input, invocation) => {
console.log(`About to run tool: ${input.toolName}`);
// Return permission decision and optionally modify args
return {
permissionDecision: "allow", // "allow", "deny", or "ask"
modifiedArgs: input.toolArgs, // Optionally modify tool arguments
additionalContext: "Extra context for the model",
};
},
// Called after each tool execution
onPostToolUse: async (input, invocation) => {
console.log(`Tool ${input.toolName} completed`);
// Optionally modify the result or add context
return {
additionalContext: "Post-execution notes",
};
},
// Called when user submits a prompt
onUserPromptSubmitted: async (input, invocation) => {
console.log(`User prompt: ${input.prompt}`);
return {
modifiedPrompt: input.prompt, // Optionally modify the prompt
};
},
// Called when session starts
onSessionStart: async (input, invocation) => {
console.log(`Session started from: ${input.source}`); // "startup", "resume", "new"
return {
additionalContext: "Session initialization context",
};
},
// Called when session ends
onSessionEnd: async (input, invocation) => {
console.log(`Session ended: ${input.reason}`);
},
// Called when an error occurs
onErrorOccurred: async (input, invocation) => {
console.error(`Error in ${input.errorContext}: ${input.error}`);
return {
errorHandling: "retry", // "retry", "skip", or "abort"
};
},
},
});Available hooks:
onPreToolUse- Intercept tool calls before execution. Can allow/deny or modify arguments.onPostToolUse- Process tool results after execution. Can modify results or add context.onUserPromptSubmitted- Intercept user prompts. Can modify the prompt before processing.onSessionStart- Run logic when a session starts or resumes.onSessionEnd- Cleanup or logging when session ends.onErrorOccurred- Handle errors with retry/skip/abort strategies.
Error Handling
try {
const session = await client.createSession();
await session.send({ prompt: "Hello" });
} catch (error) {
console.error("Error:", error.message);
}Requirements
- Node.js >= 18.0.0
- GitHub Copilot CLI installed and in PATH (or provide custom
cliPath)
License
MIT
