envpkt
v0.7.1
Published
Credential lifecycle and fleet management for AI agents
Maintainers
Readme
envpkt
Credentials your agents actually understand.
Structured metadata for every secret — capabilities, constraints, expiration, and fleet health — so agents operate within their boundaries instead of flying blind.
Every credential in your system gets an envpkt.toml entry describing what service it authenticates to, what it's allowed to do, when it expires, and how to rotate it. Your agents query this metadata via MCP to understand their operating constraints. Your operators audit credential health across entire agent fleets. The secrets themselves stay where they belong — in your secrets manager, encrypted at rest, or injected at runtime — never in the agent's conversation context.
MCP Integration
envpkt ships an MCP server that gives AI agents structured awareness of their credentials. Add it to Claude, Cursor, VS Code, or any MCP-compatible client:
{
"mcpServers": {
"envpkt": {
"command": "envpkt",
"args": ["mcp"]
}
}
}Tools
| Tool | Description |
| ------------------ | ------------------------------------------------------- |
| getPacketHealth | Get overall health status with per-secret audit results |
| listCapabilities | List agent and per-secret capabilities |
| getSecretMeta | Get metadata for a specific secret by key |
| checkExpiration | Check expiration status and days remaining |
| getEnvMeta | Get metadata for environment defaults and drift status |
Resources
| URI | Description |
| ----------------------- | --------------------------------- |
| envpkt://health | Current credential health summary |
| envpkt://capabilities | Agent and secret capabilities |
The MCP server exposes metadata only — it does not have access to secret values. See Security Model for details.
Security Model
envpkt operates a three-tier trust model. Each tier has different guarantees, and we're explicit about what each one protects against.
Tier 1: MCP metadata (agent-facing) — The MCP server never returns raw credential values. This isn't a policy choice — architecturally, the server reads envpkt.toml which contains metadata (service names, capabilities, expiration dates, rotation URLs) but never plaintext secrets. The agent gets structured awareness of its constraints without any secret material entering the LLM context window. Prompt injection attacks cannot leak what isn't there.
Tier 2: Runtime injection (process-facing) — boot() resolves secrets (from sealed packets, fnox, or environment variables) and injects them into process.env at startup, outside the LLM context. This is the same trust model as every Node.js application that reads from .env, except now secrets are encrypted at rest, scoped per-agent, and auditable. This is defense-in-depth against prompt injection — the most common attack vector — but it is not a hard boundary against agents with code execution capabilities.
Tier 3: Shell-level agents — Agents with shell access (Claude Code, Devin, etc.) can read environment variables directly. Prevention isn't possible at this tier. envpkt provides encrypted storage, scoped access, and audit trails — because when prevention isn't possible, visibility is what matters.
Quick Start
Start where your credentials already are — environment variables — and graduate to encrypted, per-agent-scoped metadata.
# Install
npm install -g envpkt
# Auto-discover credentials from your shell environment
envpkt env scan
# Scaffold envpkt.toml from discovered credentials
envpkt env scan --write
# Audit credential health
envpkt audit
# Check for drift between envpkt.toml and live environment
envpkt env check
# Scan a directory tree of agents
envpkt fleetThe envpkt.toml File
Every project gets one envpkt.toml that describes its credentials. Here's a minimal example:
#:schema https://raw.githubusercontent.com/jordanburke/envpkt/main/schemas/envpkt.schema.json
version = 1
[secret.API_KEY]
service = "stripe"And a more complete one:
#:schema https://raw.githubusercontent.com/jordanburke/envpkt/main/schemas/envpkt.schema.json
version = 1
[identity]
name = "billing-service"
consumer = "agent"
description = "Payment processing agent"
capabilities = ["charge", "refund"]
expires = "2027-01-01"
[lifecycle]
stale_warning_days = 90
require_expiration = true
require_service = true
[secret.STRIPE_SECRET_KEY]
service = "stripe"
purpose = "Process customer payments and manage subscriptions"
capabilities = ["charges:write", "subscriptions:write"]
created = "2026-01-15"
expires = "2027-01-15"
rotation_url = "https://dashboard.stripe.com/apikeys"
source = "vault"
[secret.DATABASE_URL]
service = "postgres"
purpose = "Read/write access to the billing database"
capabilities = ["SELECT", "INSERT", "UPDATE"]
created = "2026-02-01"
expires = "2026-08-01"
rotation_url = "https://wiki.internal/runbooks/rotate-db-creds"
source = "vault"For non-secret configuration defaults (runtime mode, log levels, etc.), use [env.*]:
[env.NODE_ENV]
value = "production"
purpose = "Runtime environment mode"
comment = "Override to 'development' for local testing"
[env.LOG_LEVEL]
value = "info"
purpose = "Application log verbosity"See examples/ for more configurations.
Sealed Packets
Sealed packets embed age-encrypted secret values directly in envpkt.toml. This makes your config fully self-contained — no external secrets backend needed at runtime.
Setup
# Generate an age keypair
age-keygen -o identity.txt
# public key: age1ql3z7hjy54pw3hyww5ayyfg7zqgvc7w3j2elw8zmrj2kg5sfn9aqmcac8pAdd the public key to your config and the identity file to .gitignore:
[identity]
name = "my-agent"
recipient = "age1ql3z7hjy54pw3hyww5ayyfg7zqgvc7w3j2elw8zmrj2kg5sfn9aqmcac8p"
key_file = "identity.txt"The key_file path supports ~ expansion and environment variables ($VAR, ${VAR}), so you can use paths like ~/keys/identity.txt or $KEYS_DIR/identity.txt. Relative paths are resolved from the config file's directory. When omitted, envpkt falls back to ENVPKT_AGE_KEY_FILE env var, then ~/.envpkt/age-key.txt.
Seal
envpkt sealEach secret gets an encrypted_value field with age-armored ciphertext. The TOML (including ciphertext) is safe to commit.
Boot
At runtime, sealed values are automatically decrypted:
import { boot } from "envpkt"
const result = boot() // decrypts sealed values, injects into process.envMixed mode is supported — sealed values take priority, with fnox as fallback for keys without encrypted_value.
Fleet Management
When you're running multiple agents, envpkt fleet scans a directory tree for envpkt.toml files and aggregates credential health across your entire fleet.
envpkt fleet # Scan current directory (depth 3)
envpkt fleet -d /opt/agents # Scan specific directory
envpkt fleet --depth 5 # Increase scan depth
envpkt fleet --format json # JSON output
envpkt fleet --status critical # Filter agents by health statusShared Catalog
When multiple agents consume the same secrets, a shared catalog prevents metadata duplication. Define secret metadata once in a central file, then have each agent reference it.
Catalog file (infra/envpkt.toml)
version = 1
[lifecycle]
stale_warning_days = 90
require_expiration = true
[secret.DATABASE_URL]
service = "postgres"
purpose = "Primary application database"
capabilities = ["SELECT", "INSERT", "UPDATE", "DELETE"]
rotation_url = "https://wiki.internal/runbooks/rotate-db"
source = "vault"
created = "2025-11-01"
expires = "2026-11-01"
[secret.REDIS_URL]
service = "redis"
purpose = "Caching and session storage"
created = "2025-11-01"
expires = "2026-11-01"Agent file (agents/pipeline/envpkt.toml)
version = 1
catalog = "../../infra/envpkt.toml"
[identity]
name = "data-pipeline"
consumer = "agent"
secrets = ["DATABASE_URL", "REDIS_URL"]
# Optional: narrow the catalog definition for this agent
[secret.DATABASE_URL]
capabilities = ["SELECT"]Resolve to a flat config
envpkt resolve -c agents/pipeline/envpkt.tomlThis produces a self-contained config with catalog metadata merged in and agent overrides applied. The resolved output has no catalog reference — it's ready for deployment.
Merge rules
- Each field in the agent's
[secret.KEY]override replaces the catalog field (shallow merge) - Omitted fields keep the catalog value
identity.secretsis the source of truth for which keys the agent needs
How envpkt Compares
The agentic credential space is splitting into approaches. Here's where envpkt fits:
| | envpkt | agent-vault | AgentSecrets | 1Password Agentic | Infisical | | --------------------------- | ----------------------------------------------------------- | --------------------------- | -------------------------- | ----------------------------- | -------------------- | | Core approach | Metadata sidecar | Git-based secret storage | Proxy injection | Browser autofill | Secret retrieval API | | What agents see | Structured metadata (capabilities, constraints, expiration) | Raw secret values | Nothing (proxy handles it) | Nothing (autofill handles it) | Raw secret values | | MCP server | Yes | Yes | No | No | Yes | | Encryption at rest | age sealed packets | Git-crypt | N/A (proxy model) | Vault encryption | Vault encryption | | Per-agent scoping | Yes (identity.secrets, capabilities) | Yes (policies) | Yes (proxy rules) | No | Yes (policies) | | Fleet health monitoring | Yes (fleet scan, aggregated audit) | No | No | No | No | | Credential metadata | Rich (purpose, capabilities, rotation, lifecycle) | Minimal | Minimal | Minimal | Moderate | | Adoption path | Scan existing env vars, add metadata incrementally | New secret storage workflow | Proxy configuration | Browser extension | API integration |
envpkt's angle: Competitors are fighting over how secrets move — retrieval vs. proxy vs. autofill. envpkt owns what secrets mean. Rate limits, expiration policies, capability scopes, rotation runbooks — structured semantics that travel with the credential. That's the layer the others don't have.
Note: This comparison reflects publicly available information. Verify current feature sets before making procurement decisions.
CLI Reference
envpkt init
Generate an envpkt.toml template in the current directory.
envpkt init # Basic template
envpkt init --from-fnox # Scaffold from fnox.toml
envpkt init --identity --name "my-agent" # Include identity section
envpkt init --catalog "../infra/envpkt.toml" # Reference a shared catalog
envpkt init --identity --name "bot" --capabilities "read,write" --expires "2027-01-01"envpkt audit
Check credential health against lifecycle policies. Automatically resolves catalog references.
envpkt audit # Table output
envpkt audit --format json # JSON output
envpkt audit --expiring 14 # Show secrets expiring within 14 days
envpkt audit --status expired # Filter by status
envpkt audit --strict # Exit non-zero on any non-healthy secret
envpkt audit -c path/to/envpkt.toml # Specify config pathExit codes: 0 = healthy, 1 = degraded, 2 = critical.
envpkt resolve
Resolve catalog references and output a flat, self-contained config.
envpkt resolve -c agent.toml # Output resolved TOML to stdout
envpkt resolve -c agent.toml --format json # Output as JSON
envpkt resolve -c agent.toml -o resolved.toml # Write to file
envpkt resolve -c agent.toml --dry-run # Preview without writingConfigs without a catalog field pass through unchanged.
envpkt inspect
Display a structured view of an envpkt.toml file. Automatically resolves catalog references.
envpkt inspect # Current directory
envpkt inspect -c path/to/envpkt.toml # Specific file
envpkt inspect --format json # Raw JSON dump
envpkt inspect --resolved # Show resolved view (catalog merged)
envpkt inspect --secrets # Show secret values from env (masked)
envpkt inspect --secrets --plaintext # Show secret values in plaintextThe --secrets flag reads values from environment variables matching each secret key. By default values are masked (pos•••••yapp). Add --plaintext to display full values.
envpkt exec
Run a pre-flight audit, inject secrets from fnox into the environment, then execute a command.
envpkt exec -- node server.js # Audit then run
envpkt exec --skip-audit -- npm start # Skip the audit
envpkt exec --strict -- ./deploy.sh # Abort if audit is not healthy
envpkt exec --profile staging -- ... # Use a specific fnox profileenvpkt seal
Encrypt secret values into envpkt.toml using age. Sealed values are safe to commit to git — only the holder of the private key can decrypt them.
envpkt seal # Seal all secrets in envpkt.toml
envpkt seal -c path/to/envpkt.toml # Specify config path
envpkt seal --profile staging # Use a specific fnox profile for value resolutionRequires identity.recipient (age public key) in your config. Values are resolved via cascade:
- fnox (if available)
- Environment variables (e.g.
OPENAI_API_KEYin your shell) - Interactive prompt (asks you to paste each value)
After sealing, each secret gets an encrypted_value field. At boot time, envpkt exec or boot() automatically decrypts sealed values using the identity.key_file path (or the default ~/.envpkt/age-key.txt).
See examples/sealed-agent.toml for a complete example.
envpkt env scan
Auto-discover credentials from your shell environment. Matches env vars against ~45 known services (exact name), ~13 generic suffix patterns (*_API_KEY, *_SECRET, *_TOKEN, etc.), and ~29 value shape patterns (sk-*, ghp_*, AKIA*, postgres://, etc.).
envpkt env scan # Table output with confidence icons
envpkt env scan --format json # JSON output
envpkt env scan --write # Write/append to envpkt.toml
envpkt env scan --dry-run # Preview TOML that would be written
envpkt env scan --include-unknown # Include vars with no inferred serviceConfidence levels:
- High (●) — exact name match or recognized value prefix
- Medium (◐) — generic suffix pattern with derived service name
envpkt env check
Bidirectional drift detection between envpkt.toml and the live shell environment. Checks both directions: TOML keys missing from env, and credential-shaped env vars not tracked in TOML.
envpkt env check # Table output
envpkt env check --format json # JSON output
envpkt env check --strict # Exit non-zero on any drift
envpkt env check -c path/to/envpkt.toml # Specify config pathenvpkt env export
Output export KEY='VALUE' statements for sourcing secrets into the current shell. Secrets are resolved via sealed packets and/or fnox — the same pipeline as envpkt exec, but instead of spawning a subprocess, the output is designed to be eval'd.
# Source secrets into the current shell
eval "$(envpkt env export)"
# Use a specific fnox profile
eval "$(envpkt env export --profile staging)"
# Specify config path
eval "$(envpkt env export -c path/to/envpkt.toml)"Add to your shell startup (e.g. ~/.zshrc or ~/.bashrc) for automatic secret loading. envpkt's config discovery chain finds your config automatically — no platform-specific shell logic needed:
eval "$(envpkt env export 2>/dev/null)"envpkt shell-hook
Output a shell function that runs envpkt audit --format minimal whenever you cd into a directory. envpkt's config discovery chain automatically finds config files beyond CWD (see Config Resolution), so the hook works even in directories without a local envpkt.toml.
# Add to your .zshrc
eval "$(envpkt shell-hook zsh)"
# Add to your .bashrc
eval "$(envpkt shell-hook bash)"envpkt mcp
Start the envpkt MCP server (stdio transport) for AI agent integration.
envpkt mcpConfig Resolution
Commands that read envpkt.toml resolve the config path via a priority chain:
- Explicit flag —
-c path/to/envpkt.toml - Environment variable —
ENVPKT_CONFIG - Discovery chain — searches in order:
./envpkt.toml(current working directory)- Custom paths in
ENVPKT_SEARCH_PATH(colon-separated) ~/.envpkt/envpkt.toml(user home)- Cloud storage paths (OneDrive, iCloud, Dropbox, Google Drive)
When a config is discovered outside CWD, envpkt prints where it loaded from to stderr:
envpkt: loaded /Users/you/.envpkt/envpkt.tomlENVPKT_SEARCH_PATH
Prepend custom search locations (colon-separated paths to envpkt.toml files):
export ENVPKT_SEARCH_PATH="$HOME/OneDrive/.envpkt/envpkt.toml:/custom/path/envpkt.toml"These are searched after CWD but before the built-in candidate paths. Useful for corporate OneDrive names, Google Drive with email in the path, or any non-standard location.
Library API
envpkt is also available as a TypeScript library with a functional programming API built on functype. All functions return Either<Error, Result> or Option<T> — no thrown exceptions.
import { boot, bootSafe, loadConfig, computeAudit, scanFleet, resolveConfig } from "envpkt"
// Boot API — load config, resolve catalog, audit, inject secrets
const result = boot({ configPath: "envpkt.toml", inject: true })
console.log(result.audit.status) // "healthy" | "degraded" | "critical"
console.log(result.configSource) // "flag" | "env" | "cwd" | "search"
// Safe variant returns Either instead of throwing
const safe = bootSafe({ configPath: "envpkt.toml" })
safe.fold(
(err) => console.error("Boot failed:", err._tag),
(result) => console.log(`${result.injected.length} secrets injected`),
)
// Load and audit directly
const config = loadConfig("envpkt.toml")
config.fold(
(err) => console.error("Failed:", err._tag),
(config) => {
const audit = computeAudit(config)
audit.secrets.forEach((s) => {
s.days_remaining.fold(
() => console.log(`${s.key}: no expiration set`),
(days) => console.log(`${s.key}: ${days} days remaining`),
)
})
},
)
// Fleet scan
const fleet = scanFleet("/opt/agents", { maxDepth: 3 })
console.log(`${fleet.total_agents} agents, ${fleet.total_secrets} secrets`)Framework Integration
boot() runs before your agent framework initializes, making it compatible with any framework:
import { boot } from "envpkt"
// Resolve and inject credentials before agent startup
const result = boot({ configPath: "envpkt.toml", inject: true })
console.log(`${result.audit.status} — ${result.injected.length} secrets loaded`)
// Now start your agent framework — process.env is populated
// Works with LangChain, CrewAI, AutoGen, or any framework that reads from process.envPacket Formatting API
import { formatPacket, maskValue } from "envpkt"
// formatPacket produces a human-readable text summary of a resolved config
const text = formatPacket(resolveResult)
// With secret values (masked by default)
const masked = formatPacket(resolveResult, {
secrets: { DATABASE_URL: "postgres://user:pass@host/db" },
})
// DATABASE_URL → postgres = pos•••••t/db
// With plaintext secret values
const plain = formatPacket(resolveResult, {
secrets: { DATABASE_URL: "postgres://user:pass@host/db" },
secretDisplay: "plaintext",
})Environment Scan/Check API
import { envScan, envCheck, generateTomlFromScan, matchEnvVar } from "envpkt"
// Scan process.env for credentials
const scan = envScan(process.env)
console.log(`Found ${scan.discovered.size} credentials (${scan.high_confidence} high confidence)`)
scan.discovered.forEach((m) => {
const svc = m.service.fold(
() => "unknown",
(s) => s,
)
console.log(` ${m.envVar} → ${svc} (${m.confidence})`)
})
// Generate TOML blocks from scan results
const toml = generateTomlFromScan(scan.discovered.toArray())
// Check drift between config and live env
import { loadConfig } from "envpkt"
loadConfig("envpkt.toml").fold(
(err) => console.error(err),
(config) => {
const check = envCheck(config, process.env)
if (!check.is_clean) {
console.log(`${check.missing_from_env} missing, ${check.untracked_credentials} untracked`)
}
},
)
// Match a single env var
matchEnvVar("OPENAI_API_KEY", "sk-test123").fold(
() => console.log("Not a credential"),
(m) => console.log(`Matched: ${m.confidence} confidence`),
)Seal API
import { ageEncrypt, ageDecrypt, sealSecrets, unsealSecrets } from "envpkt"
// Encrypt a single value
const encrypted = ageEncrypt("sk-my-api-key", "age1ql3z7hjy...")
encrypted.fold(
(err) => console.error("Encrypt failed:", err.message),
(ciphertext) => console.log(ciphertext), // -----BEGIN AGE ENCRYPTED FILE-----
)
// Decrypt a single value
const decrypted = ageDecrypt(ciphertext, "/path/to/identity.txt")
// Seal all secrets in a config's meta
const sealed = sealSecrets(config.meta, { OPENAI_API_KEY: "sk-..." }, recipientPublicKey)
// Unseal all encrypted_value entries
const values = unsealSecrets(config.meta, "/path/to/identity.txt")
values.fold(
(err) => console.error("Unseal failed:", err.message),
(secrets) => console.log(secrets), // { OPENAI_API_KEY: "sk-..." }
)Catalog Resolution API
import { loadConfig, resolveConfig } from "envpkt"
import { dirname } from "node:path"
const configPath = "agents/pipeline/envpkt.toml"
loadConfig(configPath).fold(
(err) => console.error(err),
(config) => {
resolveConfig(config, dirname(configPath)).fold(
(err) => console.error("Catalog error:", err._tag),
(result) => {
console.log("Resolved keys:", result.merged)
console.log("Overridden:", result.overridden)
// result.config is the flat, self-contained config
},
)
},
)Schema
envpkt.toml is validated against a JSON Schema. Editors with TOML + JSON Schema support will provide autocompletion and validation when the #:schema directive is present on line 1.
The schema is published at:
- npm:
envpkt/schema(importable via package exports) - GitHub:
schemas/envpkt.schema.json
Secret Metadata Fields
Each [secret.<KEY>] section describes a secret:
| Tier | Fields | Description |
| --------------- | ----------------------------------------------- | ------------------------------------------- |
| Scan-first | service, expires, rotation_url | Key health indicators for audit |
| Context | purpose, comment, capabilities, created | Why this secret exists and what it grants |
| Operational | rotates, rate_limit, model_hint, source | Runtime and provisioning info |
| Sealed | encrypted_value | Age-encrypted secret value (safe to commit) |
| Enforcement | required, tags | Filtering, grouping, and policy |
Identity
The optional [identity] section identifies the consumer of these credentials:
[identity]
name = "data-pipeline-agent"
consumer = "agent" # agent | service | developer | ci
description = "ETL pipeline processor"
capabilities = ["read-s3", "write-postgres"]
expires = "2027-01-01"
services = ["aws", "postgres"]
secrets = ["DATABASE_URL", "AWS_KEY"] # When using a catalogLifecycle Policy
The optional [lifecycle] section configures audit behavior:
[lifecycle]
stale_warning_days = 90 # Flag secrets older than N days without updates
require_expiration = true # Require expires on all secrets
require_service = true # Require service on all secretsfnox Integration
envpkt integrates with fnox for secret resolution:
envpkt init --from-fnoxscaffolds[secret.*]entries fromfnox.tomlenvpkt auditdetects orphaned keys (in envpkt but not in fnox, or vice versa)envpkt execinjects fnox secrets into the subprocess environment
Development
pnpm install
pnpm validate # format + lint + typecheck + test + build:schema + build
pnpm test # Run tests only
pnpm dev # Watch mode
pnpm demo # Regenerate demo HTML renders in examples/demo/See examples/demo/ for a walkthrough of the catalog system with 3 agents, including styled HTML renders of the inspect output in all 3 display modes (no secrets, masked, plaintext).
License
Apache-2.0
