@x12i/logxer
v4.3.6
Published
Standardized logging gateway for Node.js with PII sanitization, correlation trails, Shadow Logging for test/debug capture, scoping with text filters, story output, troubleshooting integration, and multi-format output (JSON/YAML/text)
Maintainers
Readme
@x12i/logxer
A standardized logging gateway for Node.js applications. Flexible multi-transport logging with console, file, and unified-logger outputs; ENV-first configuration; PII/credentials sanitization; YAML/JSON/table/text formats; per-run Shadow Logging for test/debug capture; and an official queryable runtime log method for debug tooling (getJobLogs).
🚀 Env-Ready Component (ERC 2.0)
This component supports zero-config initialization via environment variables and is compliant with the Env-Ready Component Standard (ERC 2.0).
Package-level .env contract (libraries)
For per-package log thresholds, Logxer follows {PREFIX}_LOGS_LEVEL (canonical). {PREFIX}_LOG_LEVEL remains supported when _LOGS_LEVEL is not set in the environment. Omitting both keys defaults to warn (warn and error only); set {PREFIX}_LOGS_LEVEL=off to silence. Cross-cutting options (console, file, format) stay host-level. Full detail: docs/package-usage.md (published on npm). Helpers: resolvePackageLogsLevel, parsePackageLogsLevelString, packageLogsLevelEnvKey.
ERC 2.0 Compliance
- ✅ Auto-Discovery: Zero-config initialization from environment variables
- ✅ Complete Documentation: All environment variables documented (including dependencies)
- ✅ Type Safety: Automatic type coercion and validation
- ✅ Manifest: Auto-generated
erc-manifest.jsonwith all requirements - ✅ Example File: Auto-generated
.env.examplewith all transitive requirements - ✅ Dependency Tracking: Documents both ERC and non-ERC dependencies
Quick Start (Zero-Config Mode)
# 1. Install the package
npm install @x12i/logxer
# 2. Set environment variables (replace MY_APP with your prefix)
export MY_APP_LOG_TO_CONSOLE=true
export MY_APP_LOGS_LEVEL=warn
export MY_APP_LOG_FORMAT=json
# 3. Use with zero config!
import { createLogxer } from '@x12i/logxer';
const logxer = createLogxer(
{ packageName: 'MY_APP', envPrefix: 'MY_APP' }
); // Auto-discovers from process.envAdvanced Mode (Programmatic Configuration)
import { createLogxer } from '@x12i/logxer';
const logxer = createLogxer(
{ packageName: 'MY_APP', envPrefix: 'MY_APP' },
{
logToFile: true,
logFilePath: '/var/log/myapp.log',
logLevel: 'info'
}
);Queryable runtime logs (debug tooling)
Logxer exposes an official query method for runtime observability UIs:
const res = await logxer.getJobLogs({
jobId,
graphId,
nodeId,
level: ["warn", "error"],
limit: 500,
});- The returned
linesare structured envelopes (not preformatted strings) and include all metadata needed for filtering/grouping in a UI. - The backing store is in-process (TTL + max entries). It does not read from Mongo/GCS/files.
- Full guidance:
docs/runtime-observability.md.
Environment Variables
Note: This component uses dynamic environment variable prefixes based on packageConfig.envPrefix. Replace {PREFIX} with your actual prefix (e.g., MY_APP, API_SERVICE).
See .env.example for the complete list of required and optional variables with descriptions. Generate it by running:
npm run generate-ercDependencies
- ✅ nx-config2 (ERC 2.0) - Configuration engine
- ℹ️ @x-developer/unified-logger (non-ERC) - Requirements manually documented
- ℹ️ nx-troubleshooting (non-ERC, optional) - Requirements manually documented
Features
- ✅ Console output (default)
- ✅ File output (optional)
- ✅ Unified-logger output (Papertrail/UDP/Console via
@x-developer/unified-logger) - ✅ Dual/triple output (console + file + unified-logger)
- ✅ Environment variable configuration
- ✅ Custom logger injection (Winston, Pino, etc.)
- ✅ Package-specific prefixes
- ✅ Five log levels:
verbose,debug,info,warn,error - ✅ Multiple formats:
text,json,yaml(console & file; unified stays JSON) - ✅ PII/Credentials sanitization (auto-detect & mask, opt-in)
- ✅ Dual trails: operation (Depth) & thread (Causal), plus
runId,jobId,correlationId,sessionId - ✅ OpenTelemetry context (
traceId,spanId) when available - ✅ Routing metadata (control outputs per entry)
- ✅ Recursion safety (prevent circular/unified feedback)
- ✅ TypeScript support
- ✅ Shadow Logging (per-run capture) with TTL & forced-verbose, great for tests
- ✅ Scoping – Derive focused subsets of logs with text filters and correlation keys
- ✅ Story Output – Human-readable narrative format built automatically from log entries
- ✅ Troubleshooting Integration – Wire in
nx-troubleshootingfor intelligent error-to-solution matching - ✅ DebugLogAbstract – Semantic log categories (
STATE|INTENT|EVENT|TRACE|ANOMALY) - ✅ RuntimeIdentity – Structured runtime context attached per log call
Debug Log Abstract
Logxer encourages categorizing your debug logs into five semantic roles to make log streams instantly readable and manageable. Import the DebugLogAbstract object and attach a debugKind to your log metadata:
- STATE (The Data): Snapshots of values at a given moment (inputs, outputs, current state).
logger.debug('Current user object', { debugKind: DebugLogAbstract.STATE, user }); - INTENT (The Future): Declarations of what the code is about to do — highly useful for finding where a script hangs.
logger.info('Attempting database connection...', { debugKind: DebugLogAbstract.INTENT }); - EVENT (The Past): Confirmations that an action successfully finished.
logger.info('Batch processing complete.', { debugKind: DebugLogAbstract.EVENT }); - TRACE (The Location): Breadcrumbs used purely to map the execution flow.
logger.verbose('Entered init() function', { debugKind: DebugLogAbstract.TRACE }); - ANOMALY (The Unexpected): Everything that deviates from the happy path (errors, warnings, fallbacks).
logger.warn('API timeout, using cached data', { debugKind: DebugLogAbstract.ANOMALY });
Runtime Identity
Logxer provides a structured RuntimeIdentity object to track the exact context of the executing unit (e.g., service name, version, host, region, pid).
You can configure this globally when creating the logger, or pass it on individual log calls. It is forwarded to all sinks and heavily aids filtering in observability tools.
const runtimeIdentity: RuntimeIdentity = {
service: 'payment-processor',
version: '2.1.0',
env: 'production',
host: 'pod-abcd-1234',
region: 'us-east-1'
};
// Option A: Set it globally during initialization
const logger = createLogger(
{ packageName: 'PAYMENTS', envPrefix: 'PAY_SRV' },
{ runtimeIdentity }
);
// Option B: Attach it per log call
logger.info('Payment processed', { runtimeIdentity, paymentId: '123' });Installation
npm install @x12i/logxerFor troubleshooting integration:
npm install @x12i/logxer nx-troubleshootingQuick Start
import { createLogger } from 'logs-gateway';
const logger = createLogger(
{ packageName: 'MY_APP', envPrefix: 'MY_APP' },
{
logToFile: true,
logFilePath: '/var/log/myapp.log',
logLevel: 'info', // verbose|debug|info|warn|error
logFormat: 'json', // text|json|yaml (yaml: console/file only)
enableUnifiedLogger: true,
unifiedLogger: {
transports: { papertrail: true },
service: 'my-app',
env: 'production'
},
// Optional: per-run Shadow Logging defaults (can also be enabled at runtime)
shadow: { enabled: false, ttlMs: 86_400_000 }, // 1 day
// Optional: Scoping & Troubleshooting
scoping: {
enabled: true,
errorScoping: { enabled: true, windowMsBefore: 60_000, windowMsAfter: 30_000 },
buffer: { maxEntries: 5000, preferShadow: true }
},
troubleshooting: {
enabled: true,
narrativesPath: './metadata/troubleshooting.json',
output: { formats: ['markdown'], emitAsLogEntry: true }
}
}
);
// Standard usage
logger.verbose('Very detailed info', { step: 'init', debugKind: DebugLogAbstract.TRACE });
logger.debug('Debug info', { data: 'x', debugKind: DebugLogAbstract.STATE });
logger.info('Application initialized', { version: '1.0.0', debugKind: DebugLogAbstract.EVENT });
logger.warn('Deprecated feature used', { debugKind: DebugLogAbstract.ANOMALY });
logger.error('Error occurred', { error: new Error('boom'), debugKind: DebugLogAbstract.ANOMALY });Automatic Application Identification
logs-gateway automatically detects and includes your application's package name and version in every log entry. This is done by reading the consuming project's package.json file (the project using logs-gateway, not logs-gateway itself).
How It Works
When you create a logger instance, logs-gateway automatically:
- Searches up the directory tree from
process.cwd()to find the nearestpackage.json - Extracts the
nameandversionfields from that file - Includes them in all log entries as
appNameandappVersion
This happens automatically - no configuration needed! The detection is cached after the first call for performance.
Example
If your project's package.json contains:
{
"name": "my-awesome-app",
"version": "2.1.0"
}Then all log entries will automatically include:
{
"timestamp": "2025-01-15T10:30:45.123Z",
"package": "MY_APP",
"level": "INFO",
"message": "Application initialized",
"appName": "my-awesome-app",
"appVersion": "2.1.0",
"data": { ... }
}Benefits
- Traceability: Know exactly which application version generated each log entry
- Debugging: Filter logs by application version in centralized logging systems
- Deployment Tracking: Identify which deployments are running in production
- Zero Configuration: Works automatically without any API changes
Important Notes
- The detection searches from
process.cwd()(the working directory where your app runs) - It stops at the first
package.jsonthat is not innode_modules - If no
package.jsonis found,appNameandappVersionare simply omitted (no error) - The result is cached after first detection for performance
Overview: Scoping, Story Output & Troubleshooting
This extension adds four major capabilities:
Runtime Filtering (logger-debug.json) – Filter logs at the source before they're written. Place
logger-debug.jsonat your project root to filter logs by identity or application name. This reduces noise at runtime and complements post-processing scoping.Scoping – Given a verbose log stream, derive a focused subset of logs relevant to a problem or question. Scopes can be:
- Error-centric (anchor on a specific
errorentry) - Run/Correlation-centric (anchor on
runId,correlationId, etc.) - Text-based (filter by message/data content)
- Narrative-based (optional, driven by "scoping narratives")
Scoping always uses verbose logs if available, regardless of the current log level.
- Error-centric (anchor on a specific
Story vs Full Data Output – A scope can be returned as:
- Full data: structured
ScopedLogViewwith all entries - Story: human-readable narrative built automatically from entries using a generic
scopeRecordhelper - Or both
- Full data: structured
Troubleshooting Integration – Wire in
nx-troubleshootingso that errors/scopes:- Are matched to troubleshooting narratives, and
- Produce troubleshooting artifacts (Markdown/JSON/text) as another log output channel (
troubleshooting)
The same scopeRecord helper is also exported as a generic tool for scoping arbitrary JSON records (not just logs).
Configuration
Via Constructor
const logger = createLogger(
{ packageName: 'MY_PACKAGE', envPrefix: 'MY_PACKAGE', debugNamespace: 'my-pkg' },
{
// Outputs
logToConsole: true, // default: true
logToFile: false, // default: false
logFilePath: '/var/log/app.log', // required if logToFile
enableUnifiedLogger: false, // default: false
unifiedLogger: { /* ... */ },
// Behavior
logLevel: 'info', // verbose|debug|info|warn|error (default: info)
logFormat: 'text', // text|json|yaml (default: text)
defaultSource: 'application', // fallback source tag
// Sanitization (opt-in)
sanitization: {
enabled: false, // default: false
maskWith: '[REDACTED]',
keysDenylist: ['authorization','password','secret','api_key'],
fieldsHashInsteadOfMask: ['userId'],
detectJWTs: true
// + other detectors & guardrails
},
// Trails & tracing (optional; safe no-ops if not used)
trails: {
enableDepthTrail: true, // operation trail
enableThreadTrail: true, // causal/thread trail
injectHeaders: true, // for HTTP/queues adapters
extractHeaders: true
},
tracing: { enableOtelContext: true },
// Shadow Logging (per-run capture; can also be toggled at runtime)
shadow: {
enabled: false, // default: false
format: 'json', // json|yaml (default: json)
directory: './logs/shadow', // default path
ttlMs: 86_400_000, // 1 day
forceVerbose: true, // capture all levels for the runId
respectRoutingBlocks: true, // honor _routing.blockOutputs: ['shadow'|'file']
includeRaw: false, // also write unsanitized (dangerous; tests only)
rollingBuffer: { maxEntries: 0, maxAgeMs: 0 } // optional retro-capture
},
// Scoping (opt-in)
scoping: {
enabled: false, // default: false
errorScoping: {
enabled: true, // default: true if scoping.enabled
levels: ['error'], // default: ['error']
windowMsBefore: 30_000, // default: 30_000
windowMsAfter: 30_000 // default: 30_000
},
runScoping: {
enabled: true, // default: true if scoping.enabled
defaultWindowMsBeforeFirstError: 60_000,
defaultWindowMsAfterLastEntry: 30_000
},
narrativeScoping: {
enabled: false, // default: false
narrativesPath: './metadata/log-scopes.json',
envPrefix: 'NX_SCOPE'
},
buffer: {
maxEntries: 0, // default: 0 => disabled if Shadow is enough
maxAgeMs: 0,
includeLevels: ['verbose','debug','info','warn','error'],
preferShadow: true // default: true
}
},
// Troubleshooting (opt-in, requires nx-troubleshooting)
troubleshooting: {
enabled: false, // default: false
narrativesPath: './metadata/troubleshooting.json',
envPrefix: 'NX_TROUBLE',
loggingConfig: { /* optional */ },
engine: undefined, // optional DI
output: {
formats: ['markdown'], // default: ['markdown']
writeToFileDir: undefined,
attachToShadow: false,
emitAsLogEntry: false,
callback: undefined
}
}
}
);Via Environment Variables
Note: Environment variable names use a dynamic prefix based on packageConfig.envPrefix. Replace {PREFIX} in the examples below with your actual prefix (e.g., MY_APP, API_SERVICE).
Destinations: MongoDB
Logxer can optionally write logs to MongoDB (best-effort; if Mongo output fails, primary console/file/unified outputs still work).
Common “it doesn’t write to Mongo” causes
- Your
.envis not loaded intoprocess.env: Logxer reads Mongo settings fromprocess.envat runtime. If your process never loads.env, thenMONGO_URI(and related keys) will be missing and the Mongo sink stays disabled.- Recommended (with
@x12i/env): run your app through the CLI so.envis loaded before Node starts:
- Recommended (with
npx x12i-env run --env-file .env -- node dist/index.jsOr ensure your app loads environment variables before constructing Logxer (any approach is fine as long as
process.envis populated early).MongoDB min level default is
warn: by default, onlywarnanderrorare written to MongoDB. If you only emitinfo/debug, it will look like “nothing is saved”.- Set
logxer_mongo_db_logs_level=info(ordebug/verbose) to persist lower levels.
- Set
MongoDB “init” record (connectivity probe)
When the MongoDB sink is enabled (a Mongo URI is present in process.env), Logxer attempts to write a one-time init record to the MongoDB logs collection. This is meant to make “is Mongo logging wired correctly?” obvious even if your app hasn’t emitted any warn/error yet.
MongoDB env vars
- URI resolution (in order):
logxer_mongo_uri(preferred)mongo_logs_uri(fallback)mongo_uri(fallback)MONGO_LOGXER_URI(legacy fallback)MONGO_URI(legacy fallback)
- Database name:
logxer_mongo_db(if set, wins; default:logxer)MONGO_LOGXER_DB(legacy fallback)- Logxer does not infer database name from the URI path by default
- MongoDB min level:
logxer_mongo_db_logs_level(default:warn)
Note: Cloud Storage is intentionally not supported in this repo’s log pipeline. MongoDB is the persistence target.
MongoDB log fetching helpers (public API)
If you’re using the MongoDB sink, Logxer also exposes helper functions to query logs back out of MongoDB:
fetchMachineLogs(...): fetch logs for the same client host (defaults to this machine’s hostname; filter isruntimeIdentity.hostonly, not pid)fetchGlobalLogs(...): fetch logs across all machines/processesfetchSessionLogs({ sessionId, ... }): fetch logs for a local session (filters bysessionId+ sessionstart..endwindow + session station identity)getStationRuntimeIdentity(): returns theruntimeIdentityobject used to tag logs from this machine/process
Local sessions (MongoDB) — local-sessions collection
Logxer can persist a simple “local session” record in MongoDB and then query logs back for that session.
startLocalSession({ desc? }):- inserts a record into the
local-sessionscollection - sets
status/start/endautomatically (status: "active",start: now,end: undefined) - returns
{ sessionId }
- inserts a record into the
endLocalSession(sessionId):- updates the session record to
status: "ended"and setsend: now
- updates the session record to
fetchSessionLogs({ sessionId, ... }):- loads the session record and filters logs by:
sessionId(exact match)timestampin ([start, end]) (ifendis not set yet, “now” is used)runtimeIdentity.hostfrom the session’s stored identity (host only)
- loads the session record and filters logs by:
Cleanup helpers:
deleteMachineLogsAll(...): delete all logs for this machine host (defaults to this host;runtimeIdentity.hostonly)deleteMachineLogsOlderThan(timestamp, ...): delete logs older than timestamp for this machine hostdeleteGlobalLogsAll(...): delete all logs (all machines/processes)deleteGlobalLogsOlderThan(timestamp, ...): delete logs older than timestamp globally
Both fetch functions support:
fromTimestamp: ISO start time filter (inclusive)all: true: ignore time filtering and return newest logslimit: maximum records to return (hard-capped)
import {
fetchMachineLogs,
fetchGlobalLogs,
fetchSessionLogs,
deleteMachineLogsAll,
deleteMachineLogsOlderThan,
deleteGlobalLogsAll,
deleteGlobalLogsOlderThan,
getStationRuntimeIdentity,
startLocalSession,
endLocalSession
} from '@x12i/logxer';
const station = getStationRuntimeIdentity();
// 1) Machine logs (default: current host — same-client scope by hostname)
const mine = await fetchMachineLogs({ fromTimestamp: new Date(Date.now() - 60_000).toISOString(), limit: 200 });
// 2) Global logs (all machines), newest first
const all = await fetchGlobalLogs({ all: true, limit: 500 });
// 3) Local session: create → run work → end → fetch logs for that session
const { sessionId } = await startLocalSession({ desc: 'my debug session' });
// ... run work while adding meta.sessionId = sessionId to your log calls ...
await endLocalSession(sessionId);
const sessionLogs = await fetchSessionLogs({ sessionId, limit: 500 });
// 4) Cleanup: delete older-than
await deleteMachineLogsOlderThan(new Date(Date.now() - 24 * 60 * 60 * 1000).toISOString());
await deleteGlobalLogsOlderThan(new Date(Date.now() - 7 * 24 * 60 * 60 * 1000).toISOString());
// 5) Cleanup: delete-all (use with care)
await deleteMachineLogsAll();
// await deleteGlobalLogsAll(); // DANGEROUS: wipes the entire collection# Console & file
{PREFIX}_LOG_TO_CONSOLE=true|false
{PREFIX}_LOG_TO_FILE=true|false
{PREFIX}_LOG_FILE=/path/to/log
# Unified-logger
{PREFIX}_LOG_TO_UNIFIED=true|false
# Level & format (per-package threshold: canonical _LOGS_LEVEL; legacy _LOG_LEVEL if _LOGS_LEVEL unset)
{PREFIX}_LOGS_LEVEL=off|none|silent|error|warn|info|debug|verbose
{PREFIX}_LOG_LEVEL=verbose|debug|info|warn|error
{PREFIX}_LOG_FORMAT=text|json|yaml|table
# Console output options
{PREFIX}_SHOW_FULL_TIMESTAMP=true|false # Show full ISO timestamp (default: false)
{PREFIX}_CONSOLE_PACKAGES_SHOW=package1,package2 # Only show these packages in console (default: show all)
{PREFIX}_CONSOLE_PACKAGES_HIDE=package1,package2 # Hide these packages in console (default: show all)
# Debug namespace → enables verbose+debug for that namespace
DEBUG=my-pkg,other-*
# Sanitization (subset shown)
{PREFIX}_SANITIZE_ENABLED=true|false
{PREFIX}_SANITIZE_KEYS_DENYLIST=authorization,token,secret,api_key,password
# Trails/tracing
{PREFIX}_TRACE_OTEL=true|false
{PREFIX}_TRAILS_DEPTH=true|false
{PREFIX}_TRAILS_THREAD=true|false
{PREFIX}_TRAILS_INJECT=true|false
{PREFIX}_TRAILS_EXTRACT=true|false
# Shadow Logging
{PREFIX}_SHADOW_ENABLED=true|false
{PREFIX}_SHADOW_FORMAT=json|yaml
{PREFIX}_SHADOW_DIR=/var/log/myapp/shadow
{PREFIX}_SHADOW_TTL_MS=86400000
{PREFIX}_SHADOW_FORCE_VERBOSE=true|false
{PREFIX}_SHADOW_RESPECT_ROUTING=true|false
{PREFIX}_SHADOW_INCLUDE_RAW=false
{PREFIX}_SHADOW_BUFFER_ENTRIES=0
{PREFIX}_SHADOW_BUFFER_AGE_MS=0
# Scoping
{PREFIX}_SCOPING_ENABLED=true|false
{PREFIX}_SCOPING_ERROR_ENABLED=true|false
{PREFIX}_SCOPING_ERROR_WINDOW_MS_BEFORE=30000
{PREFIX}_SCOPING_ERROR_WINDOW_MS_AFTER=30000
{PREFIX}_SCOPING_BUFFER_ENTRIES=5000
{PREFIX}_SCOPING_BUFFER_AGE_MS=300000
{PREFIX}_SCOPING_BUFFER_PREFER_SHADOW=true|false
# Troubleshooting
{PREFIX}_TROUBLESHOOTING_ENABLED=true|false
{PREFIX}_TROUBLESHOOTING_NARRATIVES_PATH=./metadata/troubleshooting.json
{PREFIX}_TROUBLESHOOTING_OUTPUT_FORMATS=markdown,json
{PREFIX}_TROUBLESHOOTING_OUTPUT_EMIT_AS_LOG_ENTRY=true|false
# Unified-logger dependencies (non-ERC, manually documented)
# Required when unified-logger papertrail transport is enabled:
PAPERTRAIL_HOST=logs.papertrailapp.com
PAPERTRAIL_PORT=12345
# Required when unified-logger udpRelay transport is enabled:
UDP_RELAY_HOST=127.0.0.1
UDP_RELAY_PORT=514Default min level:
info.DEBUG=: enables bothverboseanddebugfor matching namespaces.ERC 2.0 Note: Generate a complete
.env.examplefile with all variables by runningnpm run generate-erc.
Core Types
LogEntry
Internal normalized log shape:
export interface LogEntry {
timestamp: string; // ISO-8601
level: 'verbose' | 'debug' | 'info' | 'warn' | 'error';
package: string;
message: string;
source?: string; // e.g. 'application', 'auth-service'
data?: Record<string, any>; // user metadata, error, ids, etc.
// Automatic application identification (from consuming project's package.json)
appName?: string; // Auto-detected from package.json "name" field
appVersion?: string; // Auto-detected from package.json "version" field
// Correlation / trails / tracing
runId?: string;
jobId?: string;
correlationId?: string;
sessionId?: string;
operationId?: string;
parentOperationId?: string;
operationName?: string;
threadId?: string;
traceId?: string;
spanId?: string;
// Routing
_routing?: RoutingMeta;
// Optional scope metadata (for future/advanced use)
scope?: ScopedMetadata;
}Runtime Filtering (logger-debug.json)
Place a logger-debug.json file at your project root to filter logs at runtime:
{
"scoping": {
"status": "enabled",
"filterIdentities": ["src/auth.ts:login", "src/payment.ts:processPayment"],
"filteredApplications": ["my-app", "other-app"],
"between": [
{
"action": "include",
"exactMatch": false,
"searchLog": false,
"startIdentities": ["src/api.ts:handleRequest"],
"endIdentities": ["src/api.ts:handleRequestEnd"]
}
]
}
}Behavior:
- When
status: "enabled", logs are filtered before being written to any output (console, file, unified-logger, shadow) - A log is included if its
identitymatches any entry infilterIdentitiesOR itsappNamematches any entry infilteredApplicationsOR it falls within an active "between" range (OR logic) - If all filters are empty or missing, all logs are included (no filtering)
- File is auto-discovered from
process.cwd()(searches up directory tree like package.json) - Configuration is loaded once at startup (not reloaded dynamically)
- If file is not found or invalid, all logs are shown (graceful fallback)
Examples:
// Filter by identity only - only show logs from specific code locations
{
"scoping": {
"status": "enabled",
"filterIdentities": ["src/auth.ts:login", "src/payment.ts:processPayment"]
}
}// Filter by application only - only show logs from specific apps
{
"scoping": {
"status": "enabled",
"filteredApplications": ["my-app"]
}
}// Filter by both (OR logic - matches if identity OR appName matches)
{
"scoping": {
"status": "enabled",
"filterIdentities": ["src/auth.ts:login"],
"filteredApplications": ["my-app"]
}
}// Between rules - stateful range-based filtering
{
"scoping": {
"status": "enabled",
"between": [
{
"action": "include",
"exactMatch": false,
"searchLog": false,
"startIdentities": ["src/api.ts:handleRequest"],
"endIdentities": ["src/api.ts:handleRequestEnd"]
},
{
"action": "exclude",
"exactMatch": true,
"searchLog": false,
"startIdentities": ["src/db.ts:query"],
"endIdentities": ["src/db.ts:queryEnd"]
},
{
"action": "include",
"exactMatch": false,
"searchLog": true,
"startIdentities": [],
"endIdentities": ["src/init.ts:complete"]
},
{
"action": "include",
"exactMatch": true,
"searchLog": true,
"startIdentities": ["Payment started"],
"endIdentities": ["Payment completed"]
}
]
}
}Between Rules:
- Stateful filtering: Tracks active ranges across log calls
action:"include"to show logs within range,"exclude"to hide logs within rangeexactMatch:true: Exact string match (case sensitive)false: Partial substring match (case insensitive, default)
searchLog:true: Search entire log (message + identity + all meta fields stringified)false: Search only identity field (default)
startIdentities: Array of patterns that activate the range. Empty array means range starts from the beginningendIdentities: Array of patterns that deactivate the range. Empty array means range never ends- Multiple ranges: Can overlap and are tracked independently. Uses OR logic (if ANY include rule is active, include; if ANY exclude rule is active, exclude)
- Range behavior:
- When a log matches a start identity, the range becomes active
- When a log matches an end identity, the range becomes inactive
- If a log matches both start and end identities, the range state toggles
- Ranges with empty
startIdentitiesare active from the first log - Ranges with empty
endIdentitiesnever close once activated
Integration with Existing Scoping:
logger-debug.json→ Runtime filtering (reduces noise at source, filters before writing)scopeLogs()→ Post-processing scoping (analyzes already-written logs)- Both can be used together for maximum control
ScopeCriteria
Scoping criteria defines which logs belong to a scope. It supports:
- Correlation keys
- Time windows
- Levels
- Sources
- Text matching on message and data
export interface ScopeCriteria {
// Correlation / keys
runId?: string;
correlationId?: string;
sessionId?: string;
threadId?: string;
traceId?: string;
// Time window bounds
fromTimestamp?: string; // ISO-8601
toTimestamp?: string; // ISO-8601
// Window relative to an anchor (error or first/last entry)
windowMsBefore?: number; // relative to anchor timestamp
windowMsAfter?: number;
// Levels
levelAtLeast?: 'verbose' | 'debug' | 'info' | 'warn' | 'error';
includeLevels?: ('verbose'|'debug'|'info'|'warn'|'error')[];
// Source filters
sources?: string[]; // e.g. ['api-gateway','payments-service']
// TEXT FILTERS
/**
* Match logs whose message OR data (stringified) contains ANY of the given strings (case-insensitive).
* - string: single substring
* - string[]: log must contain at least one of them
*/
textIncludesAny?: string | string[];
/**
* Match logs whose message OR data (stringified) contains ALL of the given substrings (case-insensitive).
*/
textIncludesAll?: string[];
/**
* Optional RegExp filter over the combined text (message + JSON-stringified data).
* If provided as string, it is treated as a new RegExp(text, 'i').
*/
textMatches?: RegExp | string;
// Scope metadata filters (if used)
scopeTags?: string[]; // must include all provided tags
// Custom predicate for in-process advanced filtering
predicate?: (entry: LogEntry) => boolean;
}Text filtering behavior:
- Combine
entry.messageand a JSON string ofentry.datainto one string (e.g."Payment failed {...}"). - Apply:
textIncludesAny– inclusive OR.textIncludesAll– inclusive AND.textMatches– regex test.
- All text filtering is case-insensitive by default.
- Text filters are ANDed with other criteria (correlation, time, etc.).
ScopedLogView (full data)
export interface ScopedLogView {
id: string; // e.g. 'scope:runId:checkout-42'
criteria: ScopeCriteria;
entries: LogEntry[]; // sorted by timestamp ascending
summary: {
firstTimestamp?: string;
lastTimestamp?: string;
totalCount: number;
errorCount: number;
warnCount: number;
infoCount: number;
debugCount: number;
verboseCount: number;
uniqueSources: string[];
};
}scopeRecord – Generic Tool
The scopeRecord helper is generic (not log-specific): given any JSON record, it produces:
- A human-readable text description ("story" of the record).
- A structured description (fields, truncation info).
This function is:
- Used internally to build scope stories from
LogEntrys / aggregated records. - Exported publicly so other code can reuse it for non-log use cases.
import { scopeRecord } from 'logs-gateway';
// Example usage
const record = {
userId: '123',
action: 'payment',
amount: 100.50,
timestamp: '2025-01-01T10:00:00Z'
};
const result = scopeRecord(record, {
label: 'Payment Event',
maxFieldStringLength: 200,
excludeKeys: ['password', 'token']
});
console.log(result.text);
// Output: "Payment Event\n User ID: 123\n Action: payment\n Amount: 100.5\n Timestamp: 2025-01-01T10:00:00Z"
console.log(result.structured);
// Output: { label: 'Payment Event', fields: [...], ... }Types:
export interface AutoScopeRecordOptions {
label?: string;
formatting?: ScopingFormattingOptions;
maxFields?: number;
maxFieldStringLength?: number;
includeKeys?: string[];
excludeKeys?: string[];
skipNullish?: boolean;
header?: string;
footer?: string;
}
export interface ScopeRecordResult {
text: string; // Human-readable text output
structured: StructuredScopedPayload; // Structured representation
}Scope Story Output
To let a scope answer with full data or story format:
export type ScopeOutputMode = 'raw' | 'story' | 'both';
export interface ScopeStoryOptions {
recordOptions?: AutoScopeRecordOptions;
maxEntries?: number;
includeEntryHeader?: boolean;
}
export interface ScopeLogsResult {
view: ScopedLogView; // always present
story?: ScopedLogStory; // present if mode = 'story' or 'both'
}API Reference
createLogger(packageConfig, userConfig?) → LogsGateway
LogsGateway Methods
Standard Logging
verbose(message, data?)debug(message, data?)info(message, data?)warn(message, data?)error(message, data?)isLevelEnabled(level)– threshold check (namespace DEBUG forces verbose+debug)getConfig()– effective resolved config
Scoping
scopeLogs(criteria, options?)– Scope logs by criteria, return full data and/or story
const result = await logger.scopeLogs({
runId: 'checkout-42',
textIncludesAny: 'timeout',
levelAtLeast: 'debug'
}, {
mode: 'both',
storyOptions: {
maxEntries: 50,
includeEntryHeader: true,
recordOptions: {
label: 'Log Entry',
maxFieldStringLength: 200
}
}
});
console.log(result.view.summary);
console.log(result.story?.text);Troubleshooting
troubleshootError(error, context?, options?)– Error-centric troubleshooting
const { scope, reports } = await logger.troubleshootError(
new Error('Missing connections configuration'),
{
config: { /* app config */ },
query: { requestId: req.id },
operation: 'checkout'
},
{
formats: ['markdown'],
generateScopeStory: true,
storyOptions: { /* ... */ }
}
);troubleshootScope(scope, options?)– Scope-centric troubleshooting
const { scope: scopeResult, reports } = await logger.troubleshootScope(
scopeView, // or ScopeCriteria or scope id string
{
formats: ['markdown', 'json'],
generateScopeStory: true
}
);scopeByNarratives(options?)– Narrative-based scoping (optional)
Shadow Controller
logger.shadow.enable(runId, opts?)logger.shadow.disable(runId)logger.shadow.isEnabled(runId)logger.shadow.listActive()logger.shadow.export(runId, outPath?, compress?) → Promise<string>logger.shadow.readIndex(runId) → Promise<ShadowIndex>logger.shadow.cleanupExpired(now?) → Promise<number>
(Shadow writes sidecar files; primary transports unaffected.)
Usage Examples
1) Error → Scoped logs (text filter) → Story + Troubleshooting
import { createLogger, scopeRecord } from 'logs-gateway';
const logger = createLogger(
{ packageName: 'PAYMENTS', envPrefix: 'PAY' },
{
logToConsole: true,
logFormat: 'json',
shadow: {
enabled: true,
format: 'json',
directory: './logs/shadow',
ttlMs: 86400000,
forceVerbose: true
},
scoping: {
enabled: true,
errorScoping: {
enabled: true,
windowMsBefore: 60_000,
windowMsAfter: 30_000
},
buffer: {
maxEntries: 5000,
maxAgeMs: 300_000,
includeLevels: ['verbose','debug','info','warn','error'],
preferShadow: true
}
},
troubleshooting: {
enabled: true,
narrativesPath: './metadata/troubleshooting.json',
output: {
formats: ['markdown'],
emitAsLogEntry: true,
writeToFileDir: './logs/troubleshooting'
}
}
}
);
async function handleCheckout(req: any) {
const runId = `checkout-${Date.now()}`;
logger.info('Checkout started', { runId });
logger.verbose('Preparing payment context', { runId });
try {
// ...
throw new Error('Missing connections configuration');
} catch (error) {
// Direct troubleshooting call
const { scope, reports } = await logger.troubleshootError(error, {
config: { /* app config */ },
query: { requestId: req.id },
operation: 'checkout'
}, {
formats: ['markdown'],
generateScopeStory: true,
storyOptions: {
maxEntries: 50,
includeEntryHeader: true,
recordOptions: {
label: 'Log Entry',
maxFieldStringLength: 200,
excludeKeys: ['password', 'token']
}
}
});
// Scope includes full data + story
console.log(scope?.view.summary);
console.log(scope?.story?.text);
// Reports contain troubleshooting text
return {
ok: false,
troubleshooting: reports.map(r => r.rendered)
};
}
}2) Manual scoping by text ("include any log that mentions X")
// Scope all logs in the last 5 minutes that mention "timeout" anywhere:
const timeoutScope = await logger.scopeLogs({
levelAtLeast: 'debug',
fromTimestamp: new Date(Date.now() - 5 * 60 * 1000).toISOString(),
textIncludesAny: 'timeout' // message or data, case-insensitive
}, {
mode: 'both',
storyOptions: {
includeEntryHeader: true,
recordOptions: { label: 'Timeout Log' }
}
});
console.log(timeoutScope.view.summary);
console.log(timeoutScope.story?.text);3) Web Application
// src/logger.ts
import { createLogger, LoggingConfig, LogsGateway } from 'logs-gateway';
export function createAppLogger(config?: LoggingConfig): LogsGateway {
return createLogger(
{ packageName: 'WEB_APP', envPrefix: 'WEB_APP', debugNamespace: 'web-app' },
config
);
}
// src/index.ts
const logger = createAppLogger({ logFormat: 'json' });
logger.info('Web application initialized', { version: '1.0.0' });
async function handleRequest(request: any) {
logger.debug('Handling request', { requestId: request.id });
// ...
logger.info('Request processed', { responseTime: 42, runId: request.runId });
}4) Shadow Logging (per-run capture)
Capture everything for a specific runId to a side file (forced-verbose), then fetch it — perfect for tests.
const logger = createLogger(
{ packageName: 'API', envPrefix: 'API' },
{ shadow: { enabled: true, format: 'yaml', ttlMs: 86_400_000 } }
);
const runId = `test-${Date.now()}`;
// Turn on capture for this run (can be mid-execution)
logger.shadow.enable(runId);
logger.info('test start', { runId });
logger.verbose('deep details', { runId, step: 1 });
logger.debug('more details', { runId, step: 2 });
// ... run test ...
// Export & stop capturing
await logger.shadow.export(runId, './artifacts'); // returns exported path
logger.shadow.disable(runId);Shadow files are stored per-run (JSONL or YAML multi-doc), rotated by size/age, and removed by TTL (default 1 day). If
forceVerbose=true, all levels for thatrunIdare captured even when global level is higher.
Correlation & Trails
Attach correlation fields freely on each call:
- Job/Correlation:
runId,jobId,correlationId,sessionId - Depth / Operation trail:
operationId,parentOperationId,operationName,operationPath,operationStep - Thread / Causal trail:
threadId,eventId,causationId,sequenceNo,partitionKey,attempt,shardId,workerId - Tracing:
traceId,spanId(if OpenTelemetry context is active)
Every log entry automatically merges the current trail context. The library provides optional adapters for HTTP and queue systems to inject/extract headers across process boundaries.
Routing Metadata & Recursion Safety
interface RoutingMeta {
allowedOutputs?: string[]; // e.g. ['unified-logger','console']
blockOutputs?: string[]; // e.g. ['unified-logger','file','shadow','troubleshooting']
reason?: string;
tags?: string[];
}- Gateway internal logs (
source: 'logs-gateway-internal') never reach unified-logger. _routinglets you allow/block specific outputs per entry.- Shadow Logging honors
_routing.blockOutputsby default (configurable). - Troubleshooting output can be blocked via
_routing.blockOutputs: ['troubleshooting'].
Log Formats
Text
[2025-01-15T10:30:45.123Z] [MY_APP] [INFO] Application initialized {"version":"1.0.0"}JSON
{
"timestamp": "2025-01-15T10:30:45.123Z",
"package": "MY_APP",
"level": "INFO",
"message": "Application initialized",
"source": "application",
"data": {"version": "1.0.0"}
}YAML (console/file; unified stays JSON)
---
timestamp: 2025-01-15T10:30:45.123Z
package: MY_APP
level: INFO
message: Application initialized
source: application
appName: my-awesome-app
appVersion: "2.1.0"
data:
version: "1.0.0"Note: YAML is human-friendly but slower; prefer JSON for ingestion.
PII/Credentials Sanitization (opt-in)
Enable to auto-detect and mask common sensitive data (JWTs, API keys, passwords, emails, credit cards, cloud keys, etc.). Supports key-based rules (denylist/allowlist), hashing specific fields, depth/size/time guardrails, and a truncation flag.
const logger = createLogger(
{ packageName: 'MY_APP', envPrefix: 'MY_APP' },
{
sanitization: {
enabled: true,
maskWith: '[REDACTED]',
keysDenylist: ['authorization','password','secret','api_key'],
fieldsHashInsteadOfMask: ['userId'],
detectJWTs: true
}
}
);Shadow Logging (Per-Run Debug Capture)
Shadow Logging allows you to capture all logs for a specific runId to a separate file in JSON or YAML format, with raw (unsanitized) data, regardless of the global log level. This is ideal for debugging tests, CI runs, or specific production workflows.
Key Features
- Per-runId capture: Enable capture for specific run identifiers
- Verbose by default: Captures all log levels for the target runId
- Raw data: Bypasses sanitization to preserve original values (⚠️ use with caution)
- Optional rolling buffer: Capture logs from before shadow was enabled ("after-the-fact")
- TTL cleanup: Automatically manage storage with time-to-live expiration
- No interference: Shadow writes are async and won't affect primary logging
Configuration
Constructor Options
interface ShadowConfig {
enabled?: boolean; // default: false
format?: 'json' | 'yaml'; // default: 'json'
directory?: string; // default: './logs/shadow'
ttlMs?: number; // default: 86400000 (1 day)
respectRoutingBlocks?: boolean; // default: true
rollingBuffer?: {
maxEntries?: number; // default: 0 (disabled)
maxAgeMs?: number; // default: 0 (disabled)
};
}Runtime API
logger.shadow.enable(runId, opts?)
Enable shadow capture for a specific runId. Optionally override config for this run.
logger.shadow.enable('test-run-123');
// Or with custom options
logger.shadow.enable('test-run-yaml', {
format: 'yaml',
ttlMs: 3600000 // 1 hour
});logger.shadow.disable(runId)
Stop capturing logs for a runId and finalize the shadow file.
logger.shadow.isEnabled(runId)
Check if shadow capture is active for a runId.
logger.shadow.listActive()
List all currently active shadow captures.
logger.shadow.export(runId, outPath?)
Copy the shadow file to a destination path.
logger.shadow.cleanupExpired(now?)
Delete expired shadow files based on TTL. Returns number of deleted runs.
Important Considerations
⚠️ Security: Shadow captures raw, unsanitized data. This means passwords, API keys, and PII will be preserved in shadow files. Only use shadow logging in secure environments (development, CI, isolated test systems).
💾 Storage: Shadow files can grow quickly when capturing verbose logs. Configure appropriate ttlMs values and regularly run cleanupExpired().
🚫 Default OFF: Shadow logging is disabled by default and must be explicitly enabled via config or environment variables.
Troubleshooting Integration
The troubleshooting integration uses nx-troubleshooting to match errors to solutions. See the nx-troubleshooting documentation for details on creating troubleshooting narratives.
Key Features
- Intelligent Error Matching – Matches errors to solutions using multiple strategies
- Flexible Probe System – Built-in probes plus extensible custom probes
- Template Variables – Dynamic solution messages with
{{variable}}syntax - Multiple Output Formats – Markdown, JSON, and plain text formatting
- Automatic Scoping – Errors automatically trigger scoped log collection
Example Troubleshooting Narrative
{
"id": "missing-connections-config",
"title": "Missing Connections Configuration",
"description": "The application config is missing the required 'connections' object...",
"symptoms": [
{
"probe": "config-check",
"params": { "field": "connections" },
"condition": "result.exists == false"
}
],
"solution": [
{
"type": "code",
"message": "Add a 'connections' object to your config:",
"code": "{\n \"connections\": { ... }\n}"
}
]
}Backwards Compatibility
- All new behavior is opt-in:
scoping.enabledandtroubleshooting.enableddefault tofalse.
- If
nx-troubleshootingis not installed:troubleshooting.enabledmust remainfalseor initialization fails clearly.
- If neither shadow nor buffer is configured:
scopeLogscan still filter currently available logs if in-memory buffer is enabled; otherwise, scopes may be empty.
scopeRecordis pure and can be used independently anywhere.
Environment Variables (summary)
| Key | Description | Default |
| ---------------------- | -------------------------------------- | --------- |
| {P}_LOG_TO_CONSOLE | Enable console output | true |
| {P}_LOG_TO_FILE | Enable file output | false |
| {P}_LOG_FILE | Log file path | — |
| {P}_LOG_TO_UNIFIED | Enable unified-logger | false |
| {P}_LOGS_LEVEL | Per-package threshold (canonical); omit both this and {P}_LOG_LEVEL → warn | warn |
| {P}_LOG_LEVEL | Legacy level — used only if {P}_LOGS_LEVEL is not set in the environment | — |
| {P}_LOG_FORMAT | text\|json\|yaml\|table | table |
| {P}_SHOW_FULL_TIMESTAMP | Show full ISO timestamp in console | false |
| {P}_CONSOLE_PACKAGES_SHOW | Comma-separated packages to show (console only) | (show all) |
| {P}_CONSOLE_PACKAGES_HIDE | Comma-separated packages to hide (console only) | (show all) |
| DEBUG | Namespace(s) enabling verbose+debug | — |
| {P}_SANITIZE_ENABLED | Turn on sanitization | false |
| {P}_TRACE_OTEL | Attach traceId/spanId if available | true |
| {P}_TRAILS_* | Toggle trails/header adapters | see above |
| {P}_SHADOW_* | Shadow Logging controls | see above |
| {P}_SCOPING_* | Scoping controls | see above |
| {P}_TROUBLESHOOTING_* | Troubleshooting controls | see above |
| MONGO_LOGXER_URI | Mongo URI for Logxer Mongo sink (preferred; if missing uses MONGO_URI) | — |
| MONGO_URI | Mongo URI fallback (used only if MONGO_LOGXER_URI is missing) | — |
| MONGO_LOGXER_DB | Mongo DB name for log storage (if missing: uses DB in URI; if neither: defaults to logxer) | logxer |
Development
npm install
npm run build
npm run dev
npm run cleanLicense
MIT
Tip for tests: set a runId at the beginning of each test (e.g., runId: 'ci-<suite>-<timestamp>'), enable Shadow Logging, run, then export() the captured logs for artifacts/triage.
