npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2026 – Pkg Stats / Ryan Hefner

logfx

v1.0.0

Published

Tiny, pretty logging for JS. Colors, emojis, namespaces. Zero deps.

Readme

logfx

Beautiful, colorful console logging with emojis, levels & namespaces

npm version npm downloads build node bundle size license

Features

Core Logging:

  • Colorful output with emoji prefixes
  • Namespaces to organize logs by module
  • Log levelsdebug, info, success, warn, error + custom levels
  • Color themes — dracula, monokai, or custom
  • Timestamps — optional time display
  • Context — attach metadata to all logs
  • Universal — works in Node.js and browsers
  • Tiny — zero dependencies, ~3KB gzipped

Reliability & Performance:

  • High Performance — handles 250k+ logs/sec (comparable to Pino)
  • Retry logic — exponential backoff with jitter
  • Circuit breaker — prevent cascading failures
  • Dead letter queue — persist failed logs
  • Multi-region failover — automatic endpoint switching
  • Lazy evaluation — optimize performance
  • Async logging — buffer and batch logs

Security & Privacy:

  • PII detection — automatic redaction of emails, SSN, credit cards, phone, IP, JWT
  • Custom patterns — define your own sensitive data patterns
  • Field redaction — hide sensitive data by key or path
  • Masking functions — partial redaction for debugging

Observability:

  • Trace context — W3C TraceParent format support
  • OpenTelemetry — seamless integration via logfx-otel
  • Request ID tracking — correlate logs across services
  • Log sampling — reduce volume in production

Transports:

  • Console — pretty or JSON output
  • File — with rotation support
  • Webhook — HTTP endpoint with retry and failover
  • Beacon — browser sendBeacon for reliable page unload

Integrations:

  • Framework middleware — Express, Fastify, Next.js
  • Sentry — error tracking via logfx-sentry
  • Datadog — APM via logfx-datadog
  • Elasticsearch — search and analytics via logfx-elasticsearch
  • Cloud — AWS CloudWatch, GCP, Azure via logfx-cloudwatch, logfx-google-cloud, logfx-azure
  • Slack, Loki, Papertrail, Splunk — logfx-slack, logfx-loki, logfx-papertrail, logfx-splunk
  • Honeycomb, Logtail — logfx-honeycomb, logfx-logtail
  • TypeScript — full type support

Installation

npm install logfx

Quick Start

import { log } from 'logfx'

log.debug('Debugging info', { detailed: true })
log.info('Server started', { port: 3000 })
log.success('User created!')
log.warn('Memory usage high', { usage: '85%' })
log.error('Connection failed', new Error('Timeout'))

Output:

🔍 DEBUG   Debugging info { detailed: true }
💡 INFO    Server started { port: 3000 }
✅ SUCCESS User created!
⚠️ WARN    Memory usage high { usage: '85%' }
🔴 ERROR   Connection failed Error: Timeout
    at ...

Errors include full stack traces when available.

Namespaced Loggers

import { logger } from 'logfx'

const authLog = logger('auth')
authLog.info('User login attempt')   // 💡 INFO [auth] User login attempt
authLog.success('Login successful')  // ✅ SUCCESS [auth] Login successful

Output:

💡 INFO  [auth] User login attempt
✅ SUCCESS [auth] Login successful
💡 INFO  [database] Connecting...
✅ SUCCESS [database] Connected

Express Middleware

Automatic request logging for Express apps:

import express from 'express'
import { expressLogger } from 'logfx/middleware'

const app = express()

app.use(expressLogger())

app.get('/users', (req, res) => {
  req.log.info('Fetching users')
  res.json({ users: [] })
})

Output:

💡 INFO [http] Incoming request { method: 'GET', path: '/users', requestId: 'abc123' }
💡 INFO [http] Fetching users { requestId: 'abc123' }
💡 INFO [http] Request completed { method: 'GET', path: '/users', status: 200, durationMs: 45 }

Middleware Options

app.use(expressLogger({
  namespace: 'api',               // Custom namespace
  skip: (req) => req.path === '/health',  // Skip health checks
  includeHeader: true,           // Add X-Request-Id header (default: true)
  headerName: 'X-Trace-Id',      // Custom header name
  getId: (req) => req.headers['x-custom-id']  // Custom ID extractor
}))

Each request gets:

  • req.log - scoped logger with requestId
  • req.requestId - unique request identifier
  • Automatic timing and status code logging

Fastify Plugin

import Fastify from 'fastify'
import { fastifyLogger } from 'logfx/middleware'

const app = Fastify()
app.register(fastifyLogger())

app.get('/users', async (request, reply) => {
  request.log.info('Fetching users')
  return { users: [] }
})

Color Themes

Choose from built-in themes or use the default:

const log = createLogger({ theme: 'dracula' })
// or
const log = createLogger({ theme: 'monokai' })

Available themes: default, dracula, monokai

Issue Detection

Catch common logging mistakes in development:

const log = createLogger({ detectIssues: true })

log.info('User data', { user: undefined })
// ⚠️ Warning: undefined value detected in log data

log.info('Login', { password: 'secret123' })
// ⚠️ Warning: potential password in log data

Helps catch bugs before they reach production. Disabled by default.

Configuration

import { createLogger } from 'logfx'

const log = createLogger({
  namespace: 'api',
  level: 'warn',      // only show warn and error
  timestamp: true,
  enabled: true,
})

| Option | Type | Default | Description | |--------|------|---------|-------------| | namespace | string | - | Prefix for logs | | level | LogLevel | 'debug' | Minimum level to display | | timestamp | boolean | false | Show timestamps | | enabled | boolean | true | Enable/disable logging | | format | 'pretty' \| 'json' | auto | Output format (auto-detects based on NODE_ENV) | | theme | 'default' \| 'dracula' \| 'monokai' | 'default' | Color theme | | detectIssues | boolean | false | Warn about undefined values and passwords | | transports | Transport[] | - | Custom transports | | context | object | - | Metadata added to all logs | | redact | RedactOptions | - | Field redaction config | | sampling | SamplingOptions | - | Log sampling rates | | async | boolean | false | Enable async buffered logging | | buffer | BufferOptions | - | Buffer size and flush interval |

Auto-Detection

logfx automatically picks the right format for your environment:

  • Development (NODE_ENV !== 'production') → pretty output with colors
  • Production (NODE_ENV === 'production') → JSON format

Override with explicit format option:

const log = createLogger({ format: 'json' })  // Force JSON
const log = createLogger({ format: 'pretty' })  // Force pretty

Transports

Send logs to multiple destinations:

import { createLogger, transports } from 'logfx'

const log = createLogger({
  transports: [
    transports.console({ format: 'pretty' }),
    transports.file({ path: './logs/app.log' }),
    transports.webhook({ url: 'https://your-api.com/logs' }),
  ]
})

JSON Output for Production

Structured JSON output for log aggregation services:

const log = createLogger({
  transports: [
    transports.console({ format: 'json' })
  ]
})

log.info('User login', { userId: 123 })
// {"timestamp":"2025-12-17T...","level":"info","message":"User login","userId":123}

File transport format option:

transports.file({ 
  path: './logs/app.log',
  format: 'json'  // or 'pretty' for plain text (default: 'json')
})

Available Transports

| Transport | Description | |-----------|-------------| | console | Pretty or JSON output to stdout | | file | Write to file (Node.js only) | | webhook | POST logs to HTTP endpoint with retry and failover | | beacon | Browser-only, reliable delivery on page unload |

Webhook Transport Options

transports.webhook({
  url: 'https://your-api.com/logs',
  method: 'POST',  // default, or 'PUT'
  headers: { 'Authorization': 'Bearer token' },
  batchSize: 10,  // default: 10 logs per batch
  flushInterval: 5000,  // default: 5 seconds
  maxBufferSize: 100,  // default: 10x batchSize, drops oldest when full
  timeout: 30000,  // default: 30 seconds
  
  // Retry configuration
  retry: {
    maxRetries: 3,  // default: 3
    initialDelay: 1000,  // default: 1s
    maxDelay: 30000,  // default: 30s
    backoff: 'exponential',  // default: 'exponential', or 'linear', 'fixed'
    retryOn: [500, 502, 503, 504, 'ECONNRESET', 'ETIMEDOUT']  // default
  }
})

Batching: Logs are sent automatically when the batch is full or on the flush interval. Oldest logs are dropped if the buffer exceeds maxBufferSize.

Retry Logic: Failed requests are automatically retried with exponential backoff. Retries occur on network errors and 5xx status codes by default.

Beacon Transport (Browser Only)

Reliable log delivery on page unload using the Beacon API:

transports.beacon({
  url: '/api/logs',
  maxPayloadSize: 64000,  // 64KB limit for sendBeacon
  events: {
    beforeunload: true,      // Flush on page close
    visibilitychange: true,  // Flush when tab hidden
    pagehide: true          // Flush on page hide
  }
})

Use Cases:

  • Single Page Applications (SPAs)
  • Analytics and user behavior tracking
  • Error reporting on page close
  • Session end logging

Benefits:

  • Guaranteed delivery even when page closes
  • Non-blocking (doesn't delay page unload)
  • Automatic fallback to fetch if sendBeacon unavailable

Ecosystem Integrations

logfx provides officially supported packages for popular observability platforms.

| Package | Description | |---------|-------------| | logfx-otel | OpenTelemetry span events | | logfx-sentry | Sentry error tracking | | logfx-datadog | Datadog APM | | logfx-elasticsearch | Elasticsearch bulk API | | logfx-cloudwatch | AWS CloudWatch Logs | | logfx-google-cloud | Google Cloud Logging | | logfx-azure | Azure Monitor Log Analytics | | logfx-slack | Slack webhooks | | logfx-loki | Grafana Loki | | logfx-papertrail | Papertrail | | logfx-splunk | Splunk HEC | | logfx-honeycomb | Honeycomb Events API | | logfx-logtail | Logtail (Better Stack) |

Sentry (logfx-sentry)

Automatically capture errors and attach standard logs as breadcrumbs/context in Sentry.

npm install logfx-sentry @sentry/node
import { createLogger } from 'logfx'
import { sentryTransport } from 'logfx-sentry'
import * as Sentry from '@sentry/node'

Sentry.init({ dsn: "YOUR_DSN" })

const log = createLogger({
  transports: [
    sentryTransport({ 
      minLevel: 'warn',        // Only send warnings and errors
      captureContext: true     // Attach log metadata to Sentry
    })
  ]
})

Elasticsearch (logfx-elasticsearch)

Ship logs directly to your ELK stack using the highly-optimized Bulk API. Inherits logfx's core circuit breaker, retry, and batching logic.

npm install logfx-elasticsearch
import { createLogger } from 'logfx'
import { elasticsearchTransport } from 'logfx-elasticsearch'

const log = createLogger({
  transports: [
    elasticsearchTransport({
      node: 'https://es-cluster.example.com:9200',
      index: 'app-logs',
      auth: { apiKey: 'your-api-key' },
      batchSize: 100,          // Send logs in chunks of 100
      flushInterval: 5000,     // Or every 5 seconds
      circuitBreaker: { enabled: true } // Don't hang if ES goes down
    })
  ]
})

Datadog (logfx-datadog)

Send logs to Datadog's HTTP intake.

Cloud (logfx-cloudwatch, logfx-google-cloud, logfx-azure)

Ship logs to AWS CloudWatch, Google Cloud Logging, or Azure Monitor Log Analytics.

npm install logfx-cloudwatch      # AWS
npm install logfx-google-cloud    # GCP
npm install logfx-azure            # Azure
import { cloudwatchTransport } from 'logfx-cloudwatch'
import { googleCloudTransport } from 'logfx-google-cloud'
import { azureTransport } from 'logfx-azure'

Slack, Loki, Papertrail, Splunk

Ship logs to Slack webhooks, Grafana Loki, Papertrail, or Splunk HEC.

npm install logfx-slack
npm install logfx-loki
npm install logfx-papertrail
npm install logfx-splunk

Honeycomb, Logtail

Ship logs to Honeycomb Events API or Logtail (Better Stack).

npm install logfx-honeycomb
npm install logfx-logtail

OpenTelemetry (logfx-otel)

Automatically inject Trace and Span IDs into your logs.

Context

Attach metadata to all logs from a logger:

const log = createLogger({
  context: {
    service: 'api-gateway',
    version: '1.2.0',
    env: process.env.NODE_ENV
  },
  transports: [transports.console({ format: 'json' })]
})

log.info('Request received', { path: '/users' })
// {"service":"api-gateway","version":"1.2.0","env":"production","path":"/users",...}

Child loggers inherit and can extend context:

const requestLog = log.child('request', { 
  context: { requestId: 'req-123' } 
})
requestLog.info('Processing')
// Includes service, version, env, AND requestId

Field Redaction

Automatically hide sensitive data:

import { createLogger, maskEmail, maskCreditCard } from 'logfx'

const log = createLogger({
  redact: {
    keys: ['password', 'token', 'apiKey'],
    paths: ['user.email', 'config.secret'],
    censor: '[HIDDEN]',  // default: '[REDACTED]'
    
    // Automatic PII detection
    patterns: ['email', 'ssn', 'creditCard', 'phone', 'ip', 'jwt'],
    
    // Custom patterns
    customPatterns: [
      { name: 'apiKey', regex: /sk_(live|test)_[a-zA-Z0-9]+/g }
    ],
    
    // Custom redaction logic
    custom: (key, value) => {
      if (key === 'email') return maskEmail(String(value))
      if (key === 'cardNumber') return maskCreditCard(String(value))
      return value
    }
  },
  transports: [transports.console({ format: 'json' })]
})

log.info('User login', { username: 'john', password: 'secret123' })
// {"username":"john","password":"[HIDDEN]",...}

log.info('Contact [email protected] with card 4532-1234-5678-9010')
// Automatically redacts email and credit card

Log Sampling

Reduce log volume by sampling:

const log = createLogger({
  sampling: {
    debug: 0.1,   // 10% of debug logs
    info: 0.5,    // 50% of info logs
    warn: 1.0,    // 100% of warnings
    error: 1.0    // 100% of errors (never sample errors)
  },
  transports: [transports.console()]
})

Async Logging

Buffer logs and flush in batches for better performance:

const log = createLogger({
  async: true,
  buffer: {
    size: 100,          // flush after 100 logs
    flushInterval: 5000 // or every 5 seconds
  },
  transports: [transports.file({ path: './app.log' })]
})

// Graceful shutdown
process.on('SIGTERM', async () => {
  await log.flush()
  await log.close()
  process.exit(0)
})

Extended Features

Import only what you need:

import { box, table, diff, time, timeEnd, badge } from 'logfx'

| Import | Size | What it does | |--------|------|--------------| | Core (log) | ~2 KB | Basic logging | | time/timeEnd | +80 bytes | Performance timing | | box | +350 bytes | ASCII boxes for banners | | table | +300 bytes | Pretty-print data tables | | diff | +450 bytes | Compare objects | | Everything | ~3.4 KB | All features |

Timers

time('api-call')
await fetchData()
timeEnd('api-call')  // ⏱️ api-call: 245.32ms

Boxes

box('Server Started!', { title: '🚀 My App', borderColor: 'green' })
╭─ 🚀 My App ─────────────────╮
│  Server Started!            │
╰─────────────────────────────╯

Tables

import { table } from 'logfx'

const users = [
  { name: 'John', role: 'Admin', active: true },
  { name: 'Jane', role: 'User', active: false },
]

table(users)

Output:

┌─────────┬─────────┬─────────┐
│ name    │ role    │ active  │
├─────────┼─────────┼─────────┤
│ John    │ Admin   │ true    │
│ Jane    │ User    │ false   │
└─────────┴─────────┴─────────┘

Diff

diff({ name: 'John', age: 25 }, { name: 'Jane', age: 25, email: '[email protected]' })
Changes:
  ~ name: "John" → "Jane"
  + email: "[email protected]"

All-in-One

import { createExtendedLogger } from 'logfx'

const log = createExtendedLogger()
log.box('Ready!')
log.table(data)
log.diff(before, after)

API

// Core
log.debug(...args)
log.info(...args)
log.success(...args)
log.warn(...args)
log.error(...args)
log.child(namespace, options?)
log.setEnabled(bool)
log.setLevel(level)
log.flush()   // flush buffered logs
log.close()   // flush and close transports

// Extended
time(label) / timeEnd(label)
count(label) / countReset(label)
group(label) / groupEnd()
assert(condition, ...args)
box(message, options?)
table(data)
diff(before, after, label?)
badge(text, color?)

📚 Articles & Blog Posts

Learn more about logfx:

License

MIT