npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2026 – Pkg Stats / Ryan Hefner

@respan/tracing

v1.0.45

Published

TypeScript support for Respan SDK

Downloads

39

Readme

KeywordsAI Tracing SDK

A lightweight OpenTelemetry-based tracing SDK for KeywordsAI, built with minimal dependencies and optional instrumentation support. Inspired by Openllmetry

Features

  • Lightweight Core: Minimal dependencies for browser and Node.js compatibility
  • Optional Instrumentations: Install only the instrumentations you need
  • OpenTelemetry Native: Built directly on OpenTelemetry without wrapper dependencies
  • Decorator Pattern: Easy-to-use decorators for workflows, tasks, agents, and tools
  • Dynamic Loading: Instrumentations are loaded on-demand
  • Manual Instrumentation: Support for manual instrumentation (Next.js compatible)
  • Span Management: Full control over spans with getClient() API
  • Multi-Processor Routing: Route spans to multiple destinations
  • Span Buffering: Manual control over span export timing
  • KeywordsAI Parameters: Add customer identifiers and trace group identifiers

Installation

Core Package

npm install @keywordsai/tracing

Optional Instrumentations

Install only the instrumentations you need:

# OpenAI
npm install @traceloop/instrumentation-openai

# Anthropic
npm install @traceloop/instrumentation-anthropic

# Azure OpenAI
npm install @traceloop/instrumentation-azure

# AWS Bedrock
npm install @traceloop/instrumentation-bedrock

# Cohere
npm install @traceloop/instrumentation-cohere

# LangChain
npm install @traceloop/instrumentation-langchain

# LlamaIndex
npm install @traceloop/instrumentation-llamaindex

# Vector Databases
npm install @traceloop/instrumentation-pinecone
npm install @traceloop/instrumentation-chromadb
npm install @traceloop/instrumentation-qdrant

# Other providers
npm install @traceloop/instrumentation-together
npm install @traceloop/instrumentation-vertexai

Quick Start

Method 1: Dynamic Instrumentation (Recommended for Node.js)

import { KeywordsAITelemetry } from '@keywordsai/tracing';
import OpenAI from 'openai';

// Initialize the SDK
const keywordsAi = new KeywordsAITelemetry({
    apiKey: process.env.KEYWORDSAI_API_KEY,
    baseURL: process.env.KEYWORDSAI_BASE_URL,
    appName: 'my-app'
});

// Enable instrumentations you need
await keywordsAi.enableInstrumentation('openai');

const openai = new OpenAI();

// Use decorators to trace your functions
const generateJoke = async () => {
    return await keywordsAi.withTask(
        { name: 'joke_generation' },
        async () => {
            const completion = await openai.chat.completions.create({
                messages: [{ role: 'user', content: 'Tell me a joke' }],
                model: 'gpt-3.5-turbo'
            });
            return completion.choices[0].message.content;
        }
    );
};

Method 2: Manual Instrumentation (Recommended for Next.js)

import { KeywordsAITelemetry } from '@keywordsai/tracing';
import OpenAI from 'openai';
import Anthropic from '@anthropic-ai/sdk';

// Manual instrumentation - pass the actual imported modules
const keywordsAi = new KeywordsAITelemetry({
    apiKey: process.env.KEYWORDSAI_API_KEY,
    baseURL: process.env.KEYWORDSAI_BASE_URL,
    appName: 'my-app',
    // Specify modules to instrument manually
    instrumentModules: {
        openAI: OpenAI,
        anthropic: Anthropic,
        // Add other modules as needed
    }
});

// Wait for initialization (optional but recommended)
await keywordsAi.initialize();

// Create clients - they will be automatically instrumented
const openai = new OpenAI();
const anthropic = new Anthropic();

// Use decorators to trace your functions
const generateContent = async () => {
    return await keywordsAi.withWorkflow(
        { name: 'content_generation', version: 1 },
        async () => {
            const result = await openai.chat.completions.create({
                messages: [{ role: 'user', content: 'Generate content' }],
                model: 'gpt-3.5-turbo'
            });
            return result.choices[0].message.content;
        }
    );
};

When to Use Each Method

Dynamic Instrumentation

  • Best for: Standard Node.js applications, serverless functions
  • Pros: Simple setup, automatic loading
  • Cons: May not work in all bundling environments

Manual Instrumentation

  • Best for: Next.js, Webpack bundled apps, environments with import restrictions
  • Pros: Works in all environments, explicit control, better for tree-shaking
  • Cons: Requires importing modules explicitly

API Reference

KeywordsAITelemetry

Constructor Options

interface KeywordsAIOptions {
    appName?: string;                    // App name for traces
    apiKey?: string;                     // KeywordsAI API key
    baseURL?: string;                    // KeywordsAI base URL
    disableBatch?: boolean;              // Disable batching for development
    logLevel?: "debug" | "info" | "warn" | "error";
    traceContent?: boolean;              // Log prompts and completions
    tracingEnabled?: boolean;            // Enable/disable tracing
    silenceInitializationMessage?: boolean;
    
    // Advanced options
    resourceAttributes?: Record<string, string>;  // Custom resource attributes
    spanPostprocessCallback?: (span: any) => void;  // Span postprocessing callback
    
    // Manual instrumentation modules
    instrumentModules?: {
        openAI?: typeof OpenAI;
        anthropic?: typeof Anthropic;
        azureOpenAI?: typeof AzureOpenAI;
        cohere?: typeof Cohere;
        bedrock?: typeof BedrockRuntime;
        google_vertexai?: typeof VertexAI;
        google_aiplatform?: typeof AIPlatform;
        pinecone?: typeof Pinecone;
        together?: typeof Together;
        langchain?: {
            chainsModule?: typeof ChainsModule;
            agentsModule?: typeof AgentsModule;
            toolsModule?: typeof ToolsModule;
            runnablesModule?: typeof RunnableModule;
            vectorStoreModule?: typeof VectorStoreModule;
        };
        llamaIndex?: typeof LlamaIndex;
        chromadb?: typeof ChromaDB;
        qdrant?: typeof Qdrant;
    };
}

Methods

  • initialize() - Manually initialize tracing (returns Promise)
  • isInitialized() - Check if tracing has been initialized
  • enableInstrumentation(name: string) - Enable a specific instrumentation (dynamic method)
  • enableInstrumentations(names: string[]) - Enable multiple instrumentations (dynamic method)
  • addProcessor(config: ProcessorConfig) - Add a processor for routing spans
  • getClient() - Get the client API for span management
  • getSpanBufferManager() - Get the span buffer manager
  • shutdown() - Flush and shutdown tracing

Decorators

withWorkflow

Trace high-level workflows:

await keywordsAi.withWorkflow(
    { name: 'my_workflow', version: 1 },
    async () => {
        // Your workflow logic
    }
);

withTask

Trace individual tasks:

await keywordsAi.withTask(
    { name: 'my_task' },
    async () => {
        // Your task logic
    }
);

withAgent

Trace agent operations:

await keywordsAi.withAgent(
    { name: 'my_agent', associationProperties: { type: 'assistant' } },
    async () => {
        // Your agent logic
    }
);

withTool

Trace tool usage:

await keywordsAi.withTool(
    { name: 'my_tool' },
    async () => {
        // Your tool logic
    }
);

Decorator Configuration

interface DecoratorConfig {
    name: string;                        // Required: Name of the operation
    version?: number;                    // Optional: Version number
    associationProperties?: Record<string, string>; // Optional: Additional metadata
    traceContent?: boolean;              // Optional: Override trace content setting
    inputParameters?: unknown[];         // Optional: Custom input parameters
    suppressTracing?: boolean;           // Optional: Suppress tracing for this operation
    processors?: string | string[];      // Optional: Route to specific processor(s)
}

Advanced Features

Span Management with getClient()

Get full control over your spans with the client API:

import { KeywordsAITelemetry, getClient } from '@keywordsai/tracing';

const kai = new KeywordsAITelemetry({ apiKey: 'your-key' });
await kai.initialize();

await kai.withTask({ name: 'process_data' }, async () => {
    const client = getClient();
    
    // Get current trace and span IDs
    const traceId = client.getCurrentTraceId();
    const spanId = client.getCurrentSpanId();
    console.log(`Trace: ${traceId}, Span: ${spanId}`);
    
    // Update span with KeywordsAI parameters
    client.updateCurrentSpan({
        keywordsaiParams: {
            customerIdentifier: 'user-123',
            traceGroupIdentifier: 'data-pipeline',
            metadata: {
                version: '1.0',
                environment: 'production'
            }
        }
    });
    
    // Add events to track progress
    client.addEvent('validation_started', { records: 1000 });
    
    // Your processing logic here
    
    client.addEvent('validation_completed', { status: 'success' });
    
    // Record exceptions
    try {
        // risky operation
    } catch (error) {
        client.recordException(error as Error);
        throw error;
    }
});

Available Client Methods:

  • getCurrentTraceId() - Get the current trace ID
  • getCurrentSpanId() - Get the current span ID
  • updateCurrentSpan(options) - Update span attributes, name, status, or KeywordsAI params
  • addEvent(name, attributes?) - Add an event to the current span
  • recordException(exception) - Record an exception on the current span
  • isRecording() - Check if the span is recording
  • getTracer() - Get the tracer for manual span creation
  • flush() - Force flush all pending spans

Multi-Processor Routing

Route spans to different destinations based on processor names:

import { KeywordsAITelemetry } from '@keywordsai/tracing';

const kai = new KeywordsAITelemetry({ apiKey: 'your-key' });

// Add a debug processor (in addition to default KeywordsAI processor)
kai.addProcessor({
    exporter: new YourCustomExporter(),
    name: 'debug',
    filter: (span) => span.attributes['environment'] === 'development'
});

// Route specific spans to debug processor
await kai.withTask(
    { name: 'debug_task', processors: 'debug' },
    async () => {
        // This span goes to the debug processor
    }
);

// Route to multiple processors
await kai.withTask(
    { name: 'important_task', processors: ['debug', 'analytics'] },
    async () => {
        // This span goes to both processors
    }
);

// Default behavior - no processors attribute
await kai.withTask(
    { name: 'normal_task' },
    async () => {
        // This span goes to the default KeywordsAI processor
    }
);

Processor Configuration:

interface ProcessorConfig {
    exporter: SpanExporter;              // The span exporter to use
    name: string;                        // Processor identifier for routing
    filter?: (span: ReadableSpan) => boolean;  // Optional custom filter
    priority?: number;                   // Optional priority (higher = processed first)
}

Span Buffering for Manual Control

Buffer spans and control when they're exported:

import { KeywordsAITelemetry } from '@keywordsai/tracing';

const kai = new KeywordsAITelemetry({ apiKey: 'your-key' });
const manager = kai.getSpanBufferManager();

// Create a buffer (spans won't be auto-exported)
const buffer = manager.createBuffer('workflow-123');

// Add spans to the buffer
buffer.createSpan('validation', {
    status: 'success',
    duration_ms: 10
});

buffer.createSpan('processing', {
    status: 'success',
    duration_ms: 100
});

// Get buffered spans (they're transportable!)
const spans = buffer.getAllSpans();
console.log(`Collected ${spans.length} spans`);

// Conditionally process based on business logic
const isSuccessful = true;  // Your business logic
const isPremiumUser = true; // Your business logic

if (isSuccessful && isPremiumUser) {
    // Export to KeywordsAI
    await manager.processSpans(spans);
} else {
    // Discard spans
    buffer.clearSpans();
}

Use Cases for Span Buffering:

  • Backend systems that need delayed span export
  • Conditional export based on business logic
  • Batch processing of spans
  • Async span creation (create spans after execution)
  • Experiment tracking with selective export

SpanBuffer Methods:

  • createSpan(name, attributes?, kind?) - Create a span in the buffer
  • getAllSpans() - Get all buffered spans as a transportable array
  • getSpanCount() - Get the number of buffered spans
  • clearSpans() - Discard all buffered spans without exporting

KeywordsAI-Specific Parameters

Add customer and trace group identifiers to your spans:

import { getClient } from '@keywordsai/tracing';

await kai.withWorkflow({ name: 'user_workflow' }, async () => {
    const client = getClient();
    
    client.updateCurrentSpan({
        keywordsaiParams: {
            // Group traces by customer
            customerIdentifier: 'user-123',
            
            // Organize traces into groups
            traceGroupIdentifier: 'onboarding-flow',
            
            // Add custom metadata
            metadata: {
                plan: 'premium',
                region: 'us-east-1',
                version: '2.1.0'
            }
        }
    });
});

These parameters help you:

  • Group traces by customer for user-level analytics
  • Organize traces into logical groups (experiments, features, etc.)
  • Add custom metadata for filtering and analysis

Available Instrumentations

The following instrumentations can be enabled dynamically:

  • openai - OpenAI API calls
  • anthropic - Anthropic API calls
  • azure - Azure OpenAI API calls
  • bedrock - AWS Bedrock API calls
  • cohere - Cohere API calls
  • langchain - LangChain operations
  • llamaindex - LlamaIndex operations
  • pinecone - Pinecone vector database
  • chromadb - ChromaDB vector database
  • qdrant - Qdrant vector database
  • together - Together AI API calls
  • vertexai - Google Vertex AI API calls

Troubleshooting Instrumentation

Error: "instrumentation failed to initialize"

This usually means the instrumentation package is missing. Install it:

# For Anthropic
npm install @traceloop/instrumentation-anthropic

# For OpenAI
npm install @traceloop/instrumentation-openai

Error: "Cannot read properties of undefined (reading 'prototype')"

This means you're using manual instrumentation but passed the wrong module. Make sure to:

// ✅ CORRECT - Pass the class/module itself
instrumentModules: {
    anthropic: Anthropic  // The imported class
}

// ❌ WRONG - Don't pass an instance
instrumentModules: {
    anthropic: new Anthropic()  // This won't work
}

Dynamic instrumentation not working in Next.js/Webpack?

Use manual instrumentation instead:

// Instead of this:
await kai.enableInstrumentation('anthropic');

// Use this:
const kai = new KeywordsAITelemetry({
    instrumentModules: {
        anthropic: Anthropic
    }
});

Anthropic spans not appearing?

Known Issue: @traceloop/[email protected] doesn't work with @anthropic-ai/[email protected]+

Tested Working Versions:

  • @anthropic-ai/sdk@^0.20.0 to @^0.41.0 - Full tracing support with all metrics
  • @anthropic-ai/[email protected]+ - No spans created (breaking changes)

Official Support: The instrumentation package officially supports SDK 0.9.1 and later, but breaking changes in SDK v0.50.0 prevent spans from being created. Last confirmed working version: 0.41.0.

Solutions:

  1. Use a compatible version (recommended - use latest working version):

    npm install @anthropic-ai/sdk@^0.41.0
    # or
    yarn add @anthropic-ai/sdk@^0.41.0

    Or use any version between 0.20.0 and 0.41.0

  2. Use OpenAI instead - Fully supported with latest SDK versions

What Works with 0.20.x:

  • ✅ All token metrics (gen_ai.usage.prompt_tokens, gen_ai.usage.completion_tokens)
  • ✅ Full request/response content tracing
  • ✅ Model information and metadata
  • ✅ Proper span hierarchy within workflows

Status: SDK v0.50.0 introduced breaking changes that broke instrumentation compatibility. Waiting for @traceloop to update their package.

Spans not showing up?

  1. Check that you're using decorators (withTask, withWorkflow, etc.)
  2. Verify API key is set: process.env.KEYWORDSAI_API_KEY
  3. Enable debug logging: logLevel: 'debug'
  4. Check network requests to KeywordsAI endpoint

Environment Variables

  • KEYWORDSAI_API_KEY: Your KeywordsAI API key
  • KEYWORDSAI_BASE_URL: KeywordsAI base URL (default: https://api.keywordsai.co)
  • KEYWORDSAI_APP_NAME: Default app name
  • KEYWORDSAI_TRACE_CONTENT: Enable/disable content tracing (default: true)

Provider-Specific Examples

OpenAI

Method 1: Dynamic Instrumentation (Simple)

import { KeywordsAITelemetry } from '@keywordsai/tracing';
import OpenAI from 'openai';

const kai = new KeywordsAITelemetry({
    apiKey: process.env.KEYWORDSAI_API_KEY,
    appName: 'openai-app'
});

// Enable OpenAI instrumentation dynamically
await kai.enableInstrumentation('openai');

const openai = new OpenAI();

await kai.withTask({ name: 'chat' }, async () => {
    const response = await openai.chat.completions.create({
        model: 'gpt-4',
        messages: [{ role: 'user', content: 'Hello!' }]
    });
    console.log(response.choices[0].message.content);
});

Method 2: Manual Instrumentation (Next.js/Webpack)

import { KeywordsAITelemetry } from '@keywordsai/tracing';
import OpenAI from 'openai';

const kai = new KeywordsAITelemetry({
    apiKey: process.env.KEYWORDSAI_API_KEY,
    appName: 'openai-app',
    instrumentModules: {
        openAI: OpenAI  // Pass the OpenAI class
    }
});

await kai.initialize();

const openai = new OpenAI();

await kai.withTask({ name: 'chat' }, async () => {
    const response = await openai.chat.completions.create({
        model: 'gpt-4',
        messages: [{ role: 'user', content: 'Hello!' }]
    });
    console.log(response.choices[0].message.content);
});

Anthropic (Claude)

import { KeywordsAITelemetry } from '@keywordsai/tracing';
import Anthropic from '@anthropic-ai/sdk';

// Initialize with Anthropic instrumentation
const kai = new KeywordsAITelemetry({
    apiKey: process.env.KEYWORDSAI_API_KEY,
    appName: 'anthropic-app',
    instrumentModules: {
        anthropic: Anthropic  // Pass the Anthropic class
    }
});

await kai.initialize();

// Create Anthropic client (will be auto-instrumented)
const anthropic = new Anthropic({
    apiKey: process.env.ANTHROPIC_API_KEY
});

// Use with decorators
await kai.withTask({ name: 'generate_text' }, async () => {
    const response = await anthropic.messages.create({
        model: 'claude-3-haiku-20240307',
        max_tokens: 1024,
        messages: [{ role: 'user', content: 'Hello, Claude!' }]
    });
    
    console.log(response.content[0].text);
});

// Streaming example
await kai.withTask({ name: 'stream_text' }, async () => {
    const stream = await anthropic.messages.create({
        model: 'claude-3-haiku-20240307',
        max_tokens: 1024,
        messages: [{ role: 'user', content: 'Count to 10' }],
        stream: true
    });
    
    for await (const event of stream) {
        if (event.type === 'content_block_delta' && 
            event.delta.type === 'text_delta') {
            process.stdout.write(event.delta.text);
        }
    }
});

Note: Make sure to install the Anthropic instrumentation:

npm install @anthropic-ai/sdk @traceloop/instrumentation-anthropic

Complete Examples

Example 1: Full Workflow with Span Management

import { KeywordsAITelemetry, getClient } from '@keywordsai/tracing';
import OpenAI from 'openai';

const kai = new KeywordsAITelemetry({
    apiKey: process.env.KEYWORDSAI_API_KEY,
    appName: 'my-app',
    resourceAttributes: {
        environment: 'production',
        version: '1.0.0'
    }
});

await kai.initialize();
const openai = new OpenAI();

await kai.withWorkflow({ name: 'process_user_request', version: 1 }, async () => {
    const client = getClient();
    
    // Set customer context
    client.updateCurrentSpan({
        keywordsaiParams: {
            customerIdentifier: 'user-123',
            traceGroupIdentifier: 'onboarding'
        }
    });
    
    // Track progress with events
    client.addEvent('validation_started');
    
    // Nested task
    await kai.withTask({ name: 'validate_input' }, async () => {
        // Validation logic
    });
    
    client.addEvent('ai_processing_started');
    
    // LLM call (auto-instrumented)
    const response = await openai.chat.completions.create({
        model: 'gpt-3.5-turbo',
        messages: [{ role: 'user', content: 'Process this' }]
    });
    
    client.addEvent('ai_processing_completed', {
        tokens: response.usage?.total_tokens
    });
    
    return response.choices[0].message.content;
});

Example 2: Backend Workflow with Span Buffering

import { KeywordsAITelemetry } from '@keywordsai/tracing';

const kai = new KeywordsAITelemetry({ apiKey: 'your-key' });
const manager = kai.getSpanBufferManager();

// Ingest workflow results from backend
async function ingestWorkflow(workflowResult: any, orgId: string) {
    const buffer = manager.createBuffer(`workflow-${workflowResult.id}`);
    
    // Create spans from workflow results
    buffer.createSpan('workflow_execution', {
        organization_id: orgId,
        input: workflowResult.input,
        output: workflowResult.output,
        duration_ms: workflowResult.duration
    });
    
    for (const step of workflowResult.steps) {
        buffer.createSpan(`step_${step.name}`, {
            input: step.input,
            output: step.output,
            duration_ms: step.duration
        });
    }
    
    // Get transportable spans
    const spans = buffer.getAllSpans();
    
    // Conditionally export based on business logic
    const isPremium = orgId.includes('premium');
    
    if (isPremium) {
        await manager.processSpans(spans);
        console.log('Exported spans for premium org');
    } else {
        buffer.clearSpans();
        console.log('Skipped spans for free org');
    }
}

Example 3: Multi-Destination Routing

import { KeywordsAITelemetry } from '@keywordsai/tracing';
import { FileExporter, AnalyticsExporter } from './exporters';

const kai = new KeywordsAITelemetry({ apiKey: 'your-key' });

// Add debug file exporter
kai.addProcessor({
    exporter: new FileExporter('./debug-spans.jsonl'),
    name: 'debug'
});

// Add analytics exporter with filter
kai.addProcessor({
    exporter: new AnalyticsExporter(),
    name: 'analytics',
    filter: (span) => !span.name.includes('test')
});

// Route to default KeywordsAI processor
await kai.withTask(
    { name: 'production_task' },
    async () => { /* goes to KeywordsAI */ }
);

// Route to debug processor
await kai.withTask(
    { name: 'debug_task', processors: 'debug' },
    async () => { /* goes to file */ }
);

// Route to multiple processors
await kai.withTask(
    { name: 'important_task', processors: ['debug', 'analytics'] },
    async () => { /* goes to file + analytics */ }
);

Browser Compatibility

The core package is designed to work in both Node.js and browser environments. However, some instrumentations may be Node.js only.

Testing Builds

Before publishing, test the built package:

npm run test:build

This builds, packs, installs, and tests the package exactly as users will receive it.

Examples Directory

Check out the examples/ directory for more comprehensive examples:

  • span-management-example.ts - Full span management with getClient()
  • multi-processor-example.ts - Multi-processor routing examples
  • span-buffer-example.ts - Span buffering patterns
  • basic-usage.ts - Basic usage patterns
  • advanced-tracing-example.ts - Advanced tracing patterns

Migration from v1.0.x

All new features are backward compatible. Existing code will continue to work without modifications:

  • Default processor is automatically configured
  • New methods are additive (optional)
  • No breaking changes to existing APIs

To use new features, simply import and use them:

import { getClient } from '@keywordsai/tracing';  // New in v1.1.0

License

Apache-2.0