npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2026 – Pkg Stats / Ryan Hefner

@traceai/llamaindex

v0.1.1

Published

OpenTelemetry instrumentation for LlamaIndex

Readme

@traceai/llamaindex

OpenTelemetry instrumentation for LlamaIndex. This package provides automatic tracing and monitoring for your LlamaIndex applications.

Installation

npm install @traceai/llamaindex
# or
yarn add @traceai/llamaindex
# or
pnpm add @traceai/llamaindex

Quick Start

// Set up environment variables
process.env["FI_API_KEY"] = "your-api-key";
process.env["FI_SECRET_KEY"] = "your-secret-key";
process.env["OPENAI_API_KEY"] = "your-openai-api-key";

import { register, ProjectType } from "@traceai/fi-core";
import { diag, DiagConsoleLogger, DiagLogLevel } from "@opentelemetry/api";
import { OpenAI, OpenAIEmbedding } from "@llamaindex/openai";
import * as LlamaIndex from "llamaindex";
import { LlamaIndexInstrumentation } from "@traceai/llamaindex";

// Enable OpenTelemetry internal diagnostics (optional, for debugging)
diag.setLogger(new DiagConsoleLogger(), DiagLogLevel.INFO);

async function main() {
  // 1. Configure LlamaIndex Settings
  const { Settings } = LlamaIndex;
  Settings.llm = new OpenAI({ model: "gpt-3.5-turbo" });
  Settings.embedModel = new OpenAIEmbedding({ model: "text-embedding-ada-002" });

  // 2. Register FI Core TracerProvider
  const tracerProvider = register({
    projectName: "your-project-name",
    projectType: ProjectType.OBSERVE,
    setGlobalTracerProvider: true,
  });

  // 3. Initialize and register LlamaIndex Instrumentation
  const instrumentation = new LlamaIndexInstrumentation({});
  instrumentation.manuallyInstrument(LlamaIndex);

  // 4. Use LlamaIndex as normal
  const { Document, VectorStoreIndex } = LlamaIndex;
  
  // Create documents
  const documents = [
    new Document({ text: "LlamaIndex is a data framework for LLM applications." }),
    new Document({ text: "OpenTelemetry provides observability for applications." }),
  ];

  // Build vector store index
  const index = await VectorStoreIndex.fromDocuments(documents);

  // Query the index
  const queryEngine = index.asQueryEngine();
  const response = await queryEngine.query({
    query: "What is LlamaIndex?"
  });

  console.log("Response:", response.toString());

  // 5. Don't forget to shutdown the tracer provider when done
  try {
    await tracerProvider.shutdown();
    console.log("Tracer provider shut down successfully.");
  } catch (error) {
    console.error("Error shutting down tracer provider:", error);
  }
}

main().catch(console.error);

Environment Variables

The following environment variables are required for telemetry:

FI_API_KEY=your_api_key
FI_SECRET_KEY=your_secret_key
OPENAI_API_KEY=your_openai_api_key

Features

  • Automatic tracing of LlamaIndex operations including:
    • Vector store operations
    • Query engine execution
    • Retrieval operations
    • Embedding generation
    • LLM chat completions
  • Support for both ESM and CommonJS modules
  • Compatible with LlamaIndex.js
  • Integration with TraceAI's observability platform

Instrumented Operations

This instrumentation automatically traces:

  • Vector Store Operations: Document indexing, vector storage
  • Query Engine: Query execution and response generation
  • Retrieval: Document retrieval and similarity search
  • Embeddings: Text embedding generation
  • LLM Operations: Chat completions and text generation

Peer Dependencies

This package requires the following peer dependencies:

  • llamaindex: >=0.1.0

Development

# Install dependencies
pnpm install

# Build the package
pnpm build

# Run tests
pnpm test

# Type checking
pnpm lint

Support

For support, please open an issue in our GitHub repository.

Check your TraceAI dashboard at https://app.futureagi.com/dashboard/observe for traces!