@revenium/middleware
v1.1.2
Published
Unified Revenium middleware for AI provider usage tracking - OpenAI, Anthropic, Google, Perplexity, LiteLLM, fal.ai
Readme
Revenium Middleware for Node.js
Unified TypeScript middleware for automatic AI usage tracking across multiple providers
A professional-grade Node.js middleware that integrates with OpenAI, Azure OpenAI, Anthropic, Google (GenAI + Vertex AI), Perplexity, LiteLLM, and fal.ai to provide automatic usage tracking, billing analytics, and metadata collection. Features Go-aligned API patterns, sub-path imports for tree-shaking, and ESM + CJS dual output.
Features
- Multi-Provider Support - OpenAI, Azure OpenAI, Anthropic, Google GenAI, Google Vertex AI, Perplexity, LiteLLM, fal.ai
- Go-Aligned API - Consistent
Initialize()/GetClient()pattern across providers - Sub-Path Imports - Tree-shakeable
@revenium/middleware/openai,/anthropic, etc. - Tool Metering - Track custom tool and external API calls with
meterTool()andreportToolCall() - Fire-and-Forget - Metering never blocks your application flow
- Streaming Support - Handles regular and streaming requests for all providers
- ESM + CJS - Dual output with full TypeScript type definitions
- Automatic .env Loading - Loads environment variables automatically
Supported Providers
| Provider | Sub-Path Import | API Pattern |
| ---------------- | ------------------------------------ | --------------------------------------------------------- |
| OpenAI | @revenium/middleware/openai | Initialize() / GetClient() |
| Azure OpenAI | @revenium/middleware/openai | Initialize() / GetClient() (auto-detected) |
| Anthropic | @revenium/middleware/anthropic | initialize() / configure() / auto-init on import |
| Google GenAI | @revenium/middleware/google/genai | GoogleGenAIController / GoogleGenAIService |
| Google Vertex AI | @revenium/middleware/google/vertex | VertexAIController / VertexAIService |
| Perplexity | @revenium/middleware/perplexity | Initialize() / GetClient() |
| LiteLLM | @revenium/middleware/litellm | initialize() / configure() / enable() / disable() |
| fal.ai | @revenium/middleware/fal | Initialize() / GetClient() |
| Tool Metering | @revenium/middleware/tools | meterTool() / reportToolCall() |
Getting Started
Installation
npm install @revenium/middlewareInstall the provider SDK you need as a peer dependency:
npm install openai # For OpenAI / Azure OpenAI / Perplexity
npm install @anthropic-ai/sdk # For Anthropic
npm install @google/genai # For Google GenAI
npm install google-auth-library # For Google Vertex AI
npm install @fal-ai/client # For fal.aiConfiguration
Create a .env file in your project root. See .env.example for all available options.
Minimum required:
REVENIUM_METERING_API_KEY=hak_your_revenium_api_key_here
REVENIUM_METERING_BASE_URL=https://api.revenium.aiPlus the API key for your chosen provider (OPENAI_API_KEY, ANTHROPIC_API_KEY, etc.).
Quick Start - OpenAI
import { Initialize, GetClient } from "@revenium/middleware/openai";
Initialize();
const client = GetClient();
const response = await client.chat.completions.create({
model: "gpt-4o-mini",
messages: [{ role: "user", content: "Hello!" }],
});Quick Start - Anthropic
import "@revenium/middleware/anthropic";
import Anthropic from "@anthropic-ai/sdk";
const client = new Anthropic();
const response = await client.messages.create({
model: "claude-sonnet-4-20250514",
max_tokens: 1024,
messages: [{ role: "user", content: "Hello!" }],
});Quick Start - Google GenAI
import { GoogleGenAIController } from "@revenium/middleware/google/genai";
const controller = new GoogleGenAIController({
reveniumApiKey: process.env.REVENIUM_METERING_API_KEY!,
});
const response = await controller.generateContent({
model: "gemini-2.0-flash",
contents: "Hello!",
});Quick Start - Azure OpenAI
import { Initialize, GetClient } from "@revenium/middleware/openai";
Initialize();
const client = GetClient();
const response = await client.chat.completions.create({
model: "my-deployment-name",
messages: [{ role: "user", content: "Hello!" }],
});Azure is auto-detected when AZURE_OPENAI_API_KEY and AZURE_OPENAI_ENDPOINT are set.
Quick Start - Google Vertex AI
import { VertexAIController } from "@revenium/middleware/google/vertex";
const controller = new VertexAIController({
reveniumApiKey: process.env.REVENIUM_METERING_API_KEY!,
});
const response = await controller.generateContent({
model: "gemini-2.0-flash",
contents: "Hello!",
});Quick Start - Perplexity
import { Initialize, GetClient } from "@revenium/middleware/perplexity";
Initialize();
const client = GetClient();
const response = await client.chat.completions.create({
model: "sonar",
messages: [{ role: "user", content: "Hello!" }],
});Quick Start - fal.ai
Ensure FAL_KEY and REVENIUM_METERING_API_KEY are set in your environment before initializing.
import { Initialize, GetClient } from "@revenium/middleware/fal";
Initialize();
const fal = GetClient();
// Image generation (with cost attribution metadata)
const image = await fal.subscribe(
"fal-ai/flux/schnell",
{
input: { prompt: "a futuristic cityscape at sunset" },
},
{ subscriber: { id: "user_123" }, traceId: "req_abc789" },
);
console.log(image.data.images[0].url);
// Video generation
const video = await fal.subscribe("fal-ai/kling-video/v2/master/text-to-video", {
input: { prompt: "ocean waves crashing on rocks", duration: 5 },
});
console.log(video.data.video.url);
// Audio generation (text-to-speech)
const audio = await fal.subscribe("fal-ai/kokoro/american-english", {
input: { prompt: "Hello from Revenium!", voice: "af_heart" },
});
console.log(audio.data.audio.url);
// LLM via OpenRouter
const chat = await fal.subscribe("openrouter/router", {
input: { prompt: "Explain quantum computing", model: "google/gemini-2.5-flash" },
});
console.log(chat.data.output);The middleware automatically detects the media type from the endpoint ID and routes metering data to the correct Revenium endpoint. The optional metadata parameter enables cost attribution per subscriber, organization, or trace.
Quick Start - LiteLLM
import { initialize } from "@revenium/middleware/litellm";
initialize();API Reference
OpenAI
Go-aligned client pattern with Azure auto-detection:
| Function | Description |
| --------------------- | --------------------------------------------------------- |
| Initialize(config?) | Initialize middleware from environment or explicit config |
| GetClient() | Get the wrapped OpenAI client instance |
| Configure(config) | Alias for Initialize() for programmatic configuration |
| IsInitialized() | Check if middleware is initialized |
| Reset() | Reset the global client (useful for testing) |
Anthropic
Auto-initializes on import. Manual control available:
| Function | Description |
| -------------------- | --------------------------------------------------- |
| initialize() | Explicitly initialize middleware |
| configure(config) | Set configuration and patch Anthropic |
| patchAnthropic() | Enable request interception |
| unpatchAnthropic() | Disable request interception |
| isInitialized() | Check initialization status |
| getStatus() | Get detailed status including circuit breaker state |
| reset() | Reset middleware and circuit breaker |
Google GenAI / Vertex AI
Controller and service pattern:
| Export | Description |
| ---------------------------------------------- | ------------------------------------- |
| GoogleGenAIController / VertexAIController | Main controller for API calls |
| GoogleGenAIService / VertexAIService | Service implementation |
| trackGoogleUsageAsync() | Manual usage tracking |
| mapGoogleFinishReason() | Map finish reasons to standard format |
Perplexity
Same Go-aligned client pattern as OpenAI:
| Function | Description |
| --------------------- | --------------------------------------------------------- |
| Initialize(config?) | Initialize middleware from environment or explicit config |
| GetClient() | Get the wrapped Perplexity client instance |
| Configure(config) | Alias for Initialize() for programmatic configuration |
| IsInitialized() | Check if middleware is initialized |
| Reset() | Reset the global client (useful for testing) |
fal.ai
Enterprise wrapper for fal.ai's multi-modal platform (images, video, audio, LLM) with automatic metering:
| Function | Description |
| --------------------- | --------------------------------------------------------- |
| Initialize(config?) | Initialize middleware from environment or explicit config |
| GetClient() | Get the wrapped fal.ai client instance |
| Configure(config) | Alias for Initialize() for programmatic configuration |
| IsInitialized() | Check if middleware is initialized |
| Reset() | Reset the global client (useful for testing) |
Client Methods:
| Method | Description |
| ----------------------------------------------- | -------------------------------------------------------------------- |
| fal.subscribe(endpointId, options, metadata?) | Submit to queue and wait for result (recommended for most use cases) |
| fal.run(endpointId, options, metadata?) | Execute directly and wait for result (low-latency models) |
| fal.stream(endpointId, options, metadata?) | Stream partial results (real-time LLM or progress tracking) |
| fal.queue | Access the underlying queue client directly |
| fal.realtime | Access the underlying realtime client directly |
| fal.storage | Access the underlying storage client directly |
| fal.getUnderlyingClient() | Get the raw FalClient instance (not metered) |
The metadata parameter is optional on all methods and enables cost attribution (e.g., { subscriber: { id: '...' }, organizationName, traceId }). It does not affect the fal.ai payload. See Metadata Fields for all supported options.
Media Type Routing:
| Media Type | Metering Endpoint | Detection Examples | Billing Metric |
| ---------- | ----------------- | ------------------------------------- | ----------------------------------- |
| IMAGE | /ai/images | flux, stable-diffusion, recraft, sdxl | Per image (+ resolution) |
| VIDEO | /ai/video | kling-video, veo, sora, runway, luma | Seconds of video |
| AUDIO | /ai/audio | kokoro, chatterbox, whisper, f5-tts | Characters (TTS) / minutes (transcription) / seconds (generation) |
| CHAT | /ai/completions | openrouter | Token usage (input/output/total) |
Media type is detected via a two-phase approach: first by regex matching on the endpoint ID, then corrected by inspecting the response structure (e.g., presence of images, video, audio_url, or usage fields).
Fallback: Unknown endpoints default to IMAGE metering. A warning is logged automatically for unrecognized endpoints.
LiteLLM
HTTP client patching for LiteLLM proxy:
| Function | Description |
| --------------------------- | ------------------------------------- |
| initialize() | Initialize from environment variables |
| configure(config) | Set configuration explicitly |
| enable() | Enable HTTP client patching |
| disable() | Disable HTTP client patching |
| isMiddlewareInitialized() | Check initialization status |
| getStatus() | Get status including proxy URL |
| reset() | Reset all state |
Tool Metering
Track custom tool and external API calls. Available from any provider sub-path or directly via @revenium/middleware/tools.
import { meterTool, setToolContext } from "@revenium/middleware/tools";
setToolContext({
agent: "my-agent",
traceId: "session-123",
});
const result = await meterTool(
"weather-api",
async () => {
return await fetch("https://api.example.com/weather");
},
{
operation: "get_forecast",
outputFields: ["temperature", "humidity"],
},
);Functions
| Function | Description |
| ---------------------------------- | ------------------------------------------------------------------------- |
| meterTool(toolId, fn, metadata?) | Wrap a function with automatic metering (timing, success/failure, errors) |
| reportToolCall(toolId, report) | Manually report an already-executed tool call |
| setToolContext(ctx) | Set context for all subsequent tool calls |
| getToolContext() | Get current context |
| clearToolContext() | Clear context |
| runWithToolContext(ctx, fn) | Run function with scoped context (uses AsyncLocalStorage) |
Tool Metadata Options
| Field | Description |
| ---------------------- | ----------------------------------------------------- |
| operation | Tool operation name (e.g., "search", "scrape") |
| outputFields | Array of field names to auto-extract from result |
| usageMetadata | Custom metrics (e.g., tokens, results count) |
| agent | Agent identifier (inherited from context) |
| traceId | Trace identifier (inherited from context) |
| organizationName | Organization name (inherited from context) |
| productName | Product name (inherited from context) |
| subscriberCredential | Subscriber credential string (inherited from context) |
| workflowId | Workflow identifier (inherited from context) |
| transactionId | Transaction identifier (inherited from context) |
Metadata Fields
All fields are optional and can be set per-request via usageMetadata:
| Field | Type | Description |
| ----------------------- | ------ | ------------------------------------------------------ |
| traceId | string | Unique identifier for session or conversation tracking |
| taskType | string | Type of AI task (e.g., "chat", "embedding") |
| agent | string | AI agent or bot identifier |
| organizationName | string | Organization or company name |
| productName | string | Product or feature name |
| subscriptionId | string | Subscription plan identifier |
| responseQualityScore | number | Custom quality rating (0.0-1.0) |
| subscriber.id | string | Unique user identifier |
| subscriber.email | string | User email address |
| subscriber.credential | object | Authentication credential (name and value) |
Trace Visualization Fields
Environment variables for distributed tracing and analytics:
| Environment Variable | Description |
| -------------------------------- | -------------------------------------------------------------------------- |
| REVENIUM_ENVIRONMENT | Deployment environment (production, staging, development) |
| REVENIUM_REGION | Cloud region (auto-detected from AWS/Azure/GCP if not set) |
| REVENIUM_CREDENTIAL_ALIAS | Human-readable credential name |
| REVENIUM_TRACE_TYPE | Categorical identifier (alphanumeric, hyphens, underscores, max 128 chars) |
| REVENIUM_TRACE_NAME | Human-readable label for trace instances (max 256 chars) |
| REVENIUM_PARENT_TRANSACTION_ID | Parent transaction reference for distributed tracing |
| REVENIUM_TRANSACTION_NAME | Human-friendly operation label |
| REVENIUM_RETRY_NUMBER | Retry attempt number (0 for first attempt) |
Configuration Options
Common Environment Variables
| Variable | Required | Description |
| ---------------------------- | -------- | ---------------------------------------------------------- |
| REVENIUM_METERING_API_KEY | Yes | Revenium API key (starts with hak_) |
| REVENIUM_METERING_BASE_URL | No | Revenium API endpoint (default: https://api.revenium.ai) |
| REVENIUM_DEBUG | No | Enable debug logging (true/false) |
| REVENIUM_PRINT_SUMMARY | No | Terminal summary (true, human, json, false) |
| REVENIUM_TEAM_ID | No | Team ID for cost display in terminal summary |
| REVENIUM_CAPTURE_PROMPTS | No | Enable prompt capture (true/false) |
Provider-Specific Variables
| Variable | Provider | Description |
| -------------------------------- | ------------- | ------------------------------------------------- |
| OPENAI_API_KEY | OpenAI | OpenAI API key |
| AZURE_OPENAI_API_KEY | Azure OpenAI | Azure OpenAI API key |
| AZURE_OPENAI_ENDPOINT | Azure OpenAI | Azure resource endpoint URL |
| AZURE_OPENAI_API_VERSION | Azure OpenAI | API version (default: 2024-02-15-preview) |
| ANTHROPIC_API_KEY | Anthropic | Anthropic API key |
| GOOGLE_API_KEY | Google GenAI | Google AI Studio API key |
| GOOGLE_CLOUD_PROJECT | Google Vertex | GCP project ID |
| GOOGLE_APPLICATION_CREDENTIALS | Google Vertex | Path to service account key file |
| GOOGLE_CLOUD_LOCATION | Google Vertex | GCP region (default: us-central1) |
| PERPLEXITY_API_KEY | Perplexity | Perplexity API key |
| LITELLM_PROXY_URL | LiteLLM | LiteLLM proxy URL (e.g., http://localhost:4000) |
| LITELLM_API_KEY | LiteLLM | LiteLLM proxy API key |
| FAL_KEY | fal.ai | fal.ai API key |
See .env.example for the complete list with all optional configuration.
Troubleshooting
No tracking data appears
- Verify environment variables are set correctly in
.env - Enable debug logging:
REVENIUM_DEBUG=true - Check console for
[Revenium]log messages - Verify your
REVENIUM_METERING_API_KEYis valid
Client not initialized error
- Make sure you call
Initialize()beforeGetClient() - Check that your
.envfile is in the project root - Verify
REVENIUM_METERING_API_KEYis set
Azure OpenAI not working
- Verify all Azure environment variables are set (see
.env.example) - Check that
AZURE_OPENAI_ENDPOINTandAZURE_OPENAI_API_KEYare correct - Ensure you're using a valid deployment name in the
modelparameter
Debug Mode
Enable detailed logging:
REVENIUM_DEBUG=trueTesting
npm test # Run all tests
npm run test:core # Run core module tests
npm run test:openai # Run OpenAI tests
npm run test:anthropic # Run Anthropic tests
npm run test:google # Run Google tests
npm run test:perplexity # Run Perplexity tests
npm run test:litellm # Run LiteLLM tests
npm run test:fal # Run fal.ai tests
npm run test:integration # Run integration tests
npm run test:coverage # Run tests with coverageRequirements
- Node.js 18+
- TypeScript 5.0+ (for TypeScript projects)
- At least one provider SDK installed as peer dependency
Contributing
See CONTRIBUTING.md
Code of Conduct
Security
See SECURITY.md
License
This project is licensed under the MIT License - see the LICENSE file for details.
Support
- Website: www.revenium.ai
- Documentation: docs.revenium.io
- Issues: Report bugs or request features
- Email: [email protected]
Built by Revenium
