npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2026 – Pkg Stats / Ryan Hefner

eth-parallel-event-fetcher

v1.6.0

Published

Parallel event fetching library for Ethereum contracts

Readme

Ethereum Parallel Event Fetcher

CI npm version License: MIT

A TypeScript library for parallel fetching of Ethereum logs with configurable concurrency, rate limiting, and block range chunking.

Features

  • Parallel Processing: Fetch logs across multiple block ranges simultaneously
  • Automatic Retry Logic: Built-in retry mechanism with exponential backoff
  • Progress Tracking: Real-time progress reporting and callbacks
  • Configurable: Extensive configuration options for performance tuning
  • TypeScript Support: Full type safety and intellisense
  • Environment Variables: Configuration via environment variables

Installation

npm install eth-parallel-event-fetcher

Quick Start

import { GenericEventFetcher } from 'eth-parallel-event-fetcher';
import { ethers } from 'ethers';

// Create provider
const provider = new ethers.providers.JsonRpcProvider('your-rpc-url');

// Create fetcher with provider (required)
const fetcher = new GenericEventFetcher(provider);

// Fetch logs
const logs = await fetcher.getLogs({
  address: '0x...', // Contract address (single address only)
  topics: ['0x...'], // Event topic filters
  fromBlock: 1000000,
  toBlock: 1010000,
});

// Process logs as needed
logs.forEach((log) => {
  console.log(`Block ${log.blockNumber}: ${log.transactionHash}`);
});

Configuration

Constructor Configuration

import { GenericEventFetcher, FetcherConfig } from 'eth-parallel-event-fetcher';

const config: FetcherConfig = {
  // Concurrency and performance
  concurrency: 10, // Number of parallel requests (1-50)
  chunkSize: 5000, // Blocks per chunk (100-100,000)

  // Retry configuration
  maxRetries: 3, // Number of retry attempts (0-10)
  initialRetryDelay: 1000, // Initial delay in ms (100-30,000)

  // Rate limiting
  rateLimitPerSecond: 10, // Requests per second (1-1,000)

  // Progress tracking
  showProgress: true, // Show console progress
  progressCallback: (completed, total, chunk) => {
    console.log(`Progress: ${completed}/${total} chunks`);
  },

  // Error handling
  continueOnError: true, // Continue on individual chunk errors
  maxLogsPerChunk: 9500, // Max logs per chunk to avoid truncation
};

// Provider is required as first parameter
const fetcher = new GenericEventFetcher(provider, config);

Environment Variables

Configure the library using environment variables:

# Concurrency and performance
ETH_FETCHER_CONCURRENCY=10
ETH_FETCHER_CHUNK_SIZE=5000

# Retry configuration
ETH_FETCHER_MAX_RETRIES=3
ETH_FETCHER_INITIAL_RETRY_DELAY=1000

# Rate limiting
ETH_FETCHER_RATE_LIMIT_PER_SECOND=10

# Progress tracking
ETH_FETCHER_SHOW_PROGRESS=true

# Error handling
ETH_FETCHER_CONTINUE_ON_ERROR=true
ETH_FETCHER_MAX_LOGS_PER_CHUNK=9500

Default Configuration

const DEFAULT_CONFIG = {
  concurrency: 4,
  chunkSize: 10000,
  maxRetries: 3,
  initialRetryDelay: 1000,
  rateLimitPerSecond: 10,
  showProgress: false,
  continueOnError: true,
  maxLogsPerChunk: 10000,
};

API Reference

GenericEventFetcher

Constructor

new GenericEventFetcher(provider: ethers.providers.Provider, config?: Partial<FetcherConfig>)

Parameters:

  • provider (required): An ethers.js provider instance
  • config (optional): Configuration options

Methods

getLogs()
async getLogs(
  filter: LogFilter,
  options?: GetLogsOptions
): Promise<ethers.providers.Log[]>

Parameters:

  • filter: Ethereum log filter

    • address?: string - Contract address (single address only)
    • topics?: Array<string | Array<string> | null> - Topic filters
    • fromBlock?: number | string - Starting block
    • toBlock?: number | string - Ending block
    • blockHash?: string - Specific block hash
  • options: Additional fetch options (overrides constructor config)

    • chunkSize?: number
    • concurrency?: number
    • maxRetries?: number
    • initialRetryDelay?: number
    • rateLimitPerSecond?: number
    • showProgress?: boolean
    • onProgress?: (completed: number, total: number, chunk: [number, number]) => void
    • continueOnError?: boolean
    • maxLogsPerChunk?: number

Returns: Array of ethers.providers.Log objects

updateConfig()
updateConfig(newConfig: Partial<FetcherConfig>): void

Update the fetcher configuration at runtime.

getConfig()
getConfig(): Required<FetcherConfig>

Get the current configuration.

Types

LogFilter

interface LogFilter {
  address?: string; // Contract address (single address only)
  topics?: Array<string | Array<string> | null>; // Topic filters
  fromBlock?: number | string; // Starting block
  toBlock?: number | string; // Ending block
  blockHash?: string; // Specific block hash
}

GetLogsOptions

interface GetLogsOptions {
  chunkSize?: number;
  concurrency?: number;
  maxRetries?: number;
  initialRetryDelay?: number;
  rateLimitPerSecond?: number;
  showProgress?: boolean;
  onProgress?: (completed: number, total: number, chunk: [number, number]) => void;
  continueOnError?: boolean;
  maxLogsPerChunk?: number;
}

Usage Examples

Basic Log Fetching

import { GenericEventFetcher } from 'eth-parallel-event-fetcher';
import { ethers } from 'ethers';

// Setup
const provider = new ethers.providers.JsonRpcProvider(RPC_URL);
const fetcher = new GenericEventFetcher(provider, { concurrency: 8 });

// Fetch Transfer event logs
const transferTopic = '0xddf252ad1be2c89b69c2b068fc378daa952ba7f163c4a11628f55a4df523b3ef';
const logs = await fetcher.getLogs({
  address: '0xa0b86991c6218b36c1d19d4a2e9eb0ce3606eb48', // USDC contract
  topics: [transferTopic],
  fromBlock: 18000000,
  toBlock: 18100000,
});

console.log(`Found ${logs.length} Transfer logs`);

Advanced Log Processing

// Fetch with progress tracking
const logs = await fetcher.getLogs(
  {
    address: CONTRACT_ADDRESS,
    topics: [EVENT_TOPIC],
    fromBlock: 18000000,
    toBlock: 18100000,
  },
  {
    onProgress: (completed, total) => {
      console.log(`Progress: ${completed}/${total} chunks`);
    },
  }
);

// Process logs with ethers
const iface = new ethers.utils.Interface(ABI);
const transfers = logs.map((log) => {
  const parsed = iface.parseLog(log);
  return {
    from: parsed.args.from,
    to: parsed.args.to,
    value: ethers.utils.formatEther(parsed.args.value),
    blockNumber: log.blockNumber,
    transactionHash: log.transactionHash,
  };
});

Using Block Tags

// Use block tags instead of numbers
const logs = await fetcher.getLogs({
  address: CONTRACT_ADDRESS,
  fromBlock: 'earliest',
  toBlock: 'latest',
});

Error Handling

try {
  const logs = await fetcher.getLogs(filter, options);
} catch (error) {
  if (error.message.includes('truncation detected')) {
    // Too many logs in a chunk
    console.log('Chunk too large, reducing chunk size...');

    // Retry with smaller chunks
    const updatedFetcher = new GenericEventFetcher(provider, {
      ...fetcher.getConfig(),
      chunkSize: 1000,
    });
    const logs = await updatedFetcher.getLogs(filter, options);
  } else {
    console.error('Fetch failed:', error);
  }
}

Dynamic Configuration Updates

const fetcher = new GenericEventFetcher(provider, { concurrency: 4 });

// Update configuration based on network conditions
if (networkIsSlow) {
  fetcher.updateConfig({
    concurrency: 2,
    chunkSize: 1000,
    maxRetries: 5,
  });
}

Performance Tips

1. Optimize Chunk Size

Smaller chunks (1,000-5,000 blocks):

  • Better for contracts with many events
  • Reduces risk of hitting log limits
  • More granular progress tracking

Larger chunks (10,000-50,000 blocks):

  • Better for sparse events
  • Fewer total requests
  • Faster for simple queries

2. Adjust Concurrency

// For rate-limited endpoints
const fetcher = new GenericEventFetcher(provider, {
  concurrency: 2,
  rateLimitPerSecond: 5,
});

// For high-performance nodes
const fetcher = new GenericEventFetcher(provider, {
  concurrency: 20,
  rateLimitPerSecond: 100,
});

3. Handle Large Date Ranges

// Split very large ranges into multiple calls
const BATCH_SIZE = 100000; // blocks per batch
const allLogs = [];

for (let from = startBlock; from < endBlock; from += BATCH_SIZE) {
  const to = Math.min(from + BATCH_SIZE - 1, endBlock);

  const logs = await fetcher.getLogs({
    address: CONTRACT_ADDRESS,
    fromBlock: from,
    toBlock: to,
  });

  allLogs.push(...logs);
}

4. Monitor Progress

const fetcher = new GenericEventFetcher(provider, {
  showProgress: true,
  progressCallback: (completed, total, chunk) => {
    const percentage = ((completed / total) * 100).toFixed(2);
    const [fromBlock, toBlock] = chunk;
    console.log(`${percentage}% complete. Current chunk: ${fromBlock}-${toBlock}`);
  },
});

Troubleshooting

Common Issues

1. "Log response size exceeded" Error

Problem: Too many logs in a single chunk.

Solution: Reduce chunk size

const fetcher = new GenericEventFetcher(provider, {
  chunkSize: 1000, // Smaller chunks
  maxLogsPerChunk: 5000, // Lower limit
});

2. Rate Limiting Errors

Problem: Too many requests per second.

Solution: Reduce concurrency and add rate limiting

const fetcher = new GenericEventFetcher(provider, {
  concurrency: 2,
  rateLimitPerSecond: 5,
  initialRetryDelay: 2000,
});

3. Memory Issues with Large Result Sets

Problem: Too many logs in memory at once.

Solution: Process in batches

// Process logs in smaller batches
const BATCH_SIZE = 50000;
for (let i = 0; i < totalBlocks; i += BATCH_SIZE) {
  const logs = await fetcher.getLogs({
    fromBlock: startBlock + i,
    toBlock: Math.min(startBlock + i + BATCH_SIZE, endBlock),
  });

  // Process and persist logs immediately
  await processLogs(logs);

  // Clear logs from memory
  logs.length = 0;
}

4. Timeout Errors

Problem: RPC endpoint timing out.

Solution: Reduce chunk size and concurrency

const fetcher = new GenericEventFetcher(provider, {
  chunkSize: 2000,
  concurrency: 3,
  maxRetries: 5,
  initialRetryDelay: 3000,
});

Testing

Run the test suite:

# Install dependencies
npm install

# Run all tests
npm test

# Run with coverage
npm run test:coverage

# Run specific test file
npm test -- fetcher.test.ts

Contributing

Contributions are welcome! Please feel free to submit a Pull Request.

  1. Fork the repository
  2. Create your feature branch (git checkout -b feature/AmazingFeature)
  3. Commit your changes (git commit -m 'Add some AmazingFeature')
  4. Push to the branch (git push origin feature/AmazingFeature)
  5. Open a Pull Request

License

This project is licensed under the MIT License - see the LICENSE file for details.