npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2024 – Pkg Stats / Ryan Hefner

vercel-llm-api

v0.3.1

Published

A reverse engineered Node.js API wrapper for the Vercel AI Playground, which provides free access to many large language models without needing an account.

Downloads

20

Readme

Node.js Vercel LLM API

This is a reverse engineered API wrapper for the Vercel AI Playground, which allows for free access to many LLMs, including OpenAI's ChatGPT, Cohere's Command Nightly, as well as some open source models.

Also a JavaScript implementation of the ading2210's vercel-llm-api Python library.

Table of Contents

Table of contents generated with markdown-toc

Features

  • Download the available models
  • Generate text
  • Generate chat messages
  • Set custom parameters
  • Stream the responses

Limitations

  • No auth support
  • Can't use "pro" or "hobby" models

Installation

You can install this library by running the following command:

npm install vercel-llm-api

Documentation

Note that the entire library requires the use of async/await.

Using the Client

To use this library, simply require('vercel-llm-api') and create a Client instance. You can specify custom Axios request configurations as an argument.

See here for the Axios request config.

const { Client } = require('vercel-llm-api'),
  client = new Client();

client.on('ready', async () => {
  // the client is ready to do whatever
});

Note that the following examples assume client is the name of your Client instance and that it is inside an async function.

Downloading the Available Models

The client downloads the available models upon initialization, and stores them in client.models.

>>> console.log(client.models)

{
  "anthropic:claude-instant-v1": { 
    "id": "anthropic:claude-instant-v1", // the model's id
    "provider": "anthropic",             // the model's provider
    "providerHumanName": "Anthropic",    // the provider's display name
    "makerHumanName": "Anthropic",       // the maker of the model
    "minBillingTier": "hobby",           // the minimum billing tier needed to use the model
    "parameters": {                      // an object of optional parameters that can be passed to the generate function
      "temperature": {                   // the name of the parameter
        "value": 1,                      // the default value for the parameter
        "range": [0, 1]                  // a range of possible values for the parameter
      },
      ...
    }
    ...
  }
}

Note that, since there is no auth yet, if a model has the "minBillingTier" property present, it can't be used.

A list of model IDs is also available in client.model_ids.

>>> console.log(client.model_ids)
[
  "anthropic:claude-instant-v1", // locked to hobby tier; unusable
  "anthropic:claude-v1",         // locked to hobby tier; unusable
  "replicate:replicate/alpaca-7b",
  "replicate:stability-ai/stablelm-tuned-alpha-7b",
  "huggingface:bigscience/bloom",
  "huggingface:bigscience/bloomz",
  "huggingface:google/flan-t5-xxl",
  "huggingface:google/flan-ul2",
  "huggingface:EleutherAI/gpt-neox-20b",
  "huggingface:OpenAssistant/oasst-sft-4-pythia-12b-epoch-3.5",
  "huggingface:bigcode/santacoder",
  "cohere:command-medium-nightly",
  "cohere:command-xlarge-nightly",
  "openai:gpt-4",                // locked to pro tier; unusable
  "openai:code-cushman-001",
  "openai:code-davinci-002",
  "openai:gpt-3.5-turbo",
  "openai:text-ada-001",
  "openai:text-babbage-001",
  "openai:text-curie-001",
  "openai:text-davinci-002",
  "openai:text-davinci-003"
]

An Object of default parameters for each model can be found at client.model_params.

>>> console.log(client.model_defaults)
{
  "anthropic:claude-instant-v1": {
    "temperature": 1,
    "maximumLength": 200,
    "topP": 1,
    "topK": 1,
    "presencePenalty": 1,
    "frequencyPenalty": 1,
    "stopSequences": [
      "\n\nHuman:"
    ]
  },
  ...
}

Generating Text

To generate some text, use the client.generate function, which accepts the following arguments:

  • model - The ID of the model you want to use.
  • prompt - Your prompt.
  • params - An Object of optional parameters. See the previous section for how to find these.

The function returns the newly generated text as a ReadableStream.

await client.generate('openai:gpt-3.5-turbo', 'Summarize the GNU GPL v3');

Generating Chat Messages

To generate chat messages, use the client.chat function, which accepts the following arguments:

  • model - The ID of the model you want to use.
  • messages - A list of messages. The format for this is identical to how you would use the official OpenAI API.
  • params - An Object of optional parameters. See the "Downloading the Available Models" section for how to find these.

The function returns the newly generated text as a ReadableStream.

const messages = [
  {"role": "system", "content": "You are a helpful assistant."},
  {"role": "user", "content": "Who won the world series in 2020?"},
  {"role": "assistant", "content": "The Los Angeles Dodgers won the World Series in 2020."},
  {"role": "user", "content": "Where was it played?"}
];

await client.chat('openai:gpt-3.5-turbo', messages);

Using StreamHandler

StreamHandler is a utility function to handle the returned ReadableStream of the instantiated Client's chat and generate functions.

StreamHandler accepts the following arguments:

  • stream - The ReadableStream.
  • callback - An optional callback to process each chunk of the stream.

...and returns an Array of Strings.

const { Client, StreamHandler } = require('vercel-llm-api'),
  client = new Client();

client.on('ready', async () => {
  const messages = [
    {"role": "system", "content": "You are a helpful assistant."},
    {"role": "user", "content": "Who won the world series in 2020?"},
    {"role": "assistant", "content": "The Los Angeles Dodgers won the World Series in 2020."},
    {"role": "user", "content": "Where was it played?"}
  ];
  
  const stream = await client.chat('openai:gpt-3.5-turbo', messages),
    response = await StreamHandler(stream);

  console.log(response); // returns [ "The", " 2020", " World", " Series", " was", " played", ... ]
});

Miscellaneous

Listening to Debug Messages

If you want to show the debug messages, simply listen to the debug event of the Client instance.

const { Client } = require('vercel-llm-api'),
  client = new Client();

client.on('debug', console.log);