npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

Iโ€™ve always been into building performant and accessible sites, but lately Iโ€™ve been taking it extremely seriously. So much so that Iโ€™ve been building a tool to help me optimize and monitor the sites that I build to make sure that Iโ€™m making an attempt to offer the best experience to those who visit them. If youโ€™re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, ๐Ÿ‘‹, Iโ€™m Ryan Hefnerย  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If youโ€™re interested in other things Iโ€™m working on, follow me on Twitter or check out the open source projects Iโ€™ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soonโ€“ish.

Open Software & Tools

This site wouldnโ€™t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you ๐Ÿ™

ยฉ 2025 โ€“ย Pkg Stats / Ryan Hefner

@llamafarm/llamafarm

v0.1.0

Published

๐ŸŒพ Plant and harvest AI models, agents, and databases into single deployable binaries

Downloads

9

Readme

๐ŸŒพ LLaMA Farm CLI

Deploy AI models, agents, and databases into single deployable binaries - no cloud required.

Installation

npm install -g @llamafarm/llamafarm

Quick Start

# Deploy a model
llamafarm plant llama3-8b

# Deploy with optimization
llamafarm plant llama3-8b --optimize

# Deploy to specific target
llamafarm plant mistral-7b --target raspberry-pi

# Development/Testing (no model download)
llamafarm plant llama3-8b --mock

Complete Workflow Example

# 1. Plant - Configure your AI deployment
llamafarm plant llama3-8b \
  --device mac-arm \
  --agent chat-assistant \
  --rag \
  --database vector

# 2. Bale - Compile to single binary
llamafarm bale ./.llamafarm/llama3-8b \
  --device mac-arm \
  --optimize

# 3. Harvest - Deploy anywhere
llamafarm harvest llama3-8b-mac-arm-v1.0.0.bin --run

# Or just copy and run directly (no dependencies needed!)
./llama3-8b-mac-arm-v1.0.0.bin

Features

  • ๐ŸŽฏ One-Line Deployment - Deploy complex AI models with a single command
  • ๐Ÿ“ฆ Zero Dependencies - Compiled binaries run anywhere
  • ๐Ÿ”’ 100% Private - Your data never leaves your device
  • โšก Lightning Fast - 10x faster than traditional deployments
  • ๐Ÿ’พ 90% Smaller - Optimized models use fraction of original size

Commands

plant

Deploy a model to create a standalone binary.

llamafarm plant <model> [options]

Options:
  --target <platform>    Target platform (mac, linux, windows, raspberry-pi)
  --optimize            Enable size optimization
  --agent <name>        Include an agent
  --rag                 Enable RAG pipeline
  --database <type>     Include database (vector, sqlite)

Examples

# Basic deployment
llamafarm plant llama3-8b

# Deploy with RAG and vector database
llamafarm plant mixtral-8x7b --rag --database vector

# Deploy optimized for Raspberry Pi
llamafarm plant llama3-8b --target raspberry-pi --optimize

# Deploy with custom agent
llamafarm plant llama3-8b --agent customer-service

bale

๐ŸŽฏ The Baler - Compile your deployment into a single executable binary.

llamafarm bale <project-dir> [options]

Options:
  --device <platform>   Target platform (mac, linux, windows, raspberry-pi)
  --output <path>       Output binary path
  --optimize <level>    Optimization level (none, standard, max)
  --sign               Sign the binary for distribution
  --compress           Extra compression (slower but smaller)

The Baler packages everything into a single binary:

  • ๐Ÿง  Quantized model (GGUF format)
  • ๐Ÿค– Agent configuration & code
  • ๐Ÿ—„๏ธ Embedded vector database
  • ๐ŸŒ Web UI
  • ๐Ÿš€ Node.js runtime
  • ๐Ÿ”ง Platform-specific optimizations

Supported Platforms:

  • mac / mac-arm / mac-intel - macOS with Metal acceleration
  • linux / linux-arm - Linux with CUDA support
  • windows - Windows with DirectML/CUDA
  • raspberry-pi - Optimized for ARM devices
  • jetson - NVIDIA Jetson edge devices

Typical Binary Sizes:

  • 7B models: 4-8GB (depending on quantization)
  • 13B models: 8-13GB
  • Mixtral: 25-45GB

Bale Examples

# Standard compilation
llamafarm bale ./.llamafarm/llama3-8b --device mac-arm

# Optimized for size
llamafarm bale ./.llamafarm/llama3-8b --device raspberry-pi --optimize max --compress

# Enterprise deployment with signing
llamafarm bale ./.llamafarm/mixtral --device linux --sign --output production.bin

harvest

Deploy and run a compiled binary.

llamafarm harvest <binary-or-url> [options]

Options:
  --run                 Run immediately after deployment
  --daemon             Run as background service
  --port <number>      Override default port
  --verify             Verify binary integrity

Configuration

Create a llamafarm.yaml file for advanced configurations:

name: my-assistant
base_model: llama3-8b
plugins:
  - vector_search
  - voice_recognition
data:
  - path: ./company-docs
    type: knowledge
optimization:
  quantization: int8
  target_size: 2GB

Then build:

llamafarm build

Requirements

  • Node.js 18+
  • 8GB RAM (minimum)
  • 10GB free disk space

Documentation

For full documentation, visit https://docs.llamafarm.ai

Support

Baler FAQ

Q: Can I run the binary on a different OS than where I compiled it? A: No, you need to compile for each target platform. Use --device to specify the target.

Q: How much disk space do I need? A: During compilation, you need ~3x the final binary size. The final binary is typically 4-8GB for 7B models.

Q: Can I update the model without recompiling? A: No, the model is embedded in the binary. This ensures zero dependencies but means updates require recompilation.

Q: Does the binary need internet access? A: No! Everything runs completely offline once deployed.

License

MIT ยฉ LLaMA Farm Team