@memberjunction/ai-prompts
v5.2.0
Published
MemberJunction: AI Prompt Execution and Management
Keywords
Readme
@memberjunction/ai-prompts
Advanced AI prompt execution engine for MemberJunction. Provides hierarchical template composition, intelligent model selection with failover, parallel execution with judge-based result selection, structured output validation with retry, comprehensive execution tracking, and streaming support. This is the primary interface for executing AI prompts in the MemberJunction framework.
Architecture
graph TD
subgraph "@memberjunction/ai-prompts"
PR["AIPromptRunner"]
style PR fill:#2d8659,stroke:#1a5c3a,color:#fff
EP["ExecutionPlanner"]
style EP fill:#7c5295,stroke:#563a6b,color:#fff
PEC["ParallelExecutionCoordinator"]
style PEC fill:#7c5295,stroke:#563a6b,color:#fff
PE["ParallelExecution"]
style PE fill:#7c5295,stroke:#563a6b,color:#fff
end
subgraph "Execution Pipeline"
T["1. Template Rendering<br/>Handlebars + System Placeholders"]
style T fill:#b8762f,stroke:#8a5722,color:#fff
MS["2. Model Selection<br/>Default / Specific / ByPower"]
style MS fill:#b8762f,stroke:#8a5722,color:#fff
EX["3. LLM Execution<br/>With Streaming & Caching"]
style EX fill:#b8762f,stroke:#8a5722,color:#fff
VAL["4. Output Validation<br/>JSON Schema + Retry"]
style VAL fill:#b8762f,stroke:#8a5722,color:#fff
TRK["5. Execution Tracking<br/>AIPromptRun Records"]
style TRK fill:#b8762f,stroke:#8a5722,color:#fff
end
PR --> EP
PR --> PEC
PEC --> PE
PR --> T
T --> MS
MS --> EX
EX --> VAL
VAL --> TRK
subgraph Dependencies
AI["@memberjunction/ai<br/>BaseLLM"]
style AI fill:#2d6a9f,stroke:#1a4971,color:#fff
ACP["@memberjunction/ai-core-plus<br/>AIPromptParams"]
style ACP fill:#2d6a9f,stroke:#1a4971,color:#fff
AIE["@memberjunction/aiengine<br/>AIEngine"]
style AIE fill:#2d6a9f,stroke:#1a4971,color:#fff
TMPL["@memberjunction/templates<br/>TemplateEngine"]
style TMPL fill:#2d6a9f,stroke:#1a4971,color:#fff
CRED["@memberjunction/credentials<br/>CredentialEngine"]
style CRED fill:#2d6a9f,stroke:#1a4971,color:#fff
end
AI --> PR
ACP --> PR
AIE --> PR
TMPL --> PR
CRED --> PRInstallation
npm install @memberjunction/ai-promptsKey Features
Hierarchical Template Composition
Build complex prompts from reusable sub-templates with unlimited nesting depth:
import { AIPromptRunner } from '@memberjunction/ai-prompts';
import { AIPromptParams, ChildPromptParam } from '@memberjunction/ai-core-plus';
const runner = new AIPromptRunner();
// Parent template uses {{ analysis }} and {{ summary }} placeholders
const params = new AIPromptParams();
params.prompt = parentPrompt;
params.childPrompts = [
new ChildPromptParam(analysisParams, 'analysis'),
new ChildPromptParam(summaryParams, 'summary')
];
params.data = { userInput: 'complex data to process' };
const result = await runner.ExecutePrompt(params);Execution order:
- Child prompts render depth-first (children before parents)
- Sibling prompts at each level execute in parallel
- Child results replace placeholders in parent template
- Final composed prompt executes as a single LLM call
Model Selection Strategies
Three strategies for selecting which AI model executes a prompt:
| Strategy | Description |
|---|---|
| Default | Uses the AI configuration to determine the model based on priority and availability |
| Specific | Uses explicitly associated models from the AIPromptModels table |
| ByPower | Selects the highest PowerRank model matching the prompt's model type |
Model selection precedence (highest to lowest):
AIPromptParams.override-- Runtime model/vendor overrideAIPromptParams.modelSelectionPrompt-- Alternate prompt for model config- Prompt's own model configuration (strategy + associations)
Parallel Execution with Judging
Execute prompts across multiple models simultaneously and select the best result:
- Configurable execution groups with different models
- AI judge prompt evaluates and ranks results
- Automatic selection of best result based on judge scoring
- Full tracking of all parallel results
Output Validation and Retry
Automatic validation of AI outputs with configurable retry:
- JSON schema validation against
OutputExampledefinitions - Automatic JSON repair via JSON5 parsing and LLM-based repair
- Configurable retry count with the original or repaired prompts
- Validation syntax cleaning (removes
?,*,:typemarkers from JSON keys) - Detailed validation attempt tracking
Streaming Support
Real-time streaming of LLM responses:
const params = new AIPromptParams();
params.prompt = myPrompt;
params.onStreaming = (chunk) => {
process.stdout.write(chunk.content);
};
const result = await runner.ExecutePrompt(params);Execution Tracking
Every prompt execution creates an AIPromptRun record with:
- Model and vendor used
- Template rendering results
- Token usage (prompt + completion)
- Cost tracking
- Execution time
- Parent/child relationships for hierarchical prompts
- Agent run linkage via
agentRunId
Credential Resolution
Hierarchical credential resolution for API keys:
AIPromptParams.credentialId(per-request override)AIPromptModel.CredentialID(prompt-model specific)AIModelVendor.CredentialID(model-vendor specific)AIVendor.CredentialID(vendor default)AIPromptParams.apiKeys[](legacy runtime keys)AI_VENDOR_API_KEY__<DRIVER>environment variables (legacy)
Failover
When a model fails due to rate limiting, authentication errors, or other transient issues, the runner can automatically retry with alternate models from the selection candidates.
Usage
Basic Prompt Execution
import { AIPromptRunner } from '@memberjunction/ai-prompts';
import { AIPromptParams } from '@memberjunction/ai-core-plus';
import { AIEngine } from '@memberjunction/aiengine';
// Get prompt from metadata
await AIEngine.Instance.Config(false, contextUser);
const prompt = AIEngine.Instance.Prompts.find(p => p.Name === 'Summarize Content');
const runner = new AIPromptRunner();
const params = new AIPromptParams();
params.prompt = prompt;
params.data = { content: documentText, maxLength: 500 };
params.contextUser = contextUser;
const result = await runner.ExecutePrompt(params);
if (result.success) {
console.log(result.result); // Parsed/validated result
console.log(result.promptTokens); // Input tokens used
console.log(result.completionTokens); // Output tokens generated
console.log(result.executionTimeMS); // Execution duration
}With Progress Tracking
params.onProgress = (progress) => {
console.log(`[${progress.step}] ${progress.percentage}% - ${progress.message}`);
};With Effort Level
params.effortLevel = 85; // High effort for thorough analysis (1-100 scale)With Runtime Model Override
params.override = {
modelId: 'specific-model-id',
vendorId: 'specific-vendor-id'
};Dependencies
@memberjunction/ai-- Core AI abstractions (BaseLLM, ChatParams)@memberjunction/ai-core-plus-- AIPromptParams, AIPromptRunResult, extended entities@memberjunction/ai-engine-base-- AIEngineBase metadata cache@memberjunction/aiengine-- AIEngine server-side operations@memberjunction/core-- MJ framework core@memberjunction/core-entities-- Generated entity classes@memberjunction/credentials-- Credential resolution@memberjunction/templates-- Template rendering engine@memberjunction/templates-base-types-- Template base typesjson5-- Lenient JSON parsing for repair
