junr-test1-plugin-minimax
v0.1.1
Published
MiniMax AI model plugin for XpertAI platform
Readme
Xpert Plugin: MiniMax
Overview
@xpert-ai/plugin-minimax connects MiniMax AI models to the XpertAI platform. The plugin integrates MiniMax's OpenAI-compatible API so XpertAI agents can leverage MiniMax's large language models, text embeddings, and text-to-speech capabilities.
Core Features
- Ships
MiniMaxModule, which registers the NestJS provider strategy, lifecycle hooks, and configuration schema required by the plugin runtime. - Implements
MiniMaxLargeLanguageModel, a LangChain-powered adapter built onChatOAICompatReasoningModelthat supports streaming chat completions, function calling, and token accounting callbacks for agent telemetry. - Provides
MiniMaxTextEmbeddingModel, a custom implementation that handles MiniMax's specific embedding API format with support for document and query embedding types. - Exposes
MiniMaxTTSModel, which supports streaming text-to-speech synthesis with multiple voice options and audio formats. - Shares a console-ready
manifest.yamlthat drives the XpertAI UI forms (icons, help links, credential prompts) for quick operator onboarding.
Installation
npm install @xpert-ai/plugin-minimaxPeer Dependencies: Ensure your host service also provides
@xpert-ai/plugin-sdk,@nestjs/common,@nestjs/config,@metad/contracts,@langchain/core,@langchain/openai,zod, andtslib. Refer topackage.jsonfor exact versions.
Enabling in XpertAI
- Add the plugin to the service dependencies so Node.js can resolve
@xpert-ai/plugin-minimax. - Declare the plugin before bootstrapping the XpertAI server:
PLUGINS=@xpert-ai/plugin-minimax - In the XpertAI admin console (or config file), create a model provider pointing to
minimax, then add individual models that map to the specific MiniMax model variants you want to use.
Credentials & Model Configuration
The plugin schema backs the form fields you see in the console:
| Field | Description |
| --- | --- |
| api_key | Required. Your MiniMax API Key from platform.minimaxi.com. |
| group_id | Required. Your MiniMax Group ID from platform.minimaxi.com. |
| base_url | Optional. Base URL for API requests (defaults to https://api.minimaxi.com). Useful for custom endpoints or proxy configurations. |
During validation, the plugin checks that both api_key and group_id are provided and validates the base URL format if specified.
Supported Models
Large Language Models (LLM)
MiniMax-M2- Latest MiniMax M2 modelMiniMax-M2-Stable- Stable version of M2 modelabab7-chat-preview- Preview version of abab7 chat modelabab6.5-chat- abab6.5 chat modelabab6.5s-chat- Speed-optimized abab6.5 chat modelabab6.5t-chat- Turbo version of abab6.5 chat modelabab6-chat- abab6 chat modelabab5.5-chat- abab5.5 chat modelabab5.5s-chat- Speed-optimized abab5.5 chat modelabab5-chat- abab5 chat modelminimax-text-01- Text generation modelminimax-m1- MiniMax M1 model
Text Embedding Models
embo-01- MiniMax embedding modeltext-embedding-ada-002- OpenAI-compatible embedding model
Text-to-Speech Models
speech-01- Standard TTS modelspeech-01-hd- High-definition TTS modelspeech-01-turbo- Turbo TTS modelspeech-02- Speech-02 modelspeech-02-hd- High-definition Speech-02 modelspeech-02-turbo- Turbo Speech-02 modeltts-1- TTS-1 modeltts-1-hd- High-definition TTS-1 model
Model Capabilities
- Conversational Models:
MiniMaxLargeLanguageModelmerges provider credentials with per-model overrides, enables streaming, and registers token usage callbacks so agent telemetry stays accurate. - Embedding Models:
MiniMaxTextEmbeddingModelimplements custom embedding logic to handle MiniMax's specific API format, supporting both document and query embedding types. - TTS Models:
MiniMaxTTSModelsupports streaming text-to-speech synthesis with configurable voice settings, audio formats (mp3, wav, opus, aac, flac), and playback speed.
Development & Debugging
From the repo root, run Nx commands for this package:
cd xpertai
npx nx build models/minimax
npx nx lint models/minimaxArtifacts land in xpertai/models/minimax/dist. Jest settings live in jest.config.ts.
License
This plugin is distributed under the MIT License located at the repository root.
