AI Provider for exo
Connect WordPress to exo — run frontier AI models locally on your device cluster.
Description
Section titled “Description”This plugin registers exo as an AI provider in WordPress 7’s AI Client SDK and Connectors page.
exo connects all your devices into an AI cluster, enabling you to run frontier models locally. It exposes an OpenAI-compatible API that this plugin connects to.
Features
Section titled “Features”- Registers exo as a WordPress AI provider
- OpenAI-compatible text generation via exo’s chat completions API
- Auto-detect active models from your running exo cluster
- Capability detection — displays model capabilities (Text, Code, Thinking, Vision) as badges
- “Connect & Detect” / “Save & Re-detect” connector flow
- Optional API key authentication with secure storage
- Configurable endpoint URL (default:
http://localhost:52415) - Settings integrated into WordPress 7’s Connectors page
- Full i18n support with Norwegian Bokmål (nb_NO) translation
Choosing a Model
Section titled “Choosing a Model”exo exposes every model in its catalog, but only models actively loaded on your cluster will respond. Use Connect & Detect on the Connectors page to discover which models are running.
Recommended: Instruct models — These produce clean, usable output for WordPress AI features (title generation, content suggestions, etc.).
| Model | Size | Notes |
|---|---|---|
Llama-3.2-3B-Instruct-8bit | ~3 GB | Fast, great for short tasks |
Meta-Llama-3.1-8B-Instruct-4bit | ~4 GB | Good balance of speed and quality |
Llama-3.3-70B-Instruct-4bit | ~35 GB | Best quality, needs a larger cluster |
Avoid: Reasoning/thinking models (Qwen3.5, DeepSeek, GLM, Nemotron-Nano) — These models spend most of their token budget on internal chain-of-thought, producing slow responses with minimal visible output. They may return their thinking process instead of a clean answer.
To load a model on your cluster:
exo run mlx-community/Llama-3.2-3B-Instruct-8bitRequirements
Section titled “Requirements”- WordPress 7.0 or later
- PHP 8.3 or later
- A running exo cluster
Installation
Section titled “Installation”- Download
ai-provider-for-exo.zip - Upload via
Plugins → Add New → Upload Plugin - Activate via
WordPress Admin → Plugins - Go to Settings → Connectors and configure your exo endpoint.
- Click Connect & Detect to auto-discover active models and capabilities.
Configuration
Section titled “Configuration”The plugin can be configured via the Connectors page or environment variables:
| Setting | Environment Variable | Default |
|---|---|---|
| Endpoint URL | EXO_ENDPOINT | http://localhost:52415 |
| API Key | EXO_API_KEY | (none) |
| Model | EXO_MODEL | (auto-detected) |
You can also define these as constants in wp-config.php:
define( 'EXO_ENDPOINT', 'http://localhost:52415' );define( 'EXO_API_KEY', 'your-key-here' );define( 'EXO_MODEL', 'model-name' );Development
Section titled “Development”# Install dependenciescomposer installnpm install
# Buildnpm run build
# Watchnpm start
# Run testscomposer test # PHPUnitnpm test # Vitest
# i18nnpm run i18nChangelog
Section titled “Changelog”See CHANGELOG.md for a detailed list of changes.
License
Section titled “License”GPL-2.0-or-later. See LICENSE.
📦 Source: soderlind/ai-provider-for-exo · Edit on GitHub