Virtual Context

v5.0.6

Plugin integrates OpenClaw to virtual-context.com and provides a virtual context window of 50M+ tokens for unlimited memory, reduces upstream costs by compressing the LLM payload, and increases LLM reasoning by providing only relevant context with memory retrieval tools.

openclaw-plugin-virtual-context·runtime virtual-context·by @yursil
Code Pluginsource linkedCommunity code plugin. Review compatibility and verification before install.

README

Virtual Context Plugin for OpenClaw

virtual-context.com — OS-style memory for LLMs. Less context. Better answers.

Virtual Context lets your agents run with unlimited context windows while sending only what matters to the LLM. Conversations are compressed, organized, and indexed automatically. When context is needed, it's retrieved semantically and injected into the payload. The result: unlimited memory, lower token costs, and better reasoning from models that see clean, relevant context instead of raw history.

This plugin provides deep OpenClaw integration via the Virtual Context REST API. For other frameworks, the transparent proxy requires zero code changes.

What It Does

  • Prepare — before each LLM call, sends your messages to the Virtual Context cloud. Gets back an compressed payload with relevant historical context injected.
  • Tools — registers retrieval tools (vc_expand_topic, vc_find_quote, vc_recall_all, vc_query_facts, vc_remember_when, vc_find_session) that the LLM can call to pull in more context on demand.
  • Ingest — after each LLM response, sends the assistant's reply to the cloud for tagging and indexing.

Installation

openclaw plugins install clawhub:virtual-context

Configuration

In openclaw.json:

{
  "plugins": {
    "entries": {
      "virtual-context": {
        "enabled": true,
        "config": {
          "vcKey": "vc-your-key-here",
          "baseUrl": "https://api.virtual-context.com",
          "providers": ["openai-direct/gpt-5.4"],
          "debug": false
        }
      }
    }
  }
}

Config Options

OptionTypeDefaultDescription
vcKeystringrequiredYour Virtual Context API key
baseUrlstringhttps://api.virtual-context.comVC REST API base URL
providersstring[]allProvider/model pairs to activate for. Empty = all providers. Example: ["openai-direct/gpt-5.4"]
debugbooleanfalseEnable verbose logging of REST API calls and payloads

How It Works

  1. Bootstrap — on startup, fetches tool definitions from /api/v1/tools/definitions and registers them as OpenClaw tools
  2. Before each LLM call — calls /api/v1/context/prepare with the full message history. The cloud returns an compressed payload with context injected, old turns trimmed, and tools added. The plugin replaces the messages in-place.
  3. After each LLM response — calls /api/v1/context/ingest with the assistant's reply text for tagging and compaction
  4. On tool calls — when the LLM requests a VC tool, the plugin calls /api/v1/tools/{name} and returns the result

Provider Filtering

By default, the plugin activates for all providers. Use the providers config to restrict it to specific provider/model combinations. The plugin reads the current model from the session store at runtime, so it correctly handles /model switches.

Security and Access

This plugin is transparent about what it accesses. Here is the full list:

Network calls (to your configured baseUrl):

  • Sends conversation messages to /api/v1/context/prepare before each LLM call
  • Sends assistant reply text to /api/v1/context/ingest after each LLM response
  • Fetches tool definitions from /api/v1/tools/definitions at startup
  • Calls /api/v1/tools/{name} when the LLM requests a retrieval tool

Local filesystem reads:

  • Reads ~/.openclaw/agents/<agentId>/sessions/sessions.json to determine the current model for provider filtering. This is a read-only access to OpenClaw's session store, used because the before_prompt_build hook does not expose the active model in its context. No writes.

Payload modification:

  • Replaces the message array in-place with the compressed payload returned by the cloud
  • Can override the system prompt if the cloud returns one (VC manages the full payload to compress it)

Tool registration:

  • Registers tools dynamically from definitions fetched from the cloud at startup

Debug logging (opt-in, off by default):

  • When debug: true, logs message previews, API responses, and payload sizes to the gateway log. Disable in production.

What it does NOT do:

  • Does not write to any local files (except gateway logs via the logger)
  • Does not access files outside the session store
  • Does not send data to any endpoint other than your configured baseUrl
  • Does not store credentials or API keys beyond what is in your openclaw.json config

Getting a vcKey

Sign up at virtual-context.com to get your API key. Free tier available. Pro ($19/mo) for unlimited.

Learn More

Capabilities

configSchema
Yes
Executes code
Yes
HTTP routes
0
Plugin kind
lifecycle
Runtime ID
virtual-context

Compatibility

Built With Open Claw Version
2026.3.24
Min Gateway Version
2026.3.24
Plugin Api Range
>=2026.3.24
Plugin Sdk Version
2026.3.24

Verification

Tier
source linked
Scope
artifact only
Summary
Validated package structure and linked the release to source metadata.
Commit
efefba6e9f96
Tag
main
Provenance
No
Scan status
clean

Tags

latest
5.0.6