The agent harness performance optimization system. Skills, instincts, memory, security, and research-first development for Claude Code, Codex, Opencode, Cursor and beyond.
This integration supports injecting some dynamic content, presently the date and time, into the active Conversation Agent prompt when making a request.
This was added as it is beneficial for the model to be grounded with this context in its role as an assistant, and was previously added to the system prompt by Home Assistant itself before later being removed due to negative effects on prompt caching
and performance.
This was previously always-on but has been extracted as an experimental configuration option as this is not a once-size-fits-all for all models.
To this end I have provided a number of options so that users can try them out and select the one that works best, or disable entirely if none work well, for their chosen model.
The available options are:
Tool Result:
The date and time are inserted as a Tool Call Result message to the model, before the current user message.
As long as the model does not reject it, this is the recommended method to use and produces the most reliable results during testing.
Assistant:
The date and time are inserted as an additional Assistant message to the model, before the current user message.
In cases where the Tool Call Result role method does not work for a model, this is the next recommended to test with.
User:
The date and time are inserted as an additional User message to the model, before the current user message.
Recommended only where neither the System nor Assistant injection methods work for the model, but may not produce desirable results.
Some models have been known to repeat the date/time back to the user without request.
Disabled (no selection):
If your model simply refuses to work well with any method, simply remove the value from the configuration option to disable this again.
Experimental: Retrieval Augmented Generation (RAG) with Weaviate
Retrieval Augmented Generation is used to pre-feed your LLM messages with related data to provide contextually relevant information to the model based on the user input message.
This integration supports connecting your Agent to a Weaviate vector database server.
Once configured, user messages to the Agent will be queried against the Weaviate database first, and the result data pre-emptively injected into the current conversation as contextual data for the Agent to utilise in their response.
This is not a general-purpose "memory" for the Agent: content is only provided to the Agent if it matches on the current user input message to the model.
A pre-made docker-compose.yml is provided in the weaviate directory of this repository.
Weaviate Cloud is not supported: there is no free tier available and its cheapest pricing plan isn't attractive for personal/home use, and so I don't anticipate demand for this.
Reconfigure your LLM Server entity (not the Agent entity) in Home Assistant.
Expand the Weaviate configuration section and fill in the details server address and API key (homeassistant if using the supplied docker-compose.yml).
Optional: Reconfigure your AI Agent entities in Home Assistant.
This is only needed if you wish to change the default Weaviate values on a per-agent basis:
Object class name: Defaults to Homeassistant, can be changed if you want a different data store for the Agent. The integration will handle creating the required object class within Weaviate if it does not already exist.
Maximum number of results to use: Defaults to 2.
Result score threshold: Defaults to 0.9.
Hybrid search alpha: Defaults to 0.5. Balances the hybrid result scoring between 0 (fully text-matched) and 1 (fully vectorised) matching.
Managing Data
Self-hosted Weaviate does not come with a front-end to manage data at all at this current point in time.
I have included a simple NodeJS-based WebApp server within the /weaviate directory of this repository, that can be used to connect to your local Weaviate instance and view, query, and manage the data in your object class.
This is also setup into the supplied docker-compose.yml and exposed on port 9090 by default.
This tool supports the following basic functionality:
Connect to your server and list the available object classes.
View the available entry data in each class.
Add new entry data to a class.
Perform vector and hybrid searches against an object class.
This is not a general-purpose Weaviate management tool, rather it is purpose-built specifically for use with this integration and the object classes that it creates.
Notes
Only the current generations user message is queried in the database, no prior user messages are included.
Search results are used for the current user/assistant turns only (including multiturn tool usages), and do not carry forward to subsequent user/assistant turns.
Objects are stored as 2 pieces of data: the query, and the content:
The query is what is vectorised and the user inputs searched against.
The content is the main content to be provided to be fed to th
A Claude Code plugin that automatically captures everything Claude does during your coding sessions, compresses it with AI (using Claude's agent-sdk), and injects relevant context back into future sessions.