Heym: self-hosted AI workflow runtime built beyond n8n

person using macbook pro on table

n8n is a solid tool. So is Zapier. So is Make. They were built for deterministic, rule-based automation, and they do that job well. The problem is that AI-native workflows are not deterministic or rule-based. The moment you add agents that reason, documents that need retrieval, or approval steps before anything consequential fires, those platforms stop being a foundation and start being something you fight against.

That is the problem the developer of Heym ran into. Their response was to build a new runtime from scratch rather than keep patching a tool that was not designed for the job.

️ What Heym Is

Heym is a self-hosted, source-available AI workflow automation platform. It ships as a Docker Compose stack and runs entirely on your own infrastructure. No data leaves your stack.

The stated design goal is a single runtime for everything an AI workflow needs: agents, document retrieval, human-in-the-loop approval checkpoints, observability, scheduling, and the ability to expose any workflow as a callable tool for external AI assistants.

man in gray sweater standing beside wall

How the Execution Engine Works

The workflow engine builds a directed acyclic graph from the canvas and runs independent nodes concurrently using a thread pool. Streaming mode emits events as each node completes, so the frontend updates in real time rather than waiting for the full run to finish.

Agent nodes run a full tool-calling loop. They can execute Python tools, connect to external MCP servers, delegate to sub-agents, and call other Heym workflows as tools. When context usage approaches 80% of the model window, the engine automatically compresses history. Long-running agents don’t silently fail mid-task because the context filled up.

Human-in-the-Loop as a First-Class Primitive

This is where Heym makes the strongest case for itself. Most automation tools treat human review as an afterthought, a webhook you fire and then poll. Heym treats it as a design primitive.

The HITL node pauses execution at any point in the workflow, generates a public one-time review URL, and waits. A reviewer can accept, edit, or refuse the output without needing a Heym account. When they respond, execution resumes from an exact stored snapshot of where it paused. The same run can pause more than once.

For workflows that touch anything with real consequences, drafted emails, generated reports, data transformations feeding downstream systems, this matters. You get the speed of automation with a human gate exactly where you need it.

Built-in Vector Retrieval

Document retrieval in Heym is not a separate service called via API. It is native to the runtime. You upload documents, create vector stores, and wire semantic search directly into the workflow canvas. The entire retrieval pipeline runs inside a single workflow and shows up in a single trace.

The developer’s argument here is straightforward: bolting retrieval on from outside creates two systems to maintain and two places to debug, with no unified trace across either. Native retrieval eliminates that split.

black and gray laptop computer

MCP Server Included

Every Heym instance runs a built-in MCP server. Any workflow you build can be exposed as a tool that Claude Desktop, Cursor, or any MCP-compatible client can call directly. Agent nodes can also connect to external MCP servers as tool sources, so capabilities flow in both directions: into Heym from external tools, and out of Heym to external AI assistants.

Observability and Evals

The Traces tab logs every execution automatically. The Evals tab lets you build test suites and run evaluations across multiple models simultaneously with configurable scoring. According to the developer, this is how they debug their own workflows, not a feature added for marketing purposes.

If you can’t inspect what your AI workflow did, why it did it, and when, you can’t trust it in production. Heym’s position is that observability has to be native, not bolted on after the fact.

️ The Stack

  • Frontend: Vue 3 with TypeScript and Vue Flow
  • Backend: Python and FastAPI with async SQLAlchemy
  • Storage: PostgreSQL 16
  • Deployment: Docker Compose

Pro Tip

If you are self-hosting anything AI-related and want a single audit trail across agents, retrieval, and approvals, native observability in the runtime is worth more than you might expect. Debugging a three-system trace is a different job entirely than reading one.

Pricing and Status

Heym is currently at v0.0.1. It is actively developed and released under MIT plus Commons Clause, which means source-available rather than fully open source. Free to self-host. The source is on GitHub.

This is early software. If you’re building AI workflows and find yourself writing more glue code than actual workflow logic, it’s worth a look. The architecture decisions, particularly around HITL as a primitive and native retrieval, are the right instincts for where agent-based automation is heading.

Stay on top of AI & Automation with BizStack Newsletter
BizStack  —  Entrepreneur’s Business Stack
Logo