Deep Research — April 2026 · Top 7 platforms for running lightweight, single-purpose AI agents with memory, wallet integration, and per-user isolation
This research evaluates cloud platforms for deploying lightweight AI agents — one agent per user, calling external LLM APIs (not running models locally), with embedded state/memory, wallet integration (Turnkey, Dynamic, Privy), and strict security isolation. No GPU needed — the LLM provider handles inference.
The seven platforms below are ranked by fit for this specific architecture: stateful per-user agents, TypeScript-first, minimal attack surface, cost-efficient at scale (1K–100K users), with the ability to integrate crypto wallet SDKs server-side.
TL;DR: Cloudflare Durable Objects is the clear #1 for this architecture. Trigger.dev is the best complement for background jobs. Hetzner is the escape hatch if you outgrow serverless economics.
| Platform | Per-User Isolation | Embedded State | Hibernate / $0 Idle | TypeScript | Self-Hostable | Free Tier | Est. Cost @ 1K Users |
|---|---|---|---|---|---|---|---|
| 1. Cloudflare DO | Native (1 DO = 1 user) | SQLite per DO | Yes (WebSocket hibernate) | First-class | No | Yes (free plan) | $5–25/mo |
| 2. Trigger.dev | Via concurrency keys | Checkpointing | Yes | First-class | Apache 2.0 | $5/mo free credits | $10–50/mo |
| 3. Fly.io | 1 Machine per user (manual) | Volumes ($0.15/GB) | Yes (machine suspend) | Any runtime | No | Trial only (2h) | $30–150/mo |
| 4. Railway | Shared process | Volumes/Postgres | App sleeping (limited) | Any runtime | No | $5 trial credit | $35–80/mo |
| 5. Modal | Per-function sandboxing | Ephemeral | Yes (scale to zero) | Python only | No | $30/mo credits | $30–100/mo |
| 6. Hetzner VPS | Container-level (DIY) | Full disk (NVMe) | Always-on | Any runtime | Your infra | No | €4–15/mo |
| 7. DigitalOcean | Droplet or App Platform | Managed Postgres | Always-on | Any runtime | No | $200 credit (60 days) | $24–96/mo |
Stateful serverless at the edge — one Durable Object per user, embedded SQLite, hibernation, $0 when idle
Includes 10M requests, 30M CPU-ms. DO requests: $0.15/M beyond 1M included. SQLite reads: $0.001/M rows. Writes: $1/M rows. Storage: $0.75/GB-month. Hibernated DOs cost $0 in compute.
One Durable Object per user = one agent with private SQLite (10GB cap). Agents SDK provides AIChatAgent class with message persistence, tool calling, MCP support, WebSocket hibernate. Single-threaded execution per DO prevents race conditions on wallet operations.
Open-source TypeScript background jobs with retries, queues, and no timeouts — ideal as an execution layer alongside CF
Free: $5/mo credits, 10 concurrent runs, unlimited tasks. Paid from $10/mo. Per-run: compute at $0.0000338/sec (small-1x) + $0.000025/invocation. Self-host: Apache 2.0, unlimited runs.
Not a replacement for CF DOs but a powerful complement. Use for: long-running swap execution workflows, scheduled portfolio rebalancing, multi-step Turnkey wallet operations. Checkpointing means a crashed task resumes from last step. Concurrency keys enable per-user queuing.
Micro VMs close to users — suspend when idle, persistent volumes, multi-region by default
Pay-as-you-go. Shared-CPU 256MB: ~$0.0027/hr ($1.94/mo always-on). Volumes: $0.15/GB-mo. IPv4: $2/app/mo. Egress: $0.02/GB. No free tier for new users (2h trial). Machine suspend stops compute billing.
You could run 1 Fly Machine per user-agent. Each Machine has its own filesystem, can suspend/resume, and supports persistent volumes. Unlike DOs, you get a full Linux environment — run any binary, any language. But managing 10K+ Machines per user is operationally heavier than DOs.
Push code, get a running service — usage-based billing on your own hardware, not resold AWS
Hobby: $5/mo subscription + $5 included credit. Pro: $20/mo + $20 credit. Usage: CPU $0.000077/vCPU-sec, Memory $0.0000039/GB-sec, Volumes $0.25/GB-mo, Egress $0.05/GB. Runs on Railway's own hardware (US, EU, Asia).
Good for running a single monolithic agent service that handles all users. Not per-user isolation — you'd run one Docker container with your agent code and use an external database (Railway Postgres) for per-user state. App sleeping reduces idle costs. Git-push deploys are fast.
Serverless Python compute — decorators as infrastructure, GPU access, scale to zero
CPU: $0.0000131/core-sec base (×3.75 after region + non-preempt multipliers in US). GPU: A10G ~$1.10/hr, A100 ~$3.73/hr. Team: $250/mo + $100 credits. Sandboxes force non-preemptible pricing (3× premium). Scale to zero when idle.
Python-first, not TypeScript. Great for ML pipelines but wrong ecosystem for your TS-based Telegram bot architecture. No persistent per-user state (ephemeral containers). Sandboxes exist but at 3× premium cost. Best for batch processing or inference, not conversational agents.
Bare metal value — €4/mo for 2 vCPU / 4GB RAM / 40GB NVMe / 20TB traffic. Your infra, your rules.
CX23 (2 shared vCPU, 4GB, 40GB NVMe): €3.99/mo. CPX22 (2 vCPU, 4GB, 80GB NVMe): €7.99/mo. CCX13 (2 dedicated vCPU, 8GB, 80GB NVMe): €14.86/mo. EU includes 20TB traffic. US/SG: 1TB. Hourly billing, capped at monthly rate.
Self-host anything: run Docker containers, Trigger.dev workers, OpenClaw instances, Coolify, custom agent processes. Full root access. You build the isolation layer (Docker/podman per user, or single process with DB-backed state). Best price-to-compute ratio in the market.
Developer-friendly managed platform — from simple App Platform to Gradient AI agent infrastructure
Basic Droplet: $4/mo (512MB). $12/mo (2GB). App Platform: usage-based from $5/mo. Managed Postgres from $15/mo. Gradient AI: serverless inference with single API key across OpenAI/Anthropic/Meta. $200 free credit for 60 days.
Two paths: (1) Droplet as a VPS like Hetzner but with managed services (Postgres, Redis, monitoring). (2) Gradient AI for building/deploying agents with built-in knowledge bases and evals. More expensive than Hetzner for raw compute but easier to manage.
Primary Stack (MVP): Cloudflare Workers + Durable Objects + Agents SDK for the per-user agent core. One DO per user. SQLite for memory. AI Gateway for LLM routing. KV for rate limiting and token allowlists. This is your Openclaw architecture — it's validated.
Background Jobs (Phase 2+): Trigger.dev for durable swap workflows. When a user says "swap 100 USDC to NEAR", the DO dispatches to Trigger.dev which handles the multi-step process (quote → approve → sign via Turnkey → broadcast → confirm) with checkpointing and retries. Self-host on Hetzner to own the infra.
Scale Escape Hatch (Phase 3+): Hetzner VPS for high-volume background processing, admin dashboards, monitoring, and self-hosted Trigger.dev workers. €8–15/mo for specs that would cost $50+ on DO or AWS.
What NOT to use: Modal (wrong language — Python only). Railway (no per-user isolation). Fly.io (operational overhead of managing VMs > benefit over DOs). DigitalOcean (2–3× cost premium over Hetzner for raw compute). All of these are good platforms — just not the right fit for your specific "one lightweight agent per user" architecture.
Most agent hosting platforms in 2026 assume a shared-process model — one service handles all users, state lives in an external database. Your requirement (one isolated agent per user with embedded state, wallet keys never in shared memory, minimal attack surface) maps almost exclusively to Cloudflare Durable Objects.
The closest alternative would be Fly.io Machines (one VM per user with suspend), but the operational overhead of managing thousands of VMs exceeds the benefit when your agents are lightweight TS processes calling external APIs. DOs give you the same isolation model at 1/10th the operational cost.
The real decision isn't "which platform" — it's "what complements Cloudflare DOs" for the parts DOs can't handle well: long-running workflows (Trigger.dev), heavy background compute (Hetzner), and managed databases for analytics/reporting (DigitalOcean or Supabase, which you already use for auth).