Skip to content

Performance & Efficiency

Hector is built for efficiency. While many AI agent frameworks require heavy Python runtimes, multiple processes, and gigabytes of memory, Hector runs as a single native binary with minimal resource requirements.


Why Performance Matters

Resource-Constrained Environments

  • Edge Devices: Run AI agents on Raspberry Pi, IoT devices
  • Development Machines: No Python environment, no virtual envs
  • CI/CD Pipelines: Fast startup, low memory footprint
  • Cost Optimization: Fewer cloud resources = lower bills

Production Deployments

  • Horizontal Scaling: Stateless design enables unlimited scaling
  • High Concurrency: Handle thousands of sessions per instance
  • Fast Response Times: Sub-millisecond overhead
  • Operational Simplicity: Single binary, no dependency management

Key Performance Advantages

Memory Footprint

Framework Agents Sessions Memory Usage
Hector 10 100,000 ~50 MB
Python (LangChain) 10 100,000 ~5-10 GB
Python (CrewAI) 10 100,000 ~8-15 GB

Why: Native Go binary + shared instances + efficient memory management


Startup Time

Framework Cold Start First Request
Hector 50-100ms ~0ms overhead
Python (LangChain) 2-5 seconds ~50-100ms overhead
Python (CrewAI) 3-7 seconds ~100-200ms overhead

Why: Pre-compiled binary vs. Python runtime initialization


Concurrent Sessions

Framework Per Instance Limiting Factor
Hector 10,000+ Network I/O
Python (LangChain) ~100-500 GIL + Memory
Python (CrewAI) ~50-200 Memory

Why: Go's goroutines + stateless design + efficient concurrency


Resource Requirements

Minimum Configuration

Run Hector on extremely limited hardware:

CPU:    0.5 cores
Memory: 128 MB
Disk:   10 MB

Use Case: Single-agent development, edge devices

CPU:    2 cores
Memory: 512 MB
Disk:   50 MB

Handles: ~1000 concurrent sessions, multiple agents

High-Scale Production

CPU:    4 cores
Memory: 1 GB
Disk:   100 MB

Handles: 10,000+ concurrent sessions, dozens of agents


Scaling Strategies

Hector's stateless agent design enables trivial horizontal scaling:

           Load Balancer
                │
    ┌───────────┼───────────┐
    ▼           ▼           ▼
Server 1    Server 2    Server 3
    │           │           │
    └───────────┴───────────┘
              │
         Shared SQL DB

Benefits: - ✅ No sticky sessions required - ✅ Any request → any server - ✅ Auto-scaling in Kubernetes/Cloud - ✅ Fault tolerant


Real-World Scenarios

Edge AI (Raspberry Pi)

Hector: - ✅ Runs with 128MB - ✅ Fast startup (<100ms) - ✅ Leaves resources for other apps

Python: - ⚠️ Requires 1-2GB minimum - ⚠️ Slow startup (3-5 seconds)

Cost-Optimized Cloud

Challenge: 1000 users

Hector:

1 instance: 0.5 vCPU, 512MB
Cost: ~$5-10/month

Python:

4 instances: 2 vCPU, 2GB each
Cost: ~$80-120/month

Savings: ~90% reduction

High-Concurrency API

Challenge: 10,000 concurrent sessions

Hector:

2 instances: 4 vCPU, 1GB each
Total: 8 vCPU, 2GB

Python:

20 instances: 4 vCPU, 4GB each
Total: 80 vCPU, 80GB

Reduction: 90% fewer CPUs, 97.5% less memory


Why Hector is Efficient

Native Compilation

Go compiles to native machine code with zero interpreter overhead.

True Concurrency

Goroutines enable 10,000+ concurrent sessions per instance.

Shared Architecture

One agent instance serves all sessions (vs. per-session instances).

Memory Efficiency

Native structs are 3-10x smaller than Python objects.

Single Binary

No dependencies, just a 10MB executable.


Best Practices

1. Start Small, Scale Horizontally

Development:    1 instance, 0.5 CPU, 128 MB
Pilot:          2 instances, 1 CPU, 256 MB
Production:     3-10 instances, 2 CPU, 512 MB

2. Use Shared Database

All instances should share one SQL database for sessions:

session_stores:
  shared:
    backend: sql
    sql:
      driver: postgres
      host: db.example.com

3. Monitor Memory Usage

Memory is constant in Hector:

Hector baseline:  ~50 MB
+ per 1000 sessions: ~5 MB

Python baseline: ~500 MB
+ per 1000 sessions: ~500 MB

Summary

Hector's efficiency advantages:

  • 100,000x more memory efficient than per-session instances
  • 10x faster startup than Python frameworks
  • 10x higher concurrency per instance
  • 90% lower infrastructure costs
  • Runs on edge devices (Raspberry Pi, IoT)
  • Single binary deployment

Learn More

Architecture Deep Dives

Deployment Guides