Koyeb icon

Koyeb

Koyeb is a next-generation serverless platform that lets developers deploy and scale applications, containers, and ML models globally without managing infrastructure. It provides instant API endpoints, native autoscaling (including scale-to-zero), and access to GPUs and accelerators across 50+ locations. Koyeb supports Git-based and CLI deployments, managed Postgres with pgvector, and pay-per-use pricing with a forever free tier.

  • Serverless containers — Deploy production-grade containers with zero configuration using a Git push or the CLI; Koyeb provisions and scales instances automatically.
  • GPU & accelerator support — Run models and inference on NVIDIA, AMD, Intel, and other accelerators; choose instance types and pay per second for compute.
  • Global edge locations — Deploy to 50+ regions to reduce latency and improve availability for worldwide users.
  • Autoscaling & scale-to-zero — Use built-in autoscaling for high throughput and negligible cold starts, and scale to zero to avoid charges when idle.
  • Managed Postgres + pgvector — Store, index, and search embeddings with a fully managed serverless Postgres instance.
  • Instant API endpoints & logs — Deploy an API endpoint in seconds and troubleshoot with real-time logs and instance access.
  • Pay-per-use pricing & free tier — Start for free with the Hobby plan; upgrade to Pro or Scale for included compute credits and priority support.

To get started, sign up on the web console, connect a Git repository or use the CLI to push your app, pick instance hardware (CPU/GPU), and deploy — Koyeb handles routing, autoscaling, and global distribution.

No discussions yet

Be the first to start a discussion about Koyeb

Demo Video for Koyeb

Developer

Koyeb builds a developer-first serverless platform that deploys APIs, full-stack apps, and ML inference globally without infrastructure…read more

Pricing and Plans

(Freemium)

Hobby

Free

Forever free tier with a single free web service and a small managed Postgres instance for development and low-traffic workloads.

  • One free web Service (512MB RAM, 0.1 vCPU, 2GB SSD in supported regions)
  • One free PostgreSQL database (limited active time and storage)
  • Join without a credit card in supported regions

Pro

Popular
$29/month

Monthly plan for teams that need predictable billing and included compute credits.

  • Includes $10 of compute usage per month
  • Support during business hours
  • Access to larger instance types and priority resources

Scale

$299/month

Plan for production-scale deployments with higher included compute and enhanced support.

  • Includes $100 of compute usage per month
  • Additional support and direct Slack cross connection
  • Access to largest instances for production workloads

System Requirements

Operating System
Any OS with a modern web browser
Memory (RAM)
No local requirements (cloud-based)
Processor
Any modern CPU
Disk Space
No local storage required (cloud-based)

AI Capabilities

Model-deployment
Inference-serving
Hardware-accelerators