# Deployment Overview Mastra applications can be deployed to any Node.js-compatible environment. You can deploy a Mastra server, integrate with an existing web framework, deploy to cloud providers, or use Mastra Cloud for managed hosting. ## Runtime support Mastra can run against any of these runtime environments: - Node.js `v22.13.0` or later - Bun - Deno - Cloudflare ## Deployment Options ### Mastra Server Mastra provides a [server](https://mastra.ai/docs/server/mastra-server/llms.txt) powered by Hono that can be deployed independently. Use the `mastra build` command to build your application and deploy the output to your preferred VM, container, or PaaS platform. Use this option when you need full control over your infrastructure, long-running processes, or WebSocket connections. The [Mastra server deployment guide](https://mastra.ai/docs/deployment/mastra-server/llms.txt) provides more details. ### Monorepo Deploy a Mastra server as part of a monorepo setup, following the same approach as standalone deployment. Read about [monorepo deployment](https://mastra.ai/docs/deployment/monorepo/llms.txt). ### Cloud Providers Mastra applications can be deployed to cloud providers and serverless platforms. Mastra includes optional built-in deployers for Vercel, Netlify, and Cloudflare to automate the build and deployment process. Use this option for auto-scaling, minimal infrastructure management, or when you're already using one of these platforms. - [Amazon EC2](https://mastra.ai/guides/deployment/amazon-ec2/llms.txt) - [AWS Lambda](https://mastra.ai/guides/deployment/aws-lambda/llms.txt) - [Azure App Services](https://mastra.ai/guides/deployment/azure-app-services/llms.txt) - [Cloudflare](https://mastra.ai/guides/deployment/cloudflare-deployer/llms.txt) - [Digital Ocean](https://mastra.ai/guides/deployment/digital-ocean/llms.txt) - [Netlify](https://mastra.ai/guides/deployment/netlify-deployer/llms.txt) - [Vercel](https://mastra.ai/guides/deployment/vercel-deployer/llms.txt) ### Web Framework When Mastra is integrated with a web framework, it deploys alongside your application using the framework's standard deployment process. The guides below cover framework-specific configuration requirements for deployment. Use these guides when adding Mastra to an existing Next.js or Astro application. - [With Next.js on Vercel](https://mastra.ai/docs/deployment/web-framework/llms.txt) - [With Astro on Vercel](https://mastra.ai/docs/deployment/web-framework/llms.txt) - [With Astro on Netlify](https://mastra.ai/docs/deployment/web-framework/llms.txt) ### Mastra Cloud We're building Mastra Cloud to be the easiest place to deploy and observe your Mastra agents. It's currently in beta. Learn more in the [Mastra Cloud docs](https://mastra.ai/docs/mastra-cloud/overview/llms.txt). ## Workflow Runners Mastra workflows run using the built-in execution engine by default. For production workloads requiring managed infrastructure, workflows can also be deployed to specialized platforms like [Inngest](https://www.inngest.com) that provide step memoization, automatic retries, and real-time monitoring. Visit the [Workflow Runners guide](https://mastra.ai/docs/deployment/workflow-runners/llms.txt) for execution options and the [Inngest deployment guide](https://mastra.ai/guides/deployment/inngest/llms.txt) for setup instructions.