Blue-Green Deployment
A deployment strategy that maintains two identical production environments, blue and green. One environment serves live traffic while the other receives the new deployment. Traffic is switched atomically once the new version is verified, enabling instant rollback.
Blue-green deployment works by having two complete production environments behind a load balancer or DNS switch. The blue environment runs the current production version while the green environment receives the new deployment. After smoke tests verify the green environment, traffic is switched from blue to green. If problems emerge, traffic can be instantly reverted to blue, which still runs the known-good version.
For AI product teams, blue-green deployment is valuable for model updates because it provides instant rollback capability. If a new model version produces unexpected outputs in production, switching back to the previous version takes seconds rather than the minutes or hours needed to redeploy. Growth teams appreciate blue-green deployments because experiment rollbacks are equally fast: if a feature flag change causes metric degradation, the entire deployment can be reverted while the team investigates. The main cost is maintaining two full production environments, which doubles infrastructure expenses during the deployment window. For expensive GPU-equipped inference infrastructure, teams may optimize by sharing the model serving layer and only blue-greening the application code.
Related Terms
Content Delivery Network
A geographically distributed network of proxy servers that caches and delivers content from locations closest to end users. CDNs reduce latency, improve load times, and absorb traffic spikes by serving content from edge nodes rather than a single origin server.
Edge Computing
A distributed computing paradigm that processes data closer to the source of generation rather than in a centralized data center. Edge computing reduces latency, conserves bandwidth, and enables real-time processing for latency-sensitive applications.
Serverless Computing
A cloud execution model where the provider dynamically manages server allocation and scaling. Developers deploy functions or containers without provisioning infrastructure, paying only for actual compute time consumed rather than reserved capacity.
Function as a Service
A serverless computing category where developers deploy individual functions that execute in response to events. FaaS platforms like AWS Lambda, Google Cloud Functions, and Azure Functions handle all infrastructure management, scaling each function independently.
Platform as a Service
A cloud computing model that provides a complete development and deployment environment without managing underlying infrastructure. PaaS offerings like Heroku, Vercel, and Google App Engine handle servers, storage, networking, and runtime configuration.
Infrastructure as a Service
A cloud computing model that provides virtualized computing resources over the internet. IaaS offerings like AWS EC2, Google Compute Engine, and Azure Virtual Machines give teams full control over servers, storage, and networking without owning physical hardware.