We compare the performance of the community, NGINX Open Source, and NGINX Plus Ingress Controllers in a dynamic Kubernetes cloud environment. As the number of Pod replicas scales up and down, only the NGINX Plus Ingress Controller doesn't incur high latencies.
As more and more enterprises run containerized apps in production, Kubernetes continues to solidify its position as the standard tool for container orchestration. At the same time, demand for cloud computing has been pulled forward by a couple of years because work-at-home initiatives prompted by the COVID‑19 pandemic have accelerated the growth of Internet traffic. Companies are working rapidly to upgrade their infrastructure because their customers are experiencing major network outages and overloads.
To achieve the required level of performance in cloud‑based microservices environments, you need rapid, fully dynamic software that harnesses the scalability and performance of the next‑generation hyperscale data centers. Many organizations that use Kubernetes to manage containers depend on an NGINX‑based Ingress controller to deliver their apps to users.
We announce eight solutions that unleash developer speed without sacrificing the control infrastructure teams require. Get the details on NGINX Controller, App Protect, Ingress Controller, Service Mesh, and Analytics Cloud Service; F5 DNS Cloud Services; and Red Hat integrations.
Release 1.8.0 of the NGINX Ingress Controller for Kubernetes introduces NGINX App Protect integration, extensibility mechanisms for NGINX Ingress resources, URI rewrites and request and response header modification, policies and IP address-based ACLs, and more.
With NGINX Plus Ingress Controller for Kubernetes release 1.8.0, NGINX App Protect can be embedded in the Ingress Controller. This puts WAF protection closer to applications, which is crucial in modern app environments like Kubernetes. It also enables automation and reduces complexity and cost.
Easy and Robust Single Sign-on with OpenID Connect and NGINX ingress Controller. With the release of NGINX Ingress Controller 1.10.0, we are happy to announce a major enhancement: a technology preview of OpenID Connect (OIDC) authentication.
The demo aims at running an application in Kubernetes behind a Cloud-managed public load balancer also known as an HTTP(s) load balancer which is also known as an Ingress resource in Kubernetes dictionary. For this demo, I will be using Google Kubernetes Engine. Also, instead of using a default ingress controller that GCP makes of its own, I will be creating an NGINX ingress controller which will be used by the Ingress resource. Using this NGINX ingress controller we will be allowing IP addresses and eventually blocking others from accessing our application running in GKE. Before we start with the implementation, let us get some of our prerequisites revised.