Performance considerations for NodePort vs. ClusterIP vs. Headless Service on Kubernetes Performance considerations for NodePort vs. ClusterIP vs. Headless Service on Kubernetes kubernetes kubernetes

Performance considerations for NodePort vs. ClusterIP vs. Headless Service on Kubernetes


I've put more detailed info on the each of the connection forwarding types and how the services are forwarded down under the headings belowfor context to my answers.

If I understanding the networking correctly, it seems that a Headless Service requires less hops and would hence be (slightly) faster?

Not substantially faster. The "extra hop" is the packet traversing local lookup tables which it traverses anyway so not a noticeable difference. The destination pod is still going to be the same number of actual network hops away.

If you have 1000's of services that run on a single pod and could be headless then you might use that to limit the number of iptables NAT rules and speed rule processing up (see iptables v ipvs below).

Is < a headless service not load balanced > correct? And would this still hold when called through the external (or internal) ALB?

Yes it is correct, the client (or ALB) would need to implement the load balancing across the Pod IP's.

Is there any difference in performance for NodePort vs ClusterIP?

A NodePort has a possible extra network hop from the entry node to the node running the pod. Assuming the ClusterIP ranges are routed to the correct node (and routed at all)

If you happen to be using a service type: LoadBalancer this behaviour can change by setting [.spec.externalTrafficPolicy to Local][https://kubernetes.io/docs/concepts/services-networking/service/#aws-nlb-support] which means traffic will only be directed to a local pod.

Finally, what is the most elegant/performant way of using internal services from outside of the cluster

I would say use the AWS ALB Ingress Controller with the alb.ingress.kubernetes.io/target-type: ip annotation. The k8s config from the cluster will be pushed out to the ALB via the ingress controller and address pods directly without traversing any connection forwarding or extra hops. All cluster reconfig will be automatically pushed out.

There is a little bit of latency for config to get to the ALB compared to cluster kube-proxy reconfiguration. Something like a rolling deployment might not be as seamless as the updates arrive after a pod is gone. The ALB's are equipped to handle the outage themselves, eventually.

Kubernetes Connection Forwarding

There is a kube-proxy process running on each node which manages how and where connections are forwared. There are 3 options for how kube-proxy does that: Userspace proxy, iptables or IPVS. Most clusters will be on iptables and that will cater for the vast majority of use cases.

Userspace proxy

The forwarding is via a process that runs in userspace to terminate and forward the connections. It's slow. It's unlikely you are using it, don't use it.

iptables

iptables forwards connections in kernel via NAT, which is fast. This is most common setup and will cover 90% of use cases. New connections are shared evenly between all nodes running pods for a service.

IPVS

Runs in kernel, it is fast and scalable. If you shift a traffic to a large number of apps this might improve the forwarding performance. It also supports different service load balancing modes:

- rr: round-robin- lc: least connection (smallest number of open connections)- dh: destination hashing- sh: source hashing- sed: shortest expected delay- nq: never queue

Access to services

My explanations are iptables based as I haven't done much detailed work with ipvs clusters yet. I'm gonna handwave the ipvs complexity away and say it's basically the same as iptables, just with faster rule processing as the number of rules increases on huge clusters (i.e number of pods/services/network policies).

I'm also ignoring the userspace proxy in the description, due to the overhead just don't use it.

The basic thing to understand is a "Service ClusterIP" is a virtual construct in the cluster that only exists as rule for where the traffic should go. Every node maintains this rule mapping of all ClusterIP/port to PodIP/port (via kube-proxy)

Nodeport

ALB routes to any node, The node/nodeport forwards the connection to a pod handling the service. This could be a remote pod which would involve sending traffic back out over the "wire".

ALB > wire > Node > Kernel Forward to SVC ( > wire if remote node ) > Pod

ClusterIP

Using the ClusterIP for direct access depends on the Service cluster IP ranges being routed to the correct node. Sometimes they aren't routed at all.

ALB > wire > Node > Kernel Forward to SVC > Pod

The "Kernel Forward to SVC" step can be skipped with an ALB annotation without using a headless service.

Headless Service

Again, Pod IP's aren't always addressable from outside the cluster depending on the network setup. You should be fine on EKS.

ALB > wire > Node > Pod

Note

I'll suffix this with requests are probably looking at < 1ms of additional latency if a connection is forwarded to a node in a VPC. Enhanced networking instances at the low end of that. Inter availability-zone comms might be a tad higher than intra-AZ. If you happened to have a geographically separated cluster it might increase the importance of controlling traffic flow. For example having a tunnelled calico network that actually jumped over a number of real networks.


what is the most elegant/performant way of using internal services from outside of the cluster (where we don't have access to the Kubernetes DNS) but within the same VPC?

For this to achieve, I think you should have a look at a Service Mesh. For example, Istio(https://istio.io). It handles your internal service calls manually so that the call doesn't have to go through Kubernetes DNS. Please have a look at Istio's docs (https://istio.io/docs) for more info.

Also, you can have a look at Istio at EKS (https://aws.amazon.com/blogs/opensource/getting-started-istio-eks)


Headless service will not have any load balancing at L4 layer but if you use it behind an ALB you are getting load balancing at L7 layer.

Nodeport internally uses cluster IP but because your request may randomly be routed to a pod on another host when it could have been routed to a pod on the same host, avoiding that extra hop out to the network. Nodeport is generally a bad idea for production usage.

IMHO best way to access internal services from outside of the cluster will be using ingress.

You can use nginx as ingress controller where you deploy the nginx ingress controller on your cluster and expose it via a LoadBalancer type service using ALB. Then you can configure path or host based routing using ingress api to route traffic between backend kubernetes services.