Skip to Content
Skip Breadcrumb

In most enterprise environments are a lot systems separated via firewalls and the CSO does not want to allow the whole PaaS to access to the external Services.


With the described egress solutions in this blog is it possible to keep the separation AND all benefits of the PaaS can be used.

  • High Availability
  • Scale out
  • Reproducible images/router

In the old days, around one year ago, was no egress router part in OpenShift.

Nowadays you have several options with different techniques.
The egress router is as a dedicated outgoing border to destinations outside of the PaaS.

This schema shows a possible setup.


Current options

I will describe the following solutions:

  • builtin via iproute2 package
  • builtin via squid
  • external via a generic haproxy image
  • Your own Solution. Out of scope

Okay what’s the difference ❓

Solution Multi home configmap Environment handycaps Source Documentation
iproute2 partly partly based on macvlan
squid partly partly
haproxy ✔️ full partly requires knowledge about haproxy

Every solution have there valid use case.

Node selector

All above solutions can be placed on dedicated node with a “nodeSelector” Assigning Pods to Specific Nodes in the deploymentconfig a project.

❗ | A dedicated “nodeSelector” is only possible when the project “annotations” key “” is set to a empty string

iproute2 (kernel based router)

Red Hat calls this redirect mode and documented it at Deploying an Egress Router Pod in Redirect Mode.
This is the “easiest” Option and one of the fastest because the traffic is passed by the kernel.

There are some pitfalls when you use a this option so please read the hints very carefully Using an Egress Router to Allow External Resources to Recognize Pod Traffic

The default configuration and script is not prepared for a multi-homed and multi-route solution.
If you want to use another router-script in then Image then you will need to rebuild the Image from the existing Solution.

⚠️ | The egress router and the app MUST be in the same VNID (project/namespace)

squid (http proxy)

Red Hat call this http-proxy mode and documented it at Deploying an Egress Router HTTP Proxy Pod.

The default configuration and script is not prepared for a multi-homed solution. If you want to use another router-script in then Image then you will need to rebuild the Image from the existing Solution one you can build your own Solution.

This egress router CAN be in another project because it’s just a “normal” Service with a running pod.
This option can easily be used as a generic http proxy with white- an black- lists acls.

For the full syntax of the acl please take a look into the squid reference manual

haproxy (generic proxy)

This option is a generic tcp proxy based on haproxy 1.8.
You can use this solution like any other app due to the fact that I use the standard features from OpenShift.

I have created a Docker image which you can use out of the box.
haproxy18 which is based on this source haproxy18-centos.

There is also a Dockerfile for rhel7 which you can use to build your own haproxy with RHEL7.


This Image have the following componentes.

The version in place is visible in the Dockerfile


HAProxy have the possibility to use LUA scripts since version 1.6.
LUA can be used at several stages. I have add this feature here just in case I will need it.


I have described the Socklog in this blog post syslog in a container world

haproxy 1.8

The main component for this egress solution. I strongly suggest to read the detailed Dokumentation, due to the huge amount of feautures in haproxy.


The haproxy_exporter offers a interface to prometheus to get the haproxy statistics into prometheus.

Setup in Openshift

I describe here the following scenario.


I create a new project for the example, in case you can’t create a new project then you can use the current one.


  • Internet Access
  • docker hub is useable
  • default domain is set

In short

Here is a short example which you can copy and paste, if the prerequirement are fulfilled.

# oc new-app test-haproxy
# oc process -f \
    -p PROXY_SERVICE=test-scraper \
    -p SERVICE_NAME=tst-scr-svc \
    -p SERVICE_TCP_PORT=8443 \
    -p SERVICE_DEST_PORT=443 \
    -p \
    | oc create -f -
# oc rsh -c test-scraper $( oc get po -o jsonpath='{.items[*] }')
sh-4.2$ curl -v  http://test-scraper:${SERVICE_TCP_PORT}

logs from haproxy

oc logs -c test-scraper $( oc get po -o jsonpath='{.items[*] }') -f
00000001:public_tcp.accept(0007)=0009 from []

logs from haproxy-exporter

oc logs -c haproxy-exporter $( oc get po -o jsonpath='{.items[*] }') -f

time="2017-09-28T17:20:01+02:00" level=info msg="Starting haproxy_exporter (version=0.8.0, branch=HEAD, revision=4ce06e84e1701827f2706fd58b1e1320a52e3967)" source="haproxy_exporter.go:476"
time="2017-09-28T17:20:01+02:00" level=info msg="Build context (go=go1.8.3, user=root@27187aec7434, date=20170824-21:39:12)" source="haproxy_exporter.go:477"
time="2017-09-28T17:20:01+02:00" level=info msg="Listening on :9101" source="haproxy_exporter.go:502"

Prometheus Stats

The template creates a route test-scraper which is reachable as any other route in the cluster. You can configure now the prometheus to scrape the metrics from the generic router.


  • 2018-11-06 update to haproxy 1.8 links and images

You can contact me for any further questions and orders