ipman

Artifact Hub

IPMan - IPSec Connection Manager for Kubernetes

logo-512x512

IPMan is a Kubernetes operator that simplifies the management of IPSec connections, enabling secure communication between your Kubernetes workloads and the outside world. It automates the setup and configuration of IPSec VPN tunnels using StrongSwan, making it easy to expose pods to external networks securely.

What IPMan Does

Installation

Prerequisites

Installing with Helm

# Add the repository
helm repo add ipman https://dialohq.github.io/ipman

# Install the chart
helm install ipman ipman/ipman -n ipman-system --create-namespace

Usage

Step 1: Create a Secret with Pre-shared Key

IPMan requires a secret for IPSec authentication:

kubectl create secret generic ipsec-secret -n default --from-literal=example=yourpresharedkey

Step 2: Create a charon group

Charon groups can contain many IPSec connections. Usually a charon group will look like this:

apiVersion: ipman.dialo.ai/v1
kind: CharonGroup
metadata:
  name: charongroup1
  namespace: default
spec:
  hostNetwork: true
  nodeName: node1

Here we specify that the other side of the VPN connections points to an IP address assigned to a host interface on one of our nodes node1.

For example we could have a enp0s1 interface with an address 192.168.10.201 on node1 the next steps assume this is the case.

note: Even though we specify a nodeName here, it’s only for the public IP. Workload pods that will communicate through this VPN can be on any node. Consult the Architecture section for a visual explanation.

Step 3: Create an IPSecConnection

Create an IPSecConnection Custom Resource (CR) to establish a VPN connection:

apiVersion: ipman.dialo.ai/v1
kind: IPSecConnection
metadata:
  name: example-connection
  namespace: ipman-system
spec:
  name: "example"
  remoteAddr: 192.168.10.204
  remoteId: 192.168.10.204
  localAddr: 192.168.10.201
  localId: 192.168.10.201
  secretRef:
    name: "ipsec-secret"
    namespace: default
    key: "example"
  groupRef:
    name: charongroup1
    namespace: default
  children:
    example-child:
      name: "example-child"
      extra:
        esp_proposals: aes256-sha256-ecp256
        start_action: start
        dpd_action: restart
      local_ips:
        - "10.0.2.0/24"
      remote_ips:
        - "10.0.1.0/24"
      xfrm_ip: "10.0.2.1/24"
      vxlan_ip: "10.0.2.2/24"
      if_id: 101
      ip_pools:
        primary:
          - "10.0.2.3/24"
          - "10.0.2.4/24"

This CR looks a lot like StrongSwan configuration file, with following added fields:

  1. secretRef This is the substitute of secrets section of the StrongSwan config file. You point it at the secret created in step 1 which contains the PSK.
  2. groupRef links this connection to a group we defined in step 2.
  3. xfrm_ip and vxlan_ip These are largely arbitrary with the exception that they have to be in the subnet defined in local_ips. For most use cases you can choose them arbitrarily and make sure they don’t conflict between connections and you will be good to go.
  4. if_id This has to be unique within a single node since it specifies the ID of an xfrm interface, StrongSwan, and the Linux Kernel use to route IPSec packets.
  5. ip_pools This is the list of IPs which will be given out to pods that are supposed to be in the VPN. So again they have to be IPs defined in local_ips. They are split into pools. Here we name our pool primary but you can use any name. This helps when you share multiple services with the other side of the VPN. You may want to have a pool service1 and service2 and in each you would put IPs that the other side of the VPN expects these services to be at.

Step 3: Deploy Workloads Using the VPN Connection

To route workload traffic through the VPN tunnel, add specific annotations to your Pods or Deployments. These annotations tell IPMan to allocate IPs from the configured pools and set up the necessary routing.

apiVersion: apps/v1
kind: Deployment
metadata:
  name: example-app
spec:
  template:
    metadata:
      annotations:
        ipman.dialo.ai/childName: "example-child"            # Must match the child name in IPSecConnection
        ipman.dialo.ai/ipmanName: "example-connection"       # Must match the IPSecConnection name
        ipman.dialo.ai/poolName: "primary"                   # IP pool to use (defined in IPSecConnection)
    spec:
      # Your pod spec here

The operator will automatically:

  1. Allocate IPs from the specified pool
  2. Set up routing for your workloads
  3. Configure bridge FDB entries for communication

If your app requires a specific IP to bind to and you have multiple IPs in a pool you don’t necessarily know which pod will get which IP. To help with that there is an env var set in all worker pods named VXLAN_IP so in this example the pod could get the IP 10.0.2.3/24 from the pool and the env var will contain the value 10.0.2.3.

Configuration Reference

Troubleshooting

If you encounter issues with your IPSec connections:

  1. Check the IPSecConnection status:
    kubectl get ipsecconnection -n ipman-system
    kubectl describe ipsecconnection example-connection -n ipman-system
    
  2. Check the operator logs:
    kubectl logs -n ipman-system -l app=ipman-controller
    
  3. Verify pod annotations match the IPSecConnection configuration

Monitoring

IPMan now supports monitoring via Prometheus. To enable monitoring:

  1. Set global.monitoring.enabled to true in your Helm values
  2. Set global.monitoring.release to the name you’ve given your Prometheus operator (e.g., if installed via helm install kps prometheus-community/kube-prometheus-stack, set it to “kps”)

See helm/values.yaml for more configuration options.

Architecture

image

Description

IPMan ensures secure connectivity between remote sites and workload pods by injecting secondary interfaces tied to the local encryption domain’s network. Inbound traffic arrives at the host’s network interface, is forwarded through a dedicated XFRM interface, and routed within an isolated VXLAN segment for enhanced security and segmentation. Charon, the IKE daemon from StrongSwan, operates on user-specified nodes, with instance counts driven by the IPsec configuration. Each child connection runs in a dedicated pod, equipped with its own XFRM interface and VXLAN segment. This design enables flexible workload deployment across the cluster, abstracting the underlying physical infrastructure for seamless scalability. Only ports 500 (IKE) and 4500 (NAT-traversal/IPsec) are exposed for secure communication. Charon and the restctl service, which manage the Charon socket and XFRM interface configuration, operate within the host network namespace, exposing only sockets mounted in the host filesystem. Control sockets, accessible via proxies, facilitate cluster-wide management without requiring additional open ports.

Acknowledgments

Special thanks to LarsTi for his ipsec_exporter repository, which we’ve adapted for our use case.

TODO

License

This project is licensed under the MIT License.