-
Notifications
You must be signed in to change notification settings - Fork 92
Avoid hard-coded values for API server address #288
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: master
Are you sure you want to change the base?
Conversation
Avoid hard-coded value of 127.0.0.1 and 6443, as the API server can listen on another address when used outside of k3s. When used from another go program using the `chart` package, host and ports should be passed as argument. When used standalone, host and port should be derived from the kubeconfig and/or master-url options.
| return "", "", err | ||
| } | ||
|
|
||
| return u.Hostname(), u.Port(), nil |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
You cannot assume that the in-cluster apiserver port is the same as the actual port on the host. For example, on both K3s and RKE2, the port for the in-cluster kubernetes service is 443, and that's what you'll see here if you look at the client config - but the apiserver is actually listening on 6443 on the server nodes.
You need to understand the difference between service ports, and service endpoints. Bootstrap charts need to access the local apiserver endpoint directly, as kube-proxy and the rest of the CNI cannot be relied on to be functional yet.
root@systemd-node-001:/# kubectl get service kubernetes -o wide
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE SELECTOR
kubernetes ClusterIP 10.43.0.1 <none> 443/TCP 21h <none>
root@systemd-node-001:/# kubectl get endpointslice kubernetes -o wide
NAME ADDRESSTYPE PORTS ENDPOINTS AGE
kubernetes IPv4 6443 172.17.0.4 21hThere was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes, I understood that.
I tried the patch with minikube and k3s by using helm-controller outside the cluster. In that case it worked since the kube-config file is set with direct api endpoint and not the in-cluster service.
I forgot to try the in-cluster image. I suppose in that case the code will use the in-cluster service and fail.
I will try and report back.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes, doing this the "right way" such that it will work in all possible environments is non-trivial, which is why I opted to just fix it the easy way in K3s.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I couldn’t find a use-case where the helm controller can connect to the API but not the helm bootstrap chart pod.
If the helm controller is using the in-cluster api and has access to the custom resources, then the pod running the bootstrap chart will be able to connect as well to the in-cluster API.
If the helm controller is configured to connect to the in-cluster API and the in-cluster API is not available, then the helm controller will never see the custom resource and will never start any pod anyways.
So using the in-cluster API for bootstrap charts when helm controller is using in-cluster API shouldn’t be a problem?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
helm-controller runs embedded in the K3s and RKE2 supervisor process, and always talks to the local apiserver endpoint, not the in-cluster service endpoint.
Users can also deploy it as a daemonset that runs with host network on control-plane nodes, and point it to the local apiserver endpoint. This would be necessary if it was to be used for bootstrap charts on other Kubernetes distributions.
Running it against the in-cluster service endpoint will never work for bootstrap charts, as you cannot rely on the cluster network being up yet.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
helm-controller runs embedded in the K3s and RKE2 supervisor process, and always talks to the local apiserver endpoint, not the in-cluster service endpoint.
That use-case is not related to the PR, as host and port will be given by the supervisor process.
Users can also deploy it as a daemonset that runs with host network on control-plane nodes, and point it to the local apiserver endpoint. This would be necessary if it was to be used for bootstrap charts on other Kubernetes distributions.
In that case, the PR works since bootstrap charts will re-use host and port of local api server endpoint given by the user.
Running it against the in-cluster service endpoint will never work for bootstrap charts, as you cannot rely on the cluster network being up yet.
The only case in which the PR will use in-cluster service endpoint for bootstrap charts is if helm-controller is using in-cluster service endpoint to get the custom resource.
Either network is up, helm-controller can get custom resources, and bootstrap charts will work using in-cluster api.
Or network is down, helm-controller can’t get custer resources, and no bootstrap charts will ever be started.
I’ll make more tests and report logs, but I still don’t see what I’ve missed.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I couldn’t find a failing use-case. See below for configuration I tested and logs.
Could you provide a way to make the PR fail? I have no problem updating the PR to use hard-coded values in specific situations, but I’d rather limit using 127.0.0.1 to situations where re-using host and port from the controller actually fail.
Test 1: embedded in k3s
curl -sfL https://get.k3s.io | sh -s - server --cluster-cidr=2001:db8:42::/56 --service-cidr=2001:db8:43::/112 --flannel-backend=none
sudo systemctl stop k3s
sudo cp bin/k3s /usr/local/bin/
sudo systemctl start k3s
sudo kubectl --kubeconfig /etc/rancher/k3s/k3s.yaml apply -f - <<EOF
apiVersion: helm.cattle.io/v1
kind: HelmChart
metadata:
name: apache
namespace: kube-system
spec:
repo: https://charts.bitnami.com/bitnami
chart: apache
targetNamespace: kube-system
bootstrap: true
valuesContent: |-
service:
type: ClusterIP
ingress:
enabled: true
hostname: www.example.com
metrics:
enabled: true
EOF
sudo kubectl --kubeconfig /etc/rancher/k3s/k3s.yaml describe pod -n kube-system helm-install-apache | grep KUBERNETES_SERVICE
KUBERNETES_SERVICE_HOST: ::1
KUBERNETES_SERVICE_PORT: 6444
sudo kubectl --kubeconfig /etc/rancher/k3s/k3s.yaml logs -n kube-system -l helmcharts.helm.cattle.io/chart=apache --tail=50 | grep STATUS:
STATUS: deployed
Test 2: out of cluster
curl -sfL https://get.k3s.io | sh -s - server --cluster-cidr=2001:db8:42::/56 --service-cidr=2001:db8:43::/112 --flannel-backend=none --disable-helm-controller
sudo bin/helm-controller --kubeconfig /etc/rancher/k3s/k3s.yaml
sudo kubectl --kubeconfig /etc/rancher/k3s/k3s.yaml apply -f - <<EOF
apiVersion: helm.cattle.io/v1
kind: HelmChart
metadata:
name: apache
namespace: kube-system
spec:
repo: https://charts.bitnami.com/bitnami
chart: apache
targetNamespace: kube-system
bootstrap: true
valuesContent: |-
service:
type: ClusterIP
ingress:
enabled: true
hostname: www.example.com
metrics:
enabled: true
EOF
sudo kubectl --kubeconfig /etc/rancher/k3s/k3s.yaml describe pod -n kube-system helm-install-apache | grep KUBERNETES_SERVICE
KUBERNETES_SERVICE_HOST: ::1
KUBERNETES_SERVICE_PORT: 6443
sudo kubectl --kubeconfig /etc/rancher/k3s/k3s.yaml logs -n kube-system -l helmcharts.helm.cattle.io/chart=apache --tail=50 | grep STATUS:
STATUS: deployed
Test 3: in-cluster helm-controller using local api server endpoint
curl -sfL https://get.k3s.io | sh -s - server --cluster-cidr=2001:db8:42::/56 --service-cidr=2001:db8:43::/112 --flannel-backend=none --disable-helm-controller
sudo kubectl --kubeconfig /etc/rancher/k3s/k3s.yaml apply -f - <<EOF
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: helm-controller
rules:
- apiGroups:
- '*'
resources:
- '*'
verbs:
- '*'
---
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
name: helm-controller
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: ClusterRole
name: helm-controller
subjects:
- kind: ServiceAccount
name: helm-controller
namespace: kube-system
---
apiVersion: v1
kind: ServiceAccount
metadata:
name: helm-controller
namespace: kube-system
---
apiVersion: apps/v1
kind: Deployment
metadata:
name: helm-controller
namespace: kube-system
labels:
app: helm-controller
spec:
replicas: 1
selector:
matchLabels:
app: helm-controller
template:
metadata:
labels:
app: helm-controller
spec:
hostNetwork: true
containers:
- name: helm-controller
image: codeberg.org/jip149/helm-controller:dev
command: ["helm-controller"]
env:
- name: KUBECONFIG
value: /kubeconfig
- name: NODE_NAME
valueFrom:
fieldRef:
fieldPath: spec.nodeName
- name: JOB_CLUSTER_ROLE
valueFrom:
fieldRef:
fieldPath: spec.serviceAccountName
volumeMounts:
- mountPath: /kubeconfig
name: kubeconfig
volumes:
- name: kubeconfig
hostPath:
path: /etc/rancher/k3s/k3s.yaml
type: File
tolerations:
- effect: NoSchedule
key: node.kubernetes.io/not-ready
- effect: NoSchedule
key: node.kubernetes.io/network-unavailable
- effect: NoSchedule
key: node.cloudprovider.kubernetes.io/uninitialized
operator: Equal
value: "true"
- key: CriticalAddonsOnly
operator: Exists
- effect: NoExecute
key: node-role.kubernetes.io/etcd
operator: Exists
- effect: NoSchedule
key: node-role.kubernetes.io/control-plane
operator: Exists
serviceAccountName: helm-controller
EOF
sudo kubectl --kubeconfig /etc/rancher/k3s/k3s.yaml apply -f - <<EOF
apiVersion: helm.cattle.io/v1
kind: HelmChart
metadata:
name: apache
namespace: kube-system
spec:
repo: https://charts.bitnami.com/bitnami
chart: apache
targetNamespace: kube-system
bootstrap: true
valuesContent: |-
service:
type: ClusterIP
ingress:
enabled: true
hostname: www.example.com
metrics:
enabled: true
EOF
sudo kubectl --kubeconfig /etc/rancher/k3s/k3s.yaml describe pod -n kube-system helm-install-apache | grep KUBERNETES_SERVICE
KUBERNETES_SERVICE_HOST: ::1
KUBERNETES_SERVICE_PORT: 6443
sudo kubectl --kubeconfig /etc/rancher/k3s/k3s.yaml logs -n kube-system -l helmcharts.helm.cattle.io/chart=apache --tail=50 | grep STATUS:
STATUS: deployed
Test 4: in-cluster helm-controller using in-cluster api service
curl -sfL https://get.k3s.io | sh -s - server --cluster-cidr=2001:db8:42::/56 --service-cidr=2001:db8:43::/112 --flannel-backend=none --disable-helm-controller
sudo kubectl --kubeconfig /etc/rancher/k3s/k3s.yaml apply -f - <<EOF
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: helm-controller
rules:
- apiGroups:
- '*'
resources:
- '*'
verbs:
- '*'
---
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
name: helm-controller
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: ClusterRole
name: helm-controller
subjects:
- kind: ServiceAccount
name: helm-controller
namespace: kube-system
---
apiVersion: v1
kind: ServiceAccount
metadata:
name: helm-controller
namespace: kube-system
---
apiVersion: apps/v1
kind: Deployment
metadata:
name: helm-controller
namespace: kube-system
labels:
app: helm-controller
spec:
replicas: 1
selector:
matchLabels:
app: helm-controller
template:
metadata:
labels:
app: helm-controller
spec:
hostNetwork: true
containers:
- name: helm-controller
image: codeberg.org/jip149/helm-controller:dev
command: ["helm-controller"]
env:
- name: NODE_NAME
valueFrom:
fieldRef:
fieldPath: spec.nodeName
- name: JOB_CLUSTER_ROLE
valueFrom:
fieldRef:
fieldPath: spec.serviceAccountName
tolerations:
- effect: NoSchedule
key: node.kubernetes.io/not-ready
- effect: NoSchedule
key: node.kubernetes.io/network-unavailable
- effect: NoSchedule
key: node.cloudprovider.kubernetes.io/uninitialized
operator: Equal
value: "true"
- key: CriticalAddonsOnly
operator: Exists
- effect: NoExecute
key: node-role.kubernetes.io/etcd
operator: Exists
- effect: NoSchedule
key: node-role.kubernetes.io/control-plane
operator: Exists
serviceAccountName: helm-controller
EOF
sudo kubectl --kubeconfig /etc/rancher/k3s/k3s.yaml apply -f - <<EOF
apiVersion: helm.cattle.io/v1
kind: HelmChart
metadata:
name: apache
namespace: kube-system
spec:
repo: https://charts.bitnami.com/bitnami
chart: apache
targetNamespace: kube-system
bootstrap: true
valuesContent: |-
service:
type: ClusterIP
ingress:
enabled: true
hostname: www.example.com
metrics:
enabled: true
EOF
sudo kubectl --kubeconfig /etc/rancher/k3s/k3s.yaml describe pod -n kube-system helm-install-apache | grep KUBERNETES_SERVICE
KUBERNETES_SERVICE_HOST: 2001:db8:43::1
KUBERNETES_SERVICE_PORT: 6443
sudo kubectl --kubeconfig /etc/rancher/k3s/k3s.yaml logs -n kube-system -l helmcharts.helm.cattle.io/chart=apache --tail=50 | grep STATUS:
STATUS: deployed
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I made more tests, this time by disabling both flannel and kube-proxy.
This is now the behavior I described above. In-cluster api service cannot be reached, but the problem is not with the bootstrap chart, it’s just that helm-controller does not start (see logs below).
I still can’t find a situation where using a hard-coded value is better than connecting to host and port used by helm-controller.
Test 5: in-cluster helm-controller using in-cluster api service without kube-proxy
curl -sfL https://get.k3s.io | sh -s - server --cluster-cidr=2001:db8:42::/56 --service-cidr=2001:db8:43::/112 --flannel-backend=none --disable-helm-controller --disable-kube-proxy
sudo kubectl --kubeconfig /etc/rancher/k3s/k3s.yaml apply -f - <<EOF
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: helm-controller
rules:
- apiGroups:
- '*'
resources:
- '*'
verbs:
- '*'
---
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
name: helm-controller
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: ClusterRole
name: helm-controller
subjects:
- kind: ServiceAccount
name: helm-controller
namespace: kube-system
---
apiVersion: v1
kind: ServiceAccount
metadata:
name: helm-controller
namespace: kube-system
---
apiVersion: apps/v1
kind: Deployment
metadata:
name: helm-controller
namespace: kube-system
labels:
app: helm-controller
spec:
replicas: 1
selector:
matchLabels:
app: helm-controller
template:
metadata:
labels:
app: helm-controller
spec:
hostNetwork: true
containers:
- name: helm-controller
image: codeberg.org/jip149/helm-controller:dev
command: ["helm-controller"]
env:
- name: NODE_NAME
valueFrom:
fieldRef:
fieldPath: spec.nodeName
- name: JOB_CLUSTER_ROLE
valueFrom:
fieldRef:
fieldPath: spec.serviceAccountName
tolerations:
- effect: NoSchedule
key: node.kubernetes.io/not-ready
- effect: NoSchedule
key: node.kubernetes.io/network-unavailable
- effect: NoSchedule
key: node.cloudprovider.kubernetes.io/uninitialized
operator: Equal
value: "true"
- key: CriticalAddonsOnly
operator: Exists
- effect: NoExecute
key: node-role.kubernetes.io/etcd
operator: Exists
- effect: NoSchedule
key: node-role.kubernetes.io/control-plane
operator: Exists
serviceAccountName: helm-controller
EOF
sudo kubectl --kubeconfig /etc/rancher/k3s/k3s.yaml logs -n kube-system -l app=helm-controller
time="2025-10-18T05:46:40Z" level=warning msg="unable to list existing CRDs: failed to list crds: Get \"https://[2001:db8:43::1]:443/apis/apiextensions.k8s.io/v1/customresourcedefinitions\": dial tcp [2001:db8:43::1]:443: connect: network is unreachable"
time="2025-10-18T05:46:40Z" level=info msg="Creating embedded CRD helmcharts.helm.cattle.io"
time="2025-10-18T05:46:40Z" level=fatal msg="failed to create crd 'helmcharts.helm.cattle.io': Post \"https://[2001:db8:43::1]:443/apis/apiextensions.k8s.io/v1/customresourcedefinitions\": dial tcp [2001:db8:43::1]:443: connect: network is unreachable"
Test 6: in-cluster helm-controller using local api server endpoint without kube-proxy
curl -sfL https://get.k3s.io | sh -s - server --cluster-cidr=2001:db8:42::/56 --service-cidr=2001:db8:43::/112 --flannel-backend=none --disable-helm-controller --disable-kube-proxy
sudo kubectl --kubeconfig /etc/rancher/k3s/k3s.yaml apply -f - <<EOF
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
name: helm-controller
rules:
- apiGroups:
- '*'
resources:
- '*'
verbs:
- '*'
---
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
name: helm-controller
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: ClusterRole
name: helm-controller
subjects:
- kind: ServiceAccount
name: helm-controller
namespace: kube-system
---
apiVersion: v1
kind: ServiceAccount
metadata:
name: helm-controller
namespace: kube-system
---
apiVersion: apps/v1
kind: Deployment
metadata:
name: helm-controller
namespace: kube-system
labels:
app: helm-controller
spec:
replicas: 1
selector:
matchLabels:
app: helm-controller
template:
metadata:
labels:
app: helm-controller
spec:
hostNetwork: true
containers:
- name: helm-controller
image: codeberg.org/jip149/helm-controller:dev
command: ["helm-controller"]
env:
- name: KUBECONFIG
value: /kubeconfig
- name: NODE_NAME
valueFrom:
fieldRef:
fieldPath: spec.nodeName
- name: JOB_CLUSTER_ROLE
valueFrom:
fieldRef:
fieldPath: spec.serviceAccountName
volumeMounts:
- mountPath: /kubeconfig
name: kubeconfig
volumes:
- name: kubeconfig
hostPath:
path: /etc/rancher/k3s/k3s.yaml
type: File
tolerations:
- effect: NoSchedule
key: node.kubernetes.io/not-ready
- effect: NoSchedule
key: node.kubernetes.io/network-unavailable
- effect: NoSchedule
key: node.cloudprovider.kubernetes.io/uninitialized
operator: Equal
value: "true"
- key: CriticalAddonsOnly
operator: Exists
- effect: NoExecute
key: node-role.kubernetes.io/etcd
operator: Exists
- effect: NoSchedule
key: node-role.kubernetes.io/control-plane
operator: Exists
serviceAccountName: helm-controller
EOF
sudo kubectl --kubeconfig /etc/rancher/k3s/k3s.yaml apply -f - <<EOF
apiVersion: helm.cattle.io/v1
kind: HelmChart
metadata:
name: apache
namespace: kube-system
spec:
repo: https://charts.bitnami.com/bitnami
chart: apache
targetNamespace: kube-system
bootstrap: true
valuesContent: |-
service:
type: ClusterIP
ingress:
enabled: true
hostname: www.example.com
metrics:
enabled: true
EOF
sudo kubectl --kubeconfig /etc/rancher/k3s/k3s.yaml describe pod -n kube-system helm-install-apache | grep KUBERNETES_SERVICE
KUBERNETES_SERVICE_HOST: ::1
KUBERNETES_SERVICE_PORT: 6443
sudo kubectl --kubeconfig /etc/rancher/k3s/k3s.yaml logs -n kube-system -l helmcharts.helm.cattle.io/chart=apache --tail=50 | grep STATUS:
STATUS: deployed
Avoid hard-coded value of 127.0.0.1 and 6443, as the API server can listen on another address when used outside of k3s.
When used from another go program using the
chartpackage, host and ports should be passed as argument.When used standalone, host and port should be derived from the kubeconfig and/or master-url options.
I’m not sure how version number shoud be changed, since this breaks API for chart package.
It could also be relevant to add tests?
Related to k3s-io/k3s#12947 and #287