You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Inspecting system
Inspecting Certificates
Inspecting services
Service snap.microk8s.daemon-cluster-agent is running
Service snap.microk8s.daemon-containerd is running
Service snap.microk8s.daemon-kubelite is running
Service snap.microk8s.daemon-k8s-dqlite is running
Service snap.microk8s.daemon-apiserver-kicker is running
Copy service arguments to the final report tarball
Inspecting AppArmor configuration
Gathering system information
Copy processes list to the final report tarball
Copy disk usage information to the final report tarball
Copy memory usage information to the final report tarball
Copy server uptime to the final report tarball
Copy openSSL information to the final report tarball
Copy snap list to the final report tarball
Copy VM name (or none) to the final report tarball
Copy current linux distribution to the final report tarball
Copy asnycio usage and limits to the final report tarball
Copy inotify max_user_instances and max_user_watches to the final report tarball
Copy network configuration to the final report tarball
Inspecting kubernetes cluster
Inspect kubernetes cluster
Inspecting dqlite
Inspect dqlite
cp: cannot stat '/var/snap/microk8s/6364/var/kubernetes/backend/localnode.yaml': No such file or directory
WARNING: Maximum number of inotify user watches is less than the recommended value of 1048576.
Increase the limit with:
echo fs.inotify.max_user_watches=1048576 | sudo tee -a /etc/sysctl.conf
sudo sysctl --system
Building the report tarball
Report tarball is at /var/snap/microk8s/6364/inspection-report-20231217_102515.tar.gz
Can you suggest a fix?
Are you interested in contributing with a fix?
The text was updated successfully, but these errors were encountered:
Having this issue also. Fresh install of 22.04 + microk8s 1.29/stable from snap... here are the odd bits: removing the microk8s snap and reinstalling it resulted in the same problem before installing the host-access addon, but notably I did not explicitly disable the add-on before removing the snap--therefore I am inferring this is doing something somewhere to the host networking stack? Further, after I figured out what was going on, I disabled host-access, removed 1.29/stable, and then installed 1.28/stable... now when I enable host-access it is doing the same thing on 1.28/stable.
If this is helpful:
user@myhost:~$ microk8s kubectl get node -o wide
NAME STATUS ROLES AGE VERSION INTERNAL-IP EXTERNAL-IP OS-IMAGE KERNEL-VERSION CONTAINER-RUNTIME
myhost Ready <none> 174m v1.28.3 10.0.2.1 <none> Ubuntu 22.04.3 LTS 5.15.0-92-generic containerd://1.6.15
EDIT: should have mentioned this was a after sudo microk8s enable host-access:ip=10.0.2.1
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.
Summary
Failed to check container logs when host-access is enabled on microk8s 1.29/stable on ubuntu 22 due to below error
Error from server: Get "https://10.0.1.1:10250/containerLogs/container-registry/registry-6c9fcc695f-bkrbl/registry?follow=true": tls: failed to verify certificate: x509: certificate is valid for 192.168.1.14 not 10.0.1.1
Process finished with exit code 1
What Should Happen Instead?
I should be able to check logs. The same thing works on microk8s snap 1.27/stable channel.
Reproduction Steps
Error from server: Get "https://10.0.1.1:10250/containerLogs/container-registry/registry-6c9fcc695f-bkrbl/registry?follow=true": tls: failed to verify certificate: x509: certificate is valid for 192.168.1.14, 172.29.0.1, 172.18.0.1, 172.20.0.1, 172.21.0.1, 172.17.0.1, 2401:4900:1c71:9f73:1bb0:83dd:e889:c0be, 2401:4900:1c71:9f73:911e:b5e7:aaff:782e, not 10.0.1.1
Process finished with exit code 1
Introspection Report
Inspecting system
Inspecting Certificates
Inspecting services
Service snap.microk8s.daemon-cluster-agent is running
Service snap.microk8s.daemon-containerd is running
Service snap.microk8s.daemon-kubelite is running
Service snap.microk8s.daemon-k8s-dqlite is running
Service snap.microk8s.daemon-apiserver-kicker is running
Copy service arguments to the final report tarball
Inspecting AppArmor configuration
Gathering system information
Copy processes list to the final report tarball
Copy disk usage information to the final report tarball
Copy memory usage information to the final report tarball
Copy server uptime to the final report tarball
Copy openSSL information to the final report tarball
Copy snap list to the final report tarball
Copy VM name (or none) to the final report tarball
Copy current linux distribution to the final report tarball
Copy asnycio usage and limits to the final report tarball
Copy inotify max_user_instances and max_user_watches to the final report tarball
Copy network configuration to the final report tarball
Inspecting kubernetes cluster
Inspect kubernetes cluster
Inspecting dqlite
Inspect dqlite
cp: cannot stat '/var/snap/microk8s/6364/var/kubernetes/backend/localnode.yaml': No such file or directory
WARNING: Maximum number of inotify user watches is less than the recommended value of 1048576.
Increase the limit with:
echo fs.inotify.max_user_watches=1048576 | sudo tee -a /etc/sysctl.conf
sudo sysctl --system
Building the report tarball
Report tarball is at /var/snap/microk8s/6364/inspection-report-20231217_102515.tar.gz
Can you suggest a fix?
Are you interested in contributing with a fix?
The text was updated successfully, but these errors were encountered: