-
Notifications
You must be signed in to change notification settings - Fork 282
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Cilium Hubble Server not initialize after restore etcd snapshot #5122
Comments
That's not a restore from snapshot; all you did was reset the etcd cluster membership to a single node. Did you want to actually restore from a snapshot? |
rke2 server I also run the command above, is it the right way to restore cluster from etcd snapshot? |
Yes, restoring from a snapshot requires passing the path to the snapshot to restore, or the filename if using s3. Once it finishes, you should get additional instructions on what to do on the other servers to rejoin them. |
Thanks for your answer, I follow these step but still got the error related to cilium Hubble. The least option should be reinstall the whole cluster, right? |
That seems like overkill... have you looked at logs for all the containers in that pod? The error indicates that there is another prior failure that you need to resolve. Something else is failing to create that socket file. |
Thanks for your response. I reinstalled Cilium using Helm, and it worked. |
I encountered a similar issue after rejoining a worker node to the Kubernetes cluster.
after my cilium status is okay |
Environmental Info:
RKE2 Version:
rke2 version v1.26.9+rke2r1
Node(s) CPU architecture, OS, and Version:
5.4.0-167-generic #184-Ubuntu SMP Tue Oct 31 09:21:49 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux
Cluster Configuration:
1 master - 8 workers
Describe the bug:
![image](https://private-user-images.githubusercontent.com/75193335/289236506-df50207f-6482-4dbd-98ec-02896a3fedf8.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3Mzk0NjEwNDksIm5iZiI6MTczOTQ2MDc0OSwicGF0aCI6Ii83NTE5MzMzNS8yODkyMzY1MDYtZGY1MDIwN2YtNjQ4Mi00ZGJkLTk4ZWMtMDI4OTZhM2ZlZGY4LnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNTAyMTMlMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjUwMjEzVDE1MzIyOVomWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPWQ0YTZhOGQ3Njk5ZDNmMDc3ZmJlNTM0MDQzNzExMjNjZDU4ZWUxN2NmNzY1MGRkNjI1MWNmYzVhMjUyMDljNjAmWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0In0.C4IRZ9BpZCDGNr90lEOZAS8JdL3pzZlFt3P8dhzgHis)
Cilium Hubble Server not initialize after I restore etcd snapshot
Steps To Reproduce:
I got some problem with master node, so that I restore etcd config from etcd snapshot. I run the following commands:
systemctl stop rke2-server
rke2 server --cluster-reset
systemctl start rke2-server
After that, I check cilium status and encounter with those warning
![image](https://private-user-images.githubusercontent.com/75193335/289236678-eba8904f-ee9a-4173-9385-32bffdc4812b.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3Mzk0NjEwNDksIm5iZiI6MTczOTQ2MDc0OSwicGF0aCI6Ii83NTE5MzMzNS8yODkyMzY2NzgtZWJhODkwNGYtZWU5YS00MTczLTkzODUtMzJiZmZkYzQ4MTJiLnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNTAyMTMlMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjUwMjEzVDE1MzIyOVomWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPTVjN2QyNDEyZjQ0YzhkMDBkYzM1ZTRiNjlkMzgzMmE2MmYzODljMzYyNDdhMWI1MmQ2NzI4YTI4ZWY0MjcxNGYmWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0In0.oWOL4_ACu_WNMd_niG1uCRKXr3wrhfH5MDsQ_pzvY8U)
Expected behavior:
Everything work well especially CNI.
Actual behavior:
Cilium Hubble Server not initialize
Additional context / logs:
Check Hubble status in Cillium pod's log:
Cilium pod's log:
The text was updated successfully, but these errors were encountered: