Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Verify CP Subsystem usage on Kubernetes #262

Closed
leszko opened this issue Oct 16, 2020 · 3 comments · Fixed by #289
Closed

Verify CP Subsystem usage on Kubernetes #262

leszko opened this issue Oct 16, 2020 · 3 comments · Fixed by #289

Comments

@leszko
Copy link

leszko commented Oct 16, 2020

No description provided.

@leszko
Copy link
Author

leszko commented Dec 29, 2020

I checked and currently, Hazelcast CP Subsystem does not work correctly in Kubernetes. This is because the assumption of CP Subsystem is that the Hazelcast cluster is stable (stable IP addresses) and if any member is restarted it should be manually removed from the cluster as described in Hazelcast Reference Manual: CP Subsystem’s Fault Tolerance Capabilities. In Kubernetes however, restarting PODs is an automatic (and expected) operation.

I tested what happens with the enabled CP Subsystem when the rolling upgrade is performed. The member fails with the following logs and CP Sybsystem Group never recovers.

@sancar
Copy link

sancar commented Dec 29, 2020

Hi Rafal, Did you enable CP Persistence as I suggested here to the user hazelcast/hazelcast#15299 (comment) ?

@leszko
Copy link
Author

leszko commented Dec 29, 2020

I tried it now, and actually, yes, it solves the issue. I updated the doc here: #289

Additionally, we need to support it in the Helm Chart, created an issue here: hazelcast/charts#176

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants