Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add a param to enable spawning spark cluster pods on specific Kubernetes nodes #273

Open
kimcie opened this issue Jan 22, 2020 · 1 comment
Labels
enhancement New feature or request good first issue Good for newcomers

Comments

@kimcie
Copy link

kimcie commented Jan 22, 2020

Description:
Currently it is not possible to deploy spark cluster on selected Kubernetes nodes. All spark pods are scheduled randomly.

It would be nice if the spark-operator had a feature to choose Kubernetes nodes where to schedule spark cluster. I imagine this would involve adding new param to spark cluster config and then having the operator map this value for example to nodeSelector of spark pod manifests.

@jkremser
Copy link
Member

Hello Mikołaj, would you be interested in contributing this feature? It can be a nice first issue to nail as a new contributor and it shouldn't be super difficult as you've mentioned :) Adding the pod(Anti)Affinity field into spark cluster, probably for both master and worker, or just one global argument for both would be a good start. There is the json schema where it needs to be added and then Java classes are generated from those json schemas. Then the mapping from the java object to the fabric8 client "fluent" api.

@jkremser jkremser added enhancement New feature or request good first issue Good for newcomers labels Jan 22, 2020
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request good first issue Good for newcomers
Projects
None yet
Development

No branches or pull requests

2 participants