Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Parallelization Within Tasks #25

Open
davidcereal opened this issue Aug 27, 2020 · 0 comments
Open

Parallelization Within Tasks #25

davidcereal opened this issue Aug 27, 2020 · 0 comments

Comments

@davidcereal
Copy link

davidcereal commented Aug 27, 2020

Hello,

Is there is any local Joblib parallelization possible within parallel_backend. Consider the following slightly modified example from the README. The only thing changed is the SVC estimator is instantiated with n_jobs=2.

register_spark() # register spark backend

clf = svm.SVC(kernel='linear', C=1, n_jobs=2)

with parallel_backend('spark', n_jobs=3):
  scores = cross_val_score(clf, iris.data, iris.target, cv=5)

If there are 2 cores per task (spark.task.cpus=2), would the SVC estimation be parallelized across 2 cores within each of the 3 tasks? If not, is there any way to achieve this?

Thanks in advance!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant