Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[jvm-packages] [pyspark] Make QDM optional based on cuDF check #8471

Merged
merged 6 commits into from Nov 27, 2022
Merged
Show file tree
Hide file tree
Changes from 3 commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Jump to
Jump to file
Failed to load files.
Diff view
Diff view
13 changes: 13 additions & 0 deletions python-package/xgboost/compat.py
Expand Up @@ -43,6 +43,7 @@ def lazy_isinstance(instance: Any, module: str, name: str) -> bool:
pandas_concat = None
PANDAS_INSTALLED = False


# sklearn
try:
from sklearn.base import BaseEstimator as XGBModelBase
Expand Down Expand Up @@ -72,6 +73,18 @@ def lazy_isinstance(instance: Any, module: str, name: str) -> bool:
XGBStratifiedKFold = None


def is_cudf_installed():
"""Check cuDF installed or not"""
# Checking by `importing` instead of check `importlib.util.find_spec("cudf") is not None`
WeichenXu123 marked this conversation as resolved.
Show resolved Hide resolved
# because user might install cudf successfully but importing cudf raises issues (e.g. saying
# running on mismatched cuda version)
try:
import cudf
return True
except ImportError:
return False


class XGBoostLabelEncoder(LabelEncoder):
"""Label encoder with JSON serialization methods."""

Expand Down
10 changes: 9 additions & 1 deletion python-package/xgboost/spark/core.py
Expand Up @@ -37,6 +37,7 @@

import xgboost
from xgboost import XGBClassifier, XGBRanker, XGBRegressor
from xgboost.compat import is_cudf_installed

from .data import (
_read_csr_matrix_from_unwrapped_spark_vec,
Expand Down Expand Up @@ -755,7 +756,8 @@ def _fit(self, dataset):
k: v for k, v in train_call_kwargs_params.items() if v is not None
}
dmatrix_kwargs = {k: v for k, v in dmatrix_kwargs.items() if v is not None}
use_qdm = booster_params.get("tree_method", None) in ("hist", "gpu_hist")

use_hist = booster_params.get("tree_method", None) in ("hist", "gpu_hist")

def _train_booster(pandas_df_iter):
"""Takes in an RDD partition and outputs a booster for that partition after
Expand All @@ -769,6 +771,12 @@ def _train_booster(pandas_df_iter):

gpu_id = None

# If cuDF is not installed, then using DMatrix instead of QDM,
# because without cuDF, DMatrix performs better than QDM.
# Note: Checking `is_cudf_installed` in spark worker side because
# spark worker might has different python environment with driver side.
use_qdm = use_hist and is_cudf_installed()
WeichenXu123 marked this conversation as resolved.
Show resolved Hide resolved

if use_qdm and (booster_params.get("max_bin", None) is not None):
dmatrix_kwargs["max_bin"] = booster_params["max_bin"]

Expand Down
4 changes: 2 additions & 2 deletions python-package/xgboost/spark/data.py
Expand Up @@ -5,7 +5,7 @@
import numpy as np
import pandas as pd
from scipy.sparse import csr_matrix
from xgboost.compat import concat
from xgboost.compat import CUDF_INSTALLED, concat
WeichenXu123 marked this conversation as resolved.
Show resolved Hide resolved

from xgboost import DataIter, DMatrix, QuantileDMatrix

Expand Down Expand Up @@ -81,7 +81,7 @@ def _fetch(self, data: Optional[Sequence[pd.DataFrame]]) -> Optional[pd.DataFram
if not data:
return None

if self._device_id is not None:
if self._device_id is not None and CUDF_INSTALLED:
import cudf # pylint: disable=import-error
import cupy as cp # pylint: disable=import-error

Expand Down