Scalable, Portable and Distributed Gradient Boosting (GBDT, GBRT or GBM) Library, for Python, R, Java, Scala, C++ and more. Runs on single machine, Hadoop, Spark, Flink and DataFlow

Installers

  • linux-ppc64le v1.5.0
  • osx-arm64 v1.5.0
  • linux-64 v1.5.0
  • linux-aarch64 v1.5.0
  • linux-s390x v1.5.0
  • osx-64 v1.5.0
  • win-64 v1.5.0

conda install

To install this package run one of the following:
conda install -c anaconda py-xgboost

Description

XGBoost is an optimized distributed gradient boosting library designed to be highly efficient, flexible and portable. It implements machine learning algorithms under the Gradient Boosting framework. XGBoost provides a parallel tree boosting (also known as GBDT, GBM) that solve many data science problems in a fast and accurate way. The same code runs on major distributed environment (Hadoop, SGE, MPI) and can solve problems beyond billions of examples.


© 2022 Anaconda, Inc. All Rights Reserved. (v2.37.2 bfd12c9f) Legal | Privacy Policy