-
-
Notifications
You must be signed in to change notification settings - Fork 8.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[RFC] Version 0.82 release candidate #4201
Conversation
Any change in particular you'd like us to try testing @hcho3 ? |
One thing I am interested in is, how experimental is our external memory implementation? We can't mark it beta forever... |
@thvasilo I'm mainly soliciting feedback with regard to the release notes. After a short while, committers will approve the release. |
@trivialfis I'd like to see external memory implemented for |
LGTM. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks great!
|
@CodingCat Wait, I thought depthwise had fewer sync steps than lossguide? |
nvm, I am brain-damaged..... |
With 5 approvals, I am going to release 0.82. Thanks all! |
When would binary wheels for 0.82 be available? (especially multi-gpu) |
@hlbkin It is already available on PyPI. Run |
When will CRAN get a new release of the R package? |
There's a recent change in R-devel which causes failure in unittests.
I'll patch 0.82 with a fix and submit to CRAN before March 20th.
Best regards,
Tong He
Kodi Arfer <[email protected]> 于2019年3月8日周五 上午6:18写道:
… When will CRAN get a new release of the R package?
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
<#4201 (comment)>, or mute
the thread
<https://github.com/notifications/unsubscribe-auth/ABOugE2LmEOSRgZQK1JaPA7wnxz_Qhowks5vUnFMgaJpZM4bbzn5>
.
|
@Kodiologist 0.82 is now released on CRAN: https://cran.r-project.org/web/packages/xgboost/index.html |
v0.82 (2019.03.03)
This release is packed with many new features and bug fixes.
Roadmap: better performance scaling for multi-core CPUs (#3957)
hist
algorithm for multi-core CPUs has been under investigation (Call for contribution: improve multi-core CPU performance of 'hist' #3810). Performance optimizations for Intel CPUs #3957 marks an important step toward better performance scaling, by using software pre-fetching and replacing STL vectors with C-style arrays. Special thanks to @Laurae2 and @SmirnovEgorRu.New feature: Distributed Fast Histogram Algorithm (
hist
) (#4011, #4102, #4140, #4128)hist
algorithm in distributed setting. Special thanks to @CodingCat. The benefits include:approx
, allowing for future improvementNew feature: Multi-Node, Multi-GPU training (#4095)
New feature: Multiple validation datasets in XGBoost4J-Spark (#3904, #3910)
eval_sets
or callsetEvalSets
over aXGBoostClassifier
orXGBoostRegressor
, you can pass in multiple evaluation datasets typed as aMap
fromString
toDataFrame
. Special thanks to @CodingCat.New feature: Additional metric functions for GPUs (#3952)
rmse
,mae
,logloss
,poisson-nloglik
,gamma-deviance
,gamma-nloglik
,error
,tweedie-nloglik
. Special thanks to @trivialfis and @RAMitchell.n_gpus
parameter.New feature: Column sampling at individual nodes (splits) (#3971)
colsample_bynode
parameter, which represents the fraction of columns sampled at each node. This parameter is set to 1.0 by default (i.e. no sampling per node). Special thanks to @canonizer.colsample_bynode
parameter works cumulatively with othercolsample_by*
parameters: for example,{'colsample_bynode':0.5, 'colsample_bytree':0.5}
with 100 columns will give 25 features to choose from at each split.Major API change: consistent logging level via
verbosity
(#3982, #4002, #4138)verbosity
to 0 (silent), 1 (warning), 2 (info), and 3 (debug). This is useful for controlling the amount of logging outputs. Special thanks to @trivialfis.silent
anddebug_verbose
are now deprecated.Major bug fix: external memory (#4040, #4193)
Major bug fix: early stopping fixed in XGBoost4J and XGBoost4J-Spark (#3928, #4176)
earlyStoppingSteps
away from the best iteration. If there are multiple evaluation sets, only the last one is used to determinate early stop.Major bug fix: infrequent features should not crash distributed training (#4045)
Performance improvements
gpu_hist
(Improve update position function #3895)gpu_hist
(GPU performance logging/improvements #3945)Bug-fixes
gpu_id
when running multiple XGBoost processes on a multi-GPU machine (Fix specifying gpu_id, add tests. #3851)hist
(Prevent empty quantiles in fast hist #4155)hist
aware of query groups when running a ranking task (Make `HistCutMatrix::Init' be aware of groups. #4115). For ranking task, query groups are weighted, not individual instances.LOG(FATAL)
macro (Upgrade rabit. #4159)PATH
environment variable (handle $PATH not being set in python library #3845)coef_
andintercept_
signature to be compatible withsklearn.RFECV
(Fix coef_ and intercept_ signature to be compatible with sklearn.RFECV #3873)self.booster
attribute, for backward compatibility (Fix #3894: Allow loading pickles without self.booster attributes #3938, Fix #3894: Allow loading pickles without self.booster attributes (redux) #3944)rawPredictionCol
inXGBoostClassificationModel
([jvm-packages] Fix vector size of 'rawPredictionCol' in XGBoostClassificationModel #3932)setEvalSets
([jvm-packages] fix return type of setEvalSets #4105)getMaxLeaves
([jvm-packages] minor fix of params #4114)API changes
single_precision_histogram
to use single-precision histograms for thegpu_hist
algorithm (Single precision histograms on GPU #3965)trees_to_df()
method to dump decision trees as Pandas data frame (Added trees_to_df() method for Booster class #4153)xgb_model
option toXGBClassifier
, to load previously saved model (enable xgb_model in scklearn XGBClassifier and test. #4092)DMatrix
is now deprecated (Deprecation warning for lists passed into DMatrix #3970)Maintenance: Refactor C++ code for legibility and maintainability
hist
algorithm code and add unit tests (Add unittests for updaters. #3836)gpu_hist
(Minor refactor of split evaluation in gpu_hist #3889)TreeModel
andRegTree
classes (Combine TreeModel and RegTree #3995)gpu_exact
andgpu_coord_descent
(Use Span in GPU exact updater. #4020, Use Span in gpu coordinate. #4029)Maintenance: testing, continuous integration, build system
std::regex
since it's not supported by GCC 4.8.x (In lint check, disallow std::regex since it's not supported by GCC 4.8.x #3870)DeprecationWarning
when using Python collections (Address deprecation of Python ABC. #3909)hist
(Prevent empty quantiles in fast hist #4155)gpu_exact
algorithm (Use Span in GPU exact updater. #4020)gpu_hist
(Fix gpu_hist apply_split test. #4158)Usability Improvements
gblinear
is selected (Remove unnecessary warning when 'gblinear' is selected #3888)max_depth
parameter (fix doc about max_depth #4078)num_parallel_tree
(Document num_parallel_tree. #4022)Booster
object (Prevent training without setting up caches. #4066)benchmark_tree.py
to comply with Python style convention (reformat benchmark_tree.py to get rid of lint errors #4126)objectiveTrait
([jvm-packages] fix comments in objectiveTrait #4174)Acknowledgement
Contributors (in no particular order): Jiaming Yuan (@trivialfis), Hyunsu Cho (@hcho3), Nan Zhu (@CodingCat), Rory Mitchell (@RAMitchell), Yanbo Liang (@yanboliang), Andy Adinets (@canonizer), Tong He (@hetong007), Yuan Tang (@terrytangyuan)
First-time Contributors (in no particular order): Jelle Zijlstra (@JelleZijlstra), Jiacheng Xu (@jiachengxu), @ajing, Kashif Rasul (@kashif), @theycallhimavi, Joey Gao (@pjgao), Prabakaran Kumaresshan (@nixphix), Huafeng Wang (@huafengw), @lyxthe, Sam Wilkinson (@scwilkinson), Tatsuhito Kato (@stabacov), Shayak Banerjee (@shayakbanerjee), Kodi Arfer (@Kodiologist), @KyleLi1985, Egor Smirnov (@SmirnovEgorRu), @tmitanitky, Pasha Stetsenko (@st-pasha), Kenichi Nagahara (@keni-chi), Abhai Kollara Dilip (@abhaikollara), Patrick Ford (@pford221), @hshujuan, Matthew Jones (@mt-jones), Thejaswi Rao (@teju85), Adam November (@anovember)
First-time Reviewers (in no particular order): Mingyang Hu (@mingyang), Theodore Vasiloudis (@thvasilo), Jakub Troszok (@troszok), Rong Ou (@rongou), @Denisevi4, Matthew Jones (@mt-jones), Jeff Kaplan (@jeffdk)