- Versions 0.16.* had a bug in python applier with categorical features for applying on more than 128 documents.
- It is now possible to use pairwise modes for datasets without groups
- 1.8x Evaluation speed on asymmetrical trees
- Renamed column
Feature Index
toFeature Id
in prettified output of python methodget_feature_importance()
, because it supports feature names now - Renamed option
per_float_feature_binarization
(--per-float-feature-binarization
) toper_float_feature_quantization
(--per-float-feature-quantization
) - Removed parameter
inverted
from pythoncv
method. Addedtype
parameter instead, which can be set toInverted
- Method
get_features()
now works only for datasets without categorical features
- A new multiclass version of AUC metric, called
AUC Mu
, which was proposed by Ross S. Kleiman on NeurIPS 2019, link - Added time series cv
- Added
MeanWeightedTarget
infstat
- Added
utils.get_confusion_matrix()
- Now feature importance can be calculated for non-symmetric trees
- Removed
get_group_id()
andget_features()
methods ofPool
class
- Added
PredictionDiff
type ofget_feature_importance()
method, which is a new method for model analysis. The method shows how the features influenced the fact that among two samples one has a higher prediction. It allows to debug ranking models: you find a pair of samples ranked incorrectly and you look at what features have caused that. - Added
plot_predictions()
method
model.set_feature_names()
method in Python- Added stratified split to parameter search methods
- Support
catboost.load_model()
from CPU snapshots for numerical-only datasets CatBoostClassifier.score()
now supportsy
asDataFrame
- Added
sampling_frequency
,per_float_feature_binarization
,monotone_constraints
parameters toCatBoostClassifier
andCatBoostRegresssor
- 2x speedup of multi-classification mode
- Fixed
score()
for multiclassification, #924 - Fixed
get_all_params()
function, #926
- Clear error messages when a model cannot be saved
- parameter
fold_count
is now calledcv
ingrid_search()
andrandomized_search
- cv results are now returned from
grid_search()
andrandomized_search()
inres['cv_results']
field
- R-language function
catboost.save_model()
now supports PMML, ONNX and other formats - Parameter
monotone_constraints
in python API allows specifying numerical features that the prediction shall depend on monotonically
- Fixed
eval_metric
calculation for training with weights (in release 0.16 evaluation of a metric that was equal to an optimized loss did not use weights by default, so overfitting detector worked incorrectly)
- Added option
verbose
togrid_search()
andrandomized_search()
- Added tutorial on
grid_search()
andrandomized_search()
-
MultiClass
loss has now the same sign as Logloss. It had the other sign before and was maximized, now it is minimized. -
CatBoostRegressor.score
now returns the value of$R^2$ metric instead of RMSE to be more consistent with the behavior of scikit-learn regressors. - Changed metric parameter
use_weights
default value to false (except for ranking metrics)
- It is now possible to apply model on GPU
- We have published two new realworld datasets with monotonic constraints,
catboost.datasets.monotonic1()
andcatboost.datasets.monotonic2()
. Before that there was onlycalifornia_housing
dataset in open-source with monotonic constraints. Now you can use these two to benchmark algorithms with monotonic constraints. - We've added several new metrics to catboost, including
DCG
,FairLoss
,HammingLoss
,NormalizedGini
andFilteredNDCG
- Introduced efficient
GridSearch
andRandomSearch
implementations. get_all_params()
Python function returns the values of all training parameters, both user-defined and default.- Added more synonyms for training parameters to be more compatible with other GBDT libraries.
- AUC metric is computationally very expensive. We've implemented parallelized calculation of this metric, now it can be calculated on every iteration (or every k-th iteration) about 4x faster.
- We've improved our command-line tutorial, now it has examples of files and more information.
- Automatic
Logloss
orMultiClass
loss function deduction forCatBoostClassifier.fit
now also works if the training dataset is specified asPool
or filename string. - And some other fixes
- Function
get_feature_statistics
is replaced bycalc_feature_statistics
- Scoring function
Correlation
is renamed toCosine
- Parameter
efb_max_conflict_fraction
is renamed tosparse_features_conflict_fraction
- Models can be saved in PMML format now.
Note: PMML does not have full categorical features support, so to have the model in PMML format for datasets with categorical features you need to use set
one_hot_max_size
parameter to some large value, so that all categorical features are one-hot encoded
- Feature names can be used to specify ignored features
- Fixed restarting of CV on GPU for datasets without categorical features
- Fixed learning continuation errors with changed dataset (PR #879) and with model loaded from file (#884)
- Fixed NativeLib for JDK 9+ (PR #857)
- restored parameter
fstr_type
in Python and R interfaces
- cv is now stratified by default for
Logloss
,MultiClass
andMultiClassOneVsAll
. - We have removed
border
parameter ofLogloss
metric. You need to usetarget_border
as a separate training parameter now. CatBoostClassifier
now runsMultiClass
if more than 2 different values are present in training dataset labels.model.best_score_["validation_0"]
is replaced withmodel.best_score_["validation"]
if a single validation dataset is present.get_object_importance
function parameterostr_type
is renamed totype
in Python and R.
- Tree visualisation by @karina-usmanova.
- New feature analysis: plotting information about how a feature was used in the model by @alexrogozin12.
- Added
plot
parameter toget_roc_curve
,get_fpr_curve
andget_fnr_curve
functions fromcatboost.utils
. - Supported prettified format for all types of feature importances.
- Rust applier by @shuternay.
- DotNet applier by @17minutes.
- One-hot encoding for categorical features in CatBoost CoreML model by Kseniya Valchuk and Ekaterina Pogodina.
- Expectile Regression by @david-waterworth.
- Huber loss by @atsky.
- Speed up of shap values calculation for single object or for small number of objects by @Lokutrus.
- Cheap preprocessing and no fighting of overfitting if there is little amount of iterations (since you will not overfit anyway).
- Prediction of leaf indices.
- Rust tutorial by @shuternay.
- C# tutorial.
- Leaf indices.
- Tree visualisation tutorial by @karina-usmanova.
- Google Colab tutorial for regression in catboost by @col14m.
And a set of fixes for your issues.
- Add
has_header
parameter toCatboostEvaluation
class.
- Change output feature indices separator (
:
to;
) in theCatboostEvaluation
class.
- Changed default value for
--counter-calc-method
option toSkipTest
- Add guid to trained models. You can access it in Python using
get_metadata
function, for exampleprint catboost_model.get_metadata()['model_guid']
- Compatibility with glibc 2.12
- Improved embedded documentation
- Improved warning and error messages
-
GPU training now supports several tree learning strategies, selectable with
grow_policy
parameter. Possible values:SymmetricTree
-- The tree is built level by level untilmax_depth
is reached. On each iteration, all leaves from the last tree level will be split with the same condition. The resulting tree structure will always be symmetric.Depthwise
-- The tree is built level by level untilmax_depth
is reached. On each iteration, all non-terminal leaves from the last tree level will be split. Each leaf is split by condition with the best loss improvement.Lossguide
-- The tree is built leaf by leaf untilmax_leaves
limit is reached. On each iteration, non-terminal leaf with best loss improvement will be split.
Note: grow policies
Depthwise
andLossguide
currently support only training and prediction modes. They do not support model analysis (like feature importances and SHAP values) and saving to different model formats like CoreML, ONNX, and JSON.- The new grow policies support several new parameters:
max_leaves
-- Maximum leaf count in the resulting tree, default 31. Used only forLossguide
grow policy. Warning: It is not recommended to set this parameter greater than 64, as this can significantly slow down training.min_data_in_leaf
-- Minimum number of training samples per leaf, default 1. CatBoost will not search for new splits in leaves with sample count less thanmin_data_in_leaf
. This option is available forLossguide
andDepthwise
grow policies only.
Note: the new types of trees will be at least 10x slower in prediction than default symmetric trees.
-
GPU training also supports several score functions, that might give your model a boost in quality. Use parameter
score_function
to experiment with them. -
Now you can use quantization with more than 255 borders and
one_hot_max_size
> 255 in CPU training.
- It is now possible to use
save_borders()
function to write borders to a file after training. - Functions
predict
,predict_proba
,staged_predict
, andstaged_predict_proba
now support applying a model to a single object, in addition to usual data matrices.
- Impressive speedups for sparse datsets. Will depend on the dataset, but will be at least 2--3 times for sparse data.
- Python-package class attributes don't raise exceptions now. Attributes return
None
if not initialized. - Starting from 0.13 we have new feature importances for ranking modes. The new algorithm for feature importances shows how much features contribute to the optimized loss function. They are also signed as opposed to feature importances for not ranking modes which are non negative. This importances are expensive to calculate, thus we decided to not calculate them by default during training starting from 0.14. You need to calculate them after training.
- Fixed a bug in shap values that was introduced in v0.13
- Impressive speedup of CPU training for datasets with predominantly binary features (up to 5-6x).
- Speedup prediction and shap values array casting on large pools (issue #684).
- We've introduced a new type of feature importances -
LossFunctionChange
. This type of feature importances works well in all the modes, but is especially good for ranking. It is more expensive to calculate, thus we have not made it default. But you can look at it by selecting the type of feature importance. - Now we support online statistics for categorical features in
QuerySoftMax
mode on GPU. - We now support feature names in
cat_features
, PR #679 by @infected-mushroom - thanks a lot @infected-mushroom! - We've intoduced new sampling_type
MVS
, which speeds up CPU training if you use it. - Added
classes_
attribute in python. - Added support for input/output borders files in python package. Thank you @necnec for your PR #656!
- One more new option for working with categorical features is
ctr_target_border_count
. This option can be used if your initial target values are not binary and you do regression or ranking. It is equal to 1 by default, but you can try increasing it. - Added new option
sampling_unit
that allows to switch sampling from individual objects to entire groups. - More strings are interpreted as missing values for numerical features (mostly similar to pandas' read_csv).
- Allow
skip_train
property for loss functions in cv method. Contributed by GitHub user @RakitinDen, PR #662, many thanks. - We've improved classification mode on CPU, there will be less cases when the training diverges.
You can also try to experiment with new
leaf_estimation_backtracking
parameter. - Added new compare method for visualization, PR #652. Thanks @Drakon5999 for your contribution!
- Implemented
__eq__
method forCatBoost*
python classes (PR #654). Thanks @daskol for your contribution! - It is now possible to output evaluation results directly to
stdout
orstderr
in command-line CatBoost incalc
mode by specifyingstream://stdout
orstream://stderr
in--output-path
parameter argument. (PR #646). Thanks @towelenee for your contribution! - New loss function - Huber. Can be used as both an objective and a metric for regression. (PR #649). Thanks @atsky for your contribution!
- Changed defaults for
one_hot_max_size
training parameter for groupwise loss function training. SampleId
is the new main name for formerDocId
column in input data format (DocId
is still supported for compatibility). Contributed by GitHub user @daskol, PR #655, many thanks.- Improved CLI interface for cross-validation: replaced
-X/-Y
options with--cv
, PR #644. Thanks @tswr for your pr! eval_metrics
:eval_period
is now clipped by total number of trees in the specified interval. PR #653. Thanks @AntPon for your contribution!
- Thanks to @ws171913 we made necessary changes to prepare catboost for CRAN integration, PR #715. This is in progress now.
- R interface for cross-validation contributed by GitHub user @brsoyanvn, PR #561 -- many thanks @brsoyanvn!
- We've added new tutorial for GPU training on Google Colaboratory.
We have also done a list of fixes and data check improvements. Thanks @brazhenko, @Danyago98, @infected-mushroom for your contributions.
- Fixed loading of
epsilon
dataset into memory - Fixed multiclass learning on GPU for >255 classes
- Improved error handling
- Some other minor fixes
- Fixed Python compatibility issue in dataset downloading
- Added
sampling_type
parameter forYetiRankPairwise
loss
- Support saving models in ONNX format (only for models without categorical features).
- Added new dataset to our
catboost.datasets()
-- dataset epsilon, a large dense dataset for binary classification. - Speedup of Python
cv
on GPU. - Fixed creation of
Pool
frompandas.DataFrame
withpandas.Categorical
columns.
- Class weights are now taken into account by
eval_metrics()
,get_feature_importance()
, andget_object_importance()
. In previous versions the weights were ignored. - Parameter
random-strength
for pairwise training (PairLogitPairwise
,QueryCrossEntropy
,YetiRankPairwise
) is not supported anymore. - Simultaneous use of
MultiClass
andMultiClassOneVsAll
metrics is now deprecated.
cv
method is now supported on GPU.- String labels for classes are supported in Python.
In multiclassification the string class names are inferred from the data.
In binary classification for using string labels you should employ
class_names
parameter and specify which class is negative (0) and which is positive (1). You can also useclass_names
in multiclassification mode to pass all possible class names to the fit function. - Borders can now be saved and reused.
To save the feature quantization information obtained during training data
preprocessing into a text file use cli option
--output-borders-file
. To use the borders for training use cli option--input-borders-file
. This functionanlity is now supported on CPU and GPU (it was GPU-only in previous versions). File format for the borders is described here. - CLI option
--eval-file
is now supported on GPU.
- Some cases in binary classification are fixed where training could diverge
- A great speedup of the Python applier (10x)
- Reduced memory consumption in Python
cv
function (times fold count)
- Added speed benchmarks for CPU and GPU on a variety of different datasets.
- Added benchmarks of different ranking modes. In this tutorial we compare different ranking modes in CatBoost, XGBoost and LightGBM.
- Added tutorial for applying model in Java.
- Added benchmarks of SHAP values calculation for CatBoost, XGBoost and LightGBM. The benchmarks also contain explanation of complexity of this calculation in all the libraries.
We also made a list of stability improvements and stricter checks of input data and parameters.
And we are so grateful to our community members @canorbal and @neer201 for their contribution in this release. Thank you.
- Pure GPU implementation of NDCG metric
- Enabled LQ loss function
- Fixed NDCG metric on CPU
- Added
model_sum
mode to command line interface - Added SHAP values benchmark (#566)
- Fixed
random_strength
forPlain
boosting (#448) - Enabled passing a test pool to caret training (#544)
- Fixed a bug in exporting the model as python code (#556)
- Fixed label mapper for multiclassification custom labels (#523)
- Fixed hash type of categorical features (#558)
- Fixed handling of cross-validation fold count options in python package (#568)
- Accelerated formula evaluation by ~15%
- Improved model application interface
- Improved compilation time for building GPU version
- Better handling of stray commas in list arguments
- Added a benchmark that employs Rossman Store Sales dataset to compare quality of GBDT packages
- Added references to Catboost papers in R-package CITATION file
- Fixed a build issue in compilation for GPU
- Fixed a bug in model applicator
- Fixed model conversion, #533
- Returned pre 0.11 behaviour for
best_score_
andevals_result_
(issue #539) - Make valid RECORD in wheel (issue #534)
- Changed default border count for float feature binarization to 254 on CPU to achieve better quality
- Fixed random seed to
0
by default - Support model with more than 254 feature borders or one hot values when doing predictions
- Added model summation support in python: use
catboost.sum_models()
to sum models with provided weights. - Added json model tutorial json_model_tutorial.ipynb
- Bugfix for #518
In python 3 some functions returned dictionaries with keys of type bytes
- particularly eval_metrics and get_best_score. These are fixed to have keys of type str
.
- New metric NumErrors:greater_than=value
- New metric and objective L_q:q=value
- model.score(X, y) - can now work with Pool and labels from Pool
- Added EvalResult output after GPU catboost training
- Supported prediction type option on GPU
- Added
get_evals_result()
method andevals_result_
property to model in python wrapper to allow user access metric values - Supported string labels for GPU training in cmdline mode
- Many improvements in JNI wrapper
- Updated NDCG metric: speeded up and added NDCG with exponentiation in numerator as a new NDCG mode
- CatBoost doesn't drop unused features from model after training
- Write training finish time and catboost build info to model metadata
- Fix automatic pairs generation for GPU PairLogitPairwise target
- Fixed Python 3 support in
catboost.FeaturesData
- 40% speedup QuerySoftMax CPU training
- 2x Speedup pairwise loss functions
- For all the people struggling with occasional NaNs in test datasets - now we only write warnings about it
- We set up default loss_function in
CatBoostClassifier
andCatBoostRegressor
- Catboost write
Warning
andError
logs to stderr
- In R package we have changed parameter name
target
tolabel
in methodsave_pool()
- We don't support Python 3.4 anymore
- CatBoostClassifier and CatBoostRegressor
get_params()
method now returns only the params that were explicitly set when constructing the object. That means that CatBoostClassifier and CatBoostRegressor get_params() will not contain 'loss_function' if it was not specified. This also means that this code:
model1 = CatBoostClassifier()
params = model1.get_params()
model2 = CatBoost(params)
will create model2 with default loss_function RMSE, not with Logloss. This breaking change is done to support sklearn interface, so that sklearn GridSearchCV can work.
- We've removed several attributes and changed them to functions. This was needed to avoid sklearn warnings:
is_fitted_
=>is_fitted()
metadata_
=>get_metadata()
- We removed file with model from constructor of estimator. This was also done to avoid sklearn warnings.
- Now it is possible to save model in json format.
- We have added Java interface for CatBoost model
- We now have static linkage with CUDA, so you don't have to install any particular version of CUDA to get catboost working on GPU.
- We implemented both multiclass modes on GPU, it is very fast.
- It is possible now to use multiclass with string labels, they will be inferred from data
- Added
use_weights
parameter to metrics. By default all metrics, except for AUC use weights, but you can disable it. To calculate metric value without weights, you need to set this parameter to false. Example: Accuracy:use_weights=false. This can be done only for custom_metrics or eval_metric, not for the objective function. Objective function always uses weights if they are present in the dataset. - We now use snapshot time intervals. It will work much faster if you save snapshot every 5 or 10 minutes instead of saving it on every iteration.
- Reduced memory consumption by ranking modes.
- Added automatic feature importance evaluation after completion of GPU training.
- Allow inexistent indexes in ignored features list
- Added new metrics:
LogLikelihoodOfPrediction
,RecallAt:top=k
,PrecisionAt:top=k
andMAP:top=k
. - Improved quality for multiclass with weighted datasets.
- Pairwise modes now support automatic pairs generation (see tutorial for that).
- Metric
QueryAverage
is renamed to a more clearAverageGain
. This is a very important ranking metric. It shows average target value in top k documents of a group. Introduced parameterbest_model_min_trees
- the minimal number of trees the best model should have.
- We now support sklearn GridSearchCV: you can pass categorical feature indices when constructing estimator. And then use it in GridSearchCV.
- We added new method to utils - building of ROC curve:
get_roc_curve
. - Added
get_gpu_device_count()
method to python package. This is a way to check if your CUDA devices are available. - We implemented automatical selection of decision-boundary using ROC curve. You can select best classification boundary given the maximum FPR or FNR that you allow to the model. Take a look on
catboost.select_threshold(self, data=None, curve=None, FPR=None, FNR=None, thread_count=-1)
. You can also calculate FPR and FNR for each boundary value. - We have added pool slicing:
pool.slice(doc_indices)
- Allow GroupId and SubgroupId specified as strings.
- GPU support in R package. You need to use parameter
task_type='GPU'
to enable GPU training. - Models in R can be saved/restored by means of R: save/load or saveRDS/readRDS
- New way of loading data in Python using FeaturesData structure. Using FeaturesData will speed up both loading data for training and for prediction. It is especially important for prediction, because it gives around 10 to 20 times python prediction speedup.
- Training multiclass on CPU ~ 60% speedup
- Training of ranking modes on CPU ~ 50% speedup
- Training of ranking modes on GPU ~ 50% speedup for datasets with many features and not very many objects
- Speedups of metric calculation on GPU. Example of speedup on our internal dataset: training with - AUC eval metric with test dataset with 2kk objects is speeded up 7sec => 0.2 seconds per iteration.
- Speedup of all modes on CPU training.
We also did a lot of stability improvements, and improved usability of the library, added new parameter synonyms and improved input data validations.
Thanks a lot to all people who created issues on github. And thanks a lot to our contributor @pukhlyakova who implemented many new useful metrics!
- Fixed #403 bug in cuda train submodule (training crashed without evaluation set)
- Fixed exception propagation on pool parsing stage
- Add support of string
GroupId
andSubgroupId
in python-package - Print real class names instead of their labels in eval output
- We removed calc_feature_importance parameter from Python and R. Now feature importance calculation is almost free, so we always calculate feature importances. Previously you could disable it if it was slowing down your training.
- We removed Doc type for feature importances. Use Shap instead.
- We moved thread_count parameter in Python get_feature_importance method to the end.
In this release we added several very powerfull ranking objectives:
- PairLogitPairwise
- YetiRankPairwise
- QueryCrossEntropy (GPU only)
Other ranking improvements:
- We have made improvements to our existing ranking objectives QuerySoftMax and PairLogit.
- We have added group weights support.
- Improvement for datasets with weights
- Now we automatically calculate a good learning rate for you in the start of training, you don't have to specify it. After the training has finished, you can look on the training curve on evaluation dataset and make ajustments to the selected learning rate, but it will already be a good value.
- Several speedups for GPU training.
- 1.5x speedup for applying the model.
- Speed up multi classificaton training.
- 2x speedup for AUC calculation in eval_metrics.
- Several speedups for eval_metrics for other metrics.
- 100x speed up for Shap values calculation.
- Speedup for feature importance calculation. It used to be a bottleneck for GPU training previously, now it's not.
- We added possibility to not calculate metric on train dataset using
MetricName:hints=skip_train~false
(it might speed up your training if metric calculation is a bottle neck, for example, if you calculate many metrics or if you calculate metrics on GPU). - We added possibility to calculate metrics only periodically, not on all iterations. Use metric_period for that. (previously it only disabled verbose output on each iteration).
- Now we disable by default calculation of expensive metrics on train dataset. We don't calculate AUC and PFound metrics on train dataset by default. You can also disable calculation of other metrics on train dataset using
MetricName:hints=skip_train~true
. If you want to calculate AUC or PFound on train dataset you can useMetricName:hints=skip_train~false
. - Now if you want to calculate metrics using eval_metrics or during training you can use metric_period to skip some iterations. It will speed up eval_metrics and it might speed up training, especially GPU training.
Note that the most expensive metric calculation is AUC calculation, for this metric and large datasets it makes sense to use metric_period.
If you only want to see less verbose output, and still want to see metric values on every iteration written in file, you can use
verbose=n
parameter - Parallelization of calculation of most of the metrics during training
- It is possible now to calculate and visualise custom_metric during training on GPU.
Now you can use our Jupyter visualization, CatBoost viewer or TensorBoard the same way you used it for CPU training. It might be a bottleneck, so if it slows down your training use
metric_period=something
andMetricName:hints=skip_train~false
- We switched to CUDA 9.1. Starting from this release CUDA 8.0 will not be supported
- Support for external borders on GPU for cmdline
- We added support of feature combinations to our Shap values implementation.
- Added Shap values for MultiClass and added an example of it's usage to our Shap tutorial.
- Added pretified parameter to get_feature_importance(). With
pretified=True
the function will return list of features with names sorted in descending order by their importance. - Improved interfaces for eval-feature functionality
- Shap values support in R-package
- It is possible now to save any metainformation to the model.
- Empty values support
- Better support of sklearn
- feature_names_ for CatBoost class
- Added silent parameter
- Better stdout
- Better diagnostic for invalid inputs
- Better documentation
- Added a flag to allow constant labels
We added many new metrics that can be used for visualization, overfitting detection, selecting of best iteration of training or for cross-validation:
- BierScore
- HingeLoss
- HammingLoss
- ZeroOneLoss
- MSLE
- MAE
- BalancedAccuracy
- BalancedErrorRate
- Kappa
- Wkappa
- QueryCrossEntropy
- NDCG
- Saving model as C++ code
- Saving model with categorical features as Python code
Added make files for binary with CUDA and for Python package
We created a new repo with tutorials, now you don't have to clone the whole catboost repo to run Jupyter notebook with a tutorial.
We have also a set of bugfixes and we are gratefull to everyone who has filled a bugreport, helping us making the library better.
This release contains contributions from CatBoost team. We want to especially mention @pukhlyakova who implemented lots of useful metrics.
- New model method
get_cat_feature_indices()
in Python wrapper. - Minor fixes and stability improvements.
- We fixed bug in CatBoost. Pool initialization from
numpy.array
andpandas.dataframe
with string values that can cause slight inconsistence while using trained model from older versions. Around 1% of cat feature hashes were treated incorrectly. If you expirience quality drop after update you should consider retraining your model.
- Algorithm for finding most influential training samples for a given object from the 'Finding Influential Training Samples for Gradient Boosted Decision Trees' paper is implemented. This mode for every object from input pool calculates scores for every object from train pool. A positive score means that the given train object has made a negative contribution to the given test object prediction. And vice versa for negative scores. The higher score modulo - the higher contribution.
See
get_object_importance
model method in Python package andostr
mode in cli-version. Tutorial for Python is available here. More details and examples will be published in documentation soon. - We have implemented new way of exploring feature importance - Shap values from paper. This allows to understand which features are most influent for a given object. You can also get more insite about your model, see details in a tutorial.
- Save model as code functionality published. For now you could save model as Python code with categorical features and as C++ code w/o categorical features.
- Fix
_catboost
reinitialization issues #268 and #269. - Python module
catboost.util
extended withcreate_cd
. It creates column description file. - Now it's possible to load titanic and amazon (Kaggle Amazon Employee Access Challenge) datasets from Python code. Use
catboost.datasets
. - GPU parameter
use_cpu_ram_for_cat_features
renamed togpu_cat_features_storage
with posible valuesCpuPinnedMemory
andGpuRam
. Default isGpuRam
.
This release contains contributions from CatBoost team.
As usual we are grateful to all who filed issues or helped resolve them, asked and answered questions.
- GPU: New
DocParallel
mode for tasks without categorical features and or with categorical features and—max-ctr-complextiy 1
. Provides best performance for pools with big number of documents. - GPU: Distributed training on several GPU host via MPI. See instruction how to build binary here.
- GPU: Up to 30% learning speed-up for Maxwell and later GPUs with binarization level > 32
- Hotfixes for GPU version of python wrapper.
- Python wrapper: added methods to download datasets titanic and amazon, to make it easier to try the library (
catboost.datasets
). - Python wrapper: added method to write column desctiption file (
catboost.utils.create_cd
). - Made improvements to visualization.
- Support non-numeric values in
GroupId
column. - Tutorials section updated.
- Fixed problems with eval_metrics (issue #285)
- Other fixes
- Changed parameter order in
train()
function to be consistant with other GBDT libraries. use_best_model
is set to True by default ifeval_set
labels are present.
- New ranking mode
YetiRank
optimizesNDGC
andPFound
. - New visualisation for
eval_metrics
andcv
in Jupyter notebook. - Improved per document feature importance.
- Supported
verbose
=int
: ifverbose
> 1,metric_period
is set to this value. - Supported type(
eval_set
) = list in python. Currently supporting only singleeval_set
. - Binary classification leaf estimation defaults are changed for weighted datasets so that training converges for any weights.
- Add
model_size_reg
parameter to control model size. Fixctr_leaf_count_limit
parameter, also to control model size. - Beta version of distributed CPU training with only float features support.
- Add
subgroupId
to Python/R-packages. - Add groupwise metrics support in
eval_metrics
.
This release contains contributions from CatBoost team.
We are grateful to all who filed issues or helped resolve them, asked and answered questions.
boosting_type
parameter valueDynamic
is renamed toOrdered
.- Data visualisation functionality in Jupyter Notebook requires ipywidgets 7.x+ now.
query_id
parameter renamed togroup_id
in Python and R wrappers.- cv returns pandas.DataFrame by default if Pandas installed. See new parameter
as_pandas
.
- CatBoost build with make file. Now it’s possible to build command-line CPU version of CatBoost under Linux with make file.
- In column description column name
Target
is changed toLabel
. It will still work with previous name, but it is recommended to use the new one. eval-metrics
mode added into cmdline version. Metrics can be calculated for a given dataset using a previously trained model.- New classification metric
CtrFactor
is added. - Load CatBoost model from memory. You can load your CatBoost model from file or initialize it from buffer in memory.
- Now you can run
fit
function using file with dataset:fit(train_path, eval_set=eval_path, column_description=cd_file)
. This will reduce memory consumption by up to two times. - 12% speedup for training.
- JSON output data format is changed.
- Python whl binaries with CUDA 9.1 support for Linux OS published into the release assets.
- Added
bootstrap_type
parameter toCatBoostClassifier
andRegressor
(issue #263).
This release contains contributions from newbfg and CatBoost team.
We are grateful to all who filed issues or helped resolve them, asked and answered questions.
- BETA version of distributed mulit-host GPU via MPI training
- Added possibility to import coreml model with oblivious trees. Makes possible to migrate pre-flatbuffers model (with float features only) to current format (issue #235)
- Added QuerySoftMax loss function
- Fixed GPU models bug on pools with both categorical and float features (issue #241)
- Use all available cores by default
- Fixed not querywise loss for pool with
QueryId
- Default float features binarization method set to
GreedyLogSum
- Hotfix for critical bug in Python and R wrappers (issue #238)
- Added stratified data split in CV
- Fix
is_classification
check and CV for Logloss
- Fixed critical bugs in formula evaluation code (issue #236)
- Added scale_pos_weight parameter
- 25% speedup of the model applier
- 43% speedup for training on large datasets.
- 15% speedup for
QueryRMSE
and calculation of querywise metrics. - Large speedups when using binary categorical features.
- Significant (x200 on 5k trees and 50k lines dataset) speedup for plot and stage predict calculations in cmdline.
- Compilation time speedup.
- Industry fastest applier implementation.
- Introducing new parameter
boosting-type
to switch between standard boosting scheme and dynamic boosting, described in paper "Dynamic boosting". - Adding new bootstrap types
bootstrap_type
,subsample
. UsingBernoulli
bootstrap type withsubsample < 1
might increase the training speed. - Better logging for cross-validation, added parameter
logging_level
andmetric_period
(should be set in training parameters) to cv. - Added a separate
train
function that receives the parameters and returns a trained model. - Ranking mode
QueryRMSE
now supports default settings for dynamic boosting. - R-package pre-build binaries are included into release.
- We added many synonyms to our parameter names, now it is more convenient to try CatBoost if you are used to some other library.
- Fix for CPU
QueryRMSE
with weights. - Adding several missing parameters into wrappers.
- Fix for data split in querywise modes.
- Better logging.
- From this release we'll provide pre-build R-binaries
- More parallelisation.
- Memory usage improvements.
- And some other bug fixes.
This release contains contributions from CatBoost team.
We are grateful to all who filed issues or helped resolve them, asked and answered questions.
- We've made single document formula applier 4 times faster!
model.shrink
function added in Python and R wrappers.- Added new training parameter
metric_period
that controls output frequency. - Added new ranking metric
QueryAverage
. - This version contains an easy way to implement new user metrics in C++. How-to example is provided.
- Stability improvements and bug fixes
As usual we are grateful to all who filed issues, asked and answered questions.
Cmdline:
- Training parameter
gradient-iterations
renamed toleaf-estimation-iterations
. border
option removed. If you want to specify border for binary classification mode you need to specify it in the following way:loss-function Logloss:Border=0.5
- CTR parameters are changed:
- Removed
priors
,per-feature-priors
,ctr-binarization
; - Added
simple-ctr
,combintations-ctr
,per-feature-ctr
; More details will be published in our documentation.
- Removed
Python:
- Training parameter
gradient_iterations
renamed toleaf_estimation_iterations
. border
option removed. If you want to specify border for binary classification mode you need to specify it in the following way:loss_function='Logloss:Border=0.5'
- CTR parameters are changed:
- Removed
priors
,per_feature_priors
,ctr_binarization
; - Added
simple_ctr
,combintations_ctr
,per_feature_ctr
; More details will be published in our documentation.
- Removed
- In Python we added a new method
eval_metrics
: now it's possible for a given model to calculate specified metric values for each iteration on specified dataset. - One command-line binary for CPU and GPU: in CatBoost you can switch between CPU and GPU training by changing single parameter value
task-type CPU
orGPU
(task_type 'CPU', 'GPU' in python bindings). Windows build still contains two binaries. - We have speed up the training up to 30% for datasets with a lot of objects.
- Up to 10% speed-up of GPU implementation on Pascal cards
- Stability improvements and bug fixes
As usual we are grateful to all who filed issues, asked and answered questions.
FlatBuffers model format: new CatBoost versions wouldn’t break model compatibility anymore.
- Training speedups: we have speed up the training by 33%.
- Two new ranking modes are available:
PairLogit
- pairwise comparison of objects from the input dataset. Algorithm maximises probability correctly reorder all dataset pairs.QueryRMSE
- mix of regression and ranking. It’s trying to make best ranking for each dataset query by input labels.
- We have fixed a bug that caused quality degradation when using weights < 1.
Verbose
flag is now deprecated, please uselogging_level
instead. You could set the following levels:Silent
,Verbose
,Info
,Debug
.- And some other bugs.
This release contains contributions from: avidale, newbfg, KochetovNicolai and CatBoost team.
We are grateful to all who filed issues or helped resolve them, asked and answered questions.
GPU CUDA support is available. CatBoost supports multi-GPU training. Our GPU implementation is 2 times faster then LightGBM and more then 20 times faster then XGBoost one. Check out the news with benchmarks on our site.
Stability improvements and bug fixes
This release contains contributions from: daskol and CatBoost team.
We are grateful to all who filed issues or helped resolve them, asked and answered questions.
- R library interface significantly changed
- New model format: CatBoost v0.2 model binary not compatible with previous versions
- Cross-validation parameters changes: we changed overfitting detector parameters of CV in python so that it is same as those in training.
- CTR types: MeanValue => BinarizedTargetMeanValue
- Training speedups: we have speed up the training by 20-30%.
- Accuracy improvement with categoricals: we have changed computation of statistics for categorical features, which leads to better quality.
- New type of overfitting detector:
Iter
. This type of detector was requested by our users. So now you can also stop training by a simple criterion: if after a fixed number of iterations there is no improvement of your evaluation function. - TensorBoard support: this is another way of looking on the graphs of different error functions both during training and after training has finished. To look at the metrics you need to provide
train_dir
when training your model and then run"tensorboard --logdir={train_dir}"
- Jupyter notebook improvements: for our Python library users that experiment with Jupyter notebooks, we have improved our visualisation tool. Now it is possible to save image of the graph. We also have changed scrolling behaviour so that it is more convenient to scroll the notebook.
- NaN features support: we also have added simple but effective way of dealing with NaN features. If you have some NaNs in the train set, they will be changed to a value that is less than the minimum value or greater than the maximum value in the dataset (this is configurable), so that it is guaranteed that they are in their own bin, and a split would separates NaN values from all other values. By default, no NaNs are allowed, so you need to use option
nan_mode
for that. When applying a model, NaNs will be treated in the same way for the features where NaN values were seen in train. It is not allowed to have NaN values in test if no NaNs in train for this feature were provided. - Snapshotting: we have added snapshotting to our Python and R libraries. So if you think that something can happen with your training, for example machine can reboot, you can use
snapshot_file
parameter - this way after you restart your training it will start from the last completed iteration. - R library tutorial: we have added tutorial
- Logging customization: we have added
allow_writing_files
parameter. By default some files with logging and diagnostics are written on disc, but you can turn it off using by setting this flag to False. - Multiclass mode improvements: we have added a new objective for multiclass mode -
MultiClassOneVsAll
. We also addedclass_names
param - now you don't have to renumber your classes to be able to use multiclass. And we have added two new metrics for multiclass:TotalF1
andMCC
metrics. You can use the metrics to look how its values are changing during training or to use overfitting detection or cutting the model by best value of a given metric. - Any delimeters support: in addition to datasets in
tsv
format, CatBoost now supports files with any delimeters
Stability improvements and bug fixes
This release contains contributions from: grayskripko, hadjipantelis and CatBoost team.
We are grateful to all who filed issues or helped resolve them, asked and answered questions.