catboost v0.24.2 Release Notes
Release Date: 2020-10-07 // about 4 years ago-
Uncertainty prediction
- ๐ Supported uncertainty prediction for classification models.
- ๐ Fixed RMSEWithUncertainty data uncertainty prediction - now it predicts variance, not standard deviation.
๐ New functionality
- ๐ Allow categorical feature counters for
MultiRMSE
loss function. group_weight
parameter added tocatboost.utils.eval_metric
method to allow passing weights for object groups. Allows correctly match weighted ranking metrics computation when group weights present.- ๐ Faster non-owning deserialization from memory with less memory overhead - moved some dynamically computed data to model file, other data is computed in lazy manner only when needed.
Experimental functionality
- ๐ Supported embedding features as input and linear discriminant analysis for embeddings preprocessing. Try adding your embeddings as new columns with embedding values array in Pandas.Dataframe and passing corresponding column names to
Pool
constructor orfit
function withembedding_features=['EmbeddingFeaturesColumnName1, ...]
parameter. Another way of adding your embedding vectors is new type of column in Column Description fileNumVector
and adding semicolon separated embeddings column to your XSV file:ClassLabel\t0.1;0.2;0.3\t...
.
Educational materials
- Published new tutorial on uncertainty prediction.
๐ Bugfixes:
- โฌ๏ธ Reduced GPU memory usage in multi gpu training when there is no need to compute categorical feature counters.
- Now CatBoost allows to specify
use_weights
for metrics whenauto_class_weights
parameter is set. - Correctly handle NaN values in
plot_predictions
function. - ๐ Fixed floating point precision drop releated bugs during Multiclass training with lots of objects in our case, bug was triggered while training on 25mln objects on single GPU card.
- Now
average
parameter is passed to TotalF1 metric while training on GPU. - โ Added class labels checks
- Disallow feature remapping in model predict when there is empty feature names in model.