Version 0.19#
Version 0.19.2#
July, 2018
This release is exclusively in order to support Python 3.7.
Version 0.19.1#
October 23, 2017
This is a bug-fix release with some minor documentation improvements and enhancements to features released in 0.19.0.
Note there may be minor differences in TSNE output in this release (due to #9623), in the case where multiple samples have equal distance to some sample.
Changelog#
API changes#
- Reverted the addition of - metrics.ndcg_scoreand- metrics.dcg_scorewhich had been merged into version 0.19.0 by error. The implementations were broken and undocumented.
- return_train_scorewhich was added to- model_selection.GridSearchCV,- model_selection.RandomizedSearchCVand- model_selection.cross_validatein version 0.19.0 will be changing its default value from True to False in version 0.21. We found that calculating training score could have a great effect on cross validation runtime in some cases. Users should explicitly set- return_train_scoreto False if prediction or scoring functions are slow, resulting in a deleterious effect on CV runtime, or to True if they wish to use the calculated scores. #9677 by Kumar Ashutosh and Joel Nothman.
- correlation_modelsand- regression_modelsfrom the legacy gaussian processes implementation have been belatedly deprecated. #9717 by Kumar Ashutosh.
Bug fixes#
- Avoid integer overflows in - metrics.matthews_corrcoef. #9693 by Sam Steingold.
- Fixed a bug in the objective function for - manifold.TSNE(both exact and with the Barnes-Hut approximation) when- n_components >= 3. #9711 by @goncalo-rodrigues.
- Fix regression in - model_selection.cross_val_predictwhere it raised an error with- method='predict_proba'for some probabilistic classifiers. #9641 by James Bourbeau.
- Fixed a bug where - datasets.make_classificationmodified its input- weights. #9865 by Sachin Kelkar.
- model_selection.StratifiedShuffleSplitnow works with multioutput multiclass or multilabel data with more than 1000 columns. #9922 by Charlie Brummitt.
- Fixed a bug with nested and conditional parameter setting, e.g. setting a pipeline step and its parameter at the same time. #9945 by Andreas Müller and Joel Nothman. 
Regressions in 0.19.0 fixed in 0.19.1:
- Fixed a bug where parallelised prediction in random forests was not thread-safe and could (rarely) result in arbitrary errors. #9830 by Joel Nothman. 
- Fix regression in - model_selection.cross_val_predictwhere it no longer accepted- Xas a list. #9600 by Rasul Kerimov.
- Fixed handling of - model_selection.cross_val_predictfor binary classification with- method='decision_function'. #9593 by Reiichiro Nakano and core devs.
- Fix regression in - pipeline.Pipelinewhere it no longer accepted- stepsas a tuple. #9604 by Joris Van den Bossche.
- Fix bug where - n_iterwas not properly deprecated, leaving- n_iterunavailable for interim use in- linear_model.SGDClassifier,- linear_model.SGDRegressor,- linear_model.PassiveAggressiveClassifier,- linear_model.PassiveAggressiveRegressorand- linear_model.Perceptron. #9558 by Andreas Müller.
- Dataset fetchers make sure temporary files are closed before removing them, which caused errors on Windows. #9847 by Joan Massich. 
- Fixed a regression in - manifold.TSNEwhere it no longer supported metrics other than ‘euclidean’ and ‘precomputed’. #9623 by Oli Blum.
Enhancements#
- Our test suite and - utils.estimator_checks.check_estimatorcan now be run without Nose installed. #9697 by Joan Massich.
- To improve usability of version 0.19’s - pipeline.Pipelinecaching,- memorynow allows- joblib.Memoryinstances. This make use of the new- utils.validation.check_memoryhelper. #9584 by Kumar Ashutosh
- Made a FutureWarning in SGD-based estimators less verbose. #9802 by Vrishank Bhardwaj. 
Code and Documentation Contributors#
With thanks to:
Joel Nothman, Loic Esteve, Andreas Mueller, Kumar Ashutosh, Vrishank Bhardwaj, Hanmin Qin, Rasul Kerimov, James Bourbeau, Nagarjuna Kumar, Nathaniel Saul, Olivier Grisel, Roman Yurchak, Reiichiro Nakano, Sachin Kelkar, Sam Steingold, Yaroslav Halchenko, diegodlh, felix, goncalo-rodrigues, jkleint, oliblum90, pasbi, Anthony Gitter, Ben Lawson, Charlie Brummitt, Didi Bar-Zev, Gael Varoquaux, Joan Massich, Joris Van den Bossche, nielsenmarkus11
Version 0.19#
August 12, 2017
Highlights#
We are excited to release a number of great new features including
neighbors.LocalOutlierFactor for anomaly detection,
preprocessing.QuantileTransformer for robust feature transformation,
and the multioutput.ClassifierChain meta-estimator to simply account
for dependencies between classes in multilabel problems. We have some new
algorithms in existing estimators, such as multiplicative update in
decomposition.NMF and multinomial
linear_model.LogisticRegression with L1 loss (use solver='saga').
Cross validation is now able to return the results from multiple metric
evaluations. The new model_selection.cross_validate can return many
scores on the test data as well as training set performance and timings, and we
have extended the scoring and refit parameters for grid/randomized
search to handle multiple metrics.
You can also learn faster.  For instance, the new option to cache
transformations in pipeline.Pipeline makes grid
search over pipelines including slow transformations much more efficient.  And
you can predict faster: if you’re sure you know what you’re doing, you can turn
off validating that the input is finite using config_context.
We’ve made some important fixes too.  We’ve fixed a longstanding implementation
error in metrics.average_precision_score, so please be cautious with
prior results reported from that function.  A number of errors in the
manifold.TSNE implementation have been fixed, particularly in the
default Barnes-Hut approximation.  semi_supervised.LabelSpreading and
semi_supervised.LabelPropagation have had substantial fixes.
LabelPropagation was previously broken. LabelSpreading should now correctly
respect its alpha parameter.
Changed models#
The following estimators and functions, when fit with the same data and parameters, may produce different models from the previous version. This often occurs due to changes in the modelling logic (bug fixes or enhancements), or in random sampling procedures.
- cluster.KMeanswith sparse X and initial centroids given (bug fix)
- cross_decomposition.PLSRegressionwith- scale=True(bug fix)
- ensemble.GradientBoostingClassifierand- ensemble.GradientBoostingRegressorwhere- min_impurity_splitis used (bug fix)
- gradient boosting - loss='quantile'(bug fix)
- ensemble.IsolationForest(bug fix)
- feature_selection.SelectFdr(bug fix)
- linear_model.RANSACRegressor(bug fix)
- linear_model.LassoLars(bug fix)
- linear_model.LassoLarsIC(bug fix)
- manifold.TSNE(bug fix)
- neighbors.NearestCentroid(bug fix)
- semi_supervised.LabelSpreading(bug fix)
- semi_supervised.LabelPropagation(bug fix)
- tree based models where - min_weight_fraction_leafis used (enhancement)
- model_selection.StratifiedKFoldwith- shuffle=True(this change, due to #7823 was not mentioned in the release notes at the time)
Details are listed in the changelog below.
(While we are trying to better inform users by providing this information, we cannot assure that this list is complete.)
Changelog#
New features#
Classifiers and regressors
- Added - multioutput.ClassifierChainfor multi-label classification. By Adam Kleczewski.
- Added solver - 'saga'that implements the improved version of Stochastic Average Gradient, in- linear_model.LogisticRegressionand- linear_model.Ridge. It allows the use of L1 penalty with multinomial logistic loss, and behaves marginally better than ‘sag’ during the first epochs of ridge and logistic regression. #8446 by Arthur Mensch.
Other estimators
- Added the - neighbors.LocalOutlierFactorclass for anomaly detection based on nearest neighbors. #5279 by Nicolas Goix and Alexandre Gramfort.
- Added - preprocessing.QuantileTransformerclass and- preprocessing.quantile_transformfunction for features normalization based on quantiles. #8363 by Denis Engemann, Guillaume Lemaitre, Olivier Grisel, Raghav RV, Thierry Guillemot, and Gael Varoquaux.
- The new solver - 'mu'implements a Multiplicate Update in- decomposition.NMF, allowing the optimization of all beta-divergences, including the Frobenius norm, the generalized Kullback-Leibler divergence and the Itakura-Saito divergence. #5295 by Tom Dupre la Tour.
Model selection and evaluation
- model_selection.GridSearchCVand- model_selection.RandomizedSearchCVnow support simultaneous evaluation of multiple metrics. Refer to the Specifying multiple metrics for evaluation section of the user guide for more information. #7388 by Raghav RV
- Added the - model_selection.cross_validatewhich allows evaluation of multiple metrics. This function returns a dict with more useful information from cross-validation such as the train scores, fit times and score times. Refer to The cross_validate function and multiple metric evaluation section of the userguide for more information. #7388 by Raghav RV
- Added - metrics.mean_squared_log_error, which computes the mean square error of the logarithmic transformation of targets, particularly useful for targets with an exponential trend. #7655 by Karan Desai.
- Added - metrics.dcg_scoreand- metrics.ndcg_score, which compute Discounted cumulative gain (DCG) and Normalized discounted cumulative gain (NDCG). #7739 by David Gasquez.
- Added the - model_selection.RepeatedKFoldand- model_selection.RepeatedStratifiedKFold. #8120 by Neeraj Gangwar.
Miscellaneous
- Validation that input data contains no NaN or inf can now be suppressed using - config_context, at your own risk. This will save on runtime, and may be particularly useful for prediction time. #7548 by Joel Nothman.
- Added a test to ensure parameter listing in docstrings matches the function/class signature. #9206 by Alexandre Gramfort and Raghav RV. 
Enhancements#
Trees and ensembles
- The - min_weight_fraction_leafconstraint in tree construction is now more efficient, taking a fast path to declare a node a leaf if its weight is less than 2 * the minimum. Note that the constructed tree will be different from previous versions where- min_weight_fraction_leafis used. #7441 by Nelson Liu.
- ensemble.GradientBoostingClassifierand- ensemble.GradientBoostingRegressornow support sparse input for prediction. #6101 by Ibraim Ganiev.
- ensemble.VotingClassifiernow allows changing estimators by using- ensemble.VotingClassifier.set_params. An estimator can also be removed by setting it to- None. #7674 by Yichuan Liu.
- tree.export_graphviznow shows configurable number of decimal places. #8698 by Guillaume Lemaitre.
- Added - flatten_transformparameter to- ensemble.VotingClassifierto change output shape of- transformmethod to 2 dimensional. #7794 by Ibraim Ganiev and Herilalaina Rakotoarison.
Linear, kernelized and related models
- linear_model.SGDClassifier,- linear_model.SGDRegressor,- linear_model.PassiveAggressiveClassifier,- linear_model.PassiveAggressiveRegressorand- linear_model.Perceptronnow expose- max_iterand- tolparameters, to handle convergence more precisely.- n_iterparameter is deprecated, and the fitted estimator exposes a- n_iter_attribute, with actual number of iterations before convergence. #5036 by Tom Dupre la Tour.
- Added - averageparameter to perform weight averaging in- linear_model.PassiveAggressiveClassifier. #4939 by Andrea Esuli.
- linear_model.RANSACRegressorno longer throws an error when calling- fitif no inliers are found in its first iteration. Furthermore, causes of skipped iterations are tracked in newly added attributes,- n_skips_*. #7914 by Michael Horrell.
- In - gaussian_process.GaussianProcessRegressor, method- predictis a lot faster with- return_std=True. #8591 by Hadrien Bertrand.
- Added - return_stdto- predictmethod of- linear_model.ARDRegressionand- linear_model.BayesianRidge. #7838 by Sergey Feldman.
- Memory usage enhancements: Prevent cast from float32 to float64 in: - linear_model.MultiTaskElasticNet;- linear_model.LogisticRegressionwhen using newton-cg solver; and- linear_model.Ridgewhen using svd, sparse_cg, cholesky or lsqr solvers. #8835, #8061 by Joan Massich and Nicolas Cordier and Thierry Guillemot.
Other predictors
- Custom metrics for the - sklearn.neighborsbinary trees now have fewer constraints: they must take two 1d-arrays and return a float. #6288 by Jake Vanderplas.
- algorithm='autoin- sklearn.neighborsestimators now chooses the most appropriate algorithm for all input types and metrics. #9145 by Herilalaina Rakotoarison and Reddy Chinthala.
Decomposition, manifold learning and clustering
- cluster.MiniBatchKMeansand- cluster.KMeansnow use significantly less memory when assigning data points to their nearest cluster center. #7721 by Jon Crall.
- decomposition.PCA,- decomposition.IncrementalPCAand- decomposition.TruncatedSVDnow expose the singular values from the underlying SVD. They are stored in the attribute- singular_values_, like in- decomposition.IncrementalPCA. #7685 by Tommy Löfstedt
- decomposition.NMFnow faster when- beta_loss=0. #9277 by @hongkahjun.
- Memory improvements for method - barnes_hutin- manifold.TSNE#7089 by Thomas Moreau and Olivier Grisel.
- Optimization schedule improvements for Barnes-Hut - manifold.TSNEso the results are closer to the one from the reference implementation lvdmaaten/bhtsne by Thomas Moreau and Olivier Grisel.
- Memory usage enhancements: Prevent cast from float32 to float64 in - decomposition.PCAand- decomposition.randomized_svd_low_rank. #9067 by Raghav RV.
Preprocessing and feature selection
- Added - norm_orderparameter to- feature_selection.SelectFromModelto enable selection of the norm order when- coef_is more than 1D. #6181 by Antoine Wendlinger.
- Added ability to use sparse matrices in - feature_selection.f_regressionwith- center=True. #8065 by Daniel LeJeune.
- Small performance improvement to n-gram creation in - sklearn.feature_extraction.textby binding methods for loops and special-casing unigrams. #7567 by Jaye Doepke
- Relax assumption on the data for the - kernel_approximation.SkewedChi2Sampler. Since the Skewed-Chi2 kernel is defined on the open interval \((-skewedness; +\infty)^d\), the transform function should not check whether- X < 0but whether- X < -self.skewedness. #7573 by Romain Brault.
- Made default kernel parameters kernel-dependent in - kernel_approximation.Nystroem. #5229 by Saurabh Bansod and Andreas Müller.
Model evaluation and meta-estimators
- pipeline.Pipelineis now able to cache transformers within a pipeline by using the- memoryconstructor parameter. #7990 by Guillaume Lemaitre.
- pipeline.Pipelinesteps can now be accessed as attributes of its- named_stepsattribute. #8586 by Herilalaina Rakotoarison.
- Added - sample_weightparameter to- pipeline.Pipeline.score. #7723 by Mikhail Korobov.
- Added ability to set - n_jobsparameter to- pipeline.make_union. A- TypeErrorwill be raised for any other kwargs. #8028 by Alexander Booth.
- model_selection.GridSearchCV,- model_selection.RandomizedSearchCVand- model_selection.cross_val_scorenow allow estimators with callable kernels which were previously prohibited. #8005 by Andreas Müller .
- model_selection.cross_val_predictnow returns output of the correct shape for all values of the argument- method. #7863 by Aman Dalmia.
- Added - shuffleand- random_stateparameters to shuffle training data before taking prefixes of it based on training sizes in- model_selection.learning_curve. #7506 by Narine Kokhlikyan.
- model_selection.StratifiedShuffleSplitnow works with multioutput multiclass (or multilabel) data. #9044 by Vlad Niculae.
- Speed improvements to - model_selection.StratifiedShuffleSplit. #5991 by Arthur Mensch and Joel Nothman.
- Add - shuffleparameter to- model_selection.train_test_split. #8845 by themrmax
- multioutput.MultiOutputRegressorand- multioutput.MultiOutputClassifiernow support online learning using- partial_fit. :issue:- 8053by Peng Yu.
- Add - max_train_sizeparameter to- model_selection.TimeSeriesSplit#8282 by Aman Dalmia.
- More clustering metrics are now available through - metrics.get_scorerand- scoringparameters. #8117 by Raghav RV.
- A scorer based on - metrics.explained_variance_scoreis also available. #9259 by Hanmin Qin.
Metrics
- metrics.matthews_corrcoefnow supports multiclass classification. #8094 by Jon Crall.
- Add - sample_weightparameter to- metrics.cohen_kappa_score. #8335 by Victor Poughon.
Miscellaneous
- utils.estimator_checks.check_estimatornow attempts to ensure that methods transform, predict, etc. do not set attributes on the estimator. #7533 by Ekaterina Krivich.
- Added type checking to the - accept_sparseparameter in- sklearn.utils.validationmethods. This parameter now accepts only boolean, string, or list/tuple of strings.- accept_sparse=Noneis deprecated and should be replaced by- accept_sparse=False. #7880 by Josh Karnofsky.
- Make it possible to load a chunk of an svmlight formatted file by passing a range of bytes to - datasets.load_svmlight_file. #935 by Olivier Grisel.
- dummy.DummyClassifierand- dummy.DummyRegressornow accept non-finite features. #8931 by @Attractadore.
Bug fixes#
Trees and ensembles
- Fixed a memory leak in trees when using trees with - criterion='mae'. #8002 by Raghav RV.
- Fixed a bug where - ensemble.IsolationForestuses an incorrect formula for the average path length #8549 by Peter Wang.
- Fixed a bug where - ensemble.AdaBoostClassifierthrows- ZeroDivisionErrorwhile fitting data with single class labels. #7501 by Dominik Krzeminski.
- Fixed a bug in - ensemble.GradientBoostingClassifierand- ensemble.GradientBoostingRegressorwhere a float being compared to- 0.0using- ==caused a divide by zero error. #7970 by He Chen.
- Fix a bug where - ensemble.GradientBoostingClassifierand- ensemble.GradientBoostingRegressorignored the- min_impurity_splitparameter. #8006 by Sebastian Pölsterl.
- Fixed - oob_scorein- ensemble.BaggingClassifier. #8936 by Michael Lewis
- Fixed excessive memory usage in prediction for random forests estimators. #8672 by Mike Benfield. 
- Fixed a bug where - sample_weightas a list broke random forests in Python 2 #8068 by @xor.
- Fixed a bug where - ensemble.IsolationForestfails when- max_featuresis less than 1. #5732 by Ishank Gulati.
- Fix a bug where gradient boosting with - loss='quantile'computed negative errors for negative values of- ytrue - ypredleading to wrong values when calling- __call__. #8087 by Alexis Mignon
- Fix a bug where - ensemble.VotingClassifierraises an error when a numpy array is passed in for weights. #7983 by Vincent Pham.
- Fixed a bug where - tree.export_graphvizraised an error when the length of features_names does not match n_features in the decision tree. #8512 by Li Li.
Linear, kernelized and related models
- Fixed a bug where - linear_model.RANSACRegressor.fitmay run until- max_iterif it finds a large inlier group early. #8251 by @aivision2020.
- Fixed a bug where - naive_bayes.MultinomialNBand- naive_bayes.BernoulliNBfailed when- alpha=0. #5814 by Yichuan Liu and Herilalaina Rakotoarison.
- Fixed a bug where - linear_model.LassoLarsdoes not give the same result as the LassoLars implementation available in R (lars library). #7849 by Jair Montoya Martinez.
- Fixed a bug in - linear_model.RandomizedLasso,- linear_model.Lars,- linear_model.LassoLars,- linear_model.LarsCVand- linear_model.LassoLarsCV, where the parameter- precomputewas not used consistently across classes, and some values proposed in the docstring could raise errors. #5359 by Tom Dupre la Tour.
- Fix inconsistent results between - linear_model.RidgeCVand- linear_model.Ridgewhen using- normalize=True. #9302 by Alexandre Gramfort.
- Fix a bug where - linear_model.LassoLars.fitsometimes left- coef_as a list, rather than an ndarray. #8160 by CJ Carey.
- Fix - linear_model.BayesianRidge.fitto return ridge parameter- alpha_and- lambda_consistent with calculated coefficients- coef_and- intercept_. #8224 by Peter Gedeck.
- Fixed a bug in - svm.OneClassSVMwhere it returned floats instead of integer classes. #8676 by Vathsala Achar.
- Fix AIC/BIC criterion computation in - linear_model.LassoLarsIC. #9022 by Alexandre Gramfort and Mehmet Basbug.
- Fixed a memory leak in our LibLinear implementation. #9024 by Sergei Lebedev 
- Fix bug where stratified CV splitters did not work with - linear_model.LassoCV. #8973 by Paulo Haddad.
- Fixed a bug in - gaussian_process.GaussianProcessRegressorwhen the standard deviation and covariance predicted without fit would fail with a meaningless error by default. #6573 by Quazi Marufur Rahman and Manoj Kumar.
Other predictors
- Fix - semi_supervised.BaseLabelPropagationto correctly implement- LabelPropagationand- LabelSpreadingas done in the referenced papers. #9239 by Andre Ambrosio Boechat, Utkarsh Upadhyay, and Joel Nothman.
Decomposition, manifold learning and clustering
- Fixed the implementation of - manifold.TSNE:
- early_exaggerationparameter had no effect and is now used for the first 250 optimization iterations.
- Fixed the - AssertionError: Tree consistency failedexception reported in #8992.
- Improve the learning schedule to match the one from the reference implementation lvdmaaten/bhtsne. by Thomas Moreau and Olivier Grisel. 
- Fix a bug in - decomposition.LatentDirichletAllocationwhere the- perplexitymethod was returning incorrect results because the- transformmethod returns normalized document topic distributions as of version 0.18. #7954 by Gary Foreman.
- Fix output shape and bugs with n_jobs > 1 in - decomposition.SparseCodertransform and- decomposition.sparse_encodefor one-dimensional data and one component. This also impacts the output shape of- decomposition.DictionaryLearning. #8086 by Andreas Müller.
- Fixed the implementation of - explained_variance_in- decomposition.PCA,- decomposition.RandomizedPCAand- decomposition.IncrementalPCA. #9105 by Hanmin Qin.
- Fixed the implementation of - noise_variance_in- decomposition.PCA. #9108 by Hanmin Qin.
- Fixed a bug where - cluster.DBSCANgives incorrect result when input is a precomputed sparse matrix with initial rows all zero. #8306 by Akshay Gupta
- Fix a bug regarding fitting - cluster.KMeanswith a sparse array X and initial centroids, where X’s means were unnecessarily being subtracted from the centroids. #7872 by Josh Karnofsky.
- Fixes to the input validation in - covariance.EllipticEnvelope. #8086 by Andreas Müller.
- Fixed a bug in - covariance.MinCovDetwhere inputting data that produced a singular covariance matrix would cause the helper method- _c_stepto throw an exception. #3367 by Jeremy Steward
- Fixed a bug in - manifold.TSNEaffecting convergence of the gradient descent. #8768 by David DeTomaso.
- Fixed a bug in - manifold.TSNEwhere it stored the incorrect- kl_divergence_. #6507 by Sebastian Saeger.
- Fixed improper scaling in - cross_decomposition.PLSRegressionwith- scale=True. #7819 by jayzed82.
- cluster.SpectralCoclusteringand- cluster.SpectralBiclustering- fitmethod conforms with API by accepting- yand returning the object. #6126, #7814 by Laurent Direr and Maniteja Nandana.
- Fix bug where - sklearn.mixture- samplemethods did not return as many samples as requested. #7702 by Levi John Wolf.
- Fixed the shrinkage implementation in - neighbors.NearestCentroid. #9219 by Hanmin Qin.
Preprocessing and feature selection
- For sparse matrices, - preprocessing.normalizewith- return_norm=Truewill now raise a- NotImplementedErrorwith ‘l1’ or ‘l2’ norm and with norm ‘max’ the norms returned will be the same as for dense matrices. #7771 by Ang Lu.
- Fix a bug where - feature_selection.SelectFdrdid not exactly implement Benjamini-Hochberg procedure. It formerly may have selected fewer features than it should. #7490 by Peng Meng.
- Fixed a bug where - linear_model.RandomizedLassoand- linear_model.RandomizedLogisticRegressionbreak for sparse input. #8259 by Aman Dalmia.
- Fix a bug where - feature_extraction.FeatureHashermandatorily applied a sparse random projection to the hashed features, preventing the use of- feature_extraction.text.HashingVectorizerin a pipeline with- feature_extraction.text.TfidfTransformer. #7565 by Roman Yurchak.
- Fix a bug where - feature_selection.mutual_info_regressiondid not correctly use- n_neighbors. #8181 by Guillaume Lemaitre.
Model evaluation and meta-estimators
- Fixed a bug where - model_selection.BaseSearchCV.inverse_transformreturns- self.best_estimator_.transform()instead of- self.best_estimator_.inverse_transform(). #8344 by Akshay Gupta and Rasmus Eriksson.
- Added - classes_attribute to- model_selection.GridSearchCV,- model_selection.RandomizedSearchCV,- grid_search.GridSearchCV, and- grid_search.RandomizedSearchCVthat matches the- classes_attribute of- best_estimator_. #7661 and #8295 by Alyssa Batula, Dylan Werner-Meier, and Stephen Hoover.
- Fixed a bug where - model_selection.validation_curvereused the same estimator for each parameter value. #7365 by Aleksandr Sandrovskii.
- model_selection.permutation_test_scorenow works with Pandas types. #5697 by Stijn Tonk.
- Several fixes to input validation in - multiclass.OutputCodeClassifier#8086 by Andreas Müller.
- multiclass.OneVsOneClassifier’s- partial_fitnow ensures all classes are provided up-front. #6250 by Asish Panda.
- Fix - multioutput.MultiOutputClassifier.predict_probato return a list of 2d arrays, rather than a 3d array. In the case where different target columns had different numbers of classes, a- ValueErrorwould be raised on trying to stack matrices with different dimensions. #8093 by Peter Bull.
- Cross validation now works with Pandas datatypes that have a read-only index. #9507 by Loic Esteve. 
Metrics
- metrics.average_precision_scoreno longer linearly interpolates between operating points, and instead weighs precisions by the change in recall since the last operating point, as per the Wikipedia entry. (#7356). By Nick Dingwall and Gael Varoquaux.
- Fix a bug in - metrics.classification._check_targetswhich would return- 'binary'if- y_trueand- y_predwere both- 'binary'but the union of- y_trueand- y_predwas- 'multiclass'. #8377 by Loic Esteve.
- Fixed an integer overflow bug in - metrics.confusion_matrixand hence- metrics.cohen_kappa_score. #8354, #7929 by Joel Nothman and Jon Crall.
- Fixed passing of - gammaparameter to the- chi2kernel in- metrics.pairwise.pairwise_kernels#5211 by Nick Rhinehart, Saurabh Bansod and Andreas Müller.
Miscellaneous
- Fixed a bug when - datasets.make_classificationfails when generating more than 30 features. #8159 by Herilalaina Rakotoarison.
- Fixed a bug where - datasets.make_moonsgives an incorrect result when- n_samplesis odd. #8198 by Josh Levy.
- Some - fetch_functions in- sklearn.datasetswere ignoring the- download_if_missingkeyword. #7944 by Ralf Gommers.
- Fix estimators to accept a - sample_weightparameter of type- pandas.Seriesin their- fitfunction. #7825 by Kathleen Chen.
- Fix a bug in cases where - numpy.cumsummay be numerically unstable, raising an exception if instability is identified. #7376 and #7331 by Joel Nothman and @yangarbiter.
- Fix a bug where - base.BaseEstimator.__getstate__obstructed pickling customizations of child-classes, when used in a multiple inheritance context. #8316 by Holger Peters.
- Update Sphinx-Gallery from 0.1.4 to 0.1.7 for resolving links in documentation build with Sphinx>1.5 #8010, #7986 by Oscar Najera 
- Add - data_homeparameter to- sklearn.datasets.fetch_kddcup99. #9289 by Loic Esteve.
- Fix dataset loaders using Python 3 version of makedirs to also work in Python 2. #9284 by Sebastin Santy. 
- Several minor issues were fixed with thanks to the alerts of lgtm.com. #9278 by Jean Helie, among others. 
API changes summary#
Trees and ensembles
- Gradient boosting base models are no longer estimators. By Andreas Müller. 
- All tree-based estimators now accept a - min_impurity_decreaseparameter in lieu of the- min_impurity_split, which is now deprecated. The- min_impurity_decreasehelps stop splitting the nodes in which the weighted impurity decrease from splitting is no longer at least- min_impurity_decrease. #8449 by Raghav RV.
Linear, kernelized and related models
- n_iterparameter is deprecated in- linear_model.SGDClassifier,- linear_model.SGDRegressor,- linear_model.PassiveAggressiveClassifier,- linear_model.PassiveAggressiveRegressorand- linear_model.Perceptron. By Tom Dupre la Tour.
Other predictors
- neighbors.LSHForesthas been deprecated and will be removed in 0.21 due to poor performance. #9078 by Laurent Direr.
- neighbors.NearestCentroidno longer purports to support- metric='precomputed'which now raises an error. #8515 by Sergul Aydore.
- The - alphaparameter of- semi_supervised.LabelPropagationnow has no effect and is deprecated to be removed in 0.21. #9239 by Andre Ambrosio Boechat, Utkarsh Upadhyay, and Joel Nothman.
Decomposition, manifold learning and clustering
- Deprecate the - doc_topic_distrargument of the- perplexitymethod in- decomposition.LatentDirichletAllocationbecause the user no longer has access to the unnormalized document topic distribution needed for the perplexity calculation. #7954 by Gary Foreman.
- The - n_topicsparameter of- decomposition.LatentDirichletAllocationhas been renamed to- n_componentsand will be removed in version 0.21. #8922 by @Attractadore.
- decomposition.SparsePCA.transform’s- ridge_alphaparameter is deprecated in preference for class parameter. #8137 by Naoya Kanai.
- cluster.DBSCANnow has a- metric_paramsparameter. #8139 by Naoya Kanai.
Preprocessing and feature selection
- feature_selection.SelectFromModelnow has a- partial_fitmethod only if the underlying estimator does. By Andreas Müller.
- feature_selection.SelectFromModelnow validates the- thresholdparameter and sets the- threshold_attribute during the call to- fit, and no longer during the call to- transform. By Andreas Müller.
- The - non_negativeparameter in- feature_extraction.FeatureHasherhas been deprecated, and replaced with a more principled alternative,- alternate_sign. #7565 by Roman Yurchak.
- linear_model.RandomizedLogisticRegression, and- linear_model.RandomizedLassohave been deprecated and will be removed in version 0.21. #8995 by Ramana.S.
Model evaluation and meta-estimators
- Deprecate the - fit_paramsconstructor input to the- model_selection.GridSearchCVand- model_selection.RandomizedSearchCVin favor of passing keyword parameters to the- fitmethods of those classes. Data-dependent parameters needed for model training should be passed as keyword arguments to- fit, and conforming to this convention will allow the hyperparameter selection classes to be used with tools such as- model_selection.cross_val_predict. #2879 by Stephen Hoover.
- In version 0.21, the default behavior of splitters that use the - test_sizeand- train_sizeparameter will change, such that specifying- train_sizealone will cause- test_sizeto be the remainder. #7459 by Nelson Liu.
- multiclass.OneVsRestClassifiernow has- partial_fit,- decision_functionand- predict_probamethods only when the underlying estimator does. #7812 by Andreas Müller and Mikhail Korobov.
- multiclass.OneVsRestClassifiernow has a- partial_fitmethod only if the underlying estimator does. By Andreas Müller.
- The - decision_functionoutput shape for binary classification in- multiclass.OneVsRestClassifierand- multiclass.OneVsOneClassifieris now- (n_samples,)to conform to scikit-learn conventions. #9100 by Andreas Müller.
- The - multioutput.MultiOutputClassifier.predict_probafunction used to return a 3d array (- n_samples,- n_classes,- n_outputs). In the case where different target columns had different numbers of classes, a- ValueErrorwould be raised on trying to stack matrices with different dimensions. This function now returns a list of arrays where the length of the list is- n_outputs, and each array is (- n_samples,- n_classes) for that particular output. #8093 by Peter Bull.
- Replace attribute - named_steps- dictto- utils.Bunchin- pipeline.Pipelineto enable tab completion in interactive environment. In the case conflict value on- named_stepsand- dictattribute,- dictbehavior will be prioritized. #8481 by Herilalaina Rakotoarison.
Miscellaneous
- Deprecate the - yparameter in- transformand- inverse_transform. The method should not accept- yparameter, as it’s used at the prediction time. #8174 by Tahar Zanouda, Alexandre Gramfort and Raghav RV.
- SciPy >= 0.13.3 and NumPy >= 1.8.2 are now the minimum supported versions for scikit-learn. The following backported functions in - sklearn.utilshave been removed or deprecated accordingly. #8854 and #8874 by Naoya Kanai
- The - store_covariancesand- covariances_parameters of- discriminant_analysis.QuadraticDiscriminantAnalysishave been renamed to- store_covarianceand- covariance_to be consistent with the corresponding parameter names of the- discriminant_analysis.LinearDiscriminantAnalysis. They will be removed in version 0.21. #7998 by Jiacheng- Removed in 0.19: - utils.fixes.argpartition
- utils.fixes.array_equal
- utils.fixes.astype
- utils.fixes.bincount
- utils.fixes.expit
- utils.fixes.frombuffer_empty
- utils.fixes.in1d
- utils.fixes.norm
- utils.fixes.rankdata
- utils.fixes.safe_copy
 - Deprecated in 0.19, to be removed in 0.21: - utils.arpack.eigs
- utils.arpack.eigsh
- utils.arpack.svds
- utils.extmath.fast_dot
- utils.extmath.logsumexp
- utils.extmath.norm
- utils.extmath.pinvh
- utils.graph.graph_laplacian
- utils.random.choice
- utils.sparsetools.connected_components
- utils.stats.rankdata
 
- Estimators with both methods - decision_functionand- predict_probaare now required to have a monotonic relation between them. The method- check_decision_proba_consistencyhas been added in utils.estimator_checks to check their consistency. #7578 by Shubham Bhardwaj
- All checks in - utils.estimator_checks, in particular- utils.estimator_checks.check_estimatornow accept estimator instances. Most other checks do not accept estimator classes any more. #9019 by Andreas Müller.
- Ensure that estimators’ attributes ending with - _are not set in the constructor but only in the- fitmethod. Most notably, ensemble estimators (deriving from- ensemble.BaseEnsemble) now only have- self.estimators_available after- fit. #7464 by Lars Buitinck and Loic Esteve.
Code and Documentation Contributors#
Thanks to everyone who has contributed to the maintenance and improvement of the project since version 0.18, including:
Joel Nothman, Loic Esteve, Andreas Mueller, Guillaume Lemaitre, Olivier Grisel, Hanmin Qin, Raghav RV, Alexandre Gramfort, themrmax, Aman Dalmia, Gael Varoquaux, Naoya Kanai, Tom Dupré la Tour, Rishikesh, Nelson Liu, Taehoon Lee, Nelle Varoquaux, Aashil, Mikhail Korobov, Sebastin Santy, Joan Massich, Roman Yurchak, RAKOTOARISON Herilalaina, Thierry Guillemot, Alexandre Abadie, Carol Willing, Balakumaran Manoharan, Josh Karnofsky, Vlad Niculae, Utkarsh Upadhyay, Dmitry Petrov, Minghui Liu, Srivatsan, Vincent Pham, Albert Thomas, Jake VanderPlas, Attractadore, JC Liu, alexandercbooth, chkoar, Óscar Nájera, Aarshay Jain, Kyle Gilliam, Ramana Subramanyam, CJ Carey, Clement Joudet, David Robles, He Chen, Joris Van den Bossche, Karan Desai, Katie Luangkote, Leland McInnes, Maniteja Nandana, Michele Lacchia, Sergei Lebedev, Shubham Bhardwaj, akshay0724, omtcyfz, rickiepark, waterponey, Vathsala Achar, jbDelafosse, Ralf Gommers, Ekaterina Krivich, Vivek Kumar, Ishank Gulati, Dave Elliott, ldirer, Reiichiro Nakano, Levi John Wolf, Mathieu Blondel, Sid Kapur, Dougal J. Sutherland, midinas, mikebenfield, Sourav Singh, Aseem Bansal, Ibraim Ganiev, Stephen Hoover, AishwaryaRK, Steven C. Howell, Gary Foreman, Neeraj Gangwar, Tahar, Jon Crall, dokato, Kathy Chen, ferria, Thomas Moreau, Charlie Brummitt, Nicolas Goix, Adam Kleczewski, Sam Shleifer, Nikita Singh, Basil Beirouti, Giorgio Patrini, Manoj Kumar, Rafael Possas, James Bourbeau, James A. Bednar, Janine Harper, Jaye, Jean Helie, Jeremy Steward, Artsiom, John Wei, Jonathan LIgo, Jonathan Rahn, seanpwilliams, Arthur Mensch, Josh Levy, Julian Kuhlmann, Julien Aubert, Jörn Hees, Kai, shivamgargsya, Kat Hempstalk, Kaushik Lakshmikanth, Kennedy, Kenneth Lyons, Kenneth Myers, Kevin Yap, Kirill Bobyrev, Konstantin Podshumok, Arthur Imbert, Lee Murray, toastedcornflakes, Lera, Li Li, Arthur Douillard, Mainak Jas, tobycheese, Manraj Singh, Manvendra Singh, Marc Meketon, MarcoFalke, Matthew Brett, Matthias Gilch, Mehul Ahuja, Melanie Goetz, Meng, Peng, Michael Dezube, Michal Baumgartner, vibrantabhi19, Artem Golubin, Milen Paskov, Antonin Carette, Morikko, MrMjauh, NALEPA Emmanuel, Namiya, Antoine Wendlinger, Narine Kokhlikyan, NarineK, Nate Guerin, Angus Williams, Ang Lu, Nicole Vavrova, Nitish Pandey, Okhlopkov Daniil Olegovich, Andy Craze, Om Prakash, Parminder Singh, Patrick Carlson, Patrick Pei, Paul Ganssle, Paulo Haddad, Paweł Lorek, Peng Yu, Pete Bachant, Peter Bull, Peter Csizsek, Peter Wang, Pieter Arthur de Jong, Ping-Yao, Chang, Preston Parry, Puneet Mathur, Quentin Hibon, Andrew Smith, Andrew Jackson, 1kastner, Rameshwar Bhaskaran, Rebecca Bilbro, Remi Rampin, Andrea Esuli, Rob Hall, Robert Bradshaw, Romain Brault, Aman Pratik, Ruifeng Zheng, Russell Smith, Sachin Agarwal, Sailesh Choyal, Samson Tan, Samuël Weber, Sarah Brown, Sebastian Pölsterl, Sebastian Raschka, Sebastian Saeger, Alyssa Batula, Abhyuday Pratap Singh, Sergey Feldman, Sergul Aydore, Sharan Yalburgi, willduan, Siddharth Gupta, Sri Krishna, Almer, Stijn Tonk, Allen Riddell, Theofilos Papapanagiotou, Alison, Alexis Mignon, Tommy Boucher, Tommy Löfstedt, Toshihiro Kamishima, Tyler Folkman, Tyler Lanigan, Alexander Junge, Varun Shenoy, Victor Poughon, Vilhelm von Ehrenheim, Aleksandr Sandrovskii, Alan Yee, Vlasios Vasileiou, Warut Vijitbenjaronk, Yang Zhang, Yaroslav Halchenko, Yichuan Liu, Yuichi Fujikawa, affanv14, aivision2020, xor, andreh7, brady salz, campustrampus, Agamemnon Krasoulis, ditenberg, elena-sharova, filipj8, fukatani, gedeck, guiniol, guoci, hakaa1, hongkahjun, i-am-xhy, jakirkham, jaroslaw-weber, jayzed82, jeroko, jmontoyam, jonathan.striebel, josephsalmon, jschendel, leereeves, martin-hahn, mathurinm, mehak-sachdeva, mlewis1729, mlliou112, mthorrell, ndingwall, nuffe, yangarbiter, plagree, pldtc325, Breno Freitas, Brett Olsen, Brian A. Alfano, Brian Burns, polmauri, Brandon Carter, Charlton Austin, Chayant T15h, Chinmaya Pancholi, Christian Danielsen, Chung Yen, Chyi-Kwei Yau, pravarmahajan, DOHMATOB Elvis, Daniel LeJeune, Daniel Hnyk, Darius Morawiec, David DeTomaso, David Gasquez, David Haberthür, David Heryanto, David Kirkby, David Nicholson, rashchedrin, Deborah Gertrude Digges, Denis Engemann, Devansh D, Dickson, Bob Baxley, Don86, E. Lynch-Klarup, Ed Rogers, Elizabeth Ferriss, Ellen-Co2, Fabian Egli, Fang-Chieh Chou, Bing Tian Dai, Greg Stupp, Grzegorz Szpak, Bertrand Thirion, Hadrien Bertrand, Harizo Rajaona, zxcvbnius, Henry Lin, Holger Peters, Icyblade Dai, Igor Andriushchenko, Ilya, Isaac Laughlin, Iván Vallés, Aurélien Bellet, JPFrancoia, Jacob Schreiber, Asish Mahapatra
