IterativeImputer#

class sklearn.impute.IterativeImputer(estimator=None, *, missing_values=nan, sample_posterior=false, max_iter=10, tol=0.001, n_nearest_features=None, initial_strategy='mean', fill_value=None, imputation_order='ascending', skip_complete=false, min_value=-inf, max_value=inf, verbose=0, random_state=None, add_indicator=false, keep_empty_features=false)[source]#

Multivariate imputer that estimates each feature from all the others.

A strategy for imputing missing values by modeling each feature with missing values as a function of other features in a round-robin fashion.

Read more in the User Guide.

Added in version 0.21.

Note

This estimator is still experimental for now: the predictions and the API might change without any deprecation cycle. To use it, you need to explicitly import enable_iterative_imputer:

>>> # explicitly require this experimental feature
>>> from sklearn.experimental import enable_iterative_imputer  # noqa
>>> # now you can import normally from sklearn.impute
>>> from sklearn.impute import IterativeImputer
Parameters:
estimatorestimator object, default=BayesianRidge()

The estimator to use at each step of the round-robin imputation. If sample_posterior=True, the estimator must support return_std in its predict method.

missing_valuesint or np.nan, default=np.nan

The placeholder for the missing values. All occurrences of missing_values will be imputed. for pandas’ dataframes with nullable integer dtypes with missing values, missing_values should be set to np.nan, since pd.NA will be converted to np.nan.

sample_posteriorbool, default=false

Whether to sample from the (Gaussian) predictive posterior of the fitted estimator for each imputation. Estimator must support return_std in its predict method if set to True. Set to True if using IterativeImputer for multiple imputations.

max_iterint, default=10

Maximum number of imputation rounds to perform before returning the imputations computed during the final round. A round is a single imputation of each feature with missing values. The stopping criterion is met once max(abs(X_t - X_{t-1}))/max(abs(X[known_vals])) < tol, where X_t is X at iteration t. Note that early stopping is only applied if sample_posterior=false.

tolfloat, default=1e-3

Tolerance of the stopping condition.

n_nearest_featuresint, default=None

Number of other features to use to estimate the missing values of each feature column. Nearness between features is measured using the absolute correlation coefficient between each feature pair (after initial imputation). To ensure coverage of features throughout the imputation process, the neighbor features are not necessarily nearest, but are drawn with probability proportional to correlation for each imputed target feature. Can provide significant speed-up when the number of features is huge. If None, all features will be used.

initial_strategy{‘mean’, ‘median’, ‘most_frequent’, ‘constant’}, default=’mean’

Which strategy to use to initialize the missing values. Same as the strategy parameter in SimpleImputer.

fill_valuestr or numerical value, default=None

When strategy="constant", fill_value is used to replace all occurrences of missing_values. for string or object data types, fill_value must be a string. If None, fill_value will be 0 when imputing numerical data and “missing_value” for strings or object data types.

Added in version 1.3.

imputation_order{‘ascending’, ‘descending’, ‘roman’, ‘arabic’, ‘random’}, default=’ascending’

The order in which the features will be imputed. Possible values:

  • 'ascending': from features with fewest missing values to most.

  • 'descending': from features with most missing values to fewest.

  • 'roman': Left to right.

  • 'arabic': Right to left.

  • 'random': A random order for each round.

skip_completebool, default=false

If True then features with missing values during transform which did not have any missing values during fit will be imputed with the initial imputation method only. Set to True if you have many features with no missing values at both fit and transform time to save compute.

min_valuefloat or array-like of shape (n_features,), default=-np.inf

Minimum possible imputed value. Broadcast to shape (n_features,) if scalar. If array-like, expects shape (n_features,), one min value for each feature. The default is -np.inf.

Changed in version 0.23: Added support for array-like.

max_valuefloat or array-like of shape (n_features,), default=np.inf

Maximum possible imputed value. Broadcast to shape (n_features,) if scalar. If array-like, expects shape (n_features,), one max value for each feature. The default is np.inf.

Changed in version 0.23: Added support for array-like.

verboseint, default=0

Verbosity flag, controls the debug messages that are issued as functions are evaluated. The higher, the more verbose. Can be 0, 1, or 2.

random_stateint, RandomState instance or None, default=None

The seed of the pseudo random number generator to use. Randomizes selection of estimator features if n_nearest_features is not None, the imputation_order if random, and the sampling from posterior if sample_posterior=True. Use an integer for determinism. See the Glossary.

add_indicatorbool, default=false

If True, a MissingIndicator transform will stack onto output of the imputer’s transform. This allows a predictive estimator to account for missingness despite imputation. If a feature has no missing values at fit/train time, the feature won’t appear on the missing indicator even if there are missing values at transform/test time.

keep_empty_featuresbool, default=false

If True, features that consist exclusively of missing values when fit is called are returned in results when transform is called. The imputed value is always 0 except when initial_strategy="constant" in which case fill_value will be used instead.

Added in version 1.2.

Attributes:
initial_imputer_object of type SimpleImputer

Imputer used to initialize the missing values.

imputation_sequence_list of tuples

Each tuple has (feat_idx, neighbor_feat_idx, estimator), where feat_idx is the current feature to be imputed, neighbor_feat_idx is the array of other features used to impute the current feature, and estimator is the trained estimator used for the imputation. Length is self.n_features_with_missing_ * self.n_iter_.

n_iter_int

Number of iteration rounds that occurred. Will be less than self.max_iter if early stopping criterion was reached.

n_features_in_int

Number of features seen during fit.

Added in version 0.24.

feature_names_in_ndarray of shape (n_features_in_,)

Names of features seen during fit. Defined only when X has feature names that are all strings.

Added in version 1.0.

n_features_with_missing_int

Number of features with missing values.

indicator_MissingIndicator

Indicator used to add binary indicators for missing values. None if add_indicator=false.

random_state_RandomState instance

RandomState instance that is generated either from a seed, the random number generator or by np.random.

See also

SimpleImputer

Univariate imputer for completing missing values with simple strategies.

KNNImputer

Multivariate imputer that estimates missing features using nearest samples.

Notes

To support imputation in inductive mode we store each feature’s estimator during the fit phase, and predict without refitting (in order) during the transform phase.

features which contain all missing values at fit are discarded upon transform.

Using defaults, the imputer scales in \(\mathcal{O}(knp^3\min(n,p))\) where \(k\) = max_iter, \(n\) the number of samples and \(p\) the number of features. It thus becomes prohibitively costly when the number of features increases. Setting n_nearest_features << n_features, skip_complete=True or increasing tol can help to reduce its computational cost.

Depending on the nature of missing values, simple imputers can be preferable in a prediction context.

References

Examples

>>> import numpy as np
>>> from sklearn.experimental import enable_iterative_imputer
>>> from sklearn.impute import IterativeImputer
>>> imp_mean = IterativeImputer(random_state=0)
>>> imp_mean.fit([[7, 2, 3], [4, np.nan, 6], [10, 5, 9]])
IterativeImputer(random_state=0)
>>> X = [[np.nan, 2, 3], [4, np.nan, 6], [10, np.nan, 9]]
>>> imp_mean.transform(X)
array([[ 6.9584...,  2.       ,  3.        ],
       [ 4.       ,  2.6000...,  6.        ],
       [10.       ,  4.9999...,  9.        ]])

for a more detailed example see Imputing missing values before building an estimator or Imputing missing values with variants of IterativeImputer.

fit(X, y=None, **fit_params)[source]#

fit the imputer on X and return self.

Parameters:
Xarray-like, shape (n_samples, n_features)

Input data, where n_samples is the number of samples and n_features is the number of features.

yIgnored

Not used, present for API consistency by convention.

**fit_paramsdict

Parameters routed to the fit method of the sub-estimator via the metadata routing API.

Added in version 1.5: Only available if sklearn.set_config(enable_metadata_routing=True) is set. See Metadata Routing User Guide for more details.

Returns:
selfobject

fitted estimator.

fit_transform(X, y=None, **params)[source]#

fit the imputer on X and return the transformed X.

Parameters:
Xarray-like, shape (n_samples, n_features)

Input data, where n_samples is the number of samples and n_features is the number of features.

yIgnored

Not used, present for API consistency by convention.

**paramsdict

Parameters routed to the fit method of the sub-estimator via the metadata routing API.

Added in version 1.5: Only available if sklearn.set_config(enable_metadata_routing=True) is set. See Metadata Routing User Guide for more details.

Returns:
Xtarray-like, shape (n_samples, n_features)

The imputed input data.

get_feature_names_out(input_features=None)[source]#

Get output feature names for transformation.

Parameters:
input_featuresarray-like of str or None, default=None

Input features.

  • If input_features is None, then feature_names_in_ is used as feature names in. If feature_names_in_ is not defined, then the following input feature names are generated: ["x0", "x1", ..., "x(n_features_in_ - 1)"].

  • If input_features is an array-like, then input_features must match feature_names_in_ if feature_names_in_ is defined.

Returns:
feature_names_outndarray of str objects

Transformed feature names.

get_metadata_routing()[source]#

Get metadata routing of this object.

Please check User Guide on how the routing mechanism works.

Added in version 1.5.

Returns:
routingMetadataRouter

A MetadataRouter encapsulating routing information.

get_params(deep=True)[source]#

Get parameters for this estimator.

Parameters:
deepbool, default=True

If True, will return the parameters for this estimator and contained subobjects that are estimators.

Returns:
paramsdict

Parameter names mapped to their values.

set_output(*, transform=None)[source]#

Set output container.

See Introducing the set_output API for an example on how to use the API.

Parameters:
transform{“default”, “pandas”, “polars”}, default=None

Configure output of transform and fit_transform.

  • "default": Default output format of a transformer

  • "pandas": Dataframe output

  • "polars": Polars output

  • None: Transform configuration is unchanged

Added in version 1.4: "polars" option was added.

Returns:
selfestimator instance

Estimator instance.

set_params(**params)[source]#

Set the parameters of this estimator.

The method works on simple estimators as well as on nested objects (such as Pipeline). The latter have parameters of the form <component>__<parameter> so that it’s possible to update each component of a nested object.

Parameters:
**paramsdict

Estimator parameters.

Returns:
selfestimator instance

Estimator instance.

transform(X)[source]#

Impute all missing values in X.

Note that this is stochastic, and that if random_state is not fixed, repeated calls, or permuted input, results will differ.

Parameters:
Xarray-like of shape (n_samples, n_features)

The input data to complete.

Returns:
Xtarray-like, shape (n_samples, n_features)

The imputed input data.