You must prepare your raw data using data transforms prior to fitting a machine learning model.
This is required to ensure that you best expose the structure of your predictive modeling problem to the learning algorithms.
Applying data transforms like scaling or encoding categorical variables is straightforward when all input variables are the same type.
It can be challenging when you have a dataset with mixed types and you want to selectively apply data transforms to some, but not all, input features.
Thankfully, the scikit-learn Python machine learning library provides the ColumnTransformer that allows you to selectively apply data transforms to different columns in your dataset.
In this tutorial, you will discover how to use the ColumnTransformer to selectively apply data transforms to columns in a dataset with mixed data types.
After completing this tutorial, you will know:Let’s get started.
Use the ColumnTransformer for Numerical and Categorical Data in PythonPhoto by Kari, some rights reserved.
This tutorial is divided into three parts; they are:It is important to prepare data prior to modeling.
This may involve replacing missing values, scaling numerical values, and one hot encoding categorical data.
Data transforms can be performed using the scikit-learn library; for example, the SimpleImputer class can be used to replace missing values, the MinMaxScaler class can be used to scale numerical values, and the OneHotEncoder can be used to encode categorical variables.
For example:Sequences of different transforms can also be chained together using the Pipeline, such as imputing missing values, then scaling numerical values.
For example:It is very common to want to perform different data preparation techniques on different columns in your input data.
For example, you may want to impute missing numerical values with a median value, then scale the values and impute missing categorical values using the most frequent value and one hot encode the categories.
Traditionally, this would require you to separate the numerical and categorical data and then manually apply the transforms on those groups of features before combining the columns back together in order to fit and evaluate a model.
Now, you can use the ColumnTransformer to perform this operation for you.
The ColumnTransformer is a class in the scikit-learn Python machine learning library that allows you to selectively apply data preparation transforms.
For example, it allows you to apply a specific transform or sequence of transforms to just the numerical columns, and a separate sequence of transforms to just the categorical columns.
To use the ColumnTransformer, you must specify a list of transformers.
Each transformer is a three-element tuple that defines the name of the transformer, the transform to apply, and the column indices to apply it to.
For example:For example, the ColumnTransformer below applies a OneHotEncoder to columns 0 and 1.
The example below applies a SimpleImputer with median imputing for numerical columns 0 and 1, and SimpleImputer with most frequent imputing to categorical columns 2 and 3.
Any columns not specified in the list of “transformers” are dropped from the dataset by default; this can be changed by setting the “remainder” argument.
Setting remainder=’passthrough’ will mean that all columns not specified in the list of “transformers” will be passed through without transformation, instead of being dropped.
For example, if columns 0 and 1 were numerical and columns 2 and 3 were categorical and we wanted to just transform the categorical data and pass through the numerical columns unchanged, we could define the ColumnTransformer as follows:Once the transformer is defined, it can be used to transform a dataset.
For example:A ColumnTransformer can also be used in a Pipeline to selectively prepare the columns of your dataset before fitting a model on the transformed data.
This is the most likely use case as it ensures that the transforms are performed automatically on the raw data when fitting the model and when making predictions, such as when evaluating the model on a test dataset via cross-validation or making predictions on new data in the future.
For example:Now that we are familiar with how to configure and use the ColumnTransformer in general, let’s look at a worked example.
The abalone dataset is a standard machine learning problem that involves predicting the age of an abalone given measurements of an abalone.
You can download the dataset and learn more about it here:The dataset has 4,177 examples, 8 input variables, and the target variable is an integer.
A naive model can achieve a mean absolute error (MAE) of about 2.
363 (std 0.
092) by predicting the mean value, evaluated via 10-fold cross-validation.
We can model this as a regression predictive modeling problem with a support vector machine model (SVR).
Reviewing the data, you can see the first few rows as follows:We can see that the first column is categorical and the remainder of the columns are numerical.
We may want to one hot encode the first column and normalize the remaining numerical columns, and this can be achieved using the ColumnTransformer.
First, we need to load the dataset.
We can load the dataset directly from the URL using the read_csv() Pandas function, then split the data into two data frames: one for input and one for the output.
The complete example of loading the dataset is listed below.
Note: if you have trouble loading the dataset from a URL, you can download the CSV file with the name ‘abalone.
csv‘ and place it in the same directory as your Python file and change the call to read_csv() as follows:Running the example, we can see that the dataset is loaded correctly and split into eight input columns and one target column.
Next, we can use the select_dtypes() function to select the column indexes that match different data types.
We are interested in a list of columns that are numerical columns marked as ‘float64‘ or ‘int64‘ in Pandas, and a list of categorical columns, marked as ‘object‘ or ‘bool‘ type in Pandas.
We can then use these lists in the ColumnTransformer to one hot encode the categorical variables, which should just be the first column.
We can also use the list of numerical columns to normalize the remaining data.
Next, we can define our SVR model and define a Pipeline that first uses the ColumnTransformer, then fits the model on the prepared dataset.
Finally, we can evaluate the model using 10-fold cross-validation and calculate the mean absolute error, averaged across all 10 evaluations of the pipeline.
Tying this all together, the complete example is listed below.
Running the example evaluates the data preparation pipeline using 10-fold cross-validation.
Your specific results may vary given the stochastic learning algorithm and differences in library versions.
In this case, we achieve an average MAE of about 1.
4, which is better than the baseline score of 2.
You now have a template for using the ColumnTransformer on a dataset with mixed data types that you can use and adapt for your own projects in the future.
This section provides more resources on the topic if you are looking to go deeper.
In this tutorial, you discovered how to use the ColumnTransformer to selectively apply data transforms to columns in datasets with mixed data types.
Specifically, you learned:Do you have any questions?.Ask your questions in the comments below and I will do my best to answer.
with just a few lines of scikit-learn codeLearn how in my new Ebook: Machine Learning Mastery With PythonCovers self-study tutorials and end-to-end projects like: Loading data, visualization, modeling, tuning, and much more.
Skip the Academics.
.. More details