feature scaling machine learning code example

Example 1: what is feature scaling

'''What is feature scaling?
Feature scaling is a method used to normalize the range of independent
variables or features of data. In data processing, it is also known as 
data normalization and is generally performed during the data preprocessing
step.

Few advantages of normalizing the data are as follows:

1. It makes your training faster.
2. It prevents you from getting stuck in local optima.
3. It gives you a better error surface shape.
4. Wweight decay and bayes optimization can be done more conveniently.

Hovewer, there are few algorithms such as Logistic Regression and 
Decision Trees that are not affected by scaling of input data.

With regard to neural networks it makes a difference

Example 2: Scaling features to a range

# Scaling features to a range using MaxAbsScaler

X_train = np.array([[ 1., -1.,  2.],
                    [ 2.,  0.,  0.],
                    [ 0.,  1., -1.]])

max_abs_scaler = preprocessing.MaxAbsScaler()
X_train_maxabs = max_abs_scaler.fit_transform(X_train)
X_train_maxabs
# array([[ 0.5, -1.,  1. ],
#        [ 1. ,  0. ,  0. ],
#        [ 0. ,  1. , -0.5]])
X_test = np.array([[ -3., -1.,  4.]])
X_test_maxabs = max_abs_scaler.transform(X_test)
X_test_maxabs
# array([[-1.5, -1. ,  2. ]])
max_abs_scaler.scale_
# array([2.,  1.,  2.])

Tags:

Misc Example