MinMax Scaling
A function for minmax scaling of pandas DataFrames or NumPy arrays.
from mlxtend.preprocessing import MinMaxScaling
An alternative approach to Zscore normalization (or standardization) is the socalled MinMax scaling (often also simply called "normalization"  a common cause for ambiguities). In this approach, the data is scaled to a fixed range  usually 0 to 1. The cost of having this bounded range  in contrast to standardization  is that we will end up with smaller standard deviations, which can suppress the effect of outliers.
A MinMax scaling is typically done via the following equation:
One family of algorithms that is scaleinvariant encompasses treebased learning algorithms. Let's take the general CART decision tree algorithm. Without going into much depth regarding information gain and impurity measures, we can think of the decision as "is feature x_i >= some_val?" Intuitively, we can see that it really doesn't matter on which scale this feature is (centimeters, Fahrenheit, a standardized scale  it really doesn't matter).
Some examples of algorithms where feature scaling matters are:
 knearest neighbors with an Euclidean distance measure if want all features to contribute equally
 kmeans (see knearest neighbors)
 logistic regression, SVMs, perceptrons, neural networks etc. if you are using gradient descent/ascentbased optimization, otherwise some weights will update much faster than others
 linear discriminant analysis, principal component analysis, kernel principal component analysis since you want to find directions of maximizing the variance (under the constraints that those directions/eigenvectors/principal components are orthogonal); you want to have features on the same scale since you'd emphasize variables on "larger measurement scales" more.
There are many more cases than I can possibly list here ... I always recommend you to think about the algorithm and what it's doing, and then it typically becomes obvious whether we want to scale your features or not.
In addition, we'd also want to think about whether we want to "standardize" or "normalize" (here: scaling to [0, 1] range) our data. Some algorithms assume that our data is centered at 0. For example, if we initialize the weights of a small multilayer perceptron with tanh activation units to 0 or small random values centered around zero, we want to update the model weights "equally." As a rule of thumb I'd say: When in doubt, just standardize the data, it shouldn't hurt.
Example 1  Scaling a Pandas DataFrame
import pandas as pd
s1 = pd.Series([1, 2, 3, 4, 5, 6], index=(range(6)))
s2 = pd.Series([10, 9, 8, 7, 6, 5], index=(range(6)))
df = pd.DataFrame(s1, columns=['s1'])
df['s2'] = s2
df
s1  s2  

0  1  10 
1  2  9 
2  3  8 
3  4  7 
4  5  6 
5  6  5 
from mlxtend.preprocessing import minmax_scaling
minmax_scaling(df, columns=['s1', 's2'])
s1  s2  

0  0.0  1.0 
1  0.2  0.8 
2  0.4  0.6 
3  0.6  0.4 
4  0.8  0.2 
5  1.0  0.0 
Example 2  Scaling a NumPy Array
import numpy as np
X = np.array([[1, 10], [2, 9], [3, 8],
[4, 7], [5, 6], [6, 5]])
X
array([[ 1, 10],
[ 2, 9],
[ 3, 8],
[ 4, 7],
[ 5, 6],
[ 6, 5]])
from mlxtend.preprocessing import minmax_scaling
minmax_scaling(X, columns=[0, 1])
array([[ 0. , 1. ],
[ 0.2, 0.8],
[ 0.4, 0.6],
[ 0.6, 0.4],
[ 0.8, 0.2],
[ 1. , 0. ]])
API
minmax_scaling(array, columns, min_val=0, max_val=1)
Min max scaling of pandas' DataFrames.
Parameters

array
: pandas DataFrame or NumPy ndarray, shape = [n_rows, n_columns]. 
columns
: arraylike, shape = [n_columns]Arraylike with column names, e.g., ['col1', 'col2', ...] or column indices [0, 2, 4, ...]

min_val
:int
orfloat
, optional (default=0
)minimum value after rescaling.

min_val
:int
orfloat
, optional (default=1
)maximum value after rescaling.
Returns

df_new
: pandas DataFrame object.Copy of the array or DataFrame with rescaled columns.