paired_ttest_kfold_cv: Kfold crossvalidated paired t test
Kfold paired t test procedure to compare the performance of two models
from mlxtend.evaluate import paired_ttest_kfold_cv
Overview
Kfold crossvalidated paired ttest procedure is a common method for comparing the performance of two models (classifiers or regressors) and addresses some of the drawbacks of the resampled ttest procedure; however, this method has still the problem that the training sets overlap and is not recommended to be used in practice [1], and techniques such as the paired_ttest_5x2cv
should be used instead.
To explain how this method works, let's consider to estimator (e.g., classifiers) A and B. Further, we have a labeled dataset D. In the common holdout method, we typically split the dataset into 2 parts: a training and a test set. In the kfold crossvalidated paired ttest procedure, we split the test set into k parts of equal size, and each of these parts is then used for testing while the remaining k1 parts (joined together) are used for training a classifier or regressor (i.e., the standard kfold crossvalidation procedure).
In each kfold crossvalidation iteration, we then compute the difference in performance between A and B in each so that we obtain k difference measures. Now, by making the assumption that these k differences were independently drawn and follow an approximately normal distribution, we can compute the following t statistic with k1 degrees of freedom according to Student's t test, under the null hypothesis that the models A and B have equal performance:
Here, computes the difference between the model performances in the th iteration, , and represents the average difference between the classifier performances, .
Once we computed the t statistic we can compute the p value and compare it to our chosen significance level, e.g., . If the p value is smaller than , we reject the null hypothesis and accept that there is a significant difference in the two models.
The problem with this method, and the reason why it is not recommended to be used in practice, is that it violates an assumption of Student's t test [1]:
 the difference between the model performances () are not normal distributed because and are not independent
 the 's themselves are not independent because training sets overlap
References
 [1] Dietterich TG (1998) Approximate Statistical Tests for Comparing Supervised Classification Learning Algorithms. Neural Comput 10:1895–1923.
Example 1  Kfold crossvalidated paired t test
Assume we want to compare two classification algorithms, logistic regression and a decision tree algorithm:
from sklearn.linear_model import LogisticRegression
from sklearn.tree import DecisionTreeClassifier
from mlxtend.data import iris_data
from sklearn.model_selection import train_test_split
X, y = iris_data()
clf1 = LogisticRegression(random_state=1)
clf2 = DecisionTreeClassifier(random_state=1)
X_train, X_test, y_train, y_test = \
train_test_split(X, y, test_size=0.25,
random_state=123)
score1 = clf1.fit(X_train, y_train).score(X_test, y_test)
score2 = clf2.fit(X_train, y_train).score(X_test, y_test)
print('Logistic regression accuracy: %.2f%%' % (score1*100))
print('Decision tree accuracy: %.2f%%' % (score2*100))
Logistic regression accuracy: 97.37%
Decision tree accuracy: 94.74%
Note that these accuracy values are not used in the paired ttest procedure as new test/train splits are generated during the resampling procedure, the values above are just serving the purpose of intuition.
Now, let's assume a significance threshold of for rejecting the null hypothesis that both algorithms perform equally well on the dataset and conduct the kfold crossvalidated ttest:
from mlxtend.evaluate import paired_ttest_kfold_cv
t, p = paired_ttest_kfold_cv(estimator1=clf1,
estimator2=clf2,
X=X, y=y,
random_seed=1)
print('t statistic: %.3f' % t)
print('p value: %.3f' % p)
t statistic: 1.861
p value: 0.096
Since , we cannot reject the null hypothesis and may conclude that the performance of the two algorithms is not significantly different.
While it is generally not recommended to apply statistical tests multiple times without correction for multiple hypothesis testing, let us take a look at an example where the decision tree algorithm is limited to producing a very simple decision boundary that would result in a relatively bad performance:
clf2 = DecisionTreeClassifier(random_state=1, max_depth=1)
score2 = clf2.fit(X_train, y_train).score(X_test, y_test)
print('Decision tree accuracy: %.2f%%' % (score2*100))
t, p = paired_ttest_kfold_cv(estimator1=clf1,
estimator2=clf2,
X=X, y=y,
random_seed=1)
print('t statistic: %.3f' % t)
print('p value: %.3f' % p)
Decision tree accuracy: 63.16%
t statistic: 13.491
p value: 0.000
Assuming that we conducted this test also with a significance level of , we can reject the nullhypothesis that both models perform equally well on this dataset, since the pvalue () is smaller than .
API
paired_ttest_kfold_cv(estimator1, estimator2, X, y, cv=10, scoring=None, shuffle=False, random_seed=None)
Implements the kfold paired t test procedure to compare the performance of two models.
Parameters

estimator1
: scikitlearn classifier or regressor 
estimator2
: scikitlearn classifier or regressor 
X
: {arraylike, sparse matrix}, shape = [n_samples, n_features]Training vectors, where n_samples is the number of samples and n_features is the number of features.

y
: arraylike, shape = [n_samples]Target values.

cv
: int (default: 10)Number of splits and iteration for the crossvalidation procedure

scoring
: str, callable, or None (default: None)If None (default), uses 'accuracy' for sklearn classifiers and 'r2' for sklearn regressors. If str, uses a sklearn scoring metric string identifier, for example {accuracy, f1, precision, recall, roc_auc} for classifiers, {'mean_absolute_error', 'mean_squared_error'/'neg_mean_squared_error', 'median_absolute_error', 'r2'} for regressors. If a callable object or function is provided, it has to be conform with sklearn's signature
scorer(estimator, X, y)
; see https://scikitlearn.org/stable/modules/generated/sklearn.metrics.make_scorer.html for more information. 
shuffle
: bool (default: True)Whether to shuffle the dataset for generating the kfold splits.

random_seed
: int or None (default: None)Random seed for shuffling the dataset for generating the kfold splits. Ignored if shuffle=False.
Returns

t
: floatThe tstatistic

pvalue
: floatTwotailed pvalue. If the chosen significance level is larger than the pvalue, we reject the null hypothesis and accept that there are significant differences in the two compared models.
Examples
For usage examples, please see https://rasbt.github.io/mlxtend/user_guide/evaluate/paired_ttest_kfold_cv/