In [1]:
from sklearn_benchmarks.reporting.hp_match import HpMatchReporting
from sklearn_benchmarks.utils import default_results_directory
from pathlib import Path
import pandas as pd
pd.set_option('display.max_colwidth', None)
pd.set_option('display.max_columns', None)
pd.set_option('display.max_rows', None)

ONNX Runtime vs. scikit-learn¶

In [2]:
results_dir = default_results_directory()
In [3]:
# Parameters
results_dir = "./results/local/20220314T000704/"
In [4]:
results_dir = Path(results_dir)
In [5]:
reporting = HpMatchReporting(other_library="onnx", config="config.yml", log_scale=True, results_dir=results_dir)
reporting.make_report()

We assume here there is a perfect match between the hyperparameters of both librairies. For a given set of parameters and a given dataset, we compute the speed-up time scikit-learn / time onnx. For instance, a speed-up of 2 means that onnx is twice as fast as scikit-learn for a given set of parameters and a given dataset.

¶

onnx (1.11.0) vs. scikit-learn (1.0.2)

Speedup barplots ¶

All estimators share the following parameters: algorithm=brute.

Raw results ¶

predict

function n_samples_train n_samples n_features mean_duration_sklearn std_duration_sklearn iteration_throughput latency n_jobs n_neighbors accuracy_score_sklearn mean_duration_onnx std_duration_onnx accuracy_score_onnx speedup std_speedup sklearn_profiling onnx_profiling
0 predict 100000 1000 100 2.244 0.228 0.000 0.002 -1 1 0.676 14.968 0.018 0.676 0.150 0.150 Download Download
1 predict 100000 1 100 0.021 0.002 0.000 0.021 -1 1 0.000 0.285 0.001 0.000 0.074 0.074 Download Download
2 predict 100000 1000 100 2.865 0.089 0.000 0.003 -1 5 0.743 14.935 0.034 0.743 0.192 0.192 Download Download
3 predict 100000 1 100 0.021 0.002 0.000 0.021 -1 5 1.000 0.284 0.002 1.000 0.075 0.075 Download Download
4 predict 100000 1000 100 1.861 0.003 0.000 0.002 1 100 0.846 15.070 0.022 0.846 0.124 0.124 Download Download
5 predict 100000 1 100 0.018 0.000 0.000 0.018 1 100 1.000 0.285 0.002 1.000 0.064 0.064 Download Download
6 predict 100000 1000 100 2.621 0.048 0.000 0.003 -1 100 0.846 15.003 0.073 0.846 0.175 0.175 Download Download
7 predict 100000 1 100 0.021 0.002 0.000 0.021 -1 100 1.000 0.284 0.002 1.000 0.076 0.076 Download Download
8 predict 100000 1000 100 1.843 0.007 0.000 0.002 1 5 0.743 15.076 0.041 0.743 0.122 0.122 Download Download
9 predict 100000 1 100 0.018 0.000 0.000 0.018 1 5 1.000 0.286 0.002 1.000 0.065 0.065 Download Download
10 predict 100000 1000 100 1.140 0.004 0.001 0.001 1 1 0.676 15.176 0.057 0.676 0.075 0.075 Download Download
11 predict 100000 1 100 0.020 0.004 0.000 0.020 1 1 0.000 0.285 0.002 0.000 0.070 0.070 Download Download
12 predict 100000 1000 2 1.691 0.030 0.000 0.002 -1 1 0.845 3.651 0.007 0.845 0.463 0.463 Download Download
13 predict 100000 1 2 0.005 0.004 0.000 0.005 -1 1 1.000 0.212 0.001 1.000 0.025 0.025 Download Download
14 predict 100000 1000 2 2.447 0.024 0.000 0.002 -1 5 0.883 3.581 0.011 0.883 0.683 0.683 Download Download
15 predict 100000 1 2 0.006 0.004 0.000 0.006 -1 5 1.000 0.213 0.002 1.000 0.029 0.029 Download Download
16 predict 100000 1000 2 1.799 0.002 0.000 0.002 1 100 0.887 3.630 0.004 0.887 0.496 0.496 Download Download
17 predict 100000 1 2 0.003 0.000 0.000 0.003 1 100 1.000 0.212 0.001 1.000 0.013 0.013 Download Download
18 predict 100000 1000 2 2.465 0.021 0.000 0.002 -1 100 0.887 3.629 0.023 0.887 0.679 0.679 Download Download
19 predict 100000 1 2 0.009 0.005 0.000 0.009 -1 100 1.000 0.215 0.006 1.000 0.040 0.040 Download Download
20 predict 100000 1000 2 1.788 0.003 0.000 0.002 1 5 0.883 3.588 0.013 0.883 0.498 0.498 Download Download
21 predict 100000 1 2 0.003 0.000 0.000 0.003 1 5 1.000 0.212 0.001 1.000 0.012 0.012 Download Download
22 predict 100000 1000 2 1.029 0.002 0.000 0.001 1 1 0.845 3.634 0.005 0.845 0.283 0.283 Download Download
23 predict 100000 1 2 0.002 0.000 0.000 0.002 1 1 1.000 0.211 0.001 1.000 0.008 0.008 Download Download

¶

onnx (1.11.0) vs. scikit-learn (1.0.2)

Speedup barplots ¶

All estimators share the following parameters: learning_rate=0.01, n_iter_no_change=10.0, max_leaf_nodes=100.0, max_bins=255.0, min_samples_leaf=100.0, max_iter=300.0.

Raw results ¶

predict

function n_samples_train n_samples n_features mean_duration_sklearn std_duration_sklearn iteration_throughput latency accuracy_score_sklearn mean_duration_onnx std_duration_onnx accuracy_score_onnx speedup std_speedup sklearn_profiling onnx_profiling
0 predict 100000 1000 100 0.114 0.001 0.007 0.0 0.795 0.475 0.007 0.795 0.239 0.239 Download Download

Benchmark environment information¶

System¶

python 3.8.12 | packaged by conda-forge | (default, Jan 30 2022, 23:42:07) [GCC 9.4.0]
executable /usr/share/miniconda/envs/sklbench/bin/python
machine Linux-5.11.0-1028-azure-x86_64-with-glibc2.10

Dependencies¶

version
pip 22.0.4
setuptools 60.9.3
sklearn 1.0.2
numpy 1.22.3
scipy 1.8.0
Cython None
pandas 1.4.1
matplotlib 3.5.1
joblib 1.1.0
threadpoolctl 3.1.0

Threadpool¶

user_api internal_api prefix filepath version threading_layer architecture num_threads
0 blas openblas libopenblas /usr/share/miniconda/envs/sklbench/lib/libopenblasp-r0.3.18.so 0.3.18 pthreads SkylakeX 2
1 openmp openmp libgomp /usr/share/miniconda/envs/sklbench/lib/libgomp.so.1.0.0 None NaN NaN 2

CPU count¶

cpu_count 2
physical_cpu_count 2