In [1]:
from sklearn_benchmarks.reporting.hp_match import HpMatchReporting
from sklearn_benchmarks.utils import default_results_directory
from pathlib import Path
import pandas as pd
pd.set_option('display.max_colwidth', None)
pd.set_option('display.max_columns', None)
pd.set_option('display.max_rows', None)

ONNX Runtime vs. scikit-learn¶

In [2]:
results_dir = default_results_directory()
In [3]:
# Parameters
results_dir = "./results/local/20220313T221146/"
In [4]:
results_dir = Path(results_dir)
In [5]:
reporting = HpMatchReporting(other_library="onnx", config="config.yml", log_scale=True, results_dir=results_dir)
reporting.make_report()

We assume here there is a perfect match between the hyperparameters of both librairies. For a given set of parameters and a given dataset, we compute the speed-up time scikit-learn / time onnx. For instance, a speed-up of 2 means that onnx is twice as fast as scikit-learn for a given set of parameters and a given dataset.

¶

onnx (1.11.0) vs. scikit-learn (1.0.2)

Speedup barplots ¶

All estimators share the following parameters: algorithm=brute.

Raw results ¶

predict

function n_samples_train n_samples n_features mean_duration_sklearn std_duration_sklearn iteration_throughput latency n_jobs n_neighbors accuracy_score_sklearn mean_duration_onnx std_duration_onnx accuracy_score_onnx speedup std_speedup sklearn_profiling onnx_profiling
0 predict 100000 1000 100 1.711 0.261 0.000 0.002 -1 1 0.676 15.072 0.002 0.676 0.114 0.114 Download Download
1 predict 100000 1 100 0.018 0.001 0.000 0.018 -1 1 0.000 0.279 0.003 0.000 0.066 0.066 Download Download
2 predict 100000 1000 100 2.444 0.026 0.000 0.002 -1 5 0.743 15.089 0.045 0.743 0.162 0.162 Download Download
3 predict 100000 1 100 0.020 0.001 0.000 0.020 -1 5 1.000 0.279 0.003 1.000 0.071 0.071 Download Download
4 predict 100000 1000 100 1.960 0.003 0.000 0.002 1 100 0.846 15.356 0.082 0.846 0.128 0.128 Download Download
5 predict 100000 1 100 0.017 0.000 0.000 0.017 1 100 1.000 0.280 0.002 1.000 0.060 0.060 Download Download
6 predict 100000 1000 100 2.436 0.023 0.000 0.002 -1 100 0.846 15.296 0.093 0.846 0.159 0.159 Download Download
7 predict 100000 1 100 0.019 0.001 0.000 0.019 -1 100 1.000 0.277 0.004 1.000 0.067 0.067 Download Download
8 predict 100000 1000 100 1.941 0.011 0.000 0.002 1 5 0.743 15.311 0.019 0.743 0.127 0.127 Download Download
9 predict 100000 1 100 0.016 0.000 0.000 0.016 1 5 1.000 0.279 0.004 1.000 0.058 0.058 Download Download
10 predict 100000 1000 100 1.088 0.003 0.001 0.001 1 1 0.676 15.291 0.106 0.676 0.071 0.071 Download Download
11 predict 100000 1 100 0.017 0.002 0.000 0.017 1 1 0.000 0.278 0.003 0.000 0.062 0.062 Download Download
12 predict 100000 1000 2 1.412 0.015 0.000 0.001 -1 1 0.845 2.866 0.014 0.845 0.493 0.493 Download Download
13 predict 100000 1 2 0.004 0.001 0.000 0.004 -1 1 1.000 0.219 0.004 1.000 0.016 0.016 Download Download
14 predict 100000 1000 2 2.323 0.024 0.000 0.002 -1 5 0.883 2.802 0.013 0.883 0.829 0.829 Download Download
15 predict 100000 1 2 0.006 0.003 0.000 0.006 -1 5 1.000 0.217 0.004 1.000 0.029 0.029 Download Download
16 predict 100000 1000 2 1.899 0.005 0.000 0.002 1 100 0.887 2.870 0.017 0.887 0.662 0.662 Download Download
17 predict 100000 1 2 0.003 0.000 0.000 0.003 1 100 1.000 0.217 0.004 1.000 0.013 0.013 Download Download
18 predict 100000 1000 2 2.331 0.026 0.000 0.002 -1 100 0.887 2.859 0.027 0.887 0.815 0.815 Download Download
19 predict 100000 1 2 0.005 0.001 0.000 0.005 -1 100 1.000 0.217 0.004 1.000 0.021 0.021 Download Download
20 predict 100000 1000 2 1.884 0.002 0.000 0.002 1 5 0.883 2.807 0.016 0.883 0.671 0.671 Download Download
21 predict 100000 1 2 0.003 0.000 0.000 0.003 1 5 1.000 0.216 0.002 1.000 0.013 0.013 Download Download
22 predict 100000 1000 2 0.961 0.004 0.000 0.001 1 1 0.845 2.851 0.006 0.845 0.337 0.337 Download Download
23 predict 100000 1 2 0.002 0.000 0.000 0.002 1 1 1.000 0.217 0.003 1.000 0.008 0.008 Download Download

¶

onnx (1.11.0) vs. scikit-learn (1.0.2)

Speedup barplots ¶

All estimators share the following parameters: learning_rate=0.01, n_iter_no_change=10.0, max_leaf_nodes=100.0, max_bins=255.0, min_samples_leaf=100.0, max_iter=300.0.

Raw results ¶

predict

function n_samples_train n_samples n_features mean_duration_sklearn std_duration_sklearn iteration_throughput latency accuracy_score_sklearn mean_duration_onnx std_duration_onnx accuracy_score_onnx speedup std_speedup sklearn_profiling onnx_profiling
0 predict 100000 1000 100 0.122 0.014 0.007 0.0 0.795 0.488 0.006 0.795 0.251 0.251 Download Download

Benchmark environment information¶

System¶

python 3.8.12 | packaged by conda-forge | (default, Jan 30 2022, 23:42:07) [GCC 9.4.0]
executable /usr/share/miniconda/envs/sklbench/bin/python
machine Linux-5.11.0-1028-azure-x86_64-with-glibc2.10

Dependencies¶

version
pip 22.0.4
setuptools 60.9.3
sklearn 1.0.2
numpy 1.22.3
scipy 1.8.0
Cython None
pandas 1.4.1
matplotlib 3.5.1
joblib 1.1.0
threadpoolctl 3.1.0

Threadpool¶

user_api internal_api prefix filepath version threading_layer architecture num_threads
0 blas openblas libopenblas /usr/share/miniconda/envs/sklbench/lib/libopenblasp-r0.3.18.so 0.3.18 pthreads SkylakeX 2
1 openmp openmp libgomp /usr/share/miniconda/envs/sklbench/lib/libgomp.so.1.0.0 None NaN NaN 2

CPU count¶

cpu_count 2
physical_cpu_count 2