Rosykunai's picture
Update README.md
b2f8d91 verified
metadata
dataset_info:
  features:
    - name: MWG
      dtype: int64
    - name: NWG
      dtype: int64
    - name: KWG
      dtype: int64
    - name: MDIMC
      dtype: int64
    - name: NDIMC
      dtype: int64
    - name: MDIMA
      dtype: int64
    - name: NDIMB
      dtype: int64
    - name: KWI
      dtype: int64
    - name: VWM
      dtype: int64
    - name: VWN
      dtype: int64
    - name: STRM
      dtype: int64
    - name: STRN
      dtype: int64
    - name: SA
      dtype: int64
    - name: SB
      dtype: int64
    - name: Run_time
      dtype: float64
    - name: __index_level_0__
      dtype: int64
  splits:
    - name: train
      num_bytes: 24748672
      num_examples: 193349
  download_size: 3916323
  dataset_size: 24748672
configs:
  - config_name: default
    data_files:
      - split: train
        path: data/train-*

Information

SGEMM GPU kernel performance Data Set available for download at https://archive.ics.uci.edu/ml/datasets/SGEMM+GPU+kernel+performance

We performed some filtering on this dataset.

Here is the original information of this dataset:

This data set measures the running time of a matrix-matrix product AB = C, where all matrices have size 2048 x 2048, using a parameterizable SGEMM GPU kernel with 241600 possible parameter combinations. For each tested combination, 4 runs were performed and their results are reported as the 4 last columns. All times are measured in milliseconds.

There are 14 parameter, the first 10 are ordinal and can only take up to 4 different powers of two values, and the 4 last variables are binary. Out of 1327104 total parameter combinations, only 241600 are feasible (due to various kernel constraints). This data set contains the results for all these feasible combinations.

The experiment was run on a desktop workstation running Ubuntu 16.04 Linux with an Intel Core i5 (3.5GHz), 16GB RAM, and a NVidia Geforce GTX 680 4GB GF580 GTX-1.5GB GPU. We use the 'gemm_fast' kernel from the automatic OpenCL kernel tuning library 'CLTune' (https://github.com/CNugteren/CLTune).

  • Note: for this kind of data sets it is usually better to work with the logarithm of the running times (see e.g. Falch and Elster, 'Machine learning-based auto-tuning for enhanced performance portability of OpenCL applications', 2015).

Download

from datasets import load_dataset

dataset = load_dataset("Rosykunai/SGEMM_GPU_performance")

Reference

  • Rafael Ballester-Ripoll, Enrique G. Paredes, Renato Pajarola.
  • Sobol Tensor Trains for Global Sensitivity Analysis.
  • In arXiv Computer Science / Numerical Analysis e-prints, 2017
  • Cedric Nugteren and Valeriu Codreanu.
  • CLTune: A Generic Auto-Tuner for OpenCL Kernels.
  • In: MCSoC: 9th International Symposium on Embedded Multicore/Many-core Systems-on-Chip. IEEE, 2015