Datasets:

Modalities:
Text
Formats:
parquet
Languages:
English
Size:
< 1K
ArXiv:
Libraries:
Datasets
pandas
License:
albertvillanova's picture
Convert dataset to Parquet (#3)
7a3f932 verified
metadata
annotations_creators:
  - expert-generated
language_creators:
  - expert-generated
language:
  - en
license:
  - cc0-1.0
multilinguality:
  - monolingual
size_categories:
  - n<1K
source_datasets:
  - original
task_categories:
  - text-classification
task_ids:
  - intent-classification
paperswithcode_id: snips
pretty_name: SNIPS Natural Language Understanding benchmark
dataset_info:
  features:
    - name: text
      dtype: string
    - name: label
      dtype:
        class_label:
          names:
            '0': ComparePlaces
            '1': RequestRide
            '2': GetWeather
            '3': SearchPlace
            '4': GetPlaceDetails
            '5': ShareCurrentLocation
            '6': GetTrafficInformation
            '7': BookRestaurant
            '8': GetDirections
            '9': ShareETA
  splits:
    - name: train
      num_bytes: 19427
      num_examples: 328
  download_size: 11158
  dataset_size: 19427
configs:
  - config_name: default
    data_files:
      - split: train
        path: data/train-*
train-eval-index:
  - config: default
    task: text-classification
    task_id: multi_class_classification
    train_split: train
    col_mapping:
      text: text
      label: target
    metrics:
      - type: accuracy
        name: Accuracy
      - type: f1
        name: F1 macro
        args:
          average: macro
      - type: f1
        name: F1 micro
        args:
          average: micro
      - type: f1
        name: F1 weighted
        args:
          average: weighted
      - type: precision
        name: Precision macro
        args:
          average: macro
      - type: precision
        name: Precision micro
        args:
          average: micro
      - type: precision
        name: Precision weighted
        args:
          average: weighted
      - type: recall
        name: Recall macro
        args:
          average: macro
      - type: recall
        name: Recall micro
        args:
          average: micro
      - type: recall
        name: Recall weighted
        args:
          average: weighted

Dataset Card for Snips Built In Intents

Table of Contents

Dataset Description

Dataset Summary

Snips' built in intents dataset was initially used to compare different voice assistants and released as a public dataset hosted at https://github.com/sonos/nlu-benchmark in folder 2016-12-built-in-intents. The dataset contains 328 utterances over 10 intent classes. A related Medium post is https://medium.com/snips-ai/benchmarking-natural-language-understanding-systems-d35be6ce568d.

Supported Tasks and Leaderboards

There are no related shared tasks that we are aware of.

Languages

English

Dataset Structure

Data Instances

The dataset contains 328 utterances over 10 intent classes. Each sample looks like: {'label': 8, 'text': 'Transit directions to Barcelona Pizza.'}

Data Fields

  • text: The text utterance expressing some user intent.
  • label: The intent label of the piece of text utterance.

Data Splits

The source data is not split.

Dataset Creation

Curation Rationale

The dataset was originally created to compare the performance of a number of voice assistants. However, the labelled utterances are useful for developing and benchmarking text chatbots as well.

Source Data

Initial Data Collection and Normalization

It is not clear how the data was collected. From the Medium post: The benchmark relies on a set of 328 queries built by the business team at Snips, and kept secret from data scientists and engineers throughout the development of the solution.

Who are the source language producers?

Originally prepared by snips.ai. The Snips team has since joined Sonos in November 2019. These open datasets remain available and their access is now managed by the Sonos Voice Experience Team. Please email sve-research@sonos.com with any question.

Annotations

Annotation process

It is not clear how the data was collected. From the Medium post: The benchmark relies on a set of 328 queries built by the business team at Snips, and kept secret from data scientists and engineers throughout the development of the solution.

Who are the annotators?

[More Information Needed]

Personal and Sensitive Information

[More Information Needed]

Considerations for Using the Data

Social Impact of Dataset

[More Information Needed]

Discussion of Biases

[More Information Needed]

Other Known Limitations

[More Information Needed]

Additional Information

Dataset Curators

Originally prepared by snips.ai. The Snips team has since joined Sonos in November 2019. These open datasets remain available and their access is now managed by the Sonos Voice Experience Team. Please email sve-research@sonos.com with any question.

Licensing Information

The source data is licensed under Creative Commons Zero v1.0 Universal.

Citation Information

Any publication based on these datasets must include a full citation to the following paper in which the results were published by the Snips Team:

Coucke A. et al., "Snips Voice Platform: an embedded Spoken Language Understanding system for private-by-design voice interfaces." CoRR 2018, https://arxiv.org/abs/1805.10190

Contributions

Thanks to @bduvenhage for adding this dataset.