Stay organized with collections
Save and categorize content based on your preferences.
Description:
SAVEE (Surrey Audio-Visual Expressed Emotion) is an emotion recognition dataset.
It consists of recordings from 4 male actors in 7 different emotions, 480
British English utterances in total. The sentences were chosen from the standard
TIMIT corpus and phonetically-balanced for each emotion. This release contains
only the audio stream from the original audio-visual recording.
The data is split so that the training set consists of 2 speakers, and both the
validation and test set consists of samples from 1 speaker, respectively.
Manual download instructions: This dataset requires you to
download the source data manually into download_config.manual_dir
(defaults to ~/tensorflow_datasets/downloads/manual/):
manual_dir should contain the file AudioData.zip. This file should be under
Data/Zip/AudioData.zip in the dataset folder provided upon registration.
You need to register at
http://personal.ee.surrey.ac.uk/Personal/P.Jackson/SAVEE/Register.html in
order to get the link to download the dataset.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Missing the information I need","missingTheInformationINeed","thumb-down"],["Too complicated / too many steps","tooComplicatedTooManySteps","thumb-down"],["Out of date","outOfDate","thumb-down"],["Samples / code issue","samplesCodeIssue","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2022-12-23 UTC."],[],[],null,["# savee\n\n\u003cbr /\u003e\n\n| **Warning:** Manual download required. See instructions below.\n\n- **Description**:\n\nSAVEE (Surrey Audio-Visual Expressed Emotion) is an emotion recognition dataset.\nIt consists of recordings from 4 male actors in 7 different emotions, 480\nBritish English utterances in total. The sentences were chosen from the standard\nTIMIT corpus and phonetically-balanced for each emotion. This release contains\nonly the audio stream from the original audio-visual recording.\n\nThe data is split so that the training set consists of 2 speakers, and both the\nvalidation and test set consists of samples from 1 speaker, respectively.\n\n- **Additional Documentation** :\n [Explore on Papers With Code\n north_east](https://paperswithcode.com/dataset/savee)\n\n- **Homepage** :\n \u003chttp://kahlan.eps.surrey.ac.uk/savee/\u003e\n\n- **Source code** :\n [`tfds.datasets.savee.Builder`](https://github.com/tensorflow/datasets/tree/master/tensorflow_datasets/datasets/savee/savee_dataset_builder.py)\n\n- **Versions**:\n\n - **`1.0.0`** (default): No release notes.\n- **Download size** : `Unknown size`\n\n- **Dataset size** : `259.15 MiB`\n\n- **Manual download instructions** : This dataset requires you to\n download the source data manually into `download_config.manual_dir`\n (defaults to `~/tensorflow_datasets/downloads/manual/`): \n\n manual_dir should contain the file AudioData.zip. This file should be under\n Data/Zip/AudioData.zip in the dataset folder provided upon registration.\n You need to register at\n \u003chttp://personal.ee.surrey.ac.uk/Personal/P.Jackson/SAVEE/Register.html\u003e in\n order to get the link to download the dataset.\n\n- **Auto-cached**\n ([documentation](https://www.tensorflow.org/datasets/performances#auto-caching)):\n No\n\n- **Splits**:\n\n| Split | Examples |\n|----------------|----------|\n| `'test'` | 120 |\n| `'train'` | 240 |\n| `'validation'` | 120 |\n\n- **Feature structure**:\n\n FeaturesDict({\n 'audio': Audio(shape=(None,), dtype=int64),\n 'label': ClassLabel(shape=(), dtype=int64, num_classes=7),\n 'speaker_id': string,\n })\n\n- **Feature documentation**:\n\n| Feature | Class | Shape | Dtype | Description |\n|------------|--------------|---------|--------|-------------|\n| | FeaturesDict | | | |\n| audio | Audio | (None,) | int64 | |\n| label | ClassLabel | | int64 | |\n| speaker_id | Tensor | | string | |\n\n- **Supervised keys** (See\n [`as_supervised` doc](https://www.tensorflow.org/datasets/api_docs/python/tfds/load#args)):\n `('audio', 'label')`\n\n- **Figure**\n ([tfds.show_examples](https://www.tensorflow.org/datasets/api_docs/python/tfds/visualization/show_examples)):\n Not supported.\n\n- **Examples**\n ([tfds.as_dataframe](https://www.tensorflow.org/datasets/api_docs/python/tfds/as_dataframe)):\n\nDisplay examples... \n\n- **Citation**:\n\n @inproceedings{Vlasenko_combiningframe,\n author = {Vlasenko, Bogdan and Schuller, Bjorn and Wendemuth, Andreas and Rigoll, Gerhard},\n year = {2007},\n month = {01},\n pages = {2249-2252},\n title = {Combining frame and turn-level information for robust recognition of emotions within speech},\n journal = {Proceedings of Interspeech}\n }"]]