Skip to content

Commit

Permalink
Merge pull request #26 from SuperKogito/restructure-and-jsonify
Browse files Browse the repository at this point in the history
Restructure and jsonify
  • Loading branch information
SuperKogito authored Feb 13, 2023
2 parents 699af50 + acf5264 commit 7349aa2
Show file tree
Hide file tree
Showing 9 changed files with 849 additions and 296 deletions.
5 changes: 5 additions & 0 deletions CONTRIBUTING.md
Original file line number Diff line number Diff line change
Expand Up @@ -22,3 +22,8 @@ please feel free to add it.
* The dataset should not be provided in an active PR.
* The dataset should be available for researchers for free.
* The information about the dataset must be accessible for verification.

## How to contribute
First go to `src/` using `cd src`. Then add a the dictionary / part json data of the contributed dataset to `src/ser-datasets`.
Make sure the json is valid, then run the `python generate_files.py` to update the restructured text file, csv file and the README.
That's it, Congrats! and thank you for your contribution. Now open a PR with your changes. I will review it and then publish the results :))
195 changes: 44 additions & 151 deletions README.md

Large diffs are not rendered by default.

7 changes: 7 additions & 0 deletions src/build_project.sh
Original file line number Diff line number Diff line change
@@ -0,0 +1,7 @@
#!/bin/bash

echo "Rst and Md files generations"
python generate_files.py

echo "Make html files"
make html
3 changes: 1 addition & 2 deletions src/conf.py
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@
'sphinx.ext.coverage',
'sphinx.ext.ifconfig',
'sphinx.ext.viewcode',
'sphinx_contributors',
#'sphinx_contributors',
'crate.sphinx.csv',
]

Expand All @@ -59,7 +59,6 @@
html_theme_options = {
"github_url": "https://github.com/superkogito/ser-datasets",
"search_bar_text": "Search this site...",
"google_analytics_id": "UA-133660046-1",

"navbar_start": ["navbar-logo"],
"navbar_center": ["navbar-nav"],
Expand Down
91 changes: 91 additions & 0 deletions src/generate_files.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,91 @@

import csv
import json
from tabulate import tabulate


# load datasets
json_file_path = "ser-datasets.json"
with open(json_file_path, 'r') as j:
content = json.loads(j.read())

# init keys
keys = ["Dataset", "Year", "Content", "Emotions", "Format", "Size", "Language", "Paper", "Access", "License", "Dataset-link", "Paper-link", "License-link"]
header = ["Dataset", "Year", "Content", "Emotions", "Format", "Size", "Language", "Paper", "Access", "License"]

md_1 = """***Spoken Emotion Recognition Datasets:*** *A collection of datasets (count=42) for the purpose of emotion recognition/detection in speech.
The table is chronologically ordered and includes a description of the content of each dataset along with the emotions included.
The table can be browsed, sorted and searched under https://superkogito.github.io/SER-datasets/*
"""

md_2 = """## References
- Swain, Monorama & Routray, Aurobinda & Kabisatpathy, Prithviraj, Databases, features and classifiers for speech emotion recognition: a review, International Journal of Speech Technology, [paper](https://www.researchgate.net/publication/322602563_Databases_features_and_classifiers_for_speech_emotion_recognition_a_review#pf19)
- Dimitrios Ververidis and Constantine Kotropoulos, A State of the Art Review on Emotional Speech Databases, Artificial Intelligence & Information Analysis Laboratory, Department of Informatics Aristotle, University of Thessaloniki, [paper](http://poseidon.csd.auth.gr/papers/PUBLISHED/CONFERENCE/pdf/Ververidis2003b.pdf)
- A. Pramod Reddy and V. Vijayarajan, Extraction of Emotions from Speech-A Survey, VIT University, International Journal of Applied Engineering Research, [paper](https://www.ripublication.com/ijaer17/ijaerv12n16_46.pdf)
- Emotional Speech Databases, [document](https://link.springer.com/content/pdf/bbm%3A978-90-481-3129-7%2F1.pdf)
- Expressive Synthetic Speech, [website](http://emosamples.syntheticspeech.de/)
- Towards a standard set of acoustic features for the processing of emotion in speech, Technical university Munich, [document](https://asa.scitation.org/doi/pdf/10.1121/1.4739483)
## Contribution
- All contributions are welcome! If you know a dataset that belongs here (see [criteria](https://github.com/SuperKogito/SER-datasets/blob/master/CONTRIBUTING.md#criteria)) but is not listed, please feel free to add it. For more information on Contributing, please refer to [CONTRIBUTING.md](https://github.com/SuperKogito/SER-datasets/blob/master/CONTRIBUTING.md).
- If you notice a typo or a mistake, please [report this as an issue](https://github.com/SuperKogito/SER-datasets/issues/new) and help us improve the quality of this list.
## Disclaimer
- The mainter and the contributors try their best to keep this list up-to-date, and to only include working links (using automated verification with the help of the [urlchecker-action](https://github.com/marketplace/actions/urlchecker-action)). However, we cannot guarantee that all listed links are up-to-date. Read more in [DISCLAIMER.md](https://github.com/SuperKogito/SER-datasets/blob/master/DISCLAIMER.md).
"""


print(" -> Generate Markdown Text")
def format_md_link(label, link):
res = "[{0}]({1})".format(label, link) if "http" in link else label
return res

# tabulate
table = []
for key, values in content.items():
# add elements to row
row = [format_md_link(key, values["Dataset-link"])]
row += [values[k] for k in ["Year", "Content", "Emotions", "Format", "Size", "Language"]]
row += [format_md_link(values["Paper"], values["Paper-link"]), values["Access"], format_md_link(values["License"], values["License-link"])]

# add styles and add row to table
row = ["<sub>{0}</sub>".format(e) for e in row]
table.append(row)

table = tabulate(table, keys, tablefmt="pipe")
with open("../README.md", "w") as f:
f.write(md_1)
f.write(table)
f.write(md_2)


print(" -> Generate Restructured Text")
def format_rst_link(label, link):
res = "`{0} <{1}>`_".format(label, link) if "http" in link else label
return res

# tabulate
table = []
for key, values in content.items():
# add elements to row
row = [format_rst_link(key, values["Dataset-link"])]
row += [values[k] for k in ["Year", "Content", "Emotions", "Format", "Size", "Language"]]
row += [format_rst_link(values["Paper"], values["Paper-link"]), values["Access"]]
row += [format_rst_link(values["License"], values["License-link"])]

# format and add row to csv
table.append(row)

with open('ser-datasets.csv', 'w', encoding='UTF8', newline='') as f:
writer = csv.writer(f)

# write the header
writer.writerow(header)

# write multiple rows
writer.writerows(table)
101 changes: 1 addition & 100 deletions src/index.rst
Original file line number Diff line number Diff line change
Expand Up @@ -40,103 +40,4 @@ Disclaimer
===========

The maintainer and the contributors try their best to keep this list up-to-date, and to only include working links (using automated verification with the help of the `urlchecker-action <https://github.com/marketplace/actions/urlchecker-action>`_).
However, we cannot guarantee that all listed links are up-to-date. Read more in `DISCLAIMER.md <https://github.com/SuperKogito/SER-datasets/blob/master/DISCLAIMER.md>`_.


.. datasets
.. _`MESD`: https://data.mendeley.com/datasets/cy34mh68j9/5
.. _`MLEnd`: https://www.kaggle.com/datasets/jesusrequena/mlend-spoken-numerals
.. _`ASVP-ESD`: https://www.kaggle.com/datasets/dejolilandry/asvpesdspeech-nonspeech-emotional-utterances
.. _`ESD`: https://hltsingapore.github.io/ESD/
.. _`MuSe-CAR`: https://zenodo.org/record/4134758
.. _`MSP-Podcast corpus`: https://ecs.utdallas.edu/research/researchlabs/msp-lab/MSP-Podcast.html
.. _`emotiontts open db`: https://github.com/emotiontts/emotiontts_open_db
.. _`URDU-Dataset`: https://github.com/siddiquelatif/urdu-dataset
.. _`BAVED`: https://www.kaggle.com/a13x10/basic-arabic-vocal-emotions-dataset
.. _`VIVAE`: https://zenodo.org/record/4066235
.. _`SEWA`: https://db.sewaproject.eu/
.. _`MELD`: https://affective-meld.github.io/
.. _`ShEMO`: https://github.com/mansourehk/ShEMO
.. _`DEMoS`: https://zenodo.org/record/2544829
.. _`VERBO`: https://sites.google.com/view/verbodatabase/home
.. _`AESDD`: http://m3c.web.auth.gr/research/aesdd-speech-emotion-recognition/
.. _`Emov-DB`: https://mega.nz/#F!KBp32apT!gLIgyWf9iQ-yqnWFUFuUHg!mYwUnI4K
.. _`RAVDESS`: https://zenodo.org/record/1188976#.XrC7a5NKjOR
.. _`JL corpus`: https://www.kaggle.com/tli725/jl-corpus
.. _`CaFE`: https://zenodo.org/record/1478765
.. _`EmoFilm`: https://zenodo.org/record/1326428
.. _`ANAD`: https://www.kaggle.com/suso172/arabic-natural-audio-dataset
.. _`EmoSynth`: https://zenodo.org/record/3727593
.. _`CMU-MOSEI`: https://www.amir-zadeh.com/datasets
.. _`CMU-MOSI`: https://www.amir-zadeh.com/datasets
.. _`MSP-IMPROV`: https://ecs.utdallas.edu/research/researchlabs/msp-lab/MSP-Improv.html
.. _`CREMA-D`: https://github.com/CheyneyComputerScience/CREMA-D
.. _`Example emotion videos used in investigation of emotion perception in schizophrenia`: https://espace.library.uq.edu.au/view/UQ:446541
.. _`EMOVO`: http://voice.fub.it/activities/corpora/emovo/index.html
.. _`RECOLA`: https://diuf.unifr.ch/main/diva/recola/download.html
.. _`GEMEP corpus`: https://www.unige.ch/cisa/gemep
.. _`OGVC`: https://sites.google.com/site/ogcorpus/home/en
.. _`LEGO corpus`: https://www.ultes.eu/ressources/lego-spoken-dialogue-corpus/
.. _`SEMAINE`: https://semaine-db.eu/
.. _`SAVEE`: http://kahlan.eps.surrey.ac.uk/savee/Database.html
.. _`TESS`: https://tspace.library.utoronto.ca/handle/1807/24487
.. _`EEKK`: https://metashare.ut.ee/repository/download/4d42d7a8463411e2a6e4005056b40024a19021a316b54b7fb707757d43d1a889/
.. _`IEMOCAP`: https://sail.usc.edu/iemocap/iemocap_release.htm
.. _`Keio-ESD`: http://research.nii.ac.jp/src/en/Keio-ESD.html
.. _`EMO-DB`: http://emodb.bilderbar.info/index-1280.html
.. _`eNTERFACE05`: http://www.enterface.net/enterface05/docs/results/databases/project2_database.zip
.. _`DES`: http://kom.aau.dk/~tb/speech/Emotions/

.. license
.. _`CC BY 4.0`: https://creativecommons.org/licenses/by/4.0/
.. _`CC BY-NC-SA 4.0`: https://creativecommons.org/licenses/by-nc-sa/4.0/
.. _`CC BY-NC-ND 4.0`: https://creativecommons.org/licenses/by-nc-nd/4.0/
.. _`CC-BY license`: https://metashare.ut.ee/repository/download/4d42d7a8463411e2a6e4005056b40024a19021a316b54b7fb707757d43d1a889/
.. _`Permitted Non-commercial Re-use with Acknowledgment`: https://guides.library.uq.edu.au/deposit_your_data/terms_and_conditions
.. _`Open Database License & Database Content License`: https://github.com/CheyneyComputerScience/CREMA-D/blob/master/LICENSE.txt
.. _`CC0 1.0`: https://creativecommons.org/publicdomain/zero/1.0/
.. _`CMU-MOSEI License`: https://github.com/A2Zadeh/CMU-MultimodalSDK/blob/master/LICENSE.txt
.. _`CMU-MOSI License`: https://github.com/A2Zadeh/CMU-MultimodalSDK/blob/master/LICENSE.txt
.. _`IEMOCAP license`: https://sail.usc.edu/iemocap/Data_Release_Form_IEMOCAP.pdf
.. _`SEWA EULA`: https://db.sewaproject.eu/media/doc/eula.pdf
.. _`Meld: GPL-3.0 License`: https://github.com/declare-lab/MELD/blob/master/LICENSE

.. papers
.. _`The Mexican Emotional Speech Database (MESD): elaboration and assessment based on machine learning`: https://pubmed.ncbi.nlm.nih.gov/34891601/
.. _`Seen And Unseen Emotional Style Transfer For Voice Conversion With A New Emotional Speech Dataset`: https://arxiv.org/pdf/2010.14794.pdf
.. _`The Multimodal Sentiment Analysis in Car Reviews (MuSe-CaR) Dataset: Collection, Insights and Improvements`: https://arxiv.org/pdf/2101.06053.pdf
.. _`The MSP-Conversation Corpus`: http://www.interspeech2020.org/index.php?m=content&c=index&a=show&catid=290&id=684
.. _`Cross Lingual Speech Emotion Recognition: Urdu vs. Western Languages`: https://arxiv.org/pdf/1812.10411.pdf
.. _`Estonian Emotional Speech Corpus`: https://www.researchgate.net/publication/261724574_Estonian_Emotional_Speech_Corpus_Release_1
.. _`IEMOCAP: Interactive emotional dyadic motion capture database`: https://sail.usc.edu/iemocap/Busso_2008_iemocap.pdf
.. _`A Database of German Emotional Speech`: http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.130.8506&rep=rep1&type=pdf
.. _`SEWA DB: A Rich Database for Audio-Visual Emotion and Sentiment Research in the Wild`: https://arxiv.org/pdf/1901.02839.pdf
.. _`Documentation of the Danish Emotional Speech Database`: http://kom.aau.dk/~tb/speech/Emotions/des.pdf
.. _`EMOTIONAL SPEECH SYNTHESIS USING SUBSPACE CONSTRAINTS IN PROSODY`: http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.420.8899&rep=rep1&type=pdf
.. _`Naturalistic emotional speech collectionparadigm with online game and its psychological and acoustical assessment`: https://www.jstage.jst.go.jp/article/ast/33/6/33_E1175/_pdf
.. _`EMOVO Corpus: an Italian Emotional Speech Database`: https://core.ac.uk/download/pdf/53857389.pdf
.. _`VERBO: Voice Emotion Recognition dataBase in Portuguese Language`: https://thescipub.com/pdf/jcssp.2018.1420.1430.pdf
.. _`The eNTERFACE’05 Audio-Visual Emotion Database`: http://poseidon.csd.auth.gr/papers/PUBLISHED/CONFERENCE/pdf/Martin06a.pdf
.. _`Arabic Natural Audio Dataset`: https://data.mendeley.com/datasets/xm232yxf7t/1
.. _`Introducing the Geneva Multimodal Expression Corpus for Experimental Research on Emotion Perception`: https://www.researchgate.net/publication/51796867_Introducing_the_Geneva_Multimodal_Expression_Corpus_for_Experimental_Research_on_Emotion_Perception
.. _`Speech Emotion Recognition for Performance Interaction`: https://www.researchgate.net/publication/326005164_Speech_Emotion_Recognition_for_Performance_Interaction
.. _`MELD: A Multimodal Multi-Party Dataset for Emotion Recognition in Conversations`: https://arxiv.org/pdf/1810.02508.pdf
.. _`BEHAVIOURAL FINDINGS FROM THE TORONTO EMOTIONAL SPEECH SET`: https://www.semanticscholar.org/paper/BEHAVIOURAL-FINDINGS-FROM-THE-TORONTO-EMOTIONAL-SET-Dupuis-Pichora-Fuller/d7f746b3aee801a353b6929a65d9a34a68e71c6f/figure/2
.. _`CREMA-D: Crowd-sourced Emotional Multimodal Actors Dataset`: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4313618/
.. _`DEMoS: An Italian emotional speech corpus. Elicitation methods, machine learning, and perception`: https://link.springer.com/epdf/10.1007/s10579-019-09450-y?author_access_token=5pf0w_D4k9z28TM6n4PbVPe4RwlQNchNByi7wbcMAY5hiA-aXzXNbZYfsMDDq2CdHD-w5ArAxIwlsk2nC_26pSyEAcu1xlKJ1c9m3JZj2ZlFmlVoCZUTcG3Hq2_2ozMLo3Hq3Y0CHzLdTxihQwch5Q%3D%3D
.. _`A Parameterized and Annotated Spoken Dialog Corpus of the CMU Let’s Go Bus Information System`: http://www.lrec-conf.org/proceedings/lrec2012/pdf/333_Paper.pdf
.. _`Introducing the RECOLA Multimodal Corpus of Remote Collaborative and Affective Interactions`: https://drive.google.com/file/d/0B2V_I9XKBODhNENKUnZWNFdVXzQ/view
.. _`Multimodal Emotion Recognition`: http://personal.ee.surrey.ac.uk/Personal/P.Jackson/pub/ma10/HaqJackson_MachineAudition10_approved.pdf
.. _`The Perceived Emotion of Isolated Synthetic Audio: The EmoSynth Dataset and Results`: https://dl.acm.org/doi/10.1145/3243274.3243277
.. _`MSP-IMPROV: An Acted Corpus of Dyadic Interactions to Study Emotion Perception`: https://ecs.utdallas.edu/research/researchlabs/msp-lab/publications/Busso_2017.pdf
.. _`Multi-attention Recurrent Network for Human Communication Comprehension`: https://arxiv.org/pdf/1802.00923.pdf
.. _`Categorical vs Dimensional Perception of Italian Emotional Speech`: https://pdfs.semanticscholar.org/e70e/fcf7f5b4c366a7b7e2c16267d7f7691a5391.pdf
.. _`Multi-attention Recurrent Network for Human Communication Comprehension`: https://arxiv.org/pdf/1802.00923.pdf
.. _`ShEMO: a large-scale validated database for Persian speech emotion detection`: https://link.springer.com/article/10.1007/s10579-018-9427-x
.. _`The emotional voices database: Towards controlling the emotion dimension in voice generation systems`: https://arxiv.org/pdf/1806.09514.pdf
.. _`The Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS): A dynamic, multimodal set of facial and vocal expressions in North American English`: https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0196391
.. _`An Open Source Emotional Speech Corpus for Human Robot Interaction Applications`: https://www.isca-speech.org/archive/Interspeech_2018/pdfs/1349.pdf
.. _`The SEMAINE Database: Annotated Multimodal Records of Emotionally Colored Conversations between a Person and a Limited Agent`: https://ieeexplore.ieee.org/document/5959155
However, we cannot guarantee that all listed links are up-to-date. Read more in `DISCLAIMER.md <https://github.com/SuperKogito/SER-datasets/blob/master/DISCLAIMER.md>`_.
67 changes: 67 additions & 0 deletions src/requirements.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,67 @@
ablog==0.10.25
alabaster==0.7.12
Babel==2.10.1
beautifulsoup4==4.11.1
bleach==6.0.0
certifi==2021.10.8
cffi==1.15.1
charset-normalizer==2.0.12
commonmark==0.9.1
cryptography==39.0.1
docutils==0.17.1
entrypoints==0.4
feedgen==0.9.0
idna==3.3
imagesize==1.3.0
importlib-metadata==4.11.3
invoke==1.7.0
jeepney==0.8.0
Jinja2==3.1.2
keyring==23.6.0
latexcodec==2.0.1
lxml==4.8.0
MarkupSafe==2.1.1
nest-asyncio==1.5.6
packaging==21.3
pkginfo==1.8.3
pybtex==0.24.0
pybtex-docutils==1.0.1
pycparser==2.21
pydata-sphinx-theme==0.8.1
Pygments==2.12.0
pyparsing==3.0.8
python-dateutil==2.8.2
pytz==2022.1
PyYAML==6.0
pyzmq==24.0.1
readme-renderer==35.0
requests==2.27.1
requests-toolbelt==0.9.1
rich==12.5.1
SecretStorage==3.3.2
six==1.16.0
snowballstemmer==2.2.0
soupsieve==2.3.2.post1
Sphinx==4.5.0
sphinx-copybutton==0.5.0
sphinx-csv-filter==0.4.0
sphinx-panels==0.6.0
sphinx-sitemap==2.2.0
sphinxcontrib-applehelp==1.0.2
sphinxcontrib-bibtex==2.4.2
sphinxcontrib-devhelp==1.0.2
sphinxcontrib-htmlhelp==2.0.0
sphinxcontrib-jsmath==1.0.1
sphinxcontrib-pdfembed @ git+https://github.com/SuperKogito/sphinxcontrib-pdfembed@d75fb37f9e4a303888a61f265b568f7729826c4a
sphinxcontrib-qthelp==1.0.3
sphinxcontrib-serializinghtml==1.1.5
sphinxcontrib-tikz==0.4.16
sphinxemoji==0.2.0
sphinxext-opengraph==0.6.3
tabulate==0.9.0
tornado==6.2
tqdm==4.64.0
urllib3==1.26.9
watchdog==2.1.7
webencodings==0.5.1
zipp==3.8.0
Loading

0 comments on commit 7349aa2

Please sign in to comment.