|
--- |
|
language: |
|
- en |
|
- zh |
|
license: cc-by-nc-4.0 |
|
extra_gated_prompt: >- |
|
Subject to other terms Subject to other terms: |
|
1. The corpus is not used for commercial purposes and is only provided free of charge to “universities and research institutes” for scientific research. |
|
|
|
2、When publishing papers and applying for results, if you use this dataset, please indicate the reference: |
|
@article{ |
|
liu2024emotion, |
|
title={Emotion and Intent Joint Understanding in Multimodal Conversation: A Benchmarking Dataset}, |
|
author={Liu, Rui and Zuo, Haolin and Lian, Zheng and Xing, Xiaofen and Schuller, Bj{\"o}rn W and Li, Haizhou}, |
|
journal={arXiv preprint arXiv:2407.02751}, |
|
year={2024} |
|
} |
|
|
|
3. The final interpretation of this corpus belongs to S2LAB Lab, Inner Mongolia University, China. |
|
|
|
extra_gated_fields: |
|
First Name: text |
|
Last Name: text |
|
Date of birth: date_picker |
|
Country: country |
|
Institution: |
|
type: text |
|
placeholder: "e.g., Stanford University" |
|
description: "Please enter the full name of your institution (e.g., including 'University' or 'Institute')." |
|
Job title: |
|
type: select |
|
options: |
|
- Student |
|
- Research Graduate |
|
- AI researcher |
|
- AI developer/engineer |
|
- Reporter |
|
- Other |
|
geo: ip_location |
|
By clicking submit below, I accept the terms of the license and acknowledge that the information I provide will be collected, stored and processed by S2LAB: checkbox |
|
extra_gated_description: >- |
|
The information you provide will be collected, stored and processed by S2LAB. |
|
extra_gated_button_content: Submit |
|
extra_gated_eu_disallowed: true |
|
task_categories: |
|
- text-to-speech |
|
tags: |
|
- Multimodal and Multilingual |
|
- Dialog |
|
- Conversational Emotion and Intent Recognition |
|
pretty_name: MC-EIU |
|
size_categories: |
|
- 100M<n<1B |
|
--- |
|
|
|
# MC-EIU |
|
|
|
### Introduction |
|
|
|
This is the official repository for the MC-EIU dataset. More details can be found in the paper: [Emotion and Intent Joint Understanding in Multimodal Conversation: A Benchmarking Dataset](https://arxiv.org/abs/2407.02751). |
|
|
|
[Rui Liu *](https://ttslr.github.io/), Haolin Zuo, Zheng Lian, Xiaofen Xing, Björn W. Schuller, Haizhou Li |
|
|
|
### MC-EIU Overview |
|
|
|
Statistic of our MC-EIU. UL refers to the utterance length, DU denotes the duration per utterance, UC is the utterances per conversation, EC means the emotions per conversation, and IC means the intents per conversation. |
|
|
|
| Statistics | English | | | Mandarin | | | |
|
|-----------------------------|-----------------------|---------------------|--------------------|----------------------|---------------------|--------------------| |
|
| | Train | Valid | Test | Train | Valid | Test | |
|
| # Conversations | 2,807 | 400 | 806 | 667 | 95 | 195 | |
|
| # Utterances | 31,451 | 4,509 | 9,049 | 7,643 | 1,148 | 2,212 | |
|
| # Duration (hours) | 28.51 | 4.02 | 8.22 | 8.51 | 1.36 | 2.42 | |
|
| Avg. UL | 12.68 | 12.49 | 12.76 | 19.11 | 19.91 | 18.14 | |
|
| Avg. # of DU (seconds) | 3.26 | 3.21 | 3.27 | 4.01 | 4.26 | 3.94 | |
|
| Avg. # of UC | 11.20 | 11.27 | 11.23 | 11.46 | 12.08 | 11.34 | |
|
| Avg. # of EC | 2.58 | 2.57 | 2.60 | 2.41 | 2.54 | 2.42 | |
|
| Avg. # of IC | 3.29 | 3.86 | 3.87 | 3.18 | 3.24 | 3.10 | |
|
|
|
|
|
|
|
### Citations |
|
If you find this dataset useful in your research, please consider citing the following paper: |
|
``` |
|
@article{ |
|
liu2024emotion, |
|
title={Emotion and Intent Joint Understanding in Multimodal Conversation: A Benchmarking Dataset}, |
|
author={Liu, Rui and Zuo, Haolin and Lian, Zheng and Xing, Xiaofen and Schuller, Bj{\"o}rn W and Li, Haizhou}, |
|
journal={arXiv preprint arXiv:2407.02751}, |
|
year={2024} |
|
} |
|
``` |
|
|
|
⚠ The collected TV shows clips are all from public resources on the Internet. If there is any infringement, please contact us to delete them. (📧: zuohaolin[email protected]) |
|
|