Skip to content
This repository has been archived by the owner on Mar 8, 2023. It is now read-only.

LIST OF AUDIO+TEXT DATASETS #114

Open
nefastosaturo opened this issue Dec 17, 2020 · 10 comments
Open

LIST OF AUDIO+TEXT DATASETS #114

nefastosaturo opened this issue Dec 17, 2020 · 10 comments
Labels
dataset help wanted Extra attention is needed

Comments

@nefastosaturo
Copy link
Collaborator

nefastosaturo commented Dec 17, 2020

LIST OF ALL ITALIAN DATASETS FOUND

From issue #90 I'm putting here all the datasets that have been discovered.
Some of them are plug-and-play for Deepspeech others instead need to be created from scratch (splits up audio by sentences)

Feel free to pickup one that has not been done for checking it out.

NOTE

If one of this dataset needs a deeper analysis please do not start a discussion here but open a new issue and I will update this table with the issue reference.

DATASETS

dataset hrs url plug-n-play TODOs doing done note
MLS 279.43 h HOT!!!!
VoxForge #111 20h
  • url replace in DS import_voxforge.py script
  • fix import sys error
MAILABS 127h40m
Evalita2009 5h
MSPKA 3h
SIWIS 4.5h
SUGAR 1.5h sentences are not useful
VociParlateWikipedia #34 ?
  • sync audio with its page revision
EMOVO ~12m
  • align filename codes with their sentences
interesting for emotions (disgust, happy..)
ZIta <1hr transcriptions do not follow recordings (eg: Lett_Z_Sp1_zero.wav)
LIM_Veneti <1hr no audio files?
split-MDb ~46m
  • parse&clean the .wrd files
based on CLIPS
tg60 1h30m
  • long audio files to be split
maybe among the info files there are some timings that could be useful for splitting up?
PraTiD 1h12m
  • long audio files to be split
From CLIPS; maybe among the info files there are some timings that could be useful for splitting up?
ParlatoCinematografico ?
  • long audio files to be split
.lab files with speakers timings
PerugiaCorpusPEC ? a login is needed. License?
@nefastosaturo nefastosaturo added dataset help wanted Extra attention is needed labels Dec 17, 2020
@nshmyrev
Copy link

nshmyrev commented Dec 17, 2020

MLS from facebook has Italiano:

http://openslr.org/94/

279.43 hours

@nefastosaturo
Copy link
Collaborator Author

@nshmyrev WOW, thank you for this christmas present!!

@eziolotta
Copy link
Contributor

eziolotta commented Dec 19, 2020

I performed script for import MLS files in MITADS-Speech Datasets

https://dl.fbaipublicfiles.com/mls/mls_italian.tar.gz (14.3G zip)

Converting the .flac audio files to Wav 16KHz and doing some checks.

With sample tests, Audio are of good quality and transcripts are clean.
All clips are between 10 and 20 seconds long (specified in paper)

if my script work fine All clips <= 15 seconds (and successfully resampled)
are in total: 159.23h

Textual corpus on which speech dataset is based, includes ancient works like this:
works of Giovanni Francesco Straparola (1400),
Divina Commedia (and others) by Dante Alighieri (1300)
works of Luigi Pirandello (1900)

In some cases we find in sentences obsolete forms and terms that person using speech technologies today is unlikely to pronounce.
If we need to filter clips by Author/Work, this information is present in flac audio file

@Mte90
Copy link
Member

Mte90 commented Dec 19, 2020

I think that we should avoid this kind of ancient work (except pirandello) as we did in Mitads itself as example.

@eziolotta eziolotta mentioned this issue Jan 11, 2021
@eziolotta
Copy link
Contributor

eziolotta commented Jan 17, 2021

https://arxiv.org/pdf/2101.00390.pdf

VoxPopuli: largest open unlabelled speech dataset, totaling 100K hours in 23 languages from European Parliament.
Also contains 1.8K hours of transcribed speeches in 16 languages.

They will release the corpus at https://github.com/facebookresearch/voxpopuli under a open license.

@eziolotta
Copy link
Contributor

eziolotta commented Jan 17, 2021

Europarl-ST , multilingual corpus for speech translation of parliamentary debates.

Total 64.18h of Italian clips audio.
From approximate calculations about 30% of these (20h?) have a transcript in Italian
Most clips range duration from 1 minute to 2 minutes

https://arxiv.org/pdf/1911.03167.pdf

https://www.mllp.upv.es/europarl-st/v1.1.tar.gz (20G)

@eziolotta
Copy link
Contributor

eziolotta commented Feb 27, 2021

New Speech Dataset: Multilingual TEDx by OpenSLR
https://www.openslr.org/100/

The corpus comprises audio recordings and transcripts from TEDx Talks in 8 languages
Italian: about 123 Hours

Licence: Creative Commons Attribution-NonCommercial-NoDerivs 4.0
https://www.ted.com/about/our-organization/our-policies-terms/ted-talks-usage-policy

UPDATE:
Audio clips have an average length, ranging from 4-5 minutes to 25 minutes
This does not make it directly usable for training with deep speech

@eziolotta
Copy link
Contributor

eziolotta commented Feb 27, 2021

For Multilingual TEDx dataset:
in segments.txt file there are audio segments of each clip, they are text alignment with audio timestamps.
Through this file, audio-clips could be segmented easily, but I think we cannot redistribute dataset due to the license.

Unfortunately Deep Speech doesn't have a powerfull toolkit to split audio by aligner or vad, like Kaldi or others.
Could we try this : https://espnet.github.io/espnet/apis/espnet_bin.html#asr-align-py ?

@DanBmh
Copy link

DanBmh commented Apr 6, 2021

You can import the mTEDx dataset with corcua:
https://gitlab.com/Jaco-Assistant/corcua

Splitting the files will take some time (~2 days), but if you extend the script with a parallelization, it should work much faster:)

@nefastosaturo
Copy link
Collaborator Author

You can import the mTEDx dataset with corcua:
https://gitlab.com/Jaco-Assistant/corcua

Splitting the files will take some time (~2 days), but if you extend the script with a parallelization, it should work much faster:)

Dear DanBmh, thank you so much! I will update the table above

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
dataset help wanted Extra attention is needed
Projects
None yet
Development

No branches or pull requests

5 participants