Your search
Results 23 resources
-
A multi-speaker corpus of ultrasound images of the tongue and video images of the lips The Tongue and Lips (TaL) corpus is a multi-speaker corpus of ultrasound images of the tongue and video images of lips. This corpus contains synchronised imaging data of extraoral (lips) and intraoral (tongue) articulators from 82 native speakers of English. The TaL corpus consists of two datasets: - TaL1...
-
We introduce the Speak & Improve Corpus 2025, a dataset of L2 learner English data with holistic scores and language error annotation, collected from open (spontaneous) speaking tests on the Speak & Improve learning platform. The aim of the corpus release is to address a major challenge to developing L2 spoken language processing systems, the lack of publicly available data with high-quality...
-
The data deposited are taken from fieldwork recordings undertaken with speakers from the three fieldwork sites, Newcastle, Sunderland and Middlesbrough in the Northeast of England. From each locality, 40 informants were recorded, giving a total of 120 informants. The key data in the file All_formants_July_2018.xlsx are vowel formant frequency measurements, in Hertz, for the peripheral...
-
DECTE is an amalgamation of the existing Newcastle Electronic Corpus of Tyneside English (NECTE), created between 2001 and 2005, and NECTE2, a collection of interviews conducted in the Tyneside area since 2007. It thereby constitutes a rare example of a publicly available on-line corpus presenting dialect material spanning five decades.
-
The Buckeye Corpus of conversational speech contains high-quality recordings from 40 speakers in Columbus OH conversing freely with an interviewer. The speech has been orthographically transcribed and phonetically labeled. The audio and text files, together with time-aligned phonetic labels, are stored in a format for use with speech analysis software (Xwaves and Wavesurfer).
-
The MSP-AVW is an audiovisual whisper corpus for audiovisual speech recognition purpose. The MSP-AVW corpus contains data from 20 female and 20 male speakers. For each subject, three sessions are recorded consisting of read sentences, isolated digits and spontaneous speech. The data is recorded under neutral and whisper conditions. The corpus was collected in a 13ft x 13ft ASHA certified...
-
This 3-year project investigates language change in five urban dialects of Northern England—Derby, Newcastle, York, Leeds and Manchester. Data collection method: Linguistic analysis of speech data (conversational, word list) from samples of different northern English urban communities. Data collection consisted of interviews, which included (1) some structured questions about the interviewee...
-
Abstract The use of real-time magnetic resonance imaging (rt-MRI) of speech is increasing in clinical practice and speech science research. Analysis of such images often requires segmentation of articulators and the vocal tract, and the community is turning to deep-learning-based methods to perform this segmentation. While there are publicly available rt-MRI datasets of speech,...
-
Multi-laboratory evaluation of forensic voice comparison systems under conditions reflecting those of a real forensic case. There is increasing pressure on forensic laboratories to validate the performance of forensic analysis systems before they are used to assess strength of evidence for presentation in court (including pressure from the recently released report by the President’s Council...
-
Forensic database of voice recordings of 500+ Australian English speakers (AusEng 500+). This database contains 3899 recordings totalling 310 hours of speech from 555 Australian-English speakers. 324 female speakers: - 91 recorded in one recording session - 69 recorded in two separate recording sessions - 159 recorded in three recording sessions - 5 recorded in more than three recording...
-
The MSP-Conversation corpus contains interactions annotated with time-continuous emotional traces for arousal (calm to active), valence (negative to positive), and dominance (weak to strong). Time-continuous annotations offer the flexibility to explore emotional displays at different temporal resolutions while leveraging contextual information. Release 1.0 contains 74 conversations with...
-
Welcome to our interactive International Phonetic Association (IPA) chart website! Clicking on the IPA symbols on our charts will allow you to listen to their sounds and see vocal-organ movements imaged with ultrasound, MRI, or in animated form. To find out more about how our IPA charts were made, click on the buttons on the left-hand side of this page. The website contains two main...
-
This is a corpus of articulatory data of different forms (EMA, MRI, video, 3D scans of upper/lower jaw, audio etc.) acquired from one male British English speaker.
-
USC-EMO-MRI is an emotional speech production database which includes real-time magnetic resonance imaging data with synchronized speech audio from five male and five female actors, each producing a passage and a set of sentences in multiple repetitions, while enacting four different target emotions (neutral, happy, angry, sad). The database includes emotion quality evaluation from at least...
Explore
Audio
-
Accent/Region
(8)
- American English (1)
- Australian English (2)
- British English (4)
- World Englishes (1)
- Child Speech (1)
- Conversation (6)
- Emotional Speech (3)
- Forensic (2)
-
Language
(14)
- Bi-/Multilingual (1)
- English (14)
- L2+ (1)
- Language Learning (1)
- Mandarin (1)
- Multiple (1)
- Spanish (1)
- Multi-Speaker (11)
- Multi-Style (1)
- Speech in Noise (2)
Derived & Measured Data
Software, Processing & Utilities
Speech Production & Articulation
- Articulography (2)
- MRI (6)
- Ultrasound (2)
- Video (3)
Teaching Resources
Vocal Anatomy
- Vocal Tract (6)
Tags
- English
- audio data (17)
- adult (16)
- read speech (11)
- male (11)
- female (9)
- transcribed (5)
- MRI (5)
- conversation (5)
- articulatory data (4)
- real-time MRI (rtMRI) (4)
- spontaneous speech (4)
- English accents (3)
- British (3)
- perceptually annotated (3)
- video (3)
- interview (3)
- Newcastle (3)
- angry (2)
- emotional speech (2)
- happy (2)
- sad (2)
- articulation (2)
- American English (2)
- electromagnetic articulography (EMA) (2)
- speech production (2)
- Australian (2)
- forensic (2)
- phonetic labels (2)
- British English (2)
- rainbow passage (1)
- environmental noise (1)
- noisy audio (1)
- reverberation (1)
- Mandarin (1)
- surprise (1)
- podcast (1)
- Spanish (1)
- bilingual (1)
- child speech (1)
- child-centered audio (1)
- mother-child interaction (1)
- multimodal (1)
- volumetric MRI (1)
- jaw scans (1)
- International Phonetic Alphabet (IPA) (1)
- lip video (1)
- teaching resource (1)
- ultrasound tongue imaging (UTI) (1)
- arousal (1)
- dominance (1)
- valence (1)
- telephone (1)
- rtMRI (1)
- segmentation (1)
- Derby (1)
- Leeds (1)
- Manchester (1)
- York (1)
- audiovisual (1)
- digits (1)
- whisper (1)
- Ohio (1)
- Non-native speech (1)
- adaptation (1)
- diapix (1)
- Middlesbrough (1)
- Sunderland (1)
- L2 English (1)
- L2 speech (1)
- annotated (1)
- language learning (1)
- professional voice (1)
- silent speech (1)
- ultrasound (1)
Resource type
- Dataset (16)
- Journal Article (2)
- Report (1)
- Software (1)
- Web Page (3)