Search
Full catalogue 113 resources
-
Abstract Real-time magnetic resonance imaging (rtMRI) is a technique that provides high-contrast videographic data of human anatomy in motion. Applied to the vocal tract, it is a powerful method for capturing the dynamics of speech and other vocal behaviours by imaging structures internal to the mouth and throat. These images provide a means of studying the physiological basis for...
-
This database contains two non-contemporaneous recordings of each of 68 female speakers of Standard Chinese (a.k.a. Mandarin and Putonghua). 60 of the speakers are from north eastern China, and 8 are from southern China. Each speaker was recorded in three speaking styles: - casual telephone conversation (cnv) - information exchange task over the telephone (fax) - pseudo-police-style interview (int)
-
Multi-laboratory evaluation of forensic voice comparison systems under conditions reflecting those of a real forensic case. There is increasing pressure on forensic laboratories to validate the performance of forensic analysis systems before they are used to assess strength of evidence for presentation in court (including pressure from the recently released report by the President’s Council...
-
Forensic database of voice recordings of 500+ Australian English speakers (AusEng 500+). This database contains 3899 recordings totalling 310 hours of speech from 555 Australian-English speakers. 324 female speakers: - 91 recorded in one recording session - 69 recorded in two separate recording sessions - 159 recorded in three recording sessions - 5 recorded in more than three recording...
-
The MSP-Conversation corpus contains interactions annotated with time-continuous emotional traces for arousal (calm to active), valence (negative to positive), and dominance (weak to strong). Time-continuous annotations offer the flexibility to explore emotional displays at different temporal resolutions while leveraging contextual information. Release 1.0 contains 74 conversations with...
-
Large-scale, weakly-supervised speech recognition models, such as Whisper, have demonstrated impressive results on speech recognition across domains and languages. However, their application to long audio transcription via buffered or sliding window approaches is prone to drifting, hallucination and repetition; and prohibits batched transcription due to their sequential nature. Further,...
-
We study the capabilities of speech processing systems trained simply to predict large amounts of transcripts of audio on the internet. When scaled to 680,000 hours of multilingual and multitask supervision, the resulting models generalize well to standard benchmarks and are often competitive with prior fully supervised results but in a zero-shot transfer setting without the need for any...
-
In this paper, we first provide a review of the state-of-the-art emotional voice conversion research, and the existing emotional speech databases. We then motivate the development of a novel emotional speech database (ESD) that addresses the increasing research need. With this paper, the ESD database1 is now made available to the research community. The ESD database consists of 350 parallel...
-
Twenty five countries have Arabic as an official language, but the dialects spoken vary greatly, and even within one country different accents are heard. Many features create the impression of 'a different accent', including how particular sounds are pronounced, where stress falls in a word, and what intonation pattern is used. There is extensive prior research on the first two of these for...
-
Hi – my name is Simon King and this is my personal website for supporting my teaching. I am the Professor of Speech Processing at the University of Edinburgh, where I teach courses in speech processing and speech synthesis at advanced undergraduate and Masters level. Use of this website Students: You may use this website freely for personal use. You may download copies of the content for your...
-
Welcome to our interactive International Phonetic Association (IPA) chart website! Clicking on the IPA symbols on our charts will allow you to listen to their sounds and see vocal-organ movements imaged with ultrasound, MRI, or in animated form. To find out more about how our IPA charts were made, click on the buttons on the left-hand side of this page. The website contains two main...
-
Dynamic Dialects contains an articulatory video-based corpus of speech samples from world-wide accents of English. Videos in this corpus contain synchronised audio, ultrasound-tongue-imaging video and video of the moving lips. We are continuing to augment the database. The website contains three main resources: - A clickable Accent Map: clicking on points of the map will open up links to...
-
This is a corpus of articulatory data of different forms (EMA, MRI, video, 3D scans of upper/lower jaw, audio etc.) acquired from one male British English speaker.
-
The USC Speech and Vocal Tract Morphology MRI Database consists of real-time magnetic resonance images of dynamic vocal tract shaping during read and spontaneous speech with concurrently recorded denoised audio, and 3D volumetric MRI of vocal tract shapes during vowels and continuant consonants sustained for 7 seconds, from 17 speakers.
-
USC-EMO-MRI is an emotional speech production database which includes real-time magnetic resonance imaging data with synchronized speech audio from five male and five female actors, each producing a passage and a set of sentences in multiple repetitions, while enacting four different target emotions (neutral, happy, angry, sad). The database includes emotion quality evaluation from at least...
Explore
Audio
-
Accent/Region
(13)
- American English (2)
- Arabic (1)
- Australian English (2)
- British English (6)
- World Englishes (3)
- Child Speech (9)
- Conversation (9)
- Directed Speech (1)
- Electroglottography / Electrolaryngography (1)
- Emotional Speech (5)
- Forensic (5)
-
Language
(27)
- Arabic (1)
- Bi-/Multilingual (1)
- English (19)
- French (1)
- L2+ (1)
- Language Learning (2)
- Mandarin (3)
- Multiple (2)
- Multiple (2)
- Spanish (1)
- Multi-Speaker (18)
- Multi-Style (2)
- Pathological (9)
- Singing (2)
- Speech in Noise (3)
- Synthetic Speech (2)
Benchmarks & Validation
- Glottis (2)
Derived & Measured Data
- Formant Measurements (7)
- Fundamental Frequency (2)
- Phone-Level Alignments (1)
- Subglottal Tract (3)
- Vocal Tract (10)
- Vocal Tract Resonances (1)
- Voice Quality Measures (1)
Software, Processing & Utilities
- Articulatory Data Processing (2)
- Feature Extraction (4)
- Image and Volume Segmentation (3)
- Numerical Acoustic Modelling (3)
- Phone Apps (1)
- Speech Processing (5)
- Transcription (3)
- Utilities (4)
Speech Production & Articulation
- Articulography (2)
- Brain Imaging (1)
- MRI (11)
- Ultrasound (10)
- Video (3)
- X-Ray (1)
Teaching Resources
- 3D Models (2)
- Articulation Data (3)
- Tutorials (2)
- Videos (2)
Vocal Anatomy
- Hyoid (1)
- Larynx and Glottis (3)
- Mandible (2)
- Mechanical Properties (1)
- Vocal Tract (11)
Tags
- audio data (46)
- adult (40)
- male (33)
- female (28)
- read speech (23)
- English (23)
- transcribed (13)
- vowels (11)
- MRI (11)
- formant measurement (10)
- spontaneous speech (10)
- child speech (10)
- speech-language pathology (9)
- speech processing (7)
- video (7)
- ultrasound (7)
- teaching resource (6)
- interview (6)
- real-time MRI (rtMRI) (6)
- conversation (6)
- child (6)
- MATLAB (5)
- open-source (5)
- articulatory data (5)
- volumetric MRI (5)
- American English (5)
- vocal tract shape (5)
- segmentation (5)
- automatic speech recognition (ASR) (4)
- speech recognition (4)
- emotional speech (4)
- rtMRI (4)
- annotated (4)
- vocal tract area function (4)
- STL files (3)
- forensic (3)
- telephone (3)
- speaker diarization (3)
- audio processing (3)
- transcription (3)
- Python (3)
- English accents (3)
- British (3)
- angry (3)
- happy (3)
- older adult (3)
- sad (3)
- Mandarin (3)
- perceptually annotated (3)
- speech production (3)
- ultrasound tongue imaging (UTI) (3)
- Newcastle (3)
- DICOM (3)
- computed tomography (CT) (3)
- pathological speech (3)
- speech sound disorder (3)
- numerical acoustic modelling (3)
- source-filter model (2)
- tube model (2)
- Praat (2)
- phonetics (2)
- child-centered audio (2)
- audio (2)
- convert (2)
- file format (2)
- feature extraction (2)
- speech to text (2)
- speech activity detection (2)
- voice activity detection (2)
- whisper (2)
- synthetic speech (2)
- singing (2)
- audiovisual (2)
- articulation (2)
- multimodal (2)
- International Phonetic Alphabet (IPA) (2)
- electromagnetic articulography (EMA) (2)
- lip video (2)
- sociophonetic (2)
- Australian (2)
- phonetic labels (2)
- British English (2)
- L2 English (2)
- finite element method (FEM) (2)
- mandible (2)
- impedance (2)
- vocal tract length (2)
- subglottal tract (2)
- fundamental frequency (2)
- benchmark (2)
- glottis (2)
- videoendoscopy (2)
- multi-language (2)
- 3D print (1)
- Southern standard British English (SSBE) (1)
- map task (1)
- TextGrid (1)
- software (1)
- spectrogram (1)
- speech analysis (1)
- language development (1)
- language environment analysis (LENA) (1)
- word count estimation (1)
- record (1)
- stream (1)
- cepstral peak prominence (CPP) (1)
- harmonic-to-noise ratio (HNR) (1)
- C++ (1)
- classification (1)
- emotion recognition (1)
- speaker identification (1)
- conversational AI (1)
- overlapped speech detection (1)
- speaker embedding (1)
- anechoic (1)
- fast speech (1)
- high pitch (1)
- loud speech (1)
- low pitch (1)
- shout (1)
- slow speech (1)
- deepfake (1)
- logical access (1)
- physical access (1)
- spoof (1)
- speaker detection (1)
- two-class recognizer (1)
- rainbow passage (1)
- labelled (1)
- non-speech (1)
- environmental noise (1)
- noisy audio (1)
- reverberation (1)
- disgust (1)
- surprise (1)
- podcast (1)
- Spanish (1)
- bilingual (1)
- mother-child interaction (1)
- speech rate (1)
- syllable (1)
- syllable nuclei (1)
- consonants (1)
- jaw scans (1)
- accent map (1)
- speech synthesis (1)
- Arabic (1)
- accent variability (1)
- dialect variability (1)
- arousal (1)
- dominance (1)
- valence (1)
- Putonghua (1)
- image processing (1)
- French (1)
- Derby (1)
- Leeds (1)
- Manchester (1)
- York (1)
- digits (1)
- Ohio (1)
- Non-native speech (1)
- adaptation (1)
- diapix (1)
- Middlesbrough (1)
- Sunderland (1)
- speech acoustics (1)
- longitudinal (1)
- formant tracking (1)
- anatomy (1)
- app (1)
- larynx (1)
- typically developing (1)
- x-ray (1)
- x-ray microbeam (1)
- L2 speech (1)
- language learning (1)
- electroglottography (EGG) (1)
- intraoral pressure (1)
- validation (1)
- hyoid (1)
- antiresonance (1)
- vocal tract resonance (1)
- resonance (1)
- corner vowels (1)
- developmental trajectory (1)
- sexual dimorphism (1)
- loudness (1)
- subglottal pressure (1)
- back placement (1)
- chest resonance (1)
- classical (1)
- front placement (1)
- head resonance (1)
- open throat (1)
- roughness (1)
- tenor (1)
- vibrato (1)
- dysarthria (1)
- Amyotrophic Lateral Sclerosis (ALS) (1)
- Down syndrome (1)
- Parkinson's disease (1)
- cerebral palsy (1)
- stroke (1)
- stutter (1)
- cleft (1)
- liquids (1)
- nasals (1)
- plosives (1)
- morphometric (1)
- Lombard speech (1)
- clear speech (1)
- computer-directed speech (1)
- infant-directed speech (1)
- non-native-directed speech (1)
- speech in noise (1)
- Scottish English (1)
- coarticulation (1)
- within-speaker variability (1)
- phone duration (1)
- phone-level alignment (1)
- pitch (1)
- CAPE-V (1)
- GRBAS (1)
- clinical (1)
- voice quality (1)
- area function (1)
- vocal fold model (1)
- vocal tract transfer function (1)
- held vowel (1)
- brain activity (1)
- fMRI (1)
- vocal imitation (1)
- professional voice (1)
- silent speech (1)
- sociolinguistic (1)
- World Englishes (1)
- dyadic (1)
Resource type
- Conference Paper (1)
- Dataset (54)
- Journal Article (21)
- Preprint (2)
- Report (1)
- Software (19)
- Web Page (15)