Search
Full catalogue 113 resources
-
Corpus phonetics has become an increasingly popular method of research in linguistic analysis. With advances in speech technology and computational power, large scale processing of speech data has become a viable technique. A fair number of researchers have exploited these methods, yet these techniques still remain elusive for many. In the words of Mark Liberman, there has been “surprisingly...
-
This website provides a tutorial on R for Linguists. The tutorial provides learners with a foundation to the R programming language using RStudio. It covers the basics of the R grammar, including how to define variables, use R as a calculator, and do basic data manipulation. Building on basic data manipulation, it then covers how to work with full datasets and obtaining summary statistics...
-
The data deposited are taken from fieldwork recordings undertaken with speakers from the three fieldwork sites, Newcastle, Sunderland and Middlesbrough in the Northeast of England. From each locality, 40 informants were recorded, giving a total of 120 informants. The key data in the file All_formants_July_2018.xlsx are vowel formant frequency measurements, in Hertz, for the peripheral...
-
DECTE is an amalgamation of the existing Newcastle Electronic Corpus of Tyneside English (NECTE), created between 2001 and 2005, and NECTE2, a collection of interviews conducted in the Tyneside area since 2007. It thereby constitutes a rare example of a publicly available on-line corpus presenting dialect material spanning five decades.
-
This site allows visitors to access recordings of speakers who stutter and background details about these speakers and the conditions in which the recordings were made. The recordings are available in various formats. The main two sets of recordings were made in normal speaking conditions and the final one was made when the sound of the speaker’s voice was altered as he or she spoke. The three...
-
The current data package includes 1,090 hours of recorded speech (as .wav files) from about 1,130 participants, including those with ALS, cerebral palsy, Down syndrome, Parkinson’s disease and those who have had a stroke. The download also includes text of the original speech prompts and a transcript of the participants’ responses. A subset includes annotations describing the speech...
-
Purpose: The Speech Accessibility Project (SAP) intends to facilitate research and development in automatic speech recognition (ASR) and other machine learning tasks for people with speech disabilities. The purpose of this article is to introduce this project as a resource for researchers, including baseline analysis of the first released data package. ...
-
The Buckeye Corpus of conversational speech contains high-quality recordings from 40 speakers in Columbus OH conversing freely with an interviewer. The speech has been orthographically transcribed and phonetically labeled. The audio and text files, together with time-aligned phonetic labels, are stored in a format for use with speech analysis software (Xwaves and Wavesurfer).
-
The MSP-AVW is an audiovisual whisper corpus for audiovisual speech recognition purpose. The MSP-AVW corpus contains data from 20 female and 20 male speakers. For each subject, three sessions are recorded consisting of read sentences, isolated digits and spontaneous speech. The data is recorded under neutral and whisper conditions. The corpus was collected in a 13ft x 13ft ASHA certified...
-
This 3-year project investigates language change in five urban dialects of Northern England—Derby, Newcastle, York, Leeds and Manchester. Data collection method: Linguistic analysis of speech data (conversational, word list) from samples of different northern English urban communities. Data collection consisted of interviews, which included (1) some structured questions about the interviewee...
-
Ultrasound imaging has been widely adopted in speech research to visualize dynamic tongue movements during speech production. These images are universally used as visual feedback in interventions for articulation disorders or visual cues in speech recognition. Nevertheless, the availability of high-quality audio-ultrasound datasets remains scarce. The present study, therefore, aims to...
-
Abstract The use of real-time magnetic resonance imaging (rt-MRI) of speech is increasing in clinical practice and speech science research. Analysis of such images often requires segmentation of articulators and the vocal tract, and the community is turning to deep-learning-based methods to perform this segmentation. While there are publicly available rt-MRI datasets of speech,...
-
Abstract The study of articulatory gestures has a wide spectrum of applications, notably in speech production and recognition. Sets of phonemes, as well as their articulation, are language-specific; however, existing MRI databases mostly include English speakers. In our present work, we introduce a dataset acquired with MRI from 10 healthy native French speakers. A corpus...
-
Real-time magnetic resonance imaging (rtMRI) is a technique that provides high-contrast videographic data of human anatomy in motion. Applied to the vocal tract, it is a powerful method for capturing the dynamics of speech and other vocal behaviours by imaging structures internal to the mouth and throat. These images provide a means of studying the physiological basis for speech, singing,...
Explore
Audio
-
Accent/Region
(13)
- American English (2)
- Arabic (1)
- Australian English (2)
- British English (6)
- World Englishes (3)
- Child Speech (9)
- Conversation (9)
- Directed Speech (1)
- Electroglottography / Electrolaryngography (1)
- Emotional Speech (5)
- Forensic (5)
-
Language
(27)
- Arabic (1)
- Bi-/Multilingual (1)
- English (19)
- French (1)
- L2+ (1)
- Language Learning (2)
- Mandarin (3)
- Multiple (2)
- Multiple (2)
- Spanish (1)
- Multi-Speaker (18)
- Multi-Style (2)
- Pathological (9)
- Singing (2)
- Speech in Noise (3)
- Synthetic Speech (2)
Benchmarks & Validation
- Glottis (2)
Derived & Measured Data
- Formant Measurements (7)
- Fundamental Frequency (2)
- Phone-Level Alignments (1)
- Subglottal Tract (3)
- Vocal Tract (10)
- Vocal Tract Resonances (1)
- Voice Quality Measures (1)
Software, Processing & Utilities
- Articulatory Data Processing (2)
- Feature Extraction (4)
- Image and Volume Segmentation (3)
- Numerical Acoustic Modelling (3)
- Phone Apps (1)
- Speech Processing (5)
- Transcription (3)
- Utilities (4)
Speech Production & Articulation
- Articulography (2)
- Brain Imaging (1)
- MRI (11)
- Ultrasound (10)
- Video (3)
- X-Ray (1)
Teaching Resources
- 3D Models (2)
- Articulation Data (3)
- Tutorials (2)
- Videos (2)
Vocal Anatomy
- Hyoid (1)
- Larynx and Glottis (3)
- Mandible (2)
- Mechanical Properties (1)
- Vocal Tract (11)
Tags
- audio data (46)
- adult (40)
- male (33)
- female (28)
- read speech (23)
- English (23)
- transcribed (13)
- vowels (11)
- MRI (11)
- formant measurement (10)
- spontaneous speech (10)
- child speech (10)
- speech-language pathology (9)
- speech processing (7)
- video (7)
- ultrasound (7)
- teaching resource (6)
- interview (6)
- real-time MRI (rtMRI) (6)
- conversation (6)
- child (6)
- MATLAB (5)
- open-source (5)
- articulatory data (5)
- volumetric MRI (5)
- American English (5)
- vocal tract shape (5)
- segmentation (5)
- automatic speech recognition (ASR) (4)
- speech recognition (4)
- emotional speech (4)
- rtMRI (4)
- annotated (4)
- vocal tract area function (4)
- STL files (3)
- forensic (3)
- telephone (3)
- speaker diarization (3)
- audio processing (3)
- transcription (3)
- Python (3)
- English accents (3)
- British (3)
- angry (3)
- happy (3)
- older adult (3)
- sad (3)
- Mandarin (3)
- perceptually annotated (3)
- speech production (3)
- ultrasound tongue imaging (UTI) (3)
- Newcastle (3)
- DICOM (3)
- computed tomography (CT) (3)
- pathological speech (3)
- speech sound disorder (3)
- numerical acoustic modelling (3)
- source-filter model (2)
- tube model (2)
- Praat (2)
- phonetics (2)
- child-centered audio (2)
- audio (2)
- convert (2)
- file format (2)
- feature extraction (2)
- speech to text (2)
- speech activity detection (2)
- voice activity detection (2)
- whisper (2)
- synthetic speech (2)
- singing (2)
- audiovisual (2)
- articulation (2)
- multimodal (2)
- International Phonetic Alphabet (IPA) (2)
- electromagnetic articulography (EMA) (2)
- lip video (2)
- sociophonetic (2)
- Australian (2)
- phonetic labels (2)
- British English (2)
- L2 English (2)
- finite element method (FEM) (2)
- mandible (2)
- impedance (2)
- vocal tract length (2)
- subglottal tract (2)
- fundamental frequency (2)
- benchmark (2)
- glottis (2)
- videoendoscopy (2)
- multi-language (2)
- 3D print (1)
- Southern standard British English (SSBE) (1)
- map task (1)
- TextGrid (1)
- software (1)
- spectrogram (1)
- speech analysis (1)
- language development (1)
- language environment analysis (LENA) (1)
- word count estimation (1)
- record (1)
- stream (1)
- cepstral peak prominence (CPP) (1)
- harmonic-to-noise ratio (HNR) (1)
- C++ (1)
- classification (1)
- emotion recognition (1)
- speaker identification (1)
- conversational AI (1)
- overlapped speech detection (1)
- speaker embedding (1)
- anechoic (1)
- fast speech (1)
- high pitch (1)
- loud speech (1)
- low pitch (1)
- shout (1)
- slow speech (1)
- deepfake (1)
- logical access (1)
- physical access (1)
- spoof (1)
- speaker detection (1)
- two-class recognizer (1)
- rainbow passage (1)
- labelled (1)
- non-speech (1)
- environmental noise (1)
- noisy audio (1)
- reverberation (1)
- disgust (1)
- surprise (1)
- podcast (1)
- Spanish (1)
- bilingual (1)
- mother-child interaction (1)
- speech rate (1)
- syllable (1)
- syllable nuclei (1)
- consonants (1)
- jaw scans (1)
- accent map (1)
- speech synthesis (1)
- Arabic (1)
- accent variability (1)
- dialect variability (1)
- arousal (1)
- dominance (1)
- valence (1)
- Putonghua (1)
- image processing (1)
- French (1)
- Derby (1)
- Leeds (1)
- Manchester (1)
- York (1)
- digits (1)
- Ohio (1)
- Non-native speech (1)
- adaptation (1)
- diapix (1)
- Middlesbrough (1)
- Sunderland (1)
- speech acoustics (1)
- longitudinal (1)
- formant tracking (1)
- anatomy (1)
- app (1)
- larynx (1)
- typically developing (1)
- x-ray (1)
- x-ray microbeam (1)
- L2 speech (1)
- language learning (1)
- electroglottography (EGG) (1)
- intraoral pressure (1)
- validation (1)
- hyoid (1)
- antiresonance (1)
- vocal tract resonance (1)
- resonance (1)
- corner vowels (1)
- developmental trajectory (1)
- sexual dimorphism (1)
- loudness (1)
- subglottal pressure (1)
- back placement (1)
- chest resonance (1)
- classical (1)
- front placement (1)
- head resonance (1)
- open throat (1)
- roughness (1)
- tenor (1)
- vibrato (1)
- dysarthria (1)
- Amyotrophic Lateral Sclerosis (ALS) (1)
- Down syndrome (1)
- Parkinson's disease (1)
- cerebral palsy (1)
- stroke (1)
- stutter (1)
- cleft (1)
- liquids (1)
- nasals (1)
- plosives (1)
- morphometric (1)
- Lombard speech (1)
- clear speech (1)
- computer-directed speech (1)
- infant-directed speech (1)
- non-native-directed speech (1)
- speech in noise (1)
- Scottish English (1)
- coarticulation (1)
- within-speaker variability (1)
- phone duration (1)
- phone-level alignment (1)
- pitch (1)
- CAPE-V (1)
- GRBAS (1)
- clinical (1)
- voice quality (1)
- area function (1)
- vocal fold model (1)
- vocal tract transfer function (1)
- held vowel (1)
- brain activity (1)
- fMRI (1)
- vocal imitation (1)
- professional voice (1)
- silent speech (1)
- sociolinguistic (1)
- World Englishes (1)
- dyadic (1)
Resource type
- Conference Paper (1)
- Dataset (54)
- Journal Article (21)
- Preprint (2)
- Report (1)
- Software (19)
- Web Page (15)