Skip to main content

About

The Cornell Phonetics Lab is a group of students and faculty who are curious about speech. We study patterns in speech — in both movement and sound. We do a variety research — experiments, fieldwork, and corpus studies. We test theories and build models of the mechanisms that create patterns. Learn more about our Research. See below for information on our events and our facilities.

/

Upcoming Events


  • 27th October 2025 12:20 PM

    PhonDAWG - Phonetics Lab Data Analysis Working Group

    Sam will present a preliminary analysis of his ambisyllabicity experiment.

     

     

    Location: : B11 Morrill Hall, 159 Central Avenue, Morrill Hall, Ithaca, NY 14853-4701, USA
  • 28th October 2025 04:30 PM

    ASL Linguistics Lecture Series: Dr. Naomi Caselli

    The Cornell Linguistics Department proudly presents Dr. Naomi Caselli, Associate Professor, Boston University.   Dr. Caselli will lecture on:  "ASL Vocabulary Acquisition!"

     

    ASL/English Interpretation provided, light refreshments to follow. 

     

    Abstract:

     

    This talk addresses four central questions in research on deaf children’s language acquisition:

    1. Can children with hearing parents develop age-expected sign-language vocabularies?

    2. How do parent sign language skills relate to child vocabulary outcomes?

    3. Does learning sign language interfere with spoken language development?

    4. Does early access to sign language support long-term academic achievement?

     

    These questions are at the center of longstanding debates in deaf education, with significant implications for families, educators, and policymakers.

     

    Drawing on recent population-level studies and emerging tools such as the ASL CDI 2.0, this presentation situates these issues within the broader evidence base on language deprivation, access, and development.

     

     

     

    Speaker Bio:

     

    Dr. Naomi Caselli is an associate professor of Deaf education, director of the Deaf Center, and the director of the AI and Education initiative at Boston University.

     

    She is hearing, and her first languages are American Sign Language (ASL) and English. She leads a research team that works to make research on language—across education, computer science, linguistics, psychology, and medicine—inclusive of sign languages, and to ensure all deaf children have access to language.

     

    Her research is centered on three questions:

    • How does early language experience shape how deaf children learn language?
    • How is the sign language lexicon structured, learned, and processed?
    • How can we responsibly use AI to make the world more accessible to sign language users?

     

    She is internationally recognized for creating open-access sign language databases that support research and education.

     

    Location: 106 Morrill Hall, Cornell University, 159 Central Avenue, Morrill Hall, Ithaca, NY 14853-4701, USA
  • 29th October 2025 12:20 PM

    Phonetics Lab Meeting

    Annabelle will lead a discussion on this paper on Whispered speech.  

     

    "Auditory speaker discrimination by forensic phoneticians and naive listeners in voiced and whispered speech" October 2015, International Journal of Speech Language and the Law 22(2):229-248, DOI:10.1558/ijsll.v22i2.23101

     

    Abstract:

     

    In whispered speech some important cues to a speaker’s identity (e.g. fundamental frequency, intonation) are inevitably absent.

     

    In the present study we investigated listeners’ ability to discriminate between speakers in short utterances in voiced and whispered speech.

     

    The performances of a group of 11 forensic phoneticians and a group of 22 naive listeners were compared in a binary forced-choice speaker discrimination task, with 48 same-speaker and 60 different-speaker pairs of short speech samples (≤ 3 s) in each test. Listeners were asked to say whether the two voice samples in each pair were produced by the same or different speakers, and to give a certainty rating.

     

    The results showed that speaker discrimination is more difficult in whispered than in voiced speech, and that while the phoneticians were only slightly better than the naive listeners in voiced speech, the gap widened in whispered speech.

     

    Phoneticians were more cautious in their responses, but also more accurate than naive listeners. When unsure, the phoneticians tended to say two utterances came from different speakers, whereas naive listeners tended to say two utterances came from the same speaker.

     

    Results support the view that trained phoneticians may have an advantage over naive listeners in auditory speaker discrimination when the signal is degraded.

     

    Location: B11 Morrill Hall, 159 Central Avenue, Morrill Hall, Ithaca, NY 14853-4701, USA
  • 30th October 2025 04:30 PM

    Colloquium Talk Series - Marten van Schijndel

    The Cornell Linguistics Department is proud to present Dr. Marten van Schijndel, Associate Professor of Linguistics at Cornell University, who will give a talk titled:  "Largely Language Modeling".

     

    Abstract:

     

    Language modeling is an old task which has gained popularity as the backbone of the current AI boom. This is, in part, because the task is simple to implement at scale, and in part because the method has achieved good success at a number of language tasks.

     

    However, the learned representations of neural language models are opaque, making it difficult to assess what they encode. Many current discussions about language models center around their emergent properties; their apparent ability to transcend the language modeling task and instead achieve human-like language understanding.

     

    While language models work well for natural language processing tasks, I argue that language model representations differ in key ways from human linguistic representations.

     

    I discuss multiple studies from my group that identify weaknesses in these models and in the ways they are used which produce mismatches with human linguistic response patterns.

     

    While I provide methods for mitigating some of these weaknesses, caution should be employed when using these models since they are largely language models.

     

     

    Location: 106 Morrill Hall, Cornell University, 159 Central Avenue, Morrill Hall, Ithaca, NY 14853-4701, USA

Facilities

The Cornell Phonetics Laboratory (CPL) provides an integrated environment for the experimental study of speech and language, including its production, perception, and acquisition.

Located in Morrill Hall, the laboratory consists of six adjacent rooms and covers about 1,600 square feet. Its facilities include a variety of hardware and software for analyzing and editing speech, for running experiments, for synthesizing speech, and for developing and testing phonetic, phonological, and psycholinguistic models.

Web-Based Phonetics and Phonology Experiments with LabVanced

 

The Phonetics Lab licenses the LabVanced software for designing and conducting web-based experiments.

 

Labvanced has particular value for phonetics and phonology experiments because of its:

 

  • *Flexible audio/video recording capabilities and online eye-tracking.
  • *Presentation of any kind of stimuli, including audio and video
  • *Highly accurate response time measurement    
  • *Researchers can interactively build experiments with LabVanced's graphical task builder, without having to write any code.

 

Students and Faculty are currently using LabVanced to design web experiments involving eye-tracking, audio recording, and perception studies.  

 

Subjects are recruited via several online systems:

 

 

 

 

Computing Resources

 

The Phonetics Lab maintains two Linux servers that are located in the Rhodes Hall server farm:

 

  • Lingual -  This Ubuntu Linux web server hosts the Phonetics Lab Drupal websites, along with a number of event and faculty/grad student HTML/CSS websites.  

 

  • Uvular - This Ubuntu Linux dual-processor, 24-core, two GPU server is the computational workhorse for the Phonetics lab, and is primarily used for deep-learning projects.

 

In addition to the Phonetics Lab servers, students can request access to additional computing resources of the Computational Linguistics lab:

 

  • *Badjak - a Linux GPU-based compute server with eight NVIDIA GeForce RTX 2080Ti GPUs

 

  • *Compute server #2 - a Linux GPU-based compute server with eight NVIDIA  A5000 GPUs

 

  • *Oelek  - a Linux NFS storage server that supports Badjak. 

 

These servers, in turn, are nodes in the G2 Computing Cluster, which currently consists of 195 servers (82 CPU-only servers and 113 GPU servers) consisting of ~7400 CPU cores and 698 GPUs.

 

The G2 Cluster uses the SLURM Workload Manager for submitting batch jobs  that can run on any available server or GPU on any cluster node. 

 

 

 

 

Articulate Instruments - Micro Speech Research Ultrasound System

We use this Articulate Instruments Micro Speech Research Ultrasound System to investigate how fine-grained variation in speech articulation connects to phonological structure.

 

The ultrasound system is portable and non-invasive, making it ideal for collecting articulatory data in the field.

 

 

BIOPAC MP-160 System

The Sound Booth Laboratory has a BIOPAC MP-160 system for physiological data collection.   This system supports two BIOPAC Respiratory Effort Transducers and their associated interface modules.

Language Corpora

  • The Cornell Linguistics Department has more than 915 language corpora from the Linguistic Data Consortium (LDC), consisting of high-quality text, audio, and video corpora in more than 60 languages.    In addition, we receive three to four new language corpora per month under an LDC license maintained by the Cornell Library.

 

 

  • These and other corpora are available to Cornell students, staff, faculty, post-docs, and visiting scholars for research in the broad area of "natural language processing", which of course includes all ongoing Phonetics Lab research activities.   

 

  • This Confluence wiki page - only available to Cornell faculty & students -  outlines the corpora access procedures for faculty supervised research.

 

Speech Aerodynamics

Studies of the aerodynamics of speech production are conducted with our Glottal Enterprises oral and nasal airflow and pressure transducers.

Electroglottography

We use a Glottal Enterprises EG-2 electroglottograph for noninvasive measurement of vocal fold vibration.

Real-time vocal tract MRI

Our lab is part of the Cornell Speech Imaging Group (SIG), a cross-disciplinary team of researchers using real-time magnetic resonance imaging to study the dynamics of speech articulation.

Articulatory movement tracking

We use the Northern Digital Inc. Wave motion-capture system to study speech articulatory patterns and motor control.

Sound Booth

Our isolated sound recording booth serves a range of purposes--from basic recording to perceptual,  psycholinguistic, and ultrasonic experimentation. 

 

We also have the necessary software and audio interfaces to perform low latency real-time auditory feedback experiments via MATLAB and Audapter.