Sound waves are the foundation of speech and language. These vibrations travel through air, carrying information about frequency, amplitude, and wavelength. Understanding these properties is crucial for grasping how we produce and perceive speech.
Speech production and perception involve complex processes. From the vocal cords to the brain, our bodies work together to create and interpret sounds. Tools like waveforms and spectrograms help linguists analyze these intricate acoustic patterns in speech.
Acoustic Properties of Sound
Properties of sound waves
- Sound waves propagate as longitudinal waves through a medium (typically air) caused by vibrations of a sound source (vocal cords)
- Frequency measures wave cycles per second in Hertz (Hz) determining pitch perception (low frequencies = low pitch)
- Amplitude refers to maximum wave displacement from rest position relating to sound intensity or loudness measured in decibels (dB)
- Wavelength represents distance between consecutive wave peaks or troughs measured in meters (m) inversely proportional to frequency
- Speed of sound equation $c = f \lambda$ where c ≈ 343 m/s in air at 20℃, f is frequency, λ is wavelength
Speech Production and Perception
Articulatory vs acoustic characteristics
- Vowels characterized by formants (resonant frequencies) with F1 related to tongue height and F2 to tongue advancement (high F1 = low tongue position)
- Consonants produce distinct acoustic patterns: stops create brief silence then burst, fricatives generate turbulent noise, nasals show anti-formants
- Voice onset time (VOT) measures delay between stop release and vocal fold vibration distinguishing voiced from voiceless stops (longer VOT = voiceless)
- Source-filter theory explains speech production: larynx as sound source, vocal tract as modifying filter
- Coarticulation causes overlapping articulatory gestures resulting in acoustic transitions between adjacent sounds (anticipatory or carryover)
Process of speech perception
- Outer ear collects and funnels sound waves, ear canal resonates specific frequencies
- Middle ear transmits vibrations via ossicles (malleus, incus, stapes) amplifying signals
- Inner ear's cochlea converts mechanical vibrations to electrical signals, basilar membrane organized tonotopically
- Auditory nerve transmits signals to brain for processing in auditory cortex
- Categorical perception allows phoneme distinction despite acoustic variations (ba vs pa)
- Top-down processing utilizes context and linguistic knowledge in speech understanding
- McGurk effect demonstrates visual information's influence on auditory perception (seeing "ga" while hearing "ba" results in perceiving "da")
- Waveforms display time-domain representation with amplitude vs time useful for syllable boundaries and VOT
- Spectrograms show time-frequency representation with intensity as color/darkness revealing formants and transitions
- Formant analysis tracks frequencies to identify vowels and analyze quality (plotting F1 vs F2)
- VOT measurement identifies time between stop release and voicing onset (positive VOT for aspirated stops)
- Pitch tracking analyzes fundamental frequency (F0) contours for intonation and stress patterns
- Intensity analysis measures relative loudness of speech segments (stress and prominence)
- Praat software offers various analysis and visualization options for speech research (formant tracking, pitch analysis)