As often happens in a thread there is confusion because people are using words with different ideas of what they mean. My definition of "pitch" is set out in post 11 and I distinguish between definite and indefinite pitch. Sobakus on the other hand uses "pitch" to refer solely to definite pitch.
It's clear to me that, unfortunately, you have insufficient linguistic knowledge to talk about speech sound perception on linguistic science's terms; as a result, you're trying to fill that gap by introducing into it
musical pitch, which leads to wide-ranging confusion. The goal of my replies was to show this and to counteract this confusion. I care deeply about explaining things in an precise, understandable and scientifically correct way, and the effect of your replies in this thread has unfortunately been the opposite on all these counts. I think people who talk with authority on scientific matters on a public forum should take extra caution so as not to spread confusing or confused information.
I must emphatically repeat that speech sounds
are not perceived as pitch, neither absolute nor indefinite. They're perceived as combinations of articulatory features when one's grammar allows their perception; otherwise they're perceived as strange, foreign speech sounds that the listener will struggle to repeat; nevertheless they aren't heard as pitch.
Again, a number of
frequencies, not pitches can be distinguished in vowels. These
individual frequencies will be heard as pitches only when artificially synthesised
in isolation. These perceived pitches might be absolute (definite) or indefinite. But when produced
in combination, humans listeners perceive them as a vowel that they can try and reproduce.
The thread is headed: Why do languages sound different? The answer is that all sounds can be analysed in terms of: how loud they are; how long they last; and their frequencies. It is the different ways that loudness, duration and frequencies combine that explains the difference between one language and another.
Sounds as physical phenomena can indeed be analysed in this way. This is a subject that
physics deals with;
linguistics deals with this only when conducting
instrumental studies. Humans are unable to perceive absolute loudness, duration and frequencies. Birds potentially do that when they imitate human speech, phone rings, dogs and car alarms.
This is what my own not insignificant knowledge of phonetics and phonology tells me. If you disagree with this thesis, please provide any linguistic work that describes human speech perception using these physical terms. It is my firm and well-founded understanding that human perceive
combinations of loudness, duration and frequencies as
speech segments.
The elements of loudness and frequency at the syllable level are relative. In a tonal language the pitch of the tones is not fixed, but varies according to the speaker and to the intonation of the utterance. (I never said that Chinese does not have sentence intonation.)
You cannot deny saying here that Spanish has something that Chinese doesn't, that being "prosodic pitch". Otherwise you would have gone on talking about Chinese instead of introducing Spanish:
Pitch can operate at two levels. In Chinese the pitch assigned to a word establishes its meaning. In Spanish it operates at the prosodic level to, for example, distinguish a statement from a question. Statements and "yes or no" questions may contain the same words in the same order, but the rise in pitch at the end tells you when it is a question.
—
A Spanish speaker distinguishes between gano "I win" stressed on the first syllable and ganó "he won" (stressed on the last syllable) whether he speaks softly or shouts and whether or not he is emphasising the word. Whilst every utterance is a complete entity it can still be analysed at both the syllable and utterance levels depending on what you are looking at.
I have to say it bothers me when I'm being talked past in this manner after I have gone out of my way to explain the issue and illustrate it with diagrams and links to references and publications. Please re-read this:
All the levels listed here, from Syllables to Utterance, are studied holistically by Prosodic Phonology. […] Pitch movement operates on all these prosodic levels - the fundamental frequency (F0) changes over individual Syllables, Feet, Prosodic Words, Phonological and Intonational Phrases, as well as whole Utterances.
You're trying to draw a distinction between "prosodic pitch" which according to you is "prosodic", and "syllable pitch" which isn't prosodic but "phonemic". I've been trying to show that the theory of Prosodic Phonology draws no such distinction. If you wish to contest this statement - and like any scientific-theoretical statement, it
can be contested - you should do so by adducing evidence and referencing scientific publications.
I would be very happy to have that discussion, but I would ask that you stop talking past your interlocutor and trying to introduce musical theory into a linguistic discussion in order to fill your gaps in understanding. I don't pretend for a moment that Prosodic Phonology to be the be-all and end-all descriptive framework for this, but the alternative framework must be a linguistic one, and you must rely on your knowledge and understanding of both the alternative theory and of Prosodic Phonology in order to challenge someone who operates with knowledge and understanding of PPh.
One example of evidence that would disagree with PPh's treatment of lexical tone together with utterance intonation is if you're able to demonstrate that tone processing inside the brain for speakers of Chinese happens in a completely separate manner from intonation processing for speakers of Spanish. I have found a couple of papers that point to a directly opposite conclusion and support PPh's holistic treatment of tone and intonation as part of Prosody. I will quote from the abstract of
Chien, Friederici, Hartwigsen, Sammler (2020). Neural correlates of intonation and lexical tone in tonal and non‐tonal language speakers:
Tone processing overlapped with intonation processing in left fronto‐parietal areas, in both groups, but evoked additional activity in bilateral temporo‐parietal semantic regions and subcortical areas in Mandarin speakers only. Together, these findings confirm cross‐linguistic commonalities in the neural implementation of intonation processing but dissociations for semantic processing of tone only in tonal language speakers.
Li et al. (2010) is a study that finds clear separation between
tones and segments, disagreeing with your treatment of tones as "phonemic":
In direct contrasts between phonological units, tones, relative to consonants and rhymes, yield increased activation in frontoparietal areas of the right hemisphere. This finding indicates that the cortical circuitry subserving lexical tones differs from that of consonants or rhymes.
Both of these studies are of perception. Left-hemisphere damage leading to aphasia doesn't appear to lead to difficulty in
tone production in speakers of tonal languages (
Müller 2015), which is evidence that this happens in the right hemisphere, same as utterance intonation and singing.
I would again ask that you avoid referencing musical pitch perception in our linguistic discussion.