Quick viewing(Text Mode)

The Phonology and Phonetics of Speech Prosody: Between Acoustics and Interpretation

The Phonology and Phonetics of Speech Prosody: Between Acoustics and Interpretation

The and of Speech Prosody: Between and Interpretation

Daniel Hirst

CNRS UMR 6057 Parole et Langage, Université de Provence [email protected]

that the former parachutist in question was Spaggiari himself Abstract rather than being someone else who had escaped from prison The way in which prosody contributes to meaning is still, in his company. Similar interpretations could be obtained by today, a poorly understood process corresponding to a different readings of the English translation. mapping between two levels of representation for neither of There is, today, no general consensus on the way in which there is any general consensus. It is argued that which the prosody of an contributes to its meaning. annotation of prosody generally consists in describing both The last six chapters of [7], constituting about half of the prosodic function and prosodic form, but that it would be book, give a well-documented account of the various ways preferable to clearly distinguish the two levels. One in which the problem of the meaning of has been elementary annotation system for prosodic function: IF- approached in the literature. The fact that intonation annotation, is, it is argued, sufficient to capture at least those meaning can be approached in so many different ways, for aspects of prosodic function which influence syntactic results which, as Couper-Kuhlen (p209) admits: interpretation. The annotation of prosodic form can be "are rather modest indeed" carried out automatically by means of an F0 modelling seems to indicate that we have still not got properly started algorithm, MOMEL, and an automatic coding scheme, on the analysis of intonational meaning. As Cruttenden [8] INTSINT. The resulting annotation is underdetermined by puts it: the IF-annotation, but defining mapping rules between "it is not yet even clear what sorts of meanings are representations of function and representation of form could involved." p.184 provide an interesting means of establishing an enriched Surprisingly, in the literature on prosody, there are functional annotation system through analysis by synthesis. fewer publications addressing this subject as compared to phonological, phonetic or acoustic analyses for example, 1. Introduction despite the fact that nearly everyone agrees that this is the central question in the field. Everybody agrees that prosody contributes to the meaning of There are a number of possible explanations for this at an utterance. In fact, when there is a discrepancy between first sight rather surprising fact. The simplest one is that the prosody of the utterance and its overt semantic content researchers tend to be specialists in one specific domain and we usually trust the prosody rather the semantics. So when that the interaction between phonology and interpretation, as someone says: implied in the title of this paper, requires a knowledge of two It's so exciting! fields, phonology on the one hand and syntax, semantics and with a bored of , we tend to believe he is bored pragmatics on the other. Few people working in the field of even though he has overtly said the opposite. Similarly if the syntax, semantics or pragmatics have detailed knowledge sentence: about prosody and the reverse is just as true. He's got nice handwriting…! A further explanation comes from the fact that is pronounced with a falling-rising nucleus on 'handwriting', phonological representation and syntactic, semantic or the utterance is quite likely to be interpreted as a criticism pragmatic interpretation tend to be very theory dependent. even though it is overtly a compliment. To convince an audience that your explanation of the way in For a 'real life' example of the complex way in which which intonation contributes to meaning is correct, you have prosody can contribute to the interpretation of an utterance, to be able to convince them that both your phonology and the following sentence French sentence was pronounced by your interpretation are right. 1 a newsreader on the French national radio France Inter : Not only there is no consensus on the explanation for the Il semble que les policiers sont à deux doigts way that prosody contributes to meaning, there is not even a d'arreter Spaggiari, mais il faudra qu'il real consensus on the way that prosody should be fassent vite pour trouver la cachette de represented phonologically, nor on the way in which we l'ancien parachutiste. should represent the sort of meanings that prosody (It seems that the police are on the point of contributes. arresting Spaggiari but they'll have to act In the rest of this paper I look at the question of prosodic quickly to find the hiding place of the former annotation and the distinction between prosodic function and parachutist) prosodic form. I conclude that a clear separation between these two would be highly desirable and could lead to new Reading this sentence, the first interpretation which insights into the way in which a phonological representation springs to mind is that the police are looking for two escaped of prosody can be mapped onto a functional representation prisoners, one named Spaggiari and the other a former parachutist. The intonation used by the speaker, however, with a falling pitch on 'vite' and a low flat pitch on 'pour trouver la cachette de l'ancien parachutiste', made it clear

1 France Inter, Informations 13-14. 12 mars 1977 2. Prosodic annotation. 2.2. ToBI or not ToBI Ten years after the ToBI standard for prosodic 2.1. The ToBI system: a standard for prosodic annotation of American English was first proposed, one of annotation. the co-authors of the original paper, Colin Wightman [35], In an attempt to meet this need for consensus in prosodic presented a critical evaluation of the usefulness for speech representation, a group of linguists and engineers, mostly technology of this annotation system. American, came to an agreement over a system of One of the major aims of the ToBI project was to representation for the prosody of American English [32]. provide a system which would have a high level of inter- This system, which they called ToBI, an acronym for Tones transcriber agreement. Wightman notes that, while and Break Indices, proposed to represent the prosody of an considerable cross-transcriber agreement has indeed been utterance by means of an alphabet of discrete symbols found for the identification of prominences and boundaries, representing the different pitch accents which had been the agreement is far less consistent for the type of pitch described in American English, decomposed into sequences accent or the type of boundary. He claims furthermore that of symbols H and L (for high and low tones respectively), much of the motivation for large-scale manual transcription together with a scalar representation of the degree of of speech corpora is today obviated by the general separation between consecutive words, going from 0 availability of software capable of extracting prosodic (absence of break) to 4 (major intonation unit break). Within information automatically from the speech signal as well as each pitch accent, one tone symbol is accompanied by the by the massive increase in size of computer memory which [*] indicating that the tone in question is directly makes it possible today to carry out on personal computers associated with the accented of the word. Besides analyses of large corpora, which ten years ago could only the break indices, boundaries are also marked by the have been performed by mainframe computers in specialised presence of a "boundary tone", again either H or L. These institutes. are distinguished from the tones belonging to the pitch Wightman's conclusion is that because of its labour- accents by the presence of a diacritic symbol: [-] for the so- intensive cost, manual transcription should be reserved for called " accent" (in fact a phrase boundary tone) and those aspects of prosody which untrained listeners actually [%] for major intonation boundaries. hear: he formulates this as the maxim: This system was proposed, and accepted, as a standard "Transcribe what you hear!" for the description of the prosody of American English This maxim could be interpreted in a number of ways. and rapidly became the most widely used Listeners obviously hear many different things, including prosodic annotation system in the world. Although the some of which they are not consciously aware. It has been system was originally designed uniquely for the description shown, for example, [14], [[15] that listeners may be of American English, it has rapidly been adapted for a generally incapable, under normal listening conditions, of number of other dialects and including German, distinguishing utterances produced by Italian, Japanese and Chinese. It is worth noting, however, implementing very fine differences in the way in which co- that the authors of the system themselves warn against using articulation phenomena are handled. When the same ToBI indiscriminantly to describe other languages. utterances are heard in adverse conditions, such as with The principal authors of ToBI, insist on the fact that the heavy background noise, however, there is a considerable system was based on a detailed analysis and a particular difference in performance on intelligibility tests between the theory of the intonation system of American English . To two sets of utterances. Very fine details of acoustic describe the prosody of a using this system it is information, then, can be seen to contribute to the robustness indispensable to begin with an exhaustive inventory and of and comprehension. It seems difficult phonological analysis of the possible pitch accents and to claim that listeners do not hear these differences, even boundary tones of the language. though they may be unaware of doing so. Despite these warnings, a number of publications have One way of interpreting Wightman's maxim would be to used an adaptation of ToBI to describe the pitch patterns of say that manual transcription should be reserved for those languages for which there is not yet a complete phonological aspects of prosody that contribute to the listener's description of the intonation system. In doing so they interpretation of the utterance. Transcribers, in other words, attempt to use ToBI as the prosodic equivalent of the should be attentive to prosodic function rather than to International Phonetic Alphabet. prosodic form. In this way, the transcriber is required to The authors of ToBI (ToBI 1999) clearly state (on the perform a task of linguistic interpretation rather than a meta- official ToBI website [33], that: linguistic task of phonetic analysis. As is well known in Note: ToBI is not an International Phonetic psycholinguistic studies, meta-linguistic tasks performed by Alphabet for prosody. Because intonation and untrained subjects entail considerable problems of prosodic organization differ from language to interpretation. language, and often from dialect to dialect In this paper, I suggest that a systematic distinction within a language, there are many different between function and form is a highly desirable aspect of a ToBI systems, each one specific to a language prosodic annotation system. I outline some specific variety and the community of researchers proposals in this area and suggest that such a multi-level working on that language variety. system of annotation could be of interest for speech synthesis and automatic speech recognition as well as for fundamental research into the linguistic analysis of speech prosody. 3. Function and Form in prosody on prosodic function rather than on prosodic form. This suggests that a first approximation for an annotation scheme Like all linguistic phenomena, prosody has both function for prosodic function could be to adapt ToBI by dropping the and form. In many systems of prosodic annotation, perhaps tonal specification and keeping only the boundaries and most, the two levels of representation are intimately prominences. For an adaptation of this type cf. [36]). We can intertwined. There are, however, a number of reasons for call this toneless ToBi or StarBI annotation. distinguishing these levels. A number of years ago, I proposed a slightly more First of all, many prosodic functions seem to be quasi- elaborate functional annotation system [16] with a system universal (Hirst & Di Cristo 1998), in nearly all languages taking into account four degrees of prominence (unstressed, prosody contributes in some way to lexical identity (via tone, stressed nuclear and emphatic) and two types of prosodic 2 quantity and accent) , expressing prominence, boundaries, boundary terminal and non-terminal. An example of this non-finality etc. In the same way, prosodic forms are type of annotation is the following which includes all the certainly universal all languages use rising and falling pitch, symbols: ' [], ° [nucleus], | [boundary], + [non-terminal longer and shorter segments, etc. The mapping between form boundary], || [terminal boundary] __ [emphasis]: and function, however, is certainly not universal. If it were | If you 'can't °lift it + 'ask °Peter to 'help you || so, we should expect all languages to use the same prosodic Corresponding to an utterance like the following: forms to express the same meanings and this is clearly not 250 the case. A second reason is that studying the relationship between prosodic form and function becomes rather circular if a clear distinction between the two levels is not made. To take a fairly trivial example, if we were to make use of a prosodic annotation system that distinguished two rising if you can’t lift it ask Peter to help you

0 intonation contours, calling one a continuation rise and the 0.30570 2.67218 other an interrogative rise, there would be little point in Time (s) 0 examining the correlation between the distribution of the two Ð0.3593 patterns with respect to syntactic or pragmatic criteria unless 0 2.67218 Time (s) the two patterns could be distinguished entirely on the basis of their formal characteristics. Halliday [13], for example Figure 1. Signal and F0 of the utterance "If you can't describes the difference between continuation rises and lift it, ask Peter to help you." interrogative rises as follows: The difference, though gradual, is best I refer to this system, to which I return below, as IF regarded as phonetic overlap (…) the one annotation (which can be glossed as either Intonative being merely lower than the other (…) But Features, the title of the 1977 book, or alternatively as the meanings are fairly distinct. In most cases Intonation Functions). In Hirst (op cit), I argue that this the speaker is clearly using one or the other; annotation is sufficient to account for all those aspects of but sometimes one meets an instance which prosodic representation which contribute directly or could be either. p21 indirectly to syntactic interpretation. Halliday is clearly basing the distinction between the two The original formulation of the IF annotation system was types of rises on prosodic function although he presents presented as a set of distinctive features. Within the them as if they were distinct prosodic forms. framework of non-linear phonology, assuming that segments are grouped into higher level phonological constituents, Such confusion of prosodic form and prosodic function 3 is far more widespread than is commonly realised and this specifically into , units , tonal units and mixing of levels can only be prejudicial to linguistic intonation units, only [± emphatic] and [± terminal] need to analysis. The use of such hybrid annotation is not restricted be retained as features. The utterance above could then be to any particular school or tradition of prosodic analysis. transcribed: Annotation systems developed in the British school, for example (such as [30], [13], [9], [8], [7]), used different symbols to annotate similar types of pitch movement (rising, falling, falling-rising etc) depending on the prosodic function of the pitch movement considered as a pre-nuclear accent or a nucleus. The ToBI annotation system also combines representations of prosodic form (H, L) with representations of prosodic function (- * %).

4. Representing prosodic function which is formally equivalent to the linear IF transcription. I mentioned above that inter-transcriber agreement is in general far higher when transcribers are asked to concentrate

2 French in this respect is rather exceptional in that lexical 3 For arguments in favour of the tonal unit as a phonological representations need to include neither tone, accent nor constituent for English and French cf[17], for the rhythm unit quantity. cf. [27], [4]. 5. Representing prosodic form. 5.2. Surface phonological representation As Wightman [35] observed, the need for manual The prosodic annotation alphabet INTSINT was based annotation of prosodic form is far less evident today with the on the descriptions of the surface patterns of the intonation widespread availability of automatic algorithms for pitch of twenty languages [22] and was used in that volume for the extraction and stylisation. My colleagues and I have description of nine languages (British English, Spanish, suggested [23] that the mapping between prosodic form and European Portuguese, , French, prosodic function should involve a number of different Romanian, Bulgarian, Moroccan and Japanese). levels of representation including a level of phonetic Intonation patterns are analysed as consisting of a representation, consisting of scalar values directly related to sequence of tonal segments, defined in one of two ways: the acoustic signal, and a level of surface phonology, which either globally with respect to the speaker's pitch range unlike the phonetic representation, codes the prosodic form (Top, Mid or Bottom) or locally with respect to the as a sequence of discrete symbols but which are still directly preceding target (Higher, Same or Lower) with an iterative related to the acoustic signal. We also propose a more variant of these locally defined targets (Upstepped, abstract underlying phonological representation of prosodic Downstepped) assuming that an iterative tone can be form, which we assume is more directly related to the followed by the same tone whereas a non-iterative tone representation of prosodic function. In the next sections I cannot and furthermore that the iterative tones correspond to outline briefly the nature of these different levels of a smaller pitch interval than the non-iterative ones. representation. This transcription system, originally designed as a tool for linguists transcribing the intonation of utterances of 5.1. Phonetic representation different languages, was intended to provide at least a first approximation to a prosodic equivalent of the International The MOMEL algorithm developped at the LPL [24], Phonetic Alphabet. As we saw above this is specifically not [23], provides an automatic phonetic representation of a the case for the ToBI system. curve. The algorithm is often In the case of INTSINT, it was intended from the first referred to as a stylisation of fundamental frequency but it that the transcription should be convertible to and from a should more properly be called a model since it consists in sequence of target points. A first version of an algorithm for factoring the raw fundamental frequency curve into two converting between Momel and INTSINT was described in components without any loss of information. These are a [23]. An extension of the system to annotate duration and macroprosodic component, consisting of a continuous timing has also been proposed [19]. smooth curve (represented as a quadratic spline function) A simpler and more robust algorithm has since been corresponding to the linguistic function of the contour, and a developed [20]. In this version, target points are coded on microprosodic component consisting of deviations from the the basis of two speaker dependent parameters: key and macroprosodic curve caused by the nature of the phonematic range. Given these, the absolute tones are defined as the (voiced/unvoiced obstruent, sonorant, etc) (cf limits of the speaker's pitch range (Top and Bottom) [10]). The output of the algorithm is a sequence of target assumed to be symmetrical around the central value (Mid). points which are sufficient to define the macroprosodic The relative tones are then defined by an interval between component of the fundamental frequency when used as input the preceding target point (Pi-1) and the two extreme values to a quadratic spline function. taken as an asymptote for these targets as in the following: Momel is currently available in a number of Pi = Pi-1 + c.(A-Pi ) implementations in various speech-analysis environments where A is either T, (for H and U) or B (for L and D) and including Mes (Unix) [12], SFS (Windows) [26] as well as where c is set at 0.5 for the non-iterative targets H and L and the multi-platform system Praat [3] in the form of a script [1] at 0.25 for the iterative targets U and D. calling Momel as an external C program. The software is This algorithm, applied to the targets of the French and freely available for non-commercial non-military research. English passages of the Eurom1 corpus [6], was optimised A recent evaluation of the algorithm [5] was made using over the parameter space: recordings of the continuous passages of the Eurom1 corpus key = mean ± 50 (in Hz) for five languages (English, German, Spanish, French, range [0.5, 2.5] (in octaves). Italian) in all a total of 5 hours of speech. The evaluation Interestingly, the mean optimal range parameter estimated a global precision of 97.6% by comparison with resulting from this analysis was not significantly different manually corrected target estimation. Compared to the from 1.0 octave. It remains to be seen, however, how far this 46982 target points provided by the automatic analysis, 3179 result is due to the nature of the EUROM1 corpus which was were added manually by the correctors and 1107 removed. analysed (40 passages consisting each of 5 semantically The algorithm gave only slightly worse results (93.4% connected sentences) and whether it can be generalised to precision) when applied to a corpus of spontaneous spoken other speech styles and other (particularly non-European) French. The majority of these corrections involved languages. systematic errors, in particular before pauses, which an The symbolic coding of the F0 target points obviously improvement of the algorithm should eliminate. entails some loss of information with respect to the original The output of the algorithm as a sequence of target data, unlike the Momel analysis which is entirely reversible. points is particularly suitable for interpretation as a sequence The loss of information is, however, quite small as can of tonal segments such as the INTSINT representation be seen from Figure 2 which illustrates the output from the described below, but the relatively theory-neutral nature of optimised INTSINT coding compared to the original target the modelling, together with its reversibility, has allowed the points for a complete five sentence passage from the Eurom1 algorithm to be used as input for other types of annotation corpus. including ToBI [34], [28] and the Fujisaki model [29]. 200 180 160 140 120 100 Hz 80 60 40 20 0 original coded time (s)

Figure 2. Coding of the F0 targets from a passage from the Eurom1 corpus showing the original target points(lozenges) estimated by the Momel algorithm and the target points(squares) derived from the optimised INTSINT coding.

Work in progress involves the elaboration of mapping 6. From prosodic function to prosodic form. rules between IF annotation and INTSINT annotation for both English and French. Converting IF to INTSINT is fairly As I mentioned above, Momel and INTSINT provide simple. An implementation for French, is described in [11]. reversible representations of intonation patterns since not The inverse mapping, however, is not currently feasible, only can they be derived automatically from the acoustic since IF in its present state can generate only a subset of signal but it is also possible to convert a sequence of possible and observed INTSINT patterns. INTSINT symbols, together with two speaker/utterance Thus for example [+emphatic] in British English, will dependent parameters key and range, into a sequence of generally correspond to a high falling nuclear pitch accent target points which can then be converted to a smoothed when followed by a [+terminal] boundary but to a rising- fundamental frequency curve. falling nuclear pitch accent when followed by a non-terminal The example given above: boundary. There are, however, a number of secondary | If you 'can't °lift it + 'ask °Peter to 'help you || characteristics which often, but not always, accompany would be converted to : emphatic nuclear pitch accents. The high falling pattern, for If you can't lift it ask Peter to help you. example, is often preceded by an upstepping head. This M H B H M B T B B corresponds to the global pattern which has sometimes been which, in turn, can be converted to an appropriate sequence called the "surprise/redundancy" pattern. The effect of the of tonal targets as input to a speech synthesis system such as: upstepping head is to reinforce the fact that the final fall is If you can't lift it ask Peter to help you. higher than the preceding accent. This characteristic, while 135 163 95 143 135 95 191 95 95 very common for emphatic terminal pitch patterns, is by no which can finally be used to produce a continuous F0 curve: means the only possibility and seems to represent a separate 250 choice on the part of the speaker. The fact that an upstepping head begins with a low accent may furthermore be reinforced by a high onset for any preceding unstressed syllables (high pre-head). Once again, while this is a common characteristic of upstepping heads it is by no means necessary and it is not restricted to 0 this context, either. 0 2.67218 Time (s) Similarly, a falling-rising nuclear pattern (emphatic non- terminal) is frequently preceded, in British English, by a sequence of falling pitch patterns on the pre-nuclear accents (the head). This is in fact the only context, in British English, (contrary to American and Scottish English) where this type ifyoucan’t lift it ask Peter to help you of pattern is fairly systematic although it is by no means 0 2.67218 Time (s) impossible in other contexts. In [18] I argue that surface phonological representations Figure 3. Continuous F0 curve produced from a for non-emphatic and emphatic intonation patterns in British quadratic spline function defined by a sequence of 9 English can be derived from rather abstract underlying target points. phonological representations which are quite naturally related to the prosodic structure represented in the IF annotation. In this approach, two prosodically very different References languages such as English and French can be characterised by means of a small number of abstract prosodic parameters. French, like other Romance languages would be [1] Auran, C., 2003. Momel and Intsint package. characterised as having a right-headed Tonal Unit while http://www.lpl.univ-aix.fr/~auran/ English, like other Germanic languages would have a left- [2] Auran, C., C. Bouzon & D.J.Hirst, 2004. The Aix- headed Tonal Unit. The underlying tonal template for Marsec nproject: an evolutive database of spoken British French, in this analysis, would be the sequence [L H] English. in Bel & Marlien (eds) Proceedings of Speech whereas in English the underlying sequence would be [H L]. Prosody 2004. It seems furthermore that the hierarchy of prosodic [3] Boersma, P., & Weenink, D., 1995-2003. Praat: a constituents is different in English and French. It seems that system for doing phonetics by computer. English possesses a rhythm unit which is at a lower level htttp://www.fon.hum.uva.nl/praat/ than the tonal unit (as suggested by Jassem over 50 years ago [4] Bouzon, C & D.J.Hirst, 2004. and prosodic [27] and confirmed in recent work [4] using a large (5 1/2 structure in British English. in Bel & Marlien (eds) hour) corpus of spoken English [2]). In French it seems more Proceedings of Speech Prosody 2004. appropriate to consider the tonal unit as bei,ng on a lower [5] Campione, E., 2001. Etiquetage prosodique semi- level than the rhythm unit. One of the results of this automatique de corpus oraux : algorithmes et parametrisation of the phonology of prosodic systems might méthodologie. Doctoral thesis.. Aix-en-Provence: be that for French, unlike for English, there is no Université de Provence.. distinctivity for the "nuclear" pitch accent since the [6] Chan, D., Fourcin, A., Gibbon, D., Granström, B., possibility of the nuclear accent in English occuring on a Huckvale, M., Kokkinas, G., Kvale, L., Lamel, L., non-final stress (without emphasis) is a consequence of the Lindberg, L., Moreno, A., Mouropoulos, J., Senia, F., possibility of grouping several rhythm units into a single Trancoso, I., Veld, C., & Zeiliger, J., 1995. EUROM: a tonal unit, which is not possible in French apart from in resource for the EU. Proceedings of emphatic patterns or with post-posed sentence adverbs. the 4th European Conference on Speech Communication and Speech Tecnology, Eurospeech '95, 7. From prosodic form to prosodic function. (Madrid) 1, 867-880. [7] Couper-Kuhlen, E., 1986. An Introduction to English The ultimate aim of describing the prosody of natural Prosody. London: Arnold. language utterances is to provide a deeper understanding of [8] Cruttenden, A., 1986. Intonation. Cambridge: the way in which prosody contributes to the interpretation of Cambridge University Press. these utterances. Such a goal clearly has implications for [9] Crystal, D. 1969., Prosodic Systems and Intonation in speech technology since, as Wightman [35] notes, despite English. Cambridge: Cambridge University Press. the considerable research invested in the transcription of [10] Di Cristo, A. & Hirst, D.J., 1986. Modelling French prosody of various different languages in the last decade, the micromelody. Analysis and synthesis. Phonetica,. 43, actual implementation of prosody in TTS or ASR 11-30. applications is remarkably limited. Paradoxically, as [11] Di Cristo, A.; Di Cristo, P. & Véronis, J., 1997. A Ostendorf has noted [31], speech technology is even more in metrical model of rhythm and intonation for French text- need of prosodic aids than human speakers, in both t-speech synthesis. in A. Botinis (ed.) Intonation: production and perception, since computers have far less Theory, Models and Applications (ESCA), 83-86. knowledge of the world than humans to help them to [12] Espesser, R., 1996. MES : Un environnement de interpret utterances. traitement du signal. Proceedings XXIe Journées The under-determination of the INTSINT representation d'Etude sur la Parole 1996 June 10-14 : Avignon, with respect to IF annotation, suggests a strategy of analysis France, p. 447.. by synthesis which seems rather promising. [13] Halliday, M.A.K, 1967. Intonation and in In this approach, a preliminary IF annotation is used to British English. The Hague: Mouton. generate an INTSINT representation, which is then [14] Hawkins, S., and Slater, A., 1994. Spread of CV and V- compared to the annotation derived from the actual to-V coarticulation in British English: Implications for recording. Systematic differences can then be used to either the intelligibility of synthetic speech. ICSLP 94 correct the mapping rules or to extend the IF annotation (Proceedings of the 1994 International Conference on system which in its present state is obviously rather spoken Language Processing) 1, 57-60. rudimentary. In the emphatic examples which we discussed [15] Heid, S. and Hawkins, S., 1999. Synthesizing systematic above, for example, we might decide that the high falling variation at boundaries between and obstruents. nuclear pattern, the upstepping head and the high pre-head In J.J. Ohala, Y. Hasegawa, M. Ohala, D. Granville, and constitute three independent choices for the speaker with A.C. Bailey (eds.), Proceedings of the XIVth respect to the emphatic nature of the utterance. International Congress of Phonetic Sciences. University This in turn suggests a number of experimental of California, Berkeley, CA. 1, 511-514. paradigms to examine the orthogonality of such subsets of [16] Hirst, D.J. 1977., Intonative features. A syntactic intonation patterns which we intend to explore in more detail approach to English intonation. (= Janua Linguarum in future work. series minor 139), Mouton, The Hague. [17] Hirst, D.J. 1988., Tonal units as phonological constituents: the evidence from French and English intonation. in H. Van der Hulst & N. Smith (eds) 1988, Autosegmental studies in pitch accent. Foris, Dordrecht, 151-165. [18] Hirst, D.J., 1998. Intonation in British English. in Hirst & Di Cristo (eds.) 1998. [19] Hirst, D.J., 1999., The symbolic coding of segmental duration and tonal alignment. An extension to the INTSINT system. Proceedings ICSLP '99. [20] Hirst, D.J. 2000., Optimising the INTSINT coding of F0 targets for multi-lingual speech synthesis. ISCA Workshop: Prosody 2000 Speech recognition and Synthesis (Kraków, October 2000). [21] Hirst, D.J. & Di Cristo, A., 1998. A survey of intonation systems. in Hirst & Di Cristo (eds) 1998, . [22] Hirst, D.J. & Di Cristo, A., (eds). 1998. Intonation systems. A survey of twenty languages. Cambridge University Press. Cambridge. [23] Hirst, D.J., Di Cristo, A. & Espesser, R., 2000. Levels of representation and levels of analysis for the description of intonation systems. in M. Horne (ed) 2000, 51-87. [24] Hirst, D.J. & Espesser, R., 1993. Automatic modelling of fundamental frequency curves using a quadratic spline function. Travaux de l'Institut de Phonétique d'Aix, 15, 71-85. [25] Horne, M., (ed) Prosody: Theory and Experiment. Studies Presented to Gösta Bruce. Dordrecht: Kluwer Academic Publishers. [26] Huckvale, M., 2000. Speech Filing System. Tools for speech research . http://www.phon.ucl.ac.uk/resource/sfs/ [27] Jassem, W. 1952., Intonation of Conversational English (Educated Southern British). Wroclaw, Wroclawskie Towarzystwo Naukow. [28] Maghbouleh, A., 1998. ToBI accent type recognition. Proceedings ICSLP '98. [29] Mixdorff, H. 1999., A novel approach to the fully automatic extraction of Fujisaki model parameters. ICASSP 1999. [30] OConnor, J.D. and Arnold, G.F., 1961. Intonation of Colloquial English. London: Longman (2nd edition, 1973). [31] Ostendorf, M. 2000., Prosodic boundary detection. in M. Horne (ed) 2000, 263-279. [32] Silverman, K., Beckman, M., Pitrelli, J., Ostendorf, M., Wightman, C., Price, P., Pierrehumbert, J., Hirschberg, J., 1992. ToBI: a standard for labelling English prosody. Proceedings ICSLP92, 2, 867- 870, Banff, Canada. [33] TobI 1999. http://www.ling.ohio-state.edu/~tobi/ [34] Wightman, C. & Campbell, N., 1995. Improved labeling of prosodic structure. IEEE Trans. on Speech and Audio Processing. [35] Wightman, C. 2002. ToBI or not ToBI? in B.Bel & I.Marlien (eds) Proceedings of Speech Prosody 2002, Aix en Provence. [36] Wightman, C. W., Syrdal, A. K. et al., 2000. Perceptually based automatic prosody labeling and prosodically enriched unit selection improve concatenative speech synthesis, in Proceedings ICSLP, vol. 2, pp. 7174.