Spoken languages have been classified into stress-timed, syllable-timed, and mora-timed languages. English and Dutch have been regarded as stress-timed languages; French, Italian, and Spanish as syllable-timed languages (Abercrombie, Reference Abercrombie1967; Lloyd James, Reference Lloyd James1940; Pike, Reference Pike1945); and Japanese as a mora-timed language (Benguerel, Reference Benguerel1999; Bloch, Reference Bloch1950; Han, Reference Han1962; Hoequist, Reference Hoequist1983a, Reference Hoequist1983b; Ladefoged, Reference Ladefoged1975; Port, Dalby, & O’Dell, Reference Port, Dalby and O’Dell1987). This traditional classification was based on the assumption that the unit of rhythm (i.e., foot in stress-timed languages, syllable in syllable-timed languages, and mora in mora-timed languages) occurs at regular intervals. Although the isochrony of the rhythm units has not been empirically supported in later studies (e.g., Arvaniti, Reference Arvaniti2012; Dauer, Reference Dauer1983; Roach, Reference Roach and Crystal1982; Wenk & Wiolland, Reference Wenk and Wioland1982), these rhythm types are still believed to be perceptually distinguishable even by infants (Nazzi, Bertoncini, & Mehler, Reference Nazzi, Bertoncini and Mehler1998; Ramus, Dupoux, & Mehler, Reference Ramus, Dupoux and Mehler2003).
In order to quantify such rhythmic differences, a variety of rhythm metrics have been developed. Ramus et al. (Reference Ramus, Nespor and Mehler1999) proposed three interval measures: %V (the proportion of vocalic intervals in a sentence), ΔV (the standard deviation of vocalic intervals), and ΔC (the standard deviation of consonantal intervals). They showed that because stress-timed languages permit more complex syllable structures and have vowel reduction in unstressed syllables, they can have lower %V and higher ΔV and ΔC compared to syllable-timed languages. Another widely used rhythm metric is the pairwise variability index (PVI hereafter) proposed in Low, Grabe, and Nolan (Reference Low, Grabe and Nolan2000). The PVI measures the level of variability in syllable duration by calculating the average difference in duration between two successive syllables, with stress-timed languages expected to have higher PVI scores than syllable-timed languages. Grabe and Low (Reference Grabe, Low, Gussenhoven and Warner2002) proposed raw PVI for consonantal intervals (rPVI-C) measured without speech rate normalization, and normalized PVI for vocalic intervals (nPVI-V) to control for speech rate variation in vocalic intervals. In addition, rate-normalized standard deviations, VarcoV and VarcoC, were developed in Dellwo (Reference Dellwo, Karnowski and Szigeti2006).
These rhythm metrics have been widely used for various languages. Despite the popularity, there have been large discrepancies between studies. For example, different studies have classified the same languages into different rhythm types, or some studies have failed to find significant differences even between prototypical languages such as English and French (e.g., Baltazani, Reference Baltazani2007; Grabe & Low, Reference Grabe, Low, Gussenhoven and Warner2002; White & Mattys, Reference White and Mattys2007; see Arvaniti, Reference Arvaniti2012, for a review). Arvaniti (Reference Arvaniti2012) tested the performance of the rhythm metrics (ΔC, %V, rPVI, nPVI, VarcoC, and VarcoV) on English, German, Greek, Italian, Spanish, and Korean. Her results showed that the metric scores were affected by various types of “noise” in the data such as elicitation methods (e.g., read or spontaneous speech), the complexity of syllable structure, and interspeaker variation. This suggests that the rhythm metrics may not be robust enough to quantify rhythmic differences between languages.
The aim of the present study was to find reliable metrics that can quantify rhythmic characteristics of non-native speakers. Previous research has also investigated the rhythm of non-native speech using those rhythm metrics or acoustic correlates of stress (i.e., pitch, intensity, duration, or vowel quality). Non-native speakers whose native language has been described as a syllable-timed or mora-timed language (e.g., Malay, Vietnamese, Spanish, and Japanese) were shown to speak English with characteristics of their first language rhythm (e.g., Carter, Reference Carter2004, Reference Carter2005a, Reference Carter, Gess and Rubin2005b; Mochizukisudo & Fokes, Reference Mochizukisudo and Fokes1985; Mochizukisudo & Kiritani, Reference Mochizukisudo and Kiritani1991; Nguyen, Reference Nguyen2003; White & Mattys, Reference White and Mattys2007). Similarly, Singapore English and Hong Kong English were found to have smaller variability in syllable durations than British English as measured by the rhythm metrics, due to the influence of Singaporean Mandarin and Cantonese, respectively (Deterding, Reference Deterding2001; Low et al., Reference Low, Grabe and Nolan2000; Setter, Reference Setter2006).
However, it remains uncertain whether the rhythm metrics can be used to evaluate the rhythmic characteristics of non-native learners with varying levels of proficiency. For example, Stockmal, Markus, and Bond (Reference Stockmal, Markus and Bond2005) showed that ΔC and PVI-C can differentiate low-proficiency and high-proficiency Russian learners of Latvian. In this case, the authors suggested that the high variability of the rhythm scores among nonproficient learners was due to their imperfect fluency rather than the influence of their native language (L1). In contrast, Chen and Zechner (Reference Chen and Zechner2011) did not find any significant correlation between rhythm scores and proficiency levels for non-native speakers of English whose L1 was Mandarin, although adding the rhythm metrics to their automatic English speech scoring system significantly improved the performance of the system (i.e., higher agreement between machine-predicted scores and human scores) compared to a model that was only based on nonrhythm features.
It also remains debated how much speech rhythm contributes to the perception of non-native speech. For example, some studies have found that fluency-related characteristics such as pause duration, pause frequency, and speech rate are stronger predictors of proficiency or foreign accentedness than rhythm-related characteristics such as the degree of stress-timing (e.g., Trofimovich & Baker, Reference Trofimovich and Baker2006; Iwashita, Brown, McNamara, & O’Hagan, Reference Iwashita, Brown, McNamara and O’Hagan2008). Similarly, Jenkins (Reference Jenkins2000) excluded English stress-timed rhythm from her lingua franca core by describing it as a “non-core feature,” whereas nuclear stress was described as one of the most important factors affecting intelligibility in communication among non-native speakers (“core feature”) due to its role in conveying important or new information in the sentence. Although quantifying the relative contribution of rhythm to the perception of second language (L2) speech compared to other aspects of speech is beyond the scope of the present study, it is important to establish stress metrics that can quantify non-native speakers’ rhythm proficiency reliably.
The present study investigated the production of speech rhythm by Korean learners of English. Some studies have suggested that Korean learners’ English rhythm is more syllable timed than stress timed due to the influence of their L1 rhythm (e.g., Jang, Reference Jang2008; Kim, Flynn, & Oh, Reference Kim, Flynn and Oh2007; Lee & Kim, Reference Lee and Kim2005). However, it has been debated in previous research whether Korean is a stress-timed or syllable-timed language, with some studies suggesting that it does not belong to any rhythm class (see Jeon Reference Jeon, Brown and Yeon2015, for a brief review).Footnote 1 Korean rhythm has some characteristics of a stress-timed language in that Korean permits syllable codas and syllable weight plays an important role in shaping Korean rhythm patterns (cf. Lee, Reference Lee1990). However, Korean can also be more similar to a syllable-timed language in that Korean stress is not easily perceived due to the lack of vowel reduction, which is one of the most important distinctive properties of stress-timed languages (Dauer, Reference Dauer1983). Lee (Reference Lee2011) suggested that low-proficiency Korean learners tend to place sentence stress on most words in a sentence including grammatical words and use strong vowels in unstressed syllables, which gives the impression of syllable-timed rhythm, or even word-timed rhythm. In contrast, the rhythm of high-proficiency speakers sounds closer to stress-timed rhythm, suggesting that the rhythm of non-native speakers with the same L1 background can vary depending on their proficiency.
It remains unclear whether the rhythm metrics are suitable for quantifying the differences in rhythm between Korean learners of English. Jang (Reference Jang2008) found significant correlations between Korean learners’ English proficiency and some of the rhythm metrics, such as %V, VarcoV, and nPVI-V, but only for a subset of sentences, whereas other metrics, such as the proportion of the duration of function words in a sentence,Footnote 2 articulation rate, and the number of silence intervals, significantly correlated with proficiency for most of the sentences that were tested. Although the limitations of the rhythm metrics have been previously reported (e.g., Arvaniti, Reference Arvaniti2012; Wiget et al., Reference Wiget, White, Schuppler, Grenon, Rauch and Mattys2010), we investigated whether the rhythm metrics can capture rhythmic differences between Korean learners of English in a large-scale study and compared these with the metrics that we will propose in this study.
It should be noted that the rhythm metrics described above are only derived from durational patterns of speech.Footnote 3 However, Fuchs (Reference Fuchs2014) has developed nPVI-V metrics that calculate the degree of variability in loudness and simultaneous variability in loudness and duration of vocalic intervals. While his new metrics were shown to be effective in showing rhythmic differences between Indian and British English, very few studies have used acoustic correlates other than duration to quantify rhythmic characteristics. Temporal characteristics measured by the previous rhythm metrics can only partially account for perceived rhythm because English rhythm is associated with the alternation of stressed and unstressed syllables. That is, the rhythmic beat of stressed syllables is manifested not only in prolonged duration but also in higher (or sometimes lower) pitch and stronger intensity in English.
Rhythmic beats that are perceived on the utterance level are referred to as sentence stress, whose placement was called “accentuation” in Gimson (Reference Gimson1980). For example, grammatical words do not usually bear sentence stress when produced in a sentence. A syllable receiving sentence stress and any following unstressed syllables comprises a foot in English (Abercrombie, Reference Abercrombie1967). Placing sentence stress on most words in a sentence including grammatical words can thus lead to more syllable-timed rhythm patterns. Sentence stress in this paper is also different from “pitch accent,” which is the stress involving salient pitch prominence that is caused by an important intonational event in intonational phonology (cf. Pierrehumbert, 2000; Silverman et al., Reference Silverman, Beckman, Pitrelli, Ostendorf, Wightman, Price and Hirschberg1992). Sentence stresses cause the impression of stress-timed rhythm whereas pitch accent serves to lead a pitch pattern (e.g., head and nuclear tone in the British tradition) in intonation. In Jassem’s (Reference Jassem1999) prosodic model, sentence stress, which was termed “tertiary accent,” was clearly distinguished from word stress (“potential for accent”) and pitch accent (“primary or secondary accents”).
Furthermore, because one of the functions of sentence stress is to mark semantically important words, analyzing the speaker’s sentence stress assignment may measure what is perceived as speech rhythm more accurately than simply calculating the degree of temporal variation between intervals (i.e., the previous rhythm metrics). That is, it is also important to assess which words and syllables received sentence stress in a sentence.
As mentioned above, low-proficiency Korean learners tend to place sentence stress on most words in a sentence including grammatical words, whereas high-proficiency learners tend to place sentence stress mostly on content words as native speakers do. To verify this observation using reliable measures, we propose two sentence stress metrics: “accentuation rate” (the ratio of the number of accented words to the total number of test words) and “accentuation error rate” (the ratio of the number of sentence stress errors to the total number of words in test sentences). In the present study, low-proficiency learners were expected to speak English with higher accentuation and accentuation error rates than high-proficiency learners.
To sum up, the aim of the present study was to find robust rhythm measures to validate our hypothesis that speech rhythm of Korean learners of English varies depending on their proficiency levels. This study thus investigated whether scores of our sentence stress metrics and the rhythm metrics (i.e., interval measures and PVIs) correlated with the learners’ proficiency levels. We used 525 sentences that were extracted from a large sentence corpus of Korean learners of English. The present study enhanced the reliability of the rhythm measurements by using a large number of sentences and speakers, compared to previous studies, which often used a very small number of sentences but with more controlled syllable structure (e.g., Ramus, Nespor, & Mehler, Reference Ramus, Nespor and Mehler1999; White & Mattys, Reference White and Mattys2007).
Methods
Materials
In order to carry out the rhythm analyses, we used the Korean Learners’ English Accentuation Corpus (KLEAC), which was made to develop an automatic sentence stress prediction, detection, and feedback system (Lee et al., Reference Lee, Lee, Song, Kim, Kang, Lee and Hwang2017). This database consists of recordings of 5,500 English sentences read by 75 Korean learners, who were middle school students aged between 13 and 14. The sentences were made up of words and sentence structures that were appropriate for the students, as they were originally developed as the training corpus of an automatic speech scoring system for secondary school students in Korea.Footnote 4 The speakers of the KLEAC were learning English as an L2 both at school and at a private English institute in South Korea at the time of testing, where the average length of learning was 7–8 years. Seven phonetically trained Korean labelers marked sentence stress on the stressed syllables imposed on the recorded sentences. The annotations were partly crosschecked between the labelers at an early stage, and they had regular meetings to discuss problematic cases that they had faced in the labeling process. The labelers showed very strong interrater agreement rates (e.g., Fless’s κ was .868; see Lee et al., Reference Lee, Lee, Song, Kim, Kang, Lee and Hwang2017, for further details).
Sentence stress errors were also marked on a separate tier as shown in Figure 1. They occurred when a grammatical word was stressed by mistake, when a content word was not stressed, or when stress fell on the wrong syllable. However, it should be noted that some grammatical words like demonstratives (e.g., “this” and “that”) tend to receive sentence stress while some content words like relative adverbs (i.e., “when,” “why,” and “where”) do not (see Kingdon, Reference Kingdon1958. for details).
Sentences of the KLEAC database were also different from the materials used in previous rhythm studies (e.g., Arvaniti, Reference Arvaniti2012; Low et al., Reference Low, Grabe and Nolan2000). Specifically, syllable complexity, or the distribution of stressed and unstressed syllables within a sentence, was controlled in some of the previous studies to manipulate the durational variability of successive intervals. In contrast, the sentences used in the present study were not contrived as in the previous studies. Instead, 7 sentences were randomly selected from each of the 75 speakers (i.e., a total of 525 sentences; 97 unique sentences), which allowed for a greater number of natural sentences to be used. This large set of sentences can better represent the phonological structures of English (e.g., syllable complexity or the number of unstressed and stressed syllables), thereby increasing the reliability of the metric scores without being too sensitive to characteristics of individual sentences. Average sentence length was approximately 10 syllables.
Speaking proficiency rating task
Seven native English speakers performed the speaking proficiency rating task. They were not phonetically trained annotators, but they were all familiar with Korean-accented English and were living in South Korea at the time of testing. The raters were instructed to listen to seven sentences from each speaker and assess their overall speaking proficiency level on a 5-point scale, with 5 being the most intelligible and nativelike and 1 being the least intelligible and nativelike. The raters were given practice trials at the beginning of the task and were allowed to listen to the sentences as many times as they needed.
The proficiency scale used in this study is simplified from the 6-point Interagency Language Roundtable scale, which is the standard grading scale for measuring speaking proficiency in the United States (see http://www.govtilr.org/skills/ILRscale2.htm). Because all the Korean speakers had learned English for 7–8 years at school, we omitted the speaking 0 level (no proficiency). As the raters assessed the learners’ overall speaking proficiency, the ratings can reflect not only accentedness and intelligibility (e.g., Munro & Derwing, Reference Munro and Derwing1995) but also fluency (e.g., Tavakoli & Skehan, Reference Tavakoli, Skehan and Ellis2005).
Calculation of rhythm metrics
The interval measures used in this study were %V, ΔV, ΔC, VarcoV, and VarcoC. The pairwise variability indices (i.e., rPVI-V, rPVI-C, nPVI-V, and nPVI-C) were also used. The definition of each metric is provided in Table 1.
To calculate these rhythm metrics, vocalic and consonantal intervals of the sentences were initially autosegmented with a forced aligner that extracts the vocalic nucleus of each syllable (Mertens, Reference Mertens2004), and they were then manually modified by two phonetically trained labelers. This was performed following the conventions used in previous studies on the rhythm metrics, such as Ramus et al. (Reference Ramus, Nespor and Mehler1999) and Grabe and Low (Reference Grabe, Low, Gussenhoven and Warner2002). For example, when more than one vowel appeared consecutively, they were labeled as one vocalic interval. This also applied when segmenting consonantal intervals. Pauses were excluded from the analysis. The rhythm metrics were calculated using the software Correlatore 2.1 (Mairano & Romano, Reference Mairano, Romano, Schmid, Schwarzenbach and Studer2010), and the values of each rhythm metric were averaged across sentences for each speaker in order to perform correlation analyses between their proficiency scores and rhythm metric values.
Calculation of accentuation and accentuation error rates
Sentence stress was also analyzed using the same 525 sentences (i.e., 7 sentences per speaker) from the KLEAC database. As explained, sentence stress and sentence stress errors were annotated in the corpus. In this study, the total number of words in the sentences, the number of sentence stresses imposed, and the number of sentence stress errors were counted for each speaker to calculate the rate of accentuation and accentuation errors as shown in the equations in (1).
(1) Accentuation and accentuation error rates
a. Accentuation rate
= Total number of sentence stresses / Total number of words × 100 (%)
b. Accentuation error rate
= Total number of sentence stress errors / Total number of words × 100 (%)
We conducted correlation analyses to see if individual speakers’ accentuation and accentuation error rates correlated with their proficiency scores. All statistical analyses in this study were performed in R.
Results
Correlation between rhythm metrics and proficiency levels
Correlation analyses were performed for each speaker’s average score on each rhythm metric and their average proficiency score. The distribution of the individual speakers’ average proficiency scores is displayed in Figure 2. Even though fewer speakers received high scores over 3–4 (i.e., the data was positively skewed to some degree), their proficiency scores covered a wide range on the scale (range: 1.429 ~ 5, median: 2.571, mean: 2.866). The results of the correlation analysis showed that out of nine rhythm metrics, four rhythm metrics (ΔV, ΔC, rPVI-V, and rPVI-C) were significantly correlated with proficiency scores, as shown in Table 2. However, as displayed in Figure 3, there were negative relationships between the rhythm metrics and proficiency. That is, speakers who received higher proficiency scores (i.e., more nativelike) had lower rhythm scores, more similar to those of a syllable-timed language. This was the opposite of our expectation. However, it should be noted that the significant correlations were only found with metrics that did not control for speech rate variation, suggesting that the higher rhythm scores of low-proficiency learners were probably related to their slower speech rate. Other metrics failed to show any significant correlation with proficiency.
Correlation between accentuation and accentuation error rates and proficiency levels
Correlation analyses were also carried out for individual learners’ accentuation and accentuation error rates and proficiency scores. As shown in Figure 4, there was a significant negative relationship between learners’ accentuation rates and proficiency scores (Spearman’s ρ = –.57, p < .001), meaning the higher the learner’s proficiency score (i.e., more nativelike), the fewer sentence stresses they produced. Similarly, there was a significant negative correlation between learners’ accentuation error rates and proficiency scores (Spearman’s ρ = –.60, p < .001), which means that the higher the learner’s proficiency score was, the fewer accentuation errors they made. That is, the perceived level of Korean learners’ overall speaking proficiency was highly correlated with their accuracy in producing sentence stress, which is closely related to rhythmic patterns of speech.
Discussion and conclusion
It has previously been observed that rhythmic characteristics in English speech produced by Korean learners vary widely depending on their English proficiency, with more proficient learners producing more nativelike, stress-timed rhythm (e.g., Lee, Reference Lee2011). However, previous studies have not been able to quantify the rhythm of non-native speakers using reliable measures. The aim of this study was to find metrics that are robust enough to validate this observation. The results showed that Korean learners’ rhythmic beat placement as measured by accentuation and accentuation error rates was more nativelike as their speaking proficiency increased.
It thus appears that our method of evaluating non-native speakers’ English rhythm using sentence stress metrics can appropriately quantify rhythmic differences between non-native speakers with varying levels of speaking proficiency. The present study also demonstrates that rhythmic characteristics in non-native speech are closely related to judgments of the speaker’s speaking proficiency by native listeners, in contrast to what some previous studies have found, such as Iwashita et al. (Reference Iwashita, Brown, McNamara and O’Hagan2008). As the assessment of speaking proficiency in this paper was based on read speech, not on spontaneous conversational speech, it inevitably had some limitations in evaluating the individual learners’ speaking proficiency. According to Pearson’s technical report on Versant English Test (Pearson Education, 2011), however, there was a high degree of correlation between the scores of the Versant English Test, which is automatically assessed mostly with read and repeated speech, and those of other well-established speaking tests of English, which are evaluated with conversational speech. Hence, basing the assessment on read speech did not appear to pose a serious problem in carrying out this study. In addition, our measure of speaking proficiency can be linked to the degree of foreign accentedness or intelligibility, so it would be interesting to investigate how non-native production of rhythm or sentence stress affects accentedness and intelligibility in future studies.
In contrast, despite using a large set of sentences (i.e., 525 sentences) to calculate the rhythm metrics reliably in the present study, the pairwise variability indices and interval measures did not show the rhythmic differences between Korean learners. The negative correlation found between learners’ proficiency levels and some of the rhythm metrics was unexpected. It is unlikely that the rhythm of low-proficiency learners was more similar to that of native speakers. Given that the negative correlation was only found in raw variability indices or interval measures that do not control for speech rate variation (i.e., rPVI-V, rPVI-C, ΔV, and ΔC), it seems that the high metric scores of low-proficiency learners were driven by slower speaking rate or disfluencies in their speech. That is, long intervals in their speech might have increased the degree of absolute differences between intervals (Dellwo, Reference Dellwo, Karnowski and Szigeti2006). Similar results have been found in previous studies due to slower speaking rates of non-native speakers (e.g., Jang, Reference Jang2008; Lin & Wang, Reference Lin and Wang2007).
One may argue that the rhythm metrics might have accurately captured Korean learners’ rhythmic characteristics, which did not differ depending on proficiency. It is also possible to argue that rhythmic characteristics did not simply affect the perception of proficiency. However, such explanations would not be congruent with what we found with the sentence stress metrics; the placement of rhythmic beats by less-proficient Korean learners was more similar to that of syllable-timed rhythm or word-timed rhythm. Although it is difficult to determine the validity of the rhythm metrics solely based on the current findings, it is more likely that the rhythm metrics were not very accurate at capturing rhythmic differences between non-native learners with different levels of proficiency, as well as those between different languages (e.g., Arvaniti, Reference Arvaniti2012).
The present study suggests that sentence stress can be an effective measure for evaluating Korean learners’ English rhythm proficiency. Automatic English speech scoring systems have been developed to assess English learners’ speaking proficiency levels (e.g., Bernstein, Van Moere, & Cheng, Reference Bernstein, Van Moere and Cheng2010; Blake, Wilson, Cetto, & Pardo-Ballester, Reference Blake, Wilson, Cetto and Pardo–Ballester2008; Chandel et al., Reference Chandel, Parate, Madathingal, Pant, Rajput, Ikbal and Verma2007; Chen & Zechner, Reference Chen and Zechner2011; Johnson, Kang, & Ghanem, Reference Johnson, Kang and Ghanem2016; Nielson, Reference Nielson2011; Teixeira, Franco, Shriberg, Precoda, & Sonmez, Reference Teixeira, Franco, Shriberg, Precoda and Sonmez2000; Zechner, Higgins, & Xi, Reference Zechner, Higgins and Xi2007; Zechner, Xi, & Chen, Reference Zechner, Xi and Chen2011). However, most systems have heavily relied on fluency-related features like speech tempo and the number and duration of filled pauses, and rhythmic features have rarely been used except in Chen and Zechner (Reference Chen and Zechner2011), where rhythm metrics were found to be useful in improving the accuracy of the system.
The results of this paper suggest that incorporating our sentence stress metrics into an automatic English speech scoring system may improve its accuracy. Lee et al. (Reference Lee, Lee, Song, Kim, Kang, Lee and Hwang2017) have developed an automatic sentence stress prediction, detection, and feedback system. This system analyzes learners’ sentence stress placement using a detection model that is trained using acoustic, lexical, and syntactic features, then compares it with a reference generated by a prediction model, and offers feedback to the learners on the errors that they made. The accuracy of the prediction and detection of this system reached 96.6% and 84.1%, respectively. This level of accuracy is high enough to suggest that the system can automatically assess accentuation and accentuation error rates, and the system has been shown to be effective in improving learners’ accentedness and rhythm (Lee et al., Reference Lee, Lee, Song, Kim, Kang, Lee and Hwang2017). Furthermore, the results of the present study have important implications for the teaching of English pronunciation in general; focusing on lowering the number of sentence stress errors can be an efficient training method to help learners increase their rhythm proficiency.
Acknowledgments
This work was supported by the Seoul National University Research Grant in 2017. The KLEAC database we used for this study was developed in a research project funded by SK Telecom of South Korea. The authors would also like to thank Dr. Kiduk Yoon and other colleagues for their contribution to the development of the KLEAC database.