he design of an instrument to evaluate sotware for EFL / ESL pronunciation teaching

his study aimed at designing an instrument composed of 72 questions that evaluate the extent to which sotware programs teach pronunciation of English as a Foreign Language and/or Second Language (EFL/ESL) following the principles of the Communicative Approach (Celce-Murcia et al, 2010). he designed instrument was tested for reliability and validity. For this purpose, 46 EFL/ESL teachers used the instrument to analyze an online version of the sotware program Pronunciation Power 2. he participants’ answers to the questions were subjected to reliability and validity tests. he results of these statistical tests suggest the instrument is potentially valid for evaluating EFL/ESL pronunciation teaching sotware.


Introduction
With the advent of globalization, there is a growing need to communicate with people of diferent nationalities.his communication includes the use of oral language, and helps explain the increasing difusion of English teaching and learning throughout the world.According to the British Council (2012), a non-ministerial department of the UK government responsible for education, one out of four of the world's population speaks English with some level of competence.English is the primary language used for international contacts and business transactions.hus, English has become an international language Given these facts, sotware programs for developing English pronunciation have increasingly been launched in the market.Publishers have promoted these programs stating they are efective and of good quality.However, when it comes to the advertising discourse of English language teaching publishers, promotion is not always factual, as investigated by Carvalho (2011).Teachers and learners should not be seduced by the strong appeal of the marketing done by publishers.Instead, it is necessary to analyze English as a Foreign Language and/or Second Language (EFL/ESL) pronunciation teaching sotware programs as to their potential for developing English pronunciation.here is an unquestionable need to analyze these programs from a critical perspective using pedagogically coherent and technically elaborated criteria (Navarro, 1999).
While several studies sought to establish criteria and evaluation instruments for analyzing educational sotware programs in general, others investigated criteria and evaluation instruments for analyzing language learning sotware programs.Table 1 provides an overview of some of the studies done to establish criteria and evaluation instruments for analyzing educational sotware programs.Seven evaluation categories: 1) Traditional Techniques; 2) Strategies; 3) Metaphors; 4) Goals; 5) Picture; 6) Sound; 7) Animated screensavers.
as described by Jenkins (2000) and Walker (2010), among others.Indeed, studying English has become a basic need for acceptance and qualiication in almost all professions.Aware of the growing number of people interested in learning this language worldwide, the publishing industry has released thousands of books, CDs, DVDs and sotware programs claiming to help learners master English.In order to make their business more proitable, the publishing industry is increasingly seeking to tailor their materials to particular audiences (Assumpção Filho, 2011).
With the advent of computer technology and growing technological appeal, the use of sotware programs has been one more option among the set of materials used when teaching and prioritizing pronunciation.Levis wrote that the use of computers is almost ideally suited to learning pronunciation skills.Computers can provide individualized instruction, frequent practice through listening discrimination and focused repetition exercises, and automatic visual support that demonstrates to learners how closely their own pronunciation approximates model utterances.(2007, p.184)The analysis of these studies indicates that there are no specific criteria for evaluating pronunciation software programs, although some studies, such as Chapelle and Jamieson (2008), establish criteria for evaluating the skill of speaking in general.Therefore, there is a need for studies that attempt to establish specific criteria to evaluate programs for teaching pronunciation.There might be pronunciation software programs whose interface may look attractive but fail at reflecting solid grounded principles for teaching pronunciation.
hus, this study aimed at designing an instrument to evaluate the extent to which sotware programs teach EFL/ESL pronunciation following the principles of the Communicative Approach (Celce-Murcia, Brinton, Goodwin, & Griner, 2010).More speciically, this study focused on answering the following questions:

heoretical framework
Because no criteria for the evaluation of pronunciation sotware programs currently exist, the purpose of this study was to design an instrument to evaluate the extent to which sotware programs designed Oliveira, Costa and Moreira (2001) Four evaluation categories: 1) Interaction student-sotware programteacher; 2) Pedagogical basis; 3) Content; 4) Programming.(2006) hree evaluation categories: 1) Technological features; 2) Pedagogical features; 3) Individualized-learning features.Fino (2003) Criteria for the use of educational sotware programs were more important than evaluation criteria and the teacher.for teaching English pronunciation to FL/SL learners follow the principles of the Communicative Approach (Celce-Murcia et al, 2010). he designed instrument was tested for reliability and validity.Underlying this process was the belief that sotware programs should be based on an understanding of Second Language (SL) 1 acquisition, more speciically SL phonological acquisition, and the principles of the Communicative Approach (Celce-Murcia et al, 2010).In this section, the principles which were used to ground the instrument are briely described.

SL Phonological Acquisition
Phonology brings together a set of segments (phonemes) that can distinguish meaning.It is by combining these segments that spoken language is perceived and produced by an individual.As the acquisition of lexical, syntactic, semantic and pragmatic aspects, the acquisition of phonological aspects is also essential in the perception and production of spoken language (Pennington, 1999).
Research studies have been done to explain how SL phonological acquisition occurs.In this ield, one of the most inluential researchers is James Emil Flege. Flege (1995, 2002, 2003) developed the Speech Learning Model (SLM), which posited that an individual's perception level of phonological aspects was positively related to the accurate production of these aspects.
Another model that deals with SL phonological acquisition is the Perceptual Assimilation Model (PAM) proposed by Best (1995) and extended by Best and Tyler (2007).In the second model (PAM-L2), Best and Tyler (2007) stated that adults perceived phonological aspects of a SL by comparing the similarities and diferences of the SL with the irst language (L1), especially when they had little contact with the SL.It was through this contact that phonological aspects were perceptually learned.
he model proposed by Best and Tyler (2007) shows the relevance of perceiving distinctive language sounds to SL phonological acquisition, while the model proposed by Flege (1995Flege ( , 2002Flege ( , 2003) ) shows the relevance of perceiving and producing these sounds to SL phonological acquisition.herefore, sotware programs designed for teaching English pronunciation to Foreign Language/Second Language (FL/SL) learners should have activities that work on the perception and production of diferent phonological aspects of the English language.Figure 1 shows examples of the questions in the designed instrument that analyze the activities that develop sound perception and production (for more examples of the questions that analyze this characteristic, please refer to the Evaluation Instrument in the Appendix).

he Place of Pronunciation in a Communicative Approach
Most researchers agree that communicative competence does not necessarily mean to sound like a native speaker of a language.Instead, the goal of teaching pronunciation is to have the learners acquire intelligible pronunciation, in other words, being able to make themselves understood by others without communication breakdown (Jenkins, 2000).Pronunciation teaching in this approach is also associated with the learners performing interactive pronunciation activities in the classroom.he teacher can make use of technical explanations, but it is important that learners have the opportunity to interact with others while making use of both segmental and suprasegmental aspects of language (see Figure 2).Suprasegmental aspects are as important as segmental ones and both should be taught in a communicative approach (Celce-Murcia et al, 2010; Levis & Levelle, 2009).Celce-Murcia et al ( 2010) proposed a communicative framework for teaching pronunciation and suggested that a pronunciation lesson should be divided into ive phases: 1. DESCRIPTION AND ANALYSISoral and written illustrations of how the feature is produced and when it occurs within spoken discourse; 2. LISTENING DISCRIMINATION -focused listening practice with feedback on learners' ability to correctly discriminate the feature; 3. CONTROLLED PRACTICEoral reading of minimal-pair sentences, short dialogues, etc., with special attention paid to the highlighted feature in order to raise learner consciousness; 4. GUIDED PRACTICEstructured communication exercises, such as information-gap activities or cued dialogues, that enable the learner to monitor for the speciied feature; 5. COMMUNICATIVE PRACTICE -less structured, luency-building activities (e.g., role-play, problem solving) that require the learner to attend to both form and content of utterances.(p.45) Celce-Murcia et al ( 2010) also stated that it is essential for the learners to get systematic feedback in all these phases (see Figure 3). he use of phonetic symbols is also considered valuable in pronunciation teaching (Kelly, 2000), (see Figure 4).Apart from helping learners visually perceive pronunciation, the use of phonetic symbols can help them become more autonomous in their learning (Celce-Murcia et al, 2010).Another relevant pedagogical procedure is to contrast the learners' mother tongue sounds to the target language sounds (see Figure 5).his procedure may help the learners raise awareness of the diferences between the two languages, identify which sounds are more diicult and set strategies to work on typical errors (Dale, 2001).Exposing the learners to highly variable input leads to gains in pronunciation learning (Bybee & Torres Cacoullos, 2008;hompson, 2011). he learners should be aware of diferent regional phonological varieties to understand the English spoken by diferent speakers of English (Kelly, 2000).hey should also develop intelligible rather than native-like pronunciation (Celce-Murcia et al, 2010; Levis & Levelle, 2009).Acquiring a pronunciation that can be understood by any other speaker of English is a more realistic goal for pronunciation teaching (see Figure 6).

Ways to Best Utilize the Potential of Computer-Assisted Pronunciation Teaching (CAPT)
enrich learning and assist in the acquisition process.Levis (2007) stated that "for any teacher who thinks that pronunciation is essential, Computer-Assisted Pronunciation Teaching (CAPT) is immensely promising" (p.196).
In fact, there are many advantages to the use of CAPT.When comparing the teaching of pronunciation mediated by a phonetician or a teacher of the same area, Pennington (1996) argued that CAPT may prove to be superior in some aspects.With the computer, it is possible to analyze the users' pronunciation and give them feedback faster than a human does.Since people usually sufer from limitations such as patience, hearing and judgment, this analysis via computer can be performed more accurately and more reliably an ininite number of times (see Figure 7). he computer may present various phonological aspects in more ways than a human can.hus, the users can easily receive diferent types of input, become more conscious about their pronunciation, have a better understanding of many phonological aspects and increase their motivation to work on their pronunciation (Pennington, 1996).
In addition, the computer may be able to provide more prominent feedback than a human, and this prominence arises from the use of visual aids such as videos, animations, pictures, graphics, sounds, and other multimodal presentations that only a computer can provide (Pennington, 1996), (see Figure 8). he computer can individualize the teaching of pronunciation, allowing the users themselves to choose the phonological aspects they want to work on or suggesting which aspects the users need to improve according to the pronunciation analysis made by the computer.he users may also receive input from the computer whenever and wherever they need.he computer can also give the users the opportunity to take responsibility for their own learning, which, according to Benson (2007), is a key element in the development of learner autonomy.By developing autonomy, the learners can compensate classroom learning limitations and boost pronunciation learning without the presence of a teacher.
However, Chapelle and Jamieson (2008) advised that the users need guidance for choosing what to learn and how to learn and it is the teacher's job to provide them with the necessary guidance for selecting and evaluating instructional materials for CALL based on the teacher's expertise and experience in language teaching and learning (see Figure 9).Even with so many positive attributes that the computer has to help in the development of pronunciation, CAPT still has its limitations: many educational sotware programs are attractive, but do not meet educational requirements (Neri, Cucchiarini & Strik, 2002), while others work on pronunciation in a decontextualized way presenting only mechanical exercises with no focus on meaning (Pennington, 1996), (see Figure 10).here are sotware programs that present only a single pronunciation model, either a female or a male one and a single regional phonological variant, limiting the input received by the users as well as their perception and production.As asserted by Cardoso, Smith and Garcia Fuentes (2015) and Soler-Urzua (2012), diferent kinds of quality input can enhance learners' perception and production.hus, the learners need to be exposed to diferent models in order to be able to recognize diferent patterns of pronunciation (see Figure 11).he pronunciation feedback provided to the users should be easily understandable.he users should be able to understand through the feedback received by the computer which phonological aspects they need to work on.Although feedback can be provided in several ways by the computer, the most common ones are those that utilize visual displays such as spectrograms, waveforms and pitch tracings.Many of these visual displays require some sort of practice to help teachers build a kind of knowledge that they, in general, do not have (Levis, 2007), (see Figure 12).However, in order for the computer to analyze the users' pronunciation and provide feedback, an Automatic Speech Recognition (ASR) Mechanism is necessary.Derwing, Munro and Carbonaro (2000) tested diferent versions of ASRs and found out that these mechanisms could analyze quite accurately (95%) the pronunciation of English native speakers, but when analyzing the pronunciation of non-native speakers, the level of accuracy dropped to approximately 70%.
According to these researchers, the problem with the inaccuracy of these mechanisms was that they were not designed for non-native speakers.
A recent study by Liakin, Cardoso and Liakina (2015) tested the use of ASR for SL pronunciation instruction in a mobile-assisted learning environment.Forty-two learners were divided into three groups: 1) a group using ASR to do pronunciation activities; 2) a group not using ASR; and 3) a control group.he results of this study showed that only the learners in the group using ASR improved their pronunciation signiicantly.herefore, it is relevant to evaluate ASR mechanisms in sotware programs.he following question checks the ability of ASR devices to provide feedback (Figure 13).CAPT can also take place in language labs, allowing the teacher to monitor multiple users on diferent computers at the same time.he teacher can guide users on how to use the computer, revise their pronunciation analysis made by the computer, compare the analysis of one user with the others as well as provide feedback.
herefore, EFL/ESL pronunciation teaching sotware programs should explore technological characteristics efective for computer-mediated interactive pronunciation learning as just discussed.

Method
In order to design an instrument with questions that encompass characteristics to be observed during the evaluation of sotware programs for teaching pronunciation to EFL/ESL learners, criteria and the evaluation models reported in Table 1 were reviewed.
he following criteria were used in the design of the instrument (see Appendix): 1) the presence of pedagogical characteristics consistent with the Communicative Approach (Celce-Murcia et al, 2010) and 2) the presence of technological characteristics efective for computer-mediated interactive pronunciation learning.
Based on these criteria, 72 questions were elaborated.It is noteworthy that the instrument was divided into two parts: 1) the Descriptive Analysis, which describes the technical and pedagogical characteristics of the EFL/ESL pronunciation teaching sotware program provided by the developer/ distributor and 2) the Critical Analysis, which analyzes the extent to which the sotware program follows the principles of the Communicative Approach (Celce-Murcia et al, 2010).
To assess the extent to which the sotware program meets the characteristics described in the questions, a rating scale of 0-4 was used, where 0 indicates absence of the characteristic and 4 indicates a completely satisfying exploration of the characteristic (See Figures 1 to 13 for examples of questions in the instrument with this rating scale).he greater the overall rating of the sotware program, the more the potential to improve the pronunciation of the English language learner.
In order to test the reliability and the internal consistency of the instrument and make sure that it has some validity, 46 participants used the instrument to analyze an online version of the sotware program Pronunciation Power 2 3 , because it didn't require any installation procedure.An online version of the instrument was also used, because it allowed the researchers to get the responses provided by the informants by email. he participants were selected based on the following criteria: 1) teachers of English teaching in language schools, secondary schools or colleges, and 2) teachers graduated in English and/or with a specialization, Master's Degree and/or Ph.D. in English, linguistics and/or applied linguistics.779 teachers were invited to be part of the study by email, but only 58 sent their evaluation responses.hese 58 teachers were irst divided into two groups according to their context of teaching English -foreign language and second language teaching.hen they were grouped according to their level of expertise in English phonetics and phonology -teachers and experts.hese procedures resulted in three groups: Group 1 -Brazilian EFL teachers; Group 2 -Brazilian experts in English phonetics and phonology; and Group 3 -Foreign experts in English phonetics and phonology.
However, from the 58 teachers, only the responses sent by 46 were considered.Twelve teachers had to be discarded because 10 of them assigned 3 and/or 4 points to all questions of the instrument and two did not it into any of the three groups.hus, in this study, the responses sent by 46 teachers/participants were analyzed and the teachers were divided as follows: 27 participants in Group 1; 11 participants in Group 2; and 8 participants in Group 3. he participants did not receive any training nor had a time limit to complete the instrument.
he instrument was submitted for the analysis of an expert in English phonetics and phonology for face and content validity examination.he data collected through the responses of the participants were statistically analyzed to determine the degree of instrument validity for its purpose.hus, tests of reliability and internal consistency of the instrument were performed: Intra-Class Correlation Coeicient (ICC); One-Way Analysis of Variance (ANOVA); Cronbach's alpha coeicient; and Factor Analysis.Some procedures of descriptive statistics were used to describe and summarize the results collected from responses sent by the 46 participants of this study when analyzing the sotware program Pronunciation Power 2.

Results
he results of this study are presented in three sections.In section one, the results for face and content validity of the instrument are described; in the second section, the results for the reliability of the instrument are presented; and in section three, the results for the internal consistency of the instrument are described.

Results for Face and Content Validity of the Instrument
he researchers also had the opportunity to submit the instrument for analysis by an expert 4 in English phonetics and phonology.he criteria that guided the design of the instrument were given to this expert.Ater having analyzed and used the instrument himself, the expert assured that the instrument "on its face" seemed to be appropriate to analyze if a sotware program met the criteria.

Results for the Reliability of the Instrument
he irst statistical procedure used to analyze the level of agreement among the participants' responses was the ICC. he ICC found for the three groups of participants was .983and the level of conidence at 95% was .976 to .988.hese results indicated a high degree of correlation among the ratings of the 46 participants despite the fact that they teach English in diferent contexts and have diferent levels of expertise in English phonetics and phonology (see Table 2).he second statistical procedure performed was the One-Way ANOVA, 5 which was used to compare the responses given by the three groups of participants: Group 1 (Brazilian EFL teachers); Group 2 (Brazilian experts in English phonetics and phonology); and Group 3 (Foreign experts in English phonetics and phonology).hese three groups form the independent variable while the total rates for the 67 questions 6 of the instrument are the dependent variable.
Despite the diferences in the English teaching context and the level of expertise in English phonetics and phonology of the three groups, the results of the one-way ANOVA indicated no signiicant diference (p > .05) in the evaluation of 62 of the 67 questions for the three groups.his means that there was an agreement of 95%, p > .05, in the evaluation of 62 questions among the three groups of participants.
he ive questions with p less than .05,that is, with the lowest level of agreement, are Questions 1.1, 1.2, 1.6, 2.14 and 2.26.Concerning Questions 1.1 and 1.2 (p = .008and .004,respectively), some participants reported that the characteristics described in those questions were not found in the sotware program analyzed.herefore, they had to assign zero to those questions.he objectives and the proiciency level proposed were only described on the developer's website and could only be found by navigating to the sotware program developer's website before logging in to the program.As these pieces of information were not available in the program itself, some participants may have not navigated to the developer's website.his may have led to a signiicant diference in the evaluation of the characteristics described in those two questions.As to the other three questions -1.6, 2.14 and 2.26 -the p found, respectively .039,.048,.025,although less than .05,does not point to a large discrepancy in evaluating the characterisitics described in those three questions.
In the One-Way ANOVA, the Post Hoc Fisher's test comparing the least signiicant diference (LSD) was also performed to locate diferences in the p value <.05. he results of the post-hoc test indicated that the signiicant diference for these ive questions always occurred between the group of Brazilian EFL teachers and the group of Brazilian experts in English phonetics and phonology or the group of teachers and the group of foreign experts in English phonetics and phonology, but never between the two groups of experts (see Table 3).Perhaps Questions 1.6, 2.14 and 2.26 inquire about speciic characteristics of the program, requiring a greater knowledge of English phonetics and phonology to be analyzed.his may explain the signiicant diference found between the evaluation of the group of teachers and the group of Brazilian and foreign experts in those three questions.

Results for the Internal Consistency of the Instrument
he irst statistical procedure used to analyze the degree of internal consistency of the instrument was the Cronbach's alpha coeicient.he Cronbach's alpha coeicient obtained for the 67 questions in the instrument was equal to .918,indicating a high degree of internal consistency of the instrument.
he second statistical procedure performed was a factor analysis.hrough factor analysis, we tried to determine the existing components (factors) among the 67 questions (variables) of the instrument and the correlation of each of the questions in those components (factors).A principal components analysis extraction and promax rotation with Kaiser normalization method was used and factors with eigenvalues > 1 were retained in this analysis to reduce the amount of observed variables into a smaller number of factors.
his extraction method resulted in 18 components with initial eigenvalues > 1 with the irst ive components explaining 50.48% of the total variance of the questions of the instrument. he irst component -Factor 1with initial eigenvalue of 15.95 explained 23.81% of the total variance; the second component -Factor 2 -with initial eigenvalue of 5.63 explained 8.41%; the third component -Factor 3 -with initial eigenvalue of 4.72 explained 7.05% of the total variance; the fourth component -Factor 4 -with initial eigenvalue of 3.89 explained 5.81% of the total variance; and the ith component -Factor 5 -with initial eigenvalue of 3.60 explained 5.37% of the total variance.he other 13 components explained less than 5% of the total variance of the questions of the instrument and therefore were not retained in this analysis.
Ater the retention of ive components in the irst stage of factor analysis, we calculated the factor loading of each item of the instrument regarding these factors to verify which questions would be grouped into each of them.
Communicative Approach (Celce-Murcia et al, 2010).By analyzing each of the 72 questions of the instrument, it is possible to assess the extent to which the program presents pedagogical characteristics consistent with the principles of the Communicative Approach (Celce-Murcia et al, 2010) and technological characteristics efective for computer-mediated interactive pronunciation learning.hus, this instrument can be widely used by teachers and researchers when analyzing and/or developing EFL/ESL pronunciation teaching sotware programs.
Some limitations of this study need to be mentioned.he participants of the study, who were Brazilian EFL teachers, Brazilian experts and foreign experts in English phonetics and phonology represent only a portion of the entire population of EFL/ESL teachers, and experts.EFL/ESL teachers from other countries were not represented in this sample.Further studies may include a more diverse group of participants.
In this study, only one EFL/ESL pronunciation teaching sotware program was used to evaluate the reliability and validity of the instrument.Future studies may use more programs to evaluate the reliability and validity of the instrument.Besides, this instrument was designed to evaluate only EFL/ESL pronunciation teaching sotware programs.Since there are also websites and cell phone and tablet applications designed for teaching English pronunciation to FL/SL learners, future studies may design instruments to evaluate websites and applications as well.
he data collected through the responses of 46 participants were statistically analyzed in this study.In order to evaluate the reliability and validity of the instrument, tests of reliability and internal consistency of the instrument were performed: ICC; One-Way ANOVA; Cronbach's alpha coeicient; and Factor Analysis.he same statistical tests as well as other statistical tests may be employed with data collected from a larger sample in additional studies.
Finally, additional studies may evaluate the extent to which a group of learners improve their pronunciation when using an EFL/ESL pronunciation teaching sotware program that follows the principles of Communicative Approach (Celce-Murcia et al, 2010), then evaluate the same program using the instrument and contrast both results.

Figure 1 2
Figure 1 2 .Questions for analyzing activities for sound perception and production.

Figure 2 .
Figure 2. Questions in the instrument for analyzing the presentation of segmental and suprasegmental aspects.

Figure 3 .
Figure 3. Questions in the instrument for analyzing feedback for activities.

Figure 4 .
Figure 4. Questions in the instrument for analyzing the use of phonetic symbols and transcriptions.

Figure 5 .
Figure 5. Questions in the instrument for analyzing the contrast of the user's mother tongue sounds to the English sounds.

Figure 6 .
Figure 6.Questions in the instrument for analyzing the incorporation of highly variable input.

Figure 7 .
Figure 7. Questions in the instrument for analyzing the redo of the activities by the user.

Figure 8 .
Figure 8. Questions in the instrument for analyzing the use of visual aids.

Figure 9 .
Figure 9. Questions in the instrument for analyzing the possibility of choosing the content of the activities.

Figure 10 .
Figure 10.Questions in the instrument for analyzing the use of contextualized activities.

Figure 11 .
Figure 11.Questions in the instrument for analyzing the possibility of choosing a pronunciation model.

Figure 12 .
Figure 12.Questions in the instrument for analyzing the use of visual displays.

Figure 13 .
Figure 13.Questions in the instrument for analyzing the use of ASR.

Table 1
Studies on Criteria and Instruments for Analyzing Educational Sotware Programs

Table 2
ICC of the Instrument

Table 3
Questions with p < .05(ANOVA: Post-Hoc Fisher'sTest) Notes 1.Although in the literature the concepts of Second Language (SL) and Foreign Language (FL) are sometimes distinguished, in this paper they are used interchangeably.2. he igures in this paper will show only some examples of the questions in the instrument.For a total perception of all questions, the whole Evaluation Instrument should be analyzed in the Appendix.3.According to the publisher, Pronunciation Power 2 is designed for intermediate to advanced English learners of all ages.It presents 52 English sounds which are considered necessary to speak English clearly.It also brings hundreds of hours of instruction and practice (See References for bibliographical information on this sotware).4. he expert was Dr. John Michael Levis.Dr. Levis studies SL pronunciation and speech intelligibility, with a focus on how SL pronunciation research afects the teaching of pronunciation.He is the founder of the annual Pronunciation in Second Language Learning and Teaching conference and is the founding editor of the Journal of Second Language Pronunciation.He is co-editor for the Phonetics & Phonology section of the Encyclopedia for Applied Linguistics, and two books, Social Dynamics in Second Language Accent and the Handbook of English Pronunciation.Dr. Levis has also written articles about pronunciation teaching in a variety of journals such as Applied Linguistics, TESOL Quarterly, TESOL Journal, ELT Journal. 5. A One-Way ANOVA compares existing variations within groups, of participants with existing variations between groups checking if there is a signiicant diference between them or not.To determine the location of the diference if there is one, it is necessary to perform a post-hoc analysis besides the One-Way ANOVA.6.As the participants analyzed an online version of the sotware program Pronunciation Power 2, Questions 4.48 to 4.51 were ignored, totalling 67 questions analyzed instead of 72.