CHAPTER ONE Introduction - thuir.thu.edu.twthuir.thu.edu.tw/retrieve/3266/096thu00094001-002.pdfas...
Transcript of CHAPTER ONE Introduction - thuir.thu.edu.twthuir.thu.edu.tw/retrieve/3266/096thu00094001-002.pdfas...
1
CHAPTER ONE
Introduction
1.1 Background
The Communicative Language Teaching (CLT) approach has been widely discussed
as well as applied in numerous ways in language classrooms around the world since the
1970s. The CLT approach is a commonly recognized approach in the field of English as a
second language/foreign language (ESL/EFL) (Brown, 2001). The impact of the CLT
principles is evident in many aspects— not only in teaching/learning activities, but also in
material development. Indeed, teaching and learning materials are continuously being
reexamined to determine whether they conform to the CLT characteristics.
Meanwhile, recent advances in computer technology have dramatically changed the
learning and teaching environment in English education over the past four decades. The
relationship between technology and language teaching/learning has interested a great
number of second and/or foreign language educators. With the rapid development of
multimedia and the Internet, terms such as web-based learning, educational multimedia,
hypermedia, and courseware have been coined to refer to learning activities that involve
specific computer tools and facilities. In addition, the emergence and development of
Computer-Assisted Language Learning (CALL) has drawn a significant amount of attention
in the field of second/foreign language learning for years. Numerous CALL-related studies
have been conducted, and web-based language learning materials are increasingly being
developed. Just like other types of language teaching and learning materials, CALL materials
are expected to be learner-centered, interactive, and authentic (Brown, 2001; Levin & Bruce,
2001; Levy, 1997). According to Bradin (1999), it is important for language teachers to look
at CALL software as critically as they look at textbooks.
2
Software companies continue to produce new packages, including CD-ROMs and
online courseware, while a large number of instructional software packages are released
every year. In this web-based learning era, language teachers frequently adopt and adapt
online courseware as one of the teaching tools, for use either in or outside the classroom; one
of the advantages of online courseware is that it is not constrained by space and time.
Learners can access online courseware almost anywhere at any time they choose— outside the
classroom, at home, or at a self- learning center on campus. Learners can thus develop their
autonomy and become independent learners, one of the ultimate goals in education (McDevitt,
1997); an effective user-centered CALL tool should be able to guide learners toward this
autonomy. However, the wide selection of online courseware available creates challenges for
language educators in choosing appropriate tools to suit different instructional contexts and
various kinds of learners. Evaluators’ evaluation criteria are crucial. Therefore, appropriate
evaluation criteria and evaluation tools have become essential.
When selecting and evaluating suitable online CALL courseware, two issues play an
important role: policy considerations and evaluation criteria. Policy and decision makers,
such as stakeholders, make selections based on a series of relevant concerns. In Taiwan,
English teachers, administrators, and principals have been involved in the decision-making
process regarding English teaching materials; those who are not language teachers tend to
focus on costs as well as the designer/publisher’s reputation. In addition, government policy
affects the decision-making; for example, the Ministry of Education (MOE) may deve lop and
publish its own textbooks for schools or develop curriculum guidelines that may affect
decisions about instructional materials. Consequently, teachers often complain that the
chosen textbooks’ and/or courseware’s content do not meet teachers’ requirements. Moreover,
learners’ views and needs are frequently ignored as they are not practically involved in the
process of material selection and evaluation.
3
Evaluation criteria also play a significant role; what kinds of criteria are to be
included are quite relevant. Since CALL adopts a multidisciplinary perspective (Levy, 1997),
incorporating applied linguistics, second language acquisition (SLA), education, information
technology (IT), and computer science, it requires both pedagogical and technical criteria,
such as whether the courseware offers a clear learning goal, is based on a certain learning
theory, or helps learners achieve learning objectives. Furthermore, the degree of technical
difficulty and user- friendliness must be considered.
1.2 Statement of the Problem
In recent years, the impact of the CLT approach has encouraged language teachers to
pay more attention to learner-centeredness in language teaching and learning. Meanwhile, the
user-centeredness of CALL tools has drawn increased attention. For instance, Squires and
Preece (1996) emphasize the importance of considering the way students learn and the
usability of the courseware design in creating natural and intuitive interaction between
learners and the courseware. Heller (2005) also claims that a CALL tool should “work
flawlessly with a user- friendly design and self-explanatory user manual for instructors and
users” (p. 120). Although the literature focuses on the need for user-centered tools in CALL
(Hémard, 2003), many CALL tools do not function as they should. According to Felix (2003),
many CALL tool designs do not follow established learning theories or principles; instead,
designers focus more on technical features rather than on the evaluation of these CALL tools
with the help of users.
Consequently, adequate evaluation of CALL materials is becoming increasingly
crucial. Only can user- friendly and learner-centered courseware attract learners to use it
outside the classroom, thereby accessing its advantages. Thus, language educators and/or
institution administrators should have suitable guidelines or criteria to follow when
evaluating and selecting web-based courseware; these guidelines or criteria should lead the
4
evaluation process in the direction of learner-centeredness/user-centeredness. Since the
learner is the center of the language learning process as well as the main user of CALL
materials, evaluating programs and materials from the learner’s standpoint is necessary. In
other words, evaluation criteria for CALL courseware should be aligned with the perspective
of learner-centeredness.
1.3 Purpose of the Study
The purpose of this study is two-fold. First, the study aims to provide language
teachers with evaluation criteria for online CALL courseware based on the concept of
learner-centeredness. Second, the study expects to increase the awareness of learner-
centeredness among English teachers by using a developed learner-centered evaluation
checklist to examine the commercial courseware My English Tutor (MyET) (L Labs, 2006).
1.4 Significance of the Study
Despite the fact that many CALL-related evaluations have been conducted, limited
research exists discussing issues relevant to online CALL courseware evaluation in Taiwan—
particularly in regards to learner-centeredness. Indeed, among the few studies conducted in
Taiwan, most have focused on evaluating CD-ROMs (e.g., Chang, 1992; Lin, 2002; Su, 2003;
Yeh, 2002) or websites (e.g., Tsai, 2002) and have focused on specific issues such as
pronunciation training (e.g., Hong, 2003).
Unlike previous studies, the current study focuses on online courseware and aims to
propose a checklist based on the premises of learner-centeredness for online courseware
evaluation. As such, it seeks to make several contributions to the field of English language
teaching, including: 1) connecting the learner-centered principles with CALL research, 2)
raising awareness among language instructors of learners’ needs and learner-centered design,
and 3) providing a set of specific learner-centered evaluation criteria for online CALL
courseware that are more specific than the current general evaluation criteria.
5
1.5 Definition of Terms
The study operationally defines the following terms:
CALL Tools
This term refers to any tools used on a computer basis and aimed at enhancing
language learning, such as CD-ROMs, software, online courseware, and websites.
Courseware
Longman Dictionary of Language Teaching and Applied Linguistics (Richards, Platt,
& Platt, 1998) defines courseware as “computer programs used in computer assisted
learning” (p. 113). According to the online glossary of the Learning Technology Jargon
Buster (2006), courseware refers to “computer software and associated materials designed for
educational or training purposes.” Similarly, in the E-learning Glossary (2006), courseware
means “educational software that delivers course material and instruction via computer.” In
short, courseware is a word coined from course and software, referring to software with
teaching and learning purposes and functions. In this research, MyET is regarded as CALL
courseware as it is designed for use as an English instructional program.
Online Courseware
Online courseware is courseware used in a web-based environment— i.e., on the
Internet.
Learner/user
In this study, learner represents those who learn English as a second/foreign language
while user refers to those who use CALL courseware to study English. In the CALL
environment, the user in a broad sense may include the learner, teacher, and courseware
administrator. However, the user in this study refers only to the learner. Thus, the learner and
the user are interchangeable in this study.
6
Teacher-evaluators
This term refers to the three teachers involved in this study who conducted the
evaluations (see section 3.2.2).
Student-evaluators
This term refers to the 22 college students involved in this study who conducted the
evaluations (see section 3.2.2).
1.6 Organization of the Thesis
This thesis consists of five chapters. Chapter One states the background and purpose
of the study as well as its significance and operational definitions of terms. Chapter Two
provides an overview of the literature on learner-centeredness and CALL, emphasizing the
salience of learner-centeredness in a CALL tool. It also discusses certain issues relevant to
English teaching/learning material evaluation, including the classification of existing
evaluation approaches and evaluation criteria for ESL/EFL materials, by reviewing previous
studies related to English textbook evaluation and CALL courseware evaluation. Finally, it
presents the research questions based on the review of the literature. Chapter Three describes
the research design, including the rationale, participants, data collection and analysis
procedures, and validation in three phases. Chapter Four presents the learner-centered
evaluation checklist developed in this study as well as the results of applying the checklist to
the evaluand, MyET. In addition, it examines the effectiveness of the checklist as well as
answers and discusses the research questions. Finally, Chapter Five draws conclusions and
offers implications to language teachers, language institution stakeholders, CALL tool
designers, and material developers; it further identifies the limitations of the study and
proposes suggestions for further research in this area.
7
CHAPTER TWO
Review of the Literature
This chapter reviews the literature on learner-centeredness in education and its
relationship with CALL along with evaluation issues impacting English teaching/learning
materials. Research questions are proposed at the end of the chapter based on the review of
the literature.
2.1 Learner-centeredness in Education
Learner-centered concepts are instructional design and teaching practices based on
cognition and learning (Mynard, 2003). The concept of learner-centeredness formally
emerged in 1990 from the American Psychological Association’s (APA) Presidential Task
Force on Psychology in Education (Lambert & McCombs, 1998). One of the task force’s
purposes was to develop general principles to establish a framework for school redesign and
reform. Twelve fundamental principles about learners and learning were proposed and later
revised in 1997. Now the document includes 14 principles, which are categorized into four
domains essential to learning─metacognitive and cognitive factors, affective and
motivational factors, developmental and social factors, and individual differences factors
(McCombs, 2000) as shown in Table 2.1. A learner-centered model embracing five premises
was shaped (McCombs & Whisler, 1997)─summarized in Table 2.2.
8
Table 2.1
The learner-centered psychological principles
Cognitive and Metacognitive Factors
Principle 1: Nature of the learning process.
The learning of complex subject matter is most effective when it is an intentional process of
constructing meaning from information and experience.
Principle 2: Goals of the learning process.
The successful learner, over time and with support and instructional guidance, can create meaningful,
coherent representations of knowledge.
Principle 3: Construction of knowledge.
The successful learner can link new information with existing knowledge in meaningful ways.
Principle 4: Strategic thinking.
The successful learner can create and use a repertoire of thinking and reasoning strategies to achieve
complex learning goals.
Principle 5: Thinking about thinking.
Higher order strategies for selecting and monitoring mental operations facilitate creative and critical
thinking.
Principle 6: Context of learning.
Learning is influenced by environmental factors, including culture, technology, and instructional
practices.
Motivational and Affective Factors
Principle 7: Motivational and emotional influences on learning.
What and how much is learned is influenced by the learner’s motivation. Motivation to learn, in turn,
is influenced by the individual’s emotional states, beliefs, interests and goals, and habits of thinking.
Principle 8: Intrinsic motivation to learn.
The learner’s creativity, higher order thinking, and natural curiosity all contribute to motivation to
learn. Intrinsic motivation is stimulated by tasks of optimal novelty and difficulty, relevant to personal
interests, and providing for personal choice and control.
Principle 9: Effects of motivation on effort.
Acquisition of complex knowledge and skills requires extended learner effort and guided practice.
Without learners’ motivation to learn, the willingness to exert this effort is unlikely without coercion.
9
Table 2.1 (continued)
Developmental and Social Factors
Principle 10: Developmental influence on learning.
As individuals develop, they encounter different opportunities and experience different constraints for
learning. Learning is most effective when differential development within and across physical,
intellectual, emotional, and social domains is taken into account.
Principle 11: Social influences on learning.
Learning is influenced by social interactions, interpersonal relations, and communication with others.
Individual Differences Factors
Principle 12: Individual differences in learning.
Learners have different strategies, approaches, and capabilities for learning that are a function of prior
experience and heredity.
Principle 13: Learning and diversity.
Learning is most effective when differences in learners’ linguistic, cultural, and social backgrounds
are taken into account.
Principle 14: Standards and assessment.
Setting appropriately high and challenging standards and assessing the learner and learning progress
─including diagnostic, process, and outcome assessment─are integral parts of the learning process.
Source: McCombs, B. L. (2000). Assessing the role of educational technology in the teaching and
learning process: A learner-centered perspective. Paper presented on Secretary’s Conference on
Educational Technology 2000.
http://www.ed.gov/print/rschstat/eval/tech/techcon00/mccombs_paper.html
10
Table 2.2
Premises of the APA’s learner-centered model
Premise One
Learners are distinct and unique, and this distinctiveness and uniqueness must be taken into account.
Premise Two
Learners’ uniqueness includes their emotional states, learning rates, learning styles, abilities, talents, and
needs.
Premise Three
Learning is a constructive process, so, what is being learned should be relevant and meaningful to the
learner and be based on the learner’s prior knowledge and experience.
Premise Four
Learning occurs best in a positive environment in which the learner has positive interpersonal interactions
and feels appreciated and validated.
Premise Five
Learning is a natural process; learners are naturally curious and interested in learning about their world.
Source: Lambert, N. & McCombs, B. L. (Eds.). (1998). How students learn: Reforming schools
through learner-centered education. Washington, DC: American Psychological Association.
Educators define learner-centeredness as “the perspective that couples a focus on
individual learners [… ] with a focus on learning. This dual focus then informs and drives
educational decision making” (McCombs & Whisler, 1997, p. 9). McCombs and Whisler’s
definition is based on the understanding of the Learner-centered Psychological Principles as a
demonstration of the knowledge base on the learner and learning” (McCombs, 2000).
Bransford et al. (1999) propose that learner-centeredness refers to the environments focus ing
on knowledge, skills, attitudes, and beliefs that learners bring to an educational setting.
Besides, learner-centered environments include teachers who are aware of the knowledge,
beliefs, understandings, and cultural practices that learners construct and bring to the learning
environment. Sparks and Hirsh (1997) used the term learner-centeredness to refer to learning
new beliefs and new visions of practice corresponding to the needs of students and teachers
as learners. In short, learners must be involved in the educational decision-making process;
11
individuals’ differences should be respected and accounted for in the learning process.
In the language teaching field, Nunan (1997) explains that, in learner-centered
environments learners are actively involved in the learning process while the focus is on
the learner. In other words, language teaching is now seen as language learning, placing
the learner at the center of attention in language education.
In learner-centered instruction, the spotlight has shifted from the teacher to the
learner, concentrating on what learners need and how they learn rather than on what the
teacher wants to teach. This transformation does not devalue the teacher’s professional
role, but instead attempts to encourage learners to move toward a primary goal of
education: becoming an autonomous learner. Learner-centered teachers recognize the
importance of building on the knowledge that students have already accumulated during
the learning process. In this regard, “teachers no longer function as exclusive content
expert or authoritarian classroom managers” (Weimer, 2002, p. 14).
The Learner-centered Model (see Table 2.1 and Table 2.2) is claimed to be
research-validated as it integrates research and theory from psychology and education
(McCombs, 2000). Research (e.g., Bain, 2004; McCombs & Whisler, 1997; McCombs,
2000; Palmer, 1998) has identified the positive effects of learner-centeredness on
education. Bain’s (2004) research, for instance, focuses on college teachers who adopted
learner-centered instruc tion and concludes that these faculty have a strong faith in
student ability, creating a safe learning environment allowing students to try. In addition
to identifying the benefits of learner-centered practice, McCombs and Whisler (1997)
point out that the abundant evidence has proved that “motivation, learning, and
achievement are enhanced where learner-centered principles and practice are in place”
(p. 62). For example, in McCombs and Stiller’s study (in McCombs & Whisler, 1997),
teachers report that with a learner-centered approach there is less student disruption and
12
fewer discipline problems. McCombs (2000) further emphasizes the positive feedback
of learner-centeredness between the learner and the teacher as well as the importance of
an encouraging climate of learning both in and outside the classroom. A report provided
by the Committee on Academic Programs and Teaching (CAPT) Learner-centered Task
Force of University of Southern California (2006) also indicates the positive results of
the practice of learner-centered teaching embedded in units across the school.
2.2 Learner-centeredness and CALL
The emergence and transformation of CALL is intertwined with the trends in
educational theory and language acquisition theory at different times. Warschauer and
Kern (2000) use three metaphors to describe the three prominent stages of development
in the CALL field─a tutorial metaphor (computer-as-tutor), a construction metaphor
(computer-as-pupil), and a toolbox metaphor (computer-as-tool) :
From the 1950s to 1960s, behaviorism and the structural perspective dominated
education and language teaching, greatly impacting the development of CALL. Many
software developers built numerous drills and practice exercises consisting of grammar
and vocabulary tutorials into their programs to emphasize repetition in learning (Levy,
1997; Warschauer & Kern, 2000). The earliest CALL programs “strictly followed the
computer-as-tutor model” (Warschauer & Kern, 2000, p. 8).
During the 1970s, influenced by cognitive/constructivist views of learning, the
agency in CALL programs was shifted from the program to the learner. Learning
software designers realized that the computer was something to be controlled by the
learner instead of being a controller over the learner. More resources and functions
along with interactive tasks (interacting with computers) were incorporated into CALL
programs. This period defined the computer-as-pupil model generation.
13
Currently, the sociocognitive approach is influencing the language teaching
profession, and communicative competence has become a buzzword in this era
(Warschauer & Kern, 2000). Language educators have started to focus more on
providing authentic contexts and materials to enhance learners’ communicative ability.
Learners no longer merely interact with computers; more importantly, they interact with
other humans via computers. Under this model of computer-as-tool, computers offer
learners access to information and to other people. For example, many learning websites
and web-based learning environments provide a computer-mediated communication
(CMC) function, using computers to assist authentic communication between two or
more people. Technological advances like this have changed language learners’
activities and made autonomy and independence in learning become something
achievable (Benson & Voller, 1997).
A great number of studies demonstrate the positive and effective impact that
technology brings to learning (McCombs, 2000). Indeed, previous research has
indicated that computer-related technology can facilitate language learning by
supporting learner-centered practice in two primary aspects: 1) researching and creating
an authentic product for assessment and 2) providing an unrestricted number of sources
on most topics (Encyclopedia of Educational Technology, 2005). Recently, researchers
(e.g., Felix, 1999; Norlan, 2000) have focused on the learner/user rather than the
technology. According to Felix (1999), the focus has shifted to what the learner can do
with the technology and not what the technology can do. Such fundamental shifts have
led to increased attention to learner-centered design, user-centeredness, and/or user-
friendliness of CALL tools.
Since computer-related technology is meant to serve as a learning tool to
promote effective learning, CALL tool developers should involve the personal domain
14
of a learner and address learners’ needs in the design process. As McCombs (2000)
states, the learner-centered framework serves as a reminder that the human issues can
never be left out of the technology-supported networked learning environments. Easy-
to-use, relevance to the user’s life, meaningful learning activities, clear instructions
combined with vivid and interactive interfaces are necessary features of high-quality
courseware, as such courseware— with its learner-centered design— focuses on the tasks
and goals of learners, motivates them, and encourages them to take more responsibility
for their own learning when using the CALL tool. Consequently, when technology is
regarded as a learning tool instead of a teaching approach, it has the capacity to greatly
facilitate learning by providing individualized opportunities for students (Nolan, 2000).
In this regard, CALL tools can become more usable for learners. Usability has
been the focus of CALL courseware developers, researchers, and language educators. A
number of researchers (e.g., Felix, 2003; Heller, 2005; Hémard, 2003; Nielsen, 2000;
Rubin, 1988) have pointed out the importance of high-quality usability of CALL tools.
The usability of a CALL tool involves the features related to the user from the user’s
perspective; as such, users play a crucial role in the effectiveness of CALL courseware.
Only CALL courseware that places the learner at the center of the learning process can
effectively enhance learning.
Despite the increased attention on usability, user-centeredness, and user-
friendliness in the CALL area, little existing literature discusses the integration of
learner-centeredness into CALL tool evaluation. Moreover, few studies examine CALL
courseware from the perspective of learner-centeredness. The following section reviews
literature relevant to English material evaluation in general as well as that specific to
CALL material evaluation.
15
2.3 ESL/EFL Material Evaluation
Tomlinson (1998) defines material as “anything which is used by teachers and
learners to facilitate the learning of a language” and “anything which is deliberately
used to increase the learners’ knowledge and or experience of the language” (p. 2). This
section provides an overview of the literature related to ESL/EFL material evaluation. It
begins by discussing evaluation approaches─and thus providing some important
evaluation terms and concepts─before investigating evaluation criteria in general and
those for online courseware
in particular.
2.3.1 Evaluation Approaches
Evaluation approaches can be classified according to evaluation instruments,
objectives, timing, and the evaluator’s perspective. First, general evaluation instruments
include checklists, interviews, and evaluation frameworks and models based on certain
specific guidelines or principles. A checklist is one of the most common types of
evaluation; it provides a list of items that include important points to be examined.
Checklists vary in content, length, style, purpose, and target, but all have been
developed in an attempt to help language teachers, students, and/or even administrators
to choose appropriate materials for instructional use. Traditional checklists have
received criticisms (Oliver, 2000); for example, they assume all criteria are of equal
importance, allowing no gradation of the degree to which a specific evaluation object
meets the different criteria, and are not able to consider the contextual conditions of
learning (McDougall & Squires, 1995; Tergan & Schenkel, 2003). However, they serve
as one of the most frequently used methods for evaluations because of their ease of use.
Interviews cover basic information about the material and the reviewer’s
subjective description by offering a direct manner to obtain evaluators’ responses
16
(Reeder, Heift, Roche, Tabyanian, Schlickau, & Golz, 2004). Some researchers who
have carried out evaluations through interviews have found that many insights can be
gained in utilizing this approach (Le & Le, 1997). However, although interviews enrich
research data, one of the drawbacks is that their dynamics and flexibility may result in
inconsistency and cause the difficulty in analyzing data (Patton, 1980).
In the recent CALL field, several researchers (e.g., Chapelle, 2001; Hubbard,
1988; Squires & Preece, 1996) have proposed evaluation frameworks and/or models for
CALL tools instead of using certain evaluation criteria.
Second, the objective of evaluations can be addressed via formative and
summative evaluations. According to Trochim (2006), formative evaluations aim to
strengthen or improve the object being evaluated, while summative evaluations focus on
the effects or outcomes of the object. Lau (2000) contends that formative evaluations
investigate the feedback on any factors influencing teaching and learning processes and
then design and develop an instructional technology efficiently, whereas summative
evaluation aims to justify the implementation of the instructional technology.
Another way to distinguish evaluation approaches concerns timing— namely,
through predictive and retrospective evaluations. The former is conducted before
materials are adopted, whereas the latter functions as a reflective way to examine
learning materials, which is more empirical (Ellis, 1998). Instructors frequently use
predictive evaluation before selecting courseware. Squires and McDougall (1996) define
it as “the assessment of the quality and potential of a software application before it is
used with students” (p. 147).
Finally, evaluation approaches can be categorized in terms of the evaluator’s
perspective, classified as either judgmental or empirical (Chapelle, 2001). In a
judgmental analysis, the analysis is usually done by material developers and teachers;
17
CALL software is often the object of such evaluation. Chapelle (2001) used an example
question to explain this type of subjective analysis: “Does the software provide learners
the opportunity for interactional modifications to negotiate meaning?” (p. 53).
Responses to this evaluation question vary depending on the evaluator. On the other
hand, an empirical analysis is normally conducted by collecting data from learners, such
as by observing learners’ performance when they are using the courseware. Empirical
evaluation methods are extremely diverse, including case studies, pilot projects, project
monitoring, field studies, replicated experiments, synthetic experiments, product
simulations, and process simulations. This type of evaluation is usually time consuming.
It is important to note that judgmental and empirical evaluations are complementary
rather than mutually exclusive (Chapelle, 2001).
2.3.2 Evaluation Criteria
With a great number of language-learning materials on the market, the
responsibility of identifying which materials to use and how to use them falls upon
language teachers and school stakeholders (e.g., administrators). Numerous evaluation
methods and suggestions as well as evaluation criteria have been proposed for
evaluating learning materials; some focus on general material evaluation while others
are used for specific materials such as textbooks, websites, or online courseware. This
section reviews evaluation criteria for ESL/EFL textbooks and web-based CALL
courseware, respectively.
It is necessary to include literature on textbook evaluation in this section for
three reasons. First, although the features of textbooks differ from those of online
courseware, the learning theories and teaching methodology are/should be similar. In
addition, research on ESL/EFL textbook evaluation emerged much earlier than that of
CALL courseware evaluation; as such, many discussions have examined the relation
18
between the learner and the textbook (e.g., Cunningsworth, 1984; Ellis, 1998; Littlejohn
& Windeatt, 1989; Sheldon, 1988; Tomlinson, 1998), and these may provide insights
into CALL evaluation. Finally, teachers should evaluate web-based learning materials as
seriously as they do any print materials (Bradin, 1999; Susser & Robb, 2004).
Criteria for Textbook Evaluation
Abundant literature exists on ESL/EFL textbook evaluation. Researchers have
suggested various aspects for evaluating and selecting English materials; to a great
degree, these researchers’ evaluation criteria are compatible. For instance,
Cunningsworth (1984) provided a checklist, which includes 70 questions, encompasses
7 areas: language content, selection and grading, presentation and practice, the
development of language skills and communicative abilities, supplementary materials,
learners’ motivation, and overall evaluation (pp. 75-79). Meanwhile, Littlejohn and
Windeatt (1989) identified six perspectives as fundamental in the evaluative approach:
general or subject knowledge in materials, views on the nature and acquisition of
knowledge, views on the nature of language learning, role relations implicit in materials,
opportunities for the development of cognitive abilities, and values and attitudes in
materials.
Subsequent researchers have followed quite similar points of view as these
previous studies. For example, Ansary and Babaii (2002) collected ten ESL/EFL
material evaluation checklists, proposed by Williams (1983), Skierso (1991), Ur (1996)
and others from 1970 to 2000; these checklists served as the basis of their study. Table
2.3 summarizes the studies on ESL/EFL textbook evaluation criteria reviewed in this
section; a summary of the global categories for ESL/EFL textbook evaluation,
developed by the researcher, is provided in Table 2.4.
19
Table 2.3
Summary of studies on ESL/EFL textbook evaluation reviewed in this study
Name (Year) Theme of the Study Results
Ansary & Babaii
(2002)
Systematic Textbook Evaluation A set of universal features of
EFL/ESL textbooks proposed
Cunningsworth
(1984)
Evaluating and Selecting EFL Teaching
Materials
Seven areas and 70 questions
focusing on learners’ needs and
processes proposed
Ellis
(1988)
Empirical Evaluation of Language
Teaching Materials
Guidelines and process for evaluation
proposed
Littlejohn & Windeatt
(1989)
Material Selection Six perspectives along with 6
questions suggested
Sheldon
(1988)
Evaluating ELT Textbooks and Materials A checklist of 17 categories and 53
criteria suggested
Skierso
(1991)
Textbook Selection and Evaluation A checklist of 11 categories and 104
items developed
Tomlinson
(1998)
Materials Evaluation Principles and process of developing
criteria proposed
Ur
(1996)
Assessing a Coursebook Guidelines and a checklist of 19
criteria provided
Williams
(1983)
Developing Criteria for Textbook
Evaluation
A checklist with 28 criteria proposed
20
Table 2.4
Global categories for ESL/EFL textbook evaluation
1. Theoretical background/Approach of the textbook
2. Goals and objectives of the textbook
3. Content of the textbook
4. Activities and tasks of the textbook
5. Author(s)/Developer(s) of the textbook
6. Supplementary materials and aids of the textbook
7. Physical Make-up/Presentation of the textbook
8. Administrative Concerns
9. Overall Value
In addition to the global textbook evaluation categories in Table 2.4, several
researchers and educators have pointed out the necessity of evaluating the teaching/learning
materials from learners’ standpoints and the importance of embracing the learner in the
evaluation process. In other words, learners’ needs, learning purposes, language levels, age,
cognitive state, and learning processes should also be taken into consideration. For example,
Cunningsworth (1984) suggests that learners’ needs and processes relate to teaching materials.
Meanwhile, Littlejohn and Windeatt (1989) argue that an evaluation approach should “look
beyond goals of language learning itself” and include issues such as “learners’ perceptions of
knowledge, language learning and roles” (p. 174). After reviewing the literature on the
evaluation of learning materials in general, the following section discusses evaluation criteria
specifically related to online CALL courseware.
Criteria for Online CALL Courseware Evaluation
CALL combines technology and language education, creating the distinctive features
of new-generation learning tools such as e- learning tools. Reeder and his colleagues (2004)
21
have distinguished pedagogical traits from technical ones in e- learning media. The former
include the proximity that web-based courseware offers to “authentic or simulated linguistic
and cultural settings” (p. 256); thus, CALL programs can promote a high degree of
interaction between the program and the learner. The latter include the “extensive use of the
multimedia capacities of computers, including complex graphic elements [… and… ] fairly
sophisticated sound elements” (p. 257).
On one hand, due to the unique characteristics of CALL tools, the approaches to
evaluating CALL courseware should differ from the way in which traditional English
materials are investigated; on the other hand, the principles and criteria for textbook
evaluation and CALL courseware evaluation in fact overlap to a certain degree, revealing
their essential roles in ESL/EFL material evaluation. This can be seen in, for example, Susser
and Robb (2004), who advise a practical and comprehensive framework with five modules
and items to checkmark for evaluating websites for language skills. The first three modules
overlap with each other as well as with evaluation criteria of textbooks; the fourth and the
fifth modules concern courseware and online courseware specifically.
Current literature highlights at least two deficiencies in the current approaches to
CALL courseware evaluation: the lack of a link between courseware design and instructional
methodology and the lack of new approaches to evaluating new genres of CALL courseware.
The first problem emerges from the fact that most evaluative criteria fail to link the design of
courseware to its pedagogical methodology; few CALL courseware evaluations provide a
methodological framework (Reeder et al., 2004). A few researchers have examined this issue,
attempting to link pedagogy to CALL tool evaluation. For instance, Hubbard (1988)
developed a framework for CALL courseware development and claims that courseware
evaluation needs to be related to pedagogical principles and their relevant constituents. Using
SLA theories, Chapelle (1998) proposed seven hypotheses and corresponding suggestions for
22
the development of multimedia CALL; issues of learning and learners play an important part,
which— as she expects— have provided “a starting point for development of a complementary
relationship between SLA research and CALL practice” (p. 33). Chapelle (2001) later sheds
light on the need for improvements in CALL tool evaluation and suggests five principles of
evaluation for evaluating CALL tools.
The second problem relates to the lack of a new and adequate approach to evaluating
CALL courseware (Reeder et al., 2004). Due to the distinctiveness of CALL tools, evaluation
criteria cover a wide range of issues. A few researchers have endeavored to propose
evaluation frameworks or models, instead of checklists. For example, Reeder and his
colleagues (2004), focusing on empirical evaluations, suggest a research agenda consisting of
four steps: 1) constructing an instrument to describe CALL materials, 2) building a
theoretical framework and taxonomy, 3) developing a suite of new instruments and guidelines
for evaluation derived from the framework, and 4) testing the suite of instruments and
guidelines on CALL materials in a setting with a representative range of users (p. 265).
However, instead of specifying evaluation criteria, this model simply provides guidelines on
how to conduct an empirical evaluation of CALL courseware.
In addition to instructional concerns, technical issues play an important role in CALL
courseware design and evaluation as well. Numerous researchers (e.g., Bradin 1999; Hémard,
2003; Nielsen & Mack, 1994; Squires & Preece 1999) have incorporated technological
criteria in CALL evaluation. These technological facets take into account 1) usability
(Nielsen, 2000)— the measure of the quality of the user experience when interacting with a
website or software, 2) accessibility (Susser & Robb, 2004)— aspects that facilitate or prevent
people with disabilities from accessing and using the online courseware, and 3) functionality
(Bradin, 1999)— easy navigation, rapid loading, and the quality of multimedia design.
23
A review of the current literature on evaluation frameworks and checklists for online
CALL courseware/websites─summarized in Table 2.5─ reveals both diversity and similarity.
However, overall it can be said that the evaluation criteria proposed in these studies fall into
two main categories— pedagogical and technical— followed by several subcategories. These
are illustrated in Table 2.6.
24
Table 2.5
Summary of studies on online CALL courseware evaluation reviewed in this study
Name (Year) Theme of the Study Results
Bradin
(1999)
Instructional Aspects of Software
Evaluation
A two-step checklist with follow-up
criteria proposed
Broadbent
(2002)
Decision Making Step by Step A checklist of 40 criteria proposed
Chapelle
(2001)
Principles for CALL Evaluation Five principles of evaluation
developed
Elissavet & Economides
(2003)
An Evaluation Instrument for
Hypermedia Courseware
A checklist of 124 items presented
Hémard
(2003)
Language Learning Online:
Designing towards User
Acceptability
A checklist for web design evaluation
provided
Hubbard
(1988)
An Integrated Framework for
CALL Courseware Evaluation
An evaluation framework consisting 3
major sections offered
Jolliffe et al.
(2001)
Selecting and Reviewing Learning
Resources
A checklist of 45 criteria proposed
Khan & Vega
(1997)
Factors to Consider When
Evaluating a Web-based Instruction
Course
Thirty-six evaluation criteria with
follow-up questions indicated
Murray & Barnes
(1998)
Evaluating Multimedia Language
Learning Software from a
Pedagogical Viewpoint
A framework with 15 questions
presented
Nielsen & Mack
(1994)
Heuristic Evaluation A usability engineering method
proposed
25
Table 2.5 (continued)
Odell
(1986)
Evaluating CALL software A checklist with two 2 main
categories─technical design and
Pedagogic evaluation─offered
Reeder et al.
(2004)
Toward a Theory of E/Valuation for
Second Language Learning Media
A research agenda consisting of 4
steps suggested
Robb
(1996)
ESL Software Evaluation An evaluation form with 29 criteria
offered
Squires & Preece
(1999)
Evaluating the Potential of
Educational Software
The Jigsaw Model proposed
Susser & Robb
(2004)
Evaluations of ESL/EFL
Instructional Websites
A framework with 5 modules and
checking items advised
Table 2.6
Global categories for online CALL courseware evaluation
Pedagogical Concerns
1. Curricular Issues
2. Learning Theory
3. Teaching Methodology
4. Accuracy
5. Learners’ Choices
6. Authenticity
7. Interactivity
8. Feedback
9. Authority
10. Bias/Stereotypes
Technical Considerations
1. Usability
2. Presentation/Physical Makeup
3. Navigation
4. Speed
26
As Table 2.6 shows, global categories for online CALL evaluation fall into two areas:
pedagogical concerns and technical considerations. This section will provide further
descriptions of these categories and subcategories.
Pedagogical concerns. Pedagogical concerns involve those issues regarding language
teaching and learning in the courseware:
n Curricular issues (e.g., Broadbent, 2002; Elissavet & Economides, 2003) include
the goal, objectives, and purpose of the courseware; the target users/learners; and
the skills focused on (listening, speaking, reading, writing, or integration of certain
skills). Evaluation criteria should consider whether the scope and aims of the
courseware are explicit, whether the purpose of the courseware is stated clearly,
whether the courseware fulfils its purpose, who the intended users are, what kind of
skills are being developed, what the coverage (e.g., topics) of the courseware
includes, and whether the courseware can be integrated into a curriculum.
n Learning theory (e.g., Tomlinson, 1998) refers to the theoretical background of
learning behind the courseware, such as behaviorist, cognitivist, and sociocognitive
approaches; this must be taken into account in the courseware evaluation to ensure
that the materials provided by the courseware fall in line with the educational
system’s approach to learning.
n In addition, the teaching methodology (e.g., Susser & Robb, 2004) adopted in the
courseware is another essential criterion. Teaching methodology concerns the
approach/method the learning tool adopts and thus affects the learning outcome to a
great degree.
n Accuracy (e.g., Bradin, 1999; Robb, 1996) refers to issues such as whether the
content is presented accurately, whether the information is reliable, and whether the
information sources are salient and noteworthy.
27
n Learners’ choices (e.g., Glidenston, 1994; Robb, 1996; Shin & Wastell, 2001)
addresses the freedom given to users in the learning process, such as whether the
user is allowed to control the order of presentation and whether the user is allowed
to skip certain activities.
n The authenticity (e.g., Jolliffe et al., 2001; Murray & Barnes, 1998) of the learning
materials in the courseware— as well as whether the materials are provided in
appropriate contexts— is vital for providing meaningful input for learners.
n Taking advantage of technological advances, CALL courseware should be
interactive (e.g., Elissavet & Economides, 2003; Susser & Robb, 2004), providing
learners with opportunities to interact with the computer or with others via the
computer.
n Feedback (e.g., Broadbent, 2002; Gildenston, 1994) is a vital element in the
courseware as appropriate feedback enhances learners’ motivation. Issues like
whether error handling is meaningful and helpful and whether immediate feedback
is offered for an online language- learning tool are important.
n Authority (e.g., Reeder et al., 2004) is another essential component in courseware.
The absence of language professionals or instructors on a courseware design team
creates a gap between design and actual use. When evaluating CALL courseware,
one has to ponder whether the information on the author(s) is available and
sufficient, whether the author(s) has trustworthy education, expertise, or experience
in related fields, and whether the author(s) can be contacted for further information.
n Finally, bias or stereotypes (e.g., Gildenston, 1994; Jolliffe et al., 2001; Murray &
Barnes, 1998) in the form of the material developer’s background, knowledge, and
attitudes toward language learning and different cultures can significantly impact
the content presented in the courseware.
28
Technical issues. Most existing evaluation criteria concern four primary technical
issues: usability, presentation/physical makeup, navigation, and speed.
n Usability (e.g., Nielsen & Mack, 1994) is an umbrella term covering different but
comparable meanings in different fields and sub-fields. The terminology related
to usability has changed with various shifts in the field. In the 1980s and 1990s,
the buzzwords were functionality, usability, usability engineering, and user-
centered design; in the 2000s, the term user experience engineering started to
appear. Scholars encouraged practitioners to “look beyond traditional usability
concerns, such as ease of learning and effectiveness, to broader user issues, such
as aesthetics, collaboration, accessibility, credibility, persuasion, and pleasure”
(Wilson, 2005, p. 6). Nielsen’s (2000) well-known heuristic model provides
guidelines for software design with a high degree of usability. In short,
evaluating the usability of CALL courseware covers the following issues:
whether the user interface is easy to use, whether the information is easy to
access, whether the online courseware is fee-based, whether a user must register
a name and password before using it, whether the instructions or guide are clear
and easy-to-follow, whether the learner has choices in using the courseware (e.g.,
sound volume and text font), and whether the courseware offers different
pathways corresponding to different learning styles/needs/paces/levels.
n The presentation/physical makeup of the courseware can impact learners’ ability
to access content. Presentation/physical makeup (e.g., Robb, 1996) ranges widely
from the layout (word font, color, format, spacing, etc.) to multimedia (graphics,
sounds, audio/video files, animation). Important criteria focus on whether the
web pages or information presented is comprehensible, whether the layout of the
courseware is attractive to users and facilitative for learning, and whether the
29
format is consistent.
n Related to presentation/physical makeup is navigation (e.g., Chapelle, 2001;
Robb, 1996). The quality of navigation of online courseware plays a crucial role,
and criteria such as how easy it is to move from page to page and/or from link to
link, whether all internal and external links work properly, and whether any
incorrect links or dead ends exist need to be considered.
n The speed and ease (e.g., Broadbent, 2002) in downloading a web page is critical
for web-based courseware.
As stated earlier the criteria in the studies reviewed here can be divided into two main
categories─pedagogical and technical. These two facets are like the wings of a bird; only
when they are equal in strength can the bird succeed in flying. Thus, a balanced view of these
two evaluation schemes is necessary.
2.4 Research Questions
Recent studies and practices of CALL material evaluation have come into a new era;
consequently, researchers continue to produce new evaluation models for CALL courseware.
Yet most such models and frameworks emphasize or incorporate general aspects. Since
learners are the center of the language learning process as well as the primary users of
courseware, evaluating a program from the learners’ perspective is essential. Language
teachers and those involved in the decision-making process must keep the learner in mind
when evaluating and selecting CALL materials.
Based on the review of the literature, this study claims that the concept of learner-
centeredness in language teaching is no different from the idea of user-centeredness in CALL.
Nonetheless, few researchers have discussed this connection between learner-centeredness
and CALL evaluation. Therefore, this study aims to explore the following questions:
30
1. What are the criteria for evaluating online CALL courseware from the perspective
of learner-centeredness?
2. Can a learner-centered evaluation checklist based on such criteria assist in
evaluating the learner-centeredness in online courseware, using MyET as an example?
The methodology used to answer these questions is described in the following chapter,
and Chapter Four discusses the research results that answer these questions.
31
CHAPTER THREE
Methodology
3.1 Introduction
In order to answer the research questions identified in the previous chapter, this study
adopted primary qualitative research methods consisting of three stages; at each stage,
specific data were collected for analysis. This chapter outlines the three phases (see Table 3.1)
as well as the participants, data collection and analysis procedures, and validation of results at
each stage. The first stage involved developing a checklist aimed at collecting evaluation
criteria associated with the characteristics of learner-centeredness, followed by the
establishment of an evaluation checklist focused on learner-centeredness. The second stage
involved piloting the checklist developed in the first stage by applying the checklist in
evaluating the commercial CALL courseware MyET; the evaluation results of MyET in terms
of its learner-centeredness were analyzed during this stage. The third stage involved a meta-
evaluation─evaluating the effectiveness of the checklist itself; during this stage, evaluators’
feedback on the checklist was collected for analysis. The discussion of the validation of
results addresses how the validity of the analysis was double-checked or confirmed by
another English teacher or group of people.
32
Table 3.1
An overview of the three-stage study
Stage I. Development of the Learner-centered Evaluation Checklist
Stage I aimed to develop an evaluation checklist based on the premises of learner-
centeredness. More than 45 studies were reviewed and a checklist of 39 items was developed.
Data collection and analysis during this stage lasted from July 2006 to December
2006.
Stage II. Application of the Evaluation Checklist: Evaluation of MyET
Stage II aimed to pilot the constructed evaluation checklist developed in Stage I.
Three teachers and 22 non-English-major students from a university in northern Taiwan used
the checklist to conduct an evaluation of MyET. Some descriptive statistical analysis was
conducted on the quantitative data resulting from use of the checklist.
Data collection was done in January 2007 and data analysis was conducted from
February 2007 to April 2007.
Stage III. Meta-evaluation of the Evaluation Checklist
Stage III aimed to explore the effectiveness of the evaluation checklist. The teachers
were interviewed and the students were given three open-ended questionnaire questions in
order to elicit their feedback on the checklist.
Data collection was done in January 2007 and data analysis was conducted from
February to April 2007.
3.2 Stage I. Development of the Learner-centered Evaluation Checklist
The first stage aimed to develop an evaluation checklist based on the characteristics of
learner-centeredness. More than 45 sets of evaluation checklists, models, guidelines, and
studies were reviewed and 39 criteria generated.
3.2.1 Rationale
The rationale for creating a learner-centered checklist is based on two main concerns.
First, as mentioned in Chapter Two, although a great number of evaluation checklists and
models have been proposed in the ESL/EFL literature, most share numerous criteria in
common; moreover, their purpose is to help language educators evaluate whether materials
are adequate or not. The coverage of the evaluation criteria is generally broad and lacks the
33
basis of learning theories (Tergan & Schenkel, 2003). However, ESL/EFL materials are
expected to be learner-centered; therefore, it is necessary to incorporate this characteristic
into the evaluation criteria so that language teachers can keep learners in mind while
evaluating and selecting appropriate materials for their students. Ideally, students are
practically involved in the evaluating and selecting process; however, in reality this seems
rare in the EFL settings in Taiwan. EFL materials or textbooks are usually chosen by English
teachers without students’ involvement due to economic and efficiency issues.
In addition, the need exists for a new genre of evaluation tools for the new generation
of CALL materials. For example, experts have asserted that user- friendliness and usability are
crucial components of CALL tools. It has also been argued that most evaluation checklists
focus on technical rather than on learning and educational issues (Tergan & Schenkel, 2003).
Therefore, it is necessary to link instructional and methodological considerations to technical
designs in a CALL tool. This study attempted to develop an evaluation checklist, based on
existing software/website evaluation checklists and literature on learner-centeredness as well
as ESL/EFL material evaluation, for online CALL courseware. Language educators may thus
have another evaluation tool to assist them in evaluating and selecting suitable CALL tools
for learners.
This study adopted the form of a checklist as the evaluation tool. Although traditional
checklists have received several criticisms (see section 2.3.1), a checklist, compared to other
evaluation methods, is the most frequently used tool due to its convenience and ease (Tergan,
2003). This study implemented a checklist as the evaluation tool but at the same time it also
left room for evaluators to provide comments and feedback in order to attain more descriptive
data regarding each evaluation decision made on the checklist.
34
3.2.2 Data Collection
Since CALL is a multidisciplinary area (Levy, 1997), the range of data collection at
this stage covered a wide scope, including current literature and studies in the fields of
applied linguistics, second language acquisition (SLA), and information technology (IT). Two
primary areas of data were collected: evaluation criteria for ESL/EFL textbooks and for
multimedia learning materials— e.g., software and online courseware, from general
educational materials to English learning courseware. More than 45 sets of evaluation
checklists, models, and guidelines were collected and reviewed during this stage (see
Appendix I for the sources adopted and adapted).
3.2.3 Data Analysis
Each evaluation criterion, item, or guideline of the data collected was individually
reviewed to determine whether it conformed to the premises of learner-centeredness
discussed in section 2.1. After a review of the options, the basic format of the checklist was
established. Based on the literature review, evaluation criteria fell into two main categories:
pedagogical and technical aspects, each with specific subcategories (see Table 2.6). The
developed evaluation checklist followed the categorization.
Those criteria that seemed most able to help discriminate whether learning materials
or tools were learner-centered were selected as evaluation items for the checklist and placed
under different subcategories. For example, one criterion in a reviewed study said, “Is
feedback on learner responses encouraging?” (Son, 2003) and another said, “The hypermedia
courseware provides opportunities for interaction” (Elissavet & Economides, 2003); criteria
such as these correspond to the Premise Four of the APA’s Learner-centered Model:
“Learning occurs best in a positive environment in which the learner has positive
interpersonal interaction and feels appreciated and validated.” Thus, these criteria were
selected as items under the category feedback of the checklist. On the other hand, another
35
evaluation checklist asked, “Is the content accurate?” and “Is there information on the
author?” (Son, 2003). Although such criteria are not inadequate, they seem irrelevant to the
Premises of the APA’s learner-centered model. Thus, such questions were not included in the
evaluation checklist. The wording was modified when necessary.
3.2.4 Construction Process
The study adopted a 4-point Likert-scale ranging from 0 (very unlikely) to 3 (very
likely). Using an even number of options helps avoid evaluators’ tendency to choose the
neutral score available on an odd-numbered scale (Brown, 2000). To make the study more
meaningful, both teachers and students— the main users of the courseware— participated in
the study in order to increase the validity of the study by triangulating different data sources,
even though the checklist was aimed at language educators’ use.
Slight differences existed in the design between the checklists used by the teacher-
evaluators and those by the student-evaluators (see Appendices I and II for comparison). The
teacher-evaluators used a Chinese-English version, with headings specifying the main
categories and subcategories. The checklist for student-evaluators was modified somewhat to
be shorter and more concise; categories were not listed, and all descriptions were in Chinese.
This modification stemmed from the fact that a long checklist might exhaust student-
evaluators; moreover, descriptions in Chinese were deemed more comfortable for the
students.
In addition, time constraints made it impossible to interview all students about their
feedback on MyET. Therefore, an additional evaluation question was included in the student-
evaluator checklist— “In general, this courseware is helpful for learners; namely, it is able to
meet learners’ needs or enhance English learning” (see Appendix II)— to obtain their overall
opinion on the courseware being evaluated. At the end of the student-evaluator checklist were
three open-ended questions used for eliciting feedback on the checklist. The three open-ended
36
questions and the rationale for adding them are presented in section 3.4.3.
3.2.5 Validation
Because the data analysis and construction process were subjective, it was necessary
to obtain further validation of the analysis. After the evaluation tool was developed, a college-
level English teacher reviewed the checklist and recommended revisions to wording and
grammar; these recommendations were followed. For example, the teacher suggested
changing the word “easy-to-access” into “easy to access”. The checklist was subsequently
proposed and discussed at a meeting1 with several professors and graduate students of
Tunghai University, Taichung, Taiwan. Revisions to wording and the number of evaluation
items were again made according to the feedback and suggestions received from the meeting.
The original checklist included about 50 items; 39 items were left after these modifications.
The deleted items were considered irrelevant to learner-centeredness. In addition, a table
matching each evaluation criterion with one or more premises of the APA’s learner-centered
model developed to reveal the process of construction of the checklist (see Appendix III).
Afterwards, a colleague also helped match the criteria with the premises of learner-
centeredness in order to double check the researcher’s decisions. All the items matched were
the same except item I, 2.4; one thought it went with Premise Two, and the other thought it
corresponded to Premise Four. Otherwise, both shared agreement to a great degree. Another
university teacher looked over the table as a further check on the content of the validity of the
checklist.
3.3 Stage II. Application of the Evaluation Checklist: Evaluation of MyET
The second stage aimed to pilot the constructed evaluation checklist and to provide
the 25 participants an experience of using the checklist. The participants conducted an
1 The meeting was an informal review session held on November 3, 2006, with graduate students and professors in the Department of Foreign Languages and Literature of Tunghai University, Taichung, along with visiting faculty member Dr. Antony Kunnan.
37
evaluation of MyET using the checklist. Some descriptive statistical analysis was conducted
on the quantitative data resulting from use of the checklist.
3.3.1 Rationale
The checklist was tried out in order to attain actual experience of using the developed
checklist. The online CALL courseware MyET was chosen as the evaluand. MyET was
designed and developed by L Labs in Taiwan. Founded in 2002, the team members of L Labs
come from various countries such as Taiwan, Canada, the U.S., Japan and Korea and all hold
master or doctoral degrees. L Labs serves as a “technology provider” working with “content
providers” and English materials developers (L Labs, 2006). MyET functions as a platform
for material developers and publishers to offer target users accessible interfaces. Thus far,
several material providers— such as Studio Classroom, International Community Radio
Taipei (ICRT), and Longman (Pearson)— have adopted MyET to offer online learning courses,
although their target users and pedagogical purposes vary. For instance, General English
Proficiency Test (GEPT) Courses provided by Studio Classroom aim to familiarize learners
with the test style and model of GEPT, whereas Studio Classroom Standard Course gives
online learning materials for learners’ use in self-study to improve English proficiency.
Technologies used in this online courseware focus primarily on developing learners’
listening and speaking skills. One feature that L Labs claims differentiates the courseware
from other online courseware in Taiwan is its Automatic Speech Analysis System (ASAS),
which is designed to train learners’ pronunciation and speaking skills by analyzing and
recognizing the sounds of the user, indicating the problems of each sound and syllable as well
as providing suggestions for improvement. Dr. Paul Pimsleur ’s learning method is adopted as
the theoretical background of MyET, with the claim that the way of learning a language by
repeating sentences via the courseware is able to assist learners to build a way of learning
similar to learning a native language (L Labs, 2006). Except for a few free trial lessons, all
38
users have to pay to get an authorized number to access MyET courses (see Appendix VI for
a few sample pages of MyET).
This study adopted MyET as the evaluand for several reasons. First, it is a locally
developed online English self- instruction tool in Taiwan, although the development team
includes professionals from various countries. The target users of this CALL courseware are
EFL learners, mostly in Taiwan and other Asian countries. Second, since entering the market,
the courseware has reportedly registered over one hundred thousand users in Asia, including
more than ten local colleges and high schools and ten corporations (L Labs, 2006). It seems
to have gained popularity over the past few years, as more and more self- learning centers and
language labs in Taiwan have been offering MyET as part of their supplementary learning
materials for students. In addition, most of the online materials’ authors or providers (e.g.,
Studio Classroom and Longman) are recognized English publishers who have historically
used magazines and books as teaching and learning tools and have now expanded to include
MyET. It is assumed that these providers have evaluated the potential of MyET. Given the
diversity and scope of online courses, much remains to be seen as to whether MyET is able to
meet users’ expectations. As such, the current study adopted MyET as the evaluand and
expected that the results of evaluation would provide some practical implications for the
CALL field in Taiwan.
This study originally intended to evaluate three different courses under MyET.
However, the participants had used only two: Studio Classroom Standard Course and Let’s
Talk in English by Studio Classroom. Therefore, the study was adapted accordingly. The
difference between the two courses is merely the level of difficulty and the texts (readings
and dialogues). Since MyET serves as a platform, the two courses are presented with the
same interfaces and learning activities, such as repetitions and role-plays. The degree of
learner-centeredness— if any— of the two courses and of the overall interface design of
39
MyET was evaluated using the checklist.
3.3.2 Participants
Twenty-five participants— 3 teachers and 22 students— were involved in the second
stage of the study. The three English teachers (teacher-evaluators) work in National Taiwan
Normal University, Taipei, Taiwan; the school has adopted several MyET courses for its
English self- learning center. In addition to MyET, the self- learning center also has adopted
other courseware for students. Two of the teachers have about ten years of teaching
experience; one of them is the director of the English self- learning center. The third teacher
has taught in the school for one year. All of them have had experience using MyET as
supplementary materials in class; one teacher had his students use MyET for seven to eight
months, one had had his students use MyET for four months before the interview was
conducted, and the third one for two months. Since all three teachers had used MyET, it was
not necessary to train them on how to use the courseware. The three teachers evaluated the
Studio Classroom Standard Course, one of the courses under MyET.
The remaining 22 participants (student-evaluators) were non-English-major freshmen
at National Taiwan Normal University, from a Basic English class taught by one of the
teacher-evaluators. They were all students in the low-achievement class and had been using
Let’s Talk in English, another course under MyET, for three months. They were required to go
to the self- learning center to use Let’s Talk in English as outside-class materials. These
students might go to the self- learning center anyt ime they wanted; the times they went were
recorded. The teacher even offered incentives for those who used the courseware most. Since
different courseware offer different functions and focuses, the teacher of the class asked his
students to use MyET mainly for pronunciation and speaking practice. He expected his
students’ interest in English learning to increase via diverse learning tools with different
functions. This group of learners might represent the majority of MyET users in English self-
40
learning centers in colleges, Taiwan. Since one of the purposes of a self- learning center is to
offer various self- learning and supplementary materials for learners, low-achievement
learners, in particular, need such assistance, whereas high-achievement and highly motivated
learners tend to have developed their own ways of learning and are relatively active in
learning.
3.3.3 Data Collection
The instrument adopted at this stage was the checklist developed during Stage I. The
differences between the teacher-evaluator and student-evaluator versions were addressed
earlier in section 3.2.4. In addition to the checklist, a formal semi-structured interview with
each teacher-evaluator was conducted before the teacher-evaluators applied the checklist in
order to obtain general information about their background and concepts about learner-
centeredness as well as to ensure that they understood the objective of the evaluation and the
meaning of each criterion in the checklist. The interview consisted of three primary questions:
1) How long have you been teaching English? 2) How long have your students been using
MyET? and 3) What do you think about learner-centeredness? All interviews were audio
recorded with a tape recorder and conducted in Chinese (see Part I in Appendix IV for
transcripts, including an English translation). Notes were taken when necessary during the
interviews.
Meanwhile, data from the student-evaluators were collected from their evaluations of
MyET using the checklist. The 22 participants were asked to do the checklist during class
time. They were told that the researcher and their English teacher would like to know their
feedback on MyET. They were also reminded that three additional open-ended questions
enclosed in the checklist were concerning the evaluation checklist itself. As soon as they
completed filling out the checklist, they handed it to the researcher. Unlike with the teacher-
evaluators, no interviews were conducted with the student-evaluators.
41
3.3.4 Data Analysis
The results from the teacher-evaluators’ and student-evaluators’ ratings of MyET were
entered into and analyzed with Microsoft Excel. The evaluation results were analyzed and
categorized based on the scores received. The categories included: strengths of the
courseware, weaknesses of the courseware, and non-scored evaluation criteria. The findings
are presented and further discussed in Chapter Four. Because the courses evaluated by the
teachers and students differed, this study did not attempt to compare the evaluation results of
the two courses with each other. However, except for the difficulty of the content, all the
tasks and interfaces of the two courses were similar. Thus, it should be possible to discuss to
what degree the two groups of raters share their opinions of MyET as a whole.
3.3.5 Validation
Although the current study aimed to increase the awareness of learner-centeredness
among English educators as they evaluate courseware, learners’ opinions were collected in
order to triangulate them with teacher-evaluators’ evaluations. According to Patton (1980),
triangulation is “a powerful solution to the problem of relying too much on any single data
source or method” (p. 193). Involving the primary users’ opinions should provide further
insights into learners’ needs as well as expectations. In addition, to confirm the findings, a
colleague of the researcher helped double-check the results using Microsoft Excel to ensure
that the numbers entered and the quantitative results, including total scores and the mean of
each criterion, were accurate. This colleague also helped review the findings selected by the
researcher based on the rankings of the ratings (see section 4.2 for findings and discussion).
3.4 Stage III. Meta-evaluation of the Evaluation Checklist
The third stage aimed to investigate the effectiveness of the evaluation checklist.
Interviews and written open-ended questions were adopted respectively to elicit the
participants’ feedback on the checklist itself.
42
3.4.1 Rationale
This study’s main focus is to develop an evaluation checklist based on the premises of
learner-centeredness; as an evaluation tool, it was essential to investigate whether the
checklist was able to function as expected and planned. Thus, an evaluation of the
evaluation— a meta-evaluation (Stufflebeam, Madaus, & Kellaghan, 2000)— was necessary.
Meta-evaluation can be defined as a critical review of the evaluation design and
implementation. At the same time, since the primary purpose of the study is to reinforce the
awareness of learner-centeredness in CALL eva luation, Stage III investigated whether the
checklist was able to draw the evaluators’ attention to learner-centeredness.
3.4.2 Participants
The participants at Stage III were the same as those in Stage II (see section 3.3.2).
3.4.3 Data Collection
To obtain participants’ feedback on the effectiveness of the checklist, a semi-
structured interview with seven open-ended questions for each of the three teacher-evaluators
was conducted after they used the checklist. The initial interview questions are provided in
Table 3.2. Nonetheless, any interview— regardless of whether it is structured or semi-
structured— has its own dynamics during the process itself. Consequently, the actual
interview questions differed slightly from the proposed ones (see Part II in Appendix VI for
transcripts, including an English translation). Interviews were used to obtain data during this
stage because they are the most direct way to pursue in-depth information (McNamara, 1999).
They also allowed for flexibility in asking follow-up questions based on interviewees’
responses. All interviews were tape recorded and conducted in Chinese.
43
Table 3.2
Interview questions of the second interviews with teacher-evaluators
1. Are these evaluation questions easy to answer?
2. What can be done differently in a revised version?
3. Do you think the checklist can focus your attention on learner-centeredness?
Why or why not?
4. Have you used any other evaluation checklists for ESL/EFL materials?
If yes, what are the differences between the one or ones you used and this one?
5. What are the strengths and weaknesses of this evaluation checklist?
6. Do you think this evaluation checklist will be helpful for English educators
in evaluating and selecting a suitable online courseware? If so, how?
7. What do you think of the evaluation checklist in general?
To collect the data of the student-evaluators’ perspectives on the checklist, three
written open-ended questions were added to the checklist. The following offers the English
version of the three questions.
1. Generally speaking, do you find these evaluation items easy to answer? Why/Why
not?
2. Do you think this evaluation checklist is able to focus the attention of teachers and
students on learners’ needs and conditions?
3. In addition to these evaluation items, what concerns should be taken into account
when a teacher is evaluating and selecting suitable learning materials to meet
learners’ needs?
The rationale for adopting open-ended questions stemmed mainly from time
constraints related to interviewing all 22 student-participants. Also, conducting only selective
interviews would require determining who should be selected, which might cause another
issue affecting the study. Open-ended questions offered all participants room to provide
comments on the questions.
44
3.4.4 Data Analysis
After collecting the data through interviews with the teachers, the researcher first
uploaded the audio-recorded tape into the computer and saved it as a Microsoft.WAV file.
The interviews were then played via the Microsoft Media Player program and selectively
transcribed by the researcher herself following certain basic principles: first, only main
interview questions and follow-up questions as well as the interviewee’s responses to these
questions were transcribed. Questions and conversations unrelated to the research were
omitted— e.g., previous studies that the teacher-evaluator’s students did. Second, redundant
words/phrases, such as discourse markers— e.g., mmm (“嗯” in Chinese)— were omitted.
Third, any repetitions or answers requiring self-corrections by the speaker were omitted; only
the corrected words/phrases/sentences were kept. The transcripts were translated into English
and marked with line numbers for ease of reference (see Appendix VI). The responses from
the interviewees were reviewed and divided into several categories based on the purpose of
each interview question. The categories generated include: overall opinions, strengths of the
checklist, weaknesses of the checklist, and suggestions for further revisions.
On the other hand, to analyze data collected from the student-evaluators, certain
categories based on their responses on the three open-ended questions were set up. The
process for categorizing was as follows. First, all of the comments were listed and reviewed.
Responses embracing the same meaning, although maybe expressed differently, were
gathered together. For example, “okay” (“還好” in Chinese) and “all right” (“尚可” in Chinese)
were regarded as the same category. This process resulted in several categories (see section
4.4.2). Then, these comments were counted and converted into numbers (see Table 4.10).
3.4.5 Validation
A colleague of the researcher assisted in the review of the transcripts, comparing the
transcripts and the audio-tape to ensure that the transcripts covered all the main points of the
45
interviews. No revisions were done. Another college- level English teacher read through the
whole transcripts and marked which category each response belonged to— e.g., overall or
specific, and strengths or weaknesses— with the goal of double-checking the categorization
generated by the researcher. The teacher’s categorization and the researcher’s were exactly
the same.
3.5 Summary
This chapter has described the research design and process of the three-stage study.
The following chapter will discuss the findings, including the establishment of the learner-
centered evaluation checklist, the evaluation results of MyET and the effectiveness of the
developed checklist. In addition, the research questions are answered and discussed.
46
CHAPTER FOUR
Findings and Discussion
4.1 Introduction
This three-stage study aimed to answer the research questions proposed in Chapter
Two— namely, 1) what are the criteria for evaluating online CALL courseware from the
perspective of learner-centeredness? and 2) Can a learner-centered evaluation checklist based
on such criteria assist in evaluating the learner-centeredness in online courseware, using
MyET as an example? Chapter Three presented the research design used to answer the
questions. This chapter will state and discuss the results from each stage, including the
developed learner-centered checklist, the evaluation results on MyET, and the meta-
evaluation of the evaluation checklist. Section 4.2 will address the first research question,
section 4.3 the findings from the MyET evaluations, and section 4.4 the second research
question. Table 4.1 provides an overview of the findings and discussion.
47
Table 4.1
An overview of the findings and discussion
Stage I: Development of the Learner-centered Evaluation Checklist
1. Features of the Evaluation Checklist
2. Definitions of the Subcategories
3. Summary of Stage I
Stage II: Application of the Evaluation Checklist: Evaluation of MyET
1. Teacher-evaluators’ Evaluation Results
Strengths of the Courseware
Weaknesses of the Courseware
Non-scored Evaluation Criteria
Controversial Issues
2. Student-evaluators’ Evaluation Results
Strengths of the Courseware
Weakness of the Courseware
Non-scored Evaluation Criteria
3. Caveats about Stage II
4. Summary of Stage II
Stage III: Meta-evaluation of the Learner-centered Evaluation Checklist
1. Teacher-evaluators’ Feedback
Overall Opinions toward the Evaluation Checklist
Strengths of the Evaluation Checklist
Weaknesses of the Evaluation Checklist
Suggestions for Further Revisions
Teacher-evaluators’ Concepts on Learner-centeredness
2. Student-evaluators’ Feedback
Overall Opinions toward the Evaluation Checklist
Positivity about the Evaluation Checklist
Negativity about the Evaluation Checklist
Additional Comments
3. Summary of Stage III
48
4.2 Stage I: Development of the Learner-centered Evaluation Checklist
To answer the first research question, the learner-centered evaluation checklist, which
includes 39 evaluation criteria, was constructed. Table 4.2 highlights all the evaluation
categories and items of the checklist (see Appendix I for the checklist used by the teacher-
evaluators). To allow for a triangulation of data, another version of the checklist was
generated for student-evaluators’ use in order to attain students’ feedback (see section 3.3.5
and Appendix II for the checklist used by the student-evaluators). The versions include
certain differences: 1) neither the main categories/subcategories nor English translations were
included on student-evaluators’ checklist, for simplicity, and 2) whereas teacher-evaluators
had the chance to provide feedback on MyET during face-to-face interviews, student-
evaluators responded to an additional evaluation criterion— “In general, this courseware is
helpful for learners; namely, it is able to meet learners’ needs or enhance English learning” on
the checklist (see item No. 40 in Appendix II).
49
Table 4.2
Learner-centered evaluation checklist for online CALL courseware
I. Pedagogical Criteria
1.1 The information on the objectives is sufficient for learners. 0 1 2 3
1.2 The objectives of the course are associated with learners’
learning development.
0 1 2 3
1.3 The objectives are achievable for learners. 0 1 2 3
1
Learning
Objectives
1.4 The objectives reflect what learners need to know. 0 1 2 3
2.1 The learning activities are able to arouse learners’ interests in
participation.
0 1 2 3
2.2 The instructions of the learning activities are clear for learners. 0 1 2 3
2.3 The guidance and models of learning activities are clear for
learners.
0 1 2 3
2.4 There is a variety of learning activities for learners with varied
learning styles.
0 1 2 3
2
Learning
Activities
2.5 There is a variety of learning activities for learners with multiple
intelligences (e.g., interpersonal, mathematics… .).
0 1 2 3
3.1 The content matches learners’ English proficiency level. 0 1 2 3
3.2 The content is related to learners’ life experience. 0 1 2 3
3.3 The content covers a variety of topics that are able to attain
learners’ interests.
0 1 2 3
3.4 The content is culturally appropriate for learners. 0 1 2 3
3.5 The content accommodates learners with various learning styles. 0 1 2 3
3.6 The content helps learners build up strategic ability. 0 1 2 3
3.7 The content stimulates recall of prior learning. 0 1 2 3
3
Content
3.8 The content enhances learners’ cognitive development. 0 1 2 3
4
Skills
4.1 The course offers skill-training that is able to meet learners’
needs.
0 1 2 3
5.1 The course offers immediate and proper feedback on learners’
performance and/or responses.
0 1 2 3
5.2 The course offers feedback that is able to motivate learners. 0 1 2 3
5
Feedback
5.3 The course offers meaningful and helpful error handling which
is able to help learners improve.
0 1 2 3
50
5.4 The course offers feedback that can be saved and accessed easily
so that learners may realize what difficultie s they have in the
learning process.
0 1 2 3
5.5 The course offers an opportunity for learners to give personal
feedback on their learning experience.
0 1 2 3
6.1 Learners are allowed to choose the learning activities that they
want to use.
0 1 2 3
6.2 Learners have their control over their own pace. 0 1 2 3 6
Freedom 6.3 Learners are able to learn the course independently outside the
classroom.
0 1 2 3
7.1 Learners have opportunities to interact with the courseware. 0 1 2 3 7
Interactivity 7.2 Learners have opportunities to interact with other learners. 0 1 2 3
8.1 The course offers an entry requirement verification device to see
if the learner meets the entry conditions.
0 1 2 3
8.2 The course is able to help arouse learners’ awareness of their
own learning progress/status.
0 1 2 3
8.3 Learners are able to find where they have been (i.e. a learning
profile is offered) and where they are heading during the course.
0 1 2 3
8
Learners’
profile
8.4 The learning profile is easy to access. 0 1 2 3
II. Technical Criteria
1
Instructions
1.1 The instructions on how to use the courseware are easy for
learners to follow.
0 1 2 3
2.1 The interfaces are easy for learners to use. 0 1 2 3 2
Interfaces 2.2 The icons/symbols are intuitive and easy to understand. 0 1 2 3
3
Layout
3.1 The overall presentation (colors, font, spacing, etc.) is able to
interest learners.
0 1 2 3
4.1 It is easy to install the courseware. 0 1 2 3 4
Accessibility 4.2 It is easy to access the course to be learned. 0 1 2 3
5
Support
5.1 The technical support offered (e.g., FAQ, online support) is
sufficient.
0 1 2 3
51
4.2.1 Features of the Learner-centered Evaluation Checklist
Several features distinguish the current checklist from others. First, this checklist
focuses on learner-centeredness, and all evaluation criteria used are related to the principles,
premises, and characteristics of learner-centeredness discussed in section 2.1. For instance,
item I, 3.2 “The content is related to learners’ life experience”─adapted from Odell (1986),
Broadbent (2002), and Su (2003)─is connected to Premise Two of the APA’s Learner-
centered Model regarding learners’ exposure to interaction with others (see Table 2.2); item I,
2.1 “The learning activities are able to arouse learners’ interests in participation”─adapted
from Cunningham (1995) and Elissavet and Economides (2003)─conforms to Premise Four
of the Learner-centered Model concerning learners’ uniqueness. Table 4.3 lists examples of
the evaluation criteria echoing the premises of learner-centeredness (see Appendix III for all
the evaluation criteria matching with the Premises of the learner-centered Model). As a matter
of fact, a criterion might be corresponding to more than one premise of learner-centeredness.
52
Table 4.3
Examples of the evaluation criteria echoing the premises of the APA’s learner-centered model
Premises of Learner-centeredness Evaluation Criteria in the Checklist
1. Learners are distinct and unique, and this
distinctiveness and uniqueness must be taken into
account.
I, 2.5 There is a variety of learning activities for
learners with multiple intelligences (e.g.,
interpersonal, mathematics… .).
2. Learners’ uniqueness includes their emotional
states, learning rates, learning styles, abilities,
talents, and needs.
I, 3.5 The content accommodates learners with
various learning styles.
I, 4.1 The course offers skill-training that is able to
meet learners’ needs.
3. Learning is a constructive process, so, what is
being learned should be relevant and meaningful to
the learner and be based on the learner’s prior
knowledge and experience.
I, 8.2 The course is able to help arouse learners’
awareness of their own learning progress/status.
I, 8.3 Learners are able to find where they have been
(i.e. a learning profile is offered) and where they are
heading during the course.
4. Learning occurs best in a positive environment in
which the learner has positive interpersonal
interactions and feels appreciated and validated.
I, 7.2 Learners have opportunities to interact with
other learners.
5. Learning is a natural process; learners are
naturally curious and interested in learning about
their world. Learners must be involved in an
educational decision-making process, and
individuals’ differences should be respected and
accounted for in the learning process.
I, 3.8 The content enhances learners’ cognitive
development.
I, 6.1 Learners are allowed to choose the learning
activities that they want to use.
Note: References to the evaluation criteria are provided; for example, “I, 1.1” refers to item 1.1 in the
pedagogical category and “II, 1.1” refers to item 1.1 in the technical category.
In addition, the evaluation checklist is category-oriented. All of the criteria were
divided into two main categories— pedagogical or technical aspects— along with 13
subcategories. Such categorization offered evaluators an advanced organizer of the areas
being examined. Due to the fact that most current evaluation checklists for CALL tools pay
more attention to technical issues, it was vital to incorporate pedagogical facets into the
evaluation checklist. Within the primary pedagogical category were eight subcategories:
53
learning objectives, learning activities, content, skills, feedback, freedom, interactivity, and
learners’ profile. Furthermore, certain technical issues relevant to users’ needs were included
in the technical category, which contained five subcategories: instructions, interfaces, layout,
accessibility, and support. Each subcategory included from one to eight evaluation criteria
(see Table 4.2). Definitions of these subcategories are further provided in section 4.2.2.
Finally, a 4-point Likert scale was utilized ranging from 0 (very unlikely) to 3 (very
likely). In addition to indicating the degree to which the evaluators agreed with each criterion
description, evaluators could leave comments about their decisions in the comment column.
Open-ended comments offer insight into evaluators’ thinking in the process of evaluating and
provide further explanations of their judgments.
4.2.2 Definitions of the Subcategories
Thirteen subcategories derived from the review of studies were considered crucial in
evaluating the learner-centeredness of online CALL courseware. Table 4.4 lists the definitions
of each subcategory and its sample evaluation criteria.
54
Table 4.4
Definitions of subcategories and sample evaluation items from the evaluation checklist
I. Pedagogical Criteria Subcategory Definition Sample Evaluation Item
1. Learning objectives
Course objectives to be achieved by learners. 1.1 The information on the objectives is sufficient for learners.
2. Learning activities
Course activities provided to learners for use with the courseware.
2.4 There is a variety of learning activities for learners with various learning styles.
3. Content Content presented by the course, including reading texts and dialogues.
3.1 The content matches learners’ proficiency level.
4. Skills Skills the courseware trains learners to develop through learning activities and content.
4.1 The course offers skill-training that is able to meet learners’ needs.
5. Feedback The courseware’s feedback system/mechanism, such as correction or grading learners’ performance.
5.2 The course offers feedback that is able to motivate learners.
6. Freedom The choices and control level allowing learners to choose what to do/ not to do while using the courseware.
6.2 Learners have their control over their own pace.
7. Interactivity The interaction mechanism and activities enabling learners to interact with one another or with the courseware.
7.2 Learners have opportunities to interact with other learners.
8. Learners’ profile
The profile of a learner in using the courseware, including the learner’s background information, completed lessons, grades/performance of each lesson, etc.
8.3 Learners are able to find where they have been (i.e., a learning profile is offered) and where they are heading during the course.
II. Technical Criteria Subcategory Definition Sample Evaluation Item
1. Instructions The instructions presenting either in general or in specific lesson/activity to guide users.
1.1 The instructions on how to use the courseware are easy for learners to follow.
2. Interfaces The overall interfaces of the courseware, such as the icons and symbols.
2.2 The icons/symbols are intuitive and easy to understand.
3. Layout The overall physical presentation of the courseware, such as colors, word fonts, and spacing.
3.1 The overall presentation (colors, font, spacing, etc) is able to interest learners.
4. Accessibility The easiness and convenience of the course when used.
4.1 It is easy to install the courseware.
5. Support The technical support from the courseware company.
5.1 The technical support (e.g., FAQ, online support) is sufficient.
55
4.2.3 Summary of Stage I
As mentioned in Chapter Two, the need exists for a new means for evaluating CALL
tools─a new genre of learning materials. Recent literature has indicated the lack of the
inclusion of instructional concerns in most of the existing evaluation checklists. Thus, the
current study attempted to construct an evaluation checklist with a specific focus on learner-
centeredness. Of the 39 evaluation criteria, 32 were pedagogical criteria and 7 technical ones,
indicating two relevant points. First, the checklist emphasizes pedagogical aspects more than
technical issues. Second, the research results from Stage I revealed that most criteria— mainly
technical ones— that current evaluation checklists have adopted are not actually related to the
concept of learner-centeredness. For example, one criterion said, “The hypermedia
courseware includes information regarding how often is updated” and “The hypermedia
courseware includes information regarding its latest update” (Elissavet & Economides, 2003).
Such evaluation questions are not inadequate; however, they are irrelevant to learner-
centeredness.
Therefore, the 39 evaluation criteria in the evaluation checklist address the first
research question of this study. It is worth noting that the constructed evaluation checklist
differs from most existing evaluation checklists for CALL materials in that this checklist has
a clear purpose: attempting to raise English teachers’ awareness of learners/users and their
concerns.
4.3 Stage II: Application of the Evaluation Checklist: Evaluation of MyET
During Stage II, the learner-centered evaluation checklist was used to evaluate MyET.
Generally speaking, the evaluators held a positive attitude toward MyET. The student-
evaluators expressed higher appreciation of the courseware (mean score of 2.23 out of 3) than
the teacher-evaluators (mean score of 1.66 out of 3). This section discusses these findings
along with teacher-evaluators’ and student-evaluators’ opinions of MyET. Although initially
56
meant to be an example evaluation, this stage generated much data, and analyzing this
information provided two concrete benefits. First, the evaluation results showed the
usefulness of the checklist in some way; second, the results demonstrated MyET’s degree of
learner-centeredness, at least in an initial way.
4.3.1 Teacher-evaluators’ Evaluation Results
The total of the three evaluators’ responses for each evaluation criterion ranged from
8 (highest) to 1 (lowest), with a mean score of 1.66 out of 3. A total score above 7 is
considered high. A score of 6 might involve the scores 3, 2, and 1, in which 1 is regarded as a
low score; 7 might include 3, 3, and 1. However, a score of 8 represents 3, 3, and 2, in which
all three teacher-evaluators gave high scores. Based on the results, the following discussion
will explore the strengths and weaknesses of MyET as well as non-scored evaluation criteria.
(Note: References to the checklist evaluation criteria are provided; for example, “I, 1.1”
refers to item 1.1 in the pedagogical category and “II, 1.1” refers to item 1.1 in the technical
category.) Meanwhile, the findings obtained from the first interview during Stage II (see
section 3.3.3) regarding the teaching background of the teacher-evaluators and their concepts
about learner-centeredness will be addressed in section 4.4.
Strengths of the Courseware
Looking at the results by evaluation criterion, 6 of the 39 evaluation criteria received
the highest score (8) while another 6 attained the second highest score (7).
Table 4.5 displays the high-scoring evaluation criteria, highlighting the strengths of
the courseware examined. All of the six evaluation criteria were addressed by all three
teacher-evaluators.
57
Table 4.5
The high-scoring evaluation criteria from the teacher-evaluators’ evaluation results
Item No. Item Description Total*
I, 2.3 The guidance and models of learning activities are clear for learners. 8
I, 6.1 Learners are allowed to choose the learning activities that they want to use. 8
I, 6.2 Learners have their control over their own pace. 8
I, 6.3 Learners are able to learn the course independently outside the classroom. 8
I, 8.2 The course is able to help arouse learners’ awareness of their own learning
progress/status.
8
II, 2.2 The icons/symbols are intuitive and easy to understand. 8
Note: * Here the total means the total of the three evaluators’ responses.
According to the teachers’ evaluations, MyET’s Studio Classroom guidance and
learning activities models are clear (I, 2.3). In addition, all three criteria related to freedom (I,
6.1, 6.2, and 6.3) received the highest score, indicating consensus in regards to the
courseware’s sufficient user control of the learning process. The teacher-evaluators also
determined that MyET is able to help learners become aware of their learning status and
progress (I, 8.2). Finally, they found the icons/symbols to be intuitive and easy to understand
(II, 2.2).
Examining the findings category by category, freedom, interfaces, instructions, and
layout earned the highest averages. Interestingly, three of the five technical subcategories
received high marks, while only one of the eight pedagogical subcategories did so, which
seems to indicate that MyET’s technical features outperformed its pedagogical facets. This
finding echoes a phenomenon discussed in many current studies: most existing commercial
courseware emphasizes technical characteristics more than pedagogical considerations (Neri,
Cucchiarini, Strik, & Boves, 2002). Another possible reason for higher number of technical
criteria receiving high marks is that technical criteria may be easier to judge since they are
relatively visible, while pedagogical ones are more abstract.
58
Weaknesses of the Courseware
Looking at the results by each evaluation criterion, several evaluation criteria scored
relatively low, with only one or two points. However, a few of these low-scoring items were
not scored by all three evaluators, which in turn lowered the overall scores for the items. The
teacher-evaluators’ explanations for this (identified during interviews) will be addressed in
the next section. Two of the low-scoring items were answered by all three teacher-evaluators
and, as such, are indeed low-scoring evaluation items that reveal clear weaknesses of MyET.
Table 4.6 shows the low-scoring evaluation items answered by all teacher-evaluators. Table
4.7 lists the low-scoring evaluation items to which not all teacher-evaluators responded.
These imply possible weaknesses of MyET.
The results indicate that the evaluators did not agree that the courseware serves
certain learner-centered features— namely: 1) the content helps learners build up strategic
ability (I, 3.6) and 2) there is a variety of learning activities for learners with multiple
intelligences, e.g., interpersonal, mathematics… (I, 2.5). Other low-scoring items included, for
example, the content accommodates learners with various learning styles (I, 3.5), the course
offers feedback that can be saved and accessed easily so that learners may realize what
difficulties they have in the learning process (I, 5.4), the learning profile is easy to access (I,
8.4), and the course offers an entry requirement verification device to see if the learner meets
the entry conditions (I, 8.1).
59
Table 4.6
The low-scoring evaluation items from the teacher-evaluators’ evaluation results (responded to
by all evaluators)
Item No. Item Description Total*
I, 3.6 The content helps learners build up strategic ability. 2
I, 2.5 There is a variety of learning activities for learners with multiple intelligences
(e.g., interpersonal, mathematics… ).
3
Note: * Here the total means the total of the three evaluators’ responses
Table 4.7
The low-scoring evaluation items from the teacher-evaluators’ evaluation results (responded to
by only one or two evaluators)
Item No. Item Description Total*
I, 3.5 The content accommodates learners with various learning styles. 1
I, 5.4 The course offers feedback that can be saved and accessed easily so that
learners may realize what difficulties they have in the learning process.
1
I, 8.4 The learning profile is easy to access. 1
I, 8.1 The course offers an entry requirement verification device to see if the learner
meets the entry conditions.
2
II, 5.1 The technical support offered (e.g., FAQ, online support) is sufficient. 2
I, 4.1 The course offers skill-training that is able to meet learners’ needs. 3
I, 5.2 The course offers feedback that is able to motivate learners. 3
Note: * Here the total means the total of the three evaluators’ responses, reflecting one or two
evaluators’ non-responses.
60
When looking at the results by category, support received the lowest score (2) while
skills scored only 3. These figures may indicate that the courseware does not provide
sufficient technical support— e.g., FAQ and online support (II, 5.1)— and the skill-training
offered by the courseware may not be able to meet learners’ needs (I, 4.1). In fact, although
MyET increased its various learning courses, including reading and writing programs, it
continues to emphasize speaking skills, particular pronunciation; its learning activities are
based primarily on the Automatic Speech Analysis System (ASAS), which focuses on
improving learners’ pronunciation skills. MyET’s Chinese commercial refers to the
courseware as “My Oral-training English Tutor” (“我的口說英語家教” in Chinese). Thus, it is
not surprising that the teacher-evaluators did not think this courseware offered skill-training
that was able to meet learners’ various needs, such as reading and writing.
Non-scored Evaluation Criteria
Several evaluation items that received quite low scores do not necessarily point to
shortcomings of the courseware since they were not responded to by every teacher-evaluator.
Teacher-evaluator A did not score 14 evaluation items, and Teacher-evaluator C did not
respond to 7 items; Teacher-evaluator B answered all 39 evaluation questions. Teacher-
evaluator A wrote “unknown” or “not applicable” in several items, and Teacher-evaluator C
made question marks when she did not mark a score. Non-scored answers do not equate to
zero in this sense.
These evaluators were not able to score the items for several reasons. First, the
evaluation description was not applicable to the evaluand. For example, Teacher-evaluator A
did not think the feedback offered by the courseware could be saved and accessed easily (I,
5.4), so he marked “not applicable” in the comment column instead of checking the score 0,
indicating “very unlikely”. Second, when the evaluators perceived that a criterion was only
partially met, they tended not to score the criterion, indicating “unknown” in the comment
61
column. For instance, for the evaluation item “the content matches learners’ English
proficiency level,” Teacher-evaluator A’s idea was partially yes as he thought the English
proficiency level of his students varied, so he decided not to score it, writing “some yes, some
no” in the comment column instead. When asked the reasons for the uncertainty, Teacher-
evaluator A replied that he had kept his students, who had been using MyET, in mind while
doing the evaluation. With a specific target group of students in mind, he found it difficult to
make judgments unless he had already seen evidence of the effectiveness within the students’
skills. In other words, he did not try to predict the effectiveness of the courseware, but instead
adopted a retrospective view to conduct the evaluation (see section 2.3.1 of the literature
review). Therefore, he could not score quite a few of the evaluation questions.
Another reason for the lack of a score might be unfamiliarity with the courseware.
Although all of the teacher-evaluators had used MyET, they might not have had the chance to
try all of its functions and interfaces; thus, they could not respond to the checklist items. For
example, Teacher-evaluator C had no idea whether MyET offered an entry requirement
verification device or a learner’s file; she indicated this with a question mark in the comment
column and explained this during an interview.
Finally, the evaluators did not give a score because they interpreted the criteria
descriptions differently. Teacher-evaluator C, for instance, responded to item I, 7.1 “Learners
have opportunities to interact with the courseware” by writing her comment that “Isn’t it
interaction to use the courseware?” (“使用本身不就是互動了嗎?” in Chinese.) However, the
interactivity here refers to the interactive opportunities the courseware offers. The way that
the teacher-evaluator interpreted the criteria influenced her judgment; thus she could not
respond to this evaluation question.
These concerns led to a lack of responses, creating complexity in analyzing and
interpreting the research data. Contextual variables affect the use of a checklist. In fact,
62
unable-to-answer or not-sure phenomena are common if the checklist is not tailor-made for a
specific evaluation (Bradin, 1999; Le & Le, 1997). As the aim of the current study is to
explore the learner-centeredness in online CALL tools and develop a learner-centeredness-
based checklist for online courseware evaluation, Stage II served as an application of the
checklist developed, giving all evaluators the actual experience of using the checklist for the
subsequent meta-evaluation of its effectiveness. It should be reiterated that the statistical
results of the MyET evaluation in this study were not intended to be the main findings, nor to
be generalized.
Controversial Issues
All three teacher-evaluators agreed to a great extent on their evaluations. However,
disagreements occurred among a few evaluation criteria (i.e., I, 1.1; I, 3.7; I, 3.8; I, 5.5; I, 8.3;
II, 4.2). For example, two evaluators thought the courseware offered sufficient information on
learning objectives, while the third disagreed (I, 1.1). In addition, when responding to
whether the courseware provides an opportunity for learners to give personal feedback on
their learning experience (I, 5.5) and whether the courseware provides learners a learning
profile (I, 8.3), Teacher-evaluator A held a positive attitude whereas Teacher-evaluator B
disagreed; Teacher-evaluator C did not answer these two questions at all. Still, the findings
indicate that most of the time these evaluators shared mutual points of view on MyET, which
might imply the effectiveness of the checklist, which will be discussed in section 4.4.
4.3.2 Student-evaluators’ Evaluation Results
In addition to the 39 evaluation criteria, these student-evaluators were asked to
comment on MyET in general─“In general, this courseware is helpful for learners; namely, it
is able to meet learners’ needs or enhance English learning” (see item No. 40 in Appendix II).
Ten out of the total 22 student-evaluators thought that MyET was likely helpful for their
learning, and another 10 agreed that MyET was very likely helpful.
63
Generally speaking, the student-evaluators held a positive view of MyET. With 22
student-evaluators involved in this evaluation, the total scores should range from the two
extremes— 0 to 66— making any score below 33 a negative perception. Based on the
statistical results from student-evaluators’ evaluations, the total scores for each evaluation
criterion range from 55 (highest) to 43 (lowest), with a mean score of 2.23 out of 3. The fact
that the lowest score was 43 indicates the students’ positive attitude. (Note: item No. 40 was
not included in the calculation in order to be able to compare the student-evaluators’ mean
score with the mean score calculated from the teacher-evaluators’ evaluation results.)
Strengths of the Courseware
Table 4.8 highlights examples of the high-scoring evaluation items, which are
considered the strengths of the evaluand. For instance, the student-evaluators thought the
content of MyET’s Let’s Talk in English was related to their life experiences. Looking at the
results category by category, freedom received the highest average (52)— similar to the
teacher-evaluators’ evaluation— and learning objectives the second highest (51). This result
shows that most of the student-evaluators were satisfied with the freedom that MyET
provides learners.
64
Table 4.8
The high-scoring evaluation items from the student-evaluators’ evaluation results
Item No. Item Description Total*
I, 3.2 The content is related to learners’ life experience. 55
I, 1.1 The information on the objectives is sufficient for learners. 53
I, 1.2 The objectives of the course are associated with learners’ learning development. 53
I, 6.1 Learners are allowed to choose the learning activities that they want to use. 53
I, 6.3 Learners are able to learn the course independently outside the classroom. 52
II, 2.1 The interfaces are easy for learners to use. 52
Note: * Here the total means the total of the 22 student-evaluators’ responses
Weaknesses of the Courseware
No evaluation items scored below 33, which seems to indicate that there were no low-
scoring evaluation items from the student-evaluators’ evaluation results. Table 4.9 lists a few
examples of the lowest scores according to the student-evaluators. These results may not
actually indicate deficiencies in MyET, but might point to possible weaknesses. For example,
MyET does not provide a wide variety of activities for users with different multiple
intelligences (I, 2.5). Moreover, among the 13 subcategories, support had the lowest average
(45). This finding not only reveals a difference between the two groups of evaluators, but also
exposes a phenomenon that, as the primary user, students expect more sufficient support (e.g.,
online support) when using the courseware. Nonetheless, although these criteria received
lower scores, when divided by the total number of student-evaluators, the average (2)
indicates that the students agreed that MyET likely conforms to the criteria. On the other
hand, the subcategory with the highest average is freedom (with the average 52). Interestingly,
the teacher-evaluators’ evaluation results show the same finding.
65
Table 4.9
The low-scoring evaluation items from the student -evaluators’ evaluation results
Item No. Item Description Total*
I, 2.5 There is a variety of learning activities for learners with multiple intelligences
(e.g., interpersonal, mathematics… ).
43
I, 7.2 Learners have opportunities to interact with other learners. 44
II, 5.1 The technical support offered (e.g., FAQ, online support) is sufficient. 45
I, 3.4 The content is culturally appropriate for learners. 46
I, 3.6 The content helps learners build up strategic ability. 46
I, 5.2 The course offers feedback that is able to motivate learners. 46
I, 5.5 The course offers an opportunity for learners to give personal feedback on their
learning experience.
46
II, 4.1 It is easy to install the courseware. 46
Note: * Here the total means the total of the 22 student-evaluators’ responses
Non-scored Evaluation Criteria
Only one student among the 22 left items unanswered (a total of three unanswered
questions); the others answered all the evaluation questions. It seems that students tended to
answer all the questions they encountered.
4.3.3 Caveats about Stage II
As mentioned earlier, due to the small sample sizes, the statistical results presented
here are not aimed to be generalized and to decide the value of the evaluand; instead, the
purpose of this stage of the study is to give the participants experience in using the checklist
as well as to gather insight as to what actual learners think and need when using courseware
to learn English, since all of the participants had used the evaluand. Another interesting
finding worth mentioning is that among both groups of evaluators, few evaluators gave any
zeros when responding to the evaluation checklist. Teacher-evaluator A gave one zero, and
Teacher-evaluator C gave two; two student-evaluators gave one zero each. Whether this
phenomenon shows the courtesy of Taiwanese when they make judgments or suggests the
66
value of MyET is an issue for further study. Finally, all data were collected in January 2007
based on these evaluators’ previous experiences using the courseware. MyET may have
subsequently issued new editions.
4.3.4 Summary of Stage II
In sum, the total 25 evaluators regarded MyET in a positive light. Although the
evaluators examined two different MyET’s courses, Studio Classroom Standard Course and
Let’s Talk in English, both courses used the same platform and interfaces. Thus, the
evaluation results still present the advantages and disadvantages of the courseware itself.
Teacher-evaluators tended to examine the courseware from a thorough and critical point of
view, resulting in a lower mean (1.66) than the student-evaluators’ (2.23). Contextual issues,
such as professional background (teachers versus students) and evaluation processing time
frame (a week versus an hour), could have caused such differences. Moreover, any evaluation
process embraces subjective judgments; for example, the interpretation of “likely” in “the
course/courseware likely conforms to the description” might be different for different
evaluators.
The results and discussions of Stage II aimed to display the results from the
evaluations of MyET and are expected to provide insights for English teachers, institution
administrators (stakeholders) who may need to make decisions as to what courseware to use
in a language lab or a self- learning center, or even online courseware developers.
4.4 Stage III: Meta-evaluation of the Learner-centered Evaluation Checklist
To answer the second research question, Stage III gathered feedback on the evaluation
checklist through interviews and open-ended questions. The first interview with each teacher-
evaluator during the previous stage (see section 3.3.3) focused on their teaching background
and concepts about learner-centeredness. Two of the teachers had been teaching for
approximately ten years, while the other one for six years─one year in a college setting. The
67
second interview during Stage III, aiming to gather their feedback in regards to the evaluation
checklist, was conducted after they had used the checklist to evaluate MyET. Interview
question four─“Have you used any other evaluation checklist for ESL/EFL materials?”─
aimed to determine these teachers’ experiences using an evaluation checklist for ESL/EFL
materials. Two of them had used several such checklists previously; however, these two
teachers mentioned that the evaluation checklists they had used were for textbooks, not for
courseware. One teacher had never used such an evaluation tool.
Feedback on the checklist from teacher-evaluators and student-evaluators is discussed
in the following two sections.
4.4.1 Teacher-evaluators’ Feedback
Seven primary interview questions were asked (see Table 3.2), along with a few
follow-up questions when necessary during each interview (see Appendix VI for the
interview transcripts with English translations). References to the transcripts are marked in
parentheses, for example, “(A, 100 & 105)”, referring to Teacher-evaluator A’s comments
presented in line number 100─responses in Chinese─and line number 105─English
translation. Categorizing the responses to the interview questions, five aspects are discussed
as follows.
Overall Opinions toward the Evaluation Checklist
Based on the findings generated from interview questions one, six, and seven,
generally speaking, the three teacher-evaluators thought the checklist would be helpful for
teachers in evaluating courseware’s learner-centeredness ( A, 276-277, 280 & 281-282, 286;
B, 343 & 345; C, 475 & 478). All agreed that the developed evaluation checklist focused
teachers’ attention on learners. One felt most teachers were influenced by technical issues
when evaluating software; what attracted teachers might simply be the beauty of the
software’s design and multimedia. When asked whether the checklist would be helpful for
68
English teachers in evaluating and selecting courseware, Teacher-evaluator A replied, “Sure!
Why not? I think this checklist would help teachers see what they should pay attention to
when it comes to selecting courseware” (A, 276-277 & 281-282). He also agreed that this
checklist would focus teachers’ awareness on the need to include language
acquisition/learning issues during the evaluation (A, 279-280 & 284-286). Another evaluator
found the checklist “not bad” (B, 291 & 292). “If I have to grade it, I’ll give it a least an 85.
The questionnaire is not bad,” answered Teacher-evaluator B (B, 296 & 297). He thought the
checklist was helpful for teachers in that it offered quite a few items that most teachers might
overlook (B, 343-344 & 345-346). When asked whether the checklist was helpful for teachers
in evaluating courseware, the third teacher-evaluator replied, “I think so. With this checklist
you could be more objective. And it reminds you of some things that you’ve never thought
of ” (C, 475-476 & 478-479). She also mentioned that the checklist was able to help avoid
selecting software “by first impressions” since it provided different aspects to consider (C,
476-477 & 479-480).
It might be argued that the three teacher-evaluators gave positive comments out of
politeness. Nonetheless, since all indicated a similar attitude toward the checklist to a great
extent, it can be concluded that the learner-centered evaluation checklist serves as a good
evaluation tool for online courseware. In addition, one way to validate findings is to look for
negative evidence (Miles & Huberman, 1994). These teacher-evaluators did actually point out
the evaluation checklist’s deficiencies (see Weaknesses of the Evaluation Checklist below),
revealing that their positive responses in general were not simply made out of courtesy.
Strengths of the Evaluation Checklist
In response to interview question five, the three teacher-evaluators identified several
strengths of the checklist. First, the checklist’s understandable categorization enhances
understanding of the focus. “The most helpful part of the checklist is the categories,” one
69
teacher-evaluator said (A, 237 & 240). He further mentioned, “I should focus on my student
when I try to help him/her to learn. But do I notice some relevant points? I like the
questionnaire because it reminds me of some points” (A, 238-240 & 241-244). Second, it has
sufficient coverage of crucial evaluation facets for online CALL tools. “At least it gives
teachers quite a few criteria for evaluation; teachers may not think of so many themselves” (B,
343-344 & 345-346). According to the third teacher-evaluator, “it covers quite a lot of aspects,
allowing teachers to evaluate courseware from various aspects” (C, 448-449 & 450-452).
Finally, the checklist involves certain pedagogical concerns that might be easily ignored by a
teacher when evaluating courseware; for example, one teacher-evaluator mentioned that her
attention was drawn to the learner— her students— when she saw the evaluation criterion
concerning learning objectives. This teacher-evaluator started to think about the reason for
having her students use the courseware and whether students could accomplish something at
the end. This particular evaluation item reminded this teacher-evaluator to take her students’
uniqueness into account (C, 437 & 440-442).
Weaknesses of the Evaluation Checklist
On the other hand, these teacher-evaluators pointed out several weaknesses or defects
in the checklist when replying to interview question five. First, a few evaluation items were
not easy to answer for two of the evaluators. One possible explanation is that one teacher-
evaluator always kept his students in mind during the evaluation, adopting a retrospective
perspective in evaluating this courseware— in other words, if he could not decide his
students’ potential performance, he chose “unknown.” “The reason I can’t answer this
question is because MyET focuses on pronunciation, and we haven’t finished the course yet.
So I can’t answer it; I wrote ‘unknown’” (A, 183-185 & 189-191). As mentioned earlier, it is
a common phenomenon that in an evaluation process certain items are not easy to respond to
since the evaluation context varies. Bradin (1999) suggests that it is not necessary to finish a
70
checklist from the beginning to the end; wherever there are primary items that do not fit,
there is no need to continue the evaluation. Another possibility is the degree of familiarity
with this courseware. If the teacher-evaluator had not had a chance to test all functions of the
courseware, some evaluation criteria were difficult to rate. For example, Teacher-evaluator C
wrote several question marks due to the fact that she had not used or known of certain MyET
function, such as “learners’ profile.”
Second, the translation of two terms between the English and Chinese was not
consistent or clear. Two of the teacher-evaluators pointed out the deficiency. One problem
was the heading of the subcategory freedom (translated as “自由度” in Chinese). According to
Teacher-evaluator B, the Chinese translation does not interpret the real meanings of the
evaluation criteria under the subcategory, including students’ control over the learning
activities and learning pace (B, 315-318 & 323-326). Another error of translation is that
cognitive development should have been translated as “認知發展” instead of “智能發展” (B,
314-315 & 321-322; C, 384-386 & 388-390). This checklist was first developed in English,
and the subsequent mistakes occurred when it was translated into Chinese; more caution is
required in the translation process.
Suggestions for Further Revisions
With regard to further additions to and deletions from the evaluation checklist, none
of the three teacher-evaluators proposed any suggestions except that one expected to see
more items regarding SLA on an evaluation checklist for CALL tools (A, 258-259 & 261-
263). As for further suggestions for revision, two teacher-evaluators suggested that the
subcategory interactivity be defined first and include specific types of interactivity since it
might be confusing to evaluators (B, 351-359 & 361-372; C, 400-403 & 417-420). Different
types of interactivity exist, such as interaction between the user and the courseware and
interaction between/among users. Finally, one teacher-evaluator suggested separating the four
71
skills (listening, speaking, reading, and writing) under the subcategory skills since different
courseware may focus on different skills (C, 397-399 & 413-417). For example, MyET’s
focus is primarily on pronunciation skills. Revisions to the checklist based on these
suggestions are discussed in Chapter Five.
Teacher-evaluators’ Concepts of Learner-centeredness
Before using the checklist, teacher-evaluators were asked about their concepts of
learner-centeredness (see section 3.3.3 about the first interview). All three expressed similar
viewpoints. Teacher-evaluator A thought that learner-centeredness meant “wherever the
learner is, what he or she is learning must be relevant to himself or herself. It has to be
meaningful; if not, it will be that teachers teach whatever they feel like teaching and students
learn whatever they feel like learning” (A, 76-78 & 91-95). Teacher-evaluator B stated,
“Learner-centeredness means that the focus is on the students when it comes to learning or
preparing teaching activities, right? On one hand, as a teacher, you acknowledge the needs of
the students and try to satisfy them. On the other hand, students are placed in the center of
learning” (B, 128-130 & 138-141). In Teacher-evaluator C’s opinion, learner-centeredness is
“the design as well as the process of a curriculum has the learner as the focus. The teacher,
here, plays a role like an assistant” (C, 168 & 172-173). Although the teacher-evaluators’
descriptions seemed broad, they touched on the characteristics of learner-centeredness to
some degree. It is interesting that two of the teacher-evaluators commented that not all
learners understood what they needed in terms of learning, particularly those with lower
proficiency levels (A, 78-80 & 95-98; B, 130-136 & 141-148), implying that a teacher still
needed to intervene in learning. This idea reveals the reality of the English learning context in
Taiwan.
The teacher-evaluators’ thinking and attitude toward learner-centeredness might have
affected their evaluations. After using the checklist to evaluate MyET, when answering
72
whether the checklist helped focus the attention to learner-centeredness, Teacher evaluator A
replied, “It should!” since his curriculum design for his class focused on self- learning (A,
230-231 & 233-234). Teacher-evaluator C also agreed that the checklist was able to draw her
attention to learner-centeredness; she mentioned her experience of using the checklist, “You
mention a few questions concerning objectives in teaching, which leads me to think why I
would want my students to use such software and think if they can achieve the objectives.
That’s what the checklist did for me” (C, 437-439 & 440-442). As for Teacher-evaluator B, he
did not describe his own experience; however, he thought the checklist “should be a little bit”
able to arouse teachers’ awareness of learner-centeredness (B, 302 & 306). To sum up, based
on these teachers’ responses, it can be concluded that the use of the evaluation checklist
increased their awareness of learner-centeredness in evaluating courseware.
4.4.2 Student-evaluators’ Appreciation
Three open-ended questionnaire questions on the student-evaluators’ version of the
checklist were included to elicit feedback on its use (see section 3.4.3). Their responses were
coded as positive or negative, categorized and counted the number of responses (see section
3.4.4 about the categorization process). The results are shown in Table 4.10 (translated from
Chinese; see Appendix V for the original comments in Chinese).
73
Table 4.10
Frequency of student-evaluators’ feedback on the learner-centered evaluation checklist
Question 1: Generally speaking, do you find these evaluation items easy to answer? Why/Why not?
Response Frequency Percentage
Positive 4 18%
Fairly Positive 8 36%
Partially Positive 1 5%
Negative 3 14%
No Response 5 23%
Responses that do not fit the question 1 5%
Total 22 101%*
Question 2: Do you think this evaluation checklist is able to focus the attention on learners’ needs
and conditions?
Response Frequency Percentage
Positive 10 45%
Partially Positive 1 5%
Negative 1 5%
No Response 9 41%
Responses that do not fit the question 1 5%
Total 22 101%*
Question 3: In addition to these evaluation items, what concerns should be taken into account when
a teacher is evaluating and selecting suitable learning materials to meet learners’ needs?
Response Frequency Percentage
Learners’ proficiency levels 2 9%
Learners’ situations of learning and using of the courseware 3 14%
Learners’ Freedom 2 9%
Content of the courseware (interesting) 2 9%
No further comments (enough items.) 4 18%
No response 7 32%
Responses that do not fit the question 2 9%
Total 22 100%
Note:*Approximate value exceeds 100 because of rounding up
74
Overall Opinions toward the Evaluation Checklist
In general, the student-evaluators thought the questions were easy to answer since the
descriptions of the evaluation criteria were presented in a straightforward manner. They also
agreed that the evaluation checklist was helpful for focusing attention on learners’ needs and
situations when evaluating learning courseware. The following sections discuss the student-
evaluators’ positive and negative comments in regards to the checklist.
Positivity about the Evaluation Checklist
The findings indicate the student-evaluators’ positivity about the checklist as follows.
First, four of the 22 student-evaluators rated the checklist as easy to answer; another 8 found
it fairly easy to answer. Second, nearly 50 percent of the student-evaluators agreed that the
checklist was able to focus their attention on learners’ needs and conditions in evaluating
courseware. Although these students held a relatively positive attitude toward the checklist,
three student-evaluators thought it included too many questions, some of which were too long
or vague. This finding confirms what had been suspected: students prefer a short, concise
evaluation checklist, while teachers prefer a comprehensive one. This outcome proves that
the decision to modify the original six-page bilingual checklist into a two-page Chinese
checklist was sensible (see section 3.2.4).
Negativity about the Evaluation Checklist
The findings reveal the student-evaluators’ negative attitude toward the checklist as
follows. First, three student-evaluators rated the checklist as not easy to answer, citing the
vagueness of the evaluation item descriptions, experience with MyET (they had not being
using it for long), and the coverage of MyET functions (they had not tried all functions).
Second, one student-evaluator disagreed that the checklist functioned as it claimed since the
checklist did not reveal the reality of learners’ learning situations. In addition, a certain
percentage indicated “no response”; five did not give feedback on question one, and nine did
75
not on question two (see Table 4.10).
Additional Comments
When asked what concerns, in addition to the 39 criteria on the checklist, should be
taken into account in evaluating courseware for learners, several student-evaluators proposed
certain evaluation criteria to this learner-centeredness-based checklist. Two mentioned
learners’ proficiency levels and learning abilities, and another three listed learners’ learning
status and use of the courseware. One stated that the option to skip difficult questions in the
courseware is an important item, while another pointed out that the inability to install
courseware on students’ personal computers reduces learning opportunities; these two ideas
conform to the concept of freedom of the courseware— the degree to which a user can control
the process when using the learning tool. In addition, two student-evaluators expected the
courseware to be interesting and to provide more games; such responses hint at students’
needs and expectations for a more learner-centered learning tool.
Four irrelevant comments in total were made by the student-evaluators, revealing their
inability to understand and answer the questions. For example, two student-evaluators
responded “the repeating and recording part” (“跟著讀的部份” in Chinese) to the question: “In
addition to these evaluation items, what concerns should be taken into account when a
teacher is evaluating and selecting suitable learning materials to meet learners’ needs?” The
response seemed irrelevant or unclear.
4.4.3 Summary of Stage III
Based on all 25 evaluators’ feedback on the evaluation checklist, the second research
question has been answered; this checklist can indeed assist in focusing teachers’ (and
students’) attention on learners’ needs in evaluating courseware, although certain revisions
are required. The findings also highlight that a need does indeed exist for a new approach to
courseware evaluation in which learners’ needs are taken into serious consideration.
76
4.5 Summary
This chapter displays and discusses the findings of the three stages of this study as
well as answers the two research questions proposed in Chapter Two: an evaluation checklist
based on the premises of learner-centeredness was developed, and its effectiveness of
enhancing teachers’ awareness of learner-centeredness was also affirmed. The following
chapter will give conclusions.
77
CHAPTER FIVE
Conclusions
This final chapter draws conclusions in regards to this study. In this chapter, a revised
learner-centered evaluation checklist is proposed based on the suggestions and comments
from the evaluators. Besides, the contribution of this study, its limitations, and suggestions
for further research will be presented.
5.1 Revised Learner-centered Evaluation Checklist
Based on the teacher-evaluators’ and student-evaluators’ suggestions and comments,
several revisions were made to the evaluation checklist (see Table 5.1 for all the revised items
and Appendix VII for the revised checklist in Chinese and English). First, in the pedagogical
part, the heading skill in Category Four has been changed to skill-training; in addition, the
four skills (listening, speaking, reading, and writing) were separated into their own items,
shown as item I, 4.1, I, 4.2, I, 4.3, and I, 4.4 (see Suggestions for Further Revisions in section
4.4.1). Second, in Category Six, the heading was changed from freedom to learner control
(“學習者的自主性” in Chinese), which may correspond more with its items (see Weaknesses of
the Evaluation Checklist in section 4.4.1). Learner control/user control allows learners to take
their own path through the courseware. Hoven (1999) argues that offering a certain degree of
control to learners is desirable in order to promote learner autonomy. Little (1991) also states
that the lack of choice and freedom may lead to the problem that learners are reluctant to take
responsibility for their own learning. Third, the translation error in item I, 3.8 has been
modified from the erroneous “智能發展” to the corrected “認知發展”, which is consistent with
the English term cognitive development. Finally, regarding Category Seven, two evaluators
suggested defining the term interactivity (see Suggestions for Further Revisions in section
4.4.1); however, item I, 7.1 and I, 7.2 have actually indicated the differences between “the
interaction between the learner and the courseware” and “the interaction between/among the
78
learner and other learners.” To avoid any misunderstanding, the revised checklist includes
examples in the descriptions of these items (see Table 4.2 for comparison).
79
Table 5.1
Revised learner-centered evaluation checklist for online CALL courseware
I. Pedagogical Criteria 1.1 The information on the objectives is sufficient for learners. 0 1 2 3
1.2 The objectives of the course are associated with learners’ learning development.
0 1 2 3
1.3 The objectives are achievable for learners. 0 1 2 3
1 Learning
Objectives
1.4 The objectives reflect what learners need to know. 0 1 2 3 2.1 The learning activities are able to arouse learners’ interests in participation.
0 1 2 3
2.2 The instructions of the learning activities are clear for learners. 0 1 2 3
2.3 The guidance and models of learning activities are clear for learners.
0 1 2 3
2.4 There is a variety of learning activities for learners with varied learning styles.
0 1 2 3
2 Learning Activities
2.5 There is a variety of learning activities for learners with multiple intelligences (e.g., interpersonal, mathematics… .).
0 1 2 3
3.1 The content matches learners’ English proficiency level. 0 1 2 3
3.2 The content is related to learners’ life experience. 0 1 2 3 3.3 The content covers a variety of topics that are able to attain learners’ interests.
0 1 2 3
3.4 The content is culturally appropriate for learners. 0 1 2 3 3.5 The content accommodates learners with various learning styles. 0 1 2 3
3.6 The content helps learners build up strategic ability. 0 1 2 3 3.7 The content stimulates recall of prior learning. 0 1 2 3
3 Content
3.8 The content enhances learners’ cognitive development. 0 1 2 3
4.1 The content offers listening training that is able to meet learners’ needs.
0 1 2 3
4.2 The content offers speaking training that is able to meet learners’ needs.
0 1 2 3
4.3 The content offers reading training that is able to meet learners’ needs.
0 1 2 3
4
Skill-training
4.4 The content offers writing training that is able to meet learners’ needs.
0 1 2 3
5.1 The course offers immediate and proper feedback on learners’ performance and/or responses.
0 1 2 3
5.2 The course offers feedback that is able to motivate learners. 0 1 2 3
5.3 The course offers meaningful and helpful error handling which is able to help learners improve.
0 1 2 3
5.4 The course offers feedback that can be saved and accessed easily so that learners may realize what difficulties they have in the learning process.
0 1 2 3
5 Feedback
5.5 The course offers an opportunity for learners to give personal feedback on their learning experience.
0 1 2 3
1
3
80
6.1 Learners are allowed to choose the learning activities that they want to use.
0 1 2 3
6.2 Learners have their control over their own pace. 0 1 2 3
6
Learner Control 6.3 Learners are able to learn the course independently outside the classroom.
0 1 2 3
7.1 Learners have opportunities to interact with the courseware (e.g., role-play activities for dialog practice).
0 1 2 3 7
Interactivity 7.2 Learners have opportunities to interact with other learners (e.g., online discussion forum).
0 1 2 3
8.1 The course offers an entry requirement verification device to see if the learner meets the entry conditions.
0 1 2 3
8.2 The course is able to help arouse learners’ awareness of their own learning progress/status.
0 1 2 3
8.3 Learners are able to find where they have been (i.e. a learning profile is offered) and where they are heading during the course.
0 1 2 3
8 Learners’
profile
8.4 The learning profile is easy to access. 0 1 2 3
II. Technical Criteria 1
Instructions 1.1 The instructions on how to use the courseware are easy for learners to follow.
0 1 2 3
2.1 The interfaces are easy for learners to use. 0 1 2 3 2 Interfaces 2.2 The icons/symbols are intuitive and easy to understand. 0 1 2 3
3 Layout
3.1 The overall presentation (colors, font, spacing, etc.) is able to interest learners.
0 1 2 3
4.1 It is easy to install the courseware. 0 1 2 3 4 Accessibility 4.2 It is easy to access the course to be learned. 0 1 2 3
5 Support
5.1 The technical support offered (e.g., FAQ, online support) is sufficient.
0 1 2 3
5.2 Contributions of the Study
5.2.1 Contribution to Theory
As mentioned in Chapter Two, a gap exists between pedagogical theories and CALL
courseware design. Most courseware developers emphasize the technical features instead of
taking instructional theories into consideration. As a result, when language teachers evaluate
and select a CALL tool, the beauty of the interface as well as the integration of audio and
video features might mislead teachers to believing that only colorful Flash files and cool
sound tracks can attract and motivate learners. Current literature on CALL evaluations shows
2
4
4
81
that, although this gap has been noticed among a great number of evaluation approaches and
tools, few studies have attempted to involve specific learning principles— in this case,
learner-centeredness— in a CALL-tool evaluation checklist. The concept of learner-
centeredness in second/foreign language teaching is not new, yet CALL researchers have only
recently started paying more attention to issues of user- friendliness and promoting user-
centered designs. This study sought to fill the research gap and contribute to CALL and
evaluation research by providing an example of the application of a key concept in teaching
English as second or other language (TESOL) to the evaluation of CALL courseware.
5.2.2 Contribution to Practice
Implications for English Teachers and Stakeholders of Institutions
Due to the characteristics of evaluation, which involves a judgment along with
evidence about the worth of the evaluand, evaluation is intended to “have immediate utility
for policy shaping and is expected to be influential in short-term decision making” (Rea-
Dickins & Germaine, 1998, p. 11). The developed learner-centered evaluation checklist
supplements existing evaluation checklists for English learning materials. The learner-
centered-focused checklist provides a series of criteria to follow for language teachers and
stakeholders who make decisions about the learning materials to be used in a classroom or a
self- learning center. This evaluation checklist should also serve to remind teachers and
stakeholders of learners’ individuality and uniqueness. When they evaluate and select
courseware, learners’ needs and expectations should be taken into account. Learning
efficiency will increase a great deal once the learner is placed in the right position— at the
core of the learning process.
Implications for Courseware Developers and Material Providers
It is believed that online courseware has its advantages in assisting language learning.
Numerous English material providers are working with courseware developers in an attempt
82
to provide their learning materials on the Internet. However, Susser and Robb (2004) point
out that online courseware does not mean moving the content in print to a web-based
environment; instead, an online learning tool should follow certain principles of learning and
pay attention to interactivity and authenticity, which can be some of the advantages of a
CALL tool. Applying the checklist to MyET as an example, the study found that the learning
activities need improvement; they should be more interactive rather than focused on
repetition and drill. The evaluand seems to follow the computer-as-pupil metaphor
(Warschauer & Kern, 2000) (see section 2.2). By embracing the principles of learner-
centeredness in the process of courseware evaluation, this checklist can direct courseware
designers and material providers, balancing the design process with a focus on the learner to
avoid the overemphasis on technical characteristics.
5.3 Limitations of the Study
Three limitations need to be acknowledged and addressed regarding the present study.
First, only one courseware was evaluated. Evaluating more than one courseware would have
enabled comparisons that would have assisted and strengthened the analysis of the
effectiveness of the developed evaluation checklist. Second, the samples are small— 3
teachers and 22 students— which caused challenges in dealing with the quantitative data from
Stage II. For example, when one of the three teacher-evaluators did not respond to 14 of 39
evaluation questions, determining how to deal with these non-scored questions was a
dilemma. Therefore, the results from Stage II regarding the evaluations of MyET cannot be
generalized in regards to the strengths and weaknesses of MyET. Finally, the researcher did
not interview the student-evaluators for additional feedback on the use of the checklist due to
practical constraints. The three open-ended questions were insufficient for exploring the
student-evaluators’ opinions in depth.
83
5.4 Suggestions for Further Research
Certain avenues for further research are suggested. First, the learner-centered
evaluation checklist should be applied to more courseware with greater variety to examine its
effectiveness. Second, more participants should be involved, either quantitatively (more
teachers and students) or qualitatively (interview more evaluators to obtain more in-depth
data). In addition to English teachers and students, if possible, stakeholders, courseware
developers, and material providers should be invited to participate in the study to obtain their
feedback and— hopefully— focus their attention on the learner-centeredness in CALL
material evaluation. Moreover, in addition to collecting criteria from existing checklists,
researchers might also attempt to create several criteria of their own based on the principles
of learner-centeredness.
Furthermore, it might be interesting to see the checklist being used in different
perspectives of evaluation, both predictive and retrospective. For instance, it could be used in
the selection process before specific courseware is chosen; it can also be used as a tool to
examine whether the courseware being used is learner-centered or not. Determining the
correlation between learners’ performance using learner-centered CALL materials might be
another possibility for further research. For example, two or more courseware packages could
be compared with one another to evaluate the differences in their degree of learner-
centeredness. The learners could take a pre-test on their language performance and then, after
using both courseware packages for a while, complete a post-test to investigate their
performance; meanwhile, a comparison analysis could be conducted to determine whether
relatively learner-centered courseware enhances learners’ learning effectiveness.
5.5 Conclusion
As an English teacher for 8 years as well as a language learner for 20 years, the
researcher had used MyET for one year before the study was initiated. During this time,
84
while benefiting from the courseware, the researcher came up with specific questions
regarding the courseware design, such as the Automatic Speech Analysis System and the
design of learning tasks. Investigating more studies in CALL and the idea of user-
centeredness more deeply, the researcher became interested in finding out more about the role
of the learner— the primary user of the courseware— in CALL tool development and
evaluation. The user has increasingly been considered an important part in the process of
courseware development. The researcher felt such significance should be considered during
the process of courseware evaluation.
The present study made an effort to develop an evaluation checklist with a specific
focus— learner-centeredness— to incorporate more pedagogical aspects. Based on the
findings of this study, it can be claimed that the established learner-centered evaluation
checklist did function as expected. Nonetheless, the creation of a set of criteria for the
evaluation of CALL courseware is not without limitations. When one area receives more
attention, another might be neglected. Thus, developers and evaluators should remember that
where an evaluation is conducted and where the courseware is used are actually context-
dependent and dynamic. The development of the checklist does not mean that it will be
appropriate in all areas of language instruction; hence, adoption and adaptation of an
evaluation framework or checklist should be flexible and dependent on the context in which
the evaluation is conducted.
It is noted that, although this checklist has a specific focal point on learner-
centeredness, the study does not imply that the evaluation checklist should be solely and
exclusive ly applied. Instead, the checklist can accompany other evaluation instruments as
long as the combination will lead to effective evaluations and benefit users and the quality of
learning. Maddux, Johnson, and Willis (1997) believe that educators should “take a flexible
approach to the use of forms” (p. 344). Any type of checklists may need to be modified to suit
85
the purposes of the setting in which they will be used.
To sum up, effective learning relies on the focus of the learning process being
shifted
from the teacher to the learner. A learner-centered approach views learners as individuals
with
individual needs and rights. Only when the learner’s experience, talents, needs, and
interests are appreciated can the learner’s intrinsic motivation in learning thus be triggered
and increased. Learner-centeredness should always be considered in both the process of
developing learning material and designing courseware, as well as the process of
evaluating and selecting courseware. This study expects to contribute to the field by
putting a missing puzzle piece— awareness of learner-centeredness— into the wider
picture of the CALL field.