COMPUTER SCIENCE EDUCATION RESEARCH AT THE CROSSROADS:
A METHODOLOGICAL REVIEW OF COMPUTER
SCIENCE EDUCATION RESEARCH: 2000-2005
by
Justus J. Randolph
A dissertation submitted in partial fulfillment
of the requirements for the degree
of
DOCTOR OF PHILOSOPHY
in
Education
(Research and Evaluation)
Approved
teve Lehman, PhD
Committee Member
Stepnen Clyd'
Committee Member
iC 1
Jihj/Dorward, PhD
Committee Member
Erkki Sutinen, PhD
Committee Member
Burnflam, EdD
ean of Graduate Studies
UTAH STATE UNIVERSITY
Logan, Utah
2007
This work is licensed under the Creative Commons Attribution-Noncommercial-No
Derivative Works 3.0 United States License. To view a copy of this license, visit
http://creativecommons.Org/licenses/by-nc-nd/3.0/us/ or send a letter to Creative
Commons, 171 Second Street, Suite 300, San Francisco, California, 94105, USA.
0©@©
Ill
ABSTRACT
Computer Science Education Research at the Crossroads:
A Methodological Review of Computer
Science Education Research: 2000-2005
by
Justus J. Randolph, Doctor of Philosophy
Utah State University, 2007
Major Professor: Dr. George Julnes
Department: Psychology
Methodological reviews have been used successfully to identify research trends
and improve research practice in a variety of academic fields. Although there have been
three methodological reviews of the emerging field of computer science education
research, they lacked reliability or generalizability. Therefore, because of the capacity for
a methodological review to improve practice in computer science education and because
the previous methodological reviews were lacking, in this dissertation, a large-scale,
reliable, and generalizable methodological review of the recent research on computer
science education was conducted. The purpose of this methodological review was to have
a valid and convincing basis on which to make recommendations for the improvement of
computer science education research and to promote informed dialogue about its practice.
IV
After taking a proportional stratified sample of 352 articles from a population of
1,306 computer science education articles published from 2000 to 2005, each article was
coded in terms of their general characteristics; report elements; research methodology;
research design; independent, dependent, and mediating/moderating variables examined;
and statistical practices. A second rater coded a reliability subsample of 53 articles so that
estimates of interrater reliability could be established.
The major findings were that (a) the majority of investigations were insufficiently
controlled to make generalized causal inferences, (b) there were no differences in the
methodological quality of articles published in journals or those published in conference
proceedings, and (c) there was a decreasing yearly trend in the number of articles that
only presented anecdotal evidence and the number of articles using explanatory
descriptive (e.g., qualitative) research methods. Also, (d) it was found that the region of
the first author's affiliation covaried with proportion of articles that reported on
experimental or quasi-experimental investigations, explanatory descriptive investigations,
and on proportion of articles in which attitudes were the sole dependent measure. In
addition, several differences in research practices across the fields of computer science
education, educational technology, and education research proper were found.
(341 pages)
ACKNOWLEDGMENTS
There are several people and groups that have been instrumental in helping me
overcome the hurdles involved with writing this dissertation. This dissertation writing
process has once again affirmed my belief in the notion that nobody wins alone.
I am indebted to Dr. George Julnes for trusting in me and my vision, even in the
face of adversity, and for his consistent support and late night insights. He, more than
anyone else, helped me make sense of computer science education research. I also have to
thank the other members of my committee — Dr. Stephen Clyde, Dr. Jim Dorward, Dr.
Steve Lehman, and Dr. Erkki Sutinen — for their thoughtful feedback. An extra note of
thanks goes out to Dr. Sutinen for his local support here in Joensuu. Also, thanks to
Karen Ranson for her diligent copyediting.
Also, I am indebted to several of my colleagues at the University of Joensuu: to
Roman Bednarik for being such a conscientious interrater reliability coder, to Matti Tedre
for his keen insights on the phenomenon called computing, to Elina Hartikainen for being
a daily sounding board for my ideas, and to Jorma Sajaniemi for his advice on which
variables to concentrate.
Thanks also to Dr. Karl White for his mentorship during my stay in Logan and for
being a good interlocutor, and, of course, to Dr. Susan Friedman for teaching me how to
become a better scientist and a better person. I also have to acknowledge the secretarial
wizardry of Shannon Johnson and Merja Hyttinen — both of whom went out of their way
to help with the countless issues that arose from doing my research abroad.
VI
My sincere gratitude goes out to the ACM SIGCSE for their generous financial
support for this methodological review. Also, I would like to express my appreciation to
the Fulbright Center for Finnish- American Academic Exchange for getting me to Finland
in the first place.
Last, I have to thank my family in the U.S. for their social, emotional, and
financial support and for always being there for me. And, of course, thanks to Kaisa
Ratilainen and Wilma Sue for being so pleasant to go home to after a long day of writing.
Justus Randolph
Joensuu, Finland
October, 2006
Vll
CONTENTS
Page
ABSTRACT iii
ACKNOWLEDGMENTS v
LIST OF TABLES viii
LIST OF FIGURE xii
FNTRODUCTION 1
METHOD 47
RESULTS 65
DISCUSSION 127
CONCLUSION 173
REFERENCES 178
APPENDICES 189
CURRICULUM VITAE 319
Vlll
LIST OF TABLES
Table Page
1 Evidence Table for Themes of the Literature Review 13
2 Research Questions in Educational Technology
Methodological Reviews 24
3 Characteristics of Educational Technology Reviews
Included in the Quantitative Synthesis 25
4 The Composition of Educational Technology Metacategories 26
5 Comparison of the Proportion of Human Participants Articles in
Educational Technology and Education Proper 29
6 Comparison of Type of Methods Used in Educational Technology
and Education Proper 30
7 Description of the Electronic Search for Previous
Methodological Reviews 35
8 Sampling Frame 49
9 Number of Articles Sampled from Each Forum and Year 49
10 Interrater Reliabilities for General Characteristics Variables 67
1 1 Interrater Reliabilities for Research Methods Variables 67
12 Interrater Reliabilities for Experimental Design Variables 67
13 Interrater Reliabilities for Independent Variables 68
14 Interrater Reliabilities for Type of Dependent Variable Measured 68
15 Interrater Reliabilities for Grade Level and Undergraduate Year 69
16 Interrater Reliabilities for Mediating or Moderating Variables 69
17 Interrater Reliabilities for Type of Effect Size Reported Variables .... 69
IX
Table Page
18 Interrater Reliabilities for Type of Measure Used Variables 70
19 Interrater Reliabilities for Type of Inferential Analyses Variables 70
20 Interrater Reliabilities for Report Element Variables 71
21 Institutions with Greater Number of Articles 74
22 Proportions of Report Elements 75
23 Proportions of Articles Falling into Each of Kinmunen's Categories . . 76
24 Proportions of Articles Falling into Each of Valentine's Categories ... 76
25 Proportion of Articles Dealing with Human Participants 76
26 Proportions of Grade Level of Participants 78
27 Proportion of Undergraduate Level of Computing Curriculum 78
28 Proportion of Human Participants Articles that Provide Anecdotal
Evidence Only 78
29 Proportions of Types of Articles Not Dealing With Human
Participants 79
30 Proportion of Methodology Types Used 80
31 Proportion of Types of Methods 80
32 Proportions of Types of Experimental/Quasi-Experimental
Designs Used 80
33 Proportion of Types of Independent Variables Used 82
34 Proportions of Types of Dependent Variables Measured 82
35 Proportions of Mediating or Moderating Variables Investigated 83
36 Proportions of Types of Measures Used 84
Table Page
37 Proportions of Types of Inferential Analyses Used 85
38 Proportions of Types of Effect Sizes Reported 86
39 Crosstabulation of Anecdotal-Only Papers in Conferences
and Journals 88
40 Crosstabulation of Experimental Papers in Conferences
and Journals 89
41 Crosstabulation of Explanatory Descriptive Papers in Conferences
and Journals 89
42 Crosstabulation of Attitudes-Only Papers in Conferences
and Journals 90
43 Crosstabulation of Experimental Papers That Used Posttest-Only
Designs Exclusively 90
44 Anecdotal-Only Papers by Year 91
45 Explanatory Descriptive Papers by Year 92
46 Experimental/Quasi-Experimental Papers by Year 92
47 One-Group Posttest-Only Papers by Year 92
48 Attitudes-Only Papers by Year 93
49 Experimental Papers by Region of First Author's Affiliation 94
50 Explanatory Descriptive Papers by Region of First Author's
Affiliation 95
51 Attitudes-only Papers by Region of First Author's Affiliation 95
52 Anecdotal-Only Articles by Region of First Author's Affiliation 96
53 One-Group Posttest-Only Papers by Region of First Author's
Affiliation 96
XI
Table Page
54 The Fit of Several Logistic Regression Models for
Anecdotal-Only Papers 98
55 Summary of Regression Analysis for Predictors of Anecdotal-Only
Articles, (N= 233) 100
56 The Fit of Several Regression Models for Experimental/
Quasi-Experimental Papers 104
57 Summary of Regression Analysis for Predictors Experimental/
Quasi-Experimental Articles (N= 144), With Outliers 106
58 Summary of Regression Analysis for Predictors of Experimental/
Quasi-Experimental Articles (N= 144), With Outliers and
Without Interaction Term 107
59 The Fit of Several Logistic Repression Models for Explanatory
Descriptive Papers 110
60 Summary of Regression Analysis for Predictors of Explanatory
Descriptive Articles, (N= 143) Ill
61 The Fit of Several Logistic Regression Models for Attitudes-Only
Papers 114
62 Summary of Regression Analysis for Predictors of Attitudes-Only
Articles (N= 123), With Outliers 116
63 Summary of Regression Analysis for Predictors of Attitudes-Only
Articles (N= 99), With Outliers Removed 116
64 The Fit of Several Logistic Regression Models for One-Group
Posttest-Only Papers 119
65 Summary of Regression Analysis for Predictors of One-Group
Posttest-Only Articles for Model With Interaction Term (N = 93) 121
66 Summary of Regression Analysis for Predictors of One-Group
Posttest-Only Articles for Model Without Interaction Term (N = 93) . . 123
Xll
Table Page
67 Comparison of the Proportion of Empirical, Human Participants
Articles in Computer Science Education and Education Proper 125
68 Comparison of the Proportion of Empirical, Human Participants
Articles in Computer Science Education and Education Technology . . 125
69 Comparison of the Proportion of Empirical, Human Participants
Articles in Computer Science Education and Education Proper 126
Xlll
LIST OF FIGURES
Figure Page
1 Proportion of types of articles in educational technology journals 27
2 Proportion of types of educational technology articles by forum 27
3 Proportions of types of educational technology articles by
time period 28
4 Proportions of articles published in each forum 72
5 Expected and observed probabilities for anecdotal-only papers 98
6 Anecdotal-only papers by combined region and year 103
7 Anecdotal-only papers by combined regions 103
8 Expected and observed probabilities for experimental/
quasi-experimental papers 105
9 Experimental/quasi-experimental papers by combined region
and form type 109
10 Experimental-quasi-experimental papers by combined region
and year 109
1 1 Expected and observed probabilities for explanatory descriptive
papers Ill
12 Explanatory descriptive papers by year and region 113
13 Explanatory descriptive papers by region 113
14 Expected and observed probabilities for attitudes-only papers 115
15 Attitudes-only papers by year and combined region 118
16 Attitudes-only papers by combined region 118
XIV
Figure Page
17 Expected and observed probabilities for one-group posttest-only
articles with interaction term 119
18 One-group posttest-only articles by combined region 121
19 One-group posttest-only articles by combined region 123
INTRODUCTION
As technology comes to play an increasing role in the economic and social fabric
of humanity, the need for computer science education will also increase. Computer
science education can enable students to take part in a sociotechnological future, help
them understand the electronic world around them, and empower students to control,
rather than be controlled by, technology. Furthermore, computer science education will
help prepare students for higher education in the computing sciences and, consequently,
help remedy the projected shortage of highly trained computing specialists required to
keep economic infrastructures functional.
It is a given that research on computer science education can lead to the
improvement of computer science education. However, computer science education
research is acknowledged as being an emerging and isolated field. One way to improve
an emerging field is with a review of the research methods used in that field so that those
methods can be analyzed and improved upon.
In a methodological review, a content analysis approach is used to analyze the
research practices reported in a body of academic articles. Methodological reviews differ
from meta-analyses in that research practices, rather than research outcomes, are
emphasized. They are known to be one way to improve the research methods of a field
because they provide a solid basis on which to make recommendations for improvements
in practice. Methodological reviews have been successfully used to inform policy and
practice in fields such as educational technology and behavioral science statistics.
2
Although there have been methodological reviews of computer science education
research, they have either examined nonrepresentative samples of research articles or
have been of poor quality. Because of the benefits that can be reaped from
methodological reviews and because the previous methodological reviews of computer
science education research are lacking, I conducted a rigorous methodological review,
from a behavioral science perspective, on a representative sample of all the research
articles published in major computer science education research journals and conference
proceedings from 2000-2005.
I expect that this dissertation will make a contribution to the field by supplying a
solid ground on which to make recommendations for improvement and to promote
informed dialogue about computer science education research. If my recommendations
are heeded, I expect that computer science education research will improve, which will
improve computer science education, which will, in turn, help the technologically
oriented social and economic needs of the future be met.
The Importance of Computer Science Education
The study of the discipline of computing, defined as "the systematic study of
algorithmic processes that describe and transform information: their theory, analysis,
design, efficiency, implementation, and application" (Denning et al., 1989, p. 12) is
considered to be a key factor in preparing K-12 students, and people in general, for a
technologically oriented future (see Tucker et al., 2003, p. 4). (In this dissertation I use
the term computer science education, rather than the more general term computing
3
education, since computer science education is the term adopted by the Association for
Computing Machinery's Special Interest Group on Computer Science Education.) The
National Research Council Committee on Information Technology Literacy (NRC; 1999)
provides strong rationales for teaching students about technology and computer science.
The NRC argues that people will increasingly need to understand technology to carry out
personally meaningful and necessary tasks, such as
• Using e-mail to stay in touch with family and friends
• Pursuing hobbies
• Helping children with homework and projects
• Finding medical information or information about political candidates over the
World Wide Web (n.p.)
The NRC also argues that an informed citizenry must also be a citizenry that has a
high degree of technological fluency because many contemporary public policy debates
are associated with information technology. For example, the NRC wrote,
A person with a basic understanding of database technology can better appreciate
the risks to privacy entailed in data-mining based on his or her credit card
transactions. A jury that understands the basics of computer animation and image
manipulation may have a better understanding of what counts as "photographic
truth" in the reconstruction of a crime or accident. ... A person who understands
the structure and operation of the World Wide Web is in a better position to
evaluate and appreciate the policy issues related to the First Amendment, free
expression, and the availability of pornography on the Internet, (n.p.)
In terms of U.S. labor needs, the U.S. Department of Commerce's Office of
Technology Policy found that there was "substantial evidence that the United States is
having trouble keeping up with the demand for new information technology workers" (as
cited in Babbit, 2001, p. 21). Computer support specialist and systems administrator are
expected to be two of the fastest growing U.S. occupations during the decade from 2002
4
to 2012 (U.S. Department of Labor-Bureau of Labor Statistics, n.d.a). Also, employment
for computer systems analysts, database administrators, and computer scientists "is
expected to increase much faster than average as organizations continue to adopt
increasingly sophisticated technologies" (U.S. Department of Labor-Bureau of Labor
Statistics, n.d.b).
Computer Science Education Research Can Improve
Computer Science Education
Researchers, such as Gall, Borg, and Gall (1996), have shown that education
research contributes to the practice of education. Gall and colleagues argue that
educational research contributes four types of knowledge to the field of
education — description, prediction, improvement, and explanation — and that education
research enables practitioners to use "research knowledge about what is to inform
dialogue about what ought to be" (p. 13). They further claim that basic educational
research has been shown to influence practice even when influencing practice was not its
intention.
If Gall and colleagues (1996) are correct, in as much as computer science
education is a subset of education research proper, then computer science education
research also has the potential to make contributions to computer science education.
However, as I argue in the section below, currently the realization of that potential is
uncertain; there needs to be more research knowledge about what is to inform what ought
to be.
Computer Science Education Research Is an
Isolated, but Emerging Field
The seminal book on computer science education research (Fincher & Petre,
2004) begins with the following statement: "Computer science education research is an
emergent area and is still giving rise to a literature" (p. 1). Top computer science
education researchers like Mark Guzdial and Vicki Almstrum argue that the
interdisciplinary gap between computer science education and educational research
proper, including methods developed in the broader field of behavioral research, must be
overcome before computer science education research can be considered to be a field
which has emerged (Almstrum, Hazzan, Guzdial, & Petre, 2005). (In this dissertation, I
use the term behavioral research as a synonym for what Guzdzial, in Almstrum et al.
[2005, p. 192], calls "education, cognitive science, and learning sciences research.")
Addressing this lack of connection with behavioral research, Guzdial, in Almstrum and
colleagues (2005), wrote:
The real challenge in computer education is to avoid the temptation to re-invent
the wheel. Computers are a revolutionary human invention, so we might think that
teaching and learning about computers requires a new kind of education. That's
completely false: The basic mechanism of human learning hasn't changed in the
last 50 years.
Too much of the research in computing education ignores the hundreds of
years of education, cognitive science, and learning sciences research that have
gone before us. . . . If we want our research to have any value to the researchers
that come after us, if we want to grow a longstanding field that contributes to the
improvement of computing education, then we have to "stand on the shoulders of
giants," as Newton put it, and stop erecting ant hills that provide too little thought,
(pp. 191-192)
6
The findings from three previous methodological reviews — (a) a critical review of
Kindergarten through 12 th -grade (K-12) computer science education program evaluations,
(b) a methodological review of selected articles published in the SIGCSE Technical
Symposium Proceedings, and (c) a methodological review of the full-papers published in
the Proceedings of the Koli Calling Conference on Computer Science Education
triangulate to support the idea that computer science education research and evaluation is
indeed an emerging and isolated field. (In this dissertation, by program, I mean & project,
not software.) The findings from those three previous reviews (i.e., Randolph, 2005;
Randolph, Bednarik, & Myller, 2005; Valentine, 2004) are summarized below.
A Methodological Review of K-12 Computer Science
Education Program Evaluations
I conducted a methodological review and meta-analysis of the program evaluation
reports in computer science education, which is reported in Randolph (2005).
(Throughout this dissertation, because of the difficulties of making an external decision
about what is research and what is evaluation, I operationalize an evaluation report as a
document that the author called an evaluation, evaluation report, or a program
evaluation report.) To identify the strengths and weaknesses in K-12 computer science
education program evaluation practice, I attempted to answer the following questions:
1. What are the methodological characteristics of computer science education
program evaluations?
2. What are the demographic characteristics of computer science education
evaluation reports?
7
3. What are the evaluation characteristics of computer science education program
evaluations?
4. What is the average effect of a particular type of program on computer science
achievement?
Electronic searches of major academic databases, the Internet, and the ACM
digital library; a subsequent reference-branching hand search; and a query to over 4,000
computer science education researchers and program evaluators were the search
techniques used to collect a comprehensive sample of K-12 computer science education
program evaluations. After selecting the evaluation reports that met seven stringent
criteria for inclusion, the sample of program evaluations were then coded in four areas:
demographic characteristics, intervention characteristics, evaluation characteristics, and
findings. In all, 14 main variables were coded for: region of origin, source, decade of
publication, grade level of target participants, target population, area of computing
curriculum, program activities, outcomes measured, moderating factors examined,
measures, type of measures, type of inquiry, experimental design, and study quality.
Additionally, Cohen's d was calculated for impact on computer science achievement for
each study that reported enough information to do so. A second rater coded a portion of
the reports on the key variables to estimate levels of interrater reliability.
Frequencies and percents were calculated for each of the variables above. A
random effects, variance and within-study sample size/study-quality weighting approach
was used to combine effect sizes. Interactions were examined for type of program.
8
In all, 29 evaluation reports were included. Eight of those reports had data that
could be converted to effect sizes and were included in the meta-analytic portion of the
article, where the effect sizes were synthesized. The major findings are summarized
below:
1 . Most of the programs that were evaluated offered direct computer science
instruction to general education, high school students in North America.
2. In order of decreasing frequency, evaluators examined stakeholder attitudes,
program enrollment, academic achievement in core courses, and achievement in
computer science.
3. The most frequently used measures were, in decreasing order of frequency,
questionnaires, existing sources of data, standardized tests, and teacher-made or
researcher-made tests. Only one measure of computer science achievement, which is no
longer available, had reliability or validity estimates. The pretest-posttest design with a
control group and the one-group posttest-only design were the most frequently used
research designs.
4. No interaction between type of program and computer science achievement
improvement was found.
In terms of the link between program evaluation and computer science education,
the fact that there were so few program evaluations being done, that so few of them (i.e.,
only eight) went beyond simple program description and student attitudes, that only one
used an instrument with information about measurement reliability and validity, and that
one-group posttest-only designs were so frequently used indicate that the past K-12
9
computer science education program evaluations have had many deficiencies. As the next
review indicates, the deficiencies are not solely found in K-12 computer science
education program evaluations; there are also several deficiencies in K-12 computer
science education research as well.
A Methodological Review of Selected Articles in
SIGCSE Technical Symposium Proceedings
Valentine (2004) critically analyzed over 20 years of computer science education
conference proceedings that dealt with first-year university computer science instruction.
In that review, Valentine categorized 444 articles into six categories. The major finding
from his review was that only 21% of papers in the last 20 years of proceedings were
categorized as experimental, which was operationalized as the author of the paper making
"any attempt at assessing the 'treatment' with some scientific analysis" (p. 256). Some of
Valentine's other findings are listed below:
1 . The proportion of experimental articles had been increasing since the mid-90s.
2. The proportion of what he calls Marco Polo — / went there and I saw this —
types of papers had been declining linearly since 1984.
3. The overall number of papers being presented in the SIGCSE forum had been
steadily increasing since 1984 (as cited in Randolph, Bednarik, & Myller, 2005, p. 104).
Valentine concluded that the challenge is to increase the number of experimental
investigations in computer science education research and decrease the number of "I went
there and saw that," self-promotion, or descriptions-of-tools types of articles. The
reliability of Valentine's findings, however, is questionable; Valentine was the single
10
coder and reported no estimates of interrater agreement.
A Methodological Review of the Papers Published
in Koli Calling Conference Proceedings
Randolph, Bednarik, and Myller (2005) conducted a critical, methodological
review of all of the full-papers in the proceedings of the Koli Calling: Finnish/Baltic Sea
Conference on Computer Science Education (hereafter Koli Proceedings) from 2001 to
2004. Each paper was analyzed in terms of (a) methodological characteristics, (b) section
proportions (i.e., the proportion of literature review, methods, and program description
sections), (c) report structure, and (d) region of origin. Based on an analysis of all of the
full-papers published in four years of Koli Proceedings, their findings were that
1. The most frequently published type of paper in the Koli Proceedings was the
program (project) description.
2. Of the empirical articles reporting research that involved human participants,
exploratory descriptive (e.g., survey research) and quasi-experimental methods were the
most common.
3. The structure of the empirical papers that reported research involving human
participants deviated sharply from structures that are expected in behavioral science
papers. For example, only 50% of papers that reported research on human participants
had literature reviews; only 17% had explicitly stated research questions.
4. Most of the text in empirical papers was devoted to describing the evaluation
of the program; very little was devoted to literature reviews.
5. The Koli Calling proceedings represented mainly the work of Nordic/Baltic,
11
especially Finnish, computer science education researchers.
An additional finding was that no article reported information on the reliability or validity
of the measures that were used.
Both the Valentine (2004) and Randolph, Bednarik, and Myller (2005) reviews
converged on the finding that few computer science education research articles went beyond
describing program activities. In the rare cases when impact analysis was done, it was usually
done using anecdotal evidence or with weak research designs.
Synthesis of Findings across Methodological Reviews
When synthesizing the results of these methodological reviews, between
methodological reviews, several preliminary themes from the papers covered in the
methodological reviews emerged. They are listed below:
1 . There is a paucity of impact evaluation/research.
2. There is a proliferation of pure program descriptions.
3. There is an urgent need for reliable and valid measures of computer science
achievement.
4. Research/evaluation reports concentrate mainly on stakeholder attitudes
towards a computer science education program.
5. When experiments or quasi-experiments are conducted, the research designs
are weak.
6. There is a huge gap between how research on human participants is conducted
and reported by computer-science-grounded practitioners and by behavioral-science-
12
grounded practitioners. (Even the term evaluation is used differently by practitioners in
these two groups. See Randolph & Hartikainen, 2004.)
7. Literature reviews in computer science education research and evaluation
reports are missing or inadequate.
Table 1 shows which review provided evidence for each of the themes listed
above. In essence, the findings of the three reviews described above do in fact converge
on Fincher and Petre's (2004) hypothesis that computer science education research is an
emerging, but isolated, field.
Methodological Reviews Can Improve Research Practice
In many types of literature reviews the emphasis is on the analysis and integration
of research outcomes and on how study characteristics covary with outcomes. In fact, the
ERIC Processing Manual defines "a literature review" as an "information analysis and
synthesis, focusing on outcomes . . ." (as cited in Cooper & Hedges, 1994, p. 4). In
methodological reviews, however, the emphasis is not on research outcomes, but on the
description and analysis of research practices (see Cooper, 1988). Keselman et al. (1998)
wrote,
Reviews typically focus on summarizing the results of research in particular areas
of inquiry (e.g., academic achievement of English as a second language) as a
means of highlighting important findings and identifying gaps in the literature.
Less common, but equally important, are reviews that focus on the research
process, that is, the methods by which a research topic is addressed, including
research design and statistical analyses issues, (pp. 350-351)
13
Table 1
Evidence Table for Themes of the Literature Review
Theme
Randolph, 2005
Valentine, 2004
Randolph, Bednarik,
& Myller, 2005
1 . Paucity of
impact
research
2. Mostly program
descriptions
3. Need for
measures
4. Stakeholder
attitudes
5. Weak designs
6. Research gap
8. Lack of
literature
reviews
As an example, in a methodological review of educational researchers' ANOVA,
MANOVA, and ANCOVA analyses, Keselman and colleagues (1998) used a content
analysis approach to synthesize the statistical practices in research articles published in
major education research journals. They then compared the statistical practices of
educational researchers with the statistical practices recommended by statisticians and
made recommendations for improvement.
14
Of the variety of reasons for conducting a methodological review, two of the most
obvious reasons are to help improve methodological practice and inform editorial policy.
According to Keselman and colleagues (1998),
Methodological reviews have a long tradition (e.g., Edgington, 1964; Elmore &
Woehlke, 1988, 1998; Goodwin & Goodwin, 1985a, 1985b; West, Carmody, &
Stallings, 1983). One purpose of these reviews had been the identification of
trends in . . . practice. The documentation of such trends has a twofold purpose:
(a) it can form the basis for recommending improvement in research practice, and
(b) it can be used as a guide for the types of . . . procedures that should be taught
in methodological courses so that students have adequate skills to interpret the
published literature of a discipline and to carry out their own projects, (pp. 350-
351)
One current example of how methodological reviews can bring about improved
practice and inform policy is shown in Leland Wilkinson and the APA Task Force on
Statistical Inference's influential 1999 report — Statistical Methods in Psychology
Journals: Guidelines and Explanations (hereafter Wilkinson et al). In that report, several
of the most prominent figures in behavioral science research (e.g., Robert Rosenthal,
Jacob Cohen, Donald Rubin, Bruce Thompson, Lee Cronbach, and others) came together,
in response to the use and abuse of inferential statistics reported in Cohen (1994), to
codify best practices in inferential statistics and in statistical methods in general. In that
report, they drew heavily on methodological reviews of the statistical practices of
behavioral science researchers, such as Keselman and colleagues (1998), Kirk (1996), and
Thompson and Snyder (1998). Keselman and colleagues were interested in the ANOVA,
ANCOVA, and MANOVA practices used by educational researchers. Kirk and
Thompson and Snyder were interested in the statistical inference and reliability analyses
done by education researchers. In addition to the fields of psychological statistics,
15
methodological reviews have also been published in other fields, from program
evaluation (Lawrenz, Keiser, & Lavoir, 2003; Randolph, 2005) to special education (Test,
Fowler, Brewer, & Wood, 2005) to medical science (Clark, Anderson, & Chalmers, 2002;
Huwiler-Miintener, Juni, Junker, & Egger, 2002; Lee, Schotland, Bacchetti, & Bero,
2002).
In general terms, The Social Science Research Council (SSRC) and the National
Academy of Education's (NAE) Joint Committee on Education Research noted a lack of
and need for "data and analysis of the education research enterprise" (Ranis & Walters,
2004, p. 798). In fact the research priorities concerning the lack of data and analysis in
education research included "determination of where education research is conducted and
by whom" and "identification of the range of problems addressed and the methods used
to address them" (p. 799). Methodological reviews can help meet the need for data about
and analysis of the education research enterprise, especially regarding the research
priorities identified above.
There are two conditions that suggest the value for a methodological review to
improve practice and inform policy. The first is when there is consensus among experts
for "best practice" but actual practice is expected to fall far short of best practice. The
methodological review can identify these shortcomings and suggest policies for research
funding and publication. For example, in the Keselman and colleagues (1998) review,
they found that there was a difference between how statisticians use ANOVA and how
social science researchers use ANOVA. Thus, the rationale for the Keselman and
colleagues review was that the recommendations given by the statisticians could benefit
16
the research practices of the social science researchers. The second condition is when
there are islands of practice that can benefit from exposure to each other — for example,
when there are groups that practice research in different ways or at different levels.
In terms of the conditions for a methodological review to improve practice and
inform policy, both conditions are met for the field of computer science education. First,
there are islands of practice. As Guzdzial points out in the statement of the Association
for Computing Machinery's Special Interest Group on Computer Science Education's
(hereafter .4 CM SIGCSE) panel on 'Challenges to Computer Science Education
Research,' there are two distinct islands of practice: computer science education research
and "education, cognitive science, and learning sciences research" (Almstrum et al., 2005,
p. 192). Second, there is a call for interdisciplinary exchange between islands of practice;
actual practice in computer science education research differs from accepted practice in
"education, cognitive science and learning sciences research." The ACM SIGCSE panel
on 'Challenges to Computer Science Education Research' stated that one of the keys to
improving computer science education research is for computer science educators to look
to "education, cognitive science, and learning sciences research." This sentiment was also
stated by the computer science education panel on Import and Export to/from Computing
Science Education (Almstrum, Ginat, Hazzan, & Morely, 2002). They wrote:
Computing science education is a young discipline still in search of its research
framework. A practical approach to formulating such a framework is to adapt
useful approaches found in the research from other disciplines, both educational
and related areas. At the same time, a young discipline may also offer innovative
approaches to the older discipline, (p. 193)
17
Methodological Reviews in the Field of Educational Technology
Psychology is not the only field in which methodological reviews have been
conducted. The field of educational technology, which makes use of the software
engineering and management information systems components of computer science, has a
long history of methodological reviews, dating as far back as the mid-1970s. To make
sense of all of those reviews and to be able to compare the results of this dissertation
across fields, I conducted a review of those methodological reviews. Specifically, I
attempted to answer the following research questions:
1 . What metacategories can be used to subsume the categories used in the
previous educational technology methodological reviews?
2. What proportions of articles in the previous educational technology
methodological reviews fall into each of these categories?
3. How do those proportions of articles differ by year and type of forum?
4. How do these proportions compare with the proportions in education research
proper?
In the sections that follow I (a) present the results of a methodological review of
education proper articles (to be able to answer Question 4), (b) present the methods for
conducting this review of methodological reviews of education technology articles, and
(c) finally present the results of the review of methodological reviews of educational
technology articles.
18
The Proportions of Article Types in
Education Research Proper
Before describing the methods that were used in this review of reviews, to have a
point of reference on which this review's results can be compared and contrasted, I report
on a high-quality methodological review in the field of education research proper. In that
review, Gorard and Taylor (2004) reviewed 42 articles from the six issues published in
2001 in the British Educational Research Journal (BERJ), 28 articles from the four
issues published in 2002 in the British Journal of Educational Psychology (BJEP), and 24
articles from the four issues published in 2002 in Educational Management and
Administration. Gorard and Taylor found the following results:
Overall, across three very different [education] journals in 2002, 17 per cent of
articles were clearly or largely non-empirical (although this description includes
literature reviews, presumably based on empirical evidence), 4 percent were
empirical pieces using a combination of 'qualitative' and 'quantitative' methods
(therefore a rather rare phenomenon), 34 percent used qualitative methods alone,
and 47 percent used quantitative methods alone, (p. 141)
Because the cumulative percent above is 102, 1 rounded some figures down and assumed
then that, out of 94 articles, 15, 4, 32, and 43 articles were nonempirical, mixed,
qualitative, and quantitative, respectively.
Although Gorard and Taylor's (2004) sample of articles that were reviewed was
small, Gorard and Taylor provided convincing evidence, from a variety of sources, that
validated the proportions of nonempirical, quantitative, qualitative, and mixed-methods
articles found in their review. Those sources included
• interviews with key stakeholders from across the education field,
includingresearchers, practitioner representatives, policy makers and policy
implementers;
19
a large-scale survey of the current methodological expertise and future training
needs of UK education researchers; [and a ]
detailed analysis and breakdown of 2001 RAE [Research Assessment Exercise,
2001]. (p. 114)
Method for Conducting a Review of Methodological Reviews
In this section I explain the methods that I used for conducting this review of
methodological reviews in educational technology. It includes a description of the criteria
for inclusion and exclusion, the search strategy, coding categories, and data analysis
procedures.
Criteria for Inclusion and Exclusion
Articles were included in this review if they met six criteria, which are listed
below:
1. It was a quantitative review (e.g., a content analysis) of research practices, not
a literature review in general or a meta-analysis, which focuses on research outcomes.
2. The review dealt with the field of educational technology or distance
education.
3. The review was written in English.
4. The number of articles that were reviewed was specified.
5. The candidate review's categories were able to be subsumed under
metacategories.
6. The review's articles did not overlap with another review's articles. (When
reviews overlapped, only the most comprehensive review was taken.)
20
Search Strategy
The first step of the search strategy was to conduct an electronic search of the
academic databases Academic Search Elite, Psych Info, and ERIC, and of the Internet, via
Google. The electronic search was conducted in July 2006 using the terms educational
technology, methodological review; computer-assisted instruction, methodological
review; educational technology, review; and computer-assisted instruction, review. The
title of each entry was read to determine if it might lead to a review that would meet the
criteria for inclusion. (In cases where the review returned more than 500 entries, only the
first 500 were read.) If the title looked promising, the resulting webpage, abstract, or
entire article was read to see if the article met the criteria for inclusion.
The second step of the search strategy was to do pearl building. The references
section of the articles identified from the electronic search and the articles that were
known to me beforehand were searched. This pearl-building process was repeated until a
point of saturation was reached.
The third step of the search strategy was to compile a list of articles that met the
criteria for inclusion and to send that list out to experts in the field of educational
technology to see if there were any methodological reviews that should have been
included on the preliminary list but had not. A query was sent to the members of the
ITFORUM listserv on July 20, 2006. Eight ITFORUM members responded to the query
and suggested more articles that might meet the criteria for inclusion.
21
Coding Categories
Each of the methodological reviews that met all six criteria was coded on seven
attributes:
1. The forum from which the methodological review came;
2. The author(s) of the methodological review;
3. The year of the methodological review;
4. The forums, issues, and time periods from which the reviewed articles came;
5. The categories that each methodological review used;
6. The number of articles that were put into each of the methodological review's
categories; and
7. The research question that the review attempted to answer.
Data Analysis
In the reviews which met all six criteria for inclusion, the number of articles
which fit into each metacategory was recorded. Those results were summed to arrive at an
overall picture of how many articles, across methodological reviews, fell into each of the
metacategories. Those results were disaggregated by forum and by year. Also, the results
of this methodological review of articles from educational technology forums were
compared with the results of Gorard and Taylor's (2004) methodological review of
articles from education research journals proper. Chi-square analyses were used to
determine the likelihood of getting differences in the observed multinomial proportions as
22
large as those expected by chance. In addition to the quantitative synthesis, I also
recorded the research question that each methodological review attempted to answer.
Results of Review of Reviews
The literature search resulted in 13 methodological reviews that met at least the
first three criteria for inclusion (Alexander & Hedberg, 1994; Caffarella, 1999; Clark &
Snow, 1975; Dick & Dick, 1989; Driscoll & Dick, 1999; Higgins, Sullivan, Harper-
Marinick, & Lopez, 1989; Klein, 1997; Phipps & Merisotis, 1999; Randolph, in press;
Randolph, Bednarik, Silander, et al., 2005; Reeves, 1995; Ross & Morrison, 2004;
Williamson, Nodder, & Baker, 2001). Four of the reviews mentioned above did not meet
all six criteria for inclusion and, therefore, were not included in the current review.
Phipps and Merisotis 's review, a large scale critical review of the research on distance
learning, was excluded because it did not meet Criterion 4: it did not specify how many
articles were reviewed. Ross and Morrison's review and Alexander and Hedberg' s review
were excluded because they did not meet criterion five: Ross and Morrison categorized
by experimental design and setting, Alexander and Hedberg categorized by evaluation
design. Also, Caffarella, who did a review of educational technology dissertations, was
excluded because the categories used could not be codified with the metacategories in the
current review. Driscoll and Dick was excluded because their sample overlapped with
Klein's review, which had a more comprehensive sample. Reeves' sample of articles
from Educational Technology Research & Development was not included because it also
overlapped with Klein's review; however, Reeve's sample of Journal of Computer-Based
23
Instruction articles was included. Thus, nine methodological reviews, covering 905
articles from the last 30 years of educational technology, were included in this review of
educational technology methodological reviews. The questions that each of those
methodological reviews attempted to answer are summarized in Table 2. At a glance, the
question being asked in the major methodological reviews of the educational technology
literature was "What are the types and methodological properties of research reported in
educational technology articles?"
Table 3 presents those reviews, the forum, the years sampled, and the number of
articles reviewed. As shown in Table 3 the forums that were covered in the previous
reviews were A V Communication Review (AVCR), Educational Communication and
Technology Journal (ECTJ), Journal of Instructional Development (JID), Journal of
Computer-Based Instruction (JCBI), Educational Technology Research & Development
(ETR&D), American Journal of Distance Education (AJDE), Distance Education (DE),
Journal of Distance Education (JDE), Proceedings of the International Conference on
Advanced Learning Technologies (ICALT).
Also, of the 46 papers reviewed in Williams et al. (2001),
37 originate[d] from refereed journals or conference proceedings and the
remainder from academic websites or Government departments. ... In particular
we drew material from the conferences of the Australasian Society for Computers
in Learning in Tertiary Education (ASCILITE) and from the National Advisory
Committee for Computing Qualifications (NACCQ). (p. 568)
Table 4 shows the categories that were used in previous methodological reviews.
It shows how I grouped these categories together to arrive at the four metacategories:
quantitative, quantitative, mixed-methods, and other. The other category included
Table 2
24
Research Questions in Educational Technology Methodological Reviews
Study
Overview of research questions
Alexander &
Hedberg, 1994
Caffarella, 1999
Clark & Snow, 1975
Dick & Dick, 1989
Driscoll &Dick,
1999
Klein, 1997
Higgins et al, 1999
What, and in what proportions, evaluation models are used in
evaluations of educational technology?
How have the themes and research methods of educational technology
dissertations changed over the past 22 years?
What research designs are being reported in educational technology
journals? In what proportions?
How do the demographics, first authors, and substance of articles in two
certain educational technology journals differ?
What types of inquiry are being reported in educational technology
journals? In what proportions?
What types of articles and what topics are being published in a certain
educational technology journal? In what proportions?
What do members of a certain educational technology journal want to
read?
Phipps & Merisotis, What are the methodological characteristics of studies published in
1999 major educational technology forums?
Randolph, in press Are the same methodological deficiencies reported in Phipps &
Merisotis (1999) still present in current research?
Randolph et al.,
2005
Ross & Morrison,
2004
Reeves, 1995
Williamson et al.,
2001
What are the methodological properties of articles in the proceedings of
ICALT 2004?
What are proportions of experimental designs being used in educational
technology research?
What types of methodological orientations do published educational
technology articles take? In what proportions?
What types of research methods and pedagogical strategies are being
reported in educational technology forums?
25
Table 3
Characteristics of Educational Technology Reviews Included in the Quantitative
Synthesis
Number of articles
Review
Forum
Years covered
reviewed
Clark & Snow, 1975
AVCR
1970-1975
111
Dick & Dick, 1989
ECTJ
1982-1986
106
JID
1982-1986
88
Higgins et al, 1989
ECTJ
1986-1988
40
JID
1986-1988
50
Reeves, 1995
JCBI
1989-1994
123
Klein, 1997
TR&D
1989-1997
100
Williamson et al, 2001
Mixed
1996-2001
46
Randolph, in press
AJDE
2002
12
DE
2002
14
JDE
2002-2003
40
Randolph, 2005
ICALT
2004
175 a
Total
905
Note. AVCR = Audio Visual Communication Review, ECTJ = Educational Communication and
Technology Journal, JID = Journal of Instructional Development, JCBI = Journal of Computer-
Based Instruction, ETRD = Educational Technology Research & Development, AJDE =
American Journal of Distance Education, DE = Distance Education, JDE = Journal of Distance
Education, ICALT = International Conference on Advanced Learning Technologies .
a 175 investigations reported in 123 articles
articles that did not deal with human participants, such as literature reviews, descriptions
of tools, or theoretical papers.
Figure 1 shows the number and percentage of 905 articles that were distributed
into each metacategory. The other category is the largest category, experimental is the
26
Table 4
The Composition of Educational Technology Metacategories
Qualitative
Quantitative
Mixed methods
Other
Qualitative; critical Quantitative;
theory; explanatory experimental/quasi-
descriptive; case experimental; quasi-
studies
experimental;
exploratory
descriptive,
correlational;
causal-comparative;
classification;
descriptions;
experimental
research;
experimental study;
survey research,
empirical research;
evaluation;
correlational;
empirical,
experimental, or
evaluation;
quantitative
descriptive
Mixed methods;
triangulated; mixed
Literature reviews;
other; description
with no data;
theory, position
paper, and so
forth.; theory;
methodology;
professional
second largest category, and those categories are followed by the qualitative and mixed
methods categories.
Figure 2 shows the proportions of articles that fell into each of the different
categories in each forum. It indicates that there as considerable variability between
forums in terms of the proportions of types of articles that were published. It should be
noted that these data usually only represent a limited time span over the life of the forum.
27
n=411 /
46% |
■J.
n=40
4%
"^\ n=280
\ 31%
>
^^^ n=174
19%
D Quantitative
■ Qualitative
D Mixed methods
□ Other
Figure 1. Proportion of types of articles in educational technology journals.
1 00%
90% -
80% -
70% -
60% -
50% -
40% -
30% -
20% -
10% -
0%
~
1
□ Other
□ Mixed methods
■ Qualitative
□ Quantitative
_
_
^
* / j? / • * /
Figure 2. Proportion of types of educational technology articles by forum.
28
Figure 3 shows that the proportions of types of articles varied over each time
period. (Note that the other category was not included here so the remaining categories
could be more easily compared.) This figure shows that there were high proportions of
qualitative articles from the early 80s to early 90s, but the proportions dropped off in the
late 90s and early 00s. It is important to note when interpreting Figure 3 that forums were
not constant across time periods; some forums were sampled more heavily in different
time periods than others. Table 3 showed how many articles were sampled from each
forum each time period. The median year in a yearly range determined what time period
a review would be categorized into.
100%
90%
80% 4
70%
60%
50% ]-
40%
30%
20%
10%
0%
hUHU
W
aSJ
* S? .<& jcSS
Q* <$*
A^ ,<$> s<&> j$
oP ^ oP ^ oP J^
□ Mixed methods
■ Qualitative
□ Quantitative
Figure 3. Proportions of types of educational technology articles by time period.
29
Table 5 shows the difference between the numbers of articles dealing with human
participants in the current review of educational technology reviews and Gorard and
Taylor's (2004) methodological review of British educational research. In short,
education proper articles had, on average, 30% more articles that reported research on
human participants than in educational technology articles. The difference was
statistically significant, f(l, N= 999) = 30.21,/? < .000.
Table 6 shows, however, that the proportions of quantitative, qualitative, and
mixed-methods articles were nearly the same in educational technology and general
education-research forums. The differences were not statistically significant, % 2 (2, N =
573)= 1.41,/? = .495.
Table 5
Comparison of the Proportion of Human Participants Articles in Educational
Technology and Education Proper
Human participants Percentage Adjusted
Field Yes No Total yes residual
Ed. tech 494 411 905 54.6 -5.5
Ed. proper 79 15 94 84.0 5.5
Total 573 426 999
Note. Ed. tech. = educational technology, Ed. proper = education proper % 2 (\, N= 999) =
30.21, p<. 000.
30
Table 6
Comparison of Type of Methods Used in Educational Technology and Education Proper
Field
Type of article
Ed. tech
Ed. proper
Total
Quantitative
Qualitative
Mixed methods
Total
280 (56.7%)
174 (35.2%)
40(8.1%)
494 (100%)
43 (54.4%)
32 (40.5%)
4(5.1%)
79 (100%)
323 (56.4%)
206 (36.0%)
44 (7.7%)
573 (100%)
Note. Percentages are within Review; Ed. tech. = educational technology. Ed. proper :
education proper. %\2, N= 573) = 1.41,/? = .495.
One limitation of this review of reviews was that there were no estimates of
interrater reliability for the variables that were coded. However, that limitation is
mitigated by the fact that the coding variables were not of a subjective nature. In Table 4,
I listed all of the previous categories that had been used and made explicit how they
related to the metacategory variable. Arriving at the proportions for the metacategories
was then simply a matter of summing the number of articles that belonged to each of the
subcategories in the metacategory.
In summary, I found that most of the research in educational technology had been
quantitative, some of it qualitative, and a small percentage of it mixed methods. The
percentage of empirical papers that dealt with human participants was much higher in
education research proper than in educational technology. However, the relative
31
proportions of quantitative, qualitative, and mixed-methods articles in educational
technology and education research proper forums were about the same.
Methodological Reviews in Computer Science Proper,
Software Engineering, and Information Systems
Although ancillary to computer science education, there are three seminal
methodological reviews of the computer science literature proper that are worth
mentioning and that may help put the results of this dissertation into context. Those
reviews are Glass, Ramesh, and Vessey (2004); Tichy, Luckowicz, Prechelt, and Heinz
(1995); and Zelkowitz and Wallace (1997).
In "An Analysis of Research in Computing Disciplines," Glass et al. (2004)
reviewed 1,485 articles from a selection of journals in the fields of computer science,
software engineering, and information systems. They classified each article by topic,
research approach, research method, reference discipline, and level of analysis. Some
findings from the Glass et al. review that might be relevant to the current review are
quoted below:
CS [computer science] research methods consisted predominantly of
mathematically based Conceptual Analysis (73%). SE [software engineering] used
Conceptual Analysis that is not mathematically based (44%) with Concept
Implementation also representing a significant research method at 17%. IS
[information systems] research used predominantly five types of research
methods, the most notable being Field Study (27%), Laboratory Experiment
(Human), Conceptual Analysis (15%), and Case Study (13%). (p. 92)
In "Experimental Evaluation in Computer Science: A Quantitative Study," Tichy
et al. (1995) did a methodological review of 400 articles from
32
complete volumes of several refereed computer science journals, a conference,
and 50 titles drawn at random from all articles published by ACM [The
Association for Computing Machinery] in 1993. The journals of Optical
Engineering (OE) and Neural Computation (NC) were used for comparison, (p. 9)
They classified each article according to several attributes, such as whether it was an
empirical work or not. The major findings from the Tichy et al. review are quoted below:
Of the papers in the random sample that would require experimental validation,
40% have none at all. In journals related to software engineering, this fraction is
50%. In comparison, the fraction of papers in OE [a journal called Optical
Engineering] andNC [a journal called Neural Computing] is only 15% and 12%,
respectively. Conversely, the fraction of papers that devote one fifth or more of
their space to experimental validation is almost 70% for OE and NC, while it is a
mere 30% for the computer science (CS) random sample and 20% for software
engineering. The low ratio of validated results appears to be a serious weakness in
computer science research. This weakness should be rectified for the long-term
health of the field, (p. 9)
Zelkowitz and Wallace (1997), in "Experimental Validation in Software
Engineering," reviewed over 600 papers from the software engineering literature and 100
articles from other fields as a basis for comparison. As in the other reviews, they
classified the articles into methodological categories. Some of their findings that are
relevant to the current review are presented below:
We observed that 20% of the papers in the journal IEEE Transactions on Software
Engineering have no validation component (either experimental or theoretical).
This number is comparable to the 15 to 20% observed in other scientific
disciplines. However, about one-third of the software engineering papers had a
weak form of experimentation (assertions) where the comparable figure for other
fields was more like 5 to 10%. (p. 742)
Several things need to be noted about these reviews before using them as a basis
for comparison with computer science education research. First, it is difficult, if not
impossible, to synthesize the results of these reviews because each uses a different
33
categorization system. Second, the relevance of these reviews to the field of computer
science education is questionable; these reviews only apply to computer science
education research in as much as computer science education research was a part of the
samples of the computer science, software engineering, and information systems literature
that were reviewed. Finally, some question the validity of these reviews. For example,
Tedre (2006) argued that the Glass et al. (2004) study "may not adequately describe what
actually happens in computer science" (p. 349), that the granularity of the categories in
Glass et al.'s study is overly coarse, and that "the choice of mainstream journals may have
biased the sample of articles towards mainstream research so that alternative methods
may get lesser attention" (p. 349).
The Scope and Quality of the Previous Methodological
Reviews of Computer Science Education Research
The argument that has been developed thus far is that methodological reviews
have been used successfully to improve the methodological practices of researchers in a
variety of behavioral research fields, and the conditions appear met for methodological
reviews to also help improve the emerging methodological practices of computer science
education researchers. Although there have been several methodological reviews of
research on computer science education, I will demonstrate in the following section that
those methodological reviews are limited either in their breadth, depth, or reliability.
To identify all the past methodological reviews of computer science education, six
searches of the Internet; the ACM Digital Library; and Academic Premier, Computer
34
Source, ERIC, Psychology and Behavioral Science Collections, and PyscINFO (via Ebsco
Host) were conducted on November 29, 2005 using the keyword combinations:
"computer science education research, " "methodological review, " and "computer
science education research, " "meta-analysis. " Another six searches on January 20, 2006
were conducted using the same databases but using the keyword combinations:
"computer science education research, " "systematic review," and "computer science
education research, " "research synthesis. " The summary, title, or abstract of each record
was read to determine if it would lead to a review of the research methods in computer
science education.
In addition to the electronic searches, the table of contents of (a) the Koli Calling
Proceedings (2001-2005), (b) the ICER Proceedings 2005, (c) Computer Science
Education (volumes 8-15), and (d) the Journal of Computer Science Education Online
(the volumes published between 2001-2005) were searched. Also, a pearl-building
approach was taken to identify other reviews from the reference sections of the reviews,
including meta-analyses, found from the searches described above. Meta-analyses, or
other reviews that emphasized research outcomes rather than methods, were excluded
from this review of computer science education methodological reviews. The term meta-
analysis was included as a search term because sometimes methodological reviews are
mislabeled as meta-analyses, as was the case with Valentine's article (2004). Table 7
shows the number of records that resulted from each search.
Based on the search procedure mentioned above, I found that three
methodological reviews of computer science research (or evaluation) had been conducted
Table 7
Description of the Electronic Search for Previous Methodological Reviews
35
Search
Term
Database Records
1 "computer science education research"
"methodological review"
2 "computer science education research" "meta-
analysis"
3 "computer science education research" "systematic
review"
4 "computer science education research" "research
synthesis"
5 "computer science education research"
"methodological review"
6 "computer science education research" "meta-
analysis"
7 "computer science education research" "systematic
review"
8 "computer science education research" "research
synthesis"
9 "computer science education research"
"methodological review"
10 "computer science education research" "meta-
analysis"
1 1 "computer science education research" "systematic
review"
12 "computer science education research" "research
synthesis"
Internet
(Google)
Internet
(Google)
Internet
(Google)
Internet
(Google)
ACM library
ACM library
ACM library
ACM library
Ebsco Host
Ebsco Host
Ebsco Host
Ebsco Host
10
27
315
33
21
since computer science education research began in the early 1970s. (One review that
36
should be acknowledged, but was not classified as a methodological review is Kinnunen
[n.d.]. In that review, Kinnunen examined the subject matter of the articles published in
SIGCSE Bulletin.) Those three reviews (Randolph, 2005; Randolph, Bednarik, & Myller
2005; Valentine, 2004) were already presented in detail in the section entitled "Computer
Science Education Research is an Emerging Field, " so they will not be presented again
here. I will, however, describe their samples and map the areas of computer science
education research that have been covered. Before that, however, I will explain my
assumption of what the population of computer science education research reports
consist of.
In this dissertation, I was interested in making a generalization to the entirety of
recent research published in the major computer science education research forums. I
operationalized this as the full papers published from 2000 to 2005 as the June and
December issues of SIGCSE Bulletin [hereafter Bulletin], a computer science education
journal; Computer Science Education [hereafter CSE], a computer science education
research journal; the Journal of Computer Science Education Online, [hereafter JCSE], a
little-known computer science education journal; the Proceedings of the SIGCSE
Technical Symposium [hereafter SIGCSE]; The Proceedings of the Innovation and
Technology in Computer Science Education Conference [hereafter ITiCSE]; the Koli
Calling: Finnish/Baltic Sea Conference on Computer Science Education [hereafter Koli],
the Proceedings of the Australasian Computing Education Conference [hereafter ACE],
and the International Computer Science Education Research Workshop [hereafter ICER].
(The fall and spring issues of Bulletin are the SIGCSE and ITiCSE proceedings.) I
37
included "full papers," but excluded poster summaries, demo summaries, editorials,
conference reviews, book reviews, forewords, introductions, and prologues in the
sampling frame. The three previous methodological reviews of computer science
education research (Randolph, 2005; Randolph, Bednarik, & Myller, 2005; Valentine,
2004) only cover a very small part of the population operationalized above. Additionally,
the review that is most representative of the population of computer science education
research articles (Valentine) has serious methodological flaws.
In the Randolph, Bednarik, and Myller (2005) methodological review, a census of
the full papers published in the Proceedings of the Koli Calling Conference from 2001 to
2004 was reviewed. Although a census was conducted, the articles in the Proceedings of
the Koli Calling Conference made up only a small, marginal part of the population of
recent computer science education research articles. For example, the articles published in
the Proceedings of the Koli Calling Conference from 2001 to 2005 only accounted for
7% of the population specified above. Also, the Koli Calling Conference is a regional
conference (Finnish/Baltic) and, therefore, its proceedings are not representative of the
population of computer science education research articles as a whole. For example,
about 90% of the papers in the Randolph et al. review were of Finnish origin.
The Randolph (2005) methodological review focused on a subset of the grey
literature on computer science education — reports of evaluations of computer science
education programs. (Almost all of the program evaluation reports included in the review
of program evaluation reports were published on the Internet or in the ERIC database.) In
the methodological review section of the Randolph review, 29 program evaluation reports
38
were analyzed. Of those 29, only two of the reviewed reports had been summarized in
one of the forums included in my operationalization of the computer science education
research population. Thus, the population of the Randolph review is almost entirely
different than the population of this dissertation.
The Valentine (2004) methodological review included 444 articles that dealt with
the first year of computer science education courses and were published in the SIGCSE
Technical Symposium Proceedings from 1984 to 2003. Valentine reviewed a large
number of articles, but he sampled them from only one forum for publishing computer
science education research and excluded articles that did not deal with first-year computer
science courses. In addition to the potentially low generalizability of Valentine's sample,
the quality of the Valentine review is questionable. First, Valentine only coded one
variable for each article — he simply classified the articles into one of six categories:
Marco Polo, Tools, Experimental, Nifty, Philosophy, and John Henry. The experimental
category — operationalized as "any attempt at assessing the 'treatment' with some
scientific analysis" (Valentine, p. 256) — is so broad that it is not useful as a basis for
recommending improvements in practice. Second, Valentine coded all of the articles
himself without any measure of interrater agreement.
In conclusion, the three previous methodological reviews either lacked breadth,
depth, or reliability. Randolph, Bednarik, and Myller (2005), Randolph (2005), and, to a
lesser extent, Valentine (2004) do not represent the population of published computer
science education research. What is more, the Valentine review, which has the greatest
number of reviewed articles, has questionable reliability. Also, Valentine only coded the
39
articles in terms of one somewhat light-hearted variable. Given that fact, it is difficult to
say with certainty what the methodological practices in computer science education
research are and, consequently, it is also difficult to have a convincing basis to suggest
improvements in practice.
Purpose and Research Questions
Because the past methodological reviews of computer science education research
had limitations either in terms of their generalizability or reliability, I conducted a
replicable, reliable, methodological review of a representative sample of the research
published in the major computer science education forums over the last 6 years. This
dissertation (a) provides significantly more-representative coverage of the field of
computer science education than any of the previous reviews, (b) covers articles with
more analytical depth (with a more-refined coding sheet) than any of the previous
reviews, and (c) with a greater amount of reliability and replicability than any of the other
previous reviews. In short, this dissertation simultaneously extends the breadth, depth,
and reliability of the previous reviews.
The purpose of this methodological review was to have a valid and convincing
basis on which to make recommendations for the improvement of computer science
education research and to promote informed dialogue about its practice. If my
recommendations are heeded and dialogue increases, computer science education is
expected to improve and, consequently, help meet the social and economic needs of a
technologically oriented future.
40
To have a valid basis to recommend improvements of computer science education
research methodology, I answered the primary research question: What are the
methodological properties of research reported in articles in major computer science
education research forums from the years 2000-2005? The primary research question
can be broken down into several subquestions, which are listed below:
1 . What was the proportion of articles that reported research on human
participants?
2. Of the articles that did not report research on human participants, what types of
articles were being published and in what proportions?
3. Of the articles that did report research on human participants, what proportion
provided only anecdotal evidence for their claims?
4. Of the articles that did report research on human participants, what types of
methods were used and in what proportions?
5. Of the articles that did report research on human participants, what measures
were used, in what proportions, and was psychometric information reported?
6. Of the articles that did report research on human participants, what were the
types of independent, dependent, mediating, and moderating factors that were examined
and in what proportions?
7. Of the articles that used experimental/quasi-experimental methods, what types
of designs were used and in what proportions? Also, were participants randomly assigned
or selected?
41
8. Of the articles that reported quantitative results, what kinds of statistical
practices were used and in what proportions?
9. Of the articles that did report research on human participants, what were the
characteristics of the articles' structures?
Based on the previous methodological reviews of computer science education
research, I made predictions for seven of the nine subquestions above. This dissertation
tested those predictions on a random sample of the entire population of articles or
conference papers published in major computer science education research forums. The
predictions are listed below; the citations refer to the source(s) from which the prediction
was made.
1. Between 60% and 80% of computer science education research papers will not
report research on human participants (Randolph, 2005; Randolph, Bednarik, & Myller,
2005).
2. Of the papers that do not report research on human subjects, the majority
(about 60%) will be purely program (intervention) descriptions (Randolph, Bednarik, &
Myller, 2005; Valentine, 2004).
3. Of the articles that do report on human participants, about 15% will report only
anecdotal evidence for their claims (Randolph, Bednarik, & Myller, 2005).
4. Of the articles that report research on human participants, articles will most
frequently be reports of experiments/quasi-experiments or exploratory descriptions (e.g.,
survey research), as opposed to correlational studies, explanatory descriptive studies (e.g.,
42
qualitative types of research), causal-comparative studies, or classification studies;
(Randolph, 2005; Randolph, Bednarik, & Myller, 2005).
5. Of the articles that do report research on human participants, questionnaires,
grades, and log files will be the most frequently used types of measures. None (or very
few) of the measures will have psychometric information reported (Randolph, 2005;
Randolph, Bednarik, & Myller, 2005).
6. Of the articles that do report research on human participants, the most frequent
type of independent variable will be student instruction, the most frequent dependent
variable will be stakeholder attitudes, and the most frequent moderating variable will be
gender (Randolph, 2005; Randolph, Bednarik, & Myller, 2005).
7. Of the articles that report experiments or quasi-experiments, the one-group
posttest-only design and posttest-only with controls design will be the most frequently
used types of experimental designs. Instances of random selection or random assignment
will be rare (Randolph, 2005; Randolph, Bednarik, & Myller, 2005).
8. Of the articles that report research on human participants, about 50% of the
reports will be missing a literature review section. The vast majority will not have
explicitly stated research questions. (Randolph, Bednarik, & Myller, 2005).
In addition to answering the primary research question — What are the
methodological characteristics of the computer science education research published in
major forums between 2000 and 2005? — I conducted 15 planned contrasts to identify
islands of practice. In the contrasts, there were three comparison variables — (a) type of
publication forum: journal or conference proceedings, (b) year, and (c) region of first
43
author's institutional affiliation — crossed by five dependent variables: (a) frequency of
articles in which only anecdotal evidence was reported; (b) frequency of articles that
reported on experimental or quasi-experimental investigations; (c) frequency of articles
that reported on explanatory descriptive investigations; (d) frequency of experimental or
quasi-experimental articles that used a one-group posttest-only research design
exclusively; and (5) the frequency of articles in which attitudes were the only dependent
variable measured.
The 15 planned contrasts answered the following three secondary research
questions:
1 . Is there an association between type of publication (whether articles are
published in conferences or in journals) and frequency of articles providing only
anecdotal evidence, frequency of articles using experimental/quasi-experimental research
methods, frequency of articles using explanatory descriptive research methods, frequency
of articles in which the one-group posttest-only design was exclusively used, and
frequency of articles in which attitudes were the sole dependent variable?
2. Is there a yearly trend (from 2000-2005) in terms of the frequency of articles
providing only anecdotal evidence, frequency of articles using experimental/quasi-
experimental research methods, frequency of articles using explanatory descriptive
research methods, frequency of articles in which the one-group posttest-only design was
exclusively used, and frequency of articles in which attitudes were the sole dependent
variable?
44
3. Is there an association between the region of the first author's institutional
affiliation and frequency of articles providing only anecdotal evidence, frequency of
articles using experimental/quasi-experimental research methods, frequency of articles
using explanatory descriptive research methods, frequency of articles in which one-group
posttest-only designs were exclusively used, and frequency of articles in which attitudes
were the sole dependent variable?
Note that the primary and secondary questions that were asked here are basically
the same questions that were asked in methodological reviews in a closely related
field — educational technology (see Table 2). Also, the question regarding the statistical
practices of computer science education researchers (i.e., Subquestion 8 of the primary
research question) was aligned with the main questions that were asked in the
methodological reviews that supported the APA Task Force on Statistical Inference's
recommendations.
In addition to investigating islands of practice within the field of computer science
education, I also investigated islands of practice between the related fields of computer
science education, educational technology, and education research proper. My research
question in this area follows: How do the proportions of quantitative, qualitative, and
mixed methods articles in computer science education compare to those proportions in
the fields of educational technology and education research proper?
Tedre (2006) explained that computer science is a field that is comprised, mainly,
of three traditions: a formalist tradition, an engineering tradition, and an empirical
tradition. I predicted that this engineering tradition would make itself most evident in
45
computer science education research, and to a lesser degree in education technology
(because it also consists of an engineering component; Ely [1999], one of the key figures
in education technology, calls it a "physical sciences component"), and reflected least in
education research proper. Here I assume that the number of papers that are program
descriptions (i.e., papers that do not empirically deal with human participants) is an
indicator of the degree of the engineering and formalist traditions in the fields of
computer science education, educational technology, and education research proper.
Specifically, if my prediction is correct then I would expect to find that computer
science education research forums have the highest proportions of program descriptions
(engineering) articles (e.g., I built this thing to these specifications types of articles),
educational technology forums would have the second highest proportions of program
descriptions articles, and that education proper forums would have the lowest proportions
of program descriptions article, but would have the highest proportion of empirical
articles dealing with human participants.
Biases
My background is in behavioral science research (particularly quantitative
education-research and program evaluation); therefore, I brought the biases of a
quantitatively trained behavioral scientist into this investigation. It is my belief that when
one does education-related research on human participants the conventions, standards,
and practices of behavioral research should apply; therefore, I approached this
methodological review from a behavioral science perspective. Nevertheless, I realize that
46
computer science education and computer science education research is a maturing,
multidisciplinary field, and I acknowledge that the behavioral science perspective is just
one of many valid perspectives that one can take in analyzing computer science education
research.
47
METHOD
Neuendorfs (2002) Integrative Model of Content Analysis was used as the model
for carrying out the proposed methodological review. Neuendorfs model consists of the
following steps: (a) developing a theory and rationale, (b) conceptualizing variables, (c)
operationalizing measures, (d) developing a coding form and coding book, (e) sampling,
(f) training and determining pilot reliabilities, (g) coding, (h) calculating final reliabilities,
and (i) analyzing and reporting data.
In the following subsections, I describe how I conducted each of the steps of
Neuendorfs model. Because the rationale (the first step in Neuendorfs model) was
described earlier, I do not discuss it below.
Conceptualizing Variables, Operationalizing Measures,
and Developing a Coding Form and Coding Book
Because this methodological review was the sixth in a series of methodological
reviews I had conducted (see Randolph et al., 2004; Randolph, 2005; Randolph, in press;
Randolph, Bednarik, & Myller, 2005; Randolph, Bednarik, Silander, et al., 2005; and
Randolph & Hartikainen, 2005), most of the variables had already been conceptualized,
measures had been operationalized, and coding forms and coding books had been created
in previous reviews. A list of the articles that were sampled are included in Appendix A.
The coding form and coding book that I used for this methodological review are included
as Appendices B and C, respectively.
48
Sampling
A proportional stratified random sample of 352 articles, published between the
years 2000 and 2005, were drawn, without replacement, from the eight major peer-
reviewed computer science education publications. (That sample size, 352, out of a finite
population of 1,306 was determined a priori, through the Sample Planning Wizard [2005]
and confirmed through resampling, to be large enough to achieve a +/- 5% margin of
error with a 95% level of statistical confidence if I were to treat all variables, and levels of
variables, as dichotomous, in the most conservative case — where/* and q = .5. This power
estimate refers to the aggregate sample, not to subsamples.) The sample was stratified
according to year and source of publication. Table 8, the sampling frame, shows the
number of papers (by year and publication) that existed in the population as I
operationalized it. Table 5 shows the number of articles that were randomly sampled (by
year and publication source) from each cell of the sampling frame presented in Table 9.
The articles that were included in this sample are listed in Appendix A.
The population was operationalized in such a way that it was a construct of what
typically is accepted as mainstream computer science education research. The population
did not include the marginal, grey areas of the literature such as unpublished reports,
program evaluation reports, or other nonpeer-reviewed publications because I was not
interested in the research practices reported in the entirety of computer science education
research. Rather, I was interested in research practices reported in current, peer-reviewed,
mainstream computer science education research forums.
49
Table 8
Sampling Frame
Year/forum
2000
2001
2002
2003
2004
2005
Total
Bulletin
31
21
21
40
36
38
187
CSE
17
17
17
17
17
15
100
JCSE
2
7
5
2
2
18
KOLI
14
10
13
21
25
83
SIGCSE
78
78
74
75
02
104
501
ITICSE
45
44
42
41
46
68
286
ICER
16
16
ACE
34
48
33
115
Total
171
176
171
225
262
301
1306
Table 9
Number of Articles Sampled from Each Forum and Year
Year/forum
2000
2001
2002
2003
2004
2005
Total
Bulletin
8
6
6
11
10
10
51
CSE
5
5
5
5
5
4
29
JCSE
2
1
3
KOLI
4
3
3
6
7
23
SIGCSE
21
21
20
20
25
28
135
ITICSE
12
12
11
11
12
13
76
ICER
4
4
ACE
9
13
9
31
Total
46
48
47
60
71
80
352
In general, nonpeer-reviewed articles or poster-summary papers (i.e., papers two
or fewer pages in length) were not included in the sampling frame. In Bulletin, only the
peer-reviewed articles were included; featured columns, invited columns, and working
group reports were excluded in the sampling frame of Table 8. In CSE and JCSE,
editorials and introductions were excluded. In the SIGCSE, ITICSE, ACE, and ICER
50
forums, only full peer-reviewed papers at least three pages in length were included; panel
sessions and short papers (i.e., papers two pages or less in length) were excluded. In Koli,
research and discussion papers were included; demo and poster papers were excluded.
Training and Determining Pilot Reliabilities
In this methodological review, an interrater reliability reviewer, who had
participated in previous methodological reviews, was trained in the coding book and
coding sheet, which are included as Appendices B and C. The interrater reliability
reviewer, Roman Bednarik, was a PhD student in computer science at the University of
Joensuu. He was chosen because he had significant knowledge of computer science,
computer science education, and quantitative research methodology and because he had
participated in previous methodological reviews of computer science education or
educational technology research. (Randolph, Bednarik, & Myller, 2005; Randolph,
Bednarik, Silander, et al., 2005). AlthouGH his knowledge and previous experience in
collaborating on methodological reviews meant that he required less coder training than if
a different coder had been chosen, it also meant that he was aware of my hypotheses
about computer science education research.
Initially the interrater reliability reviewer and I read through the coding book and
coding sheet together and discussed any questions that he had about the coding book or
coding sheet. When inconsistencies or ambiguities in the coding book or coding sheet
were found in the initial training session, the coding book or coding sheet was modified
to remedy those inconsistencies or ambiguities. Then the interrater reliability reviewer
51
was given a revised version of the coding book and coding sheet and was asked to
independently code a purposive pilot sample of 10 computer science education research
articles, which were not the same articles that were included in the final reliability
subsample. The purposive sample consisted of articles that I deemed to be representative
of the different types of research methods that were to be measured, articles that were
anecdotal only, and articles that did not deal with human participants. I, the primary
coder, also coded those 10 articles. After both of us had coded the 10 articles we came
together to compare our codes and to discuss the inconsistencies or unclear directions in
the coding book and coding sheet. When we had disagreements about article codes, we
would try to determine the cause of the disagreement and I would modify the coding book
if it were the cause of the disagreement. After pilot testing and subsequent improvement
of the coding book and the coding, the final reliability subsample was coded (see the
section entitled Calculating Final Reliabilities).
Since many of the variables in the coding book were the same as in previous
reviews (specifically, Randolph, 2005; Randolph, Bednarik, & Myller, 2005; Randolph,
Bednarik, Silander, et al., 2005), many of the pilot reliabilities had already been
estimated. The variables that had been used in previous reviews and already had estimates
of interrater reliabilities were methodology category; type of article, if not dealing with
human participants; whether an experimental or quasi-experimental design was used; type
of selection and assignment; psychometric information provided; type of experimental or
quasi-experiment; structure of the paper (i.e., report elements); measures; independent
variables; dependent variables; and moderating or mediating variables. (See Randolph,
52
2005; Randolph, Bednarik, & Myller, 2005; and Randolph, Bednarik, Silander, et al.,
2005 for previous estimates and discussions of interrater reliabilities for these variables.)
In general, all of the reliabilities for these variables were, or eventually became,
acceptable or the source of the unreliability had been identified and had been remedied in
the current coding book (see Randolph, Bednarik, & Myller). The only set of variables
whose reliabilities had not been pilot tested in previous methodological reviews dealt
with statistical practices or were demographic variables. Reliabilities for the demographic
characteristics, such as name of the first author, were not estimated since they were
objective facts.
Coding
Appendices B and C, which are the coding sheet and coding book, provide
detailed information on the coding variables, their origin, and the coding procedure.
Because the complete coding sheet and coding book are included as appendices, I will
only summarize them here.
Articles were coded in terms of demographic characteristics, type of article, type
of methodology used, type of research design used, independent variables examined,
dependent and mediating measures examined, moderating variables examined, measures
used, and statistical practices. In the rest of this section I describe the variables in the
coding book and their origin and history.
The first set of variables, demographic characteristics, consisted of the following
variables:
53
The case number,
The case number category (the first two digits of the case number),
Whether it was a case used for final reliability estimates,
The name of the reviewer,
The forum from which the article came,
The type of forum from which the article came (i.e., a journal or conference
proceedings),
The year the article was published,
The volume number where the article was published,
The issue in which the article was published,
The page number on which the article began,
The number of pages,
The region of the first author's affiliation,
The university affiliation of the first author,
The number of authors, and
The last name and first initials of the first author.
The variables in the second set, type of article, are listed below:
Kinnunen's categories;
Valentine's categories;
Whether the article dealt with human participants;
If the article did not deal with human participants, what type of article it was;
and
54
• If the article did deal with human participants, whether it presented only
anecdotal evidence or not.
The Kinnunen's categories variable was derived from Kinnunen (n.d.). The
Valentine's category variable was derived from Valentine (2004). The rest of the
variables in this section were originally derived from an emergent coding technique in
Randolph, Bednarik, Silander, and colleagues (2005) and then refined and used in
Randolph, Bednarik, and Myller (2005) before being refined again and used in the current
coding book.
The third set of variables, report structure, originated in the Parts of a Manuscript
section of the Publication Manual of the American Psychological Association (2001).
The exceptions are the grade level and curriculum year varaibles, which were suggested
by committee members during the proposal defense of this dissertation. The report
structure variables are listed below:
Type of abstract,
Introduction to problem present,
Literature review present,
Purpose/rational present,
Research questions/hypotheses present,
Adequate information on participants present,
Grade level of students,
Curriculum level taught,
Information about settings present,
55
• Information about instruments present,
• Information about procedure present, and
• Information about results and discussion present.
The fourth set of variables, methodology type, was developed from Gall, Borg,
and Gall (1996) and from the Publication Manual of the American Psychological
Association (APA, 2001). The explanatory descriptive and exploratory descriptive labels
came from Yin (1988). The descriptions of these variables in the coding book evolved
into their current form though Randolph (2005, in press), Randolph, Bednarik, and Myller
(2005), and Randolph, Bednarik, Silander, and colleagues. (2005). The assignment
variable originated from Shadish, Cook, and Campbell (2002). The methodology type
variables are listed below:
• Whether the article reported on an experimental or quasi-experimental
investigation or not,
Whether the article reported on an explanatory descriptive investigation or not,
Whether the article reported on an exploratory descriptive investigation or not,
Whether the article reported on a correlational investigation or not,
Whether the article reported on a causal-comparative investigation or not,
If there was not enough information to determine what type of method was
used, and
• The type of selection used.
The fifth set of variables, experimental research designs, relate to the articles that
reported on an experimental or quasi-experimental investigation. If experimental or
56
quasi-experimental investigations were reported, the type of experimental or quasi-
experimental design was noted. These research design variables were derived from
Shadish, Cook, and Campbell (2002) and from the Publication Manual of the American
Psychological Association (APA, 2001). These variables had been previously pilot tested
in Randolph (2005; in press), Randolph, Bednarik, and Myller (2005), and Randolph,
Bednarik, Silander, and colleagues (2005), except for the multiple factor variable, which
had not been previously pilot tested. The experimental research design variables are listed
below:
• If there was enough information to determine what experimental design had
been used if one had been used,
• If the researchers used a one-group posttest-only design,
• If the researchers used a posttest with controls design,
• If the researchers used a pre/posttest without controls design,
• If the researchers used a pre/posttest with controls design,
• If the researchers conducted a repeated measures investigation,
• If the researchers used a design that involved multiple factors, and
• If the researchers used a single-case design.
The sixth set of variables dealt with the type of independent variables that were
reported. These variables were derived through an emergent coding technique from
Randolph (2005) and Randolph, Bednarik, and Myller (2005). The binary independent
variables listed in the coding book for this set of variables are listed below:
• Student instruction,
57
Teacher instruction,
Computer science fair or contest,
Mentoring,
Listening to computer science speakers,
Computer science fields, and
Other types of interventions (open variable).
The seventh set of variables in the coding book dealt with the types of dependent
variables that were measured. These variables were based on codes that emerged from
Randolph (2005) and Randolph, Bednarik, and Myller (2005). The variables in this set
are listed below:
Attitudes (including self/reports of learning),
Attendance,
Achievement in core courses,
Achievement in computer science,
Teaching practices,
Students' intentions for the future,
Program implementation,
Costs,
Socialization,
Computer use, or
Other types of dependent variables (open variable).
The eighth set of variables dealt with the types of measures that computer science
58
educators used. These measurement variables were derived from codes that emerged in
Randolph (2005) and Randolph, Bednarik, and Myller (2005). Those binary measurement
variables are listed below:
Grades,
Student diaries,
Questionnaires,
Log files,
Teacher- or researcher-made tests,
Interviews,
Direct observation,
Standardized tests,
Student work,
Focus groups,
Existing records, or
Other types of measures (open variables).
Additionally whether any sort of psychometric information was provided for the variables
involving questionnaires, teacher- or researcher-made tests, direct observation, or
standardized tests.
The ninth set of variables involved mediating or moderating variables. In the
coding book this set of variables are called Factors (Non-manipulatable variables). This
set of variables was based on codes that emerged from Randolph (2005) and Randolph,
Bednarik, and Myller (2005). Those variables are listed below:
59
Gender,
Aptitude,
Race/ethnic origin,
Nationality,
Disability,
Socioeconomic status, and
Other types of dependent variables (open variables).
The tenth and final set of variables involved statistical practices. The statistical
practices variables dealt mainly with how inferential statistics and effect sizes were used
and reported. Particular emphasis was placed on whether informationally adequate
statistics were provided for a certain type of analysis. What was considered to be an
informationally adequate set of statistics is discussed in detail in the coding book. These
variables were based on the guidelines in Informationally Adequate Statistics section of
the Publication Manual of the American Psychological Association (APA, 2001). The
variables in that set are listed below:
• Whether quantitative results were reported,
• Whether inferential statistics were reported,
• Whether parametric tests were conducted and an informationally adequate set
of statistics were reported for them,
• Whether multivariate analyses were conducted and an informationally adequate
set of statistics was reported for them,
60
• Whether correlational analyses were conducted and an informationally
adequate set of statistics was reported for them,
• Whether parametric analyses were conducted and an informationally adequate
set of statistics was reported for them, and
• Whether analyses for small samples were conducted and an informationally
adequate set of statistics was reported.
In addition to the variables related to inferential practices, there was also a set of
variables about what types of effect sizes were reported. Those variables are listed below:
• Whether an effect size was reported,
• Whether a raw difference effect size was reported,
• Whether a standardized mean difference effect size was reported,
• Whether a correlational effect size was reported,
• Whether odds ratios were reported,
• Whether odds were reported, and
• Whether some other type of effect size other than the ones above were reported
(an open variable).
In terms of the coding procedure, the primary coder (the author of this
dissertation) used the coding sheet and coding book to code a stratified random sample of
352 articles. A subsample of 53 articles was selected randomly from those 352 articles
and electronic files of those 53 articles was given to the interrater reliability coder, who
also used the coding sheet and coding book to code those 53 articles. The primary coder
and interrater reliability coder did not converse about the coding process while the coding
61
was being done. After the coding was completed the primary coder merged the two sets
of codes for the subsample and calculated interrater reliability estimates. When there were
disagreements about the coding categories, the primary coder's judgment took precedent.
Variable-by-variable instructions for the coding procedure are given in the coding book.
Calculating Final Reliabilities
According to Neuendorf (2002), a reliability subsample of between 50 and 200
units is appropriate for estimating levels of interrater agreement. In this case, a simple
random reliability subsample of 53 articles was drawn from the sample of 352 articles.
Those 53 articles were coded independently by the interrater reliability reviewer so that
interrater reliabilities could be estimated.
Because the marginal amounts of each level of variables to be coded were not
fixed, Brennan and Prediger's (1981) free-marginal kappa (Km) was used as the statistic
of interrater agreement. (By fixed, I mean that there was not a fixed number of articles
that must be assigned to given categories. The marginal distributions were free. See
Brennan & Prediger, 1981.) Values of kappa lower than .4 were considered to be
unacceptable, values between .4 and .6 were considered to be poor, values between and
including .6 and .8 were considered to be fair, and values above .8 were considered to be
good reliabilities. Confidence intervals around kappa were found through resampling.
The resampling code that was used for creating confidence intervals around Km can be
found in Appendix D.
62
Data Analysis
To answer the primary research question, I reported frequencies for each of the
multinomial variables or groups of binominal variables. Confidence intervals (95%) for
each binary variable or multinomial category were calculated through resampling (see
Good, 2001; Simon, 1997), "an alternative inductive approach to significance testing,
now becoming more popular in part because of the complexity and difficulty of applying
traditional significance tests to complex samples" (Garson, 2006, n.p). The Resampling
Stats language (1999) was used with the Grosberg's (n.d.) resampling program.
Appendix E presents an example of Resampling Stats code that was used to calculate
confidence intervals around a proportion.
To answer the research questions that involved finding islands of practice, I took
two approaches. In the first approach, I cross tabulated the data for the 15 planned
contrasts, examined the adjusted residuals, and, for categorical variables calculated % 2 (see
Agresti, 1996) and found its probability through resampling. For ordinal variables, such
as year, I calculated M 2 (see Agresti) and found its probability through resampling. The
resampling codes for calculating x 2&n d M 2 from a proportionally stratified random sample
can be found in Appendix F. In the second approach, I used logistic regression to
determine the unique effect of the three predictor variables (i.e., forum type, region of
first author's affiliation, and year) on the five binary outcome variables (i.e., anecdotal-
only paper, experimental/quasi-experimental paper, explanatory descriptive paper,
63
attitudes-only paper, or one-group posttest-only paper) and to determine if there were
interactions between the variables.
To carry out the logistic regression, with SPSS 1 1.0, 1 followed the method
described in Agresti (1996). First, I found the best fitting logistic regression model for
each outcome variable by starting with the most complex model, which had the main
effects, all two-way interactions, and the one three-way interaction (i.e., I+R+Y+F+R
*Y+R*F+Y*F+R*Y*F; where I = intercept, R = region of first author's affiliation [a
categorical variable], F = forum type [journal or conference proceeding] [a categorical
variable], and Y = year), and then reducing the complexity of the model until the point
when the less-complex model would raise the difference in the deviances between the
two models to a statistically significant level. To determine if a less-complex model was
as good fitting as the more-complex model, I took the absolute value of the difference in
the -2 Log Likehood [hereafter deviance] and degrees of freedom between each model
and used the x 2 distribution to determine if there was a statistically significant increase in
the deviance. For example, if a full model had a deviance of 286.84 and 1 1 degrees of
freedom and the model without the three-way interaction had a deviance of 289.93 and 9
degrees of freedom, the difference between models would be 1 .09 in deviance and 2
degrees of freedom. The % 2 probability associated with those values is .58. Because the
difference was not statistically significant, I concluded that the less-complex model was,
more or less, as well fitting (i.e., it had about an equal amount of deviance) as the more-
complex model. I repeated this process until I found the least complex model that had a
deviance about equal to the deviance of the next most complex model. If the best fitting
64
model was overspecified (i.e., if the continuous, year variable was not in the best- fitting
model), I included the year variable nonetheless to fix the overspecification problem and
ran both analyses, with and without the continuous variable.
I relied on several methods to determine the overall fit of the model to the data. I
used SPSS's Omnibus Test of Model Coefficents (i.e., % 2 of the difference of the selected
model and the model with only a constant), which should be statistically significant if the
chosen model is better than the model with only a constant (Agresti, 1996). I also used
SPSS's version of the Hosmer and Lemeshow test, which breaks the data set into deciles
and computes the deviation between observed and predicted values. If the model fits
appropriately, the Hosmer and Lemeshow test should not be statistically significant
(Agresti). Also, I created scatterplots of the expected and observed probabilities. If visual
inspection of the plots showed that there were outliers, I ran regression analyses with and
without the outliers removed. Finally, I also examined the regression coefficients to
determine if the model seemed to fit the data. For example, if there were exponentiated
coefficients (odds ratios) in the thousands, I would use a different model or group the data
in a different way. To illustrate, in some cases I found that I had to group some of the
regions together to get enough cases in a category for the regression coefficients to make
sense.
65
RESULTS
Complications
To eliminate a significant rounding error when automating the resampling
analysis, I had to slightly overestimate the population size so that the ratio of population-
to-sample was an integer. Without this overestimation, the rounding error caused the
resampled parameter proportions to differ significantly from the sample
proportions — sometimes the two proportions would differ by as much as 5%. The actual
population to sample ratio was 3.71/1 (or 1,306/352), but in my analysis I rounded the
ratio's numerator to the next nearest integer, 4. In terms of my analyses, my estimate of
the finite population was 1,408 (4*352) instead of 1,306. The statistical consequences are
that overestimating the population will lead to slightly conservative results (Kalton,
1983); however, in this case the differences between using a population of 1,306 and
1,408 were negligible. Using Formula 1 1 of Kalton (p. 21) to manually estimate the
confidence intervals around a proportion, in this case around the proportion of human
participants variable, the proportion of the standard error when using a population of
1,306 (1.84) to the standard error when using a population of 1,408 (1.86) was 0.99. Or,
from a different viewpoint, the length of confidence intervals when using a population
size of 1,306 was 7.30 percentage units long and when using a population size of 1408
the length of the confidence interval was 7.21 percentage units long — a 9/100%
difference in the length of the confidence intervals.
66
According to Agresti (1996), regrouping data sometimes is necessary when
working with categorical data. In this case it was necessary to group the regions of first
author's affiliations together in order for certain statistical analyses, such as logistic
regression, to work. For example, in some of the logistic regression equations I had to
group the regional categories with the fewest cases into one group, because they had so
few observations at fine levels of analysis. Specifically, I sometimes grouped some of the
region of first author's affiliation categories — Africa, Asia-Pacific/Eurasia, and Middle
East — into one category that I called Asia-Pacific/Eurasia et al. My rationale for this
grouping is that although I could no longer make distinctions between African, Asian-
Pacific/Eurasian, and Middle Eastern papers, I could still compare papers from regions of
the world that contribute the most to the English language computer science education
literature — North America, Europe, and Asia-Pacifica/Eurasia et al. — at a fine level of
detail. (There was only one paper from an African institution, and none from South
American institutions, in the analysis of the planned contrasts.)
Interrater Reliability
Tables 10 through 20 present the number of cases (out of 53) that could be used to
calculate an interrater reliability statistic, the Km, and its 95% confidence intervals. In
short, the interrater reliabilities were good or fair (i.e., greater than .6) for most variables;
however, they were lower than .60 on seven variables: Kinnunen's categories; type of
paper, if not dealing with human participants; literature review present; setting adequately
described; procedure adequately described; and results and discussion separate. Five out
of seven variables with low reliabilities concern report elements.
Table 10
Interrater Reliabilities for General Characteristics Variables
67
General characteristics
Kappa
Lower CI
95%
Upper CI
95%
Kinnunen's categories
Valentine's categories
Human participants
Anecdotal
Type of 'other'
53
.40
.27
.55
53
.62
.48
.75
53
.81
.66
.96
34
.94
.82
1.00
17
.56
.27
.80
Table 11
Interrater Reliabilities for Research Methods Variables
Research method
Kappa
Lower CI
95%
Upper CI
95%
Experimental/quasi-experimental
Random assignment
Explanatory descriptive
Exploratory descriptive
Correlational
Causal-comparative
17
.88
.65
1.00
10
.70
.40
1.00
17
.65
.29
1.00
17
.88
.65
1.00
17
1.00
17
.88
.65
1.00
Table 12
Interrater Reliabilities for Experimental Design Variables
Type of experimental design
Kappa
Lower CI
95%
Upper CI
95%
One-group posttest-only
Posttest with controls
Pretest/posttest with controls
Group repeated measures
Multiple factor
Single case
10
1.00
10
.80
.40
.10
10
.80
.40
.10
10
.80
.40
.10
10
1.00
10
1.00
Table 13
Interrater Reliabilities for Independent Variables
68
Type of independent variable used
Kappa
Lower CI
95%
Upper CI
95%
Student instruction
Teacher instruction
Mentoring
Speakers at school
Field trips
Computer science fair/contest
10
1.00
10
1.00
10
1.00
10
1.00
10
1.00
10
1.00
Table 14
Interrater Reliabilities for Type of Dependent Variable Measured
Lower CI
Upper CI
Type of dependent variable measured
n
Kappa
95%
95%
Attitudes (student or teacher)
15
1.00
Achievement in computer science
15
.60
.20
1.00
Attendance
15
.87
.60
1.00
Other
15
.72
.33
1.00
Computer use
15
.87
.60
1.00
Students' intention for future
15
1.00
Teaching practices
15
.87
.60
1.00
Achievement in core (non-cs) courses
15
1.00
Socialization
15
1.00
Program implementation
15
1.00
Costs and benefits
15
1.00
Table 15
Interrater Reliabilities for Grade Level and Undergraduate Year
Grade level of participant
Kappa
Lower CI
95%
Upper CI
95%
Grade level
Undergraduate year
.39
1.00
.02
.75
69
Table 16
Interrater Reliabilities for Mediating or Moderating Variables
Lower CI
Upper CI
Mediating or moderating variable
n
Kappa
95%
95%
Mediating/moderating factor examined
15
.71
.33
1.00
Gender
6
1.00
Nationality
6
1.00
Aptitude (in computer science)
6
1.00
Race/ethnic origin
6
1.00
Disability
6
1.00
Socioeconomic status
6
1.00
Other
6
1.00
Table 17
Interrater Reliabilities for Type of Effect Size Reported Variables
Lower CI
Upper CI
Type of effect size reported
n
Kappa
95%
95%
Effect size reported
15
1.0
Raw difference
14
1.0
Variability reported with means
9
1.0
Correlational effect size
14
1.0
Standardized mean difference
14
1.0
Odds ratio
14
1.0
Odds
14
1.0
Relative risk
14
1.0
70
Table 18
Interrater Reliabilities for Type of Measure Used Variables
Lower CI
Upper CI
Type of measure used
n
Kappa
95%
95%
Questionnaires
15
.72
.33
1.00
Reliability or validity information
6
1.00
Grades
15
.87
.60
1.00
Teacher- or researcher-made tests
15
.72
.33
1.00
Reliability or validity information
5
.60
-.19
1.00
Student work
15
.60
.20
1.00
Existing records
15
.87
.60
1.00
Log files
15
.72
.33
1.00
Standardized tests
15
.87
.60
1.00
Reliability or validity information
1
1.00
Interviews
15
.87
.60
1.00
Direct observation
15
1.00
Reliability or validity information"
Learning diaries
15
1.00
Focus groups
15
1.00
Other
15
.87
.60
1.00
a No interrater reliability cases available.
Table 19
Interrater Reliabilities or Type of Inferential Analyses Variables
Type of inferential analysis used
Kappa
Lower CI
95%
Upper CI
95%
Inferential analyses used
Parametric analysis
Measure of centrality and dispersion
reported
Correlational analysis
Sample size reported
Correlation or covariance matric reported
Nonparametric analysis
Raw data summarized
Small sample analysis
Entire data set reported 2
Multivariate analysis
Cell means reported"
Cell sample size reported"
Pooled within variance or covariance
matrix reported 8
15
1.00
4
1.00
2
1.00
4
1.00
1
1.00
1
1.00
4
1.00
1
1.00
1
1.00
1.00
No interrater reliaibility cases available.
71
Table 20
Interrater Reliabilities for Report Element Variables
Lower CI
Upper CI
Report element
n
Kappa
95%
95%
Abstract present
15
.87
.60
1.00
Problem is ontroduced
15
.87
.60
1.00
Literature review present
15
.47
.07
.87
Research questions/hypotheses stated
15
.60
.20
1.00
Purpose/rationale
15
.06
-.33
.47
Participants adequately described
15
.72
.33
1.00
Setting adequately described
15
.47
.07
.87
Instrument adequately described
1
1.00
Procedure adequately described
15
.47
.07
.87
Results and discussion separate
15
.47
.07
.87
Aggregated Results
In this subsection I present the aggregate findings. Note that in tables of groups of
binomial variables, the column marginals do not sum to the total because one or more
attributes could have applied. For example, an article could have used mixed-methods
and could have been an experimental and explanatory descriptive type of article at the
same time.
General Characteristics
Forum where article was published. Figure 4, which presents again the
information in Table 9 collapsed across years, is a pie chart of the relative proportions of
articles included in the sample, by forum. Note that Bulletin is the label for the June and
December issues of SIGCSE bulletin; CSE is the label for the journal — Computer Science
72
ACE
Figure 4. Proportions of articles published in each forum.
Education; JCSE is the label for the Journal of Computer Science Education Online;
SIGCSE is label for the Proceedings of the SIGCSE Technical Symposium, which is
published in the March Issue of SIGCSE Bulletin; ITiCSE is the label for the Proceedings
of the Innovation and Technology in Computer Science Education Conference, which is
published in the September issue of SIGCSE Bulletin; Koli is the label for the Koli
Calling: Finnish/Baltic Sea Conference on Computer Science Education; ACE is the
label for the Proceedings of the Australasian Computing Education Conference; and
ICER is the label for the International Computer Science Education Research Workshop.
The three forums that had published the most articles from 2000-2005 (SIGCSE, ITiCSE,
and Bulletin) are all publications that are published by ACM in SIGCSE Bulletin.
73
When aggregating the forums into journals or conference proceedings, 289
(76.4%) were published in conference proceedings and 83 (23.6%) were published in
journals. (In this case, Bulletin, CSE, and JCSE were considered to be journals and the
other forums were considered to be conference proceedings.)
First authors whose articles were most frequently sampled. The first author whose
articles were most frequently selected in this random sample was Ben-David Kollikant,
with four articles. Other first authors whose articles were also frequently selected were
A.T. Chamillard, Orit Hazzan, David Ginat, H. Chad Lane, and Richard Rasala, each
with three articles in the sample.
First authors ' affiliations. The authors of the articles in the selected sample
represented 242 separate institutions. Of those 242 institutions, 207 were universities or
colleges; 24 were technical universities, institutes of technology, or polytechnics; and 1 1
were other types of organizations, like research and evaluation institutes or centers. The
majority of articles have first authors whom are affiliated with organizations in the U.S.
or Canada.
Table 21 shows the 12 institutions that were most often randomly selected into the
sample. The number of articles that should correspond with the number of articles in the
population can be estimated by multiplying the number of articles in the sample for each
institution by 3.71, which is the ratio of the number of articles in the population to the
number of articles in the sample. The University of Joensuu, with 13 articles included in
the sample, was an outlier. Of those 13 articles, 1 1 were from the Koli Conference, a
conference held in a remote location near Joensuu.
74
Table 21
Institutions with Greatest Number of Articles
Number of articles
Institution
in sample
Proportion
University of Joensuu
13
3.7
Technion - Israel Institute of Technology
6
1.7
Drexel University
5
1.4
Northeasern University
5
1.4
Tel-Aviv University
5
1.4
Weizmann Institute of Science
5
1.4
Helsinki University of Technology
4
1.1
Michigan Technological University
4
1.1
Trinity College
4
1.1
University of Arizona
4
1.1
University of Technology, Sydney
4
1.1
Virginia Tech
4
1.1
Other institutions
289
82.4
Total
352
100.0
Median number of authors per articles. The median number of authors on each of
the 352 articles was 2, with a minimum of 1 and a maximum of 7. The 2.5 th and 95 th
percentiles of the median from 100,000 samples of size 352 were 5 and 5.
Median number of pages per article. Of the 349 articles that had page numbers,
the median number of pages in the sample was 5, with a minimum of 3 and a maximum
of 37. The 2.5th and 97.5th percentiles of the median from 10,000 samples of
size 349 were 5 and 5.
Report elements. Table 22 shows the proportion of articles that had or did not
have report elements that are considered by the American Psychological Association to be
needed in empirical, behavioral papers. Note that the interrater reliabilities for the
75
Table 22
Proportions of Report Elements
Report element
(of 123)
%
Lower CI Upper CI
95% 95%
Abstract present
Problem is introduced
Literature review present
Purpose/rationale stated
Research questions/hypotheses stated
Participants adequately described
Setting adequately described
Instrument adequately described 1 "
Procedure adequately described
Results and discussion separate
122
99.2
98.4
100.0
119
96.7
94.3
99.2
89
72.4
65.9
78.1
45
36.6
30.8
42.3
27
22.0
16.3
27.6
56
45.5
39.0
52.0
79
64.2
58.5
69.9
66
58.4
52.2
64.6
46
37.4
30.9
43.9
36
29.3
23.6
35.0
Note. Column marginals do not sum to 144 (or 100%) because more than one methodology type per article
was possible.
a Of 113.
literature review present, purpose/rationale stated, setting adequately described, procedure
adequately described, and results and discussion separate variables were low.
Kinnunen 's content categories. Table 23 shows how the articles were distributed
according to Kinnunen's categories for describing the content of computer science
education articles. It shows that the most frequently occurring type of content had to do
with a new way to organize a course. Note that the interrater reliability for this variable
was poor.
Valentine's research categories. Table 24 shows how the sampled articles were
distributed into Valentine's research categories. Experimental and Marco Polo were the
most frequently seen types of articles.
Human participants. Of the 352 articles in this sample, the majority of articles
dealt with human participants. See Table 25.
76
Table 23
Proportions of Articles Falling into Each ofKinnunen 's Categories
Lower CI
Upper CI
Content category
n
%
95%
95%
New way to organize a course
175
49.7
45.7
54.0
Tool
66
18.8
15.3
22.2
Other
56
15.9
13.1
19.0
Teaching programming languages
31
8.8
6.5
11.4
Paraallel computing
10
2.8
1.4
4.3
Curriculum
5
1.7
0.6
2.8
Visualization
5
1.7
0.6
2.8
Simulation
2
0.6
0.0
1.1
Total
352
100.0
Table 24
Proportions of Articles Falling into Each of Valentine's Categories
Valentine's category
%
Lower CI
95%
Upper CI
95%
Experimental
Marco Polo
Tools
Philosophy
Nifty
John Henry
Total
144
40.9
36.7
44.9
118
33.5
29.7
37.5
44
12.5
9.7
15.3
39
11.1
8.5
13.6
7
2.0
0.9
3.1
0.0
352
100.0
Table 25
Proportion of Articles Dealing with Human Participants
Human participants
%
Lower CI
95%
Upper CI
95%
Yes
No
Total
233
66.2
62.2
70.1
119
33.8
29.8
37.8
352
100.0
77
Grade level of participants . Table 26 shows the grade level of participants of the
123 articles that dealt with human participants, that were not explanatory descriptive
only, and that presented more than anecdotal evidence (hereafter these 123 articles are
called the behavioral, quantitative, and empirical articles). Bachelor's degree students
were overwhelmingly the type of participants most often investigated in the articles in
this sample.
As Table 27 shows, of the 64 Bachelor's degree participants, most were taking
first-year computer science courses at the time the study was conducted. Studies in which
the participants were not students (e.g., teachers) or the participants were of mixed
grade levels were included in the mixed level/other category. (Note that the interrater
reliability for the grade level of participants variable, but not the undergraduate year
variable, was below a kappa of .4).
Anecdotal evidence only. Of the 233 articles that dealt with human participants,
38.2% presented only anecdotal evidence. See Table 28.
Types of articles that did not deal with human participants. Of the 119 articles
that did not deal with human participants, the majority were purely descriptions of
interventions. See Table 29, which shows the proportions of those articles that were
program descriptions; theory, methodology, or philosophical papers; literature reviews; or
technical papers. (Note that the interrater reliability estimate of kappa for this variable
was below .6.)
Table 26
Proportions of Grade Level of Participants
78
Grade level of participant
%
Lower CI
95%
Upper CI
95%
Preschool
K-12
Bachelor's level
Master's level
Doctoral lavel
Mixed level/other
Total
2
2.3
0.0
5.7
5
5.7
2.3
10.2
64
72.7
64.8
80.7
1
1.1
0.0
3.4
0.0
16
18.2
11.4
25.0
88
100.0
Table 27
Proportion of Undergraduate Level of Computing Curriculum
Year of undergraduate level
computing curriculum
%
Lower CI
95%
Upper CI
95%
First year
Second year
Third year
Fourth year
Total
39
70.9
61.8
80.0
3
5.5
1.8
90.9
8
14.5
7.3
2.2
5
9.1
3.6
14.6
64
100.0
Table 28
Proportion of Human Participants Articles that Provide Anecdotal Evidence Only
Lower CI
Upper CI
Anecdotal
n
%
95%
95%
Yes
89
38.2
33.1
43.3
No
144
61.8
56.7
66.5
Total
233
100.0
Table 29
Proportions of Types of Articles Not Dealing With Human Participants
Type of article
%
Lower CI
95%
79
Upper CI
95%
Program description
Theory, methodology, or
Philosophical paper
Literature review
Technical
Total
72
60.5
53.8
67.2
36
30.3
24.4
37.0
10
8.4
5.0
11.8
1
0.8
0.0
1.7
19
100.0
Types of Research Methods and
Research Designs Used
Types of research methods used. Table 30 shows that the experimental/quasi-
experimental methodology type was the most frequently used type of methodology in the
articles that dealt with human participants and that presented more than anecdotal
evidence. Table 31 shows the proportions of quantitative articles (i.e., not explanatory
descriptive), qualitative articles (i.e., only explanatory descriptive), and mixed-methods
articles (i.e., explanatory descriptive and one or more of the following: experimental/
quasi-experimental, exploratory descriptive, correlational, causal-comparative).
In terms of the 144 studies that dealt with human participants and that presented
more than anecdotal evidence, convenience sampling of participants was used in 124
(86.1%) of the cases, purposive (nonrandom) sampling was used in 14 (9.7%) of the
cases. Random sampling was used in 6 (4.2%) of the cases.
Research designs. Table 32 shows that the most frequently used research design
was the one-group posttest-only (i.e., the ex post facto design) design. Of the 51 articles
80
Table 30
Proportion of Methodology Types Used
Lower CI
Upper CI
Methodology types
n
%
95%
95%
Experimental/quasi-experimental
93
64.6
58.3
70.8
Explanatory descriptive
38
26.4
20.8
31.3
Causal comparative
26
18.1
13.2
22.9
Correlational
15
10.4
7.0
14.6
Exploratory descriptive
11
7.6
4.2
11.1
Table 31
Proportion of Types of Methods
Type of method
%
Lower CI
95%
Upper CI
95%
Quantitative
Qualitative
Mixed
Total
107
74.3
68.1
80.2
22
15.3
10.4
20.8
15
10.4
6.3
14.6
144
100.0
Table 32
Proportions of Types of Experimental/Quasi-Experimental Designs Used
Type of experimental design
%
Lower CI
95%
Upper CI
95%
Posttest only
posttest with controls
Pretest/posttest without controls
Repeated measures
Pretest/posttest with controls
Single-subject
51
54.8
47.3
62.4
22
23.7
17.2
30.1
12
12.9
8.6
18.3
7
7.5
4.3
11.8
6
6.5
2.2
10.8
3
3.2
1.1
5.3
Note. Column marginals do not sum to 93 (or 100%) because more than one methodology type per article
was possible.
81
that used the one-group posttest-only design, 46 articles used it exclusively (i.e., they did
not use a one-group posttest-only design and a research design that incorporated a pretest
or a control of contrast group).
In the sampled articles, quasi-experimental studies were much more frequently
conducted than truly experimental studies. Of the 93 studies that used an experimental or
quasi-experimental methodology, participants self-selected into conditions in 81 (87.1%)
of the studies, participants were randomly assigned to conditions in 7 (7.5%) of the
studies, and participants were assigned to conditions purposively, but not randomly, by
the researched s) in 5 (5.4%) of the studies.
Independent, Dependent, and Moderating/
Mediating Variables Investigated
Independent variables. Table 33 shows the proportions of types of independent
variables that were investigated in the 93 articles that used an experimental/quasi-
experimental methodology. Nearly 99% of all independent variables were related to
student instruction.
Dependent variables. Table 34 shows the proportions of the different types of
dependent variables that were measured in the 123 behavioral, quantitative, and empirical
articles. Table 34 shows that attitudes and achievement in computer science were the
dependent variables that were most frequently measured. The variables project
implementation and costs and benefits, although included as categories on the coding
sheet are not included in Table 34 because there were no studies that used them as
dependent measures.
82
Table 33
Proportion of Types of Independent Variables Used
Type of independent variable used
n
(93)
%
Lower CI
95%
Upper CI
95%
Teacher instruction
Mentoring
Speakers at school
Field trips
Computer science fair/contest
92
98.9
96.8
1.0
4
4.3
2.2
6.5
2
2.2
0.0
5.3
2
2.2
0.0
5.3
1
1.1
0.0
2.2
0.0
Note. Column marginals do not sum to 93 (or 100%) because more than one type of independent variable
could have been used in each article (e.g., when there were multiple experiments).
Table 34
Proportions of Types of Dependent Variables Measured
N
Lower CI
Upper CI
Type of dependent variable measured
(of 123)
%
95%
95%
Attitudes (student or teacher)
74
60.2
53.7
66.7
Achievement in computer science
69
56.1
49.6
62.6
Attendance
26
21.1
15.5
28.3
Other
14
11.5
7.4
15.6
Computer use
5
4.1
1.6
6.5
Students' intention for future
3
2.4
0.1
4.9
Teaching practices
2
1.6
0.0
3.3
Achievement in core (non-cs) courses
1
0.8
0.0
2.4
Socialization
1
0.8
0.0
2.4
Note. Column marginals do not sum to 123 (or 100%) because more than one type of dependent variables
could have been measured.
Mediating or moderating variables examined. Of the 123 behavioral, quantitative,
and empirical articles; moderating or mediating variables were examined in 29 (23.6%).
Table 35 shows the types and proportions of moderating or mediating variables that were
examined in the sample of articles. There were many articles that examined moderating
83
Table 35
Proportions of Mediating or Moderating Variables Investigated
Mediating or moderating variable
11
Lower CI
Upper CI
investigated
(of 29)
%
95%
95%
Gender
6
20.7
13.8
27.6
Grade level"
4
13.8
6.9
20.7
Learning styles"
4
13.8
6.9
20.7
Aptitude (in computer science) 3
2
6.8
3.5
10.3
Major/minor subject 1 *
2
6.8
3.5
10.3
Race/ethnic origin
2
6.8
3.5
10.3
Age a
3.4
0.0
6.9
Amount of scaffolding provided"
3.4
0.0
6.9
Frequency of cheating"
3.4
0.0
6.9
Pretest effects"
3.4
0.0
6.9
Programming language"
3.4
0.0
6.9
Type of curriculum"
3.4
0.0
6.9
Type of institution"
3.4
0.0
6.9
Type of computing laboratory"
3.4
0.0
6.9
Type of grading (human or computer")
3.4
0.0
6.9
Self-efficacy"
3.4
0.0
6.9
Note. Column marginals do not sum to 29 (or 100%) because more than one methodology tpe per article
was possible.
"These items were not a part of the original coding categories.
or mediating variables that fit into the other category (i.e., they were not originally on the
coding sheet); those other variables were tabulated and have been incorporated into Table
35. Although included on the coding sheet, the variables — disability and socioeconomic
status — were not included in Table 34 because no study examined them as mediating or
moderating variables.
Types of Measures and Statistical Practices
Types of measures used. Table 36 shows the proportions of types of measures that
were used in the 123 behavioral, quantitative, and empirical articles. Note that
questionnaires were clearly the most frequently used type of measure. Measurement
84
Table 36
Proportions of Types of Measures Used
11
Lower CI
Upper CI
Type of measure used
(of 123)
%
95%
95%
Questionnaires
65
52.8
46.3
59.4
Grades
36
29.3
23.6
35.0
Teacher- or researcher-
•made tests
27
22.0
16.3
27.6
Student work
22
17.9
13.0
23.6
Existing records
20
16.3
11.4
21.1
Log files
15
12.2
8.1
9.2
Standardized tests
11
8.9
4.9
13.0
Interviews
8
6.5
3.3
9.8
Direct observation
4
3.3
0.8
5.7
Learning diaries
4
3.3
0.8
5.7
Focus groups
3
2.4
0.8
4.9
Note. Column marginals do not sum to 123 because more than one meaasure per article was possible.
validity or reliability data were provided for questionnaires in 1 of 65 (1.5 %) of articles,
for teacher- or researcher-made tests in 5 of 27 (18.5 %) of articles, for direct observation
(e.g., interobserver reliability) in 1 of 4 (25%) of articles, and for standardized tests in 6
of 11 (54.5%) of articles.
Type of inferential analyses used. Of the 123 behavioral, quantitative, and
empirical articles, inferential statistics were used in 44 (35.8%) of them. The other 79
articles reported quantitative results, but did not use inferential analyses. Table 37 shows
the types of inferential statistics used, their proportions, and the proportion of articles that
provided statistically adequate information along with the inferential statistics that were
reported.
Type of effect size reported. Of the 123 behavioral, quantitative, and empirical
articles, 120 (97.6%) reported some type of effect size. In the three articles that reported
85
Table 37
Proportions of Types of Inferential Analyses Used
Lower CI Upper CI
Type of inferential analysis used n % 95% 95%
Parametric analysis (of 44) 25 56.8 47.7 65.9
Measure of centrality and dispersion
Reported (of 25) 15 60.0 48.0 72.0
Correlational analysis (of 44)
13
29.5
23.3
37.2
Sample size reported (of 13)
10
76.9
53.9
92.3
Correlaction or covariance matrix reported
(of 13)
5
38.5
15.4
61.5
Nonparametric analysis (of 44)
11
25.0
13.2
31.8
Raw data summarized (of 1 1)
8
72.7
45.6
90.9
Small sample analysis (of 44)
2
4.5
0.0
9.1
Entire data set reported (of 2)
0.0
Multivariate analysis (of 44)
1
2.3
0.0
2.3
Cell means reported (of 1)
0.0
Cell sample size reported (of 1)
0.0
Pooled within variance or covariance
Matrix reported (of 1)
0.0
Note. Column marginals do not sum because more than one methodology type per article was possible.
quantitative statistics but not an effect size, those articles presented only probability
values or only reported if the result was "statistically significant" or not. Table 38
presents the types of effect sizes that were reported and their proportions. Odds, odds
ratio, or relative risk were not reported in any of the articles in this sample. Of the
articles that reported a raw difference effect size, 74 of those reported the raw difference
as a difference between means (the rest were reported as raw numbers, proportions,
means, or medians). Of the 74 articles that reported means, 29 (62.5%) did not report a
measure of dispersion along with the mean. Note that a liberal definition of a raw
86
Table 38
Proportions of Types of Effect Sizes Reported
Type of effect size reported
n
(of
Lower CI
Upper CI
1203)
%
95%
95%
117
97.5
95.0
100.0
8
6.7
3.3
6.7
6
5.0
1.7
8.3
Raw difference
Correlational effect size
Standardized mean difference
Note. Column marginals do not sum to 120 (or 100%) because more than one methodology type per
article was possible.
difference-also referred to as relative risk or a gain score — was used here. The authors
did not actually have to subtract pretest and posttest raw scores (or pretest and posttest
proportions) from one another to be considered a raw difference effect size. They simply
had to report two raw scores in such a way that a reader could subtract one from another
to get a raw difference.
Islands of Practice: Analysis of Crosstabulations
In this section I present the crosstabulated results for the 15 planned contrasts. Of
the 15 contrasts, only the contrasts that were significant at the .003 probability level and
the contrasts regarding the difference between articles published in papers and
conferences are discussed in detail here. However, I do present crosstabulations for each
of the 15 contrasts. Note that the probability level that corresponds with an overall
probability level across the 15 contrasts of .05 is .003; see Stevens, 1999.
87
Differences between Journal and Conference
Proceedings Articles
The results of these crosstabulation analyses show that there were no statistically
significant differences between journal and conference proceedings articles in terms of
several methodological attributes. Those attributes were the proportion of articles that
provided anecdotal-only evidence, the proportion of articles that used an experimental or
quasi-experimental method, the proportion of articles that used an explanatory descriptive
method, the proportions of articles that used a one-group posttest-only research design
exclusively, and the proportion of articles that examined attitudes as the only dependent
variable. However, using the logistic regression approach it was found that there was a
statistically significant difference, at the .10 alpha level, in the proportion of
experimental/ quasi-experimental articles when a forum type by region interaction term in
included in the model.
Anecdotal-only articles. Table 39 presents the frequencies and percentages of
articles that dealt with human participants but only presented anecdotal evidence. The
journal articles in this sample had 8.8% more anecdotal-only articles than conference
articles; the difference in the overall observed cell deviations from the expected cell
deviations was not statistically significant, % 2 (1, N = 233) = 1.32,/? = .251; resampled
p = .256.
In the case of Table 39, the adjusted residuals are small, which is congruent with
the finding that % 2 was not statistically significant. According to Agresti, "an adjusted
88
Table 39
Crosstabulation of Anecdotal-Only Papers in Conferences and Journals
Anecdotal
-only
Percentage
Adjusted
Forum
Yes
No
Total
yes
residual
Conference
66
116
182
36.3
-1.1
Journal
23
28
51
45.1
1.1
Total
89
144
233
38.2
residual that exceeds about 2 or 3 in absolute value indicates lack of fit (of the null
hypothesis) in that cell" (1996, pp. 31-32).
Experimental/quasi-experimental articles. Table 40 presents the frequencies and
percentages of articles that reported on experimental or quasi-experimental investigations.
Journal articles had 4.1% more experimental/quasi-experimental investigations than did
conference articles; the difference between journal articles and conference articles was
not statistically significant, %\\, N=144) = 0.16, p = .687; resampled/? = .672. (See the
logistic regression approach section for an alternate finding when a region by forum type
interaction is controlled for.)
Explanatory descriptive articles. Journal articles had 7.1% more explanatory
descriptive articles than did articles published in conference proceedings. This difference
was not statistically significant, % 2 (\, N=144) = 0.59, p = .441; resampled/? = .426. (See
Table 41.)
89
Table 40
Crosstabulation of Experimental Papers in Conferences and Journals
Experimental
Percentage
Adjusted
Forum
Yes
No
Total
yes
residual
Conference
74
42
116
63.8
-0.4
Journal
19
9
28
67.9
0.4
Total
93
51
144
64.6
Table 41
Crosstabulation of Explanatory Descriptive Papers in Conferences and Journals
Explanatory descriptive
Percentage
Adjusted
Forum
Yes No
Total
yes
residual
Conference
29 87
116
25.0
-0.8
Journal
9 19
28
32.1
0.8
Total
38 106
144
26.4
Attitudes-only articles. Table 42 indicates that journals had 5.9% less articles that
examined only attitudes than conference proceedings. The difference was not statistically
significant, % 2 (3, N= 123) = 0.31, p = .580; resampled/? = .579.
One-group posttest-only articles. Table 43 shows the proportions of conference
and journal articles that used one-group posttest-only research designs only and those that
used designs with controls. Conference proceedings had 2.6% more articles that used the
one-group posttest-only design exclusively than did journal articles. The difference was
not statistically significant, % 2 (\, N= 93) = 0.04, p = .838; resampled/? = .835.
90
Table 42
Crosstabulation of Attitudes-Only Papers in Conferences and Journals
Attitudes
-only
Percentage
Adjusted
Forum
Yes
No
Total
yes
residual
Conference
32
68
100
32.0
0.6
Journal
6
17
23
26.1
-0.6
Total
38
85
123
30.9
Table 43
Crosstabulation of Experimental Papers That Used Posttest-Only Designs Exclusively
Posttest
-only exclusively
Percentage
Adjusted
Forum
Yes
No
Total
yes
residual
Conference
37
37
74
50.0
0.2
Journal
9
10
19
47.4
-0.2
Total
46
47
93
49.5
Yearly Trends
Out of the five planned contrasts involving yearly trends, two were statistically
significant. The number of anecdotal articles and the number of explanatory descriptive
articles had decreased from 2000 to 2005. Anecdotal-only articles. Table 44 shows that
there was a decreasing trend in the number of anecdotal-only articles from 2000-2005.
The fact that the adjusted residuals in the Percentage Yes column transition, more or less,
from large positive values in 2000 to large negative values in 2005 and that the
percentages, more or less, transition from larger to smaller support the finding that there
91
Table 44
Anecdotal-Only Papers by Year
Anecdotal-only
Percentage
Adjusted
Year
Yes No
Total
yes
residual
2000
18 13
31
58.1
2.4
2001
15 15
30
50.0
1.4
2002
9 17
26
34.6
-0.4
2003
14 25
39
35.9
-0.3
2004
18 34
52
34.6
-0.6
2005
15 40
55
27.3
-1.9
Total
89 144
233
was a trend. The trend was statistically significant, M 2 (l, N = 233) = 9.00, p = .003;
resampled/? = .003.
Explanatory descriptive articles. Table 45 shows that there was a somewhat
decreasing trend in the number of explanatory descriptive articles that were published
each year. Although the trend was not consistent (2002 was an exception to the trend), it
was statistically significant, M\\, N = 144) = 11.54,/? = .001; resampled/? < .000.
Other types of articles. Crosstabulations for the types of articles where there was
not a statistically significant trend (i.e., experimental/quasi-experimental articles, one-
group posttest-only articles, and attitudes-only articles) are presented below. Table 46
shows that there was not a strong trend in the number of experimental/quasi-experimental
papers that were published each year. Likewise for Table 47, which shows the number of
one-group posttest-only articles per year, and for Table 48, which shows the number of
92
Table 45
Explanatory Descriptive Papers by Year
Explanatory descriptive
Percentage
Adjusted
Year
Yes
No
Total
yes
residual
2000
7
6
13
53.8
2.4
2001
4
11
15
26.7
0.0
2002
8
9
17
47.1
2.1
2003
7
18
25
28.0
0.2
2004
9
25
34
26.5
0.0
2005
3
37
40
7.5
-3.2
Total
38
106
144
Table 46
Experimental/Quasi-Experimental Papers by Year
Experimental
Percentage
Adjusted
Year
Yes
No
Total
yes
residual
2000
8
5
13
61.5
-0.2
2001
11
4
15
73.3
0.7
2002
10
7
17
58.8
-0.5
2003
14
11
25
56.0
-1.0
2004
22
12
34
64.7
0.0
2005
28
12
40
70.0
0.8
Total
93
51
144
Note.Af(l,N =
- 144) =
= 0.17,p =
.676;
resampled/? = .676.
Table 47
One-Group Posttest-Only Papers by Year
Anecdotal-only
Percentage
Adjusted
Year
Yes
No
Total
yes
residual
2000
6
2
8
75.0
1.5
2001
6
5
11
54.5
0.4
2002
4
6
10
40.0
-0.6
2003
4
10
14
28.6
-1.7
2004
15
7
22
68.2
2.0
2005
11
17
28
39.3
-1.3
Total
46
47
93
93
Table 48
Attitudes-Only Papers by Year
Attitudes'
-only
Percentage
Adjusted
Year
Yes
No
Total
yes
residual
2000
1
8
9
11.1
-1.3
2001
6
7
13
46.2
1.3
2002
3
9
12
25.0
-0.5
2003
5
17
22
22.7
-0.9
2004
12
17
29
41.4
1.4
2005
11
27
38
28.9
-0.3
Total
38
85
123
Note. M\\,N= 93) = 0.97, p = .326; resampled;? = .315.
attitudes-only papers by year. There was not strong evidence that there was a trend
between the years 2000 and 2005.
Region of First Author 's Affiliation
Of the five contrasts that dealt with the region of first author's affiliation, three
were statistically significant. The statistically significant findings are described below.
Experimental/quasi-experimental articles. Table 49 shows that first authors who
were affiliated with institutions in North America tend to write, and get published,
articles that used experimental or quasi-experimental articles. In contrast, first authors
who were affiliated with institutions in Europe or in the Middle East tended not to write,
or get published, experimental or quasi-experimental articles. In fact, the odds of a first
author affiliated with a North American association having published an experimental
paper were more than 3.6 times greater than a first author affiliated with a European
institution and
94
Table 49
Experimental Papers by Region of First Author 's Affiliation
Experimental/
Quasi-
■experimental
Percentage
Adjusted
Region
Yes
No
Total
yes
residual
Eurasia
20
10
30
66.7
0.3
Europe
14
16
30
49.7
-2.3
Middle East
4
9
13
30.8
-2.6
North America
54
16
70
77.1
3.1
Total
92
51
143
more than 7.5 times greater than a first author affiliated with a Middle Eastern institution.
The differences between observed and expected cell values in Table 49 were statistically
significant, % 2 (3, N= 143) = 15.54, p = .001; resampled/? < .000.
Explanatory descriptive articles. Table 50 shows that first authors who were
affiliated with a Middle Eastern institution tended to write and get published explanatory
descriptive articles. The odds of a first author affiliated with a Middle Eastern institution
having written and gotten published an explanatory descriptive articles was more than 13
times greater than the odds of their counterpart affiliated with a North American
institution having written and gotten published an explanatory descriptive article. The
differences were statistically significant, % 2 (3, N= 143) = 20.13,/? < .000; resampled
/?<.000.
Attitudes-only articles. Table 51 shows that the odds of a first author affiliated
with an institution in the Asian Pacific or Eurasia having written and published an article
in which attitudes were the sole dependent measure were more than 12 times greater than
95
Table 50
Explanatory Descriptive Papers by Region of First Author's Affiliation
Explanatory descriptive
Percentage
Adjusted
Region
Yes No
Total
yes
residual
Eurasia
5 25
30
16.7
-1.4
Europe
9 21
30
30.0
0.5
Middle East
10 3
13
76.9
4.3
North America
14 56
70
20.0
-1.7
Total
38 105
143
Table 51
Attitudes-only Papers by Region of First Author's Affiliation
Attitudes'
-only
Percentage
Adjusted
Region
Yes
No
Total
yes
residual
Eurasia
16
10
26
61.5
3.9
Europe
3
24
27
11.1
-2.5
Middle East
1
4
5
20.0
-0.5
North America
17
47
64
26.9
-1.0
Total
37
85
122
a first author affiliated with an institution in Europe. The differences were statistically
significant, x 2 (3, N = 122) = 17.39,/? = .00; resampled/? < .000.
Other types of articles. Crosstabulations for the types of articles in which there
were no statistically significant regional differences (i.e., anecdotal-only papers and one-
group posttest-only papers) are presented in Tables 52 and 53 below. (Note that the
logistic regression analysis, however, showed that region is a statistically significant
predictor of an article being an anecdotal-only article when the other factors are
controlled for.)
96
Table 52
Anecdotal-Only Articles by Region of First Author 's Affiliation
Anecdotal-only
Percentage
Adjusted
Region
Yes No
Total
yes
residual
Eurasia
10 30
40
25.0
-1.9
Europe
14 30
44
31.8
-1.0
Middle East
5 13
18
27.8
-.9
North America
59 70
129
45.7
2.7
Total
88 143
231
Note. x 2 (3,7V = 231) = 7.65, p = .054; resampledp = .059.
Table 53
One-Group Posttest-Only Papers by Region of First Author's Affiliation
One-group
posttest-only
Percentage
Adjusted
Region
Yes
No
Total
yes
residual
Eurasia
13
7
20
65.0
1.6
Europe
8
6
14
57.1
0.7
Middle East
3
1
4
75.0
1.1
North America
21
33
54
38.9
-2.3
Total
45
47
92
Note. %\3,N =
92) =
= 5.71, p = .
,127:
, resampled p =
.128.
Islands of Practice: Logistic Regression Analysis
For each of the five outcome variables (i.e., anecdotal-only papers, experimental/
quasi-experimental papers, explanatory descriptive papers, attitudes-only papers, and one-
group posttest-only papers), I present the history of model fitting, information about the
overall fit of the regression equation, and the regression equation(s) themselves. I also
97
present graphs that visually portray the best fitting model. Note that the regression
equations refer to probability of a yes (successful) outcome (i.e.,;?, not q).
On all of the outcomes besides explanatory descriptive, the African, Asia-
Pacific/Eurasian, and Middle Eastern categories were combined into a combined region
category called Asian-Pacific/Eurasian et al. I called it Asian-Pacific et al. because most
of the observations came from the Asian-Pacific/Eurasian regions. The breakdown of
articles into each region is given for each analysis below. Note that only articles that
dealt with human participants are included in these regression analyses. A South
American category was not included because there were no South American articles that
dealt with human participants in the sample.
Anecdotal-only Articles
Table 54 shows comparisons of the fit of several logistic regression models using
anecdotal-only papers, a binary variable, as the outcome. In this case the best fitting
model was Model 9: intercept + region + year + region * year.
For the anecdotal-only papers variable, the Omnibus Test of Model Coefficients
was statistically significant, % 2 (7, N= 233) = 20.74, p = .001, and the Hosmer and
Lemeshow test was not statistically significant, % 2 (7, N = 233) =2.97, p = .888, which
indicate that the overall fit of the model was appropriate. Figure 5 shows the scatterplot of
expected and observed probabilities. It has one outlier at coordinate (0.5, 0.2), which
corresponds with the three 2001 Asian-Pacific/Eurasian et al. anecdotal-only articles that
dealt with human participants. A regression analysis was conducted with those three
Table 54
98
The Fit of Several Logistic Regression Models for Anecdotal-Only Papers
Deviance
Models
Difference
Model
Predictors
(df)
compared
(df)
P
1
I+R+Y+F+R*Y+R*F+Y*F+R*Y*F
286.84(11)
_
„
_
2
I+R+Y+F+R*Y+R*F+Y*F
287.93(9)
1 & 2
1.09(2)
.58
3
I+R+Y+F+R*Y+R*F
288.32(8)
2 & 3
0.39(1)
.53
4
I+R+Y+F+R*Y+Y*F
288.01(7)
2 & 4
0.31(2)
.86
5
I+R+Y+F+R*F+Y*F
293.50(8)
2 & 5
5.57(1)
.02
6
I+R+Y+F+R*Y
288.45(6)
4& 6
0.44(1)
.51
7
I+R+Y+F+F*Y
293.50(5)
4& 7
0.00(2)
.99
8
I+R+Y+F
294.27(4)
6& 8
5.79(2)
.06
9
I+R+Y+R*Y
289.17(5)
6& 9
0.72(1)
.40
Note. I = intercept, R = region, Y = year, F = forum type.
0.8
0.7
>.
0.6
+J
!5
0.5
Si
o
a.
0.4
■o
0.3
o
u
a.
X
0.2
HI
0.1
♦
♦
♦
<♦♦
r^
0.2 0.4 0.6 0.8 1
Observed probability
« Anecdotal-only (yes)
Figure 5. Expected and observed probability for anecdotal-only papers.
99
articles removed; I do not present those results of that analysis here because they were
negligibly different from the results when the outlying data point was included.
Table 55 shows the results of regression analysis for the anecdotal-only papers.
The breakdown of the n-size of the region categories was 129, 60, and 44 for North
American, Asian-Pacific/Eurasian et al., and European articles, respectively. For the
Asian-Pacific/Eurasian et al. category, the n-sizes for each region were 40, 18, and 2 for
Asian-Pacific/Eurasian, Middle Eastern, and African articles, respectively.
The interpretation of logistic regression equations is as not as straightforward as it
is for regression with a continuous outcome variable. Therefore, I will explain the
interpretation of the items in the regression tables that are presented in this section.
The first column shows the elements that were included in the regression
equation; in the case of anecdotal-only papers those elements were a constant, year,
region of first author's affiliation, and a region by year interaction. Because region was a
categorical variable, the categories that it was comprised of — North America, Asia-
Pacific/Eurasia et al., and Europe — are displayed. They are indented under the region
label. In these regression analyses, North America was the reference group, so the
comparisons were always be between North America and one of the other regions.
The second column, labeled B, shows the log coefficient. For a continuous
variable, if the coefficient is positive, then that indicates that the odds of success (i.e., a
yes) increase as the coefficient increases, and vice versa. For example, if the coefficient
were positive for year, then that would indicate that the odds of a success would have
increased every year. For categorical variables (like regions), the comparison category has
100
Table 55
Summary of Regression Analysis for Predictors of Anecdotal-Only Articles, (N=233)
Variable
B
S.E.
Wald
df
P
Exp(B)
Year
-0.37
0.11
11.65
1
.00
.69
Region
9.65
2
.01
North America (reference group)
Asia-Pacific/Eurasia et al.
-2.24
0.79
7.95
1
.01
.11
Europe
-1.31
0.71
3.40
1
.07
.27
Region by year
5.33
2
.07
North American (reference group)
Asia-Pacific/Eurasia et al.
0.49
0.22
4.82
1
.03
1.63
Europe
0.27
0.22
1.52
1
.22
1.30
Intercept
0.85
0.35
5.88
1
.02
2.33
a greater odds of success than the reference category if the log coefficient is positive, and
vice versa. For example, if the coefficient for the Europe category were positive, that
means that the likelihood of a European article's being an anecdotal-only article would
have been greater than the likelihood of a North American article being an anecdotal-only
article. If the coefficient were negative, the opposite would be true: The likelihood of a
European article's being an anecdotal-only article would be less than the likelihood of a
North American article's being an anecdotal-only article.
The column labeled S.E. displays the standard error of the log coefficient. The
category labeled Wald shows the value of the Wald statistic, which, along with the
degrees of freedom (df) in the next column, is used to determine the statistical
significance of the coefficient.
101
Finally, since log coefficients alone cannot be easily interpreted, I have included
the exponentiated B coefficient in the last column, labeled exp(B). The value of 8 can be
interpreted as an odds ratio — for categorical variables, the ratio of the odds in the
reference category to the odds in the comparison category; for continuous variables, the
ratio of odds between subsequent quantitative units. An odds ratio of one indicates that
the odds of success are the same in both categories, an odds ratio less than one indicates
that the odds are greater in the reference category, and an odds ratio greater than one
indicates that the odds are greater in the comparison category. For example, an odds ratio
of .27; where North America is the reference category, where Europe is the comparison
category, and a success means that an article is anecdotal; would mean that the odds of a
North American article's being anecdotal would be greater than for a European
article — about 3.7 times greater because 1/.27 = 3.7. If the odds ratios in the same case
were 3.7 instead of .27, then that would mean that the odds in Europe papers were 3.7
times greater than the odds in North America papers.
So, based on the information given above, the following interpretations can be
made from Table 55.
1 . The predicted odds of an article's not being anecdotal had gotten 1 .45 (1/.69 =
1.45) times greater per year between 2000 and 2005 (i.e., there was a decrease in
anecdotal articles over time). The decrease was statistically significant.
2. The predicted odds of an article's being anecdotal were 9.1 (l/.ll =9.1) times
greater for North American articles than for Asian-Pacific/Eurasian et al. articles and 3.7
(1/.27 = 3.7) times greater for European articles. The difference between North America
102
and Asian-Pacific/Eurasian et al. categories was statistically significant, and the
difference between North American and European categories was nearly statistically
significant (p = .07).
3. There was a statistically significant interaction in the difference between the
decline in trend in anecdotal articles between North American articles and Asian-
Pacific/Eurasian et al. articles.
Figure 6 shows the percentage of anecdotal-only articles to anecdotal-only plus
nonanecdotal-only articles by region and year. The values next to each marker in a series
show the number of anecdotal articles in that region each year. In Figure 6 it is clear that
the percentage of North American anecdotal-only articles had decreased linearly between
2000 and 2005. Figure 6 also shows that the percentage of European anecdotal-only
articles had dropped 30% between 2000 and 2001 and then leveled off. It also shows that
there was considerable variability in the percentage of Asia-Pacific/Eurasian et al. articles
across years.
Figure 7 shows the proportions of anecdotal-only articles by region. As shown in
Table 55, there was a higher percentage of North American anecdotal-only articles than
the percentage of European anecdotal-only articles, which was, in turn, higher than the
percentage of Asian-Pacific/Eurasian et al. anecdotal-only articles.
Experimental/Quasi-Experimental Articles
Table 56 shows a history of model selection for the experimental/quasi-
experimental variable. The best fitting model in this case, Model 9, was: intercept +
103
(A
ai
>>
>.
c
■
re
+j
o
■a
u
<D
c
<
100
90
80
70
60
50
40
30
20
10
15
— ♦—
North America
- Europe
-Asian-
Pacific/Eurasia et
al.
------
\ 11
3
. 7 10
A
9
10
/ \
4
-.. ^
+
/
; , 1
r' 2 y
6
\3
/
\ / 2
A
i
\ U '
£.
2000
2001
2002 2003
Year
2004
2005
Figure 6. Anecdotal-only papers by combined region and year. The
value nearest to a data point shows the n-size for that data point.
^
if
.*
*
&
Region
50
n=59)
45
40
35
30
25
20
15
10
5
V)
(n-14)
in-ioj
c
o
a
LL
□ Anecdotal-only (% yes)
Figure 7. Anecdotal-only papers by combined regions.
104
Table 56
The Fit of Several Regression Models for Experimental/Quasi-Experimental Papers
Deviance
Models
Difference
Model
Predictors
(df)
compared
(df)
P
1
I+R+Y+F+R*Y+R*F+Y*
: F+R*Y*F
165.53(11)
2
I+R+Y+F+R*Y+R*F+Y*
F
167.10(9)
1 & 2
1.57(2)
.46
3
I+R+Y+F+R*Y+R*F
167.49(8)
2 & 3
0.39)1)
.53
4
I+R+Y+F+R*Y+Y*F
175.54(7)
2 & 4
8.44(2)
.01
5
I+R+Y+F+R*F+Y*F
168.93(7)
2 & 5
1.83(2)
.40
6
I+R+Y+F+R*Y
175.64(6)
3 & 6
8.15(2)
.02
7
I+R+Y+F+R*F
169.22(6)
3 & 7
1.73(2)
.42
8
I+R+Y+F
176.75(4)
7 & 8
7.53(2)
.02
9
I+R+F+R*F
169.31(5)
7 & 9
0.09(1)
.76
Note. I = intercept, R = region, Y = year, F = forum type.
region + forum type. However, I chose Model 7 over Model 9 in this case because after
running the regression equation for Model 9, it turned out that Model 9 was exactly
specified (i.e., there was perfect prediction if the continuous variable — year — was not
included). Although Model 7 was a slightly more complicated model than Model 9, it had
approximately the same deviance as Model 9. The differences between the values of the
region, journal, and journal by region coefficients were negligible between models 7 and
9, so I only present the results of Model 9 here. Figure 8 shows a scatter plot of the
expected and observed probabilities for experimental/quasi-experimental articles.
The Omnibus Test of Model Coefficients was statistically significant, % 2 (6, N =
144) 17.89,/? = .006, and the Hosmer and Lemeshow test was not statistically significant,
X 2 (8, N = 144) 1.94, p = .983, which indicate that the overall fit of the model was good.
There are three data points that I considered through visual analysis to be outliers, which
105
.Q
<o
.Q
o
■a
o
+•*
o
a>
Q.
X
LU
1
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
►* ♦ ♦♦
*T
* Experimental/quasi
experimental (yes)
0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
Observed probability
Figure 8. Expected and observed probabilities for experimental/
quasi-experimental papers.
are located approximately at coordinate (1.0, 0.6). Those data points represent the one
nonanecdotal-only journal article from Europe in 2004, the three nonanecdotal-only
journal articles from North America in 2004, and the one nonanecdotal-only journal
article from North America in 2005. 1 ran regression equations with and without those
outliers removed. The differences were minimal between the two equations so I only
include the one with outliers here. The only notable difference however was that thep-
value associated with forum type was .05 without outliers, and .09 with outliers (as
shown in Table 57).
Table 57 shows a summary of the regression analyses when run with outliers.
With outliers included, the breakdown of the n-size of the region categories was 70, 44,
106
Table 57
Summary of Regression Analysis for Predictors Experimental/Quasi-Experimental
Articles (N = 144), With Outliers
Variable
B
S.E.
Wald
df
P
Exp(B)
Year
0.04
0.12
0.09
1
.11
1.04
Region
13.66
2
.00
North America (reference group)
Asia-Pacific/Eurasia et al.
-1.50
0.48
9.66
1
.00
0.22
Europe
-1.73
0.54
10.46
1
.00
0.18
Forum type
Conference (reference group)
Journal
-1.08
0.64
2.85
1
.09
0.34
Region by forum
6.38
2
.04
Journal by North American (reference group)
Journal by Asia-Pacific/Eurasia et al.
3.10
1.29
5.64
1
.02
21.21
Journal by Europe
1.72
1.19
2.09
1
.15
5.56
Contrast
1.39
0.53
6.88
1
.01
4.00
and 40 for North American, Asian-Pacific/Eurasian et al., and European articles,
respectively. For the Asian-Pacific/Eurasian et al. category the breakdown of the n-sizes
into regions was 30, 13, and 1 for Asian-Pacific/Eurasian, Middle Eastern, and African
articles, respectively.
To illustrate the effect of the region by forum interaction, I also include the results
of the regression equation without the region by forum interaction (with the outliers
included) in 57. By comparing Tables 57 and 58 one can see that it is including the region
by forum type interaction that causes the direction to switch on the forum type variable.
Note that the model fit was statistically significantly better for the regression equation
107
Table 58
Summary of Regression Analysis for Predictors of Experimental/Quasi-Experimental
Articles (N = 144), With Outliers and Without Interaction Term
Variable
B
S.E.
Wald
df
P
Exp(B)
Year
0.30
0.11
0.08
1
.79
1.03
Region
9.56
2
.01
North America (reference group)
Asia-Pacific/Eurasia et al.
-0.94
0.42
5.02
1
.03
.39
Europe
-1.34
0.47
8.27
1
.00
.26
Forum type
Conference (reference group)
Journal
.14
0.47
0.08
1
.77
1.15
Constant
1.09
0.48
5.13
1
.02
2.97
with the interaction term than without it (see Table 56). Yet, the regression equation
without the interaction term had an overall good fit; the Omnibus Test of Model
Coefficients was significant, %\4, N= 144) = 10.49,/? = .03, and the Hosmer Lemeshow
test was not significant, x 2 (8, N = 144) = 8.45,/? = .390.
The findings from these regression analyses, which are based on the regression
equation with the outliers and interaction term left in, are listed below:
1. Region was a significant predictor of an article's being experimental/quasi-
experimental or not. Specifically, the predicted odds of a North American article's being
an experimental/quasi-experimental article were 4.6 (1/.22) times greater than an Asian-
Pacific/Eurasian et al. article's odds and 5.6 (1/.18) times greater than the odds of
European article's odds.
108
2. When controlling for the journal by region interaction, the odds of a conference
article's being an experimental/quasi-experimental article were about 2.9 times (1/.34)
greater than a journal article's odds.
3. There was a statistically significant interaction between type of forum and
region.
Figure 9 shows the percent (yes) and number of experimental/quasi-experimental
articles by forum type and region. It shows that there was a higher proportion of
experimental/quasi-experimental articles in conferences than in journals in North
American papers, but the opposite holds true for European and Asia-Pacific/Eurasia et al.
papers. An explanation for this interaction and for the fact that forum type is significant
here, but not in the crosstabulation of Table 40, is given in the discussion section. Figure
10 shows the percentage of experimental/quasi-experimental articles by combined region
and year. In Figure 10 it appears that the proportion of experimental/quasi-experimental
papers did not change significantly across years.
Explanatory Descriptive Papers
For explanatory descriptive papers, I did not combine regional categories because
the n-sizes of each category were large enough to get a sensible regression each equation.
(I did not have to group Asian-Pacific/Eurasian, Middle Eastern, and African papers
together.) I did however exclude the one African paper that was not ancecdotal-only from
this analysis. Table 59 shows the history of model fitting for explanatory descriptive
papers. Model 8 (intercept + region + year) turned out to be the best fitting model.
109
c
o
£
o
a.
x ~~
9 w
,i u
to >.
5 ss
■2" """
TO
c
£
o
a
x
LU
100
90
80
70
60
50
40
30
20
10
7
45
A.
D N "-"'"^
^^"^
3 X A
'" *
11
. North America
Europe
±.-Asian-
Pacific/Eurasia et
al.
Journal Conference
Forum type
Figure 9. Experimental/quasi-experimental papers by combined region and
forum type. The value nearest the data point shows the n-size for
that data point.
100
90
re
c
80
F
L.
70
0)
Q.
X -^
60
— <u
50
re ^p
3 °^
O -
40
re
C
30
0)
E
20
L-
11)
Q.
10
X
LU
7
10
3 y\
10 r>
/
S 2
\ 6
♦ ,o /
fi
iu~l.^
.A
\
/
/
3 2
2*
3
N ■/
■
2
iii
North America
Europe
±.-Asia-
Pacific/Eurasia et
al
2000 2001 2002 2003 2004 2005
Year
Figure 10. Experimental/quasi-experimental papers by combined region and year.
The value nearest to a data point shows the n-size for that data point.
110
Table 59
The Fit of Several Logistic Regression Models for Explanatory Descriptive Papers
Deviance
Models
Difference
Model
Predictors
(df)
compared
(df)
P
1
I+R+Y+F+R*Y+R*F+Y*F+R*Y*F
127.20(15)
2
I+R+Y+F+R*Y+R*F+Y*F
130.79(12)
1 & 2
3.59(3)
.31
3
I+R+Y+F+R*Y+R*F
131.62(11)
2 & 3
0.83(1)
.36
4
I+R+Y+F+R*Y+Y*F
135.13(9)
2 & 4
4.34(3)
.23
5
I+R+Y+F+R*F+Y*F
132.49(9)
2 & 5
1.70(3)
.64
6
I+R+Y+F
138.30(5)
3 & 6
6.68(6)
.54
7
I+R+F
147.78(4)
6& 7
9.48(1)
.00
8
I+R+Y
138.37(4)
6& 8
0.07(1)
.79
9
I+Y+F
153.89(2)
6& 9
15.59(2)
.00
10
I+R
147.78(3)
8 & 10
9.41(1)
.00
11
I+Y
153.96(1)
8 & 11
15.59(3)
.00
Note. I = intercept, R = region, Y = year, F = forum type.
Figure 1 1 shows the expected and observed probabilities for explanatory
descriptive papers. The Omnibus Test of Model Coefficients was statistically significant,
X 2 (4, N = 143) = 27.22, p = .000, and the Hosmer and Lemeshow test was not statistically
significant, x 2 (8, N = 143) = 4.99, p = .768, which indicate that the overall fit of the
model was appropriate. Through visual inspection, I did not consider any of the data
points to be outliers.
Table 60 shows the regression equation for explanatory descriptive papers. The
breakdown of the n-sizes of the region categories here was 70, 30, 30, and 13 for North
American, Asian-Pacific/Eurasian, European, and Middle Eastern articles, respectively.
The one African nonanecdotal article was not included in this analysis. For the Asian-
Ill
.Q
re
O
■a
v
+j
o
V
a.
x
HI
1
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
*
♦
♦
♦
♦ ♦
° ♦ *~
:: ^
^ — ♦ «
,^ — : —
^ Explanatory descriptve
(yes)
0.2 0.4 0.6 0.8
Observed probability
Figure 11. Expected and observed probabilities for explanatory
descriptive papers.
Table 60
Summary of Regression Analysis for Predictors of Explanatory Descriptive Articles,
(N=143)
Variable
B
S.E.
Wald
df
P
Exp(B)
Year
-0.39
0.13
8.91
13
11
11
.00
0.68
Region
13.00
.01
North America (reference group)
Asia-Pacific/Eurasia et al.
-0.17
0.59
0.08
.77
0.84
Europe
0.47
0.52
0.82
.36
1.60
Middle East
2.59
0.76
11.75
.00
13.31
Constant
-0.22
0.47
0.23
.63
0.80
112
Pacific/Eurasian et al. category the n-sizes were 20, 4, and 1 for Asian-Pacific/Eurasian,
Middle Eastern, and African articles, respectively.
The findings that relate to Table 60 are listed below:
1. Year was a significant predictor of explanatory descriptive papers. The odds of
a paper's not being an explanatory descriptive paper was 1.47 (1/.68) times greater each
year from 2000 to 20005.
2. Region was a significant predictor of a paper's being an explanatory
descriptive paper. The odds of a Middle Eastern paper's being explanatory descriptive
was over 13 times greater than the odds in a North American paper — a statistically
significant difference in this case.
Figure 12 shows the percentage and number of explanatory descriptive papers by
region. In Figure 12 there is considerable variability and low n-sizes. However, it appears
that there had been a steady decrease in the number of North American explanatory
descriptive papers from 2000 to 2005, although there was not a statistically significant
interaction between year and region. Figure 13 shows the percentage and number of
explanatory descriptive paper by region and year. The Middle Eastern category had the
greatest proportion of explanatory descriptive papers.
Attitudes -Only Papers
Table 61 shows the history of model-fitting for attitudes-only papers. The best
fitting model was actually Model 10 (intercept + region); however, I choose to keep year
in the model because Model 10 was exactly specified. That is, I decided to use Model 8
113
10
100
90
80
r 70
■s 60
o
in
u
■a
o
«
c
_2
Q.
X
HI
50
40
30
20
10
2
ji
3
▲
: V.
^
2 *.
: '2 v
■i <,
S
•i > 3
n
2"-
%v - /' ■» ■ /k
\ 3
x -#• 1
/ s * ■
•ff 0/ '*"
" 2 \\.-
u
•? 7 o
-i — * — i — ♦ — i — i-
—i — *^— i — -* —
« Asian-
Pacific/Eurasia
— ■ — Europe
. A . . Middle East
North America
2000 2001 2002 2003 2004 2005
Year
Figure 12. Explanatory descriptive papers by year and region. The value
nearest to a data point shows the n-size for that data point.
V
C
V
O
i-
0)
a.
90
80
70
60
50
40
30
20
10
( n= 14)
<F
in=9f
(n=5)
(n=10)
J-
#
J
r
#
&
Region
□ Explanatory descriptive
(% yes)
Figure 13. Explanatory descriptive papers by region.
114
Table 61
The Fit of Several Logistic Regression Models for Attitudes-Only Papers
Deviance
Models
Difference
Model
Predictors
(df)
compared
(df)
P
1
I+R+Y+F+R*Y+R*F+Y*F+R*Y*F
128.30(11)
2
I+R+Y+F+R*Y+R*F+Y*F
129.33(9)
1 & 2
1.03(2)
.60
3
I+R+Y+F+R*Y+R*F
130.07(8)
2 & 3
0.74(1)
.39
4
I+R+Y+F+R*Y+Y*F
133.11(7)
2 & 4
3.78(2)
.15
5
I+R+Y+F+R*F+Y*F
132.93(7)
2 & 5
3.60(2)
.17
6
I+R+Y+F
136.05(4)
3 & 6
5.98(4)
.20
7
I+R+F
136.08(3)
6& 7
0.03(1)
.86
8
I+R+Y
136.69(3)
6& 8
0.61(1)
.44
9
I+F+Y
151.62(2)
6& 9
15.57(2)
.00
10
I+R
136.79(2)
7 & 10
0.71(1)
.40
11
I+F
151.78(1)
7 & 11
15.70(2)
.00
12
I+Y
151.89(1)
8 & 12
15.20(2)
.00
Note. I = intercept, R = region, Y = year, F = forum type.
(intercept + region + year) rather than Model 10. 1 ran logistic regressions for both Model
10 and for Model 8 and found that the differences between them were negligible.
Figure 14 shows the expected and observed probabilities for attitudes-only papers.
The Omnibus Test of Model Coefficients was statistically significant, % 2 (3, N= 123)
15.40,/? = .002, and the Hosmer and Lemeshow test was not statistically significant, % 2 (8,
N = 123) = 7.93, p = .440, which indicates that the overall fit of the model was good.
Through visual inspection, I considered the data points at coordinates (0.7, 0.1)
and (1.0, 0.55) to be outliers. The data point at coordinate (0.7,0.1) consisted of four
articles from 2003 from the Asian-Pacific/Eurasian et al. category and the data point at
coordinate (1.0, 055) consisted of three European articles from 2001. 1 ran regression
115
.o
A
O
■c
0)
o
<D
Q.
X
HI
1
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
♦ ♦♦♦ ♦
<> *^
0.2 0.4 0.6 0.8
Observed probability
, Attitudes-only (yes)
Figure 14. Expected and observed probabilities for attitudes-only papers.
analyses with and without the outliers and, because there was an interesting difference in
the resulting regression equations, I present regression results for both.
Table 62 shows a summary of the regression analysis with outliers included and
Table 63 shows a summary of the regression analysis with the outliers excluded. With
outliers included, the breakdown of hte n-sizes of the combined region category was 64,
32, 27 for North American, Asian-Pracific/Eurasian et al., and European articles,
respectively. For the Asian-Pacific/Eurasian et al. category, the n-sizes were 26, 5, and 1
for Asian-Pacific/Eurasian, Middle Eastern, and African articles, respectively.
It was found that Region was a statistically significant predictor of an article's
being an attitudes-only paper. The predicted odds of an Asian-Pacific/Eurasian article's
being an attitudes-only article was 3.56 times higher than the predicted odds of a North
116
Table 62
Summary of Regression Analysis for Predictors of Attitudes-Only Articles (N = 123),
With Outliers
Variable B S.E. Wald df p Exp(B)
Year .04 -0.13 0.10 1 0.75 1.04
Region 13.40 2 .00
North America (reference group)
Asia-Pacific/Eurasia et al.
1.27
0.46
7.77
1
.01
3.56
Europe
-1.06
0.68
2.44
1
.12
0.35
Constant
-1.16
0.54
4.71
1
.03
0.31
Table 63
Summary of Regression Analysis for Predictors of Attitudes-Only Articles (N = 99), With
Outliers Removed
Variable
B
S.E.
Wald
df
P
Exp(B)
Year
0.13
0.14
0.79
14.09
1
2
.37
.00
1.14
Region
North America (reference group)
Asia-Pacific/Eurasia et al.
1.28
0.46
7.81
1
.01
3.59
Europe
-2.13
1.06
4.04
1
.04
0.12
Constant
-1.45
0.57
6.40
1
.01
0.23
American article's being an attitudes-only article. Also, the predicted odds of a European
article's not being an attitudes-only articles was 2.9 (1/.35) times greater than predicted
odds of a North American article's being an attitudes-only article.
117
Also, in the regression analysis with outliers excluded, the comparisons between
the odds of both North American and Asian-Pacific/Eurasian et al. papers and between
North American and European papers were statistically significant. In the regression
analysis with the outliers included, the comparison of the odds between North American
and Asian-Pacific/Eurasian et al. papers was statistically significant and the comparison
between North American and European articles was nearly statistically significant
ip = .n.)
Figure 15 shows the percentage of attitudes-only articles by year and combined
region and Figure 16 shows the percentage of attitudes-only articles only by combined
region. Those figures help illustrate the findings listed above: Namely, Asian-
Pacific/Eurasian et al. articles had the higher proportion of attitudes-only articles.
One-Group Posttest-Only Articles
Table 64 shows the history of model-fitting for the one-group posttest-only
articles. Based on Table 64, Model 9 (intercept + region + year + region by year) was the
best model.
Figure 17 shows a plot of expected and observed probabilities (using Model 9) for
one-group posttest-only articles. For Model 9, The Omnibus Test of Model Coefficients
was statistically significant, % 2 (5, N= 93) = 14.53,/? = .013, and the Hosmer and
Lemeshow test was not statistically significant, % 2 (8, N = 93) =12.15,/? = .15, which
indicate that the overall fit of the model was good.
118
to
3
>.
^S
_>.
C
o
1
(A
9
■a
3
4-1
S3
<
100
90
80
70
60
50
40
30
20
10
A
/
\
— ♦—
North America
- Europe
_ Asia-
Pacific/Eurasia et
al
2
/
------
Ik.
/
6
• '£■ N
1/ x o
,^
/
1/
1 \
jL ■
'
/5 ^
\ ®
•
/2
■ \ / 1
'. 2\ /
;' / \ 1/
/
*•'
2000
2001
2002 2003 2004
Year
2005
Figure 15. Attitudes-only papers by year and combined regions. The value
nearest to a data point shows the n-size for that data point.
60
(n=18)
50
1 40
■£ 30
o
o 20
0)
o- 10
o
(n=17)
□ Attitudes-only (% yes)
(n=3)
1
<*>
4t *'
Region
Figure 16. Attitudes-only papers by combined regions.
119
Table 64
The Fit ofSeveal Logistic Regression Models for One-Group Posttest-Only Papers
Deviance
Models
Difference
Model
Predictors
(df)
compared
(df)
P
1
I+R+Y+F+R*Y+R*F+Y*F+R*Y*F
110.95(11)
2
I+R+Y+F+R*Y+R*F+Y*F
113.00(9)
1 &2
2.05(2)
.36
3
I+R+Y+F+R*Y+R*F
113.12(8)
2&3
0.12(1)
.73
4
I+R+Y+F+R*Y+Y*F
114.24(8)
2&4
1.24(1)
.27
5
I+R+Y+F+R*F+Y*F
120.48(7)
2&5
7.48(1)
.00
6
I+R+Y+F+R*Y
114.25(6)
3&6
1.13(1)
.29
7
I+R+Y+F+R*F
120.63(6)
3&7
7.51(1)
.00
8
I+R+Y+F
121.36(4)
6&8
7.11(2)
.03
9
I+R+Y+R*Y
114.39(5)
6&9
0.14(1)
.71
10
I+R+Y
121.79(3)
9 & 10
7.40(2)
.03
Note. I = intercept, R = region, Y = year, F = forum type.
1
0.9
0.8
s 0.7
«s 0.6
n
|0-5
"g 0.4
+■1
o
g. 0.3
X
uj 0.2
0.1
(
<
*
♦
♦ ♦
i
►
♦ ♦ <
►
+ One-group posttest-onlj[
(yes)
♦ *
♦
4
♦
) 0.2 0.4 0.6 0.8
Observed probability
Figure 1 7. Expected and observed probabilities for one-group posttest-only
articles, with interaction term.
120
I considered three data points to be outliers. They were approximately at
coordinates (1.0, 0.65), (1.0, 5.5), (0.8, 3.5), and (0.55, .25); which correspond with the
two experimental Asian-Pacific/Eurasian et al. articles in 2003, with the three
experimental North American articles in 2001, with the nine experimental North
American articles in 2003, and with the three experimental European articles in 2005. 1
ran regression analyses with and without outliers and found no meaningful differences
whether outliers were included or not; therefore, I only present results here with the
outliers included. Table 65 shows a summary of the regression analysis for Model 9. The
breakdown of the n-size of the combined region category was 54, 25, 14 for North
American, Asian-Pacific/Eurasian et al., and European articles, respectively. For the
Asian-Pacific/Eurasian et al. category the n-sizes were 20, 4, and 1 for Asian-
Pacific/Eurasian, Middle Eastern, and African articles, respectively.
Table 65 shows that none of the predictor variables were significant predictors of
one-group posttest-only papers. However, the interaction of year and region was
statistically significant; specifically, there was an interaction between North American
papers by year and Asian-Pacific/Eurasian papers by year. This interaction becomes clear
from a visual examination of Figure 18, which is a graph of the percentages of one-group
posttest-only papers by region and year.
In Figure 18, it shows that, more or less, there was a decline in the number of
papers in Europe and North America. It also shows that, except for 2004, the pattern of
decline of one-group posttest-only papers in Europe was similar to the pattern of decline
in North America and that the North American series was usually slightly lower than in
121
Table 65
Summary of Regression Analysis for Predictors of One-Group Posttest-Only Articles
for Model With Interaction Term (N= 93)
Variable
B
S.E.
Wald
df
P
Exp(B)
Year
-0.21
0.18
1.44
12
11
21
.23
0.81
Region
2.99
.50
North America
11
(reference group; n = 54)
Asia-Pacific/Eurasia et al. (n = 25)
-0.76
1.12
0.47
.50
0.47
Europe (n = 14)
2.23
1.66
1.97
.16
10.22
Region by year
North American (reference group)
Asia-Pacific/Eurasia et al.
Europe
Constant
6.38
0.24 0.63
0.14
.04
0.62
0.32
3.80
.05
1.86
0.55
0.47
1.38
.24
0.58
.71
1.27
120
V
c
o
I
to
0)
S
to
o
a
a.
3
o
O)
I
0)
£
O
3
2
2
00
2
/\ 7 5
80
N . 7 .A
/
60
\ 1
' 1/
1 / \
40
^ X
'. /
- - -M
- -■ / \
' 7 1 \ 1
K
3
/ '■ \«
20
1
2 \
\/ 5
I
I I
-^-^ , ,
2000 2001 2002 2003 2004 2005
Year
North America
-■- - - Europe
. A - - Asia-
Pacific/Eurasia et
al.
Figure 18. One-group posttest-only articles by combined region. The value
nearest to a data point shows the n-size for that data point.
122
Europe. Also, Figure 18 shows that in the Asian Pacific et al. region there was an
increase, except for 2004, in one-group posttest-only papers between 2000 and 2005.
Although, Figure 18 indicates there was a difference between regions, the low n-sizes
(only 5 out of 15 data points had n-sizes above 5) could have masked the difference in
terms of finding statistical significance. Indeed, when collapsing across years, there was a
statistically significant difference between regions, as Table 66 shows.
Table 66, in which I show the results of Model 10 — the regression equation
without the interaction (i.e., intercept + region + year), shows that there was a statistically
significant difference in the proportion of one-group posttest-only articles between North
America and Asian-Pacific Eurasian et al. articles, but not between North American and
European articles. This difference is also visualized in Figure 19, where the percentages
of one-group posttest-only articles by region only are displayed. It is important to note,
however, that Model 10 is not as good a fitting model as Model 9 (with the interaction) as
Table 64. shows. Also, the Omnibus Test of Model Coefficients for Model 10,
X 2 (3, N=93) = 7.13,;? = .068, and the Hosmer and Lemeshow test, % 2 (7, N=93)= 16.91,
p= .018, show that Model 9 is a poor model for predicting one-group posttest-only
articles. Therefore, the results of Model 9 should be regarded with caution.
Comparisons Between Fields
Up to this point I have presented results within the field of computer science
education. In this section I present results concerning the proportions of empirical (i.e.,
123
Table 66
Summary of Regression Analysis for Predictors of One-Group Posttest-Only Articles
for Model Without Interaction Term (N = 93)
Variable
B
S.E.
Wald
df
P
Exp(B)
Year
-0.12
0.13
0.84
12
11
.36
.89
Region
5.85
1
.05
North America
(reference group; n = 54)
Asia-Pacific/Eurasia et al. (n = 25)
1.21
0.51
5.55
.02
3.36
Europe (n= 14)
0.68
0.61
1.23
.27
1.98
Constant
-0.05
0.52
0.10
.92
0.95
80
70
s 60
£ 50
*T 40
• 20
10
^
P*
.*
,f S
■A
J
r
r
Region
(n=21)
(n=8)
I
(n-17)
I
I
□ One-group posttest only
(% yes)
Figure 19. One-group posttest-only articles by combined region.
not anecdotal) articles dealing with human participants and proportions of quantitative,
124
qualitative, and mixed methods research between fields. Note that the proportions for the
field of education proper come from Gorard and Taylor (2004) and the proportions for the
field of educational technology come from the review of methodological reviews of
educational technology, which was presented earlier in this dissertation.
Proportions of Empirical Articles Dealing
with Human Participants
Table 67 shows that the proportions of empirical articles dealing with human
participants decreased monotonically from education proper to educational technology
and from educational technology to computer science education. Assuming that those
fields are ordinal in terms of the degree to which they have an engineering tradition
(where computer science education has the largest degree of the engineering tradition and
education proper has the least), indicated by the number of articles that do not deal with
human participants, the results of the M 1 test, indeed, showed that there was a statistically
significant linear (monotonic) relationship, M 2 (l, N = 1,351) = 52.32, p < .000. The
adjusted residuals, which ranged from 6.2 for education proper and -5.3 for computer
science education, showed that the linear relationship was pronounced.
Proportions of Types of Research
Traditions Between Fields
Table 68 shows that there was a statistically significant difference, % 2 (2, N = 638)
= 20.84,/? < .000, between the proportions of quantitative, qualitative, and mixed
methods articles in computer science education and educational technology forums. The
125
Table 67
Comparison of the Proportion of Empirical, Human Participants Articles in Computer
Science Education and Education Proper
Empirical
research with
human
participants
Percentage
Adjusted
Field
Yes
No
Total
yes
residual
Ed. Proper
79
15
94
84.0
6.2
Ed. tech.
494
411
905
54.6
1.6
CSE
144
208
352
40.9
-5.3
Total
717
634
1,351
Note. Ed. proper = education proper, Ed. tech. = educational technology, CSE = computer
science education.
Table 68
Comparison of the Proportion of Empirical, Human Participants Articles in Computer
Science Education and Education Technology
Field
Adjusted
Percentage
Percentage
residual
Method
C SE
Ed. tech.
Total
CSE
Ed. tech
(CSE)
Quantitative
107
280
387
74.3
56.7
3.8
Qualitative
22
174
196
15.3
35.2
-4.6
Mixed
15
40
55
10.4
8.1
0.9
Total
144
494
638
Note. CSE = computer science education, Ed. tech. = educational technology.
adjusted residuals show that authors of computer science education articles tended to
write, and get published, quantitative articles and tended to not write, or get published,
qualitative-only articles, compared to authors of papers published in educational
126
technology forums. The percentage of mixed-method articles in each field was about the
same however.
Table 69 shows that there was also a statistically significant difference, x 2 (2, N =
223) = 18.12,;? < .000, between the proportions of quantitative, qualitative, and mixed
methods articles between the fields of computer science education and education research
proper. The adjusted residuals show that the authors of computer science education
research articles tended to use quantitative methods and tended to not use qualitative
methods. Again, the proportions of mixed methods articles were about the same across
fields.
Table 69
Comparison of the Proportion of Empirical, Human Participants Articles in Computer
Science Education and Education Proper
Field
Adjusted
Ed.
Percentage
Percentage
residual
Method
C SE
proper
Total
CSE
Ed. proper
(CSE)
Quantitative
107
43
150
74.3
54.4
3.0
Qualitative
22
32
54
15.3
40.5
-4.2
Mixed
15
4
19
10.4
5.1
1.4
Total
144
79
223
Note. CSE = computer science education, Ed. proper = educational proper.
127
DISCUSSION
Study Limitations
One study limitation was that the interrater reliabilities were low on a small
proportion of the variables. I tried to circumvent this study limitation by not making
strong conclusions about variables with poor reliabilities or by qualifying claims that
were supported by variables with poor reliabilities.
As was mentioned in the Methods section, I recognize that I approached this
review from the viewpoint of a primarily quantitatively oriented behavioral science
researcher. I investigated most deeply the quantitative experimental articles and did not
deeply analyze articles that exclusively used explanatory descriptive modes of inquiry.
Because of the significant variety and variability of explanatory descriptive methods, I
was not confident that I could develop (or implement) a reliable system of classifying,
analyzing, and evaluating those articles. Therefore, another study limitation was that I
concentrated on experimental articles at the expense of explanatory descriptive articles.
A third limitation had to do with the coders not being blind to certain
characteristics of the articles (e.g., the institution, author, whether it came from a journal
or a conference proceeding). Therefore, coder bias was possible. However, I have reasons
to believe that coder bias did not unduly affect the results. The first is that because there
was an interrater reliability coder, the coder bias would have had to have operated in the
same direction for both coders, otherwise the interrater reliabilities would have been low.
Although it is possible that both the primary and secondary coders had the same bias, it is
128
less probable than just a single coder having the bias. Also, had there been coder bias, as I
discuss in the section on the difference between journal and conference papers, the bias
probably would have manifested itself in a way that supported the hypothesis. However,
on the variables where coder bias would have been harmful, such as the difference
between journals and conference proceedings, the results contradicted the hypothesis.
Interpretation of Descriptive Findings
My primary research question, which I addressed in terms of nine subquestions,
was- What are the methodological properties of research reported in articles in major
computer science education research forums from the years 2000-2005. A summary list of
answers to each of those research questions is given below:
1 . About one third of articles did not report research on human participants.
2. Most of the articles that did not deal with human participants were program
descriptions.
3. Nearly 40% of articles dealing with human participants only provided
anecdotal evidence.
4. Of the articles that provided more than anecdotal evidence, most articles used
experimental/quasi-experimental or explanatory descriptive methods.
5. Questionnaires were clearly the most frequently used type of measurement
instrument. Almost all of the measurement instruments that should have psychometric
information provided about them did not have psychometric information provided.
129
6. Student instruction, attitudes, and gender were the most frequent independent,
dependent, and mediating/moderating variables, respectively.
7. Of the articles that used an experimental research design, the majority used the
one-group posttest-only design.
8. When inferential statistics were used, the amount of statistical information
used was inadequate in many cases.
Because of the poor interrater reliabilities, I am hesitant about making summary
conclusions about the types of articles that did not deal with human participants (related
to Question 2) and about the question related to article structures (Question 9).
In terms of my secondary research questions about islands of practice, I conducted
15 planned contrasts. Those 15 contrasts concerned the differences between journals and
conference papers, yearly trends, and the regions of affiliation of the first authors, on the
major methodological variables: proportion of anecdotal only papers, proportion of
experimental/quasi-experimental papers, proportion of explanatory descriptive papers,
proportion of papers using a one-group posttest-only design, and proportion of papers
measuring attitudes only. The major findings abut the islands of practice and trends in
computer science education research are listed below:
9. There was no difference in major methodological characteristics between
articles published in computer science education journals and those published in peer-
reviewed conference proceedings. However, there is some evidence that there was a
slightly higher proportion of experimental/quasi-experimental articles in conference
proceedings when a region by forum type reaction is controlled for.
130
10. There was a decreasing yearly trend in the number of anecdotal-only articles
and in the number of articles that used explanatory descriptive methods.
1 1 . First authors affiliated with North American institutions tended to publish
papers in which experimental/quasi-experimental methods were used; first authors
affiliated with Middle Eastern or European institutions tended to not publish papers in
which experimental or quasi-experimental methods were used.
12. First authors affiliated with Middle Eastern institutions strongly tended to
publish explanatory descriptive articles.
13. First authors affiliated with Asian-Pacific or Eurasian institutions tended to
publish articles in which attitudes were the sole dependent variable; and
14. First authors affiliated with North American institutions tended to publish
more anecdotal-only articles than their peers in other regions. However, this proportion
had been decreasing linearly over time.
Proportion of Human Participants Articles
My prediction for the proportion of articles that would not report research on
human participants; which was based on the Randolph, Bednarik, and Myller (2005); was
between 80% and 60%. However, the proportion in the current review (33.8%) was about
30% lower than I had predicted. My explanation for this discrepancy is that the Koli
forum, on which my prediction was based, simply had a higher proportion of research that
did not deal with human participants than the computer science education research in
general.
131
Proportion of Program Description Articles
Earlier I made a prediction that the majority of articles that would not deal with
human participants would be program descriptions; that prediction was confirmed. Of the
34% of papers that did not report research on human participants, most (60%) of the
papers were purely descriptions of interventions without any analysis of the effects of the
intervention on computer science students. This proportion of articles is slightly higher,
but near, the proportion of program descriptions in other computing-related
methodological reviews in which the proportion of program descriptions was measured.
Assuming that Valentine's (2004) categories — Marco Polo and Tools — coincide with my
program description category, then Valentine's findings are similar to my own; he found
that 49% of computer science education research articles are what he called Marco Polo
or Tools articles. In addition, Tichy and colleagues (1995) found that 43% of the
computer science articles in their study were design and modeling articles, which would
be called program descriptions in my categorization system.
One of the assumptions of this dissertation is that the proportion of program
description-type articles is an indicator that the engineering tradition of computer science
(see Tedre, 2006) is an artifact in computer science education research. Although it would
be foolish to recommend an ideal proportion of program description and formalist articles
to empirical articles dealing with human participants, perhaps a statement by Ely, one of
the key figures in educational technology, can help inform the practice of computer
science education. In an article in which Ely re-examined some of his assertions about the
philosophy of educational technology made 30 years prior, he had the following to say
132
about his earlier assertion that "the behavioral science concept of instructional technology
is more valid than the physical science concept" (1999, p. 307):
The original intent of this statement [that the behavioral science concept of
instructional technology is more valid than the physical concept] was to contrast
the psychology of learning (behavioral science) with the hardware/software
aspects of technology (physical science). Using the same construct today,
behavioral science becomes psychology of learning and instruction while physical
science remains as the hardware/software configurations that deliver education
and training. The psychological concept here is often referred to as instructional
design (or sometimes, instructional systems design). There is growing evidence
that the use of instructional design procedures and processes lead to improved
learning without regard to the hardware and software that is used. Design is a
more powerful influence on learning than the system that delivers it. (p. 307)
[Italics added]
The conclusion I drew from this quote, which can also be applied to computer science
education, is that while many computer science educators may be experts at creating the
software and hardware to create automated interventions to increase the learning of
computer science, an increased emphasis should be put on the instructional design of the
intervention rather than only or primarily on the software and hardware mechanisms for
delivering the instructional intervention merits careful consideration.
One way to inform the dialogue about the distributions of research methods in
computer science education is to examine statements from authorities such as Ely or the
variety of working groups on computer science education. Another way to inform the
dialogue is to relate the research areas in computer science education to the types of
research methods that are used in it.
In terms of the types of research areas in computer science education, there are
several taxonomy systems that have been used. These include taxonomies presented in
133
Fincher and Petre (2004), Glass and colleagues (2004), and Valentine (2004). Pears,
Seidman, Eney, Kinnunen, and Malmi (2005) critically reviewed those taxonomies and
concluded that Fincher and Petre 's taxonomy of research areas was superlative because it
"corresponded best to the diversity of computing education research" (p. 154).
Fincher and Petre's 10 research areas (as cited in Pears et al.) are listed below:
1 . Student understanding.
2. Animation, visualization, and simulation.
3 . Teaching methods .
4. Assessment.
5. Educational technology.
6. Transferring professional practice to the classroom.
7. Incorporating new developments and new technologies.
8. Transferring from campus-based teaching to distance education.
9. Recruitment and retention.
10. Construction of the discipline, (p. 153)
In terms of the types of research methods that are used in fields related to
information technology, Jarvinen (2000) has proposed a useful taxonomy. In that
taxonomy of research approaches, Jarvinen first divided the variety of research
approaches into two classes: (a) approaches studying reality and (b) mathematical
approaches. Jarvinen further divided the "approaches studying reality" category into five
subcategories: (a) conceptual-analytical approaches, (b) theory-testing approaches, (c)
134
theory-creating approaches, (d) artifacts-building approaches, and (e) artifacts-evaluating
approaches.
Now, relating Jarvinen's (2000) taxonomy of research approaches to Fincher and
Petre's (2005) taxonomy of research areas, the relation between the distribution of
research approaches and the major research areas becomes clearer. From my perspective,
categories 1, 2, 3, 4, 6, 7, 8, 9, and the research component of Category 5 — educational
technology-lend themselves to empirical research with human participants. The
development component of the educational technology category, in as much as that means
the development of learning technologies, lends itself to what Jarvinen calls artifacts-
building approaches. I do not consider Fincher and Petre's "incorporating new
developments and new technologies" research area to be an area that refers to the
construction of new developments and technologies. I argue, rather, that it refers to the
implementation of technologies into the physical learning environment, which is a
research area that lends itself to empirical approaches that deal with human participants.
If the majority of research areas in Fincher and Petre's (2005) taxonomy do lend
themselves to empirical research approaches that deal with human participants, then it
would make sense to assume that the majority of research approaches would be empirical
research approaches that deals with human participants. Indeed, that was what was found
in this methodological review: Over 66% of the research papers in this review used
approaches that dealt with human participants (see Table 25). One interesting finding
though was that there was such a large proportion of reports on artifact-building (i.e.,
what I called program descriptions) given that the artifacts-building approach was directly
135
relevant in only 1 subcategory in 1 out of 10 of Fincher and Petre's categories — the
development component of the educational technology category. In fact, about 21%
(78/ 352) of the total articles sampled in this methodological review were purely program
descriptions. The conclusion that I drew from this finding was that the research areas in
Fincher and Petre's taxonomy are not equally represented in the computer science
education research literature — it seems that the development component of the
educational technology research area makes up a larger part of the computer science
education literature than the other research areas.
In fact, the development component in the computer science education research
literature makes up an even larger proportion than the developmental component in the
educational technology research literature itself. Supposing that across the fields of
educational technology and computer science education research there are equal
proportions of program/tool descriptions in the articles that do not deal with human
participants, then the proportion of program/tool descriptions in the computer science
education research literature is almost 15% higher than in the field of education
technology (see Table 69). This finding is surprising because one would assume that
computer science education is a field characterized as largely technology education, not
educational technology.
Proportions of Anecdotal-only Articles
The issue of the proliferation of anecdotal evidence in computing research,
136
especially in software engineering, was being addressed over ten years ago. Holloway
(1995) wrote:
Rarely, if ever, are [empirical claims about software engineering] augmented with
anything remotely resembling either logical or experimental evidence. Thus, one
can conclude that software engineering is based on a combination of anecdotal
experience and human authority. That is, we know that a particular technique is
good because John Doe, who is an authority in the field says that it is good
(human authority); John Doe knows that it is good because it worked for him
(anecdotal experience). Resting an entire discipline on such a shakey
epistemological foundation is absurd, but ubiquitous nonetheless, (p. 21)
As Table 28 showed, the proliferation of anecdotal evidence is also an issue for
the current computer science education research. The proportion of anecdotal-only
articles was 22.3% higher than I had predicted based on previous research.
Note that by the term anecdotal evidence in this review I have meant the informal
observation of a phenomenon by a researcher. I do not necessarily mean that humans
cannot make valid and reliable observations themselves, as happens in ethnographic
research or research in which humans operationalize and empirically observe behavior.
Also, I concur that anecdotal experience has a role in the research process-it has a role in
hypothesis generation. But, as Holloway (1995) pointed out, there are major problems to
using informal anecdotal experience as the sole means of hypothesis confirmation.
Valentine in his methodological review came to the same conclusion about the
proliferation of anecdotal evidence in the field computer science education research. In
fact, he ended his article with a call for more research not based on anecdotal experience.
Valentine (2004) wrote:
We need more [conclusions that are based on defensible research, and not mere
assumptions] of this in SIGCSE. I challenge the creators of CS1/CS2 Tools, in
137
particular to step up and prove to us that your Tool actually does what you are
claiming that it does. Do the fundamental research necessary to rest your claims
upon defensible fact. (p. 259)
This sentiment about the importance of collecting empirical data is also echoed in several
papers on computer science education research such as Clancy, Stasko, Guzdzial,
Fincher, and Dale (2001) and Holmboe, Mclver, and George (2001).
Also concerning anecdotal evidence, it is important that computer science
education researchers make claims that are congruent with the quantity and quality of
evidence that was collected. For example, if a CSE researcher were to write "Our
intervention caused students to learn more, more quickly" and the evidence that was
collected consisted only of informal, anecdotal observations, then that would surely be an
example of a mismatch between what was claimed and what, in the spirit of scientific
honesty, should have been claimed. I did not code for a mismatch between a claim and
what could have been claimed based on anecdotal evidence. However, based on my own
anecdotal experience from reviewing about one quarter of the mainstream computer
science education research published between 2000 and 2005, 1 hypothesize that this
mismatch between claim and evidence for the claim does exist and that it is even
common.
Types of Research Methods Used
I predicted that most articles that provided more than anecdotal evidence for their
claims would use experimental/quasi-experimental or exploratory descriptive methods
more than other methods. I was correct in the prediction that experimental/quasi-
138
experimental methods would be used more frequently than other methods. However, I
was wrong on the other part of the prediction; explanatory descriptive methods were used
more often than exploratory descriptive methods. Perhaps this a good sign for the state of
computer science education research; it signals a shift from the description of phenomena
to the causal explanation of phenomena.
Experimental/quasi-experimental and explanatory descriptive methods are both
methods that allow researchers to make causal inferences, and thereby confirm their
causal hypotheses (Mohr, 1999). Experimental/quasi-experimental research is predicated
on a comparison between a counterfactual and factual condition, via, what Mohr called,
factual causal reasoning. Explanatory descriptive research is predicated on what Mohr
called physical causal reasoning, or what Scriven (1976) called the Modus Operandi
Method of demonstrating causality.
To illustrate the difference between these approaches, suppose that it is a
researcher's task to prove that turning on the light switch in a room causes that room's
light to come on. Using factual causal reasoning the researcher would conduct an
experiment in which the researcher would note that when the switch is put in the "off
position, the light goes off (the factual condition); that when the switch is put in the "on"
position, the light goes on (the counterfactual condition); and that the light never goes on
unless the switch is in the on position, and vice versa — disregarding the possibility of a
burnt-out bulb. Through this factual causal reasoning process of comparing factual and
counterfactual conditions the researcher would arrive at the conclusion that turning the
switch on causes the light to go on.
139
On the other hand, if the researcher were to use physical causal reasoning to
determine if turning the switch on causes the light to come on, the process would be
entirely different. The research might tear through the walls and examine the switch, the
light, the power source, and the electrical wiring between the switch, the light, and the
power source. By knowing the theory of how electricity and circuits work, the researcher,
without ever having turned on the switch would be able to say with confidence that
turning on the switch will cause the light to come on.
At any rate, the fact that most of the research being done in computer science
education is done with types of methods that could possibly arrive at causal conclusions
(given that the research is conducted properly) is a positive sign for computer science
education research. Explanatory descriptive researchers in computer science education
use physical causal reasoning to arrive at their causal conclusions; experimental
researchers compare factual and counterfactual conditions. This fact indicates that
computer science education researchers are asking causal questions and also choosing
methods that can answer causal questions, if the method is conducted properly.
Types of Measures Used
Based on previous research I predicted that questionnaires, grades, and log files
would be the most frequently used types of measures. I was correct except that teacher- or
researcher-made tests were used more often than log files.
Another prediction was that few or none of the measures that should have had
psychometric information reported, had that information reported. This was especially
140
true of questionnaires; only 1 out of 65 articles in which questionnaires were used gave
any information about the reliability or validity of the instrument. According to
Wilkinson et al., "if a questionnaire is used to collect data, summarize the psychometric
properties of its scores with specific regard to the way the instrument is used in a
population. Psychometric properties include measures of validity, reliability and internal
validity" (1999, n.p). Obviously, the lack of psychometric information about instruments
is a clear weakness in the body of the computer science education research.
Proportions of Dependent, Independent, and
Mediating/Moderating Variables Examined
My prediction was that student instruction, attitudes, and type of course would be
the most frequently used types of independent, dependent, and mediating/moderating
variables, respectively. My prediction was correct.
Mark Guzdzial, one of the members of the working group on Challenges to
Computer Science Education Research, admits that, "We know that student opinions are
unreliable measures of learning or teaching quality" (Almstrum et al., 2005, p. 191). Yet,
this review shows that attitudes are the most frequently measured variable. In fact, 44%
of articles used attitudes as the sole independent article. While attitudes may be of
interest to computer science education researchers, as Guzdzial suggests, they are
unreliable indicators of learning or teaching quality.
Experimental Research Design Used
I was correct in my prediction that the one-group posttest-only and posttest-only
141
with control designs would be the most frequently used type of research designs. It is
important to note that the one-group posttest-only design was used more than twice as
often as the next most frequently used design, the posttest-only design with controls.
Although the one-group posttest-only design is the most common experimental
design in computer science education research, it is also probably the worst of the
experimental research designs in terms of internal validity. According to Shadish et al.
(2002), "nearly all threats to internal validity except ambiguity about temporal precedence
usually apply to this design. For example a history threat is nearly always present because
of other events might have occurred at the same time as the treatment" (p. 107). They do
argue, however, that
the [one-group posttest-only] design has merit in rare cases in which much
specific background knowledge exists about how the dependent variable behaves.
. . For valid descriptive causal inferences to result, the effects must be large
enough to stand out clearly, and either the possible alternative causes must be
known and be clearly implausible or there should be no known alternative that
could operate in the study context (Campbell, 1975). These conditions are rarely
met in the social sciences, and so this design is rarely useful in this simple form.
(P- 107)
The obvious conclusion is that the one-group posttest-only design is poor for
making causal inferences in most cases. Other designs, with pretests and/or control
groups, obviously would be better design choices if the goal is causal inference.
In terms of random selection and random assignment, I correctly predicted that
these would be rare in the computer science education research. Convenience samples
were used in 86% of articles, and students self-selected into treatment and control
conditions in 87% of the articles.
142
While some, such as Kish (1987) and Lavori, Louis, Bailar, and Polansky (1986),
are staunch advocates of the formal model of sampling (i.e., random sampling followed
by random assignment), there are others that question that model's utility. Shadish and
colleagues (2002) claim that formal sampling methods have limited utility for the
following reasons:
1 . The [formal] model is rarely relevant to making generalizations to treatments
and effects.
2. The formal model assumes that sampling occurs from a meaningful
population, though ethical, political, and logical constraints often limit random
selection to less meaningful populations.
3. The formal model assumes that random selection and its goals do not conflict
with random assignment and its goals.
4. Budget realities rarely limit the selection of units to a small and geographically
circumscribed population at a narrowly prescribed set of places and times.
5. The formal model is relevant only to generalizing to populations specified in
the original sampling plan and not to extrapolating to populations other than
those specified.
6. Random sampling makes no clear contribution to construct validty. . . (p. 348)
Shadish and colleagues (2002) concluded that "although we unambiguously
advocate [formal random sampling] when it is feasible, we cannot rely on it as an all-
purpose theory of generalized theory of causal inference. So researchers must use other
theories and tools to explore generalized causal inference of this type" (p. 348). Some of
the 'other theories and tools to explore generalized causal inference" are listed below:
1 . Assessing surface similarity-'assessing the apparent similarities between study
operations and the prototypical characteristics of the target population" (p. 357).
2. Ruling out irrelevancies-'identifying those attributes of persons, settings,
treatments, and outcome measures that are irrelevant because they do not
change a generalization" (p. 357).
3. Making discriminations-'identifying those features of persons, settings,
treatments, or outcomes that limit generalization" (p. 357).
4. Interpolating and extrapolating-'generalizing by interpolating to unsampled
143
values within the range of sampled persons, settings, treatments, and outcomes
by extrapolating beyond the sampled range (p. 366).
5. Making causal explanation-developing and testing explanatory theories about
the target of generalization (p. 366).
This same notion was expressed by Wilkinson et al. (1999). They stated:
Using a convenience sample does not automatically disqualify a study from
publication, but it harms your objectivity to try to conceal this by implying that
you used a random sample. Sometimes the case for the representativeness of a
convenience sample can be strengthened by explicit comparison of sample
characteristics with those of a defined population across a wide range of variables.
(n.p.)
The conclusion for computer science education researchers is that while random
sampling is desirable when it can be done, doing purposive sampling or at least assessing
the representativeness of a sample by examining surface similarities, ruling out
irrelevancies, making discriminations, and interpolating and extrapolating, and examining
causal explanations can be a reasonable alternative.
In terms of random assignment of participants to treatment conditions, the same
types of lessons apply. While random assignment is desirable, when it is not feasible
there are other ways to make strong causal conclusions. This is explained in Wilkinson et
al. (1999):
For research involving causal inferences, the assignment of units to levels of the
causal variable is critical. Random assignment (not to be confused with random
selection) allows for the strongest possible causal inferences free of extraneous
assumptions. If random assignment is planned, provide enough information to
show that the process for making the actual assignments is random.
For some research questions, random assignment is not feasible. In such
cases, we need to minimize effects of variables that affect the observed
relationship between a causal variable and an outcome. Such variables are
commonly called confounds or covariates. The researcher needs to attempt to
determine the relevant covariates, measure them adequately, and adjust for their
effects either by design or by analysis. If the effects of covariates are adjusted by
144
analysis, the strong assumptions that are made must be explicitly stated and, to the
extent possible, tested and justified. Describe methods used to attenuate sources of
bias, including plans for minimizing dropouts, noncompliance, and missing data.
(n.p.)
The conclusion for computer science education researchers is that when it is not
possible to randomly assign participants to experimental conditions, steps need to be
made, through design or analysis, to "minimize the effects of variables that affect the
observed relations between a causal variable and an outcomes" (Wilkinson et al., 1999,
n.p.).
Lack of Literature Reviews
I predicted that about 50% of articles sampled in the current review would lack a
literature review section. However, I am not confident about making a strong claim about
the presence or absence of literature reviews in the articles in the current review because
of the low levels of interrater agreement on this variable and on the other variables
dealing with report elements. However, I think that the fact that two raters could not
reliably agree on the presence or absence of key report elements; such as the literature
review, research questions, report elements, description of participants, description of
procedure; at least points out that these elements need to be explained more clearly. For
example, if two raters cannot agree on whether or not there is a literature review in an
academic paper, I am inclined to believe that the literature review is flawed in some way.
Assuming that the literature reviews in computer science education research
articles are indeed lacking, then it is no surprise that the ACM SIGCSE Working Group
on Challenges to Computer Science Education concluded that there is a lack of
145
accumulated evidence and a tendency for computer science educators to "reinvent the
wheel" (Almstrum et al., 2005, p. 191). Besides allowing evidence to accumulate and not
reinventing the wheel, conducting thorough literature reviews takes some of the burden
off researchers who are attempting to gather evidence for a claim since "good prior
evidence often reduces the quality needed for later evidence" (Mark, Henry, & Julnes,
2000, p. 87).
Also, one conclusion that can be drawn from the fact that the literature review and
other report elements variables had such low reliabilities is that the traditions of reporting
differ significantly between what is suggested by the American Psychological suggestion
and how most computer science education reports are structured. While not having agreed
upon structures enables alternative styles of reporting to flourish and gives authors plenty
of leeway to present their results, it makes it difficult for the reader to quickly extract
needed information from the articles. Additionally, I hypothesize that the lack of agreed
upon structures for computer science education articles leads to the omission of critical
information needed in reports of research with human participants, such as a description
of procedures and participants, especially by beginning researchers. Note that the report
element variables; such as the lack of a literature review, the lack of information about
participants or procedures, etc.; only pertained to articles that reported on investigations
with human participants and not to other types of articles, such as program descriptions or
theoretical papers, in which the report structures would obviously differ from a report of
an investigation with human participants.
146
Statistical Practices
The American Psychological Association (2001, p. 23) suggests that certain
information be provided when certain statistical analyses are used. For example when
parametric tests of location are used "a set of sufficient statistics consists of cell means,
cell sample sizes, and some measures of variability. . . . Alternately, a set of sufficient
statistics consists of cell means, along with the mean square error and degrees of freedom
associated with the effect being tested." Second, the American Psychological Association
(2001) and the American Psychological Association's Task Force on Statistical Inference
Testing (Wilkinson et al., 1999) argue that it is best practice to report an effect size in
addition to /^-values.
The results of this review showed that inferential analyses are conducted in 36%
of cases when quantitative results are reported. When computer science educators do
conduct inferential analyses, only a moderate proportion report informationally adequate
statistics. Areas of concern include reporting a measure of centrality and dispersion for
parametric analyses, reporting sample sizes and correlation or covariance matrices for
correlational analyses, and summarizing raw data when nonparametric analyses are used.
Islands of Practice
In this section I discuss where there were or were not differences in research
practices — in journals and conference proceedings, across regions, and across years. I
used two different kinds of statistical approaches-x 2 analyses of crosstabulation and
logistic regression-in my search for islands of practice. Most of the time those two
147
approaches yielded the same results, sometimes they did not. In the cases where there was
a discrepancy, I provide an explanation in this section. A summary of findings about
islands of is provided in the list below:
1. There were no difference between journals and conference proceedings in
terms of the proportions of anecdotal-only articles, explanatory descriptive articles,
attitudes-only articles, and one-group posttest-only articles. Controlling for a region by
forum type interaction, there is some evidence that the proportion of experimental/quasi-
experimental articles is greater in conferences than in journals.
2. Region was a statistically significant predictor on every outcome variable
except the proportion of one-group posttest-only articles.
a. Controlling for other factors, North American articles had a higher
proportion of anecdotal only articles than most other regions.
b. North American articles had higher proportion of experimental/quasi-
experimental articles than other regions.
c. Middle Eastern articles had a much higher proportion of explanatory
descriptive articles than articles from any other region.
d. Asian-Pacific/Eurasian articles had a higher proportion of attitudes-only
articles than did articles from other regions.
3. The proportion of anecdotal-only articles had decreased each year; the
strongest decrease was seen in North American articles. Also, the proportion of
explanatory descriptive articles had decreased every year.
148
Journal Versus Conference Papers
There has been an ongoing debate in the field of computer science education
about the relative merit that should be afforded to papers published in peer-reviewed
journals and those published in peer-reviewed conference proceedings (see Frailey, 2006;
Hodas, 2002). The outcomes of the debate about which academic publishing forums have
the most merit are important to several groups. According to Walstrom, Hardgrave and
Wilson (1995), those groups are:
• Selection, promotion, and tenure committees as they seek to secure and retain
the best possible individuals for the faculty;
• Researchers as they seek to determine appropriate outlets for their research
findings;
• Individuals seeking to identify the significant research streams in an academic
discipline;
• Journal editors and associates as they seek to raise the quality of their journal
[or conference] to the highest level possible;
• The academic discipline in question as it seeks to gain an identity of its own,
especially as it relates to a young field;
• Students of the discipline as they seek to gain an understanding of what the
discipline encompasses; and
• Librarians as they seek to wisely invest their ever-decreasing funds, (p. 93)
Particularly, the outcomes of the merit debate have serious economic
consequences for academic professionals who work in a "publish-or-perish" environment.
For example, Gill reports that "a published MIS [management information systems]
referred journal article can be worth approximately $20,000 in incremental pay, over an
assumed five-year lifetime, to a faculty member" (2001, p. 14).
In the computing sciences, the relative academic worth afforded to journal and
conference papers differs significantly from department to department. Some departments
reportedly do not accept conference proceedings in the tenure review process (Hodas,
149
2002), Grudin (2004) reported that "some departments equate two conference papers to a
journal article, or even award stature to papers in conferences that accept fewer than 25%
of submissions" (p. 12), while others assign value to each article, whether published
journal or conference proceedings, on a case-by-case basis (National Research Council,
1994). At any rate, the prevailing perception is that, generally, articles published in
archival journals receive more academic merit than articles published in conference
proceedings (National Research Council, 1994). Research conducted by the National
Research Council has shown that researchers and university administrators who believe
that journals are superior to conference proceedings believe so because of "the more
critical reviewing and permanent record of the former" (p. 138).
There has been much research done in the field of MIS on the relative qualities of
the different journal publication forums. The authors of that research (e.g.,
Katerattanakul, Han, & Hong, 2003; Rainer & Miller, 2005; Walstrom et al, 1995)
generally took a citation analysis approach or measured the perceptions of those articles.
However, that body of research is not directly applicable to this methodological review
because they compared journals with journals and they conducted the study in the field of
MIS, not computer science education.
There are a few methodological reviews of the computer science education
literature that have been published (Randolph, Bednarik, & Myller, 2006; Valentine,
2004). However, none of them specifically compared the methodological properties of
journal and conference articles. However, one study that did compare journal articles with
conference proceedings articles was conducted by the National Research Council (1994).
150
In that study they compared computer science journals and conference publications on
three variables: (a) time to publication, (b) median age of a reference, and (c) acceptance
rate. The National Research Council's findings are listed below:
1 . The median time from initial submission to publication in conference
proceedings was 7 months while in journals it was 31 months.
2. The median age of a reference (the median difference between the date of an
article's publication and the date of publication of the articles that were cited) was 3 years
for conference proceeding articles and nearly 5 years for journal articles.
3. The acceptance rate for prestigious conference proceedings, which ranged from
18 to 23%, was slightly lower that the estimated acceptance rate for journals, 25 to 30%.
Although the National Research Council study (1994) provided some interesting
results, it did not measure any construct dealing with the quality of the articles published
in each of those forums. Given that the National Research Council's findings above are
true, journal and conference articles might still differ substantially in terms of the quality
of methodological practices used, which is one claim made by those who support giving
more merit to journals.
If the variables-proportion of anecdotal-only articles, proportion of attitudes-only
articles, proportions of articles using a one-group posttest-only design only, and
proportion of experimental articles — are valid indicators of the methodological quality of
articles, the hypothesis that computer science education journal articles are more
methodologically sound than computer science education conference proceedings articles
turned out to be wrong. In fact, there is some evidence that conference proceedings have a
151
higher proportion of experimental/quasi-experimental articles than journal articles, when
a region by forum type interaction is controlled for.
Crosstabulation Tables 39 through 43 showed that there were no statistically
differences on any of the outcome variables, including the proportion of experimental/
quasi-experimental articles. When aggregating across regions and year, there is even a
slightly greater proportion of experimental/quasi-experimental journal articles than
conference articles (69.7% vs. 68.3%), see Table 40. However, using the logistic
regression approach in which the unique effect of each predictor could be estimated and
interactions could be modeled, there is evidence that the odds of a conference article's
being experimental/quasi-experimental is greater than the odds for a journal paper. There
was a statistically significant interaction between forum type and region. This interaction
helps explain the incongruence between the aggregate, crosstabulation analysis and the
logistic regression analysis.
Figure 9 shows that the proportion of experimental/quasi-experimental journal
articles is much lower than the proportion of experimental/quasi-experimental conference
papers for European and Asian-Pacific/Eurasian articles. However, the opposite is the
case for North American articles; there are more experimental/quasi-experimental
conference papers than there are experimental/quasi-experimental journal articles. My
hypothesis for why this interaction exists rests on two assumptions.
The first is that journals are less influenced by regional affects than are conference
proceedings. For example, authors who have a paper accepted at a conference are
physically expected to appear at the conference to present their results. The effect is that
152
people tend to attend, and submit papers to, conferences that are nearby. A quick glance
at the conference proceedings included in this sample will support this point. Therefore,
the research practices in a certain region will be reflected to some degree in the
conference proceedings. The same does not hold for journals or holds to a lesser degree;
authors of journal manuscripts are not expected to travel to the physical location where a
journal is published.
The second assumption is that North American researchers tend to write and get
published experimental/quasi-experimental articles more than European and Asian-
Pacific/Eurasian et al. authors. This assumption is backed up from the region section of
Table 57 and from Table 49.
Therefore, because of the greater effect of region on conference proceedings than
on journals and because of the tendency of North American researchers to do
experimental research, the interaction is not surprising. The interaction seems to be strong
enough that when included in the regression equation, it can switch the direction of the
odds ratio (i.e., the predicted odds of a conference article's being an experimental/quasi-
experimental article becomes greater than the odds of a journal article's being an
experimental/quasi-experimental article.) Whether the interaction term is included or not,
the results overall indicate that there are nonsignificant differences, or differences slightly
in favor of conferences, in terms of the proportion of experimental/quasi-experimental
articles in journals and conference proceedings. The results from both analyses indicate
that there are no statistically significant differences between journals and conference
153
proceedings in terms of the proportions of anecdotal-only, explanatory descriptive,
attitudes-only, or one-group posttest-only articles.
One limitation regarding this finding was that the coders were aware of whether
the article being coded came from a conference proceeding or from a journal. Thus, it is
plausible that experimenter bias could have come into play — the coders might have
tended to code journal articles more leniently than conference articles because of a pre-
existing belief that journal articles are more methodologically sound. Blind review was
not possible in this case because the length of the article would usually entail its status; if
the article was five pages or less, it was most likely a conference proceeding paper.
However, there is one reason that I believe that experimenter bias was not a serious threat
in this study. If there had been experimenter bias, it should have worked in favor of the
hypothesis that journal articles are more methodologically sound than conference
proceedings articles; however, that was not the case.
In terms of informing policy for the personnel evaluation of computer science
education researchers, the major implication of this finding is that it is inadvisable to
summarily give less academic merit to conference proceedings than to journal articles,
because their methodological soundness has been shown to be similar. I acknowledge,
however, that the methodological soundness of an article should not be the only way that
an article is evaluated. In essence, I agree with the Patterson, Snyder, and Ullman,
representatives of the Computing Research Association, who wrote:
For the purposes of evaluating a faculty member for promotion or tenure, there are
two critical objectives of an evaluation: (a) establish a connection between a
faculty member's intellectual contribution and the benefits claimed for it, and (b)
154
determine the magnitude and significance of the impact. Both aspects can be
documented, but it is more complicated than simply counting archival
publications. . . . Not all papers in high quality publications are of great
significance, and high quality papers can appear in lower quality venues.
Publication's indirect approach to assessing impact implies that it is useful, but
not definitive. The primary direct means of assessing impact — to document items
(a) and (b) above — is by letters of evaluation from peers. (1999, pp. A-B)
Although publication counting and using merit formulas (e.g., that two conference papers
are worth one journal article) are easy evaluation strategies, there can be no substitute for
case-by-case assessment in which a variety of factors are taken into account in the gestalt
of a faculty member's academic output.
Yearly Trends
Valentine (2004) identified several encouraging trends in computer science
education research from 1984 to 1999. First, the number of technical symposium
proceedings had been increasing each year. Second, the percentage of experimental
articles (loosely defined as the author having made "any attempt at assessing the
'treatment' with some scientific analysis" [p. 256]) had increased since the mid '90s.
Third, the percentage of Marco Polo articles (which probably would correspond with
what I called anecdotal-only articles) had shown a yearly decrease.
The findings of this methodological review show that two out of the three trends
identified by Valentine (2004), from 1984 to 1999, continued in the years from 2000 to
2005. First, as is evident from Table 5, the number of articles in the SIGCSE Technical
Symposium (and in computer science education forums in general) has still been on the
rise. Second, the decline in the number of anecdotal-only/Marco Polo articles had
155
continued to decline in the years from 2000-2005. The decline was most pronounced for
North American articles. In contrast to what Valentine found, it was not found that the
proportions of experimental articles had continued to increase into the years from 2000 to
2005. However, it is important to note here that I used a more conservative definition of
experimental than did Valentine. I assume that, in addition to true experiments or quasi-
experiments, Valentine would have included explanatory descriptive, exploratory
descriptive, correlational, and causal comparative investigations in the "experimental"
category. I, on the other hand, only included actual experiments or quasi-experiments in
the experimental category.
Region of Origin
Concerning region of first author's origin, both the crosstabulation approach and
the logistic regression approach revealed several differences in the way that computer
science education researchers from institutions in different regions conduct research:
1 . Computer science education researchers from North American institutions
tended to do experimental research, while their European and Middle Eastern
counterparts tended to not do experimental research;
2. Computer science education researchers from Middle Eastern institutions
strongly tended to do explanatory descriptive (qualitative) research;
3. North American researchers tended to do anecdotal-only research more than
their peers in other regions, but the proportions of North American anecdotal research
156
articles had been on the decline while the proportions had been stable across time for the
other regions; and
4. Computer science education researchers from Asian-Pacific or Eurasian
institutions tended to measure attitudes only.
Disentangling the relationship between the factors related to the environment that
a group of scientists work in and how they carry out their research is difficult (see
Depaepe, 2002). It is like speculating how the work of the Vienna School, for example,
would have been different had they been the Toledo (Ohio) School instead. Nonetheless,
below I describe some of my hypotheses, which might be used to inform further
investigations, about why the results may have turned out as they did.
One possible reason for the tendency for North American education researchers to
do experiments could be that the worth attributed to randomized field trials by the U.S.
Department of Education, a major source of funding for U.S. education researchers, has
something to do with the tendency of North American researchers (of whom most are
from U.S. institutions) to do experimental research. The U.S. Department of Education
(2002) made the following statement about the relative importance they give to
descriptive studies and to "rigorous field trials of specific interventions":
Descriptive implementation studies play a crucial role in understanding the impact
of policy changes, but they are no substitute for rigorous field trials of specific
interventions.
Even with high-quality fast-response surveys, annual performance data, and
descriptive studies, we still cannot answer the question on the minds of
practitioners: "What works?" To be able to make causal links between
interventions and outcomes, we need rigorous field trials, complete with random
assignment, value-added analysis of longitudinal achievement data, and distinct
interventions to study.
157
This approach might be considered "research" rather than "evaluation."
Whatever the name, the Department's evaluation agenda would be incomplete
without it. It is a fair use of evaluation dollars because federal program funds are
paying for the interventions to be studied. (Para. 24-26)
This policy is a hotly-debated topic in U.S. research and evaluation circles (see
Donaldson & Christie, 2005; Julnes & Rog, in press; or Lawrenz & Huffman, 2006).
Regardless of the propriety of this policy, the quote above shows that U.S. educational
policymakers give value and funding priority to true experiments, and, it is not surprising
then that many U.S. education researchers strive to do experimental research.
Second, the tendency of European researchers to not do experimental research is
congruent with the contemporary European decline in the popularity of the study of
quantitative research methods. Rautopuro and Vaisanen (2005); well-known Finnish,
quantitative-research-method educators; wrote the following about the state of
quantitative research methods, at least in Finland:
The level of skills in the quantitative methods seems to be worrying. In
educational science, too, the level of method used as well as how they are used in
quantitative research in all levels — from master theses to dissertations — is getting
out of hand. The students do not get excited of taking voluntary quantitative
research methods courses and therefore are not capable to use them in their own
research. Compulsory statistics courses, as well, are only a necessity for the
students and sometimes for the researcher, too. Moreover, one generation of
educational researchers, at least partially, have lost the competence of applying
quantitative research methods and because of this they have also lost the
possibility to pass on the tradition of the use of these methods, (p. 273)
If Rautopuro and Vaisanen's (2005) findings generalize to the rest of Europe (and
there is reason to believe that it does — see European Science Foundation, 2004), then it
is no surprise that there is a tendency for European computer science researchers to not do
experimental research. One possible reason for this could be that the resurgence of the
158
qualitative research tradition has had a greater influence in Europe than in North
America, according to Fielding (2005). Fielding speculated that the "American
quantitative approach was influential during this period [i.e., the resurgence of the
qualitative method since the publication of Glaser and Strauss's Discover of Grounded
Theory in 1967, Strauss and Corbin's revision of it in 1990, and Turner's influential 1981
paper on qualitative data analysis] too but qualitative methodology was arguably more
secure in the European curriculum due to the import of hermeneutics in German social
philosophy and the life history method in French and Italian sociology" (2005, para. 12).
Fielding (2005) also mentioned that qualitative research has become increasingly
legitimized and institutionalized in the European social science research curriculum since
the 1980s. One example of this institutionalization of qualitative research that Fielding
provides are the postgraduate training guidelines written by the United Kingdom's
Economic and Social Research Council (ESRC). According to Fielding those curriculum
guidelines
strongly emphasize qualitative methods and require that students understand
archival, documentary and historical data, life stories, visual images and materials,
ethnographic methods, cases studies and group discussions, at least one qualitative
software package, and a range of analytic techniques including conversation
analysis and discourse analysis. Since the guidelines are written by senior
academics, they clearly index the institutionalization of qualitative methods.
(Para. 21)
Concerning the finding that computer science education researchers affiliated with
Middle Eastern institutions tended to do explanatory descriptive research, a quick
examination of the Middle Eastern institutions from which the Middle Eastern articles
came sheds light on this finding. Three Israeli institutions accounted for over half of the
159
Middle Eastern computer science education articles. Those institutions were the Technion
- Israel Institute of Technology, the Weizmann Institute of Science, and Tel-Aviv
University, which contributed 23.1, 23.1, and 1 1.5% of the total number of Middle
Eastern computer science articles included in this sample.
One interesting finding was that North American papers had a significantly higher
proportion of anecdotal-only papers than other regions (see Figure 7), but that this
proportion had been declining over time in North American papers. As Figure 6 shows, in
2000 the proportion of North American anecdotal-only papers was about 80%; in 2005
the proportion was about equal with the proportions of other regions at about 30%.
Although I do not have any informed hypotheses about why the proportion of anecdotal-
only North American papers would have been so much higher than in other regions in
2000, 1 do have one hypothesis about why the proportion of anecdotal-only articles had
been declining steadily only in North America, besides the fact that extreme scores tend
to regress towards the mean.
Given that more than one third of the total computer science articles came from
the SIGCSE Conference Proceedings, which were held in the United States from 2000
through 2005, one possible explanation is that the decline in North American conference
papers is heavily correlated with a decline in anecdotal-only papers in SIGCSE
conference proceedings. (In fact, the Spearman correlation of the percent anecdotal-only
by year between the SIGCSE Conference Proceedings and North American articles in
general was quite high, r(6) = .87, p < .02.) In addition, that decline in the proportion of
anecdotal-only SIGCSE conference papers could be a result of the increased interest in
160
the methodological qualities of the articles published in SIGCSE Proceedings, which is
evident in recent SIGCSE Conference Proceedings articles, such as Valentine (2004), and
working group reports, such as Almstrum, Ginat, Hazzan, and Clement (2003) and
Almstrum and colleagues (2005). One flaw with this hypothesis though is that there has
also been a recent interest in the methodological quality of computer science education
research articles across the range of computer science publication forums, which is
evident in articles such as Almstrum et al. (2002); Bouvier, Lewandowski, and Scott
(2003); Carbone and Kaasbooll (1998); Clear (2001); Daniels, Petre, and Berglund
(1998); Fincher et al. (2005); Fincher and Petre (2004); Greening 1997); Lister (2005);
Pears and colleagues (2005); Pears, Daniels, and Berglund (2002); Randolph, Bednarik,
and Myller (2005), and Sandstrom and Daniels (2000), among others.
Differences Across Fields
Earlier I predicted that computer science education research would have the
greatest proportion of papers that do not empirically deal with human participants,
educational technology papers would have fewer of those papers than computer science
education papers, and that education research proper papers would have the fewest of
those types of papers. That prediction turned out to be correct. Assuming that the
proportion of papers that do not empirically deal with human participants are, more or
less, indicators of engineering and/or formalist traditions lingering in computer science
education, then, it can be said that computer science education is a field in which the
traditions of computer science research proper, especially the engineering tradition, bleed
161
through to the practice of computer science education research. Computer science
education researchers, as a whole, publish more "I engineered this intervention to certain
specifications" types of articles and less "I empirically evaluated the effects of this
intervention on student learning" types of articles than their counterparts in educational
technology. In turn, educational technologists, as a whole, publish more engineering types
of articles and less empirical types of articles than their counterparts in educational
research proper.
In terms of the proportions of qualitative, quantitative, and mixed-methods
research, computer science educators tended to use quantitative methods more frequently
and qualitative research less frequently than their counterpart researchers in educational
technology or education proper. This might come as a source of concern to the factions of
computer science education researchers who call for more qualitative research, such as
Ben-Ari, Berglund, Booth, and Holmboe (2004); Berglund, Daniels, and Pears (2006);
Hazzan, Dubinsky, Eidelman, Sakhnini, and Teif (2006) and Lister (2003).
Profile of the Average Computer Science Education Paper
From these results, it is possible to create a profile of the average computer
science education research paper. It is important to note that this profile is a synthesis of
averages; there might not actually be an average paper that has this exact profile.
Nonetheless, I include the average profile here because of the narrative efficiency in
which it can characterize what computer science education research papers, in general,
are like. The profile follows:
162
The typical computer science education research paper is a 5 -page conference
paper written by two authors. The first author is most likely affiliated with a university in
North America. If the article does not deal with human participants, then it is likely to be
a description of some kind of an intervention, such as a new tool or a new way to teach a
course. If the article does deal with human participants, then there is a 40% chance that it
is basically a description of an intervention in which only anecdotal evidence is provided.
If more than anecdotal evidence is provided the authors probably used a one-group
posttest-only design in which they gave out an attitude questionnaire, after the
intervention was implemented, to a convenience sample of first-year undergraduate
computer science students. The students were expected to report on how well they liked
the intervention or how well they thought that the intervention helped them learn. Most
likely, the authors presented raw statistics on the proportions of students who held
particular attitudes.
Recommendations
In this section I report on what I consider to be the most important evidence-based
recommendations for improving the current state of computer science education. Because
I expect that the improvements will be most likely effected by editors and reviewers
raising the bar in terms of the methodological quality of papers that get accepted for
publication, I direct these recommendations primarily to the editors and reviewers of
computer science education research forums. Also, these recommendations are relevant to
funders of computer science research; to consumers of computer science education
163
research, such as educational administrators; and, of course, to computer science
education researchers themselves.
Accept Anecdotal Experience as a Means of
Hypothesis Generation, But Not as a Sole
Means of Hypothesis Confirmation
While a field probably cannot be built entirely on anecdotal experience (although
some might not agree), that does not mean that anecdotal experience does not have an
important role in scientific inquiry — it has an important role in the generation of
hypotheses. Sometimes it is through anecdotal experience that researchers come to
formulate important hypotheses. However, because of its informality, anecdotal
experience is certainly a dubious type of evidence for hypothesis confirmation.
Not accepting anecdotal evidence as a means of hypothesis confirmation is not to
say that a human cannot make valid and reliable observations. However, there is a
significant difference between a researcher reporting that "we noticed that students
learned a lot from our program" and a researcher who reports on the results of a well-
planned qualitative inquiry or on the results of carefully controlled direct observations of
student behavior, for example. Also when anecdotal evidence is presented either as a
rationale for a hypothesis to be investigated or as evidence to confirm a hypothesis, it
should be clearly stated that anecdotal experience was the basis for that evidence.
Be Wary of Investigations That Only Measure
Students 'Self-Reports of Learning
Of course, stakeholders' reports about how much they have learned are important;
164
however, it probably is not the only dependent of variable of interest in an educational
intervention. As a measure of learning, as Guzdzial (in Almstrum et al., 2005) has
pointed out, students' opinions are poor indicators of how much learning has actually
occurred.
Insist That Authors Provide Some Kind of
Information About the Reliability and
Validity of Measures That They Use
Wilkinson et al. (1999) provided valuable advice to editors concerning this issue,
especially in "a new and rapidly growing research area" (like computer science
education). They advised,
Editors and reviewers should pay special attention to the psychometric properties
of the instrument used, and they might want to encourage revisions (even if not by
the scale's author) to prevent the accumulation of results based on relatively
invalid or unreliable measures, (n.p.)
Realize That The One-Group Posttest-Only
Research Design Is Susceptible to Almost
All Threats to Internal Validity
In the one-group posttest-only design, almost any influence could have caused the
result. For example, in a one-group posttest-only design, if the independent variable was
an automated tool to teach programming concepts and the dependent variable was the
mastery of programming concepts, it is entirely possible that, for example, students
already knew the concepts before using the tools, or that something other than the tool
(e.g., the instructor) caused the mastery of the concepts. Experimental research designs
165
that compare a factual to a counterfactual condition are much better at establishing
causality than research designs that do not.
Report Informationally Adequate Statistics
When inferential statistics are used, be sure that the author includes enough
information for the reader to understand the analysis used and to examine alternative
hypotheses for the results that were found. The American Psychological Association
(2001) gives the following guidelines:
Because analytic technique depends on different aspects of the data, it is
impossible to specify what constitutes a set of minimally adequate statistics for
every analysis. However, a minimally adequate set usually includes at least the
following: the per-cell sample size, the observed cell means (or frequencies of
cases in each category for a categorical variable), the cell standard deviations, and
an estimate of pooled within-cell variance. In the case of multi variable analytic
systems such as multivariate analyses, regression analyses, and structural equation
modeling analyses, the mean(s), sample size(s), and the variance-covariance (or
correlation) matrix or matrices are a part of a minimally adequate set of statistics,
(p. 23)
Insist that Authors Provide Sufficient Detail
about Participants and Procedures
When authors report research on human participants be sure that they include
adequate information about the participants, apparatus, and procedure. In terms of
adequately describing participants the American Psychological Association (2001)
suggests the following:
When humans participated as the subjects of the study, report the procedures for
selecting and assigning them and the agreements and payments made. . . . Report
major demographic characteristics such as sex, age, and race/ethnicity, and where
possible and appropriate, characteristics such as socio-economic status, disability
status, and sexual orientation. When a particular demographic characteristic is an
166
experimental variable or is important for the interpretation of results, describe the
group specifically-for example, in terms of national origin, level of education,
health status, and language preference .... Even when a characteristic is not an
analytic variable, reporting it may give readers a more complete understanding of
the sample and often proves useful in meta-analytic studies that incorporate the
article's results, (pp. 18-19)
In terms of the adequate level of detail for the Procedures section, the American
Psychological (2001) gives the following advice:
The subsection on procedures summarizes each step in the execution of the
research. Include the instructions to the participants, the formation of the groups,
and the specific experimental manipulations. Describe randomization,
counterbalancing, and other control features in the design. Summarize or
paraphrase instructions, unless they are unusual or compose an experimental
manipulation, in which case they may be presented verbatim. Most readers are
familiar with standard testing procedures; unless new or unique procedures are
used, do not describe them in detail.
If a language other than English is used in the collection of information, the
language should be specified. When an instrument is translated into another
language, the specific method of translation should be described (e.g., back
translation, in which a text is translated into another language and then back into
the first to ensure that it is equivalent enough that the results can be compared.)
Remember that the Method section should tell the reader what you did and
how you did it in sufficient detail so that a reader could reasonably replicate your
study. Methodological articles may defer highly detailed accounts of approaches
(e.g., derivations and details of data simulation approaches) to an appendix, (p. 20)
In short, enough information should be provided about participants so that readers can
determine generalization parameters and enough information should be provided about
the procedure that it could be independently replicated.
An Example of a High-Quality Computer Science
Education Research Article
In this section I examine in detail one article that I think is a particularly good
167
example of high quality computer science education research and evaluate it in terms of
the recommendations that I mentioned above. All though there were many high-quality
articles in the sample that would have worked for this purpose, I chose Sajaniemi and
Kuittinen's (2005) "An Experiment on Using Roles of Variables in Teaching Introductory
Programming" because it was particularly clear and well-written and is exemplary in the
areas that my recommendations relate to. (Although Jorma Sajaniemi works in the same
department as I, this did not influence my choosing this article — at least that I am aware
of. It was a random chance that this article was included in my sample in the first place.)
The article is somewhat atypical in that that it is a 25 -page journal paper (published in
Computer Science Education), whereas most computer science education research papers
are 5 -page conference papers.
To get a sense of what the article is about in general I have included the text from
entire abstract below:
Roles of variables is a new concept that captures tacit expert knowledge in a form
that can be taught in introductory programming courses. A role describes some
stereotypic use of variables, and only ten roles are needed to cover 99% of all
variables in novice-level programs.
This paper presents the results of an experiment where roles were
introduced to novices learning Pascal programming. Students were divided into
three groups that were instructed differently: in the traditional way with no
treatment of roles; using roles throughout the course; and using a role-based
program animator in addition to using roles in teaching.
The results show that students are not only able to understand the role
concept and to apply it in new situations but — more importantly — that roles
provide students a new conceptual framework that enables them to mentally
process program information in a way demonstrating good programming skills.
Moreover, the use of the animator seems to foster the adoption of role knowledge.
(P- 59)
168
According to the Publication Manual of the American Psychological Association
(American Psychological Association, 2001) the abstract of an empirical report should
describe
• the problem under investigation, in one sentence if possible;
• the participants or subjects, specifying pertinent characteristic, such as
number, type, age, sex, . . . ;
• the experimental method, including the apparatus, data-gathering procedures,
[and] complete test names. . . .;
• the findings, including statistical significance levels; and the conclusions and
the implications or applications, (p. 14).
Sajaniemi and Kuitten's abstract described most of the information that the
Publication Manual of the American Psychological Association calls for. The exceptions
were, however, that Sajaniemi and Kuitten did not include as detailed information about
participants as called for by the American Psychological Association, information about
data-gathering procedures, and information about the significance level of findings.
Overall, however, the abstract accurately summarizes the important parts of the article
and, admittedly, Sajaniemi and Kuitten may have written their article according to some
other publication manual than the Publication Manual of the American Psychological
Association.
The introduction of their article clearly introduced the problem (a need for and
lack of research on the role concept in teaching programming) and answered the
following questions (from American Psychological Association, 2001, pp. 15-16):
169
1 . Why is the problem important? (The answer could inform the teaching of
programming.)
2. How do the hypothesis and the experimental design relate to the problem?
(The hypothesis relates to a new way of teaching programming; the experimental design
allows for an examination of the effects of that way of teaching programming or learning
of programming.)
3. What are the theoretical implications of the study, and how does the study
relate to previous literature? (The study informs theories about the different theories of
teaching programming and can also inform other learning theories, such as the dual-
coding theory, the cognitive constructivism theory, and the epistemic fidelity theory; the
study relates to a new category of research on teaching of programming — software design
patterns and roles of variables.)
4. What theoretical propositions are tested, and how were they derived. (The
study tests the proposition that teaching roles of variables facilitates student learning of
programming; Sajaniemi and Kuittinen provide a detailed research history of how those
theoretical propositions were derived from previous research over the past 20 years.)
In the introduction of their article, Sajaniemi and Kuittien developed the
background of the study with a discussion of the previous literature on teaching of
programming, discussed how the theory being tested was derived, and gave a history and
description of the intervention(s) that were used. As the Publication Manual of the
American Psychological Association suggests, they cited "only works pertinent to the
specific issue and not works of only tangential or general significance" (American
170
Psychological Association, 2001, p. 16). Also, Sajaniemi and Kuitten clearly stated the
purpose of their study, "to find out the effects of using the role concept in teaching
programming to novices" (p. 60), and their research hypothesis — "introducing roles of
variables in teaching facilitates learning to program" (p. 64).
The Publication Manual of the American Psychological Association (2001)
suggests that the Method section should enable "the reader to evaluate the
appropriateness of your methods and the reliability and validity of your results. It also
permits experienced investigators to replicate the study if they so desire" (p. 17) and that
it should, in most cases, contain the following subsections: participants, apparatus, and
procedure. The Method section of Sajaniemi and Kuittinen's paper met all of those
suggestions.
The Participants section of their paper (Sajaniemi and Kuittinen called it the
Subjects section) provided detailed information about several participant variables that
could have been confounded with treatment in the experiment. Some of those participants
variables were the number of subjects; gender; performance in high school mathematics,
information technology, art; previous spreadsheet creation experience; previous
programming courses; and previous programming experience. In short, they provided
enough information about the participants that other researchers and practitioners would
be able to establish generalization parameters and, by measuring variables that were
thought to be possible confounding factors, were able to rule out a host of extraneous
threats to internal validity.
In the Apparatus section, which Sajaniemi and Kuittinen labeled the "Materials"
171
section, they provided detailed information on the measures that were used and even
provided a web link, which actually worked, to the experimental materials that were used.
The only information missing from the description of the examination was information
about previous investigations on the validity or reliability of the measurement instrument
(the examination).
In the beginning of the Method section and in the Procedure section Sajaniemi and
Kuittinen provided copious detail about the research design (a between-subject design
with the content of instruction as the between-subject factor, with researcher and grader
blinding) and study procedures used. In my opinion, they provided enough information
that other researchers could replicate the study.
In the Results section, Sajaniemi and Kuittenen did appropriate statistical analysis
and presented informationally adequate statistics for the types of analyses the
conducted-means, standard deviations, and n-sizes; correlational and raw effect sizes;
and the value of the test statistic, degrees of freedom, and probability values. And they
also presented a number of graphs to aid in the interpretation of results. The only
information that would have improved this Results section is information on the interrater
reliability estimates between graders.
In the Discussion section and Conclusion section, Sajaniemi and Kuittinen
summarized their findings, revisited their research hypotheses, and related their findings
back to the previous literature. They also outlined the implications of their study,
discussed alternative hypotheses, and commented on study limitations.
172
This article can serve as a model for other computer science researchers in how to
avoid the pitfalls common in the computer science research. First, they did a carefully
controlled and rigorous study so that evidence could be collected that could help confirm
or disconfirm their hypothesis. They used a design that is much better than the one-group
posttest-only design for ruling out threats to internal validity. They created an instrument
to measure learning instead of relying on students self-reports on whether they had
learned or not. Although they did not provide information about the psychometric
properties of their measurement instrument, they did describe the instrument in detail and
their rationale for its validity. Also, they gave readers direct access to the actual
measurement instrument that was used so that the readers could make their own
judgments about the psychometric properties of the instrument. They provided rich
enough detail of the participants, materials, and procedures used that the reader could
clearly understand what happened in the experiment and could even replicate it. Finally,
they provided informationally adequate statistics in the Results section.
It is true that they had 25 pages in which to work and that normally computer
science education research forums allow only up to 5 pages. Nevertheless, a 5-page
empirical report should also have the same elements as a 25 -page report-only the level of
detail might change. Articles such as Clark, Anderson, and Chalmers (2002); Lee et al.
(2002); and Olson et al. (2002), although in the field of medical science, are good
examples of how empirical reports can be written in such a way that they are complete,
but also very concise.
173
CONCLUSION
Summary
In this dissertation, I used a content analysis approach to conduct a
methodological review of the articles published in mainstream computer science
education forums from 2000 to 2005. Of the population of articles published during that
time a random sample of 352 articles was drawn; each article was reviewed in terms of its
general characteristics; the type of methods used; the research design used; the
independent, dependent, and mediating or moderating variables used; the measures used;
and statistical practices used. The major findings from the review are listed below:
1 . About one third of articles did not report research on human participants.
2. Most of the articles that did not deal with human participants were program
descriptions.
3. Nearly 40% of articles that dealt with human participants only provided
anecdotal evidence for their claims.
4. Of the articles that provided more than anecdotal evidence, most articles used
experimental/quasi-experimental or explanatory descriptive methods.
5. Of the articles that used an experimental research design, the majority used a
one-group posttest-only design exclusively.
6. Student instruction, attitudes, and gender were the most frequent independent,
dependent, and mediating/moderating variables, respectively.
174
7. Questionnaires were clearly the most frequently used type of measurement
instrument. Almost all of the measurement instruments that should have psychometric
information provided about them did not have psychometric information provided.
8. When inferential statistics were used, the amount of statistical information
used was inadequate in many cases.
9. There was no difference in major methodological characteristics between
articles published in computer science education journals and those published in peer-
reviewed conference proceedings. However, there is some evidence that when controlling
for the interaction between region and forum type, the odds of an article's being
experimental/quasi-experimental was higher in conference proceedings.
10. There was a decreasing yearly trend in the number of anecdotal-only articles
and in the number of articles that used explanatory descriptive methods.
1 1 . First authors affiliated with North American institutions tended to publish
papers in which experimental/quasi-experimental papers were used; first authors
affiliated with Middle Eastern or European institutions tended not to publish papers in
which experimental or quasi-experimental methods were used.
12. First authors affiliated with Middle Eastern institutions strongly tended to
publish explanatory descriptive articles.
13. First authors affiliated with Asian-Pacific or Eurasian institutions tended to
publish articles in which attitudes were the sole independent variable.
14. First authors affiliated with North American institutions tended to publish
anecdotal-only articles; however, that proportion of North American anecdotal-only
175
articles had declined linearly over time and was about equal to the proportion in other
regions by 2005.
15. Computer science education research forums published more engineering-
oriented program-description types of papers than educational technology forums
published and much more than education research proper forums published.
16. Computer science education researchers, in general, tended to use quantitative
methods and tended not to use qualitative methods more than their counterparts in
educational technology or education research proper.
Based on these findings, I made the following recommendations to editors,
reviewers, authors, funders, and consumers of computer science education research:
1. Accept anecdotal experience as a means of hypothesis generation, but not as
the sole means of hypothesis confirmation.
2. Be wary of investigations that measure only students' attitudes and self-reports
of learning as a result of an intervention.
3. Insist that authors provide some kind of information about the reliability and
validity of measures that they use.
4. Realize that the one-group posttest-only research design is susceptible to
almost all threats to internal validity.
5. Encourage authors to report informationally adequate statistics.
6. Insist that authors provide sufficient detail about participants and procedures.
176
Computer Science Education Research at the Crossroads
Based on the results of this review, I can say that what computer science educators
have so far been great at is generating a large number of informed research hypotheses,
based on anecdotal experience or on poorly designed investigations. However, they have
not systematically tested these hypotheses. This leaves computer science education at a
crossroads. To the crossroads computer science education researchers bring a
proliferation of well-informed hypotheses. What will happen to these hypotheses remains
to be seen.
One option is that these informed hypotheses will overtime, through repeated
exposure, "on the basis of 'success stories' and slick sales pitches" (Holloway, 1995, p.
20) come to be widely accepted as truths although having never been empirically verified.
That is, they will become folk conclusions. (I use the term folk conclusions instead of folk
theorems [see Harel, 1980] ox folk myths [see Denning, 1980] since the validity of the
conclusion has not yet been empirically determined.)
The consequences of accepting folk conclusions that are not actually true can be
serious. Although speaking in the context of software engineering, but which probably
still applies to some degree computing education as well, Holloway (1995) wrote:
I pray that it will not take the loss of hundreds of lives in an airplane crash, or
even the loss of millions of dollars in a financial system collapse, before we
acknowledge our ignorance and redirect our efforts away from [promoting folk
conclusions] and towards developing a valid epistemological foundation, (p. 21)
Because scientific knowledge usually develops cumulatively, if informed
hypotheses are allowed to developed into folk conclusions, then layers of folk
177
conclusions (both true and untrue) will become inexorably embedded in the cumulative
knowledge of what is known about computer science education. Computer science
education will become a field of research whose foundational knowledge is based on
conclusions that are believed to be true, but which have never been empirically verified.
Indeed, as Holloway suggests "resting an entire discipline on such a shaky
epistemological foundation is absurd . . ." (1995, p. 21). In the same vein, basing the
future of an entire discipline on such a shaky epistemological foundation is also absurd.
I am not arguing, however, that hypothesis generation or any other type of
research activity in computer science education should be abandoned altogether. There
needs to be a requisite variety of methods to draw from so that a rich variety of research
acts can be carried out. Also, hypothesis generation is inexorably tied with innovation.
What I am arguing is that the proportions of research methods being used needs to
be congruent with the current challenges and problems in computer science education. If
the ACM SIGCSE's Working Group on Challenges to Computer Science Education is
correct that the current challenges involve a lack or rigor and accumulated evidence, then
it makes sense to shift the balance from one that emphasizes anecdotal evidence and
hypothesis generation to one that emphasizes rigorous methods and hypothesis
confirmation. Coming back to the discussion of the crossroads, the sustainable path for
computer science education involves building on the hypotheses of the past and striking a
balance between innovation and experimentation in the future.
178
REFERENCES
Agresti, A. (1996). An introduction to categorical data analysis. New York: Wiley.
Alexander, S., & Hedberg, J. G. (1994). Evaluating technology-based learning: Which
model? In K. Beattie, C. McNaught, & S. Willis (Eds.) Interactive Multimedia in
higher education: Designing for change in teaching and learning (pp. 233-244).
Amsterdam: Elsevier.
Almstrum, V. L., Ginat, D., Hazzan, O., & Clement, J. (2003). Transfer to/from
computing science education: The case of science education research. In
Proceedings of the 34 th SIGCSE Technical Symposium on Computer Science
Education, 303-304. New York: ACM Press.
Almstrum, V. L., Ginat, D., Hazzan, O., & Morely, T. (2002). Import and export to/from
computing science education. The case of mathematics education research.
Proceedings of the 7 th Annual Conference on Innovation and Technology in
Computer Science Education ITiCSE '02, 193-194. New York: ACM Press.
Almstrum, V. L., Hazzan, O., Guzdzial, M., & Petre, M. (2005). Challenges to computer
science education research. Proceedings of the 36 th SIGCSE Technical Symposium
on Computer Science Education SIGCSE '05, 191-192. New York: ACM Press.
American Psychological Association. (2001). Publication manual of the American
Psychological Association (5th ed.). Washington, DC: Author.
Babbitt, T. (2001). A tale of two shortages: An analysis of the IT professional and MIS
faculty shortages. Proceedings of the 2001 ACM SIGCPR Conference on
Computer Personnel Research, 21-23. New York: ACM Press.
ih
Ben-Ari, M., Berglund, A., Booth, S., & Holmboe, C. (2004). What do we mean by
theoretically sound research in computer science education? Proceedings of the 9
Annual Conference on Innovation and Technology in Computer Science
Education ITiCSE '04, 230-231. New York: ACM Press.
Berglund, A., Daniels, M., & Pears, A. (2006). Qualitative research projects in computing
education research. Proceedings of the 8 th Australian Conference on Computing
Education, 52, 25-33. Darlinghurst, Australia: Australian Computer Society.
179
Bouvier, D., Lewandowski, G., & Scott, T. (2003). Developing a computer science
education research program. Journal of Computing Sciences in Colleges, 79(1),
218.
Brennan, R. L., & Prediger, D. J. (1981). Coefficient kappa: Some uses, misuses, and
alternatives. Educational and Psychological Measurement, 41, 687-699.
Caffarella, E. P. (1999). The major themes and trends in doctoral dissertation research in
educational technology from 1977 through 1998. Proceedings of selected research
and development papers presented at the National Convention of the Association
for Educational Communications and Technology (AECT). (ERIC Document
Reproduction Service No. ED 436178)
Campbell, D. T. (1975). "Degrees of freedom" and the case study. Comparative Political
Studies, 8, 178-193.
Carbone, A., & Kaasbooll, J. (1998). A survey of methods used to evaluation computer
science teaching. SIGCSE Bulletin, 30(3), 41-45
Clancy, M., Stasko, J., Guzdzial, M., Fincher, S., & Dale, N. (2001). Model and areas for
CS education research. Computer Science Education, 11(4), 323-341.
Clark, M., Anderson, P., & Chalmers, I. (2002). Discussion sections in reports of
controlled trials published in general medical journals. Journal of the American
Medical Association, 287(21), 2799-2800.
Clark, R. E., & Snow, R. E. (1975). Alternative designs for instructional technology
research. AV Communications Review, 23, 373- 394.
Clear, T. (2001). Thinking issues. Research paradigms and the nature and meaning of
truth. SIGCSE Bulletin, 33(2), 9-10.
Cohen, J. (1994). The earth is round (p < .05). American Psychologist, 49, 997-1003.
Cooper, H. M. (1988). Organizing knowledge synthesis: A taxonomy of literature
reviews. Knowledge in Society, 1, 104-126.
Cooper, H. M., & Hedges, L. V. (1994). Research synthesis as a scientific enterprise. In
H. Cooper & L. V. Hedges (Eds.), The handbook of research synthesis (pp. 3- 14).
New York: Russell Sage Foundation.
180
Daniels, M., Petre, M., & Berglund, A. (1998). Building a rigorous research agenda into
changes to teaching. Proceedings of the 3 rd Australasian Conference on Computer
Science Education, 203-209. New York: ACM Press.
Denning, P. J. (1980). On folk theorems, and folk myths. Communications of the ACM,
23(9), 493-494.
Denning, P. J., Comer, D. E., Gries, D., Mulder, M. C, Tucker, A., Turner, A. J., et al.
(1989). Computing as a discipline. Communications of the ACM, 32(1), 9-23.
Depaepe, M. (2002). A comparative history of the educational sciences: The
comparability of the incomparable? European Education Research Journal, 7(1),
118-122.
Dick, W., & Dick, W. D. (1989). Analytical and empirical comparisons of the Journal of
Instructional Development and Educational Communication and Technology
Journal. Educational Technology Research & Development, 37(1), 81-87.
Donaldson, S. I., & Christie, C. A. (2005). The 2004 Claremont Debate: Lipsey vs.
Scriven. Determining causality in program evaluation and applied research:
Should experimental evidence be the gold standard. Journal of Multidisciplinary
Evaluation, 3, 60-77.
Driscoll, M. P., & Dick, W. (1999). New research paradigms in instructional technology.
An inquiry. Educational Technology Research & Development, 47(2), 7-18.
Edgington, E. S. (1964). A tabulation of inferential statistics used in psychology journals.
American Psychologist, 19, 202-203.
Elmore, P. B., & Woehlke, P. J. (1988). Statistical methods employed in American
Educational Research Journal, Educational Researcher, and Review of
Educational Research from 1978 to 1987. Educational Researcher, 17(9), 19-20.
Elmore, P. B., & Woehlke, P. L. (1998, April). Twenty years of research methods
employed in American Education Research Journal, Educational Researcher, and
Review of Educational Research. Paper presented at the Annual Meeting of the
American Educational Research Association, San Diego, CA.
Ely, D. (1999). Toward a philosophy of instructional technology. Thirty years on. British
Journal of Educational Technology, 30(4), 305-310.
181
European Science Foundation. (2004, October). Quantitative methods in the social
sciences (QMSS): An ESF scientific programme. Retrieved August 24, 2006, from
http://www.esf.org/publication/193/QMSS.pdf
Fielding, N. (2005, May). The resurgence, legitmation and institutionalization of
qualitative methods. Forum Qualitative Sozialforschung / Forum: Qualitative
Social Research, 6(2), Article 32. Retrieved September 25, 2006, from
http://www.qualitative-research.net/fqs-texte/2-05/05-2-32-e.htm
Fincher, S., Lister, R., Clear, T., Robins, A., Teneberg, J., & Petre, M. (2005). Multi-
institutional, multi-national studies in CSEd research: Some design considerations
and trade-offs. Proceedings of the 2005 International Workshop on Computing
Education Research ICER '05, 1 1 1-121. New York: ACM Press.
Fincher, S., & Petre, M. (2004). Computer science education research. London: Taylor &
Francis.
Frailey, D. J. (2006, July 10). What price publications? Message posted to SIGCSE
Member Forum electronic mailing list. Retrieved October 24, 2006, from
http://listserv.acm.org/scripts/wa.exe?A2=ind0607b&L=sigcse-
members&F=&S=&P=2865
Gall, M. D., Borg, W. R., & Gall, J. P. (1996). Educational research: An introduction
(6th ed.). New York: Longman.
Garson, D. V. (2006). Sampling. Retrieved March 28, 2006, from North Carolina State
University, College of Humanites & Social Science Web site: http://www2.
chass.ncsu.edu/garson/PA765/sampling.htm
Gill, T. G. (2001). What's an MIS paper worth? (An exploratory analysis). The DATA
BASE for Advances in Information Systems, 32(2), 14-33.
Glaser, B., & Strauss, A. (1967). The discovery of grounded theory. New York: Aldine.
Glass, R. L., Ramesh, V., & Vessey, I. (2004). An analysis of research in computing
disciplines. Communications of the ACM, 47(6), 89-94.
Good, P. I. (2001). Resampling methods. A practical guide to data analysis (2nd ed.).
Boston: Birkhauser.
Goodwin, L. D., & Goodwin, W. L. (1985a). An analysis of the statistical techniques used
in the Journal of Educational Psychology, 1979-1983. Educational Psychologist,
20, 13-21.
182
Goodwin, L. D., & Goodwin, W. L. (1985b). Statistical techniques in AERJ articles,
1979-1983: The preparation of graduate students to read the educational research
literature. Educational Researcher, 14(2), 5-11.
Gorard, S., & Taylor, C. (2004). Combining methods in educational and social research.
Berkshire, England: Open University Press.
Greening, T. (1997). Paradigms for educational research in computer science.
Proceedings of the 2nd Australasian Conference on Computer Science Education,
47-5 1 . New York: ACM Press.
Grosberg, J. (n.d.). Statistics 101 [Computer Software]. Retrieved July 11, 2006, from
http://www. statistics 101 .net/index. htm
Grudin, J. (2004). Crossing the divide. ACM Transactions on Computer-Human
Interaction, 11(1), 1-25.
Harel, D. (1980). On folk theorems. Communications of the ACM, 23(1), 379-494.
Hazzan, O., Dubinsky, Y., Eidelman, L., Sakhnini, V., & Teif, M. (2006). Qualitative
research in computer science education. Proceedings of the 37th SIGCSE
Technical Symposium on Computer Science Education, 408-412. New York:
ACM Press.
Higgins, N., Sullivan, H., Harper-Marinick, M., & Lopez, C. (1999). Perspectives on
educational technology research and development. Educational Technology
Research & Development, 37(1), 7-17 '.
Hodas, J. S. (2002, November 15). Re: Request for permission: Appropriate journals
(fwd). Message posted to SIGCSE Member Forum electronic mailing list.
Retrieved October 24, 2006, from http://listserv.acm.org/scripts/
wa.exe?A2=ind0211C&L=SIGCSE/MEMBERS&P=R350&I=-3
Holloway, C. M. (1995). Software engineering and epistemology. Software Engineering
Notes, 20(2), 20-21.
Holmboe, C, Mclver, L., & George, C. (2001). Research agenda for computer science. In
G. Kododa (Ed.), Proceedings of the 13 th Annual Workshop of the Pscyhology of
Programming Interest Group, 207-223. PPIG(http://www.ppig.org)
Huwiler-Miintener, K., Juni, P., Junker, C, & Egger, M. (2002). Quality of reporting of
randomized trials as a measure of methodologic quality. Journal of the American
Medical Association, 287(21), 2801-2804.
183
Jarvinen, P. H. (2000). Research questions guiding selection of an appropriate research
method. Proceedings of the 8th European Conference on Information Systems,
124-131. London School of Economics and Political Science. http://is2.
lse.ac.uk/asp/aspecis/Default.htm
Julnes, G., & Rog, D. J. (in press). Informing federal policies on evaluation methodology:
Building the evidence for method choice in government sponsored evaluation.
New Directions for Evaluation.
Kalton, G. (1983). Introduction to survey sampling. London: Sage.
Katerattanakul, P., Han, B., & Hong, S. (2003). Objective quality ranking of computer
journals. Communications of the ACM, 46(\0), 111-114.
Keselman, H. J., Huberty, C. J., Lix, L. M., Olejnik, S., Cribbie, R., Donahue, B., et al.
(1998). Statistical practices of educational researchers. An analysis of their
ANOVA, MANOVA, and ANCOVA analyses. Review of Educational Research,
68, 350-86.
Kinnunen, P. (n.d.). Guidelines of computer science education research. Retrieved
November 29, 2005, from http://www.cs.hut.fi/Research/COMPSER/
ROLEP/seminaari-k05/S_05-nettiin/Guidelines_o f_CSE-teksti-paivi.pdf
Kirk, R. E. (1996). Practical significance. A concept whose time has come. Educational
and Psychological Measurement, 56, 746-759.
Kish, I. (1987). Statistical design for research. New York: Wiley.
Klein, J. D. (1997). ETR&D development: An analysis of content and survey of future
direction. Educational Technology Research & Development, 45(3), 57-62
Lavori, P. W., Louis, T. A., Bailar, J. C, & Polansky, H. (1986). Design of experiments:
Parallel comparisons of treatments. In J. C. Bailar & F. Mosteller (Eds.), Medical
uses of statistics (pp. 61-82.) Waltham, MA: New England Journal of Medicine.
Lawrenz, F., & Huffman, D. (2006). Methodological pluralism: The gold standard of
STEM evaluation. In D. Huffman & F. Lawrenz (Eds.), New directions for
evaluation: Critical issues in STEM evaluation (pp. 19-34). Hoboken, NJ: Wiley.
Lawrenz, F., Keiser, N., & Lavoir, B. (2003). Evaluative site visits: A methodological
review. American Journal of Evaluation, 24, 341-352.
184
Lee, K. P., Schotland, M., Bachetti, P., & Bero, L. A. (2002). Association of journal
quality indicators with methodological quality of clinical research articles. Journal
of the American Medical Association, 287(1), 2805-2808.
Lister, R. (2003). A research manifesto, and the relevance of phenomenography. ACM
SIGCSE Bulletin, 35(2), 15-16.
Lister, R. (2005). CS research: Mixed methods: Positivists are from Mars, constructivists
are from Venus. SIGCSE Bulletin, 37(4), 18-19.
Mark, M. M., Henry, G. T., & Julnes, G. (2000). Evaluation: An integrated framework
for understanding, guiding, and improving policies and programs. San Francisco:
Jossey-Bass.
Mohr, L. B. (1999). The qualitative method of impact analysis. American Journal of
Evaluation, 20(1), 69-84.
National Research Council (NRC). (1994). Academic careers for computer scientists and
engineers. Washington, DC: National Academy Press.
National Research Council Committee on Information Technology Literacy. (1999, May).
Chapter I: Why know about information technology: Being fluent with
information technology. Washington, DC: National Academy Press. Retrieved
October 12, 2004, from http://books.nap.edu/html/beingfluent/chl.html
Neuendorf, K. A. (2002) The content analysis handbook. Thousand Oaks, CA: Sage.
Olson, C. M., Rennie, D., Cook, D., Dickersin, K., Flanagin, A., Hogan, J. W., et al.
(2002). Publication bias in editorial decision making. Journal of the American
Medical Association, 287(21), 2825-2828.
Patterson, D., Snyder, L., & Ullman, J. (1999, September). Best practices memo:
Evaluating computer scientists and engineers for promotion and tenure.
Computing Research New, A-B.
Pears, A., Daniels, M., & Berglund, A. (2002). Describing computer science education
research. An academic process view. Proceedings of the Conference on
Simulation and Multimedia in Engineering Education, ICSEE 2002, 99-104.
Pears, A., Seidman, S., Eney, C, Kinnunen, P., & Malmi, L. (2005). ITiCSE 2005
working group reports: Constructing a core literature for computing education
research. SIGCSE Bulletin, 3 7(4), 152-161.
185
Phipps, R., & Merisotis, J. (1999, April). What's the difference: A review of the
contemporary research on the effectiveness of distance learning in higher
education. The Institute for higher education policy. Retrieved July 27, 2006 from
http://www.ihep.org/Pubs/PDF/Difference.pdf
Rainer, R. K. Jr., & Miller, M. D. (2005). Examining differences across journal rankings.
Communications of the ACM, 48(2), 91-94.
Randolph, J. J. (2005) A methodological review of the program evaluations in K-12
computer science education. Manuscript submitted for publication. Available
online http://www.geocities.com/justusrandolph/cse_eval_review.pdf
Randolph, J.J. (in press). What's the difference, still: A follow-up review of the
quantitative research methodology in distance learning. Informatics in Education.
Randolph, J. J., Bednarik, R., & Myller, N. (2005). A methodological review of the
articles published in the proceedings of Koli Calling 2001-2004. In T. Salakoski,
T. Mantyla, & M. Laakso (Eds.), Proceedings of the 5th Annual Finnish /Baltic
Sea Conference on Computer Science Education, 103-109. Finland: Helsinki
University of Technology Press. Retrieved March 19, 2006, from
http://www.it.utu.fi/koli05/proceedings/final_composition.b5.060207.pdf
Randolph, J. J., Bednarik, R., Silander, P., Lopez-Gonzalez, J., Myller, N., & Sutinen, E.
(2005). A critical review of research methodologies reported in the full-papers of
ICALT 2004. Proceedings of the Fifth International Conference on Advanced
Learning Technologies, 10-14. Los Alamitos, CA: IEEE Press. Available online:
http://ieeexplore.ieee.org/xpls/abs_all.jsp7isnumberf23 17&
arnumber=1508593&count=303&index=4
Randolph, J. J., & Hartikainen, E. (2005). A review of resources for K-12 computer-
science-education program evaluation. Yhtendistyvdt vai erilaistuvat oppimisenja
koulutuksen polut: Kasvatustieteen pdivien 2004 verkkojulkaisu [Electronic
Proceedings of the Finnish Education Research Days Conference 2004] (pp. 183-
193). Finland: University of Joensuu Press. Available online: http://
www.geocities.com/ justusrandolph/ review_of_resources.pdf
Randolph, J. J., Hartikainen, E., & Kahkonen, E. (2004, November). Lessons learned from
developing a procedure for the critical review of educational technology research.
Paper presented at Kasvatustieteen Paivat 2004 [Finnish Education Research Days
Conference 2004], Joensuu, Finland.
186
Ranis, S. H., & Walters, B. H. (2004). Education research as a contested enterprise: The
deliberations of the SSRC-NAE Joint Committee on Education Research.
European Education Research Journal, 3(4), 795-806.
Rautopuro, J., & Vaisanen, P. (2005). DEEP WATER? Quantitative research methods in
educational science in Finland. In M-L Julkunen (Ed.), Learning and instruction
in multiple contexts and settings III. Proceedings of the Fifth Joensuu Symposium
on Learning and Instruction. Bulletins of the Faculty of Education, 96, 273-293.
Finland: University of Joensuu Press.
Reeves, T. C. (1995). Questioning the questions of instructional technology research.
Retrieved October 19, 2004, from http://www.hbg.ps.edu/bsed/intro/docs/dean/
Resampling Stats (Version 5.0.2) [Computer software and manual]. (1999). Arlington,
VA: Resampling Stats.
Research Assessment Exercise. (2001). Retrieved July 28, 2006, from http://
www.hero.ac.uk /rae/
Ross, S. M., & Morrison, G. R. (2004). Experimental research methods. In D. J. Jonassen
(Ed.) Handbook of research on educational communications and technology (2nd
ed., pp. 1021-1043.) Mahwah, NJ: Erlbaum.
Sajaniemi, J., & Kuittinen, M. (2005). An experiment on using roles of variables in
teaching introductory programming. Computer Science Education, 15{\), 59-82.
Sample Planning Wizard [Computer software]. (2005). Stat Trek.
Sandstrom, A., & Daniels, M. (2000). Time studies as a tool for (computer science )
education research. Proceedings of the Australasian Conference on Computing
Education, 208-214. New York: ACM Press.
Scriven, M. (1976). Maximizing the power of causal investigations: The modus operandi
method. In G. V. Glass (Ed.), Evaluation studies review annual (Vol. 1, pp. 101-
1 1 8). Beverly Hills, CA: Sage.
Shadish, W. R., Cook, T. D., & Campbell, D. T. (2002). Experimental and quasi-
experimental designs for generalized causal inference. Boston: Houghton Mifflin.
Simon, J. L. (1997). Resampling: The new statistics. Arlington, VA: Resampling Stats.
Stevens, J. (1999). Intermediate statistics: A modern approach (2nd ed.). Mahwah, NJ:
Erlbaum.
187
Strauss, A., & Corbin, J. (1990). Basics of qualitative research. Newbury Park, CA: Sage.
Tedre, M. (2006). The development of computer science: A sociocultural perspective.
Dissertation, University of Joensuu. Retrieved October 24, 2006, from http://
j oypub .joensuu.fi/publications/dissertations/tedre_development/index_en.html
Test, D. W., Fowler, C. H., Brewer, D. M., & Wood, W. M. (2005). A content and
methodological review of self-advocacy intervention studies. Exceptional
Children, 72, 101-125.
Thompson, B., & Snyder, P. A. (1998). Statistical significance and reliability analyses in
recent JCD research articles. Journal of Counseling and Development, 76, 436-
441.
Tichy, W. F., Lukowicz, P., Prechelt, L., & Heinz, E. A. (1995). Experimental evaluation
in computer science. A quantitative study. Journal of Systems and Software, 28, 9-
18.
Tucker, A., Deck, F., Jones, J., McCowan, D., Stephenson, C, & Verno, A. (ACM K-12
Education Task Force Curriculum Committee). (2003, October 22). A model
curriculum for K-12 computer science: Final report of the ACM K-12 Education
Task Force Curriculum Committee. Retrieved October 20, 2004, from
http://www.acm.org/education/kl2/kl2finall022.pdf
Turner, B. (1981). Some practical aspects of qualitative data analysis. Quality and
Quantity, 15, 225-247.
U.S. Department of Education. (2002). Press release: New directions for program
evaluation at the U.S. Department of Education. Retrieved August 24, 2006, from
http://www.ed.gov/news/pressreleases/2002/04/evaluation.html
U.S. Department of Labor-Bureau of Labor Statistics, (n.d.a). Computer support
specialists and systems administrators. In Occupational outlook handbook, 2004-
2005 edition. Retrieved October 22, 2004, from http://www.bls.gov/
oco/pdf/oco/ocos268.pdf
U.S. Department of Labor-Bureau of Labor Statistics, (n.d.b). Computer systems analysts
database administrators, and computer scientists. In Occupational outlook
handbook, 2004-2005 edition. Retrieved October 22, 2004, from
http://www.bls.gov/oco/pdf/oco/ocos042.pdf
188
Valentine, D. W. (2004). CS educational research: A meta-analysis of SIGCSE technical
symposium proceedings. Proceedings of the 35 th Technical Symposium on
Computer Science Education, 255-259. New York: ACM Press.
Walstrom, K. A., Hardgrave, B. G., & Wilson, R. L. (1995). Forums for management
information systems scholars. Communications of the ACM, 38(3), 93-107.
West, C. K., Carmody, C, & Stallings, W. M. (1983). The quality of research articles in
Journal of Educational Research, 1970 and 1980. Journal of Educational
Research, 77, 70-76.
Wilkinson, L., & Task Force on Statistical Inference. (1999). Statistical methods in
psychology journals: Guidelines and explanations [Electronic version]. American
Psychologist, 54, 594-604.
Williamson, A., Nodder, C, & Baker, P. (2001). A critique of educational technology
research in New Zealand since 1996. Retrieved October 20, 2004, from
http://www.ascilite.org.av/conferences/melbourne01/pdf/papers/williamson.pdf
Zelkowitz, M. V., & Wallace, D. (1997). Experimental validation in software
engineering. Information and Software Technology, 39, 735-743.
189
APPENDICES
190
Appendix A:
A List of the Articles Included in the Sample
Abunawass, A., Lloyd, W., & Rudolph, E. (2004). COMPASS: A CS program
assessment project. ITiCSE '04: Proceedings of the 9th Annual SIGCSE Conference
on Innovation and Technology in Computer Science Education, Leeds, United
Kingdom, 127-131. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/1007996.1008031
Acharya, M., & Funderlic, R. (2003). 'Laurel and Hardy' model for analyzing process
synchronization algorithms and primitives. SIGCSE Bulletin (Association for
Computing Machinery, Special Interest Group on Computer Science Education),
35(2), 107-110.
Adams, J. C, Bauer, V., & Baichoo, S. (2003). An expanding pipeline: Gender in
mauritius. SIGCSE '03: Proceedings of the 34th SIGCSE Technical Symposium on
Computer Science Education, Reno, NV, United States. 59-63, Retrieved September
4, 2006, from http://doi.acm.Org/10.l 145/61 1892.61 1932
Adams, J., Nevison, C, & Schaller, N. C. (2000). Parallel computing to start the
millennium. SIGCSE '00: Proceedings of the Thirty-First SIGCSE Technical
Symposium on Computer Science Education, Austin, TX, United States, 65-69.
Retrieved September 4, 2006, from http://doi.acm.org/10.1145/330908.331815
Aharoni, D. (2000). Cogito, ergo sum! cognitive processes of students dealing with data
structures. SIGCSE '00: Proceedings of the Thirty-First SIGCSE Technical
Symposium on Computer Science Education, Austin, TX, United States, 26-30.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/330908.331804
Aiken, R., Kock, N., & Mandviwalla, M. (2000). Fluency in information technology: A
second course for non-CIS majors. SIGCSE '00: Proceedings of the Thirty-First
SIGCSE Technical Symposium on Computer Science Education, Austin, TX, United
States, 280-284. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/330908.331870
Alphonce, C, & Ventura, P. (2002). Object orientation in CS1-CS2 by design. ITiCSE
'02: Proceedings of the 7th Annual Conference on Innovation and Technology in
Computer Science Education, Aarhus, Denmark. 70-74, Retrieved September 4,
2006, from http://doi.acm.org/10.! 145/544414.544437
191
Aly, A. A., & Akhtar, S. (2004). Cryptography and security protocols course for
undergraduate IT students. SIGCSE Bulletin (Association for Computing Machinery,
Special Interest Group on Computer Science Education) , 36(2), 44-47.
Anderson, R., Dickey, M., & Perkins, H. (2001). Experiences with tutored video
instruction for introductory programming courses. SIGCSE '01: Proceedings of the
Thirty-Second SIGCSE Technical Symposium on Computer Science Education,
Charlotte, NC, United States, 347-351. Retrieved September 4, 2006 from
http://doi.acm.org/10.1145/364447.364619
Anttila, I., Jormanainen, I., Kannusmaki, O. and Lehtonen, J. (2001). Lego-Compiler.
Kolin Kolistelut - Koli Calling 2001, Proceedings of the First Annual Finnish
Annual Finnish Baltic Sea Conference on Computer Science Education, Koli,
Finland, 9-12. Retrieved September 4, 2006, from http://cs.joensuu.fi/
kolistelut/archive/200 l/koli_proc_200 1 .pdf
Applin, A. G. (2001). Second language acquisition and CS1. SIGCSE '01: Proceedings of
the Thirty-Second SIGCSE Technical Symposium on Computer Science Education,
Charlotte, NC, United States, 174-178. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/364447.364579
Ariga, T., & Tsuiki, H. (2001). Programming for students of information design. SIGCSE
Bulletin (Association for Computing Machinery, Special Interest Group on
Computer Science Education) , 33(4), 59-63.
Armen, C, & Morelli, R. (2005). Teaching about the risks of electronic voting
technology. ITiCSE '05: Proceedings of the 10th Annual SIGCSE Conference on
Innovation and Technology in Computer Science Education, Caparica, Portugal,
227-231. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/1067445.1067508
Aycock, J., & Uhl, J. (2005). Choice in the classroom. SIGCSE Bulletin (Association for
Computing Machinery, Special Interest Group on Computer Science Education),
37(4), 84-88.
Bagert, D., & Mead, N. (2001). Software engineering as a professional discipline.
Computer Science Education, 11(1), 73-87.
Bailey, T., & Forbes, J. (2005). Just-in-time teaching for CS0. SIGCSE '05: Proceedings
of the 36th SIGCSE Technical Symposium on Computer Science Education, St.
Louis, MO, United States, 366-370. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/1047344.1047469
192
Baldwin, D. (2000). Some thoughts on undergraduate teaching and the Ph.D. SIGCSE
Bulletin (Association for Computing Machinery, Special Interest Group on
Computer Science Education) , 32(4), 60-62.
Barker, L., & Garvin-Doxas, K. (2004). Making visible the behaviors that influence the
learning environment: A Qualitative Exploration of Computer Science Classrooms.
Computer Science Education, 14(2), 119-145.
Barker, L. J., Garvin-Doxas, K., & Roberts, E. (2005). What can computer science learn
from a fine arts approach to teaching? SIGCSE '05: Proceedings of the 36th SIGCSE
Technical Symposium on Computer Science Education, St. Louis, MO, United
States, 421-425. Retrieved September 4, 2006, from http://doi.acm.Org/10.l 145/
1047344.1047482
Beck, L. L., Chizhik, A. W., & McElroy, A. C. (2005). Cooperative learning techniques
in CS1: Design and experimental evaluation. SIGCSE '05: Proceedings of the 36th
SIGCSE Technical Symposium on Computer Science Education, St. Louis, MO,
United States, 470-474. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/1047344.1047495
Becker, B. W. (2001). Teaching CS1 with Karel the robot in Java. SIGCSE '01:
Proceedings of the Thirty-Second SIGCSE Technical Symposium on Computer
Science Education, Charlotte, NC, United States, 50-54. Retrieved September 4,
2006, from http://doi.acm.Org/10.l 145/364447.364536
Bednarik, R. & Franti, P. (2004). Survival of students with different learning preferences.
Kolin Kolistelut - Koli Calling 2004, Proceedings of the Fourth Annual Finnish
Annual Finnish Baltic Sea Conference on Computer Science Education, Koli,
Finland, 121-125. Retrieved September 4, 2006, from http://cs.joensuu.fi/
kolistelut/archive/2004/koli_proc_2004.pdf
Bennedsen, J. & Caspersen, M. E. (2005). An upcoming study of potential success factors
for an introductory model-driven programming course. Kolin Kolistelut - Koli
Calling 2005, Proceedings of the Fourth Annual Finnish Annual Finnish Baltic Sea
Conference on Computer Science Education, Koli, Finland, 166-169. Retrieved
September 4, 2006, from http://cs.joensuu.fi/kolistelut/archive/
2005/koli_proc_2005.pdf
Beyer, S., Rynes, K, Perrault, J., Hay, K, & Haller, S. (2003). Gender differences in
computer science students. SIGCSE '03: Proceedings of the 34th SIGCSE Technical
Symposium on Computer Science Education, Reno, NV, United States, 49-53.
Retrieved September 4, 2006, from http://doi.acm.org/10.! 145/ 61 1892.61 1930
193
Boehm, B., Port, D., & Winsor Brown, A. (2002). Balancing plan-driven and agile
methods in software engineering project courses. Computer Science Education,
12(3), 187-195.
Booth, S. (2001). Learning computer science and engineering in context. Computer
Science Education, 11(3), 169-188.
Bouvier, D. J. (2003). Pilot study: Living flowcharts in an introduction to programming
course. SIGCSE '03: Proceedings of the 34th SIGCSE Technical Symposium on
Computer Science Education, Reno, NV, United States, 293-295. Retrieved
September 4, 2006 from http://doi.acm.Org/10.l 145/61 1892.61 1991
Bruce, K. B., Danyluk, A., & Murtagh, T. (2001). A library to support a graphics-based
object- first approach to CS 1. SIGCSE '01: Proceedings of the Thirty -Second
SIGCSE Technical Symposium on Computer Science Education, Charlotte, NC,
United States, 6-10. Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/
364447.364527
Bruhn, R. E., & Burton, P. J. (2003). An approach to teaching Java using computers.
SIGCSE Bulletin (Association for Computing Machinery, Special Interest Group on
Computer Science Education) , 35(4), 94-99.
Bruhn, R. E., & Camp, J. (2004). Capstone course creates useful business products and
corporate-ready students. SIGCSE Bulletin (Association for Computing Machinery,
Special Interest Group on Computer Science Education) , 36(2), 87-92.
Bryant, K. C. (2004). The evaluation of courses in information systems. Sixth
Australasian Computing Education Conference (ACE2004); Conferences in
Research and Practice in Information Technology, 30, 15-23.
Buck, D., & Stucki, D. J. (2001). JKarelRobot: A case study in supporting levels of
cognitive development in the computer science curriculum. SIGCSE '01:
Proceedings of the Thirty-Second SIGCSE Technical Symposium on Computer
Science Education, Charlotte, NC, United States, 16-20. Retrieved September 4,
2006, from http //doi.acm.org/10.1 145/364447.364529
Butler, J. E., & Brockman, J. B. (2001). A web-based learning tool that simulates a
simple computer architecture. SIGCSE Bulletin (Association for Computing
Machinery, Special Interest Group on Computer Science Education) , 33(2), 47-50.
194
Carr, S., Chen, P., Jozwowski, T. R., Mayo, J., & Shene, C. (2002). Channels,
visualization, and topology editor. ITiCSE '02: Proceedings of the 7th Annual
Conference on Innovation and Technology in Computer Science Education, Aarhus,
Denmark, 106-1 10. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/544414.544448
Carr, S., & Shene, C. (2000). A portable class library for teaching multithreaded
programming. ITiCSE '00: Proceedings of the 5th Annual SIGCSE/SIGCUE ITiCSE
conference on Innovation and Technology in Computer Science Education, Helsinki,
Finland, 124-127. Retrieved September 4, 2006, from http://doi.acm.Org/10.l 145/
343048.343138
Cavalcante, R., Finley, T., & Rodger, S. H. (2004). A visual and interactive automata
theory course with JFLAP 4.0. SIGCSE '04: Proceedings of the 35th SIGCSE
Technical Symposium on Computer Science Education, Norfolk, VA, United States,
140-144. Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/
971300.971349
Cernuda del Rio, A. (2004). How not to go about a programming assignment. SIGCSE
Bulletin (Association for Computing Machinery, Special Interest Group on
Computer Science Education), 36(2), 97-100.
Chalk, P. (2000). Webworlds — web-based modeling environments for learning software
engineering. Computer Science Education, I0(\), 39-56.
Chamillard, A. T., & Joiner, J. K. (2001). Using lab practica to evaluate programming
ability. SIGCSE '01: Proceedings of the Thirty-Second SIGCSE Technical
Symposium on Computer Science Education, Charlotte, NC, United States, 159-163.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/364447.364572
Chamillard, A. T., & Merkle, L. D. (2002). Management challenges in a large
introductory computer science course. SIGCSE '02: Proceedings of the 33rd SIGCSE
Technical Symposium on Computer Science Education, Cincinnati, OH, United
States 252-256. Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/
563340.563440
Chamillard, A. T., Moore, J. A., & Gibson, D. S. (2002, February) Using graphics in an
introductory computer science course. Journal of Computer Science Education
Online. Retrieved September 4, 2006, from http://www.iste.org/Template.cfm?
Section=February&Template=/MembersOnly.cfm&ContentID=4244
195
Chandra, S. (2003). Beacond: A peer-to-peer system to teach ubiquitous computing.
SIGCSE '03: Proceedings of the 34th SIGCSE Technical Symposium on Computer
Science Education, Reno, NV, United States, 257-261. Retrieved September 4, 2006,
from http://doi.acm.Org/10.l 145/61 1892.61 1983
Chen, S., & Morris, S. (2005). Iconic programming for flowcharts, Java, turing, etc.
ITiCSE '05: Proceedings of the 10th Annual SIGCSE Conference on Innovation and
Technology in Computer Science Education, Caparica, Portugal, 104-107. Retrieved
September 4, 2006 from http://doi.acm.0rg/lO.l 145/1067445.1067477
Christensen, H. B., & Caspersen, M. E. (2002). Frameworks in CS1: A different way of
introducing event-driven programming. ITiCSE '02: Proceedings of the 7th Annual
Conference on Innovation and Technology in Computer Science Education, Aarhus,
Denmark, 75-79. Retrieved September 4, 2006, from http://doi.acm.org/ 10.1 145/
544414.544438
Cigas, J. (2003). An introductory course in network administration. SIGCSE '03:
Proceedings of the 34th SIGCSE Technical Symposium on Computer Science
Education, Reno, NV, United States, 113-116. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/611892.611946
Cinneide, M. 6. & Tynan, R. (2004). A problem-based approach to teaching design
patterns. ITiCSE-WGR '04: Working Group Reports from ITiCSE on Innovation and
Technology in Computer Science Education, Leeds, United Kingdom, 80-82.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/1044550.1041663
Clancy, M., Titterton, N., Ryan, C, Slotta, J., & Linn, M. (2003). New roles for students,
instructors, and computers in a lab-based introductory programming course. SIGCSE
'03: Proceedings of the 34th SIGCSE Technical Symposium on Computer Science
Education, Reno, NV, United States, 132-136. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/611892.611951
Clark, M. (2000). Getting participation through discussion. SIGCSE '00: Proceedings of
the Thirty-First SIGCSE Technical Symposium on Computer Science Education,
Austin, TX, United States, 129-133. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/330908.331841
Clark, N. (2005). Evaluating student teams developing unique industry projects. Seventh
Australasian Computing Education Conference (ACE2005); Conferences in
Research and Practice in Information Technology, 42, 21-30.
196
Clark, N. (2004). Peer testing in software engineering projects. Sixth Australasian
Computing Education Conference (ACE2004); Conferences in Research and
Practice in Information Technology, 30, 41-48.
Claypool, K., & Claypool, M. (2005). Teaching software engineering through game
design. ITiCSE '05: Proceedings of the 10th Annual SIGCSE Conference on
Innovation and Technology in Computer Science Education, Caparica, Portugal,
123-127. Retrieved September 4, 2006, from http://doi.acm.Org/10.l 145/
1067445.1067482
Claypool, M., Finkel, D., & Wills, C. (2001). An open source laboratory for operating
systems projects. ITiCSE '01: Proceedings of the 6th Annual Conference on
Innovation and Technology in Computer Science Education, Canterbury, United
Kingdom, 145-148. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/377435.377669
Clayton, D., & Lynch, T. (2002). Ten years of strategies to increase participation of
women in computing programs: The central queensland university experience: 1999-
-2001. SIGCSE Bulletin (Association for Computing Machinery, Special Interest
Group on Computer Science Education), 34(2), 89-93.
Cohen, R. F., Fairley, A. V., Gerry, D., & Lima, G. R. (2005). Accessibility in
introductory computer science. SIGCSE '05: Proceedings of the 36th SIGCSE
Technical Symposium on Computer Science Education, St. Louis, MO, United
States, 17-21. Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/
1047344.1047367
Combs, W., Hawkins, R., Pore, T., Schechet, A., Wahls, T., & Ziantz, L. (2005). The
course scheduling problem as a source of student projects. SIGCSE '05: Proceedings
of the 36th SIGCSE Technical Symposium on Computer Science Education, St.
Louis, MO, United States, 81-85. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/1047344.1047385
Cooper, S., Dann, W., & Pausch, R. (2003). Using animated 3D graphics to prepare
novices for CS1. Computer Science Education, 73(1), 3-30.
Coppit, D., & Haddox-Schatz, J. M. (2005). Large team projects in software engineering
courses. SIGCSE '05: Proceedings of the 36th SIGCSE Technical Symposium on
Computer Science Education, St. Louis, MO, United States, 137-141. Retrieved
September 4, 2006, from http://doi.acm.org/10.! 145/1047344.1047400
197
Countermine, T., & Pfeiffer, P. (2000). Implementing an IT concentration in a CS
department: Content, rationale, and initial impact. SIGCSE '00: Proceedings of the
Thirty-First SIGCSE Technical Symposium on Computer Science Education, Austin,
TX, United States, 275-279. Retrieved September 4, 2006 from, http://doi.acm.org/
10.1145/330908.331869
Crouch, D. B., & Schwartzman, L. (2003). Computer science accreditation: The
advantages of being different. SIGCSE '03: Proceedings of the 34th SIGCSE
Technical Symposium on Computer Science Education, Reno, NV, United States,
36-40. Retrieved September 4, 2006, from http://doi.acm.Org/10.l 145/
611892.611927
Cunningham, S. (2002). Graphical problem solving and visual communication in the
beginning computer graphics course. SIGCSE '02: Proceedings of the 33rd SIGCSE
Technical Symposium on Computer Science Education, Cincinnati, OH, United
States 181-185. Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/
563340.563410
Czajkowski, M. F., Foster, C. V., Hewett, T. T., Casacio, J. A., Regli, W. C, & Sperber,
H. A. (2001). A student project in software evaluation. ITiCSE '01: Proceedings of
the 6th Annual Conference on Innovation and Technology in Computer Science
Education, Canterbury, United Kingdom, 13-16. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/377435.377446
Daly, C, & Horgan, J. (2005). Patterns of plagiarism. SIGCSE '05: Proceedings of the
36th SIGCSE Technical Symposium on Computer Science Education, St. Louis, MO,
United States, 383-387. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/1047344.1047473
Dann, W., Dragon, T., Cooper, S., Dietzler, K., Ryan, K., & Pausch, R. (2003). Objects:
Visualization of behavior and state. ITiCSE '03: Proceedings of the 8th Annual
Conference on Innovation and Technology in Computer Science Education,
Thessaloniki, Greece, 84-88. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/961511.961537
Dantin, U. (2005). Application of personas in user interface design for educational
software. Seventh Australasian Computing Education Conference (ACE2005);
Conferences in Research and Practice in Information Technology, 4,2 239-247.
D'Antonio, L. (2003). Incorporating bioinformatics in an algorithms course. ITiCSE '03:
Proceedings of the 8th Annual Conference on Innovation and Technology in
Computer Science Education, Thessaloniki, Greece, 211-214. Retrieved September
4, 2006, from http://doi.acm.org/10.1145/961511.961569
198
Davis, T., Geist, R., Matzko, S., & Westall, J. (2004). ?????: A first step. SIGCSE '04:
Proceedings of the 35th SIGCSE Technical Symposium on Computer Science
Education, Norfolk, VA, United States, 125-129. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/971300.971345
Debray, S. (2004). Writing efficient programs: Performance issues in an undergraduate
CS curriculum. SIGCSE '04: Proceedings of the 35th SIGCSE Technical Symposium
on Computer Science Education, Norfolk, VA, United States, 275-279. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/971300.971396
Debray, S. (2002). Making compiler design relevant for students who will (most likely)
never design a compiler. SIGCSE '02: Proceedings of the 33rd SIGCSE Technical
Symposium on Computer Science Education, Cincinnati, OH, United States 341-345.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/563340.563473
Demetriadis, S., Triantfillou, E., & Pombortsis, A. (2003). A phenomenographic study of
students' attitudes toward the use of multiple media for learning. ITiCSE '03:
Proceedings of the 8th Annual Conference on Innovation and Technology in
Computer Science Education, Thessaloniki, Greece, 183-187. Retrieved September
4, 2006 from http://doi.acm.0rg/lO.l 145/96151 1.961562
DePasquale, P., Lee, J. A. N., Perez-Quinones, M. A. (2004). Evaluation of subsetting
programming language elements in a novice's programming environment. SIGCSE
'04: Proceedings of the 35th SIGCSE Technical Symposium on Computer Science
Education, Norfolk, VA, United States, 260-264. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/971300.971392
deRaadt, M., Watson, R., & Toleman, M. (2003). Language tug-of-war: Industry demand
and academic choice. Fifth Australasian Computing Education Conference
(ACE2003); Conferences in Research and Practice in Information Technology, , 20
137-142.
Dewan, P. (2005). Teaching inter-object design patterns to freshmen. SIGCSE '05:
Proceedings of the 36th SIGCSE Technical Symposium on Computer Science
Education, St. Louis, MO, United States, 482-486. Retrieved September 4, 2006,
from http://doi.acm.Org/10.l 145/1047344.1047498
Dick, M. (2005). Student interviews as a tool for assessment and learning in a systems
analysis and design course. ITiCSE '05: Proceedings of the 10th Annual SIGCSE
Conference on Innovation and Technology in Computer Science Education,
Caparica, Portugal, 24-28. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/1067445.1067456
199
Dick, M., Sheard, J., & Markham, S. (2001). Is it okay to cheat? - the views of
postgraduate students. ITiCSE '01: Proceedings of the 6th Annual Conference on
Innovation and Technology in Computer Science Education, Canterbury, United
Kingdom, 61-64. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/377435.377474
Dierbach, C, Taylor, B., Zhou, H., & Zimand, I. (2005). Experiences with a CS0 course
targeted for CS1 success. SIGCSE '05: Proceedings of the 36th SIGCSE Technical
Symposium on Computer Science Education, St. Louis, MO, United States, 317-320.
Retrieved September 4, 2006, from http://doi.acm.Org/10.l 145/1047344.1047453
Donaldson, J. L. (2001). Architecture-dependent operating system project sequence.
SIGCSE '01: Proceedings of the Thirty-Second SIGCSE Technical Symposium on
Computer Science Education, Charlotte, NC, United States, 322-326. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/364447.364613
Drummond, S., Boldyreff, C, & Ramage, M. (2001). Evaluating groupware support for
software engineering students. Computer Science Education, 11(1), 33-54.
Drury, H., Kay, J., & Losberg, W. (2003). Student satisfaction with groupwork in
undergraduate computer science : Do things get better? Fifth Australasian
Computing Education Conference (ACE2003); Conferences in Research and
Practice in Information Technology, 20, 77-85.
Edwards, S. H. (2004). Using software testing to move students from trial-and-error to
reflection-in-action. SIGCSE '04: Proceedings of the 35th SIGCSE Technical
Symposium on Computer Science Education, Norfolk, VA, United States, 26-30.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/971300.971312
Efopoulos, V., Dagdilelis, V., Evangelidis, G., & Satratzemi, M. (2005). WIPE: A
programming environment for novices. ITiCSE '05: Proceedings of the 10th Annual
SIGCSE Conference on Innovation and Technology in Computer Science Education,
Caparica, Portugal, 113-1 17. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/1067445.1067479
Egea, K. (2003). Managing the managers: Collaborative virtual teams with large staff and
student numbers. Fifth Australasian Computing Education Conference (ACE2003);
Conferences in Research and Practice in Information Technology, 20, 87-94.
Eisenberg, M. (2003). Creating a computer science canon: A course of "classic" readings
in computer science. SIGCSE '03: Proceedings of the 34th SIGCSE Technical
Symposium on Computer Science Education, Reno, NV, United States, 336-340.
Retrieved September 4, 2006, from http://doi.acm.org/10.! 145/61 1892.612002
200
Elsharnouby, T., & Shankar, A. U. (2005). Using SeSFJava in teaching introductory
network courses. SIGCSE '05: Proceedings of the 36th SIGCSE Technical
Symposium on Computer Science Education, St. Louis, MO, United States, 67-71.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/1047344.1047381
Ericson, B., Guzdial, M., & Biggers, M. (2005). A model for improving secondary CS
education. SIGCSE '05: Proceedings of the 36th SIGCSE Technical Symposium on
Computer Science Education, St. Louis, MO, United States, 332-336. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/1047344.1047460
Faux, R. (2003, April). The role of observation in computer science learning. Journal of
Computer Science Education Online. Retrieved September 4, 2006 from
http://www.iste. org/Template.cfm?Section=April&Template=/MembersOnly.cfm&C
ontentID=4223
Fekete, A. (2003). Using counter-examples in the data structures course. Fifth
Australasian Computing Education Conference (ACE2003); Conferences in
Research and Practice in Information Technology, 20, 179-186.
Felleisen, M., Findler, R., Flatt, M., & Krishnamurthi, S. (2004). The TeachScheme!
Project: Computing and programming for every student. Computer Science
Education, I4(\), 55-77.
Fenwick, J. B., Norris, C, & Wilkes, J. (2002). Scientific experimentation via the
matching game. SIGCSE '02: Proceedings of the 33rd SIGCSE Technical
Symposium on Computer Science Education, Cincinnati, OH, United States 326-330.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/563340.563469
Fernandes, E., & Kumar, A. N. (2004). A tutor on scope for the programming languages
course. SIGCSE '04: Proceedings of the 35th SIGCSE Technical Symposium on
Computer Science Education, Norfolk, VA, United States, 90-93. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/971300.971332
Filho, W. P. P. (2001). Requirements for an educational software development process.
ITiCSE '01: Proceedings of the 6th Annual Conference on Innovation and
Technology in Computer Science Education, Canterbury, United Kingdom, 65-68.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/377435.377476
201
Fincher, S., Petre, M., Tenenberg, J., Blaha, K., Bouvier, D., Chen, T-Z., et al. (2004). A
multi-national, multi-institutional study of student-generated software designs. Kolin
Kolistelut - Koli Calling 2004, Proceedings of the Fourth Annual Finnish Annual
Finnish Baltic Sea Conference on Computer Science Education, Koli, Finland, 11-
19. Retrieved September 4, 2006, from http://cs.joensuu.fi/kolistelut/archive/2004/
koli_proc_2004.pdf
Fisher, J., Lowther, J., & Shene, C. (2004). If you know b-splines well, you also know
NURBS! SIGCSE '04: Proceedings of the 35th SIGCSE Technical Symposium on
Computer Science Education, Norfolk, VA, United States, 343-347. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/971300.971420
Fleury, A. E. (2001). Encapsualtion and reuse as viewed by Java students. SIGCSE '01:
Proceedings of the Thirty-Second SIGCSE Technical Symposium on Computer
Science Education, Charlotte, NC, United States, 189-193. Retrieved September 4,
2006, from http://doi.acm.org/10.1145/364447.364582
Francioni, J. M., & Smith, A. C. (2002). Computer science accessibility for students with
visual disabilities. SIGCSE '02: Proceedings of the 33rd SIGCSE Technical
Symposium on Computer Science Education, Cincinnati, OH, United States 91-95.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/563340.563372
Frens, J. D. (2004). Taming the tiger: Teaching the next version of Java. SIGCSE '04:
Proceedings of the 35th SIGCSE Technical Symposium on Computer Science
Education, Norfolk, VA, United States, 151-155. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/971300.971356
Frieze, C, & Blum, L. (2002). Building an effective computer science student
organization: The Carnegie Mellon Women@SCS action plan. SIGCSE Bulletin
(Association for Computing Machinery, Special Interest Group on Computer Science
Education), 34(2), 74-78.
Gal-Ezer, J., Vilner, T., & Zur, E. (2004). Teaching algorithm efficiency at CS1 level: A
different approach. Computer Science Education, 14(3), 235-248.
Gal-Ezer, J., & Zeldes, A. (2000). Teaching software designing skills. Computer Science
Education, 10(\), 25-38.
Gallagher, J. C, & Perretta, S. (2002). WWW autonomous robotics: Enabling wide area
access to a computer engineering practicum. SIGCSE '02: Proceedings of the 33rd
SIGCSE Technical Symposium on Computer Science Education, Cincinnati, OH,
United States 13-17. Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/
563340.563346
202
Galpin, V. (2002). Women in computing around the world. SIGCSE Bulletin (Association
for Computing Machinery, Special Interest Group on Computer Science Education),
34(2), 94-100.
Gaona, A. L., (2000). The relevance of design in CS1. SIGCSE Bulletin (Association for
Computing Machinery, Special Interest Group on Computer Science Education),
32(2), 53-55.
Garner, S., Haden, P., & Robins, A. (2005). My program is correct but it doesn't run: A
preliminary investigation of novice programmers' problems. Seventh Australasian
Computing Education Conference (ACE2005); Conferences in Research and
Practice in Information Technology, 42, 173-180.
Gegg-Harrison, T. S. (2001). Ancient egyptian numbers: A CS-complete example.
SIGCSE '01: Proceedings of the Thirty-Second SIGCSE Technical Symposium on
Computer Science Education, Charlotte, NC, United States, 268-272. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/364447.364598
George, C. E. (2002). Using visualization to aid program construction tasks. SIGCSE '02:
Proceedings of the 33rd SIGCSE Technical Symposium on Computer Science
Education, Cincinnati, OH, United States 191-195. Retrieved September 4, 2006,
from http://doi.acm.Org/10.l 145/563340.563413
Gerdt, P. (2001). Applying computer supported collaborative writing in education. Kolin
Kolistelut - Koli Calling 2001, Proceedings of the First Annual Finnish Annual
Finnish Baltic Sea Conference on Computer Science Education, Koli, Finland, 38-
45. Retrieved September 4, 2006, from http://cs.joensuu.fi/kolistelut/archive/
2001/koli_proc_2001.pdf
Gerhardt-Powals, J., & Powals, M. H. (2000). Distance education: Law attempts to catch
up with technology (battle between copyright owners and academics). SIGCSE '00:
Proceedings of the Thirty-First SIGCSE Technical Symposium on Computer Science
Education, Austin, TX, United States, 335-342. Retrieved September 4, 2006, from
http://d0i.acm.0rg/l 0. 1 145/330908.33 1881
Gibson, J. P., & O'Kelly, J. (2005). Software engineering as a model of understanding for
learning and problem solving. ICER '05: Proceedings of the 2005 International
Workshop on Computing Education Research, Seattle, WA, United States, 87-97.
Retrieved September 4, 2006, from http://doi.acm.org/10.! 145/1089786.1089795
203
Ginat, D. (2004). Embedding instructive assertions in program design. ITiCSE '04:
Proceedings of the 9th Annual SIGCSE Conference on Innovation and Technology in
Computer Science Education, Leeds, United Kingdom, 62-66. Retrieved September
4, 2006, from http://doi.acm.org/10.1145/1007996.1008015
Ginat, D. (2003). The greedy trap and learning from mistakes. SIGCSE '03: Proceedings
of the 34th SIGCSE Technical Symposium on Computer Science Education, Reno,
NV, United States, 11-15. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/611892.611920
Ginat, D. (2002). On varying perspectives of problem decomposition. SIGCSE '02:
Proceedings of the 33rd SIGCSE Technical Symposium on Computer Science
Education, Cincinnati, OH, United States 331-335. Retrieved September 4, 2006,
from http://doi.acm.Org/10.l 145/563340.563470
Goldweber, M., Davoli, R., & Morsiani, M. (2005). The Kaya OS project and the //MPS
hardware emulator. ITiCSE '05: Proceedings of the 10th Annual SIGCSE Conference
on Innovation and Technology in Computer Science Education, Caparica, Portugal,
49-53. Retrieved September 4, 2006, from http://doi.acm.Org/10.l 145/
1067445.1067462
Golub, E. (2004). Handwritten slides on a tabletPC in a discrete mathematics course.
SIGCSE '04: Proceedings of the 35th SIGCSE Technical Symposium on Computer
Science Education, Norfolk, VA, United States, 51-55. Retrieved September 4, 2006
from http://doi.acm.org/10.1145/971300.971322
Goold, A., & Coldwell, J. (2005). Teaching ethics in a virtual classroom. ITiCSE '05:
Proceedings of the 10th Annual SIGCSE Conference on Innovation and Technology
in Computer Science Education, Caparica, Portugal, 232-236. Retrieved September
4, 2006, from http://doi.acm.org/10.1145/1067445.1067509
Gotterbarn, D., & Clear, T. (2004). Using SoDIS as a risk analysis process: A teaching
perspective. Sixth Australasian Computing Education Conference (ACE2004);
Conferences in Research and Practice in Information Technology, 30, 83-90.
Grandell, L., Peltomaki, M., & Salakoski, T. (2005). High school programming - A
beyond-syntax analysis of novice programmers' difficulties. Kolin Kolistelut — Koli
Calling 2005, Proceedings of the Fourth Annual Finnish Annual Finnish Baltic Sea
Conference on Computer Science Education, Koli, Finland, 17-24. Retrieved
September 4, 2006, from http://cs.joensuu.fi/kolistelut/archive/2005/
koli_proc_2005 .pdf
204
Grinder, M. T., Kim, S. B., Lutey, T. L., Ross, R. J., & Walsh, K. F. (2002). Loving to
learn theory: Active learning modules for the theory of computing. SIGCSE '02:
Proceedings of the 33rd SIGCSE Technical Symposium on Computer Science
Education, Cincinnati, OH, United States 371-375. Retrieved September 4, 2006,
from http://doi.acm.Org/10.l 145/563340.563488
Gruba, P., & Al-Mahmood, R. (2004). Strategies for communication skills development.
Sixth Australasian Computing Education Conference (ACE2004); Conferences in
Research and Practice in Information Technology, 30, 101-107.
Gruba, P., Moffat, A., Sondergaard, H., & Zobel, J. (2004). What drives curriculum
change? Sixth Australasian Computing Education Conference (ACE2004);
Conferences in Research and Practice in Information Technology, 30 109-1 17.
Gupta, G. K. (2001). Information technology and liberal arts. SIGCSE Bulletin
(Association for Computing Machinery, Special Interest Group on Computer Science
Education), 33(2), 55-57.
Guzdial, M. (2001). Use of collaborative multimedia in computer science classes. ITiCSE
'01: Proceedings of the 6th Annual Conference on Innovation and Technology in
Computer Science Education, Canterbury, United Kingdom, 17-20. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/377435.377452
Haberman, B., & Averbuch, H. (2002). The case of base cases: Why are they so difficult
to recognize? student difficulties with recursion. ITiCSE '02: Proceedings of the 7 th
Annual Conference on Innovation and Technology in Computer Science Education,
Aarhus, Denmark, 84-88. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/544414.544441
Hadjerrouit, S. (2001). Web-based application development: A software engineering
approach. SIGCSE Bulletin (Association for Computing Machinery, Special Interest
Group on Computer Science Education), 33(2), 31-34.
Hamelin, D. (2004). Searching the web to develop inquiry and collaborative skills.
ITiCSE-WGR '04: Working Group Reports from ITiCSE on Innovation and
Technology in Computer Science Education, Leeds, United Kingdom, 76-79.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/1044550.1041662
Hamer, J. (2004). An approach to teaching design patterns using musical composition.
ITiCSE '04: Proceedings of the 9th Annual SIGCSE Conference on Innovation and
Technology in Computer Science Education, Leeds, United Kingdom, 156-160.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/1007996.1008038
205
Harney, L. G. C. (2003). Teaching secure data communications using a game
representation. Fifth Australasian Computing Education Conference (ACE2003);
Conferences in Research and Practice in Information Technology, 20, 187-196.
Hanks, B., McDowell, C, Draper, D., & Krnjajic, M. (2004). Program quality with pair
programming in CS 1 . ITiCSE '04: Proceedings of the 9th Annual SIGCSE
Conference on Innovation and Technology in Computer Science Education, Leeds,
United Kingdom. 176-180. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/1007996.1008043
Hannon, C, Huber, M., & Burnell, L. (2005). Research to classroom: Experiences from a
multi-institutional course in smart home technologies. SIGCSE '05: Proceedings of
the 36th SIGCSE Technical Symposium on Computer Science Education, St. Louis,
MO, United States, 121-125. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/1047344.1047395
Hansen, K. M., & Ratzer, A. V. (2002). Tool support for collaborative teaching and
learning of object-oriented modeling. ITiCSE '02: Proceedings of the 7th Annual
Conference on Innovation and Technology in Computer Science Education, Aarhus,
Denmark, 146- 150. Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/
544414.544458
Harrison, S. M. (2005). Opening the eyes of those who can see to the world of those who
can't: A case study. SIGCSE '05: Proceedings of the 36th SIGCSE Technical
Symposium on Computer Science Education, St. Louis, MO, United States, 22-26.
Retrieved September 4, 2006 from http://doi.acm.Org/10.l 145/1047344.1047368
Hartley, S. J. (2000). VAlfonse, you have a messaged". SIGCSE '00: Proceedings of the
Thirty-First SIGCSE Technical Symposium on Computer Science Education, Austin,
TX, United States, 60-64. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/330908.331813
Hazzan, O. (2005). Professional development workshop for female software engineers.
SIGCSE Bulletin (Association for Computing Machinery, Special Interest Group on
Computer Science Education) , 37(2), 75-79.
Hazzan, O. (2003). Application of computer science ideas to the presentation of
mathematical theorems and proofs. SIGCSE Bulletin (Association for Computing
Machinery, Special Interest Group on Computer Science Education), 35(2), 38-42.
Hazzan, O. (2001). On the presentation of computer science problems. SIGCSE Bulletin
(Association for Computing Machinery, Special Interest Group on Computer Science
Education), 33(4), 55-58.
206
Herrmann, N., Popyack, J. L., Char, B., Zoski, P., Cera, C. D., & Lass, R. N., et al.
(2003). Redesigning introductory computer programming using multi-level online
modules for a mixed audience. SIGCSE '03: Proceedings of the 34th SIGCSE
Technical Symposium on Computer Science Education, Reno, NV, United States,
196-200. Retrieved September 4, 2006, from http://doi.acm.Org/10.l 145/
611892.611967
Hilburn, T. B., & Townhidnejad, M. (2000). Software quality: A curriculum postscript?
SIGCSE '00: Proceedings of the Thirty-First SIGCSE Technical Symposium on
Computer Science Education, Austin, TX, United States. 167 '-171, Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/330908.331848
Hislop, G., Lutz, M., Fernando Naveda, J., Michael McCracken, W., Mead, N., &
Williams, L. (2002). Integrating agile practices into software engineering courses.
Computer Science Education, 12(3), 169-185.
Hoffman, M. E., & Vance, D. R. (2005). Computer literacy: What students know and
from whom they learned it. SIGCSE '05: Proceedings of the 36th SIGCSE Technical
Symposium on Computer Science Education, St. Louis, MO, United States, 356-360.
Retrieved September 4, 2006m from http://doi.acm.Org/10.l 145/1047344.1047467
Hogan, J. M., Smith, G., & Thomas, R. C. (2005). Tight spirals and industry clients: The
modern SE education experience. Seventh Australasian Computing Education
Conference (ACE2005); Conferences in Research and Practice in Information
Technology, 42, 217-222.
Holland, D. A., Lim, A. T., & Seltzer, M. I. (2002). A new instructional operating system.
SIGCSE '02: Proceedings of the 33rd SIGCSE Technical Symposium on Computer
Science Education, Cincinnati, OH, United States 111-115. Retrieved September 4,
2006, from http://doi.acm.org/10.1145/563340.563383
Holmboe, C. (2005). The linguistics of object-oriented design: Implications for teaching.
ITiCSE '05: Proceedings of the 10th Annual SIGCSE Conference on Innovation and
Technology in Computer Science Education, Caparica, Portugal, 188-192. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/1067445.1067498
Hood, C. S., & Hood, D. J. (2005). Teaching programming and language concepts using
LEGOs; ITiCSE '05: Proceedings of the 10th Annual SIGCSE Conference on
Innovation and Technology in Computer Science Education, Caparica, Portugal, 19-
23. Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/
1067445.1067454
207
Howe, E., Thornton, M., & Weide, B. W. (2004). Components-first approaches to
CS1/CS2: Principles and practice. SIGCSE '04: Proceedings of the 35th SIGCSE
Technical Symposium on Computer Science Education, Norfolk, VA, United States,
291-295. Retrieved September 4, 2006, from http://doi.acm.Org/10.l 145/
971300.971404
Hsia, J. I., Simpson, E., Smith, D., & Cartwright, R. (2005). Taming Java for the
classroom. SIGCSE '05: Proceedings of the 36th SIGCSE Technical Symposium on
Computer Science Education, St. Louis, MO, United States, 327-331. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/1047344.1047459
Hu, C. (2003). A framework for applet animations with controls. SIGCSE Bulletin
(Association for Computing Machinery, Special Interest Group on Computer Science
Education), 35(4), 90-93.
Hunt, K. (2003). Using image processing to teach CS1 and CS2. SIGCSE Bulletin
(Association for Computing Machinery, Special Interest Group on Computer Science
Education), 35(4), 86-89.
Huth, M. (2004). Mathematics for the exploration of requirements. SIGCSE Bulletin
(Association for Computing Machinery, Special Interest Group on Computer Science
Education), 36(2), 34-39.
Hamalaien, W. (2003). Problem-based learning of theoretical computer science. Kolin
Kolistelut - Koli Calling 2003, Proceedings of the Third Annual Finnish Annual
Finnish Baltic Sea Conference on Computer Science Education, Koli, Finland, 50-
58. Retrieved September 4, 2006, from http://cs.joensuu.fi/kolistelut/archive/ 2003/
kol i_proc_2003 .pdf
Hamalainen, W., Myller, N., Pitkanen, S. H., & Lopez-Cuadrado, J. (2002). Learning by
gaming in apaptive learning system. Kolin Kolistelut - Koli Calling 2002,
Proceedings of the Second Annual Finnish Annual Finnish Baltic Sea Conference on
Computer Science Education, Koli, Finland, 22-26. Retrieved September 4, 2006,
from http://cs.joensuu.fi/kolistelut/archive/2002/koli_proc_2002.pdf
Jacobsen, C. L., & Jadud, M. C. (2005). Towards concrete concurrency: Occam-pi on the
LEGO mindstorms. SIGCSE '05: Proceedings of the 36th SIGCSE Technical
Symposium on Computer Science Education, St. Louis, MO, United States, 431-435.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/1047344.1047485
Jacobson, N. (2000). Using on-computer exams to ensure beginning students'
programming competency. SIGCSE Bulletin (Association for Computing Machinery,
Special Interest Group on Computer Science Education) , 32(4), 53-56.
208
Jacobson, N., & Thornton, A. (2004). It is time to emphasize arraylists over arrays in
java-based first programming courses. ITiCSE-WGR '04: Working Group Reports
from ITiCSE on Innovation and Technology in Computer Science Education, Leeds,
United Kingdom, 88-92. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/1044550.1041666
Jadud, M. (2005). A first look at novice compilation behaviour using BlueJ. Computer
Science Education, 75(1), 25-40.
Jalloul, G. (2000). Links: A framework for object-oriented software engineering.
Computer Science Education, 10(\), 75-93.
James, R. H. (2005). External sponsored projects: Lessons learned. SIGCSE Bulletin
(Association for Computing Machinery, Special Interest Group on Computer Science
Education), 37(2), 94-98.
Jenkins, T. (2001). The motivation of students of programming. ITiCSE '01: Proceedings
of the 6th Annual Conference on Innovation and Technology in Computer Science
Education, Canterbury, United Kingdom, 53-56. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/377435.377472
Jepson, A., & Perl, T. (2002). Priming the pipeline. SIGCSE Bulletin (Association for
Computing Machinery, Special Interest Group on Computer Science Education),
34(2), 36-39.
Jimenez-Diaz, G., Gomez-Albarran, M., et al. (2005). Software behaviour understanding
supported by dynamic visualization and role-play. ITiCSE '05: Proceedings of the
10th Annual SIGCSE Conference on Innovation and Technology in Computer
Science Education, Caparica, Portugal, 54-58. Retrieved September 4, 2006,
Retrieved September 4, 2006 from http://doi.acm.Org/10.l 145/1067445.1067464
Johnson, D., & Caristi, J. (2002). Using extreme programming in the software design
course. Computer Science Education, 12(3), 223-234.
Jones, E. L. (2001). Integrating testing into the curriculum \ — arsenic in small doses.
SIGCSE '01: Proceedings of the Thirty-Second SIGCSE Technical Symposium on
Computer Science Education, Charlotte, NC, United States, 337-341. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/364447.364617
Jones, E. L., & Allen, C. S. (2003). Repositories for CS courses: An evolutionary tale.
ITiCSE '03: Proceedings of the 8th Annual Conference on Innovation and
Technology in Computer Science Education, Thessaloniki, Greece, 119-123.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/96151 1.961546
209
Jones, M. (2005). The pedagogic opportunities of touch-screen voting. ITiCSE '05:
Proceedings of the 10th Annual SIGCSE Conference on Innovation and Technology
in Computer Science Education, Caparica, Portugal, 223-226. Retrieved September
4, 2006, from http://doi.acm.org/10.1145/1067445.1067507
Jones, R. M. (2000). Design and implementation of computer games: A capstone course
for undergraduate computer science education. SIGCSE '00: Proceedings of the
Thirty-First SIGCSE Technical Symposium on Computer Science Education, Austin,
TX, United States. 260-264, Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/330908.331866
Jormanainen, I. (2004). A visual approach for concretizing sorting algorithms. Kolin
Kolistelut - Koli Calling 2004, Proceedings of the Fourth Annual Finnish Annual
Finnish Baltic Sea Conference on Computer Science Education, Koli, Finland, 141-
145. Retrieved September 4, 2006, from http://cs.joensuu.fi/kolistelut/ archive/
2004/koli_proc_2004.pdf
Jovanovic-Dolecek, G., & Champac, V. H. (2000). CGTDEMO \ — educational software
for the central limit theorem. SIGCSE Bulletin (Association for Computing
Machinery, Special Interest Group on Computer Science Education) , 32(2), 46-48.
Joyce, D., & Young, A. L. (2004). Developing and implementing a professional doctorate
in computing. Sixth Australasian Computing Education Conference (ACE2004);
Conferences in Research and Practice in Information Technology, 30, 145-149.
Kamppuri, M., Tedre, M. & Tukianen, M. (2005). A cultural approach to interface
design. Kolin Kolistelut - Koli Calling 2005, Proceedings of the Fourth Annual
Finnish Annual Finnish Baltic Sea Conference on Computer Science Education,
Koli, Finland, 149-152. Retrieved September 4, 2006, from http://cs.joensuu.fi/
kolistelut/ archive/ 2005/koli_proc_2005.pdf
Karavita, V., Korhonen, A., Nikander, J., & Tenhunen, P. (2002). Effortless creation of
algorithm visualization. Kolin Kolistelut - Koli Calling 2002, Proceedings of the
Second Annual Finnish Annual Finnish Baltic Sea Conference on Computer Science
Education, Koli, Finland, 52-56. Retrieved September 4, 2006, from
http://cs.joensuu.fi/kolistelut/archive/2002/koli_proc_2002.pdf
Keefe, K, & Dick, M. (2004). Using extreme programming in a capstone project. Sixth
Australasian Computing Education Conference (ACE2004); Conferences in
Research and Practice in Information Technology, 30, 151-160.
210
Kellerer, W., Autenrieth, A., & Iselt, A. (2000). Experiences with evaluation of SDL-
based protocol engineering in education. Computer Science Education, 10(3), 225-
241.
Kenny, C, & Pahl, C. (2005). Automated tutoring for a database skills training
environment. SIGCSE '05: Proceedings of the 36th SIGCSE Technical Symposium
on Computer Science Education, St. Louis, MO, United States. 58-62, Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/1047344.1047377
Khuri, S., & Holzapfel, K. (2001). EVEGA: An educational visualization environment
for graph algorithms. ITiCSE '01: Proceedings of the 6th Annual Conference on
Innovation and Technology in Computer Science Education, Canterbury, United
Kingdom, 101-104. Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/
377435.377497
Khuri, S., & Hsu, H. (2000). Interactive packages for learning image compression
algorithms. ITiCSE '00: Proceedings of the 5th Annual SIGCSE/SIGCUE
ITiCSEconference on Innovation and Technology in Computer Science Education,
Helsinki, Finland, 73-76. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/343048.343081
Klassner, F. (2004). Enhancing lisp instruction with RCXLisp and robotics. SIGCSE '04:
Proceedings of the 35th SIGCSE Technical Symposium on Computer Science
Education, Norfolk, VA, United States, 214-218. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/971300.971377
Koldehofe, B., & Tsigas, P. (2001). Using actors in an interactive animation in a graduate
course on distributed system. ITiCSE '01: Proceedings of the 6th Annual Conference
on Innovation and Technology in Computer Science Education, Canterbury, United
Kingdom, 149-152. Retrieved September 4, 2006, from http://doi.acm.org/10. 1145/
377435.377670
Kolikant, Y. B. (2005). Students' alternative standards for correctness. ICER '05:
Proceedings of the 2005 International Workshop on Computing Education Research,
Seattle, WA, United States, 37-43. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/1089786.1089790
Kolikant, Y. B., Ben-Ari, M., & Pollack, S. (2000). The anthropology semaphores.
ITiCSE '00: Proceedings of the 5th Annual SIGCSE/SIGCUE ITiCSE conference on
Innovation and Technology in Computer Science Education, Helsinki, Finland, 21-
24. Retrieved September 4, 2006, from http://doi.acm.org/10.! 145/343048.343061
211
Kolikant, Y. B., & Pollack, S. (2004). Community-oriented pedagogy for in-service CS
teacher training. ITiCSE '04: Proceedings of the 9th Annual SIGCSE Conference on
Innovation and Technology in Computer Science Education, Leeds, United
Kingdom, 191-195. Retrieved September 4, 2006, from http://doi.acm.Org/10.l 145/
1007996.1008047
Kolikant, Y., & Pollack, S. (2004). Establishing computer science professional norms
among high-school students. Computer Science Education, 14{\), 21-35.
Korhonen, A., & Malmi, L. (2000). Algorithm simulation with automatic assessment.
ITiCSE '00: Proceedings of the 5th Annual SIGCSE/SIGCUE ITiCSEconference on
Innovation and Technology in Computer Science Education, Helsinki, Finland, 160-
163. Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/343048.343157
Korhonen, A., Malmi, L., Myller, N., Scheinin, P.. (2002). Does it make a difference if
students exercise on the web or in the classroom? ITiCSE '02: Proceedings of the 7th
Annual Conference on Innovation and Technology in Computer Science Education,
Aarhus, Denmark, 121-124. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/544414.544452
Kuittinen, M., & Sajaniemi, J. (2004). Teaching roles of variables in elementary
programming courses. ITiCSE '04: Proceedings of the 9th Annual SIGCSE
Conference on Innovation and Technology in Computer Science Education, Leeds,
United Kingdom, 57-61. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/1007996.1008014
Kumar, A. (2000). Dynamically generating problems on static scope. ITiCSE '00:
Proceedings of the 5th Annual SIGCSE/SIGCUE ITiCSE conference on Innovation
and Technology in Computer Science Education, Helsinki, Finland, 9-12. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/343048.343055
Laakso, M-J., Salakoski, T., Korhonen, A. & Malmi, L. (2004). Automatic assessment of
exercises for algorithms and data structures — a case study with TRAKLA2. Kolin
Kolistelut - Koli Calling 2004, Proceedings of the Fourth Annual Finnish Annual
Finnish Baltic Sea Conference on Computer Science Education, Koli, Finland, 28-
36. Retrieved September 4, 2006, from http://cs.joensuu.fi/kolistelut/
archive/2004/koli_proc_2004.pdf
Lai, Y. (2005). Teaching computer applications to pre-school teachers through problem
based learning approach. SIGCSE Bulletin (Association for Computing Machinery,
Special Interest Group on Computer Science Education) , 37(4), 89-92.
212
Laine, H. (2001). SQO-Trainer. Kolin Kolistelut - Koli Calling 2001, Proceedings of the
First Annual Finnish Annual Finnish Baltic Sea Conference on Computer Science
Education, Koli, Finland, 13-19. Retrieved September 4, 2006, from
http://cs.joensuu.fi/ kolistelut/archive/200 l/koli_proc_200 1 .pdf
Lane, H. C, & VanLehn, K. (2003). Coached program planning: Dialogue-based support
for novice program design. SIGCSE '03: Proceedings of the 34th SIGCSE Technical
Symposium on Computer Science Education, Reno, NV, United States, 148-152.
Retrieved September 4, 2006, from http://doi.acm.Org/10.l 145/61 1892.61 1955
Lane, H. C, & VanLehn, K. (2005). Intention-based scoring: An approach to measuring
success at solving the composition problem. SIGCSE '05: Proceedings of the 36th
SIGCSE Technical Symposium on Computer Science Education, St. Louis, MO,
United States, 373-377. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/1047344.1047471
Lane, H. C, & VanLehn, K. (2005). Teaching the tacit knowledge of programming to
noviceswith natural language tutoring. Computer Science Education, 15(3), 183-
201.
Lapidot, T., & Hazzan, O. (2005). Song debugging: Merging content and pedagogy in
computer science education. SIGCSE Bulletin (Association for Computing
Machinery, Special Interest Group on Computer Science Education) , 37(4), 79-83.
Lapidot, T., & Hazzan, O. (2003). Methods of teaching a computer science course for
prospective teachers. SIGCSE Bulletin (Association for Computing Machinery,
Special Interest Group on Computer Science Education), 35(4), 29-34.
La Russa, G. (2001). Concretising tools for computer science education. Kolin Kolistelut
- Koli Calling 2001, Proceedings of the First Annual Finnish Annual Finnish Baltic
Sea Conference on Computer Science Education, Koli, Finland, 4-8. Retrieved
September 4, 2006, from http://cs.joensuu.fi/kolistelut/archive/2001/
kol i_proc_200 1 .pdf
La Russa, G., Silander, P., & Rytkonen, A. (2004). Utilizing pedagogical models in web-
based CSE education. Kolin Kolistelut - Koli Calling 2004, Proceedings of the
Fourth Annual Finnish Annual Finnish Baltic Sea Conference on Computer Science
Education, Koli, Finland, 107-1 1 1. Retrieved September 4, 2006, from
http://cs.joensuu.fi/kolistelut/archive/2004/koli_proc_2004.pdf
213
Lass, R. N., Cera, C. D., Bomberger, N. T., Char, B., Popyack, J. L., & Herrmann, N., et
al. (2003). Tools and techniques for large scale grading using web-based commercial
off-the-shelf software. ITiCSE '03: Proceedings of the 8th Annual Conference on
Innovation and Technology in Computer Science Education, Thessaloniki, Greece,
168-172. Retrieved September 4, 2006, from http://doi.acm.Org/10.l 145/
961511.961558
Last, M. Z., Daniels, M., Almstrum, V. L., Erickson, C, & Klein, B. (2000). An
international student/faculty collaboration: The runestone project. ITiCSE '00:
Proceedings of the 5th Annual SIGCSE/SIGCUE ITiCSE conference on Innovation
and Technology in Computer Science Education, Helsinki, Finland, 128-131.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/343048.343140
Latu, S., & Young, A. L. (2004). Teaching ICT to pacific island background students.
Sixth Australasian Computing Education Conference (ACE2004); Conferences in
Research and Practice in Information Technology, 30, 169-175.
Learmonth, R. (2001). Flexible delivery of information systems as a core MBA subject.
ITiCSE '01: Proceedings of the 6th Annual Conference on Innovation and
Technology in Computer Science Education, Canterbury, United Kingdom, 29-32.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/377435.377459
Lee, J. A. N. (2002). Internationalization of the curriculum report of a project within
computer science. SIGCSE '02: Proceedings of the 33rd SIGCSE Technical
Symposium on Computer Science Education, Cincinnati, OH, United States 68-72.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/563340.563366
Lee, J. A. N. (2001). Teaching and learning in the 21st century: The development of
"future CS faculty". SIGCSE Bulletin (Association for Computing Machinery,
Special Interest Group on Computer Science Education) , 33(2), 25-30.
Leventhal, L., & Barnes, J. (2003). Two for one: Squeezing human-computer interaction
and software engineering into a core computer science course. Computer Science
Education, 13(3), 177-190.
Leventhal, L. M., Barnes, J., & Chao, J. (2004). Term project user interface specifications
in a United States usability engineering course: Challenges and suggestions. SIGCSE
'04: Proceedings of the 35th SIGCSE Technical Symposium on Computer Science
Education, Norfolk, VA, United States, 41-45. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/971300.971316
214
Levitin, A. (2005). Analyze that: Puzzles and analysis of algorithms. SIGCSE '05:
Proceedings of the 36th SIGCSE Technical Symposium on Computer Science
Education, St. Louis, MO, United States, 171-175. Retrieved September 4, 2006,
from http://doi.acm.Org/10.l 145/1047344.1047409
Levy, D., & Lapidot, T. (2002). Shared terminology, private syntax: The case of recursive
descriptions. ITiCSE '02: Proceedings of the 7th Annual Conference on Innovation
and Technology in Computer Science Education, Aarhus, Denmark, 89-93.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/544414.544442
Lewis, T. L., Rosson, M. B., Perez-Quinones, M. A. (2004). What do the experts say?:
Teaching introductory design from an expert's perspective. SIGCSE '04: Proceedings
of the 35th SIGCSE Technical Symposium on Computer Science Education, Norfolk,
VA, United States, 296-300. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/971300.971405
Lister, R. (2004). Teaching Java first: Experiments with a pigs-early pedagogy. Sixth
Australasian Computing Education Conference (ACE2004); Conferences in
Research and Practice in Information Technology, 30 177-183.
Lister, R. (2001). Objectives and objective assessment in CS1. SIGCSE '01: Proceedings
of the Thirty-Second SIGCSE Technical Symposium on Computer Science
Education, Charlotte, NC, United States, 292-296. Retrieved September 4, 2006,
from http://doi.acm.Org/10.l 145/364447.364605
Lowther, J., & Shene, C. (2003). Teaching B-splines is not difficult! SIGCSE '03:
Proceedings of the 34th SIGCSE Technical Symposium on Computer Science
Education, Reno, NV, United States, 381-385. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/611892.612012
Lukins, S., Levicki, A., & Burg, J. (2002). A tutorial program for propositional logic with
human/computer interactive learning. SIGCSE '02: Proceedings of the 33rd SIGCSE
Technical Symposium on Computer Science Education, Cincinnati, OH, United
States 381-385. Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/
563340.563490
Lynch, S., & Rajendran, K. (2005). Multi-agent systems design for novices. Computer
Science Education, 75(1), 41-57.
215
Maj, S. P., Veal, D., & Duley, R. (2001). A proposed new high level abstraction for
computer technology. SIGCSE '01: Proceedings of the Thirty -Second SIGCSE
Technical Symposium on Computer Science Education, Charlotte, NC, United States,
199-203. Retrieved September 4, 2006, from http://doi.acm.Org/10.l 145/
364447.364584
Manolopoulos, Y. (2005). On the number of recursive calls of recursive functions.
SIGCSE Bulletin (Association for Computing Machinery, Special Interest Group on
Computer Science Education) , 37(2), 61-64.
Marks, J., Freeman, W., & Leitner, H. (2001). Teaching applied computing without
programming: A case-based introductory course for general education. SIGCSE '01:
Proceedings of the Thirty-Second SIGCSE Technical Symposium on Computer
Science Education, Charlotte, NC, United States, 80-84. Retrieved September 4,
2006, from http://doi.acm.org/10.1145/364447.364547
Marrero, W., & Settle, A. (2005). Testing first: Emphasizing testing in early
programming courses. ITiCSE '05: Proceedings of the 10th Annual SIGCSE
Conference on Innovation and Technology in Computer Science Education,
Caparica, Portugal, 4-8. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/1067445.1067451
Matos, V., & Grasser, R. (2000). RELAX \ — the relational algebra pocket calculator
project. SIGCSE Bulletin (Association for Computing Machinery, Special Interest
Group on Computer Science Education), 32(4), 40-44.
Maurer, W. D. (2002). The comparative programming languages course: A new chain of
development. SIGCSE '02: Proceedings of the 33rd SIGCSE Technical Symposium
on Computer Science Education, Cincinnati, OH, United States 336-340. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/563340.563472
McCormick, J. W. (2005). We've been working on the railroad: A laboratory for real-time
embedded systems. SIGCSE '05: Proceedings of the 36th SIGCSE Technical
Symposium on Computer Science Education, St. Louis, MO, United States, 530-534.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/1047344.1047510
McGuffee, J. W. (2000). Defining computer science. SIGCSE Bulletin (Association for
Computing Machinery, Special Interest Group on Computer Science Education),
32(2), 74-76.
216
McKinney, D., & Denton, L. F. (2005). Affective assessment of team skills in agile CS1
labs: The good, the bad, and the ugly. SIGCSE '05: Proceedings of the 36th SIGCSE
Technical Symposium on Computer Science Education, St. Louis, MO, United
States, 465-469. Retrieved September 4, 2006, from http://doi.acm.Org/10.l 145/
1047344.1047494
Medley, M. D. (2001). Using qualitative research software for CS education research.
ITiCSE '01: Proceedings of the 6th Annual Conference on Innovation and
Technology in Computer Science Education, Canterbury, United Kingdom, 141-144.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/377435.377668
Mendes, A. J., Gomes, A., Esteves, M., Marcelino, M. J., & Bravo, C, et al. (2005).
Using simulation and collaboration in CS1 and CS2. ITiCSE '05: Proceedings of the
10th Annual SIGCSE Conference on Innovation and Technology in Computer
Science Education, Caparica, Portugal, 193-197. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/1067445.1067499
Mendes, E., Mosley, N., & Counsell, S. (2001). The cognitive flexibility theoryO: An
approach for teaching hypermedia engineering. ITiCSE '01: Proceedings of the 6th
Annual Conference on Innovation and Technology in Computer Science Education,
Canterbury, United Kingdom, 21-24. Retrieved September 4, 2006, from
http://doi.acm.org/ 10.1145/377435.377457
Michael, M. (2000). Fostering and assessing communication skills in the computer
science context. SIGCSE '00: Proceedings of the Thirty-First SIGCSE Technical
Symposium on Computer Science Education, Austin, TX, United States, 1 19-123.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/330908.331834
Miller, C. (2003). Relating theory to actual results in computer science and human-
computer ilnteraction. Computer Science Education, 13(3), 227-240.
Modak, V. D., Langan, D. D., & Hain, T. F. (2005). A pattern-based development tool for
mobile agents. SIGCSE '05: Proceedings of the 36th SIGCSE Technical Symposium
on Computer Science Education, St. Louis, MO, United States, 72-75. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/1047344.1047382
Moody, D. L., & Sindre, G. (2003). Incorporating quality assurance processes into
requirements analysis education. ITiCSE '03: Proceedings of the 8th Annual
Conference on Innovation and Technology in Computer Science Education,
Thessaloniki, Greece, 74-78. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/961511.961534
217
Moore, T. K. (2002). Bringing the enterprise into a database systems course. SIGCSE '02:
Proceedings of the 33rd SIGCSE Technical Symposium on Computer Science
Education, Cincinnati, OH, United States 262-265. Retrieved September 4, 2006,
from http://doi.acm.Org/10.l 145/563340.563443
Moritz, S. H., & Blank, G. D. (2005). A design-first curriculum for teaching Java in a
CS1 course. SIGCSE Bulletin (Association for Computing Machinery, Special
Interest Group on Computer Science Education), 37(2), 89-93.
Moritz, S. H., Wei, F., Parvez, S. M., & Blank, G. D. (2005). From objects-first to
design-first with multimedia and intelligent tutoring. ITiCSE '05: Proceedings of the
1 0th Annual SIGCSE Conference on Innovation and Technology in Computer
Science Education, Caparica, Portugal, 99-103. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/1067445.1067475
Morris, J. (2005). Algorithm animation: Using the algorithm code to drive the animation.
Seventh Australasian Computing Education Conference (ACE2005); Conferences in
Research and Practice in Information Technology, 42, 15-20.
Mosiman, S., & Hiemcke, C. (2000). Interdisciplinary capstone group project: Designing
autonomous race vehicles. SIGCSE '00: Proceedings of the Thirty-First SIGCSE
Technical Symposium on Computer Science Education, Austin, TX, United States,
270-274. Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/
330908.331868
Muller, O. (2005). Pattern oriented instruction and the enhancement of analogical
reasoning. ICER '05: Proceedings of the 2005 International Workshop on
Computing Education Research, Seattle, WA, United States, 57-67. Retrieved
September 4, 2006, from http://doi.acm.org/10.1145/1089786.1089792
Muller, O., Haberman, B., & Averbuch, H. (2004). (An almost) pedagogical pattern for
pattern-based problem-solving instruction. ITiCSE '04: Proceedings of the 9th
Annual SIGCSE Conference on Innovation and Technology in Computer Science
Education, Leeds, United Kingdom. 102-106. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/1007996.1008025
Mantyla, T. Laakso, M., Innola, E., & Salakoski, T. (2005). Student counseling with the
SOPS-TOOL Kolin Kolistelut - Koli Calling 2005, Proceedings of the Fourth
Annual Finnish Annual Finnish Baltic Sea Conference on Computer Science
Education, Koli, Finland, 1 19-124. Retrieved September 4, 2006, from
http://cs.joensuu.fi/kolistelut/archive/2005/koli_proc_2005.pdf
218
Nagappan, N., Williams, L., Ferzli, M., Wiebe, E., Yang, K., & Miller, C, et al. (2003).
Improving the CS1 experience with pair programming. SIGCSE '03: Proceedings of
the 34th SIGCSE Technical Symposium on Computer Science Education, Reno, NV,
United States, 359-362. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/611892.612006
Nair, A. S. S., & Mahalakshmi, T. (2004). Conceptualizing data structures: A pedagogic
approach. ITiCSE-WGR '04: Working Group Reports from ITiCSE on Innovation
and Technology in Computer Science Education, Leeds, United Kingdom, 97-100.
Retrieved September 4, 2006, from http://doi.acm.Org/10.l 145/1044550.1041668
Naumanen, M. (2003). A narrative perspective to students' experiences in problem-based
leaning in theory of computation. Kolin Kolistelut - Koli Calling 2003, Proceedings
of the Third Annual Finnish Annual Finnish Baltic Sea Conference on Computer
Science Education, Koli, Finland, 59-62. Retrieved September 4, 2006, from
http://cs.joensuu.fi/kolistelut/archive/2003/koli_proc_2003.pdf
Nerur, S., Ramanujan, S. & Kesh, S. (2002, April). Pedagogical issues in object
orientation. Journal of Computer Science Education Online. Retrieved September 4,
2006, from http://www.iste.org/Template.cfm?Section=Aprill&Template=/
MembersOnly.cfm&ContentID=4234
Nguyen, D., & Wong, S. B. (2001). Design patterns for sorting. SIGCSE '01: Proceedings
of the Thirty-Second SIGCSE Technical Symposium on Computer Science
Education, Charlotte, NC, United States, 263-267. Retrieved September 4, 2006,
from http://doi.acm.Org/10.l 145/364447.364597
Noonan, R. E. (2000). An object-oriented view of backtracking. SIGCSE '00:
Proceedings of the Thirty-First SIGCSE Technical Symposium on Computer Science
Education, Austin, TX, United States, 362-366. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/330908.331886
Noyes, J. L. (2002). A first course in computational science: (why a math book isn't
enough). SIGCSE '02: Proceedings of the 33rd SIGCSE Technical Symposium on
Computer Science Education, Cincinnati, OH, United States 18-22. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/563340.563348
Null, L., & Rao, K. (2005). CAMERA: Introducing memory concepts via visualization.
SIGCSE '05: Proceedings of the 36th SIGCSE Technical Symposium on Computer
Science Education, St. Louis, MO, United States, 96-100. Retrieved September 4,
2006, from http://doi.acm.org/10.! 145/1047344.1047389
219
N0rmark, K. (2000). A suite of WWW-based tools for advanced course management.
ITiCSE '00: Proceedings of the 5th Annual SIGCSE/SIGCUE ITiCSEconference on
Innovation and Technology in Computer Science Education, Helsinki, Finland, 65-
68. Retrieved September 4, 2006, from http://doi.acm.Org/10.l 145/343048.343078
Odekirk, E., Jones, D., & Jensen, P. (2000). Three semesters of CSO using java:
Assignments and experiences. ITiCSE '00: Proceedings of the 5th Annual
SIGCSE/SIGCUE ITiCSE conference on Innovation and Technology in Computer
Science Education, Helsinki, Finland, 144-147. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/343048.343148
Or-Bach, R., & Lavy, I. (2004). Cognitive activities of abstraction in object orientation:
An empirical study. SIGCSE Bulletin (Association for Computing Machinery,
Special Interest Group on Computer Science Education) , 36(2), 82-86.
Ou, K.-L., Chen, G., Liu, C, & Liu, B. (2000). Instructional instruments for web group
learning systems: The grouping, intervention, and strategy. ITiCSE '00: Proceedings
of the 5th Annual SIGCSE/SIGCUE ITiCSE conference on Innovation and
Technology in Computer Science Education, Helsinki, Finland, 69-72. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/343048.343079
Parberry, I., Roden, T., & Kazemzadeh, M. B. (2005). Experience with an industry-driven
capstone course on game programming: Extended abstract. SIGCSE '05:
Proceedings of the 36th SIGCSE Technical Symposium on Computer Science
Education, St. Louis, MO, United States, 91-95. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/1047344.1047387
Parker, J. R., & Becker, K. (2003). Measuring effectiveness of constructivist and
behaviourist assignments in CS102. ITiCSE '03: Proceedings of the 8th Annual
Conference on Innovation and Technology in Computer Science Education,
Thessaloniki, Greece, 40-44. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/961511.961526
Parkinson, A., & Redmond, J. A. (2002). Do cognitive styles affect learning performance
in different computer media? ITiCSE '02: Proceedings of the 7th Annual Conference
on Innovation and Technology in Computer Science Education, Aarhus, Denmark,
39-43. Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/
544414.544427
220
Parkinson, A., Redmond, J. A., & Walsh, C. (2004). Accommodating field-dependence:
A cross-over study. ITiCSE '04: Proceedings of the 9th Annual SIGCSE Conference
on Innovation and Technology in Computer Science Education, Leeds, United
Kingdom, 72-76. Retrieved September 4, 2006, from http://doi.acm.Org/10.l 145/
1007996.1008018
Parrish, A., Dixon, B., & Cordes, D. (2001). Binary software components in the
undergraduate computer science curriculum. SIGCSE '01: Proceedings of the Thirty-
Second SIGCSE Technical Symposium on Computer Science Education, Charlotte,
NC, United States, 332-336. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/364447.364615
Pears, A. & Olsson, H. (2004). Explanograms: low overhead multi-media learning
resources. Kolin Kolistelut - Koli Calling 2004, Proceedings of the Fourth Annual
Finnish Annual Finnish Baltic Sea Conference on Computer Science Education,
Koli, Finland, 61-1 A. Retrieved September 4, 2006, from http://cs.joensuu.fi/
kolistelut/ archive/2004/koli_proc_2004.pdf
Perrenet, J., Groote, J. F., & Kaasenbrood, E. (2005). Exploring students' understanding
of the concept of algorithm: Levels of abstraction. ITiCSE '05: Proceedings of the
10th Annual SIGCSE Conference on Innovation and Technology in Computer
Science Education, Caparica, Portugal, 64-68. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/1067445.1067467
Phoha, V. V. (2001). An interactive dynamic model for integrating knowledge
management methods and knowledge sharing technology in a traditional classroom.
SIGCSE '01: Proceedings of the Thirty-Second SIGCSE Technical Symposium on
Computer Science Education, Charlotte, NC, United States, 144-148. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/364447.364567
Pillay, N. (2004). A first course in genetic programming. ITiCSE-WGR '04: Working
Group Reports from ITiCSE on Innovation and Technology in Computer Science
Education, Leeds, United Kingdom, 93-96. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/1044550.1041667
Pillay, N. (2003). Developing intelligent programming tutors for novice programmers.
SIGCSE Bulletin (Association for Computing Machinery, Special Interest Group on
Computer Science Education) , 35(2), 78-82.
221
Pollock, L., McCoy, K., Carberry, S., Hundigopal, N., & You, X. (2004). Increasing high
school girls' self confidence and awareness of CS through a positive summer
experience. SIGCSE '04: Proceedings of the 35th SIGCSE Technical Symposium on
Computer Science Education, Norfolk, VA, United States, 185-189. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/971300.971369
Preston, J. A., & Wilson, L. (2001). Offering CS1 on-line reducing campus resource
demand while improving the learning environment. SIGCSE '01: Proceedings of the
Thirty-Second SIGCSE Technical Symposium on Computer Science Education,
Charlotte, NC, United States, 342-346. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/364447.364618
Prior, J. C, & Lister, R. (2004). The backwash effect on SQL skills grading. ITiCSE '04:
Proceedings of the 9th Annual SIGCSE Conference on Innovation and Technology in
Computer Science Education, Leeds, United Kingdom, 32-36. Retrieved September
4, 2006, from http://doi.acm.org/10.1145/1007996.1008008
Proulx, V. K. (2000). Programming patterns and design patterns in the introductory
computer science course. SIGCSE '00: Proceedings of the Thirty-First SIGCSE
Technical Symposium on Computer Science Education, Austin, TX, United States,
80-84. Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/
330908.331819
Pullen, J. M., Norris, E., & Fix, M. (2000). Teaching C++ in a multi-user virtual
environment. SIGCSE Bulletin (Association for Computing Machinery, Special
Interest Group on Computer Science Education), 32(2), 60-64.
Rajaravivarma, R. (2005). A games-based approach for teaching the introductory
programming course. SIGCSE Bulletin (Association for Computing Machinery,
Special Interest Group on Computer Science Education) , 37(4), 98-102.
Ramakrishnan, S., & Nwosu, E. (2003). DBMS course: Web based database
administration tool and class projects. SIGCSE '03: Proceedings of the 34th SIGCSE
Technical Symposium on Computer Science Education, Reno, NV, United States,
16-20. Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/
611892.611922
Randall, C, Price, B., & Reichgelt, H. (2003). Women in computing programs: Does the
incredible shrinking pipeline apply to all computing programs? SIGCSE Bulletin
(Association for Computing Machinery, Special Interest Group on Computer Science
Education), 35(4), 55-59.
222
Randolph, J. J., Bednarik, R. & Myller, N. (2005). A methodological review of the
articles published in the proceedings of Koli Calling 2001-2004. Kolin Kolistelut -
Koli Calling 2005, Proceedings of the Fourth Annual Finnish Annual Finnish
Baltic Sea Conference on Computer Science Education, Koli, Finland, 103-1 1 1.
Retrieved September 4, 2006, from http://cs.joensuu.fi/kolistelut/archive/2005/
koli_proc_2005 .pdf
Raner, M. (2000). Teaching object-orientation with the object visualization and
annotation language (OVAL). ITiCSE '00: Proceedings of the 5th Annual
SIGCSE/SIGCUE ITiCSE conference on Innovation and Technology in Computer
Science Education, Helsinki, Finland, 45-48. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/343048.343071
Rasala, R. (2000). Toolkits in first year computer science: A pedagogical imperative.
SIGCSE '00: Proceedings of the Thirty-First SIGCSE Technical Symposium on
Computer Science Education, Austin, TX, United States, 185-191. Retrieved
September 4, 2006, from http://doi.acm.org/10.1145/330908.331852
Rasala, R. (2003). Embryonic object versus mature object: Object-oriented style and
pedagogical theme. ITiCSE '03: Proceedings of the 8th Annual Conference on
Innovation and Technology in Computer Science Education, Thessaloniki, Greece,
89-93, Retrieved September 4, 2006, from http://doi.acm.Org/10.l 145/
961511.961538
Rasala, R., Raab, J., & Proulx, V. K. (2001). Java power tools: Model software for
teaching object-oriented design. SIGCSE '01: Proceedings of the Thirty-Second
SIGCSE Technical Symposium on Computer Science Education, Charlotte, NC,
United States, 297-301. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/364447.364606
Redmond, J. A., Walsh, C, & Parkinson, A. (2003). Equilibriating instructional media
for cognitive styles. ITiCSE '03: Proceedings of the 8th Annual Conference on
Innovation and Technology in Computer Science Education, Thessaloniki, Greece,
55-59. Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/
961511.961529
Reges, S. (2002). Can C\# replace Java in CS1 and CS2? ITiCSE '02: Proceedings of the
7th Annual Conference on Innovation and Technology in Computer Science
Education, Aarhus, Denmark, 4-8. Retrieved September 4, 2006, from
http://doi.acm.Org/0.1145/544414.544419
223
Reges, S. (2000). Conservatively radical Java in CS1. SIGCSE '00: Proceedings of the
Thirty-First SIGCSE Technical Symposium on Computer Science Education, Austin,
TX, United States, 85-89. Retrieved September 4, 2006, from http://doi.acm.org/
0.1145/330908.331821
Reis, C, & Cartwright, R. (2004). Taming a professional IDE for the classroom. SIGCSE
'04: Proceedings of the 35th SIGCSE Technical Symposium on Computer Science
Education, Norfolk, VA, United States, 156-160. Retrieved September 4, 2006,from
http://doi.acm.org/10.1145/971300.971357
Richards, B. (2000). Bugs as features: Teaching network protocols through debugging.
SIGCSE '00: Proceedings of the Thirty-First SIGCSE Technical Symposium on
Computer Science Education, Austin, TX, United States, 256-259. Retrieved
September 4, 2006,from http://doi.acm.Org/10.l 145/330908.331865
Richards, B., & Stull, B. (2004). Teaching wireless networking with limited resources.
SIGCSE '04: Proceedings of the 35th SIGCSE Technical Symposium on Computer
Science Education, Norfolk, VA, United States, 306-310. Retrieved September 4,
2006,from http://doi.acm.0rg/lO.l 145/971300.971408
Robbert, M. A. (2000). Enhancing the value of a project in the database course. SIGCSE
'00: Proceedings of the Thirty-First SIGCSE Technical Symposium on Computer
Science Education, Austin, TX, United States, 36-40. Retrieved September 4, 2006,
from http://doi.acm.org/10.1145/330908.331807
Robbins, S. (2003). Using remote logging for teaching concurrency. SIGCSE '03:
Proceedings of the 34th SIGCSE Technical Symposium on Computer Science
Education, Reno, NV, United States, 177-181. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/611892.611963
Robbins, S. (2001). Starving philosophers: Experimentation with monitor
synchronization. SIGCSE '01: Proceedings of the Thirty-Second SIGCSE Technical
Symposium on Computer Science Education, Charlotte, NC, United States, 317-321.
Retrieved September 4, 2006,from http://doi.acm.Org/10.l 145/ 64447.364612
Robins, A, Roundtree, J., & Roundtree, N. (2003). Learning and teaching programming:
A review and discussion. Computer Science Education, 13(2), 137-172.
Roberts, E. S., Kassianidou, M., & Irani, L. (2002). Encouraging women in computer
science. SIGCSE Bulletin (Association for Computing Machinery, Special Interest
Group on Computer Science Education), 34(2), 84-88.
224
Rodger, S. H. (2002). Introducing computer science through animation and virtual
worlds. SIGCSE '02: Proceedings of the 33rd SIGCSE Technical Symposium on
Computer Science Education, Cincinnati, OH, United States 186-190. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/563340.56341 1
Rosbottom, J., Crellin, J., & Fysh, D. (2000). A generic model for on-line learning.
ITiCSE '00: Proceedings of the 5th Annual SIGCSE/SIGCUE ITiCSEconference on
Innovation and Technology in Computer Science Education, Helsinki, Finland. 108-
111. Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/343048.343131
Rosson, M. B., Carroll, J. M., & Rodi, C. M. (2004). Case studies for teaching United
Statesbility engineering. SIGCSE '04: Proceedings of the 35th SIGCSE Technical
Symposium on Computer Science Education, Norfolk, VA, United States, 36-40.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/971300.971315
Rountree, J., Rountree, N., Robins, A., & Hannah, R. (2005). Observations of student
competency in a CS1 course. Seventh Australasian Computing Education
Conference (ACE2005); Conferences in Research and Practice in Information
Technology, 42, 145-149.
Rofiling, G., & Naps, T. L. (2002). A testbed for pedagogical requirements in algorithm
visualizations. ITiCSE '02: Proceedings of the 7th Annual Conference on Innovation
and Technology in Computer Science Education, Aarhus, Denmark, 96-100.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/544414.544446
Saastamoinen, K. (2003). Distance learning experiences from the mathematical modeling
course. Kolin Kolistelut - Koli Calling 2003, Proceedings of the Third Annual
Finnish Annual Finnish Baltic Sea Conference on Computer Science Education,
Koli, Finland, 22-26. Retrieved September 4, 2006, from http://cs.joensuu.fi/
kolistelut/ archive/2003/koli_proc_2003 .pdf
Sajaniemi, J., & Kuittinen, M. (2005). An experiment on using roles of variables in
teaching introductory programming. Computer Science Education, 15{\), 59-82.
Sanders, I. (2002). Teaching empirical analysis of algorithms. SIGCSE '02: Proceedings
of the 33rd SIGCSE Technical Symposium on Computer Science Education,
Cincinnati, OH, United States 321-325. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/563340.563468
Sanders, D. (2002). Extreme programming: The student view. Computer Science
Education, 12(3), 235-250.
225
Sanders, K. E., & McCartney, R. (2003). Program assessment tools in computer science:
A report from the trenches. SIGCSE '03: Proceedings of the 34th SIGCSE Technical
Symposium on Computer Science Education, Reno, NV, United States, 31-35.
Retrieved September 4, 2006, from http://doi.acm.Org/10.l 145/61 1892.61 1926
Sanders, K., & McCartney, R. (2004). Collected wisdom: Assessment tools for computer
science programs. Computer Science Education, 14(3), 183-203.
Sarkar, N. I., & Craig, T. M. (2004). Illustrating computer hardware concepts using PIC-
based projects. SIGCSE '04: Proceedings of the 35th SIGCSE Technical Symposium
on Computer Science Education, Norfolk, VA, United States, 270-274. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/971300.971395
Saxon, C. S. (2003). Object-oriented recursive descent parsing in C\#. SIGCSE Bulletin
(Association for Computing Machinery, Special Interest Group on Computer Science
Education), 35(4), 82-85.
Schaub, S. (2000). Teaching Java with graphics in CS1. SIGCSE Bulletin (Association for
Computing Machinery, Special Interest Group on Computer Science Education),
32(2), 71-73.
Schulte, C. (2002). A picture is worth a thousand words? - UML as a multimedia
learning tool. Kolin Kolistelut - Koli Calling 2002, Proceedings of the Second
Annual Finnish Annual Finnish Baltic Sea Conference on Computer Science
Education, Koli, Finland, 57-62. Retrieved September 4, 2006, from
http://cs.joensuu.fi/kolistelut/archive/2002/koli_proc_2002.pdf
Schulte, C. (2005). Dynamic object structures as a conceptual framework for teaching
object oriented concepts to novices. Kolin Kolistelut - Koli Calling 2005,
Proceedings of the Fourth Annual Finnish Annual Finnish Baltic Sea Conference
on Computer Science Education, Koli, Finland, 113-118. Retrieved September 4,
2006, from http://cs.joensuu.fi/kolistelut/archive/2005/koli_proc_2005 .pdf
Seppala, O., Malmi, L. & Korhonen, A. (2005). Observations on student errors in
algorithm simulation exercises. Kolin Kolistelut - Koli Calling 2005, Proceedings
of the Fourth Annual Finnish Annual Finnish Baltic Sea Conference on Computer
Science Education, Koli, Finland, 81-86. Retrieved September 4, 2006, from
http://cs.joensuu.fi/kolistelut/archive/2005/koli_proc_2005.pdf
Shaffer, C. A. (2004). Buffer pools and file processing projects for an undergraduate data
structures course. SIGCSE '04: Proceedings of the 35th SIGCSE Technical
Symposium on Computer Science Education, Norfolk, VA, United States, 175-178.
Retrieved September 4, 2006, from http://doi.acm.org/10.! 145/971300.971362
226
Shaw, K., & Dermoudy, J. (2005). Engendering an empathy for software engineering.
Seventh Australasian Computing Education Conference (ACE2005); Conferences in
Research and Practice in Information Technology, 42, 135-144.
Sheard, J. (2004). Electronic learning communities: Strategies for establishment and
management. ITiCSE '04: Proceedings of the 9th Annual SIGCSE Conference on
Innovation and Technology in Computer Science Education, Leeds, United
Kingdom, 37-41. Retrieved September 4, 2006, from http://doi.acm.Org/10.l 145/
1007996.1008009
Sheard, J., & Dick, M. (2003). Influences on cheating practice of graduate students in IT
courses: What are the factors? ITiCSE '03: Proceedings of the 8th Annual
Conference on Innovation and Technology in Computer Science Education,
Thessaloniki, Greece, 45-49. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/961511.961527
Simmonds, A. (2003). Student learning experience with an industry certification course at
university. Fifth Australasian Computing Education Conference (ACE2003);
Conferences in Research and Practice in Information Technology, 20, 143-147.
Simon, B., Anderson, R., Hoyer, C, & Su, J. (2004). Preliminary experiences with a
tablet PC based system to support active learning in computer science courses.
ITiCSE '04: Proceedings of the 9th Annual SIGCSE Conference on Innovation and
Technology in Computer Science Education, Leeds, United Kingdom, 213-217.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/1007996.1008053
Smarkusky, D., Dempsey, R., Ludka, J., & Quillettes, F. d. (2005). Enhancing team
knowledge: Instruction vs. experience. SIGCSE '05: Proceedings of the 36th
SIGCSE Technical Symposium on Computer Science Education, St. Louis, MO,
United States, 460-464. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/1047344.1047493
Soh, L., Samal, A., Person, S., Nugent, G., & Lang, J. (2005). Analyzing relationships
between closed labs and course activities in CS1. ITiCSE '05: Proceedings of the
10th Annual SIGCSE Conference on Innovation and Technology in Computer
Science Education, Caparica, Portugal, 183-187. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/1067445.1067497
Soh, L., Samal, A., Person, S., Nugent, G, & Lang, J. (2005). Designing, implementing,
and analyzing a placement test for introductory CS courses. SIGCSE '05:
Proceedings of the 36th SIGCSE Technical Symposium on Computer Science
Education, St. Louis, MO, United States, 505-509. Retrieved September 4, 2006,
from http://doi.acm.org/10.! 145/1047344.1047504
227
Solomon, A. (2003). Applying NAILS to blackboard. Fifth Australasian Computing
Education Conference (ACE2003); Conferences in Research and Practice in
Information Technology, 20, 263-266.
Sooriamurthi, R. (2001). Problems in comprehending recursion and suggested solutions.
ITiCSE '01: Proceedings of the 6th Annual Conference on Innovation and
Technology in Computer Science Education, Canterbury, United Kingdom, 25-28.
Retrieved September 4, 2006, from http://doi.acm.Org/10.l 145/377435.377458
Spooner, D. L. (2000). A bachelor of science in information technology: An
interdisciplinary approach. SIGCSE '00: Proceedings of the Thirty-First SIGCSE
Technical Symposium on Computer Science Education, Austin, TX, United States,
285-289. Retrieved September 4, 2006 from, http://doi.acm.org/10.1145/
330908.331871
Steenkiste, P. (2003). A network project course based on network processors. SIGCSE
'03: Proceedings of the 34th SIGCSE Technical Symposium on Computer Science
Education, Reno, NV, United States, 262-266. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/611892.611984
Stern, L., & Naish, L. (2002). Visual representations for recursive algorithms. SIGCSE
'02: Proceedings of the 33rd SIGCSE Technical Symposium on Computer Science
Education, Cincinnati, OH, United States 196-200. Retrieved September 4, 2006,
from http://doi.acm.org/10.1145/563340.563414
Stevenson, D. E., & Phillips, A. T. (2003). Implementing object equivalence in Java using
the template method design pattern. SIGCSE '03: Proceedings of the 34th SIGCSE
Technical Symposium on Computer Science Education, Reno, NV, United States,
278-282. Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/
611892.611987
Sweedyk, E., & Keller, R. M. (2005). Fun and games: A new software engineering
course. ITiCSE '05: Proceedings of the 10th Annual SIGCSE Conference on
Innovation and Technology in Computer Science Education, Caparica, Portugal,
138-142. Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/
1067445.1067485
Tegos, G. K, Stoyanova, D. V., & Onkov, K. Z. (2005). E-learning of trend modeling in
a web-environment. SIGCSE Bulletin (Association for Computing Machinery,
Special Interest Group on Computer Science Education) , 37(2), 70-74.
228
Tenenberg, J. (2003). A framework approach to teaching data structures. SIGCSE '03:
Proceedings of the 34th SIGCSE Technical Symposium on Computer Science
Education, Reno, NV, United States, 210-214. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/611892.611971
Thomas, P. (2003). The evaluation of electronic marking of examinations. ITiCSE '03:
Proceedings of the 8th Annual Conference on Innovation and Technology in
Computer Science Education, Thessaloniki, Greece, 50-54. Retrieved September 4,
2006, from http://doi.acm.Org/10.l 145/96151 1.961528
Thomas, P., Waugh, K., & Smith, N. (2005). Experiments in the automatic marking of
ER-diagrams. ITiCSE '05: Proceedings of the 10th Annual SIGCSE Conference on
Innovation and Technology in Computer Science Education, Caparica, Portugal,
158-162. Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/
1067445.1067490
Tikvati, A., Ben-Ari, M., & Kolikant, Y. B. (2004). Virtual trees for the Byzantine
generals algorithm. SIGCSE '04: Proceedings of the 35th SIGCSE Technical
Symposium on Computer Science Education, Norfolk, VA, United States, 392-396.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/971300.971435
Tolhurst, D., & Baker, B. (2003). A new approach to a first year undergraduate
information systems course. Fifth Australasian Computing Education Conference
(ACE2003); Conferences in Research and Practice in Information Technology, 20,
169-177.
Townsend, G. C. (2002). People who make a difference: Mentors and role models.
SIGCSE Bulletin (Association for Computing Machinery, Special Interest Group on
Computer Science Education) , 34(2), 57-61.
Traynor, C, & McKenna, M. (2003). Service learning models connecting computer
science to the community. SIGCSE Bulletin (Association for Computing Machinery,
Special Interest Group on Computer Science Education) , 35(4), 43-46.
Treu, K. (2002). To teach the unteachable class: An experimental course in web-based
application design. SIGCSE '02: Proceedings of the 33rd SIGCSE Technical
Symposium on Computer Science Education, Cincinnati, OH, United States 201-205.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/563340.563416
Truong, N., Roe, P., & Bancroft, P. (2004). Static analysis of students' Java programs.
Sixth Australasian Computing Education Conference (ACE2004); Conferences in
Research and Practice in Information Technology, 30, 317-325.
229
Turtle, S. M. (2000). A capstone course for a computer information systems major.
SIGCSE '00: Proceedings of the Thirty-First SIGCSE Technical Symposium on
Computer Science Education, Austin, TX, United States, 265-269. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/330908.331867
VanDeGrift, T. (2004). Coupling pair programming and writing: Learning about students'
perceptions and processes. SIGCSE '04: Proceedings of the 35th SIGCSE Technical
Symposium on Computer Science Education, Norfolk, VA, United States, 2-6.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/971300.971306
Vandenberg, S., & Wollowski, M. (2000). Introducing computer science using a breadth-
first approach and functional programming. SIGCSE '00: Proceedings of the Thirty-
First SIGCSE Technical Symposium on Computer Science Education, Austin, TX,
United States, 180-184. Retrieved September 4, 2006, from http://doi.acm.org/
10.1145/330908.331851
van der Veer, G., & van Vliet, H. (2003). A plea for a poor man's HCI component in
software engineering and computer science curricula; after all: The human-
computer interface is the system. Computer Science Education, 13(3), 207-225.
Vaughn, J. (2001). Teaching industrial practices in an undergraduate software
engineering course. Computer Science Education, 11(1), 21-32.
Veen, M. V., Mulder, F., & Lemmen, K. (2004). What is lacking in curriculum schemes
for computing/informatics? ITiCSE '04: Proceedings of the 9th Annual SIGCSE
Conference on Innovation and Technology in Computer Science Education, Leeds,
United Kingdom, 186-190. Retrieved September 4, 2006. from http://doi.acm.org/
10.1145/1007996.1008046
Velazquez -Iturbide, J. (2000). Recursion in gradual steps (is recursion really that
difficult?). SIGCSE '00: Proceedings of the Thirty-First SIGCSE Technical
Symposium on Computer Science Education, Austin, TX, United States, 310-314.
Retrieved September 4, 2006, from http://doi.acm.0rg/lO.l 145/330908.331876
Venables, A., & Haywood, L. (2003). Programming students need instant feedback! Fifth
Australasian Computing Education Conference (ACE2003); Conferences in
Research and Practice in Information Technology, 20, 267-272.
Ventura, P., & Ramamurthy, B. (2004). Wanted: CS1 students, no experience required.
SIGCSE '04: Proceedings of the 35th SIGCSE Technical Symposium on Computer
Science Education, Norfolk, VA, United States, 240-244. Retrieved September 4,
2006, from http://doi.acm.org/10.! 145/971300.971387
230
Vidal, J., M., & Buhler, P. (2002). Using RoboCup to teach multiagent systems and the
distributed mindset. SIGCSE '02: Proceedings of the 33rd SIGCSE Technical
Symposium on Computer Science Education, Cincinnati, OH, United States 3-7.
Retrieved September 4, 2006, from http://doi.acm.Org/10.l 145/563340.563344
von Konsky, B. R., Ivins, J., & Robey, M. (2005). Using PSP to evaluate student effort in
achieving learning - outcomes in a software engineering assignment. Seventh
Australasian Computing Education Conference (ACE2005); Conferences in
Research and Practice in Information Technology, 42, 193-201.
Wagner, P. J., Shoop, E., & Carlis, J. V. (2003). Using scientific data to teach a database
systems course. SIGCSE '03: Proceedings of the 34th SIGCSE Technical Symposium
on Computer Science Education, Reno, NV, United States, 224-228. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/61 1892.61 1975
Warren, P. (2004). Learning to program: Spreadsheets, scripting and HCI. Sixth
Australasian Computing Education Conference (ACE2004); Conferences in
Research and Practice in Information Technology, 30, 327-333.
Wells, M. A., & Brook, P. W. (2004). Conversational KM - student driven learning. Sixth
Australasian Computing Education Conference (ACE2004); Conferences in
Research and Practice in Information Technology, 30, 335-341.
Whiddett, R., Jackson, B., & Handy, J. (2000). Teaching information systems
management skills: Using integrated projects and case studies. Computer Science
Education, 10(2), 165-177.
White, K. (2003). A comprehensive CMPS II semester project. SIGCSE Bulletin
(Association for Computing Machinery, Special Interest Group on Computer Science
Education), 35(2), 70-73.
Wicentowski, R., & Newhall, T. (2005). Using image processing projects to teach CS1
topics. SIGCSE '05: Proceedings of the 36th SIGCSE Technical Symposium on
Computer Science Education, St. Louis, MO, United States, 287-291. Retrieved
September 4, 2006, from http://doi.acm.0rg/lO.l 145/1047344.1047445
Wiedenbeck, S. (2005). Factors affecting the success of non-majors in learning to
program. ICER '05: Proceedings of the 2005 International Workshop on Computing
Education Research, Seattle, WA, United States, 13-24. Retrieved September 4,
2006, from http://doi.acm.org/10.1145/1089786.1089788
231
Wilkerson, M., Griswold, W. G., & Simon, B. (2005). Ubiquitous presenter: Increasing
student access and control in a digital lecturing environment. SIGCSE '05:
Proceedings of the 36th SIGCSE Technical Symposium on Computer Science
Education, St. Louis, MO, United States, 1 16-120. Retrieved September 4, 2006,
from http://doi.acm.Org/10.l 145/1047344.1047394
Williams, L., Wiebe, E., Yang, K., Ferzli, M., & Miller, C. (2002). In support of pair
programming in the introductory computer science course. Computer Science
Education, 12(3), 197-212.
Wiseman, Y. (2005). Advanced non-distributed operating systems course. SIGCSE
Bulletin (Association for Computing Machinery, Special Interest Group on
Computer Science Education), 37(2), 65-69.
Wolz, U. (2001). Teaching design and project management with lego RCX robots.
SIGCSE '01: Proceedings of the Thirty-Second SIGCSE Technical Symposium on
Computer Science Education, Charlotte, NC, United States, 95-99. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/364447.364551
Wong, Y., Burg, J., & Strokanova, V. (2004). Digital media in computer science
curricula. SIGCSE '04: Proceedings of the 35th SIGCSE Technical Symposium on
Computer Science Education, Norfolk, VA, United States, 427-43 1 . Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/971300.971444
Woodford, K., & Bancroft, P. (2005). Multiple choice questions not considered harmful.
Seventh Australasian Computing Education Conference (ACE2005); Conferences in
Research and Practice in Information Technology, 42 109-116.
Yacef, K. (2004). Making large class teaching more adaptive with the logic-ITA. Sixth
Australasian Computing Education Conference (ACE2004); Conferences in
Research and Practice in Information Technology, 30, 343-347.
Yehezkel, C. (2002). A taxonomy of computer architecture visualizations. ITiCSE '02:
Proceedings of the 7th Annual Conference on Innovation and Technology in
Computer Science Education, Aarhus, Denmark, 101-105. Retrieved September 4,
2006, from http://doi.acm.org/10.1145/544414.544447
Yoo, S., & Hovis, S. (2004). Remote access internetworking laboratory. SIGCSE '04:
Proceedings of the 35th SIGCSE Technical Symposium on Computer Science
Education, Norfolk, VA, United States, 31 1-314. Retrieved September 4, 2006, from
http://doi.acm.org/10.1145/971300.971409
232
Yurcik, W., & Brumbaugh, L. (2001). A web-based little man computer simulator.
SIGCSE '01: Proceedings of the Thirty-Second SIGCSE Technical Symposium on
Computer Science Education, Charlotte, NC, United States, 204-208. Retrieved
September 4, 2006, from http://doi.acm.Org/10.l 145/364447.364585
Zaring, A. (2001). CCL: A subject language for compiler projects. Computer Science
Education, 11(2), 135-163.
Zelle, J. M., & Figura, C. (2004). Simple, low-cost stereographies: VR for everyone.
SIGCSE '04: Proceedings of the 35th SIGCSE Technical Symposium on Computer
Science Education, Norfolk, VA, United States, 348-352. Retrieved September 4,
2006, from http://doi.acm.Org/10.l 145/971300.971421
Zeller, A. (2000). Making students read and review code. ITiCSE '00: Proceedings of the
5th Annual SIGCSE/SIGCUE ITiCSE conference on Innovation and Technology in
Computer Science Education, Helsinki, Finland. 89-92. Retrieved September 4,
2006, from http://doi.acm.org/10.! 145/343048.343090
233
Appendix B:
Methodological Review Coding Form
DEO =
DEOO
DE000. 1 = yes, 2 = no.
DEI. (reviewer): 1 = Justus, 2 = Roman, 3 = Nikko, 4 =other
DE2. (forum): 1 = SICGSE proceedings, 2 = SIGCSE bulletin, 3 = ITICES, 4 = CSER,
5 = KOLI, 6 = ICER, 7 = JCSE, 8 = ACE.
DE3. (year): = 2000, 1 = 2001, 2 = 2002, 3 = 2003, 4 = 2004, 5 = 2005.
DE4. (volume) (three numerical digits - use zero for blank digits; e.g., Volume 1
would be 001.)
DE5. (issue) (two numerical digits)
DE6. (page) (up to four digits)
DE6a. (pages)
DE7. (region) 1 = Africa, 2 = Asian-Pacific or Eurasia, 3 = Europe, 4 = Middle East,
5 = North America, 6 = South or Central America, 7 = IMPDET
DE7a (university) Write in.
DE7b (authors) #
DE7c (name) Last name, Initials
234
DE8. (Subject) 1 = New way to organize a course, 2 = Tool, 3 = Teaching programming
language category, 4 = Curriculum, 5 = Visualization, 6 = Simulation, 7 = Parallel
computing, 8 = Other.
DE8a (Valentine) 1 = Experimental, 2 = Marco Polo, 3 = Tools, 4 = John Henry, 5 =
Philosophy, 6 = Nifty
DE9. (human participants) 1 = yes, 2 = no. (If yes, go to DE9a ; if no go to A9.)
DE9a (anecdotal) 1 = yes, 2 = no.
(if yes, gotoM21.)
Type of Papers that Did Not Report Research on Human Subjects
A9. (type of other) 1 = Literature review, 2 = Program description, 3 = Theory,
Methodology, Philosophy paper, 4 = Technical investigation, 5 = Other (if 1-4, end;
if5gotoA10)
A10 (Other other) Write in a short description (End).
Methodology Type
M21. Experimental/quasi-experimental 1 = yes, 2 = no
(If M21 = yes, go to AS5, else go to M22.)
235
AS5. (assignment) 1 = self-selection 2 = random 3 = researcher-assigned
M22. Explanatory descriptive
M23. Exploratory description
M24. Correlational
M25. Causal-comparative
M26. IMPDET or anecdotal
1 = yes, 2 = no
1 = yes, 2 = no
1 = yes, 2 = no
1 = yes, 2 = no
1 = yes, 2 = no
M27. (selection) 1 = random, 2 = intentional, 3 = convenience/preexisting
[Go to All]
Report Structure
1 . Abstract
1 = narrative, 2
2. (introduce problem)
1 = yes, 2 = no
3. (literature review)
1 = yes, 2 = no
4. (purpose/rationale)
1 = yes, 2 = no
5. (questions/hypotheses)
1 = yes, 2 = no
6. (participants)
1 = yes, 2 = no
6a (grade level)
1 = preschool
2 = k-3
3 = 4-6
4 = 7-9
236
A 16b (Undergraduate
curriculum year)
A17. (settings)
A18. (instruments)
A19. (procedure)
A20. (results and discussion)
5 = 10-12
6 = bachelor
7 = masters
8 = doctoral
9 = post-doctoral
10 = other
11 = can't determine
1 = first year
2 = second year
3 = third year
4 = fourth year
1 = yes, 2 = no
1 = yes, 2 = no, -9 = n/a
1 = yes, 2 = no
1 = yes, 2 = no
[Go to RD1, if M21 = 1, else go to II.]
Experimental Research Designs
237
RD1. (design) Was M21, marked as Yes 1 = yes, 2 = no
[if yes, RD2; If no go to II]
RD2 (postonly) posttest, no controls
RD3 (post control) posttest, with controls,
RD4 (prepost only= pretest/posttest without controls
RD5 (prepost control) pretest/posttest with controls
RD6 (repeated) group repeated measures
RD7 (multiple) multiple factor
RD1 1 (factor?) If group repeated measures,
was there an experimental between group factor?
RD8 (single) single-subject
RD9 (other) other
[ifRD9, go to RD 10]
RD 10 (explain) If other, explain
RDH (posttest only highest)
= yes, 2 = no
= yes, 2 = no
= yes, 2 = no
= yes, 2 = no
= yes, 2 = no
= yes, 2 = no
yes, 2 = no
= yes, 2 = no
= yes, 2 = no
1 = yes, 2 = no
238
Independent Variables (interventions)
II. Was an independent (manipulatable) variable used in this study? 1 = yes, 2 = no
[If yes got to 12, if no go to Dl]
12 (student instruction)
13 (teacher instruction)
14 (CS fair /contest)
15 (mentoring)
16 (Speakers at school)
17 (CS field trips)
18 (other)
= yes, 2 = no
= yes, 2 = no
= yes, 2 = no
= yes, 2 = no
= yes, 2 = no
= yes, 2 = no
= yes, 2 = no
If I8a (explain) If other, explain:
[GotoDl]
Dependent Variables
Dl (attitudes)
D2 (attendance)
D3 (core achievement)
1 = yes, 2 = no
1 = yes, 2 = no
1 = yes, 2 = no
239
D4 (CS achievement)
D5 (teaching practices)
D6 (intentions for future)
D7 (program implementation)
D8 (costs and benefits $)
D9 (socialization)
D10 (computer use)
Dll (other)
Dl la (explain) If Dl 1, explain
[Go to Ml]
1 = yes, 2 = no
yes, 2 = no
yes, 2 = no
yes, 2 = no
yes, 2 = no
yes, 2 = no
yes, 2 = no
yes, 2 = no
Measures
Ml (grades)
M2 (diary)
M3 (questionnaire)
M3a (ques. psych)
M4 (log files)
M5 (test)
M5a (test psych)
M6 (interviews)
1 = yes, 2 = no
1 = yes, 2 = no
1 = yes, 2 = no
1 = yes, 2 = no
1 = yes, 2 = no
1 = yes, 2 = no
1 = yes, 2 = no
1 = yes, 2 = no
240
M7 (direct)
M7a (direct psych)
M8 (stand. Test)
M8a (psych. Stand)
M9 (student work)
M10 (focus groups)
Mil (existing data)
M12 (other)
yes, 2 = no
yes, 2 = no
yes, 2 = no
yes, 2 = no
yes, 2 = no
yes, 2 = no
yes, 2 = no
yes, 2 = no
Ml 2a (explain) If other, explain:
[GotoFl]
Factors — ( Non-manipulatable Variables)
Fl (nm factor?) Were any nonmanipulatable factors
examined as covariates?
[If yes, go to F2; if no go to SI]
F2 (gender)
F3 (aptitude)
F4 (race/ethic origin)
F5 (nationality)
1 = yes, 2 = no
1 = yes, 2 = no
1 = yes, 2 = no
1 = yes, 2 = no
1 = yes, 2 = no
241
F6 (disability) 1 = yes, 2 = no
F7 (SES) 1 = yes, 2 = no
F8 (other) 1 = yes, 2 = no
F8a (explain) If F8, then explain:
[Go to SI]
Statistical Practices
51. (quant) Were quantitative results reported? 1 = yes, 2 = no
[If yes, go to S2; if no end.]
52. (infstats) Were inferential statistics used? 1 = yes, 2 = no
[If yes, go to S3; Else go to S8]]
53 (parametric) Parametric test of location used? 1 = yes, 2 = no
[Is yes, go to s3a; else go to s4]
S3a (means) Were cell means and cell variances
or cell means, mean square error
and degrees of freedom reported? 1 = yes, 2 = no
54 (multi) Were multivariate analyses used? 1 = yes, 2 = no
[Is yes, go to s4a; else go to s5]
S4a (means) Were cell means reported? 1 = yes, 2 = no
242
S4b (sizes) Were cell sample sizes reported?
S4c (variance) Was pooled within variance or
covariance matrix reported?
1 = yes, 2 = no
1 = yes, 2 = no
55 (correlational) Were correlational analyses done?
[Is yes, go to s5a; else go to s6]
S5a (size) Was sample size reported?
S5b (matrix) Was variance - covariance,
or correlation matrix reported ?
56 (nonparametric) Were nonparametric analyses used?
[Is yes, go to s6a; else go to s7]
S6a (raw data) Were raw data summarized?
57 (small sample) Were analyses for very small samples done?
[Is yes, go to s7a; else go to s8]
S7a (entire data set) Was entire data set reported?
58 (effect size) Was an effect size reported?
[If yes, go to S8a, else end.]
S8a (raw diff) Was there a difference in
means, proportions, medians, etc., reported?
1 = yes, 2 = no
1 = yes, 2 = no
1 = yes, 2 = no
1 = yes, 2 no
1 = yes, 2 = no
1 = yes, 2 = no
1 = yes, 2 = no
1 = yes, 2 = no
1 = yes, 2 = no
S8aa (variability) Was a measure of dispersion reported if 1 = yes, 2 = no
a mean was reported? If a mean was not reported, then -9
243
S8b (SMD) Standardized mean difference effect size
S8c (Corr.) Correlational effect size
S8d (OR) Odds ratios
S8e (odds) Odds
S8f (RR) Relative risk
S8h (other) Other
S8i (explain) Explain other
yes, 2 = no
yes, 2 = no
yes, 2 = no
yes, 2 = no
yes, 2 = no
yes, 2 = no
[end]
244
This page intentionally left blank.
245
Appendix C:
Methodological Review Coding Book
Note: Unless other wise specified, every cell of the coding datasheet must be filled in.
Use -9 to specify that a variable is not applicable. Do not leave cells blank.
DEMOGRAPHIC CHARACTERISTICS
In the variables in this section, the demographic characteristics of each study are coded.
DEO. (case) This is the case number. It will be assigned by the primary coder.
DEOO. (category) This variable corresponds with the first two digits of the case number. It
refers to Table 5; the letter corresponds with the row (forum) and the number corresponds
with the year.
DE000. (kappa) This specifies if this case was used for interrater reliability estimates. 1 =
yes, 2 = no.
DEI. (reviewer) Circle the number that corresponds with your name. If your name is not
on the list, choose other and write in your name. (Choose one.)
DE2. (forum) Circle the number of the forum in which the article was published.
(SIGCSE = SIGCSE technical symposium, Bulletin = June or December issue of SIGCSE
Bulletin, ITiCSE = Innovation and Technology in Computer Science Education
Conference, CSE = Computer Science Education, ICER = International Computer
Science Education Research Workshop, JCSE = Journal of Computer Science Education
Online, ACE = Australasian Computing Education Conference.) (Choose one.)
DE2a. (type of forum). Choose 1 if the forum where the article was published is a journal
(i.e., if the article was not meant to be presented at a conference and published in a peer-
reviewed forum, or if the title of the forum includes the term journal.). Choose 2 if the
forum where the article was published is a conference proceeding (i.e., it was meant to be
published at a conference and may or may not have been peer-reviewed.) In this case,
choose 1 if the article was published in the June or December issues of SIGCSE Bulletin,
Computer Science Education, or the Journal of Computer Science Education Online,
otherwise choose 2.
DE3. (year) Write in the year in which the article was published. 0=2000, 1=2001,
2=2002, 3=2003, 4=2004, 5=2005.
246
DE4. (volume) Write in the volume in which the article was published. Use three digits
(e.g., volume 5 = 005.) If there was not a volume number, write in 000.
DE5. (issue) Write in the issue in which the article was published. Use two digits (e.g.,
issue 2 = 02.) If there was not an issue number, write in 00.
DE6. (page) Write in the page on which the article began. Use four digits (e.g., if the
article began on page 347 = 0347.) If there was not a page number, write in 0000.
DE6a. (pages) Write in how many pages long the article was. If the article had no page
numbers write in -9.
DE7. (region) Choose the region of origin of the first author's affiliation. Choose only
one. If the regions of first author's affiliation cannot be determined, use 7 (IMPDET =
impossible to determine). (This variable was derived from previous the methodological
reviews: Randolph [2005, in press], Randolph, Bednarik, & Myller [2005] and Randolph,
Bednarik, Silander, Lopez-Gonzales, Myller, & Sutinen [2005])
DE7a. (university) Write in the name of the university or affiliation of the first author.
DE7b. (authors) Write in the number of authors.
DE7c. (name) Write in the name of the first author. Last name first and then initials,
which are followed by a period (e.g. Justus Joseph Randolph = Randolph, J. J.). Use a
hyphen if a name is hyphenated (Randolph-Ratilainen), but do not use special characters.
TYPE OF PAPER
These variables group the papers into papers that did research on human participants and
those that did not. For those that did not, they are further classified.
DE8. Subject of study. (This variable comes from a review of the subject matter
discussed in SIGCSE Bulletin articles 1990-2004 [Kinnunen, n.d.]. They were derived
using a emergent approach. Quotes are from Kinnunen, n.d.) Only choose one. If an
article could belong to more than one category, choose the category that the article
discusses the most. 'Tool' articles supersede 'new ways to teach a course,' when the new
was to teach a course includes using a new tool.
• Choose 1 if the subject of the study involved new ways to organize a course. For
example some courses might include "single new assignments" or "more drastic
changes in the course." An example is Mattis (1995).
247
• Choose 2 if the article discusses "a new tool or experiences using a new tool." An
example of a tool article is Dawson-Howe (1995)
• Choose 3 if the article discusses teaching programming languages. This includes
articles that discuss "which language is best for students as a first language and
papers that discuss about how some smaller section of a language should be
taught." An example of this type of paper is Cole (1990).
• Choose 4 if the articles discusses the CSE curriculum. These types of articles
"mainly present a new curriculum in their institution and elaborate on teachers
and students' experiences." An example of this type of article is Garland (1994).
• Choose 5 if the article discusses program visualization.
• Choose 6 if the article discusses simulation.
• Choose 7 if the article discusses parallel computing, (e.g., Schaller & Kitchen,
1995).
• Choose 8 if none of the categories above apply.
DE8a. This variable is from Valentine's (2004) methodological review. (The quotes are
all from Valentine.) Choose only one category, from the categories listed below.
1= Experimental:
If the author made any attempt at assessing the "treatment" with some scientific
analysis, I counted it as an "Experimental" presentation. . . . Please note that this
was a preemptive category, so if the presentation fit here and somewhere else (e.g.
a quantified assessment of some new Tool), it was placed here. (p. 256)
Note if experimental was selected on DE8a, then DE9 should be yes and DE9a should be
no. If DE9a (anecdotal) was yes, then DE9 should be something other than experimental
— the assumption being that informal anecdotal accounts are not appropriate empirical
analyses.
2. Marco Polo
The second category is what has been called by others "Marco Polo"
presentations: "I went there and I saw this." SIGCSE veterans recognize this as a
staple at the Symposium. Colleagues describe how their institution has tried a new
curriculum, adopted a new language or put up a new course. The reasoning is
defined, the component parts are explained, and then (and this is the giveaway for
248
this category) a conclusion is drawn like "Overall, I believe the [topic] has been a
big success." or "Students seemed to really enjoy the new [topic]", (p. 256)
3. Tools
Next there was a large collection of presentations that I classified "Tools". Among
many other things, colleagues have developed software to animate algorithms, to
help grade student programs, to teach recursion, and to provide introductory
development platforms, (p. 257)
4. John Henry
The last, and (happily) the smallest category of presentations would be "John
Henry" papers. Every now and then a colleague will describe a course that seems
so outrageously difficult (in my opinion), that one suspects it is telling us more
about the author than it is about the pedagogy of the class. To give a silly
example, I suppose you could teach CS1 as a predicate logic course in IBM 360
assembler - but why would you want to do that? (p. 257)
5. Philosophy
A third classification would be "Philosophy" where the author has made an
attempt to generate debate of an issue, on philosophical grounds, among the
broader community, (p. 257)
6. Nifty
The most whimsical category would be called "Nifty", taken from the panels that
are now a fixed feature of the TSP. Nifty assignments, projects, puzzles, games
and paradigms are the bubbles in the champagne of SIGCSE. Most of us seem to
appreciate innovative, interesting ways to teach students our abstract concepts.
Sometimes the difference between Nifty and Tools was fuzzy, but generally a
Tool would be used over the course of a semester, and a Nifty assignment was
more limited in duration, (p. 257)
DE9. (human participants) Choose yes if the article reported direct research done on
human participants - even if the reporting was anecdotal. Choose no if the authors did not
report doing research on human participants. For example, if the author wrote, "the
participants reported that they liked using the Jeliot program," then yes should be chosen.
But, if the author wrote, "in other articles, people have reported that they enjoyed using
the Jeliot program," choose no since the research was not done by directly by the author.
(If yes go directly to DE9a. If no go to A9.)
249
DE9a. (anecdotal). Choose this if the article reported on investigations on human
participants, but only provided anecdotal information. If yes on DE9 and DE9a, end. If
no, on DE9a then go to Al 1 and mark A9 and A10 as -9. This might include studies that
the author purported to be a 'qualitative study,' but mark anecdotal if there was not
evidence that a qualitative methodology was used and the authors were just informally
reporting their personal observations.
A9. (type of other) If the article did not report research on human participants, classify the
type of article that it was. Choose 1 - literature review if the article was primarily a
literature review, meta-analysis, methodological review, review of websites, review of
programs, etc. Choose 2 -program description if the article primarily described a
program/software/intervention and did not have even an anecdotal evaluation section.
Choose 3 — theory, methodology, or philosophy if the paper was primarily a theoretical
paper or discussed methodology or philosophical issues, policies, etc. For example, an
article that discussed how constructivism was important for computer science education
would go into this (3) category. Choose 4 - technical if the article was primarily a
technical computer science paper. For example, an article would go into this category if it
compared the speed of two algorithms. Finally, choose the (5) other category if the article
did not fit into any of the categories above. Use category 5 as a last resort. (If categories
1,2 3, or 4, are chosen go to Al 1. Otherwise go to A10.) (Choose only one.) (This
variable was derived from previous the methodological reviews: Randolph [in press],
Randolph, Bednarik, & Myller [2005]; and Randolph, Bednarik, Silander, et al, [2005].)
A10. (other other) If you chose category 5 on variable A9, please write a description of
the paper and describe what type of paper you think that it is.
REPORT STRUCTURE
Mn this section, which is based on the structure suggested for empirical papers by the
APA publication manual (2001, Parts of a Manuscript, pp. 10-30), you will examine the
structure of the report. Filling out the report structure is not necessary if it was an
explanatory descriptive study, since this report structure does not necessarily apply to
qualitative (explanatory descriptive) reports.
Al 1. (abstract) Choose 1 - narrative if the abstract was a short (150-250) narrative
description of the article. Choose 2 - structured if the abstract was long (450 words) and
was clearly broken up into sections. Some of the abstract section headings you might see
are 'background,' 'purpose,' 'research questions,' 'participants,' 'design,' 'procedure,'
etc. A structured abstract does not necessarily have to have these headings, but it does
have to be broken up into sections. Choose 3 -no abstract if there was not an abstract for
the paper.
250
A12. (introduce problem) choose 1 -yes if the paper had even a brief section that
described the background/need/context/problem of the article. Choose 2- no if there was
not a section that put the article in context, described the background, or explained the
importance of the subject. For example, you should choose yes if an article on gender
differences in computing began with a discussion of the gender imbalance in computer
science and engineering.
A13. (literature review) Choose 1 -yes if the author at least mentioned one piece of
previous research on the same topic or a closely related topic. Choose 2- no if the author
did not discuss previous research on the same or a closely related topic.
A14. (purpose/rationale) Choose 1 -yes if the author explicitly mentioned why the
research had been done or how the problem will be solved by the research. Choose 2- no
if the author did not give a rationale for carrying out the study.
A15. (research questions/hypotheses.) Choose 1 — yes if the author explicitly stated the
research questions or hypotheses of the paper. Choose 2- no if the author did not
explicitly state the research questions or hypotheses of the paper.
A16. (participants.) Choose 1 -yes if the author made any attempt at describing the
demographic characteristics of the participants in the study. Choose 2- no if the author
did not describe any of the characteristics of the participants in the study. (Choose 2 if the
author only described how many participants were in the study.) If yes go to A 16a. If no
go to A17 and mark -9 in A16a and A16b. Please note that this refers to the participants
that were used in the evaluation of the section, not about participants who participated in
the program in general. If they did not describe the participants in the study, you do not
have to go to a 16a and a 17a.
A16a. (grade level). Categorize articles based on the grade levels of the participants
participating in the program. If ages, but grades were not given, use the age references
below. (Grades take precedent over age when there is a conflict.) If 6, go to A16b; else go
to A17 and mark -9 in A16b.
• Choose 1 if the students were in pre-school (less than 6 years old).
Choose 2 if the participants were in grades Kindergarten to 3 (Ages 6-9).
Choose 3 if the participants were in grades 4 through 6 (ages 10- 12).
Choose 4 if the participants were in grades 7-9 (ages 13-15).
Choose 5 if the participants were in grades 10-12 (ages 16-18).
•
•
•
251
• Choose 6 if the participants were undergraduates (bachelor's level) (18-22 years
old).
• Choose 7 if the participants were studying at the graduate level (master's students)
(23-24 years old).
• Choose 8 if the students were post-graduate students (doctoral students) (25-30
years old).
• Choose 9 if the students were post-doctoral students (31 and over years old).
• Choose 10 if more than one category applies or if the category that is appropriate
is not listed here.
• Choose 11 if it is impossible to determine the grade level of the participants.
A16b. (curriculum year). If 6 in A16b, choose the year (1-4) of the corresponding
undergraduate computing curriculum that the article dealt with.
A17. (setting) Choose 1 -yes if the author made any attempt at describing the setting
where the investigation occurred. Setting includes characteristics such as type of course,
environment, type of institution, etc. Choose 2- no if the author did not describe the
setting of the study. This might include a description of participants who usually attended
a course or a description of the organization that the author was affiliated with.
A18. (instruments) Choose 1 -yes if special instruments were used to conduct the study
and they were described. (For example, if a piece of software was used to measure
student responses, then choose 1 if the software was described.) Choose 2 -no if special
instruments were used, but they were not described. Choose -9 - n/a (not applicable) if
no special instruments were used in the study.
A19. (procedure). Choose 1 -yes if the author described the procedures in enough detail
that the procedure could be replicated. (If an experiment was conducted, choose yes only
if both the control and treatment procedures were described.) Choose 2- no if the author
did not describe the procedures in enough detail that the procedure could be replicated.
For example, if the author only wrote, "we had students use our program and found that
they were pleased with its usability," then the procedure was clearly not described in
enough detail to be replicated and 2 (no) should be chosen.
A20. (results and discussion). Choose 1 -yes if there was a section/paragraph of the
article that dealt solely with results. Choose 2- no if there was not a section/paragraph
just for reporting results. For example, choose 2 (no) if the results were dispersed
throughout the procedure, discussion, and conclusion sections.
252
METHODOLOGY TYPE
In this section you will code for the type of methodology that was used. Since articles can
report multiple methods, you can choose all that apply. (These methodology types were
initially developed from Gall, Borg, and Gall (1996) and from the American
Psychological Association's publication manual (2001, pp. 7-8). Explanatory descriptive
and exploratory descriptive labels came from Yin (1988). The descriptions of variables
listed below evolved into their current from Randolph (2005, in press); Randolph,
Bednarik, and Myller (2005); and Randolph, Bednarik, Silander, et al. (2005).
M21. (experimental/quasi-experimental) If the researcher manipulated a variable and
compared a factual and counterfactual condition, the case should be deemed as
experimental or quasi-experimental. For example, if a researcher developed an
intervention then measured achievement before and after the intervention was delivered,
then an experimental or quasi-experimental methodology was used. Choose 1 -yes if the
study used an experimental or quasi-experimental methodology. Choose 2 - no if the
study did not use an experimental or quasi-experimental methodology. Note if the author
did a one-group posttest-only or retrospective posttest on an intervention that the
researcher implemented, choose experimental/quasi-experimental. The posttest in this
case might be disguised by the term 'survey.'
AS5. (assignment) Use 1 when participants knowingly self-selected into treatment and
control groups or when the participants decided the order of treatment and controls
themselves. Use 2 when participants or treatment and control conditions were assigned
randomly. (Also use 2 for an alternating treatment design.) Use 3 when the researcher
purposively assigned participants to treatment and control conditions or the order of
treatment and control conditions or in designs where participants served as their own
controls. Also use 3 when assignment was done by convenience or in existing groups.
This variable originally was based on Shadish, Cook, and Campbell's (2002) distinction
between experimental and quasi-experimental designs. They have been pilot tested in
Randolph (2005, in press); Randolph, Bednarik, and Myller (2005); and Randolph,
Bednarik, Silander, et al. (2005).
M22. (explanatory descriptive) Studies that provided deductive answers to "how"
questions by explaining the causal relationships involved in a phenomenon should be
deemed as explanatory descriptive. Studies using qualitative methods often fall into this
category. For example, if a researcher did in-depth interviews to determine the process
that expert programmers go through when debugging a piece of software, this should be
considered a study in which an explanatory descriptive methodology was used. Choose 1
- yes if the study used an explanatory descriptive methodology and choose 2 -no if it did
not. This does not include content analysis, where the researcher simply quantifies
qualitative data (e.g., the researcher classifies qualitative data into categories, then
presents the distribution of units into categories.)
253
M23. (exploratory descriptive) Studies that answered "what" or "how much" questions
but did not make any causal claims used an exploratory descriptive methodology. Pure
survey research is perhaps the most typical example of the exploratory descriptive
category, but certain kinds of case studies might qualify as exploratory descriptive
research as well. Choose 1 -yes if the study used an exploratory descriptive methodology
and choose 2 -no if it did not. Note: If the author gave a survey to the participants and
the investigation did not examine the implementation of an intervention, then you should
consider that to be exploratory descriptive survey research.
M24. (correlational) A study should be categorized as correlational if it analyzed how
continuous levels of one variable systematically covaried with continuous levels of
another variable. Studies that conducted correlational analyses, structural equation
modeling studies, factor analyses, cluster analyses, and multiple regression analyses are
examples of correlational methods. Choose 1 -yes if the study used an correlational
methodology and choose 2 -no if it did not.
M25. (causal-comparative) If researchers compared two or more groups on an inherent
variable, an article should be coded as causal-comparative. For example, if a researcher
had compared computer science achievement between boys and girls, that case would
have been classified as casual-comparative because gender is a variable that is inherent in
the group and cannot be naturally manipulated by the researcher. Choose 1 -yes if the
study used a correlational methodology and choose 2 - no if it did not.
M26. (IMPDET). Use this if not enough information was given to determine what type of
methodology(ies) were used. If M26 was yes, then end.
Examples. A researcher used a group repeated measures design with one -between factor
(gender) and two-within factors (measures, treatment condition). That investigation
should be coded as an experiment because the researcher manipulated a variable and
compared factual and counterfactual conditions (the treatment-condition within factor).
The investigation should also be classified as a causal-comparative study because of the
between factor in which two levels of a non-manipulatable variable were compared. Had
the researcher not examined the gender variable, this investigation would have only been
classified as an experiment/quasi-experiment.
A researcher did a regression analysis and regressed the number of hours using Jeliot (a
computer education piece of software) on a test of computer science achievement. In
addition, the researcher also examined a dummy variable where Jeliot was used with and
without audio feedback. Because of the multiple regression, the investigation should be
classified as correlational. Because of the manipulatable dummy variable, the
investigation should also be classified as an experimental or quasi-experimental design.
254
A researcher gave only a posttest survey to a class after they used the intervention that a
researcher had assigned. The researcher claimed that 60% of the class, after using the
intervention, had exhibited mastery on the posttest. Since the researcher claimed that
60% of the class had exhibited mastery on the posttest because of the intervention, then
the investigation should be classified as an experiment or quasi-experiment (in M21) that
used a one-group posttest-only research design (RD2). (Had the researcher did a survey,
but not measured the effects of an intervention, then it would have just been exploratory
descriptive and not a one-group posttest-only experiment.)
[Go to M27 if M21, M23, M24, or M25 = 1. Else end.]
M27. (selection) Choose 1 (random) if the sampling units were randomly selected.
Choose 2 (purposive) if the participants were purposively selected. (For example, if the
researcher chose to examine only extreme cases, this would be purposive selection.)
Choose 3 if the research chose a convenience sample or existing group. Choose 3 unless
there is evidence for random or purposive sampling.
EXPERIMENTAL RESEARCH DESIGNS
If an experimental / quasi-experimental methodology was used, classify the methodology
into research design types. Choose 1 for yes and 2 for no. If no go to li and mark the rest
of the variables in this section as -9. These designs were originally based on the
descriptions of designs in Shadish, Cook, and Campbell (2002) and in American
Psychological Association (2001, pp. 23-24). They had been previously pilot tested in
Randolph (2005, in press); Randolph, Bednarik, and Myller (2005); and Randolph,
Bednarik, Silander, et al. (2005), except for the multiple factor category.
RD1. (designs) Choose 1 if M21 was marked as yes. If so, one of the following variables
must be coded as a yes. If no, mark -9 in all of the following RD variables.
RDla. (design?) Choose 1 if RD1 was marked yes but it could not be determined what
research design was used. Choose no if the design could be determined and go on to RD2.
If yes, go II.
RD2. (post-only) Use this for the one-group posttest-only design. In the one-group
posttest-only design, the researcher only gives a posttest to a single group and tries to
make causal claims. (In this design the observed mean might be compared to an expected
mean.) This includes retrospective posttests, in which participants estimate impact
between counterfactual and factual conditions.
RD3. (post controls) Use this if the posttest with controls design was used. In the posttest
with controls design the researcher only gives a posttest to both a control and treatment
255
group. Put the regression-discontinuity design into this category too and regressions with
a dummy treatment variable into this design. (The independent T-test, regression with a
dummy variable, or univariate ANOVA analyses might be used with this research
design.)
RD4. (prepost only) Use this for the pretest/posttest without controls design. In
pretest/posttest without controls design the researcher gives a pretest and posttest to only
a treatment group. (Dependent T-tests might be used in this design.)
RD5. (prepost controls) Use this for the pretest/posttest with controls design. In the
pretest/posttest with controls design the researcher gives a pretest and posttest to both a
treatment and one or more control groups. (Independent T-tests of gain scores or
ANCOVA might be used on these designs)
RD6. (repeated) Use this for repeated measures designs. In the group repeated measures
design, the researchers use participants as their own controls and are measured over
multiple points of time or levels of treatment. (Repeated measures analysis might be used
in this design.)
RD7. (multiple) Use this for designs with multiple factors that examine interactions. If
only main effects are examined, code the research design as a control group design (like
the case in a one-way anova.)
RD8. (single) Use this for single-subject designs. In this design, a researcher uses the
logic of the repeated measures design, but only examines a few cases. (Single-case
interrupted time series designs apply to this category.)
RD9. (IMPDET) Use this if the author did not give enough information to determine
what type of experimental research design was used.
RD10. (other) Use this category if the research design was well explained but were not
RD2-RD8.
RDH. (posttest only highest) Choose 1 if the only research design was the one-group
posttest-only design (i.e., if RD2 was marked yes, and RD3 through RD10 were marked
no), otherwise mark no. This construct behind this variable is whether a researcher
compared a factual with a counterfactual occurence. It assumes here that the one-group
posttest-only design does not compare a factual with a counterfactual condition.
[Go to Ii -measures.]
256
INTERVENTION (independent variable)
For this group of variables, choose 1 -yes if the listed intervention was used in the article
and choose 2- no if the intervention was not used. Choose all that apply. These
intervention codes were based on codes that emerged in the previous methodological
reviews: Randolph, (2005) and Randolph, Bednarik, and Myller (2005).
11. (intervention) Choose 1 — yes if an intervention was used in this investigation.
Choose 2 - no if an intervention was not used. There might be an intervention in an
experimental/quasi-experimental study or in an explanatory descriptive study. But, there
would not be an intervention in a causal-comparative study, since it is examines variables
not manipulated by the researcher. Also, there would not be an intervention in an
exploratory descriptive study (e.g., survey study) since exploratory descriptive research is
described here as research on a variable that is not manipulated by the researcher.
[If II = 1, go to 12, else go to Dl and mark all I variables as -9.]
12. (student instruction) Choose yes if students were given instruction in computer science
by a human or by a computerized-tool. Otherwise, choose no.
13. (teacher instruction) Choose yes if teachers were instructed on the pedagogy of
computer science. Otherwise, choose no.
14. (CS fair/contests) Choose yes if students participated in a computer science fair or
programming contest. Otherwise, choose no.
15. (mentoring) Choose yes if students were assigned to a computer science mentor.
Otherwise, choose no.
16. (speakers) Choose yes if students listened to speakers who are computer scientists.
Otherwise, choose no.
17. (CS field trips) Choose yes if students took a field trip to a computer-science-related
site. Otherwise, choose no.
18. (other) Choose yes if an intervention other than the one mentioned here was examined.
Otherwise, choose no.
DEPENDENT VARIABLES
In this section you code the dependent variables outcomes that were examined. Choose 1
for yes and 2 for no. Choose all that apply. These dependent variables codes were based
257
on codes that emerged in the previous methodological reviews: Randolph, 2005;
Randolph, Bednarik, and Myller (2005).
Dl. (attitudes) Choose yes if student attitudes (including satisfaction, self-reports of
learning, motivation, confidence, etc.) were measured. Otherwise, choose no.
D2. (attendance) Choose yes if student attendance or enrollment in a program, including
attrition, was measured. Otherwise, choose no.
D3. (core achievement) Choose yes if achievement in core courses, but not achievement
in computer science was measured. Otherwise, choose no.
D4. (CS achievement) Choose yes if achievement in computer science was measured —
this includes CS test scores, quizzes, assignments, and number of assignments completed.
Otherwise, choose no.
D5. (teaching practices) Choose yes if teaching practices were measured. Otherwise,
choose no.
D6. (intentions for future) Choose yes if what courses, fields of study, careers, etc, that
students planned to take in the future were measured. Otherwise, choose no.
D7. (program implementation) Choose yes if how well a program / intervention was
implemented as planned (i.e., treatment fidelity) was measured. Otherwise, choose no.
D8. (costs) Choose yes if how much a certain intervention/policy/program costed was
measured. Otherwise, choose no.
D9. (socialization) Choose yes if how much students socialized with each other or with
the teacher was measured. Otherwise, choose no.
D10. (computer use) Choose yes if how much or how students used computers was
measured. Otherwise, choose no.
Dl 1. (other) Use this category for dependent variables that are not included above.
Otherwise, choose no.
Dl la. (describe) Please describe the intervention if it was 'other.'
258
MEASURES
In this section you will code what kinds of measures were used to measure the dependent
variables. For some measures you will note if psychometric information, operationalized
as the author making any attempt at reporting information about the reliability or validity
of a measure. Choose 1 for yes and 2 for no. These measures codes were based on codes
that emerged in the previous methodological reviews: Randolph (2005) and Randolph,
Bednarik, and Myller (2005). For subquestions, if the head question was yes, then the
subquestion must be either yes or no. If the head question was no, then the subquestion
must be -9. For example, if M3 was yes, M3a must either bo, yes or no. If M3 was no, then
M3a must be -9.
Ml. (grades) Choose yes if grades in a computer science class - or overall grades (like
GPA) — were a measure. Otherwise, choose no.
M2. (diary) Choose yes if a learning diary was a measure. Otherwise, choose no.
M3. (questionnaire) Choose yes if a questionnaire or survey was a measure — this
includes quantitative questionnaires that had open elements. However, if a survey had all
open questions, call it an interview (m6). Otherwise, choose no.
M3a. (ques. Psych.) Choose yes if psychometric information was given about the
survey or questionnaire. Otherwise, choose no.
M4. (log files) Choose yes if computerized log files of students' behaviors when using
computers was a measure. Otherwise, choose no.
M5. (test) Choose yes if teacher-made or researcher-made tests or quizzes were measures.
Otherwise, choose no.
M5a. (test psych) Choose yes if psychometric information was given about the test
or quiz. Otherwise, choose no.
M6. (interviews) Choose yes if interviews with students or teachers was used as a
measure — this also includes written interviews or reflection essays. Otherwise, choose
no.
M7. (direct observation) Choose yes if researchers observed strictly operationalized
behaviors. Otherwise, choose no.
M7a. (direct psych) Choose yes if reliability information (e.g., interrater
agreement) was given about the direct observation. Otherwise, choose no.
259
M8. (stand, test). Choose yes if a standardized test (in core subjects or computer science)
was a measure. Otherwise, choose no.
M8a. (psych, stand) Choose yes if psychometric information was provided for each
standardized test. Otherwise, choose no.
M9. (student work) Choose yes if exercises/assignments in computer science was a
measure - this might include portfolio work. This does not include work on tests, grades,
or standardized tests. Otherwise, choose no.
M10. (focus groups) Choose yes if focus groups, swot analysis, or the Delphi technique
were used as measures. Otherwise, choose no.
Ml 1. (existing records) Choose yes if records such as attendance data, school history, etc
were used as measures. This does not include log files. Otherwise, choose no.
M12. (other) Choose yes if there were measures that were not included above. Otherwise,
choose no.
Ml 2a. (explain other) Explain what the other measure was, if there was one. Otherwise,
choose no.
[gotoFl.]
FACTORS (non-manipulatable variables)
In this section you will examine the factors or nonmanipulatable variables that were
examined. (If they were manipulatable - they should be mentioned as an intervention.)
Choose 1 for yes and 2 for no. These factors codes were based on codes that emerged in
the previous methodological reviews: Randolph, (2005) and Randolph, Bednarik, and
Myller (2005).
Fl. (factors) Choose yes if any nonmanipulatable factors examined. [If .yes , go to F2; else
SI and F2-F8 are -9.] Otherwise, choose no.
F2. (gender) Choose yes if gender of the students or the teacher was used as a factor.
Otherwise, choose no.
F3. (aptitudes) Choose yes, for example, if the researcher made a distinction between high
and low achieving students. Otherwise, choose no.
260
F4. (race/ethnic origin) Choose yes if race/ethnic origin of participants was used as a
factor. Otherwise, choose no.
F5. (nationality) Choose yes if nationality/geographic reason/ or country of origin was
used as a factor. Otherwise, choose no.
F6. (disability) Choose yes if disability status of participants was used as a factor.
Otherwise, choose no.
F7. (SES) Choose yes if the socio-economic status of students was used as a factor.
Otherwise, choose no.
F8. (other) Use yes if a factor was examined that was not listed above. Otherwise, choose
no.
F8a. (explain other). Explain what the factor was if F8 was marked as yes. Otherwise,
choose no.
[Go to SI]
STATISTICAL PRACTICES
In this section you will code for the statistical practices used. Choose 1 for yes and 2 for
no. You can check all that apply. These categories come from the Informationally
Adequate Atatis tics section of AP A publication manual (2001, pp. 23-24))
51. (quant results) Choose yes if quantitative results were reported. Otherwise, choose
no.
[If yes, go to S2; Else end and all following S2-S7 are -9.]
52. (inf. stats) Choose yes if inferential statistics was used. [Ifjes, go to S3, Else go S8
and S3-S7 are -9)] If yes, head questions must be yes or no. If the head question was yes,
then the subquestion(s) must be yes or no. If the head question was no, then subquestions
should be marked -9.
53. (parametric) Choose yes if a parametric test of location was used. — "e.g., single-
group, multiple-group, or multiple-factor tests of means" APA [2001], p. 23. [Ifjes, go to
S3a, else go to S4]
S3a. (means) Choose yes if either cell means and (cell sizes) were reported or if
means cell variances or mean square error and degrees of freedom were reported.
Otherwise, choose no.
261
54. (multi) Choose yes if multivariate types of analyses were used. Otherwise, choose no.
[If S4 if 1, go to S4a; else go to S5]
S4a. (means) Choose yes if cell means were reported. Otherwise, choose no.
S4b. (size) Choose yes if sample sizes were reported. Otherwise, choose no.
S4c. (variance) Choose yes if pooled within variance or a covariance matrix was
reported. Otherwise, choose no.
55. (correlational analyses). Choose yes if correlational analyses were done. — "e.g.,
multiple regression analyses, factor analysis, and structural equation modeling" APA
(2001, p. 23.) Otherwise, choose no. [Ifyes, go to S5a; else go to S6]
S5a. (size) Choose yes if sample size was reported. Otherwise, choose no.
S5b. (matrix) Choose yes if a variance-covariance or correlation matrix was
reported. Otherwise, choose no.
56. (nonparametric) Choose yes if nonparametric analyses were used. Otherwise, choose
no.
[If yes, go to S6a; else go to S7]
S6a (raw data) Choose yes if raw data were summarized. Otherwise, choose no.
57. (small samples) Choose yes if analyses for small samples was done. Otherwise,
choose no.
[If yes, go to S7a; else go to S8]
S7a. (entire data set) Choose yes if the entire data set was reported. Otherwise,
choose no.
58. (effect size) Choose yes if an effect size was reported Otherwise, choose no.
[If yes, go to S8a, else end.]
S8a. (raw diff) Choose yes if there wasa difference in means, proportions,
medians reported. Otherwise, choose no. (Here authors just needed to present two or
more means or proportions. They did not actually have to subtract one from the other.
This is also includes what is called 'risk difference.')
262
S8aa. (variability) Choose yes if a mean was reported and if had a standard
deviation reported? If a median was reported, choose yes if a range was also reported.
Otherwise, choose no, unless a mean or median was not reported, then use -9 here.
S8b. (SMD) Choose yes if a standardized mean difference effect size was
reported. Otherwise, choose no.
S8c. (Corr.) Choose yes if a correlational effect size was reported. Otherwise,
choose no.
S8d. (OR) Choose yes if odds ratios were reported. Otherwise, choose no.
S8e. (odds) Choose yes if odds were reported. Otherwise, choose no.
S8f (RR) Choose yes if relative risk was reported.
S8h. (other) Choose yes if some other type of effect size not listed above was
reported. Otherwise, choose no.
S8i. (explain) If S8 was marked as yes, please explain what the effect size was.
Otherwise, choose no.
263
Coding Book References
American Psychological Association. (2001). Publication manual of the American
Psychological Association (5th ed.). Washington, DC.
Cole, J. P. (1991). WHILE loops and the anology of the single-stroke engine. SIGCSE
Bulletin, 23(3), 20-22.
Dawson-Howe, K. M. (1995) Automatic submission and administration of programming
assignments. SIGCSE Bulletin, 27(4), 51-53.
Gall, M. D., Borg, W. R., & Gall, J. P. (1996). Educational research: An introduction
(6th ed.). New York: Longman.
Garland, W., & Levsen, V. (1994). Information systems curricula in AACSB accredited
business schools. SIGCSE Bulletin, 26(2), 26-30.
Kinnunen, P. (n.d.) Guidelines of Computer Science Education Research. Retrieved
November 29, 2005 from http://www.cs.hut.fi/Research/COMPSER/
ROLEP/seminaari-k05/S_05-nettiin/Guidelines_o f_CSE-teksti-paivi.pdf
Mattis, W. E. (1995). An advanced microprocessor course with a design component.
SIGCSE Bulletin, 27(4), 60-64.
Randolph, J. J. (2005). A methodological review of the program evaluations in K-I2
computer science education. Manuscript submitted for publication.
Randolph, J.J. (in press). What's the difference, still: A follow-up review of the
quantitative research methodology in distance learning. Informatics in Education.
Randolph, J. J., Bednarik, R., & Myller, N. (2005). A methodological review of the
articles published in the proceedings of Koli Calling 2001-2004. In T. Salakoski,
T. Mantyla, & M. Laakso (Eds.), Proceedings of the 5th Annual Finnish /Baltic
Sea Conference on Computer Science Education (pp. 103-109). Finland: Helsinki
University of Technology Press. Retrieved March 19, 2006, from
http://www.it.utu.fi/koli05/proceedings/final_composition.b5.060207.pdf
264
Randolph, J. J., Bednarik, R., Silander, P., Lopez-Gonzalez, J., Myller, N., & Sutinen, E.
(2005). A critical review of research methodologies reported in the full-papers of
ICALT 2004. In Proceedings of the Fifth International Conference on Advanced
Learning Technologies (pp. 10-14). Los Alamitos, CA: IEEE Press. Available
online: http://ieeexplore.ieee. org/xpls/abs_all.jsp?isnumber=323 1 7
&arnumber=1508593&count=303&index=4
Schaller, N. C, & Kitchen, A. T. (1995). Experiences in teaching parallel computing -
Five years later. SIGCSE Bulletin, 27(3), 15-20.
Shadish, W. R., Cook, T. D., & Campbell, D. T. (2002). Experimental and quasi-
experimental designs for generalized causal inference. Boston: Houghton Mifflin.
Valentine, D. W. (2004). CS educational research: A meta-analysis of SIGCSE technical
symposium proceedings. In Proceedings of the 35 th Technical Symposium on
Computer Science Education (pp. 255-259). New York: ACM Press.
Yin, R. K. (1988). Case study research: Designs and methods, (Rev. ed.). London: Sage.
265
Appendix D:
Resampling Program for Calculating Free Marginal
Kappa and Its Confidence Intervals
'RESAMPLING PROGRAM FOR CALCULATING FREE MARGINAL KAPPA AND ITS
CONFIDENCE INTERVALS
'This section of the program, until REPEAT 10000, finds free
marginal kappa given the percent of
'observed agreement and percent of expected agreement.
'The values here are from the variable HUMAN PARTICIPANTS with an
observed agreement .906,
'an expected agreement of .50, and a sample size of 53 where
'48 cases were agreements and 5 were disagreements.
'This is the percent of observed agreement (i.e., proportion of
agreements ) .
DATA 0.90 6 po
'This is the percent expected, which is In, where n is number of
categories
DATA 0.5 pe
'The following three line are the general formula for kappa.
SUBTRACT po pe num
SUBTRACT 1 pe denom
DIVIDE num denom k
'This command prints the value of kappa
PRINT k
'The following section of the program, until END will make a
distribution of 1000 Ks
'This command repeats from the commands between URN and END
10,0000 times.
REPEAT 100 00
'This command creates an urn that represents the population.
'For the urn, the sampled values are multiplied by 7 (an
approximation of 352/52 - the population/sample ratio)
'to simulate the population size.
'In this urn l=yes and 2=no.
URN 336#1 35#2 $sam
266
'The SHUFFLE command randomizes the order of values in the urn.
SHUFFLE $sam $samp
'The TAKE command takes the first 53 values from the shuffled.
TAKE $samp 1,53 $sa
'This COUNT command then counts the number of times that the
sample of 53 had a value of 1.
COUNT $sa=l $yes
'The number of l's is divided y the sample size to arrive at a
percentage of sample agreement.
DIVIDE $yes 53 $po
'The following lines get the value of kappa for the sample.
SUBTRACT $po pe $num
SUBTRACT 1 pe $denom
DIVIDE $num $denom $k
'This command keeps score of the value outside of the loop.
SCORE $k $kappa
END
'This PERCENTILE command ranks the kappa values from each
iteration and finds the given percentiles.
PERCENTILE $kappa (2.5 50 97.5) kappa
'This command prints the percentiles.
PRINT kappa
'Note. The value of kappa for this program was .812 with 2.5, 50,
and 97.5 percentiles of .66, .81, and .96.
267
Appendix E:
Resampling Stats Code for Confidence Intervals Around a
Proportion from a Proportional Stratified Random Sample
'RESAMPLING PROGRAM TO CALCULATE CONFIDENCE INTERVALS AROUND
PROPORTIONS - UP TO 35 STRATA AND VARIABLES WITH 8 LEVELS
'This command reads data from an external data file.
READ file "C : WDocuments and Settings\\localadmin\\My
DocumentsWdissertationWwhole.dat" missing -9 cell deOOO
del de2 de3 de4 de5 de6 de6a de7 de7b de8 de8a de9 de9a a9
all al2 al3 al4 al5 al6 al6a al6b al7 al8 al9 a20 a21 m26
m21 as5 m22 m23 m24 m25 m27 rdl rdla rdh rd2 rd3 rd4 rd5 rd6
rdll rd7 rd8 rd9 il i2 i3 i4 i5a i6 i7 i8 dl d2 d3 d4file d5
d6 d7 d8 d9 dlO dll dl2 ml m2 m3 m3a m4 m5 m5a m6 m7 m7a m8
m8a m9 mlO mil ml2 fl f2 f3 f4 f5 f6 f7 f8 si s2 s3fine s3a
s4 s4a s4b s4c s5 s5a s5b s6 s6a s7 s7a s8 s8a s8aa s8b s8c
s8d s8e s8f s8h var00006 filter journal cse
'The following commands renames a variable and cleans system
missing cases.
DATA al6a var
DATA cell forum
CLEAN forum var
'The following commands count the number of times that a
case occurs in each stratum.
COUNT forum=l a
COUNT forum=2 b
COUNT forum=3 c
COUNT forum=4 d
COUNT forum=5 e
COUNT forum=6 f
COUNT forum=7 g
COUNT forum=8 h
COUNT forum=9 i
COUNT forum=10 j
COUNT forum=ll k
COUNT forum=12 1
COUNT forum=13 m
COUNT forum=14 n
COUNT forum=15 o
268
COUNT
f orum=
= 16
P
COUNT
f orum=
= 17
q
COUNT
f orum=
= 18
r
COUNT
f orum=
= 19
s
COUNT
f orum=
=2
t
COUNT
f orum=
=21
u
COUNT
f orum=
=22
V
COUNT
f orum=
=23
w
COUNT
f orum=
=24
X
COUNT
f orum=
=25
Y
COUNT
f orum=
=26
z
COUNT
f orum=
=27
aa
COUNT
f orum=
=2 8
bb
COUNT
f orum=
=2 9
cc
COUNT
f orum=
=3
dd
COUNT
f orum=
=31
ee
COUNT
f orum=
=32
ff
COUNT
f orum=
=33
gg
COUNT
f orum=
= 34
hh
COUNT
f orum=
=35
ii
'This command calculates the sample size be adding the n
size of each stratum.
ADD abcdefghij klmnopqrstuvwxyZaa
bb cc dd ee ff gg hh ii sampsize
'This command creates a range of values that correspond with
the n size of the strata.
'For example stratum b contains the values of the vector var
from a+1 to a+b.
'If the n size of stratum a is 5 and the n size of stratum b
is 6 then the values of vector var that . . .
' correspond with a are 1-5 and for b are 6-11 (a+l=6 and
a+b=ll) .
ADD
a
l
b
_b
ADD
a
b
b"
e
ADD
b
e
1
c
_b
ADD
b"
e
c
c
e
ADD
c
e
1
d"
_b
ADD
c
e
d
d~
e
ADD
d~
e
1
e
_b
ADD
d~
e
e
e
e
ADD
e
e
1
f"
Jo
ADD
e
e
f
f~
e
ADD
f"
e
1
g_
_b
ADD
f"
e
g
g
e
ADD
g_
e
l
h
_b
269
ADD
g e
h
h e
ADD
h e
1
i_b
ADD
h_e
i
i e
ADD
i e
1
j_b
ADD
i e
J
J e
ADD
j_ e
1
k b
ADD
J e
k
k e
ADD
k e
1
L b
ADD
k e
L
L_e
ADD
L e
1
m b
ADD
L_e
m
m e
ADD
m e
1
n b
ADD
m e
n
n e
ADD
n e
1
o b
ADD
n e
o
o e
ADD
o e
1
p b
ADD
o e
P
p e
ADD
P e
1
q b
ADD
p e
q
q e
ADD
q e
l
r b
ADD
q e
r
r e
ADD
r e
1
s b
ADD
r e
s
s e
ADD
s e
1
t b
ADD
s e
t
t_e
ADD
t e
1
u b
ADD
t_e
u
u e
ADD
u e
1
v b
ADD
u e
V
v e
ADD
v e
1
w b
ADD
v e
w
w e
ADD
w e
1
x b
ADD
w e
X
x e
ADD
x e
1
y b
ADD
x e
Y
y e
ADD
y_e
1
z_b
ADD
y_e
z
z e
ADD
z e
1
aa b
ADD
z e
33 33 G
ADD
aa e '.
L bb b
ADD
aa e bb bb e
ADD
bb e '.
L cc b
ADD
bb e cc cc e
ADD
cc e '.
L dd b
ADD
cc e dd dd e
ADD
dd e '.
L ee b
ADD
dd e ee ee e
ADD
ee e '.
L ff b
270
ADD ee_e ff ff_e
ADD ff_e 1 gg_b
ADD ff_e gg gg_e
ADD gg_e 1 hh_b
ADD gg_e hh hh_e
ADD hh_e 1 ii_b
ADD hh_e ii ii_e
'The following commands take the values of vector var and
breaks them into smaller vectors that. . .
' correspond with each stratum, if there n size in the
stratum is greater than zero.
IF a>0
TAKE var l,a al
END
IF b>0
TAKE var b_b,b_e a2
END
IF c>0
TAKE var c_b,c_e a3
END
IF d>0
TAKE var d_b,d_e a4
END
IF e>0
TAKE var e_b,e_e a5
END
IF f>0
TAKE var f_b, f_e a 6
END
IF g>0
TAKE var g_b,g_e bl
END
IF h>0
TAKE var h_b, h_e b2
END
IF i>0
TAKE var i_b, i_e b3
END
IF j>0
TAKE var j_b,j_e b4
END
IF k>0
TAKE var k_b, k_e b5
END
IF 1>0
TAKE var L_b,L_e b6
END
271
IF m>0
TAKE var m_b,m_e c3
END
IF n>0
TAKE var n_b,n_e c4
END
IF o>0
TAKE var o_b,o_e d2
END
IF p>0
TAKE var p_b,p_e d3
END
IF q>0
TAKE var q_b,q_e d4
END
IF r>0
TAKE var r_b,r_e d5
END
IF s>0
TAKE var s_b, s_e d6
END
IF t>0
TAKE var t_b,t_e el
END
IF u>0
TAKE var u_b,u_e e2
END
IF v>0
TAKE var v_b,v_e e3
END
IF w>0
TAKE var w_b,w_e e4
END
IF x>0
TAKE var x_b,x_e e5
END
IF y>0
TAKE var y_b,y_e e6
END
IF z>0
TAKE var z_b, z_e f 1
END
IF aa>0
TAKE var aa_b,aa_e f2
END
IF bb>0
TAKE var bb_b,bb_e f3
END
272
IF cc>0
TAKE var cc_b,cc_e f4
END
IF dd>0
TAKE var dd_b,dd_e f5
END
IF ee>0
TAKE var ee_b,ee_e f6
END
IF ff>0
TAKE var f f_b, f f_e g6
END
IF gg>0
TAKE var gg_b,gg_e h4
END
IF hh>0
TAKE var hh_b,hh_e h5
END
IF ii>0
TAKE var ii_b, ii_e h.6
END
'For each stratum, the count commands below count the number
of times that a given variable value occured in each
stratum.
'The variable can have up to eight values.
COUNT
al = l
al 1
COUNT
al=2
al 2
COUNT
al = 3
al 3
COUNT
al = 4
al 4
COUNT
al = 5
al 5
COUNT
al = 6
al 6
COUNT
al=7
al 7
COUNT
al = 8
al 8
COUNT
a2 = l
a2 1
COUNT
a2=2
a2 2
COUNT
a2=3
a2 3
COUNT
a2 = 4
a2 4
COUNT
a2 = 5
a2 5
COUNT
a2 = 6
a2 6
COUNT
a2 = 7
a2 7
COUNT
a2 = 8
a2 8
COUNT
a3 = l
a3 1
COUNT
a3=2
a3_2
COUNT
a3 = 3
a3 3
COUNT
a3 = 4
a3_4
COUNT
a3=5
a3 5
COUNT
a3 = 6
a3 6
273
COUNT a3 = 7 a3_7
COUNT a3=8 a3_8
COUNT a4=l a4_l
COUNT a4=2 a4_2
COUNT a4=3 a4_3
COUNT a4=4 a4_4
COUNT a4=5 a4_5
COUNT a4=6 a4_6
COUNT a4=7 a4_7
COUNT a4=8 a4_8
COUNT a5=l a5_l
COUNT a5=2 a5_2
COUNT a5=3 a5_3
COUNT a5=4 a5_4
COUNT a5=5 a5_5
COUNT a5=6 a5_6
COUNT a5=7 a5_7
COUNT a5=8 a5_8
COUNT a6=l a6_l
COUNT a 6=2 a6_2
COUNT a 6=3 a6_3
COUNT a6=4 a6_4
COUNT a6=5 a6_5
COUNT a 6= 6 a6_6
COUNT a6=7 a6_7
COUNT a6=8 a6_8
COUNT bl=l bl_l
COUNT bl=2 bl_2
COUNT bl=3 bl_3
COUNT bl=4 bl_4
COUNT bl=5 bl_5
COUNT bl=6 bl_6
COUNT bl=7 bl_7
COUNT bl=8 bl_8
COUNT b2=l b2_l
COUNT b2=2 b2_2
COUNT b2=3 b2_3
COUNT b2=4 b2_4
COUNT b2=5 b2_5
COUNT b2=6 b2_6
COUNT b2=7 b2_7
COUNT b2=8 b2_8
COUNT b3=l b3_l
COUNT b3=2 b3_2
COUNT b3=3 b3_3
COUNT b3=4 b3_4
COUNT b3=5 b3 5
274
COUNT b3=6 b3_6
COUNT b3=7 b3_7
COUNT b3=8 b3_8
COUNT b4=l b4_l
COUNT b4=2 b4_2
COUNT b4=3 b4_3
COUNT b4=4 b4_4
COUNT b4=5 b4_5
COUNT b4=6 b4_6
COUNT b4=7 b4_7
COUNT b4=8 b4_8
COUNT b5=l b5_l
COUNT b5=2 b5_2
COUNT b5=3 b5_3
COUNT b5=4 b5_4
COUNT b5=5 b5_5
COUNT b5=6 b5_6
COUNT b5=7 b5_7
COUNT b5=8 b5_8
COUNT b6=l b6_l
COUNT b6=2 b6_2
COUNT b6=3 b6_3
COUNT b6=4 b6_4
COUNT b6=5 b6_5
COUNT b6=6 b6_6
COUNT b6=7 b6_7
COUNT b6=8 b6_8
COUNT c3=l c3_l
COUNT c3=2 c3_2
COUNT c3=3 c3_3
COUNT c3=4 c3_4
COUNT c3=5 c3_5
COUNT c3=6 c3_6
COUNT c3=7 c3_7
COUNT c3=8 c3_8
COUNT c4=l c4_l
COUNT c4=2 c4_2
COUNT c4=3 c4_3
COUNT c4=4 c4_4
COUNT c4=5 c4_5
COUNT c4=6 c4_6
COUNT c4=7 c4_7
COUNT c4=8 c4_8
COUNT d2=l d2_l
COUNT d2=2 d2_2
COUNT d2=3 d2 3
275
COUNT
d2 =
= 4
d2
4
COUNT
d2 =
=5
d2~
"5
COUNT
6,2--
= 6
d2~
~6
COUNT
6.2--
=7
d2~
~7
COUNT
62--
= 8
d2~
"8
COUNT
63~-
= 1
d3~
"l
COUNT
63~-
=2
d3~
"2
COUNT
63~-
=3
d3~
"3
COUNT
63~-
= 4
d3~
~4
COUNT
63~-
=5
d3~
"5
COUNT
63~-
= 6
d3~
~6
COUNT
63~-
=7
d3~
~7
COUNT
63~-
= 8
d3~
"8
COUNT
6A--
= 1
d4~
"l
COUNT
6A--
=2
d4~
"2
COUNT
d4 =
=3
d4~
"3
COUNT
d4 =
= 4
d4~
~4
COUNT
d4 =
=5
d4~
"5
COUNT
d4 =
= 6
d4~
"6
COUNT
d4 =
=7
d4~
"7
COUNT
d4 =
= 8
d4~
"8
COUNT
d5=
= 1
d5~
"l
COUNT
d5 =
--2
d5~
~2
COUNT
d5 =
=3
d5~
"3
COUNT
d5 =
= 4
d5~
~4
COUNT
d5 =
=5
d5~
"5
COUNT
d5=
= 6
d5~
"6
COUNT
d5 =
=7
d5~
"7
COUNT
d5=
= 8
d5~
~8
COUNT
66=
= 1
d6~
"l
COUNT
66~-
=2
d6~
"2
COUNT
d6=
=3
d6~
"3
COUNT
d6=
= 4
d6~
~4
COUNT
d6=
= 5
d6~
"5
COUNT
d6=
= 6
d6~
"6
COUNT
d6=
=7
d6~
"7
COUNT
d6=
= 8
d6~
~8
COUNT el=l el_l
COUNT el=2 el_2
COUNT el=3 el_3
COUNT el=4 el_4
COUNT el=5 el_5
COUNT el=6 el_6
COUNT el=7 el_7
COUNT el=8 el_8
COUNT e2=l e2_l
COUNT e2=2 e2 2
276
COUNT e2=3 e2_3
COUNT e2=4 e2_4
COUNT e2=5 e2_5
COUNT e2=6 e2_6
COUNT e2=7 e2_7
COUNT e2=8 e2_8
COUNT e3=l e3_l
COUNT e3=2 e3_2
COUNT e3=3 e3_3
COUNT e3=4 e3_4
COUNT e3=5 e3_5
COUNT e3=6 e3_6
COUNT e3=7 e3_7
COUNT e3=8 e3_8
COUNT e4=l e4_l
COUNT e4=2 e4_2
COUNT e4=3 e4_3
COUNT e4=4 e4_4
COUNT e4=5 e4_5
COUNT e4=6 e4_6
COUNT e4=7 e4_7
COUNT e4=8 e4_8
COUNT e5=l e5_l
COUNT e5=2 e5_2
COUNT e5=3 e5_3
COUNT e5=4 e5_4
COUNT e5=5 e5_5
COUNT e5=6 e5_6
COUNT e5=7 e5_7
COUNT e5=8 e5_8
COUNT e6=l e6_l
COUNT e6=2 e6_2
COUNT e6=3 e6_3
COUNT e6=4 e6_4
COUNT e6=5 e6_5
COUNT e6=6 e6_6
COUNT e6=7 e6_7
COUNT e6=8 e6_8
COUNT fl=l fl_l
COUNT fl=2 fl_2
COUNT fl=3 fl_3
COUNT fl=4 fl_4
COUNT fl=5 fl_5
COUNT fl = 6 fl_6
COUNT fl=7 fl_7
COUNT fl=8 fl_8
COUNT f2=l f2 1
277
COUNT f2=2 f2_2
COUNT f2=3 f2_3
COUNT f2=4 f2_4
COUNT f2=5 f2_5
COUNT f2=6 f2_6
COUNT f2=7 f2_7
COUNT f2=8 f2_8
COUNT f3=l f3_l
COUNT f3=2 f3_2
COUNT f3=3 f3_3
COUNT f3=4 f3_4
COUNT f3=5 f3_5
COUNT f3=6 f3_6
COUNT f3=7 f3_7
COUNT f3=8 f3_8
COUNT f4=l f4_l
COUNT f4=2 f4_2
COUNT f4=3 f4_3
COUNT f4=4 f4_4
COUNT f4=5 f4_5
COUNT f4=6 f4_6
COUNT f4=7 f4_7
COUNT f4=8 f4_8
COUNT f5=l f5_l
COUNT f5=2 f5_2
COUNT f5=3 f5_3
COUNT f5=4 f5_4
COUNT f5=5 f5_5
COUNT f5=6 f5_6
COUNT f5=7 f5_7
COUNT f5=8 f5_8
COUNT f6=l f6_l
COUNT f6=2 f6_2
COUNT f6=3 f6_3
COUNT f6=4 f6_4
COUNT f6=5 f6_5
COUNT f6=6 f6_6
COUNT f6=7 f6_7
COUNT f6=8 f6_8
COUNT g6=l g6_l
COUNT g6=2 g6_2
COUNT g6=3 g6_3
COUNT g6=4 g6_4
COUNT g6=5 g6_5
COUNT g6=6 g6_6
COUNT g6=7 g6_7
COUNT g6=8 g6 8
278
COUNT h4=l h4_l
COUNT h4=2 h4_2
COUNT h4=3 h4_3
COUNT h4=4 h4_4
COUNT h4=5 h4_5
COUNT h4=6 h4_6
COUNT h4=7 h4_7
COUNT h4=8 h4_8
COUNT h5=l h5_l
COUNT h5=2 h5_2
COUNT h5=3 h5_3
COUNT h5=4 h5_4
COUNT h5=5 h5_5
COUNT h5=6 h5_6
COUNT h5=7 h5_7
COUNT h5=8 h5_8
COUNT h6=l h6_l
COUNT h6=2 h6_2
COUNT h6=3 h6_3
COUNT h6=4 h6_4
COUNT h6=5 h6_5
COUNT h6=6 h6_6
COUNT h6=7 h6_7
COUNT h6=8 h6_8
'The set and multiply commands are used to estimate the size
of the population for each stratum.
'Each case is multiplied by four, which approximates the
ration of population to sample.
SET 1 4 ratio
MULTIPLY al_l ratio al_lpop
MULTIPLY al_2 ratio al_2pop
MULTIPLY al_3 ratio al_3pop
MULTIPLY al_4 ratio al_4pop
MULTIPLY al_5 ratio al_5pop
MULTIPLY al_6 ratio al_6pop
MULTIPLY al_7 ratio al_7pop
MULTIPLY al_8 ratio al_8pop
MULTIPLY a2_l ratio a2_lpop
MULTIPLY a2_2 ratio a2_2pop
MULTIPLY a2_3 ratio a2_3pop
MULTIPLY a2_4 ratio a2_4pop
MULTIPLY a2_5 ratio a2_5pop
MULTIPLY a2_6 ratio a2_6pop
MULTIPLY a2_7 ratio a2_7pop
MULTIPLY a2_8 ratio a2_8pop
MULTIPLY a3 1 ratio a3 lpop
279
MULTIPLY a3_2 ratio a3_2pop
MULTIPLY a3_3 ratio a3_3pop
MULTIPLY a3_4 ratio a3_4pop
MULTIPLY a3_5 ratio a3_5pop
MULTIPLY a3_6 ratio a3_6pop
MULTIPLY a3_7 ratio a3_7pop
MULTIPLY a3_8 ratio a3_8pop
MULTIPLY a4_l ratio a4_lpop
MULTIPLY a4_2 ratio a4_2pop
MULTIPLY a4_3 ratio a4_3pop
MULTIPLY a4_4 ratio a4_4pop
MULTIPLY a4_5 ratio a4_5pop
MULTIPLY a4_6 ratio a4_6pop
MULTIPLY a4_7 ratio a4_7pop
MULTIPLY a4_8 ratio a4_8pop
MULTIPLY a5_l ratio a5_lpop
MULTIPLY a5_2 ratio a5_2pop
MULTIPLY a5_3 ratio a5_3pop
MULTIPLY a5_4 ratio a5_4pop
MULTIPLY a5_5 ratio a5_5pop
MULTIPLY a5_6 ratio a5_6pop
MULTIPLY a5_7 ratio a5_7pop
MULTIPLY a5_8 ratio a5_8pop
MULTIPLY a6_l ratio a6_lpop
MULTIPLY a6_2 ratio a6_2pop
MULTIPLY a6_3 ratio a6_3pop
MULTIPLY a6_4 ratio a6_4pop
MULTIPLY a6_5 ratio a6_5pop
MULTIPLY a6_6 ratio a6_6pop
MULTIPLY a6_7 ratio a6_7pop
MULTIPLY a6_8 ratio a6_8pop
MULTIPLY bl_l ratio bl_lpop
MULTIPLY bl_2 ratio bl_2pop
MULTIPLY bl_3 ratio bl_3pop
MULTIPLY bl_4 ratio bl_4pop
MULTIPLY bl_5 ratio bl_5pop
MULTIPLY bl_6 ratio bl_6pop
MULTIPLY bl_7 ratio bl_7pop
MULTIPLY bl_8 ratio bl_8pop
MULTIPLY b2_l ratio b2_lpop
MULTIPLY b2_2 ratio b2_2pop
MULTIPLY b2_3 ratio b2_3pop
MULTIPLY b2_4 ratio b2_4pop
MULTIPLY b2_5 ratio b2_5pop
MULTIPLY b2_6 ratio b2_6pop
MULTIPLY b2_7 ratio b2_7pop
MULTIPLY b2 8 ratio b2 8pop
280
MULTIPLY b3_l ratio b3_lpop
MULTIPLY b3_2 ratio b3_2pop
MULTIPLY b3_3 ratio b3_3pop
MULTIPLY b3_4 ratio b3_4pop
MULTIPLY b3_5 ratio b3_5pop
MULTIPLY b3_6 ratio b3_6pop
MULTIPLY b3_7 ratio b3_7pop
MULTIPLY b3_8 ratio b3_8pop
MULTIPLY b4_l ratio b4_lpop
MULTIPLY b4_2 ratio b4_2pop
MULTIPLY b4_3 ratio b4_3pop
MULTIPLY b4_4 ratio b4_4pop
MULTIPLY b4_5 ratio b4_5pop
MULTIPLY b4_6 ratio b4_6pop
MULTIPLY b4_7 ratio b4_7pop
MULTIPLY b4_8 ratio b4_8pop
MULTIPLY b5_l ratio b5_lpop
MULTIPLY b5_2 ratio b5_2pop
MULTIPLY b5_3 ratio b5_3pop
MULTIPLY b5_4 ratio b5_4pop
MULTIPLY b5_5 ratio b5_5pop
MULTIPLY b5_6 ratio b5_6pop
MULTIPLY b5_7 ratio b5_7pop
MULTIPLY b5_8 ratio b5_8pop
MULTIPLY b6_l ratio b6_lpop
MULTIPLY b6_2 ratio b6_2pop
MULTIPLY b6_3 ratio b6_3pop
MULTIPLY b6_4 ratio b6_4pop
MULTIPLY b6_5 ratio b6_5pop
MULTIPLY b6_6 ratio b6_6pop
MULTIPLY b6_7 ratio b6_7pop
MULTIPLY b6_8 ratio b6_8pop
MULTIPLY c3_l ratio c3_lpop
MULTIPLY c3_2 ratio c3_2pop
MULTIPLY c3_3 ratio c3_3pop
MULTIPLY c3_4 ratio c3_4pop
MULTIPLY c3_5 ratio c3_5pop
MULTIPLY c3_6 ratio c3_6pop
MULTIPLY c3_7 ratio c3_7pop
MULTIPLY c3_8 ratio c3_8pop
MULTIPLY c4_l ratio c4_lpop
MULTIPLY c4_2 ratio c4_2pop
MULTIPLY c4_3 ratio c4_3pop
MULTIPLY c4_4 ratio c4_4pop
MULTIPLY c4_5 ratio c4_5pop
MULTIPLY c4_6 ratio c4_6pop
MULTIPLY c4 7 ratio c4 7pop
281
MULTIPLY c4_8 ratio c4_8pop
MULTIPLY d2_l ratio d2_lpop
MULTIPLY d2_2 ratio d2_2pop
MULTIPLY d2_3 ratio d2_3pop
MULTIPLY d2_4 ratio d2_4pop
MULTIPLY d2_5 ratio d2_5pop
MULTIPLY d2_6 ratio d2_6pop
MULTIPLY d2_7 ratio d2_7pop
MULTIPLY d2_8 ratio d2_8pop
MULTIPLY d3_l ratio d3_lpop
MULTIPLY d3_2 ratio d3_2pop
MULTIPLY d3_3 ratio d3_3pop
MULTIPLY d3_4 ratio d3_4pop
MULTIPLY d3_5 ratio d3_5pop
MULTIPLY d3_6 ratio d3_6pop
MULTIPLY d3_7 ratio d3_7pop
MULTIPLY d3_8 ratio d3_8pop
MULTIPLY d4_l ratio d4_lpop
MULTIPLY d4_2 ratio d4_2pop
MULTIPLY d4_3 ratio d4_3pop
MULTIPLY d4_4 ratio d4_4pop
MULTIPLY d4_5 ratio d4_5pop
MULTIPLY d4_6 ratio d4_6pop
MULTIPLY d4_7 ratio d4_7pop
MULTIPLY d4_8 ratio d4_8pop
MULTIPLY d5_l ratio d5_lpop
MULTIPLY d5_2 ratio d5_2pop
MULTIPLY d5_3 ratio d5_3pop
MULTIPLY d5_4 ratio d5_4pop
MULTIPLY d5_5 ratio d5_5pop
MULTIPLY d5_6 ratio d5_6pop
MULTIPLY d5_7 ratio d5_7pop
MULTIPLY d5_8 ratio d5_8pop
MULTIPLY d6_l ratio d6_lpop
MULTIPLY d6_2 ratio d6_2pop
MULTIPLY d6_3 ratio d6_3pop
MULTIPLY d6_4 ratio d6_4pop
MULTIPLY d6_5 ratio d6_5pop
MULTIPLY d6_6 ratio d6_6pop
MULTIPLY d6_7 ratio d6_7pop
MULTIPLY d6_8 ratio d6_8pop
MULTIPLY el_l ratio el_lpop
MULTIPLY el_2 ratio el_2pop
MULTIPLY el_3 ratio el_3pop
MULTIPLY el_4 ratio el_4pop
MULTIPLY el 5 ratio el 5pop
282
MULTIPLY el_6 ratio el_6pop
MULTIPLY el_7 ratio el_7pop
MULTIPLY el_8 ratio el_8pop
MULTIPLY e2_l ratio e2_lpop
MULTIPLY e2_2 ratio e2_2pop
MULTIPLY e2_3 ratio e2_3pop
MULTIPLY e2_4 ratio e2_4pop
MULTIPLY e2_5 ratio e2_5pop
MULTIPLY e2_6 ratio e2_6pop
MULTIPLY e2_7 ratio e2_7pop
MULTIPLY e2_8 ratio e2_8pop
MULTIPLY e3_l ratio e3_lpop
MULTIPLY e3_2 ratio e3_2pop
MULTIPLY e3_3 ratio e3_3pop
MULTIPLY e3_4 ratio e3_4pop
MULTIPLY e3_5 ratio e3_5pop
MULTIPLY e3_6 ratio e3_6pop
MULTIPLY e3_7 ratio e3_7pop
MULTIPLY e3_8 ratio e3_8pop
MULTIPLY e4_l ratio e4_lpop
MULTIPLY e4_2 ratio e4_2pop
MULTIPLY e4_3 ratio e4_3pop
MULTIPLY e4_4 ratio e4_4pop
MULTIPLY e4_5 ratio e4_5pop
MULTIPLY e4_6 ratio e4_6pop
MULTIPLY e4_7 ratio e4_7pop
MULTIPLY e4_8 ratio e4_8pop
MULTIPLY e5_l ratio e5_lpop
MULTIPLY e5_2 ratio e5_2pop
MULTIPLY e5_3 ratio e5_3pop
MULTIPLY e5_4 ratio e5_4pop
MULTIPLY e5_5 ratio e5_5pop
MULTIPLY e5_6 ratio e5_6pop
MULTIPLY e5_7 ratio e5_7pop
MULTIPLY e5_8 ratio e5_8pop
MULTIPLY e6_l ratio e6_lpop
MULTIPLY e6_2 ratio e6_2pop
MULTIPLY e6_3 ratio e6_3pop
MULTIPLY e6_4 ratio e6_4pop
MULTIPLY e6_5 ratio e6_5pop
MULTIPLY e6_6 ratio e6_6pop
MULTIPLY e6_7 ratio e6_7pop
MULTIPLY e6_8 ratio e6_8pop
MULTIPLY fl_l ratio fl_lpop
MULTIPLY fl_2 ratio fl_2pop
MULTIPLY fl_3 ratio fl_3pop
MULTIPLY fl 4 ratio fl 4pop
283
MULTIPLY fl_5 ratio fl_5pop
MULTIPLY fl_6 ratio fl_6pop
MULTIPLY fl_7 ratio fl_7pop
MULTIPLY fl_8 ratio fl_8pop
MULTIPLY f2_l ratio f2_lpop
MULTIPLY f2_2 ratio f2_2pop
MULTIPLY f2_3 ratio f2_3pop
MULTIPLY f2_4 ratio f2_4pop
MULTIPLY f2_5 ratio f2_5pop
MULTIPLY f2_6 ratio f2_6pop
MULTIPLY f2_7 ratio f2_7pop
MULTIPLY f2_8 ratio f2_8pop
MULTIPLY f3_l ratio f3_lpop
MULTIPLY f3_2 ratio f3_2pop
MULTIPLY f3_3 ratio f3_3pop
MULTIPLY f3_4 ratio f3_4pop
MULTIPLY f3_5 ratio f3_5pop
MULTIPLY f3_6 ratio f3_6pop
MULTIPLY f3_7 ratio f3_7pop
MULTIPLY f3_8 ratio f3_8pop
MULTIPLY f4_l ratio f4_lpop
MULTIPLY f4_2 ratio f4_2pop
MULTIPLY f4_3 ratio f4_3pop
MULTIPLY f4_4 ratio f4_4pop
MULTIPLY f4_5 ratio f4_5pop
MULTIPLY f4_6 ratio f4_6pop
MULTIPLY f4_7 ratio f4_7pop
MULTIPLY f4_8 ratio f4_8pop
MULTIPLY f5_l ratio f5_lpop
MULTIPLY f5_2 ratio f5_2pop
MULTIPLY f5_3 ratio f5_3pop
MULTIPLY f5_4 ratio f5_4pop
MULTIPLY f5_5 ratio f5_5pop
MULTIPLY f5_6 ratio f5_6pop
MULTIPLY f5_7 ratio f5_7pop
MULTIPLY f5_8 ratio f5_8pop
MULTIPLY f6_l ratio f6_lpop
MULTIPLY f6_2 ratio f6_2pop
MULTIPLY f6_3 ratio f6_3pop
MULTIPLY f6_4 ratio f6_4pop
MULTIPLY f6_5 ratio f6_5pop
MULTIPLY f6_6 ratio f6_6pop
MULTIPLY f6_7 ratio f6_7pop
MULTIPLY f6_8 ratio f6_8pop
MULTIPLY g6_l ratio g6_lpop
MULTIPLY g6_2 ratio g6_2pop
MULTIPLY g6_3 ratio g6_3pop
284
MULTIPLY g6_4 ratio g6_4pop
MULTIPLY g6_5 ratio g6_5pop
MULTIPLY g6_6 ratio g6_6pop
MULTIPLY g6_7 ratio g6_7pop
MULTIPLY g6_8 ratio g6_8pop
MULTIPLY
h4
1
ratio
h4
lpop
MULTIPLY
h4~
~2
ratio
h4"
2pop
MULTIPLY
h4~
"3
ratio
h4~
3pop
MULTIPLY
h4~
~4
ratio
h4~
4pop
MULTIPLY
h4~
"5
ratio
h4"
5pop
MULTIPLY
h4~
"6
ratio
h4"
6pop
MULTIPLY
h4~
~7
ratio
h4"
7pop
MULTIPLY
h4~
~8
ratio
h4"
8pop
MULTIPLY
h5~
"l
ratio
h5"
_lpop
MULTIPLY
h5~
"2
ratio
h5"
2pop
MULTIPLY
h5~
"3
ratio
h5"
3pop
MULTIPLY
h5~
"4
ratio
h5"
4pop
MULTIPLY
h5~
"5
ratio
h5"
5pop
MULTIPLY
h5~
"6
ratio
h5"
6pop
MULTIPLY
h5~
~7
ratio
h5"
7pop
MULTIPLY
h5~
"8
ratio
h5"
8pop
MULTIPLY
h6~
"l
ratio
h6"
lpop
MULTIPLY
h6~
~2
ratio
h6"
2pop
MULTIPLY
h6~
"3
ratio
h6~
3pop
MULTIPLY
h6~
"4
ratio
h6"
4pop
MULTIPLY
h6~
"5
ratio
h6"
5pop
MULTIPLY
h6~
"6
ratio
h6"
6pop
MULTIPLY
h6~
"7
ratio
h6"
_7pop
MULTIPLY
h6~
~8
ratio
h6~
8pop
'The following commands create an urn
estimates the size and proportions of
population .
'Each urn should have four times more
corresponding sampled stratum,
' but in the same proportions as
URN al_lpop#l al_2pop#2 al_3pop#3 al_
al_6pop#6 al_7pop#7 al_8pop#8 alu
URN a2_lpop#l a2_2pop#2 a2_3pop#3 a2_
a2_6pop#6 a2_7pop#7 a2_8pop#8 a2u
URN a3_lpop#l a3_2pop#2 a3_3pop#3 a3_
a3_6pop#6 a3_7pop#7 a3_8pop#8 a3u
URN a4_lpop#l a4_2pop#2 a4_3pop#3 a4
a4_6pop#6 a4_7pop#7 a4_8pop#8 a4u
URN a5_lpop#l a5_2pop#2 a5_3pop#3 a5_
a5_6pop#6 a5_7pop#7 a5_8pop#8 a5u
URN a6 lpop#l a6 2pop#2 a6 3pop#3 a6
for each stratum that
values in the
values than the
the sample.
4pop#4 al_5pop#5
4pop#4 a2_5pop#5
4pop#4 a3_5pop#5
4pop#4 a4_5pop#5
4pop#4 a5_5pop#5
4pop#4 a6 5pop#5
285
a6_6pop#6 a6_7pop#7 a6_8pop#8 a6u
URN bl_lpop#
bl_6pop#6 bl
URN b2_lpop#
b2_6pop#6 b2
URN b3_lpop#
b3_6pop#6 b3
URN b4_lpop#
b4_6pop#6 b4
URN b5_lpop#
b5_6pop#6 b5
URN b6_lpop#
b6 6pop#6 b6
1 bl_2pop#2
_7pop#7 blj
1 b2_2pop#2
_7pop#7 b2_i
1 b3_2pop#2
_7pop#7 b3_S
1 b4_2pop#2
_7pop#7 b4_i
1 b5_2pop#2
_7pop#7 b5_l
1 b6_2pop#2
7pop#7 b6 f
bl_3pop#3 bl_
!pop#8 blu
b2_3pop#3 b2_
!pop#8 b2u
b3_3pop#3 b3_
!pop#8 b3u
b4_3pop#3 b4
lpop#8 b4u
b5_3pop#3 b5_
!pop#8 b5u
b6_3pop#3 b6_
!pop#8 b6u
URN c3_lpop#l c3_2pop#2 c3_3pop#3 c3_
c3_6pop#6 c3_7pop#7 c3_8pop#8 c3u
URN c4_lpop#l c4_2pop#2 c4_3pop#3 c4
c4 6pop#6 c4 7pop#7 c4 8pop#8 c4u
URN d2_lpop#
d2_6pop#6 d2
URN d3_lpop#
d3_6pop#6 d3
URN d4_lpop#
d4_6pop#6 d4
URN d5_lpop#
d5_6pop#6 d5
URN d6_lpop#
d6_6pop#6 d6
URN el_lpop#
el_6pop#6 el
URN e2_lpop#
e2_6pop#6 e2
URN e3_lpop#
e3_6pop#6 e3
URN e4_lpop#
e4_6pop#6 e4
URN e5_lpop#
e5_6pop#6 e5
URN e6_lpop#
e6 6pop#6 e6
1 d2_2pop#2
_7pop#7 d2_
1 d3_2pop#2
_7pop#7 d3_
1 d4_2pop#2
_7pop#7 d4_
1 d5_2pop#2
_7pop#7 d5_
1 d6_2pop#2
_7pop#7 d6_
1 el_2pop#2
_7pop#7 el_
1 e2_2pop#2
_7pop#7 e2_
1 e3_2pop#2
_7pop#7 e3_
1 e4_2pop#2
_7pop#7 e4_
1 e5_2pop#2
_7pop#7 e5_
1 e6_2pop#2
7pop#7 e6
d2_3pop#3 d2_
8pop#8 d2u
d3_3pop#3 d3_
8pop#8 d3u
d4_3pop#3 d4
8pop#8 d4u
d5_3pop#3 d5_
8pop#8 d5u
d6_3pop#3 d6_
8pop#8 d6u
el_3pop#3 el_
8pop#8 elu
e2_3pop#3 e2_
8pop#8 e2u
e3_3pop#3 e3_
8pop#8 e3u
e4_3pop#3 e4
8pop#8 e4u
e5_3pop#3 e5_
8pop#8 e5u
e6_3pop#3 e6_
8pop#8 e6u
URN fl_lpop#l fl_2pop#2 fl_3pop#3 fl
fl_6pop#6 fl_7pop#7 fl_8pop#8 flu
URN f2_lpop#l f2_2pop#2 f2_3pop#3 f2
f2 6pop#6 f2 7pop#7 f2 8pop#8 f2u
4pop#4 bl_5pop#5
4pop#4 b2_5pop#5
4pop#4 b3_5pop#5
4pop#4 b4_5pop#5
4pop#4 b5_5pop#5
4pop#4 b6_5pop#5
4pop#4 c3_5pop#5
4pop#4 c4_5pop#5
4pop#4 d2_5pop#5
4pop#4 d3_5pop#5
4pop#4 d4_5pop#5
4pop#4 d5_5pop#5
4pop#4 d6_5pop#5
4pop#4 el_5pop#5
4pop#4 e2_5pop#5
4pop#4 e3_5pop#5
4pop#4 e4_5pop#5
4pop#4 e5_5pop#5
4pop#4 e6_5pop#5
4pop#4 fl_5pop#5
4pop#4 f2 5pop#5
286
URN f3_lpop#l f3_2pop#2 f3_3pop#3 f3_4pop#4 f3_5pop#5
f3_6pop#6 f3_7pop#7 f3_8pop#8 f3u
URN f4_lpop#l f4_2pop#2 f4_3pop#3 f4_4pop#4 f4_5pop#5
f4_6pop#6 f4_7pop#7 f4_8pop#8 f4u
URN f5_lpop#l f5_2pop#2 f5_3pop#3 f5_4pop#4 f5_5pop#5
f5_6pop#6 f5_7pop#7 f5_8pop#8 f5u
URN f6_lpop#l f6_2pop#2 f6_3pop#3 f6_4pop#4 f6_5pop#5
f6_6pop#6 f6_7pop#7 f6_8pop#8 f6u
URN g6_lpop#l g6_2pop#2 g6_3pop#3 g6_4pop#4 g6_5pop#5
g6_6pop#6 g6_7pop#7 g6_8pop#8 g6u
URN h4_lpop#l h4_2pop#2 h4_3pop#3 h4_4pop#4 h4_5pop#5
h4_6pop#6 h4_7pop#7 h4_8pop#8 h4u
URN h5_lpop#l h5_2pop#2 h5_3pop#3 h5_4pop#4 h5_5pop#5
h5_6pop#6 h5_7pop#7 h5_8pop#8 h5u
URN h6_lpop#l h6_2pop#2 h6_3pop#3 h6_4pop#4 h6_5pop#5
h6 6pop#6 h6 7pop#7 h6 8pop#8 h6u
'The following command repeats every command until the final
end 10, 000 times .
REPEAT 10000
'The following command randomizes the order of values in the
urns .
SHUFFLE alu $alus
SHUFFLE a2u $a2us
SHUFFLE a3u $a3us
SHUFFLE a4u $a4us
SHUFFLE a5u $a5us
SHUFFLE a6u $a6us
SHUFFLE blu $blus
SHUFFLE b2u $b2us
SHUFFLE b3u $b3us
SHUFFLE b4u $b4us
SHUFFLE b5u $b5us
SHUFFLE b6u $b6us
SHUFFLE c3u $c3us
SHUFFLE c4u $c4us
SHUFFLE d2u $d2us
SHUFFLE d3u $d3us
SHUFFLE d4u $d4us
SHUFFLE d5u $d5us
SHUFFLE d6u $d6us
SHUFFLE elu $elus
SHUFFLE e2u $e2us
SHUFFLE e3u $e3us
SHUFFLE e4u $e4us
SHUFFLE e5u $e5us
287
SHUFFLE e6u $e6us
SHUFFLE flu $flus
SHUFFLE f2u $f2us
SHUFFLE f3u $f3us
SHUFFLE f4u $f4us
SHUFFLE f5u $f5us
SHUFFLE f6u $f6us
SHUFFLE g6u $g6us
SHUFFLE h4u $h4us
SHUFFLE h5u $h5us
SHUFFLE h6u $h6us
'The following commands take a n sized sample from each
urn .
IF a>0
TAKE $alus l,a $als
END
IF b>0
TAKE $a2us l,b $a2s
END
IF c>0
TAKE $a3us l,c $a3s
END
IF d>0
TAKE $a4us l,d $a4s
END
IF e>0
TAKE $a5us l,e $a5s
END
IF f>0
TAKE $a6us l,f $a6s
END
IF g>0
TAKE $blus l,g $bls
END
IF h>0
TAKE $b2us l,h $b2s
END
IF i>0
TAKE $b3us l,i $b3s
288
END
IF j>0
TAKE $b4us l,j $b4s
END
IF k>0
TAKE $b5us l,k $b5s
END
IF 1>0
TAKE $b6us 1,L $b6s
END
IF m>0
TAKE $c3us l,m $c3s
END
IF n>0
TAKE $c4us l,n $c4s
END
IF o>0
TAKE $d2us l,o $d2s
END
IF p>0
TAKE $d3us l,p $d3s
END
IF q>0
TAKE $d4us l,q $d4s
END
IF r>0
TAKE $d5us l,r $d5s
END
IF s>0
TAKE $d6us l,s $d6s
END
IF t>0
TAKE $elus l,t $els
END
IF u>0
TAKE $e2us l,u $e2s
289
END
IF v>0
TAKE $e3us l,v $e3s
END
IF w>0
TAKE $e4us l,w $e4s
END
IF x>0
TAKE $e5us l,x $e5s
END
IF y>0
TAKE $e6us l,y $e6s
END
IF z>0
TAKE $flus l,z $fls
END
IF aa>0
TAKE $f2us l,aa $f2s
END
IF bb>0
TAKE $f3us l,bb $f3s
END
IF cc>0
TAKE $f4us l,cc $f4s
END
IF dd>0
TAKE $f5us l,dd $f5s
END
IF ee>0
TAKE $f6us l,ee $f6s
END
IF ff>0
TAKE $g6us l,ff $g6s
END
IF gg>0
TAKE $h4us 1 , gg $h4s
END
290
IF hh>0
TAKE $h5us 1 , hh $h5s
END
IF ii>0
TAKE $h6us l,ii $h6s
END
'The following command concates all of the samples into one
vector, which is the same size as the aggregate sample.
CONCAT $als $a2s $a3s $a4s $a5s $a6s $bls $b2s $b3s $b4s
$b5s $b6s $c3s $c4s $d2s $d3s $d4s $d5s $d6s $els $e2s $e3s
$e4s $e5s $e6s $fls $f2s $f3s $f4s $f5s $f6s $g6s $h4s $h5s
$h6s $ re samp
'The following commands count the number of times that a
given value appeared in the resampled sample.
COUNT $resamp=l $ re samp 1
COUNT $resamp=2 $resamp2
COUNT $resamp=3 $resamp3
COUNT $resamp=4 $resamp4
COUNT $resamp=5 $resamp5
COUNT $resamp=6 $ re samp 6
COUNT $resamp=7 $resamp7
COUNT $resamp=8 $ re samp 8
'These commands create a proportion for each variable value.
DIVIDE $resampl sampsize $propl
DIVIDE $resamp2 sampsize $prop2
DIVIDE $resamp3 sampsize $prop3
DIVIDE $resamp4 sampsize $prop4
DIVIDE $resamp5 sampsize $prop5
DIVIDE $resamp6 sampsize $prop6
DIVIDE $resamp7 sampsize $prop7
DIVIDE $resamp8 sampsize $prop8
'These commands keeps track of the resampled proportions for
each iteration.
SCORE $propl $prol
SCORE $prop2 $pro2
SCORE $prop3 $pro3
SCORE $prop4 $pro4
SCORE $prop5 $pro5
SCORE $prop6 $pro6
SCORE $prop7 $pro7
SCORE $prop8 $pro8
END
291
'This command ranks the 10,000 scores from each iteration
and displays the 2.5th, 50th, and 97.5th percentiles.
PERCENTILE $prol
PERCENTILE $pro2
PERCENTILE $pro3
PERCENTILE $pro4
PERCENTILE $pro5
PERCENTILE $pro6
PERCENTILE $pro7
PERCENTILE $pro8
'This command prints those percentiles.
PRINT sampsize percvl_l percvl_2 percvl_3 percvl_4 percvl_5
percvl 6 percvl 7 percvl 8
2.
.5
50
97.
.5)
percvl
1
2.
.5
50
97.
.5)
percvl
2
2.
.5
50
97.
.5)
percvl
3
2.
.5
50
97.
.5)
percvl
4
2.
.5
50
97.
.5)
percvl
5
2.
.5
50
97.
.5)
percvl
6
2.
.5
50
97.
.5)
percvl
7
2.
.5
50
97.
.5)
percvl
8
292
Appendix F:
Resampling Program for Calculating % 2 and M 2
for a Proportional Stratified Random Sample
'RESAMPLING PROGRAM TO CALCULATE CONFIDENCE INTERVALS AROUND
PROPORTIONS - UP TO 35 STRATA AND VARIABLES WITH 8 LEVELS
'This command reads data from an external data file.
READ file "C : WDocuments and Settings\\localadmin\\My
DocumentsWdissertationWwhole.dat" missing -9 cell deOOO
del de2 de3 de4 de5 de6 de6a de7 de7b de8 de8a de9 de9a a9
all al2 al3 al4 al5 al6 al6a al6b al7 al8 al9 a20 a21 m26
m21 as5 m22 m23 m24 m25 m27 rdl rdla rdh rd2 rd3 rd4 rd5 rd6
rdll rd7 rd8 rd9 il i2 i3 i4 i5 i6 i7 i8 dl d2 d3 d4file d5
d6 d7 d8 d9 dlO dll dl2 ml m2 m3 m3a m4 m5 m5a m6 m7 m7a m8
m8a m9 mlO mil ml2 fl f2 f3 f4 f5 f6 f7 f8 si s2 s3fine s3a
s4 s4a s4b s4c s5 s5a s5b s6 s6a s7 s7a s8 s8a s8aa s8b s8c
s8d s8e s8f s8h var00006 filter journal cse
'The following commands renames a variable and cleans system
missing cases.
DATA m21 var
DATA m21 varchi
DATA cell forum
DATA journal comp
data journal compm2
CLEAN forum var varchi comp compm2
'This commmand calculates the correlation between the
comparison and observation variables.
corr compm2 varchi cor
square cor scor
print cor scor
'These commands enables a vector to be split into groups.
count varchi=l sampyes
count varchi=2 sampno
add sampyes 1 yesbegin
add sampyes sampno nobegin
print sampyes
293
'These commands recodes the values of the variables into
prime numbers
'so that the vectors can be combined into unique values.
RECODE varchi = 11 varchi
RECODE varchi = 1 13 varchi
RECODE varchi = 2 17 varchi
RECODE varchi = 3 19 varchi
RECODE varchi = 4 23 varchi
RECODE comp = 41 comp
RECODE comp = 1 43 comp
RECODE comp = 2 47 comp
RECODE comp = 3 53 comp
RECODE comp = 4 59 comp
RECODE comp = 5 61 comp
RECODE comp = 6 67 comp
RECODE comp = 7 71 comp
MULTIPLY comp varchi combined
COUNT
combined
= 451
cvOO
COUNT
combined
=533
cvOl
COUNT
combined
= 697
cv02
COUNT
combined
=779
cv03
COUNT
combined
= 943
cv04
COUNT
combined
= 473
cvlO
COUNT
combined
= 559
evil
COUNT
combined
= 731
cvl2
COUNT
combined
= 817
cvl3
COUNT
combined
= 989
cvl4
COUNT
combined
= 517
cv2
COUNT
combined
= 611
cv21
COUNT
combined
= 799
cv22
COUNT
combined
= 893
cv2 3
COUNT
combined
= 1081
. cv2 4
COUNT
combined
= 583
cv30
COUNT
combined
= 689
cv31
COUNT
combined
= 901
cv32
COUNT
combined
= 1007
' cv33
COUNT
combined
=121?
> cv34
COUNT
combined
= 649
cv4
COUNT
combined
= 767
cv41
COUNT
combined
=1003
! cv42
COUNT
combined
= 1121
. cv43
COUNT
combined
= 1357
' cv44
COUNT
combined
= 671
cv50
294
COUNT combined =7 93 cv51
COUNT combined =1037 cv52
COUNT combined =1159 cv53
COUNT combined =1403 cv54
COUNT combined =737 cv60
COUNT combined =871 cv61
COUNT combined =1139 cv62
COUNT combined =1273 cv63
COUNT combined =1541 cv64
COUNT combined =7 81 cv7
COUNT combined =923 cv71
COUNT combined =1207 cv72
COUNT combined =1349 cv73
COUNT combined =1633 cv74
'These commands find the row, column and grand marginals to
get vectors of expected and observed values.
ADD cvOl cv02 rowl
ADD evil cvl2 row2
ADD cvOl evil coll
ADD cv02 cvl2 col2
ADD cvOl cv02 evil cvl2 grand
MULTIPLY rowl coll mrowlcoll
MULTIPLY rowl col2 mrowlcol2
MULTIPLY row2 coll mrow2coll
MULTIPLY row2 col2 mrow2col2
DIVIDE mrowlcoll grand ecvOl
DIVIDE mrowlcol2 grand ecv02
DIVIDE mrow2coll grand ecvll
DIVIDE mrow2col2 grand ecvl2
CONCAT ecvOl ecv02 ecvll ecvl2 expected
PRINT expected
CONCAT cvOl cv02 evil cvl2 observed
PRINT observed
'This command calculates chi square for the sample.
CHISQUARE observed expected chi
PRINT chi
'The following commands count the number of times that a
case occurs in each stratum.
COUNT forum=l a
295
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
COUNT
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
f orum=
2
=3
■4
=5
6
1
■8
9
40
41
12
43
4 4
45
4 6
4 7
48
4 9
20
■21
22
2 3
■24
2 5
2 6
21
28
2 9
■■3
■31
32
33
: 34
^35
1
k
1
m
n
o
P
q
r
s
t
u
V
w
X
Y
z
aa
bb
cc
dd
ee
ff
gg
hh
ii
'This command calculates the sample size be adding the n
size of each stratum.
ADD abcdefghij klmnopqrstuvwxyZaa
bb cc dd ee ff gg hh ii sampsize
subtract sampsize 1 nsize
print nsize
multiply nsize scor m2
print m2
'This command creates a range of values that correspond with
the n size of the strata.
'For example stratum b contains the values of the vector var
296
from a+1 to a+b.
'If the n size of stratum a is 5 and the n size of stratum b
is 6 then the values of vector var that . . .
' correspond with a are 1-5 and for b are 6-11 (a+l=6 and
a+b=ll) .
ADD a 1 b_b
ADD a b b_e
ADD b_e 1 c_b
ADD b_e c c_e
ADD c_e 1 d_b
ADD c_e d d_e
ADD d_e 1 e_b
ADD d_e e e_e
ADD e_e 1 f_b
ADD e_e f f_e
ADD f_e 1 g_b
ADD f_e g g_e
ADD g_e 1 h_b
ADD g_e h h_e
ADD h_e 1 i_b
ADD h_e i i_e
ADD i_e 1 j_b
ADD i_e j j_e
ADD j_e 1 k_b
ADD j_e k k_e
ADD k_e 1 L_b
ADD k_e L L_e
ADD L_e 1 m_b
ADD L_e m m_e
ADD m_e 1 n_b
ADD m_e n n_e
ADD n_e 1 o_b
ADD n_e o o_e
ADD o_e 1 p_b
ADD o_e p p_e
ADD p_e 1 g_b
ADD p_e g g_e
ADD g_e 1 r_b
ADD g_e r r_e
ADD r_e 1 s_b
ADD r_e s s_e
ADD s~e 1 tjo
ADD s_e t t_e
ADD t_e 1 u_b
ADD t_e u u_e
ADD u_e 1 v_b
ADD u e v v e
297
ADD v_e 1 w_b
ADD v_e w w_e
ADD w_e 1 x_b
ADD w_e x x_e
ADD x_e 1 y_b
ADD x_e y y_e
ADD y_e 1 z_b
ADD y_e z z_e
ADD z_e 1 aa_b
ADD z_e aa aa_e
ADD aa_e 1 bb_b
ADD aa_e bb bb_e
ADD bb_e 1 cc_b
ADD bb_e cc cc_e
ADD cc_e 1 dd_b
ADD cc_e dd dd_e
ADD dd_e 1 ee_b
ADD dd_e ee ee_e
ADD ee~e 1 ff_b
ADD ee_e ff ff_e
ADD ff_e 1 gg_b
ADD ff_e gg gg_e
ADD gg_e 1 hh_b
ADD gg_e hh hh_e
ADD hh_e 1 ii_b
ADD hh_e ii ii_e
'The following commands take the values of vector var and
breaks them into smaller vectors that. . .
1 correspond with each stratum, if there n size in the
stratum is greater than zero.
IF a>0
TAKE var l,a al
END
IF b>0
TAKE var b_b,b_e a2
END
IF c>0
TAKE var c_b,c_e a3
END
IF d>0
TAKE var d_b,d_e a4
END
IF e>0
TAKE var e_b,e_e a5
END
IF f>0
TAKE var f b, f e a 6
298
END
IF g>0
TAKE var g_b,g_e bl
END
IF h>0
TAKE var h_b, h_e b2
END
IF i>0
TAKE var i_b, i_e b3
END
IF j>0
TAKE var j_b,j_e b4
END
IF k>0
TAKE var k_b, k_e b5
END
IF 1>0
TAKE var L_b,L_e b6
END
IF m>0
TAKE var m_b,m_e c3
END
IF n>0
TAKE var n_b,n_e c4
END
IF o>0
TAKE var o_b,o_e d2
END
IF p>0
TAKE var p_b,p_e d3
END
IF q>0
TAKE var q_b,q_e d4
END
IF r>0
TAKE var r_b,r_e d5
END
IF s>0
TAKE var s_b, s_e d6
END
IF t>0
TAKE var t_b,t_e el
END
IF u>0
TAKE var u_b,u_e e2
END
IF v>0
TAKE var v b,v e e3
299
END
IF w>0
TAKE var w_b,w_e e4
END
IF x>0
TAKE var x_b,x_e e5
END
IF y>0
TAKE var y_b,y_e e6
END
IF z>0
TAKE var z_b, z_e f 1
END
IF aa>0
TAKE var aa_b,aa_e f2
END
IF bb>0
TAKE var bb_b,bb_e f3
END
IF cc>0
TAKE var cc_b,cc_e f4
END
IF dd>0
TAKE var dd_b,dd_e f5
END
IF ee>0
TAKE var ee_b,ee_e f6
END
IF ff>0
TAKE var f f_b, f f_e g6
END
IF gg>0
TAKE var gg_b,gg_e h4
END
IF hh>0
TAKE var hh_b,hh_e h.5
END
IF ii>0
TAKE var ii_b, ii_e h.6
END
'For each stratum, the count commands below count the number
of times that a given variable value occured in each
stratum.
'The variable can have up to eight values.
300
COUNT al = l al_l
COUNT al=2 al_2
COUNT al=3 al_3
COUNT al=4 al_4
COUNT al=5 al_5
COUNT al=6 al_6
COUNT al=7 al_7
COUNT al=8 al_8
COUNT a2=l a2_l
COUNT a2=2 a2_2
COUNT a2=3 a2_3
COUNT a2=4 a2_4
COUNT a2=5 a2_5
COUNT a2=6 a2_6
COUNT a2=7 a2_7
COUNT a2=8 a2_8
COUNT a3=l a3_l
COUNT a3=2 a3_2
COUNT a3=3 a3_3
COUNT a3=4 a3_4
COUNT a3=5 a3_5
COUNT a3=6 a3_6
COUNT a3=7 a3_7
COUNT a3=8 a3_8
COUNT a4=l a4_l
COUNT a4=2 a4_2
COUNT a4=3 a4_3
COUNT a4=4 a4_4
COUNT a4=5 a4_5
COUNT a4=6 a4_6
COUNT a4=7 a4_7
COUNT a4=8 a4_8
COUNT a5=l a5_l
COUNT a5=2 a5_2
COUNT a5=3 a5_3
COUNT a5=4 a5_4
COUNT a5=5 a5_5
COUNT a5=6 a5_6
COUNT a5=7 a5_7
COUNT a5=8 a5_8
COUNT a6=l a6_l
COUNT a 6=2 a6_2
COUNT a6=3 a6_3
COUNT a6=4 a6_4
COUNT a6=5 a6_5
COUNT a 6= 6 a6_6
COUNT a6=7 a6_7
COUNT a6=8 a6 8
301
COUNT
bl=
= 1
bl
1
COUNT
bl=
=2
bl"
2
COUNT
bl=
= 3
bl"
3
COUNT
bl=
= 4
bl"
4
COUNT
bl=
=5
bl"
5
COUNT
bl=
= 6
bl"
6
COUNT
bl=
=7
bl"
7
COUNT
bl=
= 8
bl"
8
COUNT
b2=
--1
b2~
1
COUNT
b2 =
=2
b2~
2
COUNT
b2 =
=3
b2~
3
COUNT
b2 =
= 4
b2~
4
COUNT
b2 =
= 5
b2~
5
COUNT
b2 =
= 6
b2"
6
COUNT b2=7 b2_7
COUNT b2=8 b2_8
COUNT b3=l b3_l
COUNT b3=2 b3_2
COUNT b3=3 b3_3
COUNT b3=4 b3_4
COUNT b3=5 b3_5
COUNT b3=6 b3_6
COUNT b3=7 b3_7
COUNT b3=8 b3_8
COUNT b4=l b4_l
COUNT b4=2 b4_2
COUNT b4=3 b4_3
COUNT b4=4 b4_4
COUNT b4=5 b4_5
COUNT b4=6 b4_6
COUNT b4=7 b4_7
COUNT b4=8 b4_8
COUNT b5=l b5_l
COUNT b5=2 b5_2
COUNT b5=3 b5_3
COUNT b5=4 b5_4
COUNT b5=5 b5_5
COUNT b5=6 b5_6
COUNT b5=7 b5_7
COUNT b5=8 b5_8
COUNT b6=l b6_l
COUNT b6=2 b6_2
COUNT b6=3 b6_3
COUNT b6=4 b6_4
COUNT b6=5 b6_5
COUNT b6=6 b6_6
COUNT b6=7 b6_7
COUNT b6=8 b6 8
302
COUNT c3=l c3_l
COUNT c3=2 c3_2
COUNT c3=3 c3_3
COUNT c3=4 c3_4
COUNT c3=5 c3_5
COUNT c3=6 c3_6
COUNT c3=7 c3_7
COUNT c3=8 c3_8
COUNT c4 = l c4_l
COUNT c4=2 c4_2
COUNT c4=3 c4_3
COUNT c4=4 c4_4
COUNT c4=5 c4_5
COUNT c4=6 c4_6
COUNT c4=7 c4_7
COUNT c4=8 c4_8
COUNT d2=l d2_l
COUNT d2=2 d2_2
COUNT d2=3 d2_3
COUNT d2=4 d2_4
COUNT d2=5 d2_5
COUNT d2=6 d2_6
COUNT d2=7 d2_7
COUNT d2=8 d2_8
COUNT d3=l d3_l
COUNT d3=2 d3_2
COUNT d3=3 d3_3
COUNT d3=4 d3_4
COUNT d3=5 d3_5
COUNT d3=6 d3_6
COUNT d3=7 d3_7
COUNT d3=8 d3_8
COUNT d4=l d4_l
COUNT d4=2 d4_2
COUNT d4=3 d4_3
COUNT d4=4 d4_4
COUNT d4=5 d4_5
COUNT d4=6 d4_6
COUNT d4=7 d4_7
COUNT d4=8 d4_8
COUNT d5=l d5_l
COUNT d5=2 d5_2
COUNT d5=3 d5_3
COUNT d5=4 d5_4
COUNT d5=5 d5_5
COUNT d5=6 d5_6
COUNT d5=7 d5 7
303
COUNT d5=8 d5_8
COUNT d6=l d6_l
COUNT d6=2 d6_2
COUNT d6=3 d6_3
COUNT d6=4 d6_4
COUNT d6=5 d6_5
COUNT d6=6 d6_6
COUNT d6=7 d6_7
COUNT d6=8 d6_8
COUNT el=l el_l
COUNT el=2 el_2
COUNT el=3 el_3
COUNT el=4 el_4
COUNT el=5 el_5
COUNT el=6 el_6
COUNT el=7 el_7
COUNT el=8 el_8
COUNT e2=l e2_l
COUNT e2=2 e2_2
COUNT e2=3 e2_3
COUNT e2=4 e2_4
COUNT e2=5 e2_5
COUNT e2=6 e2_6
COUNT e2=7 e2_7
COUNT e2=8 e2_8
COUNT e3=l e3_l
COUNT e3=2 e3_2
COUNT e3=3 e3_3
COUNT e3=4 e3_4
COUNT e3=5 e3_5
COUNT e3=6 e3_6
COUNT e3=7 e3_7
COUNT e3=8 e3_8
COUNT e4=l e4_l
COUNT e4=2 e4_2
COUNT e4=3 e4_3
COUNT e4=4 e4_4
COUNT e4=5 e4_5
COUNT e4=6 e4_6
COUNT e4=7 e4_7
COUNT e4=8 e4_8
COUNT e5=l e5_l
COUNT e5=2 e5_2
COUNT e5=3 e5_3
COUNT e5=4 e5_4
COUNT e5=5 e5_5
COUNT e5=6 e5 6
304
COUNT e5=7 e5_7
COUNT e5=8 e5_8
COUNT e6=l e6_l
COUNT e6=2 e6_2
COUNT e6=3 e6_3
COUNT e6=4 e6_4
COUNT e6=5 e6_5
COUNT e6=6 e6_6
COUNT e6=7 e6_7
COUNT e6=8 e6_8
COUNT fl=l fl_l
COUNT fl=2 fl_2
COUNT fl=3 fl_3
COUNT fl = 4 fl_4
COUNT fl=5 fl_5
COUNT fl = 6 fl_6
COUNT fl=7 fl_7
COUNT fl=8 fl_8
COUNT f2=l f2_l
COUNT f2=2 f2_2
COUNT f2=3 f2_3
COUNT f2=4 f2_4
COUNT f2=5 f2_5
COUNT f2=6 f2_6
COUNT f2=7 f2_7
COUNT f2=8 f2_8
COUNT f3=l f3_l
COUNT f3=2 f3_2
COUNT f3=3 f3_3
COUNT f3=4 f3_4
COUNT f3=5 f3_5
COUNT f3=6 f3_6
COUNT f3=7 f3_7
COUNT f3=8 f3_8
COUNT f4=l f4_l
COUNT f4=2 f4_2
COUNT f4=3 f4_3
COUNT f4=4 f4_4
COUNT f4=5 f4_5
COUNT f4=6 f4_6
COUNT f4=7 f4_7
COUNT f4=8 f4_8
COUNT f5=l f5_l
COUNT f5=2 f5_2
COUNT f5=3 f5_3
COUNT f5=4 f5_4
COUNT f5=5 f5 5
305
COUNT f5=6 f5_6
COUNT f5 = 7 f5_7
COUNT f5=8 f5_8
COUNT f6=l f6_l
COUNT f6=2 f6_2
COUNT f6=3 f6_3
COUNT f6=4 f6_4
COUNT f6=5 f6_5
COUNT f6=6 f6_6
COUNT f6=7 f6_7
COUNT f6=8 f6_8
COUNT g6=l g6_l
COUNT g6=2 g6_2
COUNT g6=3 g6_3
COUNT g6=4 g6_4
COUNT g6=5 g6_5
COUNT g6=6 g6_6
COUNT g6=7 g6_7
COUNT g6=8 g6_8
COUNT h4=l h4_l
COUNT h4=2 h4_2
COUNT h4=3 h4_3
COUNT h4=4 h4_4
COUNT h4=5 h4_5
COUNT h4=6 h4_6
COUNT h4=7 h4_7
COUNT h4=8 h4_8
COUNT h5=l h5_l
COUNT h5=2 h5_2
COUNT h5=3 h5_3
COUNT h5=4 h5_4
COUNT h5=5 h5_5
COUNT h5=6 h5_6
COUNT h5=7 h5_7
COUNT h5=8 h5_8
COUNT h6=l h6_l
COUNT h6=2 h6_2
COUNT h6=3 h6_3
COUNT h6=4 h6_4
COUNT h6=5 h6_5
COUNT h6=6 h6_6
COUNT h6=7 h6_7
COUNT h6=8 h6_8
'The set and multiply commands are used to estimate the size
of the population for each stratum.
306
'Each case is multiplied by four, which approximates the
ration of population to sample.
SET 1 4 ratio
MULTIPLY al_l ratio al_lpop
MULTIPLY al_2 ratio al_2pop
MULTIPLY al_3 ratio al_3pop
MULTIPLY al_4 ratio al_4pop
MULTIPLY al_5 ratio al_5pop
MULTIPLY al_6 ratio al_6pop
MULTIPLY al_7 ratio al_7pop
MULTIPLY al_8 ratio al_8pop
MULTIPLY a2_l ratio a2_lpop
MULTIPLY a2_2 ratio a2_2pop
MULTIPLY a2_3 ratio a2_3pop
MULTIPLY a2_4 ratio a2_4pop
MULTIPLY a2_5 ratio a2_5pop
MULTIPLY a2_6 ratio a2_6pop
MULTIPLY a2_7 ratio a2_7pop
MULTIPLY a2_8 ratio a2_8pop
MULTIPLY a3_l ratio a3_lpop
MULTIPLY a3_2 ratio a3_2pop
MULTIPLY a3_3 ratio a3_3pop
MULTIPLY a3_4 ratio a3_4pop
MULTIPLY a3_5 ratio a3_5pop
MULTIPLY a3_6 ratio a3_6pop
MULTIPLY a3_7 ratio a3_7pop
MULTIPLY a3_8 ratio a3_8pop
MULTIPLY a4_l ratio a4_lpop
MULTIPLY a4_2 ratio a4_2pop
MULTIPLY a4_3 ratio a4_3pop
MULTIPLY a4_4 ratio a4_4pop
MULTIPLY a4_5 ratio a4_5pop
MULTIPLY a4_6 ratio a4_6pop
MULTIPLY a4_7 ratio a4_7pop
MULTIPLY a4_8 ratio a4_8pop
MULTIPLY a5_l ratio a5_lpop
MULTIPLY a5_2 ratio a5_2pop
MULTIPLY a5_3 ratio a5_3pop
MULTIPLY a5_4 ratio a5_4pop
MULTIPLY a5_5 ratio a5_5pop
MULTIPLY a5_6 ratio a5_6pop
MULTIPLY a5_7 ratio a5_7pop
MULTIPLY a5_8 ratio a5_8pop
MULTIPLY a6_l ratio a6_lpop
MULTIPLY a6_2 ratio a6_2pop
MULTIPLY a6_3 ratio a6_3pop
MULTIPLY a6 4 ratio a6 4pop
307
MULTIPLY a6_5 ratio a6_5pop
MULTIPLY a6_6 ratio a6_6pop
MULTIPLY a6_7 ratio a6_7pop
MULTIPLY a6_8 ratio a6_8pop
MULTIPLY bl_l ratio bl_lpop
MULTIPLY bl_2 ratio bl_2pop
MULTIPLY bl_3 ratio bl_3pop
MULTIPLY bl_4 ratio bl_4pop
MULTIPLY bl_5 ratio bl_5pop
MULTIPLY bl_6 ratio bl_6pop
MULTIPLY bl_7 ratio bl_7pop
MULTIPLY bl_8 ratio bl_8pop
MULTIPLY b2_l ratio b2_lpop
MULTIPLY b2_2 ratio b2_2pop
MULTIPLY b2_3 ratio b2_3pop
MULTIPLY b2_4 ratio b2_4pop
MULTIPLY b2_5 ratio b2_5pop
MULTIPLY b2_6 ratio b2_6pop
MULTIPLY b2_7 ratio b2_7pop
MULTIPLY b2_8 ratio b2_8pop
MULTIPLY b3_l ratio b3_lpop
MULTIPLY b3_2 ratio b3_2pop
MULTIPLY b3_3 ratio b3_3pop
MULTIPLY b3_4 ratio b3_4pop
MULTIPLY b3_5 ratio b3_5pop
MULTIPLY b3_6 ratio b3_6pop
MULTIPLY b3_7 ratio b3_7pop
MULTIPLY b3_8 ratio b3_8pop
MULTIPLY b4_l ratio b4_lpop
MULTIPLY b4_2 ratio b4_2pop
MULTIPLY b4_3 ratio b4_3pop
MULTIPLY b4_4 ratio b4_4pop
MULTIPLY b4_5 ratio b4_5pop
MULTIPLY b4_6 ratio b4_6pop
MULTIPLY b4_7 ratio b4_7pop
MULTIPLY b4_8 ratio b4_8pop
MULTIPLY b5_l ratio b5_lpop
MULTIPLY b5_2 ratio b5_2pop
MULTIPLY b5_3 ratio b5_3pop
MULTIPLY b5_4 ratio b5_4pop
MULTIPLY b5_5 ratio b5_5pop
MULTIPLY b5_6 ratio b5_6pop
MULTIPLY b5_7 ratio b5_7pop
MULTIPLY b5_8 ratio b5_8pop
MULTIPLY b6_l ratio b6_lpop
MULTIPLY b6_2 ratio b6_2pop
MULTIPLY b6 3 ratio b6 3pop
308
MULTIPLY b6_4 ratio b6_4pop
MULTIPLY b6_5 ratio b6_5pop
MULTIPLY b6_6 ratio b6_6pop
MULTIPLY b6_7 ratio b6_7pop
MULTIPLY b6_8 ratio b6_8pop
MULTIPLY c3_l ratio c3_lpop
MULTIPLY c3_2 ratio c3_2pop
MULTIPLY c3_3 ratio c3_3pop
MULTIPLY c3_4 ratio c3_4pop
MULTIPLY c3_5 ratio c3_5pop
MULTIPLY c3_6 ratio c3_6pop
MULTIPLY c3_7 ratio c3_7pop
MULTIPLY c3_8 ratio c3_8pop
MULTIPLY c4_l ratio c4_lpop
MULTIPLY c4_2 ratio c4_2pop
MULTIPLY c4_3 ratio c4_3pop
MULTIPLY c4_4 ratio c4_4pop
MULTIPLY c4_5 ratio c4_5pop
MULTIPLY c4_6 ratio c4_6pop
MULTIPLY c4_7 ratio c4_7pop
MULTIPLY c4_8 ratio c4_8pop
MULTIPLY d2_l ratio d2_lpop
MULTIPLY d2_2 ratio d2_2pop
MULTIPLY d2_3 ratio d2_3pop
MULTIPLY d2_4 ratio d2_4pop
MULTIPLY d2_5 ratio d2_5pop
MULTIPLY d2_6 ratio d2_6pop
MULTIPLY d2_7 ratio d2_7pop
MULTIPLY d2_8 ratio d2_8pop
MULTIPLY d3_l ratio d3_lpop
MULTIPLY d3_2 ratio d3_2pop
MULTIPLY d3_3 ratio d3_3pop
MULTIPLY d3_4 ratio d3_4pop
MULTIPLY d3_5 ratio d3_5pop
MULTIPLY d3_6 ratio d3_6pop
MULTIPLY d3_7 ratio d3_7pop
MULTIPLY d3_8 ratio d3_8pop
MULTIPLY d4_l ratio d4_lpop
MULTIPLY d4_2 ratio d4_2pop
MULTIPLY d4_3 ratio d4_3pop
MULTIPLY d4_4 ratio d4_4pop
MULTIPLY d4_5 ratio d4_5pop
MULTIPLY d4_6 ratio d4_6pop
MULTIPLY d4_7 ratio d4_7pop
MULTIPLY d4_8 ratio d4_8pop
MULTIPLY d5 1 ratio d5 lpop
309
MULTIPLY d5_2 ratio d5_2pop
MULTIPLY d5_3 ratio d5_3pop
MULTIPLY d5_4 ratio d5_4pop
MULTIPLY d5_5 ratio d5_5pop
MULTIPLY d5_6 ratio d5_6pop
MULTIPLY d5_7 ratio d5_7pop
MULTIPLY d5_8 ratio d5_8pop
MULTIPLY d6_l ratio d6_lpop
MULTIPLY d6_2 ratio d6_2pop
MULTIPLY d6_3 ratio d6_3pop
MULTIPLY d6_4 ratio d6_4pop
MULTIPLY d6_5 ratio d6_5pop
MULTIPLY d6_6 ratio d6_6pop
MULTIPLY d6_7 ratio d6_7pop
MULTIPLY d6_8 ratio d6_8pop
MULTIPLY el_l ratio el_lpop
MULTIPLY el_2 ratio el_2pop
MULTIPLY el_3 ratio el_3pop
MULTIPLY el_4 ratio el_4pop
MULTIPLY el_5 ratio el_5pop
MULTIPLY el_6 ratio el_6pop
MULTIPLY el_7 ratio el_7pop
MULTIPLY el_8 ratio el_8pop
MULTIPLY e2_l ratio e2_lpop
MULTIPLY e2_2 ratio e2_2pop
MULTIPLY e2_3 ratio e2_3pop
MULTIPLY e2_4 ratio e2_4pop
MULTIPLY e2_5 ratio e2_5pop
MULTIPLY e2_6 ratio e2_6pop
MULTIPLY e2_7 ratio e2_7pop
MULTIPLY e2_8 ratio e2_8pop
MULTIPLY e3_l ratio e3_lpop
MULTIPLY e3_2 ratio e3_2pop
MULTIPLY e3_3 ratio e3_3pop
MULTIPLY e3_4 ratio e3_4pop
MULTIPLY e3_5 ratio e3_5pop
MULTIPLY e3_6 ratio e3_6pop
MULTIPLY e3_7 ratio e3_7pop
MULTIPLY e3_8 ratio e3_8pop
MULTIPLY e4_l ratio e4_lpop
MULTIPLY e4_2 ratio e4_2pop
MULTIPLY e4_3 ratio e4_3pop
MULTIPLY e4_4 ratio e4_4pop
MULTIPLY e4_5 ratio e4_5pop
MULTIPLY e4_6 ratio e4_6pop
MULTIPLY e4_7 ratio e4_7pop
MULTIPLY e4 8 ratio e4 8pop
310
MULTIPLY e5_l ratio e5_lpop
MULTIPLY e5_2 ratio e5_2pop
MULTIPLY e5_3 ratio e5_3pop
MULTIPLY e5_4 ratio e5_4pop
MULTIPLY e5_5 ratio e5_5pop
MULTIPLY e5_6 ratio e5_6pop
MULTIPLY e5_7 ratio e5_7pop
MULTIPLY e5_8 ratio e5_8pop
MULTIPLY e6_l ratio e6_lpop
MULTIPLY e6_2 ratio e6_2pop
MULTIPLY e6_3 ratio e6_3pop
MULTIPLY e6_4 ratio e6_4pop
MULTIPLY e6_5 ratio e6_5pop
MULTIPLY e6_6 ratio e6_6pop
MULTIPLY e6_7 ratio e6_7pop
MULTIPLY e6_8 ratio e6_8pop
MULTIPLY fl_l ratio fl_lpop
MULTIPLY fl_2 ratio fl_2pop
MULTIPLY fl_3 ratio fl_3pop
MULTIPLY fl_4 ratio fl_4pop
MULTIPLY fl_5 ratio fl_5pop
MULTIPLY fl_6 ratio fl_6pop
MULTIPLY fl_7 ratio fl_7pop
MULTIPLY fl_8 ratio fl_8pop
MULTIPLY f2_l ratio f2_lpop
MULTIPLY f2_2 ratio f2_2pop
MULTIPLY f2_3 ratio f2_3pop
MULTIPLY f2_4 ratio f2_4pop
MULTIPLY f2_5 ratio f2_5pop
MULTIPLY f2_6 ratio f2_6pop
MULTIPLY f2_7 ratio f2_7pop
MULTIPLY f2_8 ratio f2_8pop
MULTIPLY f3_l ratio f3_lpop
MULTIPLY f3_2 ratio f3_2pop
MULTIPLY f3_3 ratio f3_3pop
MULTIPLY f3_4 ratio f3_4pop
MULTIPLY f3_5 ratio f3_5pop
MULTIPLY f3_6 ratio f3_6pop
MULTIPLY f3_7 ratio f3_7pop
MULTIPLY f3_8 ratio f3_8pop
MULTIPLY f4_l ratio f4_lpop
MULTIPLY f4_2 ratio f4_2pop
MULTIPLY f4_3 ratio f4_3pop
MULTIPLY f4_4 ratio f4_4pop
MULTIPLY f4_5 ratio f4_5pop
MULTIPLY f4_6 ratio f4_6pop
MULTIPLY f4 7 ratio f4 7pop
311
MULTIPLY f4_8 ratio f4_8pop
MULTIPLY f5_l ratio f5_lpop
MULTIPLY f5_2 ratio f5_2pop
MULTIPLY f5_3 ratio f5_3pop
MULTIPLY f5_4 ratio f5_4pop
MULTIPLY f5_5 ratio f5_5pop
MULTIPLY f5_6 ratio f5_6pop
MULTIPLY f5_7 ratio f5_7pop
MULTIPLY f5_8 ratio f5_8pop
MULTIPLY f6_l ratio f6_lpop
MULTIPLY f6_2 ratio f6_2pop
MULTIPLY f6_3 ratio f6_3pop
MULTIPLY f6_4 ratio f6_4pop
MULTIPLY f6_5 ratio f6_5pop
MULTIPLY f6_6 ratio f6_6pop
MULTIPLY f6_7 ratio f6_7pop
MULTIPLY f6_8 ratio f6_8pop
MULTIPLY g6_l ratio g6_lpop
MULTIPLY g6_2 ratio g6_2pop
MULTIPLY g6_3 ratio g6_3pop
MULTIPLY g6_4 ratio g6_4pop
MULTIPLY g6_5 ratio g6_5pop
MULTIPLY g6_6 ratio g6_6pop
MULTIPLY g6_7 ratio g6_7pop
MULTIPLY g6_8 ratio g6_8pop
MULTIPLY h4_l ratio h4_lpop
MULTIPLY h4_2 ratio h4_2pop
MULTIPLY h4_3 ratio h4_3pop
MULTIPLY h4_4 ratio h4_4pop
MULTIPLY h4_5 ratio h4_5pop
MULTIPLY h4_6 ratio h4_6pop
MULTIPLY h4_7 ratio h4_7pop
MULTIPLY h4_8 ratio h4_8pop
MULTIPLY h5_l ratio h5_lpop
MULTIPLY h5_2 ratio h5_2pop
MULTIPLY h5_3 ratio h5_3pop
MULTIPLY h5_4 ratio h5_4pop
MULTIPLY h5_5 ratio h5_5pop
MULTIPLY h5_6 ratio h5_6pop
MULTIPLY h5_7 ratio h5_7pop
MULTIPLY h5_8 ratio h5_8pop
MULTIPLY h6_l ratio h6_lpop
MULTIPLY h6_2 ratio h6_2pop
MULTIPLY h6_3 ratio h6_3pop
MULTIPLY h6_4 ratio h6_4pop
MULTIPLY h6 5 ratio h6 5pop
312
MULTIPLY h6_6 ratio h6_6pop
MULTIPLY h6_7 ratio h6_7pop
MULTIPLY h6 8 ratio h6 8pop
'The followi
estimates th
population .
'Each urn sh
correspondin
' but in
URN al_lpop#
al_6pop#6 al
URN a2_lpop#
a2_6pop#6 a2
URN a3_lpop#
a3_6pop#6 a3
URN a4_lpop#
a4_6pop#6 a4
URN a5_lpop#
a5_6pop#6 a5
URN a6_lpop#
a6_6pop#6 a6
URN bl_lpop#
bl_6pop#6 bl
URN b2_lpop#
b2_6pop#6 b2
URN b3_lpop#
b3_6pop#6 b3
URN b4_lpop#
b4_6pop#6 b4
URN b5_lpop#
b5_6pop#6 b5
URN b6_lpop#
b6 6pop#6 b6
ng commands create an u
e size and proportions
ould have f
g sampled s
the same p
1 al_2pop#2
_7pop#7 al_
1 a2_2pop#2
_7pop#7 a2_
1 a3_2pop#2
_7pop#7 a3_
1 a4_2pop#2
_7pop#7 a4_
1 a5_2pop#2
_7pop#7 a5_
1 a6_2pop#2
_7pop#7 a6_
1 bl_2pop#2
_7pop#7 blj
1 b2_2pop#2
_7pop#7 b2_i
1 b3_2pop#2
_7pop#7 b3_S
1 b4_2pop#2
_7pop#7 b4_i
1 b5_2pop#2
_7pop#7 b5_i
1 b6_2pop#2
7pop#7 b6 f
our times mo
tratum,
roportions a
al_3pop#3 a
8pop#8 alu
a2_3pop#3 a
8pop#8 a2u
a3_3pop#3 a
8pop#8 a3u
a4_3pop#3 a
8pop#8 a4u
a5_3pop#3 a
8pop#8 a5u
a6_3pop#3 a
8pop#8 a6u
URN d2_lpop#l d2_2pop#2
d2_6pop#6 d2_7pop#7 d2_i
URN d3_lpop#l d3_2pop#2
d3_6pop#6 d3_7pop#7 d3_S
URN d4_lpop#l d4_2pop#2
d4_6pop#6 d4_7pop#7 d4_i
URN d5 lpop#l d5 2pop#2
rn for each stratum that
of values in the
re values than the
s the sample.
l_4pop#4 al_5pop#5
2_4pop#4 a2_5pop#5
3_4pop#4 a3_5pop#5
4_4pop#4 a4_5pop#5
5_4pop#4 a5_5pop#5
6 4pop#4 a6 5pop#5
bl_3pop#3 bl
!pop#8 blu
b2_3pop#3 b2
ipop#8 b2u
b3_3pop#3 b3
!pop#8 b3u
b4_3pop#3 b4
lpop#8 b4u
b5_3pop#3 b5
!pop#8 b5u
b6_3pop#3 b6
!pop#8 b6u
URN c3_lpop#l c3_2pop#2 c3_3pop#3 c3_
c3_6pop#6 c3_7pop#7 c3_8pop#8 c3u
URN c4_lpop#l c4_2pop#2 c4_3pop#3 c4
c4_6pop#6 c4_7pop#7 c4_8pop#8 c4u
d2_3pop#3 d2
ipop#8 d2u
d3_3pop#3 d3
!pop#8 d3u
d4_3pop#3 d4
lpop#8 d4u
d5 3pop#3 d5
4pop#4 bl_5pop#5
4pop#4 b2_5pop#5
4pop#4 b3_5pop#5
4pop#4 b4_5pop#5
4pop#4 b5_5pop#5
4pop#4 b6_5pop#5
4pop#4 c3_5pop#5
4pop#4 c4_5pop#5
4pop#4 d2_5pop#5
4pop#4 d3_5pop#5
4pop#4 d4_5pop#5
4pop#4 d5 5pop#5
313
d5_6pop#6 d5_7pop#7 d5_8pop#8 d5u
URN d6_lpop#l d6_2pop#2 d6_3pop#3 d6_4pop#4 d6_5pop#5
d6 6pop#6 d6 7pop#7 d6 8pop#8 d6u
URN el_lpop#
el_6pop#6 el
URN e2_lpop#
e2_6pop#6 e2
URN e3_lpop#
e3_6pop#6 e3
URN e4_lpop#
e4_6pop#6 e4
URN e5_lpop#
e5_6pop#6 e5
URN e6_lpop#
e6_6pop#6 e6
URN fl_lpop#
fl_6pop#6 fl
URN f2_lpop#
f2_6pop#6 f2
URN f3_lpop#
f3_6pop#6 f3
URN f4_lpop#
f4_6pop#6 f4
URN f5_lpop#
f5_6pop#6 f5
URN f6_lpop#
f6_6pop#6 f6
1 el_2pop#2
_7pop#7 el_
1 e2_2pop#2
_7pop#7 e2_
1 e3_2pop#2
_7pop#7 e3_
1 e4_2pop#2
_7pop#7 e4_
1 e5_2pop#2
_7pop#7 e5_
1 e6_2pop#2
_7pop#7 e6_
1 fl_2pop#2
_7pop#7 fl_
1 f2_2pop#2
_7pop#7 f2_
1 f3_2pop#2
_7pop#7 f3_
1 f4_2pop#2
_7pop#7 f4_
1 f5_2pop#2
_7pop#7 f5_
1 f6_2pop#2
7pop#7 f6
el_3pop#3 el_
8pop#8 elu
e2_3pop#3 e2_
8pop#8 e2u
e3_3pop#3 e3_
8pop#8 e3u
e4_3pop#3 e4
8pop#8 e4u
e5_3pop#3 e5_
8pop#8 e5u
e6_3pop#3 e6_
8pop#8 e6u
fl_3pop#3 fl
8pop#8 flu
f2_3pop#3 f2
8pop#8 f2u
f3_3pop#3 f3
8pop#8 f3u
f4_3pop#3 f4
8pop#8 f4u
f5_3pop#3 f5
8pop#8 f5u
f6_3pop#3 f6
8pop#8 f6u
4pop#4 el_5pop#5
4pop#4 e2_5pop#5
4pop#4 e3_5pop#5
4pop#4 e4_5pop#5
4pop#4 e5_5pop#5
4pop#4 e6_5pop#5
4pop#4 fl_5pop#5
4pop#4 f2_5pop#5
4pop#4 f3_5pop#5
4pop#4 f4_5pop#5
4pop#4 f5_5pop#5
4pop#4 f6 5pop#5
URN g6_lpop#l g6_2pop#2 g6_3pop#3 g6_4pop#4 g6_5pop#5
g6_6pop#6 g6_7pop#7 g6_8pop#8 g6u
URN h4_lpop#l h4_2pop#2 h4_3pop#3 h4_4pop#4 h4_5pop#5
h4_6pop#6 h4_7pop#7 h4_8pop#8 h4u
URN h5_lpop#l h5_2pop#2 h5_3pop#3 h5_4pop#4 h5_5pop#5
h5_6pop#6 h5_7pop#7 h5_8pop#8 h5u
URN h6_lpop#l h6_2pop#2 h6_3pop#3 h6_4pop#4 h6_5pop#5
h6_6pop#6 h6_7pop#7 h6_8pop#8 h6u
'The following command repeats every command until the final
end 10, 000 times .
REPEAT 10000
'The following command randomizes the order of values in the
urns .
SHUFFLE alu $alus
SHUFFLE a2u $a2us
SHUFFLE a3u $a3us
314
SHUFFLE a4u $a4us
SHUFFLE a5u $a5us
SHUFFLE a6u $a6us
SHUFFLE blu $blus
SHUFFLE b2u $b2us
SHUFFLE b3u $b3us
SHUFFLE b4u $b4us
SHUFFLE b5u $b5us
SHUFFLE b6u $b6us
SHUFFLE c3u $c3us
SHUFFLE c4u $c4us
SHUFFLE d2u $d2us
SHUFFLE d3u $d3us
SHUFFLE d4u $d4us
SHUFFLE d5u $d5us
SHUFFLE d6u $d6us
SHUFFLE elu $elus
SHUFFLE e2u $e2us
SHUFFLE e3u $e3us
SHUFFLE e4u $e4us
SHUFFLE e5u $e5us
SHUFFLE e6u $e6us
SHUFFLE flu $flus
SHUFFLE f2u $f2us
SHUFFLE f3u $f3us
SHUFFLE f4u $f4us
SHUFFLE f5u $f5us
SHUFFLE f6u $f6us
SHUFFLE g6u $g6us
SHUFFLE h4u $h4us
SHUFFLE h5u $h5us
SHUFFLE h6u $h6us
'The following commands take a n sized sample from each
urn .
IF a>0
TAKE $alus l,a $als
END
IF b>0
TAKE $a2us l,b $a2s
END
IF c>0
TAKE $a3us l,c $a3s
END
IF d>0
TAKE $a4us l,d $a4s
315
END
IF e>0
TAKE $a5us l,e $a5s
END
IF f>0
TAKE $a6us l,f $a6s
END
IF g>0
TAKE $blus l,g $bls
END
IF h>0
TAKE $b2us l,h $b2s
END
IF i>0
TAKE $b3us l,i $b3s
END
IF j>0
TAKE $b4us l,j $b4s
END
IF k>0
TAKE $b5us l,k $b5s
END
IF 1>0
TAKE $b6us 1,L $b6s
END
IF m>0
TAKE $c3us l,m $c3s
END
IF n>0
TAKE $c4us l,n $c4s
END
IF o>0
TAKE $d2us l,o $d2s
END
IF p>0
TAKE $d3us l,p $d3s
316
END
IF q>0
TAKE $d4us l,q $d4s
END
IF r>0
TAKE $d5us l,r $d5s
END
IF s>0
TAKE $d6us l,s $d6s
END
IF t>0
TAKE $elus l,t $els
END
IF u>0
TAKE $e2us l,u $e2s
END
IF v>0
TAKE $e3us l,v $e3s
END
IF w>0
TAKE $e4us l,w $e4s
END
IF x>0
TAKE $e5us l,x $e5s
END
IF y>0
TAKE $e6us l,y $e6s
END
IF z>0
TAKE $flus l,z $fls
END
IF aa>0
TAKE $f2us l,aa $f2s
END
IF bb>0
TAKE $f3us l,bb $f3s
317
END
IF cc>0
TAKE $f4us l,cc $f4s
END
IF dd>0
TAKE $f5us l,dd $f5s
END
IF ee>0
TAKE $f6us l,ee $f6s
END
IF ff>0
TAKE $g6us l,ff $g6s
END
IF gg>0
TAKE $h4us 1 , gg $h4s
END
IF hh>0
TAKE $h5us 1 , hh $h5s
END
IF ii>0
TAKE $h6us l,ii $h6s
END
'The following command concates all of the samples into one
vector, which is the same size as the aggregate sample.
CONCAT $als $a2s $a3s $a4s $a5s $a6s $bls $b2s $b3s $b4s
$b5s $b6s $c3s $c4s $d2s $d3s $d4s $d5s $d6s $els $e2s $e3s
$e4s $e5s $e6s $fls $f2s $f3s $f4s $f5s $f6s $g6s $h4s $h5s
$h6s $all
'The following commands find the expected and observed and
the value of chi sqaure for each of 10,000 resamples.
SHUFFLE $all sfalse
TAKE sfalse l,sampyes $a
TAKE sfalse yesbegin, nobegin $b
COUNT $a=l $cv01
COUNT $a=2 $cv02
COUNT $b=l $cvll
COUNT $b=2 $cvl2
318
ADD $cv01 $cv02 $rowl
ADD $cvll $cvl2 $row2
ADD $cv01 $cvll $coll
ADD $cv02 $cvl2 $col2
ADD $cv01 $cv02 $cvll $cvl2 $grand
MULTIPLY $rowl $coll $mrowlcoll
MULTIPLY $rowl $col2 $mrowlcol2
MULTIPLY $row2 $coll $mrow2coll
MULTIPLY $row2 $col2 $mrow2col2
DIVIDE $mrowlcoll $grand $ecv01
DIVIDE $mrowlcol2 $grand $ecv02
DIVIDE $mrow2coll $grand $ecvll
DIVIDE $mrow2col2 $grand $ecvl2
CONCAT $ecv01 $ecv02 $ecvll $ecvl2 $expected
CONCAT $cv01 $cv02 $cvll $cvl2 $observed
CHISQUARE $observed $expected $chi
SCORE $chi schi
'The following commands generate a distribution of null
hypothesis correlations
'to compare M2 to.
GENERATE sampsize 1,2 arand
GENERATE sampsize 1,2 brand
CORR arand brand $cor
SQUARE $cor $scor
MULTIPLY $scor nsize $m2
SCORE $m2 $sm2
END
COUNT schi >= chi kid
DIVIDE kid 10000 prob
print prob
count $sm2 >= m2 kl
divide kl 10000 probm2
print probm2
319
CURRICULUM VITAE
Justus J. Randolph
University of Joensuu
Department of Computer Science and Statistics
P.O. BOX 111
FIN-80 101, Joensuu
Finland
justusrandolph@yahoo.com
http://www.cs.joensuu.fi/~jrandolp/
+358 13 251 7924 (office)
+358 13 251 7955 (fax)
+358 44 0500 325 (mobile)
Education
Utah State University, Logan, UT 2001 -Present
PhD in education research & evaluation (in progress)
Dissertation: A Methodological Review of the Computer Science Education Research:
2000-2005
Utah State University, Logan, UT 2001 - Present
Administrative/Supervisory Certificate in education (in progress) - State of Utah
Framingham State College, Framingham, MA 2001
MEd in international education
Thesis: The Effects of Response Cards on Participation and Academic Achievement: A
Systematic Replication with Polish Students in the ESL Classroom
Hawthorne University, Salt Lake City, UT 1998
Teaching of English as a Foreign Language Certificate
Weber State University, Ogden, UT 1998
BIS in English, art history, and philosophy
320
Research and Evaluation Experience
Planning Officer - University of Joensuu, Department of Computer Science - September
2004 - Present.
Aid in the planning implementation of the International Multidisciplinary PhD
School in Educational Technology (IMPDET), curriculum development,
grantwriting, research, and teaching.
Researcher/Evaluator - University of Joensuu, Department of Computer Science - June
2004- August 2004.
Researcher, evaluation facilitator of Kids' Club, and article reviewer
Research Methodologist / Evaluation Facilitator (research internship) - University of
Joensuu, Department of Applied Education/ Department of Computer Science - October
2003 - May 2004.
Research methodologist for educational programs and research projects;
evaluation facilitator for the Development Project in Technology Education and
Learning Door
Principal Investigator - Mount Logan Middle School, Jan. 2003 - June 2003.
Evaluation of Literacy Program, Numeracy Program, and Small Schools Program,
decision-making assistance to school community councils and building internal
evaluation capacity
Graduate Research Assistant - Utah State University, Jan. 2003 - July 2003.
National Science Foundation grant providing evaluation assistance and evaluation
capacity building to other math and science instruction grantees - evaluation
research and synthesis
Graduation Research Assistant - Utah State University, Nov. 2001- July 2003.
Center for Disease Control sponsored study on newborn hearing screening
-database creation, data input, analysis, reporting, and supervision of other
assistants
Evaluator - Utah State University, Sep. 2001- July 2003.
USU College of Education Interdepartmental Doctoral Program Evaluation -
evaluation design, instrument creation, data collection, data analysis, and
reporting
SPSS Consultant - Hearing Head Start, March 2003- July 2003.
Hearing Head Start grant to research newborn hearing screening protocols -
database and syntax creation
321
Graduate Research Assistant - Utah State University, Jan 2002 - June 2002.
Utah Work Initiative Network evaluation - data analysis
Graduate Research Assistant - Utah State University, June 2001- Dec. 2001.
Idaho and Utah Universal Newborn Hearing Screening Evaluations - data input,
analysis, and reporting
Evaluation Consultant - Worldwide Institute of Research and Evaluation, Jan, 2003.
Evaluation of Junior Achievement Mentoring Program (UPS Headquarters,
Atlanta, GA) - data collection
Research Consultant - Utah State University, Dec. 2001 - Aug. 2003.
Assistance for graduate students with research designs and data analysis for theses
and dissertations
Teaching/ Administration Experience
Planning officer - University of Joensuu, Joensuu, Finland, September, 2004 - present.
Design, evaluation, and administration of a PhD program in educational
technology.
Lecturer - University of Joensuu, Finland, Fall, 2005.
Taught an online PhD course in academic writing.
Intern Principal - Mount Logan Middle School & Hillcrest Elementary School, Logan,
Utah, Jan. 2003 -June 2003.
Educational administration and supervision.
Teaching Assistant - Utah State University, 2001- 2003.
"Measurement, Design, and Analysis II", "Intro to Education Research", "Intro to
Program Evaluation", and "Proposal Development"
English Instructor - American Academy of English/Kaplan, Poland, 1998-2001.
English instruction and test preparation for students ages 9 - adult
Journal Publications
Randolph, J. J. & Eronen, P. J. (in press). Developing the Learning Door: A case study in
youth participatory program planning. Evaluation and Program Planning.
Randolph, J. J. (in press). Meta-analysis of the effects of response cards on student
achievement, participation, and intervals of off-task behavior. Journal of Positive
Behavior Interventions.
322
Randolph, J. J. (in press). What's the difference, still? A follow-up critique of the
contemporary quantitative research methodology in distance learning. Informatics
in Education.
Randolph, J. J., Virnes, M., Jormainen, I., & Eronen, P. J. (2006). The effects of a
computer-assisted interview tool on data quality. Educational Technology &
Society, 9(3), 195-205. Available online: http://www.ifets.info/journals/
93/1 7.pdf
Randolph, J. J. & Edmondson, R. S. (2005). Using the Binomial Effect Size Display
(BESD) to present the magnitude of effect sizes to the evaluation audience.
Practical Assessment, Research & Evaluation, 10(\4). Available online:
http://pareonline.net/pdf/vl0nl4.pdf
Randolph, J. J. (2005). Teacher and student satisfaction with response cards: A case study
in the Finnish as a foreign language classroom. Journal of Language and
Learning, 3(1), 53-66. ISSN 1740-4983. Available online
http://www.shakespeare.uk.net/journal/jllearn/3_l/randolph.pdf
Randolph, J. J. (2005). Teaching EFL students about the use of non-biased language.
Journal of Language and Learning 3(2), 261-267. ISSN 1740-4983. Available
online http://www.shakespeare.uk.net/journal/jllearn/3_2/randolph.pdf
Other Peer-Reviewed Publications
Randolph, J. J., Bednarik, R. & Myller, N. (2005). A methodological review of the
articles published in the proceedings of Koli Calling 2001-2004. In Proceedings
of the 5 th Annual Finnish /Baltic Sea Conference on Computer Science
Education (pp. 103-109). Finland: Helsinki University of Technology Press.
Available online: http://cs.joensuu.fi/~jrandolp/articles/koli_review.pdf
Randolph, J. J. (2005). A quantitative synthesis of the research on response cards on
participation, academic achievement, behavioral disruptions, and student
preference. In M-L. Julkunen (Ed.), Bulletins of the Faculty of Education, No. 96.
Learning and Instruction in Multiple Contexts and Settings: Proceedings of the
Fifth Joensuu Symposium on Learning and Instruction (pp. 149-165). Finland:
University of Joensuu Press. Available online: http://cs.joensuu.fi/~jrandolp/
articles/julis_rc.pdf
323
Randolph, J. J., Virnes, M., & Eronen P. J. (2005). A model for designing and evaluating
teacher training programs in technology education. In J-P.Courtiat, C, Davarakis,
& T. Villemur (Eds.), Technology Enhanced Learning: IFIP TC3 Technology
Enhanced Learning Workshop (pp. 69-79). New York: Springer. (ISBN:
0387240462) Available online: http://www.springerlink.com/
openurl.asp?genre=article&eissn= 1861 -2288&volume= 171 &spage=69
Randolph, J. J., Bednarik, R., Silander, P., Lopez-Gonzalez, J., Myller, N., & Sutinen, E.
(2005). A critical review of research methodologies reported in the full papers of
ICALT 2004. In Proceedings of the Fifth International Conference on Advanced
Learning Technologies (pp. 10-14). Los Alamitos, CA: IEEE Press. Available
online: http://ieeexplore.ieee.org/xpls/abs_all.jsp7isnumber
=32317&arnumber=1508593&count=303&index=4
Randolph, J. J. (2004). Guidelines for reducing language bias in the computing sciences:
Lessons learned from a sister science. In Proceedings of the 4th Annual Finnish /
Baltic Sea Conference on Computer Science Education (pp. 161-163). Finland:
Helsinki University of Technology Press. Available online: http://www.cs.hut.
fi/u/archie/koli04/TKOA42.pdf
Randolph, J. J., & Hartikainen, E. (2005). A review of resources for K-12 computer-
science-education program evaluation. In Yhtendistyvdt vai erilaistuvat oppimisen
ja koulutuksen polut: Kasvatustieteen paivien 2004 verkkojulkaisu [Electronic
Proceedings of the Finnish Research Days Conference 2004] (pp. 183-193).
Finland: University of Joensuu Press. Available online: http://www.geocities.com/
justusrandolph/reviewofresources.pdf
Manuscripts Submitted for Publication
Randolph, J. J. (2005). A methodological review of the program evaluations in K-12
computer science education. Manuscript submitted for publication.
Randolph, J. J. & Eronen, P.J. (2004). Program and evaluation planning lite: Planning in
the real world. Manuscript submitted for publication.
Randolph, J. J. (2005). Free-marginal multirater kappa: An alternative to Fleiss 'fixed-
marginal multirater kappa. Manuscript submitted for publication
Duveskog, M., Randolph, J J. Sutinen, E. & Vesisenaho, M. (2005). Beyond Taboos:
Exploring HIV/AIDS attitudes through contextualized learning technologies.
Manuscript submitted for publication.
324
Presentations and Non-Peer-Reviewed Publications
Randolph, J. J. (2006). Planning and Evaluating Programs in Computer Science
Education. Unpublished report prepared for the ACM Special Interest Group on
Computer Science Eduation. (192 p.)
Randolph, J. J. (2006) Free-marginal multirater kappa: An alternative to Fleiss' fixed-
marginal multirater kappa. Paper resented at the Joensuu University Learning and
Instruction Symposium 2005, Joensuu, Finland, October 14- 15th, 2005. (ERIC
Document Reproduction Service No. ED490661)
Randolph, J.J. (2005). Using the binomial distribution to confirm trends in repeated-
measures data sets that are sparse and have few cases. Presentation given at the
Joensuu University Learning and Instruction Symposium 2005, Joensuu, Finland,
October 14-15th, 2005. Presentation available online: http://www.geocities.
com/justusrandolph/sparse.ppt
Randolph, J.J. (2005). What's the difference, still: A follow-up review of the quantitative
research methodology in distance learning. Paper presented at the Joensuu
University Learning and Instruction Symposium 2005, Joensuu, Finland, October
14-15th, 2005. (ERIC Document Reproduction Service No. ED490662)
Presentation available online: http://www.geocities.com/justusrandolph/
differencejulis.ppt
Randolph, J. J. (2005, January). Teaching English as a foreign language students about
nonbiased language. Yours Truly 2005: Annual Bulletin of the Association of
Teachers of English in Finland, 16-17.
Randolph, J. J. & Eronen, P.J. (2004). Program and evaluation planning lite: Planning in
the real world. Paper presented at Kasvatustieten Pdivdt 2004 [Finnish Education
Research Days 2004], Joensuu, Finland, November 25th-26th, 2004. (ERIC
Document Reproduction Service No. ED490461)
Randolph, J. J. & Hartikainen, E. (2004). A review of resources for K-12 computer-
science-education program-evaluation. Paper presented at Kasvatustieten Pdivdt
2004 [Finnish Education Research Days 2004] , Joensuu, Finland, November
25th-26th, 2004. Presentation available online: http://www.geocities.com/
justusrandolph/reviewofresources.pdf
325
Randolph, J. J., Hartikainen, E., & Kahkonen, E. (2004). Lessons learned from
developing a methodology for the critical review of educational technology
research. Presentation given at Kasvatustieten Pdivdt 2004 [Finnish Education
Research Days 2004] , Joensuu, Finland, November 25th-26th, 2004. Presentation
available online: http://www.geocities.com/justusrandolph/
reviewmethodology.pdf
Randolph, J. J., (2004). Guidelines for reducing language bias in the computing sciences:
Language lessons learned from a sister science. Poster presented at the 4th annual
Finnish/Baltic Sea Conference on Computer Science Education, Koli, Finland,
October 1st -3rd, 2004.
Randolph, J. J. (2004). A model for designing and evaluating teacher training programs in
technology education. Paper presented at the Technology Enhanced Learning
Workshop at the IFIP World Computer Conference, Toulouse, France, August
22nd -27th, 2004. Presentation available online: http://www.geocities.com/
justusrandolph/dpet_presentation.ppt
Randolph, J. J. (2004). Getting students to talk about human rights and social justice:
Comparing Democratic Socialist platforms in the U.S. and Finland. Presentation
given at the Association of Teachers of English in Finland Seminar: Beyond
Stereotypes, Myths and Prejudices: Teaching about America, Helsinki, Finland,
January 30, 2004. Presentation available online: http://www.geocities.com/
justusrandolph/humanrights .ppt
Randolph, J. J. (2004). Teaching EFL students politically correct language in academic
and professional situations. Presentation given at the Association of Teachers of
English in Finland Seminar : Beyond Stereotypes, Myths and Prejudices:
Teaching about America, Helsinki, Finland, January 30, 2004. Presentation
available online: http://www.geocities.corn/justusrandolph/politically_correct.ppt
Randolph, J. J. (2004). A quantitative synthesis of the research on response cards on
participation, academic achievement, behavioral disruptions, and student
preference. Presentation given at the Joensuu University Learning and Instruction
Symposium. Joensuu, Finland October 24th and 25th, 2003. Presentation available
online: http://www.geocities.com/justusrandolph/quant_synth.ppt Paper available
online: http://www.geocities.com/justusrandolph/quantitative_synthesis.pdf
Randolph, J. J. (2003). The effects of response cards on participation and academic
achievement with Polish students in the English as a Foreign Language classroom.
Presentation given at the Interlearn Conference: Multidisciplinary Approaches to
Learning, Helsinki, Finland, December 4th and 5th, 2003.
http://www.geocities.com/justusrandolph/responsecard_np_pres.pdf
326
Nolan, R. K., Randolph, J. J., Kircalli, C, Adams, J. E., & Johnson, D. N. (2003). Case
study and meta-evaluation of PRAXIS: A collaborative evaluation and
organizational learning model for taking informed action. Presentation given at
the annual meeting of the American Evaluation Association Conference, Reno,
Nevada, November 7th and 8th, 2003.
Randolph, J. J. & Edmondson, R. S. (2003). The binomial effect size display: An
approach to presenting measures of effect size to the evaluation audience. Poster
presented at the American Evaluation Association Conference, Reno, Nevada,
November 7th and 8th, 2003
Randolph, J. J. (2003). The PRAXIS model: A participatory approach to school
improvement. Presentation given at the Annual Meeting of the Utah Rural Schools
Association, Cedar City, Utah, July 8th -1 1th. Presentation available online:
http://www.geocities.com/justusrandolph/PRAXIS_components.ppt
Awards and Honors
ACM SIGCSE Special Projects Grant, 2006, $4,990
ACM SIGCSE Special Projects Grant, 2005, $4,750
IFIP World Computer Congress Student Fellowship, 2004, 500 €
Fulbright Student Grant, University of Joensuu, Finland, 2003-2004, 10,500 €
Vice President for Research Fellowship, Utah State University, 2001-2002, $12,000
Cum Laude Honors, Weber State University, 1998
Academic Scholarship, Weber State University, 1995 -1998
Early College Scholarship, Weber State University, 1994
Other Professional Activities
Session Chair of Educational Modeling Languages, Interactive Learning Systems, and
Educational Paradigms. IEEE International Conference on Advanced Learning
Technologies. Joensuu, Finland, August 30th -September 1st, 2004.
327
Reviewer for the Frontiers in Education Conference: Pedagogies and Technologies for the
Emerging Global Economy.
Reviewer for American Education Research Association 2006 meeting: Education
Research in the Public Interest.
Affiliations
ACM Computer Science Education SIG
American Evaluation Association
American Psychological Association
Finnish Evaluation Society
Fulbright Alumni Association
International Society for Technology in Education
Languages
English - native language
Polish - upper intermediate spoken fluency and intermediate reading/writing proficiency
Finnish - intermediate
Research Interests
Research and evaluation methodology, international/multicultural education, social
justice in education, and technology education.