Introduction To Research Methodology
Introduction To Research Methodology
Sources of Knowledge
The major ways we learn can be classified into experience, expert opinion, and
reasoning.
Experience
The idea here is that knowledge comes from experience. Historically, this view was
called empiricism (i.e., original knowledge comes from experience). The term empirical
means "based on observation, experiment, or experience."
Expert Opinion
Because we don’t want to and don’t have time to conduct research on everything,
people frequently rely on expert opinion as they learn about the world. Note, however,
that if we rely on an expert’s opinion it is important to make sure that the expert is an
expert in the specific area under discussion and we should check to see if the expert has
a vested interest in the issue.
Reasoning.
Historically, this idea was called rationalism (i.e., original knowledge comes from
thought and reasoning). There are two main forms of reasoning:
• Deductive reasoning (i.e., the process of drawing a specific conclusion from a set of
premises).
• Inductive reasoning (i.e., reasoning from the particular to the general).
Dynamics of science.
Science has many distinguishing characteristics:
• Science is progressive.
• Science is rational.
• Science is creative.
• Science is dynamic.
• Science is open.
• Science is "critical."
• Science is never-ending.
Scientific Methods
1
There are many scientific methods. The two major methods are the inductive method
and the deductive method.
• The deductive method involves the following three steps:
1. State the hypothesis (based on theory or research literature).
2. Collect data to test the hypothesis.
3. Make decision to accept or reject the hypothesis.
Virtually any application of science includes the use of both the deductive and the
inductive approaches to the scientific method either in a single study or over time. The
inductive method is as “bottom up” method that is especially useful for generating
theories and hypotheses; the deductive method is a “top down” method that is
especially useful for testing theories and hypotheses.
Theory
The word "theory" most simply means "explanation." Theories explain "How" and
"Why" something operates as it does. Some theories are highly developed and
encompass a large terrain (i.e., "big" theories or "grand" theories); others theories are
"smaller" theories or briefer explanations.
Objectives of Research
There are five major objectives of research.
1. Exploration. This is done when we are trying to generate ideas about something.
2. Description. This is done when we want to describe the characteristics of something
or some phenomenon.
3. Explanation. This is done when we want to show how and why a phenomenon
operates as it does. If we are interested in causality, we are usually interested in
explanation.
4. Prediction. This is our objective when our primary interest is in making accurate
predictions. Note that the advanced sciences make much more accurate predictions
than the newer social and behavioral sciences.
5. Influence. This objective is a little different. It involves the application of research
results to impact the world. A demonstration program is an example of this.
One convenient and useful way to classify research is into exploratory research,
descriptive research, explanatory research, predictive research, and demonstration
research.
2
GENERAL METHODOLOGY
Researchers have long debated the relative value of qualitative and quantitative inquiry.
Phenomenological inquiry, or qualitative research, uses a naturalistic approach that
seeks to understand phenomena in context-specific settings. Logical positivism, or
quantitative research, uses experimental methods and quantitative measures to test
hypothetical generalizations. Each represents a fundamentally different inquiry
paradigm, and researcher actions are based on the underlying assumptions of each
paradigm.
Qualitative research, broadly defined, means "any kind of research that produces
findings not arrived at by means of statistical procedures or other means of
quantification". Where quantitative researchers seek causal determination, prediction,
and generalization of findings, qualitative researchers seek instead illumination,
understanding, and extrapolation to similar situations. Qualitative analysis results in a
different type of knowledge than does quantitative inquiry.
There is a kind of continuum that moves from the fictional that is "true"—the novel for
example—to the highly controlled and quantitatively described scientific experiment.
Work at either end of this continuum has the capacity to inform significantly. Qualitative
research and evaluation are located toward the fictive end of the continuum without
being fictional in the narrow sense of the term.
Cronbach (1980) claims that statistical research is not able to take full account of the
many interaction effects that take place in social settings. He gives examples of several
empirical "laws" that do not hold true in actual settings to illustrate this point. Cronbach
(1980) states that "the time has come to exorcise the null hypothesis," because it ignores
effects that may be important, but that are not statistically significant. Qualitative
inquiry accepts the complex and dynamic quality of the social world.
However, it is not necessary to pit these two paradigms against one another in a
competing stance. Patton (1990) advocates a "paradigm of choices" that seeks
"methodological appropriateness as the primary criterion for judging methodological
quality." This will allow for a "situational responsiveness" that strict adherence to one
paradigm or another will not (p. 39). Furthermore, some researchers believe that
qualitative and quantitative research can be effectively combined in the same research
project. For example, by using both quantitative and qualitative data, a study of
technology-based materials for the elementary classroom gave insights that neither type
of analysis could provide alone. Some differences between the two can be summarized
as follows:
3
Phenomenological Positivistic
Inductive Deducto-hypothetico
verificative
Holistic
Particularistic
Subjective/insider centered
(emic) Objective/outsider centered
(etic)
Process oriented
Outcome oriented
Anthropological worldview
Natural science worldview
Relative lack of control
Attempt to control variables
Goal: understanding actor’s
view Goal: finding facts & causes
Explanatory Confirmatory
Strengths
Precision and control. Control is achieved through the sampling and design, and
precision through quantitative and reliable measurement.
Statistical analyses. Since the data are quantitative, they permit statistical analyses.
In total, this approach provides answers which have a much firmer basis than the
layperson’s common sense or intuition or opinion.
Weaknesses
4
Inability to cope with the complexity of human beings. Human beings are far more
complex than the inert matter that is studied in physical sciences. This arises because
the human is not only acted on by a plethora of environmental forces, but can interpret
and respond to these forces in active way.
Assumption that facts are true and the same for all people all the time. This
approach fails to account of people’s unique ability to interpret their experiences,
construct their own meanings and act on these.
Production of trivial findings. This is due to the restriction on and the controlling of
variables and produces an artificial situation, the results of which have no bearing on
real life.
Strengths
Close association with both participants and activities within the settings. This
allows the researcher to see and document the qualities of interaction too often missed
by the scientific, more positivistic inquiries.
Insider’s view of the field. This can reveal subtleties and complexities that can go
undetected through the use of more standardized measures.
Weaknesses
Much time required for data collection, analysis, and interpretation. There is a
critical need for the researcher to spend a considerable amount of time in order to
examine, holistically and aggregately, the interactions, reactions, and activities.
5
Reactive effects on subjects. These are due to the intimacy of participant-observer
relationships within the setting.
ETHNOGRAPHY
One major approach to qualitative research is ethnography (i.e., the discovery and
description of the culture of a group of people).
• Here is the foundational question in ethnography: What are the cultural characteristics
of this group of people or of this cultural scene?
• Because ethnography originates in the discipline of Anthropology, the concept of
culture is of central importance.
• Culture is the system of shared beliefs, values, practices, language, norms, rituals, and
material things that group members use to understand their world.
• One can study micro cultures (e.g., such as the culture in a classroom) as well as macro
cultures (e.g., such as the United States of America culture).
Here are some more concepts that are commonly used by ethnographers:
• Ethnocentrism (i.e., judging others based on our cultural standards). We must avoid
this problem if we are to be a successful ethnographer!
• Emic perspective (i.e., the insider's perspective) and emic terms (i.e., specialized
words used by people in a group).
• Etic perspective (i.e., the external, social scientific view) and etic terms (i.e., outsider's
words or specialized words used by social scientists).
• Going native (i.e., identifying so completely with the group being studied that we are
unable to be objective).
• Holism (i.e., the idea that the whole is greater than the sum of its parts; it involves
describing the group as a whole unit, in addition to its parts and their
interrelationships).
The final ethnography (i.e., the report) should provide a rich and holistic description of
the culture of the group under study.
The term ethnography refers both to a research process and the product of that process.
The product is a written account, that is, the ethnography of what was studied.
Ethnographic research sometimes goes by other names, such as naturalistic research or
qualitative research. In the field of education, the ethnographic research process can be
described as the process of providing holistic and scientific descriptions of educational
systems, processes and phenomena within their specific contexts.
6
Although ethnographic research may at times involve quantitative procedures, it is
considered a part of qualitative research. As such, its epistemology, that is, its origins
and methods, is based on the epistemology of qualitative research.
The phenomenological approach emphasizes that the meaning of reality is, in essence,
in the “eyes and minds of the beholders,” the way the individuals being studied perceive
their experiences.
Ethnographic research takes a general and holistic perspective. Hypotheses are more
likely to emerge from the data than to be formulated prior to the research.
ACTION RESEARCH
7
research requires teachers to systematize what they do naturally in classroom teaching
situations when faced with a problem or dilemma: Ask relevant questions; devise a
plan; collect, analyze, and take conclusions from the data; determine and implement an
appropriate course of action; evaluate the results; and resume the cycle. It is a critical,
reflective process. Action research is thus a process in which participants examine
their own educational practice systematically and carefully using the techniques of
research. It is based on the following assumptions:
teachers and principals work best on problems they have identified for
themselves;
teachers and principals become more effective when encouraged to examine and
assess their own work and then consider ways of working differently;
1. It is not the usual things teachers do when they think about their teaching. Action
Research is systematic and involves collecting evidence on which to base rigorous
reflection.
2. It is not just problem-solving. Action Research involves problem-posing, not just
problem-solving. It does not start from a view of problems as pathologies. It is
motivated by a quest to improve and understand the world by changing it and
learning how to improve it from the effects of the changes made.
4. It is not the scientific method applied to teaching. Action Research is not just about
hypothesis-testing or about using data to come to conclusions. It is concerned with
changing situations, not just interpreting them. It takes the researcher into view.
Action Research is a systematically-evolving process of changing both the researcher
and the situations in which he or she works. The natural and historical sciences do
not have this aim.
8
such participatory research is carried out by stakeholders who are perhaps in the best
position to develop solutions, improvements, and insights. This type of research differs
from experimental research in which the investigator remains distanced from the
subjects or clients. As Osborn (2000) noted, “Action researchers, among others,
abandon the pretense of neutrality, choosing instead to focus on research results that
are trustworthy.” (p. 122).
That is, it must have credibility, the ability to communicate so that the audience can also
apply it and validate the findings. Thus, classroom action research incorporates a
number of repeated, cyclical steps (Bailey, Curtis, & Nunan, 2001; van Lier, 1996;
Wallace, 1998). The aspects of the cycle (initial definition of a problem, preliminary
inquiry, reflection, hypothesis, planning, intervention, observation, data collection, data
analysis/reflection, revision of intervention, and so on) are often not clear-cut. Bailey,
Curtis, and Nunan (2001) pointed out that it is sometimes impossible to distinguish one
phase of the process from the one preceding or following it. For instance, reflection
would ideally occur at all stages, and the preliminary inquiry may actually be continual.
The purpose and practice of action research move along a continuum, from individual
research carried out for individual, professional purposes, to collaborative research
across disciplines for improvement of praxis, to politicized research for support, for
change in consciousness, and improvement of infrastructure and delivery of services.
Osborn (2000) advocated such support through dialogue at professional conferences in
presentations of teachers’ action research. He said that, in order for such research to
bridge gaps between theory and practice, the classroom culture must frame and
engender that research. He further advocated what he conceptualizes as “critical action
research,” because it would be a tool for language educators at their level and in their
environment. He discussed the notion that cultural power often implies economic and
political power, and this reality should drive the focus of critical action research. The
focus, he recommended, should be on three areas—the implicit or explicit (and thus
economic and political) struggles within language program and practice; the illustration
of cultural struggle for students; and how “praxis can challenge consent in the
hegemonic relationships” supporting the struggles (p.122).
Illustrating the benefits and constraints: Burns (1999) discussed several advantages for
teachers who conduct action research engaging closely with their own practice—self-
awareness and personal insight; better understanding of the institutional climate;
ability to contribute to institutional solutions; and acting with the authority of
professional “knowers” rather than as the recipients (p. 14-18). Further benefits would
include a decreasing sense of isolation and enrichment of content and delivery.
Bailey, Curtis, and Nunan (2001) discussed five difficulties that teachers might face in
attempting to carry out their research:
1. administrative withholding of time or support to do the research, or lack of
recognition for the accomplishment;
2. administrative undermining of the project;
3. criticism or subversion from peers;
4. detraction from or interference with teaching duties; and
5. lack of technical skills to implement research.
9
Anderson and Heer (1999:16) have proposed five tentative tests of validity which may
enable explanations of practitioner research to acquire the 'status' of the
generalisations expected of 'research':
Outcome validity: the extent to which actions lead to a resolution of the problem that
led to the study (Did it solve the problem?)
Process validity: the extent to which problems are framed and solved in a manner that
permits ongoing learning by the individual or system (Was the activity educative and
informative?)
Democratic validity: the extent to which research is done in collaboration with all
parties who have a stake in the problem under investigation (Was the research
undertaken in collaboration with all involved with the problem under investigation?)
Catalytic validity: the extent to which the research enables the participants to
understand reality so as to transform it (Did the research transform the realities of
those involved?), and
Dialogic validity: the support for the findings accorded by a 'peer review' of colleague
practitioners (Could the research be discussed with peers in different settings?)
Will you develop and implement a new strategy or approach to address your
question? If so, what will it be?
Will you focus your study on existing practices? If so, which particular ones?
What types of data should you try to collect in order to answer your question?
How will you ensure that you have multiple perspectives?
What resources exist and what information from others might be useful in
helping you to frame your question, decide on types of data to collect, or to help
you in interpreting your findings?
10
What can you learn from the data? What patterns, insights, and new
understandings can you find?
What meaning do these patterns, insights, and new understandings have for your
practice? for your students?
How will you write about what you have learned so that the findings will be
useful to you and to others?
Action research is emergent. The process takes place gradually. Its cyclic nature helps
responsiveness. It also aids rigour. The early cycles are used to help decide how to
conduct the later cycles. In the later cycles, the interpretations developed in the early
cycles can be tested and challenged and refined.
The use of language also makes the whole process more accessible to participants.
They can develop enough understanding to become co-researchers in many situations.
One crucial step in each cycle consists of critical reflection. The researcher and others
involved first recollect and then critique what has already happened. The increased
understanding which emerges from the critical reflection is then put to good use in
designing the later steps.
The cycle best known is probably that of Stephen Kemmis and his colleagues at Deakin
University. The steps are:
plan --> act --> observe --> reflect (and then --> plan etc.)
11
The reflection leads on to the next stage of planning. The "planning" is not a separate
and prior step; it is embedded in the action and reflection. Short, multiple cycles allow
greater rigour to be achieved.
As change is intended to result, effective action research depends upon the agreement
and commitment of those affected by it. This is usually generated by involving them
directly in the research process. In many instances, researchers try to involve them as
equal partners.
CONTENT ANALYSIS
Definitions
Content analysis is
Whatever the variation in definitions, all convey the notion that content analysis is a
rigorous process with a definite structure to it.
Words and sentences are important human artifacts. What people say and write
can be a source of evidence about individual and social processes. For example: A
psychologist may administer the Thematic Apperception Test (TAT) because analysis of
the content present in stories produced from the TAT is thought to reflect the subject's
social history, values, needs, defenses, life concerns, wishes to impress psychologist.
Such data can be used to infer characteristics of a particular subject, or by
comparing stories told by subjects who vary with respect to social class, ethnicity, etc. a
researcher may extend inferences to general groups.
Content analysis techniques can be applied to the content of any:
Book
Magazine
Newspaper
Photograph
Cartoon
Comic strip
interviews
motion picture
meeting minutes
12
Radio/TV broadcasts
political documents
individual stories
While many projects are focused on the message produced, their object may be to
find links between the message and other parts of the environment. In such studies,
content analysis is considered a tool to be used in combination with other techniques.
Another objective may be to study communications in relation their environment.
One example of such work is the study of content related to gatekeepers. Gatekeepers,
by saying yes or no to messages that come to them along the chain, play one of the most
important roles in social communication. In a news agency for example, the editor is
responsible for the greatest number of decisions about the passage of news along the
chain.
13
Content analysis is context sensitive and therefore able to process symbolic
forms. In experimental research, language tends not to be treated as symbolic
phenomena.
Content analysis can cope with large volumes of data.
Text is an excellent vehicle for studying long term changes in attitudes, concerns,
and styles.
Stages of Development
14
5. Revise coding rules
6. Return to code samples check reliability until acceptable
7. Code all text
8. Assess reliability
Reliability and validity problems grow out of the ambiguity of word meanings and
category or variable definitions.
Reliability
Stability -- Extent to which the results of content classification are invariant over
time.
Reproducibility (intercoder consistency) -- Extent to which content classification
produces the same results when the same text is coded by more than one person.
Comparability -- Extent to which the classification of text corresponds to a
standard or norm. Except for training purposes, standard codings are
infrequently established for texts.
Validity
EXPERIMENTAL RESEARCH
Experimental design
The paradigm for scientific method in research is the true experiment or randomised
control trial (RCT). Experimental designs are set up to allow the greatest amount of
control possible so that causality may be examined closely.
15
The three essential elements of experimental design are:
Control : The experimenter introduces one or more controls over the experimental
situation.
The classic example is the before-after design or pre-test post-test design. This is
perhaps the most commonly used experimental design. Comparison of pre-test scores
allows the researcher to evaluate how effective the randomisation of the sample is in
providing equivalent groups. The treatment is fully under control of the researcher. The
dependent variable is measured twice during the study (before and after the
manipulation of the independent variable).
In some studies the dependent variable cannot be measured before the treatment. Here
we would use an approach known as the post-test only design. We may also wish to use
this approach where pre-test sensitisation may occur. Subjects’ post-test response may
be partly due to learning from, or as a reaction to, the pre-test. In these instances the
pre-test phase can be eliminated; however doing so removes the possibility of applying
some very powerful statistical analyses.
Quasi-experimental design.
Beyond discovering causal relationships, experimental research further seeks out how
much cause will produce how much effect; in technical terms, how the independent
variable will affect the dependent variable. Experimental research also looks into the
effects of removing something.
16
Treatment and Hypothesis
The term treatment refers to either removing or adding a stimulus in order to measure
an effect. Experimental researchers want to know how varying levels of treatment will
effect what they are studying. As such, researchers often have an idea, or hypothesis,
about what effect will occur when they cause something. Few experiments are
performed where there is no idea of what will happen. From past experiences in life or
from the knowledge we possess in our specific field of study, we know how some
actions cause other reactions. Experiments confirm or reconfirm this fact.
Causality
Experimentation becomes more complex when the causal relationships they seek aren't
clear.
Even though we have tried to remove all extraneous variables, results may appear
merely coincidental. Since we want to prove a causal relationship in which a single
variable is responsible for the effect produced, the experiment would produce stronger
proof if the results were replicated in larger treatment and control groups.
Selecting groups entails assigning subjects in the groups of an experiment in such a way
that treatment and control groups are comparable in all respects except the application
of the treatment. Groups can be created in two ways: matching and randomization.
Matching may be problematic, though, because it can promote a false sense of security
by leading [the experimenter] to believe that [the] experimental and control groups
were really equated at the outset, when in fact they were not equated on a host of
variables. It would be difficult to have equal groupings.
1. Reliability
Are the results of the experiment repeatable?
If the experiment were done the same way again, would it produce the same
results?
Reliability is a requirement before the validity of the experiment can be
established.
2. Internal validity
Accuracy or truth-value
Does the research design lead to true statements?
17
Did the independent variable cause the effects in the dependent variable?
In experimental research, this usually means eliminating alternative hypotheses.
In the example evaluating a new psychotherapy, the issue of internal validity is
whether the psychotherapy really was the causal factor in improving
participants' mental health.
3. External validity
Generalizability
Can the results be applied in another setting or to another population of research
participants?
Hypotheses
Then we hold the possibility that there is no effect of the independent variable on the
dependent variable or that the differences observed are due to chance only. This second
hypothesis is the null hypothesis. The first step in experimental research, then, is ruling
out chance. Put another way, we set up an experimental design that will allow us to
reject the null hypothesis. If we can confidently reject the null hypothesis, then we gain
confidence in the research hypothesis.
Types of Hypotheses
2. Null hypothesis states that differences are due to chance or that there are no
differences between treatments (used in statistical analysis).
But what if we don't have all of those? Do we still have an experiment? Not a true
experiment in the strictest scientific sense of the term, but we can have a quasi-
experiment, an attempt to uncover a causal relationship, even though the researcher
cannot control all the factors that might affect the outcome.
18
treatment and control groups may not be randomized or matched, or there may be no
control group. The researcher is limited in what he or she can say conclusively.
Since we're mentioning the subject of statistics, note that experimental or quasi-
experimental research cannot state beyond a shadow of a doubt that a single cause will
always produce any one effect. They can do no more than show a probability that one
thing causes another. The probability that a result is the due to random chance is an
important measure of statistical analysis and in experimental research.
19
SURVEY RESEARCH
Survey research is primarily quantitative, but qualitative methods are sometimes used,
too.
Once in a while, a researcher may be able to gather data from all members of a
population. For example, if we want to know what a neighbourhood thinks about a local
land use issue, we may be able to measure all residents of the neighbourhood if it is not
too big. However, most of the time, the population is so large that researchers must
sample only a part of the population and make conclusions about the population based
on the sample. Because of this, gaining a representative sample is crucial in survey
research.
Descriptive designs are also known as ex post facto studies. This literally means "from
after the fact". The term is used to identify that the research in question has been
conducted after the variations in the independent variable has occurred naturally.
20
Two kinds of design may be identified in ex post facto research – the correlational
study and the criterion group (comparative) study. The basic purpose of the
correlational study is to determine the relationship between variables. However the
significant difference from experimental and quasi-experimental design is that causality
cannot be established due to lack of manipulation of independent variables.
Correlation does not prove causation. Examples include many studies of lung cancer.
The researcher begins with a sample of those who have already developed the disease
and a sample of those who have not. The researcher then looks for differences between
the two groups in antecedents, behaviours or conditions such as smoking habits.
In the criterion group (comparative) study, the researcher sets out to discover
possible causes for a phenomenon being studied by comparing the subjects in which the
variable is present with similar subjects in whom it was absent. If, for example, a
researcher chooses a design to investigate factors contributing to teacher effectiveness,
the criterion group, the effective teachers, and its counterpart, a group not showing
the characteristics of the criterion group, are identified by measuring the differential
effects of the groups on classes of children. The researcher may then examine some
variable or event, i.e. the training, skills and personality of the groups to discover what
might ‘cause’ only some teachers to be effective.
Correlational research designs are founded on the assumption that reality is best
described as a network of interacting and mutually-causal relationships. Everything
affects--and is affected by--everything else. This web of relationships is not linear, as in
experimental research.
Thus, the dynamics of a system--how each part of the whole system affects each other
part--is more important than causality. As a rule, correlational designs do not indicate
causality. However, some correlational designs such as path analysis and cross-lagged
panel designs, do permit causal statements. Correlational research is quantitative.
Bivariate correlation
The relationship between two variables is measured. The relationship has a degree and
a direction.
The degree of relationship (how closely they are related) is usually expressed as a
number between -1 and +1, the so-called correlation coefficient. A zero correlation
indicates no relationship. As the correlation coefficient moves toward either -1 or +1,
the relationship gets stronger until there is a "perfect correlation" at either extreme.
The direction of the relationship is indicated by the "-" and "+" signs. A negative
correlation means that as scores on one variable rise, scores on the other decrease. A
21
positive correlation indicates that the scores move together, both increasing or both
decreasing.
A student's grade and the amount of studying done, for example, are generally
positively correlated. Stress and health, on the other hand, are generally negatively
correlated.
If there is a correlation between two variables, and we know the score on one, the
second score can be predicted. Regression refers to how well we can make this
prediction. As the correlation coefficients approach either -1 or +1, our predictions get
better. For example, there is a relationship between stress and health. If we know my
stress score, we can predict my future health status score.
Multiple regression
This extends regression and prediction by adding several more variables. The
combination gives us more power to make accurate predictions. What we are trying to
predict is called the criterion variable. What we use to make the prediction, the known
variables, are called predictor variables.
If we know not only our stress score, but also a health behaviour score (how well we
take care of ourselves) and how our health has been in the past (whether we are
generally healthy or ill), we can more closely predict our health status. Thus, there are 3
predictors--stress, health behaviour, and previous health status--and one criterion--
future health.
SAMPLING
IN QUANTITATIVE AND QUALITATIVE RESEARCH
Sampling refers to drawing a sample (a subset) from a population (the full set). The
usual goal in sampling is to produce a representative sample (i.e., a sample that is
similar to the population on all characteristics, except that it includes fewer people
because it is a sample rather than the complete population). Metaphorically, a perfect
representative sample would be a "mirror image" of the population from which it was
selected (again, except that it would include fewer people).
22
Terminology Used in Sampling
The two major types of sampling in quantitative research are random sampling and
nonrandom sampling.
The former produces representative samples.
The latter does not produce representative samples.
Simple random samples are not the only equal probability sampling method (EPSEM). It
is the most basic and well known, however. How do we draw a simple random sample?
One way is to put all the names from your population into a hat and then select a subset
(e.g., pull out 100 names from the hat). Sampling experts recommend random sampling
"without replacement" rather than random sampling "with replacement" because the
former is a little more efficient in producing representative samples (i.e., it requires
slightly fewer people and is therefore a little cheaper).
Systematic Sampling
23
Systematic sampling is the second type of random sampling.
It is an equal probability sampling method (EPSEM).
Remember simple random sampling was also an EPSEM.
When we get to the end of our sampling frame we will have all the people to be included
in our sample. One potential (but rarely occurring) problem is called periodicity (i.e.,
there is a cyclical pattern in the sampling frame). It could occur when we attach several
ordered lists to one another (e.g., if we take lists from multiple teachers who have all
ordered their lists on some variable such as IQ). On the other hand, stratification within
one overall list is not a problem at all (e.g., if we have one list and have it ordered by
gender or by IQ). Basically, if we are attaching multiple lists to one another, there could
be a problem. It would be better to reorganize the lists into one overall list (i.e.,
sampling frame).
Stratified Random Sampling
24
For disproportional stratified sampling, we might randomly select 50 females
and 50 males from the population.
In this type of sampling we randomly select clusters rather than individual type units in
the first stage of sampling. A cluster has more than one unit in it (e.g., a school, a
classroom, a team). We discuss two types of cluster sampling in the chapter, one-stage
and two-stage (note that more stages are possible in multistage sampling).
The other major type of sampling used in quantitative research is nonrandom sampling
(i.e., when we do not use one of the random sampling techniques). There are four main
types of nonrandom sampling:
The first type of nonrandom sampling is called convenience sampling (i.e., it
simply involves using the people who are the most available or the most easily
selected to be in our research study).
The second type of nonrandom sampling is called quota sampling (i.e., it
involves setting quotas and then using convenience sampling to obtain those
quotas). A set of quotas might be given to us as follows: find 25 African American
males, 25 European American males, 25 African American females, and 25
European American females. We use convenience sampling to actually find the
people, but we must make sure we have the right number of people for each
quota.
The third type of nonrandom sampling is called purposive sampling (i.e., the
researcher specifies the characteristics of the population of interest and then
25
locates individuals who match those characteristics). For example, we might
decide that we want to only include "boys who are in the 7th grade and have
been diagnosed with ADHD" in our research study. We would then, try to find 50
students who meet our "inclusion criteria" and include them in our research
study.
The fourth type of nonrandom sampling is called snowball sampling (i.e., each
research participant is asked to identify other potential research participants
who have a certain characteristic). We start with one or a few participants, ask
them for more, find those, ask them for some, and continue until we have a
sufficient sample size. This technique might be used for a hard to find population
(e.g., where no sampling frame exists). For example, we might want to use
snowball sampling if we wanted to do a study of people in our city who have a lot
of power in the area of educational policy making (in addition to the already
known positions of power, such as the school board and the school system
superintendent).
Would we like to know the answer to the question "How big should my sample be?"
Try to get as big of a sample as we can for our study (i.e., because the bigger the
sample the better).
If our population is size 100 or less, then include the whole population rather
than taking a sample (i.e., don't take a sample; include the whole population).
Look at other studies in the research literature and see how many they are
selecting.
For an exact number, just look at the table at the end of this handout which
shows recommended sample sizes.
In particular, note that we will need larger samples under these circumstances:
When the population is very heterogeneous.
When we want to breakdown the data into multiple categories.
When we want a relatively narrow confidence interval (e.g., note that the
estimate that 75% of teachers support a policy plus or minus 4% is narrower
than the estimate of 75% plus or minus 5%).
When we expect a weak relationship or a small effect.
When we use a less efficient technique of random sampling (e.g., cluster
sampling is less efficient than proportional stratified sampling).
When we expect to have a low response rate. The response rate is the percentage
of people in our sample who agree to be in our study.
N stands for the size of the population. n stands for the size of the recommended
sample. The sample sizes are based on the 95 percent confidence level.
N n N n N n N n N n
26
10 10 110 86 300 169 950 274 4,500 354
15 14 120 92 320 175 1,000 278 5,000 357
20 19 130 97 340 181 1,100 285 6,000 361
25 24 140 103 360 186 1,200 291 7,000 364
30 28 150 108 380 191 1,300 297 8,000 367
35 32 160 113 400 196 1,400 302 9,000 368
40 36 170 118 420 201 1,500 306 10,000 370
45 40 180 123 440 205 1,600 310 15,000 375
50 44 190 127 460 210 1,700 313 20,000 377
55 48 200 132 480 214 1,800 317 30,000 379
60 52 210 136 500 217 1,900 320 40,000 380
65 56 220 140 550 226 2,000 322 50,000 381
70 59 230 144 600 234 2,200 327 75,000 382
75 63 240 148 650 242 2,400 331 100,000 384
80 66 250 152 700 248 2,600 335 250,000 384
85 70 260 155 750 254 2,800 338 500,000 384
90 73 270 159 800 260 3,000 341 1,000,000 384
95 76 280 162 850 265 3,500 346 10,000,000 384
100 80 290 165 900 269 4,000 351 500,000,00 384
0
Krejcie, R. V. and Morgan, D. W. 1970. Determining Sample Size for Research Activities.
Educational and Psychological Measurement, 30, 607-610
(In Isaac, S. and Michael, W. B. 1981. Handbook in Research and Evaluation. San Diego:
EdITS publishers)
27
DATA COLLECTION INSTRUMENTS
Tests
Tests are commonly used in research to measure achievement, and performance.
Sometimes, a researcher must develop a new test to measure the specific knowledge,
skills, behavior, or cognitive activity that is being studied.
Questionnaires
28
A questionnaire is a self-report data collection instrument that is filled out by research
participants. Questionnaires are usually paper-and-pencil instruments, but they can also
be placed on the web for participants to go to and “fill out.” Questionnaires are
sometimes called survey instruments, but the actual questionnaire should not be called
“the survey.” The word “survey” refers to the process of using a questionnaire or
interview protocol to collect data. For example, we might do a survey of teacher
attitudes about inclusion; the instrument of data collection should be called the
questionnaire or the survey instrument.
One of the well-known types is a Likert Scale because the summated rating scale was
pretty much invented by the famous social psychologist named Rensis Likert.
Strengths and Weaknesses of Questionnaires
Strengths of questionnaires
• Good for measuring attitudes and eliciting other content from research
participants.
• Inexpensive (especially mail questionnaires and group administered
questionnaires).
• Can provide information about participants’ internal meanings and ways of
thinking.
• Can administer to probability samples.
• Quick turnaround.
• Can be administered to groups.
• Perceived anonymity by respondent may be high.
• Moderately high measurement validity (i.e., high reliability and validity) for well
constructed and validated questionnaires.
• Closed-ended items can provide exact information needed by researcher.
• Open-ended items can provide detailed information in respondents’ own words.
• Ease of data analysis for closed-ended items.
• Useful for exploration as well as confirmation.
Weaknesses of questionnaires
• Usually must be kept short.
• Reactive effects may occur (e.g., interviewees may try to show only what is socially
desirable).
• Nonresponse to selective items.
• People filling out questionnaires may not recall important information and may
lack self-awareness.
• Response rate may be low for mail and email questionnaires.
• Open-ended items may reflect differences in verbal ability, obscuring the issues of
interest.
• Data analysis can be time consuming for open-ended items.
• Measures need validation.
29
Interviews
In an interview, the interviewer asks the interviewee questions (in-person or over the
telephone).
- Anything else?
- Any other reason?
- What do you mean?
Quantitative interviews:
Qualitative interviews
Strengths of interviews
• Good for measuring attitudes and most other contents of interest.
• Allows probing and posing of follow-up questions by the interviewer.
30
• Can provide in-depth information.
• Can provide information about participants’ internal meanings and ways of
thinking.
• Closed-ended interviews provide exact information needed by researcher.
• Telephone and e-mail interviews provide very quick turnaround.
• Moderately high measurement validity (i.e., high reliability and validity) for well
constructed and tested interview protocols.
• Can use with probability samples.
• Relatively high response rates are often attainable.
• Useful for exploration as well as confirmation.
Weaknesses of interviews
• In-person interviews usually are expensive and time consuming.
• Reactive effects (e.g., interviewees may try to show only what is socially desirable).
• Investigator effects may occur (e.g., untrained interviewers may distort data
because of personal biases and poor interviewing skills).
• Interviewees may not recall important information and may lack self-awareness.
• Perceived anonymity by respondents may be low.
• Data analysis can be time consuming for open-ended items.
• Measures need validation.
Focus Groups
A focus group is a situation where a focus group moderator keeps a small and
homogeneous group (of 6-12 people) focused on the discussion of a research topic or
issue.
• Focus group sessions generally last between one and three hours and they are
recorded using audio and/or videotapes.
• Focus groups are useful for exploring ideas and obtaining in-depth information
about how people think about an issue.
31
• Difficult to generalize results if small, unrepresentative samples of participants are
used.
• May include large amount of extra or unnecessary information.
• Measurement validity may be low.
• Usually should not be the only data collection methods used in a study.
• Data analysis can be time consuming because of the open-ended nature of the data.
Observation
There are two important forms of observation: quantitative observation and qualitative
observation.
1) Quantitative observation involves standardization procedures, and it produces
quantitative data.
• Complete participant (i.e., becoming a full member of the group and not informing
the participants that we are studying them).
• Participant-as-Observer (i.e., spending extensive time "inside" and informing the
participants that we are studying them).
• Observer-as-Participant (i.e., spending a limited amount of time "inside" and
informing them that we are studying them).
32
• Complete Observer (i.e., observing from the "outside" and not informing that
participants that we are studying them).
Secondary/Existing Data
Secondary data (i.e., data originally used for a different purpose) are contrasted with
primary data (i.e., original data collected for the new research study).
The most commonly used secondary data are documents, physical data, and archived
research data.
2. Physical data (are any material thing created or left by humans that might provide
information about a phenomenon of interest to a researcher).
33
3. Archived research data (i.e., research data collected by other researchers for other
purposes, and these data are save often in tape form or CD form so that others might
later use the data).
CONCEPTS OF MEASUREMENT
Nominal scale
• Nominal measurement consists of assigning items to groups or categories.
• No quantitative information is conveyed and no ordering of the items is implied.
• Examples: religious preference, race, and gender
Ordinal Scale
• Measurements with ordinal scales are ordered.
34
• However, the intervals between the numbers are not necessarily equal.
• For example, on a five-point Likert scale, the difference between 2 and 3 may not
represent the same difference as the difference between 4 and 5.
Interval Scale
• There is no "true" zero point for interval scales.
• On interval measurement scales, one unit on the scale represents the same
magnitude on the trait or characteristic being measured across the whole range
of the scale.
• For example, on an interval scale of temperature, the difference between 10 and
11 degrees centigrade is the same as the difference between 11 and 12 degrees
centigrade.
Ratio scale
• There is a "true" zero point for ratio scales.
• The difference between points on the scale is precise (as in the measurement of
height and weight)
• The scale starts at zero.
• For example, height and weight start at zero (People cannot weigh less than
0.00kg and cannot be less than 0.00mm in length/height.
DESCRIPTIVE STATISTICS
Median
The median is the value of the middle item of a distribution (In the sequence
above there are eleven values and the value occurring in the middle position is
the value "4"; thus 4 is the median average for this sequence).
The median value shows nothing about the other values in the sequence e.g., if
the eight year old was replaced by a 15 year old, and one of the 2 year olds was
replaced by a baby of just 2 days old, the median would not have changed at all.
Mode
The mode (modal value) is the point of maximum frequency (In the sequence of
numbers above, the value "2" occurs most frequently and is thus the modal
value).
35
Sometimes there is more than one peak (If the values above referred to the age
of children, and the eight year old child was replaced by a six year old child, there
would be two modal values, one of the value "2" and the other of the value "6".
The distribution would then be considered "bi-modal").
The modal value does not reflect the extremes of the sequence.
Standard Deviation
Standard Deviation: a measure of spread
A measure of variation that indicates the typical distance between the scores of a
distribution and the mean.
It is determined by taking the square root of the average of the squared
deviations in a given distribution.
It can be used to indicate the proportion of data within certain ranges of scale
values when the distribution conforms closely to the normal curve.
The SD says how far away numbers on a list are from their average.
36
INFERENTIAL STATISTICS
Correlation
t-test
ANOVA
The independent variables are nominal, and the dependent variable is interval.
37
Correlation:
Pearson
One
Predictor Regression
Independent Independent
Groups Samples t-test
Interval
Data
Between Two Dependent
Groups Groups Repeated
Measures t-test
Analysis of
Differences Independent
Independent
Groups
Type of Samples ANOVA
Data
Between Dependent Repeated
Multiple Groups Groups Measures ANOVA
38