Presentation of Selected Research Methodology

Transcription

Presentation of Selected Research Methodology
Presentation of Selected
Research Methodology
Activity – Evaluating and
Interpreting Research Literature

After reading the materials titled “Evaluating and
Interpreting Research Literature, answer the
following questions:
1.
Have any of your peers, colleagues, or instructors ever stated
that a study “proves” something? If so, briefly describe what he
or she said, and in light of reading the materials provided, would
you be cautious about believing such a statement? Why?
(answer in one-two paragraphs)
According to the reading materials, what is the primary and
secondary purpose for preparing a literature review?
What do quantitatively oriented researchers emphasize when
sampling that qualitatively researchers do not emphasize?
Name two examples of common sampling flaws.
2.
3.
4.
Activity – Evaluating and
Interpreting Research Literature
1.
2.
3.
4.
5.
Name a trait, other than the ones mentioned in this chapter that
you think are inherently difficult to measure. Why?
Briefly explain why a highly reliable measuring instrument can
be invalid.
If a common well-known IQ scale is considered to have
adequate reliability and validity, does this mean the scale has
no flaws? If not, briefly explain why.
To study causality, what do researchers need to do? Why?
If a difference is statistically significant, does this mean the
difference is large? If not, what does the fact that a difference is
statistically significant tell you? What else can you look at to
indicate the magnitude of a difference?
Review
 In your last assignment you developed a
research question
 Your research question guides your design for
the research study methodology
 This session will help you better formulate and
articulate the research design and
methodology section of your paper
Brief Description of Methodology
and Research Design
 This section will define how you are going to
address the research question or questions
 You will be able to…
 Present an overview of the methodology
 Describe the appropriateness of the methodology
 Explain the rationale for selecting the methodology
Research Design Overview





Nonexperimental Research Designs
Quasi-experimental Research Designs
Experimental Research Designs
Qualitative Research Designs
Program Evaluation
Nonexperimental Research Designs
 Descriptive
 Purpose is to create a detailed description of some
phenomenon
 Example: Satisfaction levels of employers with
business graduates’ job skills
 Causal-Comparative
 Purpose is to compare two or more groups in order
to explore possible causes or effects of a
phenomenon
 Example: Effects of type of classroom (inclusion vs.
non-inclusion) on academic achievement
Nonexperimental Research Designs
(cont’d)
 Correlational
 Purpose is to determine the strength and
direction between two variables
Quasi-Experimental Design
Non-Equivalent Control Group
1.
The treatment group is composed of students already in the
program through self- or external selection. Random assignment
to treatment and control groups is not possible. Another group
that is as similar to the treatment group as possible is selected
for a control group. It does not receive the program.
2.
Administer the pre-test to both groups. Initial differences can be
adjusted later by statistical means.
3.
Expose the treatment group to the program while withholding it
from the control group.
4.
Administer the post-test to both groups.
5.
If there is a difference which favors the treatment group, you can
be fairly confident (though less so than with random assignment)
that it was due to the program. If there is no difference, i.e., if the
scores for both groups remained the same or changed equally
(up or down), that indicates the program is probably not
effective.
Quasi-Experimental Designs:
Non-Equivalent Control Group
Treatment Group (T)
Gets Program (X).
Control Group (C )
Does Not.
Treatment Group
Already Determined
T
Pre-Test
Select Similar
Control Group
C
Pre-Test
X
Post-Test
Post-Test
Controlling Non-Equated Variables Through Statistical
Analysis (or Making Non-Equivalent Groups Similar)
Often times we want to evaluate the effectiveness of a
program that is already in place, and we are not able to
construct a treatment and a control group.
For example, suppose we wanted to evaluate the effectiveness
of public schools vs.. private schools on academic
achievement; and we looked at the average NAEP math
scores for 4th grade students in public and private schools and
found the following:
240
230
220
210
200
190
180
Public
Private
But what does the picture look like when we
control for SES?
When we compare public and private students of the
same SES, we find there is little difference in their
achievement. But because there are more high SES
students in private schools, the overall comparison is
misleading. (For the precise data on these comparisons, see
Lubienski and Lubienski, “A New Look at Public and Private Schools.”
Phi Delta Kappan, May 2005.)
300
250
200
Public
Private
150
100
50
0
Low SES
Mid SES
High SES
Quasi-Experimental Design:
Interrupted Time Series Design: multiple historical
measures only on a treatment group before and after
its exposure to the program.
In situations where a control group is not possible, if (1) data on
the treatment group can be obtained for several periods both
before and after the participants are exposed to the program,
and (2) there is change in scores immediately following the
implementation of the program, and (3) there is a continuation
of the change in subsequent time periods, that is considered
good evidence that the intervention produced the change.
Time
1
Experimental Group
0
2
0
3
0
X
4
X
5
X
6
New program (X) is introduced.
X
Campbell’s Example of the
Interrupted Time Series Design—1
Decline in Highway Deaths per Million
Vehicle Miles Before and After Crackdown
35
30
25
20
15
10
5
Year -2
Year -1
Crackdown
Year
Year +1
Year +2
Year +3
Campbell’s Example of the
Interrupted Time Series Design—2
Decline in Highway Deaths per Million
Vehicle Miles Before and After Crackdown
35
30
25
20
15
10
5
Year -2
Year -1
Crackdown
Year
Year +1
Year +2
Year +3
Simple Before-After
Scores
60
50
40
30
20
10
0
Program
1
2
3 Testing Times
4
5
6
Time Series Showing No Program Impact
Program
80
60
40
20
0
1
2
3
4
5
6
5
6
Time Series Showing Program Impact
60
Program
40
20
0
1
2
3
4
Experimental Research Designs
 Involve the introduction of an intervention
by the researcher to determine a causeand-effect relationship
 Strongest type of design (pre  INT 
post)!
 To yield valid findings, these studies must
be rigorous!
Experimental Research
Designs (cont’d)
 Single-Group Designs
 All individuals in the study receive the treatment
 Example: pre  WebAchiever  post
 Threats to validity
 Control-Group Designs
 Strongest type of design!!
 Experimental group: pre  INT  post
 Control group: pre  NO  post
Experimental Research
Designs (cont’d)
 Quasi-Experimental Designs
 Random assignment of subjects is not
possible (e.g., using an entire classroom,
etc.)
 Biggest problem
Experimental Research
Designs (cont’d)
 Single Case Designs
 Involves the intense study of one individual
(or more than one individual treated as
single group).
Validity Issues
 Can the change in the posttest be
attributed only to the experimental
treatment that was manipulated by the
researcher?
 Must be able to control extraneous
variables that could have undue
influence!
Validity Issues (cont’d)
 Internal Validity
 The extent to which to which extraneous
variables have been controlled by the
researcher, so that any observed effect can
be attributed to the treatment variable.
 External Validity
 Extent to which the findings can be applied
to individuals and settings beyond those
studied
Qualitative Research Designs
 Case Study
 Researcher collects intensive data about
particular instances of a phenomenon and
seek to understand each instance in its own
terms and in its own context
 Historical Research
 Understanding the present condition by
shedding light on the past
Program Evaluation
 Making judgments about the merit, value,
or worth of a program
 A valuable tool in program management
and policy analysis
 Usually initiated by someone’s need for a
decision to be made concerning policy,
management, or political strategy.
Components Addressed in
Research Methodology
(See Syllabus for Details)






Participants
Instruments
Procedures
Limitations
Anticipated Outcomes
References
Participants

This section should include the following elements:
(a) the target population or sample to which it is
hoped the findings will be applicable should be
defined, consistent with the Statement of Problem
and the Research Question(s), (b) the population
from which the sample will actually be drawn should
be specified. This should also include demographic
information such as age, gender, ethnicity etc., (c)
procedures for selecting the sample should be
outlined, including justification for the sampling
method, (d) the implications for the generalizability of
findings from the sample to the accessible population
and then to the target population should be
addressed.
Procedures

Procedures – the procedures section will be based
directly on the research questions. That is, this is the
“how-to” section of the study and will introduce the
design of the research and how the data will be
collected based on the questions of interest. This
section should include the approach (i.e., design) to
conducting the research (e.g., experimental, quasiexperimental, survey, historical, or ethnographic) and
the appropriate procedures to be followed. For
example, for an experimental or quasi-experimental
study, the proposal should indicate how participants
will be assigned to treatments and how the research
will be conducted to ensure internal and external
validity. If an evaluation project is proposed, the
model to be followed should be specified.
Instruments
 Examples of data-gathering instruments
include standardized tests, teacher-made
tests, questionnaires, interview guides,
psychological tests, or field-study logs
 Indicate the source (literature citation) of
the instrument and cite appropriately
 Include validity and reliability information
Limitations
 Limitations are conditions, restrictions or
constraints that may affect the validity of the
project outcomes
 A limitation is a weakness or shortcoming in
the project that could not be avoided or
corrected and is acknowledged in the final
report
 Common limitations are the lack of reliability of
measuring instruments and the restriction of
the project to a particular organization setting
Anticipated Outcomes
 Description of expected results of the
study
 Detail the importance of conducting the
study as well as possible impact on
practice and theory
Research Question
Activity
 State your research question
 What are the key variables?
 What are the Conceptual definitions of
your key variables?
 What are the Operational definitions of
your key variables?
Conceptual and Operational
Variables
Concepts (or constructs) = terms that refer to
the characteristics of an event, situation, or
group being studied
We need to clearly specify how we define
each and every concept in our research
studies.
Two Types of Definitions
Conceptual Definition
...the “empirical definition” of a construct
Examples of Conceptual
Definitions
Cognitive dissonance = “the unpleasant state
of psychological arousal resulting from an
inconsistency within one’s important attitudes,
beliefs, or behaviors”
Self-esteem = “a person’s overall selfevaluation or sense of self worth”
Aggression = “behavior intended to injure
another”
Stereotype = “a belief about the personal
attributes of a group of people”
Conceptual definitions offer general, abstract
characterizations of psychological constructs.
This is exactly why we need operational
definitions!
Operational definition
...the precise specification of how a
concept is measured or manipulated in a
particular study
Operational Definitions
How can we operationalize “aggression”?
-punching another’s face?
-hitting another’s arm?
-spreading rumors about another?
-verbally insulting another?
-throwing glass at another?
-etc.
* The more specific, the better.
Operational Definitions
Why are they necessary/important?
They force us to think carefully and empirically
in precise and specific terms.
They make the concept public; they allow for
replication.
Measurable.