When a researcher
chooses to create their own survey instrument, it is appropriate to run an
exploratory factor analysis to assess for potential subscales within the
instrument. However, it is seemingly
unnecessary to run an exploratory factor analysis (EFA) on an already
established instrument. In the case of an already established instrument,
typically, a Cronbach’s alpha is the acceptable way to assess reliability.
Request
To request a blog written on a specific topic, please email James@StatisticsSolutions.com with your suggestion. Thank you!
Showing posts with label Survey instrument. Show all posts
Showing posts with label Survey instrument. Show all posts
Monday, November 12, 2012
Thursday, November 8, 2012
Recoding
Survey items can be worded with a positive or negative direction:
·
Positively worded: e.g., I know that I am
welcomed at my child’s school, I feel that I am good at my job, Having a
wheelchair helps, etc…
·
Negatively worded: e.g., I feel isolated at my
child’s school, I am not good at my job, having a wheelchair is a hindrance,
etc…
·
Likert scaled responses can vary: e.g., 1 =
never, 2= sometimes, 3 = always; OR
1 = strongly disagree, 2 = disagree, 3 = neutral, 4 = agree, 5 = strongly
agree
·
When creating a composite score from specific
survey items, we want to make sure we are looking at the responses in the same
manner. If we have survey items that are
not all worded in the same direction, we need to re-code the responses. E.g.: I
want to make a composite score called “helpfulness” from the following survey
items :
o
5-point Likert scaled, where 5 = always 4 = almost always 3 = sometimes, 2 = almost never 1 = never
1.
I like to tutor at school
2.
I am usually asked by my friends to help with
homework
3.
I typically do homework in a group setting
4.
I do not go over my homework with others
In this example, survey items 1
– 3 are all positively worded, but survey item 4 is not. When creating the composite score, we wish to
make sure that we are examining the coded responses the same way. In this case, we’d have to re-code the
responses to survey item 4 to make sure that all responses for the score
“helpfulness” are correctly interpreted; the recoded responses for survey item
4 are: 1 = always, 2 = almost always, 3 = sometimes, 4 = almost never, 5 =
never.
Now, all responses that are
scored have the same direction and thus, can be interpreted correctly: positive
responses for “helpfulness” have higher values and negative responses for
“helpfulness” have lower values.
·
Also, you may wish to change the number of
responses. For example, you may wish to
dichotomize or trichotomize the responses.
In the example above, you can trichotomize the responses by recoding
responses “always” and “almost always” to 3 = high, “sometimes” to 2 =
sometimes, and “almost never” and “never” to 1 = low. However, please be advised to make sure that
you have sound reason to alter the number of responses.
Tuesday, September 4, 2012
Creating and Validating an Instrument
To determine if an appropriate
instrument is available, a researcher can search literature and commercially
available databases to find something suitable to the study. If it is determined that there are no
instruments available that measure the variables in a study, there are four
rigorous phases for developing an instrument that accurately measures the
variables of interest (Creswell, 2005). Those
four phases are: planning, construction, quantitative evaluation, and
validation. Each phase consists of
several steps that must be taken to fully satisfy the requirements for
fulfilling a phase.
The
first phase is planning and the first step of planning includes identifying the
purpose of the test and the target group.
In this step, the researcher should identify the purpose of the test,
specify the content area to be studied, and identify the target group. The second step of phase one is to, again,
review the literature to be certain no instruments already exist for the
evaluation of the variables of interest.
Several areas to look for existing instruments include the ERIC website (www.eric.ed.gov), Mental Measurements Yearbook (Impara & Plake, 1999), and Tests in Print (Murphy, Impara, &
Plake, 1999). Once the researcher is
certain no other instruments exist, the researcher should review the literature
to determine the operational definitions of the constructs that are to be
measured. This can be an arduous task
because operationalizing a variable does not automatically indicate good
measurement and therefore the researcher must review multiple literatures to
determine an accurate and meaningful construct.
From this information, the researcher should develop open ended
questions to present to a sample that is representative of the target
group. The open ended questions aid the
researcher in determining areas of concern around the constructs to be
measured. The responses to the open
ended questions and the review of the literature should be used in unison to
create and modify accurate measures of the constructs.
The
second phase is construction and it begins with identifying the objectives of
the instrument and developing a table of specifications. Those specifications should narrow the
purpose and identify the content areas.
In the specification process, each variable should be associated with a
concept and an overarching theme (Ford, http://www.blaiseusers.org/2007/papers/Z1%20-%20Survey%20Specifications%20Mgmt%20at%20Stats%20Canada.pdf). Once the table of specification is completed,
the researcher can write the items in the instrument. The researcher must determine the format to be
used, ie. Likert scale, multiple choice, etc.
The format of the questions should be determined by the type of data
that needs to be collected. Depending on
the financial resources of the research project, experts within the field may
be hired to write the items. Once the
items are written, they need to be reviewed for clarity, formatting, acceptable
response options, and wording. After
several reviews of the questions, they should be presented to peers and
colleagues in the format the instrument is to be administered. The peers and colleagues should match the
items with the specification table and if there are not exact matches,
revisions must be made. An instrument is
content valid when the items adequately reflect the process and content
dimensions of the objectives of the instrument (Benson & Clark, 1982). Again, the instrument should be distributed
to a sample that is representative of the target group. This time the group should take the survey
and critique the quality of the individual items and overall instrument.
Phase
three is quantitative evaluation and includes administration of a pilot study
to a representative sample. It may be
helpful to ask the participants for feedback to allow for further refinement of
the instrument. The pilot study provides
quantitative data that the researcher can test for internal consistency by
conducting Cronbach’s alphas. The
reliability coefficient can range from 0.00 to 1.00, with values of 0.70 or
higher indicating acceptable reliability (George and Mallery, 2003). If the instrument is going to be used to
predict future behavior, the instrument needs to be administered to the same
sample at two different time periods and the responses will need to be
correlated to determine if there is concurrent validity. These measurements can be examined to aid the
researcher in making informed decisions about revisions to the instrument.
Phase
four is validation. In this phase the
researcher should conduct a quantitative pilot study and analyze the data. It may be helpful to ask the participants for
feedback to allow for further refinement of the instrument. The pilot study provides quantitative data
that the researcher can test for internal consistency by conducting Cronbach’s
alphas. To establish validity, the
researcher must determine which concept of validity is important. The three types of validity include content,
criterion-related, and construct. Content
validity is the extent to which the questions on a survey are representative of
the questions that could be asked to assess a particular construct. To examine content validity, the researcher
should consult two to three experts.
Criterion-referenced validity is used when the researcher wants to
determine if the scores from an instrument are a good predictor of an expected
outcome. In order to assess this type of
validity, the researcher must be able to define the expected outcome. A correlation coefficient of a .60 or above
will indicate a significant, positive relationship (Creswell, 2005). Construct validity is established by
determining if the scores recorded by an instrument are meaningful,
significant, useful, and have a purpose.
In order to determine if construct validity has been achieved, the scores
need to be assessed statistically and practically. This can be done by comparing the
relationship of a question from the scale to the overall scale, testing a
theory to determine if the outcome supports the theory, and by correlating the
scores with other similar or dissimilar variables. The use of similar instruments is referred to
as convergent validity and the use of dissimilar instruments is divergent
validity.
References
Creswell,
J. W. (2005). Educational research:
Planning, conducting, and evaluating quantitative and qualitative research
(2nd ed.). Upper Saddle River, NJ: .Pearson Education, Inc.
George,
D. & Mallery, P. (2003). SPSS for
Windows step by step: a simple guide and reference, 11.0 update (4th
ed.). Boston, MA: Allyn and Bacon.
Murphy,
L. L., Impara, J. C., & Plake, B. S. (Eds.). (1999)
Subscribe to:
Posts (Atom)