Академический Документы
Профессиональный Документы
Культура Документы
Hi, Marcie,
Do you have examples for reporting the results from Likert Scale
survey? I am looking for a simple and straightforward example for 690
students.
I have my cultural studies data and report, but my study was a little
complicated.
I'm not sure what you mean by a weighted mean--in this situation. What would be
the weighting?
However, plenty of people treat ordinal data as continuous --- especially when
survey data are supplemented with other measures.
Attached is a small survey report I just finished --- part of a blended learning
initiative I'm working on for Academic Affairs and ITS. You're welcome to use it!
M.
A typical test item in a Likert scale is a statement, the respondent is asked to indicate
their degree of agreement with the statement. Traditionally a five-point scale is used,
however many psychometricians advocate using a seven or nine point scale.
2
1. Strongly disagree
2. Disagree
3. Neither agree nor disagree
4. Agree
5. Strongly agree
Likert scaling is a bipolar scaling method, measuring either positive and negative
response to a statement. Sometimes Likert scales are used in a forced choice method
where the middle option of "Neither agree nor disagree" is not available. Likert scales
may be subject to distortion from several causes. Respondents may avoid using
extreme response categories (central tendency bias); agree with statements as
presented (acquiescence response bias); or try to portray themselves or their group in
a more favorable light (social desirability bias).
After the questionnaire is completed, each item may be analyzed separately or item
responses may be summed to create a score for a group of items. Hence, Likert scales are
often called summative scales.
Responses to a single Likert item are normally treated as ordinal data, because, especially
when using only five levels, one cannot assume that respondents perceive the difference
between adjacent levels as equidistant. When treated as ordinal data, Likert
responses can be analyzed using non-parametric tests, such as the Mann-
Whitney test, the Wilcoxon signed-rank test, and the Kruskal-Wallis test.[1]
When responses to several Likert items are summed, they may be treated as interval data measuring a latent
variable. If the summed responses are normally distributed, parametric statistical tests such as the analysis
of variance can be applied.
Examples:
• Attitudes toward Computer (20 questions, but Each participant gets one score,
summed)
• Some of the personality test (same way)
Data from Likert scales are sometimes reduced to the nominal level by combining all
agree and disagree responses into two categories of "accept" and "reject". The Cochran
Q, or McNemar-Test are common statistical procedures used after this transformation.
3
Data set: the one posted on Course Website: Cultural differences in online Learning
Data: (Part of the Raw Data-SurveyMonkey will give you a numerical version)
Culture QS
Open-Ended Q1Wisd Q2Resp Q3Equ Q4rulescond Q5expe Q6formalman
Response om ect al uct rts ner
Chinese 2 1 2 2 1 5
Chinese 2 2 2 1 1 4
Chinese 2 2 2 2 2 2
Chinese 1 1 2 1 1 2
Chinese 1 1 2 2 1 2
Chinese 3 1 2 2 4 4
Chinese 1 1 1 1 2 4
American 2 2 4 2 1 3
American 1 1 3 1 2 2
American 2 4 4 2 3 4
American 1 1 4 1 1 1
american 2 2 3 1 1 3
American 1 1 2 2 1 4
American 2 2 3 2 2 2
american 2 2 2 2 3 5
American 2 2 4 1 2 1
American 2 2 1 4 1 4
1) Participants’ perceptions on teacher and teaching in general (pre-survey): Item 4 on the pre-
survey assessed participants’ perceptions and expectations on teacher and teaching in general.
The three questions that are closely related to sense of Power Distance were analyzed
inferentially with the Kruskal-Wallis Analysis of Variance test, with cultural identity being the
independent variable. The results indicate that:
a) there were significant differences in participants’ perceptions about being equal with their
instructor. The Korean group had the highest mean rank (45.53) on a scale of 1 (strongly agree) to
5 (strongly disagree). By contrast, the Anglo-American group had the lowest mean rank (29.77)
and therefore perceived their instructors more as equals.
b) There was no significant difference in participants’ perceptions about rules of conduct in
online classes. The Chinese group had the lowest mean rank (29.36), an indication of a stronger
agreement about implementing specific rules of conduct. This result aligned with some of their
narrative comments about “feeling lost” and hoping for more guidance.
And c) There was highly significant differences in their perceptions on course conduct. Again the
Chinese had the lowest mean rank, an indication of a stronger agreement about conducting
courses in a formal manner.
2) Post-survey: approaching superior and peer when completing individual assignments and
team work: Other Responses to the post survey that reflect the impact of Power Distance include:
a) Learners’ comfort level in approaching the instructor/facilitator/TA for help with individual
assignments and/or teamwork; and b) Their comfort level in approaching the peers for help with
individual assignments and/or teamwork. Participants rated their comfort level from very
comfortable (1), to somewhat comfortable (2), uncomfortable (3), and to very uncomfortable (4).
The lower their mean rating, the higher their comfort level. Kruskal-Wallis Analysis of Variance
was used again to compare the mean differences in participants’ ranking of comfort level in
approaching “superior” or their peers, when completing individual assignments and team work if
applicable.
[Note: Because the regular Mean of Likert Scale does not make much sense, I skipped the
descriptive Analysis and Went right into Inferential Analysis—Analysis of Variance using
the Non-Parametric Kruskal-Wallis statistic]
The high Kruskal-Wallis statistic (24.8) and the small p value (<0.0001) again indicates highly
significant difference in participants’ comfort level in approaching peers for help with team work.
The Korean group (mean rank = 48.75) contributed greatly to this difference. However, the
statistical power might have been reduced in this test because of the 17 missing rating values
from the Korean group. As mentioned in the curriculum analysis, many of the Korean courses did
not involve team work and many chose “non applicable” for this survey question.
7
Summary: Influence of power distance evidenced by the four tests: Conforming to the existing
findings about Power Distance, the American group (mainly Anglo-American) had the lowest
PDI score, while the Chinese group had the highest PDI score. Possibly because of their sense of
PDI, the American group felt the most comfortable in approaching their instructors for help,
while the Korean group felt most uncomfortable in doing so. Chinese students, because of their
large class size, did not have much opportunity to interact with the instructors. Still, their reported
comfort level in approaching the instructors was low. As to approaching their peers for help, the
Chinese group felt the most comfortable in completing both individual assignments and team
work, the American group felt comfortable, while the Korean group felt the least comfortable in
completing both individual assignments and teamwork. Again, the Koreans’ cultural perceptions
on CMC might have influenced their ratings here. As some of the Korean participants
commented, peers or classmates online can be “strangers.” As to the high comfort level of the
Chinese, it is worth noting that most of these Chinese students worked in self-formed teams and
they therefore were comfortable about approaching their peers for help.
The four Kruskal-Wallis analyses on the post-survey items had revealing results. Although there
was no significant difference in the three groups’ comfort level in approaching superiors for help
with team work, there were significant differences in their rating for approaching superiors in
individual assignments, and there were highly significant differences in their levels of comfort in
approaching peers for help with individual assignments and with team work. Power Distance
indeed affected students’ ways in approaching instructors and their peers. By contrast, individuals
were able to overcome their sense of Power Distance when working as a group. In other words,
individuals became “braver” when working as a team to approach their instructors for help.
Generating the Items. next, you have to create the set of potential scale items.
These should be items that can be rated on a 1-to-5 or 1-to-7 Disagree-Agree
response scale. Sometimes you can create the items by yourself based on your
intimate understanding of the subject matter. But, more often than not, it's helpful
to engage a number of people in the item creation step. For instance, you might
use some form of brainstorming to create the items. It's desirable to have as
large a set of potential items as possible at this stage, about 80-100 would be
best.
Rating the Items. The next step is to have a group of judges rate the items.
Usually you would use a 1-to-5 rating scale where:
8
Administering the Scale. You're now ready to use your Likert scale. Each
respondent is asked to rate each item on some response scale. For instance,
they could rate each item on a 1-to-5 response scale where:
1. = strongly disagree
2. = disagree
3. = undecided
4. = agree
5. = strongly agree
There are a variety possible response scales (1-to-7, 1-to-9, 0-to-4). All of these
odd-numbered scales have a middle value is often labeled Neutral or Undecided.
It is also possible to use a forced-choice response scale with an even number of
responses and no middle neutral or undecided choice. In this situation, the
respondent is forced to decide whether they lean more towards the agree or
disagree end of the scale for each item.
The final score for the respondent on the scale is the sum of their ratings for all of the
items (this is why this is sometimes called a "summated" scale). On some scales, you will
have items that are reversed in meaning from the overall direction of the scale. These are
called reversal items. You will need to reverse the response value for each of these
items before summing for the total. That is, if the respondent gave a 1, you make it a 5; if
they gave a 2 you make it a 4; 3 = 3; 4 = 2; and, 5 = 1.
Here's an example of a ten-item Likert Scale that attempts to estimate the level of
self esteem a person has on the job. Notice that this instrument has no center or
neutral point -- the respondent has to declare whether he/she is in agreement or
disagreement with the item.
INSTRUCTIONS: Please rate how strongly you agree or disagree with each of
the following statements by placing a check mark in the appropriate box.
a type of survey question where respondents are asked to rate the level at which they
agree or disagree with a given statement. For example:
The DESCRIPTIVES procedure in SPSS produces means and standard deviations for
variables. It also prints the minimum and maximum value. Likert scale questions are
appropriate to print means for since the number that is coded can give us a feel for
which direction the average answer is. The standard deviation is also important as it
10
give us an indication of the average distance from the mean. A low standard deviation
would mean that most observations cluster around the mean. A high standard deviation
would mean that there was a lot of variation in the answers. A standard deviation of 0 is
obtained when all responses to a question are the same. The following code produces
descriptive statistics of columns 1 to 20. The minimum and maximum value tell us the
range of answers given by our survey population.
descriptives
variables = q1 to q20
Valid
Variable Mean Std Dev Minimum Maximum N Label