SlideShare une entreprise Scribd logo
1  sur  43
Research
Methods for
Organizational
Studies
- 2ed, 2004
LIU YING
copyrightⓒ 2013 All rights reserved by
LIU YING
1
3. Measurement Foundations
Validity and Validation
• Construct Definitions
– Construct Domain
– Nomological Networks
• Construct Definition Illustration
• Construct Validity Challenges
– Random Errors
– Systematic Errors
– Scores Are Critical
• Construct Validation
– Content Validity
– Reliability
• Types of Reliability
• Reliability and Construct Validity
– Convergent Validity
– Discriminant Validity
– Criterion-Related Validity
– Investigating Nomological Networks
• Summary
• For Review
– Terms to Know
– Things to Know
copyrightⓒ 2013 All rights reserved by
LIU YING
2
Construct Definitions
Measurement produces numerical values that are designed to
summarize characteristics of cases under study. A measure
is an instrument to record such scores. Construct valid measures
yield numerical values that accurately represent the
characteristic.
The most useful conceptual definitions have two elements.
 Construct Domain: First, useful conceptual definitions
identify the nature of the construct by specifying its meaning.
This element explains what a researcher has in mind for
the construct;
 Nomological Networks: The second element also should
specify how the construct of interest relates to other
constructs in a broader web of relationships
copyrightⓒ 2013 All rights reserved by
LIU YING
3
Construct Definition Illustration
copyrightⓒ 2013 All rights reserved by
LIU YING
4
Construct Validity Challenges
• Construct Validity Challenges
– Random Errors
• The more items, the more successfully this type
of random error is controlled.
– Systematic Errors
• A measure is Contaminated or Deficient
– Scores Are Critical
• Systematic variance is sometimes called true
score variance
copyrightⓒ 2013 All rights reserved by
LIU YING
5
ScoresAreCritical
copyrightⓒ 2013 All rights reserved by
LIU YING
6
Scores Are Critical
• Exhibit 3.2 also shows three sources of construct
invalidity:
– 1. Scores on a measure may be less than construct
valid because of deficiency. In the example, observed
scores are deficient because the measure does not
capture satisfaction with computer durability.
– 2. Scores may be less than construct valid because of
systematic contamination. In the example, observed
scores are contaminated because the measure includes
satisfaction with the purchasing experience.
– 3. Finally, scores on a measure are less than construct
valid to the extent that they include random errors.
copyrightⓒ 2013 All rights reserved by
LIU YING
7
Construct Validation
Because construct validity cannot be assessed directly, it cannot
be directly established. However, there are procedures available
to help researchers develop construct valid measures and to
help evaluate those measures once developed. Six such
procedures are described here.
• Content Validity
• Reliability
– Types of Reliability
– Reliability and Construct Validity
• Convergent Validity
• Discriminant Validity
• Criterion-Related Validity
• Investigating Nomological Networks
copyrightⓒ 2013 All rights reserved by
LIU YING
8
Content Validity
• A measure is content valid when its items art judged to accurately
reflect the domain of the construct as defined conceptually.
Content validation ordinarily has experts in the subject matter of
interest provide assessments of content validity.
• As a part of development, the researcher has a panel of experts in
computer programming review the measure for its content. Content
validation of this sort provides information about potential
systematic errors in measures.
• Content validation can help improve the items that form a
measure. Nevertheless, it is not sufficient for construct validity. In
particular, content validation procedures may not provide
information about potential deficiency, nor can subject matter
experts provide much information about random errors that may
be present in the scores that are obtained.
copyrightⓒ 2013 All rights reserved by
LIU YING
9
Reliability
• Types of Reliability
– Internal consistency: reliability refers to the similarity of item
scores obtained on a measure that has multiple items. It can be
assessed when items are intended to measure a single construct. In
the computer example (Exhibit 3.1), satisfaction is measured with six
items. The internal consistency of that questionnaire can be estimated if
scores are available from a set of cases.
– Interrater reliability: indicates the degree to which a group of
observers or raters provide consistent evaluations. For example, the
observers may be a group of international judges who are asked to
evaluate ice skaters performing in a competition. In this case, the judges
serve as measurement repetitions just as the items serve as repetitions
in the computer satisfaction questionnaire. High reliability is obtained
when the judges agree on the evaluation of each
– Stability reliability: refers to the consistency of measurement results
across time. Here measurement repetitions refer to time periods (a measure
is administered more than once).
• Reliability and Construct Validity
*** There are three common contexts in which researchers seek to assess the reliability of measurement.
Chapter 17 describes statistical procedures for estimating these types of reliability.
copyrightⓒ 2013 All rights reserved by
LIU YING
10
Reliability and Construct Validity
Reliability speaks only to a measure's freedom from
random errors. It does not address systematic errors
involving contamination or deficiency. Reliability is
thus necessary for construct validity but not
sufficient. It is necessary because unreliable
variance must be construct invalid. It is not sufficient
because systematic variance may be contaminated
and because reliability simply does not account for
deficiency. In short, reliability addresses only
whether scores are consistent; it does not address
whether scores capture a particular construct as
defined conceptually.
copyrightⓒ 2013 All rights reserved by
LIU YING
11
Convergent Validity
• Convergent validity is present when there is a
high correspondence between scores from two
or more different measures of the same
construct. Convergent validity is important
because it must be present if scores from both
measures are construct valid. But convergent
validity is not sufficient for construct validity any
more than is reliability.
copyrightⓒ 2013 All rights reserved by
LIU YING
12
Convergent Validity
The area crossed with vertical lines shows the
proportion of variance in scores from the two measures
that is convergent. However, only the area also crossed
with horizontal lines shows common construct valid
variance. The area covered only by vertical lines shows
where the two measures share variance that
represents contamination from a construct validity
perspective. Convergent validity also does not address
whether measures are deficient. Nor does it provide
construct validity information about the proportion of
variance in the two measures that do not converge.
Exhibit 3.3 shows that more of the variance unique to
the measure A overlaps with construct variance than
variance from measure B.
Despite these limitations, evidence of convergent
validity is desirable. If two measures that are designed
to measure the same construct do not converge, at
least one of them is not construct valid. Alternatively, if
they do converge, circumstantial evidence is obtained
that they may both be construct valid. Evidence of
convergent validity adds to a researcher's confidence
in the construct validity of measures.
copyrightⓒ 2013 All rights reserved by
LIU YING
13
Discriminant Validity
• Discriminant validity is inferred when scores
from measures of different constructs do not
converge. It thus provides information about
whether scores from a measure of a construct are
unique rather than contaminated by other
constructs.
• Proposed constructs should provide
contributions beyond constructs already in the
research domain. Consequently, measures of
proposed constructs should show evidence of
discriminant validity with measures of existing
constructs.
copyrightⓒ 2013 All rights reserved by
LIU YING
14
Criterion-Related Validity
• Criterion-related validity is present when scores on a
measure are related to scores on another measure that better
reflects the construct of interest. It differs from convergent
validity, where scores from the measures are assumed to be
equivalent representations of the construct. In criterion-
related validity the criterion measure is assumed to have
greater construct validity than the measure being developed
or investigated.
• Why not just use the criterion measure if it has greater
construct validity?
• Historically, researchers used the term to describe
relationships between a construct (represented by the
measure under consideration) and a measure of another
construct that is thought to be conceptually related to the first.
copyrightⓒ 2013 All rights reserved by
LIU YING
15
Investigating Nomological
Networks
• Nomological networks have been described as
relationships between a construct under
measurement consideration and other
constructs.
• Researchers with a conceptual orientation are
interested in the conceptual relationship
between the independent and dependent
variable constructs.
• Specifically, the measurement researcher
assumes the conceptual relationship is true (line
a).
copyrightⓒ 2013 All rights reserved by
LIU YING
16
Investigating Nomological
Networks
copyrightⓒ 2013 All rights reserved by
LIU YING
17
Investigating Nomological
Networks
• A relationship observed between one measured construct and
a measure of another provides only limited evidence for
construct validity. Thus, researchers usually seek to create
more elaborate nomological networks. These may include
several variables that are expected to vary with a measure of
the construct. It also may include variables that are expected
not to vary with it to show evidence of discriminant validity.
• Evidence for construct validity mounts as empirical research
supports relationships expected from a nomological network.
The richer the network and the more support, the
greater a researcher's confidence that the measure is
capturing variance that is construct valid.
copyrightⓒ 2013 All rights reserved by
LIU YING
18
• Summary
• For Review
– Terms to Know
– Things to Know
copyrightⓒ 2013 All rights reserved by
LIU YING
19
4. Measurement Applications
Research Questionnaires
Questionnaire Decisions
• Alternatives to Questionnaire
Construction
• Secondary Data
• Questionnaires Developed by
Others
• Questionnaire Type
• Self-Reports Versus
Observations
• Interviews Versus Written
Questionnaires
Questionnaire Construction
• Content Domain
• Items
• Item Wording
• Item Sequence
• Scaling
Questionnaire Response Styles
• Self-Reports
• Observations
• Implications for Questionnaire
Construction and Use
Pilot Testing
Summary
For Review
• Terms to Know
• Things to Know
Part II Suggested Readings
copyrightⓒ 2013 All rights reserved by
LIU YING
20
Measurement Applications
Research Questionnaires
• Questionnaires are measuring instruments that ask
individuals to answer a set of questions.
• If the questions ask for information about the individual
respondents, they are called self-report
questionnaires. Information obtained in self-report
questionnaires include biographical
information, attitudes, opinions, and knowledge.
• Individuals may complete self-report questionnaires
by responding to written questions or to questions
shown on a computer terminal. Self-reports also
may be obtained through an interview in which
another individual (the interviewer) asks the questions
verbally and is responsible for recording responses.
copyrightⓒ 2013 All rights reserved by
LIU YING
21
Questionnaire Decisions
• Alternatives to Questionnaire
Construction
• Secondary Data
• Questionnaires Developed by Others
• Questionnaire Type
• Self-Reports Versus Observations
• Interviews Versus Written Questionnaires
copyrightⓒ 2013 All rights reserved by
LIU YING
22
Questionnaire Decisions
• Constructing a questionnaire is time-consuming
and challenging. It is particularly challenging
when abstract constructs are measured.
• There are two additional questions to address if
it is decided that a questionnaire must be
developed to carry out a research project.
– One, should information be obtained with a written
questionnaire or an interview?
– Two, if the questionnaire is designed to obtain
information about individuals, should the
questionnaire obtain it from outside observers or
from individuals reporting on themselves?
copyrightⓒ 2013 All rights reserved by
LIU YING
23
Alternatives to Questionnaire
Construction
Choosing measures depends foremost on the topic a researcher seeks to investigate. Given a
topic, a starting point is to see if the data you are interested in studying may already be
available. If not, a second step is to see if a measure(s) is available that will serve your
research interest.
• Secondary Data
– secondary data: data collected for some other purpose. Such data are available from many
sources.
– internal purposes, external requirements, by outside organizations (industry trade associations and
organizations, individuals)
• Questionnaires Developed by Others
• Another researcher may have
– already developed a questionnaire that addresses your research questions. Although data may not
be available, a questionnaire may be available that you can use to collect your own data.
(Questionnaires measuring constructs relating to many individual characteristics such as
ability, personality, and interests are readily available. Questionnaires are also available for measuring
characteristics of individuals interacting with organizations, such as employee satisfaction. A good method
for finding these measures is to examine research reports on topics related to your research interests.)
– If suitable for your research interests, questionnaires already constructed are obviously
advantageous in the time and effort they save. They are especially attractive if construct validation
research as described in the previous chapter has already been performed.
copyrightⓒ 2013 All rights reserved by
LIU YING
24
Questionnaire Type
Often secondary data or questionnaires already
developed are simply not viable options. This is
necessarily true when a researcher chooses
to investigate a construct that has not been
previously defined.
• Self-Reports Versus Observations
• Interviews Versus Written Questionnaires
copyrightⓒ 2013 All rights reserved by
LIU YING
25
Self-Reports Versus Observations
• Researchers are often curious about relationships that include
behaviors or characteristics of individuals interacting with
organizations.
• Some constructs require that the information be measured with
responses provided by research participants. (Attitudes and
opinions, intentions, interests, and preferences)
• However, there are other constructs that can be measured either
internally through self-reports or externally by observation. These
constructs typically involve overt behaviors, characteristics of
individuals that can be observed directly. Observations are typically
preferred when constructs can be assessed directly. External
observers are more likely to provide consistent assessments
across research participants. Furthermore, external observers may
be less likely to bias responses in a way that characterizes the
behavior in a favorable light.
copyrightⓒ 2013 All rights reserved by
LIU YING
26
Interviews Versus Written
Questionnaires
• A distinction is sometimes drawn between the development
of interviews and the development of questionnaires. This
distinction is largely unwarranted. The difference
between the two procedures resides primarily in the way
information is obtained from research participants.
Interviews elicit information verbally; questionnaires elicit
information in written form. The same care must be taken in
developing interview questions and response formats as is
taken in developing questionnaires.
• A case can be made that interviews allow greater flexibility.
Interviewers can follow up on answers with questions that
probe respondents' thinking in greater depth. Interviewers can
record responses and interviewee behaviors that are not
available as formal questionnaire responses.
copyrightⓒ 2013 All rights reserved by
LIU YING
27
Interviews Versus Written
Questionnaires
• These are differences that make interviews attractive in the early
stages of instrument development. Interviews can help researchers
refine questions to be asked and the response formats to be used.
However, when finalized, when a researcher is ready to collect data that
will be used to investigate the main research expectations, a typical
interview schedule will look much like a typical questionnaire.
• The decision about whether to use an interview or questionnaire as the final
measurement instrument depends on other criteria. Assuming the same
care in construction, questionnaires usually are less expensive to
administer. The decision to use an interview or a questionnaire also must
take account of respondents' abilities and motivations. Reading
abilities among some members of heterogeneous populations may make
the use of questionnaires problematic. Interviews may also be
advantageous from a motivational perspective. The interaction
that takes place between the interviewer and interviewee may be used
advantageously to motivate participation and complete responses.
copyrightⓒ 2013 All rights reserved by
LIU YING
28
Interviews Versus Written
Questionnaires
• Interaction between interviewers and interviewees
also poses dangers for interviews. There is a
greater risk that the administration of questions
differs from interview to interview.
Furthermore, because there is
interaction, interviewee responses may be
influenced by the particular individual conducting
the interview.
• It is generally desirable to use questionnaires
when possible. The importance of uniformity in
questions and response coding favors
questionnaires. When interviews are used, it is
important that they be conducted as systematically
as possible.
copyrightⓒ 2013 All rights reserved by
LIU YING
29
Questionnaire Construction
• Content Domain
• Items
• Item Wording
• Item Sequence
• Scaling
copyrightⓒ 2013 All rights reserved by
LIU YING
30
Questionnaire Construction
• Questionnaires, whether administered in
written form or through interviews, have
two essential characteristics. First, they
have items designed to elicit information of
research interest. Second, they have a
protocol for recording responses.
copyrightⓒ 2013 All rights reserved by
LIU YING
31
Questionnaire
Construction
copyrightⓒ 2013 All rights reserved by
LIU YING
32
Content Domain
• A properly designed study will identify the variables to be measured
by the time questionnaire development becomes an issue. If one or
more constructs are included, these should be carefully defined as
described in chapter 3. Items should follow closely from the
definitions.
• Typically, researchers also want to obtain additional information from
their questionnaires. At the very least, information will be sought
about personal descriptive characteristics of the questionnaire
respondents.
• Interesting side issues are likely to occur while the questionnaire is
being developed. As a consequence, it is often tempting to add
items that are not central to the research investigation. Resist this
temptation. Attend to developing a set of items that focus directly
and unequivocally on your research topic. Diverting attention to
related items and issues will likely reduce the quality of items that
are essential. Furthermore, response rates inevitably decline as
questionnaire length increases.
copyrightⓒ 2013 All rights reserved by
LIU YING
33
Items
• Item wording and the arrangement of items obviously affect
the responses obtained.
• Item Wording
– 1. Keep the respondent in mind.
– 2. Make it simple.
– 3. Be specific.
– 4. Be honest.
• Item Sequence
– The way items are ordered in a questionnaire is constrained by
the type of items included. For example, order is of little
consequence if items are all similar. However, order can
influence
– It is helpful to start a questionnaire with items that participants
find interesting and that are easy to complete.
– Ask for demographic information last.
copyrightⓒ 2013 All rights reserved by
LIU YING
34
Scaling
• An open-ended response format permits respondents to
answer questions in their own words. They are sometimes
used on small groups early in the questionnaire development
process to make sure the full range of potential responses is
captured. They also are sometimes used in
interviews, particularly when the questions are designed to
elicit complex responses.
• Closed-ended response formats in which respondents are
asked to choose the one category that most closely applies
to them. Closed-ended responses are easy to complete;
they are also easy to code reliably.
• Categories with equal intervals are attractive for conducting
statistical analyses on scores, as discussed in part IV.
copyrightⓒ 2013 All rights reserved by
LIU YING
35
Scaling
copyrightⓒ 2013 All rights reserved by
LIU YING
36
RESEARCH HIGHLIGHT 4.1
words to avoid in questionnaires
• Absolutes. Words expressing absolutes such as
always, never, everyone, and all create logical problems because
statements including item are almost always
• And. The word and usually signals that the Item is getting at two
ideas not one—a double-barreled question. Double-barreled
questions are problematic because responses may differ depending
on which "barrel" is considered.
• You, You is problematic if there can be any question about whether
it refers to the respondent or to a group the respondent represents
(e.g., an organization).
• Adjectives to describe quantity. Words such as
occasionally, sometimes, frequently, and often mean different things
to different people. One person's occasionally may be
equivalent, numerically, to another person's frequently. Use
numerical values when you want to obtain numerical information.
copyrightⓒ 2013 All rights reserved by
LIU YING
37
Questionnaire Response Styles
Chapter 3 noted that scores are critical for establishing the
construct validity of measures. This is a reminder that the value of
information obtained from questionnaires is determined by the quality
of scores obtained. Items and scaling formats alone, no matter how
elegant, do not guarantee successful questionnaire outcomes.
Research has established that characteristics of the individuals
completing questionnaires and the environments in which they
complete them often affect the scores obtained. Some of these
characteristics have been studied in situations involving self-reports;
others have been
studied on observational ratings.
• Self-Reports
• Observations
• Implications for Questionnaire Construction and Use
copyrightⓒ 2013 All rights reserved by
LIU YING
38
Self-Reports
• Two tendencies that influence self-report
responses have received substantial
attention. Social desirability refers to the
tendency to present oneself in a publicly
favorable light. For example, a socially desirable response expresses
approval for a public policy (e.g., the Supreme Court's decision on abortion) because the
respondent believes others approve. Response acquiescence
or yea-saying is a tendency to agree with a
statement regardless of its content. Of the
two, social desirability appears to be a more
general problem for questionnaire responses.
copyrightⓒ 2013 All rights reserved by
LIU YING
39
Observations
A number of response styles have also been identified when
individuals are asked to make observations about some object. These
include
• leniency error, a tendency to systematically provide a more
favorable response than is warranted.
• Severity error, a tendency to systematically provide less favorable
responses than warranted, is less frequent. Alternatively,
• central tendency error is present if an observer clusters
responses in the middle of a scale when more variable responses
should be recorded.
• Halo error is present when an observer evaluates an object in an
undifferentiated manner.
For example, a student may provide favorable evaluations to an instructor on all dimensions of teaching effectiveness because the
instructor is effective on one dimension of teaching.
copyrightⓒ 2013 All rights reserved by
LIU YING
40
Implications for Questionnaire
Construction and Use
Self-report and observer errors are difficult to identify in practice.
Furthermore, attempts to control for either self-report response styles
or observational errors through questionnaire construction have
only limited success.
• Forced-choice scales are designed to provide respondents with
choices that appear to be equal in social desirability or equal in
favorability. Behaviorally anchored observation or rating scales
(see Research Highlight 4.2) are designed to yield more
accurate ratings by providing respondents with meaningful scale
anchors to help generate scores that are less susceptible to
rating errors.
• Unfortunately, research investigations comparing formats on
common self-report and observational problems have not found one
format to be systematically better than others.
copyrightⓒ 2013 All rights reserved by
LIU YING
41
Pilot Testing
• No matter how much care is used, questionnaire construction remains an imprecise research
procedure. Before using a questionnaire for substantive research, it is essential to obtain
information by pilot testing the questionnaire on individuals similar to those who will be
asked to complete it as a part of the substantive research. Two types of pilot tests are
desirable.
• One type asks individuals, preferably like those who will complete the final
questionnaire, to provide their interpretation and understanding of each item. This
assessment will help identify errors in assumptions about participants' frames of reference.
It also helps identify items that are difficult to understand. Pilot tests of this sort will almost
always lead to changes in the design of a research questionnaire. These changes may
help increase response rates, reduce missing data, and obtain more valid responses on
the final questionnaire.
• A second type of pilot test is more like a regular research study; a large number of
respondents are desirable. Data from this type of pilot test are used to see if scores
behave as expected. Are average scores reasonable? Do scores on items vary as
expected? Analyses assessing relationships among items are also useful in this
type of pilot test. For example, internal consistency reliability of multi-item measures
can be assessed by a procedure described in chapter 17. Indeed, this second type of pilot
test can be viewed as an important step in construct validation as described in the last
chapter. However, its preliminary nature must be emphasized. Changes in items will
almost always be suggested the first time scores from a new questionnaire are analyzed.
copyrightⓒ 2013 All rights reserved by
LIU YING
42
Summary
For Review
• Terms to Know
• Things to Know
Part II Suggested Readings
copyrightⓒ 2013 All rights reserved by
LIU YING
43

Contenu connexe

Similaire à Part 2

Reliability and Validity.pptx
Reliability and Validity.pptxReliability and Validity.pptx
Reliability and Validity.pptxSuhailBhat50
 
Reliability and validity ppt
Reliability and validity pptReliability and validity ppt
Reliability and validity pptsurendra poudel
 
Research Methodology3_Measurement.pptx
Research Methodology3_Measurement.pptxResearch Methodology3_Measurement.pptx
Research Methodology3_Measurement.pptxAamirMaqsood8
 
reliablity and validity in social sciences research
reliablity and validity  in social sciences researchreliablity and validity  in social sciences research
reliablity and validity in social sciences researchSourabh Sharma
 
JC-16-23June2021-rel-val.pptx
JC-16-23June2021-rel-val.pptxJC-16-23June2021-rel-val.pptx
JC-16-23June2021-rel-val.pptxsaurami
 
Research Methodology 7
Research Methodology   7Research Methodology   7
Research Methodology 7ayat_ismail
 
Doing observation and Data Analysis for Qualitative Research
Doing observation and Data Analysis for Qualitative ResearchDoing observation and Data Analysis for Qualitative Research
Doing observation and Data Analysis for Qualitative ResearchAhmad Johari Sihes
 
Topic 7 measurement in research
Topic 7   measurement in researchTopic 7   measurement in research
Topic 7 measurement in researchDhani Ahmad
 
Validity, reliability & Internal validity in Researches
Validity, reliability & Internal validity in ResearchesValidity, reliability & Internal validity in Researches
Validity, reliability & Internal validity in ResearchesMunther Magableh
 
Lesson 11 Understanding Data and Ways to Systematically Collect Data.ppt
Lesson 11 Understanding Data and Ways to Systematically Collect Data.pptLesson 11 Understanding Data and Ways to Systematically Collect Data.ppt
Lesson 11 Understanding Data and Ways to Systematically Collect Data.pptGerfelChan1
 
Questionnaire and Instrument validity
Questionnaire and Instrument validityQuestionnaire and Instrument validity
Questionnaire and Instrument validitymdanaee
 
Validity.pptx
Validity.pptxValidity.pptx
Validity.pptxrupasi13
 
PURGING OF UNTRUSTWORTHY RECOMMENDATIONS FROM A GRID
PURGING OF UNTRUSTWORTHY RECOMMENDATIONS FROM A GRIDPURGING OF UNTRUSTWORTHY RECOMMENDATIONS FROM A GRID
PURGING OF UNTRUSTWORTHY RECOMMENDATIONS FROM A GRIDijngnjournal
 

Similaire à Part 2 (20)

Validity of instrument
Validity of instrumentValidity of instrument
Validity of instrument
 
Measurement.pptx
 Measurement.pptx Measurement.pptx
Measurement.pptx
 
Reliability and Validity.pptx
Reliability and Validity.pptxReliability and Validity.pptx
Reliability and Validity.pptx
 
Reliability and validity ppt
Reliability and validity pptReliability and validity ppt
Reliability and validity ppt
 
Research Methodology3_Measurement.pptx
Research Methodology3_Measurement.pptxResearch Methodology3_Measurement.pptx
Research Methodology3_Measurement.pptx
 
reliablity and validity in social sciences research
reliablity and validity  in social sciences researchreliablity and validity  in social sciences research
reliablity and validity in social sciences research
 
JC-16-23June2021-rel-val.pptx
JC-16-23June2021-rel-val.pptxJC-16-23June2021-rel-val.pptx
JC-16-23June2021-rel-val.pptx
 
Research Methodology 7
Research Methodology   7Research Methodology   7
Research Methodology 7
 
Validity and Reliability.pdf
Validity and Reliability.pdfValidity and Reliability.pdf
Validity and Reliability.pdf
 
Validity and Reliability.pdf
Validity and Reliability.pdfValidity and Reliability.pdf
Validity and Reliability.pdf
 
Doing observation and Data Analysis for Qualitative Research
Doing observation and Data Analysis for Qualitative ResearchDoing observation and Data Analysis for Qualitative Research
Doing observation and Data Analysis for Qualitative Research
 
MBA-12-02
MBA-12-02MBA-12-02
MBA-12-02
 
Topic 7 measurement in research
Topic 7   measurement in researchTopic 7   measurement in research
Topic 7 measurement in research
 
Validity, reliability & Internal validity in Researches
Validity, reliability & Internal validity in ResearchesValidity, reliability & Internal validity in Researches
Validity, reliability & Internal validity in Researches
 
Lesson 11 Understanding Data and Ways to Systematically Collect Data.ppt
Lesson 11 Understanding Data and Ways to Systematically Collect Data.pptLesson 11 Understanding Data and Ways to Systematically Collect Data.ppt
Lesson 11 Understanding Data and Ways to Systematically Collect Data.ppt
 
unit 2.6.pptx
unit 2.6.pptxunit 2.6.pptx
unit 2.6.pptx
 
Questionnaire and Instrument validity
Questionnaire and Instrument validityQuestionnaire and Instrument validity
Questionnaire and Instrument validity
 
Validity
ValidityValidity
Validity
 
Validity.pptx
Validity.pptxValidity.pptx
Validity.pptx
 
PURGING OF UNTRUSTWORTHY RECOMMENDATIONS FROM A GRID
PURGING OF UNTRUSTWORTHY RECOMMENDATIONS FROM A GRIDPURGING OF UNTRUSTWORTHY RECOMMENDATIONS FROM A GRID
PURGING OF UNTRUSTWORTHY RECOMMENDATIONS FROM A GRID
 

Plus de Ying Liu

首尔大韩国语语法Topik考试语法合集 刘赢整理
首尔大韩国语语法Topik考试语法合集 刘赢整理首尔大韩国语语法Topik考试语法合集 刘赢整理
首尔大韩国语语法Topik考试语法合集 刘赢整理Ying Liu
 
Chapter36b
Chapter36bChapter36b
Chapter36bYing Liu
 
Chapter36a
Chapter36aChapter36a
Chapter36aYing Liu
 

Plus de Ying Liu (20)

首尔大韩国语语法Topik考试语法合集 刘赢整理
首尔大韩国语语法Topik考试语法合集 刘赢整理首尔大韩国语语法Topik考试语法合集 刘赢整理
首尔大韩国语语法Topik考试语法合集 刘赢整理
 
Chapter38
Chapter38Chapter38
Chapter38
 
Chapter36b
Chapter36bChapter36b
Chapter36b
 
Chapter36a
Chapter36aChapter36a
Chapter36a
 
Chapter35
Chapter35Chapter35
Chapter35
 
Chapter34
Chapter34Chapter34
Chapter34
 
Chapter33
Chapter33Chapter33
Chapter33
 
Chapter32
Chapter32Chapter32
Chapter32
 
Chapter31
Chapter31Chapter31
Chapter31
 
Chapter30
Chapter30Chapter30
Chapter30
 
Chapter29
Chapter29Chapter29
Chapter29
 
Chapter28
Chapter28Chapter28
Chapter28
 
Chapter27
Chapter27Chapter27
Chapter27
 
Chapter26
Chapter26Chapter26
Chapter26
 
Chapter25
Chapter25Chapter25
Chapter25
 
Chapter24
Chapter24Chapter24
Chapter24
 
Chapter23
Chapter23Chapter23
Chapter23
 
Chapter22
Chapter22Chapter22
Chapter22
 
Chapter21
Chapter21Chapter21
Chapter21
 
Chapter20
Chapter20Chapter20
Chapter20
 

Dernier

GenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdfGenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdflior mazor
 
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, AdobeApidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobeapidays
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerThousandEyes
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Miguel Araújo
 
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUnderstanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUK Journal
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024Rafal Los
 
Artificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : UncertaintyArtificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : UncertaintyKhushali Kathiriya
 
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...apidays
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)Gabriella Davis
 
MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MIND CTI
 
Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...apidays
 
presentation ICT roal in 21st century education
presentation ICT roal in 21st century educationpresentation ICT roal in 21st century education
presentation ICT roal in 21st century educationjfdjdjcjdnsjd
 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...apidays
 
Manulife - Insurer Innovation Award 2024
Manulife - Insurer Innovation Award 2024Manulife - Insurer Innovation Award 2024
Manulife - Insurer Innovation Award 2024The Digital Insurer
 
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FMESafe Software
 
2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...Martijn de Jong
 
Top 10 Most Downloaded Games on Play Store in 2024
Top 10 Most Downloaded Games on Play Store in 2024Top 10 Most Downloaded Games on Play Store in 2024
Top 10 Most Downloaded Games on Play Store in 2024SynarionITSolutions
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024The Digital Insurer
 

Dernier (20)

GenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdfGenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdf
 
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, AdobeApidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
 
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUnderstanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024
 
Artificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : UncertaintyArtificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : Uncertainty
 
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)
 
MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024
 
Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...
 
presentation ICT roal in 21st century education
presentation ICT roal in 21st century educationpresentation ICT roal in 21st century education
presentation ICT roal in 21st century education
 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
 
Manulife - Insurer Innovation Award 2024
Manulife - Insurer Innovation Award 2024Manulife - Insurer Innovation Award 2024
Manulife - Insurer Innovation Award 2024
 
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
 
2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...
 
Top 10 Most Downloaded Games on Play Store in 2024
Top 10 Most Downloaded Games on Play Store in 2024Top 10 Most Downloaded Games on Play Store in 2024
Top 10 Most Downloaded Games on Play Store in 2024
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024
 

Part 2

  • 1. Research Methods for Organizational Studies - 2ed, 2004 LIU YING copyrightⓒ 2013 All rights reserved by LIU YING 1
  • 2. 3. Measurement Foundations Validity and Validation • Construct Definitions – Construct Domain – Nomological Networks • Construct Definition Illustration • Construct Validity Challenges – Random Errors – Systematic Errors – Scores Are Critical • Construct Validation – Content Validity – Reliability • Types of Reliability • Reliability and Construct Validity – Convergent Validity – Discriminant Validity – Criterion-Related Validity – Investigating Nomological Networks • Summary • For Review – Terms to Know – Things to Know copyrightⓒ 2013 All rights reserved by LIU YING 2
  • 3. Construct Definitions Measurement produces numerical values that are designed to summarize characteristics of cases under study. A measure is an instrument to record such scores. Construct valid measures yield numerical values that accurately represent the characteristic. The most useful conceptual definitions have two elements.  Construct Domain: First, useful conceptual definitions identify the nature of the construct by specifying its meaning. This element explains what a researcher has in mind for the construct;  Nomological Networks: The second element also should specify how the construct of interest relates to other constructs in a broader web of relationships copyrightⓒ 2013 All rights reserved by LIU YING 3
  • 4. Construct Definition Illustration copyrightⓒ 2013 All rights reserved by LIU YING 4
  • 5. Construct Validity Challenges • Construct Validity Challenges – Random Errors • The more items, the more successfully this type of random error is controlled. – Systematic Errors • A measure is Contaminated or Deficient – Scores Are Critical • Systematic variance is sometimes called true score variance copyrightⓒ 2013 All rights reserved by LIU YING 5
  • 6. ScoresAreCritical copyrightⓒ 2013 All rights reserved by LIU YING 6
  • 7. Scores Are Critical • Exhibit 3.2 also shows three sources of construct invalidity: – 1. Scores on a measure may be less than construct valid because of deficiency. In the example, observed scores are deficient because the measure does not capture satisfaction with computer durability. – 2. Scores may be less than construct valid because of systematic contamination. In the example, observed scores are contaminated because the measure includes satisfaction with the purchasing experience. – 3. Finally, scores on a measure are less than construct valid to the extent that they include random errors. copyrightⓒ 2013 All rights reserved by LIU YING 7
  • 8. Construct Validation Because construct validity cannot be assessed directly, it cannot be directly established. However, there are procedures available to help researchers develop construct valid measures and to help evaluate those measures once developed. Six such procedures are described here. • Content Validity • Reliability – Types of Reliability – Reliability and Construct Validity • Convergent Validity • Discriminant Validity • Criterion-Related Validity • Investigating Nomological Networks copyrightⓒ 2013 All rights reserved by LIU YING 8
  • 9. Content Validity • A measure is content valid when its items art judged to accurately reflect the domain of the construct as defined conceptually. Content validation ordinarily has experts in the subject matter of interest provide assessments of content validity. • As a part of development, the researcher has a panel of experts in computer programming review the measure for its content. Content validation of this sort provides information about potential systematic errors in measures. • Content validation can help improve the items that form a measure. Nevertheless, it is not sufficient for construct validity. In particular, content validation procedures may not provide information about potential deficiency, nor can subject matter experts provide much information about random errors that may be present in the scores that are obtained. copyrightⓒ 2013 All rights reserved by LIU YING 9
  • 10. Reliability • Types of Reliability – Internal consistency: reliability refers to the similarity of item scores obtained on a measure that has multiple items. It can be assessed when items are intended to measure a single construct. In the computer example (Exhibit 3.1), satisfaction is measured with six items. The internal consistency of that questionnaire can be estimated if scores are available from a set of cases. – Interrater reliability: indicates the degree to which a group of observers or raters provide consistent evaluations. For example, the observers may be a group of international judges who are asked to evaluate ice skaters performing in a competition. In this case, the judges serve as measurement repetitions just as the items serve as repetitions in the computer satisfaction questionnaire. High reliability is obtained when the judges agree on the evaluation of each – Stability reliability: refers to the consistency of measurement results across time. Here measurement repetitions refer to time periods (a measure is administered more than once). • Reliability and Construct Validity *** There are three common contexts in which researchers seek to assess the reliability of measurement. Chapter 17 describes statistical procedures for estimating these types of reliability. copyrightⓒ 2013 All rights reserved by LIU YING 10
  • 11. Reliability and Construct Validity Reliability speaks only to a measure's freedom from random errors. It does not address systematic errors involving contamination or deficiency. Reliability is thus necessary for construct validity but not sufficient. It is necessary because unreliable variance must be construct invalid. It is not sufficient because systematic variance may be contaminated and because reliability simply does not account for deficiency. In short, reliability addresses only whether scores are consistent; it does not address whether scores capture a particular construct as defined conceptually. copyrightⓒ 2013 All rights reserved by LIU YING 11
  • 12. Convergent Validity • Convergent validity is present when there is a high correspondence between scores from two or more different measures of the same construct. Convergent validity is important because it must be present if scores from both measures are construct valid. But convergent validity is not sufficient for construct validity any more than is reliability. copyrightⓒ 2013 All rights reserved by LIU YING 12
  • 13. Convergent Validity The area crossed with vertical lines shows the proportion of variance in scores from the two measures that is convergent. However, only the area also crossed with horizontal lines shows common construct valid variance. The area covered only by vertical lines shows where the two measures share variance that represents contamination from a construct validity perspective. Convergent validity also does not address whether measures are deficient. Nor does it provide construct validity information about the proportion of variance in the two measures that do not converge. Exhibit 3.3 shows that more of the variance unique to the measure A overlaps with construct variance than variance from measure B. Despite these limitations, evidence of convergent validity is desirable. If two measures that are designed to measure the same construct do not converge, at least one of them is not construct valid. Alternatively, if they do converge, circumstantial evidence is obtained that they may both be construct valid. Evidence of convergent validity adds to a researcher's confidence in the construct validity of measures. copyrightⓒ 2013 All rights reserved by LIU YING 13
  • 14. Discriminant Validity • Discriminant validity is inferred when scores from measures of different constructs do not converge. It thus provides information about whether scores from a measure of a construct are unique rather than contaminated by other constructs. • Proposed constructs should provide contributions beyond constructs already in the research domain. Consequently, measures of proposed constructs should show evidence of discriminant validity with measures of existing constructs. copyrightⓒ 2013 All rights reserved by LIU YING 14
  • 15. Criterion-Related Validity • Criterion-related validity is present when scores on a measure are related to scores on another measure that better reflects the construct of interest. It differs from convergent validity, where scores from the measures are assumed to be equivalent representations of the construct. In criterion- related validity the criterion measure is assumed to have greater construct validity than the measure being developed or investigated. • Why not just use the criterion measure if it has greater construct validity? • Historically, researchers used the term to describe relationships between a construct (represented by the measure under consideration) and a measure of another construct that is thought to be conceptually related to the first. copyrightⓒ 2013 All rights reserved by LIU YING 15
  • 16. Investigating Nomological Networks • Nomological networks have been described as relationships between a construct under measurement consideration and other constructs. • Researchers with a conceptual orientation are interested in the conceptual relationship between the independent and dependent variable constructs. • Specifically, the measurement researcher assumes the conceptual relationship is true (line a). copyrightⓒ 2013 All rights reserved by LIU YING 16
  • 17. Investigating Nomological Networks copyrightⓒ 2013 All rights reserved by LIU YING 17
  • 18. Investigating Nomological Networks • A relationship observed between one measured construct and a measure of another provides only limited evidence for construct validity. Thus, researchers usually seek to create more elaborate nomological networks. These may include several variables that are expected to vary with a measure of the construct. It also may include variables that are expected not to vary with it to show evidence of discriminant validity. • Evidence for construct validity mounts as empirical research supports relationships expected from a nomological network. The richer the network and the more support, the greater a researcher's confidence that the measure is capturing variance that is construct valid. copyrightⓒ 2013 All rights reserved by LIU YING 18
  • 19. • Summary • For Review – Terms to Know – Things to Know copyrightⓒ 2013 All rights reserved by LIU YING 19
  • 20. 4. Measurement Applications Research Questionnaires Questionnaire Decisions • Alternatives to Questionnaire Construction • Secondary Data • Questionnaires Developed by Others • Questionnaire Type • Self-Reports Versus Observations • Interviews Versus Written Questionnaires Questionnaire Construction • Content Domain • Items • Item Wording • Item Sequence • Scaling Questionnaire Response Styles • Self-Reports • Observations • Implications for Questionnaire Construction and Use Pilot Testing Summary For Review • Terms to Know • Things to Know Part II Suggested Readings copyrightⓒ 2013 All rights reserved by LIU YING 20
  • 21. Measurement Applications Research Questionnaires • Questionnaires are measuring instruments that ask individuals to answer a set of questions. • If the questions ask for information about the individual respondents, they are called self-report questionnaires. Information obtained in self-report questionnaires include biographical information, attitudes, opinions, and knowledge. • Individuals may complete self-report questionnaires by responding to written questions or to questions shown on a computer terminal. Self-reports also may be obtained through an interview in which another individual (the interviewer) asks the questions verbally and is responsible for recording responses. copyrightⓒ 2013 All rights reserved by LIU YING 21
  • 22. Questionnaire Decisions • Alternatives to Questionnaire Construction • Secondary Data • Questionnaires Developed by Others • Questionnaire Type • Self-Reports Versus Observations • Interviews Versus Written Questionnaires copyrightⓒ 2013 All rights reserved by LIU YING 22
  • 23. Questionnaire Decisions • Constructing a questionnaire is time-consuming and challenging. It is particularly challenging when abstract constructs are measured. • There are two additional questions to address if it is decided that a questionnaire must be developed to carry out a research project. – One, should information be obtained with a written questionnaire or an interview? – Two, if the questionnaire is designed to obtain information about individuals, should the questionnaire obtain it from outside observers or from individuals reporting on themselves? copyrightⓒ 2013 All rights reserved by LIU YING 23
  • 24. Alternatives to Questionnaire Construction Choosing measures depends foremost on the topic a researcher seeks to investigate. Given a topic, a starting point is to see if the data you are interested in studying may already be available. If not, a second step is to see if a measure(s) is available that will serve your research interest. • Secondary Data – secondary data: data collected for some other purpose. Such data are available from many sources. – internal purposes, external requirements, by outside organizations (industry trade associations and organizations, individuals) • Questionnaires Developed by Others • Another researcher may have – already developed a questionnaire that addresses your research questions. Although data may not be available, a questionnaire may be available that you can use to collect your own data. (Questionnaires measuring constructs relating to many individual characteristics such as ability, personality, and interests are readily available. Questionnaires are also available for measuring characteristics of individuals interacting with organizations, such as employee satisfaction. A good method for finding these measures is to examine research reports on topics related to your research interests.) – If suitable for your research interests, questionnaires already constructed are obviously advantageous in the time and effort they save. They are especially attractive if construct validation research as described in the previous chapter has already been performed. copyrightⓒ 2013 All rights reserved by LIU YING 24
  • 25. Questionnaire Type Often secondary data or questionnaires already developed are simply not viable options. This is necessarily true when a researcher chooses to investigate a construct that has not been previously defined. • Self-Reports Versus Observations • Interviews Versus Written Questionnaires copyrightⓒ 2013 All rights reserved by LIU YING 25
  • 26. Self-Reports Versus Observations • Researchers are often curious about relationships that include behaviors or characteristics of individuals interacting with organizations. • Some constructs require that the information be measured with responses provided by research participants. (Attitudes and opinions, intentions, interests, and preferences) • However, there are other constructs that can be measured either internally through self-reports or externally by observation. These constructs typically involve overt behaviors, characteristics of individuals that can be observed directly. Observations are typically preferred when constructs can be assessed directly. External observers are more likely to provide consistent assessments across research participants. Furthermore, external observers may be less likely to bias responses in a way that characterizes the behavior in a favorable light. copyrightⓒ 2013 All rights reserved by LIU YING 26
  • 27. Interviews Versus Written Questionnaires • A distinction is sometimes drawn between the development of interviews and the development of questionnaires. This distinction is largely unwarranted. The difference between the two procedures resides primarily in the way information is obtained from research participants. Interviews elicit information verbally; questionnaires elicit information in written form. The same care must be taken in developing interview questions and response formats as is taken in developing questionnaires. • A case can be made that interviews allow greater flexibility. Interviewers can follow up on answers with questions that probe respondents' thinking in greater depth. Interviewers can record responses and interviewee behaviors that are not available as formal questionnaire responses. copyrightⓒ 2013 All rights reserved by LIU YING 27
  • 28. Interviews Versus Written Questionnaires • These are differences that make interviews attractive in the early stages of instrument development. Interviews can help researchers refine questions to be asked and the response formats to be used. However, when finalized, when a researcher is ready to collect data that will be used to investigate the main research expectations, a typical interview schedule will look much like a typical questionnaire. • The decision about whether to use an interview or questionnaire as the final measurement instrument depends on other criteria. Assuming the same care in construction, questionnaires usually are less expensive to administer. The decision to use an interview or a questionnaire also must take account of respondents' abilities and motivations. Reading abilities among some members of heterogeneous populations may make the use of questionnaires problematic. Interviews may also be advantageous from a motivational perspective. The interaction that takes place between the interviewer and interviewee may be used advantageously to motivate participation and complete responses. copyrightⓒ 2013 All rights reserved by LIU YING 28
  • 29. Interviews Versus Written Questionnaires • Interaction between interviewers and interviewees also poses dangers for interviews. There is a greater risk that the administration of questions differs from interview to interview. Furthermore, because there is interaction, interviewee responses may be influenced by the particular individual conducting the interview. • It is generally desirable to use questionnaires when possible. The importance of uniformity in questions and response coding favors questionnaires. When interviews are used, it is important that they be conducted as systematically as possible. copyrightⓒ 2013 All rights reserved by LIU YING 29
  • 30. Questionnaire Construction • Content Domain • Items • Item Wording • Item Sequence • Scaling copyrightⓒ 2013 All rights reserved by LIU YING 30
  • 31. Questionnaire Construction • Questionnaires, whether administered in written form or through interviews, have two essential characteristics. First, they have items designed to elicit information of research interest. Second, they have a protocol for recording responses. copyrightⓒ 2013 All rights reserved by LIU YING 31
  • 32. Questionnaire Construction copyrightⓒ 2013 All rights reserved by LIU YING 32
  • 33. Content Domain • A properly designed study will identify the variables to be measured by the time questionnaire development becomes an issue. If one or more constructs are included, these should be carefully defined as described in chapter 3. Items should follow closely from the definitions. • Typically, researchers also want to obtain additional information from their questionnaires. At the very least, information will be sought about personal descriptive characteristics of the questionnaire respondents. • Interesting side issues are likely to occur while the questionnaire is being developed. As a consequence, it is often tempting to add items that are not central to the research investigation. Resist this temptation. Attend to developing a set of items that focus directly and unequivocally on your research topic. Diverting attention to related items and issues will likely reduce the quality of items that are essential. Furthermore, response rates inevitably decline as questionnaire length increases. copyrightⓒ 2013 All rights reserved by LIU YING 33
  • 34. Items • Item wording and the arrangement of items obviously affect the responses obtained. • Item Wording – 1. Keep the respondent in mind. – 2. Make it simple. – 3. Be specific. – 4. Be honest. • Item Sequence – The way items are ordered in a questionnaire is constrained by the type of items included. For example, order is of little consequence if items are all similar. However, order can influence – It is helpful to start a questionnaire with items that participants find interesting and that are easy to complete. – Ask for demographic information last. copyrightⓒ 2013 All rights reserved by LIU YING 34
  • 35. Scaling • An open-ended response format permits respondents to answer questions in their own words. They are sometimes used on small groups early in the questionnaire development process to make sure the full range of potential responses is captured. They also are sometimes used in interviews, particularly when the questions are designed to elicit complex responses. • Closed-ended response formats in which respondents are asked to choose the one category that most closely applies to them. Closed-ended responses are easy to complete; they are also easy to code reliably. • Categories with equal intervals are attractive for conducting statistical analyses on scores, as discussed in part IV. copyrightⓒ 2013 All rights reserved by LIU YING 35
  • 36. Scaling copyrightⓒ 2013 All rights reserved by LIU YING 36
  • 37. RESEARCH HIGHLIGHT 4.1 words to avoid in questionnaires • Absolutes. Words expressing absolutes such as always, never, everyone, and all create logical problems because statements including item are almost always • And. The word and usually signals that the Item is getting at two ideas not one—a double-barreled question. Double-barreled questions are problematic because responses may differ depending on which "barrel" is considered. • You, You is problematic if there can be any question about whether it refers to the respondent or to a group the respondent represents (e.g., an organization). • Adjectives to describe quantity. Words such as occasionally, sometimes, frequently, and often mean different things to different people. One person's occasionally may be equivalent, numerically, to another person's frequently. Use numerical values when you want to obtain numerical information. copyrightⓒ 2013 All rights reserved by LIU YING 37
  • 38. Questionnaire Response Styles Chapter 3 noted that scores are critical for establishing the construct validity of measures. This is a reminder that the value of information obtained from questionnaires is determined by the quality of scores obtained. Items and scaling formats alone, no matter how elegant, do not guarantee successful questionnaire outcomes. Research has established that characteristics of the individuals completing questionnaires and the environments in which they complete them often affect the scores obtained. Some of these characteristics have been studied in situations involving self-reports; others have been studied on observational ratings. • Self-Reports • Observations • Implications for Questionnaire Construction and Use copyrightⓒ 2013 All rights reserved by LIU YING 38
  • 39. Self-Reports • Two tendencies that influence self-report responses have received substantial attention. Social desirability refers to the tendency to present oneself in a publicly favorable light. For example, a socially desirable response expresses approval for a public policy (e.g., the Supreme Court's decision on abortion) because the respondent believes others approve. Response acquiescence or yea-saying is a tendency to agree with a statement regardless of its content. Of the two, social desirability appears to be a more general problem for questionnaire responses. copyrightⓒ 2013 All rights reserved by LIU YING 39
  • 40. Observations A number of response styles have also been identified when individuals are asked to make observations about some object. These include • leniency error, a tendency to systematically provide a more favorable response than is warranted. • Severity error, a tendency to systematically provide less favorable responses than warranted, is less frequent. Alternatively, • central tendency error is present if an observer clusters responses in the middle of a scale when more variable responses should be recorded. • Halo error is present when an observer evaluates an object in an undifferentiated manner. For example, a student may provide favorable evaluations to an instructor on all dimensions of teaching effectiveness because the instructor is effective on one dimension of teaching. copyrightⓒ 2013 All rights reserved by LIU YING 40
  • 41. Implications for Questionnaire Construction and Use Self-report and observer errors are difficult to identify in practice. Furthermore, attempts to control for either self-report response styles or observational errors through questionnaire construction have only limited success. • Forced-choice scales are designed to provide respondents with choices that appear to be equal in social desirability or equal in favorability. Behaviorally anchored observation or rating scales (see Research Highlight 4.2) are designed to yield more accurate ratings by providing respondents with meaningful scale anchors to help generate scores that are less susceptible to rating errors. • Unfortunately, research investigations comparing formats on common self-report and observational problems have not found one format to be systematically better than others. copyrightⓒ 2013 All rights reserved by LIU YING 41
  • 42. Pilot Testing • No matter how much care is used, questionnaire construction remains an imprecise research procedure. Before using a questionnaire for substantive research, it is essential to obtain information by pilot testing the questionnaire on individuals similar to those who will be asked to complete it as a part of the substantive research. Two types of pilot tests are desirable. • One type asks individuals, preferably like those who will complete the final questionnaire, to provide their interpretation and understanding of each item. This assessment will help identify errors in assumptions about participants' frames of reference. It also helps identify items that are difficult to understand. Pilot tests of this sort will almost always lead to changes in the design of a research questionnaire. These changes may help increase response rates, reduce missing data, and obtain more valid responses on the final questionnaire. • A second type of pilot test is more like a regular research study; a large number of respondents are desirable. Data from this type of pilot test are used to see if scores behave as expected. Are average scores reasonable? Do scores on items vary as expected? Analyses assessing relationships among items are also useful in this type of pilot test. For example, internal consistency reliability of multi-item measures can be assessed by a procedure described in chapter 17. Indeed, this second type of pilot test can be viewed as an important step in construct validation as described in the last chapter. However, its preliminary nature must be emphasized. Changes in items will almost always be suggested the first time scores from a new questionnaire are analyzed. copyrightⓒ 2013 All rights reserved by LIU YING 42
  • 43. Summary For Review • Terms to Know • Things to Know Part II Suggested Readings copyrightⓒ 2013 All rights reserved by LIU YING 43