Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
net/publication/247737198
CITATIONS READS
76 901
3 authors:
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
Cross cultural adaptation of the Beliefs About Psychological Services scale for Iceland View project
http://tcp.sagepub.com
Published by:
http://www.sagepublications.com
On behalf of:
Additional services and information for The Counseling Psychologist can be found
at:
Subscriptions: http://tcp.sagepub.com/subscriptions
Reprints: http://www.sagepub.com/journalsReprints.nav
Permissions: http://www.sagepub.com/journalsPermissions.nav
Citations http://tcp.sagepub.com/cgi/content/refs/36/2/188
Concerns about the cross-cultural validity of constructs are discussed, including equiv-
alence, bias, and translation procedures. Methods to enhance equivalence are
described, as are strategies to evaluate and minimize types of bias. Recommendations
for translat-ing instruments are also presented. To illustrate some challenges of cross-
cultural coun-seling research, translation procedures employed in studies published in
five counseling journals are evaluated. In 15 of 615 empirical articles, a translation of
instruments was performed. In 9 studies, there was some effort to enhance and evaluate
equivalence between language versions of the measures employed. In contrast, 2
studies did not report using thorough translation and verification procedures, and 4
studies employed a mod-erate degree of rigorousness. Suggestions for strengthening
translation methodologies and enhancing the rigor of cross-cultural counseling
research are provided. To conduct cross-culturally valid research and deliver
culturally appropriate services, counseling psychologists must generate and rely on
methodologically sound cross-cultural studies. This article provides a schema for
performing such studies.
188
Downloaded from http://tcp.sagepub.com by SJO TEMP 2008 on December 7, 2008
Ægisdóttir et al. / CROSS-CULTURAL VALIDITY 189
have tremendous potential to enhance the basic core of the science and
practice of counseling psychology, both domestically and internationally”
(p. 147). He also predicted, “In the future, counseling psychology will no
longer be defined as counseling psychology within the United States, but
rather, the parameters of counseling psychology will cross many countries
and many cultures” (Heppner, 2006, p. 170).
Although an international focus in counseling is important, there are
many challenges (cf. Douce, 2004; Heppner, 2006; Pedersen, 2003). This
article discusses methodological challenges, especially as related to the
translation and adaptation of instruments for use in international and cross-
cultural studies and their link to equivalence and bias. While there has been
discussion in the counseling psychology literature about the benefits and
challenges of cross-cultural counseling and the risks of simply applying
Western theories and strategies cross-culturally, we were unable to locate
publications in our literature detailing how to perform cross-culturally valid
research. There is literature, however, in other areas of psychology (e.g.,
cross-cultural, social, international) that addresses these topics. This article
draws from this literature to introduce counseling psychologists to some
concepts, methods, and issues when conducting cross-cultural research. We
also extend this literature by discussing the potential use of cross-cultural
methodologies in counseling research.
As a way to illustrate some challenges of cross-cultural research, we also
examine, analyze, and evaluate translation practices employed in five
prominent counseling journals to determine the translation procedures
counseling researchers have used and the methods employed to minimize
bias and evaluate equivalence. Finally, we offer recommendations about
translation methodology and ways to increase validity in cross-cultural
counseling research.
Equivalence
Equivalence is a key concept in cross-cultural psychology. It addresses
the question of comparability of observations (test scores) across cultures
(van de Vijver, 2001). Several definitions or forms of equivalence have been
reported. Lonner (1985), for instance, discussed four types: functional,
concep-tual, metric, and linguistic. Functional equivalence refers to the
function the behavior under study (e.g., counselor empathy) has in different
cultures. If similar behaviors or activities (e.g., smiling) have different
functions in var-ious cultures, their parameters cannot be used for cross-
cultural comparison (Jahoda, 1966; Lonner, 1985). In comparison,
conceptual equivalence refers to the similarity in meaning attached to a
behavior or concept (Lonner, 1985; Malpass & Poortinga, 1986). Certain
behaviors and concepts (e.g., help seeking) may vary in meaning across
cultures. Metric equivalence refers to psycho-metric properties of the tool
(e.g., Self-Directed Search) used to measure the same construct across
cultures. It is assumed if psychometric data from two or more cultural
groups have the same structure (Malpass & Poortinga, 1986). Finally,
linguistic equivalence has to do with wording of items (form, meaning, and
structure) in different language versions of an instrument, the reading
difficulty of the items, and the “naturalness” of the items in the trans-lated
form (Lonner, 1985; van de Vijver & Leung, 1997).
Van de Vijver and his colleagues (van de Vijver, 2001; van de Vijver &
Leung, 1997) also discussed four types of equivalence representing a hier-
archical order from absence to higher degree of equivalence. The first type,
construct nonequivalence, refers to constructs (e.g., cultural syndromes)
being so dissimilar across cultures they cannot be compared. Under these
circumstances, no link exists between the constructs. The next three types of
equivalence demonstrate some equivalence with the higher level in the
Downloaded from http://tcp.sagepub.com by SJO TEMP 2008 on December 7, 2008
192 THE COUNSELING PSYCHOLOGIST / March 2008
scale for one group relative to the other, limiting raw score comparability
between the groups. For example, a raw score of 28 on the Minnesota
Multiphasic Personality Inventory 2 MacAndrew Alcohol Scale-Revised
(Butcher, Dahlstrom, Graham, Tellegen, & Kaemmer, 2001) does not mean
the same thing for women as it does for men. For women, this score indi-
cates more impulsiveness and greater risk for substance abuse than it does
for men (Greene, 2000). A less clear example but extremely important to
cross-cultural research involves two language versions of the same psycho-
logical instrument. Here the origins of the two language versions of the scale
may appear the same (both versions include the same interval rating scale
for the items). This assumption, however, may be threatened if the two
cultural groups responding to this measure vary in their familiarity with
Likert-type answer formats (method bias; see later). Because of the differ-
ential familiarity with this type of stimuli, the origin of the measurement unit
is not the same for both groups. Similarly, if the two cultural groups vary in
response style (e.g., acquiescence), a score of 2 on a 5-point scale may not
mean the same for both groups. In these examples, the source or the origin
of the scale is different in the two language versions, compro-mising valid
cross-cultural comparison.
Finally, and at the highest level of equivalence, is scalar equivalence or
full score comparability. Equivalent instruments at the scalar level measure a
con-cept with the same interval or ratio scale across cultures, and the origins
of the scales are the same. Therefore, at this level, bias has been ruled out,
and direct cross-cultural comparisons of average scores on an instrument can
be made (e.g., van de Vijver & Leung, 1997).
According to van de Vijver (2001), it can be difficult to discern if
measures are equivalent at the measurement-unit or scalar level. This
challenge is observed in comparison of scale scores between cultural groups
responding to the same language version of an instrument as well as
between different language versions of a measure. As an example of this
difficulty, when using the same language version of an instrument, racial
differences in intelligence test scores can be interpreted as representing true
differences in intelligence (scalar equivalence has been reached) and as an
artifact of the measures (measurement-unit equivalence has been reached).
In the latter, the measurement units are the same, but they have different
origins because of various biases, hindering valid comparisons across
different racial groups. In this instance, valid comparisons at the ratio level
(comparing mean scores) cannot be done. Higher levels of equivalence are
more diffi-cult to establish. It is, for instance, easier to show that an
instrument mea-sures the same construct across cultures (construct
equivalence) by showing a similar factor structure and nomological
networks than it is to demon-strate the instruments’ numerical comparability
(scalar equivalence). The
Downloaded from http://tcp.sagepub.com by SJO TEMP 2008 on December 7, 2008
194 THE COUNSELING PSYCHOLOGIST / March 2008
higher the level of equivalence, though, the more detailed analysis can be
performed on cross-cultural similarities and differences (van de Vijver,
2001; van de Vijver & Leung, 1997).
Levels of equivalence for measures used in cross-cultural counseling
research should be established and reported in counseling psychology pub-
lications. It is not until the equivalence of the concepts under study have
been determined that a meaningful cross-cultural comparison can be made.
Without demonstrated equivalence, numerous rival hypotheses (e.g., poor
translation) may account for observed cross-cultural differences.
Bias
Another important concept in cross-cultural research is bias. Bias
negatively influences equivalence and refers to nuisance factors, limiting the
compara-bility or scalar equivalence of observations (test scores) across
cultural groups (van de Vijver, 2001; van de Vijver & Leung, 1997; van de
Vijver & Poortinga, 1997). Typical sources of bias are construct, method,
and item bias. A con-struct bias occurs when the construct measured as a
whole (e.g., intelli-gence) is not identical across cultural groups. Potential
sources for this type of bias are when there is different coverage of the
construct across cultures (i.e., not all relevant behavioral domains are
sampled), an incomplete overlap of how the construct is defined across
cultures, and when the appropriate-ness of item content differs between two
language versions of an instrument (cf. van de Vijver & Leung, 1997; van de
Vijver & Poortinga, 1997). A serious construct bias equates to construct
nonequivalence.
Even though a construct is well represented in multilingual versions of a
scale (construct equivalence, e.g., similar factor structure, and there is no
construct bias, e.g., complete coverage of construct), bias may still exist in
the scores, resulting in measurement-unit or scalar nonequivalence (van de
Vijver & Leung, 1997). This may be a result of method bias. Method bias
can stem from characteristics of the instrument or from its administration
(van de Vijver, 2001; van de Vijver & Leung, 1997; van de Vijver &
Poortinga, 1997). Possible sources of this bias are differential response
styles (e.g., social desirability) across cultures (e.g., Johnson, Kulesa, Cho,
& Shavitt, 2005), variations in familiarity with the type of stimuli or scale
across cultures, communication problems between investigators and partici-
pants, and differences in physical conditions under which the instrument is
administered across cultures. Method bias can also limit cross-cultural com-
parisons when samples drawn from different cultures are not comparable
(e.g., prior experiences).
Item bias may also exist, posing a threat to cross-cultural comparison
(scalar equivalence). This type of bias refers to measurement at the item
level.
Downloaded from http://tcp.sagepub.com by SJO TEMP 2008 on December 7, 2008
Ægisdóttir et al. / CROSS-CULTURAL VALIDITY 195
This bias has several potential sources. It can result from poor translation or
poor item formulation (e.g., complex wording) and because item content
may not be equally relevant or appropriate for the cultural groups being
compared (e.g., Malpass & Poortinga, 1986; van de Vijver & Poortinga,
1997). An item on an instrument is considered biased if persons from
different cultures having the same standing on the underlying characteristic
(trait or state) measured yield different average item scores on the
instrument.
Finally, bias can be considered uniform and nonuniform. A uniform bias
refers to any type of bias affecting all score levels on an instrument equally
(van de Vijver and Leung, 1997). For instance, when measuring persons’
intelligence, the scale may be accurate for one group but may consistently
reflect 10 points too much for another group. The 10-point difference would
appear at different intelligence levels (a true score of 90 would be 100, and a
true score of 120 would be 130). A nonuniform bias is any type of bias
differentially affecting different score levels. In measuring persons’
intelligence, the scale may again be accurate for one group, but for the other
group, 10 points are recorded as 12 points. The difference in measured
intelligence for persons whose true score is 90 would be a score of 108 (18-
point difference), whereas for persons whose true score is 110, the differ-
ence is 22 points (a score of 132). The distortion is greater at higher levels
on the scale. Nonuniform bias is considered a greater threat in cross-cultural
comparisons than uniform bias, as it influences the origin and measurement
unit (scale) of a scale. Uniform bias affects only the origin of a scale (cf. van
de Vijver, 1998, 2001).
MEASUREMENT APPROACHES
listed two common strategies employed when using preexisting measures for
multilingual groups. First is the applied approach, where an instrument goes
through a literal translation of items. Item content is not changed to a new
cultural context, and the linguistic and psychological appropriateness of the
items are assumed. It is also assumed there is no need to change the
instrument to avoid bias. According to van de Vijver (2001), this is the most
common technique in cross-cultural research on multilingual groups. The
second strategy is adaptation, where some items may be literally translated,
while others require modification of wording and content to enhance the
appropriateness to a new cultural context (van de Vijver & Leung, 1997).
This technique is chosen if there is concern with construct bias.
Of the three approaches just mentioned (assembly, application, and
adaptation), the application strategy is the easiest and least cumbersome in
terms of money, time, and effort. This technique may also offer high levels
of equivalence (measurement-unit and scalar equivalence), and it can make
the comparison to results of other studies using the same instrument possi-
ble. This approach may not be useful, however, when the characteristic
behaviors or attitudes (e.g., obedience and being a good daughter or son)
associated with the construct (e.g., filial piety) differ across cultures (lack of
construct equivalence and high construct bias) (e.g., Ho, 1996). In such
instances, the assembly or adaptation strategy may be needed. With the
assembly approach (emic), researchers may focus on the construct validity
of the instrument (e.g., factor analysis, divergent and convergent validity),
not on direct cross-cultural comparisons. When adaptation of an instrument
is needed in which some items are literally translated, whereas others are
changed or added, cross-cultural comparisons may be challenging, as direct
comparisons of total scores may not be feasible because all items are not
identical. Only scores on identical items can be compared using mean score
comparisons (Hambleton, 2001). The application technique (etic) to trans-
lation most easily allows for a direct comparison of test scores using t tests
or ANOVA because of potential scalar equivalence. For such comparisons
to be valid, however, an absence of bias needs to be demonstrated.
The applied approach and to some degree the adaptation strategy focus
on capturing the etics, or the qualities of concepts common across cultures.
Yet cultural researchers have criticized it. Berry (1989), for instance, labeled
this practice “imposed etics,” claiming that by using the etic approach,
researchers fail to capture the culturally specific aspects of a construct and
may erroneously assume the construct exists and functions similarly across
cultures (cf. Adamopolous & Lonner, 2001). The advantage of the etic over
the emic strategy, however, is that the etic technique provides the ability to
make cross-cultural comparisons, whereas in the emic approach, cross-
cultural comparison is more difficult and not as direct.
reliability, item-total scale correlations, and item means and variations pro-
vides initial information about instruments’ psychometric properties. A sta-
tistical comparison between two independent reliability coefficients can be
performed (cf. van de Vijver & Leung, 1997). If the coefficients are signifi-
cantly different from each other, the source of the difference should be
examined. This may indicate item or construct bias. Additionally, item-total
scale correlations may indicate construct bias and nonequivalence, and
method bias (e.g., administration differences, differential social desirability,
differential familiarity with instrumentation). Finally, item score distribution
may suggest biased items and, therefore, information about equivalence. For
instance, an indicator (e.g., item or scale) showing variation in one cultural
group but not the other may represent an emic concept (Johnson, 1998).
Therefore, comparing these statistics across different language versions of
an instrument will offer preliminary data about the instruments’ equivalence
(e.g., construct, measurement unit, and scalar; van de Vijver & Leung, 1997;
conceptual and measurement; Lonner, 1985).
Construct (van de Vijver & Leung, 1997), conceptual, and measurement
equivalence (Lonner, 1985) can also be assessed at the scale level. Here,
exploratory and confirmatory factor analysis, multidimensional scaling
techniques, and cluster analysis can be used (e.g., van de Vijver & Leung,
1997). These techniques provide information about whether the construct is
structurally similar across cultures and if the same meaning is attached to the
construct. For instance, in confirmatory factor analysis, hypotheses about the
factor structure of a measure, such as the number of factors, load-ings of
variables on factors, and correlations among factors, can be tested.
Numerous fit indices can be used to evaluate the fit of the model to the data.
Scalar or full score equivalence is more difficult to establish than con-
struct and measurement-unit equivalence, and various biases may threaten
this level of equivalence. Item bias, for instance, influences scalar
equivalence. Item bias can be ascertained by studying the distribution of
item scores for all cultural groups (cf. van de Vijver & Leung, 1997). Item
response theory (IRT), in which differential item functioning (DIF) is
examined, may be used for this purpose. In IRT, it is assumed item
responses are related to an underlying or latent trait using a logistic curve
known as item characteristic curve (ICC). The ICCs for each selected
parameter (e.g., item difficulty or popularity) are compared for every item in
each cultural group using chi-square statistics. Items differing between
cultural groups are eliminated before cross-cultural comparisons are made
(e.g., Hambleton & Swaminathan, 1985; van de Vijver & Leung, 1997).
Item bias can also be examined by using ANOVA. The item score is treated
as the dependent vari-able, and the cultural group (e.g., two levels) and score
levels (levels depen-dent on number of scale items and number of
participants scoring at each
Downloaded from http://tcp.sagepub.com by SJO TEMP 2008 on December 7, 2008
200 THE COUNSELING PSYCHOLOGIST / March 2008
level) are the independent variables. Main effects for culture and the inter-
action between culture and score level are then examined. Significant effects
indicate biased items (cf. van de Vijver & Leung, 1997). Logistic regression
can also be used for this purpose using the same type of independent and
dependent variables. Additionally, multiple-group SEM invariance analy-ses
(MCFA) and multiple group mean and covariance structures analysis
(MACS) also provide information about biased items or indicators (e.g.,
Byrne, 2004; Cheung & Rensvold, 2000; Little, 1997, 2000), with the
MACS method also providing information about mean differences between
groups on latent constructs (e.g., Ployhart & Oswald, 2004).
Finally, factors contributing to method bias can be assessed and statisti-
cally held constant when measuring constructs across cultures, given that
valid measures are available. A measure of social desirability may, for
instance, be used to partially control for method bias. Also, gross national
product per capita may be used to control for method bias, as it has been
found to correlate with social desirability (e.g., Van Hemert, van de Vijver,
Poortinga, & Georgas, 2002) and acquiescence (Johnson et al., 2005).
Furthermore, personal experience variables potentially influencing the con-
struct under study differentially across cultures may serve as covariates.
Translation Methodology
Employing a proper translation methodology is extremely important to
increase equivalence between multilingual versions of an instrument and the
measures’ cross-cultural validity. About a decade ago, van de Vijver and
Hambleton (1996) published practical guidelines for translating psycholog-
ical tests that were based on standards set forth in 1993 by the International
Test Commission (ITC). The guidelines covered best practices in regard to
context, development, administration, and the interpretation of psychologi-
cal instruments (cf. Hambleton & de Jong, 2003; van de Vijver, 2001; van
de Vijver & Hambleton, 1996; van de Vijver & Leung, 1997). The context
guidelines emphasized the importance of minimizing construct, method, and
item bias and the need to assess, instead of assume, construct similarity
across cultural groups before embarking on instrument translation. The
development guidelines referred to the translation process itself, while the
administration guidelines suggested ways to minimize method bias. Finally,
the interpretation guidelines recommended caution when explaining score
differences unless alternative hypotheses had been ruled out and equivalence
between original and translated measures had been ensured (van de Vijver &
Hambleton, 1996). Counseling psychologists should review these guidelines
when designing cross-cultural research projects and prior to translating and
adapting psychological instruments for such research.
Prior to the development of the ITC standards, Brislin et al. (1973) and
Brislin (1986) had written extensively about translation procedures. The
following paragraphs outline the common translation methods that Brislin et
al. summarized with connotations to the ITC guidelines (e.g., Hambleton &
de Jong, 2003; van de Vijver & Hambleton, 1996). Additional methods to
enhance equivalence of translated scales are also mentioned.
At other times, the original language version of the scale is also changed to
ensure equivalence, a process known as decentering (Brislin et al., 1973).
Adequate back translation does not guarantee a good translation of a scale,
as this procedure often leads to literal translation at the cost of readability
and naturalness of the translated version. To minimize this, a team of back
translators with a combined expertise in psychology and linguistics may be
used (van de Vijver & Hambleton, 1996). It is also important to note that in
addition to the test items, test instructions need to go through a thorough
translation/back-translation process.
METHOD
Sample
The sample consisted of published studies employing translated instru-
ments in their data collection. To be included in this project, an integral part
of the study’s methodology had to be a translation of one or more entire
instrument or some subset of items from an instrument. Furthermore, the
target instrument could not have been translated or evaluated the same way
in earlier studies. Additionally, the included studies had to either compare
responses from persons from more than one culture (nationality) or inves-
tigate a psychological concept using a non-U.S. or non-English-speaking
sample of participants. Studies for this investigation were sampled from five
counseling journals (Journal of Counseling Psychology [JCP], Journal of
Counseling and Development [JCD], Journal of Multicultural Counseling
and Development [JMCD], Measurement and Evaluation in Counseling and
Development [MECD], and The Counseling Psychologist [TCP]) thought to
publish articles relevant to non-English-speaking cultures, eth-nic groups,
and/or countries. To assess for more recent trends in the litera-ture, only
articles published between the years 2000 and 2005 were included in our
sample. We assumed recent studies (i.e., studies published since 2000)
would provide a good representation of current translation and verification
practices employed by counseling researchers. From 2000 to 2005, a total of
615 empirical articles were published in the targeted jour-nals. Of these
articles, 15 included translation as a part of their methodol-ogy. Therefore,
2.4% of the empirical articles published in these five counseling journals
incorporated a translation process.
Procedure
The 15 identified studies were coded by (a) publication source (e.g.,
TCP), (b) year of publication (e.g., 2001), (c) construct investigated and
name of scale translated, (d) translation methodology used (single person,
committee, bilinguals), (e) whether the translated version of the scale was
pilot tested (yes or no) before main data collection, (f) number of partici-
pants used for pilot testing, (g) psychometric properties reported and statis-
tics used to evaluate the translated measure’s equivalence to the original
scale, and (h) number of participants from which the psychometric data were
gathered. Two of the current authors coded information from the arti-cles
independently. If disagreements arose in the coding (e.g., relevant
psychometrics for equivalence evaluation), these were resolved through
consensus agreement between the coders.
Downloaded from http://tcp.sagepub.com by SJO TEMP 2008 on December 7, 2008
206 THE COUNSELING PSYCHOLOGIST / March 2008
RESULTS
Table 1 lists results found for each of the 15 studies. Three of the
included studies used a structured or semistructured interview-test protocol.
In 3 studies, of which one included a semistructured test protocol, an
English-language instrument was developed and then translated to another
language. Furthermore, in 9 studies, one or more preexisting measures (the
entire instrument or subset of items) were translated into a language other
than English. In the 15 studies, a range of constructs was examined, includ-
ing persons’ counseling orientations (e.g., help-seeking attitudes, counsel-
ing expectations), adjustment (e.g., acculturation), and maladjustment (e.g.,
psychological stress). A diversity of cultural groups was represented in the
15 studies as well (see Table 1).
1. Shin, Berkson, & Psychological Immigrants from Six items from theanalysis English to Korean Committee Yes
No N/A ATSPPH: Factor
Crittenden (2000); help-seeking Korea Attitudes Toward AAS: Cronbach's alpha
JMCD attitudes; Seeking Professional (N=110 Korean immigrants
traditional values Psychological Help in U.S.)
(ATSPPH);
A
c
c
u
l
t
u
r
a
t
i
o
n
A
t
t
i
t
u
d
e
S
c
a
l
e
,
(
A
A
S
)
p
r
i
o
r
t
r
a
n
s
l
a
t
i
o
n
;
V
i
g
n
e
t
t
e
s
d
e
v
e
l
o
p
e
d
i
n
E
n
g
l
i
s
h
2. Engels, Finkenauer, Parental attachment; Dutch adolescents Parent and Peer English to Dutch Committee Yes
No N/A Cronbach's alpha (N=412
Meeus, & Dekovic Relational
(researchers) Attachment
Dutch (IPPA); (researchers);
(2001); JCP competence; Perceived Competence unclear what adolescents)
Self-esteem; Scale for Children; instruments were
Depression Self-Esteem Scale; translated in study
Depressive Mood List
3. Chung & Bemak Anxiety;
Pilot Southeastern
N/A Asian Health Opinion
Exploratory Survey
factory analysis English to Committee Yes
(2002); JCD depression; refugees (interview) Vietnamese, interviews for Vietnamese
(N=867),
psychosocial Khmer, Laotian Cambodian (N=590), and
dysfunction
symptoms
4. Kasturirangan & Culture Latino
N/A women A semiprofessor
structuredof foreign English to Spanish. Not No
Pilot interview; no Latina
Nutt-Williams Domestic
English violence
version interview served
language protocolas an auditor No discussion of reported comparison
(2003); JMCD developed by thetranslation translation method between of protocol
to ensure proper
administered to researchers: Two
interviews in English,
(n=3) Latina
women seven in Spanish
rior to data
ollection
5. Asner-Self & Attributional
No style Immigrants
N/A from The Attributional
Cronbach's alpha, Style
principle English to Committee Yes
Schreiber (2004); Central America Questionnaire (ASQ) Spanish components analysis (N=89
MECD
mmigrants in U.S.)
6. Torres & Rollock Acculturation-related Immigrants
Not reportedfrom Cultural Adjustment English to Spanish Committee Yes
No for Cronbach's alpha (N=86
(2004); MECD challenges Central & South Difficulties Checklist the 10% of the Hispanic immigrants). 90%
America (CADC) sample that of the sample responded to
esponded to
his version
r
(
TABLE 1: (continued)
Assigned
Number, Approach
Citation, Type of
Back Instrument to
and Journal Construct Sample Name Translation Translation
Translation Pretest Original Target
Equivalence studies
14. Chang & Wellness
Yes (n = 3): Immigrants from
None reported for The
NoneWellness
reportedEvaluation
for a larger English to Korean Single translator No
Myers (2003); Korea of Lifestyle (WEL) whose
Bilingual exam- a larger sample sampletranslations
(N not reported)
MECD were edited
inees by first
took both (N not
author.
reported)
Discrepancies
resolved between
translator and
editor upon mutual
in
agreement
etween
nglish and
orean version
15. Mallinckrodt & Adult attachment
No Int'l students from
Split-half The Experiences
Used bilinguals (nin= 30 English to Chinese Committee Yes
Wang (2004); JCP Taiwan Close Relationships reliability, Taiwanese international
Scale (ECRS)
college students) to evaluate
lpha (N = 399
.S. college
tudents)
v
r
c
210 THE COUNSELING PSYCHOLOGIST / March 2008
Interpretation of Results
The current results are consistent with Mallinckrodt and Wang (2004),
who discovered in their review of articles published in two counseling jour-
nals (JCP and TCP) that few studies in counseling psychology have inves-
tigated multilingual or international groups or employed translation
methods. Additionally, consistent with these investigators, we found in
many instances, counseling researchers used inadequate procedures to verify
equivalence between language versions of an instrument. For example, our
analyses
Downloaded from http://tcp.sagepub.com by SJO TEMP 2008 on December 7, 2008
Ægisdóttir et al. / CROSS-CULTURAL VALIDITY 213
RECOMMENDATIONS
Translation Practices
Several steps are essential for a valid translation. Based on our and
Brislin and colleagues’ (Brislin, 1986; Brislin et al., 1973) review of
common translation methods and the ITC guidelines (e.g., Hambleton, 2001;
van de Vijver & Hambleton, 1996), the best translation procedure involves
several steps as
1. Independent translation from two or more persons familiar with the target lan-
guage and culture and intent of the scale
2. Documentation of comparisons of translations and agreement on the best
translation
3. Rewriting of translated items to fit grammatical structure of target language
4. Independent back translation of translated measure into original language (one or
more persons)
5. Comparison of original and back-translated versions, focusing on appropriate-
ness, clarity, meaning (e.g., use rating scales)
6. Changes to the translated measure based on prior comparison. Changed items go
through the translation/back-translation iteration until satisfactory
7. If concepts or ideas do not translate well, deciding what version of the original
version of a scale should be used for cross-cultural comparison (original, back
translated, or decentered)
8. Pretest of translated instrument on an independent sample (bilinguals or target
language group). Check for clarity, appropriateness, and meaning
9. Assessment of the scale’s reliability and validity, absence of bias, and equiva-
lence to the original-language version of the scale
outlined in Table 2. All but the last step outlined in this table help to
minimize item and construct bias and therefore may increase scalar
equivalence between language versions of a measure (ITC development
guidelines). The last step or recommendation refers to verifying cross-
cultural validity of measures (i.e., absence of bias and equivalence; ITC
interpretation guidelines).
enhances the validity of the findings. Third, when method bias is not
expected but there is a potential for construct bias while the use of a
preexisting mea-sure is considered feasible, researchers should consider
collecting emic items to be included in the instrument when studying an etic
construct (e.g., Brislin, 1976; Oh & Neville, 2004). This approach will
enhance construct equiva-lence by limiting construct bias and will provide
culture-specific information to aid theory development. Fourth, when emic
scales are available in the cul-tures of interest to assess an etic construct and
cross-cultural comparisons are sought, the convergence approach should be
considered. With this approach, all instruments are translated and
administered to each cultural group. Then, items and scales shared across
cultures are used for cross-cultural compar-isons, whereas nonshared items
provide information about the unique aspect of the construct in each culture
(e.g., van de Vijver, 1998). This approach will enhance construct
equivalence, it may deepen the current understanding of cultural and cross-
cultural dimensions of a construct, and it may aid theory development.
Finally, Triandis’s (1972, 1976) suggestion can be considered. With this
procedure, instruments are simultaneously assembled in each cul-ture to
measure the etic construct (e.g., subjective well-being). With this approach,
most or all types of biases can be minimized and equivalence enhanced, as
no predetermined stimuli are used. Factor analyses can be per-formed to
identify etic constructs for cross-cultural comparisons.
CONCLUSION
REFERENCES
Adamopolous, J., & Lonner, W. J. (2001). Culture and psychology at a crossroad: Historical
perspective and theoretical analysis. In D. Matsumoto (Ed.), The handbook of culture and
psychology (pp. 11-34). New York: Oxford University Press.
American Psychological Association. (2003). Guidelines on multicultural education, training,
research, practice, and organizational change for psychologists. American Psychologist,
58, 377-402.
Arrendondo, P., Toporek, R., Brown, S. P., Jones, J., Locke, D. C., Sanchez, J., et al. (1996).
Operationalization of the multicultural counseling competencies. Journal of Multicultural
Counseling and Development, 24, 42-78.
Asner-Self, K. K., & Marotta, S. A. (2005). Developmental indices among Central American
immigrants exposed to way-related trauma: Clinical implications for counselors. Journal
of Counseling and Development, 83, 162-171.
Asner-Self, K. K., & Schreiber, J. B. (2004). A factor analytic study of the attributional style
questionnaire with Central American immigrants. Measurement and Evaluation in
Counseling & Development, 37, 144-153.
Berry, J. W. (1969). On cross-cultural comparability. International Journal of Psychology, 4,
119-128.
Berry, J. W. (1989). Imposed etics-emics-derived etics: The operationalization of a compelling
idea. International Journal of Psychology, 26, 721-735.
Betz, N. E. (2005). Enhancing research productivity in counseling psychology: Reactions to
three perspectives. The Counseling Psychologist, 33, 358-366.
Brislin, R. W. (1970). Back-translation for cross-cultural research. Journal of Cross-Cultural
Psychology, 1, 185-216.
Brislin, R. W. (1976). Comparative research methodology: Cross cultural studies. International
Journal of Psychology, 11, 213-229.
Jahoda, G. (1966). Geometric illusions and environment: A study in Ghana. British Journal of
Psychology, 57, 193-199.
Johnson, T., Kulesa, P., Cho, Y. I., & Shavitt, S. (2005). The relation between culture and
response styles: Evidence from 19 countries. Journal of Cross-Cultural Psychology,
36, 264-277. Johnson, T. P. (1998). Approaches to equivalence in cross-cultural and cross-
national survey research. In ZUMA (Centrum fur Unfragen Methoden und Analysen)-
Nachrichten Spezial Band 3: Cross-Cultural Survey Equivalence (pp. 1-40).
Retrieved from http://www.gesis
.org/Publikationen/Zeitschriften/ZUMA_Nachrichten_spezial/zn-sp-3-inhalt.htm
Kasturirangan, A., & Nutt-Williams, E. (2003). Counseling Latina battered women: A qualita-
tive study of the Latina perspective. Journal of Multicultural Counseling and Development,
31, 162-178.
Kim, U. (2001). Culture, science, and indigenous psychologies. In D. Matsumoto (Ed.), The
handbook of culture and psychology (pp. 51-76). New York: Oxford University Press.
Leong, F. T. L., & Blustein, D. L. (2000). Toward a global vision of counseling psychology.
The Counseling Psychologist, 28, 5-9.
Leong, F. T. L., & Ponterotto, J. G. (2003). A proposal for internationalizing counseling
psychology in the United States: Rationale, recommendations, and challenges. The
Counseling Psychologist, 31, 381-395.
Leung, S. A. (2003). A journey worth traveling: Globalization of counseling psychology. The
Counseling Psychologist, 31, 412-419.
Little, T. D. (1997). Mean and covariance structures (MACS) analyses of cross-cultural data:
Practical and theoretical issues. Multivariate Behavioral Research, 32, 53-76. Little, T. D.
(2000). On the comparability of constructs in cross-cultural research: A critique of
Cheung and Rensvold. Journal of Cross-Cultural Psychology, 31, 213-219. Lonner, W. J.
(1985). Issues in testing and assessment in cross-cultural counseling. The Counseling
Psychologist, 13, 599-614.
Lonner, W. J., & Berry, J. W. (Eds.). (1986). Field methods in cross-cultural research. Beverly
Hills, CA: Sage.
Mallinckrodt, B., & Wang, C.-C. (2004). Quantitative methods for verifying semantic equiva-
lence of translated research instruments: A Chinese version of the Experiences in Close
Relationships Scale. Journal of Counseling Psychology, 51, 368-379.
Malpass, R. S., & Poortinga, Y. H. (1986). Strategies for design and analysis. In J. W. Berry &
J. W. Lonner (Eds.), Cross-cultural research and methodology series: Vol. 8. Field meth-
ods in cross-cultural research (pp. 47-83). Beverly Hills, CA: Sage.
McCrae, R. R., & Costa, P. T. (1997). Personality trait structure as a human universal.
American Psychologist, 52, 509-516.
Oh, E., & Neville, H. (2004). Development and validation of the Korean rape myth acceptance
scale. The Counseling Psychologist, 32, 301-331.
Pedersen, P. B. (1991). Counseling international students. The Counseling Psychologist, 19,
10-58. Pedersen, P. B. (2003). Culturally biased assumptions in counseling psychology.
The Counseling Psychologist, 31, 396-403.
Ployhart, R. E., & Oswald, F. L. (2004). Applications of mean and covariance structure
analysis: Integrating correlational and experimental approaches. Organizational
Research Methods, 7, 27-65.
Poasa, K. H., Mallinckrodt, B., & Suzuki, L. A. (2000). Causal attributions for problematic
family interactions: A qualitative, cultural comparison of Western Samoa, American Samoa,
and the United States. The Counseling Psychologist, 28, 32-60.
Ponterotto, J. G., Casas, J. M., Suzuki, L. A., & Alexander, C. M. (Eds.). (1995). Handbook of
multicultural counseling (2nd ed.). Thousand Oaks, CA: Sage.
Shin, J. Y., Berkson, G., & Crittenden, K. (2000). Informal and professional support for
solving psychological problems among Korean-speaking immigrants. Journal of
Multicultural Counseling and Development, 28, 144-159.
Downloaded from http://tcp.sagepub.com by SJO TEMP 2008 on December 7, 2008
Ægisdóttir et al. / CROSS-CULTURAL VALIDITY 219