Measurement, 6: 147–154, 2008 Copyright © Taylor & Francis Group, LLC ISSN 1536-6367 print / 1536-6359 online DOI: 10.1080/15366360802122006

Putting Concepts and Constructs into Practice: A Reply to Cervone and Caldwell, Haig, Kane, Mislevy, and Rupp Keith A. Markus John Jay College of Criminal Justice

The commentary has greatly enriched the discussion initiated by the three target articles. The distinction between constructs and concepts contributes to both the top-down and bottom-up aspects of the dialectic between measurement theory and practice. The distinction also illuminates the abstraction from observation to manifest variables. However, the semantic analysis of variables in terms of ordered pairs of individuals and values of variables does not seek to describe a procedure for defining constructs as part of the test development process. As such, uncertainty about population membership does not pose a pragmatic constraint on the application of the distinction between concepts and constructs. Finally, meaning and reference in the context of test development can best be understood as jointly determined by both how the world is and the nature of the vocabularies chosen to describe it. Thus, both of these factors bear on the definition of specific constructs and concepts measured by individual tests. Distinguishing constructs from concepts can help clarify and advance discourse on testing and measurement across a wide range of domains. Key words: construct, measure, test, validity, variable

I want to thank all of the authors of the commentaries for enriching this set of articles with their contributions. I found the comments from Hood, Kyngdon, and Maraun and Halpin valuable, but for practical reasons have chosen to focus Correspondence should be addressed to Keith A. Markus, Department of Psychology, John Jay College of Criminal Justice, The City University of New York, 445 W59th Street, New York, NY 10019. E-mail: [email protected]

148

MARKUS

on the comments by Cervone and Caldwell, Haig, Kane, Mislevy, and Rupp in my reply. I want to offer some brief clarifications and expand on some of the issues raised in these commentaries. Cervone and Caldwell: Can All of These Islands Fit Beneath One Sky? Cervone and Caldwell (2008) helpfully articulated the diversity and complexity of empirical research programs and the challenges that these pose for a general measurement theory. Consider a dichotomous item from an object recognition task in a semantic memory experiment: the participant is shown an image of a vase and asked to name what he or she sees. The researcher deems the answer correct or incorrect. In one context, the researcher may simply throw away data for which participants give the wrong answer and justify it as being outside the scope of the study. In another context, the items might be summed to measure error rate or response accuracy overall. In another context, the item might measure vocabulary or English proficiency. In yet another, it might measure cognitive load induced by a secondary task. Cervone and Caldwell seem on target in suggesting that general measurement theory tends to adopt a stance rooted in individual differences and to a lesser extent the measurement of outcomes familiar to the sorts of classical experimental designs taught in introductory research methods courses. I suspect that this, in large part, represents a lamp post approach in which measurement theorists start with the simplest cases and work toward the more complex ones, in an attempt to develop a general theory of measurement. Cervone and Caldwell rightly pointed out some of the complexities apt to get lost in such an approach and the need for measurement theory to attend to the diversity of contemporary research in psychology and the behavioral sciences. General systematization always carries the risk of oversimplification or exclusion through idealization, and in visiting the varied islands that make up the archipelago of behavioral science research, measurement theory does well to steer a path between the Scylla of oversimplification and the Charybdis of exclusionary idealization. However, pausing to envision the opposite extreme also sheds some light on the situation. In many ways the history of measurement theory involves diverse procedures developed as ad hoc solutions to practical problems leading to puzzles or stumbling blocks, which later lead to their reformulation or reconceptualization under an integrative framework that helps to resolve the puzzles or avoid the stumbling blocks introduced by ad hoc solutions. If every island in the archipelago developed its own local measurement theory independent of the others, this would replace the dangers of general systematization with equal and opposite dangers of parochialism and the absence of coordinated effort. A

REJOINDER

149

purely bottom up approach would replace conjoint measurement theory with disjoint measurement theory. Ideally, local applications should inform and guide general measurement theory, and general measurement theory should shape the conceptualization of local measurement issues and practices. Both directions of influence contribute to the larger project. If one considers the various uses of the vase item, the distinction drawn between constructs and concepts contributes to both the top-down and the bottom-up project. Clear thinking about measurement at both levels requires a clear distinction between extensional and intensional objects of measurement and between actualities and possibilities as identity criteria for variables. Even the most basic interpretation of the response as indicative of correct identification of the vase image as a vase touches on issues of how the response might have differed had things been otherwise. These issues, in turn, figure centrally in the interpretation of responses from different participants in different conditions. These considerations also bear on an issue raised in the introduction, where Haig and Borsboom (2008) characterized the application of the term measurement to categories as an undesirable feature of common use. This characterization seems to artificially limit measurement theory in undesirable ways. For instance, it seems very awkward to build a measurement theory, that includes the use of the above item as a measure of cognitive load on a foundation that does not acknowledge the item as a measure of whether the participant correctly identifies the vase. Haig: Viewing the World as Variables Although Haig (2008) chose not to comment on my article directly, both of our pieces share a common theme: how researchers translate the world into variables. Haig describes manifest variables as abstracted from observational evidence, and the distinction between constructs and concepts bears on the extent and character of that abstraction. Kane: Formality, Informality, and Misinformality In his commentary, Kane (2008) revisited issues from his influential work on validity arguments (Kane, 1992) involving the characterization of informal argument and its role in validity theory. While I find much to admire in this work and its contributions to the understanding of validity arguments, I also bristle a bit at the identification of logic and formalization only with axiomatic deductive systems. In practice, formalization applies to informal reasoning and can offer a valuable tool for the advancement of that area as well. It would be unfortunate if Kane’s important contributions in applying the theory of informal

150

MARKUS

argument to test validity resulted in a failure to make use of formalization to advance the very program that he advocates. Moreover, one should not conflate the formal/informal distinction with the distinction between deductive and ampliative inference (comprising inductive and abductive inference). Much statistical reasoning involves formal ampliative reasoning, and much informal reasoning is nonetheless deductive in nature. Kane suggested that a definition of a construct in terms of a set of values of members of a population fails to provide an informative construct definition and would not be helpful to psychologists and educators. This observation rests on a misunderstanding and perhaps a lack of clarity in my article. Of course, anyone who works with data does name columns of numbers with variable names, and software developers would be surprised to learn that this feature of their products was of no help to psychologists and educators. Beyond this rather trivial practice, however, the analysis in my paper was never intended as a method for generating construct definitions used in test development or test use. Instead, it was intended as a semantic analysis of what I termed constructs and concepts.1 I believe that greater familiarity with the tools of possible world semantics would be useful for methodologists and test theorists. However, a clear distinction between constructs and concepts in the daily practice of test development and use does not require that test developers, psychologists, or educators define their concepts in terms of vectors of values paired with members of a population. A further passage from Kane’s commentary helps to develop the above point. Kane suggested that test developers would not interpret an English test and algebra test as measuring the same construct if they correlated perfectly in a given population. I suspect not. However, here we touch on the point at which informality shades off into misinformality—informality gone awry. In classical deductive logic, anything follows from a contradiction, and contradictory premises are thus devastating to inference, but mere appeal to informal reasoning does not warrant indifference to inconsistencies in the basis for the validity argument. If test developers would not interpret these tests as measuring the same construct, but nonetheless appeal to a basic measurement theory that has this implication, then this points to a problem in the validity argument that requires attention. In the terminology developed in my article, it seems to me that the two tests do measure the same construct, but that they measure different concepts. The intuitive resistance to this conclusion reflects the internal tensions within the common use of the term construct to mean either or both. The fact that different assumptions commonly adopted by test developers pull in opposite directions on this point indicates the utility to test developers of 1 It seems worth noting by way of clarification that although the formal apparatus is Fregean in origin, the semantics are anything but Fregean (Burge, 2005).

REJOINDER

151

minding this distinction, even when working with informal arguments. Kane’s final example seems to suggest that he envisions test developers getting by with just concepts, but calling them constructs. In my article I tried to make clear why test developers need both. Test developers regularly confront situations in which the same concept corresponds to different constructs—as in cases of differential prediction—and a vocabulary that fails to incorporate this distinction inhibits effective communication about such situations. Nonetheless, I agree that content and development issues play an important role in disambiguating concepts from one another and fixing the interpretations of test scores. The target article underplayed this important aspect of test validity as a consequence of placing its emphasis elsewhere. Rupp: Semantics, Pragmatics, and Pragmatics Among many interesting things, Rupp (2008) suggested a focus on the process by which the meanings of constructs are negotiated through consensus building. I see improving the richness and elaboration of the vocabulary available to discuss the meaning of constructs and concepts as valuable to enhancing the efficacy of such processes. Indeed, equivocation between concepts and constructs could easily result in an artificial dissensus, which a clear distinction between the two could help avoid. One goal of clarifying our terminology is to help users of that terminology to avoid talking past one another. Rupp also suggested that measurement error, or uncertainty more generally, involving population membership introduces imprecision into construct definitions. If test developers and users defined constructs in these terms, this would indeed pose a technical concern as Rupp suggested. However, these reflections rest on the same misunderstanding brought to light in Kane’s (2008) commentary. Aside from naming columns of numbers in a data file, no test developer or user should attempt to define a construct by listing or otherwise enumerating a set of ordered pairs of cases and values. The semantic analysis of what a construct means, on the other hand, holds independently of whether anyone could know with full certainty which cases belong to which populations. This is an analysis of what makes an assertion true, or what makes it entail another assertion. This is not a methodological prescription for writing construct definitions. The semantic analysis merely serves as a tool to clarify terms offered for use by test developers and test users. One does not need the semantic analysis to distinguish constructs from concepts, only to justify and clarify that distinction. (I also want to clarify that where Rupp translates my ideas in terms of carelessness or ignorance, these are translations of my ideas into a foreign vocabulary. I do not think in these terms, nor did I envision my article in these terms.) It is helpful to mind a distinction between the everyday sense of pragmatic1 meaning responsive to practical concerns, and the technical sense of pragmatic2

152

MARKUS

meaning performing actions with language which closely relates to the sorts of contextual constraints on meaning with which Rupp closed his commentary. This relation stems from the concrete and individual nature of actions performed. In my view, it is a mistake to think that because pragmatics2 affects meaning that language use does not involve the kind of reference that is subject to more traditional semantic analysis. Semantics and pragmatics2 do not offer competing explanations of meaning, but rather complementary aspects of meaning. As such, the distinction between constructs and concepts derived from and elucidated by semantic analysis applies to situationally specific and contextually constrained uses of these terms. It is my hope that the distinction between concepts and constructs facilitates careful attention to pragmatic1 concerns. I find it difficult to imagine pragmatic1 concerns that work at cross purposes with the distinction that I have proposed—although I recognize that failure of imagination does not in and of itself offer much support for a negative assertion. I see pragmatic1 constraints on the collection of test validity evidence, such as cost, time, and return on investment. I see pragmatic1 constraints on the degree of precision with which test developers can and should define a construct, such as limited theoretical and empirical understanding of the phenomenon, pragmatic1 idealization in the face of complexity, and limited incentive to seek precision at a level beyond that required by the purpose of the test or tests in question. However, the distinction between constructs and concepts does not require undue precision in the definition of particular constructs or concepts; the semantic analysis of the general terms “construct” and “concept” differs from the definitions-in-use of specific exemplars of these two general ideas. None of the examples given in my article involved great precision in the definition of the constructs or concepts (e.g., academic honesty, integrity). Distinguishing constructs from concepts may help test developers and researchers refine the definitions of their variables, but the distinction does not assume great precision in these definitions. The semantic analysis analyzes what assertions about constructs and concepts mean, independent of how precisely those making the assertions can describe that meaning. Mislevy: Sociocognition of What? Mislevy (2008) provided a rich overview of the application of cognitive science, or at least aspects of cognitive psychology, to test validity and test development. For the purposes of this reply, the pivotal passage appeared in the section on modelbased reasoning in which Mislevy contrasted taking individuals and their properties as presumed features of the world, with taking these as elements of the conceptual frame that a person adopts as a means of reasoning about some situation presumed as part of the world. I cannot provide an adequate summary of all the nuances of Mislevy’s presentation here, but want to focus on this key contrast.

REJOINDER

153

It seems to me that the contrast offers a false alternative. On one hand Mislevy offered the reader a naive realism that assumes that the world determines a unique vocabulary used to describe it (Markus, 1998). On the other hand, Mislevy offered the reader a naive constructivism that assumes that if names and properties are part of the vocabulary chosen to talk about the world, then they are completely independent of how the world is. I want to suggest that faced with such a choice the reader should demure from choosing either hand and instead cry foul. What, then, lies between these two alternatives? The central theme uniting much of my work can be expressed as follows: the world does not uniquely determine its own description. By this, I mean that how the world is does not suffice to render a single description uniquely true for the portion of the world that it describes. One can always generate alternative true descriptions that conflict with one another because they frame their vocabularies differently (Markus, Hawes & Thasites, in press). Nonetheless, the world does place significant constraints on which descriptions hold true for what they describe, and which vocabularies are socially sustainable (Harré & Gillett, 1994). Two different theories can define integrity in contradictory ways and yet both provide accurate descriptions of the world. But there are plenty of wrong descriptions too. It is hard to provide an account of such wrongness if one restricts names and properties to the vocabulary (or frame) and denies them any purchase on the world they are used to talk about. It seems to me that a more useful picture involves thinking of our theories as being about aspects of the world, but perceived or described in a particular form jointly determined by both how the world is and the nature of the vocabulary adopted to describe it. Contrasting true theories get at the same aspect of how the world is by means of different vocabularies, and thus represent similar or identical facts in conflicting forms (false theories conflict not just with other theories but with the facts; Quine, 1969). However, one has no epistemic access to how the world is (facts) except through specific vocabularies that give those facts specific forms (Putnam, 1981). The one thing that I found least satisfying about Mislevy’s account of a sociocognitive approach is that it seemed locked in the individual information processing metaphor, based on a metaphor of individual perception, despite an apparent intention to escape that metaphor. I found very little of the social aspect of social cognition in the account and, if anything, would encourage further movement in that direction along the lines suggested by Rupp (2008). In my view, both constructs and concepts are infused with and shaped by social context and social processes. Vocabularies develop through processes of social interaction. None of this, of course, was evident in my target article. To extend the frame metaphor, the article was focused elsewhere and these matters were off camera. As a result, I appreciate the opportunity that Mislevy’s comments have provided to clarify and develop this context.

154

MARKUS

ZOOM OUT, FADE TO BLACK I want to close by again thanking all the authors who provided commentary for enriching and extending the discussion of the topics explored in the three target articles. The history of testing involves a dialectic between practice and theory, in which practice spawns various technical solutions to practical problems, theory reconceptualizes and integrates these (sharpening the understanding of the practical problems), and practice, in turn, generates new, technical solutions. It is something of a cliché to say that there is nothing as practical as a good theory, but it seems hard to deny that this dialectic between theory and practice has historically moved the field forward. I hope that in this reply I have helped to clarify the distinction between theory and practice (specifically between test theory and test development), and have also rendered more clearly how the theoretical arguments in my paper apply to the practice of testing, test development, and test validation. The comments illustrate the breadth of domains across which a distinction between constructs and concepts can help clarify and refine discourse on testing and measurement.

REFERENCES Burge, T. (2005). Truth, thought, reason: Essays on Frege. Oxford: Clarendon. Cervone, D., & Caldwell, T. L. (2008). From measurement theory to psychological theory, in reverse. Measurement, 6 (1–2), 84–88. Haig, B. D. (2008). Aspects of latent variables theory. Measurement, 6 (1–2), 88–93. Haig, B. D. & Borsboom D. (2008). On the conceptual foundations of psychological measurement. Measurement, 6 (1–2), 1–6. Harré, R. & Gillett, G. (1994). The discursive mind. Thousand Oaks, CA: Sage Publications. Kane, M. T. (2008). The benefits and limitations of formality. Measurement, 6 (1–2), 101–108. Kane, M. T. (1992). An argument-based approach to validity. Psychological Bulletin, 112, 527–535. Markus, K. A. (1998). Science, measurement, and validity: Is completion of Samuel Messick’s synthesis possible? Social Indicators Research, 45, 7–34. Markus, K. A., Hawes, S. A., & Thasites, R. J. (in press). Abductive inferences to psychological variables: Steiger’s question and best explanations of psychopathy. Journal of Clinical Psychology. Mislevy, R. J. (2008). How cognitive science challenges the educational measurement tradition. Measurement, 6 (1–2), 124. Putnam, H. (1981). Reason, truth and history. Cambridge: Cambridge University Press. Quine, W. V. O. (1969). Ontological relativity and other essays. New York: Columbia University Press. Rupp, A. A. (2008). Lost in translation? Meaning- and decision-making in acutal and possible worlds. Measurement, 6 (1–2), 117–123.

Putting Concepts and Constructs into Practice - Taylor & Francis Online

As such, uncertainty about population membership does not pose a pragmatic constraint on the application of the distinction between concepts and constructs.

86KB Sizes 0 Downloads 178 Views

Recommend Documents

Putting Concepts and Constructs into Practice - Taylor & Francis Online
College of Criminal Justice, The City University of New York, 445 W59th ... of empirical research programs and the challenges that these pose for a general.

e-government - Taylor & Francis Online
Although the term “electronic government” is relatively new, its practice and associated literature are not. Broadly defined for the purposes of this bibliogra- phy, e-government is the use of information and communication technologies. (ICTs) by

arbuscular mycorrhizal fungi - Taylor & Francis Online
hosts or cyanobacterial partners, progress in obtain- ing multigene phylogenies has been slow and the ... Accepted for publication 27 September 2006. 1 Corresponding author. E-mail: [email protected] ..... zal fungi and patterns of host associa

Hegemony after deconstruction: the ... - Taylor & Francis Online
ABSTRACT. Hegemonic decisions institute and shape the ideological terrain in which politics occurs; different forms of decision will structure the terrain in.

Modus Tollens, Modus Shmollens - Taylor & Francis Online
tastes like garlic'' from the premises ''If a soup tastes like garlic, then there is ... a seemingly outrageous fallacy we will call Modus Shmollens: ''If p then q; it is not ...

Disability Rights and Wrongs Revisited - Taylor & Francis
Sep 14, 2005 - Burton-Smith, R., McVilly, K.R., Yazbeck, M., Parmenter, T.R. and Tsutsui, ..... in Australia: Exposing a Social Apartheid, Sydney: UNSW Press.

Deleuze and World Politics - Taylor & Francis Group
and Peter Stirk. 36 John Stuart Mill – Thought and. Influence. The saint of rationalism. Edited by Georgios Varouxakis and Paul Kelly. 37 Rethinking Gramsci. Edited by .... l'Aide aux Citoyens (Association for the Taxation of Financial ...... Willi

Disability Rights and Wrongs - Taylor & Francis Group
Campbell, J. (2003) Don't be fooled, we don't all want to kill ourselves. Online. Available. HTTP: ... Carpenter, M. (1994) Normality is Hard Work: trades unions and the politics of community care, London: Lawrence and Wishart. Carr, L. (2000) Enabl

Media Freedom and Social Capital - Taylor & Francis Online
Sanghoon Lee. Department of Economics, Hannam University, Daejeon, Republic of Korea. ABSTRACT. This article examines the relationship between media freedom and social capital by using cross-country panel data. The hypothesis of the current study is

Population and Income: Is There a Puzzle? - Taylor & Francis Online
*University of Texas Pan American, USA, **University of Kent, UK, ***Catholic University Brasilia,. Brazil. Final version received August 2005. ABSTRACT. This note presents wide evidence on the relationship between population and income for 125 count

Mycologia: Vol 109, No 2 - Taylor & Francis Online
Published online: 26 Jun 2017 ... Published online: 17 May 2017 .... A phylum-level phylogenetic classification of zygomycete fungi based on genome-scale data.

NAFP Delegated Authority - Putting It Into Practice Sep11.pdf ...
NAFP Delegated Authority - Putting It Into Practice Sep11.pdf. NAFP Delegated Authority - Putting It Into Practice Sep11.pdf. Open. Extract. Open with. Sign In.

Pollination genetics: Using molecular genetic ... - Taylor & Francis Online
Aug 10, 2008 - ABSTRACT. Pollination ecology and plant–pollinator relationships are major topics in studies of floral evolution and plant speciation. The genetic basis of floral traits is a key factor in determining the evolutionary response to pol

Adolescent coping with everyday stressors: A ... - Taylor & Francis Online
The present study compares problem-specific coping strategies and coping styles of European adolescents from seven nations. The sample consisted of 3031 adolescent participants, aged 11 to 20, from Croatia, the Czech. Republic, Germany, Italy, Norway

3 A AG Taylor S. Francis
cut programs without answering to a state bu- reaucracy." .... open system with multiple substances or pro- .... field's, but the voice was not especially audible,.

Putting Language into Language Modeling - CiteSeerX
Research in language modeling consists of finding appro- ..... L(i j l z) max l2fi j;1g v. R(i j l v) where. L(i j l z) = xyy i l] P(wl+1|x y) yzy l + 1 j] Q(left|y z). R(i j l v) =.

De-marketing: Putting Kotler and Levy's Ideas into ...
De-marketing: KoAer and Levy's Ideas in Practice. 317. wiU be attracted as ... independent contractors, and run their practices as small businesses. They are not.

Putting prey and predator into the CO2 equation ...
laboratory to elevated CO2 and released in their native habitat suffered. 5- to 8-fold ... Pseudochromis fuscus, was allowed to interact in a semi-natural system for ..... supporting information (other than missing files) should be addressed.

Putting the stress tensor (and traction vector) into ...
Revision https://github.com/peeterjoot/physicsplay commit 1fa4a520eb388b1d15e856ab9c00172ebb6a7582. Apr/8/2012 continuumstressTensorVectorForm.tex. Keywords: PHY454H1S, stress ..... [3] L.D. Landau and E.M. Lifshitz. A Course in Theoretical Physics-F

The Movius Line controversy: the state of the ... - Taylor & Francis Online
rities in these stone-tool industries, Movius (1944, 1948) considered all the industries to be representative .... assemblage from Java (now Indonesia). At the time .... the development of Acheulean technology, as noted by Swisher et al. (1994).

Community-based Research Methods - Putting Ideas into Action.pdf ...
Community-based Research Methods - Putting Ideas into Action.pdf. Community-based Research Methods - Putting Ideas into Action.pdf. Open. Extract.