This article may be confusing or unclear to readers. (January 2018) (Learn how and when to remove this template message)
An operational definition is the articulation of operationalization (or statement of procedures) used in defining the terms of a process (or set of validation tests) needed to determine the nature of an item or phenomenon (a variable, term, or object) and its properties such as duration, quantity, extension in space, chemical composition, etc. Since the degree of operationalization can vary itself, it can result in a more or less operational definition. The procedures included in definitions should be repeatable by anyone or at least by peers.
An example of operational definition of the term weight of an object, operationalized to a degree, would be the following: "weight is the numbers that appear when that object is placed on a weighing scale". According to it, the weight can be any of the numbers shown on the scale after and including the very moment the object is put on it. Clearly, the inclusion of the moment when one can start reading the numbers on the scale would make it more fully an operational definition. Nonetheless, it is still in contrast to those purely theoretical definitions.
Properties described in this manner must be sufficiently accessible so that people other than the definer may independently measure or test for them at will. An operational definition is generally designed to model a theoretical definition. The most operational definition is a process for identification of an object by distinguishing it from its background of empirical experience.
The binary version produces either the result that the object exists, or that it doesn't, in the experiential field to which it is applied. The classifier version results in discrimination between what is part of the object and what is not part of it. This is also discussed in terms of semantics, pattern recognition, and operational techniques, such as regression.
Operationalize means to put into operation or use. Operational definitions are also used to define system states in terms of a specific, publicly accessible process of preparation or validation testing, which is repeatable at will. For example, 100 degrees Celsius may be crudely defined by describing the process of heating water at sea level until it is observed to boil. An item like a brick, or even a photograph of a brick, may be defined in terms of how it can be made. Likewise, iron may be defined in terms of the results of testing or measuring it in particular ways.
Vandervert (1980/1988) described in scientific detail a simple, everyday illustration of an operational definition in terms of making a cake (i.e., its recipe is an operational definition used in a specialized laboratory known as the household kitchen). Similarly, the saying, if it walks like a duck and quacks like a duck, it must be some kind of duck, may be regarded as involving a sort of measurement process or set of tests (see duck test).
Despite the controversial philosophical origins of the concept, particularly its close association with logical positivism, operational definitions have undisputed practical applications. This is especially so in the social and medical sciences, where operational definitions of key terms are used to preserve the unambiguous empirical testability of hypothesis and theory. Operational definitions are also important in the physical sciences.
The Stanford Encyclopedia of Philosophy entry on scientific realism, written by Richard Boyd, indicates that the modern concept owes its origin in part to Percy Williams Bridgman, who felt that the expression of scientific concepts was often abstract and unclear. Inspired by Ernst Mach, in 1914 Bridgman attempted to redefine unobservable entities concretely in terms of the physical and mental operations used to measure them. Accordingly, the definition of each unobservable entity was uniquely identified with the instrumentation used to define it. From the beginning objections were raised to this approach, in large part around the inflexibility. As Boyd notes, "In actual, and apparently reliable, scientific practice, changes in the instrumentation associated with theoretical terms are routine, and apparently crucial to the progress of science. According to a 'pure' operationalist conception, these sorts of modifications would not be methodologically acceptable, since each definition must be considered to identify a unique 'object' (or class of objects)." However, this rejection of operationalism as a general project destined ultimately to define all experiential phenomena uniquely did not mean that operational definitions ceased to have any practical use or that they could not be applied in particular cases.
The special theory of relativity can be viewed as the introduction of operational definitions for simultaneity of events and of distance, that is, as providing the operations needed to define these terms.
Operational definitions are often most challenging in the fields of psychology and psychiatry, where intuitive concepts, such as intelligence need to be operationally defined before they become amenable to scientific investigation, for example, through processes such as IQ tests.
On October 15, 1970, the West Gate Bridge in Melbourne, Australia collapsed, killing 35 construction workers. The subsequent enquiry found that the failure arose because engineers had specified the supply of a quantity of flat steel plate. The word flat in this context lacked an operational definition, so there was no test for accepting or rejecting a particular shipment or for controlling quality.
In his managerial and statistical writings, W. Edwards Deming placed great importance on the value of using operational definitions in all agreements in business. As he said:
Operational, in a process context, also can denote a working method or a philosophy that focuses principally on cause and effect relationships (or stimulus/response, behavior, etc.) of specific interest to a particular domain at a particular point in time. As a working method, it does not consider issues related to a domain that are more general, such as the ontological, etc.
Science uses computing. Computing uses science. We have seen the development of Computer Science. There are not many who can bridge all three of these. One effect is that, when results are obtained using a computer, the results can be impossible to replicate if the code is poorly documented, contains errors, or if parts are omitted entirely. 
Many times, issues are related to persistence and clarity of use of variables, functions, and so forth. Also, systems dependence is an issue. In brief, length (as a standard) has matter as its definitional basis. What pray tell can be used when standards are to be computationally framed?
Hence, operational definition can be used within the realm of the interactions of humans with advanced computational systems. In this sense, one area of discourse deals with computational thinking in, and with how it might influence, the sciences. To quote the American Scientist:
One referenced project pulled together fluid experts, including some who were expert in the numeric modeling related to computational fluid dynamics, in a team with computer scientists. Essentially, it turned out that the computer guys did not know enough to weigh in as much as they would have liked. Thus, their role, to their chagrin, many times was "mere" programmer.
Some knowledge-based engineering projects experienced similarly that there is a trade-off between trying to teach programming to a domain expert versus getting a programmer to understand the intricacies of a domain. That, of course, depends upon the domain. In short, any team member has to decide which side of the coin to spend one's time.
The International Society for Technology in Education has a brochure detailing an "operational definition" of computational thinking. At the same time, the ISTE made an attempt at defining related skills.
A recognized skill is tolerance for ambiguity and being able to handle open-ended problems. For instance, a knowledge-based engineering system can enhance its operational aspect and thereby its stability through more involvement by the subject matter expert, thereby opening up issues of limits that are related to being human. As in, many times, computational results have to be taken at face value due to several factors (hence the duck test's necessity arises) that even an expert cannot overcome. The end proof may be the final results (reasonable facsimile by simulation or artifact, working design, etc.) that are not guaranteed to be repeatable, may have been costly to attain (time and money), and so forth.
In advanced modeling, with the requisite computational support such as knowledge-based engineering, mappings must be maintained between a real-world object, its abstracted counterparts as defined by the domain and its experts, and the computer models. Mismatches between domain models and their computational mirrors can raise issues that are apropos to this topic. Techniques that allow the flexible modeling required for many hard problems must resolve issues of identity, type, etc. which then lead to methods, such as duck typing. Many domains, with a numerics focus, use limit theory, of various sorts, to overcome the duck test necessity with varying degrees of success. Yet, with that, issues still remain as representational frameworks bear heavily on what we can know.
In arguing for an object-based methodology, Peter Wegner suggested that "positivist scientific philosophies, such as operationalism in physics and behaviorism in psychology" were powerfully applied in the early part of the 20th century. However, computation has changed the landscape. He notes that we need to distinguish four levels of "irreversible physical and computational abstraction" (Platonic abstraction, computational approximation, functional abstraction, and value computation). Then, we must rely on interactive methods, that have behavior as their focus (see duck test).
The thermodynamic definition of temperature, due to Nicolas Léonard Sadi Carnot, refers to heat "flowing" between "infinite reservoirs". This is all highly abstract and unsuited for the day-to-day world of science and trade. In order to make the idea concrete, temperature is defined in terms of operations with the gas thermometer. However, these are sophisticated and delicate instruments, only adapted to the national standardization laboratory.
For day-to-day use, the International Temperature Scale of 1990 (ITS) is used, defining temperature in terms of characteristics of the several specific sensor types required to cover the full range. One such is the electrical resistance of a thermistor, with specified construction, calibrated against operationally defined fixed points.
Electric current is defined in terms of the force between two infinite parallel conductors, separated by a specified distance. This definition is too abstract for practical measurement, so a device known as a current balance is used to define the ampere operationally.
Unlike temperature and electric current, there is no abstract physical concept of the hardness of a material. It is a slightly vague, subjective idea, somewhat like the idea of intelligence. In fact, it leads to three more specific ideas:
Of these, indentation hardness itself leads to many operational definitions, the most important of which are:
In all these, a process is defined for loading the indenter, measuring the resulting indentation, and calculating a hardness number. Each of these three sequences of measurement operations produces numbers that are consistent with our subjective idea of hardness. The harder the material to our informal perception, the greater the number it will achieve on our respective hardness scales. Furthermore, experimental results obtained using these measurement methods has shown that the hardness number can be used to predict the stress required to permanently deform steel, a characteristic that fits in well with our idea of resistance to permanent deformation. However, there is not always a simple relationship between the various hardness scales. Vickers and Rockwell hardness numbers exhibit qualitatively different behaviour when used to describe some materials and phenomena.
The constellation Virgo is a specific constellation of stars in the sky, hence the process of forming Virgo cannot be an operational definition, since it is historical and not repeatable. Nevertheless, the process whereby we locate Virgo in the sky is repeatable, so in this way, Virgo is operationally defined. In fact, Virgo can have any number of definitions (although we can never prove that we are talking about the same Virgo), and any number may be operational.
New academic disciplines appear in response to interdisciplinary activity at universities. An academic suggested that a subject matter area becomes a discipline there are more than a dozen or university departments using the same name for roughly the same subject matter.
|Theoretical definition||Operational definition|
|Weight: a measurement of gravitational force acting on an object||a result of measurement of an object on a newton spring scale|