CHAPTER 8

Language and Thinking Processes

Chapter outline

Key terms

Bias

Conceptual definition

Context specificity

Control

Dependent variable

Design

Emic

Etic

External validity

Independent variable

Instrumentation

Internal validity

Intervening variable

Manipulation

Operational definition

Rigor

Statistical conclusion validity

Validity

Variable

We now are ready to explore the language and thinking of researchers who use the range of designs relevant to health and human service inquiry. As discussed in previous chapters, significant philosophical differences exist between experimental-type and naturalistic research traditions. Experimental-type designs are characterized by thinking and action processes based in deductive logic and a positivist paradigm in which the researcher seeks to identify

image

a single reality through systematized observation. This reality is understood by reducing it to its parts, observing and measuring the parts, and then examining the relationship among these parts. Ultimately, the purpose of research in the experimental-type tradition is to predict what will occur in one part of the universe by knowing and observing another part.

Unlike experimental-type design, the naturalistic tradition is characterized by multiple ontological and epistemological foundations. However, naturalistic designs share common elements that are reflected in their languages and thinking processes. Researchers in the naturalistic tradition base their thinking in inductive and abductive logic and seek to understand phenomena within the context in which they are embedded. Thus, the notion of multiple realities and the attempt to characterize holistically the complexity of human experience are two elements that pervade naturalistic approaches.

Mixed-method designs draw on and integrate the language and thinking of both traditions.

Now we turn to the philosophical foundations, language, and criteria for scientific rigor for experimental-type and naturalistic traditions. “Rigor” is a term used in research that refers to procedures that enhance and are used to judge the integrity of the research design. As you read this chapter, compare and contrast the two traditions and consider the application and integration of both.

Experimental-type language and thinking processes

Within the experimental-type research tradition, there is consensus about the adequacy and scientific rigor of action processes. Thus, all designs in the experimental-type tradition share a common language and a unified perspective as to what constitutes an adequate design. Although there are many definitions of research design across the range of approaches, all types of experimental-type research share the same fundamental elements and a single, agreed-on meaning (Box 8-1).

BOX 8-1   Characteristics of Experimental-Type Research

image Based in one epistemology

image Historically accepted as the “scientific method” for discovering “fact”

image Evaluated and described by a unified and an agreed-on vocabulary that is well established

In experimental-type research, design is the plan or blueprint that specifies and structures the action processes of collecting, analyzing, and reporting data to answer a research question. As Kerlinger stated in his classic definition, design is “the plan, structure, and strategy of investigation conceived so as to obtain answers to research questions and to control variance.”1 “Plan” refers to the blueprint for action or the specific procedures used to obtain empirical evidence. “Structure” represents a more complex concept and refers to a model of the relationships among the variables of a study. That is, the design is structured in such a way as to enable an examination of a hypothesized relationship among variables. This relationship is articulated in the research question. The main purpose of the design is to structure the study so that the researcher can answer the research question.

In the experimental-type tradition, the purpose of the design is to control variances or restrict or control extraneous influences on the study. By exerting such control, the researcher can state with a degree of statistical assuredness that study outcomes are a consequence of either the manipulation of the independent variable (e.g., true experimental design) or the consequence of that which was observed and analyzed (e.g., nonexperimental design). In other words, the design provides a degree of certainty that an investigator's observations are not haphazard or random but reflect what is considered to be a true and objective reality. The researcher is thus concerned with developing the most optimal design that eliminates or controls what researchers refer to as disturbances, variances, extraneous factors, or situational contaminants. The design controls these disturbances or situational contaminants through the implementation of systematic procedures and data collection efforts, as discussed in subsequent chapters. The purpose of imposing control and restrictions on observations of natural phenomena is to ensure that the relationships specified in the research question(s) can be identified, understood, and ultimately predicted.

The element of design is what separates research from the everyday types of observations and thinking and action processes in which each of us engages. Design instructs the investigator to “do this” or “don't do that.” It provides a mechanism of control to ensure that data are collected objectively, in a uniform and consistent manner, with minimal investigator involvement or bias. The important points to remember are that the investigator remains separate from and uninvolved with the phenomena under study (e.g., to control one important potential source of disturbance or situational contaminant) and that procedures and systematic data collection provide mechanisms to control and eliminate bias.

Sequence of Experimental-Type Research

Design is pivotal in the sequence of thoughts and actions of experimental-type researchers (Figure 8-1). It stems from the thinking processes of formulating a problem statement, a theory-specific research question that emerges from scholarly literature, and hypotheses or expected outcomes. Design dictates the nature of the action processes of data collection, the conditions under which observations will be made, and, most important, the type of data analyses and reporting that will be possible.

image

Figure 8-1 Sequence of experimental-type research.

Do you recall our previous discussion on the essentials of research? In that discussion, we illustrated how a problem statement indicates the purpose of the research and the broad topic the investigator wants to address. In experimental-type inquiry the literature review guides the selection of theoretical principles and concepts of the study and provides the rationale for a research project. This rationale is based on the nature of research previously conducted and the level of theory development for the phenomena under investigation. The experimental-type researcher must develop a literature review in which the specific theory, scope of the study, research questions, concepts to be measured, nature of the relationship among concepts, and measures that will be used in the study are discussed and supported. Thus, the researcher develops a design that builds on both the ideas that have been formulated and the actions that have been conducted in ways that conform to the rules of scientific rigor.

The choice of design not only is shaped by the literature and level of theory development but is also dependent on the specific question asked and resources or the practical constraints, such as access to target populations and monetary and staff considerations. There is nothing inherently good or bad about a design. Every research study design has its particular strengths and weaknesses. The adequacy of a design is based on how well the design answers the research question that is posed. That is the most important criteria for evaluating a design. If it does not answer the research question, then the design, regardless how rigorous it may appear, is not appropriate. It is also important to identify and understand the relative strength and weakness of each design element. Methodological decisions are purposeful and should be made with full recognition of what is gained and what is not by implementing each design element.

Structure of Experimental-Type Research

Experimental-type research has a well-developed language that sets clear rules and expectations for the adequacy of design and research procedures. As you see in Table 8-1, nine key terms structure experimental-type research designs. Let us examine the meaning of each.

TABLE 8-1

Key Terms in Structuring Experimental-Type Research

Term Definition
Concept Symbolically represents observation and experience
Construct Represents a model of relationships among two or more concepts
Conceptual definition Concept expressed in words
Operational definition How the concept will be measured
Variable Operational definition of a concept assigned numerical values
Independent variable Presumed cause of the dependent variable
Intervening variable Phenomenon that has an effect on study variables
Dependent variable Phenomenon that is affected by the independent variable or is the presumed effect or outcome
Hypothesis Testable statement that indicates what the researcher expects to find

Concepts

A concept is defined as the words or ideas that symbolically represent observations and experiences. Concepts are not directly observable; rather, what they describe is observed or experienced. Concepts are “(1) tentative, (2) based on agreement, and (3) useful only to the degree that they capture or isolate something significant and definable.”2 For example, the terms “grooming” and “work” are both concepts that describe specific observable or experienced activities in which people engage on a regular basis. Other concepts, such as “personal hygiene” or “sadness,” have various definitions, each of which can lead to the development of different assessment instruments to measure the same underlying concept.

Constructs

As discussed in Chapter 6, constructs are theoretical creations based on observations but cannot be observed directly or indirectly.3 A construct can only be inferred and may represent a larger category with two or more concepts or constructs.

image The construct of “function” may be inferred from observing concepts such as grooming, personal hygiene, and work or asking individuals to appraise their function in discrete areas. However, the construct of function is not observable unless it is broken down into its component parts or concepts.

Definitions

The two basic types of definition relevant to research design are conceptual definitions and operational definitions. A conceptual definition, or lexical definition, stipulates the meaning of a concept or construct with other concepts or constructs. An operational definition stipulates the meaning by specifying how the concept is observed or experienced. Operational definitions “define things by what they do.”

image The construct of “self-care” may be conceptually defined as the activities that are necessary to care for one's bodily functions, whereas it is operationally defined as observations of an individual engaged in bathing, dressing, and grooming or individual appraisal of his or her level of difficulty in performing self-care activities.

Variables

A variable is a concept or construct to which numerical values are assigned. By definition, a variable must have more than one value even if the investigator is interested in only one condition.

image If an investigator is interested in evaluating the adequacy of self-care routines of persons with disabilities who are employed, both self-care routine and employment are variables. Self-care routine may have multiple values of relevance to the investigator, such as the level of dependence, safety and efficiency of performance, and perceived difficulty. The investigator needs to determine which of the potential values that may be attributed to self-care would be of most interest in the inquiry. That is, there are different ways to measure self-care routines, and each approach can lead to a different understanding. Likewise, employment status can have multiple values (e.g., employed vs. not employed, number of hours working outside the home), and the investigator needs to determine how to operationalize this variable to best match the study's purpose.

There are three basic types of variables: independent, intervening, and dependent. An independent variable “is the presumed cause of the dependent variable, the presumed effect.”4 Thus, a dependent variable (also referred to as “outcome” and “criterion”) refers to the phenomenon that the investigator seeks to understand, explain, or predict. The independent variable almost always precedes the dependent variable and may have a potential influence on it. The dependent variable is also referred to as the “predictor variable.” An intervening variable (also called a “confounding” or an “extraneous” variable) is a phenomenon that has an effect on the study variables but that may or may not be the object of the study.

image In the study of adequacy of self-care routines of employed individuals with disabilities, employment status represents the independent variable, whereas level of difficulty performing self-care routine represents the dependent variable. Intervening variables in this study may include any other variable that potentially influences either the independent or the dependent variable, such as family support, type and degree of disability, motivational status, and functional ability.

Investigators treat intervening variables differently depending on the research question. For example, an investigator may only be interested in examining the relationship between the independent and dependent variables and may thus statistically control or account for an intervening variable. The investigator would then examine the relationship after statistically “removing” the effect of one or more potential intervening variables. However, the investigator may want to examine the effect of an intervening variable on the relationship between the independent and dependent variables. For this question, the researcher would employ statistical techniques to determine the interrelationships.

Hypotheses

A hypothesis is defined as a testable statement that indicates what the researcher expects to find, given the theory and level of knowledge in the literature. A hypothesis is stated in such a way that it will either be verified or falsified by the research process. The researcher can develop either a directional or a nondirectional hypothesis (see Chapter 7). In a directional hypothesis, the researcher indicates whether she or he expects to find a positive relationship or an inverse relationship between two or more variables. A positive relationship is one in which both variables increase and decrease together to a greater or lesser degree.

image In a positive relationship, the researcher may hypothesize that the employed individual with a strong disability identity has greater self-esteem than the unemployed individual with a strong disability identity. Thus the hypothesis tests the positive movement of both variables.

In an inverse relationship, the variables are associated in opposite directions (i.e., as one increases, the other decreases). An inverse relationship may involve a hypothesis similar to the following:

image An individual with a disability who remains employed will demonstrate less difficulty in performing self-care routines than an individual who is unemployed.

In this statement, the expectation is that as the variable “employment” increases, difficulty in self-care will decrease.

Now that we have defined nine key terms that are essential to experimental-type designs, let us review the way they are actually used in research. Experimental-type research questions narrow the scope of the inquiry to specific concepts, constructs, or both. These concepts are then defined through the literature and are operationalized into variables that will be investigated descriptively, relationally, or predictively. The hypothesis establishes an equation or structure by which independent and dependent variables are examined and tested.

Plan of Design

The plan of an experimental-type design requires a set of thinking processes in which the researcher considers five core issues: bias, manipulation, control, validity, and reliability.

Bias

Bias is defined as the potential unintended or unavoidable effect on study outcomes. When bias is present and unaccounted for, the investigator may not be able to fully understand whether the study findings are accurate or reflect sources of bias and thus may misinterpret the results. Many factors can cause bias in a study (Box 8-2). We already discussed one source of bias, the “intervening variable.”

BOX 8-2   Possible Sources of Bias in a Study

image Selection of inappropriate instrumentation

image Sample selection process that favors a particular unintended group

image Improper training of interviewers

image Deviation from the plan and from structure of the design

Another source is instrumentation. This involves the ways in which data are obtained in experimental-type approaches. The two major sources of bias from instrumentation are inappropriate data collection procedures and inadequate questions.

image Assume you are a health and human service professional working in a community health center. Your supervisor, the agency administrator, is conducting a survey of worker satisfaction. To obtain information from employees, the supervisor decides to interview people. Knowing that this person is your superior, how likely are you to provide responses to job satisfaction questions, particularly knowing that your performance evaluation is next month?

In the situation portrayed above, the procedures for data collection are problematic and introduce bias into the study.

image Using this same example, now consider how questions are asked to obtain the data. Suppose this question is asked: “Don't you agree that this is a great place to work?” Certainly this question is poorly phrased; it implies a socially and politically correct response and in essence urges the responder to agree. Now consider this question: “Don't you think that the physical environment and the climate are excellent here?” In this case, the question is confusing and ambiguous. What does “the climate” refer to? Does it signify the weather? The emotional climate? The social climate? In this case, the responder is prompted or guided to produce a socially desirable response. It is also not clear what is being asked.

Interview questions that elicit a socially correct response and questions that are vague, unclear, or ambiguous introduce a source of bias into the study design.

Sampling is another major source of bias.

image Suppose that only the supervisor's social friends were selected as the targeted sample to participate in the survey and are viewed as representing the agency population. Such a sample would not represent the overall level of job satisfaction for all employees. In this example, the characteristics of the sample serve as potential bias or influence the outcomes of the study.

image Improper or uneven training of data collectors and interviewer “drift,” in which an interviewer strays from the interview protocol from fatigue or overfamiliarity, can bias study outcomes as well.

image In the same study of employee satisfaction, if multiple persons collected information about job satisfaction but each asked the survey questions differently or in a nonuniform manner, interviewer procedures will serve as a strong bias.

Experimental-type designs are characterized by having detailed and elaborate plans and procedures that are established before conducting the study. Thus, any deviation from the original plan for data collection and analysis may be a potential source of bias. The violation of standard procedures is an important point at which bias may be introduced into an experimental-type study.

image Suppose that the agency director decided to collect data regarding the level of job satisfaction of agency employees on Monday. Monday was selected because all employees were normally available. However, the data collection was accomplished on Tuesday because Monday was a holiday. On Tuesday, all home-based service providers were in the field and therefore were not surveyed for their degree of job satisfaction. This deviation in the plan could have significantly affected the study outcome, because a large segment of the agency employee population would have been omitted from the study.

An essential part of planning a design is to introduce systematic procedures to minimize or eliminate as many sources of bias as possible. As you read subsequent chapters, think about the potential sources of bias. Also, as you conduct your own studies or read and analyze published studies, consider findings and conclusions in light of potential bias.

image Let's say you participated in a program designed to help you more effectively manage your diabetes. The facilitator of the program is conducting an evaluation of the program's effectiveness. The facilitator asks each program participant to respond to a brief questionnaire before beginning the program and then upon its completion. Although you liked the facilitator a lot, you do not think the program was informative, nor did it change your ability to manage the disease. How likely are you to express your true feelings about the program if the facilitator is asking you the postevaluation questions? How can you design the pre- and post-data collection differently to avoid imposing bias?

Manipulation

Manipulation is defined as the action process of maneuvering the independent variable so that the effect of its presence, absence, or degree on the dependent variable can be observed. To the extent possible, experimental-type researchers attempt to isolate the independent and dependent variables and then manipulate or change the condition of the independent variable so that the cause-and-effect relationship between the variables can be examined.

image Research studies that examine the effects of different types of interventions on the well-being of caregivers of persons with Alzheimer's disease illustrate the concept of manipulation. To determine which types of interventions (e.g., skills training, social support, group counseling, education) promote positive outcomes, the most effective research plan would be to use a randomized, two-group design. In this type of design, caregivers would be evaluated with regard to well-being at baseline or before introducing an intervention. Study participants would then be randomly assigned (assigned on a chance-determined basis; see Chapter 11) to different intervention conditions and tested again after completing the interventions. Then caregivers would be asked to switch groups and participate in the intervention group that they had not experienced. They would then be tested again to evaluate change in the dependent variable, “well-being.” The interventions represent the independent variables being manipulated (received, not received). By manipulating the intervention timing and sequence, an investigator is able to examine the effect of each intervention and combinations on caregiver well-being.5

Control

Control is defined as the set of action processes that direct or manipulate factors to achieve an outcome. Control plays a critical role in experimental-type design. By controlling not only the independent variable but also other aspects of the research context, such as how study participants are assigned to groups, the relationship among the study variables can be observed. In experimental-type design, procedures to establish control are implemented to minimize the influences of extraneous variables on the outcome or dependent variable.

Two basic methods of control are typically introduced: random group assignment and control group. By randomly assigning subjects to one group or the other, the investigator attempts to develop equivalence, or eliminate subject bias, caused by inherent differences that may occur in the two groups.

image In evaluating the outcome of interventions on caregivers’ well-being, assume you do not use random assignment. Instead, you assign the first 10 volunteers to the first group, who receive the intervention involving group counseling, and the next 10 volunteers to the second group, who do not receive counseling. It is possible that the first 10 subjects know each other and agreed to participate in the study because they were highly motivated to seek help, and each person encouraged the other to join. In contrast, suppose the last 10 subjects were somewhat less motivated to participate. Familiarity with other subjects and a high motivation to seek help (two factors inherent in the subjects assigned to the first group) may influence caregiver responses to the outcome measure of “well-being.” By not using random assignment, the investigator may risk developing groups that initially differ from each other. In turn, this initial difference may make it difficult to discern what effect, if any, the independent variable has on the outcome variable.

Another method to enhance control is the use of a control group. A control group is one in which the experimental or comparative condition is absent.

image In caregiver intervention research, the experimental treatment is typically compared with a group that does not receive any intervention, sometimes referred to as a “usual care” group—that is, a group that receives only what they usually would, independent of the study. Occasionally, the experimental treatment group is compared with an “attention” control group—that is, a group that receives some form of attention to control for its occurrence in the treatment group. By comparing caregiver scores in an attention control group to those in an intervention group, the investigator is able to determine whether changes in the dependent variable resulted from (1) the passage of time, (2) the attention factor inherent in a behavioral intervention (or intervention involving interaction with a health or human service professional), or (3) the particular content of each group.

Validity

Validity is a concept that has numerous applications in experimental-type research. However, the concept as it applies to design refers to the extent to which your study answers the research questions and your findings are accurate or reflect the underlying purpose of the study. Although many classifications of validity have been developed, we discuss four fundamental types of validity, based on Campbell and Stanley's classic work.6 These are internal validity, external validity, statistical conclusion validity, and construct validity.

Internal Validity. The ability of the research design to answer the research question accurately is known as internal validity. If a design has internal validity, the investigator can state with a degree of confidence that the reported outcomes are the consequence of the relationship between the independent variable and dependent variable and not the result of extraneous factors. Campbell and Stanley6 identified seven major factors that pose a threat to the researcher's ability to determine whether the observable outcome is a function of the study or the result of external and unintended forces (Box 8-3).

BOX 8-3   Seven Threats to Internal Validity

1. History. Effect of external events on study outcomes

2. Testing. Effect of being observed or tested on the study outcome

3. Instrumentation. Extent to which the instrument is accurate in its measurement and extent to which the instrument itself may be responsible for outcomes

4. Maturation. Effect of the passage of time

5. Regression. Effect of a statistical phenomenon in which extreme scores tend to regress or cluster around the mean (average) on repeated testing occasions

6. Mortality. Effect on outcome caused by subject attrition or dropping out of a study before its completion

7. Interactive effects. Extent to which each of these threats interacts with sample selection to influence the outcome of a study

From Campbell DT, Stanley JC: Experimental and quasi-experimental design, Chicago, 1963, Rand McNally.

To illustrate these threats to validity, consider a hypothetical example.

image Suppose you are conducting a study to determine the extent to which an acquired immunodeficiency syndrome (AIDS) prevention program has reduced behavior that increases the risk of developing AIDS in an adolescent population. This study is designed to answer the following research question: “To what extent is adolescent risk behavior reduced by the experimental preventive intervention?” Before the intervention, you evaluate the extent to which subjects exhibit risk behavior. You conduct a prevention program and retest the participants. The retest shows an amazing reduction in risk behavior, and you claim that your program is a success. In doing so, you are making a claim that a relationship exists between your program (the independent variable) and risk behavior (the dependent variable). Furthermore, you are claiming that there is a causal relationship between the variables—that is, you claim that your program causes the reduction in behavior that increases the risk of developing AIDS.

Consider how each of the seven threats to internal validity affects your research outcomes.

image First, let us assume you have just learned that a celebrity has held a news conference in which he announced that he has tested positive for the human immunodeficiency virus (HIV). This announcement has come immediately after you first administered your test to your subjects but before the final testing occasion. It is possible that this historical event, rather than the independent variable, was the factor responsible for reducing risk behavior in your subjects.

However, assume that your subjects already knew about the celebrity before beginning the study. Now consider the threat of testing to the validity of your claims. As a sensitive researcher, you have decided to measure risk behavior by observing a discussion group about sexual behaviors among adolescents. Each time you hear a risk behavior, you score it as such without informing your subjects. In the first testing situation, the adolescents appear to discuss their sexual decision making openly. However, you note that they are aware of group norms and of being observed, and that they stay well within the boundaries tacitly set by the group when discussing their personal behavior and views. In the second test, you note a significant reduction in reported risk behaviors or in risky decision making.

Although the prevention program may be responsible for the reduction in risk behavior, there are alternative explanations. In the first testing situation, the adolescents may have revealed what was normative and acceptable in the group rather than accurately representing their risk potential. In the second test, the adolescents who underwent the intervention may have restricted their conversation to what they learned was desirable rather than revealing their own behavior.

Another alternative for the observed reduction in risk behavior may be the result of the testing situation itself. Participants who are aware of what is being observed and recorded may answer more cautiously. In other words, the testing procedures may pose yet another threat to the internal validity of the design. The reduction in reported risk behaviors may be a function of participants actively changing or adjusting their behaviors or thoughts on the topic as a consequence of participating in a group discussion with peers. The test itself or mode of data collection may have influenced a change in behavior independent of the effect of participating in the intervention. Therefore, it is difficult to determine whether observed change is a consequence of the test, of the intervention, or both.

This example also illustrates the threat to validity posed by instrumentation. In this example, the instrumentation may not be measuring what was intended. Instead of measuring risk behavior, the instrumentation may be a more accurate indicator of group norms related to sexual decision making.

Another way in which instrumentation poses a threat to a study is through changes that may occur within interviewers over time or with the instrument itself. For example, if you are collecting data regarding weight or blood pressure, any deviation in the calibration of a scale or blood pressure cuff from one testing occasion to the next will pose a significant threat to the validity of the data that are obtained.

Further, interviewer fatigue or any change in the way interviewers pose questions may cause a deviation in responses that will affect how the investigator interprets the findings.

image To illustrate the threat posed by maturation, assume that the AIDS prevention program is being conducted over a 1-year period. It is possible that the participants have matured in their thinking and have become more responsible in their decision making as a function of time rather than as a result of the prevention program.

The threat of regression frequently occurs when subjects with extreme scores are selected to participate in a study.

image Assume that adolescents who are sexually active participate in the study and therefore tend to report extremely risky behaviors. Those who are high scorers at the beginning of the program tend to report lower scores at the end.

This change in scores, however, may be a consequence of a statistical principle known as “statistical regression toward the mean,” in which extreme scores tend to move toward the mean on repeated testing.

image Fifty participants began the AIDS prevention program, but only 10 remained by the final testing period. It is possible that as a result of “experimental mortality” (attrition of participants before a study's conclusion), those who remained in the study may be more committed to reducing risk behavior than those who dropped out.

Thus, experimental mortality poses a threat to the interpretation of study findings. Numerous interactive effects could also confound the accuracy of the findings.

image Consider that the selected sample for the study was a population of adolescents on probation. This group was highly motivated to report a change in their risk behavior (a sampling bias) and was also sensitive to being tested because of their judiciary status. The interactive effect of the sample bias and the testing condition may wreak havoc on claims that the outcome was a result of the prevention program.

External Validity. External validity refers to the capacity to generalize findings and develop inferences from the sample to the study population stipulated in the research question. External validity answers the question of “generalizability.”

image Remember our research question: “To what extent is adolescent risk behavior reduced by the experimental preventive intervention?” In this question, the population is all adolescents. Now, assume that the investigator obtained a sample by asking for volunteers from a large group of adolescents between ages 12 and 18 years who were on probation in Philadelphia. The potential to generalize the findings from the sample to all adolescents on probation is limited by several important factors. First, the sample was a volunteer sample, making them potentially different from adolescents who did not volunteer. Second, the experiences of adolescents who live in large cities may not necessarily represent the experiences of adolescents in other geographic locations. Third, because the sample was voluntary, the researcher cannot know in what ways the sample may differ from the study population, which includes those who refused participation or did not have the opportunity to volunteer. Fourth, in selecting a volunteer urban sample on probation, the investigator limits generalizability to that select sample and will not be able to infer the experiences of the research sample to all adolescents, such as those who are not on probation, who live in rural communities, or who may be sexually active and at risk.

These factors represent some of the threats to external validity that may occur in a study. Although randomization decreases some of the threats to external validity, even when a sample is randomly selected, threats remain.7 The two threats in Box 8-4 result from the experimental condition, not from the limitations of sample selection. Consider your own behavior when being watched in a simulated situation. Certainly you act differently than you would under similar but unobserved and nonsimulated conditions.

BOX 8-4   Potential Threats to External Validity

1. Reactivity. Extent to which the subjects are responding to the condition of being part of a study and thus do not represent the population from whom they are selected

2. Realism. Extent to which the experimental conditions simulate actual life situations to which the population is exposed

The internal validity and external validity of a study are interrelated. As an investigator attempts to increase the internal validity of a study, the ability to make broad generalizations and inferences to a larger population decreases. That is, as more controls are implemented and extraneous factors are eliminated from a study design, the population to whom findings can be generalized becomes more limited. Thus, although investigators try to enhance the internal validity of a study to ensure that valid and accurate conclusions can be drawn, these efforts may limit the degree of external validity of the findings. It is a balancing act, but the investigator must err on the side of ensuring that the study has internal validity first, then external validity. If a study lacks internal validity, there would be no purpose in generalizing to a larger population. As you read further in this text, look for ways to balance internal validity and the scope of generalization that may be possible.

Statistical Conclusion Validity. Statistical conclusion validity refers to the power of your study to draw statistical conclusions. One aim of experimental-type research is to find relationships among variables and ultimately to predict the nature and direction of these relationships. Support for relationships is contained in the action process of statistical analysis. However, selection of statistical techniques is based on many considerations. Even though the expectation of making errors in determining relationships is built into the theory and practice of statistical analysis, the accuracy and potential of the statistics to support or predict a relationship among variables must be considered as a potential threat to the validity of a design.

image Assume you want to ensure that you do not overestimate the effects of your AIDS prevention program on risk behavior. You select a statistic that will be less likely to find a false relationship when there is none. However, this decision increases the possibility that you may not find a significant relationship when, in fact, one exists. Thus, your statistical testing may lead you to believe that no relationship exists when there is one or, conversely, to believe that there is a relationship when there is none.

Construct Validity. Construct validity addresses the fit between the constructs that are the focus of the study and the way in which these constructs are operationalized. As stated earlier, constructs are abstract representations of what humans observe and experience. The method used to define and measure constructs accurately is in large part a matter of opinion and consensus. It is possible for several factors related to construct validity to confound a study. First, a researcher may define a construct inappropriately.

image Although smiling may be one indication of happiness, if a researcher measured level of happiness exclusively by observing the frequency of smiling behavior, the full construct of “happiness” would not be captured. Also, smiling has different meanings in different contexts and cultures. By merely observing frequency, one might not really know the meaning of what was measured.

Poor or incomplete operational definitions can result from incomplete conceptual definitions or from inadequate translation of the construct into an observable one.

Second, when a cause-and-effect relationship between two constructs is determined, it may be difficult to define each construct exclusively (referring to both independent and dependent variables) and to isolate the effects of one on the other.

image You are a health and human service professional seeking to test a feeding program for persons with moderate to late-stage Alzheimer's disease. You design a program in which you choose certain foods and techniques to improve independent feeding. To ensure that your design can reveal cause and effect, you randomly select your sample and randomly assign them to the experimental group (which receives the program) and the control group (which does not receive the program). You measure feeding behavior by two methods: observation of independent feeding and volume of food consumed independently. You implement your feeding program with the experimental group and test all participants after the program to determine whether the experimental group scored higher on your measures of feeding than the group that did not receive the intervention. You conclude that your program was successful in improving independent feeding because the experimental group performed significantly better than the group without the intervention.

However, you begin to consider other confounding influences that cannot be separated from your two constructs—independent feeding and the experimental program. The attention of the experimenter may have influenced the improvement in the experimental group because eating is a social behavior and may change under different social conditions. Also, food preferences may have influenced the response of the groups to their respective conditions. Interactive effects of being observed and participating in an experiment (referred to as the “attention factor” or “Hawthorne effect”8) may also limit your capacity to isolate a causal relationship between the constructs.

Reliability

Reliability refers to the stability of a research design. In experimental-type design, rigorous research is shown to be well planned and properly executed when, if repeated under the same circumstances, the design would yield the same results. Investigators frequently replicate or repeat a study in the same population or in different groups to determine the extent to which the findings of one study are accurate in a broader scope. To replicate a study, the procedures, measures, and data analysis techniques must be consistent, well articulated, and appropriate to the research question. Reliability is threatened when a researcher is not consistent and changes a design in midstream, does not articulate procedures in sufficient detail for replication, or does not fully plan a sound design.

Experimental-Type Design Summary

Design is a pivotal concept in the experimental-type research process that indicates both the structure and the plan of the action processes. Experimental-type designs are developed to eliminate bias and the intrusion of unwanted factors that could confound findings and make them less credible. Five primary considerations in the plan of a design are bias, control, manipulation, validity, and reliability. Four types of validity threaten study outcomes: internal, external, statistical conclusion, and construct.

Experimental-type research is designed to minimize the threats posed by extraneous factors and bias by maximizing control over the research action process. In Chapter 9, we discuss how each type of design in the experimental-type tradition addresses these considerations in structuring and planning a design and using action processes to increase control.

Naturalistic language and thinking processes

Let us now turn our attention to the attributes of naturalistic inquiry and the implications for design with these traditions. Eight elements are common to designs in the naturalistic tradition, as listed in Box 8-5.

BOX 8-5   Eight Elements of Naturalistic Design

1. Purpose of research

2. Context of research

3. Pluralistic perspective of reality

4. Concern with transferability

5. Flexibility

6. Concern with language

7. Emic and etic perspectives

8. Interactive and analytical process

Purpose

Designs within the tradition of naturalistic research vary in purpose from developing descriptive knowledge to evolving full-fledged theories about observed or experienced phenomena. As discussed in earlier chapters, naturalistic designs tend to be preferred when no adequate theory exists to explain a human phenomenon or when the investigator believes that existing theory and explanations are not accurate, true, or complete. The structure of naturalistic designs is exploratory, enabling new insights and understandings to be revealed without the imposition of preconceived concepts, constructs, and principles.

Although the specific purpose of a study may differ, all naturalistic designs seek to describe, understand, or interpret daily life experiences and structures within the contexts in which they occur. Naturalistic inquiry has emerged through five phases of development to its current form of action-oriented social criticism, in which grand narratives are replaced by more local, small-scale theories fitted to specific problems and specific situations.9 Thus, the importance of naturalistic inquiry is now widely recognized because of its usefulness in examining and revealing phenomena that can help guide health and human service practitioners to address specific problems that emerge in the clinical context or explain why or how particular programs or interventions benefit individuals.

image In Tally's Corner,10 Liebow was interested in understanding the nature of a street-corner culture to which he does not belong. Although there had been extensive research on the population he chose to examine, these investigations had not characterized the group from the perspective of its members. Naturalistic method—in this case, ethnography—was a viable way to reveal the complexity and richness of the underlying culture of this group. Liebow gained valuable insights because his informants spoke for themselves and described their life experiences.

Context Specificity

As in Tally's Corner, to discover “new truth” rather than impose an existing frame of reference, the investigator proceeding from a naturalistic tradition must travel to the setting where the human phenomena occur or must seek information from the perspective of individuals who experience the phenomena of interest. Because of the revelatory nature of naturalistic research, this form of investigatory action is conducted in a natural context or seeks explanations of the natural context from those who experience it. Naturalistic research is therefore context-specific, and the “knowing” derived from this context is embedded in the context and does not extend beyond it. The designs in the tradition of naturalistic inquiry share this basic attribute of context specificity and development of knowledge that is grounded in or linked to the data that emerge from a particular field.

Complexity and Pluralistic Perspective of Reality

As a result of its underlying epistemology and its inductive and abductive approaches to knowing, naturalistic research assumes a pluralistic perspective of reality that characterizes complexity.

With inductive reasoning, principles emerge from seemingly unrelated information. One of its hallmarks is that the information can be organized differently by each individual who thinks about it. The end result of induction is the development of a complex set of relationships among smaller pieces of information (not the reduction of principles to their parts, as in deductive reasoning). It is therefore possible that the same information may have different meanings or that there may be pluralistic interpretations by different individuals. Adding abduction to this equation ensures that the information will be revisited and reanalyzed at several points, thereby sharpening and rendering the accuracy of the interpretation more credible and trustworthy.

image In the classic study by Maruyama11 on prison violence, the nature of violence was investigated in two separate institutional settings. The findings from each prison, although yielding some similarities, indicated cultural differences between the two prisons in the reported causes of violence and the inmates’ interpretation of the meaning of that violence. The naturalistic design was able to reveal the pluralistic nature of prison violence while shedding light on the basic elements common to both settings.

Transferability of Findings

The findings from naturalistic design are specific to the research context. It is not the desired outcome of naturalistic design to generalize findings from a small sample and apply this generalization to a larger group of persons with similar characteristics. You may ask, “Why bother doing a naturalistic study if you cannot use the research beyond the actual scope of the study?” The answer lies in one of the primary purposes of naturalistic research: generating theory. Naturalistic researchers use their methods and findings to generate theory and to reveal the unique meanings of human experience in human environments. Because the investigator assumes that current knowledge does not adequately explain the phenomenon under investigation, the outcome of naturalistic design is the emergence of explanations, principles, concepts, and theories. Liebow illustrated this point well when he stated:

There is no attempt here to describe any Negro men other than those with whom I was in direct immediate association. To what extent this descriptive and interpretive material is applicable to Negro street corner men elsewhere in the city or in other cities or to lower class men generally in this society or any other society is a matter of further and later study. This is not to suggest that we are dealing with unique or even distinctive persons and relationships. Indeed, the weight of the evidence is the other direction.10

In this passage, even though the researcher highlighted the contextual limitations of the knowledge, he also indicated that the principles revealed in his study could have relevance to other arenas. Further, he suggested that additional research is the appropriate means to ascertain the degree of fit, or what Guba12 called the “transferability of findings,” to other similar populations and contexts.

Because phenomena are context bound and, according to naturalistic principles, cannot be understood apart from the context in which they occur, naturalistic inquiry is not concerned with the issue of generalizability or external validity as articulated in experimental-type inquiry. Rather, naturalistic inquiry is concerned with understanding the richness and depth in one context and with the capacity to retain unique meanings that are lost when generalization is a goal.9 The development of “thick”13 and in-depth descriptions and interpretations of different contexts leads to the ability to transfer meanings in one context to another context. In so doing, the researcher is able to compare and contrast contexts and their elements to gain new insights as to the specific context itself.

Flexibility

The experimental-type research design provides the basic structure and plan for the thinking and action processes of the research endeavor. The design must be followed as initially developed. In naturalistic research, however, the opposite is the case; design is a fluid, flexible concept. Design labels such as “ethnography,” “life history,” and “grounded theory” suggest a particular purpose of the inquiry and orientation of the investigator. However, these labels do not refer to the specification of step-by-step procedures, a blueprint for action, or a predetermined structure to the data gathering and analytical efforts.

Rather, a characteristic of naturalistic design is flexibility. An important and expected feature of a naturalistic study is that the procedures and plans for conducting it will change as the research proceeds. The investigator may use the results of preliminary findings as guidance for planning or altering subsequent action processes. Not only do procedures change, but so do the nature of the research query, the scope of the study, and the manner by which information is obtained are constantly reformulated and realigned to fit the emerging truths as they are discovered and obtained in the inquiry. This flexibility is illustrated more fully in Chapters 17 and 20.

Language

A major shared concern in naturalistic inquiry is understanding language and meanings. This concern with language is not only relevant to the ethnographer, who studies a cultural context in which the language used is different from that of the investigator. Even within the same cultural or language context, people use and understand language differently. For some naturalistic investigators, language, symbols, and ways of expression provide the data through which the investigator comes to understand and derive meaning within each context.14 Investigators proceeding from a postmodern philosophical foundation are particularly concerned with language itself. In contrast to previous philosophies that attribute meaning to symbols, postmodern thinkers suggest that language as a set of symbols does not necessarily have any of the underlying reality that the symbols describe.15

Depending on the philosophical approach, the investigator concerned with language may engage in a rigorous and active analytical process to “translate” the meaning and structure of the context of the studied population into meanings and language structures represented in the investigator's world. In this type of inquiry, the investigator must be careful to represent the meanings and intent of expression accurately in this translation of the analytical and reporting process. As mentioned, however, not all investigators concerned with language believe that naturalistic inquiry can reveal meaning. Postmodern inquiry aims to “deconstruct” language, or unravel it to illustrate its arbitrary structure and identify its political and purposive usage. For example, a postmodern investigator might examine the text of legislation designed to protect women against domestic violence to uncover the metaphors that help retain men's dominance and political power over women.

Throughout the full range of philosophical approaches that underpin the naturalistic tradition, the description and analysis of language is a primary concern.

Emic and Etic Perspectives

Design structures vary as to their emic or etic orientation. An emic perspective refers to the insider's or informant's way of understanding and interpreting experience. This perspective is phenomenological in that experience is understood as only that which is perceived and expressed by informants. Data gathering and analytical actions are designed to enable the investigator to bring forth and report the voices of individuals as they speak and interpret their unique perceptions of their reality. The concern with the emic perspective is often the motivator for naturalistic inquiry. For example, in her classic work,16 Krefting explained why she decided to study head injury from an ethnographic perspective, highlighting the preference for an emic viewpoint shared by designs in this tradition, as follows:

Until recently, the investigation of health problems was dominated by the outsider perspective, in which important questions of etiology and treatment are identified by the professions. Studies based on this perspective assume that professionals are the authorities on what wellness is and that they alone know what questions ought be asked to investigate methods to promote and maintain wellness while preventing and treating illness… . Ethnographic studies of the experience of illness and disability … consider what these experiences look like from an insider's perspective.16

An etic orientation reflects the structural aspects or those that are external to a group; that is, the etic perspective is held by those who do not belong to the group being investigated but to those who select an analytical and epistemic lens through which to examine information. Unlike the emic perspective, in which those who experience are considered to be most knowledgeable, the etic perspective assumes that those who do not experience a phenomenon can come to know it through (1) structuring an investigation, (2) selecting a theoretical foundation that expands beyond the group being examined, and (3) through that lens, conducting the interpretation of data.17 Many investigators integrate an emic with an etic perspective. Typically, investigators will start by using an emic perspective, or a focus on the voices of individuals. Further along the process, other pieces of information are collected and analyzed to place individual articulation and expression within a social structural or systemic framework. Although she started with an emic consideration, Krefting used a range of data collection strategies to integrate her understandings with an etic viewpoint, as follows:

While gathering data, I was able to review theory and check it for pertinence. There was, then, a constant movement back and forth between the concrete data and the social science concepts that helped explain them.16

Some designs, such as phenomenology and life history, favor only an emic orientation. Frank described her preference for an emic perspective and the purpose of her life history approach as follows:

The life history of Diane DeVries represents a collaborative effort, between the subject and researcher, to produce a holistic, qualitative account that would bear on theoretical issues, but that primarily and essentially would convey a sense of the personal experience of severe congenital disability. The life history, conceived in this way, emerged from a humanistic interest in presenting the voices of people often unheard, yet whose lives were otherwise studied, and acted upon, based on data that are decontextualized and fragmented from the standpoint of the individual.18

Because the investigator enters the inquiry having bracketed, suspended, or let go of any preconceived concepts, the naturalistic researcher defers to the informant or experiencer as the “knower.” This abrogation of power by the investigator to the investigated is characteristic of naturalistic designs, to a greater or lesser degree.

image An extreme example of relinquishing control over the research process is illustrated by Maruyama's classic work.11 In the prison study, not only did the inmates function as the informants, they also planned and conducted the research in its entirety.

Where the investigator stands regarding an emic or etic perspective shapes the overall design that is chosen, as well as the specific data collection and analytical action processes that emerge within the context of the inquiry.

Gathering Information and Analysis

Analysis in naturalistic designs relies heavily on qualitative data and is an ongoing process throughout data-gathering activities. Thus, in the naturalistic tradition, data gathering and analysis are interdependent processes. In collecting information or data, the investigator engages in an active analytical process. In turn, the ongoing analytical activity frames the scope and direction of further data collection efforts. This interactive, iterative, and dynamic process is characteristic of designs within the tradition of naturalistic inquiry and is explored more fully later in this text.

Naturalistic Design Summary

The purpose of naturalistic inquiry and the nature of the concept of design in this tradition are vastly different from experimental-type research. The language and thinking processes that characterize naturalistic design are based on the notion that knowing is pluralistic and that knowledge derives from understanding multiple experiences. Thus, the thinking process is inductive and abductive; the action processes are dynamic and changing; and these processes are carried out within the actual or theoretical context in which the phenomena of interest occur or are experienced. The outcome of these thinking and action processes is the generation of theory, principles, or concepts that explain human experience in human environments and capture its complexity and uniqueness.

Within any naturalistic design, the investigator can implement the 10 essentials in various ways. Designs vary from informant-driven and no structure to researcher-driven and more structure.

The nature of qualitative inquiry has evolved in its forms, scope, and purposes.9 Naturalistic inquiry is continuing to evolve in its standards, vocabulary, and criteria for design adequacy, and it is becoming a well-respected tradition in its own right as it is increasingly used to answer questions about health and human service practice.

Integrated and mixed-method approaches

As discussed in previous chapters of this book, studies which integrate experimental-type and naturalistic traditions have increased despite skepticism about philosophical inconsistencies. The mixed-method approach for advancing knowledge has recently been referred to as the “third tradition”19 with a foundation in pragmatism.

Mixing methods is a systematic way to address the limitations of each approach, and we therefore support its use whenever possible. As discussed, however, both experimental-type and naturalistic traditions conform to diverse sets of rules, languages, and strategies for rigor. As such, the investigator who proceeds with an integrated approach must heed the important tenets of both traditions as well as the emerging tenets of a third tradition. Because of its nascence and limited epistemological underpinning, we do not yet view a mixed-method approach as a solid tradition in itself. Rather, at this point in its development, we view it as an attempt to integrate experimental-type and naturalistic approaches within a defined and purposive context and to incorporate language and thinking processes from each of the two traditions.

Summary

Experimental-type and naturalistic traditions have distinct language and thinking processes. Experimental-type researchers approach their studies with intact theory and a set of procedures. Procedures are developed and followed to eliminate the potential for factors other than those studied to be responsible for the findings. Dissimilarly, naturalistic language and thinking processes are flexible, fluid, and changing. As the investigator proceeds throughout the research design, the aim of capturing complexity from the perspective of the individuals or phenomena studied is actualized through systematic but dynamic interaction between collecting and analyzing information.

Exercises

1. Select from a journal a research article that uses an experimental-type design. Identify the independent variable or variables, the dependent variable or variables, and any intervening variables that may be confounding the study.

2. Identify the conceptual definitions and the operational definitions in the article that you selected.

3. Using the same article, determine (a) threats to validity of the study, (b) how the investigator minimized bias and enhanced control and validity, and (c) ethical issues that shaped the design.

4. Select a research article that uses naturalistic inquiry. Identify and provide evidence of the eight elements of the researcher's thinking process:

a. Purpose of research

b. Context of research

c. Pluralistic perspective of reality

d. Concern with transferability

e. Flexibility

f. Concern with language

g. Emic and etic perspectives

h. Interactive and analytical process

5. Using the same article, determine some of the ethical issues that shaped the investigator's behavior in the inquiry.

References

1. Kerlinger, F.N. Foundations of behavioral research, ed 2, New York: Holt, Rinehart, & Winston; 1973:279.

2. Wilson, J. Thinking with concepts. Cambridge, Mass: Cambridge University Press, 1966;54.

3. Babbie, E. The practice of social research, ed 12. Belmont, Calif: Wadsworth, 2009.

4. Kerlinger, F.N. Foundations of behavioral research, ed 2, New York: Holt, Rinehart, & Winston; 1973:39.

5. Gitlin, L.N., Winter, L., Corcoran, M., et al. Effects of the home environmental skill-building program on the caregiver-care recipient dyad: six-month outcomes from the Philadelphia REACH Initiative. The Gerontologist. 2003;43:532–546.

6. Campbell, D.T., Stanley, J.C. Experimental and quasi-experimental design. Chicago: Rand McNally, 1963.

7. Neuman, W.L. Social research methods: qualitative and quantitative approaches, ed 5. Boston: Allyn Bacon, 2003.

8. Babbie, E. The practice of social research, ed 12. Belmont, Calif: Wadsworth, 2009.

9. Denzin, N.K., Lincoln, Y.S. Handbook of qualitative research, ed 3. Thousand Oaks, Calif: Sage, 2003.

10. Liebow, E. Tally's corner. Boston: Little Brown, 1967.

11. Maruyama, M. Endogenous research: the prison project. In: Reason P., Rowan J., eds. Human inquiry: a sourcebook for new paradigm research. New York: Wiley & Sons, 1981.

12. Guba, E.G. Criteria for assessing the trustworthiness of naturalistic inquiries. Educ Commun Technol J. 1981;29:75–92.

13. Geertz, C. Thick description: toward an interpretative theory of culture. In: The interpretation of cultures: selected essays. New York: Basic Books; 1973.

14. Hodge, R., Kress, G. Social semiotics. Ithaca, NY: Cornell University Press, 1988.

15. Silverman, D. Doing qualitative research: a practical handbook, ed 2. Thousand Oaks, Calif: Sage, 2004.

16. Krefting, L. Reintegration into the community after head injury: the results of an ethnographic study. Occup Ther J Res. 1989;9:67–83.

17. Headland, T.N., Pike, K.L., Harris, M. Emics and etics: the insider/outsider debate. Newbury Park, Calif: Sage, 1990.

18. Frank, G. Life history model of adaptation to disability: the case of a congenital amputee. Soc Sci Med. 1984;19:639–645.

19. Tashakkori, A., Teddlie, C. Sage handbook of mixed methods social and behavioral research. Thousand Oaks, Calif: Sage, 2003.