Shared Flashcard Set

Details

Research Terminology
Terminology for writing research papers
61
Science
Graduate
02/17/2011

Additional Science Flashcards

 


 

Cards

Term

Evidence-based Practice (EBP)

Definition

An approach in which current, high quality research is integrated with practitioner expertise and client preferences and values into the process of making clinical decisions.

Term
Levels of Evidence
Definition

Poor Quality: -Opinions, Committee Reports

 

Medium Quality: -Case Studies, Well-designed quasi-experimental studies

 

High Level: -Well-designed controlled study without randomization Random controlled studies Systematic reviews or meta-analyses

Term
Scientific Approach
Definition

Systematic study of a given phenomenon:

Research question, Statement of the problem,

Hypothesis Development

Observation/data collection

Interpretation of data to accept or reject hypothesis Involves controlling as many variables as possible Is sometimes hard to apply to the study of human behavior

Term
Types of Research
Definition

Descriptive Research:

Describe populations Exploratory


Research: Find Relationships

 

Experimental: Cause & Effect

Term
Types of Research
Definition

Descriptive Research:  Describe populations


Exploratory Research: Find Relationships


Experimental: Cause & Effect

Term
Types of Literature Reviews
Definition

Narrative: Most common. Nonsystematic review.

Systemic: Method for locating, appraising & synthesizing evidence to help readers determine the best treatment practices.

Meta-analyses: Systematic review to which statistical weight is assigned to the studies included in the review to provide treatment results information.

Term
Variables
Definition

Measurable quantities that vary or change under different circumstances rather than remaining constant. Varying characteristics of a phenomenon.

Term
Types of Variables
Definition

Independent (IV) Conditions that cause changes in behavior. Presumed cause of the dependent variable.

 

Dependent (DV) What is measured to see if a change had been caused (by the IV)

Especially important in experimental research.

Used in cause-and-effect relationships.

Term
Types of Independent Variables
Definition

Active Variables:

Can be manipulated to determine the effect of one independent variable on another variable.

In an experimental study, active & IVs are the same thing.

Not all IVs are active variables which can be manipulated(e.g. age, gender, ethnicity cannot be manipulated)

Attribute Variables:  

Subject characteristics or attributes(e.g. age, gender, ethnicity)

In a descriptive study Attribute & IV are the same thing

Term

Identify the IV & the DV. State whether the IV is an active or attribute variable:

Do boys or girls have higher rates of autism?

Definition

Boy or girl : Independent attribute variable


Autism: Dependent

Term

Identify IV(s) and DV(s).

State whether IV(s) is/are active or attribute(s):

 

Are people who stutter more fluent when they decrease their speaking rate by 20%, 10% or 0%?

Definition

Population: People

Independent attribute variable  Who stutter:???

Independent Variable 20, 10, 0% One active independent variable with 3 rates

 

**not sure this is correct


Term

Identify IV(s) & DV(s).

State whether IV(s) is/are active or attribute variable(s)

 

Are underweight elementary school children who were born prematurely more likely to have lower standardized language scores than underweight elementary school children who were not born prematurely?

Definition

Population: Underweight schoolchildren


Language scores: Dependent variable


Birth status: One independent attribute (can't change) variable with 2 positions

Term
Types of variables
Definition

Categorical:

-Cannot be measured

-Named or categorized

-Uses the nominal scale of measurement (simple grouping)

Continuous:

-May be measured on a continum

-Rank ordering or numerical

-Uses the ratio, interval or ordinal form of measurement

Term
Categorical variables -Nominal
Definition

Nominal

-lowest on the measurement hierarchy

-Subjects are sorted into categories *Male/female *People who do/do not stutter *Caucasian, African American, Hispanic

-Sometimes data or participants can be coded e.g., 1=normal, 2=abnormal. The coded data is only a means of sorting the data into categories, and cannot be analyzed numerically. Lowest on the measurement hierarchy Least quantifiable Least accurate Results have less meaning

Term

Types of continuous variables

Ordinal Measures

Definition

Ordinal Measures

-Ranked data

-Very much like nominal scale Heaviest-lightest Tallest-shortest Mild, moderate, severe stuttering, hearing loss

-The intervals between variables are not equal

-Data often coded into number, (e.g., severe=1, moderate=2, mild=3) but cannot be analyzed mathematically Some numbers based on continuous variables data that is ranked. (numbers underpin the data rankings)

Term

Categorical Variables Examples

Do SLP's who hold the ASHA CCCs earn more money on average than SLPs without CCCs?


Is there a relationship between type of career and being diagnosed with a specific type of hearing loss (e.g. conductive, sensorineural, mixed)?

Definition

SLPs population 

With or without CCCs: Independent, Attribute (cannot be manipulated by researcher

Descriptive Research

 

Population: people w/ hearing loss Corralative study: Relationship, type of career might predict the hearing loss type: Categorical Type of career: Nominal, Categorical variable

Term

Ordinal Examples


Definition

Do women with mild, moderate or severe hearing loss earn as much money on average as women who have no hearing loss?

 

Is the smallest available dose of a drug (5-10mg) less ototoxic than the largest available doses (25-50mg)?

Term
Types of Continuous Variables
Definition

Interval Measures

-Equal intervals between values (2, 4, 6, 8)

-The difference between intervals is the same, (e.g., the difference between 1 and 5 is not the same as the difference between 12 and 16.

- No Meaningful zero Some Language test scores Ratings made on a Likert or Likert-type scale(Survey results: Rating teachers at the end of a term) Temperature

Term
Examples of Continuous Variables
Definition

-What are student's attitudes toward people who stutter as measured by a 7-point Likert scale in which 1=strongly agree and 7=strongly disagree?

 

-Are students' scores on a language test correlated with frequency of aggressive behaviors noted in the classroom? (Score on the test is the interval)

Term
Types of Continuous Variables
Definition

- Ratio Measures

-Highest level of measurement

-Equal intervals

-There is a true zero (e.g., the attitude being measured may not be present)

-Age -Weight Height

-Measurements of frequency, time, or distance

Term
Examples of Types of Continuous Variables
Definition

-Is the preceived age of a person's voice directly related to their chronological age? (ratio variable)

 

-Are students' scores on a language test correlated with frequency of aggressive behaviors noted in the classroom?

Term
Experimental Research
Definition

-Appropriate when investigating cause-and-effect relationships

-Characteristics Researcher has a purpose and has determined when to observe a certain phenomenon. Researcher can control the events thus observe specific behavior changes. Able to replicate. Researcher can manipulate or control conditions (independent variables) to measure the effect of the condition on behaviors (dependent variables).

Term

Bivalent Experiments

 

One IV

Two Values

 

** See graph

Definition

Study of the effects of 2 values of one independent variable on the dependent variables

(Gender is a categorical variable that has 2 levels of one independent variable).

( The effects of quiet & noise on stutters{independent variables])

Term

Multivalent Experiments

 

One IV

3 or More Values

 

**  See graph

Definition

The independent variable can be altered in a dynamic manner that allows for 3 or more levels to affect the dependent variable.

Term
Descriptive Research
Definition

-Examines group differences

Do not lead to cause effect relationships.

Describes a phenomenon.

Not to say it is inferior to experimental research.

A question of which research design is the most appropriate. (Your question needs to be descriptive or experimental)

Term
Variables in Descriptive Research
Definition

-Classification Variable

Analogous in independent variable A way of describing participants or variables.

 

-Criterion Variable Performance variables Analogous to dependent variable

Term

Example of Bivalent


(2 levels of 1 independent variable)


Descriptive Research

 

**See graph

Definition
Comparison of results on tests Criteria variable
Term

Example of Multivalent Descriptive Research

 

One IV

3 Values

 

** See graph

Definition

In study of on Normal, Hearing Impaired & Deaf subjects (Independent/Classification Variable)  One IV 3 Values


Number of words produced (Dependent/Criterion Variable)

Term
Other Variables in Descriptive Research
Definition

-Predictor Variable Change Agent/Independent Predicts the effect on the predicted variable.

 

-Predicted Variable/Dependent Variable Affected by the predictor

Term

Example of Predictor/Predicted Variables

 

 

**See graph

Definition

Scatterplot depicting the relationship between perceived age and chronological age of 175 talkers. Perceived age judged by listeners.

Predictor/Independent Variable: Chronological Age Predicted/Dependent Variable:Vocally Perceived Age

Term
Types of Validity
Definition

Internal

External

Content

Criterion

Construct

Term
Internal Validity
Definition

-Degree to which observed changes (dependent variable) were a result of the variables that the researcher manipulated or controlled (independent variables)

 

Associated with experimental, cause-and-effect research (not used on surveys)

Term
Extraneous variables which may affect internal validity Not interested in, but may affect my study
Definition

-History:The specific events occurring between the first and second measurement in addition to the experimental variable. (English language learners. We have accounted for the parents but not the siblings) On going history


-Maturation: Processes within the subjects operating as a function of the passage of time per se (not specific to the particular events). Including growing older, growing hungrier (during tests), growing more tired, etc. How do you know you caused the effect & maturation did not?

 

-Testing: The effects of taking a test upon the scores of a second (i.e., subsequent) testing Pretests help your scores on actual test. Warms you up. "Testing Effect" on internal validity. Longitudinal studies.

Term
Extraneous Variables Which May Affect Internal Validity
Definition

-Statistical regression: Occurs when groups have been selected on the basis of their extreme scores (random has more control)

 

-Differential selection of subjects: Biases may result if the participants in the control or comparison groups are different from one another(Broca's vs Wernicki's Aphasia)

Term
Extraneous Variables Which May Affect Internal Validity
Definition

-Experimental Mortality: Differential loss of subjects from the comparison or control groups


- Interaction of Factors: The possible interaction effects among two or more of the previously described jeopardizing factors. Lose participants for one reason or another . A combination of these "threats" to the study.


-Instrumentation: Changes in the calibration of a measuring instrument or changes in hte observers or scorers used may produce changes in the obtained measurements. Audiometer off by 10dB so could not get consistent scoring protocol

Term
Ensuring "Objectivity" of Measurements Made by Human Observers or Judges
Definition

- Identify specific criteria for judging/evaluating the performance/behavior of participants


- Train observers/judges in order to ensure that they are applying the criteria correctly and consistantly


- Evaluate intraobserver reliability (consistency "within" each observer or judge)

- Evaluate interobserver reliability (consistency "between" observers or judges)

Term
External Validity
Definition

- Generalization of finding across populations, settings, and times. (Applying to other populations in other studies.

- Addresses to what populations, settings, treatment variable, and measurement variable the effect can be generalized

- Is associated with the replication of studies (Want to replicate in other circumstances)

Term
Factors That Jeopardize External Validity
Definition

- Subject Selection: Biases the degree to which the subjects chosen for the study are representative of the population to which the researcher wishes to generalize. (As much as you can. Include both sexes, Pretesting effects)


Reactive or Interactive Effects of Pretesting: In which a pretest might increase or decrease the subject's sensitivity or responsiveness to the experimental variable and thus make the results obtained for a pretested population unrepresentative of the effects of the experimental variable for the un-pretested universe from which the experimental subjects were selected. (Research environment may not mirror outside world)

Term
Factors That Jeopardize External Validity
Definition

- Reactive effects of experimental arrangements, which would preclude generalization about the effect of the experimental variable upon persons being exposed to it in non-experimental settings.   (Your environment may affect how well your study generalizes)

- Multiple-treatment interference: Likely to occur whenever multiple treatments are applied to the same subjects, because the effects of prior treatments are usually not erasable. (A threat to internal validity also and a threat to generalizing to other populations.)

Term

Reliability

 

 

** See graph

Definition

- Refers to the consistency of a rater or of a measurement.

-Intra or interobserver (or judge)reliability

-Test-retest reliability Participants' test scores are compared from the first time participants take the test to the second time they take the same test.

-Does not confirm validity

Term

Quantitative Research / (Descriptive is Quantitative) 


Definition

Based on numerical measurements of observed phenamenon

Places much emphasis on control

Vary on a continum from most valid to least valid

   True experimental designs (most credible)

   Single-subject designs

   Quasiexperimental designs

   Pre-experimental designs

   Nonexperimental designs(weskest design)

  

Descriptive research/Get info out there


 

Term

Characteristics of good quantitative, experimental design

 

Definition

A good study attempts to ensure that only the IV(s) caused the change to the DV(s)


Extraneous (didn't consider how would impact results) variables must be accounted for or controlled

   Intrinsic variables

   Extrinsic variables

 

Term

Intrinsic Variables/ Inherent to the participant


Definition

Associated with the research participant or subject

May be controlled in a number of ways

  Randomization - randomly assign participants or groups

  Homogeneity - choose only subjects who have the same characteristics

  Blocking - account for extraneous variables by making them independent variables(aphasia subjects who attend support grp. are built into the study to compare effects.)

   Analysis of covariance - the effects of one variable are statistivally analyzed to compute its effects on another variable.  (One variable builds on another.  Ven diagram, overlap.  Same dependent variable.) 

 

Term

Extrinsic Variables


Definition

 

Are associated with the research environment and research protocols (researcher may have effect , with personality)

Should be controlled through systematic application of research protocol and consistency of researchers/observers

* How do you standardize protocols to avoid.  i.e. The experimenter nodding head will affect outcome.

 

Term

Other Characteristics of a Good Design

 

Definition

  • Design is appropriate for the question being asked
  • Should not result in biased data
  • Precision of measurements & procedures 
  • The power of design & its relationship to detect relationships among variables; often related to statistical analysis

 

Term

Null Hypothesis (Used in Experimental)


Definition

Scenario: Clinician hypothesizes that early intervention works better than traditional intervention when treating children with autism.  (Designs study to determine if early intervention is more effective based on language standardized test)

Clinician's goal should not be to prove her hypothesis is correct. Instead, she experiments to determine whether to reject or fail to rejectnull hypothesis.

  Null Hypothesis (Ho) = No difference between early/traditional intervention

  Alternative Hypothesis(H1): = Early intervenion more effective than tradidional intervention.

 

Term

Null Hypothesis (2)   (significance level)


Definition

Many statistical tests can be run, but the end result is a p value, or probability value.  P value is pre-determined before data is analyzed & often is .05

If p is = to or greater than .05 we fail to reject null hypothesis(i.e., early intervention is not more effective than traditional intervention). 

If p value is less than .05, we reject null hypothesis & assume our alternative hypothesis is correct (i.e., early intervention is more effective than traditional intervention).

For those who are designing an experiment or quasi-experimental study, results should be framed this way.

(Experimental design needs null hypothesis.)

 

Term

Factorial (group) designs


Definition

 

Sometimes researchers need to measure different levels of multiple variables.

Factorial designs have two or more independent variables which are manipulated simultaneously.

Researchers may analyze the main effects of independent variables separately and the interaction effects of the variables.

 

Term

Factorial (Group) Designs


Definition

 

e.g.Does language output(what's being measured) of children who use AAC devices vary according to the child's gender (IV, 2 levels) & whether child uses device at home or school.

2x2 design (e.g., there were 2 IV's, each with 2 different levels)

4 levels, 2 associated with each variable, use of device at school or at home (IV, 2 levels).

 

Term

Factorial (Group) Designs


Definition

 

e.g. Do cochlear implants versus hear aids (hear devices, IV) result in better speech intelligibility scores (DV) for men or women (gender  IV) with mild, moderate, or severe (3 levels IV) hearing loss?

This is a 2x2x3 design (e.g., 3 IV's, 2 have 2 levels. 1 has 3 levels.

(Seen in the methods section: factorial IV's)

 

Term

Multivariate(2 or more DV's) (Group) Designs


Definition

 

There are multiple DV's, just as there are multiple IV's (Bivalent, Multivalent)

Multivariate designs involve two or more DV's & provide information about the effect of the IV on each dependent variable.

Example: Do cochlear implants versus hearing aids result in (1) better speech intelligibility & (2) language proficiency scores for men & women with mild, moderate or severe hearing loss?

(Multiple dependent variables)

 

 

Term

Within-Subjects Group Designs


Definition

 

Unlike group designs, participants are not divided into separate groups. 

Each participant will experience all of the treatments or conditions.

Also called "repeated measures designs"

Can include factorial and multifactorial designs as well as multiple control groups or simpler designs (e.g., single factors)

 

Term

Within-Subjects Group Designs

 

Definition

 

Example: An SLP wants to know which form of sensory integration (bouncing on a large ball or swinging on a blanket) results in more immediate verbal output for children with autism.

She identifies 30 children with autism & trains parents research protocol.

On 1 day, child bounces on ball.  Tape record for 30 minutes after activity.

On 2nd day, swing child in blanket.  Child recorded for 30 min. after.

Researcher analyzes recordings & makes compares ball & blanket data.

All 30 kids do both treatment conditions.

Group of 30 more power than grp of 15.

Don't need peer matching cause kids act as own control.

 

Term

Within-Subjects Designs

 

Definition

 

Advantages

Requires fewer subjects

Subjects do not have to be matched

Statistics can be easier to interpret

Disadvantages

Can be less powerful ensuring the change on DV due to IV & not extraneous variables.

Carryover effects of one treatment on another

Treatment order is great concern

Counterbalance.  Reverse order of treatment with 1/2 group


 

Term

Single-Subject Designs (Respected)

 

Definition

 

Focus on behaviors of one or few participants, but NOT same a case study

Utilize repeated measures

Involve design phases

Baseline designs most common type, e.g. ABA studies.  Each letter is a phase.

 

Term

Single-Subjects Designs (cont)

 

Definition

 

Baselines - Repeated measurement of DV prior to introduction of treatment

Baseline is solid if stable, consistent

 

Probes - Repeated measurement of DV obtained during treatment

A(no treatment) B(treatment)


 

Term

Example of ABA Design

 

Definition

Want withdrawal to go down to baseline.

Show treatment works.

Shows definite trend.

Ethically could leave at 30% over baseline, better for subject

 

Term
Visual Analysis of Single-Subject Designs
Definition

Although both visual & statistical analysis can be used, single-subjext designs typically use visual analysis ofGRAPHS or DIAGRAMS.  You can look at graph & see progress.

Treatment effects are typically large enough to support conclusions without statistical analyses

Single-subject designs focus on relatively large treatment effects for individual subjects rather than small differences between groups of subjects.

What are the ethical issues of withdrawal? Ideally do 2nd B phase & end on high note.

Term
Qualitative Research - An Overview
Definition

Not numerically driven Not objective

Addresses subjective experiences of subjects rather than objective measures

Often conducted via surveys, interviews, focus grp Thematic analysis versus statistical analysis Not less credible than other research!!! May code data with #'s, but not analyzed statistically. Should not be evaluated with same valuation as quantitative research. Credibility is a qualitative value

Term
Null Hypothesis Continued
Definition
Many statistical tests can be run, but the end result is a p value, or probability value. P value is pre-determined before data is analyzed & ofter is .05
If p is = to or greater than .05 we fail to reject null hypothesis(i.e., early intervention is not more effective than traditional intervention).
If p value is less than .05, we reject null hypothesis & assume our alternative hypothesis is correct (i.e., early intervention is more effective than traditional intervention).
For those who are designing an experiment or quasi-experimental study, results should be framed this way.
(Experimental design needs null hypothesis.)
Supporting users have an ad free experience!