Chapter 3
Chapter 4
Chapter 5
Chapter 6
Chapter 6
100

What are the 4 validities? Define at least two of them!

Construct Validity- How well the operational variable reflects the conceptual variable

External Validity - How well the findings of a study generalize to a larger population

Statistical Validity - How well the data analysis supports the claims

Internal Validity - When studying relationships between variables, how well confounds are eliminated

100

Name one historical example of problematic ethics in research.

  • Tuskegee Syphilis Study

  • Milgram Obedience Study

  • Stanford Prison Experiment

  • The “Monster” Study (including children’s stuttering)

100

Explain the difference between measurement validity and measurement reliability.

Measurement Reliability - In research, reliability in measurement refers to the consistency of measurement results.

Measurement Validity - How well the variables in a study are measured ... How well the operational variable reflects the conceptual variable.

100

Give an example of a double-barreled question or a negatively-worded question!

Double-barreled question – The question is really two questions in disguise

-- Eg. “Do you agree that education is crucially important and that the core values of American society are better taught in private schools than public?”

-- If people have to answer this double-barreled question with a singular answer, it’s harder to reflect what they think about this issue as a whole.

-- There’s no way to reflect their perspective about each individual question

Negatively-worded question – The question introduces double negatives that can confuse

-- Eg. “People who do not drive with a suspended license should never be punished”

100

What is a masked design?

Masked design - In these studies, observers are not researchers themselves – they aren’t informed of the aims of the study or the expected findings

200

What are the 3 different types of claims? Define at least one of them.

Association Claims - arguments that two variables are related to one another, or that one variable can predict another

Frequency Claims - refer to just one variable and make an assertion about the levels of that variable

Causal Claims - The levels of one variable cause the levels of the other

200

What are the 3 principles of ethics in Animal Research? (Hint: Alliteration is key here)

  • Replacement – Find alternatives to animal research when possible

  • Refinement – Modify procedures to minimize animal distress

  • Reduction – Fewest animal participants possible

    • Even though having more participants increases statistical power, we shouldn’t necessarily do this for animal participants

200

What is 1 of the 2 ways that measurement validity can be assessed subjectively?

Face validity - Does the measure seem to make sense at face value?

-- Good example: “head circumference” for how well a hat fits

-- Bad example: “head circumference” for intelligence

Content validity - Does the measure cover all/many aspects of the concept or theory?

-- Better example: a test on mathematical, verbal, problem solving, and pattern recognition behaviors for intelligence

-- Worse example: a test on only mathematical behavior for intelligence

200

Explain the difference between likert scales and semantic differential scales!

Likert Scale - Participants are provided a prompt or a question and a scale of ordered responses to choose from.

Semantic Differential Questions - Participants are provided a scale that splits two descriptors and indicate the degree of “closeness” to either side

200

What is reactivity?

Reactivity - Refers to situations in which participant behavior is impacted by the mere presence of someone observing.

300

Explain the difference between a variable and a constant.

Variable

- Any factor in a study that can have multiple (two or more) possible levels (options, choices, or levels).

- A variable can vary

Constant

- A factor in a study with only one value/level. It does not vary.

300

What are the 3 core principles listed in the Belmont Report?

  • Respect for Persons
    • Eg. Informed Consent - Sometimes studies must involve deception. But, we at least have to inform the participants of the risk that may incur by participating.
  • Beneficence
    • Assess potential harm to participants and benefits of study.
      • What harm can this study do to its participants?
      • What benefits are gained as a society by conducting this study?
  • Justice
    • There must be balance between those who participate in research and those who benefit from it (the potential benefits of the study apply to everybody)
    • The sample should represent the population (diverse social backgrounds)
      • Don’t just take particular people from the population if it doesn’t serve the purposes of the study itself
300

Name the 3 types of measurement reliability. Define at least 2 of them!

Test-retest reliability – Are scores consistent each time a measure is applied?

-- Eg. If an IQ test is administered twice in one semester, are the scores consistent?

Interrater reliability – If multiple researchers are observing behavior, do they report it consistently?

-- Eg. Two researchers observe a problem-solving study. A child solves a puzzle in a certain number of steps. Do both researchers count the same number of steps?

Internal Reliability (not Internal Validity) – Is a participant consistent with themselves?

-- Eg. Related self-report questions

----> “How satisfied are you with your job?”

----> “To what extent do you experience job satisfaction?”

300

Surveys can include open-ended questions and forced choice questions. Define the pros and cons of each!

Open-ended questions - Participants answer however they like

-- Pros - Answers are rich and diverse

-- Cons - Can be very difficult to code and categorize responses (too many different answers)

----> Example: “How do you feel about this lecture?”

Forced choice questions - Participants choose one of the available answers.

-- Pros - Manageable to organize and code

-- Cons - Restrictive set of options

----> Example: “Have you enjoyed this lecture? Yes or no”

300

What are 2 of the ways that we can limit how much biases affect our behavioral observations? We talked about 3 in class!

Create an unobtrusive environment

-- Unobtrusive environment - One-way mirror, one face in the crowd, etc.

Waiting until participants become comfortable to record observations

Measure behavioral result instead of behavior itself

-- The advantage of this is that you eliminate some of that awkwardness from the researcher that can influence how the participant behaves

400

What are the 3 different ways to define a variable?

  • Conceptual definition of a variable

    • What does it refer to theoretically? How is it defined? (a definition/more precise description)

    • “Sensory intensity of the light brightness in the classroom”

  • Construct definition of a variable

    • What is it? What are we studying?

    • “Brightness of lights in the classroom”

  • Operational definition of a variable

    • How is it measured and/or manipulated in the study? How is it operationalized?

      • Dim – 500 lumens
      • Medium – 1,000 lumens

      • Bright – 1,500 lumens”

400

Name 3 of the 5 guiding, core ethical principles that the APA has set forth!

  1. Beneficence (and nonmaleficence)

  2. Justice

  3. Respect for people’s rights and dignity

    1. Same three as Belmont Report

  4. Fidelity and Responsibility

    1. Establish relationship of trust, be accountable for professional activities (clinician, teacher, researcher)

  5. Integrity

    1. Strive to be accurate, truthful, and honest

400
Define Criterion Validity. You only get 300 of the points if you don't name the paradigm associated with it.

Criterion validity - Evaluates the statistical association of a measure with a relevant behavior outcome (e.g., with correlation).

-- This is more objective in measure.

-- Eg. Someone’s degree of isolation should correlate with your BDI. We can quantitatively measure this.

-- It can be assessed via the known-groups paradigm, such that different [relevant] populations should be reflected differently by specific measures.

----> Eg. Beck Depression Inventory (BDI)


400

A shortcut means relying on a response set. What are response sets? What are the 2 kinds of response sets that we talked about in class?

Response sets - This occurs when participants respond to numerous items the same way without necessarily thinking it through. A series of answers that follow a trend that are consistent with each other.

-- Acquiescence (affirmative route) - Participants may be biased in favor of choosing the “yes” option more often than they should “or a positive option like “strongly agree”).

-- Fence Sitting (neutral route) - For Likert scales or semantic differential questions, participants may inordinately favor the middle “safe” option (like “Neutral” or “I don’t know”.)

500

What are the 3 criteria for establishing causality?

  • Covariance - As A changes, B changes

  • Temporal Precedence - The study’s method ensures that A comes first in time, before B

  • Internal Validity - The study’s method ensures that there are no plausible alternative explanations for the change in B; A is the only thing that changed

500

Instead of being told they had syphilis, what were the infected individuals told they had instead?

  • Tuskegee Syphilis Study

    • 600 black men were studied over 400 years

    • 400 infected, 200 not

    • Those infected were told they had “bad blood”, NOT syphilis

    • Penicillin treatment available in 1940s, denied to study participants

  • What principles were violated?

  1. Participants were disrespected

  2. Participants were harmed

  3. Participants were a targeted disadvantaged social group

500

What statistical measure is internal reliability often assessed with?

Internal Reliability is often assessed with Cronbach’s alpha.

-- NOT to be confused with the “alpha” (𝛼) you learned about in your statistics course

Cronbach’s alpha is a combined measure of internal correlation and the number of items.

-- Generally want a value ≥ .80.

500

Name 2 of the 3 additional self-report survey issues that we talked about in class (Hint: Think "Walter White")!

Social-desirability - Participants may respond in a way they believe paints them in a good light

-- Using the frame of “If I was a good person or the person I wanted to be, how would I answer this question?” instead of “If I am being truthful about who I am, how would I answer this question?”

“Faking bad” - Participants may respond the opposite way, inordinately emphasizing negative answers (answers that are perceived to be negative/socially undesirable by the participant)

Memory flaws - Participants may simply not remember their own behaviors accurately

-- It’s hard to translate what’s in one’s own mind onto paper, especially if it’s on a limited scale