• Front Matter
  • Welcome to PSY 362
  • How to use these notes
  • Attribution
    • Major & Additional Attributions
  • License
  • Colophon
  • Public Health Dashboards
    • Wake Forest
    • Forsyth County
      • Vaccinations
      • Case Counts
    • North Carolina
  • I Module 04
  • 1 Welcome to Norms and Meanings of Test Scores
    • 1.1 Learning Goals
    • 1.2 Module Materials
      • 1.2.1 Explore
      • 1.2.2 Engage
      • 1.2.3 Evaluate
      • 1.2.4 Estimated Video Length
  • 2 Lecture: Norms and Meaning of Test Scores
    • 2.1 Norms and Meaning of Test Scores (04.01)
    • 2.2 Statistics as Summary (04.02)
    • 2.3 Measures of Spread (04.03)
    • 2.4 Transforming Raw Scores (04.04)
    • 2.5 Developmental Norms (04.05)
    • 2.6 Within-Group Norms (04.06)
    • 2.7 Referenced Tests Types (04.07)
    • 2.8 Minimums, Cutoffs, and Expectancy Tables (04.08)
    • 2.9 Interpreting Test Scores Realistically: Remastered (04.09)
  • II Module 05a
  • 3 Meet our toolbox!
    • 3.1 R
      • 3.1.1 Install R and RStudio
      • 3.1.2 Testing testing
      • 3.1.3 Add-on packages
    • 3.2 Getting Help with R
      • 3.2.1 Further resources
  • 4 R basics and workflows
    • 4.1 Introducing Functions
    • 4.2 Introduction to Objects
    • 4.3 Workspace and working directory
      • 4.3.1 Workspace, .RData
      • 4.3.2 Working directory
    • 4.4 RStudio projects
    • 4.5 Saving Files and Protips
  • 5 Refreshers and Examples
    • 5.1 Some terms
    • 5.2 Summary Statistics
    • 5.3 Measures of Covariation
      • 5.3.1 Correlation Refresh
  • III Module 05b
  • 6 Welcome to Classical Test Theory
  • 7 Big Ideas in Classical Test Theory
    • 7.1 Modeling Classic Test Theory
    • 7.2 Model Applications
  • 8 Systematic Error and Unreliability
    • 8.1 Simulate a constant true score
    • 8.2 PISA total reading scores with simulated error and true scores based on CTT
    • 8.3 Reliability and unreliability illustrated
  • 9 Domain Sampling Theory
  • IV Module 06
  • 10 Installing epmr
  • 11 Statistical definition of reliability.
    • 11.1 Estimating reliability
    • 11.2 Split-half method
    • 11.3 Spearman Brown
    • 11.4 Standard Error of Measurement
  • 12 Interpreting reliability and unreliability
  • 13 Reliability study designs
    • 13.1 Interrater reliability
      • 13.1.1 Proportion agreement
      • 13.1.2 Kappa agreement
      • 13.1.3 Pearson correlation
    • 13.2 Summary
      • 13.2.1 Exercises
  • V Module 08
  • 14 Validity
    • 14.1 Objectives
  • 15 Overview of validity
    • 15.1 Definitions
    • 15.2 Validity examples
    • 15.3 Sources of validity evidence
  • 16 Content validity
  • 17 Criterion validity
    • 17.1 Definition
    • 17.2 Criterion examples
  • 18 Construct validity
    • 18.1 Definition
    • 18.2 Examples
  • 19 Unified validity and threats
  • 20 Summary
    • 20.1 Exercises
  • VI Module 09
  • 21 Factor Analysis
    • 21.1 Objectives
  • 22 Measurement models
  • 23 Exploratory factor analysis
    • 23.1 Steps in EFA
    • 23.2 Exploring PISA Approaches to Learning
  • 24 Confirmatory factor analysis
    • 24.1 Steps in CFA
    • 24.2 Confirming PISA Approaches to Learning
  • 25 Summary
    • 25.1 Optional Exercises
  • VII Module 10
  • 26 Item Analysis
    • 26.1 Objectives
  • 27 Preparing for item analysis
    • 27.1 Item quality
    • 27.2 Piloting
    • 27.3 Data entry
    • 27.4 Scoring
  • 28 Traditional item statistics
    • 28.1 Item difficulty
    • 28.2 Item discrimination
    • 28.3 Internal consistency
    • 28.4 Item analysis applications
  • 29 Additional analyses
    • 29.1 Option analysis
    • 29.2 Differential item functioning
  • 30 Summary
    • 30.1 Exercises
  • VIII Module 11
  • 31 Item Response Theory
    • 31.1 Learning objectives
  • 32 IRT versus CTT
    • 32.1 CTT review
    • 32.2 Comparing with IRT
  • 33 Traditional IRT models
    • 33.1 Terminology
    • 33.2 The IRT models
    • 33.3 Assumptions
  • 34 Applications
    • 34.1 In practice…
    • 34.2 Examples
  • 35 Summary
    • 35.1 Exercises
  • IX Module 12
  • 36 Practical Testing Applications
    • 36.1 Tests and decision making
      • 36.1.1 Educational decisions
      • 36.1.2 Psychological decisions
    • 36.2 Test types and features
      • 36.2.1 Cognitive and affective
      • 36.2.2 Achievement and aptitude
      • 36.2.3 Other distinctions
    • 36.3 Summary
    • 36.4 Finding test information
  • 37 Summary
    • 37.1 Exercises
  • 38 Test Evaluation
    • 38.1 Test purpose
    • 38.2 Study design
    • 38.3 Reliability
    • 38.4 Validity
    • 38.5 Scoring
    • 38.6 Test use
  • 39 Summary
    • 39.1 Exercises
  • X Homework 2
  • 40 Homework Instructions
  • XI Other Coolness
  • 41 Good Resources
    • 41.1 Learning R
    • 41.2 Psychology Scales
  • References
  • License: CC-BY-SA

Psychological Testing

41 Good Resources

41.1 Learning R

  • https://psychnerdjae.github.io/into-the-tidyverse/
  • Automatic Grading with RMarkdown example
  • Git/Github for virtual learning (from this tweet)
  • Learn-Datascience-for-Free
  • https://allisonhorst.shinyapps.io/dplyr-learnr/

41.2 Psychology Scales

  • International Personality Item Pool
  • Open-Source Psychometrics Project
  • APA PsycTests
  • Measurement Instrument Database for the Social Sciences
  • Paul Spector’s assessment archive
  • GESIS Open Access Repository for Measurement Instruments
  • REDCap’s library of scales
  • PhenX Toolkit
  • Registry of Scales and Measures: Psychological Tests, Scales, Questionnaires and Checklists