• Entry
  • Reader's guide
  • Entries A-Z
  • Subject index

The concept of interrater reliability essentially refers to the relative consistency of the judgments that are made of the same stimulus by two or more raters. In survey research, interrater reliability relates to observations that in-person interviewers may make when they gather observational data about a respondent, a household, or a neighborhood in order to supplement the data gathered via a questionnaire. Interrater reliability also applies to judgments an interviewer may make about the respondent after the interview is completed, such as recording on a 0 to 10 scale how interested the respondent appeared to be in the survey. Another example of where interrater reliability applies to survey research occurs whenever a researcher has interviewers complete a refusal report form immediately after a refusal takes ...

    • Loading...
    locked icon

    Sign in to access this content

    Get a 30 day FREE TRIAL

    • Watch videos from a variety of sources bringing classroom topics to life
    • Read modern, diverse business cases
    • Explore hundreds of books and reference titles