Measurement and Assessment in Education, 2nd edition

Published by Pearson (April 29, 2008) © 2009

  • Cecil R. Reynolds
  • Ronald B. Livingston University of Texas, Tyler
  • Victor Willson
$170.66

  • Hardcover, paperback or looseleaf edition
  • Affordable rental option for select titles
  • This text is very "user friendly" and assumes that the students, with a little preparation, can master the more technical aspects of educational assessment and gain a good understanding of the mathematical concepts needed to master measurement and assessment (Chs. 2-6).
  • Ethical principles, legal issues, and professional standards relevant to classroom assessment are covered thoroughly so that students are prepared to conduct classroom assessments in a professional and ethical manner (throughout the text, but specifically in Ch. 17).
  • An entire chapter (Ch. 15) is devoted to the use of assessments for students with disabilities to prepare students to assess the knowledge and skills of all students, including those with disabilities.
  • Contemporary issues regarding the assessment of students are covered in detail so that students are aware of important issues related to educational assessment.
  • Numerous pedagogical devices such as Exercises, Cases, and End of Chapter Problems are included throughout the text so that students can explore topics further.
  • A Student Activity Manual with over 96 pages of practice exercises that will enhance understanding is available with the text.
  • A Test Bank, Instructor's Manual and Powerpoints are available electronically by contacting your local sales representative.  

Chapter 1: Introduction to Educational Assessment

 

I. Introduction

 

II. The Language of Assessment

A. Tests, Measurement, & Assessment

            B. Types of Tests

            C. Types of Scores Interpretations

 

III. Assumptions of Educational Assessment

A. Psychological and educational constructs exist.

B. Psychological and educational constructs can be measured.

C. While we can measure constructs, our measurement is not perfect.

D. There are different ways to measure any given construct.

E. All assessment procedures have strengths and limitations.

F. Multiple sources of information should be part of the assessment process.

G. Performance on tests can be generalized to non-test behaviors.

H. Assessment can provide information that helps educators make better        educational decisions.

I. Assessments can be conducted in a fair manner.

J. Testing and assessment can benefit our educational institutions and society as a whole.

                       

IV. Participants in the Assessment Process

            A. People who develop tests.

            B. People who use tests.

            C. People who take tests.

            D. Other people involved in the assessment process.

 

V. Common Application of Educational Assessments

            A. Student Evaluation

            B. Instructional Decisions

            C. Selection, Placement, and Classification Decisions

            D. Policy Decisions

            E. Counseling and Guidance Decisions

 

VI. What Teachers Need to Know About Assessment

A. Teachers should be proficient in selecting professionally developed assessment

procedures that are appropriate for making instructional decisions.

B. Teachers should be proficient in developing assessment procedures that are

appropriate for making instructional decisions.

C. Teachers should be proficient in administering, scoring, and interpreting

professionally developed and teacher-made assessment procedures.

D. Teachers should be proficient in using assessment results when making

educational decisions.

E. Teachers should be proficient in developing valid grading procedures that

incorporate assessment information.

F. Teachers should be proficient in communicating assessment results.

G. Teachers should be proficient in recognizing unethical, illegal, and other

inappropriate uses of assessment procedures or information.

 

VII. Educational Assessment in the 21st Century

            A. Computerized Adaptive Testing (CAT) and Other Technological Advances.

            B. Authentic Assessments

            C. Educational Accountability and High-Stakes Assessment

            D. Trends in the Assessment of Students with Disabilities

 

VIII. Summary

 

Tables

 

A. Table 1.1: Major Categories of Tests

B. Table 1.2: Norm- and Criterion-Referenced Scores

C. Table 1.3: Assumptions of Educational Assessment

D. Table 1.4: Common Applications of Educational Assessments

E. Table 1.5: Teacher Competencies in Educational Assessment

 

Special Interest Topics

 

A. Special Interest Topic 1.1: Cognitive Diagnostic Assessment — Another Step Toward Unifying Assessment and Instruction

B. Special Interest Topic 1.2: Technology and Assessment in the Schools

C. Special Interest Topic 1.3: Princeton Review's Rankings of High-Stakes Testing Programs

D. Special Interest Topic 1.4: The “Nation’s Report Card”

E. Special Interest Topic 1.5: What Does the 21st Century Hold for the Assessment Profession?

 

Chapter 2: The Basic Mathematics of Measurement

 

I. The Role of Mathematics in Assessment

 

II. Scales of Measurement

            A. What is Measurement?

            B. Nominal Scales

            C. Ordinal Scales

            D. Interval Scales

            E. Ratio Scales


III. The Description of Test Scores

            A. Distributions

            B. Measures of Central Tendency

            C. Measures of Variability

 

IV. Correlation Coefficients

            A. Scatterplots

            B. Correlation and Prediction

            C. Types of Correlation Coefficients

            D. Correlation and Causality

 

V. Summary

 

Tables

Table 2.1: Common Nominal, Ordinal, Interval, & Ratio Scales

Table 2.2: Distribution of Scores for 20 Students

Table 2.3: Ungrouped Frequency Distribution

Table 2.4: Group Frequency Distribution

Table 2.5: Calculating the Standard Deviation and Variance

Table 2.6: Calculating a Pearson Correlation Coefficient

 

Figures

 

Figure 2.1: Graph of the Homework Scores

Figure 2.2: Hypothetical Distribution of Large Standardization Sample
Figure 2.3: Negatively Skewed Distribution

Figure 2.4: Positively Skewed Distribution

Figure 2.5: Bimodal Distribution

Figure 2.6: Relationship between Mean, Median, and Mode in Normal and Skewed Distributions

Figure 2.7: Three Distributions with Different Degrees of Variability

Figure 2.8: Scatterplots of Different Correlation Coefficients

 

Special Interest Topics

 

Special Interest Topic 2.1: Population Parameters and Sample Statistics

Special Interest Topic 2.2: A Public Outrage: Physicians Overcharge Their Patients

Special Interest Topic 2.3: Is the Variance Always Larger Than the Standard Deviation?

Special Interest Topic 2.4: Caution: Drawing Conclusions of Causality


Chapter 3: The Meaning of Test Scores

 

I. Introduction

 

II. Norm-Referenced & Criterion-Referenced Score Interpretations

A. Norm-Referenced Interpretations

B. Criterion-Referenced Interpretations

 

III. Norm-Referenced, Criterion-Referenced, or Both?

 

IV. Qualitative Description of Scores

 

V. Summary

 

Tables

 

Table 3.1: Transforming Raw Scores to Standard Scores

Table 3.2: Relationship of Different Standard Score Formats

Table 3.3: Converting Standard Scores From One Format to Another

Table 3.4: Characteristics of Norm-Referenced and Criterion-Referenced Scores

 

Figures

Figure 3.1: Illustration of the Normal Distribution

Figure 3.2: Normal Distribution with Mean, Standard Deviation, & Percentages.

Figure 3.3: Normal Distribution Illustrating the Relationship among Standard Scores.

  Special Interest Topics   Special Interest Topic 3.1: The “Flynn Effect.” Special Interest Topic 3.2: Whence the Normal Curve?

Special Interest Topic 3.3: Why do IQ Tests use a Mean of 100 and Standard Deviation of 15?

Special Interest Topic 3.4: The History of Stanine Scores

Special Interest Topic 3.5: Every Child on Grade Level?

 

Chapter 4: Reliability for Teachers

I. Introduction

 

II. Errors of Measurement

            A. Sources of Measurement Error

 

III. Methods of Estimating Reliability

            A. Test-Retest Reliability

            B. Alternate Form Reliability

            C. Internal Consistency Reliability

            D. Inter-Rater Reliability

            E. Reliability of Composite Scores

            F. Selecting a Reliability Coefficient

            G. Evaluating Reliability Coefficients

            H. How to Improve Reliability

            I. Special Problems in Estimating Reliability

 

IV. The Standard Error of Measurement

            A. Evaluating the Standard Error of Measurement

 

V. Reliability: Practical Strategies for Teachers

 

VI. Summary

Tables

 

Table 4.1: Major Types of Reliability

Table 4.2: Half-Test Coefficients and Corresponding Full-Test Coefficients Corrected

with the Spearman-Brown Formula

Table 4.3: Calculation of KR 20

Table 4.4: Calculation of Coefficient Alpha

Table 4.5: Calculating Inter-Rater Agreement

Table 4.6: Source of Error Variance Associated with Major Types of Reliability

Table 4.7: Reliability Expected When Increasing the Numbers of Items

Table 4.8: Standard Errors of Measurement for Values of Reliability and Standard Deviations

Table 4.9: Reliability Estimates for Tests with a Mean of 80%

 

Figures

 

Figure 4.1: Partitioning the Variance

 

Special Interest Topics

 

Special Interest Topic 4.1: Generalizability Theory

Special Interest Topic 4.2: Consistency of Classification with Mastery Tests

Special Interest Topic 4.3: A Quick Way To Estimate Reliability for Classroom Exams

 

Chapter 5: Validity for Teachers

 

I. Introduction

            A. Threats to Validity

            B. Reliability & Validity

           

II. "Types of Validity" versus "Types of Validity Evidence"


III. Types of Validity Evidence

A. Evidence Based on Test Content

B. Evidence Based on Relations to Other Variables

            C. Evidence Based on Internal Structure

            D. Evidence Based on Response Processes

            E. Evidence Based on Consequences of Testing

F. Integrating Evidence of Validity

 

IV. Validity: Practical Strategies for Teachers

 

V. Chapter Summary

 

Tables

 

Table 5.1: Tracing Historical Trends in the Concept of Validity

Table 5.2: Sources of Validity Evidence

 

Figures

 

Figure 5.1: Illustration of Item Relevance

Figure 5.2: Illustration of Content Coverage

Figure 5.3: Predictive and Concurrent Studies

Figure 5.4: Graph of a Regression Line

 

Special Interest Topic

 

Special Interest Topic 5.1: Regression, Prediction, and Your First Algebra Class

 

Chapter 6: Item Analysis for Teachers

 

I. Introduction

 

II. Item Difficulty Index (or Item Difficulty Level)

A. Special Assessment Situations and Item Difficulty

 

III. Item Discrimination

A. Item Discrimination on Mastery Tests

B. Difficulty and Discrimination on Speed Tests

 

IV. Distracter Analysis

            A. How Distracters Influence Item Difficulty and Discrimination

 

V. Item Analysis: Practical Strategies for Teachers

 

VI. Using Item Analysis to Improve Items

 

VII. Item Analysis and Performance Assessments

VIII. Qualitative Item Analysis

 

IX. Using Item Analysis to Improve Classroom Instruction

 

X. Summary

 

Tables

 

Table 6.1: Optimal p Values for Items with Varying Numbers of Choices

Table 6.2: Guidelines for Evaluating D Values

Table 6.3: Maximum D Values at Different Difficulty Levels

Table 6.4: Two Examples of Test Scoring and Item Analysis Programs

 

Special Interest Topics

 

Special Interest Topic 6.1: Item Difficulty Indexes and Power Tests

Special Interest Topic 6.2: Item Analysis for Constructed Response Items

Special Interest Topic 6.3: Developing a Test Bank

 

Chapter 7: The Initial Steps in Developing a Classroom Test:

Deciding What to Test and How to Test It

 

I. Introduction

 

II. Characteristics of Educational Objectives

 

III. Taxonomy of Educational Objectives

A. Cognitive Domain

B. Affective Domain

C. Psychomotor Domain

 

IV. Behavioral versus Nonbehavioral Educational Objectives

 

V. Writing Educational Objectives

VI. Developing a Table of Specifications

VII. Implementing the Table of Specifications and Developing an Assessment

            A. Norm-Referenced versus Criterion-Referenced Assessment

            B. Selecting which types of items to use

            C. Putting the Assessment Together

VIII. Preparing your Students and Administering the Assessment.

IX. Summary

Tables

 

Table 7.1: Bloom's Taxonomy of Educational Objectives

Table 7.2: Krathwohl’s Taxonomy of Affective Objectives

Table 7.3: Harrow’s Taxonomy of Psychomotor Objectives

Table 7.4: Learning Objectives for Chapter 2: The Basic Math of Measurement

Table 7.5: Table of Specifications for Test on Chapter 2: Based on Content Areas

Table 7.6: Table of Specifications for Test on Chapter 2: Content Areas with Percentages

Table 7.7: Strengths and Weaknesses of Selected-Response Items

Table 7.8: Strengths and Weaknesses of Constructed-Response Items

Table 7.9: Practical Suggestions for Assembling an Assessment

 

Special Interest Topics

 

A. Special Interest Topic 7.1: Suggestions for Reducing Test Anxiety.

B. Special Interest Topic 7.2: Strategies for Preventing Cheating

Chapter 8: The Development and Use of Selected-Response Items

 

I. Introduction

II. Multiple-choice Items

A. Guidelines for Developing Multiple-choice Items

B. Strengths and Weaknesses of Multiple-choice Items

 

III. True-False Items

A. Guidelines for Developing Multiple-choice Items

B. Strengths and Weaknesses of Multiple-choice Items

 

IV. Matching Items

A. Guidelines for Developing Matching Items

B. Strengths and Weaknesses of Matching Items

 

V. Summary

Tables

 

Table 8.1: Checklist for the Development of Multiple-choice Items

Table 8.2: Strengths and Weaknesses of Multiple-choice Items

Table 8.3: Checklist for the Development of True-False Items

Table 8.4: Strengths and Weaknesses of True-False Items

Table 8.5: Checklist for the Development of Matching Items

Table 8.6: Strengths and Weaknesses of Matching Items

 

 

Special Interest Topics

 

Special Interest Topic 8.1: Do Multiple-choice Items Penalize Creative Students?

Special Interest Topic 8.2: Correction for Guessing

Special Interest Topic 8.3: What research says about "Changing your answer?"

 

Chapter 9: The Development and Use of Constructed-Response Items

 

I. Introduction

II. Oral Testing: The Oral Essay as a Precursor of Constructed-Response Items

III. Essay Items

A. Purposes of Essay Items

B. Essay Items at Different Levels of Complexity

            C. Restricted-Response versus Extended-Response Essays

            D. Guidelines for Developing Essay Items

            E. Strengths and Weaknesses of Essay Items

            F. Guidelines for Scoring Essay Items

 

III. Short-Answer Items

A. Guidelines for Developing Short-Answer Items

B. Strengths and Weaknesses of Short-Answer Items

 

IV. A Final Note: Constructed-Response versus Selected-Response Items

 

V. Summary

Tables

 

Table 9.1: Purposes of Essay Testing

Table 9.2: Guidelines for the Development of Essay Items

Table 9.3: Strengths and Weaknesses of Essay Items

Table 9.4: Holistic Scoring Rubric

Table 9.5: Analytic Scoring Rubric

Table 9.6: Guidelines for Scoring Essay Items

Table 9.7: Guidelines for the Development of Short-Answer Items

Table 9.8: Strengths and Weaknesses of Short-Answer Items

 

Special Interest Topics

 

Special Interest Topic 9.1: Computer Scoring of Essay Items

 

Chapter 10: Performance Assessments & Portfolios

I. Introduction - What Are Performance Assessments?


II. Guidelines for Developing Effective Performance Assessments

A. Selecting Appropriate Performance Tasks

            B. Developing Instructions

            C. Developing Procedures for Scoring Responses

            D. Implementing Procedures to Minimize Errors in Rating

 

III. Strengths & Weaknesses of Performance Assessments

 

IV. Portfolios

 

V. Guidelines for Developing Portfolio Assessments

 

VI. Strengths & Weaknesses of Portfolios

 

VII. Summary

 

List of Tables

 

Table 10.1: Guidelines for Selecting Performance Tasks.

Table 10.2: Guidelines for Developing Instructions for Performance Assessments.

Table 10.3: Example of a Rating Scale using Verbal Descriptions

Table 10.4: Example of a Numerical Rating Scale 

Table 10.5: Example of a Graphic Rating Scale 

Table 10.6: Example of a Descriptive Graphic Rating Scale 

Table 10.7: Example of a Checklist Used with Preschool Children

Table 10.8: Guidelines for Developing and Implementing Scoring Procedures

Table 10.9: Strengths & Weaknesses of Performance Assessments

Table 10.10: Guidelines for Developing Portfolio Assessments

Table 10.11: Strengths and Weaknesses of Portfolios Assessments

 

Special Interest Topics

 

Special Interest Topic 10.1: Example of a Performance Assessment in Mathematics

Special Interest Topic 10.2: Reliability Issues in Performance Assessments

Special Interest Topic 10.3: Performance Assessments in High-Stakes Testing.

 

Chapter 11: Assigning Grades on the Basis of Classroom Assessments

 

I. Introduction

 

II. Feedback and Evaluation

A. Formal and Informal Evaluation

B. The Use of Formative Evaluation in Summative Evaluation

 

III. Reporting Student Progress: Which Symbols to Use

 

IV. The Basis for Assigning Grades

 

V. Frame of Reference

            A. Norm-Referenced Grading (Relative Grading)

            B. Criterion-Referenced Grading (Absolute Grading)

            C. Achievement in Relation to Improvement or Effort

            D. Achievement Relative to Ability

            E. Recommendations

 

VI. Combining Grades Into a Composite

 

VII. Informing Students of Grading System

 

VIII. Parent Conferences

 

IX. Summary

 

Tables

 

Table 11.1: Report Form Reflecting Achievement and Non-Achievement Factors

Table 11.2: Examples of Grading based on Effort versus Achievement

Table 11.3: Example of Weighting Different Assessment Procedures

Table 11.4: Commercially Available Grade Book Programs

Table 11.5: Example of Grading Requirements Presented to Students at the Beginning of a Test & Measurement Course

 

Special Interest Topics

 

Special Interest Topic 11.1: Brief History of Grading

Special Interest Topic 11.2: Schools No Longer Assigning Ds?

Special Interest Topic 11.3: Grading and Punishment?

Special Interest Topic 11.4: Some Thoughts on Weighting Assessment Procedures

 

Chapter 12: Standardized Achievement Tests in the Era of High-Stakes Assessment

 

I. Introduction

 

II. Group Administered Achievement Tests

            A. Commercially Developed Group Achievement Test

            B. State Developed Achievement Tests

            C. Best Practices in Achievement Testing in the Schools

 

III. Individual Achievement Tests

 

IV. Selecting an Achievement Battery

 

V. Summary

 

Tables

 

Table 12.1: Major Publishers of Standardized Group Achievement Tests

Table 12.2: Important Test-Taking Skills to Teach Students

 

Figures

 

Figure 12.1: TerraNova The Second Edition Sample Report - Individual Profile Report

Figure 12.2: TerraNova The Second Edition Sample Report - Home Report

Figure 12.3: Performance Profile for Iowa Tests of Basic Skills (ITBS)

Figure 12.4: Profile Narrative for Iowa Tests of Basic Skills (ITBS)

Figure 12.5: Score Labels for the Iowa Tests and CogAT

 

Special Interest Topics

 

Special Interest Topic 12.1: Standards-Based Assessment

Special Interest Topic 12.2: Finding Information on Standardized Tests

Special Interest Topic 12.3: Mismatch Between State Standards and Off-The-Shelf

Achievement Tests

Special Interest Topic 12.4: American Educational Research Association (AERA)

Position Statement on High-Stakes Testing.

Special Interest Topic 12.5: Why Standardized Tests Should Not be Used to Evaluate

Educational Quality

Special Interest Topic 12.6: Value-Added Assessment: A New Approach to

Educational Accountability

Special Interest Topic 12.7: Deciding Not to Test an Upset Student.

 

Chapter 13: The Use of Aptitude Tests in the Schools

I. Introduction

 

II. A Brief History of Intelligence Tests

 

III. The Use of Aptitude and Intelligence Tests in the Schools

            A. Aptitude-Achievement Discrepancies

 

IV. Aptitude and Intelligence Tests

            A. Group Aptitude/Intelligence Tests

            B. Individual Aptitude/Intelligence Tests

            C. Selecting Aptitude/Intelligence Tests

 

V. College Admission Tests

 

VI. Summary

Tables

 

Table 13.1: Organization of Major Group Intelligence/Aptitude Tests

Table 13.2: Organization of the Wechsler Intelligence Scale for Children - 4th Edition (WISC-IV)

Table 13.3: Organization of the Stanford-Binet Intelligence Scales, 5th Edition (SB5)

Table 13.4: Organization of Woodcock-Johnson III (WJ-III) Tests of Cognitive Abilities

Table 13.5: Organization of the Reynolds Intellectual Assessment Scales (RIAS)

Table 13.6: Organization of the Universal Nonverbal Intelligence Test (UNIT)

 

Figures

 

Figure 13.1: A Continuum of General Abilities

Figure 13.2: InView Sample Report - Home Report Form

Figure 13.3: Profile Narrative for the Cognitive Abilities Test (CogAT)

Figure 13.4: Combined Profile Narrative for the Iowa Tests of Basic Skills (ITBS)        and the Cognitive Abilities Test (CogAT)

Figure 13.5: Achievement/Ability Graphic Comparison of the Iowa Tests of Basic         Skills (ITBS) and the Cognitive Abilities Test (CogAT)

 

Special Interest Topics

 

Special Interest Topic 13.1:The Controversial IQ: Knows & Unknowns

Special Interest Topic 13.2: The Controversial IQ: IQ Tests & Schools

Special Interest Topic 13.3:Ability Profiles on the CogAT

 

Chapter 14: Assessment of Behavior and Personality

I. Assessing Behavior and Personality

            A. Response Sets

            B. Assessment of Behavior and Personality in the Schools

 

II. Behavior Rating Scales

A. Behavior Assessment System for Children - TRS & PRS

B.  Conners Rating Scales - Revised

C. Child Behavior Checklist & Teacher Report Form

 

III. Self-Report Measures

A. Behavior Assessment System for Children - Self-Report of Personality

B. Youth Self-Report

 

IV. Projective Techniques

            A. Projective Drawings

            B. Sentence Completion Tests

            C. Apperception Tests

            D. Inkblot Tests

           

V. Summary

Tables

Table 14.1: Composites and Scales in the TRS and PRS

Table 14.2: Composites and Scales in the SRP

Table 14.3: The Projective Debate

 

Figures

Figure 14.1: Completed Clinical and Adaptive Profile sections of a TRS

Figure 14.2: Completed Clinical and Adaptive Profile sections of a SRP

Figure 14.3: A Picture Similar to Those Used on Apperception Tests

Figure 14.4: An Inkblot Similar to Those Used on Inkblot Tests

 

Special Interest Topics

Special Interest Topic 14:1: The Handsome and Deformed Leg

Special Interest Topic 14.2: An Example of a "Fake Good" Response Set

Special Interest Topic 14.3: Popular Tests of Behavior & Personality

 

Chapter 15: Assessment Accommodations for Students with Disabilities

 

I. Introduction

 

II. Major Legislation that Impact the Assessment of Students with Disabilities

 

III. Individuals with Disabilities Education Act (IDEA)

            A. IDEA Categories of Disabilities

 

IV. Section 504

 

V. The Rationale for Accommodations

 

VI. When are accommodations not appropriate or necessary?

 

VII. Strategies for Accommodations

            A. Modifications of Presentation Format

            B. Modifications of Response Format

            C. Modifications of Timing

            D. Modification of Setting

            E. Adaptive Devices and Supports

            F. Using Only Portion of a Test

            G. Using Alternate Assessments

 

VIII. Determining What Accommodations to Provide

 

IX. Reporting Results of Modified Assessments

 

X. Summary

Tables

 

Table 15.1: Accommodations Involving Modifications of the Presentation Format

Table 15.2: Accommodations Involving Modifications of the Response Format

Table 15.3: Accommodations Involving Modifications of Timing

Table 15.4: Accommodations Involving Modifications of Setting

Table 15.5: Accommodations Involving Adaptive Devices and Supports

Table 15.6: Determining which Accommodations to Provide

 

List of Special Interest Topics

 

Special Interest Topic 15.1: Differential Effects & Assessment Accommodations?

Special Interest Topic 15.2: Allowable Accommodations in a Statewide Assessment Program

Special Interest Topic 15.3: Assessment of Students with Disabilities — Selected Legal Issues

 

Chapter 16: The Problem of Bias in Educational Assessment

I. Introduction

II. What Do We Mean By Bias?

III. Past and Present Concerns: A Brief Look

IV. The Controversy over Bias in Testing: Its Origin, What it is, and What it is Not

V. Cultural Bias and the Nature of Psychological Testing

VI. Objections to the Use of Educational Tests with Minorities

            A. Inappropriate content

            B. Inappropriate standardization samples.

            C. Examiner and language bias.

            D. Inequitable social consequences.

            E. Measurement of different constructs.

            F. Differential predictive validity.

            G. Qualitatively distinct aptitude and personality.

 

VII. The Problem of Definition in the Study of Bias

VIII. Culture Free Tests, Cultural Loading, and Cultural Bias

IX. Inappropriate Indicators of Bias: Mean Differences & Equivalent Distributions

X. Bias in Test Content

XII. Bias in Other Internal Features of Tests

XIII. Bias in Predictive and in Relation to Variables External to the Test

XV. Conclusion

List of Figures

Figure 16.1: Equal Slopes and Intercepts

Figure 16.2: Equal Slopes with Differing Intercepts

Figure 16.3: Equal Intercepts and Differing Slopes

Figure 16.4: Differing Slopes and Intercepts

 

Special Interest Topics

Special Interest Topic 16.1: Sex Differences in Intelligence

Special Interest Topic 16.2: Courtroom Controversy Over IQ Testing in the Public Schools

Special Interest Topic 16.3: Fairness & Bias — A Complex Relationship

 

Chapter 17: Best Practices in Educational Assessment

 

I. Introduction

II. Guidelines for Developing Assessments

III. Guidelines for Published Assessments

IV. Guidelines for Administering Assessments

V. Guidelines for Scoring Assessments

VI. Guidelines for Interpreting, Using, and Communicating Assessment Results

VII. Responsibilities of Test Takers

VIII. Summary

 

List of Tables

 

Table 17.1: Checklist for Developing Assessments

Table 17.2: Checklist for Selecting Published Assessments

Table 17.3: Checklist for Administering Assessments

Table 17.4 Checklist for Scoring Assessments

Table 17.5: Checklist for Interpreting, Using, and Communicating Assessment Results

Table 17.6: Responsibilities of Test Takers

 

List of Special Interest Topics

 

Special Interest Topic 17.1: Teachers Cheating?

Special Interest Topic 17.2: Steps to Prevent Student Cheating.


Appendices

Need help? Get in touch

Video
Play
Privacy and cookies
By watching, you agree Pearson can share your viewership data for marketing and analytics for one year, revocable upon changing cookie preferences. Disabling cookies may affect video functionality. More info...

Pearson eTextbook: What’s on the inside just might surprise you

They say you can’t judge a book by its cover. It’s the same with your students. Meet each one right where they are with an engaging, interactive, personalized learning experience that goes beyond the textbook to fit any schedule, any budget, and any lifestyle.