
- •Contents
- •Series Preface
- •Acknowledgments
- •Purposes and Uses of Achievement Tests
- •Diagnosing Achievement
- •Identifying Processes
- •Analyzing Errors
- •Making Placement Decisions and Planning Programs
- •Measuring Academic Progress
- •Evaluating Interventions or Programs
- •Conducting Research
- •Screening
- •Selecting an Achievement Test
- •Administering Standardized Achievement Tests
- •Testing Environment
- •Establishing Rapport
- •History and Development
- •Changes From KTEA-II to KTEA-3
- •Subtests
- •Mapping KTEA-3 to Common Core State Standards
- •Standardization and Psychometric Properties of the KTEA-3
- •Standardization
- •Reliability
- •Validity
- •Overview of the KTEA-3 Brief Form
- •Brief Form Standardization and Technical Characteristics
- •How to Administer the KTEA-3
- •Starting and Discontinuing Subtests
- •Sample, Teaching, and Practice Items
- •Recording Responses
- •Timing
- •Queries and Prompts
- •Subtest-by-Subtest Notes on Administration
- •How to Score the KTEA-3
- •Types of Scores
- •Subtest-by-Subtest Scoring Keys
- •How to Interpret the KTEA-3
- •Introduction to Interpretation
- •Step 1: Interpret the Academic Skills Battery (ASB) Composite
- •Step 2: Interpret Other Composite Scores and Subtest Scores
- •Subtest Floors and Ceilings
- •Interpretation of Composites
- •Clinical Analysis of Errors
- •Qualitative Observations
- •Using the KTEA-3 Across Multiple Administrations
- •Repeated Administrations of the Same Form
- •Administering Alternate Forms
- •Using the KTEA-3 Brief Form
- •Progress Monitoring
- •Screening for a Comprehensive Evaluation
- •KTEA-3 Score Reports
- •History and Development
- •Changes From WIAT-II to WIAT-III
- •Age Range
- •New and Modified Subtests
- •Composites
- •Administration and Scoring Rules
- •Skills Analysis
- •Intervention Goal Statements
- •New Analyses
- •New Scores
- •Validity Studies
- •Materials
- •Scoring and Reporting
- •Description of the WIAT-III
- •Subtests With Component Scores
- •Mapping WIAT-III to Common Core State Standards
- •Standardization and Psychometric Properties of the WIAT-III
- •Standardization
- •Reliability
- •Validity
- •Starting and Discontinuing Subtests
- •Sample, Teaching, and Practice Items
- •Recording Responses
- •Timing
- •Queries and Prompts
- •Subtest-by-Subtest Notes on Administration
- •How to Score the WIAT-III
- •Types of Scores
- •Score Reports
- •Subtest-by-Subtest Scoring Keys
- •Listening Comprehension
- •Early Reading Skills
- •Reading Comprehension
- •Sentence Composition
- •Word Reading and Pseudoword Decoding
- •Essay Composition
- •Numerical Operations
- •Oral Expression
- •Oral Reading Fluency
- •Spelling
- •Math Fluency—Addition, Subtraction, and Multiplication
- •Introduction to Interpretation
- •Step 1: Interpret the Composite Scores
- •Subtest Floors and Ceilings
- •Skills Analysis
- •Intervention Goal Statements
- •Qualitative Data
- •Using the WIAT-III Across Multiple Administrations
- •Linking Studies
- •Overview of the WISC-V, WISC-V Integrated, and KABC-II
- •Qualitative/Behavioral Analyses of Assessment Results
- •Identification of Specific Learning Disabilities
- •Interpretation and Use of Three New Composite Scores
- •Accommodations for Visual, Hearing, and Motor Impairments
- •Ongoing Research on Gender Differences in Writing and the Utility of Error Analysis
- •Female Advantage in Writing on KTEA-II Brief and Comprehensive Forms
- •Strengths and Weaknesses of the KTEA-3
- •Assets of the KTEA-3
- •Test Development
- •Two Forms
- •Standardization
- •Reliability and Validity
- •Administration and Scoring
- •Interpretation
- •Phonological Processing
- •KTEA-3 Flash Drive
- •Limitations of the KTEA-3
- •Test Development
- •Standardization
- •Reliability and Validity
- •Administration and Scoring
- •Test Items
- •Interpretation
- •Final Comment
- •Strengths and Weaknesses of the WIAT-III
- •Assets of the WIAT-III
- •Test Development
- •Normative Sample
- •Reliability and Validity
- •Administration and Scoring
- •Interpretation
- •Better Listening Comprehension Measure
- •Technical Manual
- •Limitations of the WIAT-III
- •Floor and Ceiling
- •Test Coverage
- •Poor Instructions for Scoring Certain Tasks
- •Item Scoring
- •Audio Recorder
- •Final Comment
- •Content Coverage of the KTEA-3 and WIAT-III
- •Case Report 1: Jenna
- •Reason for Evaluation
- •Background Information
- •Behavioral Observations
- •Assessment Procedures and Tests Administered
- •Test Results
- •Neuropsychological Implications and Diagnostic Impressions
- •Recommendations
- •Psychometric Summary for Jenna
- •Case Report 2: Oscar
- •Reason for Evaluation
- •Background Information
- •Behavioral Observations
- •Assessment Procedures and Tests Administered
- •Test Results
- •Diagnostic Summary
- •Recommendations
- •Resources
- •Psychometric Summary for Oscar
- •Case Report 3: Rob
- •Purpose of the Evaluation
- •History and Background
- •Behavioral Observations
- •Assessment Procedures and Tests Administered
- •Results
- •Summary and Diagnostic Impressions
- •Recommendations
- •Psychometric Summary for Rob
- •Q-interactive Versus Q-global
- •Equivalency Studies
- •Essential Features of Q-interactive
- •Key Terminology
- •Central Website
- •Assess Application
- •References
- •Annotated Bibliography
- •About the Authors
- •About the Digital Resources
- •Index

290 ESSENTIALS OF KTEA™-3 AND WIAT®-III ASSESSMENT
STRENGTHS AND WEAKNESSES OF THE KTEA-3
We previously reviewed (Dumont & Willis, 2010) the KTEA-II (Kaufman & Kaufman, 2004a) and concluded, “Although we have identified some concerns with the KTEA-II, we consider it to be one of the best currently available comprehensive achievement batteries” (p. 265). We find that the KTEA-3 has maintained the strengths of the KTEA-II and o ers significant improvements. The KTEA-3 o ers an e cient, thorough, practical, and statistically reliable and valid assessment of academic abilities. It is designed with features that enhance useful interpretation. The strengths and weaknesses we perceive in the KTEA-3 are listed in Rapid Reference 5.1.
Rapid Reference 5.1
.....................................................................................................................
Strengths and Weaknesses of the KTEA-3
Strengths |
Weaknesses |
|
Test Development |
|
|
•Item selection using curriculum consultants and content of widely used textbooks.
•Almost all Reading Comprehension questions are appropriately passage-dependent (see Keenan & Betjemann, 2006).
•Subtests in all eight specific learning disability areas identified in the Individuals with Disabilities Education Improvement Act (IDEA, 2004).
•Subtests in each of the three DSM-5 areas of impairment (Reading, Written Expression, and Math).
•Extensive tryout data using samples with approximate proportional representation by gender and ethnicity.
•Easel format for presenting subtest items.
•Reading-related subtests contribute to eight composites and yield five reading-related composites.
•Limited ceiling for some subtests for the oldest students in the Above Aver-
22222222111111111
age range. See Table 5.2 for specific data on highest possible subtest and composite scores for ages 21–25 and grade 12.
•Although most subtests have a ceiling that is at least 2 SD above the mean, at Grade 12, the ceilings for Nonsense Word Decoding, Reading Comprehension, Listening Comprehension, and Letter Naming Facility range from 124 to 129. At the highest age range, 12 of the 19 subtests have ceilings ≤150.
•Limited floor for some subtests for the youngest students with the lowest ability levels. For students younger than age 6, raw scores of 0 correspond to standard scores ranging from 78 to 40. For Pre-K through grade 1, raw scores of 0 correspond to standard scores ranging from 73 to 40. See Table 5.3 for specific data.
(continued)

STRENGTHS AND WEAKNESSES OF THE KTEA™-3 AND WIAT®-III 291
(Continued)
Strengths |
Weaknesses |
Test Development
•Innovative item types, especially in Reading Comprehension, Written Expression, and Phonological Processing.
•Use of extensive item analysis procedures to eliminate biased items (gender or race) and items with poor psychometric properties.
•Expanded error analysis procedures with item-level analyses for Reading Comprehension, Listening Comprehension, Written Expression, Oral Expression, Phonological Processing, Math Concepts and Applications, Math Computation, Letter and Word Recognition, Nonsense Word Decoding, and Spelling.
•Novel and stimulating artwork is used throughout the test.
•It is not until age 11:8 that all subtests (with the exception of Nonsense Word Decoding) have a complete floor (standard score 40) and ceiling (standard score of 160). Nonsense Word Decoding does not have both a complete floor and ceiling at any age.
•Although Form A and Form B generally produce similar standard scores by age or grade, some large differences do occur. For Spelling, at 5:0–5:2, a raw score of 0 corresponds to standard scores of 55 for Form A and 80 for Form B.
•Error analysis procedures may be
22222222 |
difficult for examiners lacking any back- |
111111111 |
|
|
ground in curriculum. In rare instances, |
|
examiner knowledge and judgment |
|
are required to override potentially |
|
misleading error-analysis scores. |
Standardization
• The standardization sample is well |
• The grade-norm sample has smaller |
|
stratified to match the U.S. popu- |
upper-grade norm samples than for |
|
lation and included 48 states (no |
the younger-grade samples. For grades |
|
examiners reported for Alaska or |
Pre-K–8 the sample size is 100 (for |
|
Maine). |
each form), while for grades 9–12 the |
|
• Stratification included age or grade, |
sample size is 75 (for each form). The |
|
18-year-old sample is N = 100, age 19 |
||
gender, ethnicity geographic region, |
||
special education group, and parent’s |
sample is N = 80, and N = 125 for the |
|
education level. |
sample ages 20–22 and 125 for the |
|
ages 23–25 sample. |
||
• Standardizations were conducted, |
||
• The age-norm sample has smaller adult |
||
one in the fall and one in the spring, |
||
yielding a large normative group |
norm samples than for the younger age |
|
samples. For ages 4–18 the samples |
||
(N = 2,050 for the age-norm sample |
||
and N = 2,600 for the grade-norm |
range in size from 100 to 160, while for |
|
sample). |
19–20 and 21–25 the sample sizes are |
|
75 each. |
||
|
||
|
|
|
|
(continued) |

292 ESSENTIALS OF KTEA™-3 AND WIAT®-III ASSESSMENT
(Continued)
Strengths |
Weaknesses |
Standardization
•The standardization sample included students receiving special education services.
•Very detailed information on standardization procedures and sample are included in the Technical & Interpretive Manual.
•Expanded age range down to 4 years 0 months and up to 25 years 11 months for both Forms A and B.
•There are no separate norms for postsecondary students. Percentages of 19–20-, and 21–25-year-olds in high school or dropped out, graduated from high school with no further education, began 2-year postsecondary program, and began 4-year postsecondary program roughly approximated U.S. Census data
in most cases, but the examiner still lacks norms specifically
for students enrolled in postsecondary programs.
Reliability and Validity
•Good to excellent mean split-half reliabilities of subtests and composites (range from .81 to .98, with the exception of Written and Oral Fluency which range from
.62 to .76).
•Two alternate forms are available for the KTEA-3, resulting in reduced practice effects for children who are assessed more than once.
•Alternate form reliability coefficients (adjusted for the variability of the norm group) are generally strong, ranging from .69 to .96 for the composites. Oral Fluency is the weakest, ranging from .69 to .74.
•The Technical & Interpretive Manual offers correlational data with two other tests of academic achievement and the KTEA-II as well as with a test of language (CELF-IV).
• The Oral Language subtests have
variable correlations with similar
22222222111111111
measures on the other achievement instruments. However this may be due to the very different approaches taken by different test batteries in measuring and defining oral expression and listening comprehension.
•Standard Errors of Measurement (SEm) are high for the ageand grade-norm sample for the Oral Fluency composite (ranging from 7.58–8.28) and several other subtests: Writing Fluency mean = 7.31, Associational Fluency mean = 9.28, Object Naming Facility mean = 8.31, and Letter Naming Facility
mean = 8.81.
(continued)

STRENGTHS AND WEAKNESSES OF THE KTEA™-3 AND WIAT®-III 293
(Continued)
Strengths |
Weaknesses |
Reliability and Validity
•Moderate to high correlations between most KTEA-3 and other measures of achievement for concurrent validity (.86–.90).
•Strong face validity based on use of textbook content and skill development in reading and mathematics.
•Construct validity indicated by increasing scores across grades and ages.
Administration and Scoring
• The test protocol has a user-friendly |
• Test protocol does not include a |
|
layout. |
summary page for recording scores. |
|
• Easel format simplifies administration. |
If hand-scoring the test, these pages |
|
are available for printing from the |
||
• Starting points, basal and discontinue |
||
flash drive. |
||
rules are clearly labeled on the record |
||
22222222 |
||
|
111111111 |
|
form and easels for all subtests. |
• Need to write responses phoneti- |
|
• Although no longer provided or |
cally to use error analysis for Letter & |
|
Word Recognition and Nonsense |
||
required, young children often find the |
||
Word Decoding, which can be espe- |
||
use of a puppet (Pepper) very engaging |
||
cially challenging when testing stu- |
||
in Phonological Awareness. |
||
dents who respond very quickly. |
||
• Students find the storybook format of |
||
• Hand scoring is particularly time con- |
||
Written Expression much less tedious |
||
suming. Computation forms must |
||
than a more traditionally formatted |
||
be downloaded and printed. Norms |
||
test of written expression. |
||
tables for converting raw scores to |
||
• Most test items are scored in a |
||
standard scores may be difficult to |
||
dichotomous manner. |
read and navigate. |
|
• Both age and grade norms are |
• Specific error analysis norms tables |
|
available. |
are difficult to locate quickly in |
|
• An entire 200+ page manual provides |
the PDF Manual. Some examin- |
|
many scoring examples for Reading |
ers are tempted to interpret raw |
|
and Listening Comprehension, Asso- |
error-category scores instead of |
|
ciational and Writing Fluency, and |
using norms, which can yield mis- |
|
Written and Oral Expression. |
leading results. |
|
|
|
|
|
(continued) |

294 ESSENTIALS OF KTEA™-3 AND WIAT®-III ASSESSMENT
(Continued)
Strengths |
Weaknesses |
Administration and Scoring
•Table for error analysis categories is provided on the record form for five subtests.
•Separate error analysis pages are provided for subtests that have large numbers of skill categories, thereby making the main record form less cumbersome.
•New choices for descriptive categories allows examiners to base ability levels on either a 10or 15-point scale.
•The record form for Letter & Word Recognition and Nonsense Word Decoding includes phonetic pronunciation keys and, when needed, indicates correct stress of syllables. There is ample space to record the examinee’s pronunciations.
•The flash drive provides audio files for instructing the examiner about the correct pronunciations for administration and scoring purposes.
•Even though they lack sufficient top, norms for Phonological Processing, Letter & Word Recognition, Nonsense Word Decoding, and Reading Comprehension above age 12 and grade 6 may be helpful with weak readers.
•Need to remember to use conversion table to convert points earned in a set to the raw score for Reading Comprehension, Written Expression, Listening Comprehension, Word Recognition Fluency, and Oral Expression. Conversion tables are in the PDF Manual.
•Although scoring of most of
the Oral Expression and Written Expression items seems clear to us (especially with the aid of the explanations, examples, and Glossary), questions from graduate students, colleagues, and workshop attendees suggest that the scoring can be confus-
22222222111111111 ing for some examiners.
•When Associational Fluency and Object Naming Facility are administered, they must be given out of easel order, which is easy to forget for some examiners.
•Subtests required for computing Core Composites and the Academic Skills Battery are not contained in a single easel. Only Reading and Math composites can be obtained using the materials in Stimulus Book 1.
•The undifferentiated “Average” range for the 15-point score classifi-
cation system extends from 85 to 115. This range may make sense statistically and psychologically, but we find it much too broad (more than two thirds of the population) for interpreting educational performance.
(continued)

STRENGTHS AND WEAKNESSES OF THE KTEA™-3 AND WIAT®-III 295
(Continued)
Strengths |
Weaknesses |
Interpretation
•Both composites and subtests can be interpreted due to high reliability coefficients.
•Error analysis procedures pinpoint specific skill deficits.
•Qualitative observations for 15 KTEA-3 subtests, identified by both IDEA domain as well as by potential cognitive processing weaknesses, are especially helpful.
•The composites and error analysis yield much diagnostic information directly relevant to instructional remediation.
•Subtest and composite intercorrelations are presented by grade (Pre-K–12) and age (17–18, 19–20, 21–25) in the KTEA-3 Technical & Interpretive Manual.
•Interpretive procedures were simplified by comparing composites and subtests to ASB rather than to the mean of all composites.
•Integrated normative and ipsative comparisons are included in the interpretive process.
•Pairs of KTEA-3 Comprehensive subtests (i.e., Reading Comprehension and Listening Comprehension, and Written Expression and Oral Expression) have similar formats and allow for useful comparisons that can help the examiner distinguish specific problems in reading or writing from more general language problems.
•Caution needs to be taken when interpreting the scores with a nonoptimal ceiling or a nonoptimal floor for children under age 6.
•If all of the ASB subtests are not administered, then the Composite and Subtest Analysis cannot be completed because the ASB is used in all of these comparisons. However, pairwise composite comparisons or pairwise subtest comparisons are still possible.
•Many evaluators want to know if the component subtests in each composite are significantly and uncommonly different from each other, which should require cautious interpretation of the composite score. This information is available in Appendix G of the Manual, but is buried among all the other subtest comparisons.
•The comparison between Reading Comprehension and Listening Comprehension is very useful because of the almost identical formats of the two subtests. Differences between the formats of Written Expression and Oral Expression make that comparison less persuasive.
•Insufficient information is provided in the
Technical & Interpretive Manual on how to utilize the Growth Scale Values (GSV).
•The Growth Scale Values (GSV) allow for comparisons between an individual’s change across time but do not provide statistical comparisons to expected growth (“closing the gap”).
(continued)