Assessment Glossary

Alignment – A logical connection between the curriculum and the expected outcomes.
Analytic Scoring – Scoring that divides the student work into elemental, logical parts or basic principles. Scorers evaluate student work across multiple dimensions of performance rather than from an overall impression (holistic scoring). In analytic scoring, individual scores for each dimension are determined and reported; however, an overall impression of quality may be included. (P.A. Gantt; CRESST Glossary)
Anchor(s) – A sample of student work that exemplifies a specific level of performance. Raters use anchors to score student work, usually comparing the student performance to the anchor.
Assessment – focused on measuring a performance, work product, or skill in order to offer feedback to document strengths and growth and to provide directives for improving future performance. Assessments are nonjudgmental and are designed and intended to be helpful to produce improvement.
Assessment for Accountability – The assessment of some unit (could be a program, department, or entire institution) conducted to satisfy external stakeholders. Results are summative and often compared across units. (Leskes, A., 2002)
Assessment for Improvement – Assessment that feeds directly back into revising the course, program, or institution to improve student learning results. (Leskes, A., 2002)
Assessment Plan – A document that outlines the program mission/goals, desired student learning outcomes (or objectives), learning processes (e.g., courses, activities, assignments) that contribute to students’ abilities reach the program’s outcomes (this may be shown in the form of a curriculum map), long-range timeline, and location of the mission/goals and student learning outcomes (e.g., web site, brochure, advising session). (adapted from the Northern Illinois University Assessment Glossa
Assessment Plan (specific) – A plan for a specific assessment activity/project will include the purpose or goal of particular assessment activities, how the results will be used and who will use them, brief explanation of data-collection methods and the analysis methods, an indication of which outcome(s)/objective(s) is/are addressed by each method, the intervals at which evidence is collected and reviewed, the individual(s) responsible for the collection/review of evidence and dissemination of assessment results. (adapted from the Northern Illinois University Assessment Glossary)
Authentic Assessment – Determining the level of student knowledge/skill in a particular area by evaluating his/her ability to perform a “real world” task in the way professionals in the field would perform it. Authentic assessment asks for a demonstration of the behavior the learning is intended to produce.
Benchmark – A point of reference for measurement; a standard of achievement against which to evaluate or judge performance.
Capstone Course/Experience – An upper-division class designed to help students demonstrate comprehensive learning in the major through some type of product or experience. In addition to emphasizing work related to the major, capstone experiences can require students to demonstrate how well they have mastered important learning objectives from the institution’s general studies programs. (Palomba & Banta, 1999)
Competency – The demonstration of the ability to perform a specific task or achieve specified criteria. (James Madison University Dictionary of Student Outcomes Assessment)
Course Assessment – Assessment to determine the extent to which a specific course is achieving its learning outcomes.
Criteria for Success – The minimum requirements for a program to declare itself successful. Criterion-referenced – Assessment where student performance is compared to a pre-established performance standard (and not to the performance of other students). (CRESST Glossary)
Curriculum Map – A matrix showing the coverage of each program learning outcome in each course.
Direct Measures– Direct assessment measures require individuals or entities to demonstrate their knowledge, skills, and/or a behavior that reflects achievement of the stated goal.
Embedded Assessment – Collecting data/evidence on program learning outcomes by extracting course assignments. It is a means of gathering information about student learning that is built into and a natural part of the teaching-learning process. The instructor evaluates the assignment for individual student grading purposes; the program evaluates the assignment for program assessment. When used for program assessment, typically someone other than the course instructor uses a rubric to evaluate the assignment. (Leskes, A., 2002)
Embedded Exams and Quizzes – Collecting data/evidence on program learning outcomes by extracting a course exam or quiz. Typically, the instructor evaluates the exam/quiz for individual student grading purposes; the program evaluates the exam/quiz for program assessment. Often only a section of the exam or quiz is analyzed and used for program assessment purposes. See also: Embedded Assessment.
Evaluation – focused on making a judgment or determination concerning the quality of a performance, work product or use of skills against a set of standards. Evaluations are designed for and intended to document the level of achievement that has been attained.
Focus Group – A qualitative data-collection method that relies on facilitated discussions, with 3-10 participants who are asked a series of carefully constructed open-ended questions about their attitudes, beliefs, and experiences. Focus groups are typically considered an indirect data-collection method.
Formative Assessment – Ongoing assessment that takes place during the learning process. It is intended to improve an individual student’s performance, program performance, or overall institutional effectiveness. Formative assessment is used internally, primarily by those responsible for teaching a course or developing and running a program. (Middle States Commission on Higher Education, 2007
Goals – General expectations for students. Effective goals are broadly stated, meaningful, achievable, and assessable.
Grading – The process of evaluating students, ranking them, and distributing each student’s value across a scale. Typically, grading is done at the course level.
High Stakes Assessment – Any assessment whose results have important consequences for students, teachers, programs, etc. For example, using results of assessment to determine whether a student should receive certification, graduate, or move on to the next level. Most often the instrument is externally developed, based on set standards, carried out in a secure testing situation, and administered at a single point in time. (Leskes, A., 2002)
Holistic Scoring – Scoring that emphasizes the importance of the whole and the interdependence of parts. Scorers give a single score based on an overall appraisal of a student’s entire product or performance. Used in situations where the demonstration of learning is considered to be more than the sum of its parts and so the complete final product or performance is evaluated as a whole. (P. A. Gantt)
Indirect Measures – Indirect assessment measures rely on individual perceptions to determine if a goal/objective has been achieved.
Learning outcomes – Statements that identify the knowledge, skills, or attitudes that students will be able to demonstrate, represent, or produce as a result of a given educational experience. There are three levels of learning outcomes: course, program, and institution.
Objective – Clear, concise statements that describe how students can demonstrate their mastery of program goals. (Allen, M., 2008) Note: on the Mānoa Assessment web site, “objective” and “outcome” are used interchangeably.
Outcomes – Clear, concise statements that describe how students can demonstrate their mastery of program goals. (Allen, M., 2008) Note: on the Mānoa Assessment web site, “objective” and “outcome” are used interchangeably.
Performance Assessment – The process of using student activities or products, as opposed to tests or surveys, to evaluate students’ knowledge, skills, and development. As part of this process, the performances generated by students are usually rated or scored by faculty or other qualified observers who also provide feedback to students. Performance assessment is described as “authentic” if it is based on examining genuine or real examples of students’ work that closely reflects how professionals in the field go about the task. (Palomba & Banta, 1999)
Portfolio – A type of performance assessment in which students’ work is systematically collected and carefully reviewed for evidence of learning. In addition to examples of their work, most portfolios include reflective statements prepared by students. Portfolios are assessed for evidence of student achievement with respect to established student learning outcomes and standards. (Palomba & Banta, 1999)
Program Assessment – An on-going process designed to monitor and improve student learning. Faculty: a) develop explicit statements of what students should learn (i.e., student learning outcomes); b) verify that the program is designed to foster this learning (alignment); c) collect data/evidence that indicate student attainment (assessment results); d) use these data to improve student learning (close the loop). (Allen, M., 2008)
Reliability – In the broadest sense, reliability speaks to the quality of the data collection and analysis. It may refer to the level of consistency with which observers/judges assign scores or categorize observations. In psychometrics and testing, it is a mathematical calculation of consistency, stability, and dependability for a set of measurements.
Rubric– A tool often shaped like a matrix, with criteria on one side and levels of achievement across the top used to score products or performances. Rubrics describe the characteristics of different levels of performance, often from exemplary to unacceptable. The criteria are ideally explicit, objective, and consistent with expectations for student performance. Rubrics may be used by an individual or multiple raters to judge student work. When used by multiple raters, norming takes place before scoring begins. Rubrics are meaningful and useful when shared with students before their work is judged so they better understand the expectations for their performance. Rubrics are most effective when coupled with benchmark student work or anchors to illustrate how the rubric is applied.
Student Learning Outcome – A statement of what students will be able to think, know, do, or feel because of a given educational experience.
Summative Assessment – The gathering of information at the conclusion of a course, program, or undergraduate/graduate career to improve learning or to meet accountability demands. The purposes are to determine whether or not overall goals have been achieved and to provide information on performance for an individual student or statistics about a course or program for internal or external accountability purposes. Grades are the most common form of summative assessment. (Middle States Commission on Higher Education, 2007)
Validity – Refers to whether the interpretation and intended use of assessment results are logical and supported by theory and evidence. In addition, it refers to whether the anticipated and unanticipated consequences of the interpretation and intended use of assessment results have been taken into consideration. (Standards for Educational and Psychological Testing, 1999)
Value-Added Assessment – Determining the impact or increase in learning that participating in higher education had on students during their programs of study. The focus can be on the individual student or a cohort of students. (Leskes, A., 2002). A value-added assessment plan is designed so it can reveal “value”: at a minimum, students need to be assessed at the beginning and the ending of the course/program/degree.