Why Do Assessment?

Total Page:16

File Type:pdf, Size:1020Kb

Why Do Assessment?

Handbook for the Assessment of Student Learning STATE UNIVERSITY OF NEW YORK AT DELHI JUNE 2011 2

What is assessment?

Academic assessment is the process of gathering data to improve teaching and learning. This begins with forming clear, measurable statements describing the expected outcomes of student learning, coupled with learning opportunities that provide students with the ability to achieve those outcomes. To measure learning, institutions develop measurement tools that are aligned with the student learning outcomes, systematically gathering data based on the measurement tool. The results can then be interpreted to determine how well student learning matches our curricular expectations, and help to direct change. After implementing change, the process begins again and continues cyclically so that teaching and learning continually improve.

Exams, labs, projects, papers, etc.

Rubrics Standards & benchmarks

Provide learning opportunities

Change outcomes, change assessment tools, change goals, change rubrics

Why do assessment?

SUNY Delhi faculty is comprised dedicated teachers who want their students to learn. The assessment process provides feedback on the teaching and learning process while opening up lines of communication between faculty and staff. With data, faculty and staff can determine what is and is not working in order to make informed decisions about changes to their courses and to the programs in order to expend resources wisely. For students, the assessment process helps them learn more effectively by providing clear expectations for what they should know and learn. 3

What are the components of assessment?

 Program Goals  Student Learning Outcomes  Curriculum Mapping  Assessment Tools  Rubrics  Setting Standards or Benchmarks  Gathering Evidence  Closing the Loop

Goals

What you, your program, or your college want to achieve. For example, SUNY Delhi wants its transfer students to succeed at their four-year institution of choice; the culinary department wants to offer a quality, state-of-the-art educational program; Professor Burger wants her students to appreciate Shakespeare.

Student Learning Outcomes

Student learning outcomes are goals describing what students take with them from a program or a course. Learning outcomes typically explain not only what students will know or be able to do, but why. Saying that students in your public speaking course will submit an outline for their introductory speech may be a course objective or course requirement; but the learning outcome is that students understand the importance of organization when speaking in public. (So all student learning outcomes are goals—but not all goals are student learning outcomes.)

Not all learning outcomes are course-based. Some are programmatic—and, in theory, might appear in more than one course outline in a program. In theory, all of a program’s student learning outcomes should appear at least once in the aggregate course outlines/syllabi for required courses in a program.

What are the difference between program student learning outcomes and course student learning outcomes? Program student learning outcomes are the overarching learning that occurs in multiple courses across the curriculum. They are broad descriptions of what the student should know, how they should think about the discipline, or what the student should be able to do when they finish the program. In writing these, one should also consider the institutional goals and how the program student learning outcomes support the institutional goals. As with all student learning outcomes, they should be written in terms that are measurable.

In contrast, course student learning outcomes are more specific and describe what the student should know, think or be able to do when they finish the course. They should be more detailed than program student learning outcomes and should be related to the course content, tests and the other graded work of the course. 4

Excellent guidelines and examples of writing student learning outcomes may be found at the University of Connecticut Assessment website: http://assessment.uconn.edu/docs/HowToWriteObjectivesOutcomes.pdf

Bloom’s Taxonomy, a multi-tiered classification of thinking and learning skills into six levels of complexity that range from basic levels of knowledge acquisition to more complex levels of creating and synthesizing (Bloom & Krathwohl, 1956). In the 1990’s, Bloom’s Taxonomy was revised to reflect the emphasis on student learning outcomes in that the classifications are action verbs rather than nouns. Bloom’s Taxonomy might prove useful in writing student learning outcomes that include action verbs and encourage higher order thinking and learning.

Bloom’s Taxonomy (New Version)

 Remembering: can the student recall or remember the information? o define, duplicate, list, memorize, recall, repeat, reproduce state  Understanding: can the student explain ideas or concepts? o classify, describe, discuss, explain, identify, locate, recognize, report, select, translate, paraphrase  Applying: can the student use the information in a new way? o choose, demonstrate, dramatize, employ, illustrate, interpret, operate, schedule, sketch, solve, use, write.  Analyzing: can the student distinguish between the different parts? o appraise, compare, contrast, criticize, differentiate, discriminate, distinguish, examine, experiment, question, test.  Evaluating: can the student justify a stand or decision? o appraise, argue, defend, judge, select, support, value, evaluate  Creating: can the student create new product or point of view? o assemble, construct, create, design, develop, formulate, write. 5

Objectives Describe the details of goals or student learning outcomes—or the tasks that must be done to achieve a goal (the means to the end itself—which is the learning outcome). A course objective may be to have students complete 20 hours of community service. The goal isn’t to have students complete the task: the goal is that students understand the value of working as a team and develop a sense of responsibility to their communities.

What are the difference between goals and objectives? Goals state intended outcomes in general terms and objectives state them in specific terms.

What are the difference between objectives and outcomes? Objectives are intended results of instruction while outcomes are the achieved results of what was learned. Objectives are more teacher centered and are written in terms of what the teacher intends to do and the subject matter that the teacher intends to cover. Outcomes are more student centered in that they describe what the student should learn and take away from the course or program.

Curriculum Mapping

Once program level and course level student learning outcomes have been written, a curriculum map can be written to examine how they fit together. Curriculum mapping is the process of linking particular content and skills to particular courses. Creating a curriculum map can determine where learning outcomes overlap, where they are missing or where they need improvement.

In designing a curriculum map, create a grid by writing the broad program student learning outcomes down the left side and courses and internships across the top, then check off the learning outcomes that are addressed in each course.

Course 1 Course 2 Course 3 Course 4 Student Learning x x Outcome 1 Student Learning x x x Outcome 2 Student Learning x Outcome 3

Excellent guidelines and examples of curriculum maps may be found at the University of West Florida Center for Teaching and Learning website http://uwf.edu/cutla/curriculum_maps.cfm.

Assessment Tools

Student learning can be assessed both directly and indirectly. Direct assessment is tangible, self- explanatory evidence of exactly what students have and have not learned and can be measured 6 by tools such as tests, papers, presentations and projects. Indirect assessment is less clear and less convincing than direct assessment in that it is a sign that students are probably learning. Indirect assessment includes measures of student perceptions like surveys (alumni, student, employer) and evaluations.

Questions to ask about Student Learning Outcomes and Assessment Tools:

 Is it possible to assess this outcome?  What types of activities (tests, quizzes, projects, essays) will best assess this outcome?  How important is this outcome? What percentage of the final grade should be based on this outcome?  Are the activities in this course actively measuring the outcomes?  Are the outcomes student-centered or course centered?  Do the questions on my exams directly measure the outcomes?  Do my assignments help student achieve the outcomes?  Is the time that students will spend on the assignment measuring this outcome worth it? (Is the time you’ll spend grading the assignment worth it?)  Are my outcomes too broad? Too specific?  Why am I asking students to do the assignments I am requiring? (The answers to this ought to be the learning outcomes.)

When creating the assessment tool to measure the student learning outcome, it is important to word the question or assignment in a way that clearly indicates the product that should be produced. Without a clearly written question, students may complete the assignment without learning what we want them to learn.

While research papers, essays and tests can be used to assess student learning outcomes, there are many ways that students can demonstrate their learning beyond the typical assignment. The following shows examples of assignments that can be used to address higher order thinking according to Bloom’s Taxonomy. 7 8

Rubrics

A rubric is a scoring guide that seeks to evaluate a student's performance based on the sum of a full range of criteria rather than a single numerical score.

Why Should I Use Rubrics?

 Rubrics improve student performance by clearly showing them how you will evaluate their work and what you expect.  Rubrics help students judge the quality of their own work better.  Rubrics make assessment more objective and consistent—reducing student complaints and arguments about grades.  Rubrics force the teacher to clarify criteria in specific terms, eliminating unintended vague or fuzzy goals.  Rubrics reduce the amount of time you spend evaluating student work.  Rubrics make students more confident in their own peer evaluations.  Rubrics provide useful feedback to you regarding how effective your teaching has been.  Rubrics give students specific, structured feedback about their strengths and weaknesses.  Rubrics are easy to use and easy to explain.

Types of Rubrics

 Rating Scale Rubric Uses checkboxes with a rating scale showing the degree to which required items appear in a completed assignment.  Descriptive Rubric Like rating scale rubrics, except that the checkboxes are replaced by descriptions of what must be done to earn each possible rating.

Steps to Creating an Effective Rubric

 Look for models—someone else may already have designed a rubric that could be modified for your use.  List the things you are looking for in the completed assignment.  Leave wiggle room: some percentage of the assignment should be based on “originality” or “effort,” or you risk seeing neither in the final product.  Create a rating scale with at least three levels.  Label levels with descriptive names, not numbers.  Fill in the boxes (for descriptive rubrics)  Try it out—and modify it if it isn’t giving you the results you want. 9

Example Rubrics

Descriptive Rubric for a Science Research Paper Beginning Developing Accomplished Exemplary Score 1 2 3 4

Introduction Does not give any information about Gives very little Gives too much Presents a what to expect in the report. information. information-- concise lead- more like a in to the summary. report.

Research Does not answer any questions Answers some Answers some Answers most suggested in the template. questions. questions and questions and includes a few includes many other interesting other facts. interesting facts.

Procedure Not sequential, most steps are missing Some of the Most of the steps Presents easy- or are confusing. steps are are to-follow understandable; understandable; steps which most are some lack detail are logical confusing and or are confusing. and lack detail. adequately detailed.

Data & Results Data table and/or graph missing Both complete, Both accurate, Data table and information and are inaccurate. minor some ill-formed graph neatly inaccuracies characters. completed and/or illegible and totally characters. accurate.

Conclusion Presents an illogical explanation for Presents an Presents a logical Presents a findings and does not address any of illogical explanation for logical the questions suggested in the explanation for findings and explanation template. findings and addresses some for findings addresses few of the questions. and addresses questions. most of the questions.

Grammar & Very frequent grammar and/or More than two Only one or two All grammar Spelling spelling errors. errors. errors. and spelling are correct.

Timeliness Report handed in more than one week Up to one week Up to two days Report late. late. late. handed in on time. Total 10

Checklist Rubric for Art Problem Solving Assignment

Not often Usually Always

PROBLEM FINDING (Task definition) The student makes a plan or draws a preliminary sketch. Comments: FACT FINDING (Information seeking and locating resources) The student brainstorms ideas in order to have several solutions from which to choose. Comments: SOLUTION FINDING (Synthesis: putting all the information together) The student is willing to try new things and make changes in his/her art. The student asks questions when he/she does not understand. The student listens to teacher's suggestions for improvement. The student works hard to finish project or task. Comments: EVALUATING The student looks for things he/she can improve. Comments:

Checklist Rubric for Automotive Program (courtesy Steve Tucker) 11

Note to instructor: Prepare the vehicle by setting fan on any speed except high and temp. any place but full cold, A/C control off. (award 1 point for each yes) NO SAFETY GLASSES and/or failure to follow safety rules = minus 5 points. Part 1: Performance Test

The student properly prepared the vehicle for the performance test ___yes ____no (fan on high, full cold, normal A/C)

Installed the thermometer in the correct location ___yes ____no

Correctly interpreted pressure gauge reading on low side ___yes ____no

Correctly interpreted pressure gauge reading on high side ___yes ____no

Correctly interpreted performance test results or faulty readings. ___yes ____no

Part 2: Leak detection

Probe is used in the correct position (below a line) ¼” away ___yes ____no

Circles fittings and connections in lines/hoses ___yes ____no

Moves detector probe at proper rate ( no faster than 1- 2”/second) ___yes ____no

Knows the minimum PSI required for test (minimum of 50) ___yes ____no

Knows how detector will react when a leak exists. ___yes ____no

Overall performance: excellent (all “yes” with a high level of expertise) 5 points satisfactory (2 or less “no” acceptable expertise) 3.5 points more preparation needed (three or more “no”) 2.5 points

Total points______Comments: 12

Checklist Rubric for Automotive Program (courtesy Steve Tucker)

Rating Scale and Recommendation

Automotive Heating and Air Conditioning

Name______Date______

RATING SCALE:

# 1 No knowledge of subject matter, unable to perform minimal amount of task, considerable preparation required.

#2 Inadequate Knowledge/skill. More preparation needed.

# 3 Basic understanding, minimal knowledge & skills, requires close supervision, significant preparation required.

# 4 Moderate understanding & skills, requires some guidance, more preparation required, exceeded allocated time limit.

# 5 Entry level, demonstrates excellent knowledge & skills, requires very little guidance, within allocated time limit.

Based on classroom knowledge, practical and live work performance, for each of the following the student demonstrated:

A/C system Diagnosis and repair: ____5 ____4 ____3 ____2 ____1

Refrigeration System Component Diagnosis and Repair: ____5 ____4 ____3 ____2 ____1

Heating/Ventilation Diagnosis and Repair: ____5 ____4 ____3 ____2 ____1

Cooling System Diagnosis And Repair: ____5 ____4 ____3 ____2 ____1

Control System Diagnosis and Repair: ____5 ____4 ____3 ____2 ____1 13

Refrigerant Recovery, Recycling And Handling: ____5 ____4 ____3 ____2 ____1

Instructor______Date______Rubric Template

Beginning Developing Accomplished Exemplary Score 1 2 3 4

In the shaded areas, list the aspects of the assignment you are looking for. In the boxes to the right, list the descriptions of the performances that merit those ratings. Use as many rows as you need.

More excellent examples of rubrics by subject area may be found at the Teacher Planet website www.rubrics4teachers.com.

Setting Benchmarks or Standards

Before beginning the collection of assessment results, benchmarks or standards should be set. Standards may be local (set by a program at SUNY Delhi), external (set by an external accrediting agency or other organization), or historic (improving over time). To set appropriate benchmarks, research what other programs at other colleges are doing, discuss ideas with faculty members in the program, and use examples of previous student work to inform your decision.

Competencies/Proficiencies Types of student learning outcomes that are focused on skills rather than knowledge. “Students must be able to type 90 words per minute” is a proficiency outcome.

Standards/Benchmarks 14

Specific scores used to measure success in achieving outcomes. A department might set a goal to have 75% of its students pass LITR 260 with a C or better—the “C” would be the standard we want them to be measured against.

Measurement Criteria The standards used to measure student success in meeting course objectives. Competencies and Standards or Benchmarks are common measurement criteria in skills-based courses. Measurement Criteria generally define quality in the course and specify how the instructor will determine the students’ level of understanding. Every Student Learning Outcome should be measured; the method used to measure it should appear in the measurement criteria.

Course information documents at SUNY Delhi (commonly called “syllabi” or “course outlines”) must include measurable student learning outcomes and measurement criteria as categories. Course outlines and syllabi also, by necessity, contain other things: some outlines will list benchmarks and proficiencies under “measurement criteria”; some will list objectives in the “measurable student learning outcomes” section.

Course outlines and syllabi often contain objectives and goals that are not measurable—not everything of value in academics can be set against a yardstick.

Gathering Evidence

The procedure for gathering assessment data can be overwhelming. It is important to develop a system that is both manageable and meaningful. The following are suggestions from other SUNY Delhi faculty members:

1. Don’t reinvent the wheel: Re-examine the assessment tools that you are currently using – many of them are probably already assessing the student learning outcomes of the course. If not, how can they be adapted to measure the student learning outcomes of your course? 2. Take a representative sample: In courses with multiple sections, choose a representative sample of student work to assess rather than assessing every student in every section. 3. Break it up: Rather than collecting data for every student learning outcome each semester, choose a two or three student learning outcomes to assess each semester. Complete the assessment of all student learning outcomes on a two or three year cycle. 4. Collect student work and assess it at the end of the semester: Keep all assignments used for assessment purposes rather than handing them back to students. At the end of the semester, using a grading rubric, sit down with a group of fellow faculty members and rate each piece of student work according to the rubric.

Typically results are reported as percentages. For example, 10% of students exceeded the standard, 40% of students met the standard, 20% of students approached the standard, and 30% of students did not meet the standard.

Reporting the Results 15

Results should be reported on the assessment web site once it is fully implemented. In the meantime, report results to your division dean, who will in turn share them with the Document Repository in the Library and the Deans’ Council and Provost’s Office.

Closing the Loop

One of the most difficult parts of the assessment process is to use the results to improve teaching and learning or “close the loop.” However, in order for assessment to be effective, this is the most important part of the process. Once you have gathered and analyzed the data:

1. Re-examine the instruction: Is the student learning outcome being addressed completely in the class? How can the instruction be improved to address the student learning outcome? How can the course materials like handouts, textbooks, PowerPoints, and assignments be improved? 2. Re-examine the student learning outcomes and goals: Are they appropriate for this course or should they be covered in another course? 3. Re-examine the rubric: Did the rubric accurately measure the assignment? In other words, did the grade assigned by the rubric match your “gut feeling” on what the grade should be? If not, how can it be adapted to better measure the assignment? 4. Re-examine the assessment tool: Did the assessment tool measure the student learning outcome? If not, can it be reworded so that students produce what is intended? 5. Re-examine the assessment process: Could this student learning outcome be assessed in a better way? 16

References:

Anderson, L. W., & Krathwohl, D. R. (2001). A taxonomy for learning, teaching, and assessing: A revision of Bloom’s taxonomy of educational objectives. New York: Longman.

Bloom, B. S., & Krathwohl, D. R. (1956). Taxonomy of educational objectives: The classification of educational goals by a committee of college and university examiners. Handbook I: Cognitive Domain. New York: Longman, Green.

Center for University Teaching, Learning, and Assessment. (2011). Retrieved from http://www.uwf.edu/cutla/index.cfm

Suskie, L. (2009). Assessing student learning: A common sense guide. San Francisco, CA: Jossey-Bass.

University of Connecticut Assessment Prime. (n.d.). Retrieved from http://assessment.uconn.edu/primer/index.html

Recommended publications