Program Assessment

Program Assessment Cycle

Contact us:

Meaningful program assessment follows an intentional and reflective process of design, implementation, evaluation, and reflection and revision. The following tools and resources are intended to support each stage of the program assessment cycle.

The Program Assessment Cycle: Closing the Loop on Student Learning progresses from outcomes to identifying student work, collecting and scoring evidence, analyzing and interpreting results, drawing conclusions, solicit feedback, and using data to inform future activities.

Program Outcomes

The assessment cycle begins with program faculty coming together to consider their program goals and student learning outcomes and to identify a research question about student learning.

  • Program goals are statements that identify the concepts and skills that students in your program should attain by the time they graduate, and should align with the mission of your department, your college, and the University.
  • Program learning outcomes are concrete descriptions of what students in your program should know and be able to do when they have completed or participated in an assignment, activity, and/or project. They should be specific and measurable, and they should correspond to your program goals.

All programs at URI are expected to post their program outcomes on their website* as doing so has been shown to improve student motivation and engagement, provides them a language to communicate what they have learned to others, and helps them practice metacognition (being aware of and understanding one’s own thought processes).
*Per the program assessment policy endorsed and ratified by URI’s Faculty Senate in April 2010

Identify Evidence of Student Learning

After articulating clear program outcomes and a research question, the next step is to determine in what ways required courses contribute to student learning. Each required course in the curriculum should be linked to at least one program outcome. Collaboratively creating a curriculum map will also help faculty determine which required courses are likely to provide the most appropriate evidence of student learning.

Collect, Score, and Analyze Evidence

After identifying evidence of student learning and in which courses provide that evidence, the process of collecting and analyzing student data begins. When planning evidence collection, consider this blog post from Linda Suskie about sample sizes. For example, a small sample size (e.g., 10-20) may need to include the work of all students and for a large sample size (e.g., 150) around 20% of students may be enough.

Analysis starts with setting an expected level of student achievement for each program outcome examined. Scoring assignments such as projects, papers, performances or presentations can be complex. A scoring guide or rubric provides the best option for consistent scoring between multiple faculty members. The analysis compares expected and actual levels of achievement to look for patterns of strengths and weaknesses.

Check out this downloadable ebook to see and learn how to design authentic rubric templates for effective and scalable assessment.

Interpreting Results and Determining Actions

To close the loop on student learning, the department then summarize the findings, interprets the analysis, draws conclusions and makes recommendations for change, including a timeframe for implementing these changes.

This forms the basis of a Well-Developed Program Assessment Report (PDF). See the Reporting page for details on that process. If your team runs into unexpected delays, you can request an extension (PDF).

Assessment Support