Dynamic Course Evaluation
How UHasselt University converted existing course evaluations into micro surveys
Avg. response rate
Avg. completion rate
Over the years, Hasselt University has used different student surveys to assess the educational quality across the university. Evaluating the quality of teaching, specific courses and the programmes in general.
From the fall semester 2021, StudentPulse has been used to convert existing surveys into a micro survey setup for three of the university’s faculties. Looking into differences between a traditional setup and a new approach to student feedback.
In their existing setup, Hasselt University experienced difficulties when surveying specific courses and programmes (course evaluations). The response rates and student easiness proved to be a an issue, and for programme staff a more flexible, customisable and usable platform was needed without compromising the possibility to compare data with other programmes and faculties. With implementing StudentPulse the aim was to:
The main challenge for Hasselt University was to increase student survey engagement and motivation, to disseminate feedback in an easy interpretable way, and to make sure feedback was used to improve students' experience.
Involving the organisation
To align a new setup with organisational needs, before implementing StudentPulse, an internal analysis clarifying the different types of quality assurance tools used at Hasselt University was conducted amongst students and staff to pinpoint the strengths and weaknesses of the institution’s QA instruments. Further, a selection of faculties and study programmes was made.
When the implementation started, the selected faculties were involved in the process of choosing questions within six different quality criteria. Further, the Quality Assurance Team together with the faculties decided upon the number of micro surveys (pulses) during the semester, the number of courses to be included and which colleagues to be involved.
By involving faculties, they got familiar with the new approach even before the first feedback was collected, but more importantly it made sure that the setup was adjusted to their needs, increasing the likelihood of not just collecting but also using student feedback more often and efficiently throughout the semester.
Building a custom StudentPulse-framework
StudentPulse comes with a framework of 11 verified student experience drivers and 60 verified questions, but in the case of Hasselt University it was decided to implement a custom framework instead. This allowed the institution to build their existing six quality criteria into the platform, making it possible to compare StudentPulse-data with faculties not using StudentPulse, as well as comparing new data with previous years’ data.
The framework was set up in a way where quality criteria were kept consistent across all participating courses, whereas differentiation of questions were allowed down to the programme level.
A micro-distribution strategy
To reach most students in the most efficient way, professors got notified with a survey-link and QR-code and got a reminder to make the students fill out the surveys during class. At the same time, surveys were posted on the institution's learning management system.
The process of collecting feedback is currently being automated and integrated but the idea of engaging professors remains at the core of the future distribution strategy, serving as an important prerequisite for building a strong feedback loop between students and staff.
Tracking survey participation, completion and easiness
When student survey engagement and motivation is put on top of the agenda, so is the need of measuring these concepts. To do so, student response rates, completion rates and easiness scores were reported throughout the semester. The data collection KPI’s all yielded significant improvements across programmes, but more importantly they uncovered the effect of different approaches to survey distribution as well as survey setup. Inputs that have been used to define the future best practices for the institution to work with student feedback.
Pushing survey links through learning management system and in-class QR codes.
Before, during and after quartile.
5-7 plus follow-up's on sceptical student answers
Response rate, completion rate, student easiness, comment likelihood
More students reply on a continuous basis, and more students complete the surveys.
Students' comment likelihood has been increased, and their improvement suggestions are automatically linked to each subdriver of student experience.
Surveys are taken multiple times throughout the educational period, and staff have the opportunity to adapt their classes to the feedback they received from the students.
Quality Assurance, Hasselt University
More stories for you