use case

Dynamic Course Evaluation

How UHasselt University converted existing course evaluations into intelligent check-ins







Focus area

Course evaluation


Avg. response rate


Avg. completion rate


Student easiness


Over the years, Hasselt University has used different student check-ins to assess the educational quality across the university. Evaluating the quality of teaching, specific courses and the programmes in general.

From the fall semester 2021, StudentPulse has been used to convert existing surveys into an intelligent check-in setup for three of the university’s faculties. Looking into differences between a traditional setup and a new approach to student feedback.

The Challenge

In their existing setup, Hasselt University experienced difficulties when checking-in on specific courses and programmes (course evaluations). The response rates and student easiness proved to be a an issue, and for programme staff a more flexible, customisable and usable platform was needed without compromising the possibility to compare data with other programmes and faculties. With implementing StudentPulse the aim was to:

  1. Increase engagement of students with regard to educational quality assurance.
  2. Motivate  students to fill in questionnaires related to educational quality in order to obtain reliable data.
  3. Support university staff in its work with improving  quality of teaching during the semester and not just at the end.
  4. Compare results across courses, programmes and faculties, still allowing a customised feedback setup within each of these.

The Solution

The main challenge for Hasselt University was to increase student engagement and motivation, to disseminate feedback in an easy interpretable way, and to make sure feedback was used to improve students' experience.

Involving the organisation

To align a new setup with organisational needs, before implementing StudentPulse, an internal analysis clarifying the different types of quality assurance tools used at Hasselt University was conducted amongst students and staff to pinpoint the strengths and weaknesses of the institution’s QA instruments. Further, a selection of faculties and study programmes was made.

When the implementation started, the selected faculties were involved in the process of choosing questions within six different quality criteria. Further, the Quality Assurance Team together with the faculties decided upon the number of check-ins during the semester, the number of courses to be included and which colleagues to be involved.

By involving faculties, they got familiar with the new approach even before the first feedback was collected, but more importantly it made sure that the setup was adjusted to their needs, increasing the likelihood of not just collecting but also using student feedback more often and efficiently throughout the semester.

Building a custom StudentPulse-framework

StudentPulse comes with a framework of 11 verified student experience drivers and 60 verified questions, but in the case of Hasselt University it was decided to implement a custom framework instead. This allowed the institution to build their existing six quality criteria into the platform, making it possible to compare StudentPulse-data with faculties not using StudentPulse, as well as comparing new data with previous years’ data.

The framework was set up in a way where quality criteria were kept consistent across all participating courses, whereas differentiation of questions were allowed down to the programme level.

A micro-distribution strategy

To reach most students in the most efficient way, professors got notified with a check-in link and QR-code and got a reminder to make the students fill out the check-ins during class. At the same time, check-ins were posted on the institution's learning management system.

The process of collecting feedback is currently being automated and integrated but the idea of engaging professors remains at the core of the future distribution strategy, serving as an important prerequisite for building a strong feedback loop between students and staff.

Tracking check-in participation, completion and easiness

When student engagement and motivation is put on top of the agenda, so is the need of measuring these concepts. To do so, student response rates, completion rates and easiness scores were reported throughout the semester. The data collection KPI’s all yielded significant improvements across programmes, but more importantly they uncovered the effect of different approaches to check-in distribution as well as check-in setup. Inputs that have been used to define the future best practices for the institution to work with student feedback.


Pushing check-in links through learning management system and in-class QR codes.

Intelligent check-in frequency

Before, during and after quartile.

Questions per pulse

5-7 plus follow-up's on sceptical student answers

Main KPI's

Response rate, completion rate, student easiness, comment likelihood

The results

More feedback

More students reply on a continuous basis, and more students complete the check-ins.

More improvement suggestions

Students' comment likelihood has been increased, and their improvement suggestions are automatically linked to each subdriver of student experience.

Faster feedback cycle

Check-ins are taken multiple times throughout the educational period, and staff have the opportunity to adapt their classes to the feedback they received from the students.

In using StudentPulse we were aiming to optimise the relevance of the questions asked, diminish the length of the questionnaires, update to a more user-friendly application, and implement a shorter and more transparent feedback loop with linked actions. This way we hoped to increase the response rate, to be more actionable and to react quickly when problems or challenges arise"

Franne Schepers

Quality Assurance, Hasselt University