Taking up the SLAQ
Picture from Needpix
Background
Like all British Dental Schools, the University of Sheffield
(TUoS) School of Clinical Dentistry was faced with an urgent and novel
situation in march 2020 with regard to student assessment due to the Corvid-19
pandemic and subsequent compulsory move to online teaching.
The School normally holds the examination for the 5th BDS in
late April and early May and this consists of written invigilated papers and
clinical viva examinations during some of which patients are present.
During the week beginning 13th March it became obvious that
all education needed to move to a digital online format and on 20th March the TUoS
declared the closure of all buildings. Thus we needed to arrange a modified
Statement on Methods of Assessment (SoMoA) and organise a completely new form
of assessment which assessed all the learning outcomes and satisfied the
requirements of the General Dental Council (the professional regulatory body
for dentistry).
On 23rd March the General Dental Council (GDC), Dental
Schools council (DSC) and Health Education England (HEE) issued a joint
statement providing the conditions under which students could be registered.
This provided the parameters for the final assessment of students which needed
to be met in order to ensure a registerable degree qualification.
Following TUoS policy on assessment, along with the Joint
Statement, the SoMoA was amended and approval sought and obtained from faculty.
The External Examiners were also informed of the changes.
What did we do?
There are normally two written papers, one designed
to assess the student’s knowledge of oral surgery, oral and maxillofacial
surgery, oral medicine, oral pathology, oral radiology and oral microbiology
and the other integrated clinical practice, including restorative, paediatric
and general primary care dentistry.
As the examinations needed to move to an online version, the emphasis
moved from testing knowledge per se to assessing the students’ ability to apply
this knowledge. Using educational theory relating to both open book
examinations and the long case, alongside our extensive knowledge of how the
clinical workplaces actually works we designed our new form of assessment, the
Structured Long Answer Question, which quickly became known as the SLAQ.
The SLAQ is based around a clinical case. We looked at the GDC
learning outcomes, which had already been mapped to the previous forms of the
exams and tried to ensure that the cases covered these. We checked to ensure
this by amending the structure of the real case upon which the SLAQ was based
with extra details, for example in relation to social history.
We were aware that during the examinations students would have access
to many sources of information to help them answer the question, but we felt that
trying to stop them using these sources would be both very difficult and
counterproductive. After all, in day to day practice we look things up all the
time to ensure our patients get the best and safest care.
Giving students free reign to access any resource they wanted would be
fair to every student and this would also make sure that students could “cheat
all they liked” and thus not “cheat” at all!
The only condition put on this was that they MUST reference where they
had used sources.
Our second issue was time. Most literature on open book examinations
suggest that there should be a time limit so that students could make full use
of any preparation and this helps discriminate between abilities once the
assessments are marked.
TUoS had insisted that the examinations were in a form which became
known as “take-aways”. The exam was to be held within a 24-hour window. This
effectively meant that they would not be time limited, so we decided to use a
word count to limit the amount of information the students could provide. It
helped with marking and ensured that they would need to demonstrate structure
in their thought processes. The scatter approach to examination (I’ll write
everything I can think of) was therefore limited, if not eliminated.
A formative question was composed along with a marking rubric. The
usual lecture held to prepare the students for finals was held, this time
online using Blackboard Collaborate Ultra. The SLAQ was explained to the
students and a (long) question and answer session followed. This was a very
important part of the process as student buy in and understanding was
essential. Good communication with the students appears to be crucial.
The formative question went “live” on 15th April. This was marked and
a further meeting using Blackboard Collaborate Ultra was arranged with the
students. At this point the senior staff marking the completely new format of
online examinations were pleasantly surprised with the outcome. There was an
amazing spread of marks ranging from 45% (just fail) to 90% (almost unheard of)
Not only was it discriminatory but the students we expected to do well did,
whilst those we expected to do less well also met our expectations, so instead
of just releasing the already agreed information on pass or fail, it was then agreed
we would release the students an indicative mark to aid their further study.
The actual summative examinations to replace both the written papers
and the clinical vivas were compiled, approval sought from the External
Examiners, a handbook for markers was produced and a training session held for
staff, again using Blackboard Collaborate Ultra, on 28th April.
The examination replacing the written papers was held on 29th April
and released to markers on 30th April. Each section of the SLAQ was designed so
that it could be marked separately and the final result would be the result of
the combination of marks from all the sections of both cases. This was
calculated separately, just as the answers to the previous finals were
calculated from the separate marks of each question. The questions were single
marked and then moderated, the moderator using a global mark to see if this
aligned with the combination from the examiners, which worked well and in the
only case where the result differed it did so by 1%. There were no signs of
either plagiarism or collaboration and the variation in answers was marked upon
by many of the staff.
The marks were returned and at the analysis meeting on 14th May it became
obvious that the system worked well and analysis had shown that the spread and
mean results were very comparable with the results from students over the
previous 5-year period.
How could
an online open examination work so well?
Between the end of the examination and the release of the results we
decided to ask both the students and the staff what they thought and did an in
depth, on line evaluation. We found that on the whole the students thought it
fair, authentic and actually learnt during the exam! The staff too were happy
and found they actually preferred online marking.
Unfortunately, that’s all we can share, as TUoS policy prevents us
sharing evaluation data further. Suffice it to say you will have to take our
word for it, but the case based SLAQ, designed to test understanding, didn’t
just fulfil the minimum standard of allowing us to test GDC learning outcomes
and thus ensure we qualified students safe to go on the dental register. It
exceeded all our expectations, received very positive comments from the
external Examiners and met many of the criteria of current best practice
assessment as it did indeed examine FOR and encourage further learning.
Comments
Post a Comment