Towards effective practitioner evaluation: An exploration of issues relating to skills, motivation and evidence

Jen Harvey, Martin Oliver, Janice Smith

    Research output: Contribution to journalReview articlepeer-review

    6 Citations (Scopus)

    Abstract

    Although academics are increasingly expected to undertake studies of their practice, particularly where this involves the use of learning technology, experience to date suggests that meeting this expectation has proved difficult. This paper attempts to explain this difficulty. After reviewing literature that provides a rationale for practitioner evaluation, the experiences of three projects (EFFECTS, ASTER and SoURCE) which attempted to draw on this process are described. Three main areas of difficulty are discussed: the skills and motivations of the academics involved, and the kinds of evidence (and its analysis) that 'count' for a given evaluation. This discussion leads to the identification of a number of problems that inhibit practitioner evaluation, including ambiguity in the nature and purpose of evaluation, and a general feeling that the function of evaluation has already been served through existing quality mechanisms. Finally, the possible implications are considered of some or all of the steps in the evaluation process being undertaken by an evaluator working alongside the academic.

    Original languageEnglish
    Pages (from-to)3-10
    Number of pages8
    JournalEducational Technology and Society
    Volume5
    Issue number3
    Publication statusPublished - Jul 2002

    Keywords

    • Academic roles
    • Evaluation
    • Learning technology
    • Participant evaluation

    Fingerprint

    Dive into the research topics of 'Towards effective practitioner evaluation: An exploration of issues relating to skills, motivation and evidence'. Together they form a unique fingerprint.

    Cite this