How To Evaluate Training

How To Evaluate Training

The definitive guide to thinking, tools and techniques

You’ve spent big money, time, effort, blood, sweat and tears in creating your training program … but did it work? Was there a payback, and how do you prove it?

Payback is the holy grail of training and training evaluation, and we’re now going to look at it in the next two editions:

  • The first, this one, will focus on models for evaluating training.
  • The second will focus on the methods.

We have got some associated collateral particularly focusing on the measures you might use. If you ping me an e-mail (mwoods@statius.co.uk), I’d be happy to provide you with a copy … and talking of measures, the first stage is to:

Determine the Measure

Before you evaluate the effectiveness of any training program, you need to decide what the indicators of “effectiveness” are. Is training a success when employees become better at their jobs? Or is it, perhaps, a happier, healthier company culture? Is it, maybe, both, or is it something else? The key is to start with the end in mind.

The point is, you’ll probably want to track more than one measure of training effectiveness. The more measures you include, the more information you’ll have to help you improve this or any future initiative. 

Determine the Method

A lot of big books have been written on the subject of training evaluation, and the methods for doing so, so we’re not short of techniques to choose from! 

More helpfully, there are a small number of tried and tested techniques, and we’re going to explore a few of them now.

Kirkpatrick’s Four-level Training Evaluation Model

Donald Kirkpatrick invented his four-step training evaluation model as part of his Ph.D. in 1954, but it really took off after the publication of his 1994 book, Evaluating Training Programs. So, this model has been around for some time and has, therefore, stood the test of time. Other models are often simply refinements of his ideas.

Let’s look, progressively deeper, at his four-step model:

Step 1: Reaction: Yep, our man Kirkpatrick is the guy responsible for those “happy sheets” you get after each and every training course you’ve ever been on — the objective of which is to evaluate your reaction to the training at the point of training. At this stage, whilst you may have learnt something, it won’t have been applied — you’ve not even left the (real or perhaps online) room yet! This is about attempting to understand your overall satisfaction with the learning experience.

Step 2: Learning: Measure what was learnt during training. Use assessments to measure how much knowledge and skills have changed from before to after training.

Step 3: Behaviour: This step seeks to understand whether or not (and by how much) your behaviour has changed as a result of the training undertaken. This is usually an external assessment by others using observation of performance, pre- and post-training. 

Step 4: Results: The fourth, final and most important step is to evaluate the impact of your training on results. This is often measured at different levels, for instance:

  • Individual: What was the impact on you? How did you benefit?
  • Team/department: What was the impact on your team/department? How did they benefit?
  • Company or organisational: What was the impact on your organisation? How did the organisation benefit?
Lift Company Measures 
Increased productivity — more visits per day/fewer call backs 
Increased engineer utilisation 
Improved response time 
Improved on-time delivery
Better project management (less wasted time)
Improved safety performance
Table 1

The actual results being measured would have been decided at the objective setting stage but might include issues like sales conversion, productivity, quality, efficiency and customer satisfaction. Digging into productivity, for a lift company, measures might include things like number of jobs completed per day, engineers’ rate of first-time fix, number of callbacks and the like. Kirkpatrick’s model is still in use for good reason. Its logical, staged approach is easy to apply, and once the evaluation is complete, you’ll have a deep and wide understanding of learning obtained.

The Phillips ROI Model

A guy called JJ Phillips argued for a fifth step to be added to the Kirkpatrick model to evaluate the program’s return on investment (ROI) so you’d need to measure the difference between your training “cost” and the “results” it delivered. When the results exceed the cost of undertaking the training, then you’ve got a positive ROI.

However, we’d suggest that, in practical terms, this is something you ought to be doing as part of the Kirkpatrick model anyway. 

Anderson’s Model of Learning Evaluation

The Anderson model adopts a focus that very much attempts to align the training, and its evaluation, to the strategy the company is employing.

The easiest way to explain the approach is probably by way of example.

How To Evaluate Training
Kirkpatrick’s Four-level Training Evaluation Model

Obviously, we are going to have to oversimplify a bit, but I hope you’ll get the picture. Let’s take a fictitious host lift company with sufficient resources to keep about 1,000 units and their associated customers extremely happy with the current level of service they promise (missed visits, on-time visits, repairs, modernisations and installations all completed on time). The No. 1 strategy is the relentless and obsessive focus on delivering an excellent service and the accompanying customer satisfaction metrics; the results from which the directors are justifiably proud.

However, the directors now want to rapidly expand the business, but they can’t yet take the chance of adding to costs with more staffing resources, so they know that additional pressures will be placed on staff, and the treasured customer satisfaction scores might suffer; they are worried that their (sterling) reputation may take a hit. Now, suppose that their external consultant has developed a couple of programs 1) to help the marketing team win new customers and another 2) to cut waste from the existing processes.

If this model works for you, then there are three stages:

  • Stage 1: Critically review your strategies and your training plans against one another. Does the training plan focus on strategic priorities? Looking to the example above, we can see that there was potentially a mis-match between the strategy to increase clients and the desire to deliver high-quality customer care.
  • Stage 2: Measure the contribution of the training programme to strategic ambitions. For example, a training program that helps you reduce waste could be measured by the percentage of decreased costs.

The measures listed in Table 1 are only a small sample, so the list is long.

  • Stage 3: This is where you decide the ROI, and this step depends on your approach. Examples include:
  • A comparison of the contribution identified in Step 2 to resources that are invested into the training programme, or,
  • You might ask whether the percentage of decrease in costs was big enough: Did it meet your expectations?

If the numbers do not produce an ROI that you are happy with, then it’s back to the drawing board to make improvements! The approach promotes a virtuous circle that seeks to learn, align and improve.

Conclusion

All models are just that — models. There is a great NLP (neuro linguistic programming — look it up; it’s a great toolkit) presupposition which states that “the map is not the territory.” We’re very much advocates of making the models work for you, so, in conclusion, we’d say the critical components of training evaluation are:

  • Deciding on the evaluation criteria (at the start)
  • Measuring impact at the individual, team and company levels to see if you have got payback (as per the diagram)

In the next article, we’re going to look at different methods of evaluation.

Mark Woods has a background in engineering. After completing an award winning apprenticeship he completed a degree in mechanical and production engineering. He also has an MBA from Bradford University where he studied with Professor J.S. Oakland, the world’s first Professor of Total Quality Management. During the time he spent in industry, he gained extensive experience of both implementing new technology and strategic appraisal. Mark is now a managing partner of Statius Management Services, a management consultancy specialising in performance improvement, a subject on which Mark has published a number of papers and articles.

Get more of Elevator World. Sign up for our free e-newsletter.

Please enter a valid email address.
Something went wrong. Please check your entries and try again.

Rethinking IoT Applications in the Elevator Industry

Rethinking IoT Applications in the Elevator Industry

Accelerating Decarbonization

Accelerating Decarbonization

Elevator-World---Fallback-Image

“An Exciting New Chapter”

Pay First or Risk Disappointment

Pay First or Risk Disappointment

Digitalisation’s Impact

Digitalisation’s Impact

Elevator-World---Fallback-Image

The Good, the Bad, the Strange

Champagne, Castles and Collaboration

Champagne, Castles and Collaboration

Elevator-World---Fallback-Image

Safety Appliances for Lifts: 1895(Part 3)