Kirkpatrick's four-level model is the most widely used framework for evaluating training programs. It consists of four levels - reaction, learning, behavior, and results. Each successive level builds upon the prior level and provides a more precise measurement of the training program's effectiveness. However, higher levels also require more rigorous and time-consuming analysis. Level 1 assesses participant satisfaction, level 2 assesses knowledge gain, level 3 evaluates behavior change, and level 4 looks at business impact.
Food Chain and Food Web (Ecosystem) EVS, B. Pharmacy 1st Year, Sem-II
Kirkpatricks four levels_of_evaluation_271
1. Kirkpatrick's Four Levels of Evaluation
ASSESSING TRAINING EFFECTIVENESS often entails
using the four-level model developed by Donald Kirkpatrick
(1994). According to this model, evaluation should always
begin with level one, and then, as time and budget allows,
should move sequentially through levels two, three, and four.
Information from each prior level serves as a base for the next
level's evaluation. Thus, each successive level represents a In Kirkpatrick's four-level model,
more precise measure of the effectiveness of the training each successive evaluation level
program, but at the same time requires a more rigorous and is built on information provided
time-consuming analysis. by the lower level.
Level 1 Evaluation - Reactions
Just as the word implies, evaluation at this level measures how participants in a training program
react to it. It attempts to answer questions regarding the participants' perceptions - Did they like
it? Was the material relevant to their work? This type of evaluation is often called a “smilesheet.”
According to Kirkpatrick, every program should at least be evaluated at this level to provide for
the improvement of a training program. In addition, the participants' reactions have important
consequences for learning (level two). Although a positive reaction does not guarantee learning,
a negative reaction almost certainly reduces its possibility.
Level 2 Evaluation - Learning
Assessing at this level moves the evaluation beyond
learner satisfaction and attempts to assess the extent
students have advanced in skills, knowledge, or attitude.
Measurement at this level is more difficult and laborious
than level one. Methods range from formal to informal
testing to team assessment and self-assessment. If
possible, participants take the test or assessment before
the training (pretest) and after training (post test) to
determine the amount of learning that has occurred.
Level 3 Evaluation - Transfer
To assess the amount of learning
that has occurred due to a training This level measures the transfer that has occurred in
program, level two evaluations often learners' behavior due to the training program. Evaluating
use tests conducted before training at this level attempts to answer the question - Are the
(pretest) and after training (post test). newly acquired skills, knowledge, or attitude being used
in the everyday environment of the learner? For many
trainers this level represents the truest assessment of a program's effectiveness. However,
measuring at this level is difficult as it is often impossible to predict when the change in behavior
will occur, and thus requires important decisions in terms of when to evaluate, how often to
evaluate, and how to evaluate.
Level 4 Evaluation- Results
2. Frequently thought of as the bottom line, this level
measures the success of the program in terms that
managers and executives can understand -increased
production, improved quality, decreased costs,
reduced frequency of accidents, increased sales, and
even higher profits or return on investment. From a
business and organizational perspective, this is the
overall reason for a training program, yet level four
results are not typically addressed. Determining Level four evaluation attempts to assess
results in financial terms is difficult to measure, and is training in terms of business results. In
hard to link directly with training. this case, sales transactions improved
steadily after training for sales staff
Methods for Long-Term Evaluation occurred in April 1997.
• Send post-training surveys
• Offer ongoing, sequenced training and coaching over a period of time
• Conduct follow-up needs assessment
• Check metrics (e.g., scrap, re-work, errors, etc.) to measure if participants achieved
training objectives
• Interview trainees and their managers, or their customer groups (e.g., patients, other
departmental staff)
Summary of the Kirkpatrick Model
Donald Kirkpatrick first proposed this four-pronged approach to evaluating training programs in
his 1959 doctoral dissertation.
Since then, it has become so widely used, that trainers can typically talk about it in shorthand and
understand the reference. For example, when one trainer says to another, "What are you doing
about level IV?" the other knows that the first trainer wants to understand how the second
evaluates the impact of training.
Level Name Issues Assessed at this Level
I Reaction
Assesses participants’ initial reactions to a course. This, in turn, offers insights into participants’
satisfaction with a course, a perception of value. Trainers usually assess this through a survey,
often called a "smiley sheet." Occasionally, trainers use focus groups and similar methods to
receive more specific comments (called qualitative feedback) on the courses. According to the
TRAINING magazine annual industry survey, almost 100 percent of all trainers perform "Level I"
evaluation.
II Learning
Assesses the amount of information that participants learned. Trainers usually assess this with a
criterion-referenced test. The criteria are objectives for the course: statements developed before
a course is developed that explicitly state the skills that participants should be able to perform
after taking a course. Because the objectives are the requirements for the course, a Level II
evaluation assesses conformance to requirements, or quality.
III Transfer
Assesses the amount of material that participants actually use in everyday work 6 weeks to 6
months (perhaps longer) after taking the course. This assessment is based on the objectives of
3. the course and assessed through tests, observations, surveys, and interviews with co-workers
and supervisors. Like the Level II evaluation, Level III assesses the requirements of the course
and can be viewed as a follow-on assessment of quality.
IV Impact
Assesses the financial impact of the training course on the bottom line of the organization 6
months to 2 years after the course (the actual time varies depending on the context of the
course).
For many reasons, Level IV is the most difficult level to measure. First, most training courses do
not have explicitly written business objectives, such as "this course should reduce support
expenses by 20 percent." Second, the methodology for assessing business impact is not yet
refined. Some assess this measurement by tracking business measurements, others assess by
observations, some by surveys, and still others assess by qualitative measures. Last, after six
months or more, evaluators have difficulty solely attributing changed business results to training
when changes in personnel, systems, and other factors might also have contributed to business
performance.
For the evaluation of T& D , the widely used model is KIRKPATRICK"S MODEL.
It consists of four steps-
1. Reactions
2. Learning.
3. Behaviour
4. Results
[img][/img]
1. Reactions. "Reaction may best be defined as how well the trainees liked a particular training program."
Reactions are typically measured at the end of training -- at Point 3 However, that is a summative or end-of-
course assessment and reactions are also measured during the training, even if only informally in terms of the
instructor's perceptions.
2 Learning. "What principles, facts, and techniques were understood and absorbed by the conferees?" What the
trainees know or can do can be measured during and at the end of training but, in order to say that this
knowledge or skill resulted from the training, the trainees' entering knowledge or skills levels must also be known
or measured. Evaluating learning, then, requires measurements at Points 1, 2 and 3 -- before, during and after
training
3. Behavior. Changes in on-the-job behavior. Clearly, any evaluation of changes in on-the-job behavior must
occur in the workplace itself -- at Point 4 . It should be kept in mind, however, that behavior changes are
acquired in training and they then transfer (or don't transfer) to the work place. It is deemed useful, therefore, to
assess behavior changes at the end of training and in the workplace.
4. Results. "Reduction of costs; reduction of turnover and absenteeism; reduction of grievances; increase in
quality and quantity or production; or improved morale which, it is hoped, will lead to some of the previously
stated results." These factors are also measurable in the workplace -- at Point 4
Hope you got....
evaluate.gif
Description:
Filesize: 4.07 KB
Viewed: 129 Time(s)
4. I use a simple form - on side 1 is the name of the employee, name of reporting officer, department name, title,
date and venue of the training, costs of the course, accommodation and travel, and then below this space for
justifying the training - both by the employee and the reporting officer, followed by Side 2 where the objectives
are listed along with how success for each will be measured. Then below this is space to put notes on an after-
training meeting and details of an action plan to take the lessons of the training back to the workplace - actions,
resources needed, responsible person for each action and deadline. I also use this form then to track the financial
costs, to audit the training request and delivery process and to hold reporting officers and employees to account
and see how they are progressing.
Why Measure Training Effectiveness?
Measuring the effectiveness of training programs consumes valuable time
and resources. As we know all too well, these things are in short supply in
organizations today. Why should we bother?
Many training programs fail to deliver the expected organizational benefits.
Having a well-structured measuring system in place can help you determine
where the problem lies. On a positive note, being able to demonstrate a real
and significant benefit to your organization from the training you provide can
help you gain more resources from important decision-makers.
Consider also that the business environment is not standing still. Your
competitors, technology, legislation and regulations are constantly changing.
What was a successful program yesterday may not be a cost-effective
program tomorrow. Being able to measure results will help you adapt to such
changing circumstances.
The Kirkpatrick Model
5. The most well-known and used model for measuring the effective of training
programs was developed by Donald Kirkpatrick in the late 1950s. It has since
been adapted and modified by a number of writers, however, the basic
structure has well stood the test of time. The basic structure of Kirkpatrick’s
four-level model is shown here.
Figure 1 - Kirkpatrick Model for Evaluating Effectiveness of Training
Programs
Level 4 - What organizational benefits resulted
Results from the training?
To what extent did participants change
Level 3 -
their behavior back in the workplace as
Behavior
a result of the training?
To what extent did participants improve
Level 2 -
knowledge and skills and change
Learning
attitudes as a result of the training?
Level 1 - How did participants react to the
Reaction program?
An evaluation at each level answers whether a fundamental requirement of
the training program was met. Its not that conducting an evaluation at one
level is more important that another. All levels of evaluation are important.
In fact, the Kirkpatrick model explains the usefulness of performing
evaluations at each level. Each level provides a diagnostic checkpoint for
problems at the succeeding level. So, if participants did not learn (Level 2),
participant reactions gathered at Level 1 (Reaction) will reveal the barriers to
learning. Now moving up to the next level, if participants did not use the
skills once back in the workplace (Level 3), perhaps they did not learn the
required skills in the first place (Level 2).
The difficulty and cost of conducting an evaluation increases as you move up
the levels. So, you will need to consider carefully what levels of evaluation
you will conduct for which programs. You may decide to conduct Level 1
evaluations (Reaction) for all programs, Level 2 evaluations (Learning) for
“hard-skills” programs only, Level 3 evaluations (Behavior) for strategic
programs only and Level 4 evaluations (Results) for programs costing over
$50,000. Above all else, before starting an evaluation, be crystal clear about
your purpose in conducting the evaluation.
Using the Kirkpatrick Model
How do you conduct an evaluation? Here is a quick guide on some
appropriate information sources for each level.
Level 1 (Reaction)
• completed participant feedback questionnaire
• informal comments from participants
• focus group sessions with participants
6. Level 2 (Learning)
• pre- and post-test scores
• on-the-job assessments
• supervisor reports
Level 3 (Behavior)
• completed self-assessment questionnaire
• on-the-job observation
• reports from customers, peers and participant’s manager
Level 4 (Results)
• financial reports
• quality inspections
• interview with sales manager
When considering what sources of data you will use for your evaluation, think
about the cost and time involved in collecting the data. Balance this against
the accuracy of the source and the accuracy you actually need. Will existing
sources suffice or will you need to collect new information?
Think broadly about where you can get information. Sources include:
• hardcopy and online quantitative reports
• production and job records
• interviews with participants, managers, peers, customers, suppliers and regulators
• checklists and tests
• direct observation
• questionnaires, self-rating and multi-rating
• Focus Group sessions
Once you have completed your evaluation, distribute it to the people who
need to read it. In deciding on your distribution list, refer to your previously
stated reasons for conducting the evaluation. And of course, if there were
lessons learned from the evaluation on how to make your training more
effective, act on them!
Business Performance Pty Ltd products can help you plan and implement
your training evaluation project.
Our Training Management Template Pack contains ready to go checklists,
interview and questionnaire forms for conducting Level 1 (Reaction) and
Level 3 (Behavior) evaluations.
Our eBook From Training to Enhanced Workplace Performance contains
guidance, checklists, interview and questionnaire forms for conducting Level
3 (Behavior) evaluations.
AT+D classic: How to start an Objective Evaluation of your Training Program
T+D, May, 2004 by Donald L. Kirkpatrick
This excerpt is part of a larger article with the same title that originally appeared in the May-June 1956 issue of the
7. Journal of the American Society of Training Directors, a predecessor to T+D. The article heralded Kirkpatrick's now classic
four-level evaluation model.
Most training men agree that it is important to evaluate training programs. They also feel that the evaluation should be
done by objective means. However, the typical training man uses evaluation sheets or comment sheets as the sole
measure of the effectiveness of his programs. He realizes he should do more, but he just doesn't know how to begin an
objective evaluation.
According the Raymond Katzell, a well-known authority in this field, the evaluation of a training program falls into a
hierarchy of steps that can be briefly stated as follows:
Step One. To determine how the trainees feel about the program.
Step Two. To determine how much the trainees learn in the form of increased knowledge and understanding.
Step Three. To measure the changes in the on-the-job behavior of the trainees.
Step Four. To determine the effects of these behavioral changes on objective criteria such as production, turnover,
absenteeism, and waste.
In climbing this ladder of evaluation, most trainers have completed the first step. Typically, the training director asks the
trainees to fill out evaluation sheets at the end of the program. Questions that are asked most frequently are
* How do you rate the program?
* What subject did you like best?
* What subject did you like least?
* What did you learn that you can use on the job?
* What subjects would you like to have discussed at future programs?
Usually the trainees are not asked to sign their name for fear they will not give an honest reaction.
This kind of subjective evaluation is important. It gives a good indication of how the trainees reacted to the program. If
they react favorably, the trainer can justifiably pat himself on the back and say, "I gave them a program they liked." But
he can't rightfully claim that the training program accomplished the objective, unless his objective was to give them a
program they liked.
The immediate objective of any training course can be stated in terms of the desired knowledge and understanding that
the program is trying to impart to the trainees. It is this stage of evaluation that should be undertaken as the second
step. It is much more difficult than step one and, therefore, is not undertaken by many trainers.
Among the possible methods for determining whether increased knowledge and understanding have taken place, the best
one seems to be the "before and after" paper and pencil test. If the scores on the posttest are significantly higher than on
the pretest, the course can be deemed effective.
In determining the effectiveness of the training, it is important to note that the paper and pencil test or inventory must
cover the principles and facts that are discussed in the course. If the trainer can find a test that covers this material, he
can use it. If he cannot find a suitable one, he must construct his own inventor. Some of the inventories that are
available are: How Supervise? by File and Remmers; Supervisory Inventory by Wesley Osterberg; and the Supervisory
Inventory on Human Relations constructed by this writer.
So far, then, it has been stated that a before and after test can be used to deter mine whether or not increased
knowledge and understanding have taken place. Also, that the inventory should cover the course content. In order to
determine whether or not an available test is suitable, a trainer must examine his course outline and list the principles
and facts he is trying to teach. A comparison of test items with these objectives will reveal whether or not the test can be
used. Because the construction of a test involves such factors as the choice of items, the wording of the items, the
number and type of possible response, and the sequence of items, it is far better to use an available inventory if it covers
most of the course content.
Having selected or constructed a test, the trainer should consider some "Do's" for administering it:
* Give the pretest at the start of the first class and the posttest at the close of the last session. This will minimize the
influence of factors apart from the training course.
* Have the trainee sign both the pretest and posttest. Then, the increased knowledge and understanding can be
computed for each individual.
In instructing the trainees before they take the pretest:
* Tell them it is a before and after procedure.
* Explain the purpose of the test.
* Encourage them to answer truthfully by assuring them that their scores will have no effect on their pay or status in the
company.
* Tell them to answer every question even if they have to guess. (This will be taken into account in the statistical analysis
of scores.)
* Encourage them to take their time in taking the test. This will help to motivate them to read each item carefully.
8. In analyzing the test results, there are two kinds of evaluations to be made:
* Was the entire course effective as shown by gains from pretest to posttest scores for all trainees?
* What specific facts and principles were learned as shown by changes from pretest to posttest for each item?
Summary
Training men agree that it is advisable to evaluate training courses as objectively as possible. Typically, their evaluation
consists of subjective comment sheets that are completed by the trainees at the end of the course. Providing that these
are properly administered, these evaluation sheets give a valid measure of trainee reaction to the program. However,
they do not give any evidence of benefits derived.
The first step in objectively evaluating the effectiveness of a training course is to determine whether or not the desired
facts and principles were learned by the trainees. This can be done by:
* Using a suitable paper and pencil test.
* Testing the trainees before and after the program.
* Determining the overall effectiveness of the course by comparing pretest and posttest scores for each trainee.
* Determining which specific facts and principles were learned by analyzing the changes on each test item from pretest
and posttest.
The purpose of this article is to suggest a specific technique for beginning an objective evaluation of training program.
Further efforts should be undertaken by every training man to follow up this kind of evaluation by attempting to measure
trainee change in behavior that occurs as a result of participation in the program.