SlideShare a Scribd company logo
1 of 8
Kirkpatrick's Four Levels of Evaluation

ASSESSING TRAINING EFFECTIVENESS often entails
using the four-level model developed by Donald Kirkpatrick
(1994). According to this model, evaluation should always
begin with level one, and then, as time and budget allows,
should move sequentially through levels two, three, and four.
Information from each prior level serves as a base for the next
level's evaluation. Thus, each successive level represents a        In Kirkpatrick's four-level model,
more precise measure of the effectiveness of the training           each successive evaluation level
program, but at the same time requires a more rigorous and          is built on information provided
time-consuming analysis.                                            by the lower level.

Level 1 Evaluation - Reactions

Just as the word implies, evaluation at this level measures how participants in a training program
react to it. It attempts to answer questions regarding the participants' perceptions - Did they like
it? Was the material relevant to their work? This type of evaluation is often called a “smilesheet.”
According to Kirkpatrick, every program should at least be evaluated at this level to provide for
the improvement of a training program. In addition, the participants' reactions have important
consequences for learning (level two). Although a positive reaction does not guarantee learning,
a negative reaction almost certainly reduces its possibility.

Level 2 Evaluation - Learning

                                            Assessing at this level moves the evaluation beyond
                                            learner satisfaction and attempts to assess the extent
                                            students have advanced in skills, knowledge, or attitude.
                                            Measurement at this level is more difficult and laborious
                                            than level one. Methods range from formal to informal
                                            testing to team assessment and self-assessment. If
                                            possible, participants take the test or assessment before
                                            the training (pretest) and after training (post test) to
                                            determine the amount of learning that has occurred.

                                            Level 3 Evaluation - Transfer
To assess the amount of learning
that has occurred due to a training          This level measures the transfer that has occurred in
program, level two evaluations often         learners' behavior due to the training program. Evaluating
use tests conducted before training          at this level attempts to answer the question - Are the
(pretest) and after training (post test).    newly acquired skills, knowledge, or attitude being used
                                             in the everyday environment of the learner? For many
trainers this level represents the truest assessment of a program's effectiveness. However,
measuring at this level is difficult as it is often impossible to predict when the change in behavior
will occur, and thus requires important decisions in terms of when to evaluate, how often to
evaluate, and how to evaluate.




Level 4 Evaluation- Results
Frequently thought of as the bottom line, this level
measures the success of the program in terms that
managers and executives can understand -increased
production, improved quality, decreased costs,
reduced frequency of accidents, increased sales, and
even higher profits or return on investment. From a
business and organizational perspective, this is the
overall reason for a training program, yet level four
results are not typically addressed. Determining             Level four evaluation attempts to assess
results in financial terms is difficult to measure, and is   training in terms of business results. In
hard to link directly with training.                         this case, sales transactions improved
                                                             steadily after training for sales staff
Methods for Long-Term Evaluation                             occurred in April 1997.

    •   Send post-training surveys
    •   Offer ongoing, sequenced training and coaching over a period of time
    •   Conduct follow-up needs assessment
    •   Check metrics (e.g., scrap, re-work, errors, etc.) to measure if participants achieved
        training objectives
    •   Interview trainees and their managers, or their customer groups (e.g., patients, other
        departmental staff)




Summary of the Kirkpatrick Model

Donald Kirkpatrick first proposed this four-pronged approach to evaluating training programs in
his 1959 doctoral dissertation.

Since then, it has become so widely used, that trainers can typically talk about it in shorthand and
understand the reference. For example, when one trainer says to another, "What are you doing
about level IV?" the other knows that the first trainer wants to understand how the second
evaluates the impact of training.

  Level    Name              Issues Assessed at this Level
I          Reaction
Assesses participants’ initial reactions to a course. This, in turn, offers insights into participants’
satisfaction with a course, a perception of value. Trainers usually assess this through a survey,
often called a "smiley sheet." Occasionally, trainers use focus groups and similar methods to
receive more specific comments (called qualitative feedback) on the courses. According to the
TRAINING magazine annual industry survey, almost 100 percent of all trainers perform "Level I"
evaluation.

II          Learning
Assesses the amount of information that participants learned. Trainers usually assess this with a
criterion-referenced test. The criteria are objectives for the course: statements developed before
a course is developed that explicitly state the skills that participants should be able to perform
after taking a course. Because the objectives are the requirements for the course, a Level II
evaluation assesses conformance to requirements, or quality.

III      Transfer
Assesses the amount of material that participants actually use in everyday work 6 weeks to 6
months (perhaps longer) after taking the course. This assessment is based on the objectives of
the course and assessed through tests, observations, surveys, and interviews with co-workers
and supervisors. Like the Level II evaluation, Level III assesses the requirements of the course
and can be viewed as a follow-on assessment of quality.

IV      Impact
Assesses the financial impact of the training course on the bottom line of the organization 6
months to 2 years after the course (the actual time varies depending on the context of the
course).

For many reasons, Level IV is the most difficult level to measure. First, most training courses do
not have explicitly written business objectives, such as "this course should reduce support
expenses by 20 percent." Second, the methodology for assessing business impact is not yet
refined. Some assess this measurement by tracking business measurements, others assess by
observations, some by surveys, and still others assess by qualitative measures. Last, after six
months or more, evaluators have difficulty solely attributing changed business results to training
when changes in personnel, systems, and other factors might also have contributed to business
performance.

For the evaluation of T& D , the widely used model is KIRKPATRICK"S MODEL.

It consists of four steps-

1.   Reactions
2.   Learning.
3.   Behaviour
4.   Results

[img][/img]


1. Reactions. "Reaction may best be defined as how well the trainees liked a particular training program."
Reactions are typically measured at the end of training -- at Point 3 However, that is a summative or end-of-
course assessment and reactions are also measured during the training, even if only informally in terms of the
instructor's perceptions.


2 Learning. "What principles, facts, and techniques were understood and absorbed by the conferees?" What the
trainees know or can do can be measured during and at the end of training but, in order to say that this
knowledge or skill resulted from the training, the trainees' entering knowledge or skills levels must also be known
or measured. Evaluating learning, then, requires measurements at Points 1, 2 and 3 -- before, during and after
training


3. Behavior. Changes in on-the-job behavior. Clearly, any evaluation of changes in on-the-job behavior must
occur in the workplace itself -- at Point 4 . It should be kept in mind, however, that behavior changes are
acquired in training and they then transfer (or don't transfer) to the work place. It is deemed useful, therefore, to
assess behavior changes at the end of training and in the workplace.

4. Results. "Reduction of costs; reduction of turnover and absenteeism; reduction of grievances; increase in
quality and quantity or production; or improved morale which, it is hoped, will lead to some of the previously
stated results." These factors are also measurable in the workplace -- at Point 4


Hope you got....




                                                   evaluate.gif

     Description:

     Filesize:      4.07 KB
     Viewed:        129 Time(s)
I use a simple form - on side 1 is the name of the employee, name of reporting officer, department name, title,
date and venue of the training, costs of the course, accommodation and travel, and then below this space for
justifying the training - both by the employee and the reporting officer, followed by Side 2 where the objectives
are listed along with how success for each will be measured. Then below this is space to put notes on an after-
training meeting and details of an action plan to take the lessons of the training back to the workplace - actions,
resources needed, responsible person for each action and deadline. I also use this form then to track the financial
costs, to audit the training request and delivery process and to hold reporting officers and employees to account
and see how they are progressing.




Why Measure Training Effectiveness?

Measuring the effectiveness of training programs consumes valuable time
and resources. As we know all too well, these things are in short supply in
organizations today. Why should we bother?

Many training programs fail to deliver the expected organizational benefits.
Having a well-structured measuring system in place can help you determine
where the problem lies. On a positive note, being able to demonstrate a real
and significant benefit to your organization from the training you provide can
help you gain more resources from important decision-makers.

Consider also that the business environment is not standing still. Your
competitors, technology, legislation and regulations are constantly changing.
What was a successful program yesterday may not be a cost-effective
program tomorrow. Being able to measure results will help you adapt to such
changing circumstances.

The Kirkpatrick Model
The most well-known and used model for measuring the effective of training
programs was developed by Donald Kirkpatrick in the late 1950s. It has since
been adapted and modified by a number of writers, however, the basic
structure has well stood the test of time. The basic structure of Kirkpatrick’s
four-level model is shown here.

Figure 1 - Kirkpatrick Model for Evaluating Effectiveness of Training
Programs
  Level 4 - What organizational benefits resulted
Results     from the training?

            To what extent did participants change
  Level 3 -
            their behavior back in the workplace as
Behavior
            a result of the training?

            To what extent did participants improve
  Level 2 -
            knowledge and skills and change
Learning
            attitudes as a result of the training?

  Level 1 - How did                  participants     react   to   the
Reaction    program?

An evaluation at each level answers whether a fundamental requirement of
the training program was met. Its not that conducting an evaluation at one
level is more important that another. All levels of evaluation are important.
In fact, the Kirkpatrick model explains the usefulness of performing
evaluations at each level. Each level provides a diagnostic checkpoint for
problems at the succeeding level. So, if participants did not learn (Level 2),
participant reactions gathered at Level 1 (Reaction) will reveal the barriers to
learning. Now moving up to the next level, if participants did not use the
skills once back in the workplace (Level 3), perhaps they did not learn the
required skills in the first place (Level 2).

The difficulty and cost of conducting an evaluation increases as you move up
the levels. So, you will need to consider carefully what levels of evaluation
you will conduct for which programs. You may decide to conduct Level 1
evaluations (Reaction) for all programs, Level 2 evaluations (Learning) for
“hard-skills” programs only, Level 3 evaluations (Behavior) for strategic
programs only and Level 4 evaluations (Results) for programs costing over
$50,000. Above all else, before starting an evaluation, be crystal clear about
your purpose in conducting the evaluation.

Using the Kirkpatrick Model

How do you conduct an evaluation? Here is a quick guide on some
appropriate information sources for each level.

Level 1 (Reaction)

   •   completed participant feedback questionnaire
   •   informal comments from participants
   •   focus group sessions with participants
Level 2 (Learning)

     •    pre- and post-test scores
     •    on-the-job assessments
     •    supervisor reports


Level 3 (Behavior)

     •    completed self-assessment questionnaire
     •    on-the-job observation
     •    reports from customers, peers and participant’s manager


Level 4 (Results)

     •    financial reports
     •    quality inspections
     •    interview with sales manager


When considering what sources of data you will use for your evaluation, think
about the cost and time involved in collecting the data. Balance this against
the accuracy of the source and the accuracy you actually need. Will existing
sources suffice or will you need to collect new information?

Think broadly about where you can get information. Sources include:

     •    hardcopy and online quantitative reports
     •    production and job records
     •    interviews with participants, managers, peers, customers, suppliers and regulators
     •    checklists and tests
     •    direct observation
     •    questionnaires, self-rating and multi-rating
     •    Focus Group sessions


Once you have completed your evaluation, distribute it to the people who
need to read it. In deciding on your distribution list, refer to your previously
stated reasons for conducting the evaluation. And of course, if there were
lessons learned from the evaluation on how to make your training more
effective, act on them!

Business Performance Pty Ltd products can help you plan and implement
your training evaluation project.

Our Training Management Template Pack contains ready to go checklists,
interview and questionnaire forms for conducting Level 1 (Reaction) and
Level 3 (Behavior) evaluations.

Our eBook From Training to Enhanced Workplace Performance contains
guidance, checklists, interview and questionnaire forms for conducting Level
3 (Behavior) evaluations.

AT+D classic: How to start an Objective Evaluation of your Training Program
T+D, May, 2004 by Donald L. Kirkpatrick

This excerpt is part of a larger article with the same title that originally appeared in the May-June 1956 issue of the
Journal of the American Society of Training Directors, a predecessor to T+D. The article heralded Kirkpatrick's now classic
four-level evaluation model.

Most training men agree that it is important to evaluate training programs. They also feel that the evaluation should be
done by objective means. However, the typical training man uses evaluation sheets or comment sheets as the sole
measure of the effectiveness of his programs. He realizes he should do more, but he just doesn't know how to begin an
objective evaluation.

According the Raymond Katzell, a well-known authority in this field, the evaluation of a training program falls into a
hierarchy of steps that can be briefly stated as follows:

Step One. To determine how the trainees feel about the program.

Step Two. To determine how much the trainees learn in the form of increased knowledge and understanding.

Step Three. To measure the changes in the on-the-job behavior of the trainees.

Step Four. To determine the effects of these behavioral changes on objective criteria such as production, turnover,
absenteeism, and waste.

In climbing this ladder of evaluation, most trainers have completed the first step. Typically, the training director asks the
trainees to fill out evaluation sheets at the end of the program. Questions that are asked most frequently are

* How do you rate the program?

* What subject did you like best?

* What subject did you like least?

* What did you learn that you can use on the job?

* What subjects would you like to have discussed at future programs?

Usually the trainees are not asked to sign their name for fear they will not give an honest reaction.

This kind of subjective evaluation is important. It gives a good indication of how the trainees reacted to the program. If
they react favorably, the trainer can justifiably pat himself on the back and say, "I gave them a program they liked." But
he can't rightfully claim that the training program accomplished the objective, unless his objective was to give them a
program they liked.

The immediate objective of any training course can be stated in terms of the desired knowledge and understanding that
the program is trying to impart to the trainees. It is this stage of evaluation that should be undertaken as the second
step. It is much more difficult than step one and, therefore, is not undertaken by many trainers.

Among the possible methods for determining whether increased knowledge and understanding have taken place, the best
one seems to be the "before and after" paper and pencil test. If the scores on the posttest are significantly higher than on
the pretest, the course can be deemed effective.

In determining the effectiveness of the training, it is important to note that the paper and pencil test or inventory must
cover the principles and facts that are discussed in the course. If the trainer can find a test that covers this material, he
can use it. If he cannot find a suitable one, he must construct his own inventor. Some of the inventories that are
available are: How Supervise? by File and Remmers; Supervisory Inventory by Wesley Osterberg; and the Supervisory
Inventory on Human Relations constructed by this writer.

So far, then, it has been stated that a before and after test can be used to deter mine whether or not increased
knowledge and understanding have taken place. Also, that the inventory should cover the course content. In order to
determine whether or not an available test is suitable, a trainer must examine his course outline and list the principles
and facts he is trying to teach. A comparison of test items with these objectives will reveal whether or not the test can be
used. Because the construction of a test involves such factors as the choice of items, the wording of the items, the
number and type of possible response, and the sequence of items, it is far better to use an available inventory if it covers
most of the course content.

Having selected or constructed a test, the trainer should consider some "Do's" for administering it:

* Give the pretest at the start of the first class and the posttest at the close of the last session. This will minimize the
influence of factors apart from the training course.

* Have the trainee sign both the pretest and posttest. Then, the increased knowledge and understanding can be
computed for each individual.

In instructing the trainees before they take the pretest:

* Tell them it is a before and after procedure.

* Explain the purpose of the test.

* Encourage them to answer truthfully by assuring them that their scores will have no effect on their pay or status in the
company.

* Tell them to answer every question even if they have to guess. (This will be taken into account in the statistical analysis
of scores.)

* Encourage them to take their time in taking the test. This will help to motivate them to read each item carefully.
In analyzing the test results, there are two kinds of evaluations to be made:

* Was the entire course effective as shown by gains from pretest to posttest scores for all trainees?

* What specific facts and principles were learned as shown by changes from pretest to posttest for each item?

Summary
Training men agree that it is advisable to evaluate training courses as objectively as possible. Typically, their evaluation
consists of subjective comment sheets that are completed by the trainees at the end of the course. Providing that these
are properly administered, these evaluation sheets give a valid measure of trainee reaction to the program. However,
they do not give any evidence of benefits derived.

The first step in objectively evaluating the effectiveness of a training course is to determine whether or not the desired
facts and principles were learned by the trainees. This can be done by:

* Using a suitable paper and pencil test.

* Testing the trainees before and after the program.

* Determining the overall effectiveness of the course by comparing pretest and posttest scores for each trainee.

* Determining which specific facts and principles were learned by analyzing the changes on each test item from pretest
and posttest.

The purpose of this article is to suggest a specific technique for beginning an objective evaluation of training program.
Further efforts should be undertaken by every training man to follow up this kind of evaluation by attempting to measure
trainee change in behavior that occurs as a result of participation in the program.

More Related Content

What's hot

The Benefits Of Utilizing Kirkpatrick’S Four Levels Of
The Benefits Of Utilizing Kirkpatrick’S Four Levels OfThe Benefits Of Utilizing Kirkpatrick’S Four Levels Of
The Benefits Of Utilizing Kirkpatrick’S Four Levels Of
wendystein
 
L&D_Metrics-Improvement_20150409
L&D_Metrics-Improvement_20150409L&D_Metrics-Improvement_20150409
L&D_Metrics-Improvement_20150409
Singydan
 
01 15 14 nyu fall 2014 syllabus
01 15 14 nyu fall 2014 syllabus01 15 14 nyu fall 2014 syllabus
01 15 14 nyu fall 2014 syllabus
Craig Rothenberg
 
Manpers Chapter 7 Presentation
Manpers Chapter 7 PresentationManpers Chapter 7 Presentation
Manpers Chapter 7 Presentation
Hitomi Tamura
 
Taedel301 asession4
Taedel301 asession4Taedel301 asession4
Taedel301 asession4
nicvw
 
Designing and Conducting Formative Evaluation
Designing and Conducting Formative EvaluationDesigning and Conducting Formative Evaluation
Designing and Conducting Formative Evaluation
Angel Jones
 
Facilitation of Training Transfer
Facilitation of Training TransferFacilitation of Training Transfer
Facilitation of Training Transfer
Deepika Malhotra
 

What's hot (20)

Training & Development - Validation & Implementation of Training
Training & Development - Validation & Implementation of TrainingTraining & Development - Validation & Implementation of Training
Training & Development - Validation & Implementation of Training
 
KIRKPATRICK MODEL OF EVALUATION (LEO CHANDRA)
KIRKPATRICK MODEL OF EVALUATION (LEO CHANDRA)KIRKPATRICK MODEL OF EVALUATION (LEO CHANDRA)
KIRKPATRICK MODEL OF EVALUATION (LEO CHANDRA)
 
Trg evaluation
Trg evaluationTrg evaluation
Trg evaluation
 
The Benefits Of Utilizing Kirkpatrick’S Four Levels Of
The Benefits Of Utilizing Kirkpatrick’S Four Levels OfThe Benefits Of Utilizing Kirkpatrick’S Four Levels Of
The Benefits Of Utilizing Kirkpatrick’S Four Levels Of
 
L&D_Metrics-Improvement_20150409
L&D_Metrics-Improvement_20150409L&D_Metrics-Improvement_20150409
L&D_Metrics-Improvement_20150409
 
Implementation of training
Implementation of trainingImplementation of training
Implementation of training
 
Chpater 12
Chpater 12Chpater 12
Chpater 12
 
Lorraine Ryan Notes
Lorraine Ryan NotesLorraine Ryan Notes
Lorraine Ryan Notes
 
Training Evaluation
Training EvaluationTraining Evaluation
Training Evaluation
 
Evaluating of Training – HR Training and Development Strategy
Evaluating of Training – HR Training and Development StrategyEvaluating of Training – HR Training and Development Strategy
Evaluating of Training – HR Training and Development Strategy
 
Giving training a purpose
Giving training a purposeGiving training a purpose
Giving training a purpose
 
01 15 14 nyu fall 2014 syllabus
01 15 14 nyu fall 2014 syllabus01 15 14 nyu fall 2014 syllabus
01 15 14 nyu fall 2014 syllabus
 
Manpers Chapter 7 Presentation
Manpers Chapter 7 PresentationManpers Chapter 7 Presentation
Manpers Chapter 7 Presentation
 
Taedel301 asession4
Taedel301 asession4Taedel301 asession4
Taedel301 asession4
 
Krickpatrick basic level of evaluation
Krickpatrick basic level of evaluationKrickpatrick basic level of evaluation
Krickpatrick basic level of evaluation
 
Designing and Conducting Formative Evaluation
Designing and Conducting Formative EvaluationDesigning and Conducting Formative Evaluation
Designing and Conducting Formative Evaluation
 
Kirkpatrick's model
Kirkpatrick's modelKirkpatrick's model
Kirkpatrick's model
 
Facilitation of Training Transfer
Facilitation of Training TransferFacilitation of Training Transfer
Facilitation of Training Transfer
 
Assessment policy in ada school
Assessment policy in ada schoolAssessment policy in ada school
Assessment policy in ada school
 
Training Design - Employee Training and Development PPT
Training Design - Employee Training and Development PPTTraining Design - Employee Training and Development PPT
Training Design - Employee Training and Development PPT
 

Viewers also liked

Principles Of Training
Principles Of TrainingPrinciples Of Training
Principles Of Training
s.munthe
 
9. Evaluation of Training Effectiveness
9. Evaluation of Training Effectiveness9. Evaluation of Training Effectiveness
9. Evaluation of Training Effectiveness
ISEAL Alliance
 
Kirkpatrick 4 level evaluation model
Kirkpatrick 4 level evaluation modelKirkpatrick 4 level evaluation model
Kirkpatrick 4 level evaluation model
zhumin
 
Training & Development (Train The Trainer ) Workshop
Training & Development (Train The Trainer ) WorkshopTraining & Development (Train The Trainer ) Workshop
Training & Development (Train The Trainer ) Workshop
Consultants for Business Leaders
 
Training evaluation ppt 6
Training evaluation   ppt 6Training evaluation   ppt 6
Training evaluation ppt 6
SBMC Jobs
 
Training & Development
Training & Development Training & Development
Training & Development
parags06
 

Viewers also liked (20)

Kirkpatrick model
Kirkpatrick modelKirkpatrick model
Kirkpatrick model
 
Principles Of Training
Principles Of TrainingPrinciples Of Training
Principles Of Training
 
9. Evaluation of Training Effectiveness
9. Evaluation of Training Effectiveness9. Evaluation of Training Effectiveness
9. Evaluation of Training Effectiveness
 
Training & Development A Part Of HRM Studies Er. S Sood
Training & Development A Part Of HRM Studies Er. S SoodTraining & Development A Part Of HRM Studies Er. S Sood
Training & Development A Part Of HRM Studies Er. S Sood
 
Kirkpatrick 4 level evaluation model
Kirkpatrick 4 level evaluation modelKirkpatrick 4 level evaluation model
Kirkpatrick 4 level evaluation model
 
Evaluation of Training & Development
Evaluation of Training & DevelopmentEvaluation of Training & Development
Evaluation of Training & Development
 
Training and Development
Training and DevelopmentTraining and Development
Training and Development
 
Training and Development
Training and DevelopmentTraining and Development
Training and Development
 
Training And Development
Training And DevelopmentTraining And Development
Training And Development
 
Training Evaluation
Training EvaluationTraining Evaluation
Training Evaluation
 
Principles of Training
Principles of TrainingPrinciples of Training
Principles of Training
 
Kirkpatrick's Levels of Training Evaluation - Training and Development
Kirkpatrick's Levels of Training Evaluation - Training and DevelopmentKirkpatrick's Levels of Training Evaluation - Training and Development
Kirkpatrick's Levels of Training Evaluation - Training and Development
 
Training and Development - Principles of Human Resource Management
Training and Development - Principles of Human Resource ManagementTraining and Development - Principles of Human Resource Management
Training and Development - Principles of Human Resource Management
 
Evaluation of training
Evaluation of trainingEvaluation of training
Evaluation of training
 
Training evaluation
Training evaluationTraining evaluation
Training evaluation
 
Training & Development (Train The Trainer ) Workshop
Training & Development (Train The Trainer ) WorkshopTraining & Development (Train The Trainer ) Workshop
Training & Development (Train The Trainer ) Workshop
 
Kirkpatrick4 levels
Kirkpatrick4 levelsKirkpatrick4 levels
Kirkpatrick4 levels
 
Training evaluation ppt 6
Training evaluation   ppt 6Training evaluation   ppt 6
Training evaluation ppt 6
 
Chapter 7 Training And Development
Chapter 7   Training And DevelopmentChapter 7   Training And Development
Chapter 7 Training And Development
 
Training & Development
Training & Development Training & Development
Training & Development
 

Similar to Kirkpatricks four levels_of_evaluation_271

Training &n Development Studies & Evaluation
Training &n Development Studies & EvaluationTraining &n Development Studies & Evaluation
Training &n Development Studies & Evaluation
ElsaCherian1
 
Lecture 7. student evaluation
Lecture 7. student evaluationLecture 7. student evaluation
Lecture 7. student evaluation
Shafiqur Rehman
 
AkdjcijdjcBusiness card 8 Dec 2023.pdfhjdhbcjkewhcjkejckljecjdj
AkdjcijdjcBusiness card 8 Dec 2023.pdfhjdhbcjkewhcjkejckljecjdjAkdjcijdjcBusiness card 8 Dec 2023.pdfhjdhbcjkewhcjkejckljecjdj
AkdjcijdjcBusiness card 8 Dec 2023.pdfhjdhbcjkewhcjkejckljecjdj
RagaviS16
 

Similar to Kirkpatricks four levels_of_evaluation_271 (20)

Kirkpatrick's Four-Level Training Evaluation Model
Kirkpatrick's Four-Level Training Evaluation ModelKirkpatrick's Four-Level Training Evaluation Model
Kirkpatrick's Four-Level Training Evaluation Model
 
Leadership and Management CIPD UK Assignment Sheet
Leadership and Management CIPD UK Assignment SheetLeadership and Management CIPD UK Assignment Sheet
Leadership and Management CIPD UK Assignment Sheet
 
Training Evaluation & Management
Training Evaluation & ManagementTraining Evaluation & Management
Training Evaluation & Management
 
MED07_joycepagkatipunan.pdf
MED07_joycepagkatipunan.pdfMED07_joycepagkatipunan.pdf
MED07_joycepagkatipunan.pdf
 
Training &n Development Studies & Evaluation
Training &n Development Studies & EvaluationTraining &n Development Studies & Evaluation
Training &n Development Studies & Evaluation
 
evaluation of program.pptx
evaluation of program.pptxevaluation of program.pptx
evaluation of program.pptx
 
PRESCRIPTIVE-EVALUATION-GALLA-JOHNNY-G..pptx
PRESCRIPTIVE-EVALUATION-GALLA-JOHNNY-G..pptxPRESCRIPTIVE-EVALUATION-GALLA-JOHNNY-G..pptx
PRESCRIPTIVE-EVALUATION-GALLA-JOHNNY-G..pptx
 
Unit 5- training evalutaion pptx
Unit 5- training evalutaion  pptxUnit 5- training evalutaion  pptx
Unit 5- training evalutaion pptx
 
Training evaluation
Training evaluationTraining evaluation
Training evaluation
 
Lecture 7. student evaluation
Lecture 7. student evaluationLecture 7. student evaluation
Lecture 7. student evaluation
 
Evaluating Training Outcomes.ppt
Evaluating Training Outcomes.pptEvaluating Training Outcomes.ppt
Evaluating Training Outcomes.ppt
 
The Employment Relationship.pptx
The Employment Relationship.pptxThe Employment Relationship.pptx
The Employment Relationship.pptx
 
Kirkspatrick model
Kirkspatrick modelKirkspatrick model
Kirkspatrick model
 
Chapter 11 Training Evaluation.ppt
Chapter 11 Training Evaluation.pptChapter 11 Training Evaluation.ppt
Chapter 11 Training Evaluation.ppt
 
Training program effectiveness a measuring instrument (1)
Training program effectiveness a measuring instrument (1)Training program effectiveness a measuring instrument (1)
Training program effectiveness a measuring instrument (1)
 
Training And Development Evaluation
Training And Development EvaluationTraining And Development Evaluation
Training And Development Evaluation
 
EVALUATION METHODS.pptx
EVALUATION METHODS.pptxEVALUATION METHODS.pptx
EVALUATION METHODS.pptx
 
Training effectiveness
Training effectivenessTraining effectiveness
Training effectiveness
 
AkdjcijdjcBusiness card 8 Dec 2023.pdfhjdhbcjkewhcjkejckljecjdj
AkdjcijdjcBusiness card 8 Dec 2023.pdfhjdhbcjkewhcjkejckljecjdjAkdjcijdjcBusiness card 8 Dec 2023.pdfhjdhbcjkewhcjkejckljecjdj
AkdjcijdjcBusiness card 8 Dec 2023.pdfhjdhbcjkewhcjkejckljecjdj
 
Sales White Paper: Evaluating Sales Training Programs
Sales White Paper: Evaluating Sales Training ProgramsSales White Paper: Evaluating Sales Training Programs
Sales White Paper: Evaluating Sales Training Programs
 

Recently uploaded

1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdf
QucHHunhnh
 
Making and Justifying Mathematical Decisions.pdf
Making and Justifying Mathematical Decisions.pdfMaking and Justifying Mathematical Decisions.pdf
Making and Justifying Mathematical Decisions.pdf
Chris Hunter
 
The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptx
heathfieldcps1
 
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in DelhiRussian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
kauryashika82
 

Recently uploaded (20)

Asian American Pacific Islander Month DDSD 2024.pptx
Asian American Pacific Islander Month DDSD 2024.pptxAsian American Pacific Islander Month DDSD 2024.pptx
Asian American Pacific Islander Month DDSD 2024.pptx
 
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
 
Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdf
 
ICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptx
 
Grant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy ConsultingGrant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy Consulting
 
psychiatric nursing HISTORY COLLECTION .docx
psychiatric  nursing HISTORY  COLLECTION  .docxpsychiatric  nursing HISTORY  COLLECTION  .docx
psychiatric nursing HISTORY COLLECTION .docx
 
Python Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docxPython Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docx
 
This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.
 
Making and Justifying Mathematical Decisions.pdf
Making and Justifying Mathematical Decisions.pdfMaking and Justifying Mathematical Decisions.pdf
Making and Justifying Mathematical Decisions.pdf
 
PROCESS RECORDING FORMAT.docx
PROCESS      RECORDING        FORMAT.docxPROCESS      RECORDING        FORMAT.docx
PROCESS RECORDING FORMAT.docx
 
Web & Social Media Analytics Previous Year Question Paper.pdf
Web & Social Media Analytics Previous Year Question Paper.pdfWeb & Social Media Analytics Previous Year Question Paper.pdf
Web & Social Media Analytics Previous Year Question Paper.pdf
 
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
 
On National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsOn National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan Fellows
 
The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptx
 
Energy Resources. ( B. Pharmacy, 1st Year, Sem-II) Natural Resources
Energy Resources. ( B. Pharmacy, 1st Year, Sem-II) Natural ResourcesEnergy Resources. ( B. Pharmacy, 1st Year, Sem-II) Natural Resources
Energy Resources. ( B. Pharmacy, 1st Year, Sem-II) Natural Resources
 
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in DelhiRussian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
 
How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17
 
Role Of Transgenic Animal In Target Validation-1.pptx
Role Of Transgenic Animal In Target Validation-1.pptxRole Of Transgenic Animal In Target Validation-1.pptx
Role Of Transgenic Animal In Target Validation-1.pptx
 
Food Chain and Food Web (Ecosystem) EVS, B. Pharmacy 1st Year, Sem-II
Food Chain and Food Web (Ecosystem) EVS, B. Pharmacy 1st Year, Sem-IIFood Chain and Food Web (Ecosystem) EVS, B. Pharmacy 1st Year, Sem-II
Food Chain and Food Web (Ecosystem) EVS, B. Pharmacy 1st Year, Sem-II
 

Kirkpatricks four levels_of_evaluation_271

  • 1. Kirkpatrick's Four Levels of Evaluation ASSESSING TRAINING EFFECTIVENESS often entails using the four-level model developed by Donald Kirkpatrick (1994). According to this model, evaluation should always begin with level one, and then, as time and budget allows, should move sequentially through levels two, three, and four. Information from each prior level serves as a base for the next level's evaluation. Thus, each successive level represents a In Kirkpatrick's four-level model, more precise measure of the effectiveness of the training each successive evaluation level program, but at the same time requires a more rigorous and is built on information provided time-consuming analysis. by the lower level. Level 1 Evaluation - Reactions Just as the word implies, evaluation at this level measures how participants in a training program react to it. It attempts to answer questions regarding the participants' perceptions - Did they like it? Was the material relevant to their work? This type of evaluation is often called a “smilesheet.” According to Kirkpatrick, every program should at least be evaluated at this level to provide for the improvement of a training program. In addition, the participants' reactions have important consequences for learning (level two). Although a positive reaction does not guarantee learning, a negative reaction almost certainly reduces its possibility. Level 2 Evaluation - Learning Assessing at this level moves the evaluation beyond learner satisfaction and attempts to assess the extent students have advanced in skills, knowledge, or attitude. Measurement at this level is more difficult and laborious than level one. Methods range from formal to informal testing to team assessment and self-assessment. If possible, participants take the test or assessment before the training (pretest) and after training (post test) to determine the amount of learning that has occurred. Level 3 Evaluation - Transfer To assess the amount of learning that has occurred due to a training This level measures the transfer that has occurred in program, level two evaluations often learners' behavior due to the training program. Evaluating use tests conducted before training at this level attempts to answer the question - Are the (pretest) and after training (post test). newly acquired skills, knowledge, or attitude being used in the everyday environment of the learner? For many trainers this level represents the truest assessment of a program's effectiveness. However, measuring at this level is difficult as it is often impossible to predict when the change in behavior will occur, and thus requires important decisions in terms of when to evaluate, how often to evaluate, and how to evaluate. Level 4 Evaluation- Results
  • 2. Frequently thought of as the bottom line, this level measures the success of the program in terms that managers and executives can understand -increased production, improved quality, decreased costs, reduced frequency of accidents, increased sales, and even higher profits or return on investment. From a business and organizational perspective, this is the overall reason for a training program, yet level four results are not typically addressed. Determining Level four evaluation attempts to assess results in financial terms is difficult to measure, and is training in terms of business results. In hard to link directly with training. this case, sales transactions improved steadily after training for sales staff Methods for Long-Term Evaluation occurred in April 1997. • Send post-training surveys • Offer ongoing, sequenced training and coaching over a period of time • Conduct follow-up needs assessment • Check metrics (e.g., scrap, re-work, errors, etc.) to measure if participants achieved training objectives • Interview trainees and their managers, or their customer groups (e.g., patients, other departmental staff) Summary of the Kirkpatrick Model Donald Kirkpatrick first proposed this four-pronged approach to evaluating training programs in his 1959 doctoral dissertation. Since then, it has become so widely used, that trainers can typically talk about it in shorthand and understand the reference. For example, when one trainer says to another, "What are you doing about level IV?" the other knows that the first trainer wants to understand how the second evaluates the impact of training. Level Name Issues Assessed at this Level I Reaction Assesses participants’ initial reactions to a course. This, in turn, offers insights into participants’ satisfaction with a course, a perception of value. Trainers usually assess this through a survey, often called a "smiley sheet." Occasionally, trainers use focus groups and similar methods to receive more specific comments (called qualitative feedback) on the courses. According to the TRAINING magazine annual industry survey, almost 100 percent of all trainers perform "Level I" evaluation. II Learning Assesses the amount of information that participants learned. Trainers usually assess this with a criterion-referenced test. The criteria are objectives for the course: statements developed before a course is developed that explicitly state the skills that participants should be able to perform after taking a course. Because the objectives are the requirements for the course, a Level II evaluation assesses conformance to requirements, or quality. III Transfer Assesses the amount of material that participants actually use in everyday work 6 weeks to 6 months (perhaps longer) after taking the course. This assessment is based on the objectives of
  • 3. the course and assessed through tests, observations, surveys, and interviews with co-workers and supervisors. Like the Level II evaluation, Level III assesses the requirements of the course and can be viewed as a follow-on assessment of quality. IV Impact Assesses the financial impact of the training course on the bottom line of the organization 6 months to 2 years after the course (the actual time varies depending on the context of the course). For many reasons, Level IV is the most difficult level to measure. First, most training courses do not have explicitly written business objectives, such as "this course should reduce support expenses by 20 percent." Second, the methodology for assessing business impact is not yet refined. Some assess this measurement by tracking business measurements, others assess by observations, some by surveys, and still others assess by qualitative measures. Last, after six months or more, evaluators have difficulty solely attributing changed business results to training when changes in personnel, systems, and other factors might also have contributed to business performance. For the evaluation of T& D , the widely used model is KIRKPATRICK"S MODEL. It consists of four steps- 1. Reactions 2. Learning. 3. Behaviour 4. Results [img][/img] 1. Reactions. "Reaction may best be defined as how well the trainees liked a particular training program." Reactions are typically measured at the end of training -- at Point 3 However, that is a summative or end-of- course assessment and reactions are also measured during the training, even if only informally in terms of the instructor's perceptions. 2 Learning. "What principles, facts, and techniques were understood and absorbed by the conferees?" What the trainees know or can do can be measured during and at the end of training but, in order to say that this knowledge or skill resulted from the training, the trainees' entering knowledge or skills levels must also be known or measured. Evaluating learning, then, requires measurements at Points 1, 2 and 3 -- before, during and after training 3. Behavior. Changes in on-the-job behavior. Clearly, any evaluation of changes in on-the-job behavior must occur in the workplace itself -- at Point 4 . It should be kept in mind, however, that behavior changes are acquired in training and they then transfer (or don't transfer) to the work place. It is deemed useful, therefore, to assess behavior changes at the end of training and in the workplace. 4. Results. "Reduction of costs; reduction of turnover and absenteeism; reduction of grievances; increase in quality and quantity or production; or improved morale which, it is hoped, will lead to some of the previously stated results." These factors are also measurable in the workplace -- at Point 4 Hope you got.... evaluate.gif Description: Filesize: 4.07 KB Viewed: 129 Time(s)
  • 4. I use a simple form - on side 1 is the name of the employee, name of reporting officer, department name, title, date and venue of the training, costs of the course, accommodation and travel, and then below this space for justifying the training - both by the employee and the reporting officer, followed by Side 2 where the objectives are listed along with how success for each will be measured. Then below this is space to put notes on an after- training meeting and details of an action plan to take the lessons of the training back to the workplace - actions, resources needed, responsible person for each action and deadline. I also use this form then to track the financial costs, to audit the training request and delivery process and to hold reporting officers and employees to account and see how they are progressing. Why Measure Training Effectiveness? Measuring the effectiveness of training programs consumes valuable time and resources. As we know all too well, these things are in short supply in organizations today. Why should we bother? Many training programs fail to deliver the expected organizational benefits. Having a well-structured measuring system in place can help you determine where the problem lies. On a positive note, being able to demonstrate a real and significant benefit to your organization from the training you provide can help you gain more resources from important decision-makers. Consider also that the business environment is not standing still. Your competitors, technology, legislation and regulations are constantly changing. What was a successful program yesterday may not be a cost-effective program tomorrow. Being able to measure results will help you adapt to such changing circumstances. The Kirkpatrick Model
  • 5. The most well-known and used model for measuring the effective of training programs was developed by Donald Kirkpatrick in the late 1950s. It has since been adapted and modified by a number of writers, however, the basic structure has well stood the test of time. The basic structure of Kirkpatrick’s four-level model is shown here. Figure 1 - Kirkpatrick Model for Evaluating Effectiveness of Training Programs Level 4 - What organizational benefits resulted Results from the training? To what extent did participants change Level 3 - their behavior back in the workplace as Behavior a result of the training? To what extent did participants improve Level 2 - knowledge and skills and change Learning attitudes as a result of the training? Level 1 - How did participants react to the Reaction program? An evaluation at each level answers whether a fundamental requirement of the training program was met. Its not that conducting an evaluation at one level is more important that another. All levels of evaluation are important. In fact, the Kirkpatrick model explains the usefulness of performing evaluations at each level. Each level provides a diagnostic checkpoint for problems at the succeeding level. So, if participants did not learn (Level 2), participant reactions gathered at Level 1 (Reaction) will reveal the barriers to learning. Now moving up to the next level, if participants did not use the skills once back in the workplace (Level 3), perhaps they did not learn the required skills in the first place (Level 2). The difficulty and cost of conducting an evaluation increases as you move up the levels. So, you will need to consider carefully what levels of evaluation you will conduct for which programs. You may decide to conduct Level 1 evaluations (Reaction) for all programs, Level 2 evaluations (Learning) for “hard-skills” programs only, Level 3 evaluations (Behavior) for strategic programs only and Level 4 evaluations (Results) for programs costing over $50,000. Above all else, before starting an evaluation, be crystal clear about your purpose in conducting the evaluation. Using the Kirkpatrick Model How do you conduct an evaluation? Here is a quick guide on some appropriate information sources for each level. Level 1 (Reaction) • completed participant feedback questionnaire • informal comments from participants • focus group sessions with participants
  • 6. Level 2 (Learning) • pre- and post-test scores • on-the-job assessments • supervisor reports Level 3 (Behavior) • completed self-assessment questionnaire • on-the-job observation • reports from customers, peers and participant’s manager Level 4 (Results) • financial reports • quality inspections • interview with sales manager When considering what sources of data you will use for your evaluation, think about the cost and time involved in collecting the data. Balance this against the accuracy of the source and the accuracy you actually need. Will existing sources suffice or will you need to collect new information? Think broadly about where you can get information. Sources include: • hardcopy and online quantitative reports • production and job records • interviews with participants, managers, peers, customers, suppliers and regulators • checklists and tests • direct observation • questionnaires, self-rating and multi-rating • Focus Group sessions Once you have completed your evaluation, distribute it to the people who need to read it. In deciding on your distribution list, refer to your previously stated reasons for conducting the evaluation. And of course, if there were lessons learned from the evaluation on how to make your training more effective, act on them! Business Performance Pty Ltd products can help you plan and implement your training evaluation project. Our Training Management Template Pack contains ready to go checklists, interview and questionnaire forms for conducting Level 1 (Reaction) and Level 3 (Behavior) evaluations. Our eBook From Training to Enhanced Workplace Performance contains guidance, checklists, interview and questionnaire forms for conducting Level 3 (Behavior) evaluations. AT+D classic: How to start an Objective Evaluation of your Training Program T+D, May, 2004 by Donald L. Kirkpatrick This excerpt is part of a larger article with the same title that originally appeared in the May-June 1956 issue of the
  • 7. Journal of the American Society of Training Directors, a predecessor to T+D. The article heralded Kirkpatrick's now classic four-level evaluation model. Most training men agree that it is important to evaluate training programs. They also feel that the evaluation should be done by objective means. However, the typical training man uses evaluation sheets or comment sheets as the sole measure of the effectiveness of his programs. He realizes he should do more, but he just doesn't know how to begin an objective evaluation. According the Raymond Katzell, a well-known authority in this field, the evaluation of a training program falls into a hierarchy of steps that can be briefly stated as follows: Step One. To determine how the trainees feel about the program. Step Two. To determine how much the trainees learn in the form of increased knowledge and understanding. Step Three. To measure the changes in the on-the-job behavior of the trainees. Step Four. To determine the effects of these behavioral changes on objective criteria such as production, turnover, absenteeism, and waste. In climbing this ladder of evaluation, most trainers have completed the first step. Typically, the training director asks the trainees to fill out evaluation sheets at the end of the program. Questions that are asked most frequently are * How do you rate the program? * What subject did you like best? * What subject did you like least? * What did you learn that you can use on the job? * What subjects would you like to have discussed at future programs? Usually the trainees are not asked to sign their name for fear they will not give an honest reaction. This kind of subjective evaluation is important. It gives a good indication of how the trainees reacted to the program. If they react favorably, the trainer can justifiably pat himself on the back and say, "I gave them a program they liked." But he can't rightfully claim that the training program accomplished the objective, unless his objective was to give them a program they liked. The immediate objective of any training course can be stated in terms of the desired knowledge and understanding that the program is trying to impart to the trainees. It is this stage of evaluation that should be undertaken as the second step. It is much more difficult than step one and, therefore, is not undertaken by many trainers. Among the possible methods for determining whether increased knowledge and understanding have taken place, the best one seems to be the "before and after" paper and pencil test. If the scores on the posttest are significantly higher than on the pretest, the course can be deemed effective. In determining the effectiveness of the training, it is important to note that the paper and pencil test or inventory must cover the principles and facts that are discussed in the course. If the trainer can find a test that covers this material, he can use it. If he cannot find a suitable one, he must construct his own inventor. Some of the inventories that are available are: How Supervise? by File and Remmers; Supervisory Inventory by Wesley Osterberg; and the Supervisory Inventory on Human Relations constructed by this writer. So far, then, it has been stated that a before and after test can be used to deter mine whether or not increased knowledge and understanding have taken place. Also, that the inventory should cover the course content. In order to determine whether or not an available test is suitable, a trainer must examine his course outline and list the principles and facts he is trying to teach. A comparison of test items with these objectives will reveal whether or not the test can be used. Because the construction of a test involves such factors as the choice of items, the wording of the items, the number and type of possible response, and the sequence of items, it is far better to use an available inventory if it covers most of the course content. Having selected or constructed a test, the trainer should consider some "Do's" for administering it: * Give the pretest at the start of the first class and the posttest at the close of the last session. This will minimize the influence of factors apart from the training course. * Have the trainee sign both the pretest and posttest. Then, the increased knowledge and understanding can be computed for each individual. In instructing the trainees before they take the pretest: * Tell them it is a before and after procedure. * Explain the purpose of the test. * Encourage them to answer truthfully by assuring them that their scores will have no effect on their pay or status in the company. * Tell them to answer every question even if they have to guess. (This will be taken into account in the statistical analysis of scores.) * Encourage them to take their time in taking the test. This will help to motivate them to read each item carefully.
  • 8. In analyzing the test results, there are two kinds of evaluations to be made: * Was the entire course effective as shown by gains from pretest to posttest scores for all trainees? * What specific facts and principles were learned as shown by changes from pretest to posttest for each item? Summary Training men agree that it is advisable to evaluate training courses as objectively as possible. Typically, their evaluation consists of subjective comment sheets that are completed by the trainees at the end of the course. Providing that these are properly administered, these evaluation sheets give a valid measure of trainee reaction to the program. However, they do not give any evidence of benefits derived. The first step in objectively evaluating the effectiveness of a training course is to determine whether or not the desired facts and principles were learned by the trainees. This can be done by: * Using a suitable paper and pencil test. * Testing the trainees before and after the program. * Determining the overall effectiveness of the course by comparing pretest and posttest scores for each trainee. * Determining which specific facts and principles were learned by analyzing the changes on each test item from pretest and posttest. The purpose of this article is to suggest a specific technique for beginning an objective evaluation of training program. Further efforts should be undertaken by every training man to follow up this kind of evaluation by attempting to measure trainee change in behavior that occurs as a result of participation in the program.