Building A World-Class Quality Assurance Program

Contact Center QA Guide:
Building a World-Class Quality Assurance Program
Sponsored by:
About DMG Consulting LLC DMG Consulting LLC is the leading analyst firm in the contact center and analytics markets. We are a strategic advisor to companies large and small. Our mission is to help companies build effective contact center and service environments that provide an outstanding customer experience. We achieve this goal by helping our clients leverage technology, process and people to optimize operational efficiency, sales and profits. Our actionable strategies and tactics effect change that enable companies to make strategic and tactical improvements with a rapid return on investment. Our customers include Global 2000 companies, government agencies, non‐profit organizations, and more than 150 contact center and analytics vendors. DMG Consulting LLC’s business strategists have an average of 20 years experience in customer experience management (CEM), customer relationship management (CRM), contact centers, building customer‐focused businesses and financial analysis. We understand the power of customer data and the contributions that quality management, recording, speech analytics, performance management, surveying, eLearning, coaching and workforce management systems make to the contact center and the entire enterprise. Our consulting experience with more than 2000 end‐user organizations and our hands‐
on operational, technology and financial expertise give us deep insight into what customers want and need from enterprises and vendors. All rights reserved. This report is protected by United States copyright law and may not be copied, reproduced, displayed, published, transmitted or distributed in whole or in part in any form or medium without the express written permission of DMG Consulting LLC. You may not alter or remove any trademark, copyright or other notice from this report. Reproduction in whole or in part in any form or medium without express written permission of DMG Consulting LLC is prohibited. Substantial effort went into verifying and validating the accuracy of the information contained within this Report, however, DMG Consulting LLC makes no warranties as to the accuracy or completeness of this information. DMG Consulting LLC is not liable for any damages, consequential or otherwise, arising from use of this information. Copyright © 2009 DMG Consulting LLC © 2009 DMG Consulting LLC
All rights reserved.
-i-
April 2009
Table of Contents Executive Summary .....................................................................................................1
1.
Introduction .....................................................................................................2
2.
What is Contact Center Quality Assurance? ......................................................3
3.
Quality Assurance Program Benefits ................................................................6
4.
Components of a QA Program.......................................................................... 7
5.
Getting Started ..............................................................................................10
5.1 Creating a Quality Assurance Leadership Team .............................................. 15
5.2 Staffing QA Positions .....................................................................................16
5.3 Implementation Roadmap..............................................................................18
6.
Developing a QA Evaluation Form .................................................................. 21
6.1 Form Categories.............................................................................................24
6.2 QA Evaluation Form Questions.......................................................................25
6.3 Assigning Weights to QA Evaluation Forms ....................................................26
6.4 Validation....................................................................................................... 27
6.5 Example QA Evaluation Forms .......................................................................29
7.
Quality Assurance Monitoring Criteria and Guidelines ....................................48
7.1 Quality Monitoring Criteria and Guidelines .....................................................48
7.2 Why are Quality Monitoring Criteria and Guidelines Necessary? .....................49
8.
Calibration......................................................................................................50
8.1 What is Calibration?........................................................................................50
8.2 Benefits of Calibration.................................................................................... 51
8.3 The Calibration Process ..................................................................................52
9. Quality Assurance Program Mechanics and Processes ....................................54
9.1 Transaction Selection Criteria.........................................................................54
9.2 Determining the Number and Frequency of Evaluations ................................. 55
9.3 Coaching ........................................................................................................56
9.4 Evaluation Feedback and Escalation Process ..................................................58
9.5 Ongoing Training ...........................................................................................59
9.6 Addressing Agent Performance Issues............................................................60
9.7 Rewards and Recognition ...............................................................................63
9.8 Updating Procedures/Training........................................................................64
© 2009 DMG Consulting LLC
All rights reserved.
- ii -
April 2009
9.9 Monitoring Quality Assurance Reviewers.......................................................65
9.10 Reporting ....................................................................................................... 67
9.11 QA Database .................................................................................................. 74
10.
Kicking Off the QA Program ........................................................................... 75
10.1 Agent Training ............................................................................................... 76
10.2 QA Program Pilot ........................................................................................... 78
11.
Advanced Quality Assurance Initiatives .......................................................... 79
11.1 Surveying ....................................................................................................... 79
11.2 Customer Experience Monitoring ................................................................... 79
11.3 First Call Resolution (FCR)...............................................................................80
12.
Quality Management/Liability Recording Suites .............................................81
Appendix A: Procedure Format Sample .....................................................................85
Appendix B: Policy Format Sample ............................................................................88
© 2009 DMG Consulting LLC
All rights reserved.
- iii -
April 2009
Table of Figures Figure 1: Quality Assurance Process .............................................................................5 Figure 2: Quality Assurance Program Development Roadmap....................................10 Figure 3: QA Implementation Roadmap .....................................................................19 Figure 4: Common Quality Evaluation Form Sections/Categories ...............................24 Figure 5: Financial Services Customer Service QA Evaluation Form ............................30 Figure 6: Healthcare Customer Service QA Evaluation Form.......................................34 Figure 7: Technical Support QA Evaluation Form ........................................................38 Figure 8: Sample Customer Service Quality Monitoring Form.....................................42 Figure 9: New Order Precision Quality Monitoring Form.............................................45 Figure 10: The Calibration Process............................................................................... 53 Figure 11: Coaching Methods ......................................................................................56 Figure 12: Average Contact Center QA Score: December 2008.................................... 67 Figure 13:Contact Center Average Scores by Evaluation Section: December 2008.......68 Figure 14: Average Scores by Evaluation Question: December 2008............................ 70 Figure 15: Average QA Scores by Agent: December 2008............................................ 72 Figure 16: QA Scores by Agent by Evaluation: December 2008.................................... 72 Figure 17: Average QA Scores by Agent by Category: December 2008 ........................ 73 Figure 18: QA Agent Training Outline.......................................................................... 76 Figure 19: Workforce Optimization Suites ...................................................................82 © 2009 DMG Consulting LLC
All rights reserved.
- iv -
April 2009
Executive Summary Contact center quality assurance (QA) is an ongoing business activity that provides ter QA Guide Contac t Cennter valuable insights into agent, customer and contact center service delivery performance and opportunities. The primary goal of any QA program is to verify and ensure that a company’s contact center services are being performed in a manner that meets or exceeds internal requirements and customer expectations. At a fundamental level, QA programs measure how well agents comply with internal policies and procedures and interact with customers in phone, email and chat sessions. The more advanced QA programs go well beyond these basics. They combine the results of customer satisfaction surveys with internal measurements to provide a 360‐degree view of the customer experience. Well‐designed and effective QA programs demonstrate a company’s commitment to its customers and agents, and are essential for building world‐class contact centers. This Guide is a “cookbook” for building a strong, effective and well‐received QA program. It is ideal for managers and supervisors in contact centers of any size who are either building their first quality assurance program or want to enhance an existing program. © 2009 DMG Consulting LLC
All rights reserved.
-1-
April 2009
1. Introduction Quality assurance isn’t an option for contact centers; it’s essential for the success of the nter ter QA Guide
Contac t Cen
contact center, customer and agent satisfaction, improving agent and supervisor productivity and effectiveness, and keeping management in touch with their staff’s performance. To achieve the best results and foster confidence in the program, managers must ensure that they are evaluating the right components of agent performance during customer interactions and using appropriate measurements and weights. Building an effective QA program is an iterative, multi‐step process that requires senior management support, planning and input, and buy‐in from all levels of contact center staff. Automation is helpful for formalizing, standardizing and institutionalizing the initiative, but QA a program will succeed only if the staff is on board and believes in its value. © 2009 DMG Consulting LLC
All rights reserved.
-2-
April 2009
2. What is Contact Center Quality Assurance? Contact center quality assurance, also known as Quality Management (QM), is a process where managers, supervisors and QA specialists monitor and evaluate how well agents handle customer transactions. The monitoring process includes a simultaneous review of a call and the nter ter QA Guide
Contac t Cen
system screens used to handle the interaction. (If QA is being done on an email or a chat session, it will review just the servicing screens and the content of the agent’s response.) QA can be done in a real‐time mode where the supervisor or QA specialist “live monitors” calls. For live monitoring QA, reviewers can either access calls through their automatic call distributor (ACD) technology, or can sit next to agents and watch and evaluate them as they handle calls. QA is often performed on recorded transactions. There are pros and cons to both live monitoring and using recorded transactions. When live monitoring, the QA specialist/supervisor can provide immediate feedback to the agent. This can become an effective coaching opportunity, as the feedback is provided in real‐time when the call is still fresh in the agent’s mind. On the other hand, doing QA from recorded transactions allows an organization to schedule calls, making the process more efficient for the reviewer. It allows QA specialists/supervisors to find interactions that require their attention, either because they are really good or really bad, instead of wasting time on satisfactory calls. It allows the reviewer to be able to carefully review the interaction/screens, including the ability to go back and re‐review a portion of the interaction. And, lastly, it is an anonymous process, so agents are not aware that they are being evaluated and will not necessarily be on their best behavior. When doing QA, the reviewer completes an evaluation form that measures how the agent handled each component of the call or online transaction. The form can be on paper or part of a QA system. The fundamental purpose of QA programs is to measure how well agents adhere to contact center departmental policies and procedures. Contact/call center managers have traditionally live monitored or listened to recorded interactions in order to gauge agent performance and internally evaluate the customer experience. To be effective, the program should be a formalized ongoing process designed to: © 2009 DMG Consulting LLC
All rights reserved.
-3-
April 2009
1.
Measure agent adherence to internal policies and procedures 2.
Improve consistency and quality of customer interactions across all channels (telephone, email, chat/IM, etc.) 3.
Assess business execution – detect and fix broken or inefficient nter ter QA Guide
Contac t Cen
policies, processes or operational issues throughout the company 4.
Improve agent performance 5.
Identify agent training needs 6.
Identify policies or processes that frustrate and alienate customers 7.
Maximize every customer interaction 8.
Identify business trends 9.
Improve the customer experience The underpinnings of a good QA program are consistency, accurate measurement, and a cycle of continuous feedback resulting in improvements. An effective QA program provides the contact center with a vehicle for measuring the quality and consistency of service delivery, capturing customer insights, and identifying trends, service issues and training/coaching opportunities to improve agent performance and productivity. Quality assurance is a dynamic and iterative process that must be adapted as a business changes. See Figure 1. © 2009 DMG Consulting LLC
All rights reserved.
-4-
April 2009
nter ter QA Guide
Contact Cen
Figure 1: Quality Assurance Process Source: DMG Consulting LLC, April 2009 Action Item: Identify the company’s reasons for performing contact center quality assurance. Build a program that delivers continuous feedback to the contact center and other operating departments. © 2009 DMG Consulting LLC
All rights reserved.
-5-
April 2009
3. Quality Assurance Program Benefits When quality assurance programs are well designed and used consistently, they yield great nter ter QA Guide
Contac t Cen
benefits for customers, enterprises, contact centers and agents. The benefits include: • Improved agent productivity, by reducing the average handle time of transactions • Reduced operating expenses • Better transaction quality • Increased customer satisfaction • Enhanced customer experience • Identification of business opportunities, cross‐sell, up‐sell, new products and services • Enhanced operating policies and procedures • Reduced enterprise risk • Improved agent satisfaction and reduced attrition • Automated reporting for tracking and trending Action Item: Set up a process for capturing, quantifying and reporting the benefits from your QA program. Be sure to share successes with agents, supervisors, managers and senior executives on an ongoing basis, so that everyone appreciates the program’s contributions. © 2009 DMG Consulting LLC
All rights reserved.
-6-
April 2009
4. Components of a QA Program While every company should customize their QA program to reflect their corporate culture and values, the fundamental building blocks for contact center QA are standard across all companies. Effective QA programs should include the following components: nter ter QA Guide
Contac t Cen
•
Procedures and policies: Document all existing transaction procedures and policies so that the contact center staff knows the appropriate steps for handling all types of inquiries. QA specialists/supervisors can also use the procedures and policies as a standard to make sure that they are evaluating different transactions properly. It’s important to set up a process that keeps the policies and procedures up‐to‐date. •
Quality monitoring criteria and guidelines: Specify the criteria to apply when evaluating transactions and performing evaluations. The criteria should define what QA specialists/supervisors are looking for in each type of contact center transaction. The easiest way to create criteria is to use the contact center’s documented procedures and policies and note on each the most important aspects for each type of transaction. It’s also a good idea to identify agent actions that would cause them to lose points in a quality evaluation. •
Program mechanics: Define the mechanics of the QA program, including who (manager, supervisor, team leader, QA specialist, trainer) is responsible for conducting evaluations, the number and frequency of evaluations, how many coaching sessions per agent per month, and how to select calls, emails and chat sessions for evaluation. (If a QA application is being used, most of the mechanics will be automated, including the transaction selection process.) •
Training: Establish a closed‐loop training process that addresses new content, system issues, updates and agent performance issues. The trainers and QA staff must work closely together to ensure that the staff is fully trained. (In many small/mid‐sized contact centers, the same people do both QA and training.) Prior to kicking off a QA program, all contact center staff – agents, supervisors, QA specialists, trainers, managers – must be fully trained so that they know how to © 2009 DMG Consulting LLC
All rights reserved.
-7-
April 2009
handle all types of customer interactions. If new or enhanced procedures and policies are drafted to support the QA program, they should be reviewed with the staff before starting the QA program. In addition to agent procedural training, it’s important to build a training program that introduces the new or enhanced QA program to the staff. The more informed the staff is about the program, the more effective the initiative. nter ter QA Guide
Contac t Cen
•
Coaching: Provide frequent feedback to agents about their performance. Feedback should address where agents are performing well and areas where they have opportunities to improve. Coaching is one of the critical success factors in QA programs and plays a very important role in agent satisfaction and retention. However, as it can be very challenging to provide negative feedback to agents, it’s important to train the QA staff to produce effective coaching sessions and to make sure that management delivers coaching sessions consistently. •
Calibration: Build consistency into the QA program through calibration. Calibration is the process of teaching all people involved in performing QA evaluations how to score transactions on a consistent and equitable basis. To make a QA program fair for agents, it’s essential for all QA reviewers to agree on the meaning and value of each question in a monitoring form. To achieve consistency, it’s important to run calibration sessions where all reviewers listen to the same call, score it, identify variance in scoring approaches, reconcile their differences, and set a standard measurement that all will use going forward. The only way to reach consensus is to run calibration sessions. Calibration is an ongoing process and should be run on a monthly basis. It’s also a good idea to involve agents in the calibration process so that they can appreciate the challenges associated with consistently evaluating transactions. •
Evaluation feedback: Provide a process to facilitate two‐way communication between agents and reviewers. Agents need to have a mechanism for responding to their quality evaluations so that they feel empowered and not “put upon.” They also need a formal process for filing complaints when they believe that a QA reviewer is not being fair or is not listening to their input. QA reviewers should welcome discourse, as it will ultimately yield a better and more effective program. © 2009 DMG Consulting LLC
All rights reserved.
-8-
April 2009
•
Rewards and recognition: Recognizing and rewarding top performers is essential for the success of a QA program. While agents should deliver outstanding performance because it’s nter ter QA Guide
Contac t Cen
their job, recognizing when they do encourages them to keep up the good work and motivates others to strive for recognition, as well. Rewards do not have to be large; they could include a plaque, a parking spot, lunch with the CEO, a gift card, movie tickets, etc. Action Item: Implement an “Executive QA” program, where senior managers from sales, marketing, operations and all other supporting areas sit with contact center agents as they handle calls. This has proven to create customer‐focused awareness and foster collaboration between departments. (Rapid process change is facilitated when senior executives hear first‐
hand the impact of their processes and programs on customers.) When senior managers take this program seriously, it has a very positive impact on agent morale and satisfaction because it underscores the importance of their job. © 2009 DMG Consulting LLC
All rights reserved.
-9-
April 2009
5. Getting Started Figure 2 depicts a high‐level project plan for building a QA program. This figure provides the nter ter QA Guide
Contact Cen
high‐level steps that each organization should customize to meet the needs of their company. The first step in the process is identifying and assigning resources to oversee and manage the program on an ongoing basis. This step is critical because every company needs a champion to push the program through, or it is unlikely to happen. (The project champion often becomes the manager or supervisor responsible for the program.) It takes a significant amount of work to build a contact center QA program and there must be an individual responsible for making sure that everyone involved in the program is motivated and the project stays on track and on time. Figure 2: Quality Assurance Program Development Roadmap 11
Implement the formal QA program
.
10
9
8
7
6
5
4
3
2
1
Adjust program and repeat pilot as needed
Pilot the program for 2 months and review results
Train all staff members on the Quality Assurance program
Build Quality Assurance training program
Define the coaching and feedback process
Develop quality monitoring criteria and guidelines
Test the form using real calls, conduct calibration and refine the form as needed
Develop the evaluation forms; determine section and question weights
Communicate to all staff that a formal Quality Assurance program is under development
Identify QA manager and reviewer staff
Source: DMG Consulting LLC, April 2009 © 2009 DMG Consulting LLC
All rights reserved.
- 10 -
April 2009
The 11 high‐level steps for building a QA program are explained below. It’s important to note that departmental procedures for handling all types of customer transactions and inquiries should be drafted or enhanced prior to developing the QA process. Step 1: Identify resources from the contact center to serve as the QA manager and quality nter ter QA Guide
Contac t Cen
review staff. (The QA manager will also play a major role in building or enhancing the program.) Ideally the manager should have prior experience in setting up and running a QA program and conducting contact center QA evaluations. Minimally, the QA reviewer(s) must demonstrate complete and accurate knowledge of all contact center policies, procedures and systems, as well as having excellent interpersonal and coaching skills. The number of resources dedicated to the process varies based on the size of the contact center. Step 2: The institution of a formal QA program should be communicated to all staff members as early as possible. The success of the program depends upon agent cooperation, which will be enhanced if agents are invited to participate in program development and all changes are clearly communicated. Communication should include information about why the program is being developed, roles and responsibilities, impact on agents (their reviews, raises, and incentives), and program benefits. Step 3: Develop the quality evaluation forms. Typically, managers create one for each channel supported (phone, email, chat), and/or one for different call classifications, i.e. service, sales, collections, fraud, etc. Once the number and types of forms are decided, determine the sections required in each form, i.e., opening, verification, problem resolution, communications, closing, and then write the questions for each section. Once the sections and questions are drafted, assign weights to each question and possibly to each section, reflecting their relative importance. It’s easiest to use a 100‐point scale but this is not a requirement. Step 4: Using real calls, test the evaluation forms to ensure all call components that need to be measured are captured and that the intended goals are achieved fairly and accurately. Conduct calibration sessions involving supervisors and QA specialists to ensure that everyone participating in the QA program is using the form correctly and with the same rigor, consistently
© 2009 DMG Consulting LLC
All rights reserved.
- 11 -
April 2009
and fairly across all questions. It generally takes a few calibration sessions to finalize the questions and weights for each evaluation form. Step 5: Document the quality monitoring evaluation criteria and guidelines for how to apply them. nter ter QA Guide
Contac t Cen
Step 6: Define the agent coaching and feedback process. Coaching is considered the most effective method for training adults; it should take place in a one‐on‐one personalized session with an agent. During coaching sessions, agents should receive specific and targeted feedback to learn what they are doing well as well as the areas where they can improve. This is a critical part of successfully managing agents and motivating them to improve their performance. Step 7: Build the QA training program. Training should include an in‐depth review of the quality monitoring process and QA form, evaluation criteria and metrics/measurements, frequency and number of evaluations conducted, scoring methodology, how to access, review and respond to evaluations, and how to provide feedback to improve the program. Optionally, the QA training program may also include a review of documented policies, procedures and guidelines that govern agents’ behavior in processing transactions or advising customers. Step 8: Train staff on the QA program. Roll out the training program so all agents and other relevant contact center staff – managers, supervisors, trainers, QA specialists, agents, and possibly participants from other departments – become familiar and comfortable with the new program. If possible, conduct calibration sessions with agents during the training session to demonstrate application of the quality criteria and use of the evaluation form. (Note: Be sure to ask the agents whose calls will be used during calibration for their permission.) Agents who understand how their calls are being evaluated and know what is expected of them are more likely to perform well. Step 9: Pilot the QA program and monitor results. The pilot provides an opportunity to see what works and what needs to be enhanced. It also gives QA reviewers and agents an opportunity to get accustomed to the program. © 2009 DMG Consulting LLC
All rights reserved.
- 12 -
April 2009
DMG recommends a 2‐ to 3‐month pilot to work the kinks out of the process and give everyone time to get used to the new program before going live. Step 10: Adjust the program throughout the pilot and, if changes are significant enough, it may make sense to keep the pilot going longer. This step allows time to refine the program before it nter ter QA Guide
Contac t Cen
is formally launched. Based on participant feedback and pilot results, the QA manager should revise processes, evaluation criteria, evaluation forms, metrics, mechanics, etc. The pilot should be run for as long as necessary to test all components of the QA program, including conducting calibration exercises. Step 11: Implement the QA program. Launch the program and begin to execute the QA processes on a daily basis. QA reviewers should work closely with management, supervisors and trainers to review quality assurance results and identify areas for training and development. Periodic reviews to evaluate the QA process should be conducted to ensure compliance with ever‐evolving procedures, policies and protocols. It is also a good idea to establish a reward program or agent appreciation events, so that management can show special recognition for outstanding employee performance based on QA program reviews. After the QA program is implemented, contact centers should consider doing the following activities on a monthly basis. They are not a requirement but will help to enhance agent moral and improve the results of your QA program. •
Conduct monthly calibration sessions with all supervisors/QA reviewers to ensure program consistency. •
Conduct monthly agent team meetings to review QA results and do group training, as necessary. •
Have QA reviewers and trainers meet on a monthly basis to review training needs and other improvement opportunities uncovered in the QA process. (In many contact centers, the QA specialists and trainers speak daily.) © 2009 DMG Consulting LLC
All rights reserved.
- 13 -
April 2009
•
Publish a monthly QA newsletter (or other vehicle) with quality tips to help the department achieve quality goals. The newsletter should also recognize top performers. nter ter QA Guide
Contac t Cen
Action Item: Implement a process where top‐performing agents conduct side‐by‐side training sessions with agents who are not meeting their potential and are “under ‐
performing”. Agents learn best from their peers. Using the top performers for this activity will recognize their outstanding performance and help get them on board in support of management objectives. © 2009 DMG Consulting LLC
All rights reserved.
- 14 -
April 2009
5.1 Creating a Quality Assurance Leadership Team Building an effective QA program is an iterative process that requires senior management nter ter QA Guide
Contac t Cen
support, in addition to thorough planning, input and buy‐in from all levels of contact center staff. It’s a best practice to form a QA leadership team with representation from all contact center constituents including managers, supervisors, trainers, agents, and possibly other departments, such as marketing and sales. This team could be led by the head of QA or the contact center director. This is a good way to keep the head of the contact center engaged in the QA program on an ongoing basis. Action Item: Use the quality assurance monitoring process to identify areas for business process optimization. Establish a cross‐functional team to address contact center and enterprise business process opportunities identified during the quality monitoring process. This team should work together to change business processes that upset customers. © 2009 DMG Consulting LLC
All rights reserved.
- 15 -
April 2009
5.2 Staffing QA Positions The success of any QA program is largely dependent upon the commitment and skills of the quality assurance specialist who administers the program and coaches agents. Quality assurance specialists must be highly skilled individuals who possess outstanding job knowledge and excellent nter ter QA Guide
Contac t Cen
communication, interpersonal and coaching skills. Because a vast amount of job knowledge is required for this role, most contact centers staff these positions by promoting agents who demonstrate exceptional job knowledge and call handling skills. (Other contact centers transfer a supervisor or manager into this position.) However, being an outstanding customer service or sales representative is not the same as knowing how to deliver effective one‐on‐one feedback to coach and motivate a contact center agent. The ability to coach is not innate, and generally has to be taught. QA specialists should have the following skills and knowledge: •
Job knowledge: To accurately, effectively, and objectively evaluate agents, quality assurance specialists must possess strong knowledge of all products, services, systems, processes and procedures. If quality assurance specialists are hired externally, they should be required to complete a comprehensive training course that includes in‐depth coverage of all products, services, policies procedures and guidelines that agents are required to adhere to when processing transactions or advising customers. •
System knowledge: Quality assurance specialists must be fully trained and knowledgeable about all systems that agents use to handle customer inquiries or process orders. Additionally, they must be fully trained to navigate the quality management application so that they can search, retrieve and play back calls, create evaluations, and respond to agents’ evaluation feedback. QA specialists should also be trained to access and use reports from the quality application as well as to create ad hoc reports. QA program mechanics: Quality assurance specialists must possess a strong understanding of the quality monitoring process, including how to select calls/emails/chat sessions, the number/frequency of evaluations that they are required to do, and how to coach agents. © 2009 DMG Consulting LLC
All rights reserved.
- 16 -
April 2009
•
QA criteria/calibration: To maintain measurement integrity, it’s essential for all QA specialists to uniformly apply quality monitoring criteria when evaluating calls. All QA reviewers should be thoroughly trained to complete evaluations in a consistent manner. nter ter QA Guide
Contac t Cen
The success and effectiveness of the program and its reception by agents depends upon the ability to consistently, objectively and fairly evaluate contact center interactions. To achieve these essential goals, it’s also recommended that calibration sessions be conducted with all quality reviewers on a monthly basis. •
Coaching/motivation: QA specialists must demonstrate strong interpersonal and coaching skills so that they can work one‐on‐one with agents to recognize areas of strengths as well as provide directed feedback on areas that require improvement. All quality assurance specialists should be required to take courses in effective coaching methods and motivation techniques. Action Item: When staffing QA programs, use highly respected staff members with strong product, service, system knowledge and expertise in coaching when staffing QA programs. Agents are more welcoming of feedback from people they respect. © 2009 DMG Consulting LLC
All rights reserved.
- 17 -
April 2009
5.3 Implementation Roadmap Once the leadership team is established and the quality manager and staff have been nter ter QA Guide
Contac t Cen
appointed, a detailed breakdown of all program deliverables should be developed and assigned to the responsible parties. Figure 3 provides a list of the required tasks, recommendations for ownership of each initiative, and estimated time frames for completion. The tasks reflected in Figure 3 have already been discussed in other sections of this Guide. They are repeated here in order to put them into a project plan and to show the responsible party and estimated time frame. DMG recommends using this list of initiatives to build a detailed project plan. Your project plan will likely include other steps and initiatives that are important for your organization. Once the project plan is drafted, review it and get the buy‐in for all involved parties, particularly those assigned tasks, to make sure that they are committed and able to dedicate time to the projects. It’s also a good idea to have a weekly project review meeting to assess progress and to address items that are slipping. It is ideal, but not necessary for senior contact center management to participate in the weekly progress meeting. However, if a contact center manager is not able to participate in the weekly meetings, the project manager should make an effort to keep him/her updated on the team’s progress. © 2009 DMG Consulting LLC
All rights reserved.
- 18 -
April 2009
Figure 3: QA Implementation Roadmap
Task/Initiative
Responsibility
Timeframe
Communicate to all staff that a quality assurance
program is under development and will soon be
implemented; communication should be frequent
Contact Center
Director
2 days
Promote or hire a QA specialist and quality
reviewers
Contact Center
Director
3 weeks
Develop a call monitoring evaluation form (Note:
In multi-channel contact centers, a QA form will
need to be created and tested for each channel)
Quality Assurance
Leadership Team
2 weeks per
form
Test the form using real calls to ensure that all
call elements that need to be evaluated are
captured
QA Manager and
reviewers
2 days
Document quality monitoring evaluation criteria
and guidelines for how to apply them
Quality Assurance
Leadership Team
2 weeks
Determine weights for each section of the
monitoring form
Quality Assurance
Leadership Team
1 week
Test the form(s) using real calls to validate
sections, questions, weights and ranges
Quality Assurance
Leadership Team
1 week
Hold calibration sessions with agents, supervisors
and managers to make sure everyone uses the
evaluation form the same way
Quality Assurance
Leadership Team
2 weeks
Enhance QA evaluation form based on input from
calibration team
QA Manager and
reviewers
3 days
Determine the volume of transactions (calls,
emails, chat sessions, other) to be monitored per
agent/month
Quality Assurance
Leadership Team
1 day
Decide who will be conducting the evaluations
(quality assurance specialists, supervisors, or
both)
Quality Assurance
Leadership Team
1 day
Conduct calibration sessions to ensure rating
reliability and consensus of all reviewers
Quality Assurance
Leadership Team
2 weeks
Define the agent coaching process
QA Manager, contact
center managers,
supervisors and
reviewers
1 day
Develop a process with an escalation/review
procedure for agents to provide feedback or
dispute evaluations
Quality Assurance
Leadership Team
1 day
Develop reports to support the QA process; if
using a QA application, build the reports in the
application
Quality Assurance
Leadership Team
2 weeks
Test reports
QA Manager
1 week
Status
nter ter QA Guide
Contac t Cen
Development Phase
© 2009 DMG Consulting LLC
All rights reserved.
- 19 -
April 2009
Figure 3: QA Implementation Roadmap
Task/Initiative
Responsibility
Timeframe
Train all QA specialists to use the QA application
and reports
QA Manager
2 days
Develop QA training program
QA Manager and
trainers
3 weeks
Train all contact center staff on the quality
assurance program
QA Manager and
trainers
Variable
depending on
size of staff
and program
length
Kick off the program
Contact Center
Director/QA Manager
1 day
Pilot program for 2 to 3 months and review results
Quality Assurance
Leadership Team
2 – 3 months
Revise departmental Agent Performance
Evaluation form to include quality monitoring
metrics
Quality Assurance
Leadership Team/HR
1 week
Implement QA program
Contact Center
Leadership Team
On-going
Establish a process for identifying and
recognizing agents who achieve quality
monitoring excellence
Contact Center
Leadership Team
2 weeks
Perform monthly calibration sessions
QA Manager, all QA
reviewers
Monthly/
on-going
Conduct monthly training sessions
Trainer
Monthly/
on-going
Create and issue monthly QA newsletter
QA Manager
Monthly/
on-going
Have QA staff meet with training to enhance
training programs
QA Manager and
trainers
Monthly/
on-going
Status
nter ter QA Guide
Contac t Cen
Development Phase
Action Item: Involve all levels of contact center staff in creating the program to avoid unnecessary skepticism and surprises. This will help agents appreciate the positive aspects of the QA program and speed up adoption. © 2009 DMG Consulting LLC
All rights reserved.
- 20 -
April 2009
6. Developing a QA Evaluation Form When managers think about developing a QA program, one of the first things they must consider is creating a QA evaluation form. The evaluation form is the most visible component of the program and is necessary for all QA programs, whether an organization is doing QA manually or nter ter QA Guide
Contac t Cen
using an automated system. The QA evaluation form needs to capture all interaction components that a contact center wants to measure. Its questions and weights should reflect the culture of the company and what is most important to their service strategy. In general, QA evaluation forms contain the following components: 1. Call/evaluation details Examples: Name of agent, date of transaction, reviewer name, date of evaluation, call type, customer identifier (account number, social security number, etc.) 2. Sections (skill categories) Examples: Call opening/closing, verification, product/plan knowledge, procedure knowledge, system knowledge, hold/mute/transfer, communication skills, resolution skills, etc. 3. Questions (to objectively assess skill proficiencies) Examples: Did the representative identify the caller according to the verification policy prior to releasing information? Did the representative log a summary of the call according to the policy? 4. Scoring Point values for each question and section of the evaluation form; includes point values, points available, and points earned. Point values should be assigned based on the relative importance of each section and question to the business and the customer. © 2009 DMG Consulting LLC
All rights reserved.
- 21 -
April 2009
5. Coaching comments Free‐form text box to allow reviewers to provide feedback on performance excellence or opportunities. nter ter QA Guide
Contac t Cen
6. Recommendations Free‐form text box to allow reviewers to document an action or follow‐up items for agents to complete. Examples: taking a specific eLearning course, reviewing specific procedures, recommendations, etc. 7. Acknowledgement (optional) Signature boxes so that the reviewer and agent can sign off and acknowledge that they had a discussion. After learning how to create a QA evaluation form and deciding what sections are required, it’s relatively easy to modify an existing form or draft a new one. However, creating one for the first time can be daunting. Therefore, we suggest getting started with an existing QA form and then modifying it to meet the organization’s specific needs. The easiest way to obtain sample quality evaluation forms is by asking a QA vendor, another contact center manager, or a consulting firm. (Most vendors have sample forms and are generally happy to share them, particularly if they think you are interested in purchasing their solution.) While it’s relatively easy to obtain a sample QA evaluation form, the catch is that there is no "one size fits all" form. For example, a technical help desk needs a different QA evaluation form from a sales or customer service contact center. The QA evaluation form should be customized to meet the needs of each particular contact center. © 2009 DMG Consulting LLC
All rights reserved.
- 22 -
April 2009
Here are a few best practices to employ when developing QA evaluation forms: nter ter QA Guide
Contact Cen
Involve all levels of your staff – agents, supervisors, trainers and managers. Develop a separate QA form for each channel that you will be monitoring. Test your QA form(s) using real transactions and refine them before putting them into production. Assign weights that correspond to the importance of all categories on your QA form. Importance may be relative to the channel of communication, customer and/or the business. Include a comments box for each section of the evaluation so that coaching comments on performance strengths and opportunities and agent feedback can be captured. Action Item: To modify or create a QA evaluation form, start with an existing or sample form. Then listen to a selection of different call types. While “scoring” these calls, identify form components that need to be changed or added to fit your environment. (This process applies to emails, chats or other types of transactions received by the contact center.) © 2009 DMG Consulting LLC
All rights reserved.
- 23 -
April 2009
6.1 Form Categories Figure 4 provides an example of sections or categories typically found in a QA form for calls. Coont
ntact Center QA Guide
It also includes example criteria that can be used to evaluate if an agent properly performed a particular skill. Depending on how the form is structured, sections can be established based on skills, call flow, competencies, or call segments. In Figure 4 below, the sections are based on skills. Figure 4: Common Quality Evaluation Form Sections/Categories Quality Evaluation Form
Call Opening/Closing
Provided name/company name; thanked the customer for calling or purchasing [brand] products
Verification
Obtained verification prior to releasing information according to established data security procedures
Demonstration of System Knowledge/Usage
Effectively accessed information on system; utilized all appropriate systems to obtain information
Demonstration of Product Knowledge and Information
Demonstrated thorough knowledge of product/part/services/warranty, etc.
Demonstration of Resolution Skills
Accurately diagnosed problem/issue
Hold/Transfer Procedure
Use hold/transfer effectively and only as necessary
Communication Skills
Spoke clearly and confidently
Agent First Contact Resolution (FCR)
Did the agent do everything possible to ensure a callback was not necessary?
Business Process FCR
Is there a policy/procedure/service issue that prevented First Contact Resolution?
Source: DMG Consulting LLC, April 2009 Action Item: Be sure to include a section in the evaluation form to reflect how well the agent performed in regard to compliance or disclosure regulations (if applicable). © 2009 DMG Consulting LLC
All rights reserved.
- 24 -
April 2009
6.2 QA Evaluation Form Questions After identifying the right sections/skills to include in the QA form, the next step is to create nter ter QA Guide
Contac t Cen
a concise list of questions that captures the elements associated with demonstrating competency in each skill. The goal is to write questions that allow QA reviewers to completely and objectively assess agent performance. Once the initial list of questions is developed for each section of the form, have a team of supervisors and QA reviewers test the form by trying it out on a few calls or emails. It will typically take at least ten rewrites of the questions before an evaluation form is complete. This exercise helps surface gaps and identifies overlaps and redundancies in questions. Revise the form based on test results, and retest it until the form successfully addresses all requirements. (You will know the form works when there is a place to address every aspect of a call or email.) Action Item: Test the QA form with real calls or emails. Make sure there is a place on the form to address everything that can happen in a call or email. This will surface confusing or ambiguous components of the form. © 2009 DMG Consulting LLC
All rights reserved.
- 25 -
April 2009
6.3 Assigning Weights to QA Evaluation Forms Once the form content is finalized, it’s time to select a scheme for assigning weights to the nter ter QA Guide
Contac t Cen
questions and/or sections of the evaluation. The simplest approach is to go with a 100 point scale. While there are many scoring methodologies, the most basic way is to assign a point value to each section on the form and distribute the points among the questions in that section. For example, the communications skill section may be assigned a total of 20 points, which are distributed among the four questions in the section. While points can be distributed equally throughout a section or the entire QA evaluation form, we suggest that they be assigned based on the relative importance of each section and each question. Importance may be relative to the channel of communication, customer and/or the business. After preliminary weights are distributed, the form should be tested again by evaluating and scoring actual transactions. Points are generally reconsidered and redistributed during testing. It’s also important to consider that in some interactions, not all components of the form are applicable (for example, not every call includes a hold or transfer). The scoring on the forms shown in Section 6.5 was designed to include “Not Applicable” (N/A) as a consideration; if a particular question on the form is not applicable to the interaction being evaluated, the question is marked N/A. The points are not available to be achieved, nor do they affect the total score. Overall evaluation scores are calculated as follows: Total points achieved ⁄ Total points available = QM evaluation score For example: 76 points achieved / 84 points available = 90.4% Action Item: When assigning weights to the form, it’s also a good idea to identify and define the serious errors that will result in a full and automatic failure of an evaluation form section or an entire evaluation. © 2009 DMG Consulting LLC
All rights reserved.
- 26 -
April 2009
6.4 Validation Validation of the metrics associated with each category of the quality assurance form is an important step in developing the quality assurance program. Metrics are typically validated by testing the proposed weights against baseline quality measurements that have already been nter QA Guide
Conta ct Ceenter
established. When modifying an existing QA evaluation form, the new form can be validated by using it to re‐
evaluate transactions that had previously been scored using a different QA form. If no quality measurement baseline exists, initial testing sessions should validate QA scores based on agent performance. So, top performers should receive high scores, average performers should be in the mid‐range, and poor performers should have low scores. Once the point values are validated and coming out as expected, the next step is to determine how QA evaluation scores equate to the performance ratings used by the organization in its annual review process. For example: Excellent: 89% and above Good: 80‐88% Average: 71‐79% Needs Improvement: 70% and below The scoring ranges above are typical for many contact center QA programs. The ranges should be high enough so that an excellent transaction is clearly identifiable, but not so high so that few agents can attain an excellent rating. As a result, it’s often a good idea to start with one range and then as agents get accustomed to the QA program and their quality improves, tighten the ranges. So, for example, when a QA program is first kicked off, excellent may be anything above 85%. However, since most customers are not likely to think that 85% is excellent, the range should be narrowed to anything above either 89% or 90% a few months into the program. © 2009 DMG Consulting LLC
All rights reserved.
- 27 -
April 2009
Be sure to communicate to agents exactly what is happening with the ranges – starting easy to give them a chance to become accustomed, and then bringing them to a level that will better meet customer expectations – so that the staff does not think that management is just giving them a hard time. Be sure to set expectations and communicate clearly at all stages of the program to get the nter ter QA Guide
Contac t Cen
staff’s support. Once the final program is rolled out, agents should have a two‐ to three‐month grace period to become accustomed to the program, criteria and scoring before they are held accountable for QA scores; this will minimize any claims of “unfairness” within the shop, as the staff will have had an extended period of time to adjust to the program requirements. During this period, QA evaluations and coaching sessions are performed, but are not counted toward agents’ annual evaluations. Action Item: Review and enhance your QA evaluation forms periodically —minimally, every 9 to 12 months to keep them in sync with your business and customers’ expectations. © 2009 DMG Consulting LLC
All rights reserved.
- 28 -
April 2009
6.5 Example QA Evaluation Forms This section includes 5 example QA forms. The first form, Figure 5, is for a financial services nter ter QA Guide
Contac t Cen
customer service contact center. Figure 6 is a form intended for use in an inbound healthcare customer service contact center. Figures 7, 8, 9 depict forms intended for inbound technical support, customer service with up‐sell/cross‐sell activities and new order processing, respectively. Use these forms as examples for getting started, but customize them to reflect the needs and priorities of your organization. © 2009 DMG Consulting LLC
All rights reserved.
- 29 -
April 2009
Figure 5: Financial Services Customer Service QA Evaluation Form Description: This form is for use by an inbound financial services contact center that addresses customer inquiries, questions or problems on their accounts, and/or services that are provided. nter ter QA Guide
Contac t Cen
Evaluation Details
Call Details
Representative name:
Account number or customer ID:
Call date:
Caller:
Evaluator name:
Call category:
Evaluation date:
Call type:
Greeting (3 points)
Point
Value
Used call greeting as defined in
Greeting Policy
Yes
No
Points
Available
Points
Achieved
Points
Available
Points
Achieved
Points
Available
Points
Achieved
3
Point Totals
Coaching Comments:
Verification (10 points)
Point
Value
Verified caller as defined in
Verification Policy prior to
releasing account information
7
Verified additional items
pertaining to specific caller
inquiry
3
Yes
No
N/A
Point Totals
Coaching Comments:
Product Knowledge (20 points)
Point
Value
Accurately identified account,
transaction, product or service
5
Provided complete information or
instructions in accordance with
established procedure
7
Provided accurate information or
instructions in accordance with
established procedure
8
Yes
No
N/A
© 2009 DMG Consulting LLC
All rights reserved.
- 30 -
Point Totals
April 2009
Coaching Comments:
Inquiry Resolution (23 points)
nter ter QA Guide
Contac t Cen
Point
Value
Accurately understood the nature
of the caller’s inquiry
5
Effectively/accurately resolved
inquiry/ issue in accordance with
established procedure
5
Yes
No
N/A
Points
Available
Points
Achieved
Points
Available
Points
Achieved
Points
Available
Points
Achieved
Completed fulfillment/referral/
5
follow-up as promised/required
Provided appropriate alternatives
relative to customer’s
need/situation
3
Ensured caller fully understood
explanation, process, time frames
and/or next steps
5
Point Totals
Coaching Comments:
Compliance (8 points)
Point
Value
Read mandatory disclosures as
required and applicable
Yes
No
N/A
8
Point Totals
Coaching Comments:
Cross-sell/Up-sell (6 points)
Point
Value
Recommended appropriate
product/service to meet customer
need or extend the relationship
3
Effectively tied product/service
benefit to customer situation/need
3
Yes
No
N/A
© 2009 DMG Consulting LLC
All rights reserved.
- 31 -
Point Totals
April 2009
Coaching Comments:
System Knowledge/Usage (7 points)
nter ter QA Guide
Contac t Cen
Point
Value
Effectively accessed and
utilized all appropriate systems,
screens and fields to obtain
information to resolve inquiry
5
Accurately utilized wrap-up
2
Yes
No
N/A
Points
Available
Points
Achieved
Points
Available
Points
Achieved
Points
Available
Points
Achieved
Point Totals
Coaching Comments:
Hold/Mute/Transfer (4 points)
Point
Value
Utilized hold/mute as defined in
Hold/Mute Policy
2
Performed transfer to the
correct area, only as necessary
and in accordance with the
Transfer Policy
2
Yes
No
N/A
Point Totals
Coaching Comments:
Communication Skills (14 points)
Point
Value
Maintained a courteous,
pleasant, and respectful tone
throughout the call
3
Conveyed information clearly
and confidently and in a
manner that was easily
understood
3
Demonstrated effective
listening skills
2
Expressed empathy and
concern, as appropriate
1
Efficiently managed time and
call flow (call management)
2
© 2009 DMG Consulting LLC
All rights reserved.
Yes
No
N/A
- 32 -
April 2009
Communication Skills
Point
Value
nter ter QA Guide
Contac t Cen
Demonstrated professional
language and behavior (call
etiquette)
Yes
No
N/A
Points
Available
Points
Achieved
Points
Available
Points
Achieved
3
Point Totals
Coaching Comments:
Closing (5 points)
Point
Value
Used call closing as defined in Call
Closing Policy
2
Accurately documented information
in accordance with established
procedure
3
Yes
No
N/A
Point Totals
Coaching Comments:
Total Points Available/Total Points Achieved
QM Evaluation Score
Date Reviewed with Representative
Comments:
Recommendations:
Acknowledged by Representative
Acknowledged by Reviewer
Source: DMG Consulting LLC, April 2009 © 2009 DMG Consulting LLC
All rights reserved.
- 33 -
April 2009
Figure 6: Healthcare Customer Service QA Evaluation Form Description: This form is for use by an inbound customer service organization that addresses calls about insurance claims. The calls may address program eligibility, claims status and dipositioning, nter ter QA Guide
Contac t Cen
program benefits and coverage. Evaluation Details
Call Details
Representative name:
HCID or SSN:
Call date:
Caller:
Evaluator name:
Call category:
Evaluation date:
Call type:
Greeting (5 points available)
Point
Value
Used call greeting as defined
in Greeting Policy
2
Documented caller name
and phone number
3
Yes
No
Points
Available
Points
Achieved
Points
Available
Points
Achieved
Point Totals
Coaching Comments:
Verification (10 points available)
Point
Value
Verified caller as defined in
Verification Policy
5
Verified eligibility as defined
in Eligibility Policy
3
Asked caller for member ID
or patient’s date of birth and
name
1
Asked caller for address on
file
1
Yes
No
N/A
Point Totals
Coaching Comments:
© 2009 DMG Consulting LLC
All rights reserved.
- 34 -
April 2009
Plan/Benefit Knowledge (25 points available)
nter ter QA Guide
Contac t Cen
Point
Value
Accurately identified plan
type, benefit, or claim
5
Demonstrated thorough
knowledge of plan type,
benefit or claim information
5
Provided complete
information or instructions in
accordance with established
procedure
7
Provided accurate
information or instructions in
accordance with established
procedure
8
Yes
No
N/A
Points
Available
Points
Achieved
Points
Available
Points
Achieved
Point Totals
Coaching Comments:
Inquiry Resolution (25 points available)
Point
Value
Accurately understood the
nature of the caller’s inquiry
5
Effectively/accurately
resolved inquiry/issue in
accordance with established
procedure
5
Completed
fulfillment/referral/follow-up
as promised/required
5
Provided alternatives, as
appropriate
3
Provided time frames, as
appropriate
2
Ensured caller fully
understood explanation,
process, and/or next steps
5
Yes
No
N/A
Point Totals
Coaching Comments:
© 2009 DMG Consulting LLC
All rights reserved.
- 35 -
April 2009
System Knowledge/Usage (10 points available)
Point
Yes
No
N/A
nter ter QA Guide
Contac t Cen
Value
Effectively accessed and
utilized all appropriate
systems, screens and
fields to obtain information
to resolve inquiry
8
Accurately utilized wrap-up
2
Points
Available
Points
Achieved
Points
Available
Points
Achieved
Points
Available
Points
Achieved
Point Totals
Coaching Comments:
Hold/Mute/Transfer (6 points available)
Point
Value
Utilized hold/mute as
defined in Hold/Mute Policy
2
Performed transfer as
defined in Warm Transfer
Policy
2
Transferred call to the
correct area
2
Yes
No
N/A
Point Totals
Coaching Comments:
Communication Skills (14 points available)
Point
Value
Maintained a courteous,
pleasant, and respectful
tone throughout the call
3
Conveyed information
clearly and confidently
and in a manner that was
easily understood
3
Demonstrated effective
listening skills
2
Expressed empathy and
concern, as appropriate
1
Efficiently managed time
and call flow (call
management)
2
Demonstrated
professionalism (call
etiquette)
3
© 2009 DMG Consulting LLC
All rights reserved.
Yes
No
N/A
- 36 -
April 2009
Point Totals
Coaching Comments:
Closing (5 points available)
nter ter QA Guide
Contac t Cen
Point
Value
Used call closing as defined
in Call Closing Policy
2
Accurately documented
information in accordance
with established procedure
3
Yes
No
N/A
Points
Available
Points
Achieved
Point Totals
Coaching Comments:
Total Points Available/Total Points Achieved
QM Evaluation Score
Date Reviewed with Representative
Comments:
Recommendations:
Acknowledged by Representative
Acknowledged by Reviewer
Source: DMG Consulting LLC, April 2009 © 2009 DMG Consulting LLC
All rights reserved.
- 37 -
April 2009
Figure 7: Technical Support QA Evaluation Form Description: This form is for use by an inbound customer service organization that provides technical support for manufactured products. The contact center addresses inquiries from customers and field nter ter QA Guide
Contac t Cen
service technicians. The inquiries encompass a wide range of issues such as parts, service, general product information, authorized service repair shops, warranties, detailed technical support, troubleshooting, etc. Evaluation Details
Call Details
Agent Name:
________________________
Caller Category:
__________________________
Date of Call:
________________________
Call Category:
__________________________
Evaluator Name:
________________________
Call Type:
__________________________
Date of Evaluation:
________________________
Product Type:
__________________________
Serial Number:
__________________________
Greeting/Opening (3 points)
Use of Established Call Opening
Provided name, company name,
inquired: "How may I help you?"
Point
Value
Yes
Points
Available
No
Points Achieved
3
Point Totals
Coaching Comments:
Verification (6 points available)
Verification of Owner/Registration
Information
Point
Value
Owner name/address/phone number:
3
Product serial number:
2
Date of installation:
1
Yes
No
N/A
Points Available
Points Achieved
Point Totals
Coaching Comments:
© 2009 DMG Consulting LLC
All rights reserved.
- 38 -
April 2009
System Knowledge/Usage (15 points available)
nter ter QA Guide
Contac t Cen
Demonstration of System Usage and
Navigation Skills
Point
Value
Effectively accessed information on
system
5
Utilized all appropriate systems to
obtain information
5
Documented/updated information
according to established procedure, as
needed or required
3
Accurately utilized wrap-up
1
Yes
No
N/A
Points Available
Points Achieved
Points Available
Points Achieved
Points Available
Points Achieved
Point Totals
Coaching Comments:
Product Knowledge and Information (25 points available)
Demonstration of Product
Knowledge/
Information Skills
Point
Value
Accurately identified
product/part/service/warranty issue
6
Demonstrated thorough knowledge of
product/part/services/warranty, etc.
5
Provided complete
information/instructions
7
Provided accurate
information/instructions
7
Yes
No
N/A
Point Totals
Coaching Comments:
Resolution Skills (25 available points)
Demonstration of Problem
Resolution Skills
Point
Value
Accurately diagnosed problem
5
Effectively resolved issue in
accordance with established procedure
5
Used effective negotiation skills
3
Completed fulfillment/follow-up as
promised/required
5
Provided alternatives and timeframes,
as appropriate
2
Ensured caller fully understood
explanation/process, next steps
5
Yes
No
N/A
Point Totals
Coaching Comments:
© 2009 DMG Consulting LLC
All rights reserved.
- 39 -
April 2009
Hold/Transfer Procedure (4 points available)
nter ter QA Guide
Contac t Cen
Use of Established Hold/Transfer
Procedures
Point
Value
Used hold effectively/transferred the
caller to the correct area, and only as
necessary
2
Asked permission to place called on
hold/thanked the caller for
holding/checked back at established
intervals
2
Yes
No
N/A
Points Available
Points Achieved
Points Available
Points Achieved
Points Available
Points Achieved
Point Totals
Coaching Comments:
Communication Skills (18 points)
Demonstration of Communication
Skills
Point
Value
Maintained a courteous, pleasant, and
respectful tone throughout the call
4
Spoke clearly and confidently;
explained information to the caller in a
manner that was easily understood
4
Demonstrated effective listening skills;
restated the caller inquiry to ensure
understanding
3
Expressed empathy and concern, as
appropriate
2
Established confidence and reinforced
brand
2
Efficiently managed time and call flow
(call management)
3
Yes
No
N/A
Point Totals
Coaching Comments:
Closing (4 points)
Use of Established Call Closing
Point
Value
Thanked the customer for calling or
purchasing brand products
2
Asked: "Is there anything else I can
help you with?"
2
Yes
No
Point Totals
Coaching Comments:
Total Points Available/ Total Points Achieved:
QM Evaluation Score:
© 2009 DMG Consulting LLC
All rights reserved.
- 40 -
April 2009
First Contact Resolution
Demonstration of First Contact
Resolution Skills
Yes
No
N/A
Points Available
Points Achieved
nter ter QA Guide
Contac t Cen
Did the agent/technician do everything
possible to ensure that a callback was
not necessary?
Coaching Comments:
FCR Score:
Note: First Contact Resolution (FCR) is intended to be a separate quality monitoring metric for performance measurement.
This category of the evaluation form should be based on the number of calls monitored. For example, if 5 calls per month
are monitored, each call is weighted at 20%. A technician who achieved First Contact Resolution on 3 calls would have an
FCR score of 60% for the month.
Business First Contact Resolution
Yes
No
N/A
Did the caller indicate this was a follow-up call?
Is there a policy/procedure/service issue that
prevented First Contact Resolution?
Comments:
Note: Business First Contact Resolution is not intended to be a quality metric for technicians. It is a data collection category
that should be used for business process optimization analysis.
Source: DMG Consulting LLC, April 2009 © 2009 DMG Consulting LLC
All rights reserved.
- 41 -
April 2009
Figure 8: VPI Sample Customer Service Quality Monitoring Form Description: This form is for use by an inbound customer service department that does up‐sell or nter ter QA Guide
Contac t Cen
cross‐sell. This QA form measures agent skills and processes. Sample Customer Service Quality Monitoring Form
Agent:
Group:
Evaluator:
Agent Name
Group Number
Evaluator Name
Evaluation Start Time:
Evaluation End Time:
Interaction Date:
Date and Time
Finalized Date or “In Progress”
Date and Time
Quality Skill Scores:
Overall Score:
115/115 (100%)
Greeting
5/5 (100%)
Procedural Requirements
10/10 (100%)
Communication
35/35 (100%)
Listening
10/10 (100%)
Empathy
10/10 (100%)
Upsell
10/10 (100%)
Problem Assessment & Resolution
30/30 (100%)
Call Handling
5/5 (100%)
Evaluation Comments:
Enter Evaluation Comments Here
Call Segment: Introduction, Score: 10/30 (33.33%)
Question:
Answer:
Did the call center agent greet the caller and clearly identify themselves and the company?
Yes
No
Skill:
Communication
Score:
5/5 (100%)
Question:
Did the call center agent verify and update customer information?
Answer:
Yes
No
Skill:
Procedural Requirements
Score:
5/5 (100%)
Question:
How effectively did the agent incorporate the script into their style, following the guidelines, but also
sounding natural ?
Answer:
Excellent
Very Good
Good
Fair
Poor
Choose One
Skill:
Communication
Score:
10/10 (100%)
Question:
How attentive was the agent with listening to the customer? (10 = Excellent, Poor = 0)
Answer:
0
1
2
Skill:
Communication
Score:
10/10 (100%)
3
4
5
6
7
8
9
10
Call Segment: Customer Service Criteria, Score: 45/45 (100%)
© 2009 DMG Consulting LLC
All rights reserved.
- 42 -
April 2009
Question:
nter ter QA Guide
Contac t Cen
Answer:
Did the agent refer to the customer as "Mr...", "Miss...", "Sir" or "Mam"?
Yes
No
Skill:
Communication
Score:
5/5 (100%)
Question:
How attentive was the agent in listening to the customer?
Answer:
Choose One
Very Good
Good
Fair
Poor
Skill:
Listening
Score:
10/10 (100%)
Question:
How well did the agent speak at an understandable rate and sound postive and upbeat?
Answer:
0
1
2
3
4
5
6
7
8
9
10
Skill:
Communication
Score:
10/10 (100%)
Question:
How well did the agent seem to empathize with the customer
Answer:
Choose One
Excellent
Very Good
Good
Fair
Skill:
Empathy
Score:
10/10 (100%)
Question:
Was the agent able get an order for the new promotion?
Answer:
Choose One
Skill:
Upsell
Score:
10/10 (100%)
Yes
Poor
No
Call Segment: Problem Identification, Score: 15/25 (60%)
Question:
Answer:
How effectively did the agent use probing questions to identify the customer's problems?
Choose One
Excellent
Good
Fair
Poor
Skill:
Problem Assessment & Resolution
Score:
10/10 (100%)
Question:
If the agent had to put the customer on-hold, did they come back after 45 seconds and thank them for
being patient and could continue to keep holding?
Answer:
Yes
No
Skill:
Call Handling
Score:
5/5 (100%)
Question:
Did the agent offer all possible solutions before scheduling an escalation call back?
Answer:
Yes
No
Skill:
Problem Assessment & Resolution
Score:
10/10 (100%)
Call Segment: Closing, Score: 15/15 (100%)
© 2009 DMG Consulting LLC
All rights reserved.
- 43 -
April 2009
Question:
Did the agent thank the customer and ask if there was anything else they could help them with?
Answer:
Yes
Problem Assessment & Resolution
Score:
5/5 (100%)
Question:
Did agent review the call and get customers approval of call's resolution?
Answer:
nter ter QA Guide
Contac t Cen
No
Skill:
No
Yes
Skill:
Problem Assessment & Resolution
Score:
5/5 (100%)
Question:
Did agent thank the customer for their business?
Answer:
Yes
No
Skill:
Procedural Requirements
Score:
5/5 (100%)
Source: VPI, April 2009 © 2009 DMG Consulting LLC
All rights reserved.
- 44 -
April 2009
Figure 9: VPI New Order Precision Quality Monitoring Form Description: This form is for use by any inbound sales organizations that focuses on improving the quality and outcome of its “new order” calls that are automatically identified via desktop analytics. nter ter QA Guide
Contac t Cen
© 2009 DMG Consulting LLC
All rights reserved.
- 45 -
April 2009
New Order Precision Quality Monitoring Form
nter ter QA Guide
Contac t Cen
Agent:
Group:
Evaluator:
Agent Name
Group Number
Evaluator Name
Evaluation Start Time:
Evaluation End Time:
Interaction Date:
Date and Time
Finalized Date or “In Progress”
Date and Time
Quality Skill Scores:
Overall Score:
70/70 (100%)
Greeting
10/10 (100%)
Communication
25/25 (100%)
Sales
10/10 (100%)
Upselling
10/10 (100.00%)
Closing
10/10 (100.00%)
After Call Work
5/5 (100.00%)
Evaluation Comments:
Enter Evaluation Comments Here
Call Segment: Opening, Score: 20/20 (100%)
Question:
Answer:
Did the agent use the appropriate greeting and properly identify themselves to the caller?
Yes
No
Skill:
Greeting
Question:
Did the call center agent verify and update customer information, if needed?
Answer:
Yes
No
Skill:
Communication
Question:
Was the agent's tone and word choice in line with the caller's?
Answer:
Skill:
Yes - the agent was extremely sensitive to the caller's needs
Mostly - the agent listened to the caller, but did not try to actively identify the problems for the caller
No - the agent did not attend to the caller's tone and urgency or anticipate their issues
Communication
Call Segment: Order Entry, Score: 20/20 (100%)
© 2009 DMG Consulting LLC
All rights reserved.
- 46 -
April 2009
Question:
Did the agent go through all available order options prior to taking the order information?
nter ter QA Guide
Contac t Cen
Answer:
Yes
No
Skill:
Sales
Question:
Was the customer offered the monthly promotion?
Answer:
Yes
No
Skill:
Upselling
Question:
Did the agent upsell the customer to the Gold level, if not already ordered?
Answer:
Skill:
Yes
No
Upselling
Call Segment: Closing, Score: 25/25 (100%)
Question:
Did the agent confirm shipping method and location before summarizing the order?
Answer:
Skill:
Question:
Yes
No
Proceedural Requirements
Was an order summary provided, including total estimated cost and delivery date, to customer?
Answer:
Yes
No
Skill:
Proceedural Requirements
Question:
Did the agent offer additional assistance to caller and thank them for the order?
Answer:
Yes
No
Skill:
Communication
Question:
Did the agent use the correct closing for a new order call?
Answer:
Skill:
Yes
No
Closing
Call Segment: After Call Work, Score: 5/5 (100%)
Question:
Did the agent enter the correct notes in to the summary field for a new order?
Answer:
Skill
Yes
No
After Call Work
Source: VPI, April 2009 © 2009 DMG Consulting LLC
All rights reserved.
- 47 -
April 2009
7. Quality Assurance Monitoring Criteria and Guidelines Quality assurance monitoring guidelines specify the criteria to be used by QA specialists and reviewers in determining if an agent properly demonstrated a skill during a transaction. They also nter ter QA Guide
Contac t Cen
indicate which section of the form to use when scoring the various components of the interaction. 7.1 Quality Monitoring Criteria and Guidelines The quality monitoring criteria define sections, questions and skills that are captured in the quality evaluation form. They describe how a skill should be demonstrated so that QA reviewers know how to objectively evaluate and score an interaction. For example, if a question on the evaluation form is intended to determine if an agent established rapport with a customer, the criteria must define what building customer rapport means, i.e., addressing the customer by name throughout the call or acknowledging an event that the customer mentioned in the course of the call (anniversary, birthday, vacation plans, etc.). Please keep in mind that the meaning of building a rapport at one company may be the opposite in another. While some companies may want their customers to call their customers by their first name, others may insist that only a title and the last name be used. The criteria should also reflect which section and questions in the QA evaluation form should be used to score the various call elements. While this may seem obvious, it is often quite difficult. Consider this example: A quality monitoring form has the following three questions in the communication skills section: Demonstrated effective listening skills? (3 points), Expressed empathy and concern as appropriate? (1 point), and Established rapport with the customer? (2 points). If during the course of a call a customer mentions that they were recently hospitalized and the agent does not acknowledge that, where should this issue be captured and scored in the prior three questions? The answer is that as long as it is consistently addressed in the same question, it doesn’t matter. The quality monitoring criteria must specify where the issue needs to be addressed. Action Item: Assign a member of the quality assurance review team to be responsible for overseeing the quality monitoring criteria and maintaining them, as required. © 2009 DMG Consulting LLC
All rights reserved.
- 48 -
April 2009
7.2 Why are Quality Monitoring Criteria and Guidelines Necessary? Quality monitoring criteria and guidelines take the “guesswork” out of figuring out the right nter ter QA Guide
Contac t Cen
way to evaluate and score a transaction. The guidelines standardize the use of QA evaluation forms, allowing QA reviewers to handle each transaction objectively and consistently. They also safeguard against “double‐dipping” – a term used by agents to refer to a situation where they have lost points for more than one question on a QA evaluation for the same issue. These issues are flushed out and resolved during the testing and calibration process. Action Item: Invest time up front to develop clear and concise quality monitoring criteria and guidelines. This yields a standardized approach and helps minimize misunderstandings about how to conduct QA evaluations. © 2009 DMG Consulting LLC
All rights reserved.
- 49 -
April 2009
8. Calibration Calibration, a method for building consensus and delivering a standardized evaluation tool, nter ter QA Guide
Contac t Cen
is key to the success of all QA programs. It is a process intended to ensure that QA results are valid and based on reliable measurement tools. For a QA program to be credible and produce accurate and dependable results, it’s essential that all reviewers evaluate transactions on a consistent basis. Calibration is not a one‐time event. To keep all evaluators synchronized, calibration must be done on an ongoing basis and include all people involved in conducting evaluations. It’s also a great way to build camaraderie and support for a QA initiative. Calibration should be conducted on a monthly basis. If a contact center is a multi‐site environment, all sites should participate in group‐wide calibration sessions to ensure that transactions are being evaluated consistently regardless of the site where the transaction is handled. 8.1 What is Calibration? Calibration is a process where all QA reviewers discuss how to score various types of transactions. The QA reviewers meet and review agent transactions. Every individual scores the same transactions and then scoring differences are identified. The reviewers then discuss the reasons for the differences and reach consensus. DMG recommends that agents be invited to participate in calibration sessions so that they can gain an appreciation of the effort and rigor applied to the QA process. When conducted properly, calibration sessions foster collaboration and establish consensus on how quality monitoring criteria should be applied to each question and scored on the evaluation form. Action Item: Update quality monitoring criteria based on the consensus reached in calibration sessions. © 2009 DMG Consulting LLC
All rights reserved.
- 50 -
April 2009
8.2 Benefits of Calibration When performed consistently, generally monthly, calibration is a proven approach for maintaining nter ter QA Guide
Contac t Cen
program integrity and equity, and ensuring that agents are treated fairly. Among its many benefits, calibration: •
Helps develop effective QA evaluation forms •
Teaches QA reviewers how to apply evaluation criteria and perform QA evaluations on a consistent basis •
Builds consensus among all QA reviewers •
Helps maintain an open dialogue between the QA team and management •
Keeps staff updated about changes to the program, scoring criteria and the evaluation form, as the program evolves •
Enhances agent perception of the program’s credibility and fairness •
Fosters collaboration and camaraderie among QA reviewers •
Keeps supervisors well versed in how agents are being evaluated and facilitates more effective coaching © 2009 DMG Consulting LLC
All rights reserved.
- 51 -
April 2009
8.3 The Calibration Process Figure 10 provides a general overview of how calibration works in most contact centers, nter ter QA Guide
Contac t Cen
although this process varies among companies. A set of recorded calls (and/or emails and chats in a multi‐channel environment) are tagged for calibration and sent to all QA reviewers. Each reviewer evaluates and scores the transactions. The QA manager generates a report that reflects scoring variances between the reviewers for each transaction; this must be done at the question level to be useful. (In some organizations, the scoring variances are calculated against a “master evaluation” that is designated as the department standard.) A calibration meeting is held to review scoring results and to discuss the variances. If a scoring discrepancy is uncovered, each reviewer presents a justification of their scoring. In most cases, this generates a lively discussion that requires the group to replay the transaction, or verify information in the procedure or training guide. Once all opinions are presented, the group reaches a consensus on the best way to score each question. This is likely to address how the question was rated and/or which form category should be used to capture a particular issue. This is an important part of the calibration process, as point values vary depending on where call elements are scored. The quality monitoring criteria guidelines (as well as procedure or training manuals, if applicable) should be modified or updated based on the consensus outcome. © 2009 DMG Consulting LLC
All rights reserved.
- 52 -
April 2009
Figure 10: The Calibration Process nter ter QA Guide Contact Cen
Source: DMG Consulting LLC, April 2009 Action Item: Include agents in calibration sessions. This helps them appreciate the effort management puts into accurately assessing calls and emails and fairly evaluating agent performance. © 2009 DMG Consulting LLC
All rights reserved.
- 53 -
April 2009
9. Quality Assurance Program Mechanics and Processes An important part of every QA program is determining the number of evaluations that must nter ter QA Guide
Contac t Cen
be completed to come up with either operationally or scientifically valid results. Once this number is determined, management must decide if it can afford the resources required to conduct the optimal number of evaluations. If the answer is “no,” as it is in many contact centers, then the challenge is to come up with a lower number of evaluations to perform that will still provide insight into agent performance and transaction trends. Determining the number and frequency of evaluations is called defining the program mechanics. 9.1 Transaction Selection Criteria Identifying transactions for QA is an important part of the process. There are four primary methods for selecting transactions to be reviewed. The first approach is to set up a schedule to capture transactions at pre‐defined intervals. The second approach is to randomly capture transactions. The third approach is to use business rules to identify calls that require attention. The fourth method is to use automation, such as speech or screen analytics, to identify transactions that are either really good or really bad and require management attention. Regardless of the approach used, it’s essential that the method capture transactions fairly and equitably. The capture method should be communicated to agents so that they know how transactions are selected for review. Once transactions are captured, they can be further qualified based on any number of factors, including call direction, call duration, call type (based on wrap‐up, disposition or other interaction classification mechanism), product type, etc. Date of interaction is another key factor as, depending on the total number of evaluations being conducted per agent/per month, it’s a best practice to spread out the evaluations over the course of the month. This helps to ensure that agents receive timely feedback and gives them an opportunity to make adjustments to their performance. Action Item: Set up a process to capture the best transactions for the QA team. © 2009 DMG Consulting LLC
All rights reserved.
- 54 -
April 2009
9.2 Determining the Number and Frequency of Evaluations The goal of a quality assurance program is to provide a statistically significant analysis of service delivery and the quality of customer interactions. To accurately measure service quality and establish credibility and reliability for the quality assurance process, randomly captured calls should nter ter QA Guide
Contac t Cen
be evaluated consistently for all agents, on a regular basis. Unfortunately, most contact centers do not have the resources to conduct QA on a statistically valid sample of transactions. Instead, management generally specifies a number of calls/emails/chat sessions to evaluate on a weekly and/or monthly basis. This number is based on QA resources. (To determine the number of QA sessions that can be performed in a day, calculate the amount of time it takes to do an evaluation and deliver a coaching session. Then divide this number by the working hours available per day, which is generally 6.5.) There are no industry guidelines for determining the ratio or number of calls that should be monitored per contact center on a daily, weekly or monthly basis. This number varies based on total number of agents, type of contact center (multi‐channel or multi‐skill) and transaction volume. Here is a strategy that some contact centers use to determine the number of evaluations to monitor per agent/month. Start by evaluating 10 calls per agent for a month to obtain a baseline figure. Each subsequent month, reduce the number of evaluations for each agent by 1 and compare the results/findings to the prior month. Continue this process until the variance between the results is significant. At that point, the number of calls evaluated is as low as it can go. (The most common number of calls evaluated for agents on a monthly basis is 3 to 5.) Another way contact centers address QA resource limitations is to split the number of agent evaluations that need to be completed among supervisors and quality reviewers. While it's important to have dedicated QA resources, it's also essential for line supervisors or managers to keep informed about their agents' performance. Evaluating agents is a great way to stay apprised of what’s happening with agents and the contact center. Action Item: Determine an appropriate number of transactions to evaluate for each agent on a weekly or monthly basis to give management insights into what’s happening in their contact center and to give agents a good idea of their performance. © 2009 DMG Consulting LLC
All rights reserved.
- 55 -
April 2009
9.3 Coaching Coaching is another vital component of a successful QA program. Coaching is the process of nter ter QA Guide
Contact Cen
delivering feedback on a frequent, consistent and timely basis to agents. Effective coaching sessions recognize agent strengths and assets in addition to identifying the areas where agents have improvement opportunities. Coaching sessions that are instructive, supportive and collaborative demonstrate management’s commitment to agents’ success. As depicted in Figure 11, a variety of coaching methods can be employed. The methods include personalized coaching, best practice clips, broadcast messages and real‐time feedback. Regardless of the method used, what is important is that the required number of coaching sessions is, in fact, completed. Lack of coaching or poor training is one of the primary reasons QA programs fail or do not achieve their anticipated results. It also leads to agent attrition. Figure 11: Coaching Methods Personalized Coaching
Best Practice Clip
Broadcast Messaging
Real-Time Feedback
Source: DMG Consulting LLC, April 2009 © 2009 DMG Consulting LLC
All rights reserved.
- 56 -
April 2009
Coaching Best Practices Always deliver coaching in a private setting to maintain confidentiality and avoid outside distractions and interruptions Begin coaching sessions on a positive note, highlighting agent strengths and nter ter QA Guide
Contact Cen
recognizing incremental improvements When discussing performance opportunities focus on specific examples and detail why the agent did not meet the performance standard Provide the agent with specific examples of how the situation could have been handled better, i.e., “Maybe a better way to explain this to the caller is to say,” or “This information can be found on this screen…” Encourage open dialogue by asking the agent to talk about areas where he/she is having difficulties and discuss what kind of support would be most helpful Establish incremental goals for the agent to achieve for the next coaching session Follow up as planned on any deliverables Provide informal support between coaching sessions. Drop by, listen for a few minutes and provide encouragement; a few kind words can have a significant impact End the session on a positive note Action Item: Coach agents frequently using the method that is most effective for each individual. © 2009 DMG Consulting LLC
All rights reserved.
- 57 -
April 2009
9.4 Evaluation Feedback and Escalation Process The most effective QA programs facilitate two‐way communication between reviewers and agents. Part of program development involves setting up the process for agents to receive their nter ter QA Guide
Contac t Cen
completed evaluations, review them along with any recorded transactions, and channel any questions or feedback back to the reviewer. Particularly in the early stages of the roll‐out, agents who are not accustomed to being monitored are likely to challenge low scores. It’s important to have a process that allows agents to voice their objections and escalate their concerns without trepidation of repercussions. The following process is recommended: •
Give agents their completed QA evaluation(s) and underlying transaction(s) prior to them sitting down with the QA reviewer •
If a coaching session is not already scheduled, allow agents to request a meeting with the QA reviewer to discuss their evaluation(s) •
QA reviewers must be able to explain and substantiate how the transaction was evaluated based on the recording and system information, citing written procedure and policy, or, in case of an error, adjusting the score •
If the agent still disagrees with the score and how the form was evaluated, forward the evaluation to a designated “QA arbitrator” for final disposition Challenges to evaluation scores, as a rule, do not happen consistently or frequently. When they do arise, it’s important for agents to feel comfortable raising concerns and voicing dissention. Action Item: Invite agent participation in the QA feedback process, so that they feel empowered by the process. © 2009 DMG Consulting LLC
All rights reserved.
- 58 -
April 2009
9.5 Ongoing Training Ongoing training is important for agents and QA specialists. Beyond basic introductory nter ter QA Guide
Contac t Cen
training, agents require up‐training on a consistent basis to keep abreast of changes in policies, procedures, processes, products, services and systems. Besides being critical for optimizing agent performance, training and coaching contribute significantly to agent job satisfaction, morale and motivation. Effective QA programs have a closed‐loop process with training so that issues identified during evaluations are shared with trainers on a timely basis. An important part of the quality reviewer’s responsibility is to identify trends and training opportunities to address performance gaps. The QA program should include a process for funneling recommendations to the training department and ensuring that all training, reference materials, policies and procedures are accurate and up‐to‐date. It is also recommended that the quality manager and training department hold monthly meetings to review training effectiveness and develop action plans to address any trends/issues/opportunities that are uncovered in the quality monitoring process. Action Item: Use QA to identify agent training needs. QA specialist and trainers should work closely together to ensure that agents receive the training they need to consistently deliver an outstanding customer experience. © 2009 DMG Consulting LLC
All rights reserved.
- 59 -
April 2009
new 9.6 Addressing Agent Performance Issues For a QA program to be taken seriously by agents and be effective in changing their behavior, its results must be reflected in each agent’s mid‐year and annual employee performance review. Quality assurance evaluation results should be an important component of the process for nter ter QA Guide
Contac t Cen
deciding the raises for agents. However, for this to happen, contact center management must work with the human resource group, as they generally set the format for employee reviews. Agents should never be surprised by the results of their performance reviews because they should be receiving monthly, or at least periodic coaching sessions throughout the year as part of the QA program. The following five steps will help quality reviewers and managers assist agents in meeting contact center performance goals. •
Diagnose performance issues Managers must be able to clearly articulate performance issues so that agents know what they need to change and how to fix it. When preparing an agent’s performance appraisal, review their quality assurance evaluation forms to determine where performance opportunities exist and identify any trends. For example, determine if the agent consistently has low or failing scores low in a particular category or for a particular call type. Are there multiple performance issues, i.e., deficiencies in communications skills or problems adhering to procedures, accessing information or processing transactions? Valuable insights can be gained by conducting a side‐by‐side QA session with the agent to view first‐hand where and when challenges arise or if work habits are contributing to their performance issues. Based on the results of the trend analysis and/or observations from the side‐by‐side sessions, document the underlying causes of performance problems. •
Create an action plan Once the area(s) where the agent needs help have been identified, create an action plan to address them. (Most performance reviews include a section for next steps or performance improvements.) The contact center manager should discuss the findings with the agent’s © 2009 DMG Consulting LLC
All rights reserved.
- 60 -
April 2009
supervisor, and then work with the supervisor and training department to identify the resources that can be used to support the agent's development. This may include enrolling the agent in additional training or up‐training sessions, eLearning or coaching sessions, providing reference materials or job aids, offering additional system or communication skills nter ter QA Guide
Contac t Cen
training, or regularly assigning time for the agent to sit with an outstanding colleague to learn the correct way to handle interactions. •
Communicate with the agent Depending on the structure of the organization, either the QA manager and/or the supervisor should discuss performance issues with the agent. Begin by making sure that the agent understands the quality evaluation criteria and how calls are scored. Be sure that the assessment is constructive and recognizes the agent's strengths as well as performance opportunities. Review the action plan that has been developed with the agent and emphasize that it is intended to help them succeed. (While there are exceptions, most people want to succeed but don’t always know what they need to do. Effective communication may convert weak agents into top performers.) •
Provide consistent feedback and reinforcement The most effective improvement programs provide continuous and encouraging feedback to motivate contact center agents. When an agent is struggling to perform, it's a good idea to offer him/her some extra attention. Conduct routine evaluations of the agent's calls and provide timely feedback. When needed, provide additional targeted coaching that addresses the agent's specific needs and reinforce what he/she is doing right. This feedback can be formal or informal. For example, when a supervisor is walking by, he/she could take a few minutes to stand by and listen to the agent and provide immediate feedback. Or, if the agent is remote, the supervisor could dial in and listen to some calls to provide additional feedback. Agents who know their management cares almost always perform better. © 2009 DMG Consulting LLC
All rights reserved.
- 61 -
April 2009
•
Monitor, track, and recognize improvements Monitor the agent's performance to track progress and improvement. Motivate the agent by recognizing and praising incremental improvements and continue to provide coaching and nter ter QA Guide
Contac t Cen
support for performance opportunities. Reward employees for good performance. Provide monthly recognition to employees with the best quality. Money is a nice reward, but public recognition can keep employees focused on what is important: quality and customer satisfaction. Action Item: Integrate the results of the QA program into the department’s semi‐annual and yearly review process. Address agent performance issues identified during the QA process on a timely basis with directed, constructive and positive feedback. © 2009 DMG Consulting LLC
All rights reserved.
- 62 -
April 2009
9.7 Rewards and Recognition It’s important for agents to consider QA a departmental initiative that contributes to their nter ter QA Guide
Contac t Cen
success, instead of a “gotcha” program. This can be achieved by using QA to readily identify and reward outstanding performers. Establish a process for identifying and recognizing agents who achieve quality monitoring excellence. This motivates the right behaviors and performance throughout the department. QA reviewers, in cooperation with managers and trainers, should identify and distribute best practice clips of outstanding agent interactions to all agents. Agents learn best from their peers, so in addition to providing recognition, best practice clips provide excellent examples of effective techniques for all agents to emulate. There are many ways to acknowledge top performers, but it is essential to do so on a consistent basis, and to make sure that the same people are not the ones always recognized. (Avoiding this common pitfall will help the QA program succeed.) Among the ways to recognize different agents on a monthly basis is to identify the top three monthly performers, the most improved agent from the prior month, the top three performers for the quarter, etc. Incentives are ideal, but if prohibited, then some other vehicle such as a lunch voucher, picture and write‐up in the contact center or company newsletter, inclusion in a “walk of fame,” special parking spot, lunch with the CEO, etc., can be used. Invite top performing agents to take an active role in departmental activities, such as coaching new hires, delivering an up‐training session, becoming a subject matter expert on a new initiative, or cross‐training on a new function. This motivates the right performance behaviors and keeps people engaged. Action Item: Identify and reward outstanding performers, but build a process that rewards different people on a monthly basis to broaden the appeal and acceptance of the QA program. © 2009 DMG Consulting LLC
All rights reserved.
- 63 -
April 2009
9.8 Updating Procedures/Training Quality assurance is a continuous cycle of improvement and measurement that must be nter ter QA Guide
Contac t Cen
adapted as products, processes, policies, customer expectations, systems, or business requirements change. It’s important to develop a process for ensuring that guidelines, reference and training materials are updated on a timely basis so that agents always have the right information. The QA manager or specialists should be involved in keeping procedures up‐to‐date, as should the training team. The appendix includes a sample format that managers can use to create policies and procedures. Action Item: Set up a process that keeps departmental policies, procedures and training materials up‐to‐date, based on feedback received from the QA program. © 2009 DMG Consulting LLC
All rights reserved.
- 64 -
April 2009
9.9 Monitoring Quality Assurance Reviewers Just as agents need to be evaluated on a consistent basis to monitor and measure their performance, the same holds true for quality assurance reviewers. Managers must oversee the performance of QA reviewers to make sure that they are meeting their productivity goals for the number of QA reviews and coaching sessions each month and that they are doing so at the highest nter ter QA Guide
Contac t Cen
level of quality. Exit interviews of agents have found that one of the main reasons they leave is because they do not get timely feedback (and in many cases, any feedback) from management and don’t think the organization cares about them. Calibration, which is discussed in section 8, is important for maintaining the integrity and consistency of the QA program. If managers attend some of the calibration sessions – which is a best practice – they can also identify QA specialists who are struggling with the program. Management can then address the performance issues and get the QA specialist re‐aligned with the goals of the QA program. It is also a good idea for the QA manager or the contact center manager (if they are overseeing the program) to conduct monthly or periodic audits of all reviewers. Here is how the audit process works: •
The QA manager sets an accuracy threshold that QA reviewers must maintain in order to pass the monthly audit. •
The QA manager selects a certain number of interactions per channel to include in the audit (at least the same number of interactions that are evaluated for each agent per month). •
The QA manager evaluates the interactions selected for the audit to create a baseline/master evaluation against which every QA evaluator’s performance is compared. •
The QA manager sends the group of transactions to be used for the audit to all of the QA reviewers to evaluate independently. The completed evaluations are returned to the QA manger; each evaluation is compared against the baseline evaluation for that particular interaction to identify scoring variances. © 2009 DMG Consulting LLC
All rights reserved.
- 65 -
April 2009
•
The QA manager produces a report that reflects scoring variances at the question, section and overall evaluation level; based on the established accuracy threshold, the QA manager tracks the performance of each reviewer and whether they or passed or failed the audit; the manager is nter ter QA Guide
Contac t Cen
looking to make sure that each QA reviewer is evaluating transactions consistently. •
Based on the performance opportunities uncovered in the audit process, the QA manager should reach out to various QA reviewers and provide feedback and coaching. •
Any QA reviewer who fails the monthly audit for more than three months and has been coached should be removed from performing QA evaluations until they are fully retrained. •
After retraining a QA reviewer, audit them again to determine if they are re‐aligned with the rest of the team; if they are, give them positive feedback and, if not, remove them from doing QA reviews. Action Item: Management should evaluate QA evaluators on a periodic basis to keep the QA program aligned with the needs of the business. It will also make it clear to agents that management is invested in this program and taking it very seriously. This will help reduce complaints from agents about being treated unfairly. © 2009 DMG Consulting LLC
All rights reserved.
- 66 -
April 2009
9.10 Reporting Accurate, timely and effective reporting is a requirement for any major initiative. An nter ter QA Guide
Contac t Cen
important part of setting up a QA program is to define the organization’s reporting needs. If QA is being done manually, it’s necessary to design and create the report templates. If using a QA application, work with the vendor to enhance or create custom reports if the solution does not have all the reports needed. However, regardless of the approach used, it’s a good idea to define the organization’s reporting requirements early on in the process in order to know what information needs to be collected. Below are a few reports that many organizations have found useful. Contact Center Quality Management Report for: December 2008 Average Contact Center QA Score: This report displays the average QA score for the contact center for the current month and as compared to the last 2 months. It is useful for determining the contact center’s overall quality performance. Figure 12: Average Contact Center QA Score: December 2008
Month
December
November 2008
October 2008
Average QA Score
80.5%
74.6%
81.3%
© 2009 DMG Consulting LLC
All rights reserved.
- 67 -
April 2009
100%
81.30%
74.60%
80.50%
Oct-08
Nov-08
Dec-08
80%
nter ter QA Guide
Contac t Cen
60%
40%
20%
0%
Average Scores by Evaluation Section: This report displays the monthly average scores for the contact center at the evaluation section level, and as compared to section averages over the past two months. Figure 13:Contact Center Average Scores by Evaluation Section:
December 2008
Section
December
November
October
Greeting (5 pts)
3.4
2.3
3.9
Verification (10 pts)
Plan/Benefit Knowledge (25
pts)
Inquiry Resolution (25 pts)
System Knowledge/Usage
(10 pts)
Hold/Mute/Transfer (6 pts)
Communication Skills (14
pts)
Closing (5 pts)
8.6
7.6
6.8
22.7
21.8
22.3
19.6
21.4
18.9
6.1
7.8
5.9
4.3
4.7
3.9
9.7
8.1
7.5
5.1
3.2
© 2009 DMG Consulting LLC
All rights reserved.
4.7
- 68 -
April 2009
25
20
10
5
(2
R
5
es
pt
ol
Kn
s)
ut
ow
io
n
le
(2
dg
5
e/
pt
U
H
s)
sa
ol
d/
g
e
M
(1
ut
C
0
e/
om
pt
Tr
s)
m
an
un
sf
er
ic
at
(6
io
pt
n
s)
Sk
ills
(1
4
pt
C
s)
lo
si
ng
(7
pt
s)
s)
pt
io
n
w
le
dg
e
ry
ui
In
q
em
st
Sy
Pl
an
/
Be
n
ef
it
Kn
o
Ve
rif
ic
at
re
et
in
g
(1
0
(5
pt
s)
0
G
nter ter QA Guide
Contac t Cen
15
Average Scores by Evaluation Question: This report displays the monthly average scores for the contact center at the evaluation question level, and as compared to specific question averages over the past two months. The purpose is to surface new trends or areas where training opportunities exist, and to determine if the prior month’s training was effective. © 2009 DMG Consulting LLC
All rights reserved.
- 69 -
April 2009
Figure 14: Average Scores by Evaluation Question: December 2008
December
November
October
nter ter QA Guide
Contac t Cen
Greeting (5 pts)
Used call greeting as defined in
Greeting Policy (2 pts)
Documented caller name and phone
number (3 pts)
Verification (10 pts)
Verified caller as defined in
Verification Policy (5 pts)
Verified eligibility as defined in
Eligibility Policy (3 pts)
Asked caller for Member and/or
Patient’s Date of Birth (1 pt)
Asked caller for address on file (1
pt)
Plan/Benefit Knowledge (25 pts)
Accurately identified plan type,
benefit, or claim (5 pts)
Demonstrated thorough knowledge
of plan type, benefit or claim
information (5 pts)
Provided complete and accurate
information or instructions in
accordance with established
procedure (15 pts)
Inquiry Resolution (25 pts)
Accurately understood the nature of
the caller’s inquiry (5pts)
Effectively/accurately resolved
inquiry/issue in accordance with
established procedure (5 pts)
Completed fulfillment/
referral/follow-up as
promised/required (5 pts)
Provided alternatives, as
appropriate (3 pts)
Provided time frames, as
appropriate (2 pts)
Ensured caller fully understood
explanation, process, and/or next
steps (5 pts)
© 2009 DMG Consulting LLC
All rights reserved.
- 70 -
April 2009
December
November
October
nter ter QA Guide
Contac t Cen
System Knowledge/Usage (10 pts)
Effectively accessed and utilized all
appropriate systems, screens and
fields to obtain information to
resolve inquiry (8 pts)
Accurately utilized wrap-up (2 pts)
Hold/Mute/Transfer (6 pts)
Utilized hold/mute as defined in
Hold/Mute Policy (2 pts)
Performed transfer as defined in the
Warm Transfer Policy (2 pts)
Transferred the call to the correct
area (2 pts)
Communication Skills (14 pts)
Maintained a courteous, pleasant,
and respectful tone throughout the
call (3 pts)
Conveyed information clearly and
confidently and in a manner that
was easily understood (3 pts)
Demonstrated effective listening
skills (2 pts)
Expressed empathy and concern as
appropriate (1 pt)
Efficiently managed time and call
flow (call management) (2 pts)
Demonstrated professionalism (call
etiquette) (3 pts)
Closing (5 pts)
Used call closing as defined in Call
Closing Policy (2 pts)
Accurately documented information
in accordance with established
procedure (3 pts)
© 2009 DMG Consulting LLC
All rights reserved.
- 71 -
April 2009
Average QA Scores by Agent: This report displays the average QA scores for each agent in the contact center, for the current month and prior two months. It is useful in identifying agents who are not meeting goal and individual agent performance trends. nter ter QA Guide
Contac t Cen
Figure 15: Average QA Scores by Agent: December 2008
Agent
December
November
October
Agent Name
Agent Name
Agent Name
Agent Name
Agent Name
Agent Name
Agent Name
Agent Name
QA Scores by Agent by Evaluation: This report provides a breakdown of the scores that each agent achieved for their evaluations in the current month. It is helpful in identifying very good or very poor performance or interactions that may require attention or follow‐up by the supervisor. Figure 16: QA Scores by Agent by Evaluation: December 2008
Evaluation Scores for December 2008
Agent
Evaluation 1
Evaluation 2
Evaluation 3
Evaluation 4
Agent Name
Agent Name
Agent Name
Agent Name
Agent Name
Agent Name
Agent Name
Agent Name
© 2009 DMG Consulting LLC
All rights reserved.
- 72 -
April 2009
Average QA Scores by Agent by Category: This report provides a breakdown of the average scores by category for each agent. It helps identify individual needs and areas of coaching opportunities specific to each agent or team Closing
Hold/Mute/Transfer
System
Knowledge/Usage
Inquiry Resolution
Plan/Benefit
Knowledge
Agent
Verification
Figure 17: Average QA Scores by Agent by Category: December 2008
Greeting
nter ter QA Guide
Contac t Cen
Action Item: Invest the time up front to define the reports needed to support your QA initiative. Ask all relevant constituents – contact center and QA managers, supervisors, QA specialists, trainers and possibly, marketing – what information they would like to see from the QA program, and design reports to meet these needs. © 2009 DMG Consulting LLC
All rights reserved.
- 73 -
April 2009
9.11 QA Database If an automated application is not being used for QA, it is necessary to build a database for nter ter QA Guide
Contac t Cen
managing QA data. Most organizations build this in either Access or Excel. The goal is to make it easy to input and retrieve data and to make it easy to run reports. Action Item: If you are not using a QA/recording system, build a database for tracking agent QA scores so that this data can be used to identify trends and create reports. © 2009 DMG Consulting LLC
All rights reserved.
- 74 -
April 2009
10. Kicking Off the QA Program Once the QA program is built, it’s time to kick it off. The entire contact center staff must be Cont
C
ontact Center QA Guide
trained prior to the beginning the pilot. This is very important because it sets the tone for the program. This is also where an organization can introduce their own “personality” into the program. DMG suggests coming up with a name for the QA program so that it isn’t just called “the QA program.” Some organizations run a name contest, which is another way of getting buy‐in and support from contact center staff. The kick‐off presentation should address the following topics: •
Introduce QA leaders •
Explanation of QA •
Description of its benefits •
What QA means to agents, including how it can help their careers •
Discussion of program mechanics •
Review of the QA evaluation form •
Upcoming training •
Rewards and recognition Action Item: Make the QA kick‐off a celebratory and fun affair that explains the benefits of QA for agents, customers and the greater enterprise. Use this kick‐off to get staff buy‐
in. © 2009 DMG Consulting LLC
All rights reserved.
- 75 -
April 2009
10.1 Agent Training Training is necessary to prepare agents for the QA program and the changes that it will bring to the department. Training is also a great way to alleviate apprehension about the program; the nter ter QA Guide
Contac t Cen
more information people have about the QA program, the more comfortable and welcoming they will be. The outline in Figure 18 below presents the topics that should be covered in an agent QA training program. However, every contact center should customize this basic training program to meet their needs. Additionally, going forward, a QA module should be added to the new agent hire training program. Figure 18: QA Agent Training Outline
Agenda
Item/Activity
QA and its
Benefits
Topics covered
•
•
Procedures
Time
1.5 hrs
•
•
•
•
What is QA?
How will QA benefit you and your
customers?
How was the program developed?
What are the program components?
What are the program goals?
Who are the people that will be involved?
•
•
•
•
•
What are procedures?
Why are they important?
How should you use them?
How does QA use them?
How are they modified and updated?
3.5 hrs
If the procedure guide is new, the trainer will have
to review the document and explain how it is
organized and used.
Policies
•
Role play to get accustomed to using the
procedures.
•
•
•
•
•
What are policies?
How do policies differ from procedures?
How should you use them?
How does QA use them?
How are they modified and updated?
2.5 hrs
If the policy guide is new, the trainer will have to
review the document and explain how it is
organized and used.
•
Role play to get accustomed to using the
policies.
© 2009 DMG Consulting LLC
All rights reserved.
- 76 -
April 2009
Agenda
Item/Activity
QA Form and
Criteria
Topics covered
•
•
•
•
•
•
•
•
nter ter QA Guide
Contac t Cen
Program
Mechanics
•
•
•
•
•
•
Agenda
Item/Activity
Calibration
Time
How was the form developed?
What is the form’s purpose?
How is the form used?
How is the form scored?
What are the criteria used to score each
question/section?
How will the scores be used?
How will the program work?
How many calls will be monitored per
week/month?
Who will be doing the evaluations and
coaching sessions?
How will calls/emails/chats be selected for
evaluation?
How will I receive my evaluations?
When can I review my evaluations?
How can I respond to my evaluations?
What if I disagree with my score?
Topics covered
•
•
•
•
2 hrs
2.5 hrs
Time
What is calibration?
Why is calibration necessary?
How often is calibration conducted?
Who participates in calibration?
2.5 hrs
Conduct a calibration session with agents using
mock calls to demonstrate how calls will be
evaluated and scored. (Agents will see that this is a
complicated process that requires a great deal of
skill.)
Evaluator Audits
•
•
•
Who will be audited?
Why are audits important?
How does the audit process work?
Recognition
Program
•
•
What are the recognition criteria?
How does the rewards and incentive
program work?
.5
Program Pilot
•
•
•
•
What is a program pilot?
How does the pilot work?
How long will the pilot be conducted?
Will scores during the pilot count?
1 hr
.5 hr
Action Item: Create a QA training program to teach agents about all aspects of the QA program and its benefits. Use this program to kick off QA. Also incorporate a QA module in your new agent training program. © 2009 DMG Consulting LLC
All rights reserved.
- 77 -
April 2009
10.2 QA Program Pilot After training all contact center staff and other interested parties, it’s time to conduct the nter ter QA Guide
Contac t Cen
QA program pilot. There are many opinions about pilots and their purposes, and this is certainly true for QA. If this is the organization’s initiation into QA, the pilot is necessary to give management, QA reviewers and agents a hands‐on opportunity to get acquainted with QA in a non‐confrontational environment where the QA scores are not being counted. If the organization is updating its QA program, the pilot gives everyone an opportunity to get used to the new QA criteria and evaluation forms. The timing of the pilot also varies based on the needs of the organization. If QA is new to the organization, we suggest beginning with a 2‐ to 3‐month pilot. Then, once the QA scores begin to be counted, we suggest setting a lenient QA goal so that more people will rate well and feel good about the program. Then, after a 3‐month period, during which there is a great deal of training and coaching, it’s time to begin in earnest and to set up QA goals that reflect customer expectations for great service. While this means that it will take 4 to 6 months before the QA program is in full operation, since the goal of QA is to improve agent performance and the customer experience (and by doing so to improve agent productivity), it will not delay the benefits. It will, however, allow for positive momentum to build and help the program succeed. The pilot is a great opportunity to make sure that the program achieves its goals. Throughout the pilot, QA leaders should be open and welcoming of all observations and feedback. Use the feedback and time to enhance and tweak the QA form and evaluation criteria, and address training issues. Action Item: Communicate and share pilot results on a regular basis to keep contact center staff engaged and alleviate apprehensions about the program. © 2009 DMG Consulting LLC
All rights reserved.
- 78 -
April 2009
11. Advanced Quality Assurance Initiatives This Guide describes the process for building a contact center quality assurance program. nter ter QA Guide
Contac t Cen
Once the program is in place and working well, some organizations want to go further and extend the benefits and uses of their QA program. QA programs can be enhanced by adding surveying, using QA to determine the first call resolution rate (FCR), or to do customer experience monitoring. Below is an explanation of these three business activities. 11.1 Surveying The best way to know if customers are satisfied with the quality of an organization’s products, services, processes and agents is to ask them. QA measures how well agents adhere to internal policies and procedures; it provides an internal view. Surveying captures the customer perspective; the external view. When survey feedback is combined with QA results the company learns what customers consider good service and specifically, which agents provide it. They also find out what processes and policies need to be changed. When done right, sharing customer survey information about agent performance can help improve quality. Agents find it helpful when they see survey results and read or hear customer feedback first‐hand. It helps them appreciate how their performance impacts customer satisfaction and the customer’s perception of their company. 11.2 Customer Experience Monitoring Customer experience monitoring is intended to provide insights into the total customer experience. Typically, contact center managers use this process to track and review interactions that have multiple segments, are put on hold, transferred, required conferences, or are repeat calls. Customer experience monitoring is done to evaluate the overall customer experience, not just call segments. © 2009 DMG Consulting LLC
All rights reserved.
- 79 -
April 2009
11.3 First Call Resolution (FCR)
First Call Resolution is a unique key performance indicator that measures the effectiveness, nter ter QA Guide
Contac t Cen
efficiency and customer satisfaction level of a contact center. It is the only single metric that provides a balanced view of the contact center’s overall performance. The challenge in using FCR is that it is difficult to determine which calls are fully resolved during the first contact. There are many ways to identify or calculate FCR, including asking the QA staff to determine if calls are resolved during the first contact based on the interactions that they evaluate. © 2009 DMG Consulting LLC
All rights reserved.
- 80 -
April 2009
12. Quality Management/Liability Recording Suites Quality Management/Liability Recording suites, also known as workforce optimization (WFO) nter ter QA Guide
Contac t Cen
solutions, deliver actionable insights and findings in addition to incorporating built‐in automated efficiencies to increase the effectiveness of the QA process and reviewer productivity. WFO suites are complex solutions that include 8 modules intended for contact centers. Complete workforce optimization suites include eight functional modules. See Figure 19. The core modules are recording (either time division multiplexing (TDM) or Internet Protocol (IP)‐based), and quality assurance. Recording systems log calls for regulatory compliance, protection against lawsuits, and quality review. Quality assurance applications are used to determine how well agents adhere to internal policies and procedures. The other six modules are: 1. Workforce Management – forecasts and schedules agent staffing needs 2. Agent Coaching – tools to communicate with agents to assist them in improving their performance 3. eLearning – learning management capabilities that allow training courses to be created, issued and tracked to ensure effectiveness 4. Surveying – Web and IVR‐based solutions for creating, issuing, tracking and analyzing customer feedback 5. Performance Management – solutions that help align contact center activities with enterprise goals; also provide scorecards and dashboards to improve the performance of the contact center 6. Speech Analytics – captures, structures and analyzes customer phone calls to identify the reasons for customer calls and to gather insights The value and benefits of these suites increase for the enterprise as additional modules are added. © 2009 DMG Consulting LLC
All rights reserved.
- 81 -
April 2009
WFO Suites In the last couple of years, use of workforce management, recording and, to a lesser degree, continue because these solutions offer quantifiable benefits for many operating areas in an enterprise. Figure 19: Workforce Optimization Suites Enterprise
Back Office
Speech Analytics
Performance Mgmt.
Contact Center
Surveying
Call Center
Value Proposition
nter ter QA Guide
Contact Cen
quality assurance modules has expanded into enterprise back offices and branches. This trend will eLearning
QM/Recording
SUITE
Agent Coaching
Workforce Mgmt.
Quality Assurance
Recording
Source: DMG Consulting LLC, April 2009 © 2009 DMG Consulting LLC
All rights reserved.
- 82 -
April 2009
In the last few years, the QA vendors have delivered many innovations to the market to make these applications more helpful and actionable for contact centers. These enhancements include: nter ter QA Guide
Contac t Cen
•
Automated call scoring – every recorded interaction is systemically scored, allowing rapid identification and action on calls that do not meet pre‐defined quality requirements. Instead of evaluating a random sample of calls, this enables organizations to focus limited QA resources on monitoring high‐impact calls. •
Flexible form development environments – the ability to easily create a wide variety of evaluation forms and associate the most appropriate form based on call type, category or disposition. Built‐in efficiencies include wizards for form creation or the ability to clone and modify an existing form, a wide variety of scoring methodologies, the ability to create forms based on skills or call components, the ability to add evaluator hints for scoring criteria, comment boxes to capture coaching tips, spell check, form preview and test capabilities, and the ability to create specialized forms for different types of evaluations such as agent self‐
evaluations, calibration or customer experience monitoring. •
Calibration – the ability to designate calls for calibration, automatically deploy calibration sessions to supervisors and reviewers, track completion, and report on scoring variances at the question, section or form level. •
Call categorization – the ability to systemically or manually categorize calls so that QA efforts can encompass all call types received by the contact center and evaluate how agents perform across all functions. This also allows reviewers to categorize calls and designate them for review by the training department or a cross‐functional team if a business process optimization opportunity exists. Search/retrieval/replay – the ability to search and retrieve transactions based on a wide variety of criteria and metadata. Automatically delivers transactions that a reviewer is responsible for evaluating to the desktop and tracks completion status. © 2009 DMG Consulting LLC
All rights reserved.
- 83 -
April 2009
During desktop replay, evaluators can bookmark call segments and annotate them with coaching comments that are tied to specific sections, questions or skills in the agent’s Contact Center QA Gu ide
evaluation. •
Coaching/eLearning – allows supervisors to easily create, deploy and track personalized coaching sessions that address agent‐specific performance opportunities. Includes the ability to share best practice clips with agents and the training department. eLearning capabilities allow reviewers to assign learning sessions to agents while they are performing an evaluation, or automatically deploy the courses based on performance thresholds. •
Reports and dashboards – automated summary and detailed reports that can be scheduled to run or designated as a dashboard report for tracking and trending QA results. The Market Place The QM/recording (WFO) technology sector is a global market with more than 45 competitors. There are many strong offerings available to satisfy the needs of companies large and small. The solutions, pricing and vendor specialties vary widely. This market gives prospects a wide variety of opportunities to acquire a solution that meets the specific needs of their company at a price point they can afford. Users also have many options for acquiring these solutions; they can buy, use hosting, or go with a managed care offering. Action Item: Determine if a QA/Recording or WFO suite will add value to your organization. If it will, find the right one for your company. © 2009 DMG Consulting LLC
All rights reserved.
- 84 -
April 2009
Appendix A: Procedure Format Sample nter ter QA Guide
Contact Cen
Situation: Caller is requesting a change of address.
Important: Address changes can only be requested by the primary/joint
card member.
Greeting (as defined in Greeting Policy)
Complete Verification Procedure:
•
•
•
•
Caller name
Last 4 digits of SSN
Mother’s maiden name
Date of birth
•
IF THE CALLER IS NOT THE PRIMARY/JOINT CARD MEMBER:
Access: Memo screen
Check: [ Special Instructions ] field to determine if the caller has authorization to make
changes on the account (power of attorney)
If caller is not listed as someone with authorization and/or power of attorney (POA) in the
special instruction field:
Advise: Only the primary/joint card member is authorized to request an
address change
Options:
The primary/joint card member will need to call to request the address change.
– or –
The primary/joint card member can:
Write to:
Fax to:
Process online at:
•
IF THE CALLER IS THE PRIMARY/JOINT CARD MEMBER:
Verify:
•
Current address
Check: [ Address ] field on the Customer screen
© 2009 DMG Consulting LLC
All rights reserved.
- 85 -
April 2009
Advise:
If new address is on the system:
Tell the account holder that we have the
correct address on file.
nter ter QA Guide
Contac t Cen
If not:
Access: Account Maintenance screen.
Obtain new address from the account
holder and enter the new address in the
appropriate fields.
Read back the new address to the
customer including spelling of the street
address and town to confirm accuracy.
Access: Billing screen
Check: [Last bill date] field to determine
of the next statement has already been
mailed.
If next statement has already been
mailed:
Advise: The new statement has already
been sent to the old address. You will
have a copy of the statement sent to the
new address. The cardmember should
expect to receive the copy with in 5 to 7
business days.
Access: Media Request screen
Click on Statement.
Enter in month/year of the statement that
you are requesting in the [month and
year] fields.
Hit enter to complete the transaction.
If next statement has not been
generated/mailed:
© 2009 DMG Consulting LLC
All rights reserved.
Advise: The next statement will be
mailed to the new address on file.
- 86 -
April 2009
Ask: Is there anything else I can help you with?
nter ter QA Guide
Contac t Cen
If yes, assist card member with the additional request(s).
If no, perform call closing as defined in Call Closing Policy.
Source: DMG Consulting LLC, April 2009 © 2009 DMG Consulting LLC
All rights reserved.
- 87 -
April 2009
Appendix B: Policy Format Sample nter ter QA Guide
Contac t Cen
Warm Transfer Policy
All transfers initiated from the contact center should be warm transferred by performing
the following steps:
•
Representative advises the caller that the call needs to be transferred to another
area that can assist the caller with their request
•
Representative asks the caller for permission to transfer the call
Before initiating the transfer, the representative should ask the caller if there is anything
else that they can assist the caller with
•
If yes, the representative should assist the caller with the request (if
applicable)
•
Representative asks permission from the caller to place them on hold while they
connect to the appropriate department to transfer the call
•
Representative clicks on the phone icon which will launch a separate window
•
Representative clicks “Start conference,” enters the extension number of the
department that is receiving the call, and clicks “OK”
•
When the receiving department answers the call, the representative should
provide:
•
•
Account number
•
Name of caller
•
Nature of the call
Representative clicks “Complete conference” and introduces the caller to the
representative in the department who is accepting the call, thanks the customer for
calling and clicks the “Hang-up” button to complete the transfer
Source: DMG Consulting LLC, April 2009 © 2009 DMG Consulting LLC
All rights reserved.
- 88 -
April 2009
About VPI VPI (Voice Print International, Inc.) is the premier provider of integrated interaction recording and workforce optimization solutions for enterprises, small‐ to medium‐size nter ter QA Guide
Contac t Cen
businesses, trading floors, government agencies, and first responders. Through its award‐
winning suite of solutions, VPI empowers organizations to proactively improve the customer experience, increase workforce performance, ensure compliance, and align tactical and strategic objectives across the enterprise. With the power to be proactive, organizations are equipped to actively identify and maximize opportunities and minimize risk. For more than a decade, VPI has been providing proven technology and superior service to more than 1,000 customers in over 35 countries. For more information, visit http://www.VPI‐corp.com. © 2009 DMG Consulting LLC
All rights reserved.
- 89 -
April 2009