344 results found
-
Populate links within email/customer interaction when reviewing evaluations
A few of my analysts and myself have come across the issue of hyperlinks not showing up when evaluating. Portions of the interaction that are hyperlinked look like plain, regular text while evaluating and appears as though there aren't any when there are.
This has caused unintentional mark offs for agents for not including a link to resources when there was in the Salesforce case. This has only become a recent issue. If possible, can we have the hyperlinks included (again) when evaluating? It does appear to show in the interaction itself before clicking to start evaluation but not during…6 votes -
Ability to use decimals in a workload for QA assignment
We need the ability to use decimals when dividing the estimated volume to QAs in a workload. I.e. Currently, if you have 5,000 interactions estimated to be divided among 100 QAs the difference between the allocation of 1% (50 interactions) and 2% (100 interactions) is not providing a fair distribution.
5 votes -
Ability to set a fixed number in a workload to analyst's distribution.
Currently, the analyst distribution step in a workload only lets us distribute a percentage of the estimated interaction volume to be assigned. We need to be able to set a fixed number for each QA analyst to distribute the estimated volume. That way, across multiple workloads every week we can determine their total assignment.
13 votes -
Option of replacing completed evaluations with Expert feedback/score after calibrations
Currently, when reviewing the outcomes of calibrations, they may result in score adjustments to the original evaluation and it is a manual action taken right now. If we could have a button as an option to replace the original evaluation score with the expert's updated feedback/scoring and have a banner that it was updated due to calibration #, that could potentially be more efficient. Ideally keeping the original evaluation timestamp/date/analyst associated with it.
4 votes -
Export evaluation logs (views, edits, etc) instead of having to access it one by one
Right now it's not possible to audit the logs for all evaluations at the same time, there's no visibility. We have to access each individually, and it's not scalable for reporting.
This could be fixed by implementing the possibility to export the logs, or at least view them, much like we export evaluations or view Audit logs for account creation.5 votes -
expert in a calibration session
The client would want for the expert in a calibration session, to see the answers of all the participants in the session before he sends his own evaluation.
2 votes -
See the drafted version of the evaluation they started earlier on the actual workload
Whenever the analyst is completing a workload, the client would like to see the drafted version of the evaluation they started earlier on the actual workload without needing to go to the "draft" tab. Also, to avoid confusion the client would want to have only one draft per ticket/ evaluation and not multiple for every time he saves that same evaluation as a draft.
5 votes -
automated workload
The client would like to have automated workloads. Whenever the workload does not get enough samples to match the ones requested when creating it (because no other interactions match the filter and date range requirement), these missing evaluations should be added automatically to the next trigger and therefore complete the quota of established tickets to evaluate.
1 vote -
Evaluations increase or decrease based on QA score
The client would like to have automated workloads. The number of evaluations increases when the agent's overall QA score is low and the number of evaluations decrease when the agent's overall QA score is high.
10 votes -
Feedback Options
Point Value scorecards have an option to add checkboxes for each score (unable to provide screenshot stating file is not supported - JPEG, PNG, GIF, HTML tried all)
These are a great way to tag behaviors based on the score and help analysts/supervisors see what behavior lead to that particular score. However, these checkboxes are not trackable in reporting and does not export easily to track these behaviors.2 votes -
The ability to change the scorecard in an evaluation.
For example, to change the scorecard from mail to call during an evaluation.
4 votes -
Reporting filters for Review stats
Include further filters in the reporting of review stats (by analyst, team, etc.)
2 votes -
Track Deleted Audits
Ability to pull a report showing all the evaluations that have been deleted, who deleted, when deleted and even some details of the evaluation - agent name, date of eval, quality score etc?
3 votes -
Agent performance by question group, not just section
Currently, we can pull individual agent reports on the scorecard section, but not by question group. It would be great for us to export this easily instead of needing to export all evals individually and then work with the data to get the information we need.
2 votes -
Review the Analyst Scorecard - weighting
In the "review the analyst" scorecard, allow weighting on the scorecard.
Currently, there is no way to set the weight and when comparing evaluations the score does not match when completed. This leads to questions from analysts why the score does not match when the evaluation questions do match.1 vote -
Dynamic Team Selection/Historic Data Accuracy
Would love to see the ability to adjust evaluation data so that if a monitor is submitted against the wrong team, the results can be fixed. I'd prefer that evaluations(at least in the team sense) are dynamic instead of static.
EXAMPLE - If I submit against Team B and meant to evaluate Team A, I want to be able to adjust the evaluation to show up against Team A retroactively without having to delete, submit again, and scrub the inital results from our data lake. This ensures that all of our front-end reporting matches results outside of Playvox and that…
3 votes -
Error calculated on 0 point questions
When a question has a 0 point value & the agent gets the question correct it is counted as part of the error rate. We would ask to have 0 point questions only counted as an error if the question is incorrect
2 votes -
Slider scale for Review scorecards
There is no option on a review scorecard to have a scale score (eg. 0-5) only 'points'. There are questions that need to be marked on a scale.
1 vote -
Score Breakdown in Evaluation Notification
We currently have 2 questions that zero the scorecard, but agents are finding this very demotivating and the 0% score overrides all the good they have achieved in the other 10 questions.
Would it be possible to present the score in the notification email, and within Playvox, broken down into 2 parts?
e.g.
Your quality score (Q1-10) is 80%
Your compliance score is 0%
Your overall score is 0%Your quality score (Q1-10) is 95%
Your compliance score is N/a
Your overall score is 95%8 votes -
Backlog Completion Settings
If a workload is not completed before a new assignment is distributed, analysts are unable to work on the previous items that were assigned to them and can create a gap in work that QA needs to complete. This must manually be captured and completed outside of the workload.
Would love to see the ability to complete the previous assignment and/or have the functionality to catch this and the future assignment to reflect what was missed. This will ensure that there are no gaps from a controls perspective
2 votes
- Don't see your idea?