This manual provides a comprehensive guide for recruiters using the Glider.ai platform's Insights feature to analyze assessment data, track candidate performance, and optimize recruitment processes.
1. Accessing Insights
Step 1: Log in to your Glider.ai account.
Step 2: Navigate to the Insights tab from the main dashboard.
2. Insights Dashboard Overview
The Insights section includes multiple tabs that offer various performance metrics. Each tab has customizable filters such as assessment title, status, date range, and specific question identifiers.
3. Insights Tabs Overview
3.1 Completion Status
Metrics Provided:
Total unique attempts
Assessment completion categorized as: Completed, Not Started, Reattempted, In Progress
Filters: Assessment title, status, and date range
Use Case: Identify engagement patterns and improve completion rates.
3.2 Pass vs Fail
Metrics Provided:
Candidate pass/fail counts and percentages
Month-to-month completion trends
Filters: Assessment title, status, and date range
Use Case: Evaluate assessment difficulty and adjust recruitment strategies.
3.3 Assessment Scores
Metrics Provided:
Total, maximum, and average assessment scores
Filters: Assessment title and date range
Use Case: Measure overall candidate performance.
3.4 Assessment Time Spent
Metrics Provided:
Minimum, average, and maximum time spent per assessment
Filters: Assessment title and date range
Use Case: Analyze time efficiency and adjust time limits if needed.
3.5 Assessment Attempts
Metrics Provided:
Average and maximum number of attempts per assessment
Filters: Assessment title and date range
Use Case: Identify challenging assessments and improve support.
3.6 Assessment Attempt Details
Metrics Provided:
Attempts distribution by candidate and assessment
Filters: Assessment title, question type, question ID, and date range
Use Case: Monitor specific attempts to detect trends or issues.
3.7 Assessment Difficulty Level
Metrics Provided:
Average percentage score and assigned grades
Filters: Assessment title and date range
Use Case: Balance assessment difficulty levels.
3.8 Question Attempts
Metrics Provided:
Number of attempts, percentage correct/incorrect, and skipped questions
Filters: Assessment title, difficulty level, question type, question ID, and date range
Use Case: Identify frequently missed questions.
3.9 Question Time Spent
Metrics Provided:
Maximum, minimum, and average time spent per question
Filters: Assessment title, question ID, and date range
Use Case: Analyze time spent on complex questions.
3.10 Question-Graded Scores
Metrics Provided:
Grades by question type and difficulty level
Filters: Assessment title, question ID, and date range
Use Case: Evaluate candidate strengths and weaknesses.
3.11 Question Difficulty Level
Metrics Provided:
Calculated and marked difficulty, attempts, and correct responses
Filters: Assessment title, question type, question ID, and date range
Use Case: Refine questions based on difficulty trends.
3.12 Distractor Analysis
Metrics Provided:
Distractor attractiveness measured by point biserial correlation
Filters: Assessment title, question type, question ID, and date range
Use Case: Improve question quality through distractor performance.
3.13 Glider Assessment Analysis
Metrics Provided:
Unique attempts, average scores, reliability indices
Filters: Assessment title, question type, question ID, and date range
Use Case: Comprehensive assessment performance review.
3.14 Glider Question Analysis
Metrics Provided:
Unique attempts, discrimination index, distractor analysis
Filters: Assessment title, question type, question ID, and date range
Use Case: Advanced question-level performance insights.
Was this article helpful?
That’s Great!
Thank you for your feedback
Sorry! We couldn't be helpful
Thank you for your feedback
Feedback sent
We appreciate your effort and will try to fix the article