Automatically evaluate learner responses to open-ended questions and provide instant feedback based on accuracy.
Setup Proces
Add an open-ended question
Enable AI Assessment: Check "This open-ended question has a correct answer"
Provide an "Example of a correct answer": Write an example of what a correct response should include
Create Feedback Messages: Write two responses - one for correct answers, one for incorrect answers
AI Evaluation Process
AI compares learner responses to your example in real-time
Uses semantic matching (understands meaning, not just exact words)
Considers lesson content and training data for context
Classifies responses as "correct" or "incorrect"
Automatically sends appropriate pre-written feedback to learners
Results count toward overall accuracy calculations
Best Practices for Questions
✅ Good for:
Questions with objectively correct answers (e.g., "Name three countries in Africa")
Scenarios with multiple correct approaches but clear wrong answers
Knowledge checks with definitive criteria
❌ Not suitable for:
Creative or purely subjective responses
Opinion-based questions
Best Practices for Writing Examples of Correct Answers
Use conversational, guiding language:
"Some examples might include..."
"The answer should focus on x, y, z"
"Key words the user might mention include..."
Provide criteria and multiple acceptable examples rather than one perfect answer.
Admin Controls
As an admin, you can go to Analytics > Select a Course > Engagement > Responses to:
View individual learner responses and exact text submitted
See correct/incorrect percentages like multiple choice questions
Manually override AI-evaluated decisions
Still need help?
Contact us