How to use assessments to verify capability and reinforce learning across partner certification, customer onboarding, channel quality, franchisee compliance, and employee skill development — using closed-form questions, file uploads, and video coaching to evaluate everything from knowledge to actual performance.
Why Assessments Matter
Every question you ask a learner is a tax.
It costs them time. It costs them attention. It costs them the small but real anxiety of being measured. If the question doesn't earn that cost back — by verifying capability, reinforcing memory, or producing data you'll actually act on — it should not exist.
This is the part most assessment design gets wrong. Teams add quizzes because the platform supports them, because compliance asks for "proof," or because content feels lighter without one. The result is trivia checks that don't predict performance, certification gates that pass everyone, and reports nobody reads.
Assessments in Continu are powerful and broad. They include closed-form questions for knowledge verification, file uploads for submitting work product, and video coaching where the learner records themselves performing a task and a grader scores it against criteria. That range lets you evaluate everything from recall to applied skill. Used well, assessments prove a partner is ready to sell, a customer admin is ready to deploy, a franchise operator is audit-ready, or a new hire has actually absorbed compliance training.
Used badly, they are noise that erodes trust in your program.
This guide is about designing assessments that earn their cost.
What an Assessment Actually Is
An assessment in Continu is a structured set of questions or prompts tied to a piece of content, a track, or a journey, that produces a score, a pass/fail outcome, and a record.
Strip away the mechanics and an assessment does one of two jobs.
Verify — prove the learner has the capability the program is supposed to build. This is what gates certification. The score is consequential.
Reinforce — strengthen memory through retrieval. The act of answering is the learning. The score is informational.
These are different jobs. They produce different question designs, different scoring rules, and different reporting expectations. Conflating them is the most common mistake we see.
The strategic question: is this assessment verifying capability, or is it reinforcing learning? Decide before you write the first question.
The Two Jobs, Side by Side
Verification (summative). End-of-program. High-stakes. The pass mark is a quality bar. Failure has consequences — a certification denied, a deployment blocked, an onboarding step incomplete. Question design favors application over recall. Reporting goes to compliance, partner program managers, and legal.
Reinforcement (formative). Mid-program or post-module. Low-stakes. The pass mark is generous or absent. Failure produces a teaching moment, not a denial. Question design favors recall and recognition. Reporting goes to content owners looking for confused concepts.
A well-designed program uses both. Reinforcement quizzes scattered through a track. A verification assessment at the end. The reinforcement quizzes catch confusion early; the verification assessment proves the capability is real.
A poorly-designed program uses one when it should use the other. A high-stakes verification quiz on trivia. A low-stakes reinforcement check at the certification gate.
Question and Prompt Types in Continu
Continu assessments support multiple formats. Choose the format that matches what you're trying to verify.
Closed-form text questions. Multi choice (single best answer), multi answer (multiple correct selections), and true/false. Auto-scored. Best for knowledge recall, concept recognition, and rule-following. The platform handles the scoring; the design effort is in writing good questions.
Open-ended short answer. Free text response. Reviewed by a human grader. Best when you need the learner to articulate something in their own words — explain a concept, justify a decision, summarize a process.
File upload. Learner submits a file as their answer. Best for work product — a deployment plan, a customer-facing deck, a written analysis, a worked example. The grader reviews the file against criteria.
Video coaching — screen recording. Learner records their screen demonstrating a workflow. Best for verifying applied skill in tools and systems — configuring a feature, running a report, executing a process. The grader scores the recording against criteria.
Video coaching — video recording. Learner records themselves on camera. Best for verifying customer-facing motions — delivering a pitch, handling an objection, leading a discovery conversation. The grader scores against criteria.
The most common assessment design mistake is using the wrong format for the job. Multi choice cannot verify whether someone can actually deliver a pitch. Video coaching is overkill for verifying whether someone knows a definition. Match the format to the capability you're evaluating.
Anatomy of an Assessment in Continu
Every assessment is a composition of these elements.
Questions and prompts. The full range above — closed-form text, short answer, file upload, video coaching. The question types you use shape everything else about the assessment.
Question bank and randomization. For closed-form text questions, a larger pool the assessment draws from, with randomized question order and randomized answer order. Reduces gaming, increases re-take fairness, scales across cohorts.
Scoring and pass mark. The threshold below which the learner fails. For closed-form questions, the platform scores automatically. For short answer, file upload, and video coaching, the grader's score is applied against the pass mark.
Grading criteria / rubrics. For open-form prompts (short answer, file upload, video coaching), the rubric the grader uses. Specific behaviors to look for, specific competencies to assess, specific scoring bands. Without a clear rubric, grading varies grader-to-grader and the program produces noise.
Grader assignment. For graded formats, who grades. A named grader, a grader pool, or a role (e.g., "the assigned manager"). Without a clear grader, submissions stack.
Retake policy. How many attempts. Whether there's a cooldown between attempts. Whether the learner sees correct answers between attempts. Whether the assessment locks after a final fail.
Time limit. Optional. Forces pace. Useful for closed-form verification, often hostile in reinforcement.
Certificate output. On pass, Continu can issue a certificate. The visible artifact partners frame on their walls, customers cite in audits, employees attach to their learning record.
Placement. Standalone (accessible directly), embedded in content (inline check after a video or article), or attached to a track or journey (gating progression).
These components combine into very different assessment shapes. A 5-question reinforcement quiz is one shape. A 40-question, time-limited, question-banked, randomized, certification-issuing final exam is another. A graded video coaching submission where the partner records a 90-second pitch is a third. All are assessments. They share almost nothing in design.
Best Practices
Decide the job first. Verification or reinforcement. Write it down before you write a single question. The job determines the pass mark, the retake policy, the time limit, the format mix, and what the report needs to show.
Match the format to the capability. A capability that's about knowing facts → closed-form questions. A capability that requires producing an artifact → file upload. A capability that requires demonstrating a skill in action → video coaching. Mismatched format is the most common design error.
Set the pass mark on purpose. A 70% pass mark is not a default — it is a claim that 70% of this content is critical and 30% is acceptable to miss. If the content is compliance-critical, 70% may be malpractice. If the content is exploratory, 70% may be punitive. Pick the number that reflects the actual stakes of getting it wrong on the job.
Write questions at the level of the job. If the learner's job is to recognize phishing emails, the question should show them an email and ask if it's phishing — not ask them to define "phishing." If the job is to deliver a pitch, the assessment should be a recorded pitch, not a multi-choice quiz about pitch theory.
Use question banks for anything assessed more than once. If two cohorts of partners take the same certification, they should not see the same 20 closed-form questions. A bank of 60 with random draw of 20 maintains the bar without leaking the answers between cohorts.
Write the grading rubric before you write the prompt. Especially for video coaching and file upload. What specifically counts as a strong response? What counts as a weak one? Behaviors, not adjectives. The rubric is what makes grading consistent and the result usable.
Brief the learner before a video coaching or file upload prompt. Tell them what scenario to enact or what artifact to produce, what behaviors to demonstrate, how long the recording should be, and how it will be graded. An open-form prompt without context produces wandering submissions.
Stage the difficulty of video coaching. A new hire doesn't lead with a customer-facing pitch on day two. Build up — internal demo first, peer practice next, customer-facing recording when the skill is real. The rubric can stay the same; the stakes rise.
Match retake policy to the job. Verification assessments should limit attempts and impose cooldowns — that's how you prove the capability is real. Reinforcement quizzes should allow unlimited retakes and show correct answers — that's how learning happens.
Show feedback on reinforcement, withhold answers on verification. Reinforcement quizzes are teaching tools; explanations of correct answers are essential. Verification assessments are measurement tools; revealing answers between attempts trains the learner on the test, not the job.
Plan the grading workflow before launch. A 200-partner cohort with a video coaching prompt is 200 grading actions, often 5-15 minutes each. A 500-learner cohort with a file upload is 500 grading actions. That's real labor cost. Identify the graders, set expectations on grading turnaround, and build the cadence so submissions don't sit in queue.
Give graded learners real feedback. A score with no comment is unhelpful for a video coaching or file upload submission. The grader's written feedback is where the learning happens. Build feedback expectations into the grader's role.
Pre-test before you certify. When launching a new certification, pilot it with a small group, look at item difficulty (which questions everyone gets right or wrong, which video coaching submissions stump the graders), and revise. A question 100% of pilot-group passes signals it's too easy or too obvious. A question 100% fails signals the content didn't teach it.
Tie the certificate to something real. A certificate that issues on pass should mean something — listed on a partner directory, required for a portal role, displayed on the learner's profile, shared as proof in a sales conversation. A certificate that issues but is never seen again is theater.
Anti-Patterns
The trivia check. Closed-form questions that test memory of facts the learner will look up on the job. "What year was this product released?" If the learner can Google it in five seconds, the question isn't measuring capability.
The gotcha question. Questions designed to be tricky rather than diagnostic. Double negatives. Misleadingly similar answer choices. "All of the above except B." These produce noise in your data and frustration in your learners.
Pass-mark inflation. Setting the pass mark low (50%, 60%) so everyone passes and the certification rate looks healthy. The certification then means nothing — to the learner, to the auditor, to the partner program, to the customer. A high pass-rate is a goal; a low pass-mark is a fraud.
Video coaching without a rubric. Asking learners to record a pitch with no grading criteria. Graders score inconsistently. The grading conversation becomes "I liked yours better." The program produces noise instead of signal.
Vague rubric language. "Was the pitch effective?" is not a rubric. "Did the rep open with the value proposition in under 60 seconds, name the specific customer pain, and ask a discovery question before product?" is. Specific behaviors; specific scoring bands.
Grader bottlenecks. Sending 500 video coaching submissions to one grader with a one-week SLA. The queue stacks. Learners wait for feedback. Trust in the program erodes. Either reduce the volume or expand the grader pool.
Video coaching for things a multi-choice question would verify. Asking a learner to record a video explaining a definition when a multi-choice question would do. Expensive on both sides for no extra signal. Reserve video coaching for capabilities that genuinely require demonstration.
File uploads nobody reviews. Collecting work product because "we should have them submit something" with no plan to actually look at it. Learners notice over time and submissions get less and less serious.
The unlimited-retake verification. Letting learners take a high-stakes assessment infinitely until they pass. Combined with feedback between attempts, this turns the assessment into a memorization exercise. The certificate at the end means "this person took the test enough times to memorize it."
One-shot reinforcement. The opposite — locking a low-stakes practice quiz to one attempt. Now the reinforcement tool produces anxiety instead of learning, and learners avoid it.
Hiding the cost from the learner. Long assessments with no time estimate, no progress indicator, no save-and-resume. Learners abandon them mid-way. Heavy formats (video coaching, file upload) without warning produce abandonment too.
Compliance theater. Quizzes added to satisfy a vague "we should have a quiz" expectation, with no clear job and no consequential pass mark. The quiz exists. Nobody believes it. The audit accepts it. Capability is unproven.
Question count as a proxy for rigor. A 50-question assessment is not more rigorous than a 15-question assessment. It's just longer. Rigor comes from item quality and format fit, not item volume.
Reporting that nobody reads. Building rich item-level analytics and never looking at them. Continu produces detail; Continu does not produce the discipline to act on detail. That has to come from a human.
In the Continu Architecture
Assessments are connected to nearly every other object in Continu.
- Content. Embedded checks live inside articles, videos, and other content types. The content delivers; the check verifies attention or reinforces.
- Tracks and Journeys. Assessments gate progression. A track can require passing assessment A before module B unlocks. A journey can branch based on assessment outcome.
- Workshops. A natural pairing — post-workshop video coaching submission verifies the workshop landed. The learner records the skill they just practiced; the grader confirms the capability is real.
- Smart Segmentation. Assessment results become user attributes. "Passed certification X" is a segment condition; "failed assessment Y twice" is a segment condition. This is how you build remedial cohorts and certified cohorts automatically.
- Automations. An assessment outcome can trigger downstream actions — a certification email, a remediation assignment, a manager notification, a directory listing update.
- Reporting. Assessment data feeds the certification reports, the question-level analytics, the cohort comparison reports, and the grader workload reports.
- Notifications. Assessment events (passed, failed, ready-to-grade, grading complete) drive notifications to learners, graders, and managers.
Designing an assessment in isolation misses most of what makes Continu valuable. The assessment is the data; the architecture is what turns that data into program movement.
External Audience Patterns
Partner pitch certification (video coaching). Partner records themselves delivering the pitch. Grader scores against the rubric — discovery questions, value-prop landing, objection handling. Required before the partner is approved for customer-facing motions. Re-recording allowed; coaching feedback delivered with the grade.
Partner knowledge certification (closed-form + question bank). Verification assessment with randomized closed-form questions. Pass marks high (often 80–90%). Question banks rotated annually so older cohorts don't leak answers. Failed re-certification expires the partner's eligibility for restricted activities.
Customer admin deployment readiness (file upload). Customer admin uploads their deployment plan, configuration artifact, or implementation roadmap. Customer success grades the file against the readiness rubric. Passing gates access to advanced features or production rollout.
Customer admin workflow verification (video coaching — screen recording). Customer admin records their screen performing a key workflow. Customer success grades against the workflow rubric. Used to verify the admin can actually run the system before they go live.
Channel quality bar. Tiered closed-form assessments tied to channel program tiers — bronze, silver, gold tiers require different assessment outcomes. Question banks rotated; pass marks reflect tier requirements.
Franchisee compliance. Multi-format. Closed-form on policy knowledge. File uploads of completed checklists. Video coaching where the franchise operator demonstrates an in-store procedure. Audit-grade rigor. Pass marks high (often 80–90% on operations, 100% on safety-critical).
Customer reinforcement (closed-form, low-stakes). Embedded reinforcement quizzes throughout customer education tracks. Generous retake policy. Feedback shown. The point is retention, not gatekeeping.
Member education certification. Closed-form verification often paired with file upload (a portfolio piece, a case study response). The credential has external value; the assessment quality reflects that.
Internal Audience Patterns
Sales pitch certification (video coaching). New rep records themselves delivering the pitch at the end of onboarding. Manager grades against rubric. Pass/coach decision is part of certification.
New hire compliance (closed-form). Verification with hard pass marks on safety, security, and regulatory content. Retake policy clear. Certificate retained for HR audit. Question banks rotated to prevent year-over-year answer-sharing between cohorts.
Onboarding milestone (file upload). New hire submits a worked example or analysis at 30, 60, or 90 days. Manager reviews against rubric. Used to track ramp-up that isn't captured by formal closed-form assessments.
Customer scenario response (video coaching). Internal team member responds to a realistic customer scenario in writing or on video. Grader scores against the response rubric. Used in support, sales, customer success training.
Manager development verification (video coaching). Manager records themselves practicing a difficult conversation — performance feedback, terminations, role changes. Senior leader or coach grades against the rubric. Used in leadership development to verify skill application.
Skill verification (mixed format). Role-specific skills — sales certification (video pitch), technical readiness (closed-form + file upload), manager fundamentals (closed-form + video). Tied to role progression or eligibility for specific work.
Annual recertification. Closed-form verification, with attention to question-bank rotation so the same assessment every year doesn't stop measuring anything.
Known Behaviors and Limits
Closed-form questions auto-grade; everything else requires human work. Multi choice, multi answer, and true/false are scored by the platform. Short answer, file upload, and video coaching submissions are routed to graders, who do the grading. Plan grader capacity for high-volume programs with open-form items.
Rubric quality determines grading consistency. A vague rubric produces grader-to-grader variation. A specific behavior-based rubric produces consistent grading across graders. Invest in the rubric; the program quality follows.
Video coaching submissions consume real storage. Large cohorts producing video coaching submissions accumulate storage at meaningful scale. Plan retention — how long submissions stay accessible, when they archive.
Grading turnaround is the trust signal. Learners who submit a video coaching recording or file and wait three weeks for feedback disengage from the program. A defined grading SLA — and visible communication of it — keeps trust intact.
Question-level analytics require setup and review. Item-level reporting (which closed-form questions everyone gets wrong, which video coaching submissions consistently fail) is available, but the program owner must actually pull and review the report. Continu surfaces the data; it does not interpret it.
Time limits are strict. Once enabled, the timer runs. Browser tabs closed mid-attempt, network drops, and learner-side issues all eat into the limit. Test your time limit settings with a real device and a real network before launching to a partner cohort.
Certificate templates are organization-wide. Customizing the certificate per program is done at the template level. Plan your certificate inventory before you have 40 programs each issuing slightly different certificates that confuse the directory.
Score thresholds are static per assessment. You cannot vary the pass mark by segment within a single assessment — different cohorts with different bars need different assessments.
Failed attempts persist in the record. A learner who fails before passing has both the failed and passed attempts in the system. This is a feature for audit, but design your reports to avoid double-counting attempts as completions.
Retake cooldowns are clock-time, not learner-controlled. A 24-hour cooldown means 24 hours from the failed attempt — not "the next time the learner logs in." Learners in time zones away from the program owner may experience this differently than expected.
Video coaching cannot be anonymous. The submission is the person on camera or their screen. Design accordingly — assessments with video coaching are always attributed.
Where to Go Next
- Tracks and Journeys: Designing Learning Paths — for how assessments gate progression in multi-step programs.
- Smart Segmentation: Designing Populations That Maintain Themselves — for how assessment outcomes drive segment membership.
- Automation Design Best Practices — for how to wire assessment events into downstream actions.
- Reporting: Which Report Should I Use? — for finding the right view of assessment data.
- Workshop Strategy: When and How to Use Live Learning — for the relationship between live training and post-workshop video coaching verification.
Design first. Click second. Use the format that actually verifies the capability.