Globebyte Documentation
  • AI Agents for Learning
  • Assess for Learning
    • Creating the Assess Connected App
    • Setting up Assess for Learning
    • Viewing Assessments
    • Assessment Outcomes & Validation
    • Marking
    • Best Practices
  • Tutor for Learning
    • Setting up Tutor
    • Agent Actions
      • Tutor_Mark
      • Tutor_Task
      • Tutor_Criterion
      • Tutor_SenseMaking
      • Tutor_Instruction
    • Topics
      • Tutor_Assessment
      • Tutor_Knowledge
  • Data for Learning
  • Actions for Learning
    • Creating the xAPI Actions Connected App
    • Setting up xAPI Actions
    • Creating your first xAPI Action Flow
    • xAPI Statement Data explorer
      • Metadata
      • xapiActor
      • xapiVerb
        • Verb reference
      • xapiObject
      • authority
      • xapiResult
      • xapiContext
    • Filtering xAPI Statements
    • Viewing xAPI Statements
    • Viewing xAPI Usage
    • Setting a default statement language
    • Error messages and troubleshooting
  • Experience for Learning
    • Setting up xAPI for Salesforce
    • Send xAPI from a Flow
    • Form Action fields
    • Send xAPI from Apex
    • xApiStatement Class reference
      • Actor
      • Verb
      • Object
      • Context
      • Result
      • Authority
      • Version
      • Send methods
    • Logging and defaults
  • Learning Journey Model
    • Introduction
    • Curriculums & Pathways
    • Courses & Modules
    • Pedagogies & Objectives
    • Rubrics & Criteria
    • Learning Resources
    • Assessments & Tasks
    • Learning Groups
    • Step-by-step working example
    • Activity Tracking (Advanced)
    • Additional Pedagogies Reference
    • Best Practices
    • Assess for Learning Integration
    • Data for Learning Integration
    • Object References
      • Learning Curriculum
      • Learning Pathway
      • Learning Course
      • Learning Module
      • Learning Pedagogy
      • Learning Objective
      • Learning Objective Assignment
      • Learning Rubric
      • Learning Rubric Criterion
      • Learning Rubric Model Solution
      • Learning Resource Type
      • Learning Resource
      • Learning Assessment
      • Learning Text Task
      • Learner Text Attempt
      • Learner Text Criterion Score
      • Learning Choice Task
      • Learner Choice Attempt
      • Learner Mark
      • Learning Group
      • Learner Group Membership
      • Learner Activity
      • Learner Activity Instance
      • Learner XAPIStatement
      • Developer Cheat Sheet: Key LDM Objects
  • Globebyte.com
Powered by GitBook
On this page
  • Learning Rubric Model Solution
  • Introduction
  • Properties
  • Example
  • Use Case
  1. Learning Journey Model
  2. Object References

Learning Rubric Model Solution

Learning Rubric Model Solution

Introduction

A Learning Rubric Model Solution offers an exemplary or ideal response within a rubric’s framework. This can be extremely helpful for subjective tasks like essays or open-ended problems, showing instructors and learners what top-tier work looks like and how it might be scored.

Key highlights:

  • Concrete Example: Gives a reference for educators to compare the actual submission with a model answer.

  • Confidence for Learners: Sometimes shared post-submission to help them see what “excellent” looks like.

  • Optional Score: You can assign a baseline numeric value to the model solution to indicate what near-perfect performance might earn.

Properties

Property Name
Description

LearningRubricId

Connects this model solution to the relevant rubric. If a rubric is used for essay scoring, the model solution might demonstrate how a top-level essay is structured, referencing the rubric’s criteria.

Name

A short identifier for the model solution, such as “Model Essay on Market Research.” Helps instructors pick out the correct example in a list.

Solution

A long textual explanation of the ideal answer. This can include full paragraphs, bullet points, references, or format guidelines. If you’re using external attachments, you might note them here as well.

Score

(Optional) A numeric mark associated with this solution—often the highest score possible, signaling that this is an “ideal response.”

CreatedDate

Indicates when this model solution record was created.

ModifiedDate

Reflects the last time any edits were made to this model solution.

Example

In a rubric for “Advanced Data Science Project,” the criterion “Discussion of Model Limitations” might have a model solution that succinctly explains potential biases and how they can be mitigated. You store it here so any instructor can quickly see what a perfect discussion might entail, awarding full points.

Use Case

A writing course maintains a repository of model solutions for typical essay prompts. Each is tied to a corresponding rubric. When instructors begin grading, they can reference these model solutions to stay consistent. Learners, after submission, might receive the model solution (at the instructor’s discretion) to enhance their understanding of best practices.

PreviousLearning Rubric CriterionNextLearning Resource Type

Last updated 5 months ago