Globebyte Documentation
  • AI Agents for Learning
  • Assess for Learning
    • Creating the Assess Connected App
    • Setting up Assess for Learning
    • Viewing Assessments
    • Assessment Outcomes & Validation
    • Marking
    • Best Practices
  • Tutor for Learning
    • Setting up Tutor
    • Agent Actions
      • Tutor_Mark
      • Tutor_Task
      • Tutor_Criterion
      • Tutor_SenseMaking
      • Tutor_Instruction
    • Topics
      • Tutor_Assessment
      • Tutor_Knowledge
  • Data for Learning
  • Actions for Learning
    • Creating the xAPI Actions Connected App
    • Setting up xAPI Actions
    • Creating your first xAPI Action Flow
    • xAPI Statement Data explorer
      • Metadata
      • xapiActor
      • xapiVerb
        • Verb reference
      • xapiObject
      • authority
      • xapiResult
      • xapiContext
    • Filtering xAPI Statements
    • Viewing xAPI Statements
    • Viewing xAPI Usage
    • Setting a default statement language
    • Error messages and troubleshooting
  • Experience for Learning
    • Setting up xAPI for Salesforce
    • Send xAPI from a Flow
    • Form Action fields
    • Send xAPI from Apex
    • xApiStatement Class reference
      • Actor
      • Verb
      • Object
      • Context
      • Result
      • Authority
      • Version
      • Send methods
    • Logging and defaults
  • Learning Journey Model
    • Introduction
    • Curriculums & Pathways
    • Courses & Modules
    • Pedagogies & Objectives
    • Rubrics & Criteria
    • Learning Resources
    • Assessments & Tasks
    • Learning Groups
    • Step-by-step working example
    • Activity Tracking (Advanced)
    • Additional Pedagogies Reference
    • Best Practices
    • Assess for Learning Integration
    • Data for Learning Integration
    • Object References
      • Learning Curriculum
      • Learning Pathway
      • Learning Course
      • Learning Module
      • Learning Pedagogy
      • Learning Objective
      • Learning Objective Assignment
      • Learning Rubric
      • Learning Rubric Criterion
      • Learning Rubric Model Solution
      • Learning Resource Type
      • Learning Resource
      • Learning Assessment
      • Learning Text Task
      • Learner Text Attempt
      • Learner Text Criterion Score
      • Learning Choice Task
      • Learner Choice Attempt
      • Learner Mark
      • Learning Group
      • Learner Group Membership
      • Learner Activity
      • Learner Activity Instance
      • Learner XAPIStatement
      • Developer Cheat Sheet: Key LDM Objects
  • Globebyte.com
Powered by GitBook
On this page
  • Learner Text Criterion Score
  • Introduction
  • Properties
  • Example
  • Use Case
  1. Learning Journey Model
  2. Object References

Learner Text Criterion Score

Learner Text Criterion Score

Introduction

A Learner Text Criterion Score records how a single Rubric Criterion is applied to a single learner’s Text Attempt. This allows highly granular scoring—for example, awarding separate points for “Clarity,” “Evidence,” or “Grammar” in an essay.

Key highlights:

  • Fine-Grained Evaluation: Each rubric criterion is scored independently, improving consistency and transparency.

  • Status Tracking: Just like the overall attempt, each criterion goes through statuses (Pending, Scoring, etc.).

  • Detailed Feedback: Provide rationale or instructions per criterion, letting the learner see precisely which part needs work.

Properties

Property Name
Description

LearnerTextAttemptId

Identifies which essay submission (Text Attempt) is being evaluated. This helps group all criterion scores under one master attempt.

LearningRubricCriterionId

Points to the specific criterion from the rubric, e.g., “Organization of Argument” or “Research Depth.”

Status

Tracks the cycle (Pending, Submitted, Scoring, Scored, Expired). Usually, this mirrors the attempt’s broader status but focuses on the specific criterion.

PendingDate

When evaluation of this criterion was created or flagged for scoring.

SubmittedDate

When it officially entered the scoring queue.

ScoringDate

When scoring kicked off—manually or by AI—for this particular criterion.

ScoredDate

When grading for this criterion wrapped up, storing the final sub-score.

ExpiredDate

If the attempt timed out or exceeded submission windows, the system can mark the criterion scoring as expired.

Score

The numeric point value awarded for this specific criterion, based on the rubric. E.g., 3 out of 5 for “Organization.”

ScoreReason

A textual explanation detailing why the scorer (or AI) assigned this score.

ScoreInstruction

General feedback relevant to this criterion. For instance, “Improve transitions between paragraphs.”

ScorePersonalInstruction

Personalized advice for the learner on how to do better in this specific area.

ScoreInstructionPlus1

Additional or follow-up pointers for advanced improvement (e.g., referencing sources, using better structure).

ScorePersonalInstructionPlus1

Extra personalized coaching or motivation tailored to that learner, e.g., “Use real-world data from your current workplace.”

CreatedDate

System-captured date/time of when this criterion score record was created.

ModifiedDate

Automatically updated if the score or feedback is changed.

Example

For a Text Attempt on “Sustainability Strategies,” the system checks the “Evidence and Support” criterion. The AI or instructor assigns a Score of 4 out of 5, with ScoreReason stating, “Good references but minor inconsistencies in citation style.” The ScorePersonalInstruction might advise more formal referencing.

Use Case

A college-level writing assignment uses a “Research Paper Rubric” with five criteria. Each is separately scored. The instructor sees that a particular student excelled in “Analysis of Data” but lost points in “Clarity of Thesis.” The system tracks these criterion-level scores, enabling targeted feedback that fosters continuous improvement.

PreviousLearner Text AttemptNextLearning Choice Task

Last updated 5 months ago