0.0 Executive Summary
This report covers the review and improvement of prompts and rubrics used in evaluation exercises. The goal was to reduce confusion caused by unclear instructions by using structured rubrics and clearer language. This helps ensure tasks are completed and graded more consistently.
The final result was a more reliable evaluation process with clearer expectations, measurable criteria, and less room for interpretation.
1.0 Prompt & Rubric Evaluation Exercises
1.1 Project Description
The goal of this task was to improve instructions so they are clear and easy to follow.
This was done by:
- Defining clear requirements for each task
- Removing vague or unclear wording
- Creating consistent grading criteria
This ensures that users understand what is expected and can complete tasks without confusion.
1.2 Technical Task / Troubleshooting Process
The process focused on finding weaknesses in existing prompts and improving them through clearer structure and wording.
Key Actions & Observations
- Reviewed existing prompts and identified unclear instructions
- Updated evaluation guidelines and rubric templates
- Defined clear criteria for grading performance
- Reduced complex or multi-part instructions that could cause confusion
- Ensured each task had a single, clear objective
- Documented changes to make the process repeatable
Root Cause:
Instructions were sometimes vague or included multiple goals in one prompt. This made it harder to understand what was required. The issue was resolved by simplifying and clarifying each task.
1.3 Resolution and Validation
The updated prompts and rubrics were tested to confirm they worked as expected.
| Parameter | Configuration Value |
|---|---|
| Evaluation Method | Rubric-Based Scoring |
| Control State | Enforced |
| Criteria Type | Clear and Measurable |
| Scope | Technical and Analytical Tasks |
Validation Steps
-
Applied the rubric to a sample technical report
-
Verified that scoring matched the defined criteria
-
Confirmed that results were consistent across similar submissions
-
Ensured the rubric worked across different types of tasks
2.0: CONCLUSION
2.1 Key Takeaways
- Clear instructions reduce mistakes and improve performance
- Rubrics should be specific and easy to understand
- Each task should have a single clear objective
- Consistent evaluation criteria leads to fairer results
- Testing is important to confirm that rubrics work as intended
2.2 Security Implications & Recommendations
Risk: Operational Errors
Unclear instructions can lead to incorrect actions or failed tasks.
Mitigation: Use clear, step-by-step instructions with defined outcomes.
Risk: Inconsistent Grading
Without clear criteria, different people may grade the same work differently.
Mitigation: Use standardized rubrics with measurable requirements.
Best Practices
- Keep instructions simple and direct
- Avoid combining multiple goals into one task
- Use consistent templates for all evaluations
- Validate rubrics before use
- Document changes to maintain repeatability
Framework Alignment
- Supports structured evaluation and consistent decision-making
- Aligns with best practices for clear communication and process control
- Reinforces the importance of measurable criteria in assessments