PULSE: Protective Utility and Limitation Scoring Engine
This document introduces the Protective Utility and Limitation Scoring Engine (PULSE), a comprehensive framework for evaluating the effectiveness of defensive measures against adversarial attacks on AI systems, with specific focus on language models and generative AI.
Framework Overview
PULSE provides a structured approach to measuring, quantifying, and comparing the effectiveness of security controls implemented to protect AI systems. It enables evidence-based defensive planning by systematically evaluating protection effectiveness, control limitations, and defensive coverage across the attack surface.
Core Evaluation Dimensions
PULSE evaluates defensive measures across five primary dimensions:
- Protection Effectiveness (PE): How well the defense prevents or mitigates attacks
- Coverage Completeness (CC): How comprehensively the defense addresses the attack surface
- Operational Impact (OI): How the defense affects system functionality and performance
- Implementation Maturity (IM): How well-developed and robust the implementation is
- Adaptation Capacity (AC): How well the defense adapts to evolving threats
Each dimension contains multiple components that are scored individually and combined to create dimension scores and an overall PULSE rating.
Dimension Components
1. Protection Effectiveness (PE)
Components measuring how well the defense prevents or mitigates attacks:
Component | Weight | Description | Scoring Guidance |
---|---|---|---|
PE1: Attack Prevention | 30% | Ability to prevent attacks completely | 0 (No prevention) to 10 (Complete prevention) |
PE2: Attack Detection | 25% | Ability to detect attempted attacks | 0 (No detection) to 10 (Comprehensive detection) |
PE3: Impact Reduction | 20% | Ability to reduce consequences when attacks succeed | 0 (No reduction) to 10 (Maximum reduction) |
PE4: Recovery Facilitation | 15% | Support for rapid recovery after attacks | 0 (No recovery support) to 10 (Optimal recovery) |
PE5: Attack Chain Disruption | 10% | Ability to break attack sequences | 0 (No disruption) to 10 (Complete disruption) |
2. Coverage Completeness (CC)
Components measuring how comprehensively the defense addresses the attack surface:
Component | Weight | Description | Scoring Guidance |
---|---|---|---|
CC1: Attack Vector Coverage | 25% | Range of attack vectors addressed | 0 (Very limited) to 10 (Comprehensive) |
CC2: Technique Variety Coverage | 20% | Range of attack techniques addressed | 0 (Minimal variety) to 10 (All techniques) |
CC3: Model Coverage | 20% | Range of models/versions protected | 0 (Single version) to 10 (All versions/models) |
CC4: Deployment Context Coverage | 15% | Range of deployment scenarios protected | 0 (Single context) to 10 (All contexts) |
CC5: User Scenario Coverage | 20% | Range of user interactions protected | 0 (Limited scenarios) to 10 (All scenarios) |
3. Operational Impact (OI)
Components measuring how the defense affects system functionality and performance:
Component | Weight | Description | Scoring Guidance |
---|---|---|---|
OI1: Performance Impact | 25% | Effect on system performance | 0 (Severe degradation) to 10 (No impact) |
OI2: User Experience Impact | 25% | Effect on user experience | 0 (Major disruption) to 10 (Transparent) |
OI3: Operational Complexity | 20% | Administrative/operational burden | 0 (Very complex) to 10 (Simple) |
OI4: Resource Requirements | 15% | Computing resources needed | 0 (Extensive resources) to 10 (Minimal resources) |
OI5: Compatibility Impact | 15% | Effect on system compatibility | 0 (Major incompatibilities) to 10 (Fully compatible) |
4. Implementation Maturity (IM)
Components measuring how well-developed and robust the implementation is:
Component | Weight | Description | Scoring Guidance |
---|---|---|---|
IM1: Development Status | 25% | Current state of development | 0 (Conceptual) to 10 (Production-hardened) |
IM2: Testing Thoroughness | 20% | Extent of security testing | 0 (Minimal testing) to 10 (Exhaustive testing) |
IM3: Documentation Quality | 15% | Comprehensiveness of documentation | 0 (Minimal documentation) to 10 (Comprehensive) |
IM4: Deployment Readiness | 20% | Ease of operational deployment | 0 (Difficult deployment) to 10 (Turnkey solution) |
IM5: Maintenance Status | 20% | Ongoing maintenance and support | 0 (Abandoned) to 10 (Actively maintained) |
5. Adaptation Capacity (AC)
Components measuring how well the defense adapts to evolving threats:
Component | Weight | Description | Scoring Guidance |
---|---|---|---|
AC1: Threat Evolution Response | 30% | Ability to address new attack variants | 0 (Static defense) to 10 (Automatically adaptive) |
AC2: Configuration Flexibility | 20% | Adaptability to different environments | 0 (Fixed configuration) to 10 (Highly configurable) |
AC3: Update Mechanism | 20% | Effectiveness of update processes | 0 (Manual, difficult) to 10 (Automatic, seamless) |
AC4: Learning Capability | 15% | Ability to improve from experience | 0 (No learning) to 10 (Continuous improvement) |
AC5: Feedback Integration | 15% | Incorporation of operational feedback | 0 (No feedback) to 10 (Comprehensive feedback loop) |
Scoring Methodology
PULSE uses a systematic calculation approach:
# Pseudocode for PULSE calculation
def calculate_pulse(scores):
# Calculate dimension scores
pe_score = (scores['PE1'] * 0.30 + scores['PE2'] * 0.25 + scores['PE3'] * 0.20 +
scores['PE4'] * 0.15 + scores['PE5'] * 0.10)
cc_score = (scores['CC1'] * 0.25 + scores['CC2'] * 0.20 + scores['CC3'] * 0.20 +
scores['CC4'] * 0.15 + scores['CC5'] * 0.20)
oi_score = (scores['OI1'] * 0.25 + scores['OI2'] * 0.25 + scores['OI3'] * 0.20 +
scores['OI4'] * 0.15 + scores['OI5'] * 0.15)
im_score = (scores['IM1'] * 0.25 + scores['IM2'] * 0.20 + scores['IM3'] * 0.15 +
scores['IM4'] * 0.20 + scores['IM5'] * 0.20)
ac_score = (scores['AC1'] * 0.30 + scores['AC2'] * 0.20 + scores['AC3'] * 0.20 +
scores['AC4'] * 0.15 + scores['AC5'] * 0.15)
# Calculate overall PULSE score (0-100 scale)
pulse_score = ((pe_score * 0.30) + (cc_score * 0.25) + (oi_score * 0.15) +
(im_score * 0.15) + (ac_score * 0.15)) * 10
# Determine effectiveness category
if pulse_score >= 80:
effectiveness = "Superior Defense"
elif pulse_score >= 60:
effectiveness = "Strong Defense"
elif pulse_score >= 40:
effectiveness = "Adequate Defense"
elif pulse_score >= 20:
effectiveness = "Weak Defense"
else:
effectiveness = "Ineffective Defense"
return {
"dimension_scores": {
"Protection Effectiveness": pe_score * 10,
"Coverage Completeness": cc_score * 10,
"Operational Impact": oi_score * 10,
"Implementation Maturity": im_score * 10,
"Adaptation Capacity": ac_score * 10
},
"pulse_score": pulse_score,
"effectiveness": effectiveness
}
The final PULSE score is calculated by combining the dimension scores with appropriate weights:
- Protection Effectiveness: 30%
- Coverage Completeness: 25%
- Operational Impact: 15%
- Implementation Maturity: 15%
- Adaptation Capacity: 15%
Effectiveness Classification
PULSE scores map to defensive effectiveness ratings:
Score Range | Effectiveness Rating | Description | Implementation Guidance |
---|---|---|---|
80-100 | Superior Defense | Exceptional protection with minimal limitations | Primary defense suitable for critical systems |
60-79 | Strong Defense | Robust protection with limited weaknesses | Core defense with supplementary controls |
40-59 | Adequate Defense | Reasonable protection with notable limitations | Acceptable for non-critical systems with layering |
20-39 | Weak Defense | Limited protection with significant gaps | Requires substantial enhancement or replacement |
0-19 | Ineffective Defense | Minimal protection with fundamental flaws | Not suitable as a security control |
Vector String Representation
For efficient communication, PULSE provides a compact vector string format:
PULSE:1.0/PE:7.2/CC:6.5/OI:8.1/IM:5.8/AC:4.7/SCORE:6.5
Components:
PULSE:1.0
: Framework versionPE:7.2
: Protection Effectiveness score (0-10)CC:6.5
: Coverage Completeness score (0-10)OI:8.1
: Operational Impact score (0-10)IM:5.8
: Implementation Maturity score (0-10)AC:4.7
: Adaptation Capacity score (0-10)SCORE:6.5
: Overall PULSE score (0-10)
Defense Classification Taxonomy
PULSE includes a comprehensive taxonomy for categorizing defensive measures:
Primary Categories
Top-level classification of defensive approaches:
Category Code | Name | Description | Examples |
---|---|---|---|
PRV | Preventive Controls | Controls that block attack execution | Input validation, prompt filtering |
DET | Detective Controls | Controls that identify attack attempts | Monitoring systems, anomaly detection |
MIG | Mitigative Controls | Controls that reduce attack impact | Output filtering, response limiting |
REC | Recovery Controls | Controls that support system recovery | Logging systems, state restoration |
GOV | Governance Controls | Controls that manage security processes | Testing frameworks, security policies |
Subcategories
Detailed classification within each primary category:
defense_taxonomy:
PRV: # Preventive Controls
PRV-INP: "Input Validation Controls"
PRV-FLT: "Filtering Controls"
PRV-AUT: "Authentication Controls"
PRV-BND: "Boundary Controls"
PRV-SAN: "Sanitization Controls"
DET: # Detective Controls
DET-MON: "Monitoring Controls"
DET-ANM: "Anomaly Detection Controls"
DET-PAT: "Pattern Recognition Controls"
DET-BEH: "Behavioral Analysis Controls"
DET-AUD: "Audit Controls"
MIG: # Mitigative Controls
MIG-OUT: "Output Filtering Controls"
MIG-RLM: "Rate Limiting Controls"
MIG-SEG: "Segmentation Controls"
MIG-CNT: "Content Moderation Controls"
MIG-TRC: "Truncation Controls"
REC: # Recovery Controls
REC-LOG: "Logging Controls"
REC-BKP: "Backup Controls"
REC-STA: "State Management Controls"
REC-RST: "Reset Mechanisms"
REC-REV: "Reversion Controls"
GOV: # Governance Controls
GOV-TST: "Testing Controls"
GOV-POL: "Policy Controls"
GOV-TRN: "Training Controls"
GOV-INC: "Incident Response Controls"
GOV-AUD: "Audit Controls"
Application Examples
To illustrate PULSE in action, consider these example defense assessments:
Example 1: Prompt Injection Detection System
A monitoring system designed to detect prompt injection attacks:
Dimension Component | Score | Justification |
---|---|---|
PE1: Attack Prevention | 3.0 | Detection only, limited prevention |
PE2: Attack Detection | 8.0 | Strong detection capabilities for known patterns |
PE3: Impact Reduction | 5.0 | Moderate impact reduction through alerting |
PE4: Recovery Facilitation | 7.0 | Good logging support for recovery |
PE5: Attack Chain Disruption | 4.0 | Limited disruption of attack sequences |
CC1: Attack Vector Coverage | 7.0 | Covers most prompt injection vectors |
CC2: Technique Variety Coverage | 6.0 | Addresses many but not all techniques |
CC3: Model Coverage | 8.0 | Works with most model versions |
CC4: Deployment Context Coverage | 6.0 | Supports multiple but not all deployment scenarios |
CC5: User Scenario Coverage | 7.0 | Covers most user interaction patterns |
OI1: Performance Impact | 8.0 | Minimal performance overhead |
OI2: User Experience Impact | 9.0 | Almost transparent to users |
OI3: Operational Complexity | 6.0 | Moderate configuration requirements |
OI4: Resource Requirements | 7.0 | Reasonable resource utilization |
OI5: Compatibility Impact | 8.0 | Good compatibility with existing systems |
IM1: Development Status | 7.0 | Production-ready with ongoing refinement |
IM2: Testing Thoroughness | 6.0 | Well-tested against common scenarios |
IM3: Documentation Quality | 8.0 | Comprehensive documentation |
IM4: Deployment Readiness | 7.0 | Relatively straightforward deployment |
IM5: Maintenance Status | 8.0 | Active maintenance and updates |
AC1: Threat Evolution Response | 5.0 | Moderate ability to address new variants |
AC2: Configuration Flexibility | 7.0 | Good configuration options |
AC3: Update Mechanism | 6.0 | Standard update processes |
AC4: Learning Capability | 4.0 | Limited autonomous learning |
AC5: Feedback Integration | 7.0 | Good incorporation of feedback |
Calculated PULSE score: 66.3 (Strong Defense) Vector: PULSE:1.0/PE:5.3/CC:6.8/OI:7.7/IM:7.2/AC:5.6/SCORE:6.6 Classification: DET-PAT (Detective Controls - Pattern Recognition Controls)
Example 2: Input Filtering and Sanitization System
A preventive control system designed to filter and sanitize inputs:
Dimension Component | Score | Justification |
---|---|---|
PE1: Attack Prevention | 8.0 | Strong prevention capabilities for known patterns |
PE2: Attack Detection | 6.0 | Moderate detection as a byproduct of filtering |
PE3: Impact Reduction | 7.0 | Significant impact reduction even when bypassed |
PE4: Recovery Facilitation | 4.0 | Limited recovery support |
PE5: Attack Chain Disruption | 8.0 | Effectively disrupts many attack sequences |
CC1: Attack Vector Coverage | 7.0 | Covers most input-based vectors |
CC2: Technique Variety Coverage | 6.0 | Addresses many but not all techniques |
CC3: Model Coverage | 8.0 | Compatible with most models |
CC4: Deployment Context Coverage | 7.0 | Works in most deployment scenarios |
CC5: User Scenario Coverage | 6.0 | Covers many user scenarios with some gaps |
OI1: Performance Impact | 6.0 | Noticeable but acceptable performance impact |
OI2: User Experience Impact | 5.0 | Some user experience degradation |
OI3: Operational Complexity | 5.0 | Moderately complex to configure optimally |
OI4: Resource Requirements | 7.0 | Reasonable resource utilization |
OI5: Compatibility Impact | 6.0 | Some compatibility challenges |
IM1: Development Status | 8.0 | Well-developed and mature |
IM2: Testing Thoroughness | 7.0 | Extensively tested |
IM3: Documentation Quality | 7.0 | Good documentation |
IM4: Deployment Readiness | 6.0 | Requires some deployment effort |
IM5: Maintenance Status | 8.0 | Actively maintained |
AC1: Threat Evolution Response | 7.0 | Good adaptation to new patterns |
AC2: Configuration Flexibility | 8.0 | Highly configurable |
AC3: Update Mechanism | 7.0 | Effective update processes |
AC4: Learning Capability | 5.0 | Some learning capabilities |
AC5: Feedback Integration | 6.0 | Decent feedback loops |
Calculated PULSE score: 69.8 (Strong Defense) Vector: PULSE:1.0/PE:7.0/CC:6.8/OI:5.8/IM:7.3/AC:6.8/SCORE:7.0 Classification: PRV-SAN (Preventive Controls - Sanitization Controls)
Defense Strategy Portfolio Analysis
PULSE enables systematic analysis of defense strategies:
1. Defense-in-Depth Assessment
Evaluating layered defense strategies:
Layer Analysis | Methodology | Strategic Insight | Example Finding |
---|---|---|---|
Layer Coverage | Map defenses to attack lifecycle stages | Identifies coverage gaps | 85% coverage at prevention layer, only 40% at detection layer |
Layer Effectiveness | Assess effectiveness at each layer | Reveals weak points | Strong prevention (7.2/10) but weak recovery (3.5/10) |
Layer Redundancy | Identify overlapping defenses | Highlights resource optimization opportunities | Redundant coverage in input filtering, gaps in monitoring |
Layer Independence | Analyze defense interdependencies | Identifies single points of failure | 65% of defenses depend on shared pattern database |
Layer-Specific Adaptation | Evaluate adaptation by layer | Reveals adaptation disparities | Prevention layer adapts quickly (7.8/10) but recovery adaptation is slow (4.2/10) |
2. Attack Vector Defense Analysis
Analyzing defenses by attack vector:
Vector Analysis | Methodology | Strategic Insight | Example Finding |
---|---|---|---|
Vector Coverage | Map defenses to attack vectors | Identifies unprotected vectors | Strong coverage against prompt injection (85%) but weak against data extraction (35%) |
Vector-Specific Effectiveness | Evaluate effectiveness by vector | Reveals vector-specific weaknesses | High effectiveness against direct injection (8.1/10) but poor against context manipulation (3.2/10) |
Cross-Vector Protection | Analyze protection across related vectors | Identifies systemic vulnerabilities | Protection decreases by 45% across related vectors |
Vector Evolution Response | Evaluate adaptation to vector evolution | Reveals adaptation challenges | 6-month lag in addressing new context manipulation variants |
Vector-Specific Investment | Analyze resource allocation by vector | Guides resource optimization | 60% of resources focused on vectors representing only 30% of attacks |
3. Operational Impact Analysis
Analyzing the deployment implications of defenses:
Impact Analysis | Methodology | Strategic Insight | Example Finding |
---|---|---|---|
Performance Budget Analysis | Measure cumulative performance impact | Enables impact optimization | Combined controls create 12% latency increase |
Experience Impact Assessment | Evaluate user experience effects | Identifies user friction points | Authentication controls create 80% of user friction |
Operational Overhead Calculation | Measure administrative burden | Guides operational planning | 35 person-hours per week for maintenance across controls |
Resource Utilization Analysis | Analyze resource consumption patterns | Enables resource optimization | Memory usage scales non-linearly with model size |
Cross-Control Interference | Identify negative control interactions | Prevents control conflicts | Filter bypass when used with specific monitoring controls |
Defense Evaluation Methodology
PULSE defines a structured approach to evaluating defensive measures:
1. Evaluation Process
Step-by-step methodology for defense assessment:
Process Step | Description | Key Activities | Outputs |
---|---|---|---|
Scope Definition | Define evaluation boundaries | Identify controls, contexts, and objectives | Evaluation scope document |
Baseline Testing | Establish current effectiveness | Test against baseline attack set | Baseline performance metrics |
Dimensional Evaluation | Score across PULSE dimensions | Component-by-component assessment | Dimensional scores |
Vector Testing | Test against specific attack vectors | Vector-specific effectiveness testing | Vector effectiveness profile |
Operational Assessment | Evaluate real-world implications | Performance testing, compatibility testing | Operational impact analysis |
Comparative Analysis | Compare against alternatives | Side-by-side effectiveness comparison | Comparative effectiveness report |
Limitation Mapping | Identify key limitations | Edge case testing, boundary analysis | Limitation document |
2. Evidence Collection Framework
Methodology for gathering assessment evidence:
Evidence Type | Collection Approach | Evaluation Value | Quality Criteria |
---|---|---|---|
Attack Success Rate | Controlled testing with success measurement | Quantifies prevention effectiveness | Statistical significance, reproducibility |
Detection Reliability | Detection rate measurement across scenarios | Quantifies detection effectiveness | False positive/negative rates, consistency |
Performance Metrics | Standardized performance measurement | Quantifies operational impact | Consistency, environment normalization |
Coverage Mapping | Systematic attack surface mapping | Quantifies protection completeness | Comprehensiveness, systematic approach |
Adaptation Testing | Evolutionary testing with variants | Quantifies adaptation capacity | Variant diversity, evolution realism |
3. Testing Methodology
Structured approach to defense testing:
Test Type | Methodology | Evaluation Focus | Implementation Guidance |
---|---|---|---|
Known Vector Testing | Testing against documented attacks | Baseline protection capability | Use standard attack library with controlled variables |
Novel Vector Testing | Testing against new attack patterns | Adaptation capability | Develop variations of known attacks |
Edge Case Testing | Testing against boundary conditions | Protection limitations | Identify and test boundary assumptions |
Performance Testing | Measuring operational characteristics | Operational impact | Use standardized performance measurement |
Adversarial Testing | Red team attack simulation | Real-world effectiveness | Employ skilled adversarial testers |
Integration with Risk Management
PULSE is designed to integrate with broader risk management frameworks:
1. Risk-Based Defense Selection
Using PULSE to select appropriate defenses:
Risk Level | Defense Selection Criteria | PULSE Thresholds | Implementation Approach |
---|---|---|---|
Critical Risk | Maximum effectiveness regardless of impact | PE > 8.0, CC > 7.0 | Layered implementation with redundancy |
High Risk | Strong protection with acceptable impact | PE > 7.0, OI > 6.0 | Primary with supplementary controls |
Medium Risk | Balanced protection and operational impact | PE > 6.0, OI > 7.0 | Optimized for operational efficiency |
Low Risk | Minimal impact with reasonable protection | OI > 8.0, PE > 5.0 | Lightweight implementation |
Acceptable Risk | Monitoring with minimal protection | PE > 3.0 (detection focus) | Monitoring-focused approach |
2. Defense Portfolio Optimization
Using PULSE to optimize defense investments:
Optimization Approach | Methodology | Strategic Value | Implementation Guidance |
---|---|---|---|
Effectiveness Maximization | Prioritize highest PE scores | Maximum risk reduction | Focus on highest-scoring PE controls |
Efficiency Optimization | Balance PE and OI scores | Optimal risk/impact ratio | Prioritize controls with high PE:OI ratio |
Coverage Completeness | Prioritize comprehensive CC | Eliminate protection gaps | Map controls to attack surface and eliminate gaps |
Adaptation Enhancement | Focus on high AC scores | Future-proof protection | Prioritize controls with highest AC scores |
Implementation Maturity | Emphasize high IM scores | Operational reliability | Select controls with production-ready IM scores |
3. Continuous Improvement Framework
Using PULSE for ongoing defense enhancement:
Improvement Focus | Methodology | Strategic Value | Implementation Guidance |
---|---|---|---|
Weakness Remediation | Target lowest dimension scores | Eliminate critical weaknesses | Identify and address lowest-scoring dimensions |
Balanced Enhancement | Incremental improvement across dimensions | Holistic security improvement | Establish minimum thresholds for all dimensions |
Evolutionary Adaptation | Focus on adaptation capacity | Future-proof security | Prioritize improvements to AC dimension |
Operational Optimization | Target operational impact improvements | User/performance optimization | Focus on improving OI dimension |
Vector-Specific Enhancement | Address specific attack vector weaknesses | Targeted risk reduction | Map controls to attack vectors and enhance weak areas |
Practical Applications
PULSE enables several practical security applications:
1. Defense Selection and Prioritization
Using PULSE to guide defense decisions:
Decision Scenario | Application Approach | Decision Support | Example |
---|---|---|---|
New Defense Selection | Compare PULSE scores across options | Objective comparison basis | Selected Filter A (PULSE:68) over Filter B (PULSE:52) |
Defense Upgrade Decisions | Compare new versions against current | Upgrade value assessment | Upgraded monitoring system for 15-point PULSE improvement |
Defense Retirement | Evaluate continued value of existing defenses | Lifecycle management | Retired redundant control with 35 PULSE score |
Defense Prioritization | Rank defenses by PULSE score | Resource allocation | Prioritized top three controls by PULSE ranking |
Defense Gap Analysis | Identify coverage gaps through PULSE dimensions | Strategic planning | Identified 40% coverage gap in context manipulation protection |
2. Security Architecture Design
Using PULSE to guide security architecture:
Architecture Element | Application Approach | Architecture Value | Implementation Example |
---|---|---|---|
Defense Layering | Design based on dimensional scores | Optimized protection depth | Implemented three layers with complementary dimension strengths |
Control Selection | Select controls based on PULSE profiles | Optimized control selection | Created matrix of controls mapped to dimensional requirements |
Architecture Validation | Validate design through PULSE scoring | Design verification | Verified minimum PULSE threshold across architectural elements |
Trade-off Analysis | Evaluate design trade-offs through dimension scores | Balanced design decisions | Accepted 5% OI reduction for 15% PE improvement |
Component Integration | Plan integration based on control profiles | Optimized component interaction | Designed integration based on complementary PULSE profiles |
3. Vendor Assessment
Using PULSE to evaluate security vendors:
Assessment Element | Application Approach | Assessment Value | Implementation Example |
---|---|---|---|
Product Comparison | Compare vendor offerings through PULSE | Objective comparison basis | Selected Vendor A based on superior PULSE profile |
Capability Verification | Verify vendor claims through PULSE scoring | Claims validation | Verified 85% of vendor capability claims through PULSE assessment |
Gap Identification | Identify vendor solution gaps | Due diligence enhancement | Identified 30% coverage gap in vendor solution |
Integration Assessment | Evaluate integration implications | Implementation planning | Predicted integration challenges based on OI dimension analysis |
Vendor Improvement Tracking | Track vendor progress over time | Relationship management | Tracked 25% PULSE improvement over three product versions |
For detailed implementation guidance, scoring templates, and practical assessment tools, refer to the associated documentation in this framework section.