2.2: Process Assessment Model (PAM)
Learning Objectives
After reading this section, you will be able to:
- Explain the structure and purpose of the PAM
- Describe base practices and their role in assessment
- Understand work product characteristics
- Apply PAM concepts to prepare for assessment
- Conduct assessments using the PAM's two-dimensional framework
- Interpret assessment results and plan improvements with AI support
What is the PAM?
The Process Assessment Model (PAM) provides guidance for how to assess whether processes achieve their purposes. While the PRM defines what to achieve, the PAM provides:
- Base Practices (BP): Activities that implement the process
- Work Products (WP): Tangible outputs with defined characteristics
- Assessment Indicators: Evidence of process implementation
Key distinction: The PRM asks "What should we achieve?" The PAM asks "How do we know we achieved it?" Think of the PRM as the destination and the PAM as the map that shows you how to verify you arrived.
PAM Structure
The diagram below shows the PAM's two-dimensional framework: the process dimension (which processes to assess) intersecting with the capability dimension (how mature each process is).
The PAM is built on a two-dimensional framework. Understanding both dimensions is essential for anyone preparing for or conducting an ASPICE assessment.
The Two Dimensions
| Dimension | What It Covers | Purpose |
|---|---|---|
| Process Dimension | Which processes to assess, organized into categories and groups | Defines the scope of assessment |
| Capability Dimension | How well each process is performed, from Level 0 to Level 3 | Defines the depth of assessment |
Together, these two dimensions create a matrix: every assessed process receives a capability level rating, and the PAM provides the indicators needed to make that rating objective and repeatable.
How the Dimensions Interact
When you assess a process like SWE.1, you first look at the Process Dimension to identify the base practices and work products specific to SWE.1. Then you look at the Capability Dimension to determine how well SWE.1 is managed, planned, and standardized. The process dimension tells you what to look for; the capability dimension tells you how mature it is.
Process Dimension
The Process Dimension organizes all assessable processes into a structured hierarchy. This hierarchy mirrors the PRM but adds the assessment-specific detail (base practices and work products) that assessors need.
Process Categories
| Category | Description | Example Groups |
|---|---|---|
| Primary Life Cycle Processes | Core engineering and acquisition activities | SYS, SWE, HWE, MLE |
| Supporting Life Cycle Processes | Processes that support primary processes | SUP |
| Organizational Life Cycle Processes | Management and organizational processes | MAN |
| Security Engineering Processes | Cybersecurity-specific processes (new in 4.0) | SEC |
Process Groups within Categories
| Group | ID Range | Processes | Focus |
|---|---|---|---|
| System Engineering | SYS.1 - SYS.5 | Requirements, Architecture, Integration, Qualification Testing, Validation | End-to-end system lifecycle |
| Software Engineering | SWE.1 - SWE.6 | Requirements through Qualification Testing | Software V-Model |
| Machine Learning Engineering | MLE.1 - MLE.5 | ML Requirements through Deployment | ML-specific lifecycle |
| Hardware Engineering | HWE.1 | Hardware Requirements | Hardware aspects |
| Support | SUP.1 - SUP.11 | QA, CM, Problem Resolution, Change Management, ML Data Management | Cross-cutting support |
| Management | MAN.3, MAN.5, MAN.6 | Project Management, Risk Management, Measurement | Project governance |
| Security | SEC.1 - SEC.3 | Security Requirements, Implementation, Verification | Cybersecurity engineering |
Individual Processes
Each process within a group contains its own set of base practices and output work products. For example, SWE.1 has 6 base practices (BP1-BP6) and produces work products such as the Software Requirements Specification (17-08). Every base practice maps to one or more process outcomes, creating a traceable chain from activity to evidence.
For assessors: You do not need to assess every process. The assessment scope is agreed upon before the assessment begins. Common scopes include the HIS scope (a subset of processes frequently required by automotive OEMs) or a custom scope tailored to the organization's goals.
Capability Dimension
The Capability Dimension defines how well a process is performed, managed, and institutionalized. In the PAM, this dimension uses Capability Levels 0 through 3, each characterized by Process Attributes (PA) that must be rated.
Important: While the ASPICE framework defines Levels 0-5, the PAM in ASPICE 4.0 provides detailed assessment indicators only for Levels 0-3. Levels 4 and 5 exist in the framework but are rarely assessed in practice and require organization-specific measurement models.
Capability Levels 0-3 in the PAM
| Level | Name | Process Attributes | What It Means |
|---|---|---|---|
| 0 | Incomplete | None | The process is not implemented or fails to achieve its purpose |
| 1 | Performed | PA 1.1 (Process Performance) | The process achieves its defined outcomes |
| 2 | Managed | PA 2.1 (Performance Management), PA 2.2 (Work Product Management) | The process is planned, monitored, and its work products are controlled |
| 3 | Established | PA 3.1 (Process Definition), PA 3.2 (Process Deployment) | A standard organizational process exists and is tailored for each project |
Process Attributes in Detail
| Process Attribute | Level | What It Measures |
|---|---|---|
| PA 1.1 Process Performance | 1 | Are the base practices performed? Are outcomes achieved? |
| PA 2.1 Performance Management | 2 | Are activities planned, monitored, and adjusted? Are responsibilities defined? |
| PA 2.2 Work Product Management | 2 | Are work products identified, documented, controlled, and reviewed? |
| PA 3.1 Process Definition | 3 | Is there an organizational standard process with tailoring guidelines? |
| PA 3.2 Process Deployment | 3 | Is the standard process deployed with adequate resources and competencies? |
How Capability Levels Build
Each level is cumulative. To achieve Level 2, a process must first satisfy Level 1 (outcomes achieved) and then demonstrate that it is also managed and planned. To achieve Level 3, it must satisfy Level 2 and additionally demonstrate that an organizational standard process exists and is deployed.
Practical tip: Most automotive OEMs require Capability Level 2 for their suppliers. Achieving Level 3 is a competitive differentiator and signals organizational maturity.
Rating Scale
Process attributes are rated using a four-point ordinal scale defined by ISO/IEC 33020. This scale determines whether each process attribute is achieved.
The N-P-L-F Scale
| Rating | Name | Achievement Range | Interpretation |
|---|---|---|---|
| N | Not achieved | 0% to 15% | There is little or no evidence of achievement. The attribute is essentially absent. |
| P | Partially achieved | >15% to 50% | There is some evidence of approach and some achievement. Systematic gaps exist. |
| L | Largely achieved | >50% to 85% | There is evidence of a systematic approach and significant achievement. Some weaknesses exist. |
| F | Fully achieved | >85% to 100% | There is evidence of a complete and systematic approach and full achievement. No significant weaknesses. |
Note: The percentage ranges use exclusive lower bounds (except for N). For example, exactly 50% would be rated P, while 50.1% would be rated L.
AI Assistance for Each Rating Level
| Rating | Common Situation | How AI Can Help |
|---|---|---|
| N | Missing documentation, no evidence of process execution | AI can generate initial process templates, draft baseline work products, and identify the minimum evidence set needed to move toward P |
| P | Sporadic evidence, inconsistent execution across projects | AI can scan repositories for existing artifacts that constitute evidence, highlight gaps in coverage, and suggest prioritized remediation steps |
| L | Systematic approach with identifiable weaknesses | AI can perform gap analysis against PAM indicators, identify specific weak areas, and draft improvement action items with estimated effort |
| F | Complete, systematic approach with minor refinements needed | AI can continuously monitor for regression, validate completeness of evidence, and flag early warning signs of degradation |
Level Achievement Rules
To achieve a capability level, the process attributes at that level must be rated L (Largely) or F (Fully), and all process attributes at lower levels must be rated F (Fully).
| Target Level | Requirements |
|---|---|
| Level 1 | PA 1.1 >= L |
| Level 2 | PA 1.1 = F, PA 2.1 >= L, PA 2.2 >= L |
| Level 3 | PA 1.1 = F, PA 2.1 = F, PA 2.2 = F, PA 3.1 >= L, PA 3.2 >= L |
Assessment Indicators
Assessment indicators are the specific pieces of evidence that assessors look for when rating process attributes. The PAM defines two types of indicators: Generic Practices (GP) and Generic Resources (GR).
Generic Practices (GP)
Generic Practices describe activities that apply to any process to achieve a process attribute. They are "generic" because they are not tied to a specific process like SWE.1 or SYS.2 — they work universally.
| Level | GP Category | Examples |
|---|---|---|
| 2 | GP 2.1.x (Performance Management) | Plan the process, monitor execution, define responsibilities, manage interfaces |
| 2 | GP 2.2.x (Work Product Management) | Define WP requirements, control storage, review and adjust WPs |
| 3 | GP 3.1.x (Process Definition) | Maintain standard process, determine competencies, define monitoring methods |
| 3 | GP 3.2.x (Process Deployment) | Deploy tailored process, ensure competencies, monitor defined process |
Cross-reference: For the complete listing of all generic practices with evidence examples, see 2.4 Generic Practices.
Generic Resources (GR)
Generic Resources are the infrastructure, tools, and supporting elements that enable process attribute achievement. Unlike GPs (which are activities), GRs are the resources that make those activities possible.
| GR Category | Description | Examples |
|---|---|---|
| Human resources | People with defined roles and competencies | Requirements engineer, test engineer, configuration manager |
| Tools | Software and hardware used to support processes | Requirements management tool, CI/CD pipeline, version control system |
| Infrastructure | Physical and organizational support | Development environments, review facilities, communication channels |
| Methods | Documented techniques and procedures | Review checklists, coding standards, test strategies |
How GP and GR Work Together
When an assessor evaluates PA 2.1 (Performance Management) for a given process, they look for evidence that:
- GPs are performed: Activities like planning, monitoring, and adjusting are happening (GP evidence)
- GRs are in place: The tools, people, and methods needed for those activities exist (GR evidence)
Both are needed. A plan without the tools to execute it, or tools without a plan to guide their use, will result in a lower rating.
Assessment Method
How Assessments Are Conducted
An ASPICE assessment follows a structured method defined by ISO/IEC 33002. The assessment proceeds through distinct phases:
| Phase | Activities | Key Outputs |
|---|---|---|
| 1. Planning | Define scope, schedule, team; agree on assessment constraints | Assessment plan |
| 2. Briefing | Introduce the assessment team to the project; gather initial documentation | Briefing records, initial evidence list |
| 3. Data Collection | Interview participants, review documents, examine tools and artifacts | Interview notes, evidence catalog |
| 4. Data Validation | Cross-check evidence, resolve inconsistencies, confirm findings | Validated evidence set |
| 5. Rating | Rate each process attribute using the N-P-L-F scale | Process attribute ratings |
| 6. Reporting | Document findings, strengths, weaknesses, and improvement opportunities | Assessment report |
Assessor Roles
| Role | Responsibility | Qualifications |
|---|---|---|
| Lead Assessor | Plans and directs the assessment; ensures method compliance; signs off on ratings | Certified (e.g., intacs Competent or Principal Assessor); extensive assessment experience |
| Co-Assessor | Supports evidence collection and analysis; participates in rating discussions | Assessment training; domain knowledge in assessed processes |
| Assessment Sponsor | Commissions the assessment; defines objectives and scope; receives the report | Organizational authority; understanding of ASPICE framework |
| Assessment Coordinator | Organizes logistics; schedules interviews; ensures evidence availability | Project knowledge; organizational access |
Assessment Types
| Type | Purpose | Typical Duration | Assessor Requirements |
|---|---|---|---|
| Full Assessment | Official capability determination; OEM qualification | 3-5 days on-site | Certified Lead Assessor required |
| Preliminary Assessment | Identify gaps before full assessment; internal readiness check | 1-3 days | Competent assessor recommended |
| Self-Assessment | Internal improvement baseline; team awareness building | 1-2 days | Trained assessor; may not be formally recognized |
Important: Only assessments conducted by a certified Lead Assessor following the recognized assessment method produce officially recognized capability level ratings.
AI in Assessment
AI is increasingly valuable in supporting ASPICE assessments, particularly in the labor-intensive phases of evidence collection, gap analysis, and reporting. However, the assessor's professional judgment remains the authoritative source for ratings.
AI-Assisted Evidence Collection
| Activity | Traditional Approach | AI-Assisted Approach | Benefit |
|---|---|---|---|
| Document discovery | Manual search through file systems and tools | AI crawls repositories, wikis, and tools to catalog potential evidence | Faster, more comprehensive discovery |
| Evidence mapping | Assessor manually maps documents to BPs and GPs | AI suggests mappings based on content analysis and naming conventions | Reduced preparation time |
| Completeness check | Assessor uses checklists to verify all BPs are covered | AI cross-references collected evidence against PAM indicator list and flags gaps | Fewer missed items |
| Traceability validation | Manual trace link review | AI traverses trace links and reports coverage, orphans, and inconsistencies | Higher accuracy |
AI-Assisted Analysis
| Analysis Task | How AI Helps |
|---|---|
| Work product characteristic compliance | AI checks whether documents contain required sections, identifiers, version info, and approval records |
| Base practice coverage | AI analyzes project artifacts to determine which base practices have supporting evidence |
| Trend identification | AI compares current evidence against historical assessment data to identify improvement or regression trends |
| Rating support | AI provides a preliminary rating suggestion with justification, which the assessor reviews and confirms or adjusts |
Boundaries of AI in Assessment
| AI Can | AI Cannot |
|---|---|
| Catalog and organize evidence | Replace assessor professional judgment |
| Flag gaps and inconsistencies | Assign official ratings |
| Generate preliminary analysis reports | Conduct interviews |
| Suggest improvement priorities | Understand organizational context and politics |
| Cross-reference across large document sets | Interpret ambiguous or contradictory evidence |
Principle: AI assists the assessment process; it does not replace the assessor. All ratings must be confirmed by a qualified assessor who applies professional judgment to the evidence.
Base Practices (BP)
Base practices are the fundamental activities that implement a process.
BP Characteristics
- Specific: Describe concrete activities
- Assessable: Evidence can be collected
- Comprehensive: Together, achieve all outcomes
- Process-specific: Each process has its own BPs
SWE.1 Base Practices Example (ASPICE 4.0)
| BP | Activity | Outcome Mapping |
|---|---|---|
| BP1 | Specify software requirements | O1 |
| BP2 | Structure software requirements | O2 |
| BP3 | Analyze software requirements | O3 |
| BP4 | Analyze the impact on the operating environment | O4 |
| BP5 | Ensure consistency and establish bidirectional traceability | O5, O6 |
| BP6 | Communicate agreed software requirements | O7 |
Note: ASPICE 4.0 consolidates traceability and consistency into BP5, which achieves both O5 (traceability to system requirements) and O6 (traceability to system architecture). SWE.1 has 7 outcomes in ASPICE 4.0.
BP-Outcome Mapping (ASPICE 4.0)
Each outcome is achieved through one or more base practices:
Work Products (WP)
Work products are the tangible outputs of processes.
Work Product Categories
ASPICE defines work product categories by number:
| Category | Type | Examples |
|---|---|---|
| 01-xx | Documentation | Plans, strategies |
| 04-xx | Specifications | Architecture, design |
| 08-xx | Test | Specifications, results |
| 11-xx | Code | Source, object |
| 13-xx | Analysis | Reports, reviews |
| 15-xx | Records | Minutes, decisions |
| 16-xx | Configuration | Baselines |
| 17-xx | Traceability | Records |
Work Product Characteristics
Each work product has defined characteristics:
| Characteristic | Description |
|---|---|
| Identification | Unique identifier, version |
| Content | Required content elements |
| Traceability | Links to other work products |
| Verification | Evidence of review/approval |
| Structure | Organization of content |
Example: Software Requirements Specification (17-08)
| Characteristic | Content |
|---|---|
| Identification | Document ID, version, date, status |
| Content | Functional requirements, non-functional requirements, interface requirements |
| Traceability | Links to system requirements, links to architecture |
| Verification | Review record, approval signature |
| Structure | Organized by function, by subsystem, or by priority |
Common Assessment Scenarios
Understanding how assessments play out in practice helps teams prepare effectively. Below are typical scenarios, the challenges they present, and how AI support can improve outcomes.
Scenario 1: First-Time Assessment (Supplier Qualification)
| Aspect | Details |
|---|---|
| Context | An automotive supplier has never been formally assessed and an OEM customer requires ASPICE Level 2 |
| Challenge | The team performs good engineering work but has limited formal documentation and process discipline |
| AI Support | AI scans existing repositories and tools to discover undocumented evidence; generates gap reports showing which BPs lack evidence; drafts templates for missing work products |
| Typical Outcome | Preliminary assessment reveals Level 1 for most processes with clear path to Level 2 |
Scenario 2: Re-Assessment After Improvement
| Aspect | Details |
|---|---|
| Context | A team achieved Level 1 in a previous assessment and has been working on improvements for 6 months |
| Challenge | Demonstrating that improvements are sustained and systematic, not just one-time fixes |
| AI Support | AI compares current evidence against previous assessment findings; tracks improvement actions to completion; highlights sustained practices versus temporary fixes |
| Typical Outcome | Level 2 achieved for processes where improvements were sustained across multiple project milestones |
Scenario 3: Multi-Site Assessment
| Aspect | Details |
|---|---|
| Context | An organization with development centers in multiple countries needs a consistent assessment across all sites |
| Challenge | Different sites may use different tools, naming conventions, and languages for their artifacts |
| AI Support | AI normalizes evidence across sites by mapping varied artifact names to standard PAM indicators; provides a unified evidence dashboard; translates and summarizes key documents |
| Typical Outcome | Consistent ratings across sites with site-specific improvement recommendations |
Scenario 4: Agile Development Assessment
| Aspect | Details |
|---|---|
| Context | A team uses Scrum and wants to demonstrate ASPICE compliance without reverting to a waterfall approach |
| Challenge | Evidence is distributed across sprint artifacts (backlogs, user stories, retrospectives) rather than traditional documents |
| AI Support | AI maps Jira tickets, Confluence pages, and Git history to ASPICE base practices; aggregates sprint-level evidence into process-level summaries; ensures traceability across iterative deliveries |
| Typical Outcome | ASPICE compliance demonstrated through agile artifacts with AI-generated mapping documentation |
Preparing for Assessment
Preparation is often the difference between a successful assessment and a disappointing one. The following checklist, enhanced with AI support opportunities, covers the essential preparation activities.
AI-Powered Assessment Preparation Checklist
| Step | Activity | AI Support | Done? |
|---|---|---|---|
| 1 | Define scope: Agree on which processes and capability levels will be assessed | AI can suggest scope based on OEM requirements and project characteristics | |
| 2 | Identify evidence owners: Assign a responsible person for each process in scope | AI can analyze organizational charts and project roles to suggest owners | |
| 3 | Collect existing evidence: Gather work products, plans, reports, and records for each process | AI crawls file systems, tools, and repositories to build an initial evidence catalog | |
| 4 | Map evidence to indicators: For each BP and GP, confirm that supporting evidence exists | AI performs automated mapping and flags gaps with severity ratings | |
| 5 | Perform gap analysis: Identify where evidence is missing, weak, or inconsistent | AI generates a gap report with remediation suggestions and effort estimates | |
| 6 | Remediate gaps: Create missing work products, update incomplete ones, establish missing links | AI drafts initial content for missing work products based on project data and templates | |
| 7 | Conduct dry run: Perform an internal pre-assessment using the same method as the real assessment | AI simulates assessor questions based on PAM indicators and checks answer completeness | |
| 8 | Prepare participants: Brief all interviewees on the assessment process, their role, and what to expect | AI generates role-specific briefing documents highlighting the BPs and GPs relevant to each participant | |
| 9 | Organize logistics: Schedule interview slots, book rooms, prepare evidence access for assessors | AI can generate the interview schedule based on scope and participant availability | |
| 10 | Final review: Verify all evidence is accessible, current, and properly organized | AI performs a final completeness and accessibility check across all evidence sources |
Best practice: Start preparation at least 8-12 weeks before a formal assessment. AI tools can compress the evidence collection and gap analysis phases significantly, but remediation of genuine process gaps requires real project time and cannot be shortcut.
Using the PAM for Preparation
- Identify required work products for each process
- Map base practices to project activities
- Collect evidence demonstrating BP implementation
- Review work product characteristics against actual deliverables
Evidence Collection Matrix
| Process | BP | Evidence Type | Location |
|---|---|---|---|
| SWE.1 | BP1 | SRS document | /docs/requirements/ |
| SWE.1 | BP2 | Requirements structure | Requirements tool |
| SWE.1 | BP3 | Review records | /reviews/SRS/ |
| SWE.1 | BP4 | Impact analysis | /docs/requirements/ |
| SWE.1 | BP5 | Traceability links | Requirements tool |
| SWE.1 | BP6 | Meeting minutes | /meetings/req_review/ |
Assessment Output
An ASPICE assessment produces a structured set of outputs that serve as the basis for improvement planning and, when applicable, supplier qualification decisions.
Assessment Report Contents
| Section | Description | AI Contribution |
|---|---|---|
| Assessment context | Scope, constraints, team composition, assessment dates | AI can auto-generate from assessment plan data |
| Process profiles | For each assessed process: capability level achieved and PA ratings | AI provides preliminary rating calculations from evidence analysis |
| Strengths | Areas where the organization excels; practices worth preserving | AI identifies consistently high-rated areas across processes and highlights patterns |
| Weaknesses | Areas where improvement is needed; specific gaps found | AI catalogs all gaps found during evidence analysis with traceability to specific indicators |
| Improvement opportunities | Recommended actions to address weaknesses | AI prioritizes improvements by impact and effort, suggests proven remediation approaches |
| Assessment record | Detailed evidence references supporting each rating | AI maintains a complete, linked evidence catalog throughout the assessment |
Interpreting Results with AI
| Output Element | Manual Interpretation | AI-Enhanced Interpretation |
|---|---|---|
| PA ratings (N/P/L/F) | Assessor explains each rating in the closing meeting | AI generates detailed rating justifications with linked evidence for each PA |
| Capability level | Team understands current level and gap to target | AI calculates effort estimates to reach the next level based on historical data from similar organizations |
| Improvement recommendations | Team creates an improvement plan manually | AI drafts an improvement plan with tasks, priorities, dependencies, and timeline suggestions |
| Trend analysis | Comparison with previous assessments done by reading old reports | AI overlays current and historical results in a dashboard view, highlighting progress and regression |
From Assessment to Action
The assessment report is the beginning, not the end. The real value comes from acting on the findings:
- Prioritize improvements — Focus on weaknesses that block the target capability level
- Assign ownership — Every improvement action needs a responsible person and a deadline
- Track progress — Use the assessment indicators as the measure of improvement
- Re-assess — Verify improvements through a follow-up assessment or internal review
AI advantage: AI can transform the static assessment report into a living improvement tracker that monitors evidence accumulation in real time and alerts when an improvement action's deadline approaches without sufficient progress.
Assessment Rating
Base practices are rated on a scale:
| Rating | Description |
|---|---|
| F (Fully) | >85% to 100% achieved |
| L (Largely) | >50% to 85% achieved |
| P (Partially) | >15% to 50% achieved |
| N (Not) | 0% to 15% achieved |
Note: The percentage ranges use exclusive lower bounds (except for N). For example, exactly 50% would be rated P, while 50.1% would be rated L.
Process capability level is determined by aggregating BP ratings.
PAM and AI Tools
AI can support PAM compliance in several ways:
| PAM Element | AI Support |
|---|---|
| Base practices | AI can automate/assist practice execution |
| Work products | AI can generate/validate work product content |
| Characteristics | AI can check characteristic compliance |
| Traceability | AI can suggest/validate trace links |
Important: AI support must be verified. Work products remain the responsibility of human engineers.
Implementation Checklist
Use this checklist to verify that your organization has the PAM-related foundations in place. Each item maps to a key concept from this chapter.
| # | Checklist Item | PAM Concept | Status |
|---|---|---|---|
| 1 | Process scope is defined (which processes will be assessed) | Process Dimension | |
| 2 | Target capability level is agreed upon with stakeholders | Capability Dimension | |
| 3 | Base practices for each in-scope process are identified and understood | BP identification | |
| 4 | Work products for each process are cataloged with their characteristics | WP characteristics | |
| 5 | Evidence exists (or is planned) for each base practice | Evidence collection | |
| 6 | Work products meet identification, content, traceability, verification, and structure characteristics | WP compliance | |
| 7 | Generic practices for the target level are understood and implemented | GP indicators | |
| 8 | Generic resources (people, tools, infrastructure, methods) are in place | GR indicators | |
| 9 | Rating scale (N/P/L/F) is understood by all team members | Rating scale | |
| 10 | Assessment method and assessor roles are understood | Assessment method | |
| 11 | AI tools are identified for evidence collection and gap analysis support | AI in assessment | |
| 12 | Preparation timeline allows at least 8-12 weeks before formal assessment | Assessment preparation | |
| 13 | Internal pre-assessment or dry run is scheduled | Assessment preparation | |
| 14 | Improvement tracking mechanism is in place for post-assessment actions | Assessment output |
Tip: This checklist is a starting point. Tailor it to your organization's context, scope, and target capability level. AI can help by auto-populating status fields based on evidence discovery scans.
Summary
The Process Assessment Model (PAM):
- Provides assessment guidance through base practices and work products
- Uses a two-dimensional framework: process dimension (what to assess) and capability dimension (how well)
- Base practices are assessable activities implementing processes
- Work products are tangible outputs with defined characteristics
- Generic Practices (GP) and Generic Resources (GR) serve as assessment indicators for capability levels
- Assessments follow a structured method with defined roles (Lead Assessor, Co-Assessor, Sponsor, Coordinator)
- The N-P-L-F rating scale provides objective, repeatable measurement of process attribute achievement
- Evidence of BPs and WPs demonstrates process implementation
- AI can support evidence collection, gap analysis, and reporting but cannot replace assessor judgment
- Assessment output drives targeted improvement planning
The PAM answers "How do we assess process achievement?" Capability levels (next section) answer "How mature is the process?"