Automotive SPICE®

Process Reference Model

Process Assessment Model

Version 4.0

Title:

Automotive SPICE Process Assessment / Reference Model

Author(s):

VDA Working Group 13

Version:

4.0

Date:

2023-11-29

Status:

Released

Acknowledgement

The VDA, the VDA QMC and the Project Group 13 explicitly acknowledge the high-quality work carried out by the members of the intacs® working groups. We would like to thank all involved people who have contributed to the development and publication of Automotive SPICE®.

Derivative works

You may not alter, transform, or build upon this work without the prior consent of the VDA Quality Management Center. Such consent may be given provided ISO copyright is not infringed.

The detailed descriptions contained in this document may be incorporated as part of any tool or other material to support the performance of process assessments, so that this process assessment model can be used for its intended purpose, provided that any such material is not offered for sale.

All distribution of derivative works shall be made at no cost to the recipient.

Document distribution

The Automotive SPICE® process assessment model may only be obtained by download from the www.vda-qmc.de web site. It is not permitted for the recipient to further distribute the document.

Change requests

Any problems or change requests should be reported through the defined mechanism at the www.vda-qmc.de web site.

Trademark notice

Automotive SPICE® is a registered trademark of the Verband der Automobilindustrie e.V. (VDA) For further information about Automotive SPICE® visit www.vda-qmc.de.

Document history

Version

Date

By

Notes

2.0

2005-05-04

AutoSIG / SUG

DRAFT RELEASE, pending final editorial review

2.1

2005-06-24

AutoSIG / SUG

Editorial review comments implemented

Updated to reflect changes in FDIS 15504-5

2.2

2005-08-21

AutoSIG / SUG

Final checks implemented: FORMAL RELEASE

2.3

2007-05-05

AutoSIG / SUG

Revision following CCB: FORMAL RELEASE

2.4

2008-08-01

AutoSIG / SUG

Revision following CCB: FORMAL RELEASE

2.5

2010-05-10

AutoSIG / SUG

Revision following CCB: FORMAL RELEASE

3.0

2015-07-16

VDA QMC WG13

Changes: See release notes

3.1

2017-11-01

VDA QMC WG13

Changes: See www.automotivespice.com

4.0

2023-10-20

VDA QMC WG13

Complete revision of PAM

Table of contents

Copyright notice 1

Acknowledgement 2

Derivative works 2

Document distribution 2

Change requests 2

Trademark notice 2

Document history 3

Table of contents 4

List of Figures 6

1. Introduction 8

1.1. Scope 8

1.2. Terminology 8

1.3. Abbreviations 13

2. Statement of compliance 14

3. Process capability determination 15

3.1. Process reference model 16

3.1.1. Primary life cycle processes category 17

3.1.2. Supporting life cycle processes category 20

3.1.3. Organizational life cycle processes category 20

3.2. Measurement framework 21

3.2.1. Process capability levels and process attributes 21

3.2.2. Process attribute rating 23

3.2.3. Rating and aggregation method 24

3.2.4. Process capability level model 26

3.3. Process assessment model 27

3.3.1. Assessment indicators 28

3.3.2. Understanding information Items and work products 29

3.3.3. Understanding the level of abstraction of a PAM 31

3.3.4. Why a PRM and PAM are not a lifecycle model or development process blueprint 31

4. Process reference model and performance indicators (Level 1) 33

4.1. Acquisition process group (ACQ) 34

4.1.1. ACQ.4 Supplier Monitoring 34

4.2. Supply process group (SPL) 37

4.2.1. SPL.2 Product Release 37

4.3. System engineering process group (SYS) 40

4.3.1. SYS.1 Requirements Elicitation 40

4.3.2. SYS.2 System Requirements Analysis 43

4.3.3. SYS.3 System Architectural Design 46

4.3.4. SYS.4 System Integration and Integration Verification 49

4.3.5. SYS.5 System Verification 52

4.4. Software engineering process group (SWE) 55

4.4.1. SWE.1 Software Requirements Analysis 55

4.4.2. SWE.2 Software Architectural Design 58

4.4.3. SWE.3 Software Detailed Design and Unit Construction 61

4.4.4. SWE.4 Software Unit Verification 64

4.4.5. SWE.5 Software Component Verification and Integration Verification 67

4.4.6. SWE.6 Software Verification 71

4.5. Validation process group (VAL) 74

4.5.1. VAL.1 Validation 74

4.6. Machine Learning Engineering process group (MLE) 77

4.6.1. MLE.1 Machine Learning Requirements Analysis 77

4.6.2. MLE.2 Machine Learning Architecture 80

4.6.3. MLE.3 Machine Learning Training 83

4.6.4. MLE.4 Machine Learning Model Testing 86

4.7. Hardware Engineering process group (HWE) 90

4.7.1. HWE.1 Hardware Requirements Analysis 90

4.7.2. HWE.2 Hardware Design 94

4.7.3. HWE.3 Verification against Hardware Design 97

4.7.4. HWE.4 Verification against Hardware Requirements 100

4.8. Supporting process group (SUP) 103

4.8.1. SUP.1 Quality Assurance 103

4.8.2. SUP.8 Configuration Management 106

4.8.3. SUP.9 Problem Resolution Management 110

4.8.4. SUP.10 Change Request Management 113

4.8.5. SUP.11 Machine Learning Data Management 116

4.9. Management process group (MAN) 119

4.9.1. MAN.3 Project Management 119

4.9.2. MAN.5 Risk Management 123

4.9.3. MAN.6 Measurement 126

4.10. Process improvement process group (PIM) 129

4.10.1. PIM.3 Process Improvement 129

4.11. Reuse process group (REU) 132

4.11.1. REU.2 Management of Products for Reuse 132

5. Process capability levels and process attributes 135

5.1. Process capability level 0: Incomplete process 135

5.2. Process capability Level 1: Performed process 135

5.2.1. PA 1.1 Process performance process attribute 136

5.3. Process capability Level 2: Managed process 136

5.3.1. PA 2.1 Process performance management process attribute 137

5.3.2. PA 2.2 Work product management process attribute 141

5.4. Process capability Level 3: Established process 143

5.4.1. PA 3.1 Process definition process attribute 144

5.4.2. PA 3.2 Process deployment process attribute 147

5.5. Process capability Level 4: Predictable process 150

5.5.1. PA 4.1 Quantitative analysis process attribute 150

5.5.2. PA 4.2 Quantitative control process attribute 153

5.6. Process capability Level 5: Innovating process 155

5.6.1. PA 5.1 Process innovation process attribute 155

5.6.2. PA 5.2 Process innovation implementation process attribute 157

Annex A Conformity statements 160

Annex A.1 Introduction 160

Annex A.2 Conformance to the requirements for process reference models 160

Annex A.3 Conformance to the requirements for process assessment models 160

Annex A.4 Conformance to the requirements for measurement frameworks 162

Annex B Information Item Characteristics 163

Annex C Key concepts and guidance 185

Annex C.1 The “Plug-in” concept 185

Annex C.2 “Element”, “Component”, and “Unit” 186

Annex C.3 Integration of Machine Learning Engineering Processes 187

Annex C.4 Example of an ML Architecture 189

Annex C.5 Traceability and consistency 190

Annex C.6 “Agree” and “Summarize and Communicate” 192

Annex C.7 Key Changes in Automotive SPICE 4.0 193

Terminology – “Measure” vs. “Metric” 193

Terminology – “Affected Party” (Level 1) vs. “Involved Party” (Level 2) 193

Terminology – “Verification” instead of “Testing” 194

Annex D Reference standards 195

List of Figures

Figure 1 — Process assessment model relationship 15

Figure 2 — Automotive SPICE process reference model – Overview 16

Figure 3 — Automotive SPICE process reference model – Overview4 16

Figure 4 — Relationship between assessment indicators and process capability 29

Figure 5 — Possible levels of abstraction for the term “process” 31

Figure 6 — Performing a process assessment for determining process capability 32

List of Tables

Table 1 — Terminology 12

Table 2 — Abbreviation List 14

Table 3 — Primary life cycle processes – ACQ process group 18

Table 4 — Primary life cycle processes – SPL process group 18

Table 5 — Primary life cycle processes – SYS process group 18

Table 6 — Primary life cycle processes – VAL process group 19

Table 7 — Primary life cycle processes – SWE process group 19

Table 8 — Primary life cycle processes – MLE process group 19

Table 9 — Primary life cycle processes – HWE process group 19

Table 10 — Supporting life cycle processes - SUP process group 20

Table 11 — Organizational life cycle processes - MAN process group 20

Table 12 — Organizational life cycle processes - PIM process group 20

Table 13 — Organizational life cycle processes - REU process group 20

Table 14 — Process capability levels 21

Table 15 — Process attributes 22

Table 16 — Rating scale 23

Table 17 — Rating scale percentage values 23

Table 18 — Refinement of rating scale 24

Table 19 — Refined rating scale percentage values 24

Table 20 — Capability levels and corresponding process attribute ratings 27

Table 21 — Template for the process description 33

Table B. 1 — Structure of information item characteristics (IIC) table 163

Table B. 2 — Information Item Characteristics 184

1. Introduction

1.1. Scope

Process assessment is a disciplined evaluation of an organizational unit’s processes against a process assessment model.

The Automotive SPICE process assessment model (PAM) is intended for use when performing conformant assessments of the process capability on the development of embedded automotive systems. It was developed in accordance with the requirements of ISO/IEC 33004:2015.

Automotive SPICE has its own process reference model (PRM), which was developed based on the Automotive SPICE process reference model 4.5. It was further developed and tailored considering the specific needs of the automotive industry. If processes beyond the scope of Automotive SPICE are needed, appropriate processes from other process reference models such as ISO/IEC 12207 or ISO/IEC/IEEE 15288 may be added based on the business needs of the organization.

The PRM is incorporated in this document and is used in conjunction with the Automotive SPICE process assessment model when performing an assessment.

This Automotive SPICE process assessment model contains a set of indicators to be considered when interpreting the intent of the Automotive SPICE process reference model. These indicators may also be used when implementing a process improvement program.

1.2. Terminology

Automotive SPICE follows the following precedence for use of terminology:

  • ISO/IEC 33001 for assessment related terminology

  • ISO/IEC/IEEE 24765, ISO/SAE 21434 and ISO/IEC/IEEE 29119 terminology (as contained in Annex C)

  • Terms introduced by Automotive SPICE (as contained in Annex C)

  • PMBOK® Guide – Fourth Edition

  • PAS 1883:2020

Table 1 — Terminology

Term

Origin

Description

Activity

Automotive SPICE V4.0

Execution of a task by a stakeholder or an involved party.

Application parameter

Automotive SPICE V4.0

An application parameter is a software variable containing data that can be changed at the system or software levels; they influence the system’s or software behavior and properties. The notion of application parameter is expressed in two ways:

  • The specification (including variable names, the domain value range, technical data types, default values, physical unit (if applicable), the corresponding memory maps, respectively).

  • The actual quantitative data value it receives by means of data application.

Application parameters are not requirements. They are a technical implementation solution for configurability-oriented requirements.

Approval

Automotive SPICE V4.0

Written statement that a deliverable is fit for its intended use, and compliant with defined criteria.

Baseline

Automotive SPICE V4.0

A defined and coherent set of read-only information, serving as an input information the for affected parties.

Deliverable

PMBOK® Guide

– Fourth Edition

Any unique and verifiable product, result, or capability to perform a service that must be produced to complete a process, phase, or project. Often used more narrowly in reference to an external deliverable, which is a deliverable that is subject to approval by the project sponsor or customer.

Functional requirement

ISO/IEC/IEEE 24765

A statement that identifies what a product or process must accomplish to produce required behavior and/or results.

Hardware

intacs® working group HW PAM

Assembled and interconnected electrical or electronic hardware components or parts which perform analog or digital functions or operations.

Hardware component

intacs® working group HW PAM

Logical (e.g., functional block) or physical group of hardware parts realizing a functionality, which

  • cannot be realized by any of its hardware parts alone, e.g., voltage monitoring, power supply.

  • may be organized hierarchically, i.e., a hardware component can contain lower-level hardware components.

NOTE: Depending on the application, e.g., the populated PCB, a system-on-chip, a microcontroller, or an SBC can be considered a HW component.

Hardware element

intacs® working group HW PAM

Generic term; can represent a hardware component, a hardware part, a hardware interface, or the hardware.

Hardware part

Automotive SPICE V4.0

Fundamental HW element the purpose and functionality of which cannot be further subdivided or separated.

NOTE: Examples are transistors, resistors, diodes, nonpopulated PCB

NOTE: Depending on the application, e.g., a system-on-chip, a microcontroller or an SBC can be considered a HW part.

NOTE: the term ‘unit’ is considered to apply to the software domain only. The term ‘hardware part’ can be viewed as the hardware counterpart of ‘software unit’.

Hyperparameter

Automotive SPICE V4.0

In machine learning, a hyperparameter is a parameter whose value is used to control the training of the ML model. Its value must be set between training iterations. Examples: learning rate, loss function, model depth, regularization constants.

Information need

Automotive SPICE V4.0

The need for characterizing process or product related effectiveness and efficiency (used by MAN.6 and PA 4.1).

Machine Learning (ML)

Automotive SPICE V4.0

In Automotive SPICE Machine Learning (ML) describes the ability of software to learn from specific training data and to apply this knowledge to other similar tasks.

Measure

Automotive SPICE V4.0

An activity to achieve a certain intent.

Measurement

Oxford Dictionary

“The activity to find the size, quantity or degree of something”.

Metric

Automotive SPICE V4.0

A quantitative or qualitative measurable indicator that matches defined information needs.

Operational Design Domain

PAS 1883:2020

Operational Design Domain (ODD) is operating conditions under which a given overall system or feature thereof is specifically designed to function.

This includes, but is not limited to, environmental, geographical, and time-of-day restrictions, and/or the requisite presence or absence of certain traffic or roadway characteristics.

Project

ISO/IEC/IEEE 24765

Endeavor with defined start and finish dates undertaken to create a product or service in accordance with specified resources and requirements.

Release

Automotive SPICE V4.0

A physical product delivered to a customer, including a defined set of functionalities and properties.

Regression verification

Automotive SPICE V4.0

Selective re-verification of elements to verify that modifications have not caused unintended effects.

Risk

ISO/IEC/IEEE 24765

The combination of the probability of occurrence and the consequences of a given future undesirable event.

Software component

Automotive SPICE V4.0

Software component in design and implementation-oriented processes:

The software architecture decomposes the software into software components across appropriate hierarchical levels down to the lowest-level software components in a conceptual model.

Software component in verification-oriented processes: The implementation of a SW component under verification is represented e.g., as source code, object files, library file, executable, or executable model.

Software element

Automotive SPICE V4.0

Refers to software component or software unit

Software unit

Automotive SPICE V4.0

Software unit in design and implementation-oriented processes: As a result of the decomposition of a software component, the software is decomposed into software units which are a representation of a software element, which is decided not to be further subdivided and that is a part of a software component at the lowest level, in a conceptual model.

Software unit in verification-oriented processes:

An implemented SW unit under verification is represented e.g., as source code files, or an object file.

Stakeholder requirements

Automotive SPICE V4.0

Any type of requirement for the stakeholders in the given context, e.g., customer requirement, supplier internal requirements (product-specific, platform etc.), legal requirements, regulatory requirements, statutory requirements, industry sector requirements, international standards, codes of practice etc. …

System Element

Automotive SPICE V4.0

System elements can be:

  • Logical and structural objects at the architectural and design level. System elements can be further decomposed into more fine-grained system elements of the architecture or design across appropriate hierarchical levels.

  • Physical representations of these objects, or a combination, e.g., peripherals, sensors, actuators, mechanical parts, software executables.

Task

Automotive SPICE V4.0

A definition, but not the execution, of a coherent and set of atomic actions.

Validation measure

Automotive SPICE V4.0

Validation measure can be:

  • Operational use case testing under real-life conditions

  • Highly accelerated life testing (HALT)

  • Simulations under real-life conditions

  • End user trials

  • Panel or blind tests

  • Expert panels

Verification

Automotive SPICE V4.0

Verification is confirmation through the provision of objective evidence that an element fulfils the specified requirements.

Verification measure

Automotive SPICE V4.0

Verification measure can be:

  • Test cases

  • Measurements

  • Calculations

  • Simulations

  • Reviews

  • Analyses

Note, that in particular domains certain verification measures may not be applicable, e.g., software units generally cannot be verified by means of calculations or analyses.

1.3. Abbreviations

Table 2 — Abbreviation List

BP

Base Practice

CAN

Controller Area Network

CASE

Computer-Aided Software Engineering,

CCB

Change Control Board

CPU

Central Processing Unit

ECU

Electronic Control Unit

EEPROM

Electrically Erasable Programmable Read Only Memory

EOL

End-of-Line

FMEA

Failure Mode and Effect Analysis

FTA

Fault Tree Analysis

GP

Generic Practice

GR

Generic Resource

IEC

International Electrotechnical Commission

IEEE

Institute of Electrical and Electronics Engineers

I/O

Input / Output

ISO

International Organization for Standardization

LIN

Local Interconnect Network

MISRA

Motor Industry Software Reliability Association

MOST

Media Oriented Systems Transport

ODD

Operational Design Domain

PA

Process Attribute

PAM

Process Assessment Model

PRM

Process Reference Model

PWM

Pulse Width Modulation

RAM

Random Access Memory

ROM

Read Only Memory

SPICE

Systems Process Improvement and Capability dEtermination

SUG

Spice User Group

USB

Universal Serial Bus

WP

Work Product

WPC

Work Product Characteristic

2. Statement of compliance

The Automotive SPICE process reference model and process assessment model are conformant with the ISO/IEC 33004:2015 and can be used as the basis for conducting an assessment of process capability.

An ISO/IEC 33003:2015 compliant Measurement Framework is defined in section 5.

A statement of compliance of the process assessment model and process reference model with the requirements of ISO/IEC 33004:2015 is provided in Annex A.

A statement of compliance of the measurement framework with the requirements of ISO/IEC 33003:2015 is provided in Annex A.

3. Process capability determination

The concept of process capability determination by using a process assessment model is based on a two-dimensional framework. The first dimension is provided by processes defined in a process reference model (process dimension). The second dimension consists of capability levels that are further subdivided into process attributes (capability dimension). The process attributes provide the measurable characteristics of process capability.

The process assessment model selects processes from a process reference model and supplements with indicators. These indicators support the collection of objective evidence which enable an assessor to assign ratings for processes according to the capability dimension.

The relationship is shown in Figure 1.

_images/image1.png

Figure 1 — Process assessment model relationship

3.1. Process reference model

Processes are collected into process groups according to the domain of activities they address.

These process groups are organized into 3 process categories: Primary life cycle processes, Organizational life cycle processes and Supporting life cycle processes.

For each process a purpose statement is formulated that contains the unique functional objectives of the process when performed in a particular environment. For each purpose statement a list of specific outcomes is associated, as a list of expected positive results of the process performance.

For the process dimension, the Automotive SPICE process reference model provides the set of processes shown in Figure 2.

_images/image2.png

Figure 2 — Automotive SPICE process reference model – Overview

3.1.1. Primary life cycle processes category

The primary life cycle processes category consists of processes that may apply for an acquirer of products from a supplier or may apply for product development when responding to stakeholder needs and delivering products including the engineering processes needed for specification, design, implementation, integration and verification.

The primary life cycle processes category consists of the following groups:

  • the Acquisition process group

  • the Supply process group

  • the System engineering process group

  • the Validation process group

  • the Software engineering process group

  • the Machine learning engineering process group

  • the Hardware engineering process group

The Acquisition process group (ACQ) consists of one process that is performed by the customer, or by the supplier when acting as a customer for its own suppliers, in order to acquire a product and/or service.

Table 3 — Primary life cycle processes – ACQ process group

ACQ.4

Supplier Monitoring

The Supply process group (SPL) consists of one process performed by the supplier in order to supply a product and/or a service.

Table 4 — Primary life cycle processes – SPL process group

SPL.2

Product Release

The System Engineering process group (SYS) consists of processes addressing the elicitation and management of customer and internal requirements, the definition of the system architecture and the integration and verification on the system level.

Table 5 — Primary life cycle processes – SYS process group

SYS.1

Requirements Elicitation

SYS.2

System Requirements Analysis

SYS.3

System Architectural Design

SYS.4

System Integration and Integration Verification

SYS.5

System Verification

The Validation process group (VAL) consists of one process that is performed to provide evidence that the product to be delivered satisfies the expectations for its intended use.

Table 6 — Primary life cycle processes – VAL process group

VAL.1

Validation

The Software Engineering process group (SWE) consists of processes addressing the management of software requirements derived from the system requirements, the development of the corresponding software architecture and design as well as the implementation, integration and verification of the software.

Table 7 — Primary life cycle processes – SWE process group

SWE.1

Software Requirements Analysis

SWE.2

Software Architectural Design

SWE.3

Software Detailed Design and Unit Construction

SWE.4

Software Unit Verification

SWE.5

Software Component Verification and Integration Verification

SWE.6

Software Verification

The Machine Learning Engineering process group (MLE) consists of processes addressing the management of ML requirements derived from the software requirements, the development of the corresponding ML architecture, the training of ML model, and testing of ML model against ML requirements.

Table 8 — Primary life cycle processes – MLE process group

MLE.1

Machine Learning Requirements Analysis

MLE.2

Machine Learning Architecture

MLE.3

Machine Learning Training

MLE.4

Machine Learning Model Testing

The Hardware Engineering process group (HWE) consists of processes addressing the management of hardware requirements derived from the system requirements, the development of the corresponding hardware architecture and design as well as the verification of the hardware.

Table 9 — Primary life cycle processes – HWE process group

HWE.1

Hardware Requirements Analysis

HWE.2

Hardware Design

HWE.3

Verification against Hardware Design

HWE.4

Verification against Hardware Requirements

3.1.2. Supporting life cycle processes category

The supporting life cycle processes category consists of processes that may be employed by any of the other processes at various points in the life cycle.

Table 10 — Supporting life cycle processes - SUP process group

SUP.1

Quality Assurance

SUP.8

Configuration Management

SUP.9

Problem Resolution Management

SUP.10

Change Request Management

SUP.11

Machine Learning Data Management

3.1.3. Organizational life cycle processes category

The organizational life cycle processes category consists of processes that develop process, product, and resource assets which, when used by projects in the organization, may help the organization achieve its business goals.

The organizational life cycle processes category consists of the following groups:

  • the Management process group;

  • the Process Improvement process group;

  • the Reuse process group.

The Management process group (MAN) consists of processes that may be used by anyone who manages any type of project or process within the life cycle.

Table 11 — Organizational life cycle processes - MAN process group

MAN.3

Project Management

MAN.5

Risk Management

MAN.6

Measurement

The Process Improvement process group (PIM) covers one process that contains practices to improve the processes performed in the organizational unit.

Table 12 — Organizational life cycle processes - PIM process group

PIM.3

Process Improvement

The Reuse process group (REU) covers one process to systematically exploit reuse opportunities in organization’s product portfolio.

Table 13 — Organizational life cycle processes - REU process group

REU.2

Management of Products for Reuse

3.2. Measurement framework

The measurement framework provides the necessary requirements and rules for the capability dimension. It defines a schema which enables an assessor to determine the Capability Level of a given process. These capability levels are defined as part of the measurement framework.

To enable the rating, the measurement framework provides process attributes defining a measurable property of process capability. Each process attribute is assigned to a specific capability level. The extent of achievement of a certain process attribute is represented by means of a rating based on a defined rating scale. The rules from which an assessor can derive a final capability level for a given process are represented by a process capability level model.

Automotive SPICE defines its own measurement framework.

Note: The Automotive SPICE measurement framework is an adaption of ISO/IEC 33020:2019. Text incorporated from ISO/IEC 33020 within this chapter is written in italic font and marked with a left side bar.

3.2.1. Process capability levels and process attributes

The process capability levels, and associated process attributes are described in detail in chapter 5.

Process attributes are features of a process that can be evaluated on a scale of achievement, providing a measurement of the capability of the process. They are applicable to all processes.

A capability level is characterized by one or more process attributes whose implementation result in a significant improvement in the capability to perform a process. Each attribute addresses a specific aspect of the capability level. The levels constitute a rational way of progressing through improvement of the capability of any process.

There are six capability levels as listed in Table 14, incorporating nine process attributes:

Table 14 — Process capability levels

Level 0:

Incomplete process

The process is not implemented or fails to achieve its process purpose.

Level 1:

Performed process

The implemented process achieves its process purpose

Level 2:

Managed process

The previously described performed process is now implemented in a managed fashion (planned, monitored and adjusted) and its work products are appropriately established, controlled and maintained.

Level 3:

Established process

The previously described managed process is now implemented using a defined process that is capable of achieving its process outcomes.

Level 4:

Predictable process

The previously described established process now operates predictively within defined limits to achieve its process outcomes. Quantitative management needs are identified, measurement data are collected and analyzed to identify assignable causes of variation.

Corrective action is taken to address assignable causes of variation.

Level 5:

Innovating process

The previously described predictable process is now continually improved to respond to organizational change.

Within this process assessment model, the determination of capability is based upon the nine process attributes (PA) as listed in Table 15 — Process attributes.

Table 15 — Process attributes

Attribute ID

Process Attributes

Level 0: Incomplete process

Level 1: Performed process

PA 1.1

Process performance process attribute

Level 2: Managed process

PA 2.1

Performance management process attribute

PA 2.2

Work product management process attribute

Level 3: Established process

PA 3.1

Process definition process attribute

PA 3.2

Process deployment process attribute

Level 4: Predictable process

PA 4.1

Quantitative analysis process attribute

PA 4.2

Quantitative control process attribute

Level 5: Innovating process

PA 5.1

Process innovation process attribute

PA 5.2

Process innovation implementation process attribute

3.2.2. Process attribute rating

To support the rating of process attributes, the measurement framework provides a defined rating scale with an option for refinement, different rating methods and different aggregation methods depending on the class of the assessment (e.g., required for organizational maturity assessments).

3.2.2.1. Rating scale

Within this process measurement framework, a process attribute is a measureable property of process capability. A process attribute rating is a judgement of the degree of achievement of the process attribute for the assessed process.

The rating scale is shown in Table 16 — Rating scale.

Note. The rating scale is identical to ISO/IEC 33020:2019

Table 16 — Rating scale

N

Not achieved

There is little or no evidence of achievement of the defined process attribute in the assessed process.

P

Partially achieved

There is some evidence of an approach to, and some achievement of, the defined process attribute in the assessed process. Some aspects of achievement of the process attribute may be unpredictable.

L

Largely achieved

There is evidence of a systematic approach to, and significant achievement of, the defined process attribute in the assessed process. Some weaknesses related to this process attribute may exist in the assessed process.

F

Fully achieved

There is evidence of a complete and systematic approach to, and full achievement of, the defined process attribute in the assessed process. No significant weaknesses related to this process attribute exist in the assessed process.

The ordinal scale defined above shall be understood in terms of percentage achievement of a process attribute. The corresponding percentages shall be:

Table 17 — Rating scale percentage values

N

Not achieved

0 to ≤ 15% achievement

P

Partially achieved

> 15% to ≤ 50% achievement

L

Largely achieved

> 50% to ≤ 85% achievement

F

Fully achieved

> 85% to ≤ 100% achievement

The ordinal scale may be further refined for the measures P and L as defined below.

Table 18 — Refinement of rating scale

P-

Partially achieved:

There is some evidence of an approach to, and some achievement of, the defined process attribute in the assessed process. Many aspects of achievement of the process attribute may be unpredictable.

P+

Partially achieved:

There is some evidence of an approach to, and some achievement of, the defined process attribute in the assessed process. Some aspects of achievement of the process attribute may be unpredictable.

L-

Largely achieved:

There is evidence of a systematic approach to, and significant achievement of, the defined process attribute in the assessed process. Many weaknesses related to this process attribute may exist in the assessed process.

L+

Largely achieved:

There is evidence of a systematic approach to, and significant achievement of, the defined process attribute in the assessed process. Some weaknesses related to this process attribute may exist in the assessed process.

The corresponding percentages shall be:

Table 19 — Refined rating scale percentage values

P-

Partially achieved -

> 15% to ≤ 32.5% achievement

P+

Partially achieved +

> 32.5 to ≤ 50% achievement

L-

Largely achieved -

> 50% to ≤ 67.5% achievement

L+

Largely achieved +

> 67.5% to ≤ 85% achievement

3.2.3. Rating and aggregation method

Rating and aggregation methods are taken from ISO/IEC 33020:2019, which provides the following definitions:

A process outcome is the observable result of successful achievement of the process purpose.

A process attribute outcome is the observable result of achievement of a specified process attribute.

Process outcomes and process attribute outcomes may be characterised as an intermediate step to providing a process attribute rating.

When performing rating, the rating method employed shall be specified relevant to the class of assessment. The following rating methods are defined.

The use of rating method may vary according to the class, scope and context of an assessment. The lead assessor shall decide which (if any) rating method to use. The selected rating method(s) shall be specified in the assessment input and referenced in the assessment report.

ISO/IEC 33020:2019 provides the following 3 rating methods: Rating method R1

The approach to process attribute rating shall satisfy the following conditions:

  1. Each process outcome of each process within the scope of the assessment shall be characterized for each process instance, based on validated data;

  2. Each process attribute outcome of each process attribute for each process within the scope of the assessment shall be characterized for each process instance, based on validated data;

  3. Process outcome characterizations for all assessed process instances shall be aggregated to provide a process performance attribute achievement rating;

  4. Process attribute outcome characterizations for all assessed process instances shall be aggregated to provide a process attribute achievement rating.

Rating method R2

The approach to process attribute rating shall satisfy the following conditions:

  1. Each process attribute for each process within the scope of the assessment shall be characterized for each process instance, based on validated data;

  2. Process attribute characterizations for all assessed process instances shall be aggregated to provide a process attribute achievement rating.

Rating method R3

Process attribute rating across assessed process instances shall be made without aggregation.

In principle the three rating methods defined in ISO/IEC 33020:2019 depend on

  1. whether the rating is made only on process attribute level (Rating method 3 and 2) or – with more level of detail – both on process attribute and process attribute outcome level (Rating method 1); and

  2. the type of aggregation ratings across the assessed process instances for each process

If a rating is performed for both process attributes and process attribute outcomes (Rating method 1), the result will be a process performance attribute outcome rating on level 1 and a process attribute achievement rating on higher levels.

Depending on the class, scope and context of the assessment an aggregation within one process (one-dimensional, vertical aggregation), across multiple process instances (one-dimensional, horizontal aggregation) or both (two-dimensional, matrix aggregation) is performed.

ISO/IEC 33020:2019 provides the following examples:

When performing an assessment, ratings may be summarized across one or two dimensions.

For example, when rating a

  • process attribute for a given process, one may aggregate ratings of the associated process (attribute) outcomes – such an aggregation will be performed as a vertical aggregation (one dimension).

  • process (attribute) outcome for a given process attribute across multiple process instances, one may aggregate the ratings of the associated process instances for the given process (attribute) outcome such an aggregation will be performed as a horizontal aggregation (one dimension)

  • process attribute for a given process, one may aggregate the ratings of all the process (attribute) outcomes for all the processes instances – such an aggregation will be performed as a matrix aggregation across the full scope of ratings (two dimensions)

The standard defines different methods for aggregation. Further information can be taken from ISO/IEC 33020:2019.

3.2.4. Process capability level model

The process capability level achieved by a process shall be derived from the process attribute ratings for that process according to the process capability level model defined in Table 20 — Capability levels.

The process capability level model defines the rules how the achievement of each level depends on the rating of the process attributes for the assessed and all lower levels.

As a general rule the achievement of a given level requires a largely or fully achievement of the corresponding process attributes and a full achievement of any lower lying process attribute.

Table 20 — Capability levels and corresponding process attribute ratings

Scale

Process attribute

Rating

Level 1

PA 1.1: Process performance process attribute

Largely or fully

Level 2

PA 1.1: Process performance process attribute

PA 2.1: Process performance management process attribute

PA 2.2: Work product management process attribute

Fully

Largely or fully

Largely or fully

Level 3

PA 1.1: Process performance process attribute

PA 2.1: Process performance management process attribute

PA 2.2: Work product management process attribute

PA 3.1: Process definition process attribute

PA 3.2: Process deployment process attribute

Fully

Fully

Fully

Largely or fully

Largely or fully

Level 4

PA 1.1: Process performance process attribute

PA 2.1: Process performance management process attribute

PA 2.2: Work product management process attribute

PA 3.1: Process definition process attribute

PA 3.2: Process deployment process attribute

PA 4.1: Quantitative analysis process attribute

PA 4.2: Quantitative control process attribute

Fully

Fully

Fully

Fully

Fully

Largely or fully

Largely or fully

Level 5

PA 1.1: Process performance process attribute

PA 2.1: Process performance management process attribute

PA 2.2: Work product management process attribute

PA 3.1: Process definition process attribute

PA 3.2: Process deployment process attribute

PA 4.1: Quantitative analysis process attribute

PA 4.2: Quantitative control process attribute

PA 5.1: Process innovation process attribute

PA 5.2: Process innovation implementation process attribute

Fully

Fully

Fully

Fully

Fully

Fully

Fully

Largely or fully

Largely or fully

3.3. Process assessment model

The process assessment model offers indicators in order to identify whether the process outcomes and the process attribute outcomes (achievements) are present or absent in the instantiated processes of projects and organizational units. These indicators provide guidance for assessors in accumulating the necessary objective evidence to support judgments of capability. They are not intended to be regarded as a mandatory set of checklists to be followed.

3.3.1. Assessment indicators

According to ISO/IEC 33004, a process assessment model needs to define a set of assessment indicators:

Assessment Indicators

A process assessment model shall be based on a set of assessment indicators that:

  1. explicitly address the purpose and process outcomes, as defined in the selected process reference model, of each of the processes within the scope of the process assessment model;

  2. demonstrate the achievement of the process attributes within the scope of the process assessment model;

  3. demonstrate the achievement (where relevant) of the process quality levels within the scope of the process assessment model.

The assessment indicators generally fall into three types:

  1. practices that support achievement of either the process purpose or the specific process attribute.

  2. information items and their characteristics that demonstrate the respective achievements.

  3. resources and infrastructure that support the respective achievements. [ISO/IEC 33004:2015, 6.3.1]

In this assessment model, only practices and information items are used.

Practices are representing activity-oriented indicators, where information items are representing result-oriented indicators. Both practices and information items are used for judging objective evidence to be collected and accumulated in the performance of an assessment.

As a first type of assessment indicator, practices are provided, which can be divided into two types:

1. Base practices (BP), applying to capability level 1

They provide an indication of the extent of achievement of the process outcomes. Base practices relate to one or more process outcomes, thus being always process-specific and not generic.

2. Generic practices (GP), applying to capability levels 1 to 5

They provide an indication of the extent of process attribute achievement. Generic practices relate to one or more process attribute achievements, thus applying to any process.

As a second type of assessment indicators, information items (II) including their characteristics (IIC) are provided in Annex B.

These are meant to offer a good practice and state-of-the-art knowledge guide for the assessor. Therefore, information items including their characteristics are supposed to be a quickly accessible information source during an assessment.

Information item characteristics shall not be interpreted as a required structure of a corresponding work products, which is defined by the project and organization, respectively.

Please refer to chapter 3.3.2 for understanding the difference between information items and work products.

ISO 33004:2015 requires the mapping of assessment indicators to process attributes as shown in figure 3.

The capability of a process on level 1 is only characterized by the measure of the extent to which the process outcomes are achieved. According to ISO 33003:2015, a measurement framework requires each level to reveal a process attribute. Therefore, the only process performance attribute for capability Level 1 (PA.1.1) has a single generic practice (GP 1.1.1) pointing as an editorial reference to the respective process performance indicators (see figure 3 and chapter 4).

_images/image3.jpeg

Figure 4 — Relationship between assessment indicators and process capability

The detailed mapping of base practices / indicators and generic practices / indicators to process outcomes and achievements, is provided in corresponding tables in chapter 4 and 5, respectively.

3.3.2. Understanding information Items and work products

In order to judge the presence or absence of process outcomes and process attribute achievements an assessment obtains objective evidence. All such evidence comes either from the examination of work products related to a specific output of the processes assessed, or from statements made by the performers and managers of the processes. Sources for such evidence is either repository content of the assessed processes, or testimony provided by the performers and managers of the assessed processes.

As described in chapter 3.3.1, this process assessment model provides information items serving as indicators to guide the assessor when judging a process attribute achievement.

3.3.2.1. Information items versus work products

ISO/IEC 33001 provides the following definition of the term “information item”:

information item

separately identifiable body of information that is produced, stored, and delivered for human use

Note 1 to entry: An information item can be produced in several versions during a system, software, or service life cycle. Syn: information product.

[ISO/IEC 33001:2015, 3.1.4]

Note: Human use includes the information stored, managed and processed by a tool.

One common definition of the term “work product” is:

work product

artifact resulting from the execution of a process

[ISO/IEC/IEEE 24765:2017]

Both terms are used in different context in an assessment:

  • Information items are defining relevant pieces of information used by the assessors to judge the achievement of process attributes.

  • Work products are produced by the organization assessed when performing, managing, establishing, analyzing and innovating processes.

Information items (together with their characteristics) are provided as guidance for “what to look for” when examining the work products available in the assessed organization. The extent of implementation of an information item (in line with its defined characteristics) in a related work product serves as objective evidence supporting the assessment of a particular process. A documented process and assessor judgment is needed to ensure that the process context (application domain, business purpose, development methodology, size of the organization, etc.) is considered when using this information.

Information items shall therefore not be mistaken for the work product generated by the assessed organization itself. There is no 1:1 relationship between an information item and the work product taken as sample evidence by the assessor when assessing the achievement of a process outcome and process attribute achievements. An output generated by a process may comprise multiple information item characteristics and multiple outputs may also contain the same information item characteristics.

Information item characteristics should be considered as indicators when considering whether, given the context, a work product is contributing to the intended purpose of the process. Context-sensitivity means that assessor judgment is needed to ensure that the actual context (application domain, business purpose, development methodology, size of the organization, etc.) is considered when using the information items.

3.3.2.2. Types of work products

A work product to be considered as evidence when rating a process attribute may not necessary be outputs from the processes assessed but can also be originated from other processes of the organization. Once such a work product is used in the performance of a process under assessment, it may be considered by the assessor as objective evidence.

In a lot of cases work products are comprising documentation aspects, such as specifications, reports, records, architectural designs, software code etc.

Examples of work products not comprising any documentation aspects are software binaries, raw data, or a physical electronic hardware.

3.3.3. Understanding the level of abstraction of a PAM

The term “process” can be understood at three levels of abstraction. Note that these levels of abstractions are not meant to define a strict black-or-white split, nor is it the aim to provide a scientific classification schema – the message here is to understand that, in practice, when it comes to the term “process” there are different abstraction levels, and that a PAM resides at the highest.

_images/image4.jpeg

Figure 5 — Possible levels of abstraction for the term “process”

Capturing experience acquired during product development (i.e., at the DOING level) in order to share this experience with others means creating a HOW level. However, a HOW is always specific to a particular context such as a company, an organizational unit, or a product line. For example, the HOW of a project, organizational unit, or company A is potentially not applicable as is to a project, organizational unit, or company B. However, both might be expected to adhere the principles represented by PAM indicators for process outcomes and process attribute achievements. These indicators are at the WHAT level while deciding on solutions for concrete templates, proceedings, and tooling etc. is left to the HOW level.

3.3.4. Why a PRM and PAM are not a lifecycle model or development process blueprint

A lifecycle model defines phases and activities in a logical timely order, possibly including cycles or loops, and parallelization. For example, some standards such as ISO 26262 or ISO/SAE 21434 are centered around a lifecycle model (neither of these standards in fact represents a PRM according to ISO/IEC 33004). Companies, organizational units, or projects will interpret such general lifecycle models given in standards, and then detail it out into roles, organizational interactions and interfaces, tools or tool chains, work instructions, and artifacts. Lifecycle models therefore are a concept at the HOW level (see Section 3.3.3).

In contrast, a PRM/PAM according to ISO/IEC 33004 (formerly ISO/IEC 15504-2) is at the level of the WHAT by abstracting from any HOW level, see Figure 4 in Section 3.3.3. In Automotive SPICE®, this has been, and is, indicated by the process MAN.3 Project Management requiring in BP2 “Define project life cycle”. A PRM/PAM groups a set of coherent and related characteristics of a particular technical topic and calls it ‘process’. In different terms, a process in a PRM represents a ‘distinct conceptual silo’. In this respect, a PRM/PAM

  • neither predefines, nor discourages, any order in which PRM processes or Base Practices are to be performed. Ultimately, in Automotive SPICE consistency must be fulfilled as required by the traceability/consistency Base Practices in MAN.3 or SYS.x, SWE.x, and HWE.x;

  • does not predefine any particular work product structure, or work product blueprints. For example, the process SYS.2 does not mean that there shall be exactly one system requirements specification containing everything provided by the stakeholders.

As a consequence, it is the assessor’s responsibility to perform a mapping of elements in such a HOW level to the Assessment Indicators in the PAM, see Figure 5.

_images/image5.jpeg

Figure 6 — Performing a process assessment for determining process capability

In this respect, a PRM or PAM further is not supposed to represent a product element hierarchy either.

4. Process reference model and performance indicators (Level 1)

The processes in the process dimension can be drawn from the Automotive SPICE process reference model, which is incorporated in the tables below indicated by a red bar at the left side.

Each table related to one process in the process dimension contains the process reference model (indicated by a red bar) and the process performance indicators necessary to define the process assessment model. The process performance indicators consist of base practices (indicated by a green bar) and output information items (indicated by a blue bar).

Table 21 — Template for the process description

Process reference

model

Process ID

Process name

Process purpose

Process outcomes

The individual processes are identified with a unique process identifier and a process name. A process purpose statement is provided, and process outcomes are defined to represent the process dimension of the Automotive SPICE process reference model. The background coloring of process ID’s and names are indicating the assignment to the corresponding process group.

Process performance indicators

Base Practices

A set of base practices for the process providing a definition of the activities to be performed to accomplish the process purpose and fulfill the process outcomes.

The base practice headers are summarized at the end of a process to demonstrate their relationship to the process outcomes.

Output information items

The output information items that are relevant to accomplish the process purpose and fulfill the process outcomes summarized at the end of a process to demonstrate their relationship to the process outcomes.

Note: Refer to Annex B for the characteristics of each information item.

4.1. Acquisition process group (ACQ)

4.1.1. ACQ.4 Supplier Monitoring

Need: ACQ.4 Supplier Monitoring ASPICE40_4_1_1
status: new

Process ID

ACQ.4

Process name

Supplier Monitoring

Process purpose

The purpose is to track and assess the performance of an external contract-based supplier company against agreed commitments.

Process outcomes

  1. Joint activities, as agreed between the customer and the supplier, are performed.

  2. All information, agreed upon for exchange, is communicated regularly between the customer and the supplier.

  3. Performance of the supplier is monitored against the agreements.

  4. Changes to the agreement, if needed, are negotiated between the customer and the supplier and documented in the agreement.

Base Practices

ACQ.4.BP1: Agree on and maintain joint activities, joint interfaces, and information to be exchanged. Establish and maintain an agreement on information to be exchanged, on joint activities, joint interfaces, responsibilities, type and frequency of joint activities, communications, meetings, status reports, and reviews.

ACQ.4.BP2: Exchange all agreed information. Use the defined joint interfaces between customer and supplier for the exchange of all agreed information.

ACQ.4.BP3: Review development work products with the supplier. Review development work products with the supplier on the agreed regular basis, covering technical aspects, problems and risks. Track open measures.

Note 1: see SUP.9 for management of problems

ACQ.4.BP4: Review progress of the supplier. Review progress of the supplier regarding schedule, quality, and cost on the agreed regular basis. Track open measures to closure and perform risk mitigation activities.

Note 2: see MAN.5 for management of risks

ACQ.4.BP5: Act to correct deviations. Take action when agreed objectives are not achieved. Negotiate changes to objectives and document them in the agreements.

ACQ.4 Supplier Monitoring

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Output Information Items

02-01 Commitment/Agreement

X

X

X

X

13-52 Communication evidence

X

X

X

13-09 Meeting support evidence

X

X

13-14 Progress status

X

X

13-16 Change request

X

13-19 Review evidence

X

14-02 Corrective action

X

15-51 Analysis results

X

Base Practices

BP1: Agree on and maintain joint processes, joint interfaces, and information to be exchanged

X

X

X

BP2: Exchange all agreed information

X

X

X

BP3: Review development work products with the supplier

X

X

X

BP4: Review progress of the supplier

X

X

X

BP5: Act to correct deviations

X

X

4.2. Supply process group (SPL)

4.2.1. SPL.2 Product Release

Need: SPL.2 Product Release ASPICE40_4_2_1
status: new

Process ID

SPL.2

Process name

Product Release

Process purpose

The purpose is to control the release of a product to the intended customer.

Process outcomes

  1. The contents of the product releases are determined.

  2. The release package is assembled from configured items.

  3. The release documentation is defined and produced.

  4. Release approval is performed against defined criteria.

  5. The release package is made available to the intended customer.

Base Practices

SPL.2.BP1: Define the functional content of releases. Define the functionality to be included and the release criteria for each release.

Note 1: This may include the hardware elements, software elements, and extra application parameter files (influencing the identified system functionality) that are needed for the release.

SPL.2.BP2: Define release package. Define the release as well as supporting tools and information.

Note 2: The release package may include also programming tools.

SPL.2.BP3: Ensure unique identification of releases. Ensure a unique identification of the release based upon the intended purpose and expectations of the release.

Note 3: Unique identification may be realized by a classification and numbering scheme for product releases.

SPL.2.BP4: Build the release from items under configuration control. Build the release from items under configuration control to ensure integrity.

Note 4: This practice may be supported by the SUP.8 Configuration Management Process.

SPL.2.BP5: Ensure release approval before delivery. Criteria for the release are satisfied before delivery takes place.

SPL.2.BP6: Provide a release note. A release is accompanied by information detailing key characteristics of the release.

Note 5: The release note may include information about legal aspects like relevant target markets, legislation that is considered etc. See also VAL.1 Validation.

SPL2.BP7: Communicate the type, service level and duration of support for a release.

Identify and communicate the type, service level and duration of support for a release.

SPL.2.BP8: Deliver the release package to the intended customer. Deliver the release package to the intended customer.

Note 6: The intended customer may be an internal organizational unit or an external organization.

SPL.2 Product Release

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Output Information Items

11-03 Release note

X

X

X

X

11-04 Product release package

X

X

13-06 Delivery evidence

X

X

13-13 Product release approval

X

X

18-06 Product release criteria

X

X

X

Base Practices

BP1: Define the functional content of releases

X

BP2: Define release package

X

BP3: Establish a product release classification and numbering scheme

X

BP4: Build the release from configured items

X

BP5: Ensure product release approval before delivery

X

BP6: Provide a release note

X

X

BP7: Communicate the type, service level and duration of support for a release

X

X

BP8: Deliver the release package to the intended customer

X

4.3. System engineering process group (SYS)

4.3.1. SYS.1 Requirements Elicitation

Need: SYS.1 Requirements Elicitation ASPICE40_4_3_1
status: new

Process ID

SYS.1

Process name

Requirements Elicitation

Process purpose

The purpose is to gather, analyze, and track evolving stakeholder needs and requirements throughout the lifecycle of the product and/or service to establish a set of agreed requirements.

Process outcomes

  1. Continuing communication with the stakeholder is established.

  2. Stakeholder expectations are understood, and requirements are defined and agreed.

  3. Stakeholder requirements changes arising from stakeholder needs are analyzed to enable associated risk assessment and impact management.

  4. Determination of stakeholder requirements status is ensured for all affected parties.

Base Practices

SYS.1.BP1: Obtain stakeholder expectations and requests. Obtain and define stakeholder expectations and requests through direct solicitation of stakeholder input, and through review of stakeholder business proposals (where relevant) and other documents containing inputs to stakeholder requirements, and consideration of the target operating and hardware environment.

Note 1: Documenting the stakeholder, or the source of a stakeholder requirement, supports stakeholder requirements agreement and change analysis (see BP2 and BP3).

SYS.1.BP2: Agree on requirements. Formalize the stakeholder’s expectations and requests into requirements. Reach a common understanding of the set of stakeholder requirements among affected parties by obtaining an explicit agreement from all affected parties.

Note 2: Examples of affected parties are customers, suppliers, design partners, joint venture partners, or outsourcing parties.

Note 3: The agreed stakeholder requirements may be based on feasibility studies and/or cost and schedule impact analysis.

SYS.1.BP3: Analyze stakeholder requirements changes. Analyze all changes made to the stakeholder requirements against the agreed stakeholder requirements. Assess the impact and risks, and initiate appropriate change control and mitigation actions.

Note 4: Requirements changes may arise from different sources as for instance changing technology, stakeholder needs, or legal constraints.

Note 5: Refer to SUP.10 Change Request Management, if required.

SYS.1.BP4: Communicate requirements status. Ensure all affected parties can be aware of the status and disposition of their requirements including changes and can communicate necessary information and data.

SYS.1 Requirements Elicitation

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Output Information Items

15-51 Analysis Results

X

13-52 Communication Evidence

X

X

17-00 Requirement

X

17-54 Requirement Attribute

X

X

X

Base Practices

BP1: Obtain stakeholder expectations and requests

X

BP2: Agree on requirements

X

BP3: Analyze stakeholder requirements changes

X

BP4: Communicate requirements status

X

X

4.3.2. SYS.2 System Requirements Analysis

Need: SYS.2 System Requirements Analysis ASPICE40_4_3_2
status: new

Process ID

SYS.2

Process name

System Requirements Analysis

Process purpose

The purpose is to establish a structured and analyzed set of system requirements consistent with the stakeholder requirements.

Process outcomes

  1. System requirements are specified.

  2. System requirements are structured and prioritized.

  3. System requirements are analyzed for correctness and technical feasibility.

  4. The impact of system requirements on the operating environment is analyzed.

  5. Consistency and bidirectional traceability are established between system requirements and stakeholder requirements.

  6. The system requirements are agreed and communicated to all affected parties.

Base Practices

SYS.2.BP1: Specify system requirements. Use the stakeholder requirements to identify and document the functional and non-functional requirements for the system according to defined characteristics for requirements.

Note 1: Characteristics of requirements are defined in standards such as ISO IEEE 29148, ISO 26262-8:2018, or the INCOSE Guide For Writing Requirements.

Note 2: Examples for defined characteristics of requirements shared by technical standards are verifiability (i.e., verification criteria being inherent in the requirements text), unambiguity/comprehensibility, freedom from design and implementation, and not contradicting any other requirement).

SYS.2.BP2: Structure system requirements. Structure and prioritize the system requirements.

Note 3: Examples for structuring criteria can be grouping (e.g., by functionality) or product variants identification.

Note 4: Prioritization can be done according to project or stakeholder needs via e.g., definition of release scopes. Please refer to SPL.2.BP1.

SYS.2.BP3: Analyze system requirements. Analyze the specified system requirements including their interdependencies to ensure correctness, technical feasibility, and to support project management regarding project estimates.

Note 5: See MAN.3.BP3 for project feasibility and MAN.3.BP5 for project estimates.

Note 6: Technical feasibility can be evaluated based on e.g., platform or product line, or by means of prototype development or product demonstrators.

SYS.2.BP4: Analyze the impact on the system context. Analyze the impact that the system requirements will have on elements in the relevant system context.

SYS.2.BP5: Ensure consistency and establish bidirectional traceability. Ensure consistency and establish bidirectional traceability between system requirements and stakeholder requirements.

Note 7: Bidirectional traceability supports consistency, facilitates impact analyses of change requests, and supports the demonstration of coverage of stakeholder requirements. Traceability alone, e.g., the existence of links, does not necessarily mean that the information is consistent with each other.

Note 8: There may be non-functional stakeholder requirements that the system requirements do not trace to. Examples are process requirements. Such stakeholder requirements are still subject to verification.

SYS.2.BP6: Communicate agreed system requirements and impact on the system context. Communicate the agreed system requirements, and results of the impact analysis on the system context, to all affected parties.

SYS.2 System Requirements Analysis

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Outcome 6

Output Information Items

17-00 Requirement

X

X

17-54 Requirement Attribute

X

X

15-51 Analysis Results

X

X

13-51 Consistency Evidence

X

13-52 Communication Evidence

X

Base Practices

BP1: Specify system requirements

X

BP2: Structure system requirements

X

BP3: Analyze system requirements

X

BP4: Analyze the impact on the system context

X

BP5: Ensure consistency and establish bidirectional traceability

X

BP6: Communicate agreed system requirements and impact on the system context

X

4.3.3. SYS.3 System Architectural Design

Need: SYS.3 System Architectural Design ASPICE40_4_3_3
status: new

Process ID

SYS.3

Process name

System Architectural Design

Process purpose

The purpose is to establish an analyzed system architecture, comprising static and dynamic aspects, consistent with the system requirements.

Process outcomes

  1. A system architecture is designed including a definition of the system elements with their behavior, their interfaces, their relationships, and their interactions.

  2. The system architecture is analyzed against defined criteria, and special characteristics are identified.

  3. Consistency and bidirectional traceability are established between system architecture and system requirements.

  4. The agreed system architecture and the special characteristics are communicated to all affected parties.

Base Practices

SYS.3.BP1: Specify static aspects of the system architecture. Specify and document the static aspects of the system architecture with respect to the functional and non-functional system requirements, including external interfaces and a defined set of system elements with their interfaces and relationships.

SYS.3.BP2: Specify dynamic aspects of the system architecture. Specify and document the dynamic aspects of the system architecture with respect to the functional and non-functional system requirements including the behavior of the system elements and their interaction in different system modes.

Note 1: Examples of interactions of system elements are timing diagrams reflecting inertia of mechanical components, processing times of ECUs, and signal propagation times of bus systems.

SYS.3.BP3: Analyze system architecture. Analyze the system architecture regarding relevant technical design aspects related to the product lifecycle, and to support project management regarding project estimates, and derive special characteristics for non-software system elements. Document a rationale for the system architectural design decisions.

Note 2: See MAN.3.BP3 for project feasibility and MAN.3.BP5 for project estimates.

Note 3: Examples for product lifecycle phases are production, maintenance & repair, decommissioning.

Note 4: Examples for technical aspects are manufacturability for production, suitability of pre-existing system elements to be reused, or availability of system elements.

Note 5: Examples for methods being suitable for analyzing technical aspects are prototypes, simulations, and qualitative analyses (e.g., FMEA approaches)

Note 6: Examples of design rationales are proven-in-use, reuse of a product platform or product line), a make-or-buy decision, or found in an evolutionary way (e.g., set-based design).

SYS.3.BP4: Ensure consistency and establish bidirectional traceability. Ensure consistency and establish bidirectional traceability between the elements of the system architecture and the system requirements that represent properties or characteristics of the physical end product.

Note 7: Bidirectional traceability further supports consistency, and facilitates impact analysis of change requests, and demonstration of verification coverage. Traceability alone, e.g., the existence of links, does not necessarily mean that the information is consistent with each other.

Note 8: There may be non-functional requirements that the system architectural design does not trace to. Examples are do not address, or represent, direct properties or characteristics of the physical end product. Such requirements are still subject to verification.

SYS.3.BP5: Communicate agreed system architecture. Communicate the agreed system architecture, including the special characteristics, to all affected parties.

SYS.3 System Architectural Design

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Output Information Items

04-06 System Architecture

X

13-51 Consistency Evidence

X

13-52 Communication Evidence

X

15-51 Analysis Results

X

17-57 Special Characteristics

X

Base Practices

BP1: Specify static aspects of system architecture

X

BP2: Specify dynamic aspects of system architecture

X

BP3: Analyze the system architecture

X

BP4: Ensure consistency and establish bidirectional traceability

X

BP5: Communicate agreed system architecture

X

4.3.4. SYS.4 System Integration and Integration Verification

Need: SYS.4 System Integration and Integration Verification ASPICE40_4_3_4
status: new

Process ID

SYS.4

Process name

System Integration and Integration Verification

Process purpose

The purpose is to integrate systems elements and verify that the integrated system elements are consistent with the system architecture.

Process outcomes

  1. Verification measures are specified for system integration verification of the integrated system elements based on the system architecture, including the interfaces of, and interactions between, system elements.

  2. System elements are integrated up to a complete integrated system consistent with the release scope.

  3. Verification measures are selected according to the release scope considering criteria, including criteria for regression verification.

  4. Integrated system elements are verified using the selected verification measures, and the results of the system integration verification are recorded.

  5. Consistency and bidirectional traceability are established between verification measures and the elements of the system architecture.

  6. Bidirectional traceability between verification results and verification measures is established.

  7. Results of the system integration and integration verification are summarized and communicated to all affected parties.

Base Practices

SYS.4.BP1: Specify verification measures for system integration. Specify the verification measures, based on a defined sequence and preconditions for the integration of system elements against the system static and dynamic aspects of the system architecture, including

  • techniques for the verification measures,

  • pass/fail criteria for verification measures,

  • a definition of entry and exit criteria for the verification measures, and

  • the required verification infrastructure and environment setup.

Note 1: Examples on what a verification measure may focus are the timing dependencies of the correct signal flow between interfacing system elements, or interactions between hardware and software, as specified in the system architecture. The system integration test cases may focus on

  • the correct signal flow between system items,

  • the timeliness and timing dependencies of signal flow between system items,

  • the correct interpretation of signals by all system items using an interface, and/or

  • the dynamic interaction between system items.

SYS.4.BP2: Select verification measures. Document the selection of verification measures for each integration step considering selection criteria including criteria for regression verification. The documented selection of verification measures shall have sufficient coverage according to the release scope.

Note 2: Examples for selection criteria can be prioritization of requirements, the need for regression verification (due to e.g., changes to the system architectural design or to system components), or the intended use of the delivered product release (e.g., test bench, test track, public road etc.)

SYS.4.BP3: Integrate system elements and perform integration verification. Integrate the system elements until the system is fully integrated according to the specified interfaces and interactions between the system elements, and according to the defined sequence and defined preconditions. Perform the selected system integration verification measures. Record the verification measure data including pass/fail status and corresponding verification measure data.

Note 3: Examples for preconditions for starting system integration can be successful system element verification or qualification of pre-existing system elements.

Note 4: See SUP.9 for handling verification results that deviate from expected results

SYS.4.BP4: Ensure consistency and establish bidirectional traceability. Ensure consistency and establish bidirectional traceability between verification measures and the system architecture. Establish bidirectional traceability between verification results and verification measures.

Note 5: Bidirectional traceability supports consistency, and facilitates impact analysis of change requests, and demonstration of verification coverage. Traceability alone, e.g., the existence of links, does not necessarily mean that the information is consistent with each other.

SYS.4.BP5: Summarize and communicate results. Summarize the system integration and integration verification results and communicate them to all affected parties.

Note 6: Providing all necessary information from the test case execution in a summary enables other parties to judge the consequences.

SYS.4 System Integration and Integration Verification

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Outcome 6

Outcome 7

Output Information Items

08-60 Verification Measure

X

06-50 Integration Sequence Instruction

X

03-50 Verification Measure Data

X

08-58 Verification Measure Selection Set

X

15-52 Verification Results

X

13-51 Consistency Evidence

X

X

13-52 Communication Evidence

X

11-06 Integrated System

X

Base Practices

BP1: Specify verification measures for system integration

X

BP2: Select verification measures

X

BP3: Integrate system elements and perform integration verification.

X

X

BP4: Ensure consistency and establish bidirectional traceability

X

X

BP5: Summarize and communicate results

X

4.3.5. SYS.5 System Verification

Need: SYS.5 System Verification ASPICE40_4_3_5
status: new

Process ID

SYS.5

Process name

System Verification

Process purpose

The purpose is to ensure that the system is verified to be consistent with the system requirements.

Process outcomes

  1. Verification measures are specified for system verification of the system based on the system requirements.

  2. Verification measures are selected according to the release scope considering criteria, including criteria for regression verification.

  3. The integrated system is verified using the selected verification measures and the results of system verification are recorded.

  4. Consistency and bidirectional traceability are established between verification measures and system requirements.

  5. Bidirectional traceability is established between verification results and verification measures.

  6. Verification results are summarized and communicated to all affected parties.

Base Practices

SYS.5.BP1: Specify verification measures for system verification. Specify the verification measures for system verification suitable to provide evidence for compliance with the functional and non-functional information in the system requirements, including

  • techniques for the verification measures,

  • pass/fail criteria for verification measures,

  • a definition of entry and exit criteria for the verification measures,

  • necessary sequence of verification measures, and

  • the required verification infrastructure and environment setup.

Note 1: The system verification measures may cover aspects such as thermal, environmental, robustness/lifetime, and EMC.

SYS.5.BP2: Select verification measures. Document the selection of verification measures considering selection criteria including criteria for regression verification. The selection of verification measures shall have sufficient coverage according to the release scope.

Note 2: Examples for criteria for selection can be prioritization of requirements, the need for regression verification (due to e.g., changes to the system requirements), the intended use of the delivered product release (test bench, test track, public road etc.)

SYS.5.BP3: Perform verification of the integrated system. Perform the verification of the integrated system using the selected verification measures. Record the verification results including pass/fail status and corresponding verification measure data.

Note 3: See SUP.9 for handling verification results that deviate from expected results

SYS.5.BP4: Ensure consistency and establish bidirectional traceability. Ensure consistency and establish bidirectional traceability between verification measures and system requirements. Establish bidirectional traceability between verification results and verification measures.

Note 4: Bidirectional traceability supports consistency, and facilitates impact analysis of change requests, and demonstration of verification coverage. Traceability alone, e.g., the existence of links, does not necessarily mean that the information is consistent with each other.

SYS.5.BP5: Summarize and communicate results. Summarize the system verification results and communicate them to all affected parties.

Note 5: Providing all necessary information from the test case execution in a summary enables other parties to judge the consequences.

SYS.5 System Verification

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Outcome 6

Output Information Item

08-60 Verification Measure

X

03-50 Verification Measure Data

X

08-58 Verification Measure Selection Set

X

15-52 Verification Results

X

13-51 Consistency Evidence

X

X

13-52 Communication Evidence

X

Base Practices

BP1: Specify verification measures for system verification

X

BP2: Select verification measures

X

BP3: Perform verification of the integrated system

X

BP4: Ensure consistency and establish bidirectional traceability.

X

X

BP5: Summarize and communicate results

X

4.4. Software engineering process group (SWE)

4.4.1. SWE.1 Software Requirements Analysis

Need: SWE.1 Software Requirements Analysis ASPICE40_4_4_1
status: new

Process ID

SWE.1

Process name

Software Requirements Analysis

Process purpose

The purpose is to establish a structured and analyzed set of software requirements consistent with the system requirements, and the system architecture.

Process outcomes

  1. Software requirements are specified.

  2. Software requirements are structured and prioritized.

  3. Software requirements are analyzed for correctness and technical feasibility.

  4. The impact of software requirements on the operating environment is analyzed.

  5. Consistency and bidirectional traceability are established between software requirements and system requirements.

  6. Consistency and bidirectional traceability are established between software requirements and system architecture.

  7. The software requirements are agreed and communicated to all affected parties.

Base Practices

SWE.1.BP1: Specify software requirements. Use the system requirements and the system architecture to identify and document the functional and non-functional requirements for the software according to defined characteristics for requirements.

Note 1: Characteristics of requirements are defined in standards such as ISO IEEE 29148, ISO 26262-8:2018, or the INCOSE Guide for Writing Requirements.

Note 2: Examples for defined characteristics of requirements shared by technical standards are verifiability (i.e., verification criteria being inherent in the requirements text), unambiguity/comprehensibility, freedom from design and implementation, and not contradicting any other requirement).

Note 3: In case of software-only development, the system requirements and the system architecture refer to a given operating environment. In that case, stakeholder requirements can be used as the basis for identifying the required functions and capabilities of the software.

Note 4: The hardware-software-interface (HSI) definition puts in context hardware and therefore it is an interface decision at the system design level. If such a HSI exists, then it may provide input to software requirements.

SWE.1.BP2: Structure software requirements. Structure and prioritize the software requirements.

Note 5: Examples for structuring criteria can be grouping (e.g., by functionality) or expressing product variants.

Note 6: Prioritization can be done according to project or stakeholder needs via e.g., definition of release scopes. Refer to SPL.2.BP1.

SWE.1.BP3: Analyze software requirements. Analyze the specified software requirements including their interdependencies to ensure correctness, technical feasibility, and to support project management regarding project estimates.

Note 7: See MAN.3.BP3 for project feasibility and MAN.3.BP5 for project estimates.

Note 8: Technical feasibility can be evaluated based on e.g., platform or product line, or by prototyping.

SWE.1.BP4: Analyze the impact on the operating environment. Analyze the impact that the software requirements will have on elements in the operating environment.

SWE.1.BP5: Ensure consistency and establish bidirectional traceability. Ensure consistency and establish bidirectional traceability between software requirements and system architecture. Ensure consistency and establish bidirectional traceability between software requirements and system requirements.

Note 9: Redundant traceability is not intended.

Note 10: There may be non-functional system requirements that the software requirements do not trace to. Examples are process requirements or requirements related to later software product lifecycle phases such as incident handling. Such requirements are still subject to verification.

Note 11: Bidirectional traceability supports consistency, and facilitates impact analysis of change requests, and demonstration of verification coverage. Traceability alone, e.g., the existence of links, does not necessarily mean that the information is consistent with each other.

Note 12: In case of software development only, the system requirements and system architecture refer to a given operating environment. In that case, consistency and bidirectional traceability can be ensured between stakeholder requirements and software requirements.

SWE.1.BP6: Communicate agreed software requirements and impact on the operating environment. Communicate the agreed software requirements, and the results of the analysis of impact on the operating environment, to all affected parties.

SWE.1 Software Requirements Analysis

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Outcome 6

Outcome 7

Output Information Items

17-00 Requirement

X

X

17-54 Requirement Attribute

X

15-51 Analysis Results

X

X

13-51 Consistency Evidence

X

X

13-52 Communication Evidence

X

Base Practices

BP1: Specify software requirements

X

BP2: Structure software requirements

X

BP3: Analyze software requirements

X

BP4: Analyze the impact on the operating environment

X

BP5: Ensure consistency and establish bidirectional traceability

X

X

BP6: Communicate agreed software requirements and impact on the operating environment

X

4.4.2. SWE.2 Software Architectural Design

Need: SWE.2 Software Architectural Design ASPICE40_4_4_2
status: new

Process ID

SWE.2

Process name

Software Architectural Design

Process purpose

The purpose is to establish an analyzed software architecture, comprising static and dynamic aspects, consistent with the software requirements.

Process outcomes

  1. A software architecture is designed including static and dynamic aspects.

  2. The software architecture is analyzed against defined criteria.

  3. Consistency and bidirectional traceability are established between software architecture and software requirements.

  4. The software architecture is agreed and communicated to all affected parties.

Base Practices

SWE.2.BP1: Specify static aspects of the software architecture. Specify and document the static aspects of the software architecture with respect to the functional and non-functional software requirements, including external interfaces and a defined set of software components with their interfaces and relationships.

Note 1: The hardware-software-interface (HSI) definition puts in context the hardware design and therefore is an aspect of system design (SYS.3).

SWE.2.BP2: Specify dynamic aspects of the software architecture. Specify and document the dynamic aspects of the software architecture with respect to the functional and nonfunctional software requirements, including the behavior of the software components and their interaction in different software modes, and concurrency aspects.

Note 2: Examples for concurrency aspects are application-relevant interrupt handling, preemptive processing, multi-threading.

Note 3: Examples for behavioral descriptions are natural language or semi-formal notation (e.g, SysML, UML).

SWE.2.BP3: Analyze software architecture. Analyze the software architecture regarding relevant technical design aspects and to support project management regarding project estimates. Document a rationale for the software architectural design decision.

Note 4: See MAN.3.BP3 for project feasibility and MAN.3.BP5 for project estimates.

Note 5: The analysis may include the suitability of pre-existing software components for the current application.

Note 6: Examples of methods suitable for analyzing technical aspects are prototypes, simulations, qualitative analyses.

Note 7: Examples of technical aspects are functionality, timings, and resource consumption (e.g, ROM, RAM, external / internal EEPROM or Data Flash or CPU load).

Note 8: Design rationales can include arguments such as proven-in-use, reuse of a software framework or software product line, a make-or-buy decision, or found in an evolutionary way (e.g, set-based design).

SWE.2.BP4: Ensure consistency and establish bidirectional traceability. Ensure consistency and establish bidirectional traceability between the software architecture and the software requirements.

Note 9: There may be non-functional software requirements that the software architectural design does not trace to. Examples are development process requirements. Such requirements are still subject to verification.

Note 10: Bidirectional traceability supports consistency, and facilitates impact analysis of change requests, and demonstration of verification coverage. Traceability alone, e.g, the existence of links, does not necessarily mean that the information is consistent with each other.

SWE.2.BP5: Communicate agreed software architecture. Communicate the agreed software architecture to all affected parties.

SWE.2 Software Architectural Design

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Output Information Items

04-04 Software Architecture

X

13-51 Consistency Evidence

X

13-52 Communication Evidence

X

15-51 Analysis Results

X

Base Practices

BP1: Specify static aspects of software architecture

X

BP2: Specify dynamic aspects of software architecture

X

BP3: Analyze software architecture

X

BP4: Ensure consistency and establish bidirectional traceability

X

BP5: Communicate agreed software architecture

X

4.4.3. SWE.3 Software Detailed Design and Unit Construction

Need: SWE.3 Software Detailed Design and Unit Construction ASPICE40_4_4_3
status: new

Process ID

SWE.3

Process name

Software Detailed Design and Unit Construction

Process purpose

The purpose is to establish a software detailed design, comprising static and dynamic aspects, consistent with the software architecture, and to construct software units consistent with the software detailed design.

Process outcomes

  1. A detailed design is specified including static and dynamic aspects.

  2. Software units as specified in the software detailed design are produced.

  3. Consistency and bidirectional traceability are established between software detailed design and software architecture; and consistency and bidirectional traceability are established between source code and software detailed design; and consistency and bidirectional traceability are established between the software detailed design and the software requirements.

  4. The source code and the agreed software detailed design are communicated to all affected parties.

Base Practices

SWE.3.BP1: Specify the static aspects of the detailed design. For each software component specify the behavior of its software units, their static structure and relationships, their interfaces including

  • valid data value ranges for inputs and outputs (from the application domain perspective), and

  • physical or measurement units applicable to inputs and outputs (from the application domain perspective).

Note 1: The boundary of a software unit is independent from the software unit’s representation in the source code, code file structure, or model-based implementation, respectively. It is rather driven by the semantics of the application domain perspective. Therefore, a software unit may be, at the code level, represented by a single subroutine or a set of subroutines.

Note 2: Examples of valid data value ranges with applicable physical units from the application domain perspective are ‘0..200 [m/s]’, ‘0..3.8 [A]’ or ‘1..100 [N]’. For mapping such application domain value ranges to programming language-level data types (such as unsigned Integer with a value range of 0..65535) refer to BP2.

Note 3: Examples of a measurement unit are ‘%’ or ‘‰’.

Note 4: A counter is an example of a parameter, or a return value, to which neither a physical nor a measurement unit is applicable.

Note 5: The hardware-software-interface (HSI) definition puts in context the hardware design and therefore is an aspect of system design (SYS.3).

SWE.3.BP2: Specify dynamic aspects of the detailed design. Specify and document the dynamic aspects of the detailed design with respect to the software architecture, including the interactions between relevant software units to fulfill the component’s dynamic behavior.

Note 6: Examples for behavioral descriptions are natural language or semi-formal notation (e.g, SysML, UML).

SWE.3.BP3: Develop software units. Develop and document the software units consistent with the detailed design, and according to coding principles.

Note 7: Examples for coding principles at capability level 1 are not to use implicit type conversions, only one entry and one exit point in subroutines, and range checks (design-by-contract, defensive programming). Further examples see e.g, ISO 26262-6 clause 8.4.5 together with table 6.

SWE.3.BP4: Ensure consistency and establish bidirectional traceability. Ensure consistency and establish bidirectional traceability between the software detailed design and the software architecture. Ensure consistency and establish bidirectional traceability between the developed software units and the software detailed design. Ensure consistency and establish traceability between the software detailed design and the software requirements.

Note 8: Redundancy should be avoided by establishing a combination of these approaches.

Note 9: Examples for tracing a software unit in the detailed design to a software requirement directly are communication matrices or basis software aspects such as a list of diagnosis identifiers inherent in an Autosar configuration.

Note 10: Bidirectional traceability supports consistency, and facilitates impact analysis of change requests, and demonstration of verification coverage. Traceability alone, e.g., the existence of links, does not necessarily mean that the information is consistent with each other.

SWE.3.BP5: Communicate agreed software detailed design and developed software units. Communicate the agreed software detailed design and developed software units to all affected parties.

SWE.3 Software Detailed Design and Unit Construction

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Output Information Items

04-05 Software Detailed Design

X

11-05 Software Unit

X

X

13-51 Consistency Evidence

X

13-52 Communication Evidence

X

Base Practices

BP1: Specify the static aspects of the detailed design

X

BP2: Specify the dynamic aspects of the detailed design

X

BP3: Develop software units

X

BP4: Ensure consistency and establish bidirectional traceability

X

BP5: Communicate agreed software detailed design and developed software units

X

4.4.4. SWE.4 Software Unit Verification

Need: SWE.4 Software Unit Verification ASPICE40_4_4_4
status: new

Process ID

SWE.4

Process name

Software Unit Verification

Process purpose

The purpose is to verify that software units are consistent with the software detailed design.

Process outcomes

  1. Verification measures for software unit verification are specified.

  2. Software unit verification measures are selected according to the release scope, including criteria for regression verification.

  3. Software units are verified using the selected verification measures, and results are recorded.

  4. Consistency and bidirectional traceability are established between verification measures and software units; and bidirectional traceability is established between verification results and verification measures.

  5. Results of the software unit verification are summarized and communicated to all affected parties.

Base Practices

SWE.4.BP1: Specify software unit verification measures. Specify verification measures for each software unit defined in the software detailed design, including

  • pass/fail criteria for verification measures,

  • entry and exit criteria for verification measures, and

  • the required verification infrastructure.

Note 1: Examples for unit verification measures are static analysis, code reviews, and unit testing.

Note 2: Static analysis can be done based on MISRA rulesets and other coding standards.

SWE.4.BP2: Select software unit verification measures. Document the selection of verification measures considering selection criteria including criteria for regression verification. The documented selection of verification measures shall have sufficient coverage according to the release scope.

SWE.4.BP3: Verify software units. Perform software unit verification using the selected verification measures. Record the verification results including pass/fail status and corresponding verification measure data.

Note 3: See SUP.9 for handling of verification results that deviate from expected results.

SWE.4.BP4: Ensure consistency and establish bidirectional traceability. Ensure consistency and establish bidirectional traceability between verification measures and the software units defined in the detailed design. Establish bidirectional traceability between the verification results and the verification measures.

Note 4: Bidirectional traceability supports consistency, and facilitates impact analysis of change requests, and demonstration of verification coverage. Traceability alone, e.g., the existence of links, does not necessarily mean that the information is consistent with each other.

SWE.4.BP5: Summarize and communicate results. Summarize the results of software unit verification and communicate them to all affected parties.

Note 5: Providing all necessary information from the test case execution in a summary enables other parties to judge the consequences.

SWE.4 Software Unit Verification

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Output Information Items

08-60 Verification Measure

X

03-50 Verification Measure Data

X

08-58 Verification Measure Selection Set

X

15-52 Verification Results

X

13-51 Consistency Evidence

X

13-52 Communication Evidence

X

Base Practices

BP1: Specify software unit verification measures

X

BP2: Select software unit verification measures

X

BP3: Verify software units

X

BP4: Ensure consistency and establish bidirectional traceability for software unit verification

X

BP5: Summarize and communicate results

X

4.4.5. SWE.5 Software Component Verification and Integration Verification

Need: SWE.5 Software Component Verification and Integration Verification ASPICE40_4_4_5
status: new

Process ID

SWE.5

Process name

Software Component Verification and Integration Verification

Process purpose

The purpose is to verify that software components are consistent with the software architectural design, and to integrate software elements and verify that the integrated software elements are consistent with the software architecture and software detailed design.

Process outcomes

  1. Verification measures are specified for software integration verification of the integrated software elements based on the software architecture and detailed design, including the interfaces of, and interactions between, the software components.

  2. Verification measures for software components are specified to provide evidence for compliance of the software components with the software components’ behavior and interfaces.

  3. Software elements are integrated up to a complete integrated software.

  4. Verification measures are selected according to the release scope considering criteria, including criteria for regression verification.

  5. Software components are verified using the selected verification measures, and the results of the integration verification are recorded.

  6. Integrated software elements are verified using the selected verification measures, and the results of the integration verification are recorded.

  7. Consistency and bidirectional traceability are established between verification measures and the software architecture and detailed design; and bidirectional traceability is established between verification results and verification measures.

  8. The results of software component verification and software elements integration verification are summarized and communicated to all affected parties

Base Practices

SWE.5.BP1: Specify software integration verification measures. Specify verification measures, based on a defined sequence and preconditions for the integration of software elements, against the defined static and dynamic aspects of the software architecture, including

  • techniques for the verification measures,

  • pass/fail criteria for verification measures,

  • entry and exit criteria for verification measures, and

  • the required verification infrastructure and environment setup.

Note 1: Examples on which the software integration verification measures may focus on are the correct dataflow and dynamic interaction between software components together with their timing dependencies, the correct interpretation of data by all software components using an interface, and the compliance to resource consumption objectives.

Note 2: The software integration verification measure may be supported by using hardware debug interfaces or simulation environments (e.g, Software-in-the-Loop-Simulation).

SWE.5.BP2: Specify verification measures for verifying software component behavior. Specify verification measures for software component verification against the defined software components’ behavior and their interfaces in the software architecture, including

  • techniques for the verification measures,

  • entry and exit criteria for verification measures,

  • pass/fail criteria for verification measures, and

  • the required verification infrastructure and environment setup.

Note 3: Verification measures are related to software components but not to the software units since software unit verification is addressed in the process SWE.4 Software Unit Verification.

SWE.5.BP3: Select verification measures. Document the selection of integration verification measures for each integration step considering selection criteria including criteria for regression verification. The documented selection of verification measures shall have sufficient coverage according to the release scope.

Note 4: Examples for selection criteria can be the need for continuous integration /continuous development regression verification (due to e.g, changes to the software architectural or detailed design), or the intended use of the delivered product release (e.g, test bench, test track, public road etc.).

SWE.5.BP4: Integrate software elements and perform integration verification. Integrate the software elements until the software is fully integrated according to the specified interfaces and interactions between the Software elements, and according to the defined sequence and defined preconditions. Perform the selected integration verification measures. Record the verification measure data including pass/fail status and corresponding verification measure data.

Note 5: Examples for preconditions for starting software integration are qualification of pre-existing software components, off-the-shelf software components, open-source-software, or auto-code generated software.

Note 6: Defined preconditions may allow e.g, big-bang-integration of all software components, continuous integration, as well as stepwise integration (e.g, across software units and/or software components up to the fully integrated software) with accompanying verification measures.

Note 7: See SUP.9 for handling deviations of verification results deviate expected results.

SWE.5.BP5: Perform software component verification. Perform the selected verification measures for verifying software component behavior. Record the verification results including pass/fail status and corresponding verification measure data.

Note 8: See SUP.9 for handling verification results that deviate from expected results.

SWE.5.BP6: Ensure consistency and establish bidirectional traceability. Ensure consistency and establish bidirectional traceability between verification measures and the static and dynamic aspects of the software architecture and detailed design. Establish bidirectional traceability between verification results and verification measures.

Note 9: Bidirectional traceability supports consistency, and facilitates impact analysis of change requests, and demonstration of verification coverage. Traceability alone, e.g., the existence of links, does not necessarily mean that the information is consistent with each other.

SWE.5.BP7: Summarize and communicate results. Summarize the software component verification and the software integration verification results and communicate them to all affected parties.

Note 10: Providing all necessary information from the test case execution in a summary enables other parties to judge the consequences.

SWE.5 Software Component Verification and

Integration Verification

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Outcome 6

Outcome 7

Outcome 8

Output Information Items

08-60 Verification Measure

X

X

06-50 Integration Sequence Instruction

X

03-50 Verification Measure Data

X

08-58 Verification Measure Selection Set

X

15-52 Verification Results

X

X

13-51 Consistency Evidence

X

13-52 Communication Evidence

X

01-03 Software Component

X

01-50 Integrated Software

X

Base Practices

BP1: Specify software integration verification measures

X

BP2: Specify verification measures for verifying software component behavior

X

BP3: Select verification measures

X

BP4: Integrate software elements and perform integration verification

X

X

BP5: Perform software component verification

X

BP6: Ensure consistency and establish bidirectional traceability

X

BP7: Summarize and communicate results

X

4.4.6. SWE.6 Software Verification

Need: SWE.6 Software Verification ASPICE40_4_4_6
status: new

Process ID

SWE.6

Process name

Software Verification

Process purpose

The purpose of the Software Verification process is to ensure that the integrated software is verified to be consistent with the software requirements.

Process outcomes

  1. Verification measures are specified for software verification of the software based on the software requirements.

  2. Verification measures are selected according to the release scope considering criteria, including criteria for regression verification.

  3. The integrated software is verified using the selected verification measures and the results of software verification are recorded.

  4. Consistency and bidirectional traceability are established between verification measures and software requirements; and bidirectional traceability is established between verification results and verification measures.

  5. Results of the software verification are summarized and communicated to all affected parties.

Base Practices

SWE.6.BP1: Specify verification measures for software verification. Specify the verification measures for software verification suitable to provide evidence for compliance of the integrated software with the functional and non-functional information in the software requirements, including

  • techniques for the verification measures,

  • pass/fail criteria for verification measures,

  • a definition of entry and exit criteria for the verification measures,

  • necessary sequence of verification measures, and

  • the required verification infrastructure and environment setup.

Note 1: The selection of appropriate techniques for verification measures may depend on the content of the respective software requirement (e.g, boundary values and equivalence classes for data range-oriented requirements, positive/sunny-day-test vs. negative testing such as fault injection), or on requirements-based testing vs. “error guessing based on knowledge or experience”.

SWE.6.BP2: Select verification measures. Document the selection of verification measures considering selection criteria including criteria for regression verification. The documented selection of verification measures shall have sufficient coverage according to the release scope.

Note 2: Examples for selection criteria can be prioritization of requirements, continuous development, the need for regression verification (due to e.g., changes to the software requirements), or the intended use of the delivered product release (test bench, test track, public road etc.)

SWE.6.BP3: Verify the integrated software. Perform the verification of the integrated software using the selected verification measures. Record the verification results including pass/fail status and corresponding verification measure data.

Note 3: See SUP.9 for handling verification results that deviate from expected results.

SWE.6.BP4: Ensure consistency and establish bidirectional traceability. Ensure consistency and establish bidirectional traceability between verification measures and software requirements. Establish bidirectional traceability between verification results and verification measures.

Note 4: Bidirectional traceability supports consistency, and facilitates impact analysis of change requests, and demonstration of verification coverage. Traceability alone, e.g., the existence of links, does not necessarily mean that the information is consistent with each other.

SWE.6.BP5: Summarize and communicate results. Summarize the software verification results and communicate them to all affected parties.

Note 5: Providing all necessary information from the test case execution in a summary enables other parties to judge the consequences.

SWE.6 Software Verification

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Output Information Items

08-60 Verification Measure

X

03-50 Verification Measure Data

X

08-58 Verification Measure Selection Set

X

15-52 Verification Results

X

13-51 Consistency Evidence

X

13-52 Communication Evidence

X

Base Practices

BP1: Specify verification measures for software verification

X

BP2: Select verification measures

X

BP3: Verify the integrated software

X

BP4: Ensure consistency and establish bidirectional traceability.

X

BP5: Summarize and communicate results

X

4.5. Validation process group (VAL)

4.5.1. VAL.1 Validation

Need: VAL.1 Validation ASPICE40_4_5_1
status: new

Process ID

VAL.1

Process name

Validation

Process purpose

The purpose is to provide evidence that the end product, allowing direct end user interaction, satisfies the intended use expectations in its operational target environment.

Process outcomes

  1. Validation measures are selected considering criteria for regression verification.

  2. The product is validated using the selected validation measures and the results of validation are recorded.

  3. Consistency and unidirectional traceability are established between validation measures and stakeholder requirements; and consistency and bidirectional traceability is established between validation results and validation measures.

  4. Results of the validation are summarized and communicated to all affected parties.

Base Practices

VAL.1.BP1: Specify validation measures for product validation. Specify the validation measures for the end product based on the stakeholder requirements to provide evidence that it fulfills its intended use expectations in its operational target environment, and

  • techniques for the validation measures,

  • pass/fail criteria for validation measures,

  • a definition of entry and exit criteria for the validation measures,

  • necessary sequence of validation measures, and

  • the required validation infrastructure and environment setup.

Note 1: An example for validation-relevant stakeholder requirements are homologation or legal type approval requirements. Further examples of sources of intended use expectations are technical risks (see MAN.5, SYS.3.BP4, SWE.2.BP3, HWE.2.BP6).

Note 2: Where stakeholder requirements cannot be specified comprehensively or change frequently, repeated validation of (often rapidly developed) increments in product evolution may be employed to refine stakeholder requirements, and to mitigate risks in the correct identification of needs.

Note 3: Validation may also be conducted to confirm that the product also satisfies the often less formally expressed, but sometimes overriding, attitudes, experience, and subjective tests that comprise stakeholder or end user satisfaction.

VAL.1.BP2: Select validation measures. Document the selection of validation measures considering selection criteria including criteria for regression validation. The documented selection of validation measures shall have sufficient coverage according to the release scope.

Note 4: Examples for criteria for selection can be the release purpose of the delivered product (such as test bench, test track, validation on public roads, field use by end users), homologation/ type approval, confirmation of requirements, or the need for regression due to e.g., changes to stakeholder requirements and needs.

VAL.1.BP3: Perform validation and evaluate results. Perform the validation of the integrated end product using the selected validation measures. Record the validation results including pass/fail status. Evaluate the validation results.

Note 5: Validation results can be used as a means for identifying stakeholder or system requirements e.g, in the case of mock-ups or concept studies.

Note 6: See SUP.9 for handling verification results that deviate from expected results

VAL.1.BP4: Ensure consistency and establish bidirectional traceability. Ensure consistency and establish bidirectional traceability from validation measures to the stakeholder requirements from which they are derived. Establish bidirectional traceability between validation results and validation measures.

Note 7: Examples of sources of validation measures from which they can be derived are legal requirements, homologation requirements, results of technical risk analyses, or stakeholder and system requirements (see SYS.1 and SYS.2).

Note 8: If sources of validation measures are e.g., legal or homologation requirements, then direct bidirectional traceability from those sources to the validation measures are not possible. In such a case, unidirectional traceability is sufficient.

Note 9: Bidirectional traceability supports consistency, and facilitates impact analyses of change requests, and demonstration of verification coverage. Traceability alone, e.g., the existence of links, does not necessarily mean that the information is consistent with each other.

VAL.1.BP5: Summarize and communicate results. Summarize the validation results and communicate them to all affected parties.

Note 10: Providing all necessary information from the test case execution in a summary enables other parties to judge the consequences.

VAL.1 Validation

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Output Information Items

08-59 Validation Measure

X

08-57 Validation Measure Selection Set

X

13-24 Validation Results

X

13-51 Consistency Evidence

X

13-52 Communication Evidence

X

Base Practices

BP1: Specify validation measures

X

BP2: Select validation measures

X

BP3: Perform validation and evaluate results

X

BP4: Ensure consistency and establish traceability.

X

BP5: Summarize and communicate results

X

4.6. Machine Learning Engineering process group (MLE)

4.6.1. MLE.1 Machine Learning Requirements Analysis

Need: MLE.1 Machine Learning Requirements Analysis ASPICE40_4_6_1
status: new

Process ID

MLE.1

Process name

Machine Learning Requirements Analysis

Process purpose

The purpose is to refine the machine learning-related software requirements into a set of ML requirements.

Process outcomes

  1. The ML requirements including ML data requirements are identified and specified based on the software requirements and the components of the software architecture.

  2. ML requirements are structured and prioritized.

  3. ML requirements are analyzed for correctness and verifiability.

  4. The impact of ML requirements on the ML operating environment is analyzed.

  5. Consistency and bidirectional traceability are established between ML requirements and software requirements, and between ML requirements and software architecture.

  6. The ML requirements are agreed and communicated to all affected parties.

Base Practices

MLE.1.BP1: Specify ML requirements. Use the software requirements and the software architecture to identify and specify functional and non-functional ML requirements, as well as ML data requirements specifying data characteristics (e.g., gender, weather conditions, street conditions within the ODD) and their expected distributions.

Note 1: Non-functional requirements may include relevant characteristics of the ODD and KPIs as robustness, performance, and level of trustworthiness.

Note 2: The ML data requirements are input for SUP.11 Machine Learning Data Management but also for other MLE processes.

Note 3: In case of ML development only, stakeholder requirements represent the software requirements.

MLE.1.BP2: Structure ML requirements. Structure and prioritize the ML requirements.

Note 4: Examples for structuring criteria can be grouping (e.g., by functionality) or variants identification.

Note 5: Prioritization can be done according to project or stakeholder needs via e.g., definition of release scopes. Refer to SPL.2.BP1.

MLE.1.BP3: Analyze ML requirements. Analyze the specified ML requirements including their interdependencies to ensure correctness, technical feasibility, and ability for machine learning model testing, and to support project management regarding project estimates.

Note 6: See MAN.3.BP3 for project feasibility and MAN.3.BP5 for project estimates.

MLE.1.BP4: Analyze the impact on the ML operating environment. Analyze the impact that the ML requirements will have on interfaces of software components and the ML operating environment.

Note 7: The ML operating environment is defined as the infrastructure and information which both the trained ML model and the deployed ML model need for execution.

MLE.1.BP5: Ensure consistency and establish bidirectional traceability. Ensure consistency and establish bidirectional traceability between ML requirements and software requirements and between ML requirements and the software architecture.

Note 8: Bidirectional traceability supports consistency, facilitates impact analyses of change requests, and verification coverage demonstration. Traceability alone, e.g., the existence of links, does not necessarily mean that the information is consistent with each other.

Note 9: Redundant traceability is not intended, but at least one out of the given traceability paths.

MLE.1.BP6: Communicate agreed ML requirements and impact on the operating environment. Communicate the agreed ML requirements, and the results of the impact analysis on the ML operating environment to all affected parties.

MLE.1 Machine Learning Requirements Analysis

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Outcome 6

Output Information Items

17-00 Requirement

X

X

17-54 Requirement attribute

X

X

13-52 Communication evidence

X

13-51 Consistency evidence

X

15-51 Analysis results

X

X

Base Practices

BP1: Specify ML requirements

X

BP2: Structure ML requirements

X

BP3: Analyze ML requirements

X

BP4: Analyze the impact on the ML operating environment

X

BP5: Ensure consistency and establish bidirectional traceability

X

BP6: Communicate agreed ML requirements

X

4.6.2. MLE.2 Machine Learning Architecture

Need: MLE.2 Machine Learning Architecture ASPICE40_4_6_2
status: new

Process ID

MLE.2

Process name

Machine Learning Architecture

Process purpose

The purpose is to establish an ML architecture supporting training and deployment, consistent with the ML requirements, and to evaluate the ML architecture against defined criteria.

Process outcomes

  1. A ML architecture is developed.

  2. Hyperparameter ranges and initial values are determined as a basis for the training.

  3. Evaluation of ML architectural elements is conducted.

  4. Interfaces of the ML architectural elements are defined.

  5. Resource consumption objectives for the ML architectural elements are defined.

  6. Consistency and bidirectional traceability are established between the ML architectural elements and the ML requirements.

  7. The ML architecture is agreed and communicated to all affected parties.

Base Practices

MLE.2.BP1: Develop ML architecture. Develop and document the ML architecture that specifies ML architectural elements including details of the ML model, pre- and postprocessing, and hyperparameters which are required to create, train, test, and deploy the ML model.

Note 1: Necessary details of the ML model may include layers, activation functions, and backpropagation. The level of detail of the ML model may not need to cover aspects like single neurons.

Note 2: The details of the ML model may differ between the ML model used during training and the deployed ML model.

MLE.2.BP2: Determine hyperparameter ranges and initial values. Determine and document the hyperparameter ranges and the initial values as a basis for the training.

MLE.2.BP3: Analyze ML architectural elements. Define criteria for analysis of the ML architectural elements. Analyze ML architectural elements according to the defined criteria.

Note 3: Trustworthiness and explainability might be criteria for the analysis of the ML architectural elements.

MLE.2.BP4: Define interfaces of the ML architectural elements. Determine and document the internal and external interfaces of each ML architectural element including its interfaces to related software components.

MLE.2.BP5: Define resource consumption objectives for the ML architectural elements. Determine and document the resource consumption objectives for all relevant ML architectural elements during training and deployment.

MLE.2.BP6: Ensure consistency and establish bidirectional traceability. Ensure consistency and establish bidirectional traceability between the ML architectural elements and the ML requirements.

Note 4: Bidirectional traceability supports consistency, and facilitates impact analyses of change requests, and verification coverage demonstration. Traceability alone, e.g., the existence of links, does not necessarily mean that the information is consistent with each other.

Note 5: The bidirectional traceability should be established on a reasonable level of abstraction to the ML architectural elements.

MLE.2.BP7: Communicate agreed ML architecture. Inform all affected parties about the agreed ML architecture including the details of the ML model and the initial hyperparameter values.

MLE.2 Machine Learning Architecture

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Outcome 6

Outcome 7

Output Information Items

04-51 ML architecture

X

X

X

X

X

13-52 Communication evidence

X

13-51 Consistency evidence

X

01-54 Hyperparameter

X

X

15-51 Analysis results

X

X

Base Practices

BP1: Develop ML architecture

X

BP2: Determine hyperparameter ranges and initial values.

X

BP3: Evaluate ML architectural elements

X

BP4: Define interfaces of the ML architectural elements

X

BP5: Define resource consumption objectives for the ML architectural elements

X

BP6: Ensure consistency and establish bidirectional traceability

X

BP7: Communicate agreed ML architecture

X

4.6.3. MLE.3 Machine Learning Training

Need: MLE.3 Machine Learning Training ASPICE40_4_6_3
status: new

Process ID

MLE.3

Process name

Machine Learning Training

Process purpose

The purpose is to optimize the ML model to meet the defined ML requirements.

Process outcomes

  1. A ML training and validation approach is specified.

  2. The data set for ML training and ML validation is created.

  3. The ML model, including hyperparameter values, is optimized to meet the defined ML requirements.

  4. Consistency and bidirectional traceability are established between the ML training and validation data set and the ML data requirements.

  5. Results of optimization are summarized, and the trained ML model is agreed and communicated to all affected parties.

Base Practices

MLE.3.BP1: Specify ML training and validation approach. Specify an approach which supports the training and validation of the ML model to meet the defined ML requirements. The ML training and validation approach includes

  • entry and exit criteria of the training and validation,

  • approaches for hyperparameter tuning / optimization,

  • approach for data set creation and modification, and

  • training and validation environment

Note 1: The ML training and validation approach may include random dropout and other robustification methods.

Note 2: ML validation is the optimization of the hyperparameters during Machine Learning Training (MLE.3). The term “validation” has a different meaning than VAL.1.

Note 3: The training environment should reflect the environment of the deployed model.

MLE.3.BP2: Create ML training and validation data set. Select data from the ML data collection provided by SUP.11 and assign them to the data set for training and validation of the ML model according to the specified ML training and validation approach.

Note 4: The ML training and validation data set may include corner cases, unexpected cases, and normal cases depending on the ML requirements.

Note 5: A separated data set for training and validation might not be required in some cases (e.g., kfold cross validation, no optimization of hyperparameters).

MLE.3.BP3: Create and optimize ML model. Create the ML model according to the ML architecture and train it, using the identified ML training and validation data set according to the ML training and validation approach to meet the defined ML requirements, and training and validation exit criteria.

MLE.3.BP4: Ensure consistency and establish bidirectional traceability. Ensure consistency and establish bidirectional traceability between the ML training and validation data set and the ML data requirements.

Note 6: Bidirectional traceability supports consistency and facilitates impact analyses of change requests. Traceability alone, e.g., the existence of links, does not necessarily mean that the information is consistent with each other.

MLE.3.BP5: Summarize and communicate agreed trained ML model. Summarize the results of the optimization and inform all affected parties about the agreed trained ML model.

MLE.3 Machine Learning Training

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Output Information Items

08-65 ML training and validation approach

X

03-51 ML data set

X

01-53 Trained ML model

X

01-54 Hyperparameter

X

13-51 Consistency evidence

X

13-52 Communication evidence

X

Base Practices

BP1: Specify ML training and validation approach

X

BP2: Create ML training and validation data set

X

BP3: Create and optimize ML model

X

BP4: Ensure consistency and establish bidirectional traceability

X

BP5: Summarize and communicate agreed trained ML model

X

4.6.4. MLE.4 Machine Learning Model Testing

Need: MLE.4 Machine Learning Model Testing ASPICE40_4_6_4
status: new

Process ID

MLE.4

Process name

Machine Learning Model Testing

Process purpose

The purpose is to ensure compliance of the trained ML model and the deployed ML model with the ML requirements.

Process outcomes

  1. A ML test approach is defined.

  2. A ML test data set is created.

  3. The trained ML model is tested.

  4. The deployed ML model is derived from the trained ML model and tested.

  5. Consistency and bidirectional traceability are established between the ML test approach and the ML requirements, and the ML test data set and the ML data requirements; and bidirectional traceability is established between the ML test approach and ML test results.

  6. Results of the ML model testing are summarized and communicated with the deployed ML model to all affected parties.

Base Practices

MLE.4.BP1: Specify an ML test approach. Specify an ML test approach suitable to provide evidence for compliance of the trained ML model and the deployed ML model with the ML requirements. The ML test approach includes

  • ML test scenarios with distribution of data characteristics (e.g., gender, weather conditions, street conditions within the ODD) defined by ML requirements,

  • distribution and frequency of each ML test scenario inside the ML test data set,

  • expected test result per test datum,

  • entry and exit criteria of the testing,

  • approach for data set creation and modification, and

  • the required testing infrastructure and environment setup.

Note 1: Expected test result per test datum might require labeling of test data to support comparison of output of the ML model with the expected output.

Note 2: Test datum is the smallest amount of data which is processed by the ML model into only one output. E.g., one image in photo processing or an audio sequence in voice recognition.

Note 3: Data characteristic is one property of the data that may have different expressions in the ODD. E.g., weather condition may contain expressions like sunny, foggy or rainy.

Note 4: An ML test scenario is a combination of expressions of all defined data characteristics e.g., weather conditions = sunny, street conditions = gravel road.

MLE.4.BP2: Create ML test data set. Create the ML test data set needed for testing of the trained ML model and testing of the deployed ML model from the ML data collection provided by SUP.11 considering the ML test approach. The ML test data set shall not be used for training.

Note 5: The ML test data set for the trained ML model might differ from the test data set of the deployed ML model.

Note 6: Additional data sets might be used for special purposes like assurance of safety, fairness, robustness.

MLE.4.BP3: Test trained ML model. Test the trained ML model according to the ML test approach using the created ML test data set. Record and evaluate the ML test results.

Note 7: Evaluation of test logs might include pattern analysis of failed test data to support e.g., trustworthiness.

MLE.4.BP4: Derive deployed ML model. Derive the deployed ML model from the trained ML model according to the ML architecture. The deployed ML model shall be used for testing and delivery to software integration.

Note 8: The deployed ML model will be integrated into the target system and may differ from the trained ML model which often requires powerful hardware and uses interpretative languages.

MLE4.BP5: Test deployed ML model. Test the deployed ML model according to the ML test approach using the created ML test data set. Record and evaluate the ML test results.

MLE.4.BP6: Ensure consistency and establish bidirectional traceability. Ensure consistency and establish bidirectional traceability between the ML test approach and the ML requirements, and the ML test data set and the ML data requirements; and bidirectional traceability is established between the ML test approach and ML test results.

Note 9: Bidirectional traceability supports consistency, and facilitates impact analyses of change requests, and verification coverage demonstration. Traceability alone, e.g., the existence of links, does not necessarily mean that the information is consistent with each other.

MLE.4.BP7: Summarize and communicate results. Summarize the ML test results of the ML model. Inform all affected parties about the agreed results and the deployed ML model.

MLE.4 Machine Learning Model Testing

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Outcome 6

Output Information Items

08-64 ML test approach

X

03-51 ML data set

X

13-50 ML test results

X

X

11-50 Deployed ML model

X

13-51 Consistency evidence

X

13-52 Communication evidence

X

Base Practices

BP1: Specify an ML test approach

X

BP2: Create ML test data set

X

BP3: Test trained ML model

X

BP4: Derive deployed ML model

X

BP5: Test deployed ML model

X

BP6: Ensure consistency and establish bidirectional traceability

X

BP7: Summarize and communicate results

X

4.7. Hardware Engineering process group (HWE)

4.7.1. HWE.1 Hardware Requirements Analysis

Need: HWE.1 Hardware Requirements Analysis ASPICE40_4_7_1
status: new

Process ID

HWE.1

Process name

Hardware Requirements Analysis

Process purpose

The purpose is to establish a structured and analyzed set of hardware requirements consistent with the system requirements, and the system architectural design.

Process outcomes

  1. Hardware requirements are specified.

  2. Hardware requirements are structured and prioritized.

  3. Hardware requirements are analyzed for correctness and technical feasibility.

  4. The impact of hardware requirements on the operating environment is analyzed.

  5. Consistency and bidirectional traceability are established between hardware requirements and system requirements.

  6. Consistency and bidirectional traceability are established between hardware requirements and system architectural design.

  7. The hardware requirements are agreed and communicated to all affected parties.

Base Practices

HWE.1.BP1: Specify hardware requirements. Use the system requirements, and the system architecture including interface definitions, to identify and document the functional and nonfunctional requirements of the hardware according to defined characteristics for requirements.

Note 1: Characteristics of requirements are defined in standards such as ISO IEEE 29148, ISO/IEC IEEE 24765, ISO 26262-8:2018, or the INCOSE Guide For Writing Requirements.

Note 2: Examples for defined characteristics of requirements shared by the above-mentioned standards are verifiability (i.e., verification criteria being inherent in the requirements text), unambiguity/comprehensibility, freedom from design and implementation, and not contradicting any other requirement).

Note 3: In case of hardware-only development, the system requirements and the system architecture refer to a given operating environment. In that case, stakeholder requirements can be used as the basis for identifying the required functions and capabilities of the hardware.

Note 4: The hardware-software-interface (HSI) definition puts in context software and therefore is an interface decision at the system design level. If such a HSI exists, then it may provide input to hardware requirements.

HWE.1.BP2: Structure hardware requirements. Structure and prioritize the hardware requirements.

Note 5: Examples for structuring criteria can be grouping (e.g., by functionality) or variants identification.

Note 6: Prioritization can be done according to project or stakeholder needs via e.g., definition of release scopes. Refer to SPL.2.BP1.

HWE.1.BP3: Analyze hardware requirements. Analyze the specified hardware requirements including their interdependencies to ensure correctness, technical feasibility, and to support project management regarding project estimates.

Note 7: See MAN.3.BP3 for project feasibility and MAN.3.BP5 for project estimates.

Note 8: The analyses of technical feasibility can be done based on a given hardware design (e.g., platform) or by prototype development.

HWE.1.BP4: Analyze the impact on the operating environment. Identify the interfaces between the specified hardware and other elements of the operating environment. Analyze the impact that the hardware requirements will have on these interfaces and the operating environment.

HWE.1.BP5: Ensure consistency and establish bidirectional traceability. Ensure consistency and establish traceability between hardware requirements and the system architecture. Ensure consistency and establish traceability between hardware requirements and system requirements.

Note 9: Redundant traceability is not intended.

Note 10: There may be non-functional hardware requirements that the hardware design does not trace to. Examples are development process requirements. Such requirements are still subject to verification.

Note 11: Bidirectional traceability supports consistency, and facilitates impact analysis of change requests, and demonstration of verification coverage. Traceability alone, e.g., the existence of links, does not necessarily mean that the information is consistent with each other.

Note 12: In case of hardware development only, the system requirements and system architecture refer to a given operating environment. In that case, consistency and bidirectional traceability can be ensured between stakeholder requirements and hardware requirements.

HWE.1.BP6: Communicate agreed hardware requirements and impact on the operating environment. Communicate the agreed hardware requirements and results of the analysis of impact on the operating environment to all affected parties.

HWE.1 Hardware Requirements Analysis

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Outcome 6

Outcome 7

Output Information Items

13-52 Communication Evidence

X

13-51 Consistency Evidence

X

X

17-00 Requirement

X

X

X

17-54 Requirement Attribute

X

15-51 Analysis Results

X

X

Base Practices

BP1: Specify hardware requirements

X

BP2: Structure hardware requirements

X

BP3: Analyze hardware requirements

X

BP4: Analyze the impact on the operating environment

X

BP5: Ensure consistency and establish bidirectional traceability

X

X

BP6: Communicate agreed hardware requirements

X

4.7.2. HWE.2 Hardware Design

Need: HWE.2 Hardware Design ASPICE40_4_7_2
status: new

Process ID

HWE.2

Process name

Hardware Design

Process purpose

The purpose is to provide an analyzed design, including dynamic aspects, that is consistent with the hardware requirements and suitable for manufacturing, and to derive production-relevant data.

Process outcomes

  1. A hardware architecture and hardware detailed design is developed that identifies the elements of the hardware and describes their behavior as well as their interfaces, and the dynamic interactions of the hardware elements.

  2. The hardware architecture and the hardware detailed design is analyzed, and special characteristics are identified.

  3. Consistency and bidirectional traceability are established between hardware requirements and hardware design.

  4. Hardware production data is derived from the hardware detailed design and communicated to all affected parties.

  5. Information for production test is derived from the hardware detailed design and communicated to all affected parties.

  6. The hardware architecture and hardware detailed design and the special characteristics are agreed and communicated to all affected parties.

Base Practices

HWE.2.BP1: Specify the hardware architecture. Develop the hardware architecture that identifies the hardware components. Document the rationale for the defined hardware architecture.

Note 1: Examples for aspects reflected in the hardware architecture are ground concept, supply concept, EMC concept.

Note 2: Examples for a design rationale can be implied by the reuse of a standard hardware, platform, or product line, respectively, or by a make-or-buy decision, or found in an evolutionary way.

HWE.2.BP2: Specify the hardware detailed design. Based on components identified in the hardware architecture, specify the detailed design description and the schematics for the intended hardware variants, including the interfaces between the hardware elements. Derive the hardware layout, the hardware bill of materials, and the production data.

Note 3: The identification of hardware parts and their suppliers in the hardware bill of materials may be subject to a pre-defined repository (see also IATF 16949:2016, clause 8.4.1.2.).

Note 4: Hardware detailed design may be subject to constraints such as availability of hardware parts on the market, hardware design rules, layout rules, creepage and clearance distances, compliance of HW parts with industry standards such as AEC-Q, REACH.

HWE.2.BP3: Specify dynamic aspects. Evaluate and document the dynamic behavior of the relevant hardware elements and the interaction between them.

Note 5: Not all hardware elements have dynamic behavior that needs to be described.

HWE.2.BP4: Analyze the hardware architecture and the hardware detailed design. Analyze the hardware architecture and hardware detailed design regarding relevant technical aspects, and support project management regarding project estimates. Identify special characteristics.

Note 6: Examples for technical aspects are manufacturability for production, suitability of pre-existing hardware components to be reused, or availability of hardware elements.

Note 7: Examples of methods suitable for analyzing technical aspects are simulations, calculations, quantitative or qualitative analyses such as FMEA.

HWE.2.BP5: Ensure consistency and establish bidirectional traceability. Ensure consistency and establish traceability between hardware elements and hardware requirements. Ensure consistency and establish traceability between the hardware detailed design and components of the hardware architecture.

Note 8: There may be non-functional hardware requirements that the hardware design does not trace to. Examples are development process requirements. Such requirements are still subject to verification.

Note 9: Bidirectional traceability further supports consistency, and facilitates impact analysis of change requests, and demonstration of verification coverage. Traceability alone, e.g, the existence of links, does not necessarily mean that the information is consistent with each other.

HWE.2.BP6: Communicate agreed hardware architecture and hardware detailed design. Communicate the agreed hardware architecture and the hardware detailed design, including the special characteristics and relevant production data, to all affected parties.

HWE.2 Hardware Design

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Outcome 6

Output Information Items

04-52 Hardware Architecture

X

04-53 Hardware Detailed Design

X

15-51 Analysis Results

X

13-51 Consistency Evidence

X

17-57 Special Characteristics

X

13-52 Communication Evidence

X

04-54 Hardware Schematics

X

X

X

14-54 Hardware Bill of Materials

X

X

X

04-55 Hardware Layout

X

X

X

03-54 Hardware Production Data

X

X

X

04-56 Hardware Element Interface

X

Base Practices

BP1: Specify the hardware architecture

X

X

X

BP2: Specify the hardware detailed design

X

X

X

BP3: Specify dynamic aspects

X

BP4: Analyze the hardware architecture and the hardware detailed design

X

BP5: Ensure consistency and establish bidirectional traceability

X

BP7: Communicate agreed hardware architecture and hardware detailed design

X

X

X

4.7.3. HWE.3 Verification against Hardware Design

Need: HWE.3 Verification against Hardware Design ASPICE40_4_7_3
status: new

Process ID

HWE.3

Process name

Verification against Hardware Design

Process purpose

The purpose is to ensure that the production data compliant hardware is verified to provide evidence for compliance with the hardware design.

Process outcomes

1) Verification measures are specified for verification of the hardware against the hardware design, including the interfaces between hardware elements and the dynamic aspects.

2) Verification measures are selected according to the release scope considering criteria, including criteria for regression verification.

3) Verification is performed on production data compliant samples using the selected verification measures, and verification results are recorded.

4) Consistency and bidirectional traceability are established between hardware elements and verification measures.

5) Bidirectional traceability is established between verification measures and verification results.

  1. Verification results are summarized and communicated to all affected parties.

Base Practices

HWE.3.BP1: Specify verification measures for the verification against hardware design. Specify the verification measures suitable to provide evidence for compliance of the hardware with the hardware design and its dynamic aspects. This includes

  • techniques for the verification measures,

  • pass/fail criteria for the verification measures,

  • a definition of entry and exit criteria for the verification measures,

  • necessary sequence of the verification measures, and

  • the required verification infrastructure and environment setup.

Note 1: Examples on what a verification measure may focus on are the timeliness and timing dependencies of the correct signal flow between interfacing hardware elements, interactions between hardware components.

Note 2: Measuring points can be used for stepwise testing of hardware elements.

HWE.3.BP2: Ensure use of compliant samples. Ensure that the samples used for verification against hardware design are compliant with the corresponding production data, including special characteristics. Ensure that deviations are documented and that they do not alter verification results.

Note 3: Examples of compliance are sample reports, record of visual inspection, ICT report.

HWE.3.BP3: Select verification measures. Document the selection of verification measures considering selection criteria including regression criteria. The documented selection of verification measures shall have sufficient coverage according to the release scope.

Note 4: Examples for selection criteria can be prioritization of requirements, the need for regression due to changes to the hardware design, or the intended use of the delivered hardware release (e.g., test bench, test track, public road etc.)

HWE.3.BP4: Verify hardware design. Verify the hardware design using the selected verification measures. Record the verification results including pass/fail status and corresponding verification measure output data.

Note 5: See SUP.9 for handling of non-conformances.

HWE.3.BP5: Ensure consistency and establish bidirectional traceability. Ensure consistency and establish bidirectional traceability between hardware elements and the verification measures. Establish bidirectional traceability between the verification measures and verification results.

Note 6: Bidirectional traceability supports consistency, and facilitates impact analysis of change requests, and demonstration of verification coverage. Traceability alone, e.g., the existence of links, does not necessarily mean that the information is consistent with each other.

HWE.3.BP6: Summarize and communicate results. Summarize the verification results and communicate them to all affected parties.

Note 7: Providing all necessary information from the test case execution in a summary enables other parties to judge the consequences.

HWE.3 Verification against Hardware Design

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Outcome 6

Output Information Items

08-60 Verification Measure

X

03-50 Verification Measure Data

X

08-58 Verification Measure Selection Set

X

15-52 Verification Results

X

13-51 Consistency Evidence

X

X

13-52 Communication Evidence

X

Base Practices

BP1: Specify verification measures for the verification against hardware design

X

BP2: Ensure use of compliant samples

X

BP3: Select verification measures

X

BP4: Verify hardware design

X

BP5: Ensure consistency and establish bidirectional traceability

X

X

BP6: Summarize and communicate results

X

4.7.4. HWE.4 Verification against Hardware Requirements

Need: HWE.4 Verification against Hardware Requirements ASPICE40_4_7_4
status: new

Process ID

HWE.4

Process name

Verification against Hardware Requirements

Process purpose

The purpose is to ensure that the complete hardware is verified to be consistent with the hardware requirements.

Process outcomes

  1. Verification measures are specified for verification of the hardware against the hardware requirements.

  2. Verification measures are selected considering criteria, including criteria for regression verification.

  3. Verification is performed, if applicable on production data compliant samples, using the selected verification measures, and verification results are recorded.

  4. Consistency and bidirectional traceability are established between verification measures and hardware requirements.

  5. Bidirectional traceability is established between verification measures and verification results.

  6. Verification results are summarized and communicated to all affected parties.

Base Practices

HWE.4.BP1: Specify verification measures for the verification against hardware requirements. Specify the verification measure to provide evidence for compliance with the hardware requirements. This includes

  • techniques for the verification measures,

  • pass/fail criteria for the verification measures,

  • a definition of entry and exit criteria for the verification measures,

  • necessary sequence of verification the measures, and

  • the required verification infrastructure and environment setup

Note 1: The verification measures may cover aspects such as thermal, environmental, robustness/lifetime, and EMC.

HWE.4.BP2: Ensure use of compliant samples. Ensure that the samples used for the verification against hardware requirements are compliant with the corresponding production data, including special characteristics, provided by hardware design.

Note 2: Examples of compliance are sample reports, record of visual inspection, ICT report.

HWE.4.BP3: Select verification measures. Document the selection of verification measures considering selection criteria including regression criteria. The documented selection of verification measures shall have sufficient coverage according to the release scope.

Note 3: Examples for selection criteria can be prioritization of requirements, the need for regression due to changes to the hardware requirements, or the intended use of the delivered hardware release (e.g, test bench, test track, public road etc.).

HWE.4.BP4: Verify the compliant hardware samples. Verify the compliant hardware samples using the selected verification measures. Record the verification results including pass/fail status and corresponding verification measure output data.

Note 4: See SUP.9 for handling of non-conformances.

HWE.4.BP5: Ensure consistency and establish bidirectional traceability. Ensure consistency between hardware requirements and verification measures. Establish bidirectional traceability between hardware requirements and verification measures. Establish bidirectional traceability between verification measures and verification results.

Note 5: Bidirectional traceability supports consistency, and facilitates impact analysis of change requests, and demonstration of verification coverage. Traceability alone, e.g., the existence of links, does not necessarily mean that the information is consistent with each other.

HWE.4.BP6: Summarize and communicate results. Summarize the verification results and communicate them to all affected parties.

Note 6: Providing all necessary information from the test case execution in a summary enables other parties to judge the consequences.

HWE.4 Verification against Hardware Requirements

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Outcome 6

Output Information Items

08-60 Verification Measure

X

03-50 Verification Measure Data

X

08-58 Verification Measure Selection Set

X

15-52 Verification Results

X

13-51 Consistency Evidence

X

X

13-52 Communication Evidence

X

Base Practices

BP1: Specify verification measures for the verification against hardware requirements

X

BP2: Ensure use of compliant samples

X

BP3: Select verification measures

X

BP4: Verify hardware

X

BP5: Ensure consistency and establish bidirectional traceability

X

X

BP6: Summarize and communicate results

X

4.8. Supporting process group (SUP)

4.8.1. SUP.1 Quality Assurance

Need: SUP.1 Quality Assurance ASPICE40_4_8_1
status: new

Process ID

SUP.1

Process name

Quality Assurance

Process purpose

The purpose of the Quality Assurance Process is to provide independent and objective assurance that work products and processes comply with defined criteria and that nonconformances are resolved and further prevented.

Process outcomes

  1. Quality assurance is performed independently and objectively without conflicts of interest.

  2. Criteria for the quality of work products and process performance are defined.

  3. Conformance of work products and process performance with the defined criteria and targets is verified, documented and communicated to the relevant parties.

  4. Non-conformances are tracked, resolved, and further prevented.

  5. Non-conformances are escalated to appropriate levels of management.

  6. Management ensures that escalated non-conformances are resolved.

Base Practices

SUP.1.BP1: Ensure independence of quality assurance. Ensure that quality assurance is performed independently and objectively without conflicts of interest.

Note 1: Possible inputs for evaluating the independence may be assignment to financial and/or organizational structure as well as responsibility for processes that are subject to quality assurance (no self-monitoring).

SUP.1.BP2: Define criteria for quality assurance. Define quality criteria for work products as well as for process tasks and their performance.

Note 2: Quality criteria may consider internal and external inputs such as customer requirements, standards, milestones, etc.

SUP.1.BP3: Assure quality of work products. Identify work products subject to quality assurance according to the quality criteria. Perform appropriate activities to evaluate the work products against the defined quality criteria and document the results.

Note 3: Quality assurance activities may include reviews, problem analysis and lessons learned that improve the work products for further use.

SUP.1.BP4: Assure quality of process activities. Identify processes subject to quality assurance according to the quality criteria. Perform appropriate activities to evaluate the processes against their defined quality criteria and associated target values and document the results.

NOTE 4: Quality assurance activities may include process assessments, problem analysis, regular check of methods, tools, and the adherence to defined processes, and consideration of lessons learned.

SUP.1.BP5: Summarize and communicate quality assurance activities and results. Regularly report performance, non-conformances, and trends of quality assurance activities to all affected parties.

SUP.1.BP6: Ensure resolution of non-conformances. Analyze, track, correct, resolve, and further prevent non-conformances found in quality assurance activities.

NOTE 5: Non-conformances detected in work products may be entered into the problem resolution management process (SUP.9).

NOTE 6: Non-conformances detected in the process definition or implementation may be entered into a process improvement process (PIM.3).

SUP.1.BP7: Escalate non-conformances. Escalate relevant non-conformances to appropriate levels of management and other relevant stakeholders to facilitate their resolution.

NOTE 7: The decision whether to escalate non-conformances may be based on criteria such as delay of resolution, urgency, and risk.

SUP.1

Quality Assurance

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Outcome 6

Output Information Items

16-50 Organizational structure

X

X

18-52 Escalation path

X

X

18-07 Quality criteria

X

X

X

13-52 Communication evidence

X

X

X

13-18 Quality conformance evidence

X

X

13-19 Review evidence

X

X

14-02 Corrective action

X

X

Base Practices

BP1: Ensure independence of quality assurance.

X

BP2: Define criteria for quality assurance.

X

BP3: Assure quality of work products.

X

X

BP4: Assure quality of process activities.

X

X

BP5: Summarize and communicate quality assurance activities and results.

X

X

X

BP6: Ensure resolution of non-conformances.

X

X

BP7: Escalate non-conformances.

X

X

4.8.2. SUP.8 Configuration Management

Need: SUP.8 Configuration Management ASPICE40_4_8_2
status: new

Process ID

SUP.8

Process name

Configuration Management

Process purpose

The purpose of the Configuration Management Process is to establish and maintain the integrity of relevant configuration items and baselines, and make them available to affected parties.

Process outcomes

  1. Selection criteria for configuration items are defined and applied.

  2. Configuration item properties are defined.

  3. Configuration management is established.

  4. Modifications are controlled.

  5. Baselining is applied.

  6. The status of the configuration items is recorded and reported.

  7. The completeness and consistency of the baselines is ensured.

  8. The availability of backup and recovery mechanisms is verified.

Base Practices

SUP.8.BP1: Identify configuration items. Define selection criteria for identifying relevant work products to be subject to configuration management. Identify and document configuration items according to the defined selection criteria.

NOTE 1: Configuration items are representing work products or group of work products which are subject to configuration management as a single entity.

NOTE 2: Configuration items may vary in complexity, size, and type, ranging from an entire system including all system, hardware, and software documentation down to a single element or document.

NOTE 3: The selection criteria may be applied to single work products or a group of work products.

SUP.8.BP2: Define configuration item properties. Define the necessary properties needed for the modification and control of configuration items.

NOTE 4: The configuration item properties may be defined for single configuration items or a group of items.

NOTE 5: Configuration item properties may include a status model (e.g., Under Work, Tested, Released, etc.), storage location, access rights, etc.

NOTE 6: The application of properties may be implemented by attributes of configuration items.

SUP.8.BP3: Establish configuration management. Establish configuration management mechanisms for control of identified configuration items including the configuration item properties, including mechanisms for controlling parallel modifications of configuration items.

NOTE 7: This may include specific mechanisms for different configuration item types, such as branch and merge management, or checkout control.

SUP.8.BP4: Control modifications. Control modifications using the configuration management mechanisms.

NOTE 8: This may include the application of a defined status model for configuration items.

SUP.8.BP5: Establish baselines. Define and establish baselines for internal purposes, and for external product delivery, for all relevant configuration items.

SUP.8.BP6: Summarize and communicate configuration status. Record, summarize, and communicate the status of configuration items and established baselines to affected parties in order to support the monitoring of progress and status.

NOTE 9: Regular communication of the configuration status, e.g., based on a defined status model supports project management, quality activities, and dedicated project phases such as software integration.

SUP.8.BP7: Ensure completeness and consistency. Ensure that the information about configuration items is correct and complete including configuration item properties. Ensure the completeness and consistency of baselines.

NOTE 10: Completeness and consistency of a baseline means that all required configuration items are included and consistent, and have the required status. This can be used to support e.g., project gate approval.

SUP.8.BP8: Verify backup and recovery mechanisms availability. Verify the availability of appropriate backup and recovery mechanisms for the configuration management including the controlled configuration items. Initiate measures in case of insufficient backup and recovery mechanisms.

NOTE 11: Backup and recovery mechanisms may be defined and implemented by organizational units outside the project team. This may include references to corresponding procedures or regulations.

SUP.8 Configuration Management

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Outcome 6

Outcome 7

Outcome 8

Output Information Items

18-53 Configuration item selection criteria

X

01-52 Configuration item list

X

X

X

16-03 Configuration management system

X

X

X

13-08 Baseline

X

X

14-01 Change history

X

X

X

15-56 Configuration status

X

13-51 Consistency Evidence

X

06-52 Backup and recovery mechanism information

X

Base Practices

BP1: Identify configuration items

X

BP2: Define configuration item properties

X

BP3: Establish configuration management

X

X

BP4: Control modifications

X

BP5: Establish baselines

X

BP6: Summarize and communicate configuration status

X

BP7: Ensure completeness and consistency

X

BP8: Verify backup and recovery mechanisms availability

X

4.8.3. SUP.9 Problem Resolution Management

Need: SUP.9 Problem Resolution Management ASPICE40_4_8_3
status: new

Process ID

SUP.9

Process name

Problem Resolution Management

Process purpose

The purpose of the Problem Resolution Management Process is to ensure that problems are identified, recorded, analyzed, and their resolution is managed and controlled.

Process outcomes

  1. Problems are uniquely identified, recorded and classified.

  2. Problems are analyzed and assessed to determine an appropriate solution.

  3. Problem resolution is initiated.

  4. Problems are tracked to closure.

  5. The status of problems including trends identified are reported to stakeholders.

Base Practices

SUP.9.BP1: Identify and record the problem. Each problem is uniquely identified, described and recorded. A status is assigned to each problem to facilitate tracking. Supporting information is provided to reproduce and diagnose the problem.

NOTE 1: Problems may relate to e.g., product, resources, or methods.

NOTE 2: Example values for the problem status are “new”, “solved”, “closed”, etc.

NOTE 3: Supporting information may include e.g, the origin of the problem, how it can be reproduced, environmental information, by whom it has been detected.

NOTE 4: Unique identification supports traceability to changes made as needed by the change request management process (SUP.10).

SUP.9.BP2: Determine the cause and the impact of the problem. Analyze the problem, determine its cause, including common causes if existing, and impact. Involve relevant parties. Categorize the problem.

NOTE 5: Problem categorization (e.g., light, medium, severe) may be based on severity, criticality, urgency, etc.

SUP.9.BP3: Authorize urgent resolution action. Obtain authorization for immediate action if a problem requires an urgent resolution according to the categorization.

SUP.9.BP4: Raise alert notifications. If according to the categorization the problem has a high impact on other systems or other affected parties, an alert notification needs to be raised accordingly.

SUP.9.BP5: Initiate problem resolution. Initiate appropriate actions according to the categorization to resolve the problem long-term, including review of those actions or initiate a change request. This includes synchronization and consistency with short-term urgent resolution actions, if applicable.

SUP.9.BP6: Track problems to closure. Track the status of problems to closure including all related change requests. The closure of problems is accepted by relevant stakeholders.

SUP.9.BP7: Report the status of problem resolution activities. Collect and analyze problem resolution management data, identify trends, and initiate related actions. Regularly report the results of data analysis, the identified trends and the status of problem resolution activities to relevant stakeholders.

NOTE 6: Collected data may contain information about where the problems occurred, how and when they were found, what their impacts were, etc.

SUP.9 Problem Resolution Management

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Output Information Items

13-07 Problem

X

X

X

X

15-55 Problem analysis evidence

X

15-12 Problem status

X

Base Practices

BP1: Identify and record the problem

X

X

BP2: Determine the cause and the impact of the problem

X

X

BP3: Authorize urgent resolution action

X

BP4: Raise alert notifications

X

BP5: Initiate problem resolution

X

BP6: Track problems to closure

X

X

BP7: Report the status of problem resolution activities

X

4.8.4. SUP.10 Change Request Management

Need: SUP.10 Change Request Management ASPICE40_4_8_4
status: new

Process ID

SUP.10

Process name

Change Request Management

Process purpose

The purpose of the Change Request Management Process is to ensure that change requests are recorded, analyzed, tracked, approved, and implemented.

Process outcomes

  1. Requests for changes are recorded and identified.

  2. Change requests are analyzed, dependencies and relationships to other change requests are identified, and the impact is estimated.

  3. Change requests are approved before implementation and prioritized accordingly.

  4. Bidirectional traceability is established between change requests and affected work products.

  5. Implementation of change requests is confirmed.

  6. Change requests are tracked to closure and status of change requests is communicated to affected parties.

Base Practices

SUP.10.BP1: Identify and record the change requests. The scope for application of change requests is identified. Each change request is uniquely identified, described, and recorded, including the initiator and reason of the change request. A status is assigned to each change request to facilitate tracking.

NOTE 1: Change requests may be used for changes related to e.g., product, process, methods.

NOTE 2: Example values for the change request status are “open”, “under investigation”, “implemented”, etc.

NOTE 3: The change request handling may differ across the product life cycle e.g., during prototype construction and series development

SUP.10.BP2: Analyze and assess change requests. Change requests are analyzed by relevant parties according to analysis criteria. Work products affected by the change request and dependencies to other change requests are determined. The impact of the change requests is assessed.

NOTE 4: Examples for analysis criteria are: resource requirements, scheduling issues, risks, benefits, etc.

SUP.10.BP3: Approve change requests before implementation. Change requests are prioritized and approved for implementation based on analysis results and availability of resources.

NOTE 5: A Change Control Board (CCB) is an example mechanism used to approve change requests.

NOTE 6: Prioritization of change requests may be done by allocation to releases.

SUP.10.BP4: Establish bidirectional traceability. Establish bidirectional traceability between change requests and work products affected by the change requests. In case that the change request is initiated by a problem, establish bidirectional traceability between change requests and the corresponding problem reports.

SUP.10.BP5: Confirm the implementation of change requests. The implementation of change requests is confirmed before closure by relevant stakeholders.

SUP.10.BP6: Track change requests to closure. Change requests are tracked to closure. The status of change requests is communicated to all affected parties.

NOTE 7: Examples for informing affected parties can be daily standup meetings or tool-supported workflows.

SUP.10 Change Request Management

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Outcome 6

Output Information Items

18-57 Change analysis criteria

X

13-16 Change request

X

X

X

X

X

13-51 Consistency evidence

X

Base Practices

BP1: Identify and record the change requests

X

BP2: Analyze and assess change requests

X

BP3: Approve change requests before implementation

X

BP4: Establish bidirectional traceability

X

BP5: Confirm the implementation of change requests

X

BP6: Track change requests to closure

X

4.8.5. SUP.11 Machine Learning Data Management

Need: SUP.11 Machine Learning Data Management ASPICE40_4_8_5
status: new

Process ID

SUP.11

Process name

Machine Learning Data Management

Process purpose

The purpose is to define and align ML data with ML data requirements, maintain the integrity and quality of the ML data, and make them available to affected parties.

Process outcomes

  1. A ML data management system including an ML data lifecycle is established.

  2. A ML data quality approach is developed including ML data quality criteria.

  3. Collected ML data are processed for consistency with ML data requirements.

  4. ML data are verified against defined ML data quality criteria and updated as needed.

  5. ML data are agreed and communicated to all affected parties.

Base Practices

SUP.11.BP1: Establish an ML data management system. Establish an ML data management system which supports

  • ML data management activities,

  • relevant sources of ML data,

  • ML data life cycle including a status model, and

  • interfaces to affected parties.

Note 1: Supported ML data management activities may include data collection, labeling/annotation, and structuring.

SUP.11.BP2: Develop an ML data quality approach. Develop an approach to ensure that the quality of ML data is analyzed based on defined ML data quality criteria and activities are performed to support avoidance of biases of data.

Note 2: Examples of ML data quality criteria are relevant data sources, reliability and consistency of labelling, completeness against ML data requirements.

Note 3: The ML data management system should support the quality criteria and activities of the ML data quality approach.

Note 4: Biases to avoid may include sampling bias (e.g., gender, age) and feedback loop bias.

Note 5: For creation of ML data sets see MLE.3.BP2 and MLE.4.BP2.

SUP.11.BP3: Collect ML data. Relevant sources for raw data are identified and continuously monitored for changes. The raw data is collected according to the ML data requirements.

Note 6: The identification and collection of ML data might be an organizational responsibility.

Note 7: Continuous monitoring should include the ODD and may lead to changes of the ML requirements.

SUP.11.BP4: Process ML data. The raw data are processed (annotated, analyzed, and structured) according to the ML data requirements.

SUP.11.BP5: Assure quality of ML data. Perform the activities according to the ML data quality approach to ensure that the ML data meets the defined ML data quality criteria.

Note 8: These activities may include sample-based reviews or statistical methods.

SUP.11.BP6: Communicate agreed processed ML data. Inform all affected parties about the agreed processed ML data and provide them to the affected parties.

SUP.11 Machine Learning Data Management

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Output Information Items

16-52 ML data management system

X

19-50 ML data quality approach

X

03-53 ML data

X

X

13-52 Communication evidence

X

Base Practices

BP1: Establish an ML data management system

X

BP2: Develop an ML data quality approach

X

BP3: Collect ML data

X

BP4: Process ML data

X

BP5: Assure quality of ML data

X

BP6: Communicate agreed processed ML data

X

4.9. Management process group (MAN)

4.9.1. MAN.3 Project Management

Need: MAN.3 Project Management ASPICE40_4_9_1
status: new

Process ID

MAN.3

Process name

Project Management

Process purpose

The purpose is to identify and control the activities, and establish resources necessary for a project to develop a product, in the context of the project’s requirements and constraints.

Process outcomes

  1. The scope of the work for the project is defined.

  2. The feasibility of achieving the goals of the project with available resources and constraints is evaluated.

  3. The activities and resources necessary to complete the work are sized and estimated.

  4. Interfaces within the project, and with other projects and organizational units, are identified and monitored.

  5. Plans for the execution of the project are developed, implemented and maintained.

  6. Progress of the project is monitored and reported.

  7. Adjustment is performed when project goals are not achieved.

Base Practices

MAN.3.BP1: Define the scope of work. Identify the project’s goals, motivation and boundaries.

MAN.3.BP2: Define project life cycle. Define the life cycle for the project, which is appropriate to the scope, context, and complexity of the project. Define a release scope for relevant milestones.

Note 1: This may include the alignment of the project life cycle with the customer’s development process.

MAN.3.BP3: Evaluate feasibility of the project. Evaluate the feasibility of achieving the goals of the project with respect to time, project estimates, and available resources.

Note 2: The evaluation of feasibility may consider technical constraints of the project.

MAN.3.BP4: Define and monitor work packages. Define and monitor work packages and their dependencies according to defined project life cycle and estimations.

Note 3: The structure and the size of the work packages support an adequate progress monitoring.

Note 4: Work packages may be organized in a work breakdown structure.

MAN.3.BP5: Define and monitor project estimates and resources. Define and monitor project estimates of effort and resources based on project’s goals, project risks, motivation and boundaries.

Note 5: Examples of necessary resources are budget, people, product samples, or infrastructure Note 6: Project risks (using MAN.5) may be considered.

Note 7: Estimations and resources may include engineering, management and supporting processes.

MAN.3.BP6: Define and monitor required skills, knowledge, and experience. Identify and monitor the required skills, knowledge, and experience for the project in line with the estimates and work packages.

Note 8: Training, mentoring or coaching of individuals may be applied to resolve deviations from required skills and knowledge.

MAN.3.BP7: Define and monitor project interfaces and agreed commitments. Identify and agree interfaces of the project with affected stakeholders and monitor agreed commitments. Define an escalation mechanism for commitments that are not fulfilled.

Note 9: Affected stakeholders may include other projects, organizational units, sub-contractors, and service providers.

MAN.3.BP8: Define and monitor project schedule. Allocate resources to work packages and schedule each activity of the project. Monitor the performance of activities against schedule.

MAN.3.BP9: Ensure consistency. Regularly adjust estimates, resources, skills, work packages and their dependencies, schedules, plans, interfaces, and commitments for the project to ensure consistency with the scope of work.

Note 10: This may include the consideration of critical dependencies, that are an input for risk management.

MAN.3.BP10: Review and report progress of the project. Regularly review and report the status of the project and the fulfillment of work packages against estimated effort and duration to all affected parties. Prevent recurrence of identified problems.

Note 11: Project reviews may be executed at regular intervals by the management. Project reviews may contribute to identify best practices and lessons learned.

Note 12: Refer to SUP.9 for resolution of problems

MAN.3 Project Management

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Outcome 6

Outcome 7

Output Information Items

08-53 Scope of work

X

08-54 Feasibility analysis

X

X

14-10 Work package

X

X

X

13-52 Communication evidence

X

X

13-16 Change request

X

13-51 Consistency evidence

X

X

14-02 Corrective action

X

X

18-52 Escalation path

X

X

X

08-56 Schedule

X

X

X

14-50 Stakeholder groups list

X

15-06 Project status

X

X

Base Practices

BP1: Define the scope of work

X

BP2: Define project life cycle

X

X

BP3: Evaluate feasibility of the project

X

BP4: Define and monitor work packages

X

X

X

X

BP5: Define and monitor project estimates and resources

X

X

X

BP6: Define and monitor required skills, knowledge, and experience

X

X

BP7: Define and monitor project interfaces and agreed commitments

X

X

X

BP8: Define and monitor project schedule

X

X

BP9: Ensure consistency

X

X

X

X

BP10: Review and report progress of the project

X

X

4.9.2. MAN.5 Risk Management

Need: MAN.5 Risk Management ASPICE40_4_9_2
status: new

Process ID

MAN.5

Process name

Risk Management

Process purpose

The purpose is to Regularly identify, analyze, treat and monitor process related risks and product related risks.

Process outcomes

  1. The sources of risks are identified and regularly updated.

  2. Potential undesirable events are identified as they develop during the conduct of the project.

  3. Risks are analyzed and the priority in which to apply resources to treatment of these risks is determined.

  4. Risk measures are defined, applied, and assessed to determine changes in the status of risk and the progress of the risk treatment activities.

  5. Appropriate treatment is taken to correct or avoid the impact of risk based on its priority, probability, and consequence or other defined risk threshold.

Base Practices

MAN.5.BP1: Identify sources of risks. Identify and regularly update the sources of risks with affected parties.

Note 1: Risks may include technical, economical, and schedule risks.

Note 2: Risks may include the suppliers’ deliverables and services.

Note 3: The risk sources may vary across the entire project life cycle.

MAN.5.BP2: Identify potential undesirable events. Identify potential undesirable events within the scope of the risk management for the project.

MAN.5.BP3: Determine risks. Determine the probability and severity of the undesirable events to support priorities for the mitigation of the risks.

Note 4: Different methods may be used to analyze technical risks of a system, for example, functional analysis, simulation, FMEA, FTA etc.

MAN.5.BP4: Define risk treatment options. For each risk select a treatment option to accept, mitigate, avoid, or share (transfer) the risk.

MAN.5.BP5: Define and perform risk treatment activities. Define and perform risk activities for risk treatment options.

MAN.5.BP6: Monitor risks. Regularly re-evaluate the risk related to the identified potential undesirable events to determine changes in the status of a risk and to evaluate the progress of the risk treatment activities.

Note 5: Risks of high priority may need to be communicated to and monitored by higher levels of management.

MAN.5.BP7: Take corrective action. When risk treatment activities are not effective, take appropriate corrective action.

Note 6: Corrective actions may involve reevaluation of risks, developing and implementing new mitigation concepts or adjusting the existing concepts.

MAN.5 Risk Management

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Output Information Items

15-51 Analysis results

X

X

X

X

15-09 Risk status

X

X

X

X

08-55 Risk measure

X

X

14-02 Corrective action

X

X

Base Practices

BP1: Identify sources of risks

X

BP2: Identify potential undesirable events

X

BP3: Determine risks

X

BP4: Define risk treatment options

X

X

BP5: Define and perform risk treatment activities.

X

X

BP6: Monitor risks

X

BP7: Take corrective action

X

4.9.3. MAN.6 Measurement

Need: MAN.6 Measurement ASPICE40_4_9_3
status: new

Process ID

MAN.6

Process name

Measurement

Process purpose

The purpose is to Collect and analyze data relating to the development results and processes implemented within the organization and its projects, to support effective management of the processes.

Process outcomes

  1. The measurement information needs that are necessary to evaluate the achievement of process objectives and the achievement of desired work products are identified.

  2. An appropriate set of metrics, driven by the information needs are identified and/or developed.

  3. Measurement activities are identified and performed.

  4. The required metrics are collected, stored, analyzed, and the results interpreted.

  5. Metrics are used to support decisions and provide an objective basis for communication.

Base Practices

MAN.6.BP1: Identify information needs. Identify the measurement information needs that are necessary to evaluate the achievement of process objectives and work products.

Note 1: Information needs may change over time. Therefore, the measurement process may be used in an iterative way.

MAN.6.BP2: Specify metrics. Identify and develop an appropriate set of metrics based on measurement information needs.

Note 2: Metrics may be related to processes or development results.

MAN.6.BP3: Collect and store metrics. Collect and store both base and derived metrics, including any context information necessary to verify and understand the metrics.

Note 3: Base metrics in the context of this process are directly gathered metrics like “number of defects found” or “number of lines of code”, where derived metrics are two or more metrics that are brought in relation to each other like “number of defects found per line of code”.

MAN.6.BP4: Analyze collected metrics. Analyze, interpret and review measured values to support decision-making.

MAN.6.BP5: Communicate analysis results. Communicate analysis results to all affected parties.

MAN.6.BP6: Use metrics for decision-making. Make accessible and use information from collected metrics and analysis results for any decision-making process for which it is relevant.

MAN.6 Measurement

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Output Information Items

03-03 Benchmarking data

X

X

03-04 Customer satisfaction data

X

X

03-06 Process performance information

X

X

07-51 Measurement result

X

X

X

X

15-51 Analysis results

X

X

X

Base Practices

BP1: Identify information needs

X

BP2: Specify metrics

X

X

BP3: Collect and store metrics

X

X

BP4: Analyze collected metrics

X

X

BP5: Communicate measurement information

X

BP6: Use metrics for decision-making

X

4.10. Process improvement process group (PIM)

4.10.1. PIM.3 Process Improvement

Need: PIM.3 Process Improvement ASPICE40_4_10_1
status: new

Process ID

PIM.3

Process name

Process Improvement

Process purpose

The purpose is to continually improve the organization’s effectiveness and efficiency through the processes used and ensure alignment of the processes with the business needs.

Process outcomes

  1. Commitment is established to provide resources to sustain improvement measures.

  2. Issues arising from the organization’s internal or external environment are identified as improvement opportunities and justified as reasons for change.

  3. Analysis of the current status of the existing process is performed.

  4. Improvement goals are identified and prioritized, and consequent changes to the process are defined, documented and implemented.

  5. The effects of process implementation are monitored, measured and confirmed against the identified improvement goals.

  6. Knowledge gained from the improvement is communicated within the organization.

Base Practices

PIM.3.BP1: Establish commitment. Establish commitment to support the process improvement staff, to provide resources and further enablers to sustain improvement actions.

Note 1: The process improvement process is a generic process, which can be used at all levels (e.g, organizational level, process level, project level, etc.) and which can be used to improve all processes.

Note 2: Commitment at all levels of management may support process improvement.

Note 3: Enablers for improvement measures may include trainings, methods, infrastructure, etc.

PIM.3.BP2: Identify improvement measures. Identify issues from the analysis of process performance and derive improvement opportunities with justified reasons for change.

Note 4: Analysis may include problem report trend analysis (see SUP.9), analysis from Quality Assurance and Verification results and records (see SUP.1), validation results and records, and product quality metrics like defect rate.

Note 5: Issues and improvement suggestions may be addressed by the customer.

Note 6: Sources for identification of issues may include: process assessment results, audits, customer’s satisfaction reports, measurements of organizational effectiveness/efficiency, costs of quality.

PIM.3.BP3: Establish process improvement goals. Analyze the current status of the existing processes and establish improvement goals.

Note 7: The current status of processes may be determined by process assessment.

PIM.3.BP4: Prioritize improvements. Prioritize the improvement goals and improvement measures.

PIM.3.BP5: Define process improvement measures. Process improvement measures are defined.

Note 8: Improvements may be documented in incremental steps.

PIM.3.BP6: Implement process improvement measures. Implement and apply the improvements to the processes. Update the Process documentation and train people as needed.

Note 9: Process application can be supported by establishing policies, adequate process infrastructure, process training, process coaching and tailoring processes to local needs.

Note 10: Improvements may be piloted before roll out within the organization.

PIM.3.BP7: Confirm process improvement. The effects of process implementation are monitored and measured, and the achievement of defined improvement goals is confirmed.

PIM.3.BP8: Communicate results of improvement. Knowledge gained from the improvements and progress of the improvement implementation is communicated to affected parties.

PIM.3 Process Improvement

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Outcome 6

Output Information Items

02-01 Commitment/agreement

X

06-04 Training material

X

X

07-04 Process metric

X

X

10-00 Process description

X

13-52 Communication evidence

X

13-16 Change request

X

15-51 Analysis result

X

X

X

X

15-13 Assessment/audit report

X

X

15-16 Improvement opportunity

X

X

X

16-06 Process repository

X

Base Practices

BP1: Establish commitment

X

BP2: Identify improvement measures

X

X

BP3: Establish process improvement goals

X

BP4: Prioritize improvements

X

BP5: Define process improvement measures

X

BP6: Implement process improvement measures

X

BP7: Confirm process improvement

X

BP8: Communicate results of improvement

X

4.11. Reuse process group (REU)

4.11.1. REU.2 Management of Products for Reuse

Need: REU.2 Management of Products for Reuse ASPICE40_4_11_1
status: new

Process ID

REU.2

Process name

Management of Products for Reuse

Process purpose

The purpose is to ensure that reused work products are analyzed, verified, and approved for their target context.

Process outcomes

  1. Products for reuse are selected using defined criteria.

  2. Products for reuse are analyzed for portability and interoperability.

  3. Limitations for reuse are defined and communicated.

  4. Products for reuse are verified.

  5. Products for reuse are provided to affected parties.

  6. Communication mechanism is established with the reuse product provider.

Base Practices

REU.2.BP1: Select products for reuse. Select the products to be reused using defined criteria.

Note 1: Products for reuse may be systems, hardware or software components, third party components or legacy components.

REU.2.BP2: Analyze the reuse capability of the product. Analyze the designated target architecture and the product to be reused to determine its applicability in the target architecture according to relevant criteria.

Note 2: Examples for criteria can be requirements compliance, verifiability of the product to be reused in the target architecture, or portability/interoperability.

REU.2.BP3: Define limitations for reuse. Define and communicate limitations for the products to be reused.

Note 3: Limitations may address parameters of operational environment.

REU.2.BP4: Ensure qualification of products for reuse. Provide evidence that the product for reuse is qualified for the intended use of the deliverable.

Note 4: Qualification may be demonstrated by verification evidence.

Note 5: Verification may include the appropriateness of documentation.

REU.2.BP5: Provide products for reuse. Make available the product to be reused to affected parties.

Note 6: Refer to HWE.3, SWE.5 or SYS.4 for more information on integration of hardware, software, or system components.

REU.2.BP6: Communicate information about effectiveness of reuse activities. Establish communication and notification mechanism about experiences and technical outcomes to the provider of reused products.

Note 7: The communication with the provider of a reused product may depend on whether the product is under development or not.

REU.2 Management of Products for Reuse

Outcome 1

Outcome 2

Outcome 3

Outcome 4

Outcome 5

Outcome 6

Output Information Items

04-02 Domain architecture

X

X

12-03 Reuse candidate

X

X

13-52 Communication evidence

X

15-07 Reuse analysis evidence

X

X

13-53 Qualification evidence

X

Base Practices

BP1: Select products for reuse

X

BP2: Analyze the reuse capability of the product

X

BP3: Define limitations for reuse

X

BP4: Ensure qualification of products for reuse

X

BP5: Provide products for reuse

X

BP6: Communicate information about effectiveness of reuse activities

X

5. Process capability levels and process attributes

The definition of process capability indicators for each process attribute is an integral part of a measurement framework. Process capability indicators such as generic practices and information items are the means to support the judgment of the degree of achievement of the associated process attribute.

This chapter defines the generic practices and information items and their mapping to the process attributes for each capability level defined in the measurement framework [3.2].

Note: Due to lack of a defined process attribute for process capability level 0, no generic practices and information items are defined.

Process

capability

level

Process attribute ID

Process attribute name

Process attribute scope

Process achievements

Each process attribute is identified with a unique identifier and name. A process attribute scope statement is provided, and process achievements are defined.

Process

attribute Achievement

indicators

Generic practices

A set of generic practices for the process attribute providing a definition of the activities to be performed to accomplish the process attribute scope and fulfill the process achievements.

The generic practice headers are summarized at the end of a process to demonstrate their relationship to the process attribute achievements.

Output information items

The output information items that are relevant to accomplish the process attribute scope and fulfill the process

achievements are summarized at the end of a process attribute section to demonstrate their relationship to the process achievements.

Note: Refer to Annex B for the characteristics of each information item.

Table 22 — Template for the process description

5.1. Process capability level 0: Incomplete process

The process is not implemented or fails to achieve its process purpose. At this level there is little or no evidence of any systematic achievement of the process purpose.

5.2. Process capability Level 1: Performed process

The implemented process achieves its process purpose. The following process attribute demonstrates the achievement of this level.

5.2.1. PA 1.1 Process performance process attribute

Process attribute ID

PA 1.1

Process attribute name

Process performance process attribute

Process attribute scope

The process performance process attribute is a measure of the extent to which the process purpose is achieved.

Process attribute achievements

  1. The process achieves its defined outcomes.

Generic practices

GP 1.1.1 Achieve the process outcomes Achieve the intent of the base practices.

Produce work products that evidence the process outcomes.

PA 1.1

Process performance process attribute

Achievement a

Output Information Items

Process specific information items, as described in chapter 4

X

Generic practices

GP 1.1.1 Achieve the process outcomes

X

5.3. Process capability Level 2: Managed process

The following process attributes, together with the previously defined process attribute, demonstrate the achievement of this level.

5.3.1. PA 2.1 Process performance management process attribute

Process attribute ID

PA 2.1

Process attribute name

Process performance management process attribute

Process attribute scope

The performance management process attribute is a measure of the extent to which the performance of the process is managed.

Process attribute achievements

  1. Strategy for the performance of the process is defined based on identified objectives.

  2. Performance of the process is planned.

  3. Performance of the process is monitored and adjusted to meet the planning.

  4. Needs for human resources including responsibilities and authorities for performing the process are determined.

  5. Needs for physical and material resources are determined.

  6. Persons performing the process are prepared for executing their responsibilities.

  7. Physical and material resources for performing the process are identified, made available, allocated and used.

  8. Interfaces between the involved parties are managed to ensure both effective communication and the assignment of responsibilities.

Generic practices

GP 2.1.1: Identify the objectives and define a strategy for the performance of the process.

The scope of the process activities including the management of process performance and the management of work products are determined.

Corresponding results to be achieved are determined.

Process performance objectives and associated criteria are identified.

Note 1: Budget targets and delivery dates to the customer, targets for test coverage and process lead time are examples for process performance objectives.

Note 2: Performance objectives are the basis for planning and monitoring.

Assumptions and constraints are considered when identifying the performance objectives.

Approach and methodology for the process performance is determined.

Note 3: A process performance strategy may not necessarily be document-ed specifically for each process. Elements applicable for multiple processes may be documented jointly, e.g, as part of a common project handbook or in a joint test strategy.

GP 2.1.2: Plan the performance of the process.

The planning for the performance of the process is established according to the defined objectives, criteria, and strategy.

Process activities and work packages are defined.

Estimates for work packages are identified using appropriate methods.

Note 4: Schedule and milestones are defined.

GP 2.1.3: Determine resource needs.

The required amount of human resources, and experience, knowledge and skill needs for the for process performance are determined based on the planning.

The needs for physical and material resources are determined based on the planning.

Note 5: Physical and material resources may include equipment, laboratories, materials, tools, licenses etc.

Required responsibilities and authorities to perform the process, and to manage the corresponding work products are determined.

Note 6: The definition of responsibilities and authorities does not necessarily require formal role descriptions.

GP 2.1.4: Identify and make available resources.

The individuals performing and managing the process are identified and allocated according to the determined needs.

The individuals performing and managing the process are being qualified to execute their responsibilities.

Note 7: Qualification of individuals may include training, mentoring, or coaching.

The other resources, necessary for performing the process are identified, made available, allocated and used according to the determined needs.

GP 2.1.5: Monitor and adjust the performance of the process.

Process performance is monitored to identify deviations from the planning.

Appropriate actions in case of deviations from the planning are taken.

The planning is adjusted as necessary.

GP 2.1.6: Manage the interfaces between involved parties.

The individuals and groups including required external parties involved in the process performance are determined.

Responsibilities are assigned to the relevant individuals or parties.

Communication mechanisms between the involved parties are determined.

Effective communication between the involved parties is established and maintained.

PA 2.1 Process Performance Management

Achievement 1

Achievement 2

Achievement 3

Achievement 4

Achievement 5

Achievement 6

Achievement 7

Achievement 8

Output Information Items

19-01 Process performance strategy

X

18-58 Process performance objectives

X

14-10 Work package

X

08-56 Schedule

X

X

13-14 Progress status

X

17-55 Resource needs

X

X

08-61 Resource allocation

X

X

08-62 Communication matrix

X

13-52 Communication evidence

X

Generic Practices

GP 2.1.1: Identify the objectives and define a strategy for the performance of the process

X

GP 2.1.2: Plan the performance of the process

X

GP 2.1.3: Determine resource needs

X

X

GP 2.1.4: Identify and make available resources

X

X

GP 2.1.5: Monitor and adjust the performance of the process

X

GP 2.1.6: Manage the interfaces between involved parties

X

5.3.2. PA 2.2 Work product management process attribute

Process attribute ID

PA 2.2

Process attribute name

Work product management process attribute

Process attribute scope

The work product management process attribute is a measure of the extent to which the work products produced by the process are appropriately managed.

Process attribute achievements

  1. Requirements for the work products of the process are defined.

  2. Requirements for storage and control of the work products are defined.

  3. The work products are appropriately identified, stored, and controlled.

  4. The work products are reviewed and adjusted as necessary to meet requirements.

Generic practices

GP 2.2.1 Define the requirements for the work products.

The requirements for the content and structure of the work products to be produced are defined.

Quality criteria for the work products are identified.

Appropriate review and approval criteria for the work products are defined.

Note 1: Possible sources of documentation requirements may be e.g., best practices or lessons learned from other projects, standards, organization requirements, customer requirements, etc.

Note 2: There may be types of work products for which no review or approval is required, thus then there would be no need to define the corresponding criteria.

GP 2.2.2 Define the requirements for storage and control of the work products.

Requirements for the storage and control of the work products are defined, including their identification and distribution.

Note 3: Possible sources for the identification of requirements for storage and control may be e.g., legal requirements, data policies, best practices from other projects, tool related requirements, etc.

Note 4: Examples for work product storage are files in a file system, ticket in a tool, Wiki entry, paper documents etc.

Note 5: Where status of a work product is required in base practices, this should be managed via a defined status model.

GP 2.2.3 Identify, store and control the work products.

The work products to be controlled are identified.

The work products are stored and controlled in accordance with the requirements.

Change control is established for work products.

Versioning and baselining of the work products is performed in accordance with the requirements for storage and control of the work products.

The work products including the revision status are made available through appropriate mechanisms.

GP 2.2.4 Review and adjust work products.

The work products are reviewed against the defined requirements and criteria.

Resolution of issues arising from work products reviews is ensured.

PA 2.2 Work product management process attribute

Achievement 1

Achievement 2

Achievement 3

Achievement 4

Output Information Items

17-05 Requirements for work products

X

X

18-59 Review and approval criteria for work products

X

18-07 Quality criteria

X

13-19 Review evidence

X

13-08 Baseline

X

16-00 Repository

X

Generic Practices

GP 2.2.1 Define the requirements for the work products

X

GP 2.2.2 Define the requirements for storage and control of the work products

X

GP 2.2.3 Identify, store and control the work products

X

GP 2.2.4 Review and adjust work products.

X

5.4. Process capability Level 3: Established process

The following process attributes, together with the previously defined process attributes, demonstrate the achievement of this level.

5.4.1. PA 3.1 Process definition process attribute

Process attribute ID

PA 3.1

Process attribute name

Process definition process attribute

Process attribute scope

The process definition process attribute is a measure of the extent to which a standard process is maintained to support the deployment of the defined process.

Process attribute achievements

  1. A standard process is developed, established, and maintained that describes the fundamental elements that must be incorporated into a defined process.

  2. The required inputs and the expected outputs for the standard process are defined.

  3. Roles, responsibilities, authorities, and required competencies for performing the standard process are defined.

  4. Tailoring guidelines for deriving the defined process from the standard process are defined.

  5. Required physical and material resources and process infrastructure needs are determined as part of the standard process.

  6. Suitable methods and required activities for monitoring the effectiveness, suitability and adequacy of the process are determined.

Generic practices

GP 3.1.1 Establish and maintain the standard process.

A suitable standard process is developed including required activities and their interactions.

Inputs and outputs of the standard process are defined including the corresponding entry and exit criteria to determine the interactions and sequence with other processes.

Process performance roles are identified and assigned to the standard process activities including their type of involvement, responsibilities, and authorities.

Note 1: An example for describing the involvement of the process roles in the activities is a RASI/RASIC representation.

Suitable guidance, procedures, and templates are provided to support the execution of the process as needed.

Note 2: Procedures may also include description of specific methods to be used.

Appropriate tailoring guidelines including predefined unambiguous criteria as well as predefined and unambiguous proceedings are defined based on identified deployment needs and context of the standard process.

The standard process is maintained according to corresponding feedback from the monitoring of the deployed processes.

Note 3: For guidance on how to perform process improvements see the Process Improvement process (PIM.3).

GP 3.1.2 Determine the required competencies.

Required competencies, skills, and experience for performing the standard process are determined for the identified roles.

Appropriate qualification methods to acquire the necessary competencies and skills are determined, maintained, and made available for the identified roles.

Note 4: Qualification methods are e.g., trainings, mentoring, self-study.

Note 5: Preparation includes e.g., identification or definition of trainings, mentoring concepts, selflearning material.

GP 3.1.3 Determine the required resources.

Required physical and material resources and process infrastructure needs for performing the standard process are determined.

Note 6: This may include e.g., facilities, tools, licenses, networks, services, and samples supporting the establishment of the required work environment.

GP 3.1.4 Determine suitable methods to monitor the standard process.

Methods and required activities for monitoring the effectiveness and adequacy of the standard process are determined.

Note 7: Methods and activities to gather feedback regarding the standard process may be lessons learned, process compliance checks, internal audits, management reviews, change requests, reflection of state-of-the-art such as applicable international standards, etc.

Appropriate criteria and information needed to monitor the standard process are defined.

Note 8: Information about process performance may be of qualitative or quantitative nature.

PA 3.1 Process definition process attribute

Achievement 1

Achievement 2

Achievement 3

Achievement 4

Achievement 5

Achievement 6

Output Information Items

06-51 Tailoring guideline

X

08-63 Process monitoring method

X

10-00 Process description

X

X

10-50 Role description

X

10-51 Qualification method description

X

10-52 Process resource and infrastructure description

X

Generic Practices

GP 3.1.1 Establish and maintain the standard process

X

X

X

X

GP 3.1.2 Determine the required competencies

X

GP 3.1.3 Determine the required resources

X

GP 3.1.4 Determine suitable methods to monitor the standard process

X

5.4.2. PA 3.2 Process deployment process attribute

Process attribute ID

PA 3.2

Process attribute name

Process deployment process attribute

Process attribute scope

The process deployment process attribute is a measure of the extent to which the standard process is deployed as a defined process to achieve its process outcomes.

Process attribute achievements

  1. A defined process is deployed based upon an appropriately selected and/or tailored standard process.

  2. Assignment of persons necessary for performing the defined process to roles is performed and communicated.

  3. Required education, training and experience is ensured and monitored for the person(s) assigned to the roles.

  4. Required resources for performing the defined process are made available, allocated, and maintained.

  5. Appropriate information is collected and analyzed as a basis for understanding the behavior of the process.

Generic practices

GP 3.2.1 Deploy a defined process that satisfies the context specific requirements of the use of the standard process.

The defined process is appropriately selected and/or tailored from the standard process.

Conformance of defined process with standard process requirements and tailoring criteria is verified.

The defined process is used as managed process to achieve the process outcomes.

Note 1: Changes in the standard process may require updates of the defined process.

GP 3.2.2 Ensure required competencies for the defined roles.

Human resources are allocated to the defined roles according to the required competencies and skills.

Assignment of persons to roles and corresponding responsibilities and authorities for performing the defined process are communicated.

Gaps in competencies and skills are identified, and corresponding qualification measures are initiated and monitored.

Availability and usage of the project staff are measured and monitored.

GP 3.2.3 Ensure required resources to support the performance of the defined process.

Required information to perform the defined process is made available, allocated and used.

Required physical and material resources, process infrastructure and work environment are made available, allocated and used.

Availability and usage of resources are measured and monitored.

GP 3.2.4 Monitor the performance of the defined process.

Information is collected and analyzed according to the determined process monitoring methods to understand the effectiveness and adequacy of the defined process.

Results of the analysis are made available to all effected parties and used to identify where continual improvement of the standard and/or defined process can be made.

Note 2: For guidance on how to perform process improvements see the Process Improvement process (PIM.3).

PA 3.2 Process deployment process attribute

Achievement 1

Achievement 2

Achievement 3

Achievement 4

Achievement 5

Output Information Items

10-00 Process description

X

15-54 Tailoring documentation

X

14-53 Role assignment

X

X

13-55 Process resource and infrastructure documentation

X

03-06 Process performance information

X

Generic Practices

GP 3.2.1 Deploy a defined process

X

GP 3.2.2 Ensure required competencies

X

X

GP 3.2.3 Ensure required resources

X

GP 3.2.4 Monitor the performance of the defined process

X

5.5. Process capability Level 4: Predictable process

The following process attributes, together with the previously defined process attributes, demonstrate the achievement of this level.

5.5.1. PA 4.1 Quantitative analysis process attribute

Process attribute ID

PA 4.1

Process attribute name

Quantitative analysis process attribute

Process attribute scope

The quantitative analysis process attribute is a measure of the extent to which information needs are defined, relationships between process elements are identified and data are collected.

Process attribute achievements

  1. Process information needs in support of relevant defined quantitative business goals are established.

  2. Measurable relationships between process elements that contribute to the process performance, and data collection techniques and data collection frequency, are identified.

  3. Process measurement objectives are derived from process information needs.

  4. Techniques for analyzing the collected data are selected.

  5. Quantitative control limits for process performance in support of relevant business goals are established.

  6. Results of measurement are collected, validated and reported in order to monitor the extent to which the quantitative targets/objectives for process performance are met.

Note: Information needs typically reflect management, technical, project, process or product needs.

Generic practices

GP 4.1.1 Identify business goals.

Business goals are identified that are supported by the quantitatively measured process.

GP 4.1.2 Establish process information needs.

Stakeholders of the identified business goals and the quantitatively measured process are identified, and their information needs are defined and agreed.

GP 4.1.3 Identify measurable relationships between process elements.

Identify the relationships between process elements, or sets of process elements, which contribute to the process information needs.

Note 1: Examples of process elements are work products, activities, tasks.

GP 4.1.4 Derive process measurement approach and select analysis techniques.

Based on the measurable relationships of process elements, or set of process elements, the process measurement metrics are derived to satisfy the established process information needs.

Frequency of data collection is defined.

Select analysis techniques, appropriate to collected data.

Algorithms and methods to create derived measurement results from base measures are defined, as appropriate.

Verification mechanism for base and derived measures is defined.

Note 2: Typically, the standard process definition is extended to include the collection of data for process measurement.

GP 4.1.5 Establish quantitative control limits.

Establish quantitative control limits for the derived metrics. Agreement with process stakeholders is established.

GP 4.1.6 Collect product and process measurement results through performing the defined process.

Data collection mechanisms are created for all identified metrics.

Required data is collected across process instances of within the defined frequency and recorded.

Measurement results are analyzed and reported to the identified stakeholders.

Note 3: A product measure can contribute to a process measure, e.g., the productivity of testing characterized by the number of defects found in a given timeframe in relation to the product defect rate in the field.

PA 4.1 Quantitative analysis process attribute

Achievement 1

Achievement 2

Achievement 3

Achievement 4

Achievement 5

Achievement 6

Output Information Items

18-70 Business goals

X

X

07-61 Quantitative process metric

X

X

07-62 Process analysis techniques

X

07-63 Process control limits

X

07-64 Process measurement data

X

Generic Practices

GP 4.1.1 Identify business goals

X

GP 4.1.2 Establish process information needs

X

GP 4.1.3 Identify measurable relationships between process elements

X

GP 4.1.4 Derive process measurement approach and select analysis techniques

X

X

GP 4.1.5 Establish quantitative control limits

X

GP 4.1.6 Collect product and process measurement results through performing the de-fined process

X

5.5.2. PA 4.2 Quantitative control process attribute

Process attribute ID

PA 4.2

Process attribute name

Quantitative control process attribute

Process attribute scope

The quantitative control process attribute is a measure of the extent to which objective data are used to manage process performance that is predictable.

Process attribute achievements

  1. Variations in process performance are identified.

  2. Assignable causes of process variation are determined through analysis of the collected quantitative data.

  3. Distributions that characterize the performance of the process are established.

  4. Corrective actions are taken to address assignable causes of variation.

Generic practices

GP 4.2.1 Identify variations in process performance.

Deviations in the performance of process instances from the established quantitative control limits are determined based on the collected quantitative measurement data.

GP 4.2.2 Identify causes of variation.

The determined deviations in process performance are analyzed to identify potential cause(s) of variation using the defined analysis techniques.

Distributions are used to quantitatively understand the variation of process performance under the influence of potential causes of variation.

Consequences of process variation are analyzed.

GP 4.2.3 Identify and implement corrective actions to address assignable causes.

Results are provided to those responsible for taking action.

Corrective actions are determined and implemented to address each assignable cause of variation.

Corrective action results are monitored and evaluated to determine their effectiveness.

Note 1: Assignable cause may indicate a possible problem in the defined process.

PA 4.2 Quantitative control process attribute

Achievement 1

Achievement 2

Achievement 3

Achievement 4

Output Information Items

15-57 Quantitative process analysis results

X

X

X

08-66 Measures against deviations in quantitative process analysis

X

Generic Practices

GP 4.2.1 Identify variations in process performance

X

GP 4.2.2 Identify causes of variation

X

X

GP 4.2.3 Identify and implement corrective actions to address assignable causes

X

5.6. Process capability Level 5: Innovating process

The following process attributes, together with the previously defined process attributes, demonstrate the achievement of this level.

5.6.1. PA 5.1 Process innovation process attribute

Process attribute ID

PA 5.1

Process attribute name

Process innovation process attribute

Process attribute scope

The process innovation process attribute is a measure of the extent to which changes to the process are identified from investigations of innovative approaches to the definition and deployment of the process.

Process attribute achievements

  1. Process innovation objectives are defined that support the relevant business goals.

  2. Quantitative data are analyzed to identify opportunities for innovation.

  3. Innovation opportunities derived from new technologies and process concepts are identified.

Generic practices

GP 5.1.1 Define the process innovation objectives for the process that support the relevant business goals.

New business visions and goals are analyzed to give guidance for new process objectives and potential areas of process innovation.

Quantitative and qualitative process innovation objectives are defined and documented.

GP 5.1.2 Analyze quantitative data of the process.

Common causes of variation in process performance across process instances are identified and analyzed to get a quantitative understanding of their impact.

GP 5.1.3 Identify innovation opportunities.

Identify opportunities for innovation based on the quantitative understanding of the analyzed data.

Industry best practices, new technologies and process concepts are identified and evaluated. Feedback on opportunities for innovation is actively sought.

Emergent risks are considered in evaluating improvement opportunities.

PA 5.1 Process innovation process attribute

Achievement 1

Achievement 2

Achievement 3

Output Information Items

18-80 Improvement opportunity

X

X

15-58 Common cause of variation analysis results

X

Generic Practices

GP 5.1.1 Define the process innovation objectives for the process that support the relevant business goals

X

GP 5.1.2 Analyze quantitative data of the process

X

GP 5.1.3 Identify innovation opportunities

X

5.6.2. PA 5.2 Process innovation implementation process attribute

Process attribute ID

PA 5.2

Process attribute name

Process innovation implementation process attribute

Process attribute scope

The process innovation process implementation attribute is a measure of the extent to which changes to the definition, management and performance of the process achieves the relevant process innovation objectives.

Process attribute achievements

  1. Impact of all proposed changes is assessed against the objectives of the defined process and standard process.

  2. Implementation of all agreed changes is managed to ensure that any disruption to the process performance is understood and acted upon.

  3. Effectiveness of process change on the basis of quantitative performance and innovation feedback is evaluated.

Generic practices

GP 5.2.1 Define and assess the impact of proposed changes.

Specified changes are assessed against product quality and process performance requirements and goals.

Impact of changes to other defined and standard processes is considered.

Objective priorities for process innovation are established.

Commitment to innovation is demonstrated by organizational management including other relevant stakeholders.

GP 5.2.2 Implement agreed process changes.

A mechanism is established for incorporating accepted changes into the defined and standard process(es) effectively and completely.

Process changes are implemented and effectively communicated to all affected parties.

GP 5.2.3 Evaluate the effectiveness of process change.

Performance and capability of the changed process are measured and compared with historical data.

Performance and capability of the changed process are analyzed to determine whether the process performance has improved with respect to common causes of variations.

Other feedback is recorded, such as opportunities for further innovation of the standard process.

A mechanism is available for documenting and reporting analysis results to stakeholders of standard and defined process.

PA 5.2 Process innovation implementation attribute

Achievement a

Achievement b

Achievement c

Output Information Items

18-81 Improvement evaluation results

X

X

08-66 Measures against deviations in quantitative process analysis

X

X

Generic Practices

GP 5.2.1 Define and assess the impact of proposed changes

X

GP 5.2.2 Implement agreed process changes

X

GP 5.2.3 Evaluate the effectiveness of process change

X

Annex A Conformity statements

Annex A.1 Introduction

The Automotive SPICE process assessment and process reference model are meeting the requirements for conformance defined in ISO/IEC 33004:2015. The process assessment model can be used in the performance of assessments that meet the requirements of ISO/IEC 33002:2015.

This clause serves as the statement of conformance of the process assessment and process reference models to the requirements defined in ISO/IEC 33004:2015. [ISO/IEC 33004:2015, 5.5 and 6.4]

Due to copyright reasons each requirement is only referred by its number. The full text of the requirements can be drawn from ISO/IEC 33004:2015.

Annex A.2 Conformance to the requirements for process reference models

Clause 5.3, “Requirements for process reference models”

The following information is provided in chapter 1 and 3 of this document:

  • the declaration of the domain of this process reference model;

  • the description of the relationship between this process reference model and its intended context of use; and

  • the description of the relationship between the processes defined within this process reference model.

The descriptions of the processes within the scope of this process reference model meeting the requirements of ISO/IEC 33004:2015 clause 5.4 is provided in chapter 4 of this document. [ISO/IEC 33004:2015, 5.3.1]

The relevant communities of interest and their mode of use and the consensus achieved for this process reference model is documented in the copyright notice and the scope of this document. [ISO/IEC 33004:2015, 5.3.2]

The process descriptions are unique. The identification is provided by unique names and by the identifier of each process of this document. [ISO/IEC 33004:2015, 5.3.3]

Clause 5.4, “Process descriptions”

These requirements are met by the process descriptions in chapter 4 of this document. [ISO/IEC 33004:2015, 5.4]

Annex A.3 Conformance to the requirements for process assessment models

Clause 6.1, “Introduction”

The purpose of this process assessment model is to support assessment of process capability within the automotive domain using the defined process measurement framework. [ISO/IEC 33004:2015, 6.1]

Clause 6.2, “Process assessment model scope”

The process scope of this process assessment model is defined in the process reference model included in chapter 3.1 of this document. The Automotive SPICE process reference model is satisfying the requirements of ISO/IEC 33004:2015, clause 5 as described in Annex A.2.

The process capability scope of this process assessment model is defined in the process measurement framework, which defines a process measurement framework for process capability satisfying the requirements of ISO/IEC 33003:2015. [ISO/IEC 33004:2015, 6.2]

Clause 6.3, “Requirements for process assessment models”

The Automotive SPICE process assessment model is related to process capability. [ISO/IEC 33004:2015, 6.3.1]

This process assessment model incorporates the defined process measurement framework, which satisfies the requirements of ISO/IEC 33003:2015. [ISO/IEC 33004:2015, 6.3.2]

This process assessment model is based on the Automotive SPICE Reference Model included in this document.

This process assessment model is based on the defined Measurement Framework. [ISO/IEC 33004:2015, 6.3.3]

The processes included in this process assessment model are identical to those specified in the Process Reference Model

[ISO/IEC 33004:2015, 6.3.4]

For all processes in this process assessment model all levels defined in the process measurement framework are addressed.

[ISO/IEC 33004:2015, 6.3.5]

This process assessment model defines

  • the selected process quality characteristic;

  • the selected process measurement framework;

  • the selected process reference model(s);

  • the selected processes from the process reference model(s) in chapter 3 of this document.

[ISO/IEC 33004:2015, 6.3.5 a-d]

In the capability dimension, this process assessment model addresses all of the process attributes and capability levels defined in the process measurement framework.

[ISO/IEC 33004:2015, 6.3.5 e]

Clause 6.3.1, “Assessment indicators”

Note: Due to an error in numbering in the published version of ISO/IEC 33004:2015 the following reference numbers are redundant to those stated above. To refer to the correct clauses from ISO/IEC 33004:2015, the text of clause heading is additionally specified for the following three requirements.

The Automotive SPICE process assessment model provides a two-dimensional view of process capability for the processes in the process reference model, through the inclusion of assessment indicators as defined in chapter 3.3. The assessment indicators used are:

  • Base practices and information items

[ISO/IEC 33004:2015, 6.3.1 a, “Assessment indicators”]

  • Generic practices and information items

[ISO/IEC 33004:2015, 6.3.1 b, “Assessment indicators”]

Clause 6.3.2, “Mapping process assessment models to process reference models”

The mapping of the assessment indicators to the purpose and process outcomes of the processes in the process reference model is included in the tables for each process in chapter 4.

The mapping of the assessment indicators to the process attributes in the process measurement framework including all of the process attribute achievements is included in the tables for each process attribute in chapter 5.

[ISO/IEC 33004:2015, 6.3.2, “Mapping process assessment models”]

Clause 6.3.3, “Expression of assessment results”

The process attributes and the process attribute ratings in this process assessment model are identical to those defined in the measurement framework. As a consequence, results of assessments based upon this process assessment model are expressed directly as a set of process attribute ratings for each process within the scope of the assessment. No form of translation or conversion is required.

[ISO/IEC 33004:2015, 6.3.3, “Expression of assessment results”]

Annex A.4 Conformance to the requirements for measurement frameworks

The measurement framework defined in Automotive SPICE 4.0 is an adaption of the measurement framework defined in ISO/IEC 33020:2019. The following modifications have been performed:

  • Renaming of the Process attribute titles.

  • Changes in the generic practices.

  • Assignments of indicators to process attribute achievements.

Conceptualization, Construct definition and Operationalization relevant for conformity to ISO/IEC 33003:2015 has been adopted from ISO/IEC 33020:2019.

The conformity of the Automotive SPICE Measurement Framework is thereby confirmed based on the existing conformance statement of 33020:2019.

Annex B Information Item Characteristics

Characteristics of information items are defined using the schema in table B.1. See Section 3.3.2 on the definition and explanation on how to interpret information items and their characteristics.

Table B. 1 — Structure of information item characteristics (IIC) table

Information item identifier

An identifier number for the information item which is used to reference the information item.

Information item name

Provides an example of a typical name associated with the information item characteristics. This name is provided as an identifier of the type of information item the practice or process might produce. Organizations may call these information items by different names. The name of the information item in the organization is not significant. Similarly, organizations may have several equivalent information items which contain the characteristics defined in one information item type. The formats for the information items can vary. It is up to the assessor and the organizational unit coordinator to map the actual information items produced in their organization to the examples given here.

Information item characteristics

Provides examples of the potential characteristics associated with the information item types. The assessor may use these in evaluating the samples provided by the organizational unit. It is not intended to use the listed characteristics as a checklist. Some characteristics may be contained in other work products, as it would be found appropriate in the assessed organization.

Table B. 2 — Information Item Characteristics

ID

Name

Characteristics

01-03

Software component

  • Software element in the software architecture above the software unit level.

  • Represented by a design model element or executable code such as libs or scripts and a configuration description, if applicable.

01-50

Integrated software

  • Software executable (e.g, simulator with stubbing, debug-able, object code) including:

  • application parameter files (being a technical implementation solution for configurability-oriented requirements)

  • all configured software elements

01-52

Configuration item list

  • Items under configuration control

  • The name of work products and an associated reference (to file, to tool artifact)

  • Configuration item attributes and properties

01-53

Trained ML model

  • The trained ML model is the output of the training process. It consists of the software representing the ML architecture, the set of weights which were optimized during the training, and the final set of hyperparameters.

01-54

Hyperparameter

  • Hyperparameters are used to control the ML model which has to be trained, e.g.:

    • Learn rate of training

    • Scaling of network (number of layers or neurons per layer)

    • Loss function

  • Minimum characteristics:

    • Description

    • Initial value

    • Final value upon communicating the results of the ML training

02-01

Commitment / agreement

  • Signed off by all parties involved in the commitment/agreement

  • Establishes what the commitment is for

  • Establishes the resources required to fulfill the commitment, such as:

    • time

    • people

    • budget

    • equipment

    • facilities

03-06

Process performance information

  • Measurements about defined quantitative or qualitative measurable indicators, that match defined information needs.

  • Measurement metrics for the calculation of the quantitatively or qualitatively measurable indicators

  • Data comparing process performance against expected levels

  • Examples for project performance information:

    • resource utilization against established target

    • time schedule against established target

    • activity or task completion criteria met

    • defined input and output work products available

    • process quality against quality expectations and/or criteria

    • product quality against quality expectations and/or criteria

    • highlight product performance issues, trends

  • Examples for service level performance information:

    • references any goals established

    • real time metrics related to aspects such as:

    • capacity

    • throughput

    • operational performance

    • operational service

    • service outage time

    • up time

    • job run time

03-50

Verification Measure data

  • Verification measure data are data recorded during the execution of a verification measure, e.g.:

  • for test cases: raw data, logs, traces, tool generated outputs

  • measurements: values

  • calculations: values

  • simulations: protocol

  • reviews such as optical inspections à findings record

  • analyses: values

03-51

ML data set

  • Selection of ML Data for e.g., ML model training (ML Training and Validation Data Set) or test of the trained and deployed ML model (ML Test Data Set).

03-53

ML data

  • Datum to be used for Machine Learning. The datum has to be attributed by metadata, e.g., unique ID and data characteristics. Examples:

  • Visual data like a photo or videos (but a video could also be considered as sequence of photos depending on the intended use)

  • Audio recording

  • Sensor data

  • Data created by an algorithm

  • Data might be processed to create additional data. E.g., processing could add noise, change colors or merge pictures.

03-54

Hardware production data

  • Consists of bill of materials

  • Consists of layout e.g, GERBER data

  • Specifies requirements for EOL test e.g.:

    • Test type (AOI, ICT, boundary scan)

    • Test coverage

    • Electrical loads

    • Acceptance criteria

  • In case of semiconductor development: mask data (GDS2)

04-04

Software architecture

  • A justifying rationale for the chosen architecture.

  • Individual functional and non-functional behavior of the software component

  • Settings for application parameters (being a technical implementation solution for configurability-oriented requirements)

  • Technical characteristics of interfaces for relationships between software components such as:

    • Synchronization of Processes and tasks

    • Programming language call

    • APIs

    • Specifications of SW libraries

    • Method definitions in an object- oriented class definitions or UML/SysML interface classes

    • Callback functions, “hooks”

  • Dynamics of software components and software states such as:

    • Logical software operating modes (e.g, start-up, shutdown, normal mode, calibration, diagnosis, etc.)

    • intercommunication (processes, tasks, threads) and priority

    • time slices and cycle time

    • interrupts with their priorities

    • interactions between software components

  • Explanatory annotations, e.g, with natural language, for single elements or entire diagrams/models.

04-05

Software detailed design

  • Elements of a software detailed design:

    • Control flow definition

    • Format of input/output data

    • Algorithms

    • Defined data structures

    • Justified global variables

    • Explanatory annotations, e.g, with natural language, for single elements or entire diagrams/models

  • Examples for expression languages, depending on the complexity or criticality of a software unit:

    • natural language or informal languages

    • semi-formal languages (e.g, UML, SysML)

    • formal languages (e.g, model-based approach)

04-06

System architecture

  • A justifying rationale for the chosen architecture.

  • Individual behavior of system elements

  • Interrelationships between system elements

Settings for system parameters (such as application parameters) Manual/human control actions, e.g., according to STPA

  • Interface Definitions:

    • Technical characteristics of interfaces for relationships between two system elements

  • Interfaces between system elements e.g.:

    • bus interfaces (CAN, MOST, LIN, Flexray etc.)

    • thermal influences

    • hardware-software-interfaces (HSI), see below

    • electromagnetic interfaces

    • optical interfaces

    • hardware-mechanical-interfaces (e.g., a cable satisfying both mechanical and electrical requirements, housing interface to a PCB)

    • hardware-mechanical interconnection technology such as connectors, pressfit

    • creepage and clearance distances

  • Fixations such as adhesive joints, screw bolts/fitting, riveted bolts, welding

  • System interfaces related to EE Hardware e.g.:

    • analogue or digital interfaces (PWM, I/O) and their pin configurations

    • SPI bus, I2C bus, electrical interconnections

    • placement, e.g., thermal interfaces between hardware elements (heat dissipation)

    • soldering

    • creepage and clearance distances

  • Interfaces for mechanical engineering e.g.:

    • friction

    • thermal influences

    • tolerances

    • clutches

    • fixations such as adhesive joints, screw bolts/fitting, riveted bolts, welding

    • forces (as a result of e.g., vibrations or friction)

    • placement

    • shape

    • A hardware-software interface, e.g.:

    • connector pin configurations and floating IOs for µCs/MOSFETs

    • signal scaling & resolution to be reflected by the application software

  • Mechanical-hardware interfaces e.g.

    • such as mechanical dimensioning

    • positioning of connectors

    • positioning of e.g., hall sensors in relation to the bus-bar

    • tolerances

  • Dynamics of system elements and system states:

    • Description of the system states and operation modes (startup, shutdown, sleep mode, diagnosis/calibration mode, production mode, degradation, emergency such as “limp-home”, etc.)

    • Description of the dependencies among the system components regarding the operation modes

    • Interactions between system elements such as inertia of mechanical components to be reflected by the ECU, signal propagation and processing time through the hardware and software and e.g., bus systems

  • Explanatory annotations, e.g., with natural language, for single elements or entire diagrams/models.

04-51

ML architecture

  • An ML architecture is basically a special part of a software architecture (see 04-04). Additionally

  • ML architecture describes the overall structure of the ML-based software element

  • ML architecture specifies ML architectural elements including an ML model and other ML architectural elements, provided to train, deploy, and test the ML model.

  • describes interfaces within the ML-based software element and to other software elements

  • ML architecture describes details of the ML model like used layers, activation functions, loss function, and backpropagation

  • ML architecture contains defined hyperparameter ranges and initial values for training start

  • resource consumption objectives are defined

  • ML architecture contains allocated ML requirements

04-52

Hardware architecture

  • Describes the initial floorplan and the overall hardware structure

  • Identifies the required hardware components

  • Includes the rationale for chosen options of hardware architecture

  • Identifies own developed and supplied hardware components

  • Identifies the required internal and external hardware component interfaces

  • Specifies the interfaces of the hardware components

  • Specifies the dynamic behavior

  • Identifies the relationship and dependency between hardware components

  • Describes all hardware variants to be developed

  • Describes power supply, thermal and grounding concepts

04-53

Hardware detailed design

  • Describes the interconnections between the hardware parts

  • Specifies the interfaces of the hardware parts

  • Specifies the dynamic behavior (examples are: transitions between electrical states of hardware parts, power-up and power-down sequences, frequencies, modulations, signal delays, debounce times, filters, short circuit behavior, self-protection)

  • Describes the conclusions and decisions based on e.g., analysis reports, datasheets, application notes

  • Describes the constraints for layout

04-54

Hardware Schematics

  • Identifies the hardware parts

  • Specifies the connections of the hardware parts

  • Specifies the unique identification of all hardware parts

  • Specifies unique variant identification

04-55

Hardware Layout

  • Specifies the placement of the hardware parts and labels

  • Specifies manufacturing data e.g., circuit paths (width, routing), vias, testing points, number of layers, drillings, material of the PCB, shape, soldering resist mask, PCB coating

  • Specifies a unique layout identification

04-56

Hardware element interface

  • is defined by output, input, type, and electrical characteristics including signal tolerances.

  • Examples of interfaces are

    • high level interfaces like SPI, I2C, CAN, LIN, Ethernet

    • electrical interconnections

    • thermal interfaces between hardware elements (heat dissipation)

06-04

Training material

  • Updated and available for new releases

  • Coverage of system, application, operations, maintenance as appropriate to the application

  • Course listings and availability

06-50

Integration sequence instruction

  • Identification of required physical elements (e.g., hardware, mechanical, wiring elements), and software executables and application parameters (being a technical implementation solution for configurability-oriented requirements)

  • necessary sequence or ordering of integration

  • preconditions for starting system integration

06-51

Tailoring guideline

  • Criteria for tailoring,

  • Proceeding of tailoring describing how to derive and document the defined process from the standard process including responsibility for tailoring and corresponding approval

  • Requirements for the defined process to ensure integrity and consistency of the defined process

  • Subset of process assets that is essential for the defined process

06-52

Backup and recovery

mechanism information

  • Description / confirmation of existing backup and recovery mechanisms

  • References to corresponding procedures or regulations

07-04

Process metric

  • Measurements about the process’ performance:

    • ability to produce sufficient work products

    • adherence to the process

    • time it takes to perform process

    • defects related to the process

  • Measures the impact of process change

  • Measures the efficiency of the process

07-05

Project metric

  • Monitors key processes and critical tasks, provides status information to the project on:

    • project performance against established plan

    • resource utilization against established plan

    • time schedule against established plan

    • process quality against quality expectations and/or criteria

    • product quality against quality expectations and/or criteria

    • highlight product performance problems, trends

  • Measures the results of project activities:

    • tasks are performed on schedule

    • product’s development is within the resource commitments allocated

  • References any goals established

07-06

Quality metric

  • Measures quality attributes of the work products defined:

    • functionality

    • reliability

    • usability

    • efficiency

    • maintainability

    • portability

  • Measures quality attributes of the “end customer” quality perception

Note: Refer ISO/IEC 25010 for detailed information on measurement of product quality.

07-08

Service level metric

  • Real time metrics taken while a system is operational, it measures the system’s performance or expected service level

  • Identifies aspects such as:

    • capacity

    • throughput

    • operational performance

    • operational service

    • service outage time

    • up time

    • job run time

07-51

Measurement result

Result of gathering qualitative or quantitative data, e.g.,

Process metric

  • Measurements about the process’ performance:

    • ability to produce sufficient work products

    • adherence to the process

    • time it takes to perform process

    • defects related to the process

  • Measures the impact of process change

  • Measures the efficiency of the process

Project metric

  • Monitors key processes and critical tasks, provides status information to the project on:

    • project performance against established plan

    • resource utilization against established plan

    • time schedule against established plan

    • process quality against quality expectations and/or criteria

    • product quality against quality expectations and/or criteria

    • highlight product performance problems, trends

  • Measures the results of project activities:

  • tasks are performed on schedule

  • product’s development is within the resource commitments allocated

  • References any goals established

Quality metric

  • Measures quality attributes of the work products defined:

    • functionality

    • reliability

    • usability

    • efficiency

    • maintainability

    • portability

  • Measures quality attributes of the “end customer” quality perceptionService level metric

  • Benchmarking data

  • Customer satisfaction survey

07-61

Quantitative process metric

  • Quantitatively measurable indicators that match information needs derived from business goals

  • Relation of the quantitatively measurable indicators to process elements in process descriptions or repositories and tools

  • Process measurement metrics for the calculation of the quantitatively measurable indicators, based on data from related process elements, repositories, or tools

07-62

Process analysis technique

  • Methods for statistical analysis of process. data

  • Frequency of data collection.

07-63

Process control limits

  • Quantitative control limits for the quantitative process metrics

07-64

Process measurement data

  • Data collected across process instances

  • Attributes of data, e.g., timestamps

  • Relation to process measurement metrics

  • Storage and retrieval

  • Effective controls over access

15-57

Quantitative process analysis results

  • Deviations, and distributions, of the quantitative performance of individual process instances performance from the established quantitative control limits (special causes of variations)

08-66

Measures against deviations in quantitative process analysis

  • Definition of counter measures actions to address each assignable cause of special causes of variation, or common causes of variation

  • Effective implementation of these counter measures

15-58

Common cause of variation analysis results

  • Identification of common causes

  • deviations of the quantitative performance of all process instances from the established quantitative control limits

  • distributions of the quantitative performance of all process instances within established quantitative control limits

08-53

Scope of work

  • Summary of deliverables for a project

  • Intended use for the deliverables

  • Main functions to be realized

  • Target delivery date and major milestones

  • Work products and activities that are not in scope of the project as needed

  • Target markets

  • Applicable standards and legal requirements

  • Reuse options

  • Integration of third party deliveries

08-54

Feasibility analysis

  • Statement about the ability of the project to achieve the project objectives with available resources

08-55

Risk measure

  • Identifies

    • the risk to be mitigated, avoided, or shared (transferred)

    • the activities to mitigate, avoid, or share (transfer) the risk

    • the originator of the measure

    • criteria for successful implementation

    • criteria for cancellation of activities

    • frequency of monitoring

  • Risk treatment alternatives:

    • treatment option selected- avoid/reduce/transfer

    • alternative descriptions

    • recommended alternative(s)

    • justifications

08-56

Schedule

  • Identifies the activities to be performed

  • Identifies the expected, and actual, start and completion date for required activities against progress/completion of activities

  • Identifies dependencies between activities and critical path

  • Has a mapping to scheduled resources and input data

  • Identifies resource allocation, resource workload, and critical resources

NOTE: A schedule is consistent with the defined work packages, see 1410

08-57

Validation Measure Selection Set

  • Include criteria for re-validation in the case of changes (regression).

  • Identification of validation measures, also for regression

08-58

Verification Measure Selection Set

  • Include criteria for re-verification in the case of changes (regression).

  • Identification of verification measures, also for regression testing

08-59

Validation Measure

  • A validation measure can be a test case, a measurement, a simulation, an emulation, or an end user survey

  • The specification of a validation measure includes

    • pass/fail criteria for validation measures (completion and end criteria)

    • a definition of entry and exit criteria for the validation measures, and abort and re-start criteria

  • Techniques

  • Necessary validation environment & infrastructure

  • Necessary sequence or ordering

08-60

Verification Measure

  • A verification measure can be a test case, a measurement, a calculation, a simulation, a review, an optical inspection, or an analysis

  • The specification of a verification measure includes

    • pass/fail criteria for verification measures (test completion and ending criteria)

    • a definition of entry and exit criteria for the verification measures, and abort and re-start criteria

  • Techniques (e.g., black-box and/or white-box-testing, equivalence classes and boundary values, fault injection for Functional Safety, penetration testing for Cybersecurity, back-to- back testing for modelbased development, ICT)

  • Necessary verification environment & infrastructure

  • Necessary sequence or ordering

08-61

Resource allocation

  • Detailed / named resources are allocated to process tasks

  • Overall resource workload is considered (e.g., allocation of resources to multiple projects)

NOTE: Work breakdown structure may be used to refine the detailed resource allocation

NOTE: A resource allocation may be integrated in a/ be a part of the schedule, see 08-56

NOTE: Resources to be allocated are e.g., personnel/human resources for project roles and physical and material resources such as (special/limited) equipment, tool, licenses, test hardware, test vehicle, climate chambers etc.

08-62

Communication matrix

  • List of relevant process internal / external stakeholders

  • Roles and contact information of the parties involved

  • Definition of required interfaces between stakeholders

  • Communication subject

  • Communication means and frequency

  • Documentation needs of the communication (e.g., type of communication record)

08-63

Process Monitoring Method

  • Measures including criteria for monitoring effectiveness, suitability, and adequacy of the standard process

  • Method for collecting and analyzing the monitoring measures

08-64

ML test approach

  • The ML test approach describes

  • ML test scenarios with distribution of data characteristics (e.g., gender, weather conditions, street conditions within the ODD) defined by ML requirements

  • quantity of each ML test scenario inside the test data set

  • expected test result per test datum

  • pass/fail criteria for the ML testing

  • entry and exit criteria for the ML testing

  • the required ML testing infrastructure and environment configuration

08-65

ML training and validation approach

  • The ML Training and Validation approach describes at least:

    • entry and exit criteria of the ML training

    • approaches for hyperparameter tuning / optimization to be used in the training

    • approach for data set creation and modification

    • training environment, including required training hardware (e.g., GPU, or supercomputer to be used)

    • interface adapter for provision of input data and storage of output data

    • if required, actions to organize the data set and training environment

  • The ML training and validation approach may additionally include robustification methods like random dropout

10-00

Process description

  • Process description of a standard or defined process (e.g., after tailoring), including:

  • scope and the intended use of the process

  • process activities including description and dependencies

  • entry and exit criteria such as input information needed and expected outputs for activities

  • Roles assigned to process activities (e.g., as RASIC ) or work products

  • guidelines

  • templates

  • specific methods/work instructions

10-50

Role description

  • Name/identifier (unique within the organization)

  • Assigned activities (e.g., as RASIC)

  • Responsibilities and authorities

  • Required competencies, skills, and experience

10-51

Qualification method description

  • Training courses

  • Training materials

  • Mentoring/coaching concepts

  • Self-learning material

10-52

Process resource and

infrastructure description

  • Required facilities

  • Required tools and corresponding licenses

  • Required networks

  • Required services

  • Required samples

11-03

Release note

  • Coverage for key elements (as appropriate to the application):

  • Description of what is new or changed (including features removed)

  • System information and requirements

  • Identification of conversion programs and instructions

  • Release numbering implementation may include:

    • the major release number

    • the feature release number

    • the defect repair number

    • the alpha or beta release; and the iteration within the alpha or beta release

  • Identification of the component list (version identification included):

    • hardware / software / product elements, libraries, etc.

    • associated documentation list

  • New/changed parameter information (e.g., for application parameters or global variables) and/or commands. Note that application parameters are a technical implementation solution for configurability-oriented requirements)

  • Backup and recovery information

  • List of open known problems, faults, warning information, etc.

  • Identification of verification and diagnostic procedures

  • Technical support information

  • Copyright and license information

  • The release note may include an introduction, the environmental requirements, installation procedures, product invocation, new feature identification and a list of defect resolutions, known defects and workarounds

11-04

Product release package

  • Includes the hardware/software/product

  • Includes and associated release elements such as:

    • system hardware/software/product elements

    • associated customer documentation

    • application parameter definitions defined

    • command language defined

    • installation instructions

    • release letter

11-05

Software Unit

Can be

  • a representation of a software element at the lowest level in a conceptual model, which is decided not to be further subdivided and that is a part of a software component, or

  • a representation of a software unit under verification such as commented source code, auto-code, an object file, a library, an executable, or an executable model as input to verification

11-06

Integrated System

  • Integrated product

  • Application parameter files (being a technical implementation solution for configurability-oriented requirements)

  • All configured elements for the product release are included

11-50

Deployed ML model

  • It is derived from the trained ML model (see 01-53) and is to be integrated into the target system.

  • It may differ from the trained ML model which often requires powerful hardware and uses interpretative languages.

12-03

Reuse candidate

  • Identifies the product to be reused

  • Identifies the responsible person for the products to be reused

  • Identifies the reuse goals and objectives

  • Identifies the list of reuse assets

  • Identifies the issues/risks of reusing the component including specific requirements (hardware, software, resource and other reuse components)

  • Identifies the person who will be qualifying the reuse candidate

13-06

Delivery evidence

  • Evidence of items shipped/delivered electronically to customer

  • Identification of:

    • to whom it was sent

    • address, where delivered

    • delivery date

    • receipt of delivered product

13-07

Problem

  • Identifies the submitter of the problem

  • Identifies the group/person(s) responsible for providing problem resolution

  • Includes a description of the problem

  • Identifies classification of the problem (criticality, urgency, relevance etc.)

  • Identifies the status of the problem

    • States such as “open”, “in review”, “in implementation”, “closed”, “rejected”, “cancelled”, …

    • Transitions between states with conditions and authorities

  • Identifies the expected closure date

13-08

Baseline

  • Identifies a state of one or a set of work products and artifacts which are consistent and complete

  • Basis for next process steps and/or delivery

  • Is unique and may not be changed

NOTE: This should be established before a release to identify consistent and complete delivery

13-09

Meeting support evidence

  • Agenda and minutes that are records that define:

  • purpose of meeting

  • attendees

  • date, place held

  • reference to previous minutes

  • what was accomplished

  • identifies issues raised

  • any open issues

  • next meeting if any

13-13

Product release approval

  • Content information of what is to be shipped or delivered

  • dssd

  • Identification of:

    • for whom it is intended

    • the address where to deliver

    • the date released

    • Evidence of supplier approval

13-14

Progress status

  • Status of a plan(s) (actual against planned) such as:

  • status of actual activities/work packages against planned activities/work package

  • status of actual results against established objectives/goals

  • status of actual resources allocation against planned resources

  • status of actual cost against budget estimates

  • status of actual time against planned schedule

  • status of actual quality against planned quality

  • Record of any deviations from planned activities and reason why

13-16

Change request

  • Identifies purpose of change

  • Identifies requester contact information

  • Impacted system(s)

  • Impact to operations of existing system(s) defined

  • Impact to associated documentation defined

  • Criticality of the request, due date

  • Information supporting the tracking of change requests to closure

    • progress status attribute (e.g., open, allocated, implemented, closed)

    • time stamp of status change

    • person who changed a status

    • rationale for changing a status

13-18

Quality conformance evidence

  • Identifies what tasks/activities/process produce the information

  • Identifies when the data was collected

  • Identifies source of any associated data

  • Identifies the associated quality criteria

  • Identifies any associated measurements using the information

13-19

Review evidence

  • Provides the context information about the review:

    • what was reviewed

    • lists reviewers who attended and their area of responsibility - status of the review

  • Provides information about the scope of the review:

    • checklists

    • review criteria

    • requirements

    • compliance to standards

  • Effort information about:

    • preparation time spent for the review

    • time spent in the review

  • Review findings:

    • non-conformances

    • improvement suggestions

13-24

Validation results

  • Validation data, logs, feedback, or documentation

  • Validation measure passed

  • Validation measure not passed

  • Validation measure not executed, and a rationale

  • Information about the validation execution (date, participants etc.)

  • Abstraction or summary of validation results

13-25

Verification results

  • Verification data and logs

  • Verification measure passed

  • Verification measure not passed

  • Verification measure not executed, and a rationale

  • Information about the verification execution (date, “object-underverification”, etc.)

  • Abstraction or summary of verification results

13-50

ML test results

  • Test data and logs

  • Test data with correct results

  • Test data with incorrect results

  • Test data not executed, and a rationale

  • Information about the test execution (date, participants, model version etc.)

  • Abstraction or summary of ML test results

13-51

Consistency Evidence

  • Demonstrates bidirectional traceability between artifacts or information in artifacts, throughout all phases of the life cycle, by e.g.,

    • tool links

    • hyperlinks

    • editorial references

    • naming conventions

  • Evidence that the content of the referenced or mapped information coheres semantically along the traceability chain, e.g., by

    • performing pair working or group work

    • performing by peers, e.g., spot checks

    • maintaining revision histories in documents

    • providing change commenting (via e.g., meta-information) of database or repository entries

Note: This evidence can be accompanied by e.g., Definition of Done (DoD) approaches.

13-52

Communication Evidence

  • All forms of interpersonal communication such as

  • e-mails, also automatically generated ones

  • tool-supported workflows

  • meeting, verbally or via meeting minutes (e.g., daily standups)

  • podcast

  • blog

  • videos

  • forum

  • live chat

  • wikis

  • photo protocol

13-55

Process resource and infrastructure documentation

  • Information on availability, allocation, and usage of

    • Facilities

    • Tools and corresponding licenses

    • Networks

    • Services

    • Samples

  • for non-standard and critical resources and infrastructure.

14-01

Change history

  • Historical records of all changes made to an object (document, file, software component, etc.):

  • description of change

  • version information about changed object

  • date of change

  • change requester information

  • change control record information

14-02

Corrective action

  • Identifies the initial problem

  • Identifies the ownership for completion of defined action

  • Defines a solution (series of actions to fix problem)

  • Identifies the open date and target closure date

  • Contains a status indicator

  • Indicates follow up audit actions

14-10

Work package

  • Defines activities to be performed

  • Documents ownership for activities e.g., by domains

  • Documents critical dependencies to other work packages

  • Documents input and output work products

  • Documents the critical dependencies between defined work products

  • Information needed to perform these activities

  • Estimates of effort, duration

NOTE: The work package descriptions may be integrated into the/be a part of a schedule, see 08-56

14-50

Stakeholder groups list

  • Identifies:

  • involved parties

  • weight/importance of each stakeholder group

  • representative(s) for each stakeholder group

  • information needs of each stakeholder group

14-53

Role Assignment

  • Assignment of person(s) to roles

  • required competencies vs existing competencies

  • required skills vs existing skills

  • required experience and trainings based on identified competencies / skills gap

14-54

Hardware Bill of materials

  • Uniquely identifies type, supplier, and amount of the complete set of all hardware parts of the hardware

15-06

Project status

  • Status of in regards to progress and consistency of schedule, work item content, tasks, resources (human resources, infrastructure, hardware/materials, budget), skills and competence of human resources

  • planned progress and expenditure against dates/deadlines and actual expenditure

  • reasons for variance from planned progress

  • threats to continued progress

  • issues which may affect the ability of the project to achieve its goals

  • contingency actions

15-07

Reuse analysis evidence

  • Identification of reuse opportunities

  • Identification of constraints for reuse

  • Identification of regression test cases

  • Identification of reuse infrastructure

  • Identification of known defects

15-09

Risk status

  • Identifies the status, or the change, of an identified risk:

  • risk statement

  • risk source

  • risk impact and risk probability

  • categories and risk thresholds, e.g., for prioritization or setting a status

  • risk treatment activities in progress

15-12

Problem status

  • Indicates progress of problem resolution

  • Status of problem e.g.,

  • by problem categories/classification

  • by problem resolution stage

15-13

Assessment/audit report

  • States the purpose of assessment

  • Method used for assessment

  • Requirements used for the assessment

  • Assumptions and limitations

  • Identifies the context and scope information required:

    • date of assessment

    • organizational unit assessed

    • sponsor information

    • assessment team

    • attendees

    • scope/coverage

    • assesses and information

    • assessment tool used

  • Records the result:

    • Data

    • identifies the gaps, potentials, weaknesses or non-conformances that require corrective actions

15-16

Improvement opportunity

  • Identifies what the problem is

  • Identifies what the cause of a problem is

  • Suggest what could be done to fix the problem

  • Identifies the value (expected benefit) in performing the improvement

  • Identifies the penalty for not making the improvement

15-51

Analysis Results

  • Identification of the object under analysis.

  • The analysis criteria used, e.g.:

    • selection criteria or prioritization scheme used

    • decision criteria

    • quality criteria

  • The analysis results, e.g.:

    • what was decided/selected

    • reason for the selection

    • assumptions made

    • potential negative impact

  • Aspects of the analysis may include

    • correctness

    • understandability

    • verifiability

    • feasibility

    • validity

15-52

Verification Results

  • Verification data and logs

  • Verification measure passed

  • Verification measure not passed

  • Verification measure not executed

  • information about the test execution (date, tester name etc.)

  • Abstraction or summary of verification results

15-54

Tailoring documentation

  • Applied criteria for tailoring,

  • Evidence that the defined process is tailored from the standard process according to the defined criteria

15-55

Problem analysis evidence

  • Author and involved parties

  • Date of the analysis

  • Context and root cause of the problem

  • Analysis result may include

    • Impact

    • Potential negative impact

    • Affected parties

    • Potential solution (if known)

15-56

Configuration status

  • Summary of configuration management records including relevant status

  • Analysis of the configuration management overall state

  • Identification of baselines made

16-03

Configuration management system

  • Supports the configuration management for the scope of the configuration item list contents

  • Correct configuration of products

  • Can recreate any release or test configuration

  • Ability to report configuration status

  • Has to cover all relevant tools

16-06

Process repository

  • Contains process descriptions

  • Supports multiple presentations of process assets

16-50

Organizational structure

  • Disciplinary reporting line

  • Organizational units and sub-units, if applicable

16-52

ML data management system

  • The ML data management system is part of the configuration management system (see 16-03) and

  • Supports data management activities like data collection, description, ingestion, exploration, profiling, labeling/annotation, selection, structuring and cleansing

  • Provides the data for different purposes, e.g., training, testing

  • Supports the relevant sources of ML data

17-00

Requirement

  • An expectation of functions and capabilities (e.g., non-functional requirements), or one of its interfaces

  • from a black-box perspective

  • that is verifiable, does not imply a design or implementation decision, is unambiguous, and does not introduce contradictions to other requirements.

  • A requirements statement that implies, or represents, a design or implementation decision is called “Design Constraint”.

  • Examples for requirements aspects at the system level are thermal characteristics such as

    • heat dissipation

    • dimensions

    • weight

    • materials

  • Examples of aspects related to requirements about system interfaces are

    • Connectors

  • cables

  • housing

  • Examples for requirements at the hardware level are

  • lifetime and mission profile, lifetime robustness

  • maximum price

  • storage and transportation requirements

  • functional behavior of analog or digital circuits and logic

  • quiescent current, voltage impulse responsiveness to crank, startstop, drop-out, load dump

  • temperature, maximum hardware heat dissipation

  • power consumption depending on the operating state such as sleep-mode, start-up, reset conditions

  • frequencies, modulation, signal delays, filters, control loops

  • power-up and power-down sequences, accuracy and precision of signal acquisition or signal processing time

  • computing resources such as memory space and CPU clock tolerances

  • maximum abrasive wear and shearing forces for e.g., pins or soldering joints

  • requirements resulting from lessons learned

    • safety related requirements derived from the technical safety concept

17-05

Requirements for work products

  • Requirements for content and structure, storage and control

  • Identifies documentation specific meta data, such as id, date, author information, ownership, access rights, review and approval status with, where applicable, status model and workflow, or others

  • Identifies requirements on documentation structure, e.g., table of content or figures or other formal aspects

  • May be provided by documentation templates

  • May be based on tool specific templates

  • Defines the storage location such as data repository, tool, versioning system

  • Requirements for versioning

  • Requirements for baselining

  • Distribution of the documents

  • Maintenance and disposal of the documents

  • May be specific for certain types of documents

17-54

Requirement Attribute

  • Meta-attributes that support structuring and definition of release scopes of requirements.

  • Can be realized by means of tools.

NOTE: usage of requirements attributes may further support analysis of requirements.

17-55

Resource needs

  • Identification of required resources for process performance

  • Staff including competencies, skills and authorities needs

  • Material, equipment, and infrastructure

  • Time and budget

NOTE: Needs are derived from Work Breakdown structure and schedule

17-57

Special Characteristics

  • Special Characteristics in terms of relevant standards such as IATF 16949, VDA 6.x Guidelines, ISO 26262.

  • Special Characteristics according to IATF 16949:2016-10 [15], Chapters 8.3.3.3, are product characteristics or production process parameters that may have an impact on safety or compliance with official regulations, the fit, the function, the performance or further processing of the product.

  • Special characteristics shall be verifiable according to VDA vol. 1

NOTE: A typical method for identifying and rate special characteristics is an FMEA.

18-00

Standard

  • Identification of to whom/what they apply

  • Expectations for conformance are identified

  • Conformance to requirements can be demonstrated

  • Provisions for tailoring or exception to the requirements are included

18-06

Product release criteria

  • Defines expectations for product release:

  • release type and status

  • required elements of the release

  • product completeness including documentation

  • adequacy and coverage of testing

  • limit for open defects

  • change control status

18-07

Quality criteria

  • Defines the expectations for work products and process performance

  • Including thresholds/tolerance levels, required measurements, required checkpoints

  • Defines what is an adequate work product (required elements, completeness expected, accuracy, etc.)

  • Defines what constitutes the completeness of the defined tasks

  • Defines what constitutes the performance of the defined tasks

  • Establishes expected performance attributes

18-52

Escalation path

  • Defined mechanisms to report and confirm escalation relevant issues

  • Identifies stakeholders to be included in the escalation path

  • Identifies levels of escalation

18-53

Configuration item selection criteria

  • Identify types of work products to be subject to configuration control

18-57

Change analysis criteria

  • Defines analysis criteria, such as

  • resource requirements

  • scheduling issues

  • risks

  • benefits

18-58

Process performance objectives

  • Objectives for the process of creating the process outcomes and capability level 2 achievements, and corresponding evaluation criteria

  • Assumptions and constraints, if applicable

  • Used as the basis for deriving a detailed planning

  • Examples:

    • Effort, costs, or budget targets (e.g., min/max limits)

    • Process-specific deadlines in line with milestones, or frequency of activities (o e.g., dates for deliveries to the customer, quality gates)

    • Metrics (e.g., max. number of open change requests per release, max. ratio of configuration items in status “in work” at certain milestones before next delivery / release date)

18-59

Review and approval criteria for work products

  • Specifies for each type of work products review and approval needs

  • If and when a review is required

  • Who shall review it

  • Who shall approve it

  • Review method(s) to be used - Criteria for approval

18-70

Business goals

  • Explanation of the business goals

  • Requirements for the business needs

  • Associations to other goals

  • Reasons for the existence of the goals and needs, level of degree of the need and effect on the business not having that need

  • Conditions, constraints, assumptions

  • Timeframe for achievement

  • Authorization at the highest level

18-80

Improvement opportunity

  • Cause of the improvement need, e.g.,

    • from qualitative or quantitative process performance analysis, evaluations, and monitoring

    • industry best practice review, state-of-the-art observations, market studies etc.

  • Improvement objectives derived from organizational business goals and improvement needs

  • Organizational scope

  • Process scope

  • Activities to be performed to keep all those affected by the improvement informed

  • Priorities

18-81

Improvement evaluation results

  • Operational impacts of identified changes on the product(s) and processes

  • Expected benefit

  • Conditions, constraints, assumptions

19-01

Process performance strategy

  • The operational approach to achieve the process outcomes, consistent with the Process Performance Objectives (18-58), e.g.:

    • proceedings, including the monitoring of the performance of the process

    • methodology

  • scope(s) of the strategy within the process, e.g.:

    • development sites

    • application domain-specific differences (e.g., software drivers versus. powertrain software)

    • disciplines (e.g., different configuration management approaches for software and hardware, or combined approaches)

    • options due to socio-cultural differences

19-50

ML data quality approach

  • The ML data quality approach

  • Defines Quality criteria (see 18-07) e.g., the relevant data sources, reliability and consistency of labelling, completeness against ML data requirementsv

  • Describes analysis activities of the data

  • Describes activities to ensure the quality of the data to avoid issues e.g., data bias, bad labeling

Annex C Key concepts and guidance

The following sections describe the key concepts that have been introduced in the Automotive SPICE PRM resp. PAM 3.1. They relate to the terminology described in Annex C Terminology.

Annex C.1 The “Plug-in” concept

The following figure shows the basic principle of the “plug-in” concept. The top-level comprises all system engineering processes organized in a system “V”. Depending on the product to be developed the corresponding engineering disciplines with their domain-specific processes (e.g., hardware engineering HWE or software engineering SWE) can be added to the assessment scope. All other processes such as management processes and supporting processes are domain-independent and are therefore designed in a way that they can be applied to both the system level and the domain levels.

_images/image6.jpeg

Figure C. 1 — The “Plug-in” concept

Annex C.2 “Element”, “Component”, and “Unit”

The following figure depicts the relationships between system elements, software components, and software units which are used consistently in the engineering processes. See the Terminology in Annex C to learn about the definitions of these terms.

_images/image7.jpeg

Figure C. 2 — Element, component, and unit

Annex C.3 Integration of Machine Learning Engineering Processes

The following figure shows how the Machine Learning Engineering Processes are integrated within the engineering V-Cycle perspective. Usually, the MLE “Machine Learning Engineering” processes are used in a highly iterative way.

Within the Software Architecture software elements shall be identified which need to be developed with machine learning. For these ML-based software elements the MLE processes apply, for other software components SWE.3 “Software Detailed Design & Unit Construction” and SWE.4 “Software Unit Verification” applies. After the successful testing of the ML-based software elements they need to be integrated with the other software components by applying SWE.5 “Software Integration & Integration Test”.

_images/image8.png

Figure C. 3 — Integration of MLE Processes

The second figure shows the interdependencies within the MLE “Machine Learning Engineering” process group and to the SUP.11 “Machine Learning Data Management”.

With MLE.1 “Machine Learning Requirements Analysis”, machine learning related Software requirements allocated to the ML-based software elements need to be refined into a set of ML requirements. These requirements must contain ML data requirements which are input for the SUP.11 “Machine Learning Data Management” and other ML requirements which are input for the other MLE “Machine Learning Engineering” processes.

By applying the SUP.11 “Machine Learning Data Management” process ML data with assured quality and integrity, which fulfill the ML data requirements, are collected, processed, and made available to all affected parties.

The other ML requirements should be used within the MLE.2 “Machine Learning Architecture” process to develop an ML architecture supporting training and deployment. Therefore, the ML architecture must contain all necessary ML architectural elements like hyperparameter ranges and initial values, details of the ML model, and possible other software parts which are necessary for MLE.3 “Machine Learning Training”. These other software parts should be developed according to SWE.3 “Software Detailed Design & Unit Construction” and SWE.4 “Software Unit Verification”.

Performing MLE.3 “Machine Learning Training” should start with specifying an ML training and validation approach. Based on this approach a training and validation dataset need to be created from the ML data pool provided by SUP.11 “Machine Learning Data Management” which is then used iteratively to optimize the ML model weights and hyperparameter values. When training exit criteria are reached the trained model should be agreed and communicated to all affected parties.

MLE.4 “Machine Learning Model Testing” focuses on testing the agreed trained model to ensure compliance with the ML requirements. Therefore, an ML test approach needs to be specified and an ML test dataset must be created from the provided ML data pool.

The ML test approach defines besides other details the distribution of data characteristics (e.g., sex, weather conditions, street conditions within the ODD) defined by ML requirements. The test dataset contains different test scenarios applying the required distribution of data characteristics e.g., driving during rain on a gravel road.

After successful testing the trained model, a deployed model is derived and tested as well. The deployed model will be integrated into the target system and may differ from the trained model which often requires powerful hardware and uses interpretative languages. Finally, the agreed test results and the deployed model must be communicated to all affected parties, so that the deployed model can be integrated with the other software units by applying SWE.5 “Software Integration and Integration Test”.

_images/image9.png

Figure C. 4 — Interdependencies within MLE and SUP.11

Annex C.4 Example of an ML Architecture

The following figure shows an example of an ML architecture, describing the overall structure of the ML-based software element and the interfaces within the ML-based software element and to other software elements. The ML architecture typically consists of an ML model and other ML architectural elements, which are other (classical) software components developed according to SWE.3 “Software Detailed Design & Unit Construction” and SWE.4 “Software Unit Verification” and provided to train, deploy and test the ML model. Furthermore, the ML architecture describes details of the ML model like used layers, activation functions, loss function, and backpropagation.

Figure C. 5 — Example of an ML Architecture

During training, hyperparameters (see appendix c #hyperparameters) defining the ML model will change therefore it is recommended to define ranges of hyperparameter values and initial values for training start are defined. Because developing an ML-based software element is highly iterative changes in the ML architecture might come up.

Furthermore, the ML architecture used for training can differ from the architecture of the deployed model, which will be integrated with the other software elements, these differences are part of the ML architecture as well.

Annex C.5 Traceability and consistency

_images/image11.png

Figure C. 6 (Figure C.2) — Consistency and traceability between system and software work products

_images/image12.png

Figure C. 7 (Figure C.3) — Consistency and traceability between system and hardware work products

_images/image13.jpeg

Figure C. 8 (Figure C.4) — Consistency and traceability between ML work products

Annex C.6 “Agree” and “Summarize and Communicate”

The information flow on the left side of the “V” is ensured through a base practice “Communicate agreed ‘work product x’”. The term “agreed” here means that there is a joint understanding between affected parties of what is meant by the content of the work product.

The information flow on the right side of the “V” is ensured through a base practice “Summarize and communicate results”. The term “Summarize” refers to abstracted information resulting from test executions made available to all affected parties.

These communication-oriented base practices do not require a planning-based approach, nor a formal approval, confirmation, or release, as this is targeted at by GP 2.1.6 on capability level 2. At capability level 1 the communication-oriented base practices mean that the work products (or their content) are to be disseminated to affected parties.

An overview of these aspects is shown in the following figure:

_images/image14.jpeg

Figure C. 9 — Agree, summarize and communicate

Annex C.7 Key Changes in Automotive SPICE 4.0

Terminology – “Measure” vs. “Metric”

In the English literature the term “measure” can mean both

  • to find the size, quantity, etc. of something in standard units, ‘size/quantity’” and “… to judge the importance, value or effect of something”, respectively

  • a plan of action or something done

In PAM v3.1, both meanings were inconsistently used in e.g.:

  • MAN.5.BP6

  • MAN.6

  • PIM.3.BP7 Note 10

  • work product characteristics 07-00 “Measure“ and 07-xx

In the assessment practice, this sometimes led to misconceptions with interpreting these assessment indicators and, thus, to varying assessment results. Since it was one of the objectives for Automotive SPICE 4.0 to use terminology more homogeneously, the decision was to consistently use the following terms and meaning:

  • Quantitative measurement – “metric

  • “A plan of action” – “measure”

  • “To act in an operational manner” – “action”

Terminology – “Affected Party” (Level 1) vs. “Involved Party” (Level 2)

Processes at capability level 1 use the term “affected party” in the context of BP “Communicate”. This is to indicate that for every process instance A there is a downstream process instance B that requires the technical (i.e., capability level 1) output of A as a necessary input. Otherwise, process instance B would not be able to proceed, or update its output.

In contrast, “involved party” at capability level 2 includes, but goes beyond “affected parties”. For example, there may be a stakeholder who

  • is passively kept in the information loop (e.g., a line manager, steering committee);

  • is providing input (e.g., a deadline, a particular resource) and only requiring a commitment, but no further active involvement.

Affected parties thus are a subset of involved parties.

Diagram, venn diagram Description automatically generated

Terminology – “Verification” instead of “Testing”

The former Automotive SPICE V3.1 process SUP.2 Verification has been removed in favor of advancing the respective SWE and SYS processes from pure testing to verification. The motivation for this was:

  • SUP.2 was vague in regards to what ‘verification‘ should consider in contrast to testing processes

  • Especially at the system level, testing is not the only verification approach. Rather, measurements (e.g., geometrical tolerances), calculations or analyses (e.g., strength/stress calculation using an FEM method), or simulations instead of using physical samples are other types of verification. The same is true for mechanical or hardware development. Therefore, the umbrella term verification now forms he center of those processes‘ purposes.

  • The process SWE.4 ‘Unit Verification’ has already been an exception as a SW unit can be verified coherently by means of a combination of static analysis, testing, or code reviews (a view that is also inherent in ISO 26262-6 clause 9).

Annex D Reference standards

Annex D provides a list of reference standards and guidelines that support implementation of the Automotive SPICE PAM / PRM.

Table D. 1 — Reference standards

ISO/IEC 33001:2015

Information technology – Process assessment – Concepts and terminology

ISO/IEC 33002:2015

Information technology – Process assessment –

Requirements for performing process assessment

ISO/IEC 33003:2015

Information technology – Process assessment –

Requirements for process measurement frameworks

ISO/IEC 33004:2019

Information technology – Process assessment –

Requirements for process reference, process assessment and maturity models

ISO/IEC 33020:2019

Information technology – Process assessment –

Process measurement framework for assessment of process capability

ISO/IEC/IEEE 24765:2017

Systems and software engineering – Vocabulary

ISO/IEC/IEEE 29148:2018

Systems and software engineering – Life cycle processes – Requirements engineering

INCOSE Guide for Writing Requirements

https://www.incose.org/

PAS 1883:2020

Operational design domain (ODD) taxonomy for an automated driving system (ADS)

ISO 26262:2018

Road vehicles — Functional safety, Second edition 2018-12