These tools and metrics are designed to help AI actors develop and use trustworthy AI systems and applications that respect human rights and are fair, transparent, explainable, robust, secure and safe.
MISSION KI: Quality Standard for Low-Risk AI

MISSION KI is developing a voluntary quality standard guideline for AI that strengthens the reliability and trustworthiness of AI applications and systems. The standard is based on the "Ethics Guidelines for Trustworthy AI" developed by the High-Level Expert Group (HLEG) of the European Commission. The HLEG established key principles for assessing AI trustworthiness, which also served as a foundation for the AI Act’s requirements.
The MISSION KI Quality Standard for Low-Risk AI (November 2025) is a voluntary framework developed under the MISSION KI initiative – a federal initiative and key project of Germany’s federal digitalisation strategy. The initiative was implemented from May 2023 – December 2025 by acatech – National Academy of Science and Engineering and funded by Germany’s Federal Ministry for Digital Transformation and Government Modernisation. The quality standard is aimed particularly at start-ups and small and medium sized enterprises whose AI systems fall below the EU AI Act’s high-risk threshold. It provides organisations with practical, verifiable criteria to demonstrate quality and trustworthiness, strengthen transparency for customers, investors, and public-sector clients, and build readiness for evolving regulatory expectations.
The standard structures its assessment around six quality dimensions: data quality, protection, and governance; non-discrimination; transparency; human oversight and control; reliability; and AI-specific cybersecurity. These dimensions are rooted in established European work on trustworthy AI, including the AI High-Level Expert Group’s ethics guidance, and are designed to be compatible with the EU AI Act while remaining efficient enough for voluntary use and self-assessments.
Assessments follow a step-by-step self-assessment workflow. It begins with a structured description of the use case, covering the system’s intended purpose, key components, interfaces, and operational context. This is followed by a protection needs analysis that determines how critical each requirement is for the specific application, typically categorising needs as high, moderate, or low. The organisation then evaluates the system using an assessment catalogue based on the VCIO model, progressing from quality dimensions to criteria, from criteria to indicators, and from indicators to observables. Each observable is graded on an A–D scale, where A indicates the strongest fulfilment and D indicates that the requirement is not met. The framework also links ratings to concrete expectations, such as required analyses, implemented measures, and—where relevant—explicit acceptance of residual risk.
To make results credible and repeatable, the process emphasises evidence and technical testing, encouraging documentation that supports the chosen ratings and enables verification. External validation is optional, allowing organisations to strengthen assurance when needed.
The overall outcome is a pass/fail plausibility check. For each criterion, the aggregated rating must meet or exceed the minimum level implied by its protection need; for example, a “high” protection need requires an A-level outcome. Passing indicates that the documented measures appear appropriate for the identified needs, while noting that the standard does not quantify residual risk.Finally, the standard includes an assessment report template and guidance for validity monitoring. Results apply only to the assessed system version and become invalid if the intended purpose changes, external conditions shift materially (for example through drift), or the system implementation is significantly modified.
To facilitate implementation of the quality standard, the initiative also developed a digital tool that guides through the assessment process and functions as a digital hub for documentation and access to a collection of various technical assessment tools. The digital tool is available open source on github.
The Mission KI Quality Standard for Low-Risk AI sets a voluntary, evidence-based self-assessment framework for AI providers below the EU AI Act’s high-risk threshold. It defines six quality dimensions (data governance, non-discrimination, transparency, human oversight, reliability, AI-specific cybersecurity) and a stepwise procedure: describe the use case, analyse protection needs, rate requirements via a VCIO catalogue, document tests/evidence, validate findings, issue a report, and monitor validity.
************************************************
More on MISSION KI Quality Standard for Low-AI Risk
1. What is the foundation of the MISSION KI Quality Standard based on?
1.1 — Values: In this framework, six core values have been identified to guide responsible AI development and deployment:
- Reliability
- Performance & Robustness
- Fallback Plans & General Safety
- AI-specific cyber security
- Resistance to AI-specific attacks and security
- Data quality, protection and management
- Data quality & integrity
- Protection of personal data
- Protection of proprietary data
- Data access
- Non-discrimination
- Avoidance of unjustified distortions
- Accessibility and universal design
- Stakeholder participation
- Transparency
- Traceability & documentation
- Explainability & interpretability
- External communication
- Human supervision & control
- Human capacity to act
- Human supervision
1.2 — Protection needs analysis:
The Mission KI Quality Standard relies on the protection needs analysis (SBA) as a starting point to ensure efficiency. This analysis determines the necessary protection requirements for the defined values and thus forms the basis for a targeted test. It filters out the relevant values and criteria for a use case and defines a target for the subsequent test.
The minimum standard therefore considers the variety of AI application scenarios - from energy distribution optimization and product recommendation systems to medical diagnostic tools. The relevance of the individual values varies depending on the use case.
For example, the value ‘non-discrimination’ plays a subordinate role in an AI for optimizing power distribution, as the decisions are based on technical parameters. In this case, the value of ‘transparency’ takes center stage: the AI's decisions must be comprehensible and understandable so that operators and regulatory authorities can check why certain energy distributions were made.
Regardless of the use case, the ‘reliability’ value is always subject to scrutiny, as it is considered fundamental to the quality of any AI application. The other values can be categorized as not applicable in whole or in part under certain conditions that are clearly defined in the protection requirements analysis.
2. How does the standard become auditable?
2.1 — The test criteria catalogue translates abstract values into measurable variables
The Mission KI test criteria catalogue is based on three sources in particular:
- VDE SPEC 90012,
- AI test catalogue of the Fraunhofer IAIS,
- AIC4 criteria catalogue for AI cloud services from the Federal Office for Security and Information Technology (BSI).
In order to make the Mission KI quality standard testable, the 6 abstract values were translated into a structured test procedure based on the so-called ‘VCIO’ approach (Values - Criteria - Indicators - Observables). This is divided into several levels: The values form the foundation on which specific criteria are built. Indicators and measurable variables (observables) are used to assess these criteria. The degree of fulfilment of each value is systematically determined on the basis of this structure. This methodology ensures a precise and comprehensible assessment.
In addition, test tools are developed to check the fulfilment of the observables and thus increase the reliability of the test result.
2.2 — The evaluation
Evaluations are conducted by either internal or external auditors.
At the end of the test process, an overall assessment is made for each of the six defined values. This assessment is compared with the previously determined protection requirements. An AI application passes the test if it achieves the defined test target for each value. This documents that the quality measures and their evidence sufficiently fulfil the identified protection requirements.
The successful test thus confirms that the AI application fulfills the necessary quality standards and has implemented the required protective measures. This process ensures a thorough evaluation and creates transparency regarding the trustworthiness and security of the tested AI systems.
About the tool
You can click on the links to see the associated tools
Developing organisation(s):
Tool type(s):
Objective(s):
Impacted stakeholders:
Lifecycle stage(s):
Type of approach:
Maturity:
Usage rights:
Target groups:
Target users:
Stakeholder group:
Validity:
Enforcement:
Geographical scope:
Technology platforms:
Tags:
- transparent
- trustworthy ai
- ai reliability
- ai auditing
- trustworthiness
- ai compliance
- data quality
- human-centred design
Use Cases
Would you like to submit a use case for this tool?
If you have used this tool, we would love to know more about your experience.
Add use case


























