ISTQB AI Testing CT-AI Question # 8 Topic 1 Discussion
CT-AI Exam Topic 1 Question 8 Discussion:
Question #: 8
Topic #: 1
Before deployment of an AI based system, a developer is expected to demonstrate in a test environment how decisions are made. Which of the following characteristics does decision making fall under?
Explainability in AI-based systems refers to the ease with which users can determine how the system reaches a particular result. It is a crucial aspect when demonstrating AI decision-making, as it ensures that decisions made by AI models are transparent, interpretable, and understandable by stakeholders.
Before deploying an AI-based system, a developer must validate how decisions are made in a test environment. This process falls under the characteristic of explainability because it involves clarifying how an AI model arrives at its conclusions, which helps build trust in the system and meet regulatory and ethical requirements.
ISTQB CT-AI Syllabus (Section 2.7: Transparency, Interpretability, and Explainability)
"Explainability is considered to be the ease with which users can determine how the AI-based system comes up with a particular result".
"Most users are presented with AI-based systems as ‘black boxes’ and have little awareness of how these systems arrive at their results. This ignorance may even apply to the data scientists who built the systems. Occasionally, users may not even be aware they are interacting with an AI-based system".
ISTQB CT-AI Syllabus (Section 8.6: Testing the Transparency, Interpretability, and Explainability of AI-based Systems)
"Testing the explainability of AI-based systems involves verifying whether users can understand and validate AI-generated decisions. This ensures that AI systems remain accountable and do not make incomprehensible or biased decisions".
Contrast with Other Options:
Autonomy (B): Autonomy relates to an AI system's ability to operate independently without human oversight. While decision-making is a key function of autonomy, the focus here is on demonstrating the reasoning behind decisions, which falls under explainability rather than autonomy.
Self-learning (C): Self-learning systems adapt based on previous data and experiences, which is different from making decisions understandable to humans.
Non-determinism (D): AI-based systems are often probabilistic and non-deterministic, meaning they do not always produce the same output for the same input. This can make testing and validation more challenging, but it does not relate to explaining the decision-making process.
Supporting References from ISTQB Certified Tester AI Testing Study Guide:Conclusion:Since the question explicitly asks about the characteristic under which decision-making falls when being demonstrated before deployment,explainability is the correct choicebecause it ensures that AI decisions are transparent, understandable, and accountable to stakeholders.
Contribute your Thoughts:
Chosen Answer:
This is a voting comment (?). You can switch to a simple comment. It is better to Upvote an existing comment if you don't have anything to add.
Submit