Find your template, checklist or other download to help you in you tasks.
Find your way to be trained or even get certified in TMAP.
Start typing keywords to search the site. Press enter to submit.
Test design is an important topic in TMAP. Structured testing requires careful consideration of what to test and how to test it. We use the term “test design” for the entirety of these activities, even though in some approaches (like exploratory testing) there is no actual up-front test case design involved.To prepare and perform tests, we distinguish two overall approaches:
In the more than 30 years of TMAP’s existence, we have described many different test design techniques and approaches. But with the rise of intelligent machines the focus shifts, the existing techniques and approaches described below, gain importance, and new techniques and approaches also become relevant.
Coverage-based test design that pertains to testing of AI based solutions:
Experience-based test design approaches that pertain to the testing of AI-based solutions are:
By no means do we claim that the above lists of coverage-based and experience-based test design are exhaustive nor definitive lists, since the field GenAI is rapidly evolving new techniques and approaches will very likely pop up. And please remember that the traditional test design techniques and approaches can also be used.
When preparing for testing an AI-based system, especially when it is Generative AI, a proper preparation is to create one or more charters for exploratory testing. This will be more effective than creating detailed test cases upfront, because of the probabilistic nature of the AI models. This probabilistic nature implies that it is always uncertain what the output of a process will be, therefore it is fundamentally less reliable when compared with a rule-based solution. If, however, the risks are assessed to be within acceptable range, and in balance with expected benefits (such as faster IT delivery), organizations may still decide to apply AI-based solutions. In this situation, performing well organized structured exploratory testing can be a good way to provide information about the quality level. This information helps stakeholders establish their confidence in achieving the pursued business value.
People learn all sorts of skills and have to prove to be proficient in an exam, for example to obtain a driving license. The exam checks if they know the rules and are able to apply them correctly. After they pass the examination, we trust in the future these people will always make the right decisions.As with people, we want to know whether an intelligent machine will perform well enough. Testing is, of course, the basic approach to evaluate the quality level. Since it isn’t possible to test all possibilities, we can use the approach of examination to decide if the intelligent machine is to be trusted to perform the task. As soon as the AI passes the exam it can be used in live operation. If the learning is frozen, so the intelligent machine doesn’t change its behavior after the exam, this may be a good approach.
If the intelligent machine keeps on learning, one exam is not sufficient, we should do an examination every once in a while. With people, periodic examination probably wouldn’t be feasible. But with an intelligent system a periodic exam would be possible, provided that the examination itself can also be done by an automated system. With periodic examination, a fundamental question is how often this examination should take place. That has to do with the periodicity of use of the intelligent machine, but most machines will be functioning continuously. So, there will probably be a desire to do continuous examination. Because it’s an operational system, it is better to use the term continuous monitoring. When stakeholders know the system is continuously monitored in a structured and formal way, this will support the confidence in such an operational system. Especially with closed-loop machine learning and other forms of AI systems that continuously improve their behavior, it is important to implement continuous monitoring of the results to check if the results remain within the boundaries of tolerance that were defined as “good behavior”.
[Reference: Testing in a digital age / chapter 4.5 [Ven, 2018]]
Overview