Recurring |
one_organization |
(a) The software failure incident having happened again at one_organization:
The incident of software failure has happened again with Woebot, as it required updates after struggling to handle reports of child sexual abuse. The app had been rated suitable for children but failed to recognize and flag clear breaches of law or safeguarding of children. As a result of the probe, Woebot's makers introduced an 18+ age limit for their product [79245].
(b) The software failure incident having happened again at multiple_organization:
There is no specific mention in the articles about the software failure incident happening again at multiple organizations. |
Phase (Design/Operation) |
design, operation |
(a) The software failure incident related to the design phase is evident in the articles. The chatbot apps, Woebot and Wysa, failed to properly handle reports of child sexual abuse, eating disorders, and drug use. Despite being rated suitable for children, these apps did not recognize and flag clear breaches of law or safeguarding of children, indicating a design flaw in their systems [79245].
(b) The software failure incident related to the operation phase is also apparent in the articles. Both Woebot and Wysa, designed to assist with mental health issues, failed to appropriately respond to serious situations such as reports of child sexual abuse, self-harm, eating disorders, and drug use. The operational failure was highlighted by the apps' inability to provide adequate responses and interventions when faced with distressing messages from users [79245]. |
Boundary (Internal/External) |
within_system |
(a) within_system:
The software failure incident reported in the articles is primarily within the system. The chatbot apps, Woebot and Wysa, failed to properly handle reports of child sexual abuse, eating disorders, and drug use despite being designed to flag serious or dangerous situations. The failures were related to the apps' inability to detect obvious signs of distress and appropriately respond to critical issues raised by users. For example, when testers mentioned being forced to have sex at a young age, the chatbots responded inadequately, showing a lack of capability within the system to address such serious concerns [79245]. |
Nature (Human/Non-human) |
non-human_actions, human_actions |
(a) The software failure incident in the articles can be attributed to non-human actions, specifically the limitations and shortcomings of the mental health chatbot apps Woebot and Wysa. These chatbots failed to properly handle reports of child sexual abuse, eating disorders, and drug use despite being designed to flag serious or dangerous situations automatically. The failures were evident in the chatbot responses to distressing messages, where they provided inadequate or inappropriate feedback, such as not recognizing clear signs of distress or illegal activities [79245].
(b) On the other hand, human actions also played a role in the software failure incident. The developers of Woebot and Wysa were responsible for introducing updates to address the deficiencies in their apps after the issues were highlighted. For example, Woebot's makers introduced an 18+ age limit for their product and updated their software to account for phrases used in the BBC tests. Similarly, Touchkin, the firm behind Wysa, mentioned that they were updating their app to better handle situations involving coercive sex, illegal drugs, and eating disorders. These actions indicate that human intervention was necessary to rectify the shortcomings in the chatbot apps [79245]. |
Dimension (Hardware/Software) |
software |
(a) The articles do not mention any software failure incident occurring due to contributing factors originating in hardware. Therefore, there is no information available regarding a hardware-related failure incident.
(b) The software failure incident reported in the articles is related to the inability of mental health chatbot apps, Wysa and Woebot, to effectively handle reports of child sexual abuse, eating disorders, and drug use. The failures were primarily due to shortcomings in the software's algorithms and responses, which led to inappropriate or inadequate reactions to serious situations reported by users [79245]. |
Objective (Malicious/Non-malicious) |
non-malicious |
(a) The software failure incident reported in Article 79245 is non-malicious. The failure was due to the chatbot apps, Woebot and Wysa, struggling to handle reports of child sexual abuse, eating disorders, and drug use. The apps failed to properly identify and respond to serious and dangerous situations, indicating a lack of capability rather than intentional harm to the system [79245]. |
Intent (Poor/Accidental Decisions) |
poor_decisions |
(a) The intent of the software failure incident was poor_decisions. The failure was due to poor decisions made in the design and implementation of the mental health chatbot apps, Woebot and Wysa. These apps struggled to handle reports of child sexual abuse, eating disorders, and drug use. They failed to recognize and flag serious or dangerous situations, such as clear breaches of law or safeguarding of children. The apps' responses to distressing messages were inadequate and inappropriate, showing a lack of understanding of the severity of the situations being described. Despite being recommended for children and teenagers, the apps' inability to effectively address critical issues led to updates and age restrictions being implemented to mitigate the risks associated with their use [79245]. |
Capability (Incompetence/Accidental) |
development_incompetence, accidental |
(a) The software failure incident related to development incompetence is evident in the case of the mental health chatbot apps, Woebot and Wysa. These apps failed to properly handle reports of child sexual abuse, eating disorders, and drug use. Despite being rated suitable for children and recommended by an NHS Trust, the chatbots did not recognize and flag clear breaches of law or safeguarding of children, indicating a lack of professional competence in designing the systems to handle such critical situations [79245].
(b) The software failure incident related to accidental factors is seen in the inability of the chatbot apps to effectively identify and respond to distress signals and serious situations. The responses generated by the chatbots when presented with scenarios of child sexual abuse, self-harm, and illegal activities were inadequate and sometimes inappropriate, indicating accidental shortcomings in the design and programming of the chatbot algorithms [79245]. |
Duration |
permanent |
(a) The software failure incident in the articles seems to be more of a permanent nature. Both mental health chatbot apps, Woebot and Wysa, required updates after struggling to handle reports of child sexual abuse, dealing with eating disorders, and drug use. Woebot even introduced an 18+ age limit for their product as a result of the probe, indicating a permanent change in their approach to handling such sensitive issues [79245]. |
Behaviour |
omission, value, other |
(a) crash: The articles do not mention any instances of the software crashing.
(b) omission: The software failure incident in the articles is related to omission. Both mental health chatbot apps, Woebot and Wysa, failed to properly handle reports of child sexual abuse, eating disorders, and drug use. They omitted to perform their intended functions of recognizing and flagging serious or dangerous situations, such as clear breaches of law or safeguarding of children [79245].
(c) timing: The articles do not mention any instances of the software performing its intended functions correctly but too late or too early.
(d) value: The software failure incident in the articles is related to value. The chatbot apps provided incorrect and inadequate responses to serious situations, such as suggesting rewriting negative thoughts instead of addressing the severity of the reported issues like being forced into sex at a young age or having eating disorders [79245].
(e) byzantine: The articles do not mention the software behaving with inconsistent responses and interactions.
(f) other: The other behavior exhibited by the software in this failure incident is the inability to detect and appropriately respond to critical and distressing situations, leading to a lack of proper intervention and support for users in need [79245]. |