Explaination-Providing AI (XAI) System

From GM-RKB
(Redirected from AI Explainability)
Jump to navigation Jump to search

A Explaination-Providing AI (XAI) System is an AI system that provides understandable reasons behind its decisions and actions, making its processes transparent and comprehensible to humans.



References

2024

[1] https://datascience.aero/explainability-interpretability-what-model-need/
[2] https://www.kdnuggets.com/2018/12/machine-learning-explainability-interpretability-ai.html
[3] https://datascience.stackexchange.com/questions/99808/an-example-of-explainable-but-not-interpretable-ml-model
[4] https://blogs.sas.com/content/hiddeninsights/2022/08/10/interpretability-vs-explainability-the-black-box-of-machine-learning/
[5] https://christophm.github.io/interpretable-ml-book/
[6] https://link.springer.com/chapter/10.1007/978-3-031-04083-2_2
[7] https://docs.aws.amazon.com/whitepapers/latest/model-explainability-aws-ai-ml/interpretability-versus-explainability.html
[8] https://www.ibm.com/topics/explainable-ai
[9] https://quiq.com/blog/explainability-vs-interpretability/
[10] https://www.datacamp.com/tutorial/explainable-ai-understanding-and-trusting-machine-learning-models 
[11] https://datascience.stackexchange.com/questions/70164/what-is-the-difference-between-explainable-and-interpretable-machine-learning

2024

2023

  • (Wikipedia, 2023) ⇒ https://en.wikipedia.org/wiki/Explainable_artificial_intelligence Retrieved:2023-8-9.
    • Explainable AI (XAI), also known as Interpretable AI, or Explainable Machine Learning (XML), is artificial intelligence (AI) in which humans can understand the reasoning behind decisions or predictions made by the AI. It contrasts with the "black box" concept in machine learning, where even the AI's designers cannot explain why it arrived at a specific decision. [1]

      XAI hopes to help users of AI-powered systems perform more effectively by improving their understanding of how those systems reason. XAI may be an implementation of the social right to explanation.[2] Even if there is no such legal right or regulatory requirement, XAI can improve the user experience of a product or service by helping end users trust that the AI is making good decisions. XAI aims to explain what has been done, what is being done, and what will be done next, and to unveil which information these actions are based on.[3] This makes it possible to confirm existing knowledge, challenge existing knowledge, and generate new assumptions.

       Machine learning (ML) algorithms used in AI can be categorized as white-box or black-box. White-box models provide results that are understandable to experts in the domain. Black-box models, on the other hand, are extremely hard to explain and can hardly be understood even by domain experts. XAI algorithms follow the three principles of transparency, interpretability, and explainability. A model is transparent “if the processes that extract model parameters from training data and generate labels from testing data can be described and motivated by the approach designer.”[4] Interpretability describes the possibility of comprehending the ML model and presenting the underlying basis for decision-making in a way that is understandable to humans.[5][6] Explainability is a concept that is recognized as important, but a consensus definition is not available.[4] One possibility is “the collection of features of the interpretable domain that have contributed, for a given example, to producing a decision (e.g., classification or regression)”. If algorithms fulfill these principles, they provide a basis for justifying decisions, tracking them and thereby verifying them, improving the algorithms, and exploring new facts.

      Sometimes it is also possible to achieve a high-accuracy result with a white-box ML algorithm that is interpretable .[7] This is especially important in domains like medicine, defense, finance, and law, where it is crucial to understand decisions and build trust in the algorithms.[3] Many researchers argue that, at least for supervised machine learning, the way forward is symbolic regression, where the algorithm searches the space of mathematical expressions to find the model that best fits a given dataset.[8] [9] [10]

      AI systems optimize behavior to satisfy a mathematically specified goal system chosen by the system designers, such as the command "maximize accuracy of assessing how positive film reviews are in the test dataset." The AI may learn useful general rules from the test set, such as "reviews containing the word "horrible" are likely to be negative." However, it may also learn inappropriate rules, such as "reviews containing 'Daniel Day-Lewis' are usually positive"; such rules may be undesirable if they are likely to fail to generalize outside the training set, or if people consider the rule to be "cheating" or "unfair." A human can audit rules in an XAI to get an idea of how likely the system is to generalize to future real-world data outside the test set.[11]

  1. Sample, Ian (5 November 2017). "Computer says no: why making AIs fair, accountable and transparent is crucial". The Guardian. Retrieved 30 January 2018.
  2. Edwards, Lilian; Veale, Michael (2017). “Slave to the Algorithm? Why a 'Right to an Explanation' Is Probably Not the Remedy You Are Looking For". Duke Law and Technology Review. 16: 18. SSRN 2972855.
  3. 3.0 3.1 Gunning, D.; Stefik, M.; Choi, J.; Miller, T.; Stumpf, S.; Yang, G.-Z. (2019-12-18). "XAI-Explainable artificial intelligence". Science Robotics. 4 (37): eaay7120. doi:10.1126/scirobotics.aay7120. ISSN 2470-9476. PMID 33137719.
  4. 4.0 4.1 Roscher, R.; Bohn, B.; Duarte, M. F.; Garcke, J. (2020). "Explainable Machine Learning for Scientific Insights and Discoveries". IEEE Access. 8: 42200–42216. arXiv:1905.08883. doi:10.1109/ACCESS.2020.2976199. ISSN 2169-3536.
  5. Murdoch, W. James; Singh, Chandan; Kumbier, Karl; Abbasi-Asl, Reza; Yu, Bin (2019-01-14). "Interpretable machine learning: definitions, methods, and applications". Proceedings of the National Academy of Sciences of the United States of America. 116 (44): 22071–22080. arXiv:1901.04592. Bibcode:2019arXiv190104592M. doi:10.1073/pnas.1900654116. PMC 6825274. PMID 31619572.
  6. Lipton, Zachary C. (June 2018). "The Mythos of Model Interpretability: In machine learning, the concept of interpretability is both important and slippery". Queue. 16 (3): 31–57. doi:10.1145/3236386.3241340. ISSN 1542-7730.
  7. Rudin, Cynthia (2019). "Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead". Nature Machine Intelligence. 1 (5): 206–215. arXiv:1811.10154. doi:10.1038/s42256-019-0048-x. ISSN 2522-5839. PMC 9122117. PMID 35603010.
  8. Wenninger, Simon; Kaymakci, Can; Wiethe, Christian (2022). “Explainable long-term building energy consumption prediction using QLattice". Applied Energy. Elsevier BV. 308: 118300. doi:10.1016/j.apenergy.2021.118300. ISSN 0306-2619. S2CID 245428233.
  9. Christiansen, Michael; Wilstrup, Casper; Hedley, Paula L. (2022). “Explainable "white-box" machine learning is the way forward in preeclampsia screening". American Journal of Obstetrics and Gynecology. Elsevier BV. 227 (5): 791. doi:10.1016/j.ajog.2022.06.057. ISSN 0002-9378. PMID 35779588. S2CID 250160871.
  10. Wilstup, Casper; Cave, Chris (2021-01-15), Combining symbolic regression with the Cox proportional hazards model improves prediction of heart failure deaths, Cold Spring Harbor Laboratory, doi:10.1101/2021.01.15.21249874, S2CID 231609904.
  11. "How AI detectives are cracking open the black box of deep learning". Science. 5 July 2017. Retrieved 30 January 2018.

2020