000 06169nam a22005895i 4500
001 978-3-031-63803-9
003 DE-He213
005 20240730172749.0
007 cr nn 008mamaa
008 240710s2024 sz | s |||| 0|eng d
020 _a9783031638039
_9978-3-031-63803-9
024 7 _a10.1007/978-3-031-63803-9
_2doi
050 4 _aQ334-342
050 4 _aTA347.A78
072 7 _aUYQ
_2bicssc
072 7 _aCOM004000
_2bisacsh
072 7 _aUYQ
_2thema
082 0 4 _a006.3
_223
245 1 0 _aExplainable Artificial Intelligence
_h[electronic resource] :
_bSecond World Conference, xAI 2024, Valletta, Malta, July 17-19, 2024, Proceedings, Part IV /
_cedited by Luca Longo, Sebastian Lapuschkin, Christin Seifert.
250 _a1st ed. 2024.
264 1 _aCham :
_bSpringer Nature Switzerland :
_bImprint: Springer,
_c2024.
300 _aXVII, 466 p. 149 illus., 136 illus. in color.
_bonline resource.
336 _atext
_btxt
_2rdacontent
337 _acomputer
_bc
_2rdamedia
338 _aonline resource
_bcr
_2rdacarrier
347 _atext file
_bPDF
_2rda
490 1 _aCommunications in Computer and Information Science,
_x1865-0937 ;
_v2156
505 0 _a -- Explainable AI in healthcare and computational neuroscience. -- SRFAMap: a method for mapping integrated gradients of a CNN trained with statistical radiomic features to medical image saliency maps. -- Transparently Predicting Therapy Compliance of Young Adults Following Ischemic Stroke. -- Precision medicine in student health: Insights from Tsetlin Machines into chronic pain and psychological distress. -- Evaluating Local Explainable AI Techniques for the Classification of Chest X-ray Images. -- Feature importance to explain multimodal prediction models. A clinical use case. -- Identifying EEG Biomarkers of Depression with Novel Explainable Deep Learning Architectures. -- Increasing Explainability in Time Series Classification by Functional Decomposition. -- Towards Evaluation of Explainable Artificial Intelligence in Streaming Data. -- Quantitative Evaluation of xAI Methods for Multivariate Time Series - A Case Study for a CNN-based MI Detection Model. -- Explainable AI for improved human-computer interaction and Software Engineering for explainability. -- Influenciae: A library for tracing the influence back to the data-points. -- Explainability Engineering Challenges: Connecting Explainability Levels to Run-time Explainability. -- On the Explainability of Financial Robo-advice Systems. -- Can I trust my anomaly detection system? A case study based on explainable AI.. -- Explanations considered harmful: The Impact of misleading Explanations on Accuracy in hybrid human-AI decision making. -- Human emotions in AI explanations. -- Study on the Helpfulness of Explainable Artificial Intelligence. -- Applications of explainable artificial intelligence. -- Pricing Risk: An XAI Analysis of Irish Car Insurance Premiums. -- Exploring the Role of Explainable AI in the Development and Qualification of Aircraft Quality Assurance Processes: A Case Study. -- Explainable Artificial Intelligence applied to Predictive Maintenance: Comparison of Post-hoc Explainability Techniques. -- A comparative analysis of SHAP, LIME, ANCHORS, and DICE for interpreting a dense neural network in Credit Card Fraud Detection. -- Application of the representative measure approach to assess the reliability of decision trees in dealing with unseen vehicle collision data. -- Ensuring Safe Social Navigation via Explainable Probabilistic and Conformal Safety Regions. -- Explaining AI Decisions: Towards Achieving Human-Centered Explainability in Smart Home Environments. -- AcME-AD: Accelerated Model Explanations for Anomaly Detection.
520 _aThis four-volume set constitutes the refereed proceedings of the Second World Conference on Explainable Artificial Intelligence, xAI 2024, held in Valletta, Malta, during July 17-19, 2024. The 95 full papers presented were carefully reviewed and selected from 204 submissions. The conference papers are organized in topical sections on: Part I - intrinsically interpretable XAI and concept-based global explainability; generative explainable AI and verifiability; notion, metrics, evaluation and benchmarking for XAI. Part II - XAI for graphs and computer vision; logic, reasoning, and rule-based explainable AI; model-agnostic and statistical methods for eXplainable AI. Part III - counterfactual explanations and causality for eXplainable AI; fairness, trust, privacy, security, accountability and actionability in eXplainable AI. Part IV - explainable AI in healthcare and computational neuroscience; explainable AI for improved human-computer interaction and software engineering for explainability; applications of explainable artificial intelligence.
650 0 _aArtificial intelligence.
_93407
650 0 _aNatural language processing (Computer science).
_94741
650 0 _aApplication software.
_9105388
650 0 _aComputer networks .
_931572
650 1 4 _aArtificial Intelligence.
_93407
650 2 4 _aNatural Language Processing (NLP).
_931587
650 2 4 _aComputer and Information Systems Applications.
_9105391
650 2 4 _aComputer Communication Networks.
_9105393
700 1 _aLongo, Luca.
_eeditor.
_0(orcid)
_10000-0002-2718-5426
_4edt
_4http://id.loc.gov/vocabulary/relators/edt
_9105394
700 1 _aLapuschkin, Sebastian.
_eeditor.
_0(orcid)
_10000-0002-0762-7258
_4edt
_4http://id.loc.gov/vocabulary/relators/edt
_9105395
700 1 _aSeifert, Christin.
_eeditor.
_0(orcid)
_10000-0002-6776-3868
_4edt
_4http://id.loc.gov/vocabulary/relators/edt
_9105397
710 2 _aSpringerLink (Online service)
_9105399
773 0 _tSpringer Nature eBook
776 0 8 _iPrinted edition:
_z9783031638022
776 0 8 _iPrinted edition:
_z9783031638046
830 0 _aCommunications in Computer and Information Science,
_x1865-0937 ;
_v2156
_9105400
856 4 0 _uhttps://doi.org/10.1007/978-3-031-63803-9
912 _aZDB-2-SCS
912 _aZDB-2-SXCS
942 _cEBK
999 _c88540
_d88540