In this section, the best explainable visualisations to check the predictions’ level of confidence are displayed. Specifically, this section serves as a valuable resource for evaluating the robustness of predictions and enhancing your overall comprehension of the predictive outcomes.
A textual representation can be used to facilitate AI model explainability by converting complex model outputs into human-readable text. Techniques like natural language processing and attention mechanisms help highlight key features and decision-making processes, providing concise and interpretable explanations. This promotes transparency and trust in AI systems for effective communication with end-users.
A text-based explaination is used by a XMANAI demonstrator to illustrate how changing the value of a specific feature will change the outcome of the prediction in a what-if scenario forecast. The user can define specific what-if scenario, changing value of some features and the results is visualised with a simple textual example.
XMANAI Project Coordinator
Michele Sesana – TXT
e-mail: michele.sesana@txtgroup.com
XMANAI Scientific Coordinator
Dr. Yury Glikman – Fraunhofer FOKUS
e-mail: yury.glikman@fokus.fraunhofer.de
XMANAI Technical Coordinator
Dr. Fenareti Lampathaki – SUITE5
e-mail: fenareti@suite5.eu
TXT e-solutions S.p.A.
Via Frigia 27 – 20126 Milano
t: +390225771804