Darold Davis, MCSD, is a Chartered Designer™ artist, and engineer with more than 28 years of experience working in digital interactive media creating products and services, that includes SaaS and software tools to help individuals and organizations improve existing processes and workflows.
This project aims to address gaps in machine learning (ML) interpretability with regard to visualization by investigating researchers workflows, tool usage, and challenges in understanding model behavior. Through a survey and interviews with practitioners, I identified limitations in existing visualization tools, such as fragmented workflows and insufficient support for analyzing neuron-level attributions. Based on these findings, I developed a prototype tool to visualize neuron activations and attributions, enabling deeper insights into model decision-making. This work contributes to enhancing the understanding of ML models and improving their transparency, a critical step toward ensuring the safety and reliability of advanced AI systems.
Introduction
Understanding model behavior is critical for AI safety, as opaque systems risk unintended... (read 942 more words →)