关键词:
Neurons
Neural networks
Maps
Algorithms
Human subjects
Artificial intelligence
Computer science
摘要:
Machine learning is being applied in various critical applications like healthcare. In order to be able to trust a machine learning model and to repair it once it malfunctions, it is important to be able to interpret its decision-making. For example, if a model's performance is poor on a specific subgroup (gender, race, etc), it is important to find out why and fix it. In this thesis, we examine the drawbacks of existing interpretability methods and introduce new ML interpretability algorithms that are designed to tackle some of the shortcomings. Data is the labor that trains machine learning models. It is not possible to interpret an ML model's behavior without going back to the data that trained it in the first place. A fundamental challenge is how to quantify the contribution of each source of data to the model's performance. For example, in healthcare and consumer markets, it has been suggested that individuals should be compensated for the data that they generate, but it is not clear what is an equitable valuation for individual datum. In this thesis, we discuss principled frameworks for equitable valuation of data; that is, given a learning algorithm and a performance metric that quantifies the performance of the resulting model, we try to find the contribution of individual datum. This thesis is divided in 3 sections, machine learning interpretability and fairness, data valuation, and machine learning for healthcare - all linked by the common goal of making the use of machine learning more responsible for the benefit of human beings.