Evaluating Deep Learning Biases Based on Grey-Box Testing Results

Document Type

Conference Proceeding

Publication Date

1-1-2021

Abstract

The very exciting and promising approaches of deep learning are immensely successful in processing large real world data sets, such as image recognition, speech recognition, and language translation. However, much research discovered that it has biases that arise in the design, production, deployment, and use of AI/ML technologies. In this paper, we first explain mathematically the causes of biases and then propose a way to evaluate biases based on testing results of neurons and auto-encoders in deep learning. Our interpretation views each neuron or autoencoder as an approximation of similarity measurement, of which grey-box testing results can be used to measure biases and finding ways to reduce them. We argue that monitoring deep learning network structures and parameters is an effective way to catch the sources of biases in deep learning.

Publication Title

Advances in Intelligent Systems and Computing

First Page Number

641

Last Page Number

651

DOI

10.1007/978-3-030-55180-3_48

This document is currently not available here.

Share

COinS