Bias – A Lurking Danger that Can Convert Algorithmic Systems into Discriminatory Entities
Version
Published
Date Issued
2020-10-18
Author(s)
Type
Conference Paper
Language
English
Abstract
Bias in algorithmic systems is a major cause of unfair and discriminatory decisions in the use of such systems. Cognitive bias is very likely to be reflected in algorithmic systems as humankind aims to map Human Intelligence (HI) to Artificial Intelligence (AI). An extensive literature review on the identification and mitigation of bias leads to precise measures for project teams building AI-systems. Aspects like AI-responsibility, AI-fairness and AI-safety are addressed by developing a framework that can be used as a guideline for project teams. It proposes measures in the form of checklists to identify and mitigate bias in algorithmic systems considering all steps during system design, implementation and application.
Subjects
QA75 Electronic computers. Computer science
QA76 Computer software
ISBN
978-1-61208-829-7
ISSN
2308-3492
Organization
Conference
Centric2020 - The 13th Int. Conf. on Advances in Human-oriented and Personalized Mechanisms, Technologies, and Services.
Publisher
IARIA
Submitter
Klein, Eduard
Citation apa
Gasser, T., Klein, E., & Seppänen, L. (2020). Bias – A Lurking Danger that Can Convert Algorithmic Systems into Discriminatory Entities. Centric2020 - The 13th Int. Conf. on Advances in Human-oriented and Personalized Mechanisms, Technologies, and Services. IARIA. https://doi.org/10.24451/arbor.13189
Note
Die Erlaubnis, diese Datei im ARBOR-Repository zu veröffentlichen, wurde eingeholt
File(s)![Thumbnail Image]()
Loading...
open access
Name
Bias_Gasser-Klein-Seppänen_centric_2020_1_10_30004.pdf
License
Publisher
Version
published
Size
498.75 KB
Format
Adobe PDF
Checksum (MD5)
82dcd39dcace344ce8f2cf807234faa1
