Setting the Scene: Artificial In Intelligence: Ethical Concerns - - PowerPoint PPT Presentation

setting the scene
SMART_READER_LITE
LIVE PREVIEW

Setting the Scene: Artificial In Intelligence: Ethical Concerns - - PowerPoint PPT Presentation

Setting the Scene: Artificial In Intelligence: Ethical Concerns Prof Barbara Prainsack Department of Political Science, University of Vienna Department of Global Health & Social Medicine, Kings College London Dialogue seminar with


slide-1
SLIDE 1

Setting the Scene: Artificial In Intelligence: Ethical Concerns

Prof Barbara Prainsack Department of Political Science, University of Vienna Department of Global Health & Social Medicine, King’s College London Dialogue seminar with churches, religious and philosophical organisations, European Parliament, 19 March 2019

slide-2
SLIDE 2

“Rise of the robotic workforce” (Forbes)

slide-3
SLIDE 3

Artificial intelligence techniques

E.g.

  • natural language processing
  • machine learning
  • predictive analytics
  • generative adversarial networks
  • mechatronics and robotics

All consist of material technologies and human practices

[Marvin C. 1990. When old technologies were new: Thinking about electric communication in the late nineteenth century. Oxford University Press]

slide-4
SLIDE 4

EGE Statement on AI, robotics and “autonomous” systems (2018)

  • A. Substantive values
  • 1. Human dignity
  • 2. Autonomy
  • 3. Responsibility
  • 4. Justice, equity, solidarity
  • 5. Democracy
  • 6. Rule of Law and Accountability
  • 7. Security, safety, bodily and mental integrity
  • 8. Data protection and privacy
  • 9. Sustainability
  • B. Process
  • 1. Deliberation
slide-5
SLIDE 5

What we must not forget:

  • The political economy context:

iLeviathan

  • What makes AI possible:

Datafication, platforms

  • Technologies in the service of

values, not the other way round!

[Prainsack B. 2019. Data donation: How to resist the iLeviathan. In: Jenny Krutzinna and Luciano Floridi (eds). The Ethics of Medical Data Donation. Dordrecht: Springer. 9-22.] [Van Dijck J, Poell T, De Waal M. 2018. The platform society: Public values in a connective world. Oxford University Press.]

slide-6
SLIDE 6

[Prainsack B, Buyx A. (2013). A solidarity-based approach to the governance of research biobanks. Medical Law Review 21/1: 71-91. Prainsack B, Buyx A. (2016). Thinking ethical and regulatory frameworks in medicine from the perspective of solidarity on both sides of the Atlantic. Theoretical Medicine and Bioethics 37: 489-501.] [McMahon Prainsack B, Buyx A. Big data governance needs more collective agency: The role of harm mitigation in the governance of data-rich projects. Work in progress.]

Harm mitigation bodies (H (HMB)

Legal remedies

  • Accessible only for primary data

subjects

  • Need to prove culpability and/or

causality

Harm mitigation (outside of the legal domain)

  • Three main functions: (a) collect

information on the types of harm

  • ccurring (b) relay feedback to

improve data use (c) financial support

  • Everybody who can report

significant harm can appeal to it

  • No fault needs to be proven
  • Causal link between an action of a

data processor/controller and the harm only needs to be made plausible

  • Independent governance
slide-7
SLIDE 7

Thank you for your attention

  • barbara.prainsack@univie.ac.at