data protection
play

DATA PROTECTION RIGHT Prof. dr. Mireille Hildebrandt Interfacing - PowerPoint PPT Presentation

GETTING DATA PROTECTION RIGHT Prof. dr. Mireille Hildebrandt Interfacing Law & Technology Vrije Universiteit Brussel Smart Environments, Data Protection & the Rule of Law Radboud University 21/2/17 Hildebrandt SNS seminar Stockholm


  1. GETTING DATA PROTECTION RIGHT Prof. dr. Mireille Hildebrandt Interfacing Law & Technology Vrije Universiteit Brussel Smart Environments, Data Protection & the Rule of Law Radboud University

  2. 21/2/17 Hildebrandt SNS seminar Stockholm 2

  3. what’s next? 1. 1. From m online ine to onlif ife 2. 2. Machine ne Learning rning 3. 3. Data a Protec ectio tion 21/2/17 Hildebrandt SNS seminar Stockholm 3

  4. what’s next? 1. From om on onli line to o on onli life 21/2/17 Hildebrandt SNS seminar Stockholm 4

  5. online → onli f e ■ internet: packet switching & routing, network structure, ■ world wide web: hyperlinking ■ search engines, blogs, social media, web portals ■ web platforms [network effects & filter bubbles; reputation & fake news] ■ mobile applications [moving towards IoT, wearables] ■ IoT: cyberphysical infrastructures [connected cars, smart energy grids] ■ cloud computing, fog computing & edge computing 21/2/17 Hildebrandt SNS seminar Stockholm 5

  6. 21/2/17 Hildebrandt SNS seminar Stockholm 6

  7. 21/2/17 Hildebrandt SNS seminar Stockholm 7

  8. 21/2/17 Hildebrandt SNS seminar Stockholm 8

  9. onli f e: data driven agency ■ creating added value from big data or small data ■ predicting behaviours ■ pre-empting behaviours ■ interplay of backend & frontend of computing systems ■ interfaces enable but they also hide , nudge and force [AB testing, ‘by design’ paradigms] 21/2/17 Hildebrandt SNS seminar Stockholm 9

  10. onli f e: digital unconscious Big Data Space: ce: ■ accumulation of behavioural and other data ■ mobile and polymorphous data & hypothesis spaces ■ distributed storage [once data has been shared, control becomes a challenge] ■ distributed access [access to data or to the inferences, to training set & algos] 21/2/17 Hildebrandt SNS seminar Stockholm 10

  11. 21/2/17 Hildebrandt SNS seminar Stockholm 11

  12. onli f e: digital unconscious Big Data Space: ce: the e envelop elop of big data space drives human agency, providing convenience & resilience Weiser’s calm computi uting , IBM’s auton onom omic ic computi uting :  increasing dependence on the dynamics of interacting data driven cyberphysical systems 21/2/17 Hildebrandt SNS seminar Stockholm 12

  13. 21/2/17 Hildebrandt SNS seminar Stockholm 13

  14. what’s next? 2. Machin ine Learni ning ng 21/2/17 Hildebrandt SNS seminar Stockholm 14

  15. 21/2/17 Hildebrandt SNS seminar Stockholm 15

  16. big data, open data, personal data ■ BIG – volume (but, n=all is nonsense) – variety (unstructured in sense of different formats) – velocity (real time, streaming) ■ OP OPEN EN as opposed to proprietary? reuse? repurposing? public-private? – creating added value is hard work, not evident, no guarantees for return on investment ■ PER ERSONA ONAL data: IoT will contribute to a further explosion of personal data – high risk high gain (think DPIA)? anonymisation will mostly be pseudonymisation! 21/2/17 Hildebrandt SNS seminar Stockholm 16

  17. machine Learning (ML) “ we say that a machine learns: - with respect to a particular task T, - performance metric P, and - type of experience E, if if - the system reliably improves its performance P - at task T, - following experience E .” (Tom Mitchell) http://www.cs.cmu.edu/~tom/mlbook.html 21/2/17 Hildebrandt SNS seminar Stockholm 17

  18. types of machine learning ■ super pervi vised sed (lear arning ning from om exam ample les – requi uire res s labelli elling, ng, doma main in exper ertise tise) ■ reinf inforce rcement ment (lea earnin rning g by correcti rection on - requires uires prior r doma omain in exper erti tise) se) ■ uns nsuper upervised vised (bott ttum up up, induc ucti tive e – danger nger of overfitt tting) ing) 21/2/17 Hildebrandt SNS seminar Stockholm 18

  19. 21/2/17 Hildebrandt SNS seminar Stockholm 19

  20. bias optimisation spurious correlations ■ 2. have a network trained to recognize animal faces ■ 1. present it with a picture of a flower ■ 2. run the algorithms ■ 3. check the output (see what it sees) http://www.nature.com/news/can-we-open-the-black-box-of-ai-1.20731 21/2/17 Hildebrandt SNS seminar Stockholm 20

  21. Wol olper pert: : no no free ee lu lunc nch h theor orem em Wher here d = trainin ning g set; et; f = ‘target’ input -ou outp tput ut relat ationshi ionships; s; h = hypo poth thesi esis (the he algori rith thm's m's gue uess ss for f made de in response ponse to d); ; and C = off-trai training ng- set ‘loss’ associated with f and h (‘generalization error’) How well you do is determined by how ‘aligned’ your learning algorithm P( h|d) is with the actual posterior, P(f|d). Check http://www.no-free-lunch.org 21/2/17 Hildebrandt SNS seminar Stockholm 21

  22. Wol olper pert: : no no free ee lu lunc nch h theor orem em Summary: – The bias that is necessary to mine the data will co-determine the results – This relates to the fact that the data used to train an algorithm is finite – ‘Reality’, whatever that is, escapes the inherent reduction – Data is not the same as what it refers to or what it is a trace of 21/2/17 Hildebrandt SNS seminar Stockholm 22

  23. 21/2/17 Hildebrandt SNS seminar Stockholm 23

  24. 21/2/17 Hildebrandt SNS seminar Stockholm 24

  25. trade-offs ■ NFL FL theo eorem rem – overfitting, overgeneralization ■ trainin ning g set, et, domai main n kno nowled wledge, ge, hypo poth theses ses space, ce, test st set et – accuracy, precision, speed, iteration ■ low w hanging ging frui uit t – may be cheap and/or available but not very helpfull ■ data nor algori rith thms s are object jectiv ive e – bias in the data, bias of the algos, guess what: bias in the output ■ the e more re data, a, the e larger er the e hypo poth theses es sp space, e, the e more ore pattern erns – spurious correlations, computational artefacts 21/2/17 Hildebrandt SNS seminar Stockholm 25

  26. data hoarding & obesitas ■ data obes esitas itas: : lots of data, but often incorrect, incomplete, irrelevant (low hanging fruit) – any personal data stored presents security and other risks sks (need for DPIA, DPbD) – pu purpose rpose limitati tion on is crucial: select ect before re you ou collect lect (and while, and after) ■ pattern ern obesi esitas tas: : trained algorithms can see patterns anywhere, added value? – training set and algorithms ne necessari essarily ly contain bias, this may be problematic (need for DPIA, DPbD) – purpose pu rpose limitati tion on is crucial: to prevent spurious correlations, to test t rele levance nce 21/2/17 Hildebrandt SNS seminar Stockholm 26

  27. agile and lean computing ■ agile e softw tware are developme elopment: nt: – iteration instead of waterfall – collaboration domain experts, data scientists, whoever invests – initial purpose (prediction of behaviour, example: tax office, car insurance) – granular purposing (testing specific patterns, AB testing to nudge specific behaviour) ■ lean n com omputing: uting: – less data = more effective & more efficient ■ meth ethodo dologi logica cal l integri egrity ty: – make your software testable and contestable: mathematical & empirical software verification – secure logging, open source 21/2/17 Hildebrandt SNS seminar Stockholm 27

  28. what’s next? 4. Data a Protect ection ion Law 21/2/17 Hildebrandt SNS seminar Stockholm 28

  29. pr priv ivacy acy and nd aut utonomy onomy ■ th the im impli lica cati tion ons s of of pre-empti tive co computi ting: – AB testing & nudging – pre-emption of our intent, playing with our autonomy – we become subject to decisions of data-driven agents – this choice architecture may generate manipulability 21/2/17 Hildebrandt SNS seminar Stockholm 29

  30. no non-discrimination discrimination ■ three ee type pes of bi bias: – bias inherent in any action-perception-system (APS) – bias that some would qualify as unfair – bias that discriminates on the basis of prohibited legal grounds 21/2/17 Hildebrandt SNS seminar Stockholm 30

  31. the opacity argument in ML: 1. 1. intent ntional nal conceal alment ment – trade de secre rets ts, , IP right hts, s, pub ublic c security urity 2. 2. we we have learned d to read and write, , not ot to code or do machine hine learning ing – monopoly of the new ‘clerks’, the end of democracy 3. 3. mismatc match h betwee etween mathe hematic matical al optimi miza zation tion and human an semant ntics ics – when it comes to law and justice we cannot settle for ‘computer says no’ – inspired by: Jenna Burrell, How the machine ‘thinks’: Understanding opacity in machine learning algorithms’, in Big Data ta & Society ty, January-June 2016, 1-12 21/2/17 Hildebrandt SNS seminar Stockholm 31

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend