Privacy-Preserving Prediction and Learning in Machine Learning Research

Slide Note
Embed
Share

Explore the concepts of privacy-preserving prediction and learning in machine learning research, including differential privacy, trade-offs, prediction APIs, membership inference attacks, label aggregation, classification via aggregation, and prediction stability. The content delves into the challenges, solutions, and trade-offs involved in maintaining privacy while ensuring accurate predictions and learning models.


Uploaded on Aug 03, 2024 | 0 Views


Download Presentation

Please find below an Image/Link to download the presentation.

The content on the website is provided AS IS for your information and personal use only. It may not be sold, licensed, or shared on other websites without obtaining consent from the author. Download presentation by click this link. If you encounter any issues during the download, it is possible that the publisher has removed the file from their server.

E N D

Presentation Transcript


  1. Privacy-preserving Prediction Vitaly Feldman Brain with Cynthia Dwork

  2. Privacy-preserving learning Input: dataset ? = (?1,?1), ,(??,??) Goal: given ? predict ? ? Model Differentially private learning algorithm ? ?(?) ?(? ) 2

  3. Trade-offs Linear regression in ? ? ? more data With ?-DP needs factor [Bassily,Smith,Thakurta 14] Learning a linear classifier over {0,1}? ? ? more data Needs factor [Feldman,Xiao 13] MNIST accuracy ??% with small ?,? vs 99.8% without privacy [AbadiCGMMTZ 16] 3

  4. Prediction Users need predictions not models ?1 ? ?2 ?2 ?2 ? ?? ?? Prediction API Users DP Fits many existing systems 4

  5. Attacks Black-box membership inference with high accuracy [Shokri,Stronati,Song,Shmatikov 17; LongBWBWTGC 18; SalemZFHB 18] 5

  6. Learning with DP prediction Accuracy-privacy trade-off Single prediction query Differentially private prediction : ?: ? ?? ? ? is ?-DP prediction algorithm if for every ? ?, ?(?,?) is ?-DP private w.r.t. ? 6

  7. Label aggregation [HCB 16; PAEGT 17; PSMRTE 18; BTT 18] ? ?1 ?2 ?3 ?? ? = ?? (non-DP) learning algo ? ? ? ? ? ? ? ? 2 ? 1 1 2 ? 3 ?(?) ? 1(?) 2(?) ? ? 2(?) 1(?) 3(?) Differentially private aggregation ? e.g. exponential mechanism ? ?? | ? ?? =?}|/2 7

  8. Classification via aggregation PAC model: Let ? be a class of function over ? For all distributions ? over ? {0,1} output such that w.h.p. ?? (?,?) ? ? ? Opt?? + ? ?-DP prediction ?-DP model Non-private VCdim ? ? VCdim ? ?? Rdim ? ?? ? Realizable case: Rdim ? ?? 1/3 + VCdim ? ?? VCdim ? ? ? Agnostic: Representation dimension [Beimel,Nissim,Stemmer 13] VCdim ? Rdim ? VCdim ? log|?|[KLNRS 08] For many classes Rdim ? = (VCdim ? log ? )[F.,Xiao 13] 8

  9. Prediction stability la [Bousquet,Elisseeff 02]: ?: ? ?? ? is uniformly ?-stable algorithm if for every, neighboring ?,? and ? ?, ? ?,? ? ? ,? ? Convex regression: given ? = ? ?,? For ? over ? ?, minimize: ? ? ?? = (?,?) ?[ (? ?,? ,?)] ? over convex ? ?, where (? ?,? ,?) is convex in ? for all ?,? Convex 1-Lipschitz regression over 2 ball of radius 1: ?-DP prediction ?-DP model Non-private 1 1 ?? 1 ?+? Excess loss: ? ?? ? 9

  10. Beyond aggregation Threshold functions on a line 1 ? Excess error for agnostic learning ?-DP prediction ?-DP model Non-private 1 1 1 1 ?+log? ?+ ?? ? ?? 10

  11. Conclusions Natural setting for learning with privacy Better accuracy-privacy trade-off Paper (COLT 2018): https://arxiv.org/abs/1803.10266 Open problems: o General agnostic learning o Other general approaches o Handling of multiple queries [BTT 18] 11

Related


More Related Content