Please use this identifier to cite or link to this item: https://doi.org/10.1109/ICDE.2019.00063
Title: Collecting and analyzing multidimensional data with local differential privacy
Authors: Wang, N
Xiao, X 
Yang, Y
Zhao, J
Hui, SC
Shin, H
Shin, J
Yu, G
Keywords: cs.CR
cs.CR
cs.CY
cs.DB
cs.LG
Local differential privacy, multidimensional data, stochastic gradient descent
Issue Date: 1-Apr-2019
Publisher: IEEE
Citation: Wang, N, Xiao, X, Yang, Y, Zhao, J, Hui, SC, Shin, H, Shin, J, Yu, G (2019-04-01). Collecting and analyzing multidimensional data with local differential privacy. Proceedings - International Conference on Data Engineering 2019-April : 638-649. ScholarBank@NUS Repository. https://doi.org/10.1109/ICDE.2019.00063
Abstract: © 2019 IEEE. Local differential privacy (LDP) is a recently proposed privacy standard for collecting and analyzing data, which has been used, e.g., in the Chrome browser, iOS and macOS. In LDP, each user perturbs her information locally, and only sends the randomized version to an aggregator who performs analyses, which protects both the users and the aggregator against private information leaks. Although LDP has attracted much research attention in recent years, the majority of existing work focuses on applying LDP to complex data and/or analysis tasks. In this paper, we point out that the fundamental problem of collecting multidimensional data under LDP has not been addressed sufficiently, and there remains much room for improvement even for basic tasks such as computing the mean value over a single numeric attribute under LDP. Motivated by this, we first propose novel LDP mechanisms for collecting a numeric attribute, whose accuracy is at least no worse (and usually better) than existing solutions in terms of worst-case noise variance. Then, we extend these mechanisms to multidimensional data that can contain both numeric and categorical attributes, where our mechanisms always outperform existing solutions regarding worst-case noise variance. As a case study, we apply our solutions to build an LDP-compliant stochastic gradient descent algorithm (SGD), which powers many important machine learning tasks. Experiments using real datasets confirm the effectiveness of our methods, and their advantages over existing solutions.
Source Title: Proceedings - International Conference on Data Engineering
URI: https://scholarbank.nus.edu.sg/handle/10635/156903
ISSN: 1084-4627
DOI: 10.1109/ICDE.2019.00063
Appears in Collections:Staff Publications
Elements

Show full item record
Files in This Item:
File Description SizeFormatAccess SettingsVersion 
1907.00782v1.pdf871.62 kBAdobe PDF

OPEN

Post-printView/Download

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.