scispace - formally typeset
Open AccessJournal ArticleDOI

Privacy-Preserving Distributed Linear Regression on High-Dimensional Data

Reads0
Chats0
TLDR
A hybrid multi-party computation protocol that combines Yao’s garbled circuits with tailored protocols for computing inner products is proposed, suitable for secure computation because it uses an efficient fixed-point representation of real numbers while maintaining accuracy and convergence rates comparable to what can be obtained with a classical solution using floating point numbers.
Abstract
We propose privacy-preserving protocols for computing linear regression models, in the setting where the training dataset is vertically distributed among several parties. Our main contribution is a hybrid multi-party computation protocol that combines Yao’s garbled circuits with tailored protocols for computing inner products. Like many machine learning tasks, building a linear regression model involves solving a system of linear equations. We conduct a comprehensive evaluation and comparison of different techniques for securely performing this task, including a new Conjugate Gradient Descent (CGD) algorithm. This algorithm is suitable for secure computation because it uses an efficient fixed-point representation of real numbers while maintaining accuracy and convergence rates comparable to what can be obtained with a classical solution using floating point numbers. Our technique improves on Nikolaenko et al.’s method for privacy-preserving ridge regression (S&P 2013), and can be used as a building block in other analyses. We implement a complete system and demonstrate that our approach is highly scalable, solving data analysis problems with one million records and one hundred features in less than one hour of total running time.

read more

Citations
More filters
Proceedings ArticleDOI

Cryptϵ: Crypto-Assisted Differential Privacy on Untrusted Servers

TL;DR: This work proposes Cryptε, a system and programming framework that achieves the accuracy guarantees and algorithmic expressibility of the central model without any trusted data collector like in the local model and demonstrates Cryptε's practical feasibility with extensive empirical evaluations on real world datasets.
Proceedings ArticleDOI

Federated Doubly Stochastic Kernel Learning for Vertically Partitioned Data

TL;DR: This paper uses random features to approximate the kernel mapping function and use doubly stochastic gradients to update the solutions, which are all computed federatedly without the disclosure of data to prove that FDSKL has a sublinear convergence rate, and can guarantee the data security under the semi-honest assumption.
Proceedings ArticleDOI

PrivFL: Practical Privacy-preserving Federated Regressions on High-dimensional Data over Mobile Networks

TL;DR: In this paper, the authors proposed PrivFL, a privacy-preserving system for training linear and logistic regression models and oblivious predictions in the federated setting, while guaranteeing data and model privacy as well as ensuring robustness to users dropping out.
Journal ArticleDOI

Privacy-Preserving Asynchronous Vertical Federated Learning Algorithms for Multiparty Collaborative Learning.

TL;DR: In this paper, the authors proposed an asynchronous federated stochastic gradient descent (AFSGD-VP) algorithm and its two variance reduction variants, including SVRG and SAGA on the VP data, under the condition of strong convexity and without any restrictions of staleness.
Journal ArticleDOI

Federated Learning and Differential Privacy: Software tools analysis, the Sherpa.ai FL framework and methodological guidelines for preserving data privacy

TL;DR: In this paper, the authors present the http URL Federated Learning framework that is built upon an holistic view of federated learning and differential privacy, and the definition of methodological guidelines for developing artificial intelligence services based on Federated learning.
References
More filters
Book

Numerical Optimization

TL;DR: Numerical Optimization presents a comprehensive and up-to-date description of the most effective methods in continuous optimization, responding to the growing interest in optimization in engineering, science, and business by focusing on the methods that are best suited to practical problems.
Book

Machine Learning : A Probabilistic Perspective

TL;DR: This textbook offers a comprehensive and self-contained introduction to the field of machine learning, based on a unified, probabilistic approach, and is suitable for upper-level undergraduates with an introductory-level college math background and beginning graduate students.
Book

The algebraic eigenvalue problem

TL;DR: Theoretical background Perturbation theory Error analysis Solution of linear algebraic equations Hermitian matrices Reduction of a general matrix to condensed form Eigenvalues of matrices of condensed forms The LR and QR algorithms Iterative methods Bibliography.
Book

The Algorithmic Foundations of Differential Privacy

TL;DR: The preponderance of this monograph is devoted to fundamental techniques for achieving differential privacy, and application of these techniques in creative combinations, using the query-release problem as an ongoing example.