
Almost Instance-optimal Clipping for Summation Problems in the
Shule Model of Dierential Privacy
Wei Dong
Nanyang Technological University
Singapore, Singapore
wei_dong@ntu.edu.sg
Qiyao Luo
OceanBase, Ant Group
Shanghai, China
luoqiyao.lqy@antgroup.com
Giulia Fanti
Carnegie Mellon University
Pittsburgh, United States
gfanti@andrew.cmu.edu
Elaine Shi
Carnegie Mellon University
Pittsburgh, United States
runting@gmail.com
Ke Yi
Hong Kong University of Science and
Technology
Hong Kong, Hong Kong
yike@cse.ust.hk
Abstract
Dierentially private mechanisms achieving worst-case optimal
error bounds (e.g., the classical Laplace mechanism) are well-studied
in the literature. However, when typical data are far from the worst
case, instance-specic error bounds—which depend on the largest
value in the dataset—are more meaningful. For example, consider
the sum estimation problem, where each user has an integer
𝑥
𝑖
from the domain
{
0
,
1
, . . . ,𝑈 }
and we wish to estimate
Í
𝑖
𝑥
𝑖
. This
has a worst-case optimal error of
𝑂 (𝑈 /𝜀)
, while recent work has
shown that the clipping mechanism can achieve an instance-optimal
error of
𝑂 (max
𝑖
𝑥
𝑖
· log log𝑈 /𝜀)
. Under the shue model, known
instance-optimal protocols are less communication-ecient. The
clipping mechanism also works in the shue model, but requires
two rounds: Round one nds the clipping threshold, and round
two does the clipping and computes the noisy sum of the clipped
data. In this paper, we show how these two seemingly sequential
steps can be done simultaneously in one round using just 1
+𝑜 (
1
)
messages per user, while maintaining the instance-optimal error
bound. We also extend our technique to the high-dimensional sum
estimation problem and sparse vector aggregation (a.k.a. frequency
estimation under user-level dierential privacy).
CCS Concepts
• Security and privacy → Privacy-preserving protocols.
Keywords
Dierential privacy; sum estimation
ACM Reference Format:
Wei Dong, Qiyao Luo, Giulia Fanti, Elaine Shi, and Ke Yi. 2024. Almost
Instance-optimal Clipping for Summation Problems in the Shue Model of
Dierential Privacy. In Proceedings of the 2024 ACM SIGSAC Conference on
Computer and Communications Security (CCS ’24), October 14–18, 2024, Salt
Lake City, UT, USA. ACM, New York, NY, USA, 15 pages. https://doi.org/10.
1145/3658644.3690225
This work is licensed under a Creative Commons Attribution
International 4.0 License.
CCS ’24, October 14–18, 2024, Salt Lake City, UT, USA
© 2024 Copyright held by the owner/author(s).
ACM ISBN 979-8-4007-0636-3/24/10
https://doi.org/10.1145/3658644.3690225
1 Introduction
The shue model [
10
,
12
,
20
] of dierential privacy (DP) is widely-
studied in the context of DP computation over distributed data.
The model has 3 steps: (1) Each client uses a randomizer
R(·)
to
privatize their data. (2) A trusted shuer randomly permutes the
inputs from each client and passes them to an untrusted analyzer,
which (3) conducts further analysis. Unlike the central model of DP,
where a trusted curator has access to all the data, the shue model
provides stronger privacy protection by removing the dependency
on a trusted curator. Unlike the local model, where clients send
noisy results to the analyzer directly, the addition of the trusted
shuer allows for a signicantly improved privacy-accuracy trade-
o. For problems like bit counting, shue-DP achieves an error of
𝑂 (
1
/𝜀)
with constant probability
1
[
23
], matching the best error of
central-DP, while the smallest error achievable under local-DP is
𝑂 (
√
𝑛/𝜀) [11, 13].
The summation problem, a fundamental problem with applica-
tions in statistics [
9
,
27
,
33
], data analytics [
15
,
44
], and machine
learning such as the training of deep learning models [
1
,
2
,
8
,
41
]
and clustering algorithms [
42
,
43
], has been studied in many works
under the shue model [
6
,
7
,
12
,
24
–
26
]. In this problem, each
user
𝑖 ∈ [𝑛]
:
= {
1
, . . . , 𝑛}
holds an integer
𝑥
𝑖
∈ {
0
,
1
, . . . ,𝑈 }
and
the goal is to estimate
Sum(𝐷)
:
=
Í
𝑖
𝑥
𝑖
, where
𝐷 = (𝑥
1
, . . . , 𝑥
𝑛
)
.
All of these works for sum estimation under shue-DP focus on
achieving an error of
𝑂 (𝑈 /𝜀)
. Such an error can be easily achieved
under central-DP, where the curator releases the true sum after
masking it with a Laplace noise of scale
𝑈 /𝜀
. In the shue-DP
model, besides error, another criterion that should be considered
is the communication cost. The recent shue-DP summation pro-
tocol of [
24
] both matches the error of central-DP and achieves
optimal communication. More precisely, it achieves an error that is
just 1
+𝑜 (
1
)
times that of the Laplace mechanism, while each user
just sends in expectation 1
+𝑜 (
1
)
messages, each of a logarithmic
number of bits.
However, in real applications (as well as in [24]), 𝑈 must be set
independently of the dataset; to account for all possible datasets,
it should be conservatively large. For instance, if we only know
that the
𝑥
𝑖
’s are 32-bit integers, then
𝑈 =
2
32
−
1. Then the error
1
In Section 1, all stated error guarantees hold with constant probability. We will make
the dependency on the failure probability 𝛽 more explicit in later sections.
BBAAD9C20180234D78A0072836F0B6D0E2B9B2091CC78BA0ADD98334B1332B246B4CB438915CAB0D22592008984631EBB0E921BAC1D09B911BBFC23670BE3FD6241144ADDF2589F764B7292767674B426EF7E18E714D1414D8BA819CE1F6ACC8D7E62996BE3
评论