Processing math: 3%
We use cookies to improve your experience with our site.

Hadamard Encoding Based Frequent Itemset Mining under Local Differential Privacy

Dan Zhao, Su-Yun Zhao, Hong Chen, Rui-Xuan Liu, Cui-Ping Li, Xiao-Ying Zhang

downloadPDF
赵丹, 赵素云, 陈红, 刘睿瑄, 李翠平, 张晓莹. 本地化差分隐私技术下基于哈达玛编码的频繁项集挖掘[J]. 计算机科学技术学报, 2023, 38(6): 1403-1422. DOI: 10.1007/s11390-023-1346-7
引用本文: 赵丹, 赵素云, 陈红, 刘睿瑄, 李翠平, 张晓莹. 本地化差分隐私技术下基于哈达玛编码的频繁项集挖掘[J]. 计算机科学技术学报, 2023, 38(6): 1403-1422. DOI: 10.1007/s11390-023-1346-7
Zhao D, Zhao SY, Chen H et al. Hadamard encoding based frequent itemset mining under local differential privacy. JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY 38(6): 1403−1422 Nov. 2023. DOI: 10.1007/s11390-023-1346-7.
Citation: Zhao D, Zhao SY, Chen H et al. Hadamard encoding based frequent itemset mining under local differential privacy. JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY 38(6): 1403−1422 Nov. 2023. DOI: 10.1007/s11390-023-1346-7.
赵丹, 赵素云, 陈红, 刘睿瑄, 李翠平, 张晓莹. 本地化差分隐私技术下基于哈达玛编码的频繁项集挖掘[J]. 计算机科学技术学报, 2023, 38(6): 1403-1422. CSTR: 32374.14.s11390-023-1346-7
引用本文: 赵丹, 赵素云, 陈红, 刘睿瑄, 李翠平, 张晓莹. 本地化差分隐私技术下基于哈达玛编码的频繁项集挖掘[J]. 计算机科学技术学报, 2023, 38(6): 1403-1422. CSTR: 32374.14.s11390-023-1346-7
Zhao D, Zhao SY, Chen H et al. Hadamard encoding based frequent itemset mining under local differential privacy. JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY 38(6): 1403−1422 Nov. 2023. CSTR: 32374.14.s11390-023-1346-7.
Citation: Zhao D, Zhao SY, Chen H et al. Hadamard encoding based frequent itemset mining under local differential privacy. JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY 38(6): 1403−1422 Nov. 2023. CSTR: 32374.14.s11390-023-1346-7.

本地化差分隐私技术下基于哈达玛编码的频繁项集挖掘

Hadamard Encoding Based Frequent Itemset Mining under Local Differential Privacy

Funds: This paper was supported by the National Natural Science Foundation of China under Grant Nos. 61772537, 61772536, 62072460, 62076245, and 62172424, the National Key Research and Development Program of China under Grant No. 2018YFB1004401, and Beijing Natural Science Foundation under Grant No. 4212022.
More Information
    Author Bio:

    Dan Zhao received his B.S. degree from Jiangsu University, Zhenjiang, in 2010, his M.S. degree from Guizhou University, Guizhou, in 2015, and his Ph.D. degree in computer application technology from Renmin University of China, Beijing, in 2022. He currently works in Institute of Scientific and Technical Information of China, Beijing. His main research fields are localized differential privacy and data publishing

    Su-Yun Zhao received her Ph.D. degree in computer science from The Hong Kong Polytechnic University, Hong Kong, in 2009. She works as an associate professor in Renmin University of China, Beijing. Her research interests focus on machine learning, including ensemble learning, generative adversarial learning, weak supervised learning and so on

    Hong Chen received her Bachelor’s and Master’s degrees from Renmin University of China, Beijing, in 1986 and 1989, respectively, and her Ph.D. degree in computer science from Chinese Academy of Sciences, Beijing, in 2000. She is currently with the School of Information, Renmin University of China, Beijing. Her current research interests include high-performance database systems, data warehouse and data mining, stream data management, and data management in wireless sensor networks

    Rui-Xuan Liu received her B.S. degree from China University of Petroleum, Beijing, and her Ph.D. degree from Renmin University of China, Beijing, in 2023, both in computer application technology. She currently works in Emory University, Atlanta. Her research interests include privacy protection, machine learning and data mining

    Cui-Ping Li received her B.E. and M.E. degrees from Xi’an Jiaotong University, Xi’an, in 1994 and 1997, respectively, and her Ph.D. degree in computer science from Institute of Computing Technology, Chinese Academy of Sciences, Beijing, in 2003. She is currently a professor with Renmin University of China, Beijing. Her current research interests include database systems, data warehousing, and data mining

    Xiao-Ying Zhang received her B.S. degree from Shandong University, Jinan, in 2009, and her Ph.D. degree from Renmin University of China, Beijing, in 2016, both in computer application technology. She works as a senior engineer in Renmin University of China, Beijing. Her research interests focus on data management and privacy preservation

    Corresponding author:

    Hong Chen: chong@ruc.edu.cn

  • 摘要:

    频繁项集挖掘是数据挖掘中的重要问题,而在挖掘中使用本地化差分隐私技术收集敏感信息能有效的保障用户隐私。大部分本地化差分隐私技术主要应对的是用户只有一个项值的情况,而当每个用户都有一组项值时,现有的本地化差分隐私技术使用“填充和采样”的步骤来获得频繁项集及其频率。目前最优的算法SVSM也需要平衡结果的方差和偏差来获得相对准确的结果。如何能同时降低方差和偏差是目前满足本地化差分隐私技术的频繁项集挖掘的挑战。

    为了解决这个问题,我们提出一种满足值-级别的本地化隐私方法IHFO,它是首次把哈达玛编码引入来应对用户拥有值集的情况,因为哈达玛编码能将任意多的值集编码为一个固定的向量,并同时使用频数估计和均值估计的本地化差分隐私算法对其增加扰动。

    在此基础上我们提出了一种优化的联合项目集挖掘方法(O-UISM)来进行频繁项集挖掘。O-UISM将基于SVSM中应用的PSFO和本文的IHFO结合到一个框架中,因为我们发现PSFO算法在不考虑频数准确性的前提下,寻找频繁一项集更加准确,而本文提出的IHFO对于获取频率估计更加准确,因此将两者的优势一起来使用,从而可以获得更加准确的频繁项集及其频率。

    最后我们从理论和实验上证明的在相同的隐私保护条件下,O-UISM在寻找频繁项集和项集的频率估计方面都优于之前的算法。

    Abstract:

    Local differential privacy (LDP) approaches to collecting sensitive information for frequent itemset mining (FIM) can reliably guarantee privacy. Most current approaches to FIM under LDP add “padding and sampling” steps to obtain frequent itemsets and their frequencies because each user transaction represents a set of items. The current state-of-the-art approach, namely set-value itemset mining (SVSM), must balance variance and bias to achieve accurate results. Thus, an unbiased FIM approach with lower variance is highly promising. To narrow this gap, we propose an Item-Level LDP frequency oracle approach, named the Integrated-with-Hadamard-Transform-Based Frequency Oracle (IHFO). For the first time, Hadamard encoding is introduced to a set of values to encode all items into a fixed vector, and perturbation can be subsequently applied to the vector. An FIM approach, called optimized united itemset mining (O-UISM), is proposed to combine the padding-and-sampling-based frequency oracle (PSFO) and the IHFO into a framework for acquiring accurate frequent itemsets with their frequencies. Finally, we theoretically and experimentally demonstrate that O-UISM significantly outperforms the extant approaches in finding frequent itemsets and estimating their frequencies under the same privacy guarantee.

  • Frequent itemset mining (FIM), a branch of machine learning used for applications such as weblog mining and trend analysis, has recently attracted interest from many enterprises and researchers[1-3]. FIM aims to identify the k most frequent itemsets and their frequencies. The discovery of frequent itemsets can serve valuable economic and research purposes, e.g., mining association rules[4], predicting user behavior[5], and finding correlations[6]. However, the direct publication of frequent itemsets incurs many risks, including leaking user preferences or individual transactions that include sensitive information. Therefore, to obtain statistical information while protecting user privacy is an important research direction.

    This paper reports on a study of methods for frequent itemset mining (FIM) of transaction databases while guaranteeing privacy. Differential privacy (DP)[7, 8] is an appealing and strict protection technology that can guarantee privacy even in the worst case. DP has become the de facto standard notion of privacy in research on private data analysis. A large domain and heterogeneous size are the two main characteristics in FIM[9]. FIM with central DP is a mature research field[1, 10]. In both interactive and non-interactive frameworks, centralized DP protocols for FIM seek to balance utility and privacy. However, DP requires a central trusted authority, which may leak private information. Protecting users' privacy requires the data aggregator to be trustworthy, but data aggregators lack credibility in most real-time applications. Consequently, as a decentralized technology, local differential privacy (LDP) has been considered for FIM; examples include LDPMiner[11], PrivSet[3], and set-value itemset mining (SVSM)[9].

    Motivation and Challenges. Our main motivation is to obtain more accurate statistical results than the state-of-the-art methods while guaranteeing the privacy. It is challenging to achieve these goals.

    Unbiased and Low Variance. Most existing LDP approaches have an evident limitation in FIM when each user transaction has a set of items. LDPMiner[11] dividing the privacy budget for two phases has a high variance of results. PrivSet[9] perturbs items as a subset but with a large domain. SVSM[3], a state-of-the-art protocol, should balance the bias and variance. Currently, it is still a challenging issue to obtain unbiased and low-variance frequencies for transactions with varied lengths. Thus, in this study, we develop an unbiased approach with low variance for FIM while protecting the privacy of items in each transaction.

    Contribution. Accuracy and privacy have always been considered in the academy and real world. In the FIM scenario in which a user has a set of items, the privacy (user-level) can be divided into sensitive privacy and non-sensitive privacy, and then we can sacrifice some non-sensitive privacy to obtain accurate results while guaranteeing sensitive privacy (Item-Level). It is considerable to sacrifice non-sensitive privacy of the user level to define the Item-Level LDP, which can improve the accuracy of the statistical results. Especially, if each user has only one item, the protection of the user level is equal to that of the Item-Level.

    We split user-level privacy into two parts: the detail of items y and the length li of transaction of Ti. For the particular case in which li is not sensitive and all the items should be considered as a whole, we define the Item-Level LDP for the privacy of items y (li = 1 is the traditional LDP case). We sacrifice non-sensitive privacy for more accurate results in Fig.1. This paper introduces an innovative Frequency Oracle (FO) approach, denoted as the Integrated-with-Hadamard-Transform-Based Frequency Oracle (IHFO), which leverages Hadamard encoding to achieve an unbiased and enhanced accuracy in frequency estimation while ensuring item-level local differential privacy (LDP). This approach builds upon the foundational work in privacy-preserving frequency oracles as delineated by Cormode et al.[12]. Further advancing the state of the art, Wang et al.[3] contributed additional FO methodologies, namely PSFO and PrivSet. PSFO is designed to strike a balance between minimizing variance and reducing bias, whereas PrivSet is tailored to identify frequent itemsets within a constrained error margin, particularly effective when the itemset size |I| is small.

    Figure  1.  Innovations: sacrificing non-sensitive privacy for more accurate results.

    Based on IHFO, this paper presents an approach for FIM called United Itemset Mining (UISM). Like SVSM and LDPMiner, UISM divides the dataset into two parts: one to identify frequent singleton items and the other to update the frequent itemsets and estimate their frequencies. In both parts, UISM adopts IHFO to mine frequent items. We find that IHFO performs significantly well for the second part and PSFO performs better than the proposed IHFO when finding the rankings of singleton items. Therefore, we propose an optimized version called O-UISM, which combines the advantages of PSFO (l = 1) and IHFO. O-UISM applies PSFO to search for frequent singleton items in the first part, and IHFO is used to obtain the estimated frequencies of frequent itemsets. Our experimental evaluation demonstrates the effectiveness of UISM and O-UISM. Table 1 presents an overview of FIM solutions.

    Table  1.  Comparative Overview of FIM Solutions
    FIM Method Identification Accuracy
    SVSM <PSFO (l = 1),
    PSFO (l = adapl)>[3]
    General General
    PrivSet<PSFO (l = 1), PrivSet> General General
    UISM<IHFO, IHFO> General Good
    O-UISM<PSFO (l = 1), IHFO> Good Good
    Note: Identification refers to the accuracy rankings of the frequent items/itemsets, and accuracy denotes the accuracy frequencies of the frequent items/itemsets.
    下载: 导出CSV 
    | 显示表格

    In summary, the main contributions of this paper are as follows.

    ● We define an Item-Level LDP and propose a Hadamard encoding based FO approach, namely IHFO, under the Item-Level LDP framework. The IHFO model encodes each itemset into a fixed length vector, which actually achieves the goal of unbiased FO estimating with lower variance and reducing the computation complexity.

    ● We provide an unbiased FIM approach, namely O-UISM. O-UISM combines PSFO and IHFO into a framework to achieve accuracy frequent itemsets together with their frequencies.

    ● We theoretically prove the unbiasedness and the lower variance of our protocols. Numerical experiments demonstrate that O-UISM can obtain more accurate frequent itemsets and frequencies with a lower squared error.

    Roadmap. The remainder of this paper is organized as follows. We discuss related work and the preliminary of this work in Section 2. In Section 3, we definite the Item-Level LDP and propose our approaches to address the problems. Then, we provide a theoretical analysis of privacy and accuracy in Section 4. The experimental results are presented in Section 5. Finally, Section 6 concludes the paper.

    In this section, we firstly introduce the related work, and then discuss the preliminary in this paper. Table 2 summarizes the notations used in this paper.

    Table  2.  Notations
    Notation Definition
    ε Privacy budget
    I Domain of items
    Ψ Perturbation algorithm
    n Number of users
    v,x Singleton item
    y Itemset
    \boldsymbol b_i Encoding vector of user i
    T_i Transaction of user i
    {\boldsymbol{H}}_r Hadamard matrices of order 2^r for
    every non-negative integer r
    {\boldsymbol{H}}(v) Column vector that item v maps to {\boldsymbol{H}}_r
    SI Set of frequent singleton items
    CS Set of candidate frequent itemsets
    FIS Set of frequent itemsets
    w.p. With probability
    下载: 导出CSV 
    | 显示表格

    Frequent itemset mining (FIM) is a well-studied problem in data mining. The FIM problem is identifying the frequent set of items that appear simultaneously in many users' transactions while guaranteeing privacy. Many protocols address this problem in the central DP[1, 13] and LDP settings[3, 9, 11].

    DP was first introduced by Dwork et al.[8] and is currently the de facto standard of data privacy. Centralized DP has been the subject of numerous studies, both theoretical[14] and practical[15, 16]. Without a trusted aggregator, LDP can protect user privacy during data collection, addressing privacy leakage from the root[12, 17-19] in many fields such as frequency oracle[20, 21], mean value[20, 22], FIM[3, 11], key-value pair estimation[23, 24], data monitoring[25], federated learning[26-28] and so on. Several variants of LDP and the corresponding mechanisms have been studied. Personalized LDP (PLDP)[29] divides users into different groups according to different privacy budgets; by contrast, condensed LDP[30], utility-optimized LDP[31], and input discriminative LDP (IN-LDP)[32] group the inputs based on defined rules to improve transmission accuracy. When the transactions have a set of items, user-level privacy needs to protect the transactions as a whole, whereas item-level privacy needs to protect the detailed items. The purpose of these definitions is to weaken the notion for the specialization of LDP.

    Protecting user privacy during FIM is essential, and the use of DP in FIM has previously been investigated[1, 13]. In both interactive[1, 13] and non-interactive frameworks[33], centralized DP protocols for FIM seek to balance utility and privacy. Because data aggregators are not credible in most real applications, protocols under local DP technologies are seldom applied to address this FIM problem[3, 11]. Previously, Evfimievski et al.[34] considered a more comfortable setting in which each transaction is assumed to have a fixed item length. However, this assumption cannot be applied to transactions of varied lengths. Qin et al. proposed LDPMiner[11] to find frequent singletons while satisfying \varepsilon -LDP by adopting “padding and sampling”. One natural LDP algorithm is for each user to sample only one item and use known algorithms to send, which may miss the information of most items. PrivSet[9] considers items as a whole, but the generated output range is too large. Then, the severe problem with PrivSet is that its calculation is prone to overflow when the parameters I and l are large. Furthermore, Wang et al.[3] proposed the state-of-the-art SVSM, which can identify both singletons and itemsets. However, the frequency estimation of SVSM is biased. Thus, it would be promising and desirable for a new approach that guarantees unbiased estimation and obtains more accurate frequencies under the LDP framework.

    Let there be n users and the domain of items I . Then, we obtain a set \Gamma(I) consisting of all nonempty subsets of I , i.e., \Gamma(I) = \{y|y\subseteq I, y \not\subset \emptyset \} .

    The aggregator aims to obtain the frequent items (or itemsets) of the dataset together with their frequencies. The frequent items (itemsets) can be defined by either determining the top- k frequent items (itemsets) or finding items (itemsets) whose frequencies are above a certain threshold. The frequency of any item x\in I is defined as the number of transactions that contain x , i.e., f_x: = |\{i|x\in T_i\}| . Similarly, the frequency of any itemset y \in\Gamma(I) is defined as f_y: = |\{i|y\subseteq T_i\}| .

    Definition 1 (Top- k Frequent Itemset Mining). Given n users with transaction T_i (1\leqslant i\leqslant n) from domain {\cal{V}} , a value y \in {\cal{V}} is a top- k frequent itemset if the frequency f_y = |\{i|1\leqslant i\leqslant n, T_i\in {\cal{V}}\}| is ranked among the top k frequencies of all possible itemsets.

    LDP is a local model of DP for collecting user data without a credible aggregator[11, 19, 35, 36]. An LDP algorithm \Psi ensures that the probability of one value being sent to the aggregator approximates the probability of any other values being sent. The formal privacy requirement satisfies \varepsilon -LDP as follows.

    Definition 2 ( \varepsilon -Local Differential Privacy). Given an algorithm \Psi with a domain Dom(\Psi) and a range Ran(\Psi) , for items t and t' ( t,t' \in Dom(\Psi) ), the same output s ( s \in Ran(\Psi) ) is transformed through algorithm \Psi while satisfying the inequality (1), and we deem that \Psi satisfies \varepsilon -LDP[7].

    \dfrac{{\mathbb{P}}[\Psi(t) = s]}{{\mathbb{P}}[\Psi(t') = s] }\leqslant {\rm{e}}^\varepsilon. (1)

    We can adjust the privacy budget \varepsilon to balance data availability and privacy. Moreover, LDP can provide a more robust privacy protection level than a centralized framework because each user reports only the perturbed data.

    The purpose of collection is to obtain frequency oracle (FO). FO is a core issue under the LDP framework, which has attracted a lot of theoretical and practical attention. FO approaches enable to estimate the frequency of any item/itemset in the domain. Wang et al.[19] introduced an abstract framework of FO approaches, and showed that most previously proposed approaches can be placed within it.

    The Hadamard (discrete Fourier) transform, known as the Walsh-Fourier transform, is described by an orthogonal-binary matrix {\boldsymbol{H}} with dimensions of 2^r\times 2^r (where r is any nonnegative integer). Its orthogonality has the natural advantage of classification in FO. The binarization is convenient for computer calculations and transmission, and the entry in [row, col] is

    H[row,col] = (-1)^{Count{{{\mathbb{1}}}}(row {{\&}} col)},

    where row\in[0,2^r-1],col\in[0,2^r-1] and Count{\mathbb{1}}() counts the number of 1 in a binary integer. Moreover, the fast generation of the Hadamard matrix designed by the well-known Sylvester construction is as follows.

    \boldsymbol H_{r+1} = \left(\begin{matrix} \boldsymbol H_{r} & \boldsymbol H_{r}\\ \boldsymbol H_{r} & -\boldsymbol H_{r} \end{matrix}\right) ,

    where \boldsymbol H_0 = (1) and r \geqslant 0 .

    Existing FO algorithms, such as OLH (optimized local hashing) and HTR, obtain their statistics in situations where each user transaction has only one item. Qin et al.[11] and Wang et al.[3] balanced the variance and bias to address the extension to each user transaction T_i ( i = 1,2,\ldots,n ) having a set of items v _i = \{x|x\in T_i, x\in I\} . However, SVSM works less efficiently under such conditions when the lengths of transactions vary. In this case, it is challenging to develop an approach that guarantees an unbiased and low-variance frequency estimation.

    PSFO[3]. The PSFO approach addresses the scenario in which each transaction has a set of items derived from “padding and sampling” and then applies optimized local hashing (OLH) to report the items. Finally, the frequency estimation is multiplied by the padding parameter l to obtain an accurate result. The variance in PSFO is

    Var(f_x) = n l^2\times \dfrac{4{\rm{e}}^\varepsilon}{({\rm{e}}^\varepsilon-1)^2}.

    The choice of l is important when using PSFO. If l is small, there is less variance but more bias; if l is large, there is less bias but more variance.

    PrivSet[9]. In the mechanism, the set-valued data is sanitized holistically by randomly outputting a fixed-size subset of items \hat{s} . First, each user uses “padding” to generate T_i' according to l , where T_i' contains the original T_i and ( l-|T_i| ) padding items. Second, each user generates a perturbed subset \hat{s} from a discrete probability distribution of all candidate itemsets whose weights follow (2):

    Weight(s) = \exp(\varepsilon\times u(T_i',\hat{s})), (2)

    where u(T_i',\hat{s}) = [T_i'\cap \hat{s} \neq \emptyset] . If T_i'\cap \hat{s} = \emptyset then u(T_i',\hat{s}) = 1 (otherwise 0). Therefore, the probability of each \hat{s} is {Weight(\hat{s})}/{\sum\nolimits_{s\in S}Weight(s)} .

    SVSM. By using PSFO repeatedly, SVSM is the state-of-the-art approach for solving FIM in the LDP setting, where each user transaction has a set of items. SVSM divides a database {{D}} into three parts {{D}}_{\rm{A}} , {{D}}_{\rm{B}} and {{D}}_{\rm{C}} , where {{D}}_{\rm{A}} , {{D}}_{\rm{B}} and {{D}}_{\rm{C}} are applied to obtain frequent single items, adaptive l , and a final frequent itemset, respectively. When we identify a frequent set of terms in {{D}}_{\rm{A}} , a small l = 1 does not affect the ranking of items because the bias tends to be in the same direction for all items. However, when we need to obtain the frequencies of itemsets in {{D}}_{\rm{C}} , an adaptive l from {{D}}_{\rm{B}} is used to balance the bias and variance, which could cause bias and inaccurate results. All steps adopt PSFO, which is based on the OLH algorithm. In the first part, SVSM finds the set of frequent singleton items SI to construct the candidate itemset CI at the expense of frequency accuracy by setting l = 1 . Then, the adaptive l is required for accurate frequency acquisition, which is the task of the second part. Finally, by combining CI and the adaptive l , the frequent itemsets and their respective frequencies are updated in the third part. Note that CI consists of the top- k frequent itemsets obtained through (3) based on SI ,

    f_y = \prod\limits_{x\in y,x\in SI} \dfrac{0.9\times f_x}{\max_{x'\in SI} f_{x'}}, (3)

    where y denotes any itemset and the factor of 0.9 is fixed in [3], which can lower the normalized estimates for the most frequent item.

    Our work is on frequency estimation and means estimation for FIM with LDP, which is related to these existing methods[3, 9, 11], as we must handle a variable transaction length. Moreover, when a user has a set of items, it can be treated as a small database, which can view the problem from the user level and item level. Inspired by the analysis of the user level and item level[38] in DP, although the above LDP algorithms are user-level, they are different. For example, these studies[3, 11, 34] randomly select an item and use the item-level perturbation algorithm, and PrivSet[9] perturbs a subset of items as a whole. This raises a problem that if each local user protects the privacy for each item in his/her own small database, the privacy for all items as a whole is protected. Therefore, this paper defines the Item-Level LDP and proposes two methods to address the above problems.

    In this section, we define the Item-Level LDP and propose a baseline approach called IHFO to handle the fact that each transaction has a set of items. IHFO protects all items in each transaction while guaranteeing an unbiased and low-variance frequency estimation. For frequent item/itemset mining (item mining and itemset mining are similar; thus, we discuss them together), we propose a general version, namely UISM. And then, we suggest an optimal version called O-UISM, which combines PSFO ( l = 1) and IHFO into a framework to achieve accuracy frequent itemsets together with their frequencies. The protocol structure of UISM is given in Fig.2. In the framework, the dataset is divided in two parts. The first part is applied to obtain frequent singleton items, and the second part is applied to obtain frequent itemsets with their frequencies. Both parts need to encode the pre-processed items, and then report to the aggregator with LDP perturbation. The aggregator sends the decoded results of the first part to the second part, and decodes the collection of the second part to output the frequent itemset with their frequencies. If the protocol of the first part is replaced by PSFO ( l = 1), this framework is called O-UISM.

    Figure  2.  Illustration of UISM.

    In the LDP case, the definition is too strict to guarantee the accuracy of the dataset for the user level. Because user i has multiple items T_i , one natural user level algorithm is for each user to sample only one item and use known algorithms to send, which may miss the information of most items. We consider that if |T_i|<<|I| and |T_i|>0 , the length of transaction |T_i| is not sensitive for the user, and each item in the transaction is sensitive. Then, the items in the transaction need to be protected more carefully than the number of items. Inspired from the view of specialization of LDP[29-32], we develop a loose definition of LDP, namely Item-Level \varepsilon -LDP 3, to design protocols that offer desirable statistical utility while preserving privacy.

    Definition 3 (Item-Level \varepsilon -Local Differential Privacy). Given an algorithm \Psi with a domain Dom(\Psi) and a range Ran(\Psi) , any two users i and i' with transactions T_i and T_i' report < |T_i|,\ \Psi(T_i) > and < |T_i'|,\ \Psi(T_i') > respectively. If the same output S ( S \in Ran(\Psi) ) is transformed through algorithm \Psi while satisfying the inequality (4), we assume that the algorithm \Psi satisfies \varepsilon -LDP:

    \dfrac{{\mathbb{P}}[\Psi(T_i) = S]}{{\mathbb{P}}[\Psi(T_i') = S] }\leqslant {\rm{e}}^\varepsilon. (4)

    Note that inevitably, some private data is revealed. For example, in the extreme case, |T_i| = 0 or |T_i| = |I| , implying that the user has no item or possesses all items in the item domain, respectively. However, |T_i|<<|I| and |T_i|>0 are the majority of cases in practice. To ensure broad applicability, a marginally relaxed interpretation of local differential privacy is necessitated. This research endeavors to safeguard the privacy of the user's transaction set |T_i| , albeit not to the extent of the stringent \varepsilon -LDP threshold. Consequently, the privacy assurance of our proposed protocols is positioned above the Item-Level \varepsilon -LDP.

    Traditional approaches[3, 9] use “padding and sampling” to protect privacy while satisfying \varepsilon -LDP, which discards most of the information in the transaction. IHFO uses the Hadamard transform to encode all items in each transaction. The vectors in the Hadamard transform matrix are orthogonal to each other and binarized; therefore, regardless of the number of items in the user transaction, they can be encoded into a vector with a fixed length. Next, adding LDP perturbations to this vector preserves the items in each transaction, thereby allowing datasets with varied transaction lengths to be handled. Finally, the aggregator collects all vectors and estimates the frequencies of items. The entire process of IHFO is divided into three phases: encoding, perturbation, and decoding. The detailed process of IHFO is shown in Algorithm 1 and the illustration of encoding and perturbation in IHFO is in Fig.3. For example, user i has a transaction with items \{b, e\} . After adding an extra item \perp , this transaction has three items (the length is equal to 3). In the first phase, three items are encoded to three column vectors from the Hadamard matrix (in order to facilitate understanding, we paint them as row vectors). Then, the user sums these vectors and selects one index k_i = 7 with its value b_i[7] = -3 . After perturbing the value b_i[7] by (5) to z_i[7] = -1 ( z_i[k_i]\in \{1,-1\} ), the perturbed tuple (3, <7, –1>) should be sent to the aggregator where < k_i, z_i[k_i] > satisfies LDP and ( |\{b,c\}\cup \{\perp\}|, < k_i,z_i[k_i] > ) satisfies the Item-Level LDP.

    Figure  3.  Illustration of encoding and perturbation in IHFO. (a) Process of IHFO. (b) Mapping Hadamard matrix.

    Algorithm 1. Integrated-with-Hadamard-Transform-Based Frequency Oracle: IHFO
    Require: the privacy budget \varepsilon ; Hadamard matrix {\boldsymbol{H}}_r where 2^r\geqslant |I|
    Ensure: the estimated frequency f(v) ;
    1: User Side:
    2: for each user u_i that has transaction T_i with a set of items, i = 1 to n do
    3:   Add an extra item \perp to T_i with a Bernoulli(0.5) distribution \rightarrow T_i' (if l_i = 0 , item \perp is necessary);
    4:   Encode T_i' to vectors through the Hadamard matrix mapping;
    5:   Sum the vectors to obtain the encoding vector \boldsymbol b_i ;
    6:   Randomly select an index k_i to obtain the value \boldsymbol b_i[k_i] ;
    7:   Obtain the perturbed column vector \boldsymbol z_i where z_i[j] = 0 if j\neq k_i ( j\in [0,2^r) ) and
           z_i[k_i]= \left\{ \begin{array}{ll} 1, & {\rm w.p.}\;\dfrac{1}{{\rm e}^\varepsilon + 1} + \dfrac{b_i[k_i] + l_i}{2l_i}\dfrac{{\rm e}^\varepsilon - 1}{{\rm e}^\varepsilon + 1}, \\ - 1, & {\rm otherwise.} \end{array} \right.
    8:   Send ( l_i,z_i ) to the aggregator where l_i = |T_i'| ;
    9: end for
    10: Aggregator side:
    11: The aggregator computes \hat{z} \leftarrow \sum_{i = 1}^n z_i\ l_i ;
    12: Obtain the frequency estimation of item v : f_v \leftarrow \dfrac{{\rm{e}}^{\varepsilon}+1}{{\rm{e}}^{\varepsilon}-1}\hat{z}\cdot {\boldsymbol{H}}(v)^{\rm{T}} ;
    13: return f(v) ;

    Encoding. First, each transaction is added an extra item \perp with a Bernoulli(0.5) distribution separately (if l_i = 0 , we also add \perp ). Then, we encode all items in the user transaction with a Hadamard matrix {\boldsymbol{H}}_r , where r = \lceil\log_2{|I\cup \{\perp\}|} \rceil . We assume that each item v,v\in I\cup \{\perp\} is uniquely mapped to a column vector {\boldsymbol{H}}(v) of a Hadamard matrix {\boldsymbol{H}} (unless otherwise stated, all symbols {\boldsymbol{H}} below represent {\boldsymbol{H}}_r ). Then, each user i can obtain the encoding vector \boldsymbol b_i = \sum_{v\in T_i} {\boldsymbol{H}}(v) , and the number of items l_i = |T_i| in the transaction ( T_i comprises old transactions and random \perp ). The special item \perp is a perturbation that renders it impossible for an attacker to confirm how many items are in each user transaction.

    Perturbation. User i randomly selects the k_i -th entry of \boldsymbol b_i and perturbs b_i[k_i] independently with the following distribution:

    z_i[k_i] = \left\{\begin{array}{ll}1, & {\rm w.p.} \ \dfrac{1}{{\rm{e}}^{\varepsilon}+1}+\dfrac{b_i[k_i]+l_i}{2l_i}\dfrac{{\rm{e}}^{\varepsilon}-1}{{\rm{e}}^{\varepsilon}+1},\\ -1, & {\rm otherwise.} \end{array}\right. (5)

    Finally, user i releases z_i[k_i] and l_i to the aggregator. As k_i is the position of vector \boldsymbol b_i , z_i[k_i] requires r+1 bits for communication, where r bits represent the position of k_i (i.e., any position j_i\ne k_i , z[j_i] = 0 ).

    Decoding. The aggregator collects all the noise information to obtain a vector of summation \boldsymbol z = \sum_i^n z_i\ l_i . Then, the unbiased frequency estimation of item v is given by

    f_v = \dfrac{{\rm{e}}^{\varepsilon}+1}{{\rm{e}}^{\varepsilon}-1}({\boldsymbol{H}}(v)^{\rm{T}} \cdot \boldsymbol z ). \label{Method:IHFO}

    Each user sends (z_i[k_i],l_i) to the aggregator, where z_i[k_i] represents the value of items in the transaction under the LDP perturbation and l_i represents the length of the transaction. We consider it feasible to sacrifice some privacy of the length to obtain more accurate FO (more specifically, the user reports the true length l_i with the probability of 50\% and reports the true length plus 1 (l_i+1) with the probability of 50\% ). Differently expressed, only z_i[k_i] in (z_i[k_i],l_i) satisfies LDP and protects the privacy of the items. We also analyze the privacy protection in Section 4.

    If an itemset is frequent, then each item in the itemset is also clearly frequent. However, given a particular dataset, we cannot intuitively determine which itemsets are frequent. Frequent itemsets cannot be obtained directly because even if the domain of items is not large (e.g., |I| = 1\;000 ), the number of possible itemsets will be |\Gamma(I)| = 2^{1\;000} , which is disastrous for the aggregator. Moreover, because most itemsets' frequencies are low, the aggregator must first reduce the domain of candidate itemsets. When the set of candidate itemsets is small, users will send more available data after the perturbation, which increases the accuracy of the frequency estimation. Thus, UISM divides users into two groups: FS , which is applied to generate candidate itemsets by finding the frequent singleton items, and FF , which is applied to update the frequent itemsets and estimate their frequencies, i.e.,

    \Theta(FS, FF)\to < FIS,f^{FIS} >.

    Step 1: Find Frequent Singleton Items. Candidate itemsets are generated based on frequent singleton items. The larger the number of accurate singleton items, the more accurate the candidate itemsets, which is beneficial for obtaining frequent itemsets together with their frequencies in step 3. Thus, the frequent singleton itemset SI with noisy frequencies f^{SI} is obtained by \Psi_{\rm{IHFO}} using

    {FS < SI,f^{SI} >: = \Psi_{\rm{IHFO}}(I,\varepsilon).}

    Step 2: Generate Candidate Itemsets[3]. Because each item in a frequent itemset is frequent, we construct a set according to the following formula:

    CS: = \left\{y|y \subseteq SI, p_y = \prod\limits_{v\in y} \dfrac{0.9f^{SI}_v}{\max_{y\in SI} f^{SI}_y} ,p_y\geqslant t\right\},

    where t is chosen such that |CS| = 2k . To reduce the calculation cost, t should take at least p_x under the k -th singleton item x .

    Step 3: Update Frequent Itemsets and Estimate Their Frequencies. After the domain CS is defined, each user in the second part of the dataset constructs their transaction (i.e., T_i' = \{x|x\subseteq T_i, x \subseteq CS\} ). The aggregator receives information from users via the IHFO algorithm. After the correction, the aggregator updates the top- k frequent itemsets and their respective frequencies as the released result.

    {FF< FIS,f^{FIS} >: = \Psi_{\rm{IHFO}}(CS,\varepsilon).}

    The UISM protocol addresses the challenge of set-valued inputs by using the Hadamard transform response FO approach to report after encoding all items in each transaction. This protocol can guarantee the accuracy of the frequency estimation.

    We find that if the top-rank frequent singleton items are more precise, the frequent itemsets with their frequencies are more accurately obtained through UISM. Moreover, the ranks of singleton items are more important than the frequencies in step 1. As described previously, in SVSM, PSFO ( l = 1) sacrifices the frequency accuracy to obtain a good rank. Thus, we adapt \Psi_{{\rm PSFO}\enspace(l = 1)} to find frequent singleton itemset SI , as shown in (6).

    FS< SI,f^{SI} >: = \Psi_{\rm{PSFO}}(I,\varepsilon,l = 1). (6)

    By combining the advantages of PSFO (l = 1) and IHFO, we propose O-UISM for FIM. Thus, O-UISM can more accurately obtain frequent itemsets and frequencies with a lower squared error. Note that we can apply any FO algorithm to replace PSFO (l = 1).

    We protect privacy from three aspects.

    1) Hadamard Encoding and Sampling. Hadamard encoding and sampling provide some privacy protection because even if the privacy budget is large, the aggregator can obtain minimal information from each sending value.

    2) Symbol \perp . We add \perp to provide some protection of |T_i| . The aggregator needs to guess the length of the transaction from |T_i| and |T_i|-1 .

    3) \varepsilon -LDP. We protect the sending value under the most rigorous standard of data privacy.

    Lemma 1.Special item \perp protects the items of transaction.

    Proof. A special item \perp represents an extra item added to each transaction from a Bernoulli(0.5) distribution. The aggregator cannot then distinguish the accuracy length of the transaction between |T_i| and |T_i|-1 obtained from user i . Item \perp , as a dummy item, is applied to defend the attraction. In particular, if |T_i| = 0 , we must add symbol \perp in this transaction.

    Lemma 2.The perturbation value z_i satisfies \varepsilon -LDP.

    Proof. We assume that two users have transactions T_1 and T_2 with lengths l_1 and l_2 , respectively. Let k be any position and RC be the ratio of two conditional probabilities with T_1 and T_2 transactions. The sensitivity is maximized when b_1 is maximum and b_2 is minimum (i.e., b_1[k] = l_1 , and b_2[k] = -l_2 ):

    \begin{split} RC_{\max} & = \max{\dfrac {P(z_1[k] = 1|b_1[k])}{P(z_2[k] = 1|b_2[k])}} \\& = \dfrac {P(z_1[k] = 1|b_1[k] = l_1)}{P(z_2[k] = 1|b_2[k] = -l_2)}\\& = \dfrac {\dfrac {{\rm{e}}^\varepsilon}{{\rm{e}}^\varepsilon+1}}{\dfrac {1}{{\rm{e}}^\varepsilon+1}}\\& = {\rm{e}}^\varepsilon. \qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad\quad\;\;\Box \end{split}

    Theorem 1. IHFO provides higher protection than Item-Level \varepsilon -LDP but also lower protection than \varepsilon -LDP.

    Proof. If a mechanism satisfies the LDP, it provides privacy protection at a high level. Lemma 2 proves that the value z_i in IHFO satisfies \varepsilon -LDP. If we send the actual length of the transaction, IHFO satisfies Item-Level \varepsilon -LDP. However, we also protect the length of transaction in Lemma 1. Thus, IHFO provides higher protection than Item-Level \varepsilon -LDP but also lower protection than rigorous \varepsilon -LDP.

    Here, we analyze the accuracy improvement of the proposed framework by evaluating unbiased estimation in Theorem 2 and the error bound in Lemma 3. Then, we compare the variance between IHFO and PSFO. Finally, we analyze the communication cost and computational complexity of PSFO and IHFO.

    Theorem 2.The estimated frequency f_v is unbiased.

    Proof. We assume that \hat{z} is the real summation of the encoding bits, where \hat{z} = \sum_i^n \boldsymbol b_i . Then, the real frequency of any item v is given by

    \hat{f_v} = \dfrac 1{2^r}{\boldsymbol{H}}(v)^{\rm{T}}\cdot\hat{z} = \dfrac 1{2^r}{\boldsymbol{H}}(v)^{\rm{T}}\cdot\sum\limits_i^n \boldsymbol b_i.

    Considering that each user i randomly selects the k -th entry of \boldsymbol b_i with the probability of 1/{2^r} , the mean value of the k -th entry in z can thus be expressed as

    \begin{split} {\mathbb{E}}(z[k]) &= \dfrac 1{2^r}{\mathbb{E}}\big(\sum\limits_i^nz_i[k]\ l_i \big)\\&= \dfrac 1{2^r}\dfrac{{\rm{e}}^{\varepsilon}-1}{{\rm{e}}^{\varepsilon}+1}\sum\limits_i^n b_i[k]. \end{split}

    Hence, the mean frequency of item v is as follows.

    \begin{split} {\mathbb{E}}(f_v) &= {\mathbb{E}}\big(\dfrac{{\rm{e}}^{\varepsilon}+1}{{\rm{e}}^{\varepsilon}-1}({\boldsymbol{H}}(v)^{\rm{T}}\cdot z ) \big)\\&= \dfrac{{\rm{e}}^{\varepsilon}+1}{{\rm{e}}^{\varepsilon}-1}\bigg(\sum\limits_{k = 0}^{2^r-1}\dfrac 1{2^r} \dfrac{{\rm{e}}^{\varepsilon}-1}{{\rm{e}}^{\varepsilon}+1}\sum\limits_i^n b_i[k]\cdot {\boldsymbol{H}}(v)^{\rm{T}}[k]\bigg)\\& = \dfrac 1{2^r}{\boldsymbol{H}}(v)^{\rm{T}}\cdot \sum\limits_i^n \boldsymbol b_i \\& = \hat{f_v}. \qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad\;\;\;\;\,\Box \end{split}

    Theorem 3 (Upper Bound). Let d = 2^r . As described previously, \hat{f_v} =(1/d){\boldsymbol{H}}(v)^{\rm{T}}\sum_{i = 1}^n\boldsymbol b_i and f_v = (({{\rm{e}}^{\varepsilon}+1})/({{\rm{e}}^{\varepsilon}-1})) {\boldsymbol{H}}(v)^{\rm{T}}\cdot\sum_{i = 1}^nz_i l_i . With the probability of at least 1-\beta ,

    {\max|f_v-\hat{f_v}| = O\left(\dfrac{\sqrt{\displaystyle\sum\limits_{i = 1}^n l_i^2\log{(1/\beta)}}}{n\varepsilon}\right).}

    Proof. In the IHFO algorithm, f_v is the unbiased counterpart of \hat{f_v} obtained by Theorem 2. Let \hat{t_i} = (1/d)({\boldsymbol{H}}(v)^{\rm{T}} \cdot b_i) and t_i = (({{\rm{e}}^{\varepsilon} + 1})/({{\rm{e}}^{\varepsilon} - 1}))({\boldsymbol{H}}(v)^{\rm{T}} \cdot z_i l_i) . Thus, |f_v-\hat{f_v}|\leqslant (({{\rm{e}}^\varepsilon+1})/({{\rm{e}}^\varepsilon-1}))\ l_i +1 . Then, by Bernstein's inequality,

    \begin{split}& {\mathbb{P}}[|f_v-\hat{f_v}|>\lambda]\\ =\ &{\mathbb{P}}\left[\left|\sum\limits_{i = 1}^n(t_i-\hat{t_i})\right|>n\lambda\right] \\ \leqslant\ & 2 \exp {\left( -\dfrac {n^2\lambda^2}{2\displaystyle\sum\limits_{i = 1}^nVar(t_i)+\dfrac 23n\lambda \Bigg(\dfrac{{\rm{e}}^\varepsilon+1}{{\rm{e}}^\varepsilon-1} l_i +1\Bigg)} \right)}. \end{split}\qquad (7)

    For each position k in z_i , if k\ne k_i , then z_i[k] = 0 ; the other case satisfies (5).

    Thus,

    \begin{split} Var(t_i) &= \dfrac{({\rm{e}}^\varepsilon+1)^2}{({\rm{e}}^\varepsilon-1)^2}l_i^2 Var( {\boldsymbol{H}}(v)^{\rm{T}}\cdot z_i)\\ & = \dfrac{({\rm{e}}^\varepsilon+1)^2}{({\rm{e}}^\varepsilon-1)^2}l_i^2-b_i^2[k_i]\\ &\leqslant \dfrac{({\rm{e}}^\varepsilon+1)^2}{({\rm{e}}^\varepsilon-1)^2}l_i^2\\& = O\left(\dfrac {l_i^2}{\varepsilon^2}\right). \end{split} (8)

    Substituting (8) and ({{\rm{e}}^\varepsilon+1})/({{\rm{e}}^\varepsilon-1})\times l_i +1 = O\left({l_i}/\varepsilon\right) into inequality (7), we obtain

    \begin{split} &{\mathbb{P}}[|f_v-\hat{f_v}|>\lambda] \\ \leqslant& 2 \exp{\left( -\ \dfrac {n^2\lambda^2}{O\left(\left(\displaystyle\sum\limits_{i = 1}^nl_i^2\right)\Bigg/\varepsilon^2\right) + n\lambda O\left({\displaystyle\sum\limits_{i = 1}^nl_i}\Bigg/{\varepsilon}\right)}\right)}. \end{split}

    By the union bound, there exists

    \qquad\qquad\qquad {\lambda = O\left({\sqrt{\displaystyle\sum\limits_{i = 1}^n l_i^2\log{(1/\beta)}}}\bigg/{n\varepsilon}\right).} \quad\quad\;\;\;\;\Box

    Variance. The variance in the frequency of item v in IHFO is denoted as

    \begin{split} Var(f_v) &= \dfrac{({\rm{e}}^{\varepsilon}+1)^2}{({\rm{e}}^{\varepsilon}-1)^2}\sum\limits_i^n l_i^2-\sum\limits_i^n (b_i[k_i])^2\\& \leqslant \dfrac{({\rm{e}}^{\varepsilon}+1)^2}{({\rm{e}}^{\varepsilon}-1)^2}\sum\limits_i^n l_i^2. \end{split}

    Theorem 4. The mean of Var_{\rm{IHFO}} is lower than that of Var_{\rm{PSFO}} when \varepsilon<\log{\sqrt{1440}}\approx 3.6 and

    \sigma>\dfrac{\sqrt{160{\rm{e}}^{2\varepsilon}+({\rm{e}}^{\varepsilon}-1)^2(40{\rm{e}}^{\varepsilon}-({\rm{e}}^{\varepsilon}+1)^2)}-4\sqrt{10}{\rm{e}}^{\varepsilon}}{40{\rm{e}}^{\varepsilon}-({\rm{e}}^{\varepsilon}+1)^2}\mu

    for a distribution {\cal{X}} with mean \mu and variance \sigma^2 .

    Proof. We should compare the variance of

    Var_{\rm{IHFO}} = ({({\rm{e}}^{\varepsilon}+1)^2})/{({\rm{e}}^{\varepsilon}-1)^2})\sum_{i = 1}^m p_i\times l_i^2

    and

    Var_{\rm{PSFO}}= {4 {\rm{e}}^\varepsilon}/({({\rm{e}}^{\varepsilon}-1)^2}) l^2

    where l: inf\sum_{i = 1}^l p_i\geqslant 0.9 (we omit the same n for the sake of calculation). According to Chebyshev's inequality (9) and 1- {\sigma^2}/{\omega^2} = 0.9 for a distribution with mean \mu and \sigma^2 ,

    P\{|{\cal{X}}-\mu|<\omega\}\geqslant 1- \dfrac{\sigma^2}{\omega^2}. (9)

    We obtain \omega \geqslant \sqrt{10}\sigma . Then l \geqslant \omega+\mu in PSFO.

    \begin{split} &\qquad {\mathbb{E}}(Var_{\rm{IHFO}}) = \dfrac{({\rm{e}}^{\varepsilon}+1)^2}{({\rm{e}}^{\varepsilon}-1)^2}(\mu^2+\sigma^2), \\&\qquad {\mathbb{E}}(Var_{\rm{PSFO}}) = \dfrac{4{\rm{e}}^{\varepsilon}}{({\rm{e}}^{\varepsilon}-1)^2}(\mu+\sqrt{10}\sigma)^2,\\& \quad \; f(\mu,\sigma,\varepsilon) \\& = {\mathbb{E}}(Var_{\rm{PSFO}})-{\mathbb{E}}(Var_{\rm{IHFO}})\\& = \dfrac {(-({\rm{e}}^{\varepsilon}-1)^2\mu^2+(40{\rm{e}}^{\varepsilon}-({\rm{e}}^{\varepsilon}-1)^2)\sigma^2+8\sqrt{10}{\rm{e}}^{\varepsilon}\mu\sigma)}{({\rm{e}}^{\varepsilon}-1)^2}. \\ & \end{split} (10)

    From (10), if \varepsilon<\log{\sqrt{1440}\approx 3.6} and

    \sigma > \dfrac{\sqrt{160{\rm{e}}^{2\varepsilon}+({\rm{e}}^{\varepsilon}-1)^2(40{\rm{e}}^{\varepsilon}-({\rm{e}}^{\varepsilon}+1)^2)}-4\sqrt{10}{\rm{e}}^{\varepsilon}}{40{\rm{e}}^{\varepsilon}-({\rm{e}}^{\varepsilon}+1)^2}\mu,

    the variance of IHFO is lower than that of PSFO.

    We use the mean of variance to compare, which may lose efficacy. Owing to padding and sampling, PSFO must transmit a value unrelated to the transaction, contributing little to calculating the item frequency. However, each value transferred through IHFO contributes part of the information on all items in this transaction. For example, in the practical setting, the privacy budget \varepsilon\in[1,2] , and the dataset confirms Poisson distribution {\cal{X}}(\lambda) . Thus, we set \varepsilon = \ln 3 for the general cases. We can obtain that the variance of IHFO is better when \lambda\leqslant 31, which can be applied in most real FIM cases.

    The Hadamard transform approach has been cited in many LDP papers[17, 18, 37] and is variously described, though always with the same core concept. Every user value is mapped to a column in the Hadamard matrix and then sends a random entry to the aggregator after combining with the LDP noise. Advantages exist on both the user side and the aggregator side: at the user side, the HTR achieves a good tradeoff between data availability and communication cost since each user transmits only r+1 bits (if a random number is generated by the aggregator, 1 bit is sufficient). The computational complexity of the aggregator in reconstructing the FO is O(n+r\times 2^r) , versus O(2^r\times n) for the OLH, since the aggregator needs to sum up only the collected information and then takes the linear product of the matrices.

    The communication costs of IHFO and PSFO are related to different parameters, but the aggregator's computational complexity is lower than that of the PSFO approach in Table 3. For PSFO, each user should transform O(\log_2 \sqrt n+\log_2 g) =O(\max\{\log \sqrt n, \log g\}) bits using the OLH algorithm. Then, the computational complexity of the aggregator is \varOmega (n^{\tfrac 32}) ( \sqrt n hash functions). For IHFO, each user should send three parameters: the location of the sampling, positive and negative values, and transaction length. Thus, the communication cost is O(\lceil \log_2{|I|} \rceil+\log_2 l_i+ 1 ) = O(\log{|I|}) . After summing all vectors, the aggregator needs only to calculate the inner matrix product once to obtain the frequency estimation; thus, the computational complexity is \varOmega(n) .

    Table  3.  Communication and Computation Cost
    FO Method Communication
    (Each User)
    Computation (Aggregator)
    PSFO O(\max\{\log \sqrt n, \log g\}) \varOmega (n^{3/2})
    IHFO O(\log{|I|}) \varOmega(n)
    Note: Communication cost denotes the bits in the collection that each user should send to the aggregator; computation cost denotes the computation complexity of the aggregator in correction.
    下载: 导出CSV 
    | 显示表格

    As O-UISM applies IHFO to estimate frequent itemsets with their frequencies, it also achieves the goal of unbiased estimating with lower variance and reducing the computation complexity.

    In this section, we report experiments on real datasets to validate our analysis with different protocols.

    Environment. All algorithms were implemented in Python 3.7.3, and all experiments were conducted on an Intel CoreTM i7-6700 3.40 GHz PC with 16 GB RAM. We report the average results over 10 runs.

    Datasets. We ran experiments on the following datasets.

    Online 1. This dataset contains the merchant transactions of 500000 users with 2603 categories.

    IBM 2. This dataset was generated by the IBM Synthetic Data Generation Code for Associations and Sequential Patterns with 1.8 million transactions generated over 1000 categories. The average transaction length is 5.

    Metrics. This study aims to find frequent itemsets together with their frequencies, which requires different metrics to evaluate their utilities. We adopt the normalized cumulative rank (NCR)[3], squared error (SE) and Kullback-Leibler divergence (KLD) to assess the frequent itemsets and frequencies, respectively. NCR is used to measure the accuracy ranks of the frequent items/itemsets, and SE and KLD are used to measure the accuracy frequencies of the frequent items.

    1) Normalized Cumulative Rank (NCR). The quality function with the most k values ranked is as follows: the highest ranked value has a score of k , the next one has a score of k-1 , and so on. The k -th value has a score of 1, and all the other values have scores of 0. To normalize this into a value between 0 and 1, we divide the sum of scores by the maximum possible score (i.e., {k(k+1)}/2 ).

    2) Squared Error (SE). We measure the estimation accuracy using the squared error. That is,

    Var = \dfrac{1}{|FIS\cap RS|}\sum\limits_{y\in FIS\cap RS}(f^{FIS}_y-f^{RS}_y)^2,

    where RS is the real set of frequent itemsets with frequencies f^{RS} . Note that we account only for heavy hitters that are successfully identified by the protocol (i.e., y\in FIS\cap RS ). Thus, lower variance means more accurate estimation.

    3) Kullback-Leibler Divergence (KLD). Csisz \acute{a} rs f-divergence is used to measure whether a privacy mechanism is an information-theoretic quantity, and the distributional difference is defined in (11)[39].

    D_f\big(R_{{\rm{est}}}||R_{{\rm{real}}}\big) = \int f\bigg(\dfrac{{\rm d}R_{{\rm{real}}}}{{\rm d}R_{{\rm{est}}}}\bigg){\rm d}R_{{\rm{est}}}, (11)

    where f(x) = x \log{x} . To express divergence more accurately, this paper detects data availability by (12).

    KLD = \dfrac 12 \Big(D_f\big(R_{{\rm{est}}}||R_{{\rm{real}}}\big)+D_f\big(R_{{\rm{real}}}||R_{{\rm{est}}}\big)\Big). (12)

    Parameter Settings. As described in [3], SVSM uses 50\% of all users to find frequent singleton items, 10\% to report the size of their itemsets that intersect with CS , and 40\% to update the frequent itemsets and their frequencies. Thus, we divide users into two parts, where \eta is the percentage assigned in the first part. For fairness, we set \eta = 0.5 , which implies that half of the users report frequent singleton items, while the other half update the estimated frequencies of the candidate frequent itemsets.

    Selected Approaches for Evaluation of FO. The approaches used in the evaluation are as follows.

    Real-IHFO. The first phase returns the real FO, and the second and third phases use IHFO.

    Real-PSFO. The first phase returns the real FO, and the second and third phases use PSFO.

    Sample-IHFO. The first phase uses sampling (IFHO without perturbation), and the second and third phases use IHFO.

    Sample-Sampling. Both the first and second phases use sampling to obtain FO. It denotes the upper bound of UISM.

    Selected Approaches for Evaluation of FIM. The approaches used in the evaluation are as follows.

    UIIM and UISM. Both the first and second phases use IHFO to obtain FO.

    SVIM and SVSM[3]. The first phase uses PSFO (OLH, l = 1), and the second and third phases use PSFO (OLH, l = adaptive L ) to obtain FO.

    O-UIIM and O-UISM. It is an optimal version of UISM. The first phase uses PSFO (OLH, l = 1), and the second phase uses IHFO to obtain FO.

    PrivSet. The first phase uses PSFO (OLH, l = 1), and the second and third phases use Privset to obtain FO.

    In this subsection, we mainly compare the performance for frequency oracle.

    First, we compare Real-IHFO and Real-PSFO. In Fig.4(a) and Fig.4(c), the NCR trendlines of Real-IHFO show a stable performance, with a small increase as the privacy budget increases in Online and IBM. The NCR trendlines of Real-PSFO increase rapidly as the privacy budget increases and the results are better than those of Real-IHFO when the privacy budgets are large, but the exceeded point of IBM is larger than that of Online. A higher score indicates better identification. In Fig.4(b) and Fig.4(d), the trendlines of Real-IHFO are lower than those of Real-PSFO. A lower score denotes better frequency estimation. These indicate that: 1) as privacy budgets grow, the error caused by sampling in the IHFO algorithm is greater than that caused by perturbation; 2) the increase of the data size has a greater effect on IHFO than on PSFO, since the amount of the IBM data is greater than the amount of the online data; and 3) IHFO has a lower variance of frequency estimation.

    Figure  4.  Performance under different protocols. (a) Online dataset, NCR (normalized cumulative rank), k = 64. (b) Online dataset, squared error, k = 64. (c) IBM dataset, NCR, k = 64. (d) IBM dataset, squared error, k = 64.

    Second, we compare Real-IHFO and Sample-IHFO. The SE trendlines of them are close in Fig.4(b) and Fig.4(d). However, the NCR trendlines of Real-IHFO are much higher than those of Sample-IHFO in Fig.4(a) and Fig.4(c). These indicate that it is more useful in the first stage to obtain an accurate frequency ranking than a frequency estimation. Therefore, choosing PSFO ( l = 1) has advantages in O-UISM. In addition, the NCR of Real-IHFO and Sample-Sample shows that even if the perturbed collection is adopted in the second stage, the result reflects an advantage when the privacy budget is slightly larger.

    Third, we compare Sample-IHFO and Sample-Sample. Fig.4(b) and Fig.4(d) show that the trendlines of Sample-Sample are far lower than those of all the other protocols regarding the squared error. They are also the upper bound errors that can be attained by UISM. Fig.4(a) shows the NCR on the Online dataset, Sample-Sample is always higher than that of Sample-IHFO. On the other hand, in Fig.4(c), the NCR on the IBM dataset is similar when the privacy budget is greater than 2. These indicate that the increase in the data size significantly affects IHFO.

    In summary, this experiment shows that IHFO has a lower variance of frequency estimation and the dataset size has an obvious effect on IHFO.

    In this subsection, we mainly compare the performance between our approaches and the state-of-the-art protocol for frequent item/itemset mining.

    Singleton Frequent Items. We first identify the frequent singleton items in Fig.5 and Fig.6. First, Fig.5(a) and Fig.6(a) demonstrate the trendlines of NCR scores of all protocols when k = 64. The trendlines of O-UIIM are the highest in most privacy budgets \varepsilon , and the score of SVIM is the highest at the last \varepsilon . A higher score indicates better identification. Figs.5(d) and 6(d) demonstrate the trendlines of NCR scores of all protocols when k = 32. The trendlines of O-UIIM are the highest in most privacy budgets \varepsilon , and the scores of PrivSet are the highest at the last two \varepsilon . These results suggest that O-UISM performs the best in these metrics. Second, Figs.5(b), 5(e), 6(b), and 6(e) and Figs.5(c), 5(f), 6(c), and 6(f) demonstrate the trendlines of SE and KLD, respectively. The trendlines of UIIM and O-UIIM are lower in most privacy budgets \varepsilon when k = 64 and k = 32. The reason for SVIM performing the best in several big privacy budgets is that the error caused by sampling in the IHFO algorithm is greater than that caused by perturbation. We combine SE and KLD for analysis because the frequency (SE) and distribution (KLD) of data are essential data availability standards. A lower score means a more accurate estimation. These results indicate that the UIIM and O-UIIM estimated data are more accurate. Third, NCR and SE are reviewed together in Figs.5(a) and 6(a) and Figs.5(b) and 6(b), respectively. A higher NCR and a lower SE imply that more items are identified with an estimation that is more accurate. Even though the SE score of UIIM is lower than that of O-UIIM, O-UIIM performs better because its NCR trendlines are notably higher than those of UIIM.

    Figure  5.  Singleton identification for the IBM dataset. (a) IBM dataset, NCR (normalized cumulative rank), k = 64. (b) IBM dataset, SE (squared error), k = 64. (c) IBM dataset, KLD (Kullback-Leibler divergence), k = 64. (d) IBM dataset, NCR, k = 32. (e) IBM dataset, SE, k = 32. (f) IBM dataset, KLD, k = 32.
    Figure  6.  Singleton identification for the Online dataset. (a) Online dataset, NCR, k = 64. (b) Online dataset, SE, k = 64. (c) Online dataset, KLD, k = 64. (d) Online dataset, NCR, k = 32. (e) Online dataset, SE, k = 32. (f) Online dataset, KLD, k = 32.

    Itemset Mining. We evaluate the effectiveness of our protocols on frequent itemset mining. The results are similar to the results of frequent singleton items. First, Figs.7(a) and 8(a) and Figs.7(d) and 8(d) demonstrate the trendlines of NCR scores of all protocols when k = 64 and k = 32 , respectively. The trendlines of O-UISM are the highest in most privacy budgets \varepsilon . A higher score indicates better identification. These suggest that O-UISM performs the best in these metrics. Second, Figs.7(b), 7(e), 8(b), and 8(e) and Figs.7(c), 7(f), 8(c), and 8(f) demonstrate the trendlines of SE and KLD, respectively. The trendlines of UIIM and O-UIIM are lower in most privacy budgets \varepsilon .

    Figure  7.  Itemset mining results for the IBM dataset. (a) IBM dataset, NCR, k = 64. (b) IBM dataset, SE, k = 64. (c) IBM dataset, KLD, k = 64. (d) IBM dataset, NCR, k = 32. (e) IBM dataset, SE, k = 32. (f) IBM dataset, KLD, k = 32.
    Figure  8.  Itemset mining results for the Online dataset. (a) Online dataset, NCR, k = 64. (b) Online dataset, SE, k = 64. (c) Online dataset, KLD, k = 64. (d) Online dataset, NCR, k = 32. (e) Online dataset, SE, k = 32. (f) Online dataset, KLD, k = 32.

    We can also see that the trendlines of O-UIIM start and rise quickly and level off since \varepsilon>1.75 . This indicates that O-UISM performs the best when \varepsilon is not large, and O-UISM cannot offset the error caused by the Hadamard encoding and sampling with increasing privacy budget. Moreover, UISM does not perform as well as O-UISM in NCR under different privacy budgets but performs well in SE and KLD. This indicates that the IHFO algorithm is good at updating frequent itemsets and their corresponding frequencies.

    In summary, this experiment shows that O-UISM performs the best in FIM.

    Different \eta . Fig.9 shows the impact of the dataset division \eta which means the aggregator applies the percent to obtain the rank of the items in the first part of the dataset. If \eta is small, fewer data in the first part are applied to identify frequent singleton items. If the amount of data in the second part is less, the squared error is higher under the same privacy budget. Trendlines of NCR are interwoven to different \eta . We consider \eta = 0.5 to be the appropriate choice.

    Figure  9.  (a) NCR and (b) SE of varying dataset division \eta under different \varepsilon .

    In this study, we investigated LDP (local differential privacy) approaches for FIM (frequent itemset mining) when each user transaction has a set of items whose length is varied. One natural LDP algorithm (e.g., SVSM) is for each user to sample only one item and use a known algorithm to send, which should balance the bias and variance. We sacrificed non-sensitivity privacy to develop an unbiased approach for FIM while protecting the privacy of items in each transaction. In most cases, our approach results in the lowest variance, ensuring the most accurate outcome. We theoretically and experimentally demonstrated that the proposed FIM approach, O-UISM, significantly outperforms the extant approaches in finding frequent itemsets and estimating their frequencies under the same privacy guarantee.

  • Figure  1.   Innovations: sacrificing non-sensitive privacy for more accurate results.

    Figure  2.   Illustration of UISM.

    Figure  3.   Illustration of encoding and perturbation in IHFO. (a) Process of IHFO. (b) Mapping Hadamard matrix.

    Figure  4.   Performance under different protocols. (a) Online dataset, NCR (normalized cumulative rank), k = 64. (b) Online dataset, squared error, k = 64. (c) IBM dataset, NCR, k = 64. (d) IBM dataset, squared error, k = 64.

    Figure  5.   Singleton identification for the IBM dataset. (a) IBM dataset, NCR (normalized cumulative rank), k = 64. (b) IBM dataset, SE (squared error), k = 64. (c) IBM dataset, KLD (Kullback-Leibler divergence), k = 64. (d) IBM dataset, NCR, k = 32. (e) IBM dataset, SE, k = 32. (f) IBM dataset, KLD, k = 32.

    Figure  6.   Singleton identification for the Online dataset. (a) Online dataset, NCR, k = 64. (b) Online dataset, SE, k = 64. (c) Online dataset, KLD, k = 64. (d) Online dataset, NCR, k = 32. (e) Online dataset, SE, k = 32. (f) Online dataset, KLD, k = 32.

    Figure  7.   Itemset mining results for the IBM dataset. (a) IBM dataset, NCR, k = 64. (b) IBM dataset, SE, k = 64. (c) IBM dataset, KLD, k = 64. (d) IBM dataset, NCR, k = 32. (e) IBM dataset, SE, k = 32. (f) IBM dataset, KLD, k = 32.

    Figure  8.   Itemset mining results for the Online dataset. (a) Online dataset, NCR, k = 64. (b) Online dataset, SE, k = 64. (c) Online dataset, KLD, k = 64. (d) Online dataset, NCR, k = 32. (e) Online dataset, SE, k = 32. (f) Online dataset, KLD, k = 32.

    Figure  9.   (a) NCR and (b) SE of varying dataset division \eta under different \varepsilon .

    Table  1   Comparative Overview of FIM Solutions

    FIM Method Identification Accuracy
    SVSM <PSFO (l = 1),
    PSFO (l = adapl)>[3]
    General General
    PrivSet<PSFO (l = 1), PrivSet> General General
    UISM<IHFO, IHFO> General Good
    O-UISM<PSFO (l = 1), IHFO> Good Good
    Note: Identification refers to the accuracy rankings of the frequent items/itemsets, and accuracy denotes the accuracy frequencies of the frequent items/itemsets.
    下载: 导出CSV

    Table  2   Notations

    Notation Definition
    \varepsilon Privacy budget
    I Domain of items
    \Psi Perturbation algorithm
    n Number of users
    v,x Singleton item
    y Itemset
    \boldsymbol b_i Encoding vector of user i
    T_i Transaction of user i
    {\boldsymbol{H}}_r Hadamard matrices of order 2^r for
    every non-negative integer r
    {\boldsymbol{H}}(v) Column vector that item v maps to {\boldsymbol{H}}_r
    SI Set of frequent singleton items
    CS Set of candidate frequent itemsets
    FIS Set of frequent itemsets
    w.p. With probability
    下载: 导出CSV
    Algorithm 1. Integrated-with-Hadamard-Transform-Based Frequency Oracle: IHFO
    Require: the privacy budget \varepsilon ; Hadamard matrix {\boldsymbol{H}}_r where 2^r\geqslant |I|
    Ensure: the estimated frequency f(v) ;
    1: User Side:
    2: for each user u_i that has transaction T_i with a set of items, i = 1 to n do
    3:   Add an extra item \perp to T_i with a Bernoulli(0.5) distribution \rightarrow T_i' (if l_i = 0 , item \perp is necessary);
    4:   Encode T_i' to vectors through the Hadamard matrix mapping;
    5:   Sum the vectors to obtain the encoding vector \boldsymbol b_i ;
    6:   Randomly select an index k_i to obtain the value \boldsymbol b_i[k_i] ;
    7:   Obtain the perturbed column vector \boldsymbol z_i where z_i[j] = 0 if j\neq k_i ( j\in [0,2^r) ) and
           z_i[k_i]= \left\{ \begin{array}{ll} 1, & {\rm w.p.}\;\dfrac{1}{{\rm e}^\varepsilon + 1} + \dfrac{b_i[k_i] + l_i}{2l_i}\dfrac{{\rm e}^\varepsilon - 1}{{\rm e}^\varepsilon + 1}, \\ - 1, & {\rm otherwise.} \end{array} \right.
    8:   Send ( l_i,z_i ) to the aggregator where l_i = |T_i'| ;
    9: end for
    10: Aggregator side:
    11: The aggregator computes \hat{z} \leftarrow \sum_{i = 1}^n z_i\ l_i ;
    12: Obtain the frequency estimation of item v : f_v \leftarrow \dfrac{{\rm{e}}^{\varepsilon}+1}{{\rm{e}}^{\varepsilon}-1}\hat{z}\cdot {\boldsymbol{H}}(v)^{\rm{T}} ;
    13: return f(v) ;
    下载: 导出CSV

    Table  3   Communication and Computation Cost

    FO Method Communication
    (Each User)
    Computation (Aggregator)
    PSFO O(\max\{\log \sqrt n, \log g\}) \varOmega (n^{3/2})
    IHFO O(\log{|I|}) \varOmega(n)
    Note: Communication cost denotes the bits in the collection that each user should send to the aggregator; computation cost denotes the computation complexity of the aggregator in correction.
    下载: 导出CSV
  • [1]

    Li N H, Qardaji W, Su D, Cao J N. PrivBasis: Frequent itemset mining with differential privacy. Proceedings of the VLDB Endowment, 2012, 5(11): 1340–1351. DOI: 10.14778/2350229.2350251.

    [2]

    Xiong X Y, Chen F, Huang P Z, Tian M M, Hu X F, Chen B D, Qin J. Frequent itemsets mining with differential privacy over large-scale data. IEEE Access, 2018, 6: 28877–28889. DOI: 10.1109/ACCESS.2018.2839752.

    [3]

    Wang T H, Li N H, Jha S. Locally differentially private frequent itemset mining. In Proc. the 2018 IEEE Symposium on Security and Privacy (SP), May 2018, pp.127–143. DOI: 10.1109/SP.2018.00035.

    [4]

    Agrawal R, Srikant R. Fast algorithms for mining association rules. In Proc. the 20th International Conference on Very Large Data Bases, Sept. 1994, pp.487–499. DOI: 10.5555/645920.672836.

    [5]

    Adar E, Weld D S, Bershad B N, Gribble S S. Why we search: Visualizing and predicting user behavior. In Proc. the 16th International Conference on World Wide Web, May 2007, pp.161–170. DOI: 10.1145/1242572.1242595.

    [6]

    Brin S, Motwani R, Silverstein C. Beyond market baskets: Generalizing association rules to correlations. In Proc. the 1997 ACM SIGMOD International Conference on Management of Data, Jun. 1997, pp.265–276. DOI: 10.1145/253260.253327.

    [7]

    Dwork C. Differential privacy: A survey of results. In Proc. the 5th International Conference on Theory and Applications of Models of Computation, Apr. 2008. DOI: 10.1007/978-3-540-79228-4_1.

    [8]

    Dwork C, McSherry F, Nissim K, Smith A. Calibrating noise to sensitivity in private data analysis. In Proc. the 3rd Theory of Cryptography Conference, Mar. 2006, pp.265–284. DOI: 10.1007/11681878_14.

    [9]

    Wang S W, Huang L S, Nie Y W, Wang P Z, Xu H L, Yang W. PrivSet: Set-valued data analyses with locale differential privacy. In Proc. the 2018 IEEE Conference on Computer Communications, Apr. 2018, pp.1088–1096. DOI: 10.1109/INFOCOM.2018.8486234.

    [10]

    Su S, Xu S Z, Cheng X, Li Z Y, Yang F C. Differentially private frequent itemset mining via transaction splitting. IEEE Trans. Knowledge and Data Engineering, 2015, 27(7): 1875–1891. DOI: 10.1109/TKDE.2015.2399310.

    [11]

    Qin Z, Yang Y, Yu T, Khalil I, Xiao X K, Ren K. Heavy hitter estimation over set-valued data with local differential privacy. In Proc. the 2016 ACM SIGSAC Conference on Computer and Communications Security, Oct. 2016, pp.192–203. DOI: 10.1145/2976749.2978409.

    [12]

    Cormode G, Jha S, Kulkarni T, Li N H, Srivastava D, Wang T H. Privacy at scale: Local differential privacy in practice. In Proc. the 2018 International Conference on Management of Data, May 2018, pp.1655–1658. DOI: 10.1145/3183713.3197390.

    [13]

    Zeng C, Naughton J F, Cai J Y. On differentially private frequent itemset mining. Proceedings of the VLDB Endowment, 2012, 6(1): 25–36. DOI: 10.14778/2428536.2428 539.

    [14]

    Vadhan S. The complexity of differential privacy. In Tutorials on the Foundations of Cryptography, Lindell Y (ed.), Springer, 2017, pp.347–450. DOI: 10.1007/978-3-319-57048-8_7.

    [15]

    Li N H, Lyu M, Su D, Yang W N. Differential privacy: From theory to practice. Synthesis Lectures on Information Security, Privacy, & Trust, 2016, 8(4): 1–138. DOI: 10.1007/978-3-031-02350-7.

    [16]

    Papernot N, Song S, Mironov I, Raghunathan A, Talwar K, Erlingsson Ú. Scalable private learning with PATE. arXiv: 1802.08908, 2018. https://arxiv.org/abs/1802.08908, Dec. 2023.

    [17]

    Kulkarni T, Cormode G, Srivastava D. Marginal release under local differential privacy. arXiv: 1711.02952, 2017. https://arxiv.org/abs/1711.02952, Dec. 2023.

    [18]

    Nguyên T T, Xiao X K, Yang Y, Hui S C, Shin H, Shin J. Collecting and analyzing data from smart device users with local differential privacy. arXiv: 1606.05053, 2016. https://arxiv.org/abs/1606.05053, Dec. 2023.

    [19]

    Wang T H, Blocki J, Li N H, Jha S. Locally differentially private protocols for frequency estimation. In Proc. the 26th USENIX Conference on Security Symposium, Aug. 2017, pp.729–745. DOI: 10.5555/3241189.3241247.

    [20]

    Bassily R, Smith A. Local, private, efficient protocols for succinct histograms. In Proc. the 47th Annual ACM Symposium on Theory of Computing, Jun. 2015, pp.127–135. DOI: 10.1145/2746539.2746632.

    [21]

    Wang T H, Li N H, Jha S. Locally differentially private heavy hitter identification. IEEE Trans. Dependable and Secure Computing, 2021, 18(2): 982–993. DOI: 10.1109/TDSC.2019.2927695.

    [22]

    Duchi J C, Wainwright M J, Jordan M I. Local privacy and minimax bounds: Sharp rates for probability estimation. In Proc. the 26th International Conference on Neural Information Processing Systems, Dec. 2013, pp.1529–1537. DOI: 10.5555/2999611.2999782.

    [23]

    Ye Q Q, Hu H B, Meng X F, Zheng H D. PrivKV: Key-value data collection with local differential privacy. In Proc. the 2019 IEEE Symposium on Security and Privacy (SP), May 2019, pp.317–331. DOI: 10.1109/SP.2019.00018.

    [24]

    Gu X L, Li M, Cheng Y Q, Xiong L, Cao Y. PCKV: Locally differentially private correlated key-value data collection with optimized utility. In Proc. the 29th USENIX Conference on Security Symposium, Aug. 2020, pp.967–984. DOI: 10.5555/3489212.3489267.

    [25]

    Erlingsson Ú, Feldman V, Mironov I, Raghunathan A, Talwar K, Thakurta A. Amplification by shuffling: From local to central differential privacy via anonymity. In Proc. the 30th Annual ACM-SIAM Symposium on Discrete Algorithms, Jan. 2019, pp.2468–2479. DOI: 10.1137/1.9781611975482.151.

    [26]

    Duchi J C, Jordan M I, Wainwright M J. Minimax optimal procedures for locally private estimation. Journal of the American Statistical Association, 2018, 113(521): 182–201. DOI: 10.1080/01621459.2017.1389735.

    [27]

    Wang N, Xiao X K, Yang Y, Zhao J, Hui S C, Shin H, Shin J, Yu G. Collecting and analyzing multidimensional data with local differential privacy. In Proc. the 35th International Conference on Data Engineering (ICDE), Apr. 2019, pp.638–649. DOI: 10.1109/ICDE.2019.00063.

    [28]

    Liu R X, Cao Y, Chen H, Guo R Y, Yoshikawa M. FLAME: Differentially private federated learning in the shuffle model. In Proc. the AAAI Conference on Artificial Intelligence, May 2021, pp.8688–8696. DOI: 10.1609/aaai.v35i10.17053.

    [29]

    Chen R, Li H R, Qin A K, Kasiviswanathan S P, Jin H. Private spatial data aggregation in the local setting. In Proc. the 32nd IEEE International Conference on Data Engineering (ICDE), May 2016, pp.289–300. DOI: 10.1109/ICDE.2016.7498248.

    [30]

    Gursoy M E, Tamersoy A, Truex S, Wei W Q, Liu L. Secure and utility-aware data collection with condensed local differential privacy. IEEE Trans. Dependable and Secure Computing, 2021, 18(5): 2365–2378. DOI: 10.1109/TDSC.2019.2949041.

    [31]

    Murakami T, Kawamoto Y. Utility-optimized local differential privacy mechanisms for distribution estimation. In Proc. the 28th USENIX Conference on Security Symposium, Aug. 2019, pp.1877–1894. DOI: 10.5555/3361338.3361468.

    [32]

    Gu X L, Li M, Xiong L, Cao Y. Providing input-discriminative protection for local differential privacy. In Proc. the 36th IEEE International Conference on Data Engineering (ICDE), Apr. 2020, pp.505–516. DOI: 10.1109/ICDE48307.2020.00050.

    [33]

    Han X, Wang M, Zhang X J, Meng X F. Differentially private top-k query over mapreduce. In Proc. the 4th International Workshop on Cloud Data Management, Oct. 2012, pp.25–32. DOI: 10.1145/2390021.2390027.

    [34]

    Evfimievski A, Gehrke J, Srikant R. Limiting privacy breaches in privacy preserving data mining. In Proc. the 22nd ACM SIGMOD-SIGACT-SIGART Symposium on Principles of Database Systems, Jun. 2003, pp.211–222. DOI: 10.1145/773153.773174.

    [35]

    Bassily R, Nissim K, Stemmer U, Thakurta A. Practical locally private heavy hitters. In Proc. the 31st International Conference on Neural Information Processing Systems, Dec. 2017, pp.2288–2296. DOI: 10.5555/3294771.3294989.

    [36]

    Bun M, Nelson J, Stemmer U. Heavy hitters and the structure of local privacy. In Proc. the 35th ACM SIGMOD-SIGACT-SIGAI Symposium on Principles of Database Systems, May 2018, pp.435–447. DOI: 10.1145/3196959.3196981.

    [37]

    Acharya J, Sun Z T, Zhang H. Hadamard response: Estimating distributions privately, efficiently, and with little communication. arXiv: 1802.04705, 2018. https://arxiv.org/abs/1802.04705, Dec. 2023.

    [38]

    Liu Y H, Suresh A T, Yu F, Kumar S, Riley M. Learning discrete distributions: User vs item-level privacy. arXiv: 2007.13660, 2020. https://arxiv.org/abs/2007.13660, Dec. 2023.

    [39]

    Kairouz P, Oh S, Viswanath P. Extremal mechanisms for local differential privacy. In Proc. the 27th International Conference on Neural Information Processing Systems, Dec. 2014, pp.2879–2887. DOI: 10.5555/2969033.2969148.

  • 期刊类型引用(3)

    1. Soner Aydin, Sinan Yıldırım. Bayesian Frequency Estimation under Local Differential Privacy with an Adaptive Randomized Response Mechanism. ACM Transactions on Knowledge Discovery from Data, 2025, 19(2): 1. 必应学术
    2. Guanghui Feng, Guojun Wang, Tao Peng, et al. Toward Answering Federated Spatial Range Queries Under Local Differential Privacy. International Journal of Intelligent Systems, 2024, 2024(1) 必应学术
    3. Dan Zhao, Yang You, Chuanwen Luo, et al. KSKV: Key-Strategy for Key-Value Data Collection with Local Differential Privacy. Computer Modeling in Engineering & Sciences, 2024, 139(3): 3063. 必应学术

    其他类型引用(0)

图(9)  /  表(4)
计量
  • 文章访问数:  209
  • HTML全文浏览量:  16
  • PDF下载量:  23
  • 被引次数: 3
出版历程
  • 收稿日期:  2021-01-30
  • 录用日期:  2023-02-20
  • 网络出版日期:  2023-06-20
  • 刊出日期:  2023-11-14

目录

/

返回文章
返回