May 2024
·
4 Reads
·
3 Citations
This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.
May 2024
·
4 Reads
·
3 Citations
November 2023
·
14 Reads
·
3 Citations
September 2022
·
58 Reads
·
3 Citations
Lecture Notes in Computer Science
A major component of the entire digital identity ecosystem are verifiable credentials. However, for users to have complete control and privacy of their digital credentials, they need to be able to store and manage these credentials and associated cryptographic key material on their devices. This approach has severe usability challenges including portability across devises. A more practical solution is for the users to trust a more reliable and available service to manage credentials on their behalf, such as in the case of Single Sign-On (SSO) systems and identity hubs. But the obvious downside of this design is the immense trust that the users need to place on these service providers.In this work, we introduce and formalize a credential transparency system (CTS) framework that adds strong transparency guarantees to a credential management system while preserving privacy and usability features of the system. CTS ensures that if a service provider presents any credential to an honest verifier on behalf of a user, and the user’s device tries to audit all the shows presented on the user’s behalf, the service provider will not be able to drop or modify any show information without getting caught. We define CTS to be a general framework that is compatible with a wide range of credential management systems including SSO and anonymous credential systems. We also provide a CTS instantiation and prove its security formally.KeywordsCredential transparencySSOanonymous credentialszero-knowledge setsaccumulatorszero-knowledge proofs
July 2022
·
130 Reads
With the wide availability of large pre-trained language model checkpoints, such as GPT-2 and BERT, the recent trend has been to fine-tune them on a downstream task to achieve the state-of-the-art performance with a small computation overhead. One natural example is the Smart Reply application where a pre-trained model is fine-tuned for suggesting a number of responses given a query message. In this work, we set out to investigate potential information leakage vulnerabilities in a typical Smart Reply pipeline and show that it is possible for an adversary, having black-box or gray-box access to a Smart Reply model, to extract sensitive user information present in the training data. We further analyse the privacy impact of specific components, e.g. the decoding strategy, pertained to this application through our attack settings. We explore potential mitigation strategies and demonstrate how differential privacy can be a strong defense mechanism to such data extraction attacks.
May 2022
·
10 Reads
·
80 Citations
November 2021
·
26 Reads
·
10 Citations
Lecture Notes in Computer Science
Recent new constructions of rate-1 OT [Döttling, Garg, Ishai, Malavolta, Mour, and Ostrovsky, CRYPTO 2019] have brought this primitive under the spotlight and the techniques have led to new feasibility results for private-information retrieval, and homomorphic encryption for branching programs. The receiver communication of this construction consists of a quadratic (in the sender’s input size) number of group elements for a single instance of rate-1 OT. Recently [Garg, Hajiabadi, Ostrovsky, TCC 2020] improved the receiver communication to a linear number of group elements for a single string-OT. However, most applications of rate-1 OT require executing it multiple times, resulting in large communication costs for the receiver.
June 2021
·
28 Reads
·
1 Citation
In the text processing context, most ML models are built on word embeddings. These embeddings are themselves trained on some datasets, potentially containing sensitive data. In some cases this training is done independently, in other cases, it occurs as part of training a larger, task-specific model. In either case, it is of interest to consider membership inference attacks based on the embedding layer as a way of understanding sensitive information leakage. But, somewhat surprisingly, membership inference attacks on word embeddings and their effect in other natural language processing (NLP) tasks that use these embeddings, have remained relatively unexplored. In this work, we show that word embeddings are vulnerable to black-box membership inference attacks under realistic assumptions. Furthermore, we show that this leakage persists through two other major NLP applications: classification and text-generation, even when the embedding layer is not exposed to the attacker. We show that our MI attack achieves high attack accuracy against a classifier model and an LSTM-based language model. Indeed, our attack is a cheaper membership inference attack on text-generative models, which does not require the knowledge of the target model or any expensive training of text-generative models as shadow models.
January 2021
·
49 Reads
Property inference attacks consider an adversary who has access to the trained model and tries to extract some global statistics of the training data. In this work, we study property inference in scenarios where the adversary can maliciously control part of the training data (poisoning data) with the goal of increasing the leakage. Previous work on poisoning attacks focused on trying to decrease the accuracy of models either on the whole population or on specific sub-populations or instances. Here, for the first time, we study poisoning attacks where the goal of the adversary is to increase the information leakage of the model. Our findings suggest that poisoning attacks can boost the information leakage significantly and should be considered as a stronger threat model in sensitive applications where some of the data sources may be malicious. We describe our \emph{property inference poisoning attack} that allows the adversary to learn the prevalence in the training data of any property it chooses. We theoretically prove that our attack can always succeed as long as the learning algorithm used has good generalization properties. We then verify the effectiveness of our attack by experimentally evaluating it on two datasets: a Census dataset and the Enron email dataset. We were able to achieve above attack accuracy with poisoning in all of our experiments.
January 2021
·
75 Reads
·
283 Citations
We met as a group during the Homomorphic Encryption Standardization Workshop on July 13–14, 2017, hosted at Microsoft Research in Redmond, and again during the second workshop on March 15–16, 2018 in MIT. Researchers from around the world represented government, industry, and academia. There are several research groups around the world who have made libraries for general-purpose homomorphic encryption available for applications and general-purpose use. Some examples include [40–46,47]. Most general-purpose libraries for homomorphic encryption implement schemes that are based on the ring learning-with-error (RLWE) problem, and many of them displayed common choices for the underlying rings, error distributions, and other parameters.
December 2020
·
51 Reads
·
31 Citations
Lecture Notes in Computer Science
Generating additive secret shares of a shuffled dataset - such that neither party knows the order in which it is permuted - is a fundamental building block in many protocols, such as secure collaborative filtering, oblivious sorting, and secure function evaluation on set intersection. Traditional approaches to this problem either involve expensive public-key based crypto or using symmetric crypto on permutation networks. While public-key-based solutions are bandwidth efficient, they are computation-heavy. On the other hand, constructions based on permutation networks are communication-bound, especially when the dataset contains large elements, for e.g., feature vectors in an ML context.
... Their implementation utilized proxy re-encryption [11], structure-preserving signatures [59], and zero-knowledge proofs of knowledge [11]. To provide transparency of a cloud wallet's actions, Chase et al. [33] proposed a credential transparency system that allowed an identity owner to audit the actions of their cloud wallet. ...
Reference:
SoK: Trusting Self-Sovereign Identity
September 2022
Lecture Notes in Computer Science
... Hypothesis Tests in Machine Learning Security. Statistical hypothesis testing has been recently introduced for issues in machine learning security such as formalizing membership inference attacks via likelihood ratio tests [9,41], auditing dif-ferentially private algorithms [23,30], property inference [29], attribute inference [17,54], or to proposing statistical verification methods for neural networks [6]. In this work, we also phrase the problem of provenance for fine-tuned LLMs as multiple statistical hypothesis testing, but our considered problem formulation is considerably different from those considered in prior works. ...
May 2022
... These parameters directly influence the performance, security, and practicality of FHE schemes. Crucial parameters include the polymodulus degree, ciphertext modulus, plaintext modulus, scaling factor, and operation depth, all of which must be carefully chosen to balance security requirements and computational efficiency in FHE implementations [7], [8]. Polymodulus Degree. ...
January 2021
... The efficiency and security of the OT protocol affect both the utility and security of the external high-level protocols that call the OT protocol. In order to make their outer layer protocols more efficient or secure, researchers have started to focus on designing more functional OT variants [7][8][9][10][11]. The cut-and-choose method, as a tool in secure two-party computation, can be used to normalise and constrain parties that honestly execute protocols in garbled circuits and can prevent circuit constructors from cheating by constructing faulty circuits. ...
November 2021
Lecture Notes in Computer Science
... Formally, this is equivalent to a reconstruction attack with a finite list of candidates for the reconstruction. Notable examples of these attacks are [47,52,55,66,70]. ...
June 2021
... Typically, credentials are publicly verifiable using the public key of the issuer. However, there are also constructions [CMZ14] and use-cases, e.g., private groups in the Signal messenger [CPZ20], where the issuer is identical to the verifier and verification uses the secret of the issuer. This can allow for more efficient constructions that can avoid the use of bilinear groups and thus specific pairing-friendly elliptic curves (cf. ...
October 2020
... Many protocols utilize pre-computations for improving efficiency, e.g., Beaver triples [7] for multiplication. They can be realized by a data-independent offline phase run by a semi-honest dealer T or 2PC protocols from homomorphic encryption [36] or oblivious transfer [28,52] or oblivious shuffle [12,50]. We adopt the first common approach (also called client-aided setting [4]) for simplicity. ...
December 2020
Lecture Notes in Computer Science
... After the intersection is computed, the parties can continue with the collaborative learning process using only the intersection of the datasets. Over the past few years, different hashing techniques for PSI have been proposed in Buddhavarapu et al. [14], Ion et al. [52], Chase and Miao [20], Lu and Ding [89] such as Bloom Filters and Oblivious Hashing. Bloom Filters are a probabilistic data structure that allow for efficient set membership testing, while Oblivious Hashing is a cryptographic technique that enables parties to privately compute hash functions without revealing the inputs to each other. ...
August 2020
Lecture Notes in Computer Science
... Verifiable registries. There is a rich literature on cryptographic construction of publicly verifiable data structures such as append-only logs [25,27,37,45,46,64,66], which offer a range of different trade-offs between security, performance, and usability. Our work builds on Trillian, an implementation of a Merkle tree-based log [4], which is both widely supported and provided adequate flexibility for VRLog. ...
November 2019
... Specifically, OLE can be used to generate multiplication triples which are the basic tool for securely computing multiplication gates [32]. Besides that, OLE has applications in more tasks for two-party secure computation [5,14,25,44,46] and Private Set Intersection [37]. ...
August 2019
Lecture Notes in Computer Science