Publication Types:

Sort by year:

Model Privacy: A Unified Framework to Understand Model Stealing Attack and Defense

AI SafetyManuscript
Ganghua Wang, Yuhong Yang, Jie Ding
Manuscript under review
Publication year: 2023

Abstract:

The security of machine learning models against adversarial attacks has been an increasingly important problem in modern application scenarios such as machine-learning-as-a-service and collaborative learning. The model stealing attack is a particular threat that aims to reverse-engineer a general learned model (e.g., a server-based API, an information exchange protocol, an on-chip AI architecture) from only a tiny number of query-response interactions. Consequently, the attack may steal a proprietary model in a manner that may be much more cost-effective than the model owner’s original training cost. Many modelstealing attack and defense strategies have been proposed with good empirical success. However, most existing works are heuristic, limited in evaluation metrics, and imprecise in characterizing loss and gain. This work presents a unified conceptual framework called Model Privacy for understanding and quantifying model stealing attacks and defenseModel privacy encapsulates the foundational tradeoffs regarding the usability and vulnerability of functionality of a learned model. Based on the developed concepts, we then develop fundamental limits on privacy-utility tradeoffs and their implications in various machine learning problems (e.g., those based on linear functions, polynomials, reproducing kernels, and neural networks). The studied new problems are also interesting from a theoretical perspective, as a model owner may maneuver multiple query responses jointly to maximally enhance model privacy, violating the data independence assumption that plays a critical role in classical learning theory. For example, we show that by breaking independence, a model owner can simultaneously attain a slight utility loss and a much larger privacy gain, a desirable property not achievable in independent data regimes.

Keywords:

Model-stealing attack and defense
Privacy

Demystifying Poisoning Backdoor Attacks from a Statistical Perspective

AI SafetyManuscript
Xun Xian, Ganghua Wang, Jayanth Srinivasa, Ashish Kundu, Xuan Bi, Mingyi Hong, Jie Ding
Manuscript under review
Publication year: 2023

Abstract:

The growing dependence on machine learning in real-world applications emphasizes the importance of understanding and ensuring its safety. Backdoor attacks pose a significant security risk due to their stealthy nature and potentially serious consequences. Such attacks involve embedding triggers within a learning model with the intention of causing malicious behavior when an active trigger is present while maintaining regular functionality without it. This paper evaluates the effectiveness of any backdoor attack incorporating a constant trigger, by establishing tight lower and upper boundaries for the performance of the compromised model on both clean and backdoor test data. The developed theory answers a series of fundamental but previously unsolved problems, including (1) what are the determining factors for a backdoor attack’s success, (2) what is the most effective backdoor attack, and (3) when will a human-imperceptible trigger succeed. The experimental outcomes corroborate the established theory.

Keywords:

Adversarial learning

Backdoor attack

Statistical analysis

Adaptive Continual Learning: Rapid Adaptation and Knowledge Refinement

AI ScalabilityManuscript
Jin Du, Yuhong Yang, Jie Ding
Manuscript under review
Publication year: 2023

Abstract:

Continual learning (CL) is an emerging research area aiming to emulate human learning throughout a lifetime. Most existing CL approaches primarily focus on mitigating catastrophic forgetting, a phenomenon where performance on old tasks declines while learning new ones. However, human learning involves not only re-learning knowledge but also quickly recognizing the current environment, recalling related knowledge, and refining it for improved performance. In this work, we introduce a new problem setting, Adaptive CL, which captures these aspects in an online, recurring task environment without explicit task boundaries or identities. We propose the LEARN algorithm to efficiently explore, recall, and refine knowledge in such environments. We provide theoretical guarantees from two perspectives: online prediction with tight regret bounds and asymptotic consistency of knowledge. Additionally, we present a scalable implementation that requires only first-order gradients for training deep learning models. Our experiments demonstrate that the LEARN algorithm is highly effective in exploring, recalling, and refining knowledge in adaptive CL environments, resulting in superior performance compared to competing methods.

Keywords:

Continual learning

Online streaming data

 

A Unified Framework for Inference-Stage Backdoor Defenses

AI SafetyManuscript
Xun Xian, Ganghua Wang, Jayanth Srinivasa, Ashish Kundu, Xuan Bi, Mingyi Hong, Jie Ding
Manuscript under review
Publication year: 2023

Abstract:

Backdoor attacks involve inserting poisoned samples during training, resulting in a model containing a hidden backdoor that can trigger specific behaviors without impacting performance on normal samples. These attacks are challenging to detect, as the backdoored model appears normal until activated by the backdoor trigger, rendering them particularly stealthy. In this study, we devise a unified inference-stage detection framework to defend against backdoor attacks. We first rigorously formulate the inference-stage backdoor detection problem, encompassing various existing methods, and discuss several challenges and limitations. We then propose a framework with provable guarantees on the false positive rate or the probability of misclassifying a clean sample. Further, we derive the most powerful detection rule to maximize the detection power, namely the rate of accurately identifying a backdoor sample, given a false positive rate under classical learning scenarios. Based on the theoretically optimal detection rule, we suggest a practical and effective approach for real-world applications based on the latent representations of backdoored deep nets. We extensively evaluate our method on 12 different backdoor attacks using Computer Vision (CV) and Natural Language Processing (NLP) benchmark datasets. The experimental findings align with our theoretical results. We significantly surpass the state-of-the-art methods, e.g., up to 300% improvement on the detection power as evaluated by AUCROC, over the state-of-the-art defense against advanced adaptive backdoor attacks.

Keywords

Backdoor defense

Data poisoning

A Framework for Incentivized Collaborative Learning

Decentralized AIManuscript
Xinran Wang, Qi Le, Ahmad Faraz Khan, Jie Ding, Ali Anwar
Manuscript under review
Publication year: 2023

Abstract:

Collaborations among various entities, such as companies, research labs, AI agents, and edge devices, have become increasingly crucial for achieving machine learning tasks that cannot be accomplished by a single entity alone. This is likely due to factors such as security constraints, privacy concerns, and limitations in computation resources. As a result, collaborative learning (CL) research has been gaining momentum. However, a significant challenge in practical applications of CL is how to effectively incentivize multiple entities to collaborate before any collaboration occurs. In this study, we propose ICL, a general framework for incentivized collaborative learning, and provide insights into the critical issue of when and why incentives can improve collaboration performance. Furthermore, we show the broad applicability of ICL to specific cases in federated learning, assisted learning, and multi-armed bandit with both theory and experimental results.

Keywords:

Collaborative learning

Incentives

Large Deviation Principle for the Whittaker 2d Growth Model

ManuscriptMiscellaneous
Jun Gao, Jie Ding
arXiv preprint arXiv:2009.12907
Publication year: 2020

Abstract:

The Whittaker 2d growth model is a triangular continuous Markov diffusion process that appears in many scientific contexts. It has been theoretically intriguing to establish a large deviation principle for this 2d process with a scaling factor. The main challenge is the spatiotemporal interactions and dynamics that may depend on potential sample-path intersections. We develop such a principle with a novel rate function. Our approach is based on Schider’s Theorem, contraction principle, and special treatment for intersecting sample paths.

Keywords:

Large deviation principle
Markov diffusion process

Dyson Brownian Motion as a Limit of the Whittaker 2d Growth Model

ManuscriptMiscellaneous
Jun Gao, Jie Ding
Manuscript
Publication year: 2020

Abstract:

This paper proves that a class of scaled Whittaker growth models will converge in distribution to the Dyson Brownian motion. A Whittaker 2d growth model is a continuous-time Markov diffusion process embedded on a spatial triangular array. Our result is interesting because each particle in a Whittaker 2d growth model only interacts with its neighboring particles. In contrast, each particle in the Dyson Brownian motion interacts with all the other particles. We provide two different proofs of the main result.

Keywords:

Stochastic differential equations
Dyson Brownian motion

Optimal Variable Selection in Regression Models

AI FoundationsManuscript
Jie Ding, Vahid Tarokh, Yuhong Yang
Manuscript, 2016
Publication year: 2016

Abstract:

We introduce a new criterion for variable selection in regression models and show its optimality in terms of both loss and risk under appropriate assumptions. The key idea is to impose a penalty that is nonlinear in model dimensions. In contrast to the state-of-art model selection criteria such as the Cp method, delete-1 or delete-k cross-validation, Akaike information criterion, Bayesian information criterion, the proposed method is able to achieve asymptotic loss and risk efficiency in both parametric and nonparametric regression settings, giving new insights on the reconciliation of two types of classical criteria with different asymptotic behaviors. Adaptivity and wide applicability of the new criterion are demonstrated by several numerical experiments. Unless the signal to noise ratio is very low, it performs better than some popular methods in our experimental study. An R package ‘bc’ is released that serves as a supplement to this work.

Keywords:

Regression
Subset selection
Feature selection