Publication Types:

Sort by year:

Parallel Assisted Learning

Decentralized AIJournal paper
Xinran Wang, Jiawei Zhang, Mingyi Hong, Yuhong Yang, Jie Ding
IEEE Transactions on Signal Processing
Publication year: 2023

Abstract:

In the era of big data, a population’s multimodal data are often collected and preserved by different business and government entities. These entities often have their local machine learning data, models, and tasks that they cannot share with others. Meanwhile, an entity often needs to seek assistance from others to enhance its learning quality without sharing proprietary information. How can an entity be assisted while it is assisting others? We develop a general method called parallel assisted learning (PAL) that applies to the context where entities perform supervised learning and can collate their data according to a common data identifier. Under the PAL mechanism, a learning entity that receives assistance is obligated to assist others without the need to reveal any entity’s local data, model, and learning objective. Consequently, each entity can significantly improve its particular task. The applicability of the proposed approach is demonstrated by data experiments.

Keywords:

Assisted learning

Incentive

Assisted Learning for Organizations with Limited Imbalanced Data

AI FoundationsDecentralized AIJournal paper
Cheng Chen, Jiaying Zhou, Jie Ding, Yi Zhou
Transactions on Machine Learning Research
Publication year: 2023

Abstract:

In the era of big data, many big organizations are integrating machine learning into their work pipelines to facilitate data analysis. However, the performance of their trained models is often restricted by limited and imbalanced data available to them. In this work, we develop an assisted learning framework for assisting organizations to improve their learning performance. The organizations have sufficient computation resources but are subject to stringent data-sharing and collaboration policies. Their limited imbalanced data often cause biased inference and sub-optimal decision-making. In assisted learning, an organizational learner purchases assistance service from an external service provider and aims to enhance its model performance within only a few assistance rounds. We develop effective stochastic training algorithms for both assisted deep learning and assisted reinforcement learning. Different from existing distributed algorithms that need to transmit gradients or models frequently, our framework allows the learner to only occasionally share information with the service provider, but still, obtain a model that achieves near-oracle performance as if all the data were centralized.

Keywords:

Assisted Learning

Imbalanced Data

A Framework for Incentivized Collaborative Learning

Decentralized AIManuscript
Xinran Wang, Qi Le, Ahmad Faraz Khan, Jie Ding, Ali Anwar
Manuscript under review
Publication year: 2023

Abstract:

Collaborations among various entities, such as companies, research labs, AI agents, and edge devices, have become increasingly crucial for achieving machine learning tasks that cannot be accomplished by a single entity alone. This is likely due to factors such as security constraints, privacy concerns, and limitations in computation resources. As a result, collaborative learning (CL) research has been gaining momentum. However, a significant challenge in practical applications of CL is how to effectively incentivize multiple entities to collaborate before any collaboration occurs. In this study, we propose ICL, a general framework for incentivized collaborative learning, and provide insights into the critical issue of when and why incentives can improve collaboration performance. Furthermore, we show the broad applicability of ICL to specific cases in federated learning, assisted learning, and multi-armed bandit with both theory and experimental results.

Keywords:

Collaborative learning

Incentives

SemiFL: Communication Efficient Semi-Supervised Federated Learning with Unlabeled Clients

AI FoundationsConference paperDecentralized AI
Enmao Diao, Jie Ding, Vahid Tarokh
Conference on Neural Information Processing Systems (NeurIPS)
Publication year: 2022

Abstract:

Federated Learning allows training machine learning models by using the computation and private data resources of many distributed clients such as smartphones and IoT devices. Most existing works on Federated Learning (FL) assume the clients have ground-truth labels. However, in many practical scenarios, clients may be unable to label task-specific data, e.g., due to a lack of expertise. This work considers a server that hosts a labeled dataset and wishes to leverage clients with unlabeled data for supervised learning. We propose a new Federated Learning framework referred to as SemiFL to address Semi-Supervised Federated Learning (SSFL). In SemiFL, clients have completely unlabeled data, while the server has a small amount of labeled data. SemiFL is communication efficient since it separates the training of server-side supervised data and client-side unsupervised data. We demonstrate several strategies of SemiFL that enhance efficiency and prediction and develop intuitions of why they work. In particular, we provide a theoretical understanding of the use of strong data augmentation for Semi-Supervised Learning (SSL), which can be interesting in its own right. Extensive empirical evaluations demonstrate that our communication efficient method can significantly improve the performance of a labeled server with unlabeled clients. Moreover, we demonstrate that SemiFL can outperform many existing SSFL methods, and perform competitively with the state-of-the-art FL and centralized SSL results. For instance, in standard communication efficient scenarios, our method can perform 93% accuracy on the CIFAR10 dataset with only 4000 labeled samples at the server. Such accuracy is only 2% away from the result trained from 50000 fully labeled data, and it improves about 30% upon existing SSFL methods in the communication efficient setting.

Keywords:

Federated Learning

Semi-Supervised Learning

Data augmentation theory

Unlabeled data

Self-Aware Personalized Federated Learning

Conference paperDecentralized AI
Huili Chen, Jie Ding, Eric Tramel, Shuang Wu, Anit Kumar Sahu, Salman Avestimehr, Tao Zhang
Conference on Neural Information Processing Systems (NeurIPS)
Publication year: 2022

Abstract:

In the context of personalized federated learning (FL), the critical challenge is to balance local model improvement and global model tuning when the personal and global objectives may not be exactly aligned. Inspired by Bayesian hierarchical models, we develop a self-aware personalized FL method where each client can automatically balance the training of its local personal model and the global model that implicitly contributes to other clients’ training. Such a balance is derived from the inter-client and intra-client uncertainty quantification. A larger inter-client variation implies more personalization is needed. Correspondingly, our method uses uncertainty-driven local training steps and aggregation rule instead of conventional local fine-tuning and sample size-based aggregation. With experimental studies on synthetic data, Amazon Alexa audio data, and public datasets such as MNIST, FEMNIST and Sent140, we show that our proposed method can achieve significantly improved personalization performance compared with the existing counterparts.

Keywords:

Bayesian hierarchical model

Personalized federated learning

Meta Clustering for Collaborative Learning

Decentralized AIJournal paper
Chenglong Ye, Reza Ghanadan, Jie Ding
Journal of Computational and Graphical Statistics
Publication year: 2022

Abstract:

An emerging number of learning scenarios involve a set of learners or analysts, each equipped with a unique dataset and algorithm, who may collaborate to enhance their learning performance. From a particular learner’s perspective, a careless collaboration with task-irrelevant other learners is likely to incur modeling error. A crucial challenge is to search for the most appropriate collaborators so that their data and modeling resources can be effectively leveraged. Motivated by this, we propose to study the problem of ‘meta clustering,’ where the goal is to identify subsets of relevant learners whose collaboration will improve each learner’s performance. In particular, we study the scenario where each learner performs a supervised regression, and the meta clustering aims to categorize the underlying supervised relations (between responses and predictors) instead of the private raw data. We propose a general method named Select-Exchange-Cluster (SEC) for performing such a clustering. Our method is computationally efficient as it does not require each learner to exchange their raw data. We prove that the SEC method can accurately cluster the learners into appropriate collaboration sets according to their underlying regression functions. Synthetic and real data examples show the desired performance and wide applicability of the SEC to various learning tasks.

Keywords:

Distributed computing
Fairness
Meta clustering
Regression

GAL: Gradient Assisted Learning for Decentralized Multi-Organization Collaborations

AI SafetyAI ScalabilityConference paperDecentralized AI
Enmao Diao, Jie Ding, Vahid Tarokh
36th Conference on Neural Information Processing Systems (NeurIPS 2022)
Publication year: 2022

Abstract:

Collaborations among multiple organizations, such as financial institutions, medical centers, and retail markets in decentralized settings are crucial to providing improved service and performance. However, the underlying organizations may have little interest in sharing their local data, models, and objective functions. These requirements have created new challenges for multi-organization collaboration. In this work, we propose Gradient Assisted Learning (GAL), a new method for multiple organizations to assist each other in supervised learning tasks without sharing local data, models, and objective functions. In this framework, all participants collaboratively optimize the aggregate of local loss functions, and each participant autonomously builds its own model by iteratively fitting the gradients of the overarching objective function. We also provide asymptotic convergence analysis and practical case studies of GAL. Experimental studies demonstrate that GAL can achieve performance close to centralized learning when all data, models, and objective functions are fully disclosed.

Keywords:

Assisted learning

Privacy

FedNAS: Federated Deep Learning via Neural Architecture Search

Conference paperDecentralized AI
Chaoyang He, Erum Mushtaq, Jie Ding, Salman Avestimehr
Manuscript
Publication year: 2022

Abstract:

Federated Learning (FL) is an effective learning framework used when data cannot be centralized due to privacy, communication costs, and regulatory restrictions. While there have been many algorithmic advances in FL, significantly less effort has been made to model development, and most works in FL employ predefined model architectures discovered in the centralized environment. However, these predefined architectures may not be the optimal choice for the FL setting since the user data distribution at FL users is often non-identical and independent distribution (nonIID). This well-known challenge in FL has often been studied at the optimization layer. Instead, we advocate for a different (and complementary) approach. We propose Federated Neural Architecture Search (FedNAS) for automating the model design process in FL. More specifically, FedNAS enables scattered workers to search for better architecture in a collaborative fashion to achieve higher accuracy. Beyond automating and improving FL model design, FedNAS also provides a new paradigm for personalized FL via customizing not only the model weights but also the neural architecture of each user. As such, we also compare FedNAS with representative personalized FL methods, including perFedAvg (based on meta-learning), Ditto (bi-level optimization), and local fine-tuning. Our experiments on a non-IID dataset show that the architecture searched by FedNAS can outperform the manually predefined architecture as well as existing personalized FL methods. To facilitate further research and real-world deployment, we also build a realistic distributed training system for FedNAS, which will be publicly available and maintained regularly.

 

 

Model Linkage Selection for Cooperative Learning

Decentralized AIJournal paper
Jiaying Zhou, Jie Ding, Kean Ming Tan, Vahid Tarokh
Journal of Machine Learning Research, 2021
Publication year: 2021

Abstract:

Rapid developments in data collecting devices and computation platforms produce an emerging number of learners and data modalities in many scientific domains. We consider the setting in which each learner holds a pair of parametric statistical model and a specific data source, with the goal of integrating information across a set of learners to enhance the prediction accuracy of a specific learner. One natural way to integrate information is to build a joint model across a set of learners that shares common parameters of interest. However, the parameter sharing patterns across a set of learners are not known a priori. In this paper, we propose a novel framework for integrating information across a set of learners that is robust against model misspecification and misspecified parameter sharing patterns. The main crux is to sequentially incorporate additional learners that can enhance the prediction accuracy of an existing joint model based on user-specified parameter sharing patterns across a set of learners.

Keywords:

Data integration
Distributed learning
Model linkage selection

HeteroFL: Computation and Communication Efficient Federated Learning for Heterogeneous Clients

Conference paperDecentralized AI
Enmao Diao, Jie Ding, Vahid Tarokh
International Conference on Learning Representations (ICLR)
Publication year: 2021

Abstract:

Federated Learning (FL) is a method of training machine learning models on private data distributed over a large number of possibly heterogeneous clients such as mobile phones and IoT devices. In this work, we propose a new federated learning framework named HeteroFL to address heterogeneous clients equipped with very different computation and communication capabilities. Our solution can enable the training of heterogeneous local models with varying computation complexities and still produce a single global inference model. For the first time, our method challenges the underlying assumption of existing work that local models have to share the same architecture as the global model. We demonstrate several strategies to enhance FL training and conduct extensive empirical evaluations, including five computation complexity levels of three model architecture on three datasets. We show that adaptively distributing subnetworks according to clients’ capabilities is both computation and communication efficient.

Keywords:

Federated learning
Heterogeneous clients

Assisted Learning: A Framework for Multi-Organization Learning

Conference paperDecentralized AI
Xun Xian, Xinran Wang, Jie Ding, Reza Ghanadan
Conference on Neural Information Processing Systems (NeurIPS), Spotlight, 2020
Publication year: 2020

Abstract:

In an increasing number of AI scenarios, collaborations among different organizations or agents (e.g., human and robots, mobile units) are often essential to accomplish an organization-specific mission. However, to avoid leaking useful and possibly proprietary information, organizations typically enforce stringent security constraints on sharing modeling algorithms and data, which significantly limits collaborations. In this work, we introduce the Assisted Learning framework for organizations to assist each other in supervised learning tasks without revealing any organization’s algorithm, data, or even task.
An organization seeks assistance by broadcasting task-specific but nonsensitive statistics and incorporating others’ feedback in one or more iterations to eventually improve its predictive performance. Theoretical and experimental studies, including real-world medical benchmarks, show that Assisted Learning can often achieve near-oracle learning performance as if data and training processes were centralized.

Keywords:

Assisted AI
Autonomy
MLaaS
Organization’s learning
Privacy