Alert button
Picture for Miao Pan

Miao Pan

Alert button

Eve Said Yes: AirBone Authentication for Head-Wearable Smart Voice Assistant

Sep 26, 2023
Chenpei Huang, Hui Zhong, Jie Lian, Pavana Prakash, Dian Shi, Yuan Xu, Miao Pan

Recent advances in machine learning and natural language processing have fostered the enormous prosperity of smart voice assistants and their services, e.g., Alexa, Google Home, Siri, etc. However, voice spoofing attacks are deemed to be one of the major challenges of voice control security, and never stop evolving such as deep-learning-based voice conversion and speech synthesis techniques. To solve this problem outside the acoustic domain, we focus on head-wearable devices, such as earbuds and virtual reality (VR) headsets, which are feasible to continuously monitor the bone-conducted voice in the vibration domain. Specifically, we identify that air and bone conduction (AC/BC) from the same vocalization are coupled (or concurrent) and user-level unique, which makes them suitable behavior and biometric factors for multi-factor authentication (MFA). The legitimate user can defeat acoustic domain and even cross-domain spoofing samples with the proposed two-stage AirBone authentication. The first stage answers \textit{whether air and bone conduction utterances are time domain consistent (TC)} and the second stage runs \textit{bone conduction speaker recognition (BC-SR)}. The security level is hence increased for two reasons: (1) current acoustic attacks on smart voice assistants cannot affect bone conduction, which is in the vibration domain; (2) even for advanced cross-domain attacks, the unique bone conduction features can detect adversary's impersonation and machine-induced vibration. Finally, AirBone authentication has good usability (the same level as voice authentication) compared with traditional MFA and those specially designed to enhance smart voice security. Our experimental results show that the proposed AirBone authentication is usable and secure, and can be easily equipped by commercial off-the-shelf head wearables with good user experience.

* 13 pages, 12 figures 
Viaarxiv icon

Semantic Communications with Explicit Semantic Base for Image Transmission

Aug 12, 2023
Yuan Zheng, Fengyu Wang, Wenjun Xu, Miao Pan, Ping Zhang

Semantic communications, aiming at ensuring the successful delivery of the meaning of information, are expected to be one of the potential techniques for the next generation communications. However, the knowledge forming and synchronizing mechanism that enables semantic communication systems to extract and interpret the semantics of information according to the communication intents is still immature. In this paper, we propose a semantic image transmission framework with explicit semantic base (Seb), where Sebs are generated and employed as the knowledge shared between the transmitter and the receiver with flexible granularity. To represent images with Sebs, a novel Seb-based reference image generator is proposed to generate Sebs and then decompose the transmitted images. To further encode/decode the residual information for precise image reconstruction, a Seb-based image encoder/decoder is proposed. The key components of the proposed framework are optimized jointly by end-to-end (E2E) training, where the loss function is dedicated designed to tackle the problem of nondifferentiable operation in Seb-based reference image generator by introducing a gradient approximation mechanism. Extensive experiments show that the proposed framework outperforms state-of-art works by 0.5 - 1.5 dB in peak signal-to-noise ratio (PSNR) w.r.t. different signal-to-noise ratio (SNR).

Viaarxiv icon

PATROL: Privacy-Oriented Pruning for Collaborative Inference Against Model Inversion Attacks

Jul 20, 2023
Shiwei Ding, Lan Zhang, Miao Pan, Xiaoyong Yuan

Figure 1 for PATROL: Privacy-Oriented Pruning for Collaborative Inference Against Model Inversion Attacks
Figure 2 for PATROL: Privacy-Oriented Pruning for Collaborative Inference Against Model Inversion Attacks
Figure 3 for PATROL: Privacy-Oriented Pruning for Collaborative Inference Against Model Inversion Attacks
Figure 4 for PATROL: Privacy-Oriented Pruning for Collaborative Inference Against Model Inversion Attacks

Collaborative inference has been a promising solution to enable resource-constrained edge devices to perform inference using state-of-the-art deep neural networks (DNNs). In collaborative inference, the edge device first feeds the input to a partial DNN locally and then uploads the intermediate result to the cloud to complete the inference. However, recent research indicates model inversion attacks (MIAs) can reconstruct input data from intermediate results, posing serious privacy concerns for collaborative inference. Existing perturbation and cryptography techniques are inefficient and unreliable in defending against MIAs while performing accurate inference. This paper provides a viable solution, named PATROL, which develops privacy-oriented pruning to balance privacy, efficiency, and utility of collaborative inference. PATROL takes advantage of the fact that later layers in a DNN can extract more task-specific features. Given limited local resources for collaborative inference, PATROL intends to deploy more layers at the edge based on pruning techniques to enforce task-specific features for inference and reduce task-irrelevant but sensitive features for privacy preservation. To achieve privacy-oriented pruning, PATROL introduces two key components: Lipschitz regularization and adversarial reconstruction training, which increase the reconstruction errors by reducing the stability of MIAs and enhance the target inference model by adversarial training, respectively.

Viaarxiv icon

Fed-CPrompt: Contrastive Prompt for Rehearsal-Free Federated Continual Learning

Jul 10, 2023
Gaurav Bagwe, Xiaoyong Yuan, Miao Pan, Lan Zhang

Figure 1 for Fed-CPrompt: Contrastive Prompt for Rehearsal-Free Federated Continual Learning
Figure 2 for Fed-CPrompt: Contrastive Prompt for Rehearsal-Free Federated Continual Learning
Figure 3 for Fed-CPrompt: Contrastive Prompt for Rehearsal-Free Federated Continual Learning
Figure 4 for Fed-CPrompt: Contrastive Prompt for Rehearsal-Free Federated Continual Learning

Federated continual learning (FCL) learns incremental tasks over time from confidential datasets distributed across clients. This paper focuses on rehearsal-free FCL, which has severe forgetting issues when learning new tasks due to the lack of access to historical task data. To address this issue, we propose Fed-CPrompt based on prompt learning techniques to obtain task-specific prompts in a communication-efficient way. Fed-CPrompt introduces two key components, asynchronous prompt learning, and contrastive continual loss, to handle asynchronous task arrival and heterogeneous data distributions in FCL, respectively. Extensive experiments demonstrate the effectiveness of Fed-CPrompt in achieving SOTA rehearsal-free FCL performance.

* Accepted by FL-ICML 2023 
Viaarxiv icon

AnycostFL: Efficient On-Demand Federated Learning over Heterogeneous Edge Devices

Jan 08, 2023
Peichun Li, Guoliang Cheng, Xumin Huang, Jiawen Kang, Rong Yu, Yuan Wu, Miao Pan

Figure 1 for AnycostFL: Efficient On-Demand Federated Learning over Heterogeneous Edge Devices
Figure 2 for AnycostFL: Efficient On-Demand Federated Learning over Heterogeneous Edge Devices
Figure 3 for AnycostFL: Efficient On-Demand Federated Learning over Heterogeneous Edge Devices
Figure 4 for AnycostFL: Efficient On-Demand Federated Learning over Heterogeneous Edge Devices

In this work, we investigate the challenging problem of on-demand federated learning (FL) over heterogeneous edge devices with diverse resource constraints. We propose a cost-adjustable FL framework, named AnycostFL, that enables diverse edge devices to efficiently perform local updates under a wide range of efficiency constraints. To this end, we design the model shrinking to support local model training with elastic computation cost, and the gradient compression to allow parameter transmission with dynamic communication overhead. An enhanced parameter aggregation is conducted in an element-wise manner to improve the model performance. Focusing on AnycostFL, we further propose an optimization design to minimize the global training loss with personalized latency and energy constraints. By revealing the theoretical insights of the convergence analysis, personalized training strategies are deduced for different devices to match their locally available resources. Experiment results indicate that, when compared to the state-of-the-art efficient FL algorithms, our learning framework can reduce up to 1.9 times of the training latency and energy consumption for realizing a reasonable global testing accuracy. Moreover, the results also demonstrate that, our approach significantly improves the converged global accuracy.

* Accepted to IEEE INFOCOM 2023 
Viaarxiv icon

Energy and Spectrum Efficient Federated Learning via High-Precision Over-the-Air Computation

Aug 15, 2022
Liang Li, Chenpei Huang, Dian Shi, Hao Wang, Xiangwei Zhou, Minglei Shu, Miao Pan

Figure 1 for Energy and Spectrum Efficient Federated Learning via High-Precision Over-the-Air Computation
Figure 2 for Energy and Spectrum Efficient Federated Learning via High-Precision Over-the-Air Computation
Figure 3 for Energy and Spectrum Efficient Federated Learning via High-Precision Over-the-Air Computation
Figure 4 for Energy and Spectrum Efficient Federated Learning via High-Precision Over-the-Air Computation

Federated learning (FL) enables mobile devices to collaboratively learn a shared prediction model while keeping data locally. However, there are two major research challenges to practically deploy FL over mobile devices: (i) frequent wireless updates of huge size gradients v.s. limited spectrum resources, and (ii) energy-hungry FL communication and local computing during training v.s. battery-constrained mobile devices. To address those challenges, in this paper, we propose a novel multi-bit over-the-air computation (M-AirComp) approach for spectrum-efficient aggregation of local model updates in FL and further present an energy-efficient FL design for mobile devices. Specifically, a high-precision digital modulation scheme is designed and incorporated in the M-AirComp, allowing mobile devices to upload model updates at the selected positions simultaneously in the multi-access channel. Moreover, we theoretically analyze the convergence property of our FL algorithm. Guided by FL convergence analysis, we formulate a joint transmission probability and local computing control optimization, aiming to minimize the overall energy consumption (i.e., iterative local computing + multi-round communications) of mobile devices in FL. Extensive simulation results show that our proposed scheme outperforms existing ones in terms of spectrum utilization, energy efficiency, and learning accuracy.

Viaarxiv icon

Service Delay Minimization for Federated Learning over Mobile Devices

May 19, 2022
Rui Chen, Dian Shi, Xiaoqi Qin, Dongjie Liu, Miao Pan, Shuguang Cui

Figure 1 for Service Delay Minimization for Federated Learning over Mobile Devices
Figure 2 for Service Delay Minimization for Federated Learning over Mobile Devices
Figure 3 for Service Delay Minimization for Federated Learning over Mobile Devices
Figure 4 for Service Delay Minimization for Federated Learning over Mobile Devices

Federated learning (FL) over mobile devices has fostered numerous intriguing applications/services, many of which are delay-sensitive. In this paper, we propose a service delay efficient FL (SDEFL) scheme over mobile devices. Unlike traditional communication efficient FL, which regards wireless communications as the bottleneck, we find that under many situations, the local computing delay is comparable to the communication delay during the FL training process, given the development of high-speed wireless transmission techniques. Thus, the service delay in FL should be computing delay + communication delay over training rounds. To minimize the service delay of FL, simply reducing local computing/communication delay independently is not enough. The delay trade-off between local computing and wireless communications must be considered. Besides, we empirically study the impacts of local computing control and compression strategies (i.e., the number of local updates, weight quantization, and gradient quantization) on computing, communication and service delays. Based on those trade-off observation and empirical studies, we develop an optimization scheme to minimize the service delay of FL over heterogeneous devices. We establish testbeds and conduct extensive emulations/experiments to verify our theoretical analysis. The results show that SDEFL reduces notable service delay with a small accuracy drop compared to peer designs.

* 15 pages, 9 figures 
Viaarxiv icon

Towards Fast and Accurate Federated Learning with non-IID Data for Cloud-Based IoT Applications

Jan 29, 2022
Tian Liu, Jiahao Ding, Ting Wang, Miao Pan, Mingsong Chen

Figure 1 for Towards Fast and Accurate Federated Learning with non-IID Data for Cloud-Based IoT Applications
Figure 2 for Towards Fast and Accurate Federated Learning with non-IID Data for Cloud-Based IoT Applications
Figure 3 for Towards Fast and Accurate Federated Learning with non-IID Data for Cloud-Based IoT Applications
Figure 4 for Towards Fast and Accurate Federated Learning with non-IID Data for Cloud-Based IoT Applications

As a promising method of central model training on decentralized device data while securing user privacy, Federated Learning (FL)is becoming popular in Internet of Things (IoT) design. However, when the data collected by IoT devices are highly skewed in a non-independent and identically distributed (non-IID) manner, the accuracy of vanilla FL method cannot be guaranteed. Although there exist various solutions that try to address the bottleneck of FL with non-IID data, most of them suffer from extra intolerable communication overhead and low model accuracy. To enable fast and accurate FL, this paper proposes a novel data-based device grouping approach that can effectively reduce the disadvantages of weight divergence during the training of non-IID data. However, since our grouping method is based on the similarity of extracted feature maps from IoT devices, it may incur additional risks of privacy exposure. To solve this problem, we propose an improved version by exploiting similarity information using the Locality-Sensitive Hashing (LSH) algorithm without exposing extracted feature maps. Comprehensive experimental results on well-known benchmarks show that our approach can not only accelerate the convergence rate, but also improve the prediction accuracy for FL with non-IID data.

* FL 
Viaarxiv icon

FedGreen: Federated Learning with Fine-Grained Gradient Compression for Green Mobile Edge Computing

Nov 11, 2021
Peichun Li, Xumin Huang, Miao Pan, Rong Yu

Figure 1 for FedGreen: Federated Learning with Fine-Grained Gradient Compression for Green Mobile Edge Computing
Figure 2 for FedGreen: Federated Learning with Fine-Grained Gradient Compression for Green Mobile Edge Computing
Figure 3 for FedGreen: Federated Learning with Fine-Grained Gradient Compression for Green Mobile Edge Computing
Figure 4 for FedGreen: Federated Learning with Fine-Grained Gradient Compression for Green Mobile Edge Computing

Federated learning (FL) enables devices in mobile edge computing (MEC) to collaboratively train a shared model without uploading the local data. Gradient compression may be applied to FL to alleviate the communication overheads but current FL with gradient compression still faces great challenges. To deploy green MEC, we propose FedGreen, which enhances the original FL with fine-grained gradient compression to efficiently control the total energy consumption of the devices. Specifically, we introduce the relevant operations including device-side gradient reduction and server-side element-wise aggregation to facilitate the gradient compression in FL. According to a public dataset, we investigate the contributions of the compressed local gradients with respect to different compression ratios. After that, we formulate and tackle a learning accuracy-energy efficiency tradeoff problem where the optimal compression ratio and computing frequency are derived for each device. Experiments results demonstrate that given the 80% test accuracy requirement, compared with the baseline schemes, FedGreen reduces at least 32% of the total energy consumption of the devices.

* Accepted for publication in Globecom'21 
Viaarxiv icon

To Talk or to Work: Delay Efficient Federated Learning over Mobile Edge Devices

Nov 01, 2021
Pavana Prakash, Jiahao Ding, Maoqiang Wu, Minglei Shu, Rong Yu, Miao Pan

Figure 1 for To Talk or to Work: Delay Efficient Federated Learning over Mobile Edge Devices
Figure 2 for To Talk or to Work: Delay Efficient Federated Learning over Mobile Edge Devices

Federated learning (FL), an emerging distributed machine learning paradigm, in conflux with edge computing is a promising area with novel applications over mobile edge devices. In FL, since mobile devices collaborate to train a model based on their own data under the coordination of a central server by sharing just the model updates, training data is maintained private. However, without the central availability of data, computing nodes need to communicate the model updates often to attain convergence. Hence, the local computation time to create local model updates along with the time taken for transmitting them to and from the server result in a delay in the overall time. Furthermore, unreliable network connections may obstruct an efficient communication of these updates. To address these, in this paper, we propose a delay-efficient FL mechanism that reduces the overall time (consisting of both the computation and communication latencies) and communication rounds required for the model to converge. Exploring the impact of various parameters contributing to delay, we seek to balance the trade-off between wireless communication (to talk) and local computation (to work). We formulate a relation with overall time as an optimization problem and demonstrate the efficacy of our approach through extensive simulations.

* Accepted for publication in Globecom'21 
Viaarxiv icon