Alert button
Picture for Jing Jin

Jing Jin

Alert button

Multistatic Integrated Sensing and Communication System in Cellular Networks

May 22, 2023
Zixiang Han, Lincong Han, Xiaozhou Zhang, Yajuan Wang, Liang Ma, Mengting Lou, Jing Jin, Guangyi Liu

Figure 1 for Multistatic Integrated Sensing and Communication System in Cellular Networks
Figure 2 for Multistatic Integrated Sensing and Communication System in Cellular Networks
Figure 3 for Multistatic Integrated Sensing and Communication System in Cellular Networks
Figure 4 for Multistatic Integrated Sensing and Communication System in Cellular Networks

A novel multistatic multiple-input multiple-output (MIMO) integrated sensing and communication (ISAC) system in cellular networks is proposed. It can make use of widespread base stations (BSs) to perform cooperative sensing in wide area. This system is important since the deployment of sensing function can be achieved based on the existing mobile communication networks at a low cost. In this system, orthogonal frequency division multiplexing (OFDM) signals transmitted from the central BS are received and processed by each of the neighboring BSs to estimate sensing object parameters. A joint data processing method is then introduced to derive the closed-form solution of objects position and velocity. Numerical simulation shows that the proposed multistatic system can improve the position and velocity estimation accuracy compared with monostatic and bistatic system, demonstrating the effectiveness and promise of implementing ISAC in the upcoming fifth generation advanced (5G-A) and sixth generation (6G) mobile networks.

Viaarxiv icon

IDO-VFI: Identifying Dynamics via Optical Flow Guidance for Video Frame Interpolation with Events

May 18, 2023
Chenyang Shi, Hanxiao Liu, Jing Jin, Wenzhuo Li, Yuzhen Li, Boyi Wei, Yibo Zhang

Figure 1 for IDO-VFI: Identifying Dynamics via Optical Flow Guidance for Video Frame Interpolation with Events
Figure 2 for IDO-VFI: Identifying Dynamics via Optical Flow Guidance for Video Frame Interpolation with Events
Figure 3 for IDO-VFI: Identifying Dynamics via Optical Flow Guidance for Video Frame Interpolation with Events
Figure 4 for IDO-VFI: Identifying Dynamics via Optical Flow Guidance for Video Frame Interpolation with Events

Video frame interpolation aims to generate high-quality intermediate frames from boundary frames and increase frame rate. While existing linear, symmetric and nonlinear models are used to bridge the gap from the lack of inter-frame motion, they cannot reconstruct real motions. Event cameras, however, are ideal for capturing inter-frame dynamics with their extremely high temporal resolution. In this paper, we propose an event-and-frame-based video frame interpolation method named IDO-VFI that assigns varying amounts of computation for different sub-regions via optical flow guidance. The proposed method first estimates the optical flow based on frames and events, and then decides whether to further calculate the residual optical flow in those sub-regions via a Gumbel gating module according to the optical flow amplitude. Intermediate frames are eventually generated through a concise Transformer-based fusion network. Our proposed method maintains high-quality performance while reducing computation time and computational effort by 10% and 17% respectively on Vimeo90K datasets, compared with a unified process on the whole region. Moreover, our method outperforms state-of-the-art frame-only and frames-plus-events methods on multiple video frame interpolation benchmarks. Codes and models are available at https://github.com/shicy17/IDO-VFI.

Viaarxiv icon

From ORAN to Cell-Free RAN: Architecture, Performance Analysis, Testbeds and Trials

Feb 07, 2023
Yang Cao, Ziyang Zhang, Xinjiang Xia, Pengzhe Xin, Dongjie Liu, Kang Zheng, Mengting Lou, Jing Jin, Qixing Wang, Dongming Wang, Yongming Huang, Xiaohu You, Jiangzhou Wang

Figure 1 for From ORAN to Cell-Free RAN: Architecture, Performance Analysis, Testbeds and Trials
Figure 2 for From ORAN to Cell-Free RAN: Architecture, Performance Analysis, Testbeds and Trials
Figure 3 for From ORAN to Cell-Free RAN: Architecture, Performance Analysis, Testbeds and Trials
Figure 4 for From ORAN to Cell-Free RAN: Architecture, Performance Analysis, Testbeds and Trials

Open radio access network (ORAN) provides an open architecture to implement radio access network (RAN) of the fifth generation (5G) and beyond mobile communications. As a key technology for the evolution to the sixth generation (6G) systems, cell-free massive multiple-input multiple-output (CF-mMIMO) can effectively improve the spectrum efficiency, peak rate and reliability of wireless communication systems. Starting from scalable implementation of CF-mMIMO, we study a cell-free RAN (CF-RAN) under the ORAN architecture. Through theoretical analysis and numerical simulation, we investigate the uplink and downlink spectral efficiencies of CF-mMIMO with the new architecture. We then discuss the implementation issues of CF-RAN under ORAN architecture, including time-frequency synchronization and over-the-air reciprocity calibration, low layer splitting, deployment of ORAN radio units (O-RU), artificial intelligent based user associations. Finally, we present some representative experimental results for the uplink distributed reception and downlink coherent joint transmission of CF-RAN with commercial off-the-shelf O-RUs.

Viaarxiv icon

Tensor Decomposition based Personalized Federated Learning

Aug 27, 2022
Qing Wang, Jing Jin, Xiaofeng Liu, Huixuan Zong, Yunfeng Shao, Yinchuan Li

Figure 1 for Tensor Decomposition based Personalized Federated Learning
Figure 2 for Tensor Decomposition based Personalized Federated Learning
Figure 3 for Tensor Decomposition based Personalized Federated Learning
Figure 4 for Tensor Decomposition based Personalized Federated Learning

Federated learning (FL) is a new distributed machine learning framework that can achieve reliably collaborative training without collecting users' private data. However, due to FL's frequent communication and average aggregation strategy, they experience challenges scaling to statistical diversity data and large-scale models. In this paper, we propose a personalized FL framework, named Tensor Decomposition based Personalized Federated learning (TDPFed), in which we design a novel tensorized local model with tensorized linear layers and convolutional layers to reduce the communication cost. TDPFed uses a bi-level loss function to decouple personalized model optimization from the global model learning by controlling the gap between the personalized model and the tensorized local model. Moreover, an effective distributed learning strategy and two different model aggregation strategies are well designed for the proposed TDPFed framework. Theoretical convergence analysis and thorough experiments demonstrate that our proposed TDPFed framework achieves state-of-the-art performance while reducing the communication cost.

Viaarxiv icon

3D Face Parsing via Surface Parameterization and 2D Semantic Segmentation Network

Jun 18, 2022
Wenyuan Sun, Ping Zhou, Yangang Wang, Zongpu Yu, Jing Jin, Guangquan Zhou

Figure 1 for 3D Face Parsing via Surface Parameterization and 2D Semantic Segmentation Network
Figure 2 for 3D Face Parsing via Surface Parameterization and 2D Semantic Segmentation Network
Figure 3 for 3D Face Parsing via Surface Parameterization and 2D Semantic Segmentation Network
Figure 4 for 3D Face Parsing via Surface Parameterization and 2D Semantic Segmentation Network

Face parsing assigns pixel-wise semantic labels as the face representation for computers, which is the fundamental part of many advanced face technologies. Compared with 2D face parsing, 3D face parsing shows more potential to achieve better performance and further application, but it is still challenging due to 3D mesh data computation. Recent works introduced different methods for 3D surface segmentation, while the performance is still limited. In this paper, we propose a method based on the "3D-2D-3D" strategy to accomplish 3D face parsing. The topological disk-like 2D face image containing spatial and textural information is transformed from the sampled 3D face data through the face parameterization algorithm, and a specific 2D network called CPFNet is proposed to achieve the semantic segmentation of the 2D parameterized face data with multi-scale technologies and feature aggregation. The 2D semantic result is then inversely re-mapped to 3D face data, which finally achieves the 3D face parsing. Experimental results show that both CPFNet and the "3D-2D-3D" strategy accomplish high-quality 3D face parsing and outperform state-of-the-art 2D networks as well as 3D methods in both qualitative and quantitative comparisons.

Viaarxiv icon

Light Field Depth Estimation Based on Stitched-EPI

Mar 29, 2022
Ping Zhou, Xiaoyang Liu, Jing Jin, Yuting Zhang, Junhui Hou

Figure 1 for Light Field Depth Estimation Based on Stitched-EPI
Figure 2 for Light Field Depth Estimation Based on Stitched-EPI
Figure 3 for Light Field Depth Estimation Based on Stitched-EPI
Figure 4 for Light Field Depth Estimation Based on Stitched-EPI

Depth estimation is one of the most essential problems for light field applications. In EPI-based methods, the slope computation usually suffers low accuracy due to the discretization error and low angular resolution. In addition, recent methods work well in most regions but often struggle with blurry edges over occluded regions and ambiguity over texture-less regions. To address these challenging issues, we first propose the stitched-EPI and half-stitched-EPI algorithms for non-occluded and occluded regions, respectively. The algorithms improve slope computation by shifting and concatenating lines in different EPIs but related to the same point in 3D scene, while the half-stitched-EPI only uses non-occluded part of lines. Combined with the joint photo-consistency cost proposed by us, the more accurate and robust depth map can be obtained in both occluded and non-occluded regions. Furthermore, to improve the depth estimation in texture-less regions, we propose a depth propagation strategy that determines their depth from the edge to interior, from accurate regions to coarse regions. Experimental and ablation results demonstrate that the proposed method achieves accurate and robust depth maps in all regions effectively.

* 15 pages 
Viaarxiv icon

Using calibrator to improve robustness in Machine Reading Comprehension

Feb 24, 2022
Jing Jin, Houfeng Wang

Figure 1 for Using calibrator to improve robustness in Machine Reading Comprehension
Figure 2 for Using calibrator to improve robustness in Machine Reading Comprehension
Figure 3 for Using calibrator to improve robustness in Machine Reading Comprehension
Figure 4 for Using calibrator to improve robustness in Machine Reading Comprehension

Machine Reading Comprehension(MRC) has achieved a remarkable result since some powerful models, such as BERT, are proposed. However, these models are not robust enough and vulnerable to adversarial input perturbation and generalization examples. Some works tried to improve the performance on specific types of data by adding some related examples into training data while it leads to degradation on the original dataset, because the shift of data distribution makes the answer ranking based on the softmax probability of model unreliable. In this paper, we propose a method to improve the robustness by using a calibrator as the post-hoc reranker, which is implemented based on XGBoost model. The calibrator combines both manual features and representation learning features to rerank candidate results. Experimental results on adversarial datasets show that our model can achieve performance improvement by more than 10\% and also make improvement on the original and generalization datasets.

Viaarxiv icon

Content-aware Warping for View Synthesis

Jan 22, 2022
Mantang Guo, Jing Jin, Hui Liu, Junhui Hou, Huanqiang Zeng, Jiwen Lu

Figure 1 for Content-aware Warping for View Synthesis
Figure 2 for Content-aware Warping for View Synthesis
Figure 3 for Content-aware Warping for View Synthesis
Figure 4 for Content-aware Warping for View Synthesis

Existing image-based rendering methods usually adopt depth-based image warping operation to synthesize novel views. In this paper, we reason the essential limitations of the traditional warping operation to be the limited neighborhood and only distance-based interpolation weights. To this end, we propose content-aware warping, which adaptively learns the interpolation weights for pixels of a relatively large neighborhood from their contextual information via a lightweight neural network. Based on this learnable warping module, we propose a new end-to-end learning-based framework for novel view synthesis from two input source views, in which two additional modules, namely confidence-based blending and feature-assistant spatial refinement, are naturally proposed to handle the occlusion issue and capture the spatial correlation among pixels of the synthesized view, respectively. Besides, we also propose a weight-smoothness loss term to regularize the network. Experimental results on structured light field datasets with wide baselines and unstructured multi-view datasets show that the proposed method significantly outperforms state-of-the-art methods both quantitatively and visually. The source code will be publicly available at https://github.com/MantangGuo/CW4VS.

* arXiv admin note: text overlap with arXiv:2108.07408 
Viaarxiv icon

DSSL: Deep Surroundings-person Separation Learning for Text-based Person Retrieval

Sep 12, 2021
Aichun Zhu, Zijie Wang, Yifeng Li, Xili Wan, Jing Jin, Tian Wang, Fangqiang Hu, Gang Hua

Figure 1 for DSSL: Deep Surroundings-person Separation Learning for Text-based Person Retrieval
Figure 2 for DSSL: Deep Surroundings-person Separation Learning for Text-based Person Retrieval
Figure 3 for DSSL: Deep Surroundings-person Separation Learning for Text-based Person Retrieval
Figure 4 for DSSL: Deep Surroundings-person Separation Learning for Text-based Person Retrieval

Many previous methods on text-based person retrieval tasks are devoted to learning a latent common space mapping, with the purpose of extracting modality-invariant features from both visual and textual modality. Nevertheless, due to the complexity of high-dimensional data, the unconstrained mapping paradigms are not able to properly catch discriminative clues about the corresponding person while drop the misaligned information. Intuitively, the information contained in visual data can be divided into person information (PI) and surroundings information (SI), which are mutually exclusive from each other. To this end, we propose a novel Deep Surroundings-person Separation Learning (DSSL) model in this paper to effectively extract and match person information, and hence achieve a superior retrieval accuracy. A surroundings-person separation and fusion mechanism plays the key role to realize an accurate and effective surroundings-person separation under a mutually exclusion constraint. In order to adequately utilize multi-modal and multi-granular information for a higher retrieval accuracy, five diverse alignment paradigms are adopted. Extensive experiments are carried out to evaluate the proposed DSSL on CUHK-PEDES, which is currently the only accessible dataset for text-base person retrieval task. DSSL achieves the state-of-the-art performance on CUHK-PEDES. To properly evaluate our proposed DSSL in the real scenarios, a Real Scenarios Text-based Person Reidentification (RSTPReid) dataset is constructed to benefit future research on text-based person retrieval, which will be publicly available.

* Accepted by ACM MM'21 
Viaarxiv icon